F. put a mongodb replica set on all hadoop workernodes and let the tasks query the mongodb at localhost.

(this is what I did recently with a multi GiB dataset)

--
Met vriendelijke groet,
Niels Basjes
(Verstuurd vanaf mobiel )

Op 30 dec. 2012 20:01 schreef "Jonathan Bishop" <jbishop.rwc@gmail.com> het volgende:
E. Store them in hbase...


On Sun, Dec 30, 2012 at 12:24 AM, Hemanth Yamijala <yhemanth@thoughtworks.com> wrote:
If it is a small number, A seems the best way to me.

On Friday, December 28, 2012, Kshiva Kps wrote:


Which one is current ..


What is the preferred way to pass a small number of configuration parameters to a mapper or reducer?

 

 

A.  As key-value pairs in the jobconf object.

 

B.  As a custom input key-value pair passed to each mapper or reducer.

 

C.  Using a plain text file via the Distributedcache, which each mapper or reducer reads.

 

D.  Through a static variable in the MapReduce driver class (i.e., the class that submits the MapReduce job).

 

Answer: B