F. put a mongodb replica set on all hadoop workernodes and let the tasks query the mongodb at localhost.

(this is what I did recently with a multi GiB dataset)

Met vriendelijke groet,
Niels Basjes
(Verstuurd vanaf mobiel )

Op 30 dec. 2012 20:01 schreef "Jonathan Bishop" <jbishop.rwc@gmail.com> het volgende:
E. Store them in hbase...

On Sun, Dec 30, 2012 at 12:24 AM, Hemanth Yamijala <yhemanth@thoughtworks.com> wrote:
If it is a small number, A seems the best way to me.

On Friday, December 28, 2012, Kshiva Kps wrote:

Which one is current ..

What is the preferred way to pass a small number of configuration parameters to a mapper or reducer?



A.  As key-value pairs in the jobconf object.


B.  As a custom input key-value pair passed to each mapper or reducer.


C.  Using a plain text file via the Distributedcache, which each mapper or reducer reads.


D.  Through a static variable in the MapReduce driver class (i.e., the class that submits the MapReduce job).


Answer: B