Sorry for the typo, the property name is mapred.max.split.size
HiYou need to alter the value of mapred.max.split size to a value larger than your block size to have less number of map tasks than the default.On Tue, Oct 2, 2012 at 10:04 PM, Shing Hing Man <email@example.com> wrote:
I am running Hadoop 1.0.3 in Pseudo distributed mode.
When I submit a map/reduce job to process a file of size about 16 GB, in job.xml, I have the following
dfs.block.size = 67108864
I would like to reduce mapred.map.tasks to see if it improves performance.
I have tried doubling the size of dfs.block.size. But the mapred.map.tasks remains unchanged.
Is there a way to reduce mapred.map.tasks ?
Thanks in advance for any assistance !