hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jun Young Kim <juneng...@gmail.com>
Subject Which strategy is proper to run an this enviroment?
Date Sat, 12 Feb 2011 04:07:05 GMT

I have small clusters (9 nodes) to run a hadoop here.

Under this cluster, a hadoop will take thousands of directories sequencely.

In a each dir, there is two input files to m/r. Size of input files are from
1m to 5g bytes.
In a summary, each hadoop job will take an one of these dirs.

To get best performance, which strategy is proper for us?

Could u suggest me about it?
Which configuration is best?

Ps) physical memory size is 12g of each node.

  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message