hadoop-common-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Wiki <wikidi...@apache.org>
Subject [Hadoop Wiki] Update of "MachineScaling" by enis
Date Thu, 06 Mar 2008 13:34:14 GMT
Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change notification.

The following page has been changed by enis:
http://wiki.apache.org/hadoop/MachineScaling

The comment on the change is:
Refactor ECC part, emphesize ECC 

------------------------------------------------------------------------------
  
  '''More specifics:'''
  
- Hadoop benefits greatly from ECC memory, which is not low-end. Multi-core boxes tend to
give more computation per dollar, per watt and per unit of operational maintenance. But the
highest clockrate processors tend to not be cost-effective, as do the very largest drives.
So moderately high-end commodity hardware is the most cost-effective for Hadoop today.
+ Multi-core boxes tend to give more computation per dollar, per watt and per unit of operational
maintenance. But the highest clockrate processors tend to not be cost-effective, as do the
very largest drives. So moderately high-end commodity hardware is the most cost-effective
for Hadoop today.
  
  Some users use cast-off machines that were not reliable enough for other applications. These
machines originally cost about 2/3 what normal production boxes cost and achieve almost exactly
1/2 as much. Production boxes are typically dual CPU's with dual cores.
  
  '''RAM:'''
  
  Many users find that most hadoop applications are very small in memory consumption. Users
tend to have 4-8 GB machines with 2GB probably being too little.
+ Hadoop benefits greatly from ECC memory, which is not low-end, however using ECC memory
is RECOMMENDED. see Dennis Kubes' discussion at http://mail-archives.apache.org/mod_mbox/hadoop-core-dev/200705.mbox/%3C465C3065.9050501@dragonflymc.com%3E
  

Mime
View raw message