hadoop-common-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Wiki <wikidi...@apache.org>
Subject [Hadoop Wiki] Update of "TooManyOpenFiles" by SteveLoughran
Date Fri, 16 Sep 2011 08:20:23 GMT
Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change notification.

The "TooManyOpenFiles" page has been changed by SteveLoughran:
http://wiki.apache.org/hadoop/TooManyOpenFiles?action=diff&rev1=1&rev2=2

Comment:
fix the link

  
  You can see this on Linux machines in client-side applications, server code or even in test
runs.
  
- It is caused by per-process limits on the number of files that a single user/process can
have open, which was introduced in [[http://lkml.indiana.edu/hypermail/linux/kernel/0812.0/01183.html|2.6.27].
The default value, 128, was chosen because "that should be enough". 
+ It is caused by per-process limits on the number of files that a single user/process can
have open, which was introduced in [[http://lkml.indiana.edu/hypermail/linux/kernel/0812.0/01183.html|the
2.6.27 kernel]]. The default value, 128, was chosen because "that should be enough". 
  
  In Hadoop, it isn't. To fix this log in/su/ssh as root and edit {{{/etc/sysctl.conf}}}
  

Mime
View raw message