hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "James Moore" <jamesthepi...@gmail.com>
Subject Re: Read timed out, Abandoning block blk_-5476242061384228962
Date Mon, 12 May 2008 22:53:28 GMT
On Sun, May 11, 2008 at 9:45 PM, Dhruba Borthakur <dhruba@gmail.com> wrote:
>  How much memory does your machine have and how many files does your
>  HDFS have? One possibility is that the memory pressure of the
>  map-reduce jobs causes more GC runs for the namenode process.

I'm using machines with 16G of memory, and there aren't that many
files yet - about 14,000 total, according to dfs -lsr /.  The admin
web page reports 1.32 TB used, but I'm assuming that's including
replication (set to 4).

19 machines total in the cluster (planned to have 20, but one didn't
come up initially and I haven't bothered to add it), and I'm currently
not running maps/reduces on the master.

And earlier, when I said I'm not seeing the timeouts, I spoke too
soon.  One of my last runs had 3 maps fail, with these sorts of

Failed to rename output with the exception:
java.net.SocketTimeoutException: timed out waiting for rpc response
	at org.apache.hadoop.ipc.Client.call(Client.java:514)
	at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198)
	at org.apache.hadoop.dfs.$Proxy4.delete(Unknown Source)
	at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
	at org.apache.hadoop.dfs.$Proxy4.delete(Unknown Source)
	at org.apache.hadoop.dfs.DFSClient.delete(DFSClient.java:426)
	at org.apache.hadoop.dfs.DistributedFileSystem.delete(DistributedFileSystem.java:149)
	at org.apache.hadoop.mapred.Task.saveTaskOutput(Task.java:430)
	at org.apache.hadoop.mapred.JobTracker$TaskCommitQueue.run(JobTracker.java:2013)

This is somewhat painful, since these are Nutch fetch maps, meaning
that they're getting data via HTTP from crawled sites.  Looks like
it's not a permanent problem, though - just doing some crawls twice.

The default timeouts for the RPCs seem high enough (ipc.client.timeout
== 60000).  Is there something else I should look at?  Watching 'top'
during a run, it didn't seem like the master ever went below 80%, so
it seems a little unlikely that the machine isn't responding for 60

Should I increase the memory for the datanode processes?  It's got
-Xmx2000m right now.  And is there a way to do that with the existing
scripts/configurations?  Obviously I can do it by hand, but I'm not
seeing something like the mapred.child.java.opts setting for the

James Moore | james@restphone.com

View raw message