hadoop-mapreduce-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From mdwa...@hotmail.com
Subject Re: block errors
Date Sat, 24 Jul 2010 21:20:51 GMT


Ted Yu <yuzhihong@gmail.com> wrote:

>Check the datanode log on 10.15.46.73 <http://10.15.46.73:50010/>
>
>You should increase dfs.datanode.max.xcievers
>
>On Tue, Jul 13, 2010 at 3:57 AM, Some Body <somebody@squareplanet.de> wrote:
>
>> Hi All,
>>
>> I had a MR job that processed 2000 small (<3MB ea.) files and it took 40
>> minutes on 8 nodes.
>> Since the files are small it triggerred 2000 tasks.  I packed my 2000 files
>> into a single 445MB
>> sequence file (K,V == Text,Text == <filename>,<file-content>).  The new
MR
>> job triggers 7 map
>> tasks (approx 64MB each) but  it takes even longer (49 minutes) so I'm
>> trying to figure out why.
>>
>> I noticed these errors and I'm hoping someone can shed some light on why?
>>
>> Before I ran  the job I ran hadoop fsck / and everything was healthy.
>> e.g.: no under-replicated, no corrupt blocks etc.
>>
>> ......
>> 2010-07-13 03:24:20,807 INFO org.apache.hadoop.mapred.ReduceTask:
>> GetMapEventsThread exiting
>> 2010-07-13 03:24:20,807 INFO org.apache.hadoop.mapred.ReduceTask:
>> getMapsEventsThread joined.
>> 2010-07-13 03:24:20,808 INFO org.apache.hadoop.mapred.ReduceTask: Closed
>> ram manager
>> 2010-07-13 03:24:20,808 INFO org.apache.hadoop.mapred.ReduceTask:
>> Interleaved on-disk merge complete: 7 files left.
>> 2010-07-13 03:24:20,808 INFO org.apache.hadoop.mapred.ReduceTask: In-memory
>> merge complete: 0 files left.
>> 2010-07-13 03:24:20,814 INFO org.apache.hadoop.mapred.ReduceTask: Merging 7
>> files, 2401573706 bytes from disk
>> 2010-07-13 03:24:20,815 INFO org.apache.hadoop.mapred.ReduceTask: Merging 0
>> segments, 0 bytes from memory into reduce
>> 2010-07-13 03:24:20,818 INFO org.apache.hadoop.mapred.Merger: Merging 7
>> sorted segments
>> 2010-07-13 03:24:20,827 INFO org.apache.hadoop.mapred.Merger: Down to the
>> last merge-pass, with 7 segments left of total size: 2401573678 bytes
>> 2010-07-13 03:30:42,329 INFO org.apache.hadoop.hdfs.DFSClient: Exception in
>> createBlockOutputStream java.io.IOException: Bad connect ack with
>> firstBadLink 10.15.46.73:50010
>> 2010-07-13 03:30:42,329 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning
>> block blk_4304053493083580280_260714
>> 2010-07-13 03:31:03,846 INFO org.apache.hadoop.hdfs.DFSClient: Exception in
>> createBlockOutputStream java.io.IOException: Bad connect ack with
>> firstBadLink 10.15.46.35:50010
>> 2010-07-13 03:31:03,846 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning
>> block blk_3680469905814989852_260716
>> 2010-07-13 03:31:08,233 INFO org.apache.hadoop.hdfs.DFSClient: Exception in
>> createBlockOutputStream java.io.IOException: Bad connect ack with
>> firstBadLink 10.15.46.35:50010
>> 2010-07-13 03:31:08,233 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning
>> block blk_-673505196560500372_260717
>> 2010-07-13 03:31:14,243 INFO org.apache.hadoop.hdfs.DFSClient: Exception in
>> createBlockOutputStream java.io.IOException: Bad connect ack with
>> firstBadLink 10.15.46.73:50010
>> 2010-07-13 03:31:14,243 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning
>> block blk_-7054031797345836167_260717
>> ......
>>
>>
>>
Mime
View raw message