hadoop-common-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Raghu Angadi <rang...@yahoo-inc.com>
Subject Re: still getting "is valid, and cannot be written to"
Date Tue, 28 Aug 2007 18:06:55 GMT

This means that a datanode is asked to receive a new block but the block 
already exists.

One case where I have seen this happen is when Namenode is trying to 
replicate blocks. Say replication increased from 3 to 20, Namenode asks 
might ask multiple datanodes to transfer a block to the same datanode. 
Smaller the cluster the cluster more probability that this happens.

What context are you seeing this?

Torsten Curdt wrote:
> So we now finally upgraded to 0.14 (more bug reports to come) but could 
> someone please tell me why we are still seeing this
> 2007-08-28 08:23:59,836 ERROR org.apache.hadoop.dfs.DataNode: 
> DataXceiver: java.io.IOException: Block blk_-5226553429149640206 is 
> valid, and cannot be written to.
>         at org.apache.hadoop.dfs.FSDataset.writeToBlock(FSDataset.java:515)
>         at 
> org.apache.hadoop.dfs.DataNode$DataXceiver.writeBlock(DataNode.java:822)
>         at 
> org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:727)
>         at java.lang.Thread.run(Thread.java:595)
> ...and what it actually means?
> cheers
> -- 
> Torsten

View raw message