hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Tsz Wo (Nicholas), SZE (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDFS-1606) Provide a stronger data guarantee in the write pipeline
Date Thu, 07 Apr 2011 18:22:06 GMT

    [ https://issues.apache.org/jira/browse/HDFS-1606?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13017071#comment-13017071
] 

Tsz Wo (Nicholas), SZE commented on HDFS-1606:
----------------------------------------------

Still have some FI tests failed.
- Too many open files
{noformat}
2011-04-06 22:27:23,329 WARN  datanode.DataNode (DataXceiverServer.java:run(142)) - DatanodeRegistration(127.0.0.1:41905,
storageID=DS-961198735-127.0.1.1-41905-1302128843146, infoPort=35802, ipcPort=59788):DataXceiveServer:
java.io.IOException: Too many open files
	at sun.nio.ch.ServerSocketChannelImpl.accept0(Native Method)
	at sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:152)
	at sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:84)
	at org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:135)
	at java.lang.Thread.run(Thread.java:662)
{noformat}
- It also seems that there are some timeout.

> Provide a stronger data guarantee in the write pipeline
> -------------------------------------------------------
>
>                 Key: HDFS-1606
>                 URL: https://issues.apache.org/jira/browse/HDFS-1606
>             Project: Hadoop HDFS
>          Issue Type: New Feature
>          Components: data-node, hdfs client, name-node
>    Affects Versions: 0.23.0
>            Reporter: Tsz Wo (Nicholas), SZE
>            Assignee: Tsz Wo (Nicholas), SZE
>             Fix For: 0.23.0
>
>         Attachments: h1606_20110210.patch, h1606_20110211.patch, h1606_20110217.patch,
h1606_20110228.patch, h1606_20110404.patch, h1606_20110405.patch, h1606_20110405b.patch, h1606_20110406.patch,
h1606_20110406b.patch
>
>
> In the current design, if there is a datanode/network failure in the write pipeline,
DFSClient will try to remove the failed datanode from the pipeline and then continue writing
with the remaining datanodes.  As a result, the number of datanodes in the pipeline is decreased.
 Unfortunately, it is possible that DFSClient may incorrectly remove a healthy datanode but
leave the failed datanode in the pipeline because failure detection may be inaccurate under
erroneous conditions.
> We propose to have a new mechanism for adding new datanodes to the pipeline in order
to provide a stronger data guarantee.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira

Mime
View raw message