hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Hudson (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDFS-14187) Make warning message more clear when there are not enough data nodes for EC write
Date Fri, 01 Feb 2019 00:29:00 GMT

    [ https://issues.apache.org/jira/browse/HDFS-14187?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16757826#comment-16757826
] 

Hudson commented on HDFS-14187:
-------------------------------

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #15863 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/15863/])
HDFS-14187. Make warning message more clear when there are not enough (weichiu: rev 0ab7fc92009fec2f0ab341f3d878e1b8864b8ea9)
* (edit) hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/DFSStripedOutputStream.java


> Make warning message more clear when there are not enough data nodes for EC write
> ---------------------------------------------------------------------------------
>
>                 Key: HDFS-14187
>                 URL: https://issues.apache.org/jira/browse/HDFS-14187
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: erasure-coding
>    Affects Versions: 3.1.1
>            Reporter: Kitti Nanasi
>            Assignee: Kitti Nanasi
>            Priority: Major
>             Fix For: 3.3.0
>
>         Attachments: HDFS-14187.001.patch
>
>
> When setting an erasure coding policy for which there are not enough racks or data nodes,
write will fail with the following message:
> {code:java}
> [root@oks-upgrade6727-1 ~]# sudo -u systest hdfs dfs -mkdir /user/systest/testdir
> [root@oks-upgrade6727-1 ~]# sudo -u hdfs hdfs ec -setPolicy -path /user/systest/testdir
> Set default erasure coding policy on /user/systest/testdir
> [root@oks-upgrade6727-1 ~]# sudo -u systest hdfs dfs -put /tmp/file1 /user/systest/testdir
> 18/11/12 05:41:26 WARN hdfs.DFSOutputStream: Cannot allocate parity block(index=3, policy=RS-3-2-1024k).
Not enough datanodes? Exclude nodes=[]
> 18/11/12 05:41:26 WARN hdfs.DFSOutputStream: Cannot allocate parity block(index=4, policy=RS-3-2-1024k).
Not enough datanodes? Exclude nodes=[]
> 18/11/12 05:41:26 WARN hdfs.DFSOutputStream: Block group <1> failed to write 2
blocks. It's at high risk of losing data.
> {code}
> I suggest to log a more descriptive message suggesting to use hdfs ec -verifyCluster
command to verify the cluster setup against the ec policies.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscribe@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-help@hadoop.apache.org


Mime
View raw message