hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Vinayakumar B (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDFS-9819) FsVolume should tolerate few times check-dir failed due to deletion by mistake
Date Thu, 18 Feb 2016 03:31:18 GMT

    [ https://issues.apache.org/jira/browse/HDFS-9819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15151684#comment-15151684
] 

Vinayakumar B commented on HDFS-9819:
-------------------------------------

Since HDFS-8845, deleting some of the blocks inside <bpDir>/finalized will not cause
any harm by DiskChecker, as it will not check recursively.
But if you delete <bpDir> itself, as [~kihwal] mentioned it looks really strange. This
level of accidental deletion should not be ignored.

> FsVolume should tolerate few times check-dir failed due to deletion by mistake
> ------------------------------------------------------------------------------
>
>                 Key: HDFS-9819
>                 URL: https://issues.apache.org/jira/browse/HDFS-9819
>             Project: Hadoop HDFS
>          Issue Type: Bug
>    Affects Versions: 2.7.1
>            Reporter: Lin Yiqun
>            Assignee: Lin Yiqun
>         Attachments: HDFS-9819.001.patch
>
>
> FsVolume should tolerate few times check-dir failed because sometimes we will do a delete
dir/file operation by mistake in datanode data-dirs. Then the {{DataNode#startCheckDiskErrorThread}}
will invoking checkDir method periodicity and find dir not existed, throw exception. The checked
volume will be added to failed volume list. The blocks on this volume will be replicated again.
But actually, this is not needed to do. We should let volume can be tolerated few times check-dir
failed like config {{dfs.datanode.failed.volumes.tolerated}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Mime
View raw message