hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Manoj Govindassamy (JIRA)" <j...@apache.org>
Subject [jira] [Updated] (HDFS-11340) DataNode reconfigure for disks doesn't remove the failed volumes
Date Sat, 14 Jan 2017 00:55:26 GMT

     [ https://issues.apache.org/jira/browse/HDFS-11340?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]

Manoj Govindassamy updated HDFS-11340:
--------------------------------------
    Attachment: HDFS-11340.01.patch

Attached v01 to address the following
* {{DataNode#parseChangedVolumes}} updated to rightly detect the removal of a failed volume
during Reconfigure with new conf
* {{DataNode#refreshVolumes}} updated to include failed volume count to verify for valid new
conf
* {{FsDataSetImpl#removeVolumes}} updated to prune failed volumes list at the time of reconfigure
* {{TestDataNodeVolumeFailureReporting}} updated with new unit test and corner cases around
reconfigure with failed volume removal

[~eddyxu], can you please take a look at the patch ?

> DataNode reconfigure for disks doesn't remove the failed volumes
> ----------------------------------------------------------------
>
>                 Key: HDFS-11340
>                 URL: https://issues.apache.org/jira/browse/HDFS-11340
>             Project: Hadoop HDFS
>          Issue Type: Bug
>    Affects Versions: 3.0.0-alpha1
>            Reporter: Manoj Govindassamy
>            Assignee: Manoj Govindassamy
>         Attachments: HDFS-11340.01.patch
>
>
> Say a DataNode (uuid:xyz) has disks D1 and D2. When D1 turns bad, JMX query on FSDatasetState-xyz
for "NumFailedVolumes" attr rightly shows the failed volume count as 1 and the "FailedStorageLocations"
attr has the failed storage location as "D1".
> It is possible to add or remove disks to this DataNode by running {{reconfigure}} command.
Let the failed disk D1 be removed from the conf and the new conf has only one good disk D2.
Upon running the reconfigure command for this DataNode with this new disk conf, the expectation
is DataNode would no more have "NumFailedVolumes" or "FailedStorageLocations". But, even after
removing the failed disk from the conf and a successful reconfigure, DataNode continues to
show the "NumFailedVolumes" as 1 and "FailedStorageLocations" as "D1" and it never gets reset.




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscribe@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-help@hadoop.apache.org


Mime
View raw message