hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From yaoxiaohua <yaoxiao...@outlook.com>
Subject RE: dfs.datanode.failed.volumes.tolerated change
Date Fri, 08 Jan 2016 13:54:28 GMT
Hi Kai,

Ok, I will have a look and try, thanks for your  help on this.


Best Regards,



From: Zheng, Kai [mailto:kai.zheng@intel.com] 
Sent: 2016年1月8日 18:06
To: yaoxiaohua
Subject: RE: dfs.datanode.failed.volumes.tolerated change




Please discuss this in the public email thread you initiated, so others may
also help you, in case I don’t or can’t answer your questions.





From: yaoxiaohua [mailto:yaoxiaohua@outlook.com] 
Sent: Friday, January 08, 2016 4:57 PM
To: Zheng, Kai <kai.zheng@intel.com>
Subject: RE: dfs.datanode.failed.volumes.tolerated change


I don’t change anything in dfs.datanode.data.dir property,

I just change  dfs.datanode.failed.volumes.tolerated property.

Thanks for your help.




From: Zheng, Kai [mailto:kai.zheng@intel.com] 
Sent: 2016年1月8日 16:40
To: yaoxiaohua; user@hadoop.apache.org
Subject: RE: dfs.datanode.failed.volumes.tolerated change


As far as I know, Hadoop 2.6 supports disk hot-swapping on a DataNode
without restarting the DataNode. Roughly you need to do two operations: 

1) change dfs.datanode.data.dir in the DataNode configuration to update
according to your removed/added disks;

2) let the DataNode reload its configuration.


Please google and checkout related docs for this feature. Hope this helps.





From: yaoxiaohua [mailto:yaoxiaohua@outlook.com] 
Sent: Friday, January 08, 2016 11:24 AM
To: user@hadoop.apache.org
Subject: dfs.datanode.failed.volumes.tolerated change



                The datanode process shutdown abnormally, 

                I set dfs.datanode.failed.volumes.tolerated =2 in hdfs-site.

                Because I found one disk is fail to access. 

                Then restart the datanode process, it works.


                One day later, we replace a good harddisk , and create
folder and chown to hadoop.

                Then I want to know if I don’t restart datanode process,
when the datanode can know that 

                The disk is good now?

                May I have to restart datanode process to update this?


Env:Hadoop 2.6



Best Regards,



View raw message