hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Chris Kline <ch...@rapleaf.com>
Subject Re: Cannot decommission on 16.4
Date Wed, 09 Jul 2008 14:24:48 GMT
Thanks Lohit.  The key point I missed was that dfs.hosts.exclude  
should exist in before starting the namenode.  It worked after  
restarting hdfs.

-Chris



On Jul 8, 2008, at 3:56 PM, lohit wrote:

> there are few things which aren't documented.
> - you should have defined full path of file as part of  
> dfs.hosts.exclude before starting the namenode. This file must  
> exists, could be a zero length file.
> - While the system is running, you add the hostname (fully  
> qualified) to this file and then invoke hadoop dfsadmin -refreshNodes
> - You should have enough free datanodes in the cluster so that the  
> blocks from this node can be replicated to other nodes. Eg if you  
> replication factor is 3 its good to have atleast 4 datanodes before  
> you decomission any one of them.
> - dfs.namenode.decommission.interval defines interval at which  
> namenode checks if decomission is complete, after which it removes  
> from its list and out of service.
>
> PS : IP address instead of hostname in excludes file should also work.
>
> Thanks,
> Lohit
>
> ----- Original Message ----
> From: Chris Kline <chris@rapleaf.com>
> To: core-user@hadoop.apache.org
> Sent: Tuesday, July 8, 2008 9:38:36 AM
> Subject: Cannot decommission on 16.4
>
> I followed the instruction on the wiki and searched JIRA tickets for
> more info, and still cannot decommission a node on 16.4.  I've tried
> different combinations of hostnames and IPs and nothing seems to work.
>
> Has anyone successfully decommissioned a node on 16.4?  If so, was
> there some trick?  I'm using the exclude method.
>
> -Chris


Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message