hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Rita <rmorgan...@gmail.com>
Subject Re: decommissioning node woes
Date Fri, 18 Mar 2011 11:23:35 GMT
Any help?


On Wed, Mar 16, 2011 at 9:36 PM, Rita <rmorgan466@gmail.com> wrote:

> Hello,
>
> I have been struggling with decommissioning data  nodes. I have a 50+ data
> node cluster (no MR) with each server holding about 2TB of storage. I split
> the nodes into 2 racks.
>
>
> I edit the 'exclude' file and then do a -refreshNodes. I see the node
> immediate in 'Decommiosied node' and I also see it as a 'live' node!
> Eventhough I wait 24+ hours its still like this. I am suspecting its a bug
> in my version.  The data node process is still running on the node I am
> trying to decommission. So, sometimes I kill -9 the process and I see the
> 'under replicated' blocks...this can't be the normal procedure.
>
> There were even times that I had corrupt blocks because I was impatient --
> waited 24-34 hours
>
> I am using 23 August, 2010: release 0.21.0 <http://hadoop.apache.org/hdfs/releases.html#23+August%2C+2010%3A+release+0.21.0+available>
>  version.
>
> Is this a known bug? Is there anything else I need to do to decommission a
> node?
>
>
>
>
>
>
>
> --
> --- Get your facts first, then you can distort them as you please.--
>



-- 
--- Get your facts first, then you can distort them as you please.--

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message