hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Mikkel Kamstrup Erlandsen <...@statsbiblioteket.dk>
Subject Re: Teating large numbers of slaves with scheduled downtime
Date Mon, 24 Jul 2006 11:07:25 GMT
On Mon, 2006-07-24 at 11:00 +0300, Doug Cutting wrote:
> The easiest way would be to not use anything but your reliable machines 
> as datanodes.  Alternately, for better performance, you could run two 
> DFS systems, one on all machines, and one on just the reliable machines, 
> and back one up to the other before you shutdown the "unreliable" nodes 
> each night.  Then, in the morning, restore things.
> Long-term, we hope to add a feature that permits one to remove a number 
> of nodes from DFS at once, forcing all of the blocks stored on these 
> nodes to migrate to other nodes.  But that feature has not yet been 
> implemented.

Ok, thanks. I will have to check how much of a performance hit it will
be to run the reliable boxes as the only data nodes.


View raw message