hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Aaron Kimball <aa...@cloudera.com>
Subject Re: tools for scrubbing HDFS data nodes?
Date Thu, 29 Jan 2009 01:57:57 GMT
By "scrub" do you mean delete the blocks from the node?

Read your conf/hadoop-site.xml file to determine where dfs.data.dir points,
then for each directory in that list, just rm the directory. If you want to
ensure that your data is preserved with appropriate replication levels on
the rest of your clutser, you should use Hadoop's DataNode Decommission
feature to up-replicate the data before you blow a copy away.

- Aaron

On Wed, Jan 28, 2009 at 2:10 PM, Sriram Rao <sriramsrao@gmail.com> wrote:

> Hi,
> Is there a tool that one could run on a datanode to scrub all the
> blocks on that node?
> Sriram

  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message