hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Ken Goodhope <kengoodh...@gmail.com>
Subject Re: using 'fs -put' from datanode: all data written to that node's hdfs and not distributed
Date Tue, 13 Jul 2010 00:46:27 GMT
All writes from a datanode leave one copy on the local node, one copy
on another node in the same rack, and a third on another rack if
available.

On 7/12/10, Nathan Grice <ngrice@gmail.com> wrote:
> We are trying to load data into hdfs from one of the slaves and when the put
> command is run from a slave(datanode) all of the blocks are written to the
> datanode's hdfs, and not distributed to all of the nodes in the cluster. It
> does not seem to matter what destination format we use ( /filename vs
> hdfs://master:9000/filename) it always behaves the same.
> Conversely, running the same command from the namenode distributes the files
> across the datanodes.
>
> Is there something I am missing?
>
> -Nathan
>

Mime
View raw message