hadoop-general mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Wojciech Langiewicz <wlangiew...@gmail.com>
Subject Re: cleaning up 'hadoop.tmp.dir' ?
Date Tue, 09 Nov 2010 16:28:28 GMT
W dniu 09.11.2010 17:23, Aaron Myers pisze:
> On Tue, Nov 9, 2010 at 6:29 AM, Wojciech Langiewicz
> <wlangiewicz@gmail.com>wrote:
>
>>
>> What might be causing situation where I have about 5TB in HDFS and hadoop
>> tmp dirs have about 16TB in total?
>>
>
> If indeed this is the block data of your HDFS files, then this makes perfect
> sense. HDFS by default replicates every block 3 times, so ~5TB used in HDFS
> is ~15TB raw on disk.

You are right, I wonder why didn't I though about it before.
Thanks for all the answers:)

But name of this option 'hadoop.tmp.dir' is at least a little confusing.

--
Wojciech Langiewicz

Mime
View raw message