hadoop-hdfs-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Florin P <florinp...@yahoo.com>
Subject SequenceFile with one very large value
Date Thu, 27 Oct 2011 16:01:10 GMT
 Suppose this scenario:
1. The DFS block 64MB
2. We populate a SequenceFile with a binary value that has 200MB (that represents a PDF file)
In the circumstances of above scenario:
1. How many blocks will be created on HDFS? 
2. The number of blocks will be 200MB/64MB aprox 4 blocks?
3. How many task mappers will created? It is the same number as the number of blocks?
4. If 4 mappers will be created, then one mapper will process the single value of the file,
and the other three are just created and stopped?

I look forward for your answers.
Thank you.

View raw message