hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Lei Chen" <lch...@gmail.com>
Subject How is big file got divided
Date Thu, 20 Apr 2006 06:20:50 GMT
     I am a new user of hadoop. This project looks cool.

     There is one question about the MapReduce. I want to process a big
file. To my understanding, hadoop will partition big file into block and
each block is assigned to a worker. Then, how does hadoop decide where to
cut those big files? Does it guarantee that each line in the input file will
be assigned to one block and no line will be divided into two parts in
different blocks?


  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message