hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Dennis Kubes <ku...@apache.org>
Subject Re: compressed input files
Date Thu, 02 Aug 2007 06:48:05 GMT
I don't know about your record count but the link error means that you 
don't have the right version of glibc that was used to compile the 
hadoop native libraries.  It shouldn't matter though as hadoop will fall 
back to the java versions if the native can't be used.

Dennis Kubes

Sandhya E wrote:
> Hi
> I'm trying to pass .gz files as input to hadoop, and at the end of
> mapreduce, the number of input records read from the input files is around
> 480, and when I uncompress the files, the number of input records read is
> around 3000. Why such a difference. Also there is a warning mesg during
> start of execution:
> 07/08/01 23:18:55 DEBUG util.NativeCodeLoader: Trying to load the
> custom-built native-hadoop library...
> 07/08/01 23:18:55 DEBUG util.NativeCodeLoader: Failed to load native-hadoop
> with error: java.lang.UnsatisfiedLinkError:
> /local/offline2/hadoop-0.13.0/lib/native/Linux-i386-32/libhadoop.so:
> /lib/tls/libc.so.6: version `GLIBC_2.4' not found (required by
> /local/offline2/hadoop-0.13.0/lib/native/Linux-i386-32/libhadoop.so)
> Can this be the reason.
> Many Thanks
> Sandhya

View raw message