I don't know about your record count but the link error means that you don't have the right version of glibc that was used to compile the hadoop native libraries. It shouldn't matter though as hadoop will fall back to the java versions if the native can't be used.

Dennis Kubes

Sandhya E wrote:
Hi

I'm trying to pass .gz files as input to hadoop, and at the end of
mapreduce, the number of input records read from the input files is around
480, and when I uncompress the files, the number of input records read is
around 3000. Why such a difference. Also there is a warning mesg during
start of execution:
07/08/01 23:18:55 DEBUG util.NativeCodeLoader: Trying to load the
custom-built native-hadoop library...
07/08/01 23:18:55 DEBUG util.NativeCodeLoader: Failed to load native-hadoop
with error: java.lang.UnsatisfiedLinkError:
/local/offline2/hadoop-0.13.0/lib/native/Linux-i386-32/libhadoop.so:
/lib/tls/libc.so.6: version `GLIBC_2.4' not found (required by
/local/offline2/hadoop-0.13.0/lib/native/Linux-i386-32/libhadoop.so)

Can this be the reason.

Many Thanks
Sandhya

Reply via email to