I tried building with a 32 bit install of Java. Now I get the following
error w.r.t zlib. What is the workaround for me ?
[exec] /usr/bin/make all-recursive
[exec] make[1]: Entering directory `/home/alakshman/FB-Projects/hadoop-
0.13.0/build/native/Linux-i386-32'
[exec] Making all in src/org/apache/hadoop/io/compress/zlib
[exec] make[2]: Entering directory `/home/alakshman/FB-Projects/hadoop-
0.13.0/build/native/Linux-i386-32/src/org/apache/hadoop/io/compress/zlib'
[exec] if /bin/sh ../../../../../../../libtool --tag=CC --mode=compile
gcc -DHAVE_CONFIG_H -I.
-I/home/alakshman/FB-Projects/hadoop-0.13.0/src/native/src/org/apache/hadoop/io/compress/zlib
-I../../../../../../.. -I/home/alakshman/jdk1.5.0_06/include
-I/home/alakshman/jdk1.5.0_06/include/linux
-I/home/alakshman/FB-Projects/hadoop-0.13.0/src/native/src -g -Wall -fPIC
-O2 -m32 -g -O2 -MT ZlibCompressor.lo -MD -MP -MF ".deps/ZlibCompressor.Tpo"
-c -o ZlibCompressor.lo
/home/alakshman/FB-Projects/hadoop-0.13.0/src/native/src/org/apache/hadoop/io/compress/zlib/ZlibCompressor.c;
\
[exec] then mv -f ".deps/ZlibCompressor.Tpo"
".deps/ZlibCompressor.Plo"; else rm -f ".deps/ZlibCompressor.Tpo"; exit 1;
fi
[exec] mkdir .libs
[exec] gcc -DHAVE_CONFIG_H -I. -I/home/alakshman/FB-Projects/hadoop-
0.13.0/src/native/src/org/apache/hadoop/io/compress/zlib
-I../../../../../../.. -I/home/alakshman/jdk1.5.0_06/include
-I/home/alakshman/jdk1.5.0_06/include/linux
-I/home/alakshman/FB-Projects/hadoop-0.13.0/src/native/src -g -Wall -fPIC
-O2 -m32 -g -O2 -MT ZlibCompressor.lo -MD -MP -MF .deps/ZlibCompressor.Tpo
-c
/home/alakshman/FB-Projects/hadoop-0.13.0/src/native/src/org/apache/hadoop/io/compress/zlib/ZlibCompressor.c
-fPIC -DPIC -o .libs/ZlibCompressor.o
[exec] In file included from /usr/include/features.h:337,
[exec] from /usr/include/stdio.h:28,
[exec] from /home/alakshman/FB-Projects/hadoop-0.13.0
/src/native/src/org/apache/hadoop/io/compress/zlib/ZlibCompressor.c:24:
[exec] /usr/include/gnu/stubs.h:7:27: error: gnu/stubs-32.h: No such
file or directory
[exec] make[2]: *** [ZlibCompressor.lo] Error 1
[exec] make[2]: Leaving directory `/home/alakshman/FB-Projects/hadoop-
0.13.0/build/native/Linux-i386-32/src/org/apache/hadoop/io/compress/zlib'
[exec] make[1]: *** [all-recursive] Error 1
[exec] make[1]: Leaving directory `/home/alakshman/FB-Projects/hadoop-
0.13.0/build/native/Linux-i386-32'
[exec] make: *** [all] Error 2
Thanks
A
On 6/11/07, Phantom <[EMAIL PROTECTED]> wrote:
I ran the java org.apache.hadoop.util.Platform and it says that it is an
AMD64. Your guess that I have an AMD64 bit install of Java is correct. Now
is there a 64 bit install for x86 Linux ? I don't seem to find one on the
Sun website. Does that mean I have to run this with 32 bit binaries for Java
and for Lzo and zlib too ? Please advice.
On 6/10/07, Arun C Murthy <[EMAIL PROTECTED]> wrote:
>
> There are 2 distinctive native components to hadoop:
> a) libhdfs - Jni based C bindings for the hadoop DFS java libraries.
> b) libhadoop - Native libraries for core-hadoop (for now we have lzo and
> zlib compression libraries:
> http://wiki.apache.org/lucene-hadoop/NativeHadoop as Doug pointed out).
>
> On Fri, Jun 08, 2007 at 04:14:04PM -0700, Phantom wrote:
> >I am running into this wierd build problem - I am building this on
> Fedora
> >Linux x86 64 bit machine but the build is spitting out AMD64 library.
> How
> >can I fix this ? Here is the error from the build :
> >
>
> Looking at the errors below 2 data-points:
>
> > [exec] /bin/sh ../../../../../../../libtool --tag=CC --mode=link
> gcc -g
> >-Wall -fPIC -O2 -m64 -g -O2
> -L/usr/local/jdk1.5.0_07/jre/lib/amd64/server
> >-o libnativezlib.la ZlibCompressor.lo ZlibDecompressor.lo -ldl -ljvm
> -ljvm
>
> and
>
> > [exec] Making all in src/org/apache/hadoop/io/compress/lzo
> > [exec] make[2]: Entering directory
> `/home/alakshman/FB-Projects/hadoop-
> >0.13.0/build/native/Linux-amd64-64/src/org/apache/hadoop/io/compress/lzo'
>
>
> (emphasis on -L/usr/local/jdk1.5.0_07/jre/lib/amd64/server and 0.13.0
> /build/native/Linux-amd64-64/src/org/apache/hadoop/io/compress/lzo)
>
> lead me to suspect that you have a amd64 build of the jvm installed...
> could you run org.apache.hadoop.util.PlatformName and check? Also double
> check the 'JAVA_HOME' env. variable...
>
> Arun
>
> >
> >BUILD FAILED
> >/home/alakshman/FB-Projects/hadoop-0.13.0/build.xml:285: exec returned:
> 2
> >
> >
> >On 6/8/07, Doug Cutting <[EMAIL PROTECTED]> wrote:
> >>
> >>Phantom wrote:
> >>> What are minimal requirements on my Linux machine for building
> libhdfs ?
> >>On
> >>> my Linux box I do not seem to have jni.h and what are the other
> binaries
> >>I
> >>> need for this to work ? Could someone please tell me what is the
> easiest
> >>> way
> >>> to get this done ?
> >>
> >>The following wiki page should help:
> >>
> >>http://wiki.apache.org/lucene-hadoop/NativeHadoop
> >>
> >>Doug
> >>
>