Hi Adarsh,

It appears you don't have the JVM libraries in your LD_LIBRARY_PATH.  Try this:

export 
LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$JAVA_HOME/jre/lib/amd64:$JAVA_HOME/jre/lib/amd64/server

Brian

On Apr 27, 2011, at 11:31 PM, Adarsh Sharma wrote:

> Dear all,
> 
> Today I am trying to run a simple code by following the below tutorial :-
> 
> 
> http://hadoop.apache.org/hdfs/docs/current/libhdfs.html
> 
> I followed the below steps :-
> 
> 1. Set LD_LIBRARY_PATH & CLASSPATH as :
> 
> export 
> LD_LIBRARY_PATH=/home/hadoop/project/hadoop-0.20.2/c++/Linux-amd64-64/lib:/usr/java/jdk1.6.0_18/jre/lib/amd64/libjava.so
> export 
> CLASSPATH=$CLASSPATH:$HADOOP_HOME:$HADOOP_HOME/lib:/home/hadoop/project/hadoop-0.20.2/c++/Linux-amd64-64/lib:/usr/java/jdk1.6.0_18/jre/lib/amd64
> 
> 2. write above_sample.c program & put it into $HADOOP_HOME/src/c++/libhdfs 
> directory
> 
> 3. After compiling with the below command I am facing issues as :-
> 
> 
> gcc above_sample.c -I/home/hadoop/project/hadoop-0.20.2/src/c++/libhdfs 
> -L/home/hadoop/project/hadoop-0.20.2/src/c++/libhdfs 
> -L/home/hadoop/project/hadoop-0.20.2/c++/Linux-amd64-64/lib 
> -L$HADOOP_HOME/c++/Linux-amd64-64/lib/libhdfs.so.0 -lhdfs -I$HADOOP_HOME 
> -I$HADOOP_HOME/lib /usr/java/jdk1.6.0_18/jre/lib/amd64/server/libjvm.so 
> /home/hadoop/project/hadoop-0.20.2/c++/Linux-amd64-64/lib/libhdfs.so.0 -o 
> above_sample
> 
> bash-3.2$ ./above_sample
> Error occurred during initialization of VM
> Unable to load native library: 
> /home/hadoop/project/hadoop-0.20.2/c++/Linux-amd64-64/libjava.so: cannot open 
> shared object file: No such file or directory
> 
> 
> Now when I try the below command :
> 
> gcc above_sample.c -I/home/hadoop/project/hadoop-0.20.2/src/c++/libhdfs 
> -L/home/hadoop/project/hadoop-0.20.2/src/c++/libhdfs 
> -L/home/hadoop/project/hadoop-0.20.2/c++/Linux-amd64-64/lib 
> -L$HADOOP_HOME/c++/Linux-amd64-64/lib/libhdfs.so.0 -lhdfs -I$HADOOP_HOME 
> -I$HADOOP_HOME/lib /usr/java/jdk1.6.0_18/jre/lib/amd64/server/libjvm.so 
> /home/hadoop/project/hadoop-0.20.2/c++/Linux-amd64-64/lib/libhdfs.so.0 
> /usr/java/jdk1.6.0_18/jre/lib/amd64/libjava.so -o above_sample
> /usr/bin/ld: warning: libverify.so, needed by 
> /usr/java/jdk1.6.0_18/jre/lib/amd64/libjava.so, not found (try using -rpath 
> or -rpath-link)
> /usr/java/jdk1.6.0_18/jre/lib/amd64/libjava.so: undefined reference to 
> `VerifyClassname@SUNWprivate_1.1'
> /usr/java/jdk1.6.0_18/jre/lib/amd64/libjava.so: undefined reference to 
> `VerifyClassForMajorVersion@SUNWprivate_1.1'
> /usr/java/jdk1.6.0_18/jre/lib/amd64/libjava.so: undefined reference to 
> `VerifyFixClassname@SUNWprivate_1.1'
> /usr/java/jdk1.6.0_18/jre/lib/amd64/libjava.so: undefined reference to 
> `VerifyClass@SUNWprivate_1.1'
> collect2: ld returned 1 exit status
> bash-3.2$
> 
> Can Someone guide me the steps needed to run a libhdfs simple program in 
> Hadoop Cluster
> 
> Thanks & best Regards
> Adarsh Sharma

Attachment: smime.p7s
Description: S/MIME cryptographic signature

Reply via email to