The command works correctly. But I still get the error for running the
fuse_dfs_wrapper.sh script:

./fuse_dfs: error while loading shared libraries: libhdfs.so.0: cannot open
shared object file: No such file or directory

Aastha.

On 1 August 2011 10:03, Arun C Murthy <a...@hortonworks.com> wrote:

> Run the following command:
>
> $ ant -Dcompile.native=true package
>
> Arun
>
> On Jul 31, 2011, at 9:20 PM, Aastha Mehta wrote:
>
> > Hi Arun,
> >
> > Thanks for the prompt reply. I am not sure, I understood you correctly.
> > Compile/binary/tar of what? The native files? The
> lib/native/Linux-amd64-64/
> > contains following files:
> > libhadoop.a
> > libhadoop.la
> > libhadoop.so
> > libhadoop.so.1
> > libhadoop.so.1.0.0
> >
> > This directory is present in the package itself. So, should I make a tar
> of
> > it and then provide it? I tried the following, but it failed:
> > ant -Dcompile.native=true
> > $HADOOP_HOME/lib/native/Linux-amd64-64/libhadoop.so
> >
> > The error I got is - "Target  lib/native/Linux-amd64-64/libhadoop.so does
> > not exist in the project Hadoop".
> >
> > Thanks,
> > Aastha.
> >
> > On 1 August 2011 09:44, Arun Murthy <a...@hortonworks.com> wrote:
> >
> >> <target> could be compile or binary or tar.
> >>
> >> Arun
> >>
> >> Sent from my iPhone
> >>
> >> On Jul 31, 2011, at 9:05 PM, Aastha Mehta <aasth...@gmail.com> wrote:
> >>
> >>> Hello,
> >>>
> >>> I am trying to run fuse_dfs_wrapper.sh from
> >>> hadoop-0.20.2/src/contrib/fuse_dfs/src on a 64-bit machine. I get the
> >>> following error:
> >>> ./fuse_dfs: error while loading shared libraries: libhdfs.so.0: cannot
> >> open
> >>> shared object file: No such file or directory
> >>>
> >>> I searched on the net and found a response to a similar query here -
> >>>
> >>
> https://groups.google.com/a/cloudera.org/group/cdh-user/browse_thread/thread/3db7efc10cff8bbc?pli=1
> >>>
> >>> My hadoop package contains the native files in
> >>> hadoop-0.20.2/lib/native/Linux-amd64-64/
> >>>
> >>> I followed to this link -
> >>> http://hadoop.apache.org/common/docs/current/native_libraries.html to
> >>> understand the steps to build hadoop native libraries.
> >>>
> >>> I have a small query regarding the building step. On the above link, it
> >> is
> >>> mentioned -
> >>>
> >>> "Once you installed the prerequisite packages use the standard hadoop
> >>> build.xml file and pass along the compile.native flag (set to true) to
> >> build
> >>> the native hadoop library:
> >>>
> >>> $ ant -Dcompile.native=true <target>
> >>>
> >>> You should see the newly-built library in:
> >>>
> >>> $ build/native/<platform>/lib
> >>>
> >>> where <platform> is a combination of the system-properties: ${os.name
> >>> }-${os.arch}-${sun.arch.data.model} (for example, Linux-i386-32)."
> >>>
> >>>
> >>> Could someone please tell what exactly is <target> in the first step.
> >>>
> >>>
> >>> Thanks and regards,
> >>>
> >>> Aastha.
> >>>
> >>>
> >>>
> >>>
> >>>
> >>> --
> >>> Aastha Mehta
> >>> B.E. (Hons.) Computer Science
> >>> BITS Pilani
> >>> E-mail: aasth...@gmail.com
> >>
> >
> >
> >
> > --
> > Aastha Mehta
> > B.E. (Hons.) Computer Science
> > BITS Pilani
> > E-mail: aasth...@gmail.com
>
>


-- 
Aastha Mehta
B.E. (Hons.) Computer Science
BITS Pilani
E-mail: aasth...@gmail.com

Reply via email to