I have already exported it in the env. Output of "export" command.
declare -x LD_LIBRARY_PATH="/usr/lib:/usr/local/lib:/home/jony/FreshHadoop/hadoop-0.20.2/build/libhdfs:/usr/lib/jvm/java-6-openjdk/jre/lib/i386/server/:/usr/lib/libfuse.so" Stuti ________________________________________ From: Harsh J [ha...@cloudera.com] Sent: Wednesday, January 04, 2012 5:34 PM To: hdfs-user@hadoop.apache.org Subject: Re: Mounting HDFS Stuti, Your env needs to carry this: export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/path/to/dir/where/libhdfs/files/are/present Otherwise the fuse_dfs binary won't be able to find and load it. The wrapper script does this as part of its setup if you read it. On Wed, Jan 4, 2012 at 5:29 PM, Stuti Awasthi <stutiawas...@hcl.com> wrote: > Im able to mount using command : > > fuse_dfs_wrapper.sh dfs://<server>:<port> /export/hdfs > > -----Original Message----- > From: Stuti Awasthi > Sent: Wednesday, January 04, 2012 5:24 PM > To: hdfs-user@hadoop.apache.org > Subject: RE: Mounting HDFS > > Harsh, > > Output of $file `which fuse_dfs` > > /sbin/fuse_dfs: ELF 32-bit LSB executable, Intel 80386, version 1 (SYSV), > dynamically linked (uses shared libs), for GNU/Linux 2.6.15, not stripped > > Same output for $ file /sbin/fuse_dfs > > Thanks > ________________________________________ > From: Harsh J [ha...@cloudera.com] > Sent: Wednesday, January 04, 2012 5:18 PM > To: hdfs-user@hadoop.apache.org > Subject: Re: Mounting HDFS > > Stuti, > > My original command was "file `which fuse_dfs`", and not just the which > command. > > Can you run "file /sbin/fuse_dfs"? You need the utility called 'file' > available (its mostly present). > > On 04-Jan-2012, at 5:08 PM, Stuti Awasthi wrote: > >> Hi Harsh, >> >> Currently I am using 32 bit Ubuntu11.10, Hadoop 0.20.2 >> >> Output of : $ which fuse_dfs >> /sbin/fuse_dfs >> >> I searched on net and I got this url >> "http://wiki.apache.org/hadoop/MountableHDFS" >> How can I get hdfs fuse deb or rpm packages ?? Thanks for pointing this, can >> you please guide me more on this . >> >> Thanks >> >> -----Original Message----- >> From: Harsh J [mailto:ha...@cloudera.com] >> Sent: Wednesday, January 04, 2012 4:51 PM >> To: hdfs-user@hadoop.apache.org >> Subject: Re: Mounting HDFS >> >> Stuti, >> >> What's your platform - 32-bits or 64-bits? Which one have you built libhdfs >> for? >> >> What's the output of the following? >> $ file `which fuse_dfs` >> >> FWIW, the most hassle free way to do these things today is to use proper >> packages available for your platform, instead of compiling it by yourself. >> Just a suggestion. >> >> On 04-Jan-2012, at 4:28 PM, Stuti Awasthi wrote: >> >>> Hi All, >>> >>> I am following http://wiki.apache.org/hadoop/MountableHDFS for HDFS mount. >>> I have successfully followed the steps till "Installing" and I am able >>> mount it properly. After that I am trying with "Deploying" step and >>> followed the steps: >>> >>> 1. add the following to /etc/fstab >>> fuse_dfs#dfs://hadoop_server.foo.com:9000 /export/hdfs fuse >>> -oallow_other,rw,-ousetrash 0 0 >>> >>> 2. added fuse_dfs to /sbin >>> >>> 3. export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$HADOOP_HOME/build/libhdfs >>> >>> 4. Mount using: mount /export/hdfs. >>> >>> But getting error : >>> fuse_dfs: error while loading shared libraries : libhdfs.so.0: cannot open >>> shared object file: No such file or directory. >>> >>> How to fix this ? >>> >>> Thanks >>> >>> ::DISCLAIMER:: >>> --------------------------------------------------------------------- >>> - >>> ------------------------------------------------- >>> >>> The contents of this e-mail and any attachment(s) are confidential and >>> intended for the named recipient(s) only. >>> It shall not attach any liability on the originator or HCL or its >>> affiliates. Any views or opinions presented in this email are solely those >>> of the author and may not necessarily reflect the opinions of HCL or its >>> affiliates. >>> Any form of reproduction, dissemination, copying, disclosure, >>> modification, distribution and / or publication of this message >>> without the prior written consent of the author of this e-mail is >>> strictly prohibited. If you have received this email in error please delete >>> it and notify the sender immediately. Before opening any mail and >>> attachments please check them for viruses and defect. >>> >>> --------------------------------------------------------------------- >>> - >>> ------------------------------------------------- >> > -- Harsh J