Hey Seraph,

fuse_dfs below should be a wrapper script that sets LD_LIBRARY_PATH,
JVM_LIB, JAVA_HOME and calls the binary fuse_dfs.

Thanks,
Eli

On Mon, May 10, 2010 at 12:46 PM, Seraph Imalia <ser...@eisp.co.za> wrote:
> Hi Eli,
>
> I was about to try what you suggest but I think I am on to something.
>
> With the new CDH2 I have noticed something new - I am able to mount the dfs
> both in and out of debug mode provided I use the fuse_dfs_wrapper.sh script
> to do it.
>
> But, it does not work when putting the mount into /etc/fstab.  I then tried
> setting the environment variables in the wrapper script to invalid values
> and was able to recreate the same behavior as the /etc/fstab way.
>
> So basically, it looks like things have improved with the use of CDH2 but
> now when using /etc/fstab, I suspect it is unable to access the following
> environment variables...
> export HADOOP_HOME=/opt/hadoop-0.20.1+169.68
> export JAVA_HOME=/usr/lib64/jre
>
> I have added them to the end of /etc/bash.bashrc which did not work.  What
> else can I do to make sure they exist when "mount /Volumes/hdfs" is run.
>  Here is the line in /etc/fstab...
>
> fuse_dfs#dfs://dynobuntu10:8020 /Volumes/hdfs fuse allow_other,rw,big_writes
> 0 0
>
> Regards,
> Seraph
>
>
>
> On 10 May 2010, at 2:36 AM, Eli Collins wrote:
>
>> Hey Seraph,
>>
>> fuse_impls_getattr.c  connects via  hdfsConnectAsUser so you should
>> see a log (unless its returning from a case that doesn't print an
>> error). Next step is to determine that you're actually reaching the
>> code you modified by adding a syslog to the top of the function (need
>> to make sure you're actually loading the libhdfs you've built vs an
>> older one or another one installed on your system), and then determine
>> which error case in that function you're seeing. It's strange that -d
>> would cause that path to change.
>>
>> I don't use -d on our internal cluster so I know that case can work.
>> Here's how we mount in fstab:
>>
>> fuse_dfs#dfs://<host>:8020   /exports/hdfs fuse allow_other,usetrash,rw 2
>> 0
>>
>> Thanks,
>> Eli
>>
>> On Sun, May 9, 2010 at 7:37 AM, Seraph Imalia <ser...@eisp.co.za> wrote:
>>>
>>> Hi Eli,
>>>
>>> I have changed fprintfs to syslogs and recompiled and it does not seem to
>>> show me what is happening - nothing new in the syslog.
>>> I also saw a function called hdfsConnectAsUserNewInstance - should the
>>> fprintfs be chaged to syslogs in there too?  I am not very experienced in
>>> c++ so please forgive my ignorant questions here.
>>>
>>> I also downloaded a CDH2 release and have experienced the same problem
>>> with
>>> that version too.  I downloaded it from here...
>>> http://archive.cloudera.com/docs/_choosing_a_version.html
>>>
>>> Regards,
>>> Seraph
>>>
>>>
>>> On 07 May 2010, at 5:58 PM, Eli Collins wrote:
>>>
>>>> Converting the fprintfs to syslogs in hdfsConnectAsUser in
>>>> src/c++/libhdfs/hdfs.c (and doing a clean build)  should let you see
>>>> the particular reason it can't connect. It's weird that it can connect
>>>> w/ the debug option but now w/o.
>>>>
>>>> Thanks,
>>>> Eli
>>>>
>>>>
>>>> On Thu, May 6, 2010 at 11:56 AM, Seraph Imalia <ser...@eisp.co.za>
>>>> wrote:
>>>>>
>>>>> I don't think that is happening... I have checked that before, but here
>>>>> is a
>>>>> dump of me checking again...
>>>>> r...@dynobuntu17:/opt/hadoop-0.20.2/contrib/fuse-dfs# ps aux | grep
>>>>> fuse
>>>>> root     27947  0.0  0.0   7524   892 pts/0    R+   20:49   0:00 grep
>>>>> fuse
>>>>> r...@dynobuntu17:/opt/hadoop-0.20.2/contrib/fuse-dfs#
>>>>> ./fuse_dfs_wrapper.sh
>>>>> dfs://dynobuntu10:9000 /Volumes/hdfs -obig_writes
>>>>> port=9000,server=dynobuntu10
>>>>> fuse-dfs didn't recognize /Volumes/hdfs,-2
>>>>> r...@dynobuntu17:/opt/hadoop-0.20.2/contrib/fuse-dfs# ps aux | grep
>>>>> fuse
>>>>> root     27952  0.3  0.1 2273632 10912 ?       Ssl  20:49   0:00
>>>>> fuse_dfs
>>>>> dfs://dynobuntu10:9000 /Volumes/hdfs -obig_writes
>>>>> root     27969  0.0  0.0   7524   904 pts/0    S+   20:49   0:00 grep
>>>>> fuse
>>>>> r...@dynobuntu17:/opt/hadoop-0.20.2/contrib/fuse-dfs# ls /Volumes/hdfs
>>>>> ls: cannot access /Volumes/hdfs: Input/output error
>>>>> r...@dynobuntu17:/opt/hadoop-0.20.2/contrib/fuse-dfs# tail --lines=2
>>>>> /var/log/syslog
>>>>> May  6 20:49:38 dynobuntu17 fuse_dfs: mounting dynobuntu10:9000
>>>>> May  6 20:49:45 dynobuntu17 fuse_dfs: ERROR: could not connect to
>>>>> dynobuntu10:9000 fuse_impls_getattr.c:37
>>>>> May  6 20:49:46 dynobuntu17 fuse_dfs: ERROR: could not connect to
>>>>> dynobuntu10:9000 fuse_impls_getattr.c:37
>>>>> r...@dynobuntu17:/opt/hadoop-0.20.2/contrib/fuse-dfs# umount
>>>>> /Volumes/hdfs/
>>>>> r...@dynobuntu17:/opt/hadoop-0.20.2/contrib/fuse-dfs# ps aux | grep
>>>>> fuse
>>>>> root     28063  0.0  0.0   7524   908 pts/0    S+   20:50   0:00 grep
>>>>> fuse
>>>>> r...@dynobuntu17:/opt/hadoop-0.20.2/contrib/fuse-dfs#
>>>>> The only thing I can find on the web is on the issues.apache.org where
>>>>> two
>>>>> people experienced the problem and one of them thought it had something
>>>>> to
>>>>> do with fuse itself.  If I knew what I could change on fuse I would do
>>>>> it
>>>>> -
>>>>> I am quite desperate to get this working without the debug option - I
>>>>> am
>>>>> so
>>>>> close.
>>>>> Here is the a link I
>>>>> found:
>>>>>
>>>>> https://issues.apache.org/jira/browse/HADOOP-4?focusedCommentId=12563182&page=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#action_12563182
>>>>> Regards,
>>>>> Seraph
>>>>>
>>>>> On 06 May 2010, at 5:44 PM, Eli Collins wrote:
>>>>>
>>>>> Hey Seraph,
>>>>>
>>>>> The -d is just a debug option, to print output to the terminal,
>>>>> shouldn't change the execution. I suspect something else is going on,
>>>>> perhaps you have an old fuse process running?
>>>>>
>>>>> Thanks,
>>>>> Eli
>>>>>
>>>>> On Thu, May 6, 2010 at 4:55 AM, Seraph Imalia <ser...@eisp.co.za>
>>>>> wrote:
>>>>>
>>>>> Hi,
>>>>>
>>>>> I am experiencing an annoying problem...
>>>>>
>>>>> When I run this "./fuse_dfs_wrapper.sh dfs://dynobuntu10:8020
>>>>> /Volumes/hdfs
>>>>>
>>>>> -obig_writes -d" then everything works fine.
>>>>>
>>>>> When I run this "./fuse_dfs_wrapper.sh dfs://dynobuntu10:8020
>>>>> /Volumes/hdfs
>>>>>
>>>>> -obig_writes" (i.e. remove the daemonize option) I get issues
>>>>> connecting
>>>>> to
>>>>>
>>>>> hadoop...
>>>>>
>>>>> Here is what is in /var/log/syslog after running "ls /Volumes/hdfs"
>>>>>
>>>>> May  6 12:57:16 dynobuntu17 fuse_dfs: ERROR: could not connect to
>>>>>
>>>>> dynobuntu10:8020 fuse_impls_getattr.c:37
>>>>>
>>>>> May  6 12:57:23 dynobuntu17 fuse_dfs: ERROR: could not connect to
>>>>>
>>>>> dynobuntu10:8020 fuse_impls_getattr.c:37
>>>>>
>>>>> No errors appear in the syslog when the -d option is specified.
>>>>>
>>>>> I also tried port 9000 with exactly the same results.  There is no
>>>>> firewall
>>>>>
>>>>> software installed on any of the servers.
>>>>>
>>>>> It almost appears as something is stopping the connection to a the
>>>>>
>>>>> nameserver when fuse_dfs runs as a daemon.  Please help.
>>>>>
>>>>> We are running OS:  Ubuntu 9.04 x64 (2.6.28-11-server #42-Ubuntu SMP
>>>>> Fri
>>>>> Apr
>>>>>
>>>>> 17 02:45:36 UTC 2009 GNU/Linux)
>>>>>
>>>>> Seraph
>>>>>
>>>>>
>>>>>
>>>>> On 06 May 2010, at 11:31 AM, Seraph Imalia wrote:
>>>>>
>>>>> Hi Eli,
>>>>>
>>>>> Thank you very much - applying patch HDFS-961-2.patch and re-building
>>>>>
>>>>> resolved the problem.
>>>>>
>>>>> Regards,
>>>>>
>>>>> Seraph
>>>>>
>>>>> On 05 May 2010, at 11:20 PM, Eli Collins wrote:
>>>>>
>>>>> Try using a port besides 8020 or applying the patch for HDFS-961.
>>>>>
>>>>> Thanks,
>>>>>
>>>>> Eli
>>>>>
>>>>> On Wed, May 5, 2010 at 9:02 AM, Seraph Imalia <ser...@eisp.co.za>
>>>>> wrote:
>>>>>
>>>>> I have successfully managed to mount a dfs - but here is what I am
>>>>>
>>>>> experiencing...
>>>>>
>>>>> I mounted to /Volumes/hdfs  when I run "ls /Volumes/hdfs" it works fine
>>>>>
>>>>> and
>>>>>
>>>>> returns no results.
>>>>>
>>>>> so I ran "mkdir /Volumes/hdfs/test" and it was successful and I could
>>>>>
>>>>> see
>>>>>
>>>>> the new directory exists using the web interface to browse hadoop.
>>>>>
>>>>> but now when I run "ls /Volumes/hdfs" it does this...
>>>>>
>>>>> ls: cannot access /Volumes/hdfs/p-site.x0: No such file or directory
>>>>>
>>>>> p-site.x0
>>>>>
>>>>> and on the terminal window where I mounted it, it dumps this...
>>>>>
>>>>> LOOKUP /p-site.x0
>>>>>
>>>>>  unique: 13, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 14, opcode: READDIR (28), nodeid: 1, insize: 80
>>>>>
>>>>>  unique: 14, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 15, opcode: RELEASEDIR (29), nodeid: 1, insize: 64
>>>>>
>>>>>  unique: 15, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 16, opcode: GETATTR (3), nodeid: 1, insize: 56
>>>>>
>>>>>  unique: 16, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 17, opcode: OPENDIR (27), nodeid: 1, insize: 48
>>>>>
>>>>>  unique: 17, error: 0 (Success), outsize: 32
>>>>>
>>>>> unique: 18, opcode: READDIR (28), nodeid: 1, insize: 80
>>>>>
>>>>>  unique: 18, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 19, opcode: GETATTR (3), nodeid: 1, insize: 56
>>>>>
>>>>>  unique: 19, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 42
>>>>>
>>>>> I am not sure what this debug info means?
>>>>>
>>>>> Here is everything that happened on terminal 1 and 2...
>>>>>
>>>>> terminal 1
>>>>>
>>>>> r...@dynobuntu17:/opt/hadoop-0.20.2#
>>>>>
>>>>> contrib/fuse-dfs/fuse_dfs_wrapper.sh
>>>>>
>>>>> dfs://dynobuntu10:8020 /Volumes/hdfs -d
>>>>>
>>>>> port=8020,server=dynobuntu10
>>>>>
>>>>> fuse-dfs didn't recognize /Volumes/hdfs,-2
>>>>>
>>>>> fuse-dfs ignoring option -d
>>>>>
>>>>> unique: 1, opcode: INIT (26), nodeid: 0, insize: 56
>>>>>
>>>>> INIT: 7.10
>>>>>
>>>>> flags=0x0000003b
>>>>>
>>>>> max_readahead=0x00020000
>>>>>
>>>>>  INIT: 7.8
>>>>>
>>>>>  flags=0x00000001
>>>>>
>>>>>  max_readahead=0x00020000
>>>>>
>>>>>  max_write=0x00020000
>>>>>
>>>>>  unique: 1, error: 0 (Success), outsize: 40
>>>>>
>>>>> unique: 2, opcode: GETATTR (3), nodeid: 1, insize: 56
>>>>>
>>>>>  unique: 2, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 3, opcode: OPENDIR (27), nodeid: 1, insize: 48
>>>>>
>>>>>  unique: 3, error: 0 (Success), outsize: 32
>>>>>
>>>>> unique: 4, opcode: READDIR (28), nodeid: 1, insize: 80
>>>>>
>>>>>  unique: 4, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 5, opcode: RELEASEDIR (29), nodeid: 1, insize: 64
>>>>>
>>>>>  unique: 5, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 6, opcode: GETATTR (3), nodeid: 1, insize: 56
>>>>>
>>>>>  unique: 6, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 7, opcode: LOOKUP (1), nodeid: 1, insize: 45
>>>>>
>>>>> LOOKUP /test
>>>>>
>>>>>  unique: 7, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 8, opcode: MKDIR (9), nodeid: 1, insize: 53
>>>>>
>>>>> MKDIR /test
>>>>>
>>>>>  NODEID: 2
>>>>>
>>>>>  unique: 8, error: 0 (Success), outsize: 136
>>>>>
>>>>> unique: 9, opcode: GETATTR (3), nodeid: 1, insize: 56
>>>>>
>>>>>  unique: 9, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 10, opcode: OPENDIR (27), nodeid: 1, insize: 48
>>>>>
>>>>>  unique: 10, error: 0 (Success), outsize: 32
>>>>>
>>>>> unique: 11, opcode: READDIR (28), nodeid: 1, insize: 80
>>>>>
>>>>>  unique: 11, error: 0 (Success), outsize: 120
>>>>>
>>>>> unique: 12, opcode: GETATTR (3), nodeid: 1, insize: 56
>>>>>
>>>>>  unique: 12, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 13, opcode: LOOKUP (1), nodeid: 1, insize: 50
>>>>>
>>>>> LOOKUP /p-site.x0
>>>>>
>>>>>  unique: 13, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 14, opcode: READDIR (28), nodeid: 1, insize: 80
>>>>>
>>>>>  unique: 14, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 15, opcode: RELEASEDIR (29), nodeid: 1, insize: 64
>>>>>
>>>>>  unique: 15, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 16, opcode: GETATTR (3), nodeid: 1, insize: 56
>>>>>
>>>>>  unique: 16, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 17, opcode: OPENDIR (27), nodeid: 1, insize: 48
>>>>>
>>>>>  unique: 17, error: 0 (Success), outsize: 32
>>>>>
>>>>> unique: 18, opcode: READDIR (28), nodeid: 1, insize: 80
>>>>>
>>>>>  unique: 18, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 19, opcode: GETATTR (3), nodeid: 1, insize: 56
>>>>>
>>>>>  unique: 19, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 20, opcode: LOOKUP (1), nodeid: 1, insize: 42
>>>>>
>>>>> LOOKUP /%
>>>>>
>>>>>  unique: 20, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 21, opcode: READDIR (28), nodeid: 1, insize: 80
>>>>>
>>>>>  unique: 21, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 22, opcode: RELEASEDIR (29), nodeid: 1, insize: 64
>>>>>
>>>>>  unique: 22, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 23, opcode: GETATTR (3), nodeid: 1, insize: 56
>>>>>
>>>>>  unique: 23, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 24, opcode: OPENDIR (27), nodeid: 1, insize: 48
>>>>>
>>>>>  unique: 24, error: 0 (Success), outsize: 32
>>>>>
>>>>> unique: 25, opcode: READDIR (28), nodeid: 1, insize: 80
>>>>>
>>>>>  unique: 25, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 26, opcode: GETATTR (3), nodeid: 1, insize: 56
>>>>>
>>>>>  unique: 26, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 27, opcode: LOOKUP (1), nodeid: 1, insize: 42
>>>>>
>>>>> LOOKUP /1
>>>>>
>>>>>  unique: 27, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 28, opcode: READDIR (28), nodeid: 1, insize: 80
>>>>>
>>>>>  unique: 28, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 29, opcode: RELEASEDIR (29), nodeid: 1, insize: 64
>>>>>
>>>>>  unique: 29, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 30, opcode: GETATTR (3), nodeid: 1, insize: 56
>>>>>
>>>>>  unique: 30, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 31, opcode: OPENDIR (27), nodeid: 2, insize: 48
>>>>>
>>>>>  unique: 31, error: 0 (Success), outsize: 32
>>>>>
>>>>> unique: 32, opcode: READDIR (28), nodeid: 2, insize: 80
>>>>>
>>>>>  unique: 32, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 33, opcode: RELEASEDIR (29), nodeid: 2, insize: 64
>>>>>
>>>>>  unique: 33, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 34, opcode: OPENDIR (27), nodeid: 1, insize: 48
>>>>>
>>>>>  unique: 34, error: 0 (Success), outsize: 32
>>>>>
>>>>> unique: 35, opcode: READDIR (28), nodeid: 1, insize: 80
>>>>>
>>>>>  unique: 35, error: 0 (Success), outsize: 120
>>>>>
>>>>> unique: 36, opcode: GETATTR (3), nodeid: 1, insize: 56
>>>>>
>>>>>  unique: 36, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 37, opcode: LOOKUP (1), nodeid: 1, insize: 50
>>>>>
>>>>> LOOKUP /d-site.x0
>>>>>
>>>>>  unique: 37, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 38, opcode: READDIR (28), nodeid: 1, insize: 80
>>>>>
>>>>>  unique: 38, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 39, opcode: RELEASEDIR (29), nodeid: 1, insize: 64
>>>>>
>>>>>  unique: 39, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 40, opcode: GETATTR (3), nodeid: 1, insize: 56
>>>>>
>>>>>  unique: 40, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 41, opcode: LOOKUP (1), nodeid: 1, insize: 45
>>>>>
>>>>> LOOKUP /test
>>>>>
>>>>>  NODEID: 2
>>>>>
>>>>>  unique: 41, error: 0 (Success), outsize: 136
>>>>>
>>>>> unique: 42, opcode: LOOKUP (1), nodeid: 2, insize: 47
>>>>>
>>>>> LOOKUP /test/inside
>>>>>
>>>>>  unique: 42, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 43, opcode: MKDIR (9), nodeid: 2, insize: 55
>>>>>
>>>>> MKDIR /test/inside
>>>>>
>>>>>  NODEID: 3
>>>>>
>>>>>  unique: 43, error: 0 (Success), outsize: 136
>>>>>
>>>>> unique: 44, opcode: GETATTR (3), nodeid: 2, insize: 56
>>>>>
>>>>>  unique: 44, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 45, opcode: OPENDIR (27), nodeid: 2, insize: 48
>>>>>
>>>>>  unique: 45, error: 0 (Success), outsize: 32
>>>>>
>>>>> unique: 46, opcode: READDIR (28), nodeid: 2, insize: 80
>>>>>
>>>>>  unique: 46, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 47, opcode: GETATTR (3), nodeid: 2, insize: 56
>>>>>
>>>>>  unique: 47, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 48, opcode: LOOKUP (1), nodeid: 2, insize: 42
>>>>>
>>>>> LOOKUP /test/e
>>>>>
>>>>>  unique: 48, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 49, opcode: READDIR (28), nodeid: 2, insize: 80
>>>>>
>>>>>  unique: 49, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 50, opcode: RELEASEDIR (29), nodeid: 2, insize: 64
>>>>>
>>>>>  unique: 50, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 51, opcode: GETATTR (3), nodeid: 2, insize: 56
>>>>>
>>>>>  unique: 51, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 52, opcode: OPENDIR (27), nodeid: 2, insize: 48
>>>>>
>>>>>  unique: 52, error: 0 (Success), outsize: 32
>>>>>
>>>>> unique: 53, opcode: READDIR (28), nodeid: 2, insize: 80
>>>>>
>>>>>  unique: 53, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 54, opcode: GETATTR (3), nodeid: 2, insize: 56
>>>>>
>>>>>  unique: 54, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 55, opcode: LOOKUP (1), nodeid: 2, insize: 42
>>>>>
>>>>> LOOKUP /test/e
>>>>>
>>>>>  unique: 55, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 56, opcode: READDIR (28), nodeid: 2, insize: 80
>>>>>
>>>>>  unique: 56, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 57, opcode: RELEASEDIR (29), nodeid: 2, insize: 64
>>>>>
>>>>>  unique: 57, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 58, opcode: GETATTR (3), nodeid: 2, insize: 56
>>>>>
>>>>>  unique: 58, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 59, opcode: OPENDIR (27), nodeid: 2, insize: 48
>>>>>
>>>>>  unique: 59, error: 0 (Success), outsize: 32
>>>>>
>>>>> unique: 60, opcode: READDIR (28), nodeid: 2, insize: 80
>>>>>
>>>>>  unique: 60, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 61, opcode: GETATTR (3), nodeid: 2, insize: 56
>>>>>
>>>>>  unique: 61, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 62, opcode: LOOKUP (1), nodeid: 2, insize: 42
>>>>>
>>>>> LOOKUP /test/e
>>>>>
>>>>>  unique: 62, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 63, opcode: READDIR (28), nodeid: 2, insize: 80
>>>>>
>>>>>  unique: 63, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 64, opcode: RELEASEDIR (29), nodeid: 2, insize: 64
>>>>>
>>>>>  unique: 64, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 65, opcode: GETATTR (3), nodeid: 2, insize: 56
>>>>>
>>>>>  unique: 65, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 66, opcode: OPENDIR (27), nodeid: 2, insize: 48
>>>>>
>>>>>  unique: 66, error: 0 (Success), outsize: 32
>>>>>
>>>>> unique: 67, opcode: READDIR (28), nodeid: 2, insize: 80
>>>>>
>>>>>  unique: 67, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 68, opcode: GETATTR (3), nodeid: 2, insize: 56
>>>>>
>>>>>  unique: 68, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 69, opcode: LOOKUP (1), nodeid: 2, insize: 42
>>>>>
>>>>> LOOKUP /test/e
>>>>>
>>>>>  unique: 69, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 70, opcode: READDIR (28), nodeid: 2, insize: 80
>>>>>
>>>>>  unique: 70, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 71, opcode: RELEASEDIR (29), nodeid: 2, insize: 64
>>>>>
>>>>>  unique: 71, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 72, opcode: GETATTR (3), nodeid: 2, insize: 56
>>>>>
>>>>>  unique: 72, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 73, opcode: OPENDIR (27), nodeid: 2, insize: 48
>>>>>
>>>>>  unique: 73, error: 0 (Success), outsize: 32
>>>>>
>>>>> unique: 74, opcode: READDIR (28), nodeid: 2, insize: 80
>>>>>
>>>>>  unique: 74, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 75, opcode: GETATTR (3), nodeid: 2, insize: 56
>>>>>
>>>>>  unique: 75, error: 0 (Success), outsize: 112
>>>>>
>>>>> unique: 76, opcode: LOOKUP (1), nodeid: 2, insize: 42
>>>>>
>>>>> LOOKUP /test/e
>>>>>
>>>>>  unique: 76, error: -2 (No such file or directory), outsize: 16
>>>>>
>>>>> unique: 77, opcode: READDIR (28), nodeid: 2, insize: 80
>>>>>
>>>>>  unique: 77, error: 0 (Success), outsize: 16
>>>>>
>>>>> unique: 78, opcode: RELEASEDIR (29), nodeid: 2, insize: 64
>>>>>
>>>>>  unique: 78, error: 0 (Success), outsize: 16
>>>>>
>>>>> terminal 2
>>>>>
>>>>> r...@dynobuntu17:~# ls /Volumes/hdfs/
>>>>>
>>>>> r...@dynobuntu17:~# mkdir /Volumes/hdfs/test
>>>>>
>>>>> r...@dynobuntu17:~# ls /Volumes/hdfs/
>>>>>
>>>>> ls: cannot access /Volumes/hdfs/p-site.x0: No such file or directory
>>>>>
>>>>> p-site.x0
>>>>>
>>>>> r...@dynobuntu17:~# ls /Volumes/hdfs/
>>>>>
>>>>> ls: cannot access /Volumes/hdfs/%: No such file or directory
>>>>>
>>>>> %
>>>>>
>>>>> r...@dynobuntu17:~# ls /Volumes/hdfs
>>>>>
>>>>> ls: cannot access /Volumes/hdfs/1: No such file or directory
>>>>>
>>>>> 1
>>>>>
>>>>> r...@dynobuntu17:~# ls /Volumes/hdfs/test
>>>>>
>>>>> r...@dynobuntu17:~# ls /Volumes/hdfs/
>>>>>
>>>>> ls: cannot access /Volumes/hdfs/d-site.x0: No such file or directory
>>>>>
>>>>> d-site.x0
>>>>>
>>>>> r...@dynobuntu17:~# mkdir /Volumes/hdfs/test/inside
>>>>>
>>>>> r...@dynobuntu17:~# ls /Volumes/hdfs/test
>>>>>
>>>>> ls: cannot access /Volumes/hdfs/test/e: No such file or directory
>>>>>
>>>>> e
>>>>>
>>>>> r...@dynobuntu17:~# ls /Volumes/hdfs/test
>>>>>
>>>>> ls: cannot access /Volumes/hdfs/test/e: No such file or directory
>>>>>
>>>>> e
>>>>>
>>>>> r...@dynobuntu17:~# ls /Volumes/hdfs/\test
>>>>>
>>>>> ls: cannot access /Volumes/hdfs/test/e: No such file or directory
>>>>>
>>>>> e
>>>>>
>>>>> r...@dynobuntu17:~# ls /Volumes/hdfs/\te\s\t
>>>>>
>>>>> ls: cannot access /Volumes/hdfs/test/e: No such file or directory
>>>>>
>>>>> e
>>>>>
>>>>> r...@dynobuntu17:~# ls /Volumes/hdfs/\t\e\s\t
>>>>>
>>>>> ls: cannot access /Volumes/hdfs/test/e: No such file or directory
>>>>>
>>>>> e
>>>>>
>>>>> On 05 May 2010, at 4:57 PM, Seraph Imalia wrote:
>>>>>
>>>>> Awesome! - thank you  after running apt-get install libfuse-dev I got a
>>>>>
>>>>> successful build.
>>>>>
>>>>> Thanks for your help :)
>>>>>
>>>>>
>>>>> On 05 May 2010, at 4:19 PM, Jason Venner wrote:
>>>>>
>>>>> It does not look like you have the fuse-devel package installed on your
>>>>>
>>>>> system.
>>>>>
>>>>> On Wed, May 5, 2010 at 7:14 AM, Seraph Imalia <ser...@eisp.co.za>
>>>>> wrote:
>>>>>
>>>>> Hi Jason,
>>>>>
>>>>> Thank you,  I had autoconf installed but not automake - after
>>>>> installing
>>>>>
>>>>> automake, the build went further, but this time failed with...
>>>>>
>>>>> BUILD FAILED
>>>>>
>>>>> /opt/hadoop-0.20.2/build.xml:497: The following error occurred while
>>>>>
>>>>> executing this line:
>>>>>
>>>>> /opt/hadoop-0.20.2/src/contrib/build.xml:30: The following error
>>>>>
>>>>> occurred
>>>>>
>>>>> while executing this line:
>>>>>
>>>>> /opt/hadoop-0.20.2/src/contrib/fuse-dfs/build.xml:57: exec returned: 2
>>>>>
>>>>> The output from the build is attached.
>>>>>
>>>>>
>>>>>
>>>>> We are running OS:  Ubuntu 9.04 x64 (2.6.28-11-server #42-Ubuntu SMP
>>>>> Fri
>>>>>
>>>>> Apr
>>>>>
>>>>> 17 02:45:36 UTC 2009 GNU/Linux)
>>>>>
>>>>> Regards,
>>>>>
>>>>> Seraph
>>>>>
>>>>>
>>>>> On 05 May 2010, at 3:47 PM, Jason Venner wrote:
>>>>>
>>>>> You will need to install the gnu development tool chain for your
>>>>>
>>>>> platform, a quick check on an older redhat system I have suggests the
>>>>>
>>>>> automake and autoconf rpm's will provide aclocal, automake and
>>>>>
>>>>> autoconf.
>>>>>
>>>>> The ./configure: not found error is an artifact of the earlier
>>>>> failures.
>>>>>
>>>>>
>>>>> On Wed, May 5, 2010 at 5:37 AM, Seraph Imalia <ser...@eisp.co.za>
>>>>> wrote:
>>>>>
>>>>> Hi,
>>>>>
>>>>> I have been following the following URL to mount an HDFS using
>>>>>
>>>>> fuse-dfs: http://wiki.apache.org/hadoop/MountableHDFS
>>>>>
>>>>> I have had many problems trying to build it but have managed to get
>>>>>
>>>>> through
>>>>>
>>>>> the first two build commands without build errors.  Running the last
>>>>>
>>>>> build
>>>>>
>>>>> command "ant compile-contrib -Dlibhdfs=1 -Dfusedfs=1" results in the
>>>>>
>>>>> following error:
>>>>>
>>>>> compile:
>>>>>
>>>>>  [echo] contrib: fuse-dfs
>>>>>
>>>>>  [exec] /opt/hadoop-0.20.1/src/contrib/fuse-dfs/bootstrap.sh: 18:
>>>>>
>>>>> aclocal: not found
>>>>>
>>>>>  [exec] /opt/hadoop-0.20.1/src/contrib/fuse-dfs/bootstrap.sh: 19:
>>>>>
>>>>> automake: not found
>>>>>
>>>>>  [exec] /opt/hadoop-0.20.1/src/contrib/fuse-dfs/bootstrap.sh: 20:
>>>>>
>>>>> autoconf: not found
>>>>>
>>>>>  [exec] /opt/hadoop-0.20.1/src/contrib/fuse-dfs/bootstrap.sh: 21:
>>>>>
>>>>> ./configure: not found
>>>>>
>>>>> BUILD FAILED
>>>>>
>>>>> /opt/hadoop-0.20.1/build.xml:497: The following error occurred while
>>>>>
>>>>> executing this line:
>>>>>
>>>>> /opt/hadoop-0.20.1/src/contrib/build.xml:30: The following error
>>>>>
>>>>> occurred
>>>>>
>>>>> while executing this line:
>>>>>
>>>>> /opt/hadoop-0.20.1/src/contrib/fuse-dfs/build.xml:54: exec returned:
>>>>> 127
>>>>>
>>>>> I found this link on
>>>>>
>>>>> google: http://issues.apache.org/jira/browse/HADOOP-4 which does not
>>>>>
>>>>> seem
>>>>>
>>>>> to
>>>>>
>>>>> help me.
>>>>>
>>>>> Please can you assist?
>>>>>
>>>>> Regards,
>>>>>
>>>>> Seraph
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>>
>>>>> Pro Hadoop, a book to guide you from beginner to hadoop mastery,
>>>>>
>>>>> http://www.amazon.com/dp/1430219424?tag=jewlerymall
>>>>>
>>>>> www.prohadoopbook.com a community for Hadoop Professionals
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>>
>>>>> Pro Hadoop, a book to guide you from beginner to hadoop mastery,
>>>>>
>>>>> http://www.amazon.com/dp/1430219424?tag=jewlerymall
>>>>>
>>>>> www.prohadoopbook.com a community for Hadoop Professionals
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>
>>>
>>>
>
>
>

Reply via email to