Hi all,
I'm trying to install a plugin called gfarm_hadoop that allows me to use a
filesystem called gfarm instead of HDFS (
https://sourceforge.net/projects/gfarm/files/gfarm_hadoop/).
I have used it before, but now I'm trying to install it in a new cluster
and for some reason it isn't working...
After installing gfarm 2.5.8 at /data/local3/marilia/gfarm, hadoop 1.1.2 at
/data/local3/marilia/hadoop-1.1.2 and the plugin, when I try to list the
new filesystem it works fine:
$ bin/hadoop fs -ls gfarm:///
Found 26 items
-rwxrwxrwx 1 101 2013-06-26 02:36 /foo
drwxrwxrwx - 0 2013-06-26 02:43 /home
But then when I try to run an example, the task eventually completes, but I
get " Unable to load libGfarmFSNative library" errors. Looking at the logs
message it seems to be a path problem, but I have tried almost everything
and it doesn't work.
The way I'm setting the path now is writing on conf/hadoop-env.sh the
following line:
export LD_LIBRARY_PATH=/data/local3/marilia/gfarm/lib
I have even moved all the .so files to the hadoop directory, but I still
get the same message...
Any ideas?
Thanks in advance.
Log:
$ bin/hadoop jar hadoop-examples-*.jar teragen 1000 gfarm:///inoa11
Generating 1000 using 2 maps with step of 500
13/06/27 03:57:32 INFO mapred.JobClient: Running job: job_201306270356_0001
13/06/27 03:57:33 INFO mapred.JobClient: map 0% reduce 0%
13/06/27 03:57:38 INFO mapred.JobClient: map 50% reduce 0%
13/06/27 03:57:43 INFO mapred.JobClient: Task Id :
attempt_201306270356_0001_m_000001_0, Status : FAILED
java.lang.Throwable: Child Error
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
Caused by: java.io.IOException: Task process exit with nonzero status of 1.
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
java.lang.Throwable: Child Error
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
Caused by: java.io.IOException: Task process exit with nonzero status of 1.
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
attempt_201306270356_0001_m_000001_0: java.lang.UnsatisfiedLinkError:
/data/local3/marilia/hadoop-1.1.2/lib/native/Linux-amd64-64/libGfarmFSNative.so:
libgfarm.so.1: cannot open shared object file: No such file or directory
attempt_201306270356_0001_m_000001_0: at
java.lang.ClassLoader$NativeLibrary.load(Native Method)
attempt_201306270356_0001_m_000001_0: at
java.lang.ClassLoader.loadLibrary0(ClassLoader.java:1803)
attempt_201306270356_0001_m_000001_0: at
java.lang.ClassLoader.loadLibrary(ClassLoader.java:1728)
attempt_201306270356_0001_m_000001_0: at
java.lang.Runtime.loadLibrary0(Runtime.java:823)
attempt_201306270356_0001_m_000001_0: at
java.lang.System.loadLibrary(System.java:1028)
attempt_201306270356_0001_m_000001_0: at
org.apache.hadoop.fs.gfarmfs.GfarmFSNative.<clinit>(GfarmFSNative.java:9)
attempt_201306270356_0001_m_000001_0: at
org.apache.hadoop.fs.gfarmfs.GfarmFileSystem.initialize(GfarmFileSystem.java:34)
attempt_201306270356_0001_m_000001_0: at
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1411)
attempt_201306270356_0001_m_000001_0: at
org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
attempt_201306270356_0001_m_000001_0: at
org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1429)
attempt_201306270356_0001_m_000001_0: at
org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254)
attempt_201306270356_0001_m_000001_0: at
org.apache.hadoop.fs.Path.getFileSystem(Path.java:187)
attempt_201306270356_0001_m_000001_0: at
org.apache.hadoop.mapred.FileOutputCommitter.getTempTaskOutputPath(FileOutputCommitter.java:234)
attempt_201306270356_0001_m_000001_0: at
org.apache.hadoop.mapred.Task.initialize(Task.java:522)
attempt_201306270356_0001_m_000001_0: at
org.apache.hadoop.mapred.MapTask.run(MapTask.java:353)
attempt_201306270356_0001_m_000001_0: at
org.apache.hadoop.mapred.Child$4.run(Child.java:255)
attempt_201306270356_0001_m_000001_0: at
java.security.AccessController.doPrivileged(Native Method)
attempt_201306270356_0001_m_000001_0: at
javax.security.auth.Subject.doAs(Subject.java:396)
attempt_201306270356_0001_m_000001_0: at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1149)
attempt_201306270356_0001_m_000001_0: at
org.apache.hadoop.mapred.Child.main(Child.java:249)
attempt_201306270356_0001_m_000001_0: Unable to load libGfarmFSNative
library
13/06/27 03:57:46 INFO mapred.JobClient: map 100% reduce 0%
13/06/27 03:57:47 INFO mapred.JobClient: Job complete: job_201306270356_0001
13/06/27 03:57:47 INFO mapred.JobClient: Counters: 18
13/06/27 03:57:47 INFO mapred.JobClient: Job Counters
13/06/27 03:57:47 INFO mapred.JobClient: SLOTS_MILLIS_MAPS=10846
13/06/27 03:57:47 INFO mapred.JobClient: Total time spent by all
reduces waiting after reserving slots (ms)=0
13/06/27 03:57:47 INFO mapred.JobClient: Total time spent by all maps
waiting after reserving slots (ms)=0
13/06/27 03:57:47 INFO mapred.JobClient: Launched map tasks=3
13/06/27 03:57:47 INFO mapred.JobClient: SLOTS_MILLIS_REDUCES=0
13/06/27 03:57:47 INFO mapred.JobClient: File Input Format Counters
13/06/27 03:57:47 INFO mapred.JobClient: Bytes Read=0
13/06/27 03:57:47 INFO mapred.JobClient: File Output Format Counters
13/06/27 03:57:47 INFO mapred.JobClient: Bytes Written=0
13/06/27 03:57:47 INFO mapred.JobClient: FileSystemCounters
13/06/27 03:57:47 INFO mapred.JobClient: HDFS_BYTES_READ=164
13/06/27 03:57:47 INFO mapred.JobClient: FILE_BYTES_WRITTEN=104248
13/06/27 03:57:47 INFO mapred.JobClient: Map-Reduce Framework
13/06/27 03:57:47 INFO mapred.JobClient: Map input records=1000
13/06/27 03:57:47 INFO mapred.JobClient: Physical memory (bytes)
snapshot=207736832
13/06/27 03:57:47 INFO mapred.JobClient: Spilled Records=0
13/06/27 03:57:47 INFO mapred.JobClient: CPU time spent (ms)=190
13/06/27 03:57:47 INFO mapred.JobClient: Total committed heap usage
(bytes)=401997824
13/06/27 03:57:47 INFO mapred.JobClient: Virtual memory (bytes)
snapshot=1104424960
13/06/27 03:57:47 INFO mapred.JobClient: Map input bytes=1000
13/06/27 03:57:47 INFO mapred.JobClient: Map output records=1000
13/06/27 03:57:47 INFO mapred.JobClient: SPLIT_RAW_BYTES=164
--
Marilia Melo