[ http://issues.apache.org/jira/browse/HADOOP-459?page=all ]
Doug Cutting updated HADOOP-459:
--------------------------------
Status: Open (was: Patch Available)
After applying this, 'ant clean test-libhdfs' fails with:
compile-libhdfs:
[mkdir] Created dir: /home/cutting/src/hadoop/trunk/build/libhdfs
[exec] gcc -g -Wall -O2 -fPIC -m32 -I/home/cutting/local/jdk/include
-I/home/cutting/local/jdk/include/linux -c hdfs.c -o
/home/cutting/src/hadoop/trunk/build/libhdfs/hdfs.o
[exec] hdfs.c: In function 'hdfsGetWorkingDirectory':
[exec] hdfs.c:1048: warning: pointer targets in initialization differ in
signedness
[exec] hdfs.c:1054: warning: pointer targets in passing argument 3 of
'(*env)->ReleaseStringUTFChars' differ in signedness
[exec] hdfs.c: In function 'hdfsGetHosts':
[exec] hdfs.c:1212: warning: pointer targets in assignment differ in
signedness
[exec] hdfs.c:1215: warning: pointer targets in passing argument 3 of
'(*env)->ReleaseStringUTFChars' differ in signedness
[exec] hdfs.c: In function 'getFileInfo':
[exec] hdfs.c:1412: warning: pointer targets in initialization differ in
signedness
[exec] hdfs.c:1416: warning: pointer targets in passing argument 3 of
'(*env)->ReleaseStringUTFChars' differ in signedness
[exec] make: *** No rule to make target
`/home/cutting/src/hadoop/trunk/build/libhdfs/hdfsJniHelper.o', needed by
`/home/cutting/src/hadoop/trunk/build/libhdfs/libhdfs.so.1'. Stop.
[exec] Result: 2
That said, before applying it, 'ant clean test-libhdfs' fails with:
test-libhdfs:
[delete] Deleting directory
/home/cutting/src/hadoop/trunk/build/libhdfs/tests/logs
[mkdir] Created dir:
/home/cutting/src/hadoop/trunk/build/libhdfs/tests/logs [exec]
./tests/test-libhdfs.sh
[exec] 06/11/03 11:20:33 INFO conf.Configuration: parsing
file:/home/cutting/src/hadoop/trunk/conf/hadoop-default.xml
[exec] starting namenode, logging to
/home/cutting/src/hadoop/trunk/build/libhdfs/tests/logs/hadoop-cutting-namenode-cutting-dc5100.out
[exec] starting datanode, logging to
/home/cutting/src/hadoop/trunk/build/libhdfs/tests/logs/hadoop-cutting-datanode-cutting-dc5100.out
[exec]
CLASSPATH=/home/cutting/src/hadoop/trunk/src/c++/libhdfs/tests/conf:/home/cutting/src/hadoop/trunk/conf:/home/cutting/src/hadoop/trunk/build/hadoop-0.7.3-dev.jar:/home/cutting/src/hadoop/trunk/lib/commons-cli-2.0-SNAPSHOT.jar:/home/cutting/src/hadoop/trunk/lib/commons-logging-1.0.4.jar:/home/cutting/src/hadoop/trunk/lib/commons-logging-api-1.0.4.jar:/home/cutting/src/hadoop/trunk/lib/jetty-5.1.4.jar:/home/cutting/src/hadoop/trunk/lib/junit-3.8.1.jar:/home/cutting/src/hadoop/trunk/lib/log4j-1.2.13.jar:/home/cutting/src/hadoop/trunk/lib/lucene-core-1.9.1.jar:/home/cutting/src/hadoop/trunk/lib/servlet-api.jar
LD_PRELOAD=/home/cutting/src/hadoop/trunk/build/libhdfs/libhdfs.so
/home/cutting/src/hadoop/trunk/build/libhdfs/hdfs_test
[exec] 06/11/03 11:20:40 INFO conf.Configuration: parsing
file:/home/cutting/src/hadoop/trunk/conf/hadoop-default.xml
[exec] 06/11/03 11:20:40 INFO ipc.Client:
org.apache.hadoop.io.ObjectWritableConnection culler maxidletime= 1000ms
[exec] 06/11/03 11:20:40 INFO ipc.Client:
org.apache.hadoop.io.ObjectWritable Connection Culler: starting
[exec] 06/11/03 11:20:40 INFO ipc.Client: Retrying connect to server:
localhost/127.0.0.1:23000. Already tried 1 time(s).
[exec] 06/11/03 11:20:41 INFO ipc.Client: Retrying connect to server:
localhost/127.0.0.1:23000. Already tried 2 time(s).
[exec] 06/11/03 11:20:42 INFO ipc.Client: Retrying connect to server:
localhost/127.0.0.1:23000. Already tried 3 time(s).
[exec] 06/11/03 11:20:43 INFO ipc.Client: Retrying connect to server:
localhost/127.0.0.1:23000. Already tried 4 time(s).
[exec] 06/11/03 11:20:44 INFO ipc.Client: Retrying connect to server:
localhost/127.0.0.1:23000. Already tried 5 time(s).
[exec] 06/11/03 11:20:45 INFO ipc.Client: Retrying connect to server:
localhost/127.0.0.1:23000. Already tried 6 time(s).
[exec] 06/11/03 11:20:46 INFO ipc.Client: Retrying connect to server:
localhost/127.0.0.1:23000. Already tried 7 time(s).
[exec] 06/11/03 11:20:47 INFO ipc.Client: Retrying connect to server:
localhost/127.0.0.1:23000. Already tried 8 time(s).
[exec] 06/11/03 11:20:48 INFO ipc.Client: Retrying connect to server:
localhost/127.0.0.1:23000. Already tried 9 time(s).
[exec] 06/11/03 11:20:49 INFO ipc.Client: Retrying connect to server:
localhost/127.0.0.1:23000. Already tried 10 time(s).
[exec] Exception in thread "main" java.net.ConnectException: Connection
refused
[exec] at java.net.PlainSocketImpl.socketConnect(Native Method)
[exec] at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:333)
[exec] at
java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:195)
[exec] at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:182)
[exec] at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366)
[exec] at java.net.Socket.connect(Socket.java:516)
[exec] at
org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:138)
[exec] at org.apache.hadoop.ipc.Client.getConnection(Client.java:517)
[exec] at org.apache.hadoop.ipc.Client.call(Client.java:445)
[exec] at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:164)
[exec] at org.apache.hadoop.dfs.$Proxy0.getProtocolVersion(Unknown
Source)
[exec] at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:248)
[exec] at org.apache.hadoop.dfs.DFSClient.<init>(DFSClient.java:105)
[exec] at
org.apache.hadoop.dfs.DistributedFileSystem.<init>(DistributedFileSystem.java:49)
[exec] at
org.apache.hadoop.fs.FileSystem.getNamed(FileSystem.java:104) [exec] at
org.apache.hadoop.fs.FileSystem.get(FileSystem.java:89)
[exec] Call to org.apache.hadoop.fs.FileSystem::get failed!
[exec] 06/11/03 11:20:50 INFO conf.Configuration: parsing
file:/home/cutting/src/hadoop/trunk/conf/hadoop-default.xml
[exec] ./tests/test-libhdfs.sh: line 45: 26791 Aborted
CLASSPATH=$HADOOP_CONF_DIR:$CLASSPATH
LD_PRELOAD="$LIBHDFS_BUILD_DIR/libhdfs.so" $LIBHDFS_BUILD_DIR/$HDFS_TEST
[exec] #
[exec] # An unexpected error has been detected by HotSpot Virtual Machine:
[exec] #
[exec] # SIGSEGV (0xb) at pc=0x402ff1c3, pid=26791, tid=1083387680
[exec] #
[exec] # Java VM: Java HotSpot(TM) Client VM (1.5.0_08-b03 mixed mode,
sharing)
[exec] # Problematic frame:
[exec] # V [libjvm.so+0x1a51c3]
[exec] #
[exec] # An error report file with more information is saved as
hs_err_pid26791.log
[exec] #
[exec] # If you would like to submit a bug report, please visit:
[exec] # http://java.sun.com/webapps/bugreport/crash.jsp
[exec] #
[exec] stopping datanode
[exec] no namenode to stop
BUILD SUCCESSFUL
Which is a separate issue...
> libhdfs leaks memory when writing to files
> ------------------------------------------
>
> Key: HADOOP-459
> URL: http://issues.apache.org/jira/browse/HADOOP-459
> Project: Hadoop
> Issue Type: Bug
> Components: dfs
> Affects Versions: 0.7.1, 0.7.0, 0.6.2, 0.6.1, 0.6.0, 0.5.0
> Reporter: Christian Kunz
> Assigned To: Sameer Paranjpye
> Attachments: libhdfs_fixes.txt
>
>
> hdfsWrite leaks memory when called repeatedly. The same probably applies to
> repeated reads using hdfsRead
--
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators:
http://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see: http://www.atlassian.com/software/jira