[ http://issues.apache.org/jira/browse/HADOOP-459?page=all ]

Doug Cutting updated HADOOP-459:
--------------------------------

    Status: Open  (was: Patch Available)

After applying this, 'ant clean test-libhdfs' fails with:

compile-libhdfs:
    [mkdir] Created dir: /home/cutting/src/hadoop/trunk/build/libhdfs
     [exec] gcc -g -Wall -O2 -fPIC -m32 -I/home/cutting/local/jdk/include 
-I/home/cutting/local/jdk/include/linux -c hdfs.c -o 
/home/cutting/src/hadoop/trunk/build/libhdfs/hdfs.o
     [exec] hdfs.c: In function 'hdfsGetWorkingDirectory':
     [exec] hdfs.c:1048: warning: pointer targets in initialization differ in 
signedness
     [exec] hdfs.c:1054: warning: pointer targets in passing argument 3 of 
'(*env)->ReleaseStringUTFChars' differ in signedness
     [exec] hdfs.c: In function 'hdfsGetHosts':
     [exec] hdfs.c:1212: warning: pointer targets in assignment differ in 
signedness
     [exec] hdfs.c:1215: warning: pointer targets in passing argument 3 of 
'(*env)->ReleaseStringUTFChars' differ in signedness
     [exec] hdfs.c: In function 'getFileInfo':
     [exec] hdfs.c:1412: warning: pointer targets in initialization differ in 
signedness
     [exec] hdfs.c:1416: warning: pointer targets in passing argument 3 of 
'(*env)->ReleaseStringUTFChars' differ in signedness
     [exec] make: *** No rule to make target 
`/home/cutting/src/hadoop/trunk/build/libhdfs/hdfsJniHelper.o', needed by 
`/home/cutting/src/hadoop/trunk/build/libhdfs/libhdfs.so.1'.  Stop.
     [exec] Result: 2

That said, before applying it, 'ant clean test-libhdfs' fails with:

test-libhdfs:
   [delete] Deleting directory 
/home/cutting/src/hadoop/trunk/build/libhdfs/tests/logs
    [mkdir] Created dir: 
/home/cutting/src/hadoop/trunk/build/libhdfs/tests/logs     [exec] 
./tests/test-libhdfs.sh
     [exec] 06/11/03 11:20:33 INFO conf.Configuration: parsing 
file:/home/cutting/src/hadoop/trunk/conf/hadoop-default.xml
     [exec] starting namenode, logging to 
/home/cutting/src/hadoop/trunk/build/libhdfs/tests/logs/hadoop-cutting-namenode-cutting-dc5100.out
     [exec] starting datanode, logging to 
/home/cutting/src/hadoop/trunk/build/libhdfs/tests/logs/hadoop-cutting-datanode-cutting-dc5100.out
     [exec] 
CLASSPATH=/home/cutting/src/hadoop/trunk/src/c++/libhdfs/tests/conf:/home/cutting/src/hadoop/trunk/conf:/home/cutting/src/hadoop/trunk/build/hadoop-0.7.3-dev.jar:/home/cutting/src/hadoop/trunk/lib/commons-cli-2.0-SNAPSHOT.jar:/home/cutting/src/hadoop/trunk/lib/commons-logging-1.0.4.jar:/home/cutting/src/hadoop/trunk/lib/commons-logging-api-1.0.4.jar:/home/cutting/src/hadoop/trunk/lib/jetty-5.1.4.jar:/home/cutting/src/hadoop/trunk/lib/junit-3.8.1.jar:/home/cutting/src/hadoop/trunk/lib/log4j-1.2.13.jar:/home/cutting/src/hadoop/trunk/lib/lucene-core-1.9.1.jar:/home/cutting/src/hadoop/trunk/lib/servlet-api.jar
 LD_PRELOAD=/home/cutting/src/hadoop/trunk/build/libhdfs/libhdfs.so 
/home/cutting/src/hadoop/trunk/build/libhdfs/hdfs_test
     [exec] 06/11/03 11:20:40 INFO conf.Configuration: parsing 
file:/home/cutting/src/hadoop/trunk/conf/hadoop-default.xml
     [exec] 06/11/03 11:20:40 INFO ipc.Client: 
org.apache.hadoop.io.ObjectWritableConnection culler maxidletime= 1000ms
     [exec] 06/11/03 11:20:40 INFO ipc.Client: 
org.apache.hadoop.io.ObjectWritable Connection Culler: starting
     [exec] 06/11/03 11:20:40 INFO ipc.Client: Retrying connect to server: 
localhost/127.0.0.1:23000. Already tried 1 time(s).
     [exec] 06/11/03 11:20:41 INFO ipc.Client: Retrying connect to server: 
localhost/127.0.0.1:23000. Already tried 2 time(s).
     [exec] 06/11/03 11:20:42 INFO ipc.Client: Retrying connect to server: 
localhost/127.0.0.1:23000. Already tried 3 time(s).
     [exec] 06/11/03 11:20:43 INFO ipc.Client: Retrying connect to server: 
localhost/127.0.0.1:23000. Already tried 4 time(s).
     [exec] 06/11/03 11:20:44 INFO ipc.Client: Retrying connect to server: 
localhost/127.0.0.1:23000. Already tried 5 time(s).
     [exec] 06/11/03 11:20:45 INFO ipc.Client: Retrying connect to server: 
localhost/127.0.0.1:23000. Already tried 6 time(s).
     [exec] 06/11/03 11:20:46 INFO ipc.Client: Retrying connect to server: 
localhost/127.0.0.1:23000. Already tried 7 time(s).
     [exec] 06/11/03 11:20:47 INFO ipc.Client: Retrying connect to server: 
localhost/127.0.0.1:23000. Already tried 8 time(s).
     [exec] 06/11/03 11:20:48 INFO ipc.Client: Retrying connect to server: 
localhost/127.0.0.1:23000. Already tried 9 time(s).
     [exec] 06/11/03 11:20:49 INFO ipc.Client: Retrying connect to server: 
localhost/127.0.0.1:23000. Already tried 10 time(s).
     [exec] Exception in thread "main" java.net.ConnectException: Connection 
refused
     [exec]     at java.net.PlainSocketImpl.socketConnect(Native Method)
     [exec]     at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:333)
     [exec]     at 
java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:195)
     [exec]     at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:182)
     [exec]     at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366)
     [exec]     at java.net.Socket.connect(Socket.java:516)
     [exec]     at 
org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:138)
     [exec]     at org.apache.hadoop.ipc.Client.getConnection(Client.java:517)
     [exec]     at org.apache.hadoop.ipc.Client.call(Client.java:445)
     [exec]     at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:164)
     [exec]     at org.apache.hadoop.dfs.$Proxy0.getProtocolVersion(Unknown 
Source)
     [exec]     at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:248)
     [exec]     at org.apache.hadoop.dfs.DFSClient.<init>(DFSClient.java:105)
     [exec]     at 
org.apache.hadoop.dfs.DistributedFileSystem.<init>(DistributedFileSystem.java:49)
     [exec]     at 
org.apache.hadoop.fs.FileSystem.getNamed(FileSystem.java:104)     [exec]     at 
org.apache.hadoop.fs.FileSystem.get(FileSystem.java:89)
     [exec] Call to org.apache.hadoop.fs.FileSystem::get failed!
     [exec] 06/11/03 11:20:50 INFO conf.Configuration: parsing 
file:/home/cutting/src/hadoop/trunk/conf/hadoop-default.xml
     [exec] ./tests/test-libhdfs.sh: line 45: 26791 Aborted                 
CLASSPATH=$HADOOP_CONF_DIR:$CLASSPATH 
LD_PRELOAD="$LIBHDFS_BUILD_DIR/libhdfs.so" $LIBHDFS_BUILD_DIR/$HDFS_TEST
     [exec] #
     [exec] # An unexpected error has been detected by HotSpot Virtual Machine:
     [exec] #
     [exec] #  SIGSEGV (0xb) at pc=0x402ff1c3, pid=26791, tid=1083387680
     [exec] #
     [exec] # Java VM: Java HotSpot(TM) Client VM (1.5.0_08-b03 mixed mode, 
sharing)
     [exec] # Problematic frame:
     [exec] # V  [libjvm.so+0x1a51c3]
     [exec] #
     [exec] # An error report file with more information is saved as 
hs_err_pid26791.log
     [exec] #
     [exec] # If you would like to submit a bug report, please visit:
     [exec] #   http://java.sun.com/webapps/bugreport/crash.jsp
     [exec] #
     [exec] stopping datanode
     [exec] no namenode to stop

BUILD SUCCESSFUL

Which is a separate issue...

> libhdfs leaks memory when writing to files
> ------------------------------------------
>
>                 Key: HADOOP-459
>                 URL: http://issues.apache.org/jira/browse/HADOOP-459
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.7.1, 0.7.0, 0.6.2, 0.6.1, 0.6.0, 0.5.0
>            Reporter: Christian Kunz
>         Assigned To: Sameer Paranjpye
>         Attachments: libhdfs_fixes.txt
>
>
> hdfsWrite leaks memory when called repeatedly. The same probably applies to 
> repeated reads using hdfsRead

-- 
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators: 
http://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Reply via email to