[ 
http://issues.apache.org/jira/browse/HADOOP-483?page=comments#action_12431328 ] 
            
Arun C Murthy commented on HADOOP-483:
--------------------------------------

 This is due to the fact that libhdfs relies on the hadoop setup to be up and 
running (and $CLASSPATH to be setup correctly to the running instance), which 
isn't the case after 'ant clean'. Should I setup the 'compile-libhdfs' target 
in build.xml to depend on 'package' too? (However that still won't solve the 
problem that hadoop isn't running when we need to run hdfs_test)

  All this while we relied on this to be done explicitly and hence the comment 
in build.xml which notes that test-libhdfs should be run manually.

  I'd appreciate any ideas on how to improve this further... including how I 
can go about ensuring that hadoop is running, setup correct classpaths (for 
jni) etc.

thanks,
Arun


> Document libhdfs and fix some OS specific stuff in Makefile
> -----------------------------------------------------------
>
>                 Key: HADOOP-483
>                 URL: http://issues.apache.org/jira/browse/HADOOP-483
>             Project: Hadoop
>          Issue Type: Improvement
>          Components: dfs
>            Reporter: Arun C Murthy
>         Assigned To: Arun C Murthy
>         Attachments: HADOOP-483.patch, libhdfs.patch, libhdfs.second.patch
>
>
> Document libhdfs (api docs and user manual) and fix some OS specific stuff 
> (paths to utilities like rm) in the Makefile and TestDFSCIO.java

-- 
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators: 
http://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Reply via email to