On Mar 19, 2013, at 8:10 AM, Varun Chandramouli <[email protected]> wrote:

> Hi Noah,
> 
> After removing that and running, I get this error:
> 
> varunc@varunc4-virtual-machine:~$ time /usr/local/hadoop/bin/hadoop jar 
> hadoop*examples*.jar wordcount /mnt/ceph/wc /mnt/ceph/op2
> Warning: $HADOOP_HOME is deprecated.

Also, is your data inside Hadoop located at /mnt/ceph/wc? Try running hadoop fs 
-ls /mnt/ceph/wc. Remember, Hadoop cannot talk to the VFS mount, so 
/mnt/ceph/op2 won't work. It needs to be within a Ceph URI mount.

-Noah


> 
> 13/03/19 20:35:03 INFO ceph.CephFileSystem: selectDataPool 
> path=ceph://null/app/hadoop/tmp/mapred/staging/varunc/.staging/job_201303192034_0001/job.jar
>  pool:repl=data:2 wanted=3
> 13/03/19 20:35:03 INFO mapred.JobClient: Cleaning up the staging area 
> ceph://null/app/hadoop/tmp/mapred/staging/varunc/.staging/job_201303192034_0001
> 13/03/19 20:35:03 ERROR security.UserGroupInformation: 
> PriviledgedActionException as:varunc 
> cause:org.apache.hadoop.mapreduce.lib.input.InvalidInputException: Input path 
> does not exist: ceph://null/mnt/ceph/wc
> org.apache.hadoop.mapreduce.lib.input.InvalidInputException: Input path does 
> not exist: ceph://null/mnt/ceph/wc
>         at 
> org.apache.hadoop.mapreduce.lib.input.FileInputFormat.listStatus(FileInputFormat.java:235)
>         at 
> org.apache.hadoop.mapreduce.lib.input.FileInputFormat.getSplits(FileInputFormat.java:252)
>         at 
> org.apache.hadoop.mapred.JobClient.writeNewSplits(JobClient.java:962)
>         at org.apache.hadoop.mapred.JobClient.writeSplits(JobClient.java:979)
>         at org.apache.hadoop.mapred.JobClient.access$600(JobClient.java:174)
>         at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:897)
>         at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:850)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:415)
>         at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>         at 
> org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:850)
>         at org.apache.hadoop.mapreduce.Job.submit(Job.java:500)
>         at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:530)
>         at org.apache.hadoop.examples.WordCount.main(WordCount.java:67)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:601)
>         at 
> org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:68)
>         at org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:139)
>         at 
> org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:64)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:601)
>         at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
> 
> Any ways to remove this error?
> 
> On Tue, Mar 19, 2013 at 7:39 PM, Noah Watkins <[email protected]> wrote:
> Hi Varun,
> 
> Try removing this configuration option:
> 
> > <property>
> >   <name>ceph.root.dir</name>
> >   <value>/mnt/ceph</value>
> > </property>
> 
> Hadoop running on Ceph uses the libcephfs user-space library to talk directly 
> to the file system, as opposed to running through the kernel or FUSE client. 
> This setting is which directory within the Ceph file system to use as a root. 
> It defaults to "/".
> 
> Thanks,
> -Noah
> 
> 
> Thanks
> Varun 

_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to