Hi Deveraj,

I went through multiple links all asking me to check if the
mapreduce.framework.name is set to yarn, it is along with proper pointing
to the Namenode i.e. fs.defaultFS.

But still it tries to connect to the local. I am not sure what to do,
please help me out with some pointers as I am fairly new to the coding
aspect of map-reduce.

Thanks
Divye Sheth


On Tue, Apr 8, 2014 at 7:43 PM, divye sheth <divs.sh...@gmail.com> wrote:

> Hi,
>
> I saw that pretty much after sending the email. I verified the properties
> file and it has all the correct properties even the mapred.framework.nameis 
> set to yarn. I am unable to figure out what is the cause and why it is
> connecting to local FS.
>
> Using the same configuration file I am able to run my WordCount MRV1
> example but not the code that I have written for a usecase.
>
> Thanks
> Divye Sheth
>
>
> On Tue, Apr 8, 2014 at 6:12 PM, Kavali, Devaraj 
> <devaraj.kav...@intel.com>wrote:
>
>>  As per the given exception stack trace, it is trying to use local file
>> system. Can you check whether you have configured the file system
>> configurations with HDFS?
>>
>>
>>
>> Thanks
>>
>> Devaraj K
>>
>>
>>
>> *From:* divye sheth [mailto:divs.sh...@gmail.com]
>> *Sent:* Tuesday, April 08, 2014 5:37 PM
>> *To:* user@hadoop.apache.org
>> *Subject:* Running MRV1 code on YARN
>>
>>
>>
>> Hi,
>>
>>
>>
>> I have installed Hadoop 2.2.0 along with YARN and am trying to submit a
>> MRV1 job already written to YARN.
>>
>>
>>
>> The job does not even submit and it prints the following stack trace on
>> console:
>>
>>
>>
>> 2014-04-08 16:56:11 UserGroupInformation [ERROR]
>> PriviledgedActionException as:eureka (auth:SIMPLE)
>> cause:org.apache.hadoop.util.Shell$ExitCodeException: chmod: cannot access
>> `/user/eureka54695942/.staging/job_local54695942_0001': No such file or
>> directory
>>
>>
>>
>> Exception in thread "main"
>> org.apache.hadoop.util.Shell$ExitCodeException: chmod: cannot access
>> `/user/eureka54695942/.staging/job_local54695942_0001': No such file or
>> directory
>>
>>
>>
>>         at org.apache.hadoop.util.Shell.runCommand(Shell.java:261)
>>
>>         at org.apache.hadoop.util.Shell.run(Shell.java:188)
>>
>>         at
>> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:381)
>>
>>         at org.apache.hadoop.util.Shell.execCommand(Shell.java:467)
>>
>>         at org.apache.hadoop.util.Shell.execCommand(Shell.java:450)
>>
>>         at
>> org.apache.hadoop.fs.RawLocalFileSystem.execCommand(RawLocalFileSystem.java:593)
>>
>>         at
>> org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:584)
>>
>>         at
>> org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.java:427)
>>
>>         at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:579)
>>
>>         at
>> org.apache.hadoop.mapred.JobClient.copyAndConfigureFiles(JobClient.java:786)
>>
>>         at
>> org.apache.hadoop.mapred.JobClient.copyAndConfigureFiles(JobClient.java:746)
>>
>>         at
>> org.apache.hadoop.mapred.JobClient.access$400(JobClient.java:177)
>>
>>         at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:963)
>>
>>         at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:948)
>>
>>         at java.security.AccessController.doPrivileged(Native Method)
>>
>>         at javax.security.auth.Subject.doAs(Subject.java:396)
>>
>>         at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
>>
>>         at
>> org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:948)
>>
>>         at org.apache.hadoop.mapreduce.Job.submit(Job.java:566)
>>
>>
>>
>> My question here is if you notice the staging location which it is trying
>> to clean I do not have any such user in the /user directory in hdfs. It
>> somehow appends the jobId to the username and creates staging area there.
>> Any reason for this? Please let me know what am I doing wrong. How can I
>> make sure it goes to the user that I have created i.e. eureka and not
>> eureka$JOBID.
>>
>>
>>
>> I am using CDH4.
>>
>>
>>
>> Thanks
>>
>> Divye Sheth
>>
>
>

Reply via email to