[jira] [Commented] (MAPREDUCE-5663) Add an interface to Input/Ouput Formats to obtain delegation tokens

2015-05-08 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5663?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14534851#comment-14534851
 ] 

Roman Shaposhnik commented on MAPREDUCE-5663:
-

[~sseth] I'm looking at it as part of bugbash (and also b/c krb is something 
I'm pretty familiar with). Given the discussion so far, I'm not quite sure 
where this stands. The last comment from [~tucu00] makes a lot of sense to me, 
but obviously I'm looking at this for the first time.

> Add an interface to Input/Ouput Formats to obtain delegation tokens
> ---
>
> Key: MAPREDUCE-5663
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5663
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>Reporter: Siddharth Seth
>Assignee: Michael Weng
>  Labels: BB2015-05-TBR
> Attachments: MAPREDUCE-5663.4.txt, MAPREDUCE-5663.5.txt, 
> MAPREDUCE-5663.6.txt, MAPREDUCE-5663.patch.txt, MAPREDUCE-5663.patch.txt2, 
> MAPREDUCE-5663.patch.txt3
>
>
> Currently, delegation tokens are obtained as part of the getSplits / 
> checkOutputSpecs calls to the InputFormat / OutputFormat respectively.
> This works as long as the splits are generated on a node with kerberos 
> credentials. For split generation elsewhere (AM for example), an explicit 
> interface is required.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-3202) Integrating Hadoop Vaidya with Job History UI in Hadoop 2.0

2014-01-03 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3202?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13861748#comment-13861748
 ] 

Roman Shaposhnik commented on MAPREDUCE-3202:
-

[~vitthal_gogate] I did the first pass over the patch and it looks reasonable. 
Here's a few bits of feedback I've got so far:
   * I would like to suggest that an extra JIRA (perhaps a subjira of this one) 
gets created to track *just* the changes required in existing HS webUI to 
support Vaidya. One half of the patch will go there.
   * please make sure to add unit tests for that first half of the patch
   * on this JIRA we shall keep track of things under hadoop-tools/hadoop-vaidya
   * we would also need to make sure that hadoop-tools/hadoop-vaidya bits end 
up in maven assembly

> Integrating Hadoop Vaidya with Job History UI in Hadoop 2.0 
> 
>
> Key: MAPREDUCE-3202
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-3202
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>  Components: jobhistoryserver
>Affects Versions: 2.0.0-alpha
>Reporter: vitthal (Suhas) Gogate
>Assignee: Evan (Yiming) Meng
> Fix For: trunk
>
> Attachments: MAPREDUCE-3202.patch, MAPREDUCE-3202.patch
>
>
> Hadoop Vaidya provides a detailed analysis of the M/R job in terms of various 
> execution inefficiencies and the associated remedies that user can easily 
> understand and fix. This Jira patch integrates it with Job History UI under 
> Hadoop 2.0 branch.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (MAPREDUCE-5362) clean up POM dependencies

2013-08-27 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13751399#comment-13751399
 ] 

Roman Shaposhnik commented on MAPREDUCE-5362:
-

I'm cleaning up a few build issues (Bigtop 0.8.0 related) and I was wondering 
whether you'd think it would be a good idea for me to tackle this one as well. 
Please assign it to me if you think it is.

> clean up POM dependencies
> -
>
> Key: MAPREDUCE-5362
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5362
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: build
>Affects Versions: 2.1.0-beta
>Reporter: Alejandro Abdelnur
>
> Intermediate 'pom' modules define dependencies inherited by leaf modules.
> This is causing issues in intellij IDE.
> We should normalize the leaf modules like in common, hdfs and tools where all 
> dependencies are defined in each leaf module and the intermediate 'pom' 
> module do not define any dependency.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5240) inside of FileOutputCommitter the initialized Credentials cache appears to be empty

2013-05-16 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5240?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-5240:


Status: Patch Available  (was: Reopened)

> inside of FileOutputCommitter the initialized Credentials cache appears to be 
> empty
> ---
>
> Key: MAPREDUCE-5240
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5240
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mrv2
>Affects Versions: 2.0.4-alpha
>Reporter: Roman Shaposhnik
>Assignee: Vinod Kumar Vavilapalli
>Priority: Blocker
>  Labels: 2.0.4.1
> Fix For: 2.0.5-beta, 2.0.4.1-alpha
>
> Attachments: LostCreds.java, MAPREDUCE-5240-20130512.txt, 
> MAPREDUCE-5240-20130513.txt, MAPREDUCE-5240.2.0.4.rvs.patch.txt
>
>
> I am attaching a modified wordcount job that clearly demonstrates the problem 
> we've encountered in running Sqoop2 on YARN (BIGTOP-949).
> Here's what running it produces:
> {noformat}
> $ hadoop fs -mkdir in
> $ hadoop fs -put /etc/passwd in
> $ hadoop jar ./bug.jar org.myorg.LostCreds
> 13/05/12 03:13:46 WARN mapred.JobConf: The variable mapred.child.ulimit is no 
> longer used.
> numberOfSecretKeys: 1
> numberOfTokens: 0
> ..
> ..
> ..
> 13/05/12 03:05:35 INFO mapreduce.Job: Job job_1368318686284_0013 failed with 
> state FAILED due to: Job commit failed: java.io.IOException:
> numberOfSecretKeys: 0
> numberOfTokens: 0
>   at 
> org.myorg.LostCreds$DestroyerFileOutputCommitter.commitJob(LostCreds.java:43)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.handleJobCommit(CommitterEventHandler.java:249)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.run(CommitterEventHandler.java:212)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>   at java.lang.Thread.run(Thread.java:619)
> {noformat}
> As you can see, even though we've clearly initialized the creds via:
> {noformat}
> job.getCredentials().addSecretKey(new Text("mykey"), "mysecret".getBytes());
> {noformat}
> It doesn't seem to appear later in the job.
> This is a pretty critical issue for Sqoop 2 since it appears to be DOA for 
> YARN in Hadoop 2.0.4-alpha

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5240) inside of FileOutputCommitter the initialized Credentials cache appears to be empty

2013-05-16 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5240?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-5240:


Attachment: MAPREDUCE-5240.2.0.4.rvs.patch.txt

Attaching a modified patch for branch-2.0.4

> inside of FileOutputCommitter the initialized Credentials cache appears to be 
> empty
> ---
>
> Key: MAPREDUCE-5240
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5240
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mrv2
>Affects Versions: 2.0.4-alpha
>Reporter: Roman Shaposhnik
>Assignee: Vinod Kumar Vavilapalli
>Priority: Blocker
>  Labels: 2.0.4.1
> Fix For: 2.0.5-beta, 2.0.4.1-alpha
>
> Attachments: LostCreds.java, MAPREDUCE-5240-20130512.txt, 
> MAPREDUCE-5240-20130513.txt, MAPREDUCE-5240.2.0.4.rvs.patch.txt
>
>
> I am attaching a modified wordcount job that clearly demonstrates the problem 
> we've encountered in running Sqoop2 on YARN (BIGTOP-949).
> Here's what running it produces:
> {noformat}
> $ hadoop fs -mkdir in
> $ hadoop fs -put /etc/passwd in
> $ hadoop jar ./bug.jar org.myorg.LostCreds
> 13/05/12 03:13:46 WARN mapred.JobConf: The variable mapred.child.ulimit is no 
> longer used.
> numberOfSecretKeys: 1
> numberOfTokens: 0
> ..
> ..
> ..
> 13/05/12 03:05:35 INFO mapreduce.Job: Job job_1368318686284_0013 failed with 
> state FAILED due to: Job commit failed: java.io.IOException:
> numberOfSecretKeys: 0
> numberOfTokens: 0
>   at 
> org.myorg.LostCreds$DestroyerFileOutputCommitter.commitJob(LostCreds.java:43)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.handleJobCommit(CommitterEventHandler.java:249)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.run(CommitterEventHandler.java:212)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>   at java.lang.Thread.run(Thread.java:619)
> {noformat}
> As you can see, even though we've clearly initialized the creds via:
> {noformat}
> job.getCredentials().addSecretKey(new Text("mykey"), "mysecret".getBytes());
> {noformat}
> It doesn't seem to appear later in the job.
> This is a pretty critical issue for Sqoop 2 since it appears to be DOA for 
> YARN in Hadoop 2.0.4-alpha

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5240) inside of FileOutputCommitter the initialized Credentials cache appears to be empty

2013-05-12 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5240?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655768#comment-13655768
 ] 

Roman Shaposhnik commented on MAPREDUCE-5240:
-

Vinod, thanks a million for attaching the patch! I'm building it right now in 
Bigtop and hope to run validation tests tomorrow! Will keep you posted on a 
JIRA.

> inside of FileOutputCommitter the initialized Credentials cache appears to be 
> empty
> ---
>
> Key: MAPREDUCE-5240
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5240
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mrv2
>Affects Versions: 2.0.4-alpha
>Reporter: Roman Shaposhnik
>Assignee: Vinod Kumar Vavilapalli
>Priority: Blocker
> Fix For: 2.0.5-beta
>
> Attachments: LostCreds.java, MAPREDUCE-5240-20130512.txt
>
>
> I am attaching a modified wordcount job that clearly demonstrates the problem 
> we've encountered in running Sqoop2 on YARN (BIGTOP-949).
> Here's what running it produces:
> {noformat}
> $ hadoop fs -mkdir in
> $ hadoop fs -put /etc/passwd in
> $ hadoop jar ./bug.jar org.myorg.LostCreds
> 13/05/12 03:13:46 WARN mapred.JobConf: The variable mapred.child.ulimit is no 
> longer used.
> numberOfSecretKeys: 1
> numberOfTokens: 0
> ..
> ..
> ..
> 13/05/12 03:05:35 INFO mapreduce.Job: Job job_1368318686284_0013 failed with 
> state FAILED due to: Job commit failed: java.io.IOException:
> numberOfSecretKeys: 0
> numberOfTokens: 0
>   at 
> org.myorg.LostCreds$DestroyerFileOutputCommitter.commitJob(LostCreds.java:43)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.handleJobCommit(CommitterEventHandler.java:249)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.run(CommitterEventHandler.java:212)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>   at java.lang.Thread.run(Thread.java:619)
> {noformat}
> As you can see, even though we've clearly initialized the creds via:
> {noformat}
> job.getCredentials().addSecretKey(new Text("mykey"), "mysecret".getBytes());
> {noformat}
> It doesn't seem to appear later in the job.
> This is a pretty critical issue for Sqoop 2 since it appears to be DOA for 
> YARN in Hadoop 2.0.4-alpha

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5240) inside of FileOutputCommitter the initialized Credentials cache appears to be empty

2013-05-12 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5240?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13655638#comment-13655638
 ] 

Roman Shaposhnik commented on MAPREDUCE-5240:
-

Vinod, if it is not too difficult I'd greatly appreciate a small patch for 
specifically this very issue. Thanks!

> inside of FileOutputCommitter the initialized Credentials cache appears to be 
> empty
> ---
>
> Key: MAPREDUCE-5240
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5240
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mrv2
>Affects Versions: 2.0.4-alpha
>Reporter: Roman Shaposhnik
>Assignee: Vinod Kumar Vavilapalli
>Priority: Blocker
> Fix For: 2.0.5-beta
>
> Attachments: LostCreds.java
>
>
> I am attaching a modified wordcount job that clearly demonstrates the problem 
> we've encountered in running Sqoop2 on YARN (BIGTOP-949).
> Here's what running it produces:
> {noformat}
> $ hadoop fs -mkdir in
> $ hadoop fs -put /etc/passwd in
> $ hadoop jar ./bug.jar org.myorg.LostCreds
> 13/05/12 03:13:46 WARN mapred.JobConf: The variable mapred.child.ulimit is no 
> longer used.
> numberOfSecretKeys: 1
> numberOfTokens: 0
> ..
> ..
> ..
> 13/05/12 03:05:35 INFO mapreduce.Job: Job job_1368318686284_0013 failed with 
> state FAILED due to: Job commit failed: java.io.IOException:
> numberOfSecretKeys: 0
> numberOfTokens: 0
>   at 
> org.myorg.LostCreds$DestroyerFileOutputCommitter.commitJob(LostCreds.java:43)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.handleJobCommit(CommitterEventHandler.java:249)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.run(CommitterEventHandler.java:212)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>   at java.lang.Thread.run(Thread.java:619)
> {noformat}
> As you can see, even though we've clearly initialized the creds via:
> {noformat}
> job.getCredentials().addSecretKey(new Text("mykey"), "mysecret".getBytes());
> {noformat}
> It doesn't seem to appear later in the job.
> This is a pretty critical issue for Sqoop 2 since it appears to be DOA for 
> YARN in Hadoop 2.0.4-alpha

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5240) inside of FileOutputCommitter the initialized Credentials cache appears to be empty

2013-05-11 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5240?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-5240:



As was mentioned on the Bigtop JIRA this looks related to HADOOP-8726

> inside of FileOutputCommitter the initialized Credentials cache appears to be 
> empty
> ---
>
> Key: MAPREDUCE-5240
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5240
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mrv1
>Affects Versions: 2.0.4-alpha
>Reporter: Roman Shaposhnik
>Priority: Blocker
> Fix For: 2.0.5-beta
>
> Attachments: LostCreds.java
>
>
> I am attaching a modified wordcount job that clearly demonstrates the problem 
> we've encountered in running Sqoop2 on YARN (BIGTOP-949).
> Here's what running it produces:
> {noformat}
> $ hadoop fs -mkdir in
> $ hadoop fs -put /etc/passwd in
> $ hadoop jar ./bug.jar org.myorg.LostCreds
> 13/05/12 03:13:46 WARN mapred.JobConf: The variable mapred.child.ulimit is no 
> longer used.
> numberOfSecretKeys: 1
> numberOfTokens: 0
> ..
> ..
> ..
> 13/05/12 03:05:35 INFO mapreduce.Job: Job job_1368318686284_0013 failed with 
> state FAILED due to: Job commit failed: java.io.IOException:
> numberOfSecretKeys: 0
> numberOfTokens: 0
>   at 
> org.myorg.LostCreds$DestroyerFileOutputCommitter.commitJob(LostCreds.java:43)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.handleJobCommit(CommitterEventHandler.java:249)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.run(CommitterEventHandler.java:212)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>   at java.lang.Thread.run(Thread.java:619)
> {noformat}
> As you can see, even though we've clearly initialized the creds via:
> {noformat}
> job.getCredentials().addSecretKey(new Text("mykey"), "mysecret".getBytes());
> {noformat}
> It doesn't seem to appear later in the job.
> This is a pretty critical issue for Sqoop 2 since it appears to be DOA for 
> YARN in Hadoop 2.0.4-alpha

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5240) inside of FileOutputCommitter the initialized Credentials cache appears to be empty

2013-05-11 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5240?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-5240:


Attachment: LostCreds.java

> inside of FileOutputCommitter the initialized Credentials cache appears to be 
> empty
> ---
>
> Key: MAPREDUCE-5240
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5240
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mrv1
>Affects Versions: 2.0.4-alpha
>Reporter: Roman Shaposhnik
>Priority: Blocker
> Fix For: 2.0.5-beta
>
> Attachments: LostCreds.java
>
>
> I am attaching a modified wordcount job that clearly demonstrates the problem 
> we've encountered in running Sqoop2 on YARN (BIGTOP-949).
> Here's what running it produces:
> {noformat}
> $ hadoop fs -mkdir in
> $ hadoop fs -put /etc/passwd in
> $ hadoop jar ./bug.jar org.myorg.LostCreds
> 13/05/12 03:13:46 WARN mapred.JobConf: The variable mapred.child.ulimit is no 
> longer used.
> numberOfSecretKeys: 1
> numberOfTokens: 0
> ..
> ..
> ..
> 13/05/12 03:05:35 INFO mapreduce.Job: Job job_1368318686284_0013 failed with 
> state FAILED due to: Job commit failed: java.io.IOException: 
> numberOfSecretKeys: 0
> numberOfTokens: 0
>   at 
> org.myorg.LostCreds$DestroyerFileOutputCommitter.commitJob(LostCreds.java:43)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.handleJobCommit(CommitterEventHandler.java:249)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.run(CommitterEventHandler.java:212)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>   at java.lang.Thread.run(Thread.java:619)
> {noformat}
> As you can see, even though we've clearly initialized the creds via:
> {noformat}
> job.getCredentials().addSecretKey(new Text("mykey"), "mysecret".getBytes());
> {noformat}
> It doesn't seem to appear later in the job.
> This is a pretty critical issue for Sqoop 2 since it appears to be DOA for 
> YARN in Hadoop 2.0.4-alpha

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5240) inside of FileOutputCommitter the initialized Credentials cache appears to be empty

2013-05-11 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5240?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-5240:


Description: 
I am attaching a modified wordcount job that clearly demonstrates the problem 
we've encountered in running Sqoop2 on YARN (BIGTOP-949).

Here's what running it produces:

{noformat}
$ hadoop fs -mkdir in
$ hadoop fs -put /etc/passwd in
$ hadoop jar ./bug.jar org.myorg.LostCreds
13/05/12 03:13:46 WARN mapred.JobConf: The variable mapred.child.ulimit is no 
longer used.
numberOfSecretKeys: 1
numberOfTokens: 0
..
..
..
13/05/12 03:05:35 INFO mapreduce.Job: Job job_1368318686284_0013 failed with 
state FAILED due to: Job commit failed: java.io.IOException:
numberOfSecretKeys: 0
numberOfTokens: 0
at 
org.myorg.LostCreds$DestroyerFileOutputCommitter.commitJob(LostCreds.java:43)
at 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.handleJobCommit(CommitterEventHandler.java:249)
at 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.run(CommitterEventHandler.java:212)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:619)
{noformat}

As you can see, even though we've clearly initialized the creds via:

{noformat}
job.getCredentials().addSecretKey(new Text("mykey"), "mysecret".getBytes());
{noformat}

It doesn't seem to appear later in the job.

This is a pretty critical issue for Sqoop 2 since it appears to be DOA for YARN 
in Hadoop 2.0.4-alpha

  was:
I am attaching a modified wordcount job that clearly demonstrates the problem 
we've encountered in running Sqoop2 on YARN (BIGTOP-949).

Here's what running it produces:

{noformat}
$ hadoop fs -mkdir in
$ hadoop fs -put /etc/passwd in
$ hadoop jar ./bug.jar org.myorg.LostCreds
13/05/12 03:13:46 WARN mapred.JobConf: The variable mapred.child.ulimit is no 
longer used.
numberOfSecretKeys: 1
numberOfTokens: 0
..
..
..
13/05/12 03:05:35 INFO mapreduce.Job: Job job_1368318686284_0013 failed with 
state FAILED due to: Job commit failed: java.io.IOException: 
numberOfSecretKeys: 0
numberOfTokens: 0
at 
org.myorg.LostCreds$DestroyerFileOutputCommitter.commitJob(LostCreds.java:43)
at 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.handleJobCommit(CommitterEventHandler.java:249)
at 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.run(CommitterEventHandler.java:212)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:619)
{noformat}

As you can see, even though we've clearly initialized the creds via:

{noformat}
job.getCredentials().addSecretKey(new Text("mykey"), "mysecret".getBytes());
{noformat}

It doesn't seem to appear later in the job.

This is a pretty critical issue for Sqoop 2 since it appears to be DOA for YARN 
in Hadoop 2.0.4-alpha


> inside of FileOutputCommitter the initialized Credentials cache appears to be 
> empty
> ---
>
> Key: MAPREDUCE-5240
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5240
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mrv1
>Affects Versions: 2.0.4-alpha
>Reporter: Roman Shaposhnik
>Priority: Blocker
> Fix For: 2.0.5-beta
>
> Attachments: LostCreds.java
>
>
> I am attaching a modified wordcount job that clearly demonstrates the problem 
> we've encountered in running Sqoop2 on YARN (BIGTOP-949).
> Here's what running it produces:
> {noformat}
> $ hadoop fs -mkdir in
> $ hadoop fs -put /etc/passwd in
> $ hadoop jar ./bug.jar org.myorg.LostCreds
> 13/05/12 03:13:46 WARN mapred.JobConf: The variable mapred.child.ulimit is no 
> longer used.
> numberOfSecretKeys: 1
> numberOfTokens: 0
> ..
> ..
> ..
> 13/05/12 03:05:35 INFO mapreduce.Job: Job job_1368318686284_0013 failed with 
> state FAILED due to: Job commit failed: java.io.IOException:
> numberOfSecretKeys: 0
> numberOfTokens: 0
>   at 
> org.myorg.LostCreds$DestroyerFileOutputCommitter.commitJob(LostCreds.java:43)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.handleJobCommit(CommitterEventHandler.java:249)
>   at 
> org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.run(CommitterEventHandler.java:212)
>   at 
> java.util.concur

[jira] [Created] (MAPREDUCE-5240) inside of FileOutputCommitter the initialized Credentials cache appears to be empty

2013-05-11 Thread Roman Shaposhnik (JIRA)
Roman Shaposhnik created MAPREDUCE-5240:
---

 Summary: inside of FileOutputCommitter the initialized Credentials 
cache appears to be empty
 Key: MAPREDUCE-5240
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5240
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv1
Affects Versions: 2.0.4-alpha
Reporter: Roman Shaposhnik
Priority: Blocker
 Fix For: 2.0.5-beta
 Attachments: LostCreds.java

I am attaching a modified wordcount job that clearly demonstrates the problem 
we've encountered in running Sqoop2 on YARN (BIGTOP-949).

Here's what running it produces:

{noformat}
$ hadoop fs -mkdir in
$ hadoop fs -put /etc/passwd in
$ hadoop jar ./bug.jar org.myorg.LostCreds
13/05/12 03:13:46 WARN mapred.JobConf: The variable mapred.child.ulimit is no 
longer used.
numberOfSecretKeys: 1
numberOfTokens: 0
..
..
..
13/05/12 03:05:35 INFO mapreduce.Job: Job job_1368318686284_0013 failed with 
state FAILED due to: Job commit failed: java.io.IOException: 
numberOfSecretKeys: 0
numberOfTokens: 0
at 
org.myorg.LostCreds$DestroyerFileOutputCommitter.commitJob(LostCreds.java:43)
at 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.handleJobCommit(CommitterEventHandler.java:249)
at 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler$EventProcessor.run(CommitterEventHandler.java:212)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:619)
{noformat}

As you can see, even though we've clearly initialized the creds via:

{noformat}
job.getCredentials().addSecretKey(new Text("mykey"), "mysecret".getBytes());
{noformat}

It doesn't seem to appear later in the job.

This is a pretty critical issue for Sqoop 2 since it appears to be DOA for YARN 
in Hadoop 2.0.4-alpha

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5088) MR Client gets an renewer token exception while Oozie is submitting a job

2013-03-28 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5088?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13616835#comment-13616835
 ] 

Roman Shaposhnik commented on MAPREDUCE-5088:
-

Sid, would you mind taking a look at MAPREDUCE-5117 ?

> MR Client gets an renewer token exception while Oozie is submitting a job
> -
>
> Key: MAPREDUCE-5088
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5088
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>Affects Versions: 2.0.3-alpha
>Reporter: Roman Shaposhnik
>Assignee: Daryn Sharp
>Priority: Blocker
> Fix For: 2.0.4-alpha
>
> Attachments: HADOOP-9409.patch, HADOOP-9409.patch, 
> MAPREDUCE-5088.patch, MAPREDUCE-5088.patch, MAPREDUCE-5088.txt
>
>
> After the fix for HADOOP-9299 I'm now getting the following bizzare exception 
> in Oozie while trying to submit a job. This also seems to be KRB related:
> {noformat}
> 2013-03-15 13:34:16,555  WARN ActionStartXCommand:542 - USER[hue] GROUP[-] 
> TOKEN[] APP[MapReduce] JOB[001-130315123130987-oozie-oozi-W] 
> ACTION[001-130315123130987-oozie-oozi-W@Sleep] Error starting action 
> [Sleep]. ErrorType [ERROR], ErrorCode [UninitializedMessageException], 
> Message [UninitializedMessageException: Message missing required fields: 
> renewer]
> org.apache.oozie.action.ActionExecutorException: 
> UninitializedMessageException: Message missing required fields: renewer
>   at 
> org.apache.oozie.action.ActionExecutor.convertException(ActionExecutor.java:401)
>   at 
> org.apache.oozie.action.hadoop.JavaActionExecutor.submitLauncher(JavaActionExecutor.java:738)
>   at 
> org.apache.oozie.action.hadoop.JavaActionExecutor.start(JavaActionExecutor.java:889)
>   at 
> org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:211)
>   at 
> org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:59)
>   at org.apache.oozie.command.XCommand.call(XCommand.java:277)
>   at 
> org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:326)
>   at 
> org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:255)
>   at 
> org.apache.oozie.service.CallableQueueService$CallableWrapper.run(CallableQueueService.java:175)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>   at java.lang.Thread.run(Thread.java:662)
> Caused by: com.google.protobuf.UninitializedMessageException: Message missing 
> required fields: renewer
>   at 
> com.google.protobuf.AbstractMessage$Builder.newUninitializedMessageException(AbstractMessage.java:605)
>   at 
> org.apache.hadoop.security.proto.SecurityProtos$GetDelegationTokenRequestProto$Builder.build(SecurityProtos.java:973)
>   at 
> org.apache.hadoop.mapreduce.v2.api.protocolrecords.impl.pb.GetDelegationTokenRequestPBImpl.mergeLocalToProto(GetDelegationTokenRequestPBImpl.java:84)
>   at 
> org.apache.hadoop.mapreduce.v2.api.protocolrecords.impl.pb.GetDelegationTokenRequestPBImpl.getProto(GetDelegationTokenRequestPBImpl.java:67)
>   at 
> org.apache.hadoop.mapreduce.v2.api.impl.pb.client.MRClientProtocolPBClientImpl.getDelegationToken(MRClientProtocolPBClientImpl.java:200)
>   at 
> org.apache.hadoop.mapred.YARNRunner.getDelegationTokenFromHS(YARNRunner.java:194)
>   at org.apache.hadoop.mapred.YARNRunner.submitJob(YARNRunner.java:273)
>   at 
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:392)
>   at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1218)
>   at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1215)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1439)
>   at org.apache.hadoop.mapreduce.Job.submit(Job.java:1215)
>   at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:581)
>   at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:576)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1439)
>   at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:576)
>   at 
> org.apache.oozie.action.hadoop.JavaActionExecutor.submitLauncher(JavaActionExecutor.java:723)
>   ... 10 more
> 2013-03-15 13:34:16,555  WARN ActionStartXCommand:542 - USER[hue] GROUP[-] 
> 

[jira] [Updated] (MAPREDUCE-5117) With security enabled HS delegation token renewer fails

2013-03-28 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5117?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-5117:


Attachment: yarn.log

> With security enabled HS delegation token renewer fails
> ---
>
> Key: MAPREDUCE-5117
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5117
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: security
>Affects Versions: 2.0.4-alpha
>Reporter: Roman Shaposhnik
>Priority: Blocker
> Fix For: 2.0.4-alpha
>
> Attachments: yarn.log
>
>
> It seems that the HSClientProtocolPBClientImpl should implement Closeable as 
> per the attached stack trace. The problem can be observed on a cluster 
> running the latest branch-2.0.4-alpha with MAPREDUCE-5088 applied on top. The 
> easiest way to reproduce it is to run an oozie pig job:
> {noformat}
> $ oozie job -oozie http://`hostname -f`:11000/oozie -run 
> -DjobTracker=`hostname -f`:8032 -DnameNode=hdfs://`hostname -f`:17020 
> -DexamplesRoot=examples -config /tmp/examples/apps/pig/job.properties
> {noformat}
> Please also note that I can successfully submit simple jobs (Pi/Sleep) from a 
> command line using hadoop jar command. Thus it *seems* related to 
> MAPREDUCE-5088 change.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (MAPREDUCE-5117) With security enabled HS delegation token renewer fails

2013-03-28 Thread Roman Shaposhnik (JIRA)
Roman Shaposhnik created MAPREDUCE-5117:
---

 Summary: With security enabled HS delegation token renewer fails
 Key: MAPREDUCE-5117
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5117
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: security
Affects Versions: 2.0.4-alpha
Reporter: Roman Shaposhnik
Priority: Blocker
 Fix For: 2.0.4-alpha


It seems that the HSClientProtocolPBClientImpl should implement Closeable as 
per the attached stack trace. The problem can be observed on a cluster running 
the latest branch-2.0.4-alpha with MAPREDUCE-5088 applied on top. The easiest 
way to reproduce it is to run an oozie pig job:

{noformat}
$ oozie job -oozie http://`hostname -f`:11000/oozie -run -DjobTracker=`hostname 
-f`:8032 -DnameNode=hdfs://`hostname -f`:17020 -DexamplesRoot=examples -config 
/tmp/examples/apps/pig/job.properties
{noformat}

Please also note that I can successfully submit simple jobs (Pi/Sleep) from a 
command line using hadoop jar command. Thus it *seems* related to 
MAPREDUCE-5088 change.


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-4549) Distributed cache conflicts breaks backwards compatability

2013-03-28 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-4549:


Resolution: Fixed
Status: Resolved  (was: Patch Available)

> Distributed cache conflicts breaks backwards compatability
> --
>
> Key: MAPREDUCE-4549
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-4549
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mrv2
>Affects Versions: 0.23.3, 2.0.2-alpha
>Reporter: Robert Joseph Evans
>Assignee: Robert Joseph Evans
>Priority: Blocker
> Fix For: 2.0.4-alpha, 0.23.5
>
> Attachments: MAPREDUCE-4549-trunk.patch, MR-4549-branch-0.23.txt
>
>
> I recently put in MAPREDUCE-4503 which went a bit too far, and broke 
> backwards compatibility with 1.0 in distribtued cache entries.  instead of 
> changing the behavior of the distributed cache to more closely match 1.0 
> behavior I want to just change the exception to a warning message informing 
> the users that it will become an error in 2.0

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-4820) MRApps distributed-cache duplicate checks are incorrect

2013-03-26 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-4820?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13614482#comment-13614482
 ] 

Roman Shaposhnik commented on MAPREDUCE-4820:
-

[~revans2] I think you may be absolutely right. This could be a residual effect 
of enabling OOZIE-1089 (which we probably shouldn't do now that the actual 
issue seems to be fixed). Let me poke on the Oozie side and report back. Thanks 
for your help so far!

> MRApps distributed-cache duplicate checks are incorrect
> ---
>
> Key: MAPREDUCE-4820
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-4820
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mr-am
>Affects Versions: 2.0.2-alpha
>Reporter: Alejandro Abdelnur
>Priority: Blocker
> Fix For: 2.0.4-alpha
>
> Attachments: launcher-job.conf.xml, launcher-job.logs.txt, 
> mr-job.conf.xml, mr-job.logs.txt
>
>
> This seems a combination of issues that are being exposed in 2.0.2-alpha by 
> MAPREDUCE-4549.
> MAPREDUCE-4549 introduces a check to to ensure there are not duplicate JARs 
> in the distributed-cache (using the JAR name as identity).
> In Hadoop 2 (different from Hadoop 1), all JARs in the distributed-cache are 
> symlink-ed to the current directory of the task.
> MRApps, when setting up the DistributedCache 
> (MRApps#setupDistributedCache->parseDistributedCacheArtifacts) assumes that 
> the local resources (this includes files in the CURRENT_DIR/, 
> CURRENT_DIR/classes/ and files in CURRENT_DIR/lib/) are part of the 
> distributed-cache already.
> For systems, like Oozie, which use a launcher job to submit the real job this 
> poses a problem because MRApps is run from the launcher job to submit the 
> real job. The configuration of the real job has the correct distributed-cache 
> entries (no duplicates), but because the current dir has the same files, the 
> submission fails.
> It seems that MRApps should not be checking dups in the distributed-cached 
> against JARs in the CURRENT_DIR/ or CURRENT_DIR/lib/. The dup check should be 
> done among distributed-cached entries only.
> It seems YARNRunner is symlink-ing all files in the distributed cached in the 
> current directory. In Hadoop 1 this was done only for files added to the 
> distributed-cache using a fragment (ie "#FOO") to trigger a symlink creation. 
> Marking as a blocker because without a fix for this, Oozie cannot submit jobs 
> to Hadoop 2 (i've debugged Oozie in a live cluster being used by BigTop 
> -thanks Roman- to test their release work, and I've verified that Oozie 3.3 
> does not create duplicated entries in the distributed-cache)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-4549) Distributed cache conflicts breaks backwards compatability

2013-03-26 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-4549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13614367#comment-13614367
 ] 

Roman Shaposhnik commented on MAPREDUCE-4549:
-

All the findings are attached over here: MAPREDUCE-4820

At this point I'd agree with Arun and close this one. We can track what's left 
over at MAPREDUCE-4820

> Distributed cache conflicts breaks backwards compatability
> --
>
> Key: MAPREDUCE-4549
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-4549
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mrv2
>Affects Versions: 0.23.3, 2.0.2-alpha
>Reporter: Robert Joseph Evans
>Assignee: Robert Joseph Evans
>Priority: Blocker
> Fix For: 0.23.5, 2.0.4-alpha
>
> Attachments: MAPREDUCE-4549-trunk.patch, MR-4549-branch-0.23.txt
>
>
> I recently put in MAPREDUCE-4503 which went a bit too far, and broke 
> backwards compatibility with 1.0 in distribtued cache entries.  instead of 
> changing the behavior of the distributed cache to more closely match 1.0 
> behavior I want to just change the exception to a warning message informing 
> the users that it will become an error in 2.0

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5088) MR Client gets an renewer token exception while Oozie is submitting a job

2013-03-24 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5088?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13612326#comment-13612326
 ] 

Roman Shaposhnik commented on MAPREDUCE-5088:
-

Guys, what's the latest on this patch? I've tested it in Bigtop and it seems to 
do exactly what it needs as far as unblocking Oozie is concerned. What else is 
left before we can commit it to trunk and all the relevant branches?

> MR Client gets an renewer token exception while Oozie is submitting a job
> -
>
> Key: MAPREDUCE-5088
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5088
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>Affects Versions: 2.0.3-alpha
>Reporter: Roman Shaposhnik
>Assignee: Daryn Sharp
>Priority: Blocker
> Fix For: 2.0.4-alpha
>
> Attachments: HADOOP-9409.patch, HADOOP-9409.patch, 
> MAPREDUCE-5088.patch, MAPREDUCE-5088.patch
>
>
> After the fix for HADOOP-9299 I'm now getting the following bizzare exception 
> in Oozie while trying to submit a job. This also seems to be KRB related:
> {noformat}
> 2013-03-15 13:34:16,555  WARN ActionStartXCommand:542 - USER[hue] GROUP[-] 
> TOKEN[] APP[MapReduce] JOB[001-130315123130987-oozie-oozi-W] 
> ACTION[001-130315123130987-oozie-oozi-W@Sleep] Error starting action 
> [Sleep]. ErrorType [ERROR], ErrorCode [UninitializedMessageException], 
> Message [UninitializedMessageException: Message missing required fields: 
> renewer]
> org.apache.oozie.action.ActionExecutorException: 
> UninitializedMessageException: Message missing required fields: renewer
>   at 
> org.apache.oozie.action.ActionExecutor.convertException(ActionExecutor.java:401)
>   at 
> org.apache.oozie.action.hadoop.JavaActionExecutor.submitLauncher(JavaActionExecutor.java:738)
>   at 
> org.apache.oozie.action.hadoop.JavaActionExecutor.start(JavaActionExecutor.java:889)
>   at 
> org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:211)
>   at 
> org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:59)
>   at org.apache.oozie.command.XCommand.call(XCommand.java:277)
>   at 
> org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:326)
>   at 
> org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:255)
>   at 
> org.apache.oozie.service.CallableQueueService$CallableWrapper.run(CallableQueueService.java:175)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>   at java.lang.Thread.run(Thread.java:662)
> Caused by: com.google.protobuf.UninitializedMessageException: Message missing 
> required fields: renewer
>   at 
> com.google.protobuf.AbstractMessage$Builder.newUninitializedMessageException(AbstractMessage.java:605)
>   at 
> org.apache.hadoop.security.proto.SecurityProtos$GetDelegationTokenRequestProto$Builder.build(SecurityProtos.java:973)
>   at 
> org.apache.hadoop.mapreduce.v2.api.protocolrecords.impl.pb.GetDelegationTokenRequestPBImpl.mergeLocalToProto(GetDelegationTokenRequestPBImpl.java:84)
>   at 
> org.apache.hadoop.mapreduce.v2.api.protocolrecords.impl.pb.GetDelegationTokenRequestPBImpl.getProto(GetDelegationTokenRequestPBImpl.java:67)
>   at 
> org.apache.hadoop.mapreduce.v2.api.impl.pb.client.MRClientProtocolPBClientImpl.getDelegationToken(MRClientProtocolPBClientImpl.java:200)
>   at 
> org.apache.hadoop.mapred.YARNRunner.getDelegationTokenFromHS(YARNRunner.java:194)
>   at org.apache.hadoop.mapred.YARNRunner.submitJob(YARNRunner.java:273)
>   at 
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:392)
>   at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1218)
>   at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1215)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1439)
>   at org.apache.hadoop.mapreduce.Job.submit(Job.java:1215)
>   at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:581)
>   at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:576)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1439)
>   at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:576)
>   at 
> org.apache.oozie.action.hadoop.JavaAct

[jira] [Commented] (MAPREDUCE-4820) MRApps distributed-cache duplicate checks are incorrect

2013-03-22 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-4820?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13611154#comment-13611154
 ] 

Roman Shaposhnik commented on MAPREDUCE-4820:
-

Robert, I just tried it with the tip of the branch-2.0.4-alpha and I still see 
the problem. I'm building the following: 

https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2.0.4-alpha@1457977

Where do you say it is fixed?

> MRApps distributed-cache duplicate checks are incorrect
> ---
>
> Key: MAPREDUCE-4820
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-4820
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mr-am
>Affects Versions: 2.0.2-alpha
>Reporter: Alejandro Abdelnur
>Priority: Blocker
> Fix For: 2.0.4-alpha
>
> Attachments: launcher-job.conf.xml, launcher-job.logs.txt, 
> mr-job.conf.xml, mr-job.logs.txt
>
>
> This seems a combination of issues that are being exposed in 2.0.2-alpha by 
> MAPREDUCE-4549.
> MAPREDUCE-4549 introduces a check to to ensure there are not duplicate JARs 
> in the distributed-cache (using the JAR name as identity).
> In Hadoop 2 (different from Hadoop 1), all JARs in the distributed-cache are 
> symlink-ed to the current directory of the task.
> MRApps, when setting up the DistributedCache 
> (MRApps#setupDistributedCache->parseDistributedCacheArtifacts) assumes that 
> the local resources (this includes files in the CURRENT_DIR/, 
> CURRENT_DIR/classes/ and files in CURRENT_DIR/lib/) are part of the 
> distributed-cache already.
> For systems, like Oozie, which use a launcher job to submit the real job this 
> poses a problem because MRApps is run from the launcher job to submit the 
> real job. The configuration of the real job has the correct distributed-cache 
> entries (no duplicates), but because the current dir has the same files, the 
> submission fails.
> It seems that MRApps should not be checking dups in the distributed-cached 
> against JARs in the CURRENT_DIR/ or CURRENT_DIR/lib/. The dup check should be 
> done among distributed-cached entries only.
> It seems YARNRunner is symlink-ing all files in the distributed cached in the 
> current directory. In Hadoop 1 this was done only for files added to the 
> distributed-cache using a fragment (ie "#FOO") to trigger a symlink creation. 
> Marking as a blocker because without a fix for this, Oozie cannot submit jobs 
> to Hadoop 2 (i've debugged Oozie in a live cluster being used by BigTop 
> -thanks Roman- to test their release work, and I've verified that Oozie 3.3 
> does not create duplicated entries in the distributed-cache)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-4820) MRApps distributed-cache duplicate checks are incorrect

2013-03-21 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4820?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-4820:


Attachment: launcher-job.conf.xml
launcher-job.logs.txt
mr-job.logs.txt
mr-job.conf.xml

To make the debugging easier I'm attaching logs and XML confs that I see on my 
cluster after submitting:

{noformat}
$ oozie job -oozie http://localhost:11000/oozie -run -DjobTracker=bigtop:8032 
-DnameNode=hdfs://bigtop:17020 -DexamplesRoot=examples -config 
/tmp/examples/apps/map-reduce/job.properties
{noformat}

> MRApps distributed-cache duplicate checks are incorrect
> ---
>
> Key: MAPREDUCE-4820
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-4820
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mr-am
>Affects Versions: 2.0.2-alpha
>Reporter: Alejandro Abdelnur
>Priority: Blocker
> Fix For: 2.0.4-alpha
>
> Attachments: launcher-job.conf.xml, launcher-job.logs.txt, 
> mr-job.conf.xml, mr-job.logs.txt
>
>
> This seems a combination of issues that are being exposed in 2.0.2-alpha by 
> MAPREDUCE-4549.
> MAPREDUCE-4549 introduces a check to to ensure there are not duplicate JARs 
> in the distributed-cache (using the JAR name as identity).
> In Hadoop 2 (different from Hadoop 1), all JARs in the distributed-cache are 
> symlink-ed to the current directory of the task.
> MRApps, when setting up the DistributedCache 
> (MRApps#setupDistributedCache->parseDistributedCacheArtifacts) assumes that 
> the local resources (this includes files in the CURRENT_DIR/, 
> CURRENT_DIR/classes/ and files in CURRENT_DIR/lib/) are part of the 
> distributed-cache already.
> For systems, like Oozie, which use a launcher job to submit the real job this 
> poses a problem because MRApps is run from the launcher job to submit the 
> real job. The configuration of the real job has the correct distributed-cache 
> entries (no duplicates), but because the current dir has the same files, the 
> submission fails.
> It seems that MRApps should not be checking dups in the distributed-cached 
> against JARs in the CURRENT_DIR/ or CURRENT_DIR/lib/. The dup check should be 
> done among distributed-cached entries only.
> It seems YARNRunner is symlink-ing all files in the distributed cached in the 
> current directory. In Hadoop 1 this was done only for files added to the 
> distributed-cache using a fragment (ie "#FOO") to trigger a symlink creation. 
> Marking as a blocker because without a fix for this, Oozie cannot submit jobs 
> to Hadoop 2 (i've debugged Oozie in a live cluster being used by BigTop 
> -thanks Roman- to test their release work, and I've verified that Oozie 3.3 
> does not create duplicated entries in the distributed-cache)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-4549) Distributed cache conflicts breaks backwards compatability

2013-03-21 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-4549:


 Priority: Blocker  (was: Critical)
Fix Version/s: (was: 2.0.5-beta)
   2.0.4-alpha

Making this a blocker for 2.0.4-alpha as agreed. I will keep attaching 
logs/debugging data from Oozie runs to MAPREDUCE-4820 

> Distributed cache conflicts breaks backwards compatability
> --
>
> Key: MAPREDUCE-4549
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-4549
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mrv2
>Affects Versions: 0.23.3, 2.0.2-alpha
>Reporter: Robert Joseph Evans
>Assignee: Robert Joseph Evans
>Priority: Blocker
> Fix For: 0.23.5, 2.0.4-alpha
>
> Attachments: MAPREDUCE-4549-trunk.patch, MR-4549-branch-0.23.txt
>
>
> I recently put in MAPREDUCE-4503 which went a bit too far, and broke 
> backwards compatibility with 1.0 in distribtued cache entries.  instead of 
> changing the behavior of the distributed cache to more closely match 1.0 
> behavior I want to just change the exception to a warning message informing 
> the users that it will become an error in 2.0

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-4820) MRApps distributed-cache duplicate checks are incorrect

2013-03-21 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4820?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-4820:


 Priority: Blocker  (was: Major)
Fix Version/s: (was: 2.0.5-beta)
   2.0.4-alpha

> MRApps distributed-cache duplicate checks are incorrect
> ---
>
> Key: MAPREDUCE-4820
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-4820
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mr-am
>Affects Versions: 2.0.2-alpha
>Reporter: Alejandro Abdelnur
>Priority: Blocker
> Fix For: 2.0.4-alpha
>
>
> This seems a combination of issues that are being exposed in 2.0.2-alpha by 
> MAPREDUCE-4549.
> MAPREDUCE-4549 introduces a check to to ensure there are not duplicate JARs 
> in the distributed-cache (using the JAR name as identity).
> In Hadoop 2 (different from Hadoop 1), all JARs in the distributed-cache are 
> symlink-ed to the current directory of the task.
> MRApps, when setting up the DistributedCache 
> (MRApps#setupDistributedCache->parseDistributedCacheArtifacts) assumes that 
> the local resources (this includes files in the CURRENT_DIR/, 
> CURRENT_DIR/classes/ and files in CURRENT_DIR/lib/) are part of the 
> distributed-cache already.
> For systems, like Oozie, which use a launcher job to submit the real job this 
> poses a problem because MRApps is run from the launcher job to submit the 
> real job. The configuration of the real job has the correct distributed-cache 
> entries (no duplicates), but because the current dir has the same files, the 
> submission fails.
> It seems that MRApps should not be checking dups in the distributed-cached 
> against JARs in the CURRENT_DIR/ or CURRENT_DIR/lib/. The dup check should be 
> done among distributed-cached entries only.
> It seems YARNRunner is symlink-ing all files in the distributed cached in the 
> current directory. In Hadoop 1 this was done only for files added to the 
> distributed-cache using a fragment (ie "#FOO") to trigger a symlink creation. 
> Marking as a blocker because without a fix for this, Oozie cannot submit jobs 
> to Hadoop 2 (i've debugged Oozie in a live cluster being used by BigTop 
> -thanks Roman- to test their release work, and I've verified that Oozie 3.3 
> does not create duplicated entries in the distributed-cache)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5088) MR Client gets an renewer token exception while Oozie is submitting a job

2013-03-21 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5088?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13609137#comment-13609137
 ] 

Roman Shaposhnik commented on MAPREDUCE-5088:
-

Guys, thanks a million for the patch. So far I tried the one from yesterday 
(20/Mar/13) and it completely unblocks Oozie in my Bigtop fully distributed 
testing.

I will try your latest patch today and report back.

On a related note it seems like  MAPREDUCE-4820 is biting us again -- I'll 
reopen it.

> MR Client gets an renewer token exception while Oozie is submitting a job
> -
>
> Key: MAPREDUCE-5088
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5088
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>Affects Versions: 2.0.3-alpha
>Reporter: Roman Shaposhnik
>Assignee: Daryn Sharp
>Priority: Blocker
> Fix For: 2.0.4-alpha
>
> Attachments: HADOOP-9409.patch, HADOOP-9409.patch, 
> MAPREDUCE-5088.patch
>
>
> After the fix for HADOOP-9299 I'm now getting the following bizzare exception 
> in Oozie while trying to submit a job. This also seems to be KRB related:
> {noformat}
> 2013-03-15 13:34:16,555  WARN ActionStartXCommand:542 - USER[hue] GROUP[-] 
> TOKEN[] APP[MapReduce] JOB[001-130315123130987-oozie-oozi-W] 
> ACTION[001-130315123130987-oozie-oozi-W@Sleep] Error starting action 
> [Sleep]. ErrorType [ERROR], ErrorCode [UninitializedMessageException], 
> Message [UninitializedMessageException: Message missing required fields: 
> renewer]
> org.apache.oozie.action.ActionExecutorException: 
> UninitializedMessageException: Message missing required fields: renewer
>   at 
> org.apache.oozie.action.ActionExecutor.convertException(ActionExecutor.java:401)
>   at 
> org.apache.oozie.action.hadoop.JavaActionExecutor.submitLauncher(JavaActionExecutor.java:738)
>   at 
> org.apache.oozie.action.hadoop.JavaActionExecutor.start(JavaActionExecutor.java:889)
>   at 
> org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:211)
>   at 
> org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:59)
>   at org.apache.oozie.command.XCommand.call(XCommand.java:277)
>   at 
> org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:326)
>   at 
> org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:255)
>   at 
> org.apache.oozie.service.CallableQueueService$CallableWrapper.run(CallableQueueService.java:175)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>   at java.lang.Thread.run(Thread.java:662)
> Caused by: com.google.protobuf.UninitializedMessageException: Message missing 
> required fields: renewer
>   at 
> com.google.protobuf.AbstractMessage$Builder.newUninitializedMessageException(AbstractMessage.java:605)
>   at 
> org.apache.hadoop.security.proto.SecurityProtos$GetDelegationTokenRequestProto$Builder.build(SecurityProtos.java:973)
>   at 
> org.apache.hadoop.mapreduce.v2.api.protocolrecords.impl.pb.GetDelegationTokenRequestPBImpl.mergeLocalToProto(GetDelegationTokenRequestPBImpl.java:84)
>   at 
> org.apache.hadoop.mapreduce.v2.api.protocolrecords.impl.pb.GetDelegationTokenRequestPBImpl.getProto(GetDelegationTokenRequestPBImpl.java:67)
>   at 
> org.apache.hadoop.mapreduce.v2.api.impl.pb.client.MRClientProtocolPBClientImpl.getDelegationToken(MRClientProtocolPBClientImpl.java:200)
>   at 
> org.apache.hadoop.mapred.YARNRunner.getDelegationTokenFromHS(YARNRunner.java:194)
>   at org.apache.hadoop.mapred.YARNRunner.submitJob(YARNRunner.java:273)
>   at 
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:392)
>   at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1218)
>   at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1215)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1439)
>   at org.apache.hadoop.mapreduce.Job.submit(Job.java:1215)
>   at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:581)
>   at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:576)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1439)
>   at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:576)

[jira] [Commented] (MAPREDUCE-3635) Improve Hadoop subcomponent integration in Hadoop 0.23

2013-03-08 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3635?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13597323#comment-13597323
 ] 

Roman Shaposhnik commented on MAPREDUCE-3635:
-

I think we're good.

> Improve Hadoop subcomponent integration in Hadoop 0.23
> --
>
> Key: MAPREDUCE-3635
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-3635
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: build, client
>Affects Versions: 0.23.0
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
> Fix For: 0.24.0
>
>
> Please see HADOOP-7939 for a complete description and discussion. This JIRA 
> is for patch tracking purposes only.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-4820) MRApps distributed-cache duplicate checks are incorrect

2013-01-29 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4820?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-4820:


 Priority: Blocker  (was: Critical)
Fix Version/s: 2.0.3-alpha

> MRApps distributed-cache duplicate checks are incorrect
> ---
>
> Key: MAPREDUCE-4820
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-4820
> Project: Hadoop Map/Reduce
>  Issue Type: Bug
>  Components: mr-am
>Affects Versions: 2.0.2-alpha
>Reporter: Alejandro Abdelnur
>Priority: Blocker
> Fix For: 2.0.3-alpha
>
>
> This seems a combination of issues that are being exposed in 2.0.2-alpha by 
> MAPREDUCE-4549.
> MAPREDUCE-4549 introduces a check to to ensure there are not duplicate JARs 
> in the distributed-cache (using the JAR name as identity).
> In Hadoop 2 (different from Hadoop 1), all JARs in the distributed-cache are 
> symlink-ed to the current directory of the task.
> MRApps, when setting up the DistributedCache 
> (MRApps#setupDistributedCache->parseDistributedCacheArtifacts) assumes that 
> the local resources (this includes files in the CURRENT_DIR/, 
> CURRENT_DIR/classes/ and files in CURRENT_DIR/lib/) are part of the 
> distributed-cache already.
> For systems, like Oozie, which use a launcher job to submit the real job this 
> poses a problem because MRApps is run from the launcher job to submit the 
> real job. The configuration of the real job has the correct distributed-cache 
> entries (no duplicates), but because the current dir has the same files, the 
> submission fails.
> It seems that MRApps should not be checking dups in the distributed-cached 
> against JARs in the CURRENT_DIR/ or CURRENT_DIR/lib/. The dup check should be 
> done among distributed-cached entries only.
> It seems YARNRunner is symlink-ing all files in the distributed cached in the 
> current directory. In Hadoop 1 this was done only for files added to the 
> distributed-cache using a fragment (ie "#FOO") to trigger a symlink creation. 
> Marking as a blocker because without a fix for this, Oozie cannot submit jobs 
> to Hadoop 2 (i've debugged Oozie in a live cluster being used by BigTop 
> -thanks Roman- to test their release work, and I've verified that Oozie 3.3 
> does not create duplicated entries in the distributed-cache)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-4219) make default container-executor.conf.dir be a path relative to the container-executor binary

2012-05-02 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-4219?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13266945#comment-13266945
 ] 

Roman Shaposhnik commented on MAPREDUCE-4219:
-

I tested Hadoop examples on the fully distributed secure cluster with ../conf 
baked in as a HADOOP_CONF

> make default container-executor.conf.dir be a path relative to the 
> container-executor binary
> 
>
> Key: MAPREDUCE-4219
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-4219
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: security
>Affects Versions: 0.23.1
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
> Fix For: 2.0.0
>
> Attachments: HADOOP-8332.patch.txt, MAPREDUCE-4219.patch.txt
>
>
> Currently, container-executor binary has an absolute pathname of its 
> configuration file baked in. This prevents an easy relocation of the 
> configuration files when dealing with multiple Hadoop installs on the same 
> node. It would be nice to at least allow for a relative path resolution 
> starting from the location of the container-executor binary itself. Something 
> like
> {noformat}
> ../etc/hadoop/
> {noformat}
> Thoughts?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (MAPREDUCE-4219) make default container-executor.conf.dir be a path relative to the container-executor binary

2012-05-02 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4219?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-4219:


Attachment: MAPREDUCE-4219.patch.txt

Please take a look at the new patch.

> make default container-executor.conf.dir be a path relative to the 
> container-executor binary
> 
>
> Key: MAPREDUCE-4219
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-4219
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: security
>Affects Versions: 0.23.1
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
> Fix For: 2.0.0
>
> Attachments: HADOOP-8332.patch.txt, MAPREDUCE-4219.patch.txt
>
>
> Currently, container-executor binary has an absolute pathname of its 
> configuration file baked in. This prevents an easy relocation of the 
> configuration files when dealing with multiple Hadoop installs on the same 
> node. It would be nice to at least allow for a relative path resolution 
> starting from the location of the container-executor binary itself. Something 
> like
> {noformat}
> ../etc/hadoop/
> {noformat}
> Thoughts?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] Updated: (MAPREDUCE-2314) configure files that are generated as part of the released tarball need to have executable bit set

2011-02-15 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-2314?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-2314:


  Component/s: build
Affects Version/s: 0.22.0

> configure files that are generated as part of the released tarball need to 
> have executable bit set
> --
>
> Key: MAPREDUCE-2314
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-2314
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: build
>Affects Versions: 0.22.0
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
> Attachments: MAPREDUCE-2314.patch
>
>
> Currently the configure files that are packaged in a tarball are -rw-rw-r--

-- 
This message is automatically generated by JIRA.
-
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] Commented: (MAPREDUCE-2314) configure files that are generated as part of the released tarball need to have executable bit set

2011-02-15 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-2314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12994993#comment-12994993
 ] 

Roman Shaposhnik commented on MAPREDUCE-2314:
-

To answer a chmod question -- I followed the already established pattern of not 
doing chmods in parallel.

> configure files that are generated as part of the released tarball need to 
> have executable bit set
> --
>
> Key: MAPREDUCE-2314
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-2314
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
> Attachments: MAPREDUCE-2314.patch
>
>
> Currently the configure files that are packaged in a tarball are -rw-rw-r--

-- 
This message is automatically generated by JIRA.
-
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] Commented: (MAPREDUCE-2314) configure files that are generated as part of the released tarball need to have executable bit set

2011-02-15 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-2314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12994991#comment-12994991
 ] 

Roman Shaposhnik commented on MAPREDUCE-2314:
-

$ ant veryclean test-patch -Dpatch.file=/home/rvs/MAPREDUCE-2314.patch 
-Dscratch.dir=/tmp/mapred/scratch -Dfindbugs.home=/home/rvs/findbugs-1.3.9 
-Dforrest.home=/home/rvs/src/apache-forrest-0.8
.

 [exec] -1 overall.  
 [exec] 
 [exec] +1 @author.  The patch does not contain any @author tags.
 [exec] 
 [exec] -1 tests included.  The patch doesn't appear to include any new 
or modified tests.
 [exec] Please justify why no new tests are needed 
for this patch.
 [exec] Also please list what manual steps were 
performed to verify this patch.
 [exec] 
 [exec] +1 javadoc.  The javadoc tool did not generate any warning 
messages.
 [exec] 
 [exec] +1 javac.  The applied patch does not increase the total number 
of javac compiler warnings.
 [exec] 
 [exec] +1 findbugs.  The patch does not introduce any new Findbugs 
(version 1.3.9) warnings.
 [exec] 
 [exec] +1 release audit.  The applied patch does not increase the 
total number of release audit warnings.
 [exec] 
 [exec] +1 system test framework.  The patch passed system test 
framework compile.


And given that it is a build change " -1 tests included" can be ignored

> configure files that are generated as part of the released tarball need to 
> have executable bit set
> --
>
> Key: MAPREDUCE-2314
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-2314
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
> Attachments: MAPREDUCE-2314.patch
>
>
> Currently the configure files that are packaged in a tarball are -rw-rw-r--

-- 
This message is automatically generated by JIRA.
-
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] Updated: (MAPREDUCE-2314) configure files that are generated as part of the released tarball need to have executable bit set

2011-02-10 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-2314?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-2314:


Attachment: MAPREDUCE-2314.patch

> configure files that are generated as part of the released tarball need to 
> have executable bit set
> --
>
> Key: MAPREDUCE-2314
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-2314
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
> Attachments: MAPREDUCE-2314.patch
>
>
> Currently the configure files that are packaged in a tarball are -rw-rw-r--

-- 
This message is automatically generated by JIRA.
-
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] Updated: (MAPREDUCE-2314) configure files that are generated as part of the released tarball need to have executable bit set

2011-02-10 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-2314?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-2314:


Status: Patch Available  (was: Open)

> configure files that are generated as part of the released tarball need to 
> have executable bit set
> --
>
> Key: MAPREDUCE-2314
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-2314
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
> Attachments: MAPREDUCE-2314.patch
>
>
> Currently the configure files that are packaged in a tarball are -rw-rw-r--

-- 
This message is automatically generated by JIRA.
-
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] Created: (MAPREDUCE-2314) configure files that are generated as part of the released tarball need to have executable bit set

2011-02-09 Thread Roman Shaposhnik (JIRA)
configure files that are generated as part of the released tarball need to have 
executable bit set
--

 Key: MAPREDUCE-2314
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-2314
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
Reporter: Roman Shaposhnik
Assignee: Roman Shaposhnik


Currently the configure files that are packaged in a tarball are -rw-rw-r--


-- 
This message is automatically generated by JIRA.
-
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] Commented: (MAPREDUCE-2260) Remove auto-generated native build files

2011-01-28 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-2260?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12988287#action_12988287
 ] 

Roman Shaposhnik commented on MAPREDUCE-2260:
-

To finish up my previous answer -- the command used to build hadoop-mapreduce 
was:

ant -Dcompile.native=true -Dcompile.c++=true -Djava5.home=$JAVA5_HOME 
-Dforrest.home=$FORREST_HOME -Dhadoop.conf.dir=/etc/hadoop-0.20/conf 
-Dlibrecordio=true veryclean api-report task-controller compile-c++ tar

Finally, as for testing -- examples (pipes) were executed

> Remove auto-generated native build files
> 
>
> Key: MAPREDUCE-2260
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-2260
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: build
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
> Attachments: MAPREDUCE-2260.diff
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> The repo currently includes the automake and autoconf generated files for the 
> native build. Per discussion on HADOOP-6421 let's remove them and use the 
> host's automake and autoconf. We should also do this for libhdfs and 
> fuse-dfs. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (MAPREDUCE-2260) Remove auto-generated native build files

2011-01-28 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-2260?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12988276#action_12988276
 ] 

Roman Shaposhnik commented on MAPREDUCE-2260:
-

bq. it would be good to verify that ... ant tar produces a tarball which 
contains the generated files (ie a dist style tarball).

Since one has to specify extra properties in order for native bits to be 
included in a dist style tarball here's a list of builds that were done:

cd hadoop-common ; ant -Dcompile.native=true -Dcompile.c++=true 
-Djava5.home=$JAVA5_HOME -Dforrest.home=$FORREST_HOME 
-Dhadoop.conf.dir=/etc/hadoop/conf  veryclean api-report compile-core-native tar
cd hadoop-hdfs ; ant -Dcompile.native=true -Dcompile.c++=true 
-Djava5.home=$JAVA5_HOME -Dforrest.home=$FORREST_HOME 
-Dhadoop.conf.dir=/etc/hadoop/conf -Dlibhdfs=1 veryclean api-report tar
cd hadoop-mapreduce ; 

That resulted in the following native bits being created:

c++/Linux-amd64-64/include/hadoop/Pipes.hh
c++/Linux-amd64-64/include/hadoop/SerialUtils.hh
c++/Linux-amd64-64/include/hadoop/TemplateFactory.hh
c++/Linux-amd64-64/include/hadoop/StringUtils.hh
c++/Linux-amd64-64/lib/libhadooppipes.a
c++/Linux-amd64-64/lib/libhadooputils.a
c++/Linux-amd64-64/lib/libhdfs.a
c++/Linux-amd64-64/lib/libhdfs.so.0.0.0
c++/Linux-amd64-64/lib/libhdfs.so.0
c++/Linux-amd64-64/lib/libhdfs.so
c++/Linux-amd64-64/lib/libhdfs.la
c++/lib/libhdfs.a
c++/lib/libhdfs.so.0.0.0
c++/lib/libhdfs.so.0
c++/lib/libhdfs.so
c++/lib/libhdfs.la
librecordio/librecordio.a
bin/task-controller
lib/native/Linux-amd64-64/libhadoop.so
lib/native/Linux-amd64-64/libhadoop.a
lib/native/Linux-amd64-64/libhadoop.so.1
lib/native/Linux-amd64-64/libhadoop.so.1.0.0
lib/native/Linux-amd64-64/libhadoop.la





> Remove auto-generated native build files
> 
>
> Key: MAPREDUCE-2260
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-2260
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: build
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
> Attachments: MAPREDUCE-2260.diff
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> The repo currently includes the automake and autoconf generated files for the 
> native build. Per discussion on HADOOP-6421 let's remove them and use the 
> host's automake and autoconf. We should also do this for libhdfs and 
> fuse-dfs. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (MAPREDUCE-2260) Remove auto-generated native build files

2011-01-18 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-2260?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-2260:


Attachment: MAPREDUCE-2260.diff

> Remove auto-generated native build files
> 
>
> Key: MAPREDUCE-2260
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-2260
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: build
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
> Attachments: MAPREDUCE-2260.diff
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> The repo currently includes the automake and autoconf generated files for the 
> native build. Per discussion on HADOOP-6421 let's remove them and use the 
> host's automake and autoconf. We should also do this for libhdfs and 
> fuse-dfs. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (MAPREDUCE-2260) Remove auto-generated native build files

2011-01-18 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-2260?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik updated MAPREDUCE-2260:


Status: Patch Available  (was: Open)

I'm attaching a first cut at the patch. Please let me know what do you think 
about the approach taken.

Notes:

1. since none of the C++ projects is imported (they are all developed as part 
of the Hadoop mapreduce workspace)
 it seems that we all would be better off defining a single common umbrella 
C++ project with a single autotools 
 metastructure and 5 different subdirectories corresponding to pipes, 
pipes-examples, task-controller, utils and librecordio.
 We don't really gain anything by splitting C++ sources at the level of Ant.

2. the build (at least the release build) now depends on the entire autoconf


> Remove auto-generated native build files
> 
>
> Key: MAPREDUCE-2260
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-2260
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: build
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> The repo currently includes the automake and autoconf generated files for the 
> native build. Per discussion on HADOOP-6421 let's remove them and use the 
> host's automake and autoconf. We should also do this for libhdfs and 
> fuse-dfs. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Assigned: (MAPREDUCE-2260) Remove auto-generated native build files

2011-01-17 Thread Roman Shaposhnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-2260?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Roman Shaposhnik reassigned MAPREDUCE-2260:
---

Assignee: Roman Shaposhnik

> Remove auto-generated native build files
> 
>
> Key: MAPREDUCE-2260
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-2260
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: build
>Reporter: Roman Shaposhnik
>Assignee: Roman Shaposhnik
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> The repo currently includes the automake and autoconf generated files for the 
> native build. Per discussion on HADOOP-6421 let's remove them and use the 
> host's automake and autoconf. We should also do this for libhdfs and 
> fuse-dfs. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (MAPREDUCE-2260) Remove auto-generated native build files

2011-01-12 Thread Roman Shaposhnik (JIRA)
Remove auto-generated native build files


 Key: MAPREDUCE-2260
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-2260
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: build
Reporter: Roman Shaposhnik


The repo currently includes the automake and autoconf generated files for the 
native build. Per discussion on HADOOP-6421 let's remove them and use the 
host's automake and autoconf. We should also do this for libhdfs and fuse-dfs. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.