[jira] [Commented] (HADOOP-13812) Upgrade Tomcat to 6.0.48

2016-11-29 Thread John Zhuge (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13812?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707816#comment-15707816
 ] 

John Zhuge commented on HADOOP-13812:
-

Thx.

> Upgrade Tomcat to 6.0.48
> 
>
> Key: HADOOP-13812
> URL: https://issues.apache.org/jira/browse/HADOOP-13812
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 2.6.0
>Reporter: John Zhuge
>Assignee: John Zhuge
>Priority: Blocker
> Fix For: 2.8.0, 2.7.4, 3.0.0-alpha2, 2.6.6
>
> Attachments: HADOOP-13812.001.patch, HADOOP-13812.branch-2.6.patch
>
>
> KMS and HttpFS currently uses Tomcat 6.0.44, propose to upgrade to the latest 
> version is 6.0.48.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13812) Upgrade Tomcat to 6.0.48

2016-11-29 Thread Akira Ajisaka (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13812?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707803#comment-15707803
 ] 

Akira Ajisaka commented on HADOOP-13812:


Now branch-2.6 is not maintained for Jenkins precommit job, so there are many 
-1's.

> Upgrade Tomcat to 6.0.48
> 
>
> Key: HADOOP-13812
> URL: https://issues.apache.org/jira/browse/HADOOP-13812
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 2.6.0
>Reporter: John Zhuge
>Assignee: John Zhuge
>Priority: Blocker
> Fix For: 2.8.0, 2.7.4, 3.0.0-alpha2, 2.6.6
>
> Attachments: HADOOP-13812.001.patch, HADOOP-13812.branch-2.6.patch
>
>
> KMS and HttpFS currently uses Tomcat 6.0.44, propose to upgrade to the latest 
> version is 6.0.48.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13848) Missing auth-keys.xml prevents detecting test code build problem

2016-11-29 Thread John Zhuge (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13848?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

John Zhuge updated HADOOP-13848:

Priority: Blocker  (was: Minor)

> Missing auth-keys.xml prevents detecting test code build problem
> 
>
> Key: HADOOP-13848
> URL: https://issues.apache.org/jira/browse/HADOOP-13848
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3, fs/swift, test
>Reporter: John Zhuge
>Assignee: John Zhuge
>Priority: Blocker
>
> Both hadoop-aws and hadoop-openstack require the existence of file 
> {{src/test/resources/auth-keys.xml}} to run the tests. With the design of the 
> pom.xml, the non-existence of auth-keys.xml also prevents building the test 
> code. Unfortunately this leads to delayed detection of build problems in test 
> code, e.g., introduced by a mistake in backports.
> {code}
> 
>   tests-off
>   
> 
>   src/test/resources/auth-keys.xml
> 
>   
>   
> true
>   
> 
> 
>   tests-on
>   
> 
>   src/test/resources/auth-keys.xml
> 
>   
>   
> false
>   
> 
> {code}
> Section {{Skipping by Default}} in 
> http://maven.apache.org/surefire/maven-surefire-plugin/examples/skipping-test.html
>  proposes a solution. Any time you want to run tests, you must do 2 things 
> instead of 1:
> * Copy auth-keys.xml to src/test/resources
> * Run {{mvn install}} with the extra {{-DskipTests=false}}
> Would like the community to weigh in on this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13848) Missing auth-keys.xml prevents detecting test code build problem

2016-11-29 Thread John Zhuge (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13848?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

John Zhuge updated HADOOP-13848:

Description: 
Both hadoop-aws and hadoop-openstack require the existence of file 
{{src/test/resources/auth-keys.xml}} to run the tests. With the design of the 
pom.xml, the non-existence of auth-keys.xml also prevents building the test 
code. Unfortunately this leads to delayed detection of build problems in test 
code, e.g., introduced by a mistake in backports.
{code}

  tests-off
  

  src/test/resources/auth-keys.xml

  
  
true
  


  tests-on
  

  src/test/resources/auth-keys.xml

  
  
false
  

{code}

Section {{Skipping by Default}} in 
http://maven.apache.org/surefire/maven-surefire-plugin/examples/skipping-test.html
 proposes a solution. Any time you want to run tests, you must do 2 things 
instead of 1:
* Copy auth-keys.xml to src/test/resources
* Run {{mvn install}} with the extra {{-DskipTests=false}}

Would like the community to weigh in on this.

  was:
Both hadoop-aws and hadoop-openstack require the existence of file 
{{src/test/resources/auth-keys.xml}} to run the tests. With the design of the 
pom.xml, the non-existence of auth-keys.xml also prevents building the test 
code. Unfortunately this leads to delayed detection of build problems in test 
code, e.g., introduced by a mistake in backports.
{code}

  tests-off
  

  src/test/resources/auth-keys.xml

  
  
true
  


  tests-on
  

  src/test/resources/auth-keys.xml

  
  
false
  

{code}

Section {{Skipping by Default}} in 
http://maven.apache.org/surefire/maven-surefire-plugin/examples/skipping-test.html
 proposes a second solution. Any time you want to run tests, you must do 2 
things instead of 1:
* Copy auth-keys.xml to src/test/resources
* Run {{mvn install}} with the extra {{-DskipTests=false}}

Would like the community to weigh in on this.


> Missing auth-keys.xml prevents detecting test code build problem
> 
>
> Key: HADOOP-13848
> URL: https://issues.apache.org/jira/browse/HADOOP-13848
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3, fs/swift, test
>Reporter: John Zhuge
>Assignee: John Zhuge
>Priority: Minor
>
> Both hadoop-aws and hadoop-openstack require the existence of file 
> {{src/test/resources/auth-keys.xml}} to run the tests. With the design of the 
> pom.xml, the non-existence of auth-keys.xml also prevents building the test 
> code. Unfortunately this leads to delayed detection of build problems in test 
> code, e.g., introduced by a mistake in backports.
> {code}
> 
>   tests-off
>   
> 
>   src/test/resources/auth-keys.xml
> 
>   
>   
> true
>   
> 
> 
>   tests-on
>   
> 
>   src/test/resources/auth-keys.xml
> 
>   
>   
> false
>   
> 
> {code}
> Section {{Skipping by Default}} in 
> http://maven.apache.org/surefire/maven-surefire-plugin/examples/skipping-test.html
>  proposes a solution. Any time you want to run tests, you must do 2 things 
> instead of 1:
> * Copy auth-keys.xml to src/test/resources
> * Run {{mvn install}} with the extra {{-DskipTests=false}}
> Would like the community to weigh in on this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13675) Bug in return value for delete() calls in WASB

2016-11-29 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13675?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707722#comment-15707722
 ] 

Hadoop QA commented on HADOOP-13675:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
14s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
 1s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
17s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
14s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
21s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
13s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
26s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
13s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
15s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
14s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
14s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
0m 11s{color} | {color:orange} hadoop-tools/hadoop-azure: The patch generated 4 
new + 43 unchanged - 0 fixed = 47 total (was 43) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
17s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
 9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
30s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  1m 
22s{color} | {color:green} hadoop-azure in the patch passed. {color} |
| {color:red}-1{color} | {color:red} asflicense {color} | {color:red}  0m 
15s{color} | {color:red} The patch generated 1 ASF License warnings. {color} |
| {color:black}{color} | {color:black} {color} | {color:black} 13m 43s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | HADOOP-13675 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12840990/HADOOP-13675.002.patch
 |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux 2fbda7bed30b 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed 
Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / aeecfa2 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/11163/artifact/patchprocess/diff-checkstyle-hadoop-tools_hadoop-azure.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/11163/testReport/ |
| asflicense | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/11163/artifact/patchprocess/patch-asflicense-problems.txt
 |
| modules | C: hadoop-tools/hadoop-azure U: hadoop-tools/hadoop-azure |
| Console output | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/11163/console |
| Powered by | Apache Yetus 0.4.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> Bug in return value for delete() calls in WASB
> --
>
> Key: HADOOP-13675
> URL: https://issues.apache.org/jira/browse/HADOOP-13675
> Project: Hadoop Common
>  Issue Type: Bug
>  

[jira] [Comment Edited] (HADOOP-13675) Bug in return value for delete() calls in WASB

2016-11-29 Thread Dushyanth (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13675?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707688#comment-15707688
 ] 

Dushyanth edited comment on HADOOP-13675 at 11/30/16 6:17 AM:
--

[~liuml07] Thanks. I will be taking over the responsibility of uploading the 
second patch and addressing any issues for this Jira going forward. I have 
attached Hadoop-13675.002.patch that Jameel had worked on, kindly review the 
patch.


was (Author: dchickabasapa):
[~liuml07] Thanks. I will be taking over the responsibility of uploading the 
second patch and addressing any issues for this Jira going forward. I have 
attached Hadoop-13675.002.patch that Jameel had worked on.

> Bug in return value for delete() calls in WASB
> --
>
> Key: HADOOP-13675
> URL: https://issues.apache.org/jira/browse/HADOOP-13675
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: azure, fs/azure
>Affects Versions: 2.8.0
>Reporter: Dushyanth
>Assignee: Dushyanth
> Fix For: 2.9.0
>
> Attachments: HADOOP-13675.001.patch, HADOOP-13675.002.patch
>
>
> Current implementation of WASB does not correctly handle multiple 
> threads/clients calling delete on the same file. The expected behavior in 
> such scenarios is only one of the thread should delete the file and return 
> true, while all other threads should receive false. However in the current 
> implementation even though only one thread deletes the file, multiple clients 
> incorrectly get "true" as the return from delete() call..



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-13675) Bug in return value for delete() calls in WASB

2016-11-29 Thread Dushyanth (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13675?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707688#comment-15707688
 ] 

Dushyanth edited comment on HADOOP-13675 at 11/30/16 6:17 AM:
--

[~liuml07] Thanks. I will be taking over the responsibility of uploading the 
second patch and addressing any issues for this Jira going forward. I have 
attached Hadoop-13675.002.patch that Jameel had worked on.


was (Author: dchickabasapa):
[~liuml07] Thanks. I will be taking over the responsibility of uploading the 
second patch and addressing any issues for this Jira going forward. I am 
attaching Patch-2 that Jameel had worked on.

> Bug in return value for delete() calls in WASB
> --
>
> Key: HADOOP-13675
> URL: https://issues.apache.org/jira/browse/HADOOP-13675
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: azure, fs/azure
>Affects Versions: 2.8.0
>Reporter: Dushyanth
>Assignee: Dushyanth
> Fix For: 2.9.0
>
> Attachments: HADOOP-13675.001.patch, HADOOP-13675.002.patch
>
>
> Current implementation of WASB does not correctly handle multiple 
> threads/clients calling delete on the same file. The expected behavior in 
> such scenarios is only one of the thread should delete the file and return 
> true, while all other threads should receive false. However in the current 
> implementation even though only one thread deletes the file, multiple clients 
> incorrectly get "true" as the return from delete() call..



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13675) Bug in return value for delete() calls in WASB

2016-11-29 Thread Dushyanth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13675?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dushyanth updated HADOOP-13675:
---
Attachment: HADOOP-13675.002.patch

> Bug in return value for delete() calls in WASB
> --
>
> Key: HADOOP-13675
> URL: https://issues.apache.org/jira/browse/HADOOP-13675
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: azure, fs/azure
>Affects Versions: 2.8.0
>Reporter: Dushyanth
>Assignee: Dushyanth
> Fix For: 2.9.0
>
> Attachments: HADOOP-13675.001.patch, HADOOP-13675.002.patch
>
>
> Current implementation of WASB does not correctly handle multiple 
> threads/clients calling delete on the same file. The expected behavior in 
> such scenarios is only one of the thread should delete the file and return 
> true, while all other threads should receive false. However in the current 
> implementation even though only one thread deletes the file, multiple clients 
> incorrectly get "true" as the return from delete() call..



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13675) Bug in return value for delete() calls in WASB

2016-11-29 Thread Dushyanth (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13675?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707688#comment-15707688
 ] 

Dushyanth commented on HADOOP-13675:


[~liuml07] Thanks. I will be taking over the responsibility of uploading the 
second patch and addressing any issues for this Jira going forward. I am 
attaching Patch-2 that Jameel had worked on.

> Bug in return value for delete() calls in WASB
> --
>
> Key: HADOOP-13675
> URL: https://issues.apache.org/jira/browse/HADOOP-13675
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: azure, fs/azure
>Affects Versions: 2.8.0
>Reporter: Dushyanth
>Assignee: Dushyanth
> Fix For: 2.9.0
>
> Attachments: HADOOP-13675.001.patch
>
>
> Current implementation of WASB does not correctly handle multiple 
> threads/clients calling delete on the same file. The expected behavior in 
> such scenarios is only one of the thread should delete the file and return 
> true, while all other threads should receive false. However in the current 
> implementation even though only one thread deletes the file, multiple clients 
> incorrectly get "true" as the return from delete() call..



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-10930) HarFsInputStream should implement PositionedReadable with thead-safe.

2016-11-29 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10930?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707686#comment-15707686
 ] 

Hadoop QA commented on HADOOP-10930:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
15s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red}  0m  
0s{color} | {color:red} The patch doesn't appear to include any new or modified 
tests. Please justify why no new tests are needed for this patch. Also please 
list what manual steps were performed to verify this patch. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
27s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  9m 
43s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
30s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m  
7s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
20s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
34s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
48s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
41s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  9m 
42s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  9m 
42s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
0m 30s{color} | {color:orange} hadoop-common-project/hadoop-common: The patch 
generated 1 new + 42 unchanged - 0 fixed = 43 total (was 42) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 2 line(s) that end in whitespace. Use git 
apply --whitespace=fix <>. Refer https://git-scm.com/docs/git-apply 
{color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m 
45s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
28s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 14m 
13s{color} | {color:green} hadoop-common in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
38s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 55m 30s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | HADOOP-10930 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12659535/HADOOP-10930.001.patch
 |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux e4002e8c7066 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed 
Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / aeecfa2 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/11162/artifact/patchprocess/diff-checkstyle-hadoop-common-project_hadoop-common.txt
 |
| whitespace | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/11162/artifact/patchprocess/whitespace-eol.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/11162/testReport/ |
| modules | C: hadoop-common-project/hadoop-common U: 
hadoop-common-project/hadoop-common |
| Console output | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/11162/console |
| Powered by | Apache Yetus 0.4.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> HarFsInputStream should implement 

[jira] [Assigned] (HADOOP-13675) Bug in return value for delete() calls in WASB

2016-11-29 Thread Dushyanth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13675?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dushyanth reassigned HADOOP-13675:
--

Assignee: Dushyanth  (was: Jameel Naina Mohamed)

> Bug in return value for delete() calls in WASB
> --
>
> Key: HADOOP-13675
> URL: https://issues.apache.org/jira/browse/HADOOP-13675
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: azure, fs/azure
>Affects Versions: 2.8.0
>Reporter: Dushyanth
>Assignee: Dushyanth
> Fix For: 2.9.0
>
> Attachments: HADOOP-13675.001.patch
>
>
> Current implementation of WASB does not correctly handle multiple 
> threads/clients calling delete on the same file. The expected behavior in 
> such scenarios is only one of the thread should delete the file and return 
> true, while all other threads should receive false. However in the current 
> implementation even though only one thread deletes the file, multiple clients 
> incorrectly get "true" as the return from delete() call..



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-13848) Missing auth-keys.xml prevents detecting test code build problem

2016-11-29 Thread John Zhuge (JIRA)
John Zhuge created HADOOP-13848:
---

 Summary: Missing auth-keys.xml prevents detecting test code build 
problem
 Key: HADOOP-13848
 URL: https://issues.apache.org/jira/browse/HADOOP-13848
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs/s3, fs/swift, test
Reporter: John Zhuge
Assignee: John Zhuge
Priority: Minor


Both hadoop-aws and hadoop-openstack require the existence of file 
{{src/test/resources/auth-keys.xml}} to run the tests. With the design of the 
pom.xml, the non-existence of auth-keys.xml also prevents building the test 
code. Unfortunately this leads to delayed detection of build problems in test 
code, e.g., introduced by a mistake in backports.
{code}

  tests-off
  

  src/test/resources/auth-keys.xml

  
  
true
  


  tests-on
  

  src/test/resources/auth-keys.xml

  
  
false
  

{code}

Section {{Skipping by Default}} in 
http://maven.apache.org/surefire/maven-surefire-plugin/examples/skipping-test.html
 proposes a second solution. Any time you want to run tests, you must do 2 
things instead of 1:
* Copy auth-keys.xml to src/test/resources
* Run {{mvn install}} with the extra {{-DskipTests=false}}

Would like the community to weigh in on this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13600) S3a rename() to copy files in a directory in parallel

2016-11-29 Thread Sahil Takiar (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13600?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707670#comment-15707670
 ] 

Sahil Takiar commented on HADOOP-13600:
---

Sounds good Steve. I rebased the PR and created a separate TransferManager just 
for copies. Haven't looked at the testing yet.

> S3a rename() to copy files in a directory in parallel
> -
>
> Key: HADOOP-13600
> URL: https://issues.apache.org/jira/browse/HADOOP-13600
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.7.3
>Reporter: Steve Loughran
>Assignee: Sahil Takiar
>
> Currently a directory rename does a one-by-one copy, making the request 
> O(files * data). If the copy operations were launched in parallel, the 
> duration of the copy may be reducable to the duration of the longest copy. 
> For a directory with many files, this will be significant



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-10930) HarFsInputStream should implement PositionedReadable with thead-safe.

2016-11-29 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10930?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707573#comment-15707573
 ] 

Hudson commented on HADOOP-10930:
-

FAILURE: Integrated in Jenkins build Hadoop-trunk-Commit #10912 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/10912/])
HADOOP-10930. Refactor: Wrap Datanode IO related operations. Contributed (arp: 
rev aeecfa24f4fb6af289920cbf8830c394e66bd78e)
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/SimulatedFSDataset.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/ReplicaOutputStreams.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/LocalReplicaInPipeline.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestSimulatedFSDataset.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsVolumeImpl.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/extdataset/ExternalDatasetImpl.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DNConf.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestFileAppend.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/LocalReplica.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/FsDatasetSpi.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/ReplicaInputStreams.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/extdataset/ExternalReplicaInPipeline.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/BlockPoolSlice.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockRecovery.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataStorage.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/ReplicaInPipeline.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetAsyncDiskService.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockSender.java


> HarFsInputStream should implement PositionedReadable with thead-safe.
> -
>
> Key: HADOOP-10930
> URL: https://issues.apache.org/jira/browse/HADOOP-10930
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 2.6.0
>Reporter: Yi Liu
>Assignee: Yi Liu
>  Labels: BB2015-05-TBR
> Attachments: HADOOP-10930.001.patch
>
>
> {{PositionedReadable}} definition requires the implementations for its 
> interfaces should be thread-safe.
> HarFsInputStream doesn't implement these interfaces with tread-safe, this 
> JIRA is to fix this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13837) Process check bug in hadoop_stop_daemon of hadoop-functions.sh

2016-11-29 Thread Weiwei Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13837?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Weiwei Yang updated HADOOP-13837:
-
Target Version/s: 3.0.0-alpha2

> Process check bug in hadoop_stop_daemon of hadoop-functions.sh
> --
>
> Key: HADOOP-13837
> URL: https://issues.apache.org/jira/browse/HADOOP-13837
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: scripts
>Reporter: Weiwei Yang
>Assignee: Weiwei Yang
> Attachments: HADOOP-13837.01.patch, HADOOP-13837.02.patch, 
> HADOOP-13837.03.patch, HADOOP-13837.04.patch, check_proc.sh
>
>
> Always get {{ERROR: Unable to kill ...}} after {{Trying to kill with kill 
> -9}}, see following output of stop-yarn.sh
> {code}
> : WARNING: nodemanager did not stop gracefully after 5 seconds: 
> Trying to kill with kill -9
> : ERROR: Unable to kill 18097
> {code}
> hadoop_stop_daemon doesn't check process liveness correctly, this bug can be 
> reproduced by the script easily. kill -9 would need some time to be done, 
> directly check process existence right after mostly will fail.
> {code}
> function hadoop_stop_daemon
> {
> ...
>   kill -9 "${pid}" >/dev/null 2>&1
> fi
> if ps -p "${pid}" > /dev/null 2>&1; then
>   hadoop_error "ERROR: Unable to kill ${pid}"
> else
>   ...
> }
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-13597) Switch KMS from Tomcat to Jetty

2016-11-29 Thread John Zhuge (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13597?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15704286#comment-15704286
 ] 

John Zhuge edited comment on HADOOP-13597 at 11/30/16 2:58 AM:
---

Map {{sbin/kms.sh}} to {{bin/hadoop kms}}:
| kms.sh run | hadoop kms |
| kms.sh start | hadoop --daemon start kms |
| kms.sh stop | hadoop --daemon stop kms |
| kms.sh status | hadoop --daemon status kms |




was (Author: jzhuge):
Map {{sbin/kms.sh}} to {{bin/hadoop kms}}:
| kms.sh run | hadoop kms |
| kms.sh start | hadoop kms --daemon start |
| kms.sh stop | hadoop kms --daemon stop |
| kms.sh status | hadoop kms --daemon status |



> Switch KMS from Tomcat to Jetty
> ---
>
> Key: HADOOP-13597
> URL: https://issues.apache.org/jira/browse/HADOOP-13597
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: kms
>Affects Versions: 2.6.0
>Reporter: John Zhuge
>Assignee: John Zhuge
> Attachments: HADOOP-13597.001.patch
>
>
> The Tomcat 6 we are using will reach EOL at the end of 2017. While there are 
> other good options, I would propose switching to {{Jetty 9}} for the 
> following reasons:
> * Easier migration. Both Tomcat and Jetty are based on {{Servlet 
> Containers}}, so we don't have change client code that much. It would require 
> more work to switch to {{JAX-RS}}.
> * Well established.
> * Good performance and scalability.
> Other alternatives:
> * Jersey + Grizzly
> * Tomcat 8
> Your opinions will be greatly appreciated.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Assigned] (HADOOP-13847) KMSWebApp should close KeyProviderCryptoExtension

2016-11-29 Thread John Zhuge (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13847?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

John Zhuge reassigned HADOOP-13847:
---

Assignee: John Zhuge

> KMSWebApp should close KeyProviderCryptoExtension
> -
>
> Key: HADOOP-13847
> URL: https://issues.apache.org/jira/browse/HADOOP-13847
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Reporter: Anthony Young-Garner
>Assignee: John Zhuge
>
> KeyProviderCryptoExtension should be closed so that all KeyProviders are also 
> closed. See related HADOOP-13838.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13449) S3Guard: Implement DynamoDBMetadataStore.

2016-11-29 Thread Aaron Fabbri (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13449?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707310#comment-15707310
 ] 

Aaron Fabbri commented on HADOOP-13449:
---

This fixes that failed unit test for me:

{code}
--- 
a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/MockS3ClientFactory.java
+++ 
b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/MockS3ClientFactory.java
@@ -41,9 +41,8 @@
   @Override
   public AmazonDynamoDBClient createDynamoDBClient(
   URI uri, com.amazonaws.regions.Region region) throws IOException {
-final DefaultS3ClientFactory factory = new DefaultS3ClientFactory();
-factory.setConf(getConf());
-return factory.createDynamoDBClient(uri, region);
+throw new IOException("Purposely failing to create DynamoDB client"
+  + " for unit test.");
   }
{code}

Also noticed a spot we need to fix the exception thrown (supposed to be an 
IOException):

{code}
@Override
public AmazonDynamoDBClient createDynamoDBClient(URI fsUri, Region region)
throws IOException {
{code}
...
{code}
  String msg = "Incorrect DynamoDB endpoint: "  + endPoint;
  LOG.error(msg, e);
  throw new IllegalArgumentException(msg, e);
}
  }
{code}

I have a number of integration test failures I'll be working through next.  
BTW, I'm happy to submit a follow-up (v11) patch with these things if that 
would help, just shout.

> S3Guard: Implement DynamoDBMetadataStore.
> -
>
> Key: HADOOP-13449
> URL: https://issues.apache.org/jira/browse/HADOOP-13449
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Reporter: Chris Nauroth
>Assignee: Mingliang Liu
> Attachments: HADOOP-13449-HADOOP-13345.000.patch, 
> HADOOP-13449-HADOOP-13345.001.patch, HADOOP-13449-HADOOP-13345.002.patch, 
> HADOOP-13449-HADOOP-13345.003.patch, HADOOP-13449-HADOOP-13345.004.patch, 
> HADOOP-13449-HADOOP-13345.005.patch, HADOOP-13449-HADOOP-13345.006.patch, 
> HADOOP-13449-HADOOP-13345.007.patch, HADOOP-13449-HADOOP-13345.008.patch, 
> HADOOP-13449-HADOOP-13345.009.patch, HADOOP-13449-HADOOP-13345.010.patch
>
>
> Provide an implementation of the metadata store backed by DynamoDB.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13449) S3Guard: Implement DynamoDBMetadataStore.

2016-11-29 Thread Aaron Fabbri (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13449?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707232#comment-15707232
 ] 

Aaron Fabbri commented on HADOOP-13449:
---

Congrats on the clean jenkins run [~liuml07].

I'm running some tests with this configured as the MetadataStore today.  First 
thing I'm noticing is a failure in {{TestS3AGetFileStatus}}.  This led me to 
this part of the v10 patch:

{code}
+public class MockS3ClientFactory extends Configured implements S3ClientFactory 
{
+
+  @Override
+  public AmazonDynamoDBClient createDynamoDBClient(
+  URI uri, com.amazonaws.regions.Region region) throws IOException {
+final DefaultS3ClientFactory factory = new DefaultS3ClientFactory();
+factory.setConf(getConf());
+return factory.createDynamoDBClient(uri, region);
+  }
{code}

I believe the goal of the mock s3 client is to be able to run non-integration 
(unit) tests without S3 configured.  It looks like you are creating an actual 
S3 client from the mock client.  Doesn't this break the ability of unit tests 
to run without S3?

It seems like all the unit tests should only use MetadataStores which can run 
locally (Null or LocalMetadataStore).  So, maybe we do not need this code at 
all.  Maybe MockS3ClientFactory#createDynamoDBClient() just throws a runtime 
exception "Failing to create DynamoDB for unit test", and then we fall back to 
the NullMetadataStore automatically in S3Guard#getMetadataStore()?

I'm also wondering if, instead of having S3ClientFactory expose a 
createDynamoDBClient() method, we should just add getters to S3ClientFactory 
(getAwsConfig() and maybe getCredentials()), and then move 
createDynamoDBClient() to inside the DynamoDBMetadataStore? The 
DynamoDBMetadataStore can then call the getters on the client to get what it 
needs to construct a DynamoDB client.  The goal here would be to keep dynamodb 
specifics encapsulated in DynamoDBMetadataStore.  This would allow, for 
example, removing the Dynamo dependency from s3a if we ever want to create a 
separate submodule for the DynamoDBMetadataStore.




> S3Guard: Implement DynamoDBMetadataStore.
> -
>
> Key: HADOOP-13449
> URL: https://issues.apache.org/jira/browse/HADOOP-13449
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Reporter: Chris Nauroth
>Assignee: Mingliang Liu
> Attachments: HADOOP-13449-HADOOP-13345.000.patch, 
> HADOOP-13449-HADOOP-13345.001.patch, HADOOP-13449-HADOOP-13345.002.patch, 
> HADOOP-13449-HADOOP-13345.003.patch, HADOOP-13449-HADOOP-13345.004.patch, 
> HADOOP-13449-HADOOP-13345.005.patch, HADOOP-13449-HADOOP-13345.006.patch, 
> HADOOP-13449-HADOOP-13345.007.patch, HADOOP-13449-HADOOP-13345.008.patch, 
> HADOOP-13449-HADOOP-13345.009.patch, HADOOP-13449-HADOOP-13345.010.patch
>
>
> Provide an implementation of the metadata store backed by DynamoDB.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13818) TestLocalFileSystem#testSetTimes fails

2016-11-29 Thread Yiqun Lin (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13818?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707194#comment-15707194
 ] 

Yiqun Lin commented on HADOOP-13818:


The Jenkins result seems clean. I agree with the opinion that we should 
minimise use of the call {{getFileStatus}} since these calls as potentially 
very expensive in some FS. What do you think, [~ajisakaa]?

> TestLocalFileSystem#testSetTimes fails
> --
>
> Key: HADOOP-13818
> URL: https://issues.apache.org/jira/browse/HADOOP-13818
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: test
> Environment: Mac OS Sierra, OpenJDK 8u122-ea
>Reporter: Akira Ajisaka
>Assignee: Yiqun Lin
>Priority: Minor
> Attachments: HADOOP-13818.001.patch, HADOOP-13818.002.patch, 
> HADOOP-13818.003.patch
>
>
> {noformat}
> Running org.apache.hadoop.fs.TestLocalFileSystem
> Tests run: 20, Failures: 1, Errors: 0, Skipped: 1, Time elapsed: 4.887 sec 
> <<< FAILURE! - in org.apache.hadoop.fs.TestLocalFileSystem
> testSetTimes(org.apache.hadoop.fs.TestLocalFileSystem)  Time elapsed: 0.084 
> sec  <<< FAILURE!
> java.lang.AssertionError: expected:<23456000> but was:<1479176144000>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.fs.TestLocalFileSystem.checkTimesStatus(TestLocalFileSystem.java:391)
>   at 
> org.apache.hadoop.fs.TestLocalFileSystem.testSetTimes(TestLocalFileSystem.java:414)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-12329) io.file.buffer.size is only for SequenceFiles

2016-11-29 Thread Andres Perez (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12329?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707138#comment-15707138
 ] 

Andres Perez commented on HADOOP-12329:
---

I think this can be easily be solved by removing the first setence and 
reorganizing a little bit the description to look like this:

{code:xml}
Determines how much data is buffered during read and write 
operations. The size of this buffer should probably be a multiple of 
hardware page size (4096 on Intel x86).
{code}



> io.file.buffer.size is only for SequenceFiles
> -
>
> Key: HADOOP-12329
> URL: https://issues.apache.org/jira/browse/HADOOP-12329
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: documentation
>Affects Versions: 2.7.1
>Reporter: Kun Yan
>Priority: Trivial
>  Labels: documentation
>
> the core-site.xml  io.file.buffer.size  description:Size of read/write buffer 
> used in SequenceFiles.
> This parameter is only for SequenceFiles? I search for the other one issue as 
> that is not true? If not only affect SequenceFiles could let users to ignore 
> this parameter.URL 
> location:http://hadoop.apache.org/docs/r2.7.1/hadoop-project-dist/hadoop-common/ClusterSetup.html



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13597) Switch KMS from Tomcat to Jetty

2016-11-29 Thread Xiao Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13597?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707117#comment-15707117
 ] 

Xiao Chen commented on HADOOP-13597:


Thanks for working on this [~jzhuge].

Some review comments aside from your todos:
- Suggest to name the new config key names _KEY, to be more consistent with 
HDFS/Common way.
- Nice we're loading envvars for compatibility. It'll make upgrades less 
painful. OTOH, could you add some text in the {{Loaded env}} logs to indicate 
this is a deprecated way, and user should user configuration files instead?
- {{kms-default.xml}} has some duplicate key names 
(hadoop.kms.authentication.signer.secret.provider.zookeeper.XXX).
- This exists before your patch, but good to fix along: kms-default.xml has 
some escaped chars () in configuration.
- {{hadoop.kms.authentication.signer.secret.provider.zookeeper.auth.type}} 
defaults to {{none}} before. We should keep that unchanged. The {{kerberos}} 
default value also doesn't work, it should be {{sasl}}
- We should add basic testing to {{KMSHttpServer}}
- Had a quick run, kms.log seems record every request/response. By default this 
shouldn't be logged.
- After clean install kms.log appears to be at TRACE level. Suggest to make it 
INFO.

> Switch KMS from Tomcat to Jetty
> ---
>
> Key: HADOOP-13597
> URL: https://issues.apache.org/jira/browse/HADOOP-13597
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: kms
>Affects Versions: 2.6.0
>Reporter: John Zhuge
>Assignee: John Zhuge
> Attachments: HADOOP-13597.001.patch
>
>
> The Tomcat 6 we are using will reach EOL at the end of 2017. While there are 
> other good options, I would propose switching to {{Jetty 9}} for the 
> following reasons:
> * Easier migration. Both Tomcat and Jetty are based on {{Servlet 
> Containers}}, so we don't have change client code that much. It would require 
> more work to switch to {{JAX-RS}}.
> * Well established.
> * Good performance and scalability.
> Other alternatives:
> * Jersey + Grizzly
> * Tomcat 8
> Your opinions will be greatly appreciated.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13449) S3Guard: Implement DynamoDBMetadataStore.

2016-11-29 Thread Mingliang Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13449?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15707088#comment-15707088
 ] 

Mingliang Liu commented on HADOOP-13449:


I merged again from the {{trunk}} locally and resolved minor conflicts with 
[HADOOP-13823]. If you guys support the merge, I'll push it to the remote 
feature branch.

> S3Guard: Implement DynamoDBMetadataStore.
> -
>
> Key: HADOOP-13449
> URL: https://issues.apache.org/jira/browse/HADOOP-13449
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Reporter: Chris Nauroth
>Assignee: Mingliang Liu
> Attachments: HADOOP-13449-HADOOP-13345.000.patch, 
> HADOOP-13449-HADOOP-13345.001.patch, HADOOP-13449-HADOOP-13345.002.patch, 
> HADOOP-13449-HADOOP-13345.003.patch, HADOOP-13449-HADOOP-13345.004.patch, 
> HADOOP-13449-HADOOP-13345.005.patch, HADOOP-13449-HADOOP-13345.006.patch, 
> HADOOP-13449-HADOOP-13345.007.patch, HADOOP-13449-HADOOP-13345.008.patch, 
> HADOOP-13449-HADOOP-13345.009.patch, HADOOP-13449-HADOOP-13345.010.patch
>
>
> Provide an implementation of the metadata store backed by DynamoDB.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13675) Bug in return value for delete() calls in WASB

2016-11-29 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13675?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu updated HADOOP-13675:
---
Assignee: (was: Dushyanth)

> Bug in return value for delete() calls in WASB
> --
>
> Key: HADOOP-13675
> URL: https://issues.apache.org/jira/browse/HADOOP-13675
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: azure, fs/azure
>Affects Versions: 2.8.0
>Reporter: Dushyanth
> Fix For: 2.9.0
>
> Attachments: HADOOP-13675.001.patch
>
>
> Current implementation of WASB does not correctly handle multiple 
> threads/clients calling delete on the same file. The expected behavior in 
> such scenarios is only one of the thread should delete the file and return 
> true, while all other threads should receive false. However in the current 
> implementation even though only one thread deletes the file, multiple clients 
> incorrectly get "true" as the return from delete() call..



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13675) Bug in return value for delete() calls in WASB

2016-11-29 Thread Mingliang Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13675?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706925#comment-15706925
 ] 

Mingliang Liu commented on HADOOP-13675:


Hi [~jameeln], I have added you to the Hadoop Contributor on this site; and 
assigned this JIRA to you. [~dchickabasapa] please feel free to re-assign to 
yourself after Jameel uploads the new patch.

> Bug in return value for delete() calls in WASB
> --
>
> Key: HADOOP-13675
> URL: https://issues.apache.org/jira/browse/HADOOP-13675
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: azure, fs/azure
>Affects Versions: 2.8.0
>Reporter: Dushyanth
>Assignee: Jameel Naina Mohamed
> Fix For: 2.9.0
>
> Attachments: HADOOP-13675.001.patch
>
>
> Current implementation of WASB does not correctly handle multiple 
> threads/clients calling delete on the same file. The expected behavior in 
> such scenarios is only one of the thread should delete the file and return 
> true, while all other threads should receive false. However in the current 
> implementation even though only one thread deletes the file, multiple clients 
> incorrectly get "true" as the return from delete() call..



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13675) Bug in return value for delete() calls in WASB

2016-11-29 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13675?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu updated HADOOP-13675:
---
Assignee: Jameel Naina Mohamed

> Bug in return value for delete() calls in WASB
> --
>
> Key: HADOOP-13675
> URL: https://issues.apache.org/jira/browse/HADOOP-13675
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: azure, fs/azure
>Affects Versions: 2.8.0
>Reporter: Dushyanth
>Assignee: Jameel Naina Mohamed
> Fix For: 2.9.0
>
> Attachments: HADOOP-13675.001.patch
>
>
> Current implementation of WASB does not correctly handle multiple 
> threads/clients calling delete on the same file. The expected behavior in 
> such scenarios is only one of the thread should delete the file and return 
> true, while all other threads should receive false. However in the current 
> implementation even though only one thread deletes the file, multiple clients 
> incorrectly get "true" as the return from delete() call..



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13824) FsShell can suppress the real error if no error message is present

2016-11-29 Thread Wei-Chiu Chuang (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13824?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706668#comment-15706668
 ] 

Wei-Chiu Chuang commented on HADOOP-13824:
--

[~jzhuge] thanks for the patch. I reviewed it and have just spotted two very 
tiny nits. Both are subjective to my personal preference:
* {{StringContains#containsString}} may be replaced by 
{{{GenericTestUtils#assertMatches}}.
* I feel {{SystemErrCapture}} should go into {{GenericTestUtils}}. There's a 
{{LogCapturer}} class in {{GenericTestUtils}} serving the similar purpose for 
logs. But that's purely my personal preference.

> FsShell can suppress the real error if no error message is present
> --
>
> Key: HADOOP-13824
> URL: https://issues.apache.org/jira/browse/HADOOP-13824
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 2.7.1, 2.7.3
>Reporter: Rob Vesse
>Assignee: John Zhuge
>  Labels: supportability
> Attachments: HADOOP-13824.001.patch, HADOOP-13824.002.patch
>
>
> The {{FsShell}} error handling assumes in {{displayError()}} that the 
> {{message}} argument is not {{null}}. However in the case where it is this 
> leads to a NPE which results in suppressing the actual error information 
> since a higher level of error handling kicks in and just dumps the stack 
> trace of the NPE instead.
> e.g.
> {noformat}
> Exception in thread "main" java.lang.NullPointerException
>   at org.apache.hadoop.fs.FsShell.displayError(FsShell.java:304)
>   at org.apache.hadoop.fs.FsShell.run(FsShell.java:289)
>   at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>   at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>   at org.apache.hadoop.fs.FsShell.main(FsShell.java:340)
> {noformat}
> This is deeply unhelpful because depending on what the underlying error was 
> there may be no stack dumped/logged for it (as HADOOP-7114 provides) since 
> {{FsShell}} doesn't explicitly dump traces for {{IllegalArgumentException}} 
> which appears to be the underlying cause of my issue.  Line 289 is where 
> {{displayError()}} is called for {{IllegalArgumentException}} handling and 
> that catch clause does not log the error.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13846) S3A to implement rename(final Path src, final Path dst, final Rename... options)

2016-11-29 Thread Mingliang Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706483#comment-15706483
 ] 

Mingliang Liu commented on HADOOP-13846:


This proposal makes a lot of sense. My only concern is the FS#rename/3 is 
marked deprecated, and users may think we don't support this in any future.

> S3A to implement rename(final Path src, final Path dst, final Rename... 
> options)
> 
>
> Key: HADOOP-13846
> URL: https://issues.apache.org/jira/browse/HADOOP-13846
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
>Reporter: Steve Loughran
>
> S3a now raises exceptions on invalid rename operations, but these get lost. I 
> plan to use them in my s3guard committer HADOOP-13786.
> Rather than just make innerRename() private, S3A could implement 
> {{FileSystem.rename(final Path src, final Path dst, final Rename... 
> options)}} and so have an exception-raising rename which can be called 
> without going more into the internals. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-13847) KMSWebApp should close KeyProviderCryptoExtension

2016-11-29 Thread Anthony Young-Garner (JIRA)
Anthony Young-Garner created HADOOP-13847:
-

 Summary: KMSWebApp should close KeyProviderCryptoExtension
 Key: HADOOP-13847
 URL: https://issues.apache.org/jira/browse/HADOOP-13847
 Project: Hadoop Common
  Issue Type: Bug
  Components: kms
Reporter: Anthony Young-Garner


KeyProviderCryptoExtension should be closed so that all KeyProviders are also 
closed. See related HADOOP-13838.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13840) Implement getUsed() for ViewFileSystem

2016-11-29 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706437#comment-15706437
 ] 

Hadoop QA commented on HADOOP-13840:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
16s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  6m 
55s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 10m 
41s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
30s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
10s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
16s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
29s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
45s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
42s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 10m 
37s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 10m 
37s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
30s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m  
1s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
19s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
38s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} javadoc {color} | {color:red}  0m 
49s{color} | {color:red} hadoop-common-project_hadoop-common generated 3 new + 
0 unchanged - 0 fixed = 3 total (was 0) {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  8m 
34s{color} | {color:green} hadoop-common in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
32s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 48m 27s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | HADOOP-13840 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12840911/HADOOP-13840.02.patch 
|
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux 3b1ea88171bc 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed 
Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / b407d53 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| javadoc | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/11161/artifact/patchprocess/diff-javadoc-javadoc-hadoop-common-project_hadoop-common.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/11161/testReport/ |
| modules | C: hadoop-common-project/hadoop-common U: 
hadoop-common-project/hadoop-common |
| Console output | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/11161/console |
| Powered by | Apache Yetus 0.4.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> Implement getUsed() for ViewFileSystem
> --
>
> Key: HADOOP-13840
> URL: https://issues.apache.org/jira/browse/HADOOP-13840
> Project: Hadoop Common
>  Issue Type: Task
>  Components: viewfs
>Affects Versions: 3.0.0-alpha1
>Reporter: Manoj Govindassamy
>Assignee: 

[jira] [Commented] (HADOOP-13600) S3a rename() to copy files in a directory in parallel

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13600?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706370#comment-15706370
 ] 

Steve Loughran commented on HADOOP-13600:
-

sahil, unless you object, I'm assigning this to you; I'll help with review & 
test. I do like the idea of fault injection via a new s3 client though.

BTW, filed HADOOP-13846; implementing the {{FileSystem.rename(final Path src, 
final Path dst, final Rename... options)}} method. That can come after; I'm 
just thinking it may be the way to expose exception raising in rename() 
operations.

> S3a rename() to copy files in a directory in parallel
> -
>
> Key: HADOOP-13600
> URL: https://issues.apache.org/jira/browse/HADOOP-13600
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.7.3
>Reporter: Steve Loughran
>Assignee: Sahil Takiar
>
> Currently a directory rename does a one-by-one copy, making the request 
> O(files * data). If the copy operations were launched in parallel, the 
> duration of the copy may be reducable to the duration of the longest copy. 
> For a directory with many files, this will be significant



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13846) S3A to implement rename(final Path src, final Path dst, final Rename... options)

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706366#comment-15706366
 ] 

Steve Loughran commented on HADOOP-13846:
-

Requirements

# rename/3 behaviour to be specified fully. It's better than rename/2 here. 
# contract tests
# S3a to implement with contract tests
# have rename/2 forward to the new rename, catching IOEs and downgrading to 
return codes, the way it now does. Are behaviours consistent enough to do that?

> S3A to implement rename(final Path src, final Path dst, final Rename... 
> options)
> 
>
> Key: HADOOP-13846
> URL: https://issues.apache.org/jira/browse/HADOOP-13846
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
>Reporter: Steve Loughran
>
> S3a now raises exceptions on invalid rename operations, but these get lost. I 
> plan to use them in my s3guard committer HADOOP-13786.
> Rather than just make innerRename() private, S3A could implement 
> {{FileSystem.rename(final Path src, final Path dst, final Rename... 
> options)}} and so have an exception-raising rename which can be called 
> without going more into the internals. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-13846) S3A to implement rename(final Path src, final Path dst, final Rename... options)

2016-11-29 Thread Steve Loughran (JIRA)
Steve Loughran created HADOOP-13846:
---

 Summary: S3A to implement rename(final Path src, final Path dst, 
final Rename... options)
 Key: HADOOP-13846
 URL: https://issues.apache.org/jira/browse/HADOOP-13846
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 2.8.0
Reporter: Steve Loughran


S3a now raises exceptions on invalid rename operations, but these get lost. I 
plan to use them in my s3guard committer HADOOP-13786.

Rather than just make innerRename() private, S3A could implement 
{{FileSystem.rename(final Path src, final Path dst, final Rename... options)}} 
and so have an exception-raising rename which can be called without going more 
into the internals. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-13786) add output committer which uses s3guard for consistent O(1) commits to S3

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13786?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706351#comment-15706351
 ] 

Steve Loughran edited comment on HADOOP-13786 at 11/29/16 7:51 PM:
---

For the curious, my WiP, in the s3guard branch, is here: 
https://github.com/steveloughran/hadoop/tree/s3guard/HADOOP-13786-committer

I've been tweaking the mapreduce code to add ability to switch to a subclass of 
FileOutputCommitter in FileOutputFormat (And transitively, all children), by 
way of a factory. The S3a factory dynamically chooses the committer based on 
the destination FS. There's currently no difference in the codebase apart from 
logging of operation durations. This means that we can switch the committer 
behind all output formats to using the s3a one (and also allowing anyone else 
to do the same with their own FOF committer subclass). 

I can see that ParquetOutputFormat has its own committer, 
{{ParquetOutputCommitter}}, as used by {{ParquetOutputFormat}} and so, I 
believe spark's {{dataframe.parquet.write()}} pipeline; my IDE isn't 
highlighting much else.

Now, could I get away with just modifying the base FOF committer?


for exception reporting it could use {{FileSystem.rename(final Path src, final 
Path dst, final Rename... options)}}; S3A to impl that with exceptions. That's 
the "transitory" method for use between FS and FileContext, which does raise 
exceptions; the default is non-atomic and eventually gets to {{rename(src, 
dest)}}, except for HDFS which does a full implementation. But: what if the 
semantics of that rename() (not yet in the FS spec, AFAIK) are different from 
what callers expect in some of the corner cases? And so commits everywhere 
break? That would not be good. And it would remove the ability to use any 
private s3a/s3guard calls if we wanted some (e.g. get a lock on a directory), 
unless they went in to the standard FS APIs.

Making the output factory pluggable would avoid such problems, though logging 
duration might be nice all round. That said, given the time to rename, its 
somewhat unimportant when using real filesystems.





was (Author: ste...@apache.org):
For the curious, my WiP, in the s3guard branch, is here: 
https://github.com/steveloughran/hadoop/tree/s3guard/HADOOP-13786-committer

I've been tweaking the mapreduce code to add ability to switch to a subclass of 
FileOutputCommitter in FileOutputFormat (And transitively, all children), by 
way of a factory. The S3a factory dynamically chooses the committer based on 
the destination FS. There's currently no difference in the codebase apart from 
logging of operation durations. This means that we can switch the committer 
behind all output formats to using the s3a one (and also allowing anyone else 
to do the same with their own FOF committer subclass). 

I can see that ParquetOutputFormat has its own committer, 
{{ParquetOutputCommitter}}, as used by {{ParquetOutputFormat}} and so, I 
believe spark's {{dataframe.parquet.write()}} pipeline; my IDE isn't 
highlighting much else.

Now, could I get away with just modifying the base FOF committer?


for exception reporting it could use {{FileSystem.rename(final Path src, final 
Path dst, final Rename... options)}}; S3A to impl that with exceptions. That's 
the "transitory" method for use between FS and FileContext, which does raise 
exceptions; the default is non-atomic and eventually gets to {{rename(src, 
dest)}}, except for HDFS which does a full implementation. But: what if the 
semantics of that rename() (not yet in the FS spec, AFAIK) are different from 
what callers expect in some of the corner cases? And so commits everywhere 
break? That would not be good. And it would remove the ability to use any 
private s3a/s3guard calls if we wanted some (e.g. get a lock on a directory), 
unless they went in to the standard FS APIs.

Making the output factory pluggable would avoid such problems, though logging 
duration might be nice all round. That said, given the time to rename, its 
somewhat unimportant when using real filesystems.




> add output committer which uses s3guard for consistent O(1) commits to S3
> -
>
> Key: HADOOP-13786
> URL: https://issues.apache.org/jira/browse/HADOOP-13786
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-alpha2
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>
> A goal of this code is "support O(1) commits to S3 repositories in the 
> presence of failures". Implement it, including whatever is needed to 
> demonstrate the correctness of the algorithm. (that is, assuming that s3guard 
> provides a consistent view of the presence/absence of blobs, show that we can 
> commit directly).
> I consider ourselves free to 

[jira] [Commented] (HADOOP-13786) add output committer which uses s3guard for consistent O(1) commits to S3

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13786?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706351#comment-15706351
 ] 

Steve Loughran commented on HADOOP-13786:
-

For the curious, my WiP, in the s3guard branch, is here: 
https://github.com/steveloughran/hadoop/tree/s3guard/HADOOP-13786-committer

I've been tweaking the mapreduce code to add ability to switch to a subclass of 
FileOutputCommitter in FileOutputFormat (And transitively, all children), by 
way of a factory. The S3a factory dynamically chooses the committer based on 
the destination FS. There's currently no difference in the codebase apart from 
logging of operation durations. This means that we can switch the committer 
behind all output formats to using the s3a one (and also allowing anyone else 
to do the same with their own FOF committer subclass). 

I can see that ParquetOutputFormat has its own committer, 
{{ParquetOutputCommitter}}, as used by {{ParquetOutputFormat}} and so, I 
believe spark's {{dataframe.parquet.write()}} pipeline; my IDE isn't 
highlighting much else.

Now, could I get away with just modifying the base FOF committer?


for exception reporting it could use {{FileSystem.rename(final Path src, final 
Path dst, final Rename... options)}}; S3A to impl that with exceptions. That's 
the "transitory" method for use between FS and FileContext, which does raise 
exceptions; the default is non-atomic and eventually gets to {{rename(src, 
dest)}}, except for HDFS which does a full implementation. But: what if the 
semantics of that rename() (not yet in the FS spec, AFAIK) are different from 
what callers expect in some of the corner cases? And so commits everywhere 
break? That would not be good. And it would remove the ability to use any 
private s3a/s3guard calls if we wanted some (e.g. get a lock on a directory), 
unless they went in to the standard FS APIs.

Making the output factory pluggable would avoid such problems, though logging 
duration might be nice all round. That said, given the time to rename, its 
somewhat unimportant when using real filesystems.




> add output committer which uses s3guard for consistent O(1) commits to S3
> -
>
> Key: HADOOP-13786
> URL: https://issues.apache.org/jira/browse/HADOOP-13786
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-alpha2
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>
> A goal of this code is "support O(1) commits to S3 repositories in the 
> presence of failures". Implement it, including whatever is needed to 
> demonstrate the correctness of the algorithm. (that is, assuming that s3guard 
> provides a consistent view of the presence/absence of blobs, show that we can 
> commit directly).
> I consider ourselves free to expose the blobstore-ness of the s3 output 
> streams (ie. not visible until the close()), if we need to use that to allow 
> us to abort commit operations.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13794) JSON.org license is now CatX

2016-11-29 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13794?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HADOOP-13794:
-
Affects Version/s: (was: 3.0.0-alpha2)

I'm going to remove the 3.0.0-alpha2 affects version since the AWS and jackson 
deps have been upgraded in trunk

This is also just a stay of execution. We can keep spinning out maintenance 
releases until April, but we really need to get 2.8 (and 2.9?) out so we can 
retire the 2.6 and 2.7 lines.

> JSON.org license is now CatX
> 
>
> Key: HADOOP-13794
> URL: https://issues.apache.org/jira/browse/HADOOP-13794
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.8.0, 2.7.4, 2.6.6
>Reporter: Sean Busbey
>Priority: Blocker
>
> per [update resolved legal|http://www.apache.org/legal/resolved.html#json]:
> {quote}
> CAN APACHE PRODUCTS INCLUDE WORKS LICENSED UNDER THE JSON LICENSE?
> No. As of 2016-11-03 this has been moved to the 'Category X' license list. 
> Prior to this, use of the JSON Java library was allowed. See Debian's page 
> for a list of alternatives.
> {quote}
> We have a test-time transitive dependency on the {{org.json:json}} artifact 
> in trunk and branch-2. AFAICT, this test time dependency doesn't get exposed 
> to downstream at all (I checked assemblies and test-jar artifacts we publish 
> to maven), so it can be removed or kept at our leisure. keeping it risks it 
> being promoted out of test scope by maven without us noticing. We might be 
> able to add an enforcer rule to check for this.
> We also distribute it in bundled form through our use of the AWS Java SDK 
> artifacts in trunk and branch-2. Looking at the github project, [their 
> dependency on JSON.org was removed in 
> 1.11|https://github.com/aws/aws-sdk-java/pull/417], so if we upgrade to 
> 1.11.0+ we should be good to go. (this might be hard in branch-2.6 and 
> branch-2.7 where we're on 1.7.4)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13840) Implement getUsed() for ViewFileSystem

2016-11-29 Thread Manoj Govindassamy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Manoj Govindassamy updated HADOOP-13840:

Attachment: HADOOP-13840.02.patch

Thanks [~ste...@apache.org]. Yes, assertEquals can give us verbose information 
on the comparisons done. Updated the patch with Assert.assertTrue replaced with 
assertEquals.

> Implement getUsed() for ViewFileSystem
> --
>
> Key: HADOOP-13840
> URL: https://issues.apache.org/jira/browse/HADOOP-13840
> Project: Hadoop Common
>  Issue Type: Task
>  Components: viewfs
>Affects Versions: 3.0.0-alpha1
>Reporter: Manoj Govindassamy
>Assignee: Manoj Govindassamy
> Attachments: HADOOP-13840.01.patch, HADOOP-13840.02.patch
>
>
> ViewFileSystem doesn't override {{FileSystem#getUSed()}}. So, when file 
> system used space is queried for slash root "/" paths, the default 
> implementations tries to run the {{getContentSummary}} which crashes on 
> seeing unexpected mount points under slash. 
> ViewFileSystem#getUsed() is not expected to collate all the space used from 
> all the mount points configured under "/". Proposal is to avoid invoking 
> FileSystem#getUsed() and throw NotInMountPointException until LinkMergeSlash 
> is supported.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-13811) s3a: getFileStatus fails with com.amazonaws.AmazonClientException: Failed to sanitize XML document destined for handler class

2016-11-29 Thread Luke Miner (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13811?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706266#comment-15706266
 ] 

Luke Miner edited comment on HADOOP-13811 at 11/29/16 7:23 PM:
---

Turns out I still had {{hadoop-aws:2.7.3}} in my spark conf file. I ended up 
including the {{hadoop-aws-2.9.0-SNAPSHOT.jar}} that I built using the 
instructions you gave above. I also bumped up amazon's {{aws-java-sdk}} to 
{{1.11.57}}. I'm still seeing the same error, only on a different line number 
now. Oddly, it also seems to be telling me that I should be using 
{{S3AFileSystem}} instead of {{S3FileSystem}}.

{code}
S3FileSystem is deprecated and will be removed in future releases. Use 
NativeS3FileSystem or S3AFileSystem instead.
16/11/29 19:00:19 INFO MemoryStore: Block broadcast_0 stored as values in 
memory (estimated size 177.7 KB, free 366.1 MB)
16/11/29 19:00:19 INFO MemoryStore: Block broadcast_0_piece0 stored as bytes in 
memory (estimated size 21.0 KB, free 366.1 MB)
16/11/29 19:00:19 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory on 
10.0.229.45:52703 (size: 21.0 KB, free: 366.3 MB)
16/11/29 19:00:19 INFO SparkContext: Created broadcast 0 from textFile at 
json2pq.scala:130
Exception in thread "main" java.lang.NumberFormatException: For input string: 
"100M"
at 
java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
at java.lang.Long.parseLong(Long.java:441)
at java.lang.Long.parseLong(Long.java:483)
at org.apache.hadoop.conf.Configuration.getLong(Configuration.java:1320)
at 
org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:234)
at 
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2904)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:101)
at 
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2941)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2923)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:390)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:356)
at 
org.apache.hadoop.mapred.FileInputFormat.singleThreadedListStatus(FileInputFormat.java:265)
at 
org.apache.hadoop.mapred.FileInputFormat.listStatus(FileInputFormat.java:236)
at 
org.apache.hadoop.mapred.FileInputFormat.getSplits(FileInputFormat.java:322)
at org.apache.spark.rdd.HadoopRDD.getPartitions(HadoopRDD.scala:202)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:248)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:246)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:246)
at 
org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:248)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:246)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:246)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1957)
at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:928)
at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:358)
at org.apache.spark.rdd.RDD.collect(RDD.scala:927)
at Json2Pq$.main(json2pq.scala:130)
at Json2Pq.main(json2pq.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:738)
at 
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
{code}

I did also try to run it using {{hadoop-aws-3.0.0-alpha1.jar}} that is 
currently on central, but got this error instead. Perhaps because I'm running 
off a hadoop 2.9 snapshot.

{code}
Exception in thread "main" java.lang.IllegalArgumentException: Error while 
instantiating 'org.apache.spark.sql.internal.SessionState':
at 
org.apache.spark.sql.SparkSession$.org$apache$spark$sql$SparkSession$$reflect(SparkSession.scala:965)
at 

[jira] [Comment Edited] (HADOOP-13811) s3a: getFileStatus fails with com.amazonaws.AmazonClientException: Failed to sanitize XML document destined for handler class

2016-11-29 Thread Luke Miner (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13811?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706266#comment-15706266
 ] 

Luke Miner edited comment on HADOOP-13811 at 11/29/16 7:22 PM:
---

Turns out I still had {{hadoop-aws:2.7.3}} in my spark conf file. I ended up 
including the {{hadoop-aws-2.9.0-SNAPSHOT.jar}} that I built using the 
instructions you gave above. I also bumped up amazon's {{aws-java-sdk}} to 
{{1.11.57}}. I'm still seeing the same error, only on a different line number 
now. Oddly, it also seems to be telling me that I should be using 
{{S3AFileSystem}} instead of {{S3FileSystem}}.

{code}
S3FileSystem is deprecated and will be removed in future releases. Use 
NativeS3FileSystem or S3AFileSystem instead.
16/11/29 19:00:19 INFO MemoryStore: Block broadcast_0 stored as values in 
memory (estimated size 177.7 KB, free 366.1 MB)
16/11/29 19:00:19 INFO MemoryStore: Block broadcast_0_piece0 stored as bytes in 
memory (estimated size 21.0 KB, free 366.1 MB)
16/11/29 19:00:19 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory on 
10.0.229.45:52703 (size: 21.0 KB, free: 366.3 MB)
16/11/29 19:00:19 INFO SparkContext: Created broadcast 0 from textFile at 
json2pq.scala:130
Exception in thread "main" java.lang.NumberFormatException: For input string: 
"100M"
at 
java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
at java.lang.Long.parseLong(Long.java:441)
at java.lang.Long.parseLong(Long.java:483)
at org.apache.hadoop.conf.Configuration.getLong(Configuration.java:1320)
at 
org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:234)
at 
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2904)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:101)
at 
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2941)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2923)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:390)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:356)
at 
org.apache.hadoop.mapred.FileInputFormat.singleThreadedListStatus(FileInputFormat.java:265)
at 
org.apache.hadoop.mapred.FileInputFormat.listStatus(FileInputFormat.java:236)
at 
org.apache.hadoop.mapred.FileInputFormat.getSplits(FileInputFormat.java:322)
at org.apache.spark.rdd.HadoopRDD.getPartitions(HadoopRDD.scala:202)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:248)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:246)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:246)
at 
org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:248)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:246)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:246)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1957)
at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:928)
at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:358)
at org.apache.spark.rdd.RDD.collect(RDD.scala:927)
at Json2Pq$.main(json2pq.scala:130)
at Json2Pq.main(json2pq.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:738)
at 
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
{code}

I did also try to run it using {{hadoop-aws-3.0.0-alpha1.jar}} that is 
currently on central, but got this error instead. Perhaps because I'm running 
off a hadoop 2.9 snapshot.

{code}
Exception in thread "main" java.lang.IllegalArgumentException: Error while 
instantiating 'org.apache.spark.sql.internal.SessionState':
at 
org.apache.spark.sql.SparkSession$.org$apache$spark$sql$SparkSession$$reflect(SparkSession.scala:965)
at 

[jira] [Commented] (HADOOP-13811) s3a: getFileStatus fails with com.amazonaws.AmazonClientException: Failed to sanitize XML document destined for handler class

2016-11-29 Thread Luke Miner (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13811?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706266#comment-15706266
 ] 

Luke Miner commented on HADOOP-13811:
-

Turns out I still had {{hadoop-aws:2.7.3}} in my spark conf file. I ended up 
including the {{hadoop-aws-2.9.0-SNAPSHOT.jar}} that I built using the 
instructions you gave above. I also bumped up amazon's {{aws-java-sdk}} to 
{{1.11.57}}. I'm still seeing the same error, only on a different line number 
now. Oddly, it also seems to be telling me that I should be using 
{{S3AFileSystem}} instead of {{S3FileSystem}}.

{code}
S3FileSystem is deprecated and will be removed in future releases. Use 
NativeS3FileSystem or S3AFileSystem instead.
16/11/29 19:00:19 INFO MemoryStore: Block broadcast_0 stored as values in 
memory (estimated size 177.7 KB, free 366.1 MB)
16/11/29 19:00:19 INFO MemoryStore: Block broadcast_0_piece0 stored as bytes in 
memory (estimated size 21.0 KB, free 366.1 MB)
16/11/29 19:00:19 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory on 
10.0.229.45:52703 (size: 21.0 KB, free: 366.3 MB)
16/11/29 19:00:19 INFO SparkContext: Created broadcast 0 from textFile at 
json2pq.scala:130
Exception in thread "main" java.lang.NumberFormatException: For input string: 
"100M"
at 
java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
at java.lang.Long.parseLong(Long.java:441)
at java.lang.Long.parseLong(Long.java:483)
at org.apache.hadoop.conf.Configuration.getLong(Configuration.java:1320)
at 
org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:234)
at 
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2904)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:101)
at 
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2941)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2923)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:390)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:356)
at 
org.apache.hadoop.mapred.FileInputFormat.singleThreadedListStatus(FileInputFormat.java:265)
at 
org.apache.hadoop.mapred.FileInputFormat.listStatus(FileInputFormat.java:236)
at 
org.apache.hadoop.mapred.FileInputFormat.getSplits(FileInputFormat.java:322)
at org.apache.spark.rdd.HadoopRDD.getPartitions(HadoopRDD.scala:202)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:248)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:246)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:246)
at 
org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:248)
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:246)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.rdd.RDD.partitions(RDD.scala:246)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1957)
at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:928)
at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:358)
at org.apache.spark.rdd.RDD.collect(RDD.scala:927)
at Json2Pq$.main(json2pq.scala:130)
at Json2Pq.main(json2pq.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:738)
at 
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
{code}

I did also try to run it using {{hadoop-aws-3.0.0-alpha1.jar}} that is 
currently on central, but got this error instead. Perhaps because I'm running 
off a hadoop 2.9 snapshot.

{code}
Exception in thread "main" java.lang.IllegalArgumentException: Error while 
instantiating 'org.apache.spark.sql.internal.SessionState':
at 
org.apache.spark.sql.SparkSession$.org$apache$spark$sql$SparkSession$$reflect(SparkSession.scala:965)
at 
org.apache.spark.sql.SparkSession.sessionState$lzycompute(SparkSession.scala:110)
at 

[jira] [Commented] (HADOOP-13650) S3Guard: Provide command line tools to manipulate metadata store.

2016-11-29 Thread Allen Wittenauer (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706113#comment-15706113
 ] 

Allen Wittenauer commented on HADOOP-13650:
---

Based upon this patch, you don't need separate scripts/separate command.  
There's not enough here to warrant polluting the shell path.  (The single case 
statement is a dead give away.)  Use a shell profile to add the s3guard command 
to bin/hadoop dynamically.  See, e.g., distcp, hadoop-archives, etc. 

Also: 

{code}
if declare -f hadoop_subcommand_
{code}

(plus a few other spots)

This isn't bin/hadoop, so doing this will conflict.

{code}
+hadoop_debug "Appending hadoop-aws jars"
+hadoop_add_classpath "${HADOOP_TOOLS_HOME}/${HADOOP_TOOLS_LIB_JARS_DIR}/*"
{code}

This adds way way way more than you need and should be specifically avoided in 
trunk. So don't do that. Besides, if hadoop-aws is enabled as a supported tool, 
the AWS bits are already present in the classpath.  If you want to forcibly 
load them regardless of the HADOOP_OPTIONAL_TOOLS setting, then call 
_hadoop-aws_hadoop_classpath directly.  If the hadoop-aws support has been 
already added, hadoop_add_classpath will skip adding the jars a second time.

> S3Guard: Provide command line tools to manipulate metadata store.
> -
>
> Key: HADOOP-13650
> URL: https://issues.apache.org/jira/browse/HADOOP-13650
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Reporter: Lei (Eddy) Xu
>Assignee: Lei (Eddy) Xu
> Attachments: HADOOP-13650-HADOOP-13345.000.patch, 
> HADOOP-13650-HADOOP-13345.001.patch, HADOOP-13650-HADOOP-13345.002.patch
>
>
> Similar systems like EMRFS has the CLI tools to manipulate the metadata 
> store, i.e., create or delete metadata store, or {{import}}, {{sync}} the 
> file metadata between metadata store and S3. 
> http://docs.aws.amazon.com//ElasticMapReduce/latest/ReleaseGuide/emrfs-cli-reference.html
> S3Guard should offer similar functionality. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-13597) Switch KMS from Tomcat to Jetty

2016-11-29 Thread John Zhuge (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13597?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15704129#comment-15704129
 ] 

John Zhuge edited comment on HADOOP-13597 at 11/29/16 6:08 PM:
---

Patch 001
- KMSHttpServer based on HttpServer2
- Redirect MiniKMS to KMSHttpServer
- Add kms-default.xml
- Add Jetty properties including SSL properties
- Convert hadoop-kms from war to jar
- Rewrite kms.sh to use Hadoop shell script framework
- Obsolete HTTP admin port for the Tomcat Manager GUI which does not seem to 
work anyway
- Obsolete {{kms.sh version}} that prints Tomcat version

TESTING DONE
- All hadoop-kms unit tests. MiniKMS equals full KMS.
- Non-secure REST APIs
- Non-secure “hadoop key” commands
- SSL REST APIs
- kms.sh run/start/stop/status
- JMX works
- /logs works

TODO
- Set HTTP request header size by env KMS_MAX_HTTP_HEADER_SIZE
- Add static web content /index.html
- More ad-hoc testing
- Integration testing
- Update docs: index.md.vm

TODO in new JIRAs:
- Integrate with Hadoop SSL server configuration
- Full SSL server configuration: 
includeProtocols/excludeProtocols/includeCipherSuites/excludeCipherSuites, etc.
- Design common Http server configuration. Common properties in 
“-site.xml” with config prefix, e.g., “hadoop.kms.”.
- Design HttpServer2 configuration-based builder
- Share web apps code in Common, HDFS, and YARN

My private branch: https://github.com/jzhuge/hadoop/tree/HADOOP-13597.001


was (Author: jzhuge):
Patch 001
- KMSHttpServer based on HttpServer2
- Redirect MiniKMS to KMSHttpServer
- Add kms-default.xml
- Add Jetty properties including SSL properties
- Convert hadoop-kms from war to jar
- Rewrite kms.sh to use Hadoop shell script framework
- Obsolete HTTP admin port for the Tomcat Manager GUI which does not seem to 
work anyway
- Obsolete {{kms.sh version}} that prints Tomcat version

TESTING DONE
- All hadoop-kms unit tests which exercise the full KMS instead of MiniKMS
- Non-secure REST APIs
- Non-secure “hadoop key” commands
- SSL REST APIs
- kms.sh run/start/stop/status

TODO
- Set HTTP request header size by env KMS_MAX_HTTP_HEADER_SIZE
- Add static web content /index.html
- More ad-hoc testing
- Integration testing
- Update docs: index.md.vm

TODO in new JIRAs:
- Integrate with Hadoop SSL server configuration
- Full SSL server configuration: 
includeProtocols/excludeProtocols/includeCipherSuites/excludeCipherSuites, etc.
- Design common Http server configuration. Common properties in 
“-site.xml” with config prefix, e.g., “hadoop.kms.”.
- Design HttpServer2 configuration-based builder
- JMX not working, existing issue
- Share web apps code in Common, HDFS, and YARN

My private branch: https://github.com/jzhuge/hadoop/tree/HADOOP-13597.001

> Switch KMS from Tomcat to Jetty
> ---
>
> Key: HADOOP-13597
> URL: https://issues.apache.org/jira/browse/HADOOP-13597
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: kms
>Affects Versions: 2.6.0
>Reporter: John Zhuge
>Assignee: John Zhuge
> Attachments: HADOOP-13597.001.patch
>
>
> The Tomcat 6 we are using will reach EOL at the end of 2017. While there are 
> other good options, I would propose switching to {{Jetty 9}} for the 
> following reasons:
> * Easier migration. Both Tomcat and Jetty are based on {{Servlet 
> Containers}}, so we don't have change client code that much. It would require 
> more work to switch to {{JAX-RS}}.
> * Well established.
> * Good performance and scalability.
> Other alternatives:
> * Jersey + Grizzly
> * Tomcat 8
> Your opinions will be greatly appreciated.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13836) Securing Hadoop RPC using SSL

2016-11-29 Thread Arun Suresh (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706026#comment-15706026
 ] 

Arun Suresh commented on HADOOP-13836:
--

Thanks for pointing out IPv6 [~steve_l]

I assume it should just work, considering this comes into play only at Socket 
creation. My understanding is that SSL/TLS certification authentication is 
based on the DNS names of the entities involved, so again, I assume it should 
just work, if DNS resolution works correctly.

But yes, we will try to verify it. My understanding is that the HADOOP-11890 
branch had most of the IPv6 based changes. If this branch is uptodate, 
[~kartheek],  we can maybe test it against that branch. Any more pointers to 
verifying if this would work with IPv6 is welcome. (cc: [~eclark], [~nkedel])



> Securing Hadoop RPC using SSL
> -
>
> Key: HADOOP-13836
> URL: https://issues.apache.org/jira/browse/HADOOP-13836
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: ipc
>Reporter: kartheek muthyala
>Assignee: kartheek muthyala
> Attachments: HADOOP-13836.patch
>
>
> Today, RPC connections in Hadoop are encrypted using Simple Authentication & 
> Security Layer (SASL), with the Kerberos ticket based authentication or 
> Digest-md5 checksum based authentication protocols. This proposal is about 
> enhancing this cipher suite with SSL/TLS based encryption and authentication. 
> SSL/TLS is a proposed Internet Engineering Task Force (IETF) standard, that 
> provides data security and integrity across two different end points in a 
> network. This protocol has made its way to a number of applications such as 
> web browsing, email, internet faxing, messaging, VOIP etc. And supporting 
> this cipher suite at the core of Hadoop would give a good synergy with the 
> applications on top and also bolster industry adoption of Hadoop.
> The Server and Client code in Hadoop IPC should support the following modes 
> of communication
> 1.Plain 
> 2. SASL encryption with an underlying authentication
> 3. SSL based encryption and authentication (x509 certificate)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-13836) Securing Hadoop RPC using SSL

2016-11-29 Thread Arun Suresh (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15706026#comment-15706026
 ] 

Arun Suresh edited comment on HADOOP-13836 at 11/29/16 6:02 PM:


Thanks for pointing out IPv6 [~steve_l]

I assume it should just work, considering this comes into play only at Socket 
creation. My understanding is that SSL/TLS certificate authentication is based 
on the DNS names of the entities involved, so again, I assume it should just 
work, if DNS resolution works correctly.

But yes, we will try to verify it. My understanding is that the HADOOP-11890 
branch had most of the IPv6 based changes. If this branch is uptodate, 
[~kartheek],  we can maybe test it against that branch. Any more pointers to 
verifying if this would work with IPv6 is welcome. (cc: [~eclark], [~nkedel])




was (Author: asuresh):
Thanks for pointing out IPv6 [~steve_l]

I assume it should just work, considering this comes into play only at Socket 
creation. My understanding is that SSL/TLS certification authentication is 
based on the DNS names of the entities involved, so again, I assume it should 
just work, if DNS resolution works correctly.

But yes, we will try to verify it. My understanding is that the HADOOP-11890 
branch had most of the IPv6 based changes. If this branch is uptodate, 
[~kartheek],  we can maybe test it against that branch. Any more pointers to 
verifying if this would work with IPv6 is welcome. (cc: [~eclark], [~nkedel])



> Securing Hadoop RPC using SSL
> -
>
> Key: HADOOP-13836
> URL: https://issues.apache.org/jira/browse/HADOOP-13836
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: ipc
>Reporter: kartheek muthyala
>Assignee: kartheek muthyala
> Attachments: HADOOP-13836.patch
>
>
> Today, RPC connections in Hadoop are encrypted using Simple Authentication & 
> Security Layer (SASL), with the Kerberos ticket based authentication or 
> Digest-md5 checksum based authentication protocols. This proposal is about 
> enhancing this cipher suite with SSL/TLS based encryption and authentication. 
> SSL/TLS is a proposed Internet Engineering Task Force (IETF) standard, that 
> provides data security and integrity across two different end points in a 
> network. This protocol has made its way to a number of applications such as 
> web browsing, email, internet faxing, messaging, VOIP etc. And supporting 
> this cipher suite at the core of Hadoop would give a good synergy with the 
> applications on top and also bolster industry adoption of Hadoop.
> The Server and Client code in Hadoop IPC should support the following modes 
> of communication
> 1.Plain 
> 2. SASL encryption with an underlying authentication
> 3. SSL based encryption and authentication (x509 certificate)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13600) S3a rename() to copy files in a directory in parallel

2016-11-29 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13600?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-13600:

Assignee: Sahil Takiar  (was: Steve Loughran)

> S3a rename() to copy files in a directory in parallel
> -
>
> Key: HADOOP-13600
> URL: https://issues.apache.org/jira/browse/HADOOP-13600
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.7.3
>Reporter: Steve Loughran
>Assignee: Sahil Takiar
>
> Currently a directory rename does a one-by-one copy, making the request 
> O(files * data). If the copy operations were launched in parallel, the 
> duration of the copy may be reducable to the duration of the longest copy. 
> For a directory with many files, this will be significant



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13845) s3a to instrument duration of HTTP calls

2016-11-29 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13845?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-13845:

Priority: Minor  (was: Major)

> s3a to instrument duration of HTTP calls
> 
>
> Key: HADOOP-13845
> URL: https://issues.apache.org/jira/browse/HADOOP-13845
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.9.0
>Reporter: Steve Loughran
>Priority: Minor
>
> HADOOP-13844 proposes pulling out the swift duration classes for reuse; this 
> patch proposes instrumenting s3a with it.
> One interesting question: what to do with the values. For now, they could 
> just be printed, but it might be interesting to include in FS stats collected 
> at the end of a run. However, those are all assumed to be simple counters 
> where merging is a matter of addition. These are more metrics



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-13845) s3a to instrument duration of HTTP calls

2016-11-29 Thread Steve Loughran (JIRA)
Steve Loughran created HADOOP-13845:
---

 Summary: s3a to instrument duration of HTTP calls
 Key: HADOOP-13845
 URL: https://issues.apache.org/jira/browse/HADOOP-13845
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 2.9.0
Reporter: Steve Loughran


HADOOP-13844 proposes pulling out the swift duration classes for reuse; this 
patch proposes instrumenting s3a with it.

One interesting question: what to do with the values. For now, they could just 
be printed, but it might be interesting to include in FS stats collected at the 
end of a run. However, those are all assumed to be simple counters where 
merging is a matter of addition. These are more metrics



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Issue Comment Deleted] (HADOOP-13204) Über-jira: S3a phase III: scale and tuning

2016-11-29 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13204?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-13204:

Comment: was deleted

(was: Patch 008; in sync with branch-2)

> Über-jira: S3a phase III: scale and tuning
> --
>
> Key: HADOOP-13204
> URL: https://issues.apache.org/jira/browse/HADOOP-13204
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Affects Versions: 2.8.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>
> S3A Phase III work; post 2.8. 
> Areas could include
> * customisation
> * performance enhancement
> * management and metrics
> * error/failure handling
> And of course any bugs which surface



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13204) Über-jira: S3a phase III: scale and tuning

2016-11-29 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13204?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-13204:

Attachment: (was: HADOOP-13614-branch-2-008.patch)

> Über-jira: S3a phase III: scale and tuning
> --
>
> Key: HADOOP-13204
> URL: https://issues.apache.org/jira/browse/HADOOP-13204
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Affects Versions: 2.8.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>
> S3A Phase III work; post 2.8. 
> Areas could include
> * customisation
> * performance enhancement
> * management and metrics
> * error/failure handling
> And of course any bugs which surface



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-13844) Move swift duration stats tracking to hadoop common utils

2016-11-29 Thread Steve Loughran (JIRA)
Steve Loughran created HADOOP-13844:
---

 Summary: Move swift duration stats tracking to hadoop common utils
 Key: HADOOP-13844
 URL: https://issues.apache.org/jira/browse/HADOOP-13844
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs/azure, fs/oss, fs/s3, fs/swift
Affects Versions: 2.9.0
Reporter: Steve Loughran
Assignee: Steve Loughran
Priority: Minor


Swift FS has some simple duration counting classes which are used to track HTTP 
performance, with incremental creation of mean, min, max and variance. These 
are not moving averages, simply ongoing over time. They've proved useful in 
detecting which operations are slowest against objects stores. (it turned out 
to be delete throttling in test teardown BTW, showing the value of the 
instrumentation)

By moving them into utils, all the object stores can build up some stats on 
cost of HTTP operations.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13836) Securing Hadoop RPC using SSL

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705730#comment-15705730
 ] 

Steve Loughran commented on HADOOP-13836:
-

This is all going to work once the IPV6 patch is merged in, right?

> Securing Hadoop RPC using SSL
> -
>
> Key: HADOOP-13836
> URL: https://issues.apache.org/jira/browse/HADOOP-13836
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: ipc
>Reporter: kartheek muthyala
>Assignee: kartheek muthyala
> Attachments: HADOOP-13836.patch
>
>
> Today, RPC connections in Hadoop are encrypted using Simple Authentication & 
> Security Layer (SASL), with the Kerberos ticket based authentication or 
> Digest-md5 checksum based authentication protocols. This proposal is about 
> enhancing this cipher suite with SSL/TLS based encryption and authentication. 
> SSL/TLS is a proposed Internet Engineering Task Force (IETF) standard, that 
> provides data security and integrity across two different end points in a 
> network. This protocol has made its way to a number of applications such as 
> web browsing, email, internet faxing, messaging, VOIP etc. And supporting 
> this cipher suite at the core of Hadoop would give a good synergy with the 
> applications on top and also bolster industry adoption of Hadoop.
> The Server and Client code in Hadoop IPC should support the following modes 
> of communication
> 1.Plain 
> 2. SASL encryption with an underlying authentication
> 3. SSL based encryption and authentication (x509 certificate)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13836) Securing Hadoop RPC using SSL

2016-11-29 Thread kartheek muthyala (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705693#comment-15705693
 ] 

kartheek muthyala commented on HADOOP-13836:


[~antkou], 

I liked your suggestion of using "hadoop.rpc.socket.factory.class.default" to 
specify our own SSLSocketFactory class on the Client side. For this I just have 
to move the existing SSL socket creation related changes from Client.java to a 
separate file. Will consider this in my next patch.

We have considered org.apache.hadoop.security.ssl.SSLFactory for creating 
SSLEngine on Server. But it doesn't has  SSLSocketChannel on top of SSLEngine, 
like niossl library does. Having this support would make us stick to the 
existing SocketChannel Server design. 



> Securing Hadoop RPC using SSL
> -
>
> Key: HADOOP-13836
> URL: https://issues.apache.org/jira/browse/HADOOP-13836
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: ipc
>Reporter: kartheek muthyala
>Assignee: kartheek muthyala
> Attachments: HADOOP-13836.patch
>
>
> Today, RPC connections in Hadoop are encrypted using Simple Authentication & 
> Security Layer (SASL), with the Kerberos ticket based authentication or 
> Digest-md5 checksum based authentication protocols. This proposal is about 
> enhancing this cipher suite with SSL/TLS based encryption and authentication. 
> SSL/TLS is a proposed Internet Engineering Task Force (IETF) standard, that 
> provides data security and integrity across two different end points in a 
> network. This protocol has made its way to a number of applications such as 
> web browsing, email, internet faxing, messaging, VOIP etc. And supporting 
> this cipher suite at the core of Hadoop would give a good synergy with the 
> applications on top and also bolster industry adoption of Hadoop.
> The Server and Client code in Hadoop IPC should support the following modes 
> of communication
> 1.Plain 
> 2. SASL encryption with an underlying authentication
> 3. SSL based encryption and authentication (x509 certificate)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-13843) S3Guard, MetadataStore to support atomic create(path, overwrite=false)

2016-11-29 Thread Steve Loughran (JIRA)
Steve Loughran created HADOOP-13843:
---

 Summary: S3Guard, MetadataStore to support atomic create(path, 
overwrite=false)
 Key: HADOOP-13843
 URL: https://issues.apache.org/jira/browse/HADOOP-13843
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.0.0-alpha2
Reporter: Steve Loughran


Support atomically enforced file creation. Current s3a can do a check in 
create() and fail if there is something there, but a new entry only gets 
created at the end of the PUT; during the entire interval between that check 
and the close() of the stream, there's nothing to stop other callers creating 
an object.

Proposed: s3afs can do a check + create a 0 byte file at the path; that'd need 
some {{putNoOverwrite(DirListingMetadata)}} call in MetadataStore, followed by 
a PUT of an 0-byte file to S3. That will increase cost of file creation, though 
at least with the MD store, the cost of the initial getFileStatus() check is 
down.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13786) add output committer which uses s3guard for consistent O(1) commits to S3

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13786?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705650#comment-15705650
 ] 

Steve Loughran commented on HADOOP-13786:
-


Looking at the atomic commit problem, I'm starting to worry that I may need 
some directory leasing, which would imply a lock field in the table. Just 
warning you there. The problem is that while a rename is in progress, you wan't 
to stop anything trying to PUT data anywhere under the two paths being worked 
on.



> add output committer which uses s3guard for consistent O(1) commits to S3
> -
>
> Key: HADOOP-13786
> URL: https://issues.apache.org/jira/browse/HADOOP-13786
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-alpha2
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>
> A goal of this code is "support O(1) commits to S3 repositories in the 
> presence of failures". Implement it, including whatever is needed to 
> demonstrate the correctness of the algorithm. (that is, assuming that s3guard 
> provides a consistent view of the presence/absence of blobs, show that we can 
> commit directly).
> I consider ourselves free to expose the blobstore-ness of the s3 output 
> streams (ie. not visible until the close()), if we need to use that to allow 
> us to abort commit operations.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13650) S3Guard: Provide command line tools to manipulate metadata store.

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705643#comment-15705643
 ] 

Steve Loughran commented on HADOOP-13650:
-

* {{S3GuardTool.main}} needs to catch, log all exceptions. Ideally return a 
different exit code for an exception than from a usage command; this makes 
testing/diags easier.
* be good to have ~aw to review the bash scripts. Are these testable? At the 
very least, something that goes as far as a usage command in the java code can 
do basic installation testing.
* we have the option of JCommander here for arg parsing. just a thought.
* might be good to have the option of printing the diff out in a way that's 
easy to parse downstream. Is the current output machine parseable?
* For the comparison, a {{listFiles(recursive=true)}} is much faster to list s3 
buckets...but I'm not sure how well it will work here, because the iterator 
returned from the metastore may return differently ordered results from the FS; 
you don't want to build up large structures in memory. Even so, I'm not sure 
how well the existing code can handle a path with a few million blobs directly 
under a path. some s3 buckets do have this, apparently.
* Maybe an operation to verify that the metastore is in sync with s3, return a 
specific error code to indicate this is not the case. That's something you can 
use at installation, preflight, etc.


> S3Guard: Provide command line tools to manipulate metadata store.
> -
>
> Key: HADOOP-13650
> URL: https://issues.apache.org/jira/browse/HADOOP-13650
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Reporter: Lei (Eddy) Xu
>Assignee: Lei (Eddy) Xu
> Attachments: HADOOP-13650-HADOOP-13345.000.patch, 
> HADOOP-13650-HADOOP-13345.001.patch, HADOOP-13650-HADOOP-13345.002.patch
>
>
> Similar systems like EMRFS has the CLI tools to manipulate the metadata 
> store, i.e., create or delete metadata store, or {{import}}, {{sync}} the 
> file metadata between metadata store and S3. 
> http://docs.aws.amazon.com//ElasticMapReduce/latest/ReleaseGuide/emrfs-cli-reference.html
> S3Guard should offer similar functionality. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-13650) S3Guard: Provide command line tools to manipulate metadata store.

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705643#comment-15705643
 ] 

Steve Loughran edited comment on HADOOP-13650 at 11/29/16 3:42 PM:
---

* {{S3GuardTool.main}} needs to catch, log all exceptions. Ideally return a 
different exit code for an exception than from a usage command; this makes 
testing/diags easier.
* be good to have [~aw] to review the bash scripts. Are these testable? At the 
very least, something that goes as far as a usage command in the java code can 
do basic installation testing.
* we have the option of JCommander here for arg parsing. just a thought.
* might be good to have the option of printing the diff out in a way that's 
easy to parse downstream. Is the current output machine parseable?
* For the comparison, a {{listFiles(recursive=true)}} is much faster to list s3 
buckets...but I'm not sure how well it will work here, because the iterator 
returned from the metastore may return differently ordered results from the FS; 
you don't want to build up large structures in memory. Even so, I'm not sure 
how well the existing code can handle a path with a few million blobs directly 
under a path. some s3 buckets do have this, apparently.
* Maybe an operation to verify that the metastore is in sync with s3, return a 
specific error code to indicate this is not the case. That's something you can 
use at installation, preflight, etc.



was (Author: ste...@apache.org):
* {{S3GuardTool.main}} needs to catch, log all exceptions. Ideally return a 
different exit code for an exception than from a usage command; this makes 
testing/diags easier.
* be good to have ~aw to review the bash scripts. Are these testable? At the 
very least, something that goes as far as a usage command in the java code can 
do basic installation testing.
* we have the option of JCommander here for arg parsing. just a thought.
* might be good to have the option of printing the diff out in a way that's 
easy to parse downstream. Is the current output machine parseable?
* For the comparison, a {{listFiles(recursive=true)}} is much faster to list s3 
buckets...but I'm not sure how well it will work here, because the iterator 
returned from the metastore may return differently ordered results from the FS; 
you don't want to build up large structures in memory. Even so, I'm not sure 
how well the existing code can handle a path with a few million blobs directly 
under a path. some s3 buckets do have this, apparently.
* Maybe an operation to verify that the metastore is in sync with s3, return a 
specific error code to indicate this is not the case. That's something you can 
use at installation, preflight, etc.


> S3Guard: Provide command line tools to manipulate metadata store.
> -
>
> Key: HADOOP-13650
> URL: https://issues.apache.org/jira/browse/HADOOP-13650
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Reporter: Lei (Eddy) Xu
>Assignee: Lei (Eddy) Xu
> Attachments: HADOOP-13650-HADOOP-13345.000.patch, 
> HADOOP-13650-HADOOP-13345.001.patch, HADOOP-13650-HADOOP-13345.002.patch
>
>
> Similar systems like EMRFS has the CLI tools to manipulate the metadata 
> store, i.e., create or delete metadata store, or {{import}}, {{sync}} the 
> file metadata between metadata store and S3. 
> http://docs.aws.amazon.com//ElasticMapReduce/latest/ReleaseGuide/emrfs-cli-reference.html
> S3Guard should offer similar functionality. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13812) Upgrade Tomcat to 6.0.48

2016-11-29 Thread John Zhuge (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13812?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705634#comment-15705634
 ] 

John Zhuge commented on HADOOP-13812:
-

[~ajisakaa] So many -1's with 2.6 patch?

> Upgrade Tomcat to 6.0.48
> 
>
> Key: HADOOP-13812
> URL: https://issues.apache.org/jira/browse/HADOOP-13812
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 2.6.0
>Reporter: John Zhuge
>Assignee: John Zhuge
>Priority: Blocker
> Fix For: 2.8.0, 2.7.4, 3.0.0-alpha2, 2.6.6
>
> Attachments: HADOOP-13812.001.patch, HADOOP-13812.branch-2.6.patch
>
>
> KMS and HttpFS currently uses Tomcat 6.0.44, propose to upgrade to the latest 
> version is 6.0.48.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13421) Switch to v2 of the S3 List Objects API in S3A

2016-11-29 Thread Thomas Demoor (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705631#comment-15705631
 ] 

Thomas Demoor commented on HADOOP-13421:


Thanks for reaching out [~ste...@apache.org] ([~PieterReuse] discussed with 
me). 

This API call is quite recent (May 4th 2016) so there are a lot of legacy 
systems which will not support this. Our latest release does but I think if we 
don't make this optional the majority of other s3 endpoints will brake. 

Making this optional seems relatively low cost. I quickly checked and we don't 
seem to use the {{marker}} functionality in the V1 implementation which is 
replaced by {{start-after}} in V2.

> Switch to v2 of the S3 List Objects API in S3A
> --
>
> Key: HADOOP-13421
> URL: https://issues.apache.org/jira/browse/HADOOP-13421
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
>Reporter: Steven K. Wong
>Priority: Minor
>
> Unlike [version 
> 1|http://docs.aws.amazon.com/AmazonS3/latest/API/RESTBucketGET.html] of the 
> S3 List Objects API, [version 
> 2|http://docs.aws.amazon.com/AmazonS3/latest/API/v2-RESTBucketGET.html] by 
> default does not fetch object owner information, which S3A doesn't need 
> anyway. By switching to v2, there will be less data to transfer/process. 
> Also, it should be more robust when listing a versioned bucket with "a large 
> number of delete markers" ([according to 
> AWS|https://aws.amazon.com/releasenotes/Java/0735652458007581]).
> Methods in S3AFileSystem that use this API include:
> * getFileStatus(Path)
> * innerDelete(Path, boolean)
> * innerListStatus(Path)
> * innerRename(Path, Path)
> Requires AWS SDK 1.10.75 or later.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13812) Upgrade Tomcat to 6.0.48

2016-11-29 Thread John Zhuge (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13812?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

John Zhuge updated HADOOP-13812:

Release Note: Tomcat 6.0.46 starts to filter weak ciphers. Some old SSL 
clients may be affected. It is recommended to upgrade the SSL client. Run the 
SSL client against https://www.howsmyssl.com/a/check to find out its TLS 
version and cipher suites.  (was: Tomcat 6.0.46 starts to filter weak ciphers. 
Some old SSL clients may be affected. It is recommended to upgrade the SSL 
client.)

> Upgrade Tomcat to 6.0.48
> 
>
> Key: HADOOP-13812
> URL: https://issues.apache.org/jira/browse/HADOOP-13812
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 2.6.0
>Reporter: John Zhuge
>Assignee: John Zhuge
>Priority: Blocker
> Fix For: 2.8.0, 2.7.4, 3.0.0-alpha2, 2.6.6
>
> Attachments: HADOOP-13812.001.patch, HADOOP-13812.branch-2.6.patch
>
>
> KMS and HttpFS currently uses Tomcat 6.0.44, propose to upgrade to the latest 
> version is 6.0.48.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13836) Securing Hadoop RPC using SSL

2016-11-29 Thread kartheek muthyala (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705614#comment-15705614
 ] 

kartheek muthyala commented on HADOOP-13836:


[~drankye] , very good questions. Here are some of my responses. Correct me if 
I am wrong. I will soon post a design doc with all these details

What's the scenarios, requirements and use cases you have in mind for this 
support (other than Kerberos)?
- Avoiding man in the middle attacks, through proper SSL Connection handshake 
before even exchanging the data.
- Better encryption over the wire.
- Ability to get trusted third party validation through Versign, goDaddy etc, 
which improves industrial adoption for sensitive data exchange.
- We can extend the same cipher suite to encrypt data on flight and rest.

What interfaces will be taken care of by this: RPC/commands, REST, web, JDBC 
and etc.
- For now we are supporting interfaces that derive from the hadoop.ipc.Server 
and hadoop.ipc.Client classes. So, primarily RPC.

How authentication will be considered? Still simple or some mechanisms over 
SSL/TLS?
- Today we enabled client to authenticate with the servers on connection. So 
configuring keystore is a must on the server. And with a configured KeyManager 
we can decide on what authentication credentials should be sent to the remote 
host for authentication during SSL handshake.

How would you manage credentials (X.509 certificates) for Hadoop services and 
maybe clients?
- The current work requires both server and client to be installed with 
keystore and truststore, and configured through ssl-client.xml and 
ssl-server.xml.  

What's the exact SSL/TLS versions to support and how to configure such with the 
cipher suite options?
- Currently we are supporting TLSv1.2 as a default. Because the cipher suite 
hasn't changed between TLSv1 and TLSv1.2, TLSv1 also should be supported. Given 
that the SSLContext varies for different versions of SSL/TLS, we can provide an 
interface for deriving this SSLContext depending upon the version of TLS/SSL 
configured.

> Securing Hadoop RPC using SSL
> -
>
> Key: HADOOP-13836
> URL: https://issues.apache.org/jira/browse/HADOOP-13836
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: ipc
>Reporter: kartheek muthyala
>Assignee: kartheek muthyala
> Attachments: HADOOP-13836.patch
>
>
> Today, RPC connections in Hadoop are encrypted using Simple Authentication & 
> Security Layer (SASL), with the Kerberos ticket based authentication or 
> Digest-md5 checksum based authentication protocols. This proposal is about 
> enhancing this cipher suite with SSL/TLS based encryption and authentication. 
> SSL/TLS is a proposed Internet Engineering Task Force (IETF) standard, that 
> provides data security and integrity across two different end points in a 
> network. This protocol has made its way to a number of applications such as 
> web browsing, email, internet faxing, messaging, VOIP etc. And supporting 
> this cipher suite at the core of Hadoop would give a good synergy with the 
> applications on top and also bolster industry adoption of Hadoop.
> The Server and Client code in Hadoop IPC should support the following modes 
> of communication
> 1.Plain 
> 2. SASL encryption with an underlying authentication
> 3. SSL based encryption and authentication (x509 certificate)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13589) S3Guard: Allow execution of all S3A integration tests with S3Guard enabled.

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13589?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705598#comment-15705598
 ] 

Steve Loughran commented on HADOOP-13589:
-

now we've got multiple back ends (and dynamo soon), life is complex. I'd like 
to have the option to mandate s3guard, which if set goes to local 
(authoritative = ? ), but allow us to switch to dynamodb as well.

{{S3ATestUtils.getTestProperty()}} will grab a property for a test, giving a 
sysprop which != "unset" the ability to override anything in the XML test 
configurations. This would let you 

# have a property {{-Ds3guard}}
# which will enable an {{s3guard}} profile to set what we want

if we modify the surefire/failsafe profiles to pass down a set of system 
properties, which are all set to {{unset}} by default then any of those set in 
the s3guard profile, or set on the maven command line will be picked up and 
override any in the XML files. 

The other thing to consider is junit profiles. HADOOP-13834 covers that idea as 
a way of enabling scale tests, rather than the contrived bits in 
{{AbstractS3aScaleTests}}. I haven't got that working, and haven't really put 
in the time to understand why —it felt like one of those interesting wastes of 
large amounts of time. A thought though. 





> S3Guard: Allow execution of all S3A integration tests with S3Guard enabled.
> ---
>
> Key: HADOOP-13589
> URL: https://issues.apache.org/jira/browse/HADOOP-13589
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Reporter: Chris Nauroth
>Assignee: Steve Loughran
>
> With S3Guard enabled, S3A must continue to be functionally correct.  The best 
> way to verify this is to execute our existing S3A integration tests in a mode 
> with S3A enabled.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13812) Upgrade Tomcat to 6.0.48

2016-11-29 Thread John Zhuge (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13812?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705577#comment-15705577
 ] 

John Zhuge commented on HADOOP-13812:
-

Thanks [~ajisakaa] for the review and commit.

> Upgrade Tomcat to 6.0.48
> 
>
> Key: HADOOP-13812
> URL: https://issues.apache.org/jira/browse/HADOOP-13812
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 2.6.0
>Reporter: John Zhuge
>Assignee: John Zhuge
>Priority: Blocker
> Fix For: 2.8.0, 2.7.4, 3.0.0-alpha2, 2.6.6
>
> Attachments: HADOOP-13812.001.patch, HADOOP-13812.branch-2.6.patch
>
>
> KMS and HttpFS currently uses Tomcat 6.0.44, propose to upgrade to the latest 
> version is 6.0.48.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13836) Securing Hadoop RPC using SSL

2016-11-29 Thread Arun Suresh (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13836?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun Suresh updated HADOOP-13836:
-
Assignee: kartheek muthyala

> Securing Hadoop RPC using SSL
> -
>
> Key: HADOOP-13836
> URL: https://issues.apache.org/jira/browse/HADOOP-13836
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: ipc
>Reporter: kartheek muthyala
>Assignee: kartheek muthyala
> Attachments: HADOOP-13836.patch
>
>
> Today, RPC connections in Hadoop are encrypted using Simple Authentication & 
> Security Layer (SASL), with the Kerberos ticket based authentication or 
> Digest-md5 checksum based authentication protocols. This proposal is about 
> enhancing this cipher suite with SSL/TLS based encryption and authentication. 
> SSL/TLS is a proposed Internet Engineering Task Force (IETF) standard, that 
> provides data security and integrity across two different end points in a 
> network. This protocol has made its way to a number of applications such as 
> web browsing, email, internet faxing, messaging, VOIP etc. And supporting 
> this cipher suite at the core of Hadoop would give a good synergy with the 
> applications on top and also bolster industry adoption of Hadoop.
> The Server and Client code in Hadoop IPC should support the following modes 
> of communication
> 1.Plain 
> 2. SASL encryption with an underlying authentication
> 3. SSL based encryption and authentication (x509 certificate)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13332) Remove jackson 1.9.13 and switch all jackson code to 2.x code line

2016-11-29 Thread Akira Ajisaka (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13332?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705399#comment-15705399
 ] 

Akira Ajisaka commented on HADOOP-13332:


bq. What uses jersey-json? ATS rest client?
Used by JAXBContextResolver.java in the following modules:
* hadoop-yarn-server-nodemanager
* hadoop-yarn-server-resourcemanager
* hadoop-yarn-server-applicationhistoryservice
* hadoop-yarn-common
* hadoop-mapreduce-client-app
* hadoop-mapreduce-client-hs

> Remove jackson 1.9.13 and switch all jackson code to 2.x code line
> --
>
> Key: HADOOP-13332
> URL: https://issues.apache.org/jira/browse/HADOOP-13332
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: build
>Affects Versions: 2.8.0
>Reporter: PJ Fanning
>Assignee: Akira Ajisaka
> Attachments: HADOOP-13332.00.patch, HADOOP-13332.01.patch, 
> HADOOP-13332.02.patch, HADOOP-13332.03.patch
>
>
> This jackson 1.9 code line is no longer maintained. Upgrade
> Most changes from jackson 1.9 to 2.x just involve changing the package name.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13332) Remove jackson 1.9.13 and switch all jackson code to 2.x code line

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13332?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705250#comment-15705250
 ] 

Steve Loughran commented on HADOOP-13332:
-

hbase is only used for ATS -hbase tests, AFAIK. What uses jersey-json? ATS rest 
client?

> Remove jackson 1.9.13 and switch all jackson code to 2.x code line
> --
>
> Key: HADOOP-13332
> URL: https://issues.apache.org/jira/browse/HADOOP-13332
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: build
>Affects Versions: 2.8.0
>Reporter: PJ Fanning
>Assignee: Akira Ajisaka
> Attachments: HADOOP-13332.00.patch, HADOOP-13332.01.patch, 
> HADOOP-13332.02.patch, HADOOP-13332.03.patch
>
>
> This jackson 1.9 code line is no longer maintained. Upgrade
> Most changes from jackson 1.9 to 2.x just involve changing the package name.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13842) Update jackson from 1.9.13 to 2.x in hadoop-maven-plugins

2016-11-29 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13842?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705246#comment-15705246
 ] 

Hadoop QA commented on HADOOP-13842:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
16s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red}  0m  
0s{color} | {color:red} The patch doesn't appear to include any new or modified 
tests. Please justify why no new tests are needed for this patch. Also please 
list what manual steps were performed to verify this patch. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
13s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
13s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
11s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
15s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
10s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
25s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
10s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
12s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
 9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
13s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
 9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
1s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
33s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m  
9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  0m 
11s{color} | {color:green} hadoop-maven-plugins in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
16s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 13m 17s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | HADOOP-13842 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12840867/HADOOP-13842.01.patch 
|
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  xml  findbugs  checkstyle  |
| uname | Linux 74a3b1c9a081 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed 
Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 00096dc |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/11160/testReport/ |
| modules | C: hadoop-maven-plugins U: hadoop-maven-plugins |
| Console output | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/11160/console |
| Powered by | Apache Yetus 0.4.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> Update jackson from 1.9.13 to 2.x in hadoop-maven-plugins
> -
>
> Key: HADOOP-13842
> URL: https://issues.apache.org/jira/browse/HADOOP-13842
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: build
>Reporter: Akira 

[jira] [Commented] (HADOOP-13600) S3a rename() to copy files in a directory in parallel

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13600?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705228#comment-15705228
 ] 

Steve Loughran commented on HADOOP-13600:
-

yeah, looking at my patch I'm not sure if there's much of it worth preserving. 
I'd realised it had got over complex. How about you sync your patch up with 
HADOOP-13823 and we start from there. Elsewhere (HADOOP-13786) I'm doing a 
special committer for s3, which is going to directly call innerRename() & so 
relay failures up immediately.

Failures on renames is going to be fun;
#  I can think of a test for that but it will be fairly brittle: create two 
large files, have a thread size of 1 for the copies, delete the 2nd while the 
first copy is in progress but after the listing has started. Or: start playing 
with mocks via the plug in AWS client code; have something which explicitly 
triggers failures on copy requests.
# I do want at least the first copy failure to be propagated up to the caller; 
usually it is that first failure which is most interesting.

> S3a rename() to copy files in a directory in parallel
> -
>
> Key: HADOOP-13600
> URL: https://issues.apache.org/jira/browse/HADOOP-13600
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.7.3
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>
> Currently a directory rename does a one-by-one copy, making the request 
> O(files * data). If the copy operations were launched in parallel, the 
> duration of the copy may be reducable to the duration of the longest copy. 
> For a directory with many files, this will be significant



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13332) Remove jackson 1.9.13 and switch all jackson code to 2.x code line

2016-11-29 Thread Akira Ajisaka (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13332?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705216#comment-15705216
 ] 

Akira Ajisaka commented on HADOOP-13332:


Filed HADOOP-13842 and attached a patch there.
Now jersey-json, avro, hbase-server, and hbase-client depend on jackson 1.9.x. 
We cannot remove the setting in hadoop-project pom.xml.
{noformat:title=mvn dependency:tree}
[INFO] +- com.sun.jersey:jersey-json:jar:1.19:compile
[INFO] |  +- org.codehaus.jettison:jettison:jar:1.1:compile
[INFO] |  +- com.sun.xml.bind:jaxb-impl:jar:2.2.3-1:compile
[INFO] |  |  \- javax.xml.bind:jaxb-api:jar:2.2.11:compile (version managed 
from 2.2.2)
[INFO] |  +- org.codehaus.jackson:jackson-core-asl:jar:1.9.13:compile (version 
managed from 1.9.2)
[INFO] |  +- org.codehaus.jackson:jackson-mapper-asl:jar:1.9.13:compile 
(version managed from 1.9.2)
[INFO] |  +- org.codehaus.jackson:jackson-jaxrs:jar:1.9.13:compile (version 
managed from 1.9.2)
[INFO] |  \- org.codehaus.jackson:jackson-xc:jar:1.9.13:compile (version 
managed from 1.9.2)
{noformat}

> Remove jackson 1.9.13 and switch all jackson code to 2.x code line
> --
>
> Key: HADOOP-13332
> URL: https://issues.apache.org/jira/browse/HADOOP-13332
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: build
>Affects Versions: 2.8.0
>Reporter: PJ Fanning
>Assignee: Akira Ajisaka
> Attachments: HADOOP-13332.00.patch, HADOOP-13332.01.patch, 
> HADOOP-13332.02.patch, HADOOP-13332.03.patch
>
>
> This jackson 1.9 code line is no longer maintained. Upgrade
> Most changes from jackson 1.9 to 2.x just involve changing the package name.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13842) Update jackson from 1.9.13 to 2.x in hadoop-maven-plugins

2016-11-29 Thread Akira Ajisaka (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13842?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira Ajisaka updated HADOOP-13842:
---
Priority: Minor  (was: Major)

> Update jackson from 1.9.13 to 2.x in hadoop-maven-plugins
> -
>
> Key: HADOOP-13842
> URL: https://issues.apache.org/jira/browse/HADOOP-13842
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: build
>Reporter: Akira Ajisaka
>Assignee: Akira Ajisaka
>Priority: Minor
> Attachments: HADOOP-13842.01.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13842) Update jackson from 1.9.13 to 2.x in hadoop-maven-plugins

2016-11-29 Thread Akira Ajisaka (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13842?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira Ajisaka updated HADOOP-13842:
---
Status: Patch Available  (was: Open)

> Update jackson from 1.9.13 to 2.x in hadoop-maven-plugins
> -
>
> Key: HADOOP-13842
> URL: https://issues.apache.org/jira/browse/HADOOP-13842
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: build
>Reporter: Akira Ajisaka
>Assignee: Akira Ajisaka
> Attachments: HADOOP-13842.01.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13842) Update jackson from 1.9.13 to 2.x in hadoop-maven-plugins

2016-11-29 Thread Akira Ajisaka (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13842?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira Ajisaka updated HADOOP-13842:
---
Attachment: HADOOP-13842.01.patch

> Update jackson from 1.9.13 to 2.x in hadoop-maven-plugins
> -
>
> Key: HADOOP-13842
> URL: https://issues.apache.org/jira/browse/HADOOP-13842
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: build
>Reporter: Akira Ajisaka
>Assignee: Akira Ajisaka
> Attachments: HADOOP-13842.01.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-13842) Update jackson from 1.9.13 to 2.x in hadoop-maven-plugins

2016-11-29 Thread Akira Ajisaka (JIRA)
Akira Ajisaka created HADOOP-13842:
--

 Summary: Update jackson from 1.9.13 to 2.x in hadoop-maven-plugins
 Key: HADOOP-13842
 URL: https://issues.apache.org/jira/browse/HADOOP-13842
 Project: Hadoop Common
  Issue Type: Sub-task
Reporter: Akira Ajisaka
Assignee: Akira Ajisaka






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-13841) S3a auth exception to link to a wiki page on the problem

2016-11-29 Thread Steve Loughran (JIRA)
Steve Loughran created HADOOP-13841:
---

 Summary: S3a auth exception to link to a wiki page on the problem
 Key: HADOOP-13841
 URL: https://issues.apache.org/jira/browse/HADOOP-13841
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 2.8.0
Reporter: Steve Loughran
Priority: Minor


Currently when there's an auth problem, the user gets a stack trace followed by 
"Invalid arguments: Unable to load AWS credentials from any provider in the 
chain"

this is explained in the AWS troubleshooting docs, but unless the user knows 
about that, they aren't going to make much progress. Better to have 
{{S3AUtils.translateException()}} point to a wiki page on the topic



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13812) Upgrade Tomcat to 6.0.48

2016-11-29 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13812?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705118#comment-15705118
 ] 

Hadoop QA commented on HADOOP-13812:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
18s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
1s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red}  0m  
0s{color} | {color:red} The patch doesn't appear to include any new or modified 
tests. Please justify why no new tests are needed for this patch. Also please 
list what manual steps were performed to verify this patch. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
15s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 12m 
48s{color} | {color:green} branch-2.6 passed {color} |
| {color:red}-1{color} | {color:red} compile {color} | {color:red}  1m 
42s{color} | {color:red} root in branch-2.6 failed with JDK v1.8.0_111. {color} 
|
| {color:red}-1{color} | {color:red} compile {color} | {color:red}  1m 
48s{color} | {color:red} root in branch-2.6 failed with JDK v1.7.0_121. {color} 
|
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m  
2s{color} | {color:green} branch-2.6 passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
53s{color} | {color:green} branch-2.6 passed {color} |
| {color:red}-1{color} | {color:red} javadoc {color} | {color:red}  0m 
12s{color} | {color:red} hadoop-hdfs-httpfs in branch-2.6 failed with JDK 
v1.8.0_111. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
35s{color} | {color:green} branch-2.6 passed with JDK v1.7.0_121 {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
13s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
22s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} compile {color} | {color:red}  1m 
34s{color} | {color:red} root in the patch failed with JDK v1.8.0_111. {color} |
| {color:red}-1{color} | {color:red} javac {color} | {color:red}  1m 34s{color} 
| {color:red} root in the patch failed with JDK v1.8.0_111. {color} |
| {color:red}-1{color} | {color:red} compile {color} | {color:red}  1m 
45s{color} | {color:red} root in the patch failed with JDK v1.7.0_121. {color} |
| {color:red}-1{color} | {color:red} javac {color} | {color:red}  1m 45s{color} 
| {color:red} root in the patch failed with JDK v1.7.0_121. {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
48s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
45s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 2048 line(s) that end in whitespace. Use 
git apply --whitespace=fix <>. Refer 
https://git-scm.com/docs/git-apply {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m 
55s{color} | {color:red} The patch 35 line(s) with tabs. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
0s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:red}-1{color} | {color:red} javadoc {color} | {color:red}  0m 
11s{color} | {color:red} hadoop-hdfs-httpfs in the patch failed with JDK 
v1.8.0_111. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
32s{color} | {color:green} the patch passed with JDK v1.7.0_121 {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  0m  
8s{color} | {color:green} hadoop-project in the patch passed with JDK 
v1.7.0_121. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  6m  
7s{color} | {color:green} hadoop-common in the patch passed with JDK 
v1.7.0_121. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  1m 
22s{color} | {color:green} hadoop-kms in the patch passed with JDK v1.7.0_121. 
{color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  3m 
13s{color} | {color:green} hadoop-hdfs-httpfs in the patch passed with JDK 
v1.7.0_121. {color} |
| {color:red}-1{color} | {color:red} asflicense {color} | {color:red}  0m 
32s{color} | {color:red} The patch generated 164 ASF License warnings. {color} |
| {color:black}{color} | {color:black} {color} | 

[jira] [Commented] (HADOOP-13332) Remove jackson 1.9.13 and switch all jackson code to 2.x code line

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13332?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705112#comment-15705112
 ] 

Steve Loughran commented on HADOOP-13332:
-

and there was me thinking you were done...

> Remove jackson 1.9.13 and switch all jackson code to 2.x code line
> --
>
> Key: HADOOP-13332
> URL: https://issues.apache.org/jira/browse/HADOOP-13332
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: build
>Affects Versions: 2.8.0
>Reporter: PJ Fanning
>Assignee: Akira Ajisaka
> Attachments: HADOOP-13332.00.patch, HADOOP-13332.01.patch, 
> HADOOP-13332.02.patch, HADOOP-13332.03.patch
>
>
> This jackson 1.9 code line is no longer maintained. Upgrade
> Most changes from jackson 1.9 to 2.x just involve changing the package name.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13840) Implement getUsed() for ViewFileSystem

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705110#comment-15705110
 ] 

Steve Loughran commented on HADOOP-13840:
-

can you switch to {{assertEquals()}} for the equality test in the assertion; it 
provides meaningful details on a failure *provided the expected value comes 
first(

{code}
assertEquals("space used", usedSpceByPathViaTargetFs ,usedSpcByPathViaViewFs)
{code}

> Implement getUsed() for ViewFileSystem
> --
>
> Key: HADOOP-13840
> URL: https://issues.apache.org/jira/browse/HADOOP-13840
> Project: Hadoop Common
>  Issue Type: Task
>  Components: viewfs
>Affects Versions: 3.0.0-alpha1
>Reporter: Manoj Govindassamy
>Assignee: Manoj Govindassamy
> Attachments: HADOOP-13840.01.patch
>
>
> ViewFileSystem doesn't override {{FileSystem#getUSed()}}. So, when file 
> system used space is queried for slash root "/" paths, the default 
> implementations tries to run the {{getContentSummary}} which crashes on 
> seeing unexpected mount points under slash. 
> ViewFileSystem#getUsed() is not expected to collate all the space used from 
> all the mount points configured under "/". Proposal is to avoid invoking 
> FileSystem#getUsed() and throw NotInMountPointException until LinkMergeSlash 
> is supported.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13826) S3A Deadlock in multipart copy due to thread pool limits.

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13826?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705101#comment-15705101
 ] 

Steve Loughran commented on HADOOP-13826:
-

they're kept consistent for a reason, not just because it simplifies 
configuration in general.

S3 stores separate parts in separate files; you should get better performance 
when reading parts separately; that is, for max speed you should set s3a  block 
size == upload partition size. By doing a copy with copy partition size == 
upload part size, we hope to preserve that performance on later reads. Who 
knows, maybe it will even help copy performance.

what would be ideal would be to know the part size of an object; HADOOP-13261 
proposed adding a custom header for this. However, time spent looking at split 
calculation performance has convinced me that a new header would be useless 
there; the overhead of querying the objects makes it too expensive. We could 
start uploading it though, and maybe use it for a copy. still expensive though; 
a 400mS HEAD would be about 2MB of copy bandwidth based on my (ad-hoc) 
measurements of copy B/W of 6 MB/s

> S3A Deadlock in multipart copy due to thread pool limits.
> -
>
> Key: HADOOP-13826
> URL: https://issues.apache.org/jira/browse/HADOOP-13826
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.7.3
>Reporter: Sean Mackrory
> Attachments: HADOOP-13826.001.patch, HADOOP-13826.002.patch
>
>
> In testing HIVE-15093 we have encountered deadlocks in the s3a connector. The 
> TransferManager javadocs 
> (http://docs.aws.amazon.com/AWSJavaSDK/latest/javadoc/com/amazonaws/services/s3/transfer/TransferManager.html)
>  explain how this is possible:
> {quote}It is not recommended to use a single threaded executor or a thread 
> pool with a bounded work queue as control tasks may submit subtasks that 
> can't complete until all sub tasks complete. Using an incorrectly configured 
> thread pool may cause a deadlock (I.E. the work queue is filled with control 
> tasks that can't finish until subtasks complete but subtasks can't execute 
> because the queue is filled).{quote}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13261) save partition split size on multipart uploads

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705095#comment-15705095
 ] 

Steve Loughran commented on HADOOP-13261:
-

it'd be expensive to query though, as you do not get this information back on a 
LIST; you'd need to do a HEAD on the file —way too expensive to use in split 
calculation. 

Where it could be used is in copy and perhaps an s3-specific distcp, where the 
partition size could be propagated

> save partition split size on multipart uploads
> --
>
> Key: HADOOP-13261
> URL: https://issues.apache.org/jira/browse/HADOOP-13261
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0
>Reporter: Steve Loughran
>Priority: Minor
>
> On multipart uploads, save the split size as a metadata value. This would 
> allow split calculation optimized for the partitions to be performed in some 
> bulk operation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-13811) s3a: getFileStatus fails with com.amazonaws.AmazonClientException: Failed to sanitize XML document destined for handler class

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13811?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705056#comment-15705056
 ] 

Steve Loughran edited comment on HADOOP-13811 at 11/29/16 11:45 AM:


you are getting an exception on line 248 of S3a, where it's calling {{partSize 
= conf.getLong(MULTIPART_SIZE, DEFAULT_MULTIPART_SIZE)}}.

Now, that line only exists on branch-2.7.x; it's not on the branch-2 line, 
where we use {{getLongBytes()}} to support byte unit measurement s (M, K, P, T, 
G). Which means that somehow you are getting to 2.7 binaries in your jobs.

You could make that error move off that line by changing the option 
{{fs.s3a.multipart.size}} to {{104857600}}. But you've still got that mix of 
binaries there, and that means your problem isn't going to go away.


was (Author: ste...@apache.org):
you are getting an exception on line 248 of S3a, where it's calling {{partSize 
= conf.getLong(MULTIPART_SIZE, DEFAULT_MULTIPART_SIZE)}}.

Now, that line only exists on branch-2.7.x; it's not on the branch-2 line, 
where we use getLongBytes() to support byte unit measurement s (M, K, P, T, G). 
Which means that somehow you are getting to 2.7 binaries in your jobs.

You could make that error move off that line by changing the option 
{{fs.s3a.multipart.size}} to {{104857600}}. But you've still got that mix of 
binaries there, and that means your problem isn't going to go away.

> s3a: getFileStatus fails with com.amazonaws.AmazonClientException: Failed to 
> sanitize XML document destined for handler class
> -
>
> Key: HADOOP-13811
> URL: https://issues.apache.org/jira/browse/HADOOP-13811
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0, 2.7.3
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>
> Sometimes, occasionally, getFileStatus() fails with a stack trace starting 
> with {{com.amazonaws.AmazonClientException: Failed to sanitize XML document 
> destined for handler class}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13811) s3a: getFileStatus fails with com.amazonaws.AmazonClientException: Failed to sanitize XML document destined for handler class

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13811?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15705056#comment-15705056
 ] 

Steve Loughran commented on HADOOP-13811:
-

you are getting an exception on line 248 of S3a, where it's calling {{partSize 
= conf.getLong(MULTIPART_SIZE, DEFAULT_MULTIPART_SIZE)}}.

Now, that line only exists on branch-2.7.x; it's not on the branch-2 line, 
where we use getLongBytes() to support byte unit measurement s (M, K, P, T, G). 
Which means that somehow you are getting to 2.7 binaries in your jobs.

You could make that error move off that line by changing the option 
{{fs.s3a.multipart.size}} to {{104857600}}. But you've still got that mix of 
binaries there, and that means your problem isn't going to go away.

> s3a: getFileStatus fails with com.amazonaws.AmazonClientException: Failed to 
> sanitize XML document destined for handler class
> -
>
> Key: HADOOP-13811
> URL: https://issues.apache.org/jira/browse/HADOOP-13811
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.8.0, 2.7.3
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>
> Sometimes, occasionally, getFileStatus() fails with a stack trace starting 
> with {{com.amazonaws.AmazonClientException: Failed to sanitize XML document 
> destined for handler class}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13812) Upgrade Tomcat to 6.0.48

2016-11-29 Thread Akira Ajisaka (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13812?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira Ajisaka updated HADOOP-13812:
---
   Resolution: Fixed
 Hadoop Flags: Incompatible change,Reviewed  (was: Incompatible change)
Fix Version/s: 2.6.6
   3.0.0-alpha2
   2.7.4
   2.8.0
   Status: Resolved  (was: Patch Available)

Committed this to branch-2.6 as well. Thanks [~jzhuge] for the contribution!

> Upgrade Tomcat to 6.0.48
> 
>
> Key: HADOOP-13812
> URL: https://issues.apache.org/jira/browse/HADOOP-13812
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 2.6.0
>Reporter: John Zhuge
>Assignee: John Zhuge
>Priority: Blocker
> Fix For: 2.8.0, 2.7.4, 3.0.0-alpha2, 2.6.6
>
> Attachments: HADOOP-13812.001.patch, HADOOP-13812.branch-2.6.patch
>
>
> KMS and HttpFS currently uses Tomcat 6.0.44, propose to upgrade to the latest 
> version is 6.0.48.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13812) Upgrade Tomcat to 6.0.48

2016-11-29 Thread Akira Ajisaka (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13812?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira Ajisaka updated HADOOP-13812:
---
Attachment: HADOOP-13812.branch-2.6.patch

Committed this to trunk, branch-2, branch-2.8, branch-2.7.
Now cherry-picking to branch-2.6. I had to rebase the patch for branch-2.6, so 
uploading this.

> Upgrade Tomcat to 6.0.48
> 
>
> Key: HADOOP-13812
> URL: https://issues.apache.org/jira/browse/HADOOP-13812
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 2.6.0
>Reporter: John Zhuge
>Assignee: John Zhuge
>Priority: Blocker
> Attachments: HADOOP-13812.001.patch, HADOOP-13812.branch-2.6.patch
>
>
> KMS and HttpFS currently uses Tomcat 6.0.44, propose to upgrade to the latest 
> version is 6.0.48.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13794) JSON.org license is now CatX

2016-11-29 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15704983#comment-15704983
 ] 

Steve Loughran commented on HADOOP-13794:
-

Andrew Wang has pointed to: 
https://lists.apache.org/thread.html/bb18f942ce7eb83c11438303c818b885810fb76385979490366720d5@%3Clegal-discuss.apache.org%3E

bq. If you have been using it, and have done so in a *release*, AND there has 
been NO pushback from your community/eco-system, you have a temporary exclusion 
from the Cat-X classification thru April 30, 2017. At that point in time, ANY 
and ALL usage of these JSON licensed artifacts are DISALLOWED. You must either 
find a suitably licensed replacement, or do without. There will be NO 
exceptions.

This means that Hadoop 2.6-2.8 can continue to ship the older libs —with that 
disclaimer; with the jackson/aws SDK updates there's no issue in 2.9+ trunk. 
This gives us some headroom to come up with some shading.

What it doesn't do is let projects downstream, which weren't bundling the AWS 
SDK, to start doing so.

> JSON.org license is now CatX
> 
>
> Key: HADOOP-13794
> URL: https://issues.apache.org/jira/browse/HADOOP-13794
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.8.0, 2.7.4, 3.0.0-alpha2, 2.6.6
>Reporter: Sean Busbey
>Priority: Blocker
>
> per [update resolved legal|http://www.apache.org/legal/resolved.html#json]:
> {quote}
> CAN APACHE PRODUCTS INCLUDE WORKS LICENSED UNDER THE JSON LICENSE?
> No. As of 2016-11-03 this has been moved to the 'Category X' license list. 
> Prior to this, use of the JSON Java library was allowed. See Debian's page 
> for a list of alternatives.
> {quote}
> We have a test-time transitive dependency on the {{org.json:json}} artifact 
> in trunk and branch-2. AFAICT, this test time dependency doesn't get exposed 
> to downstream at all (I checked assemblies and test-jar artifacts we publish 
> to maven), so it can be removed or kept at our leisure. keeping it risks it 
> being promoted out of test scope by maven without us noticing. We might be 
> able to add an enforcer rule to check for this.
> We also distribute it in bundled form through our use of the AWS Java SDK 
> artifacts in trunk and branch-2. Looking at the github project, [their 
> dependency on JSON.org was removed in 
> 1.11|https://github.com/aws/aws-sdk-java/pull/417], so if we upgrade to 
> 1.11.0+ we should be good to go. (this might be hard in branch-2.6 and 
> branch-2.7 where we're on 1.7.4)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13812) Upgrade Tomcat to 6.0.48

2016-11-29 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13812?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15704967#comment-15704967
 ] 

Hudson commented on HADOOP-13812:
-

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #10905 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/10905/])
HADOOP-13812. Upgrade Tomcat to 6.0.48. Contributed by John Zhuge. (aajisaka: 
rev 00096dcc0c23af38e697cd7c6c29e3f00d744e54)
* (edit) hadoop-common-project/hadoop-kms/src/site/markdown/index.md.vm
* (edit) 
hadoop-hdfs-project/hadoop-hdfs-httpfs/src/site/markdown/ServerSetup.md.vm
* (edit) hadoop-project/pom.xml


> Upgrade Tomcat to 6.0.48
> 
>
> Key: HADOOP-13812
> URL: https://issues.apache.org/jira/browse/HADOOP-13812
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 2.6.0
>Reporter: John Zhuge
>Assignee: John Zhuge
>Priority: Blocker
> Attachments: HADOOP-13812.001.patch
>
>
> KMS and HttpFS currently uses Tomcat 6.0.44, propose to upgrade to the latest 
> version is 6.0.48.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org