[jira] [Commented] (HADOOP-14521) KMS client needs retry logic

2017-06-12 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14521?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047443#comment-16047443
 ] 

Hadoop QA commented on HADOOP-14521:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
18s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 2 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
18s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 15m 
42s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 13m 
21s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
54s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m  
3s{color} | {color:green} trunk passed {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  1m 
24s{color} | {color:red} hadoop-common-project/hadoop-common in trunk has 19 
extant Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
39s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
16s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
31s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 10m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 10m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
54s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m  
7s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
1s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
46s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  8m  
4s{color} | {color:green} hadoop-common in the patch passed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 63m 29s{color} 
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
42s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}153m 47s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.hdfs.TestAclsEndToEnd |
|   | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure070 |
|   | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure150 |
|   | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure080 |
|   | hadoop.hdfs.TestDFSStripedInputStreamWithRandomECPolicy |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:14b5c93 |
| JIRA Issue | HADOOP-14521 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12872794/HDFS-11804-trunk-7.patch
 |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  xml  |
| uname | Linux da44da7a9e61 3.13.0-116-generic #163-Ubuntu SMP Fri Mar 31 
14:13:22 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / bec79ca |
| Default Java | 1.8.0_131 |
| findbugs | v3.1.0-RC1 |
| findbugs | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12525/artifact/patchprocess/branch-findbugs-hadoop-common-project_hadoop-common-warnings.html
 |
| unit | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12525/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt
 |
|  

[jira] [Commented] (HADOOP-14503) Make RollingAverages a mutable metric

2017-06-12 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14503?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047393#comment-16047393
 ] 

Hudson commented on HADOOP-14503:
-

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #11862 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/11862/])
HADOOP-14503. Make RollingAverages a mutable metric. Contributed by (arp: rev 
8633ef8e10a78883fbd6bf197007dc5191bf4535)
* (add) 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MutableRollingAverages.java
* (delete) 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/RollingAverages.java
* (add) 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/lib/MetricsTestHelper.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
* (edit) 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MutableMetricsFactory.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/metrics/TestDataNodeOutlierDetectionViaMetrics.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tools/TestHdfsConfigFields.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDataNodePeerMetrics.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java
* (delete) 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/lib/TestRollingAverages.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/metrics/DataNodePeerMetrics.java
* (add) 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/lib/TestMutableRollingAverages.java
* (edit) 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MetricsRegistry.java


> Make RollingAverages a mutable metric
> -
>
> Key: HADOOP-14503
> URL: https://issues.apache.org/jira/browse/HADOOP-14503
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Reporter: Hanisha Koneru
>Assignee: Hanisha Koneru
> Fix For: 3.0.0-alpha4
>
> Attachments: HADOOP-14503.001.patch, HADOOP-14503.002.patch, 
> HADOOP-14503.003.patch, HADOOP-14503.004.patch, HADOOP-14503.005.patch, 
> HADOOP-14503.006.patch, HADOOP-14503.007.patch
>
>
> RollingAverages metric extends on MutableRatesWithAggregation metric and 
> maintains a group of rolling average metrics. This class should be allowed to 
> register as a metric with the MetricSystem.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14469) FTPFileSystem#listStatus get currentPath and parentPath at the same time, causing recursively list action endless

2017-06-12 Thread Hongyuan Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14469?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongyuan Li updated HADOOP-14469:
-
Summary: FTPFileSystem#listStatus get currentPath and parentPath at the 
same time, causing recursively list action endless  (was: 
FTPFileSystem#listStatus get currentPath and parentPath at the same time, 
causing recursively list action stuck)

> FTPFileSystem#listStatus get currentPath and parentPath at the same time, 
> causing recursively list action endless
> -
>
> Key: HADOOP-14469
> URL: https://issues.apache.org/jira/browse/HADOOP-14469
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs, tools/distcp
>Affects Versions: 3.0.0-alpha3
> Environment: ftp build by windows7 + Serv-U_64 12.1.0.8 
> code runs any os
>Reporter: Hongyuan Li
>Assignee: Hongyuan Li
>Priority: Critical
> Attachments: HADOOP-14469-001.patch, HADOOP-14469-002.patch, 
> HADOOP-14469-003.patch
>
>
> for some ftpsystems, liststatus method will return new Path(".") and new 
> Path(".."), thus causing list op looping.for example, Serv-U
> We can see the logic in code below:
> {code}
>   private FileStatus[] listStatus(FTPClient client, Path file)
>   throws IOException {
> ……
> FileStatus[] fileStats = new FileStatus[ftpFiles.length];
> for (int i = 0; i < ftpFiles.length; i++) {
>   fileStats[i] = getFileStatus(ftpFiles[i], absolute);
> }
> return fileStats;
>   }
> {code}
> {code}
> public void test() throws Exception{
> FTPFileSystem ftpFileSystem = new FTPFileSystem();
> ftpFileSystem.initialize(new 
> Path("ftp://test:123456@192.168.44.1/;).toUri(),
> new Configuration());
> FileStatus[] fileStatus  = ftpFileSystem.listStatus(new Path("/new"));
> for(FileStatus fileStatus1 : fileStatus)
>   System.out.println(fileStatus1);
> }
> {code}
> using test code below, the test results list below
> {code}
> FileStatus{path=ftp://test:123456@192.168.44.1/new; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14431-002.patch; 
> isDirectory=false; length=2036; replication=1; blocksize=4096; 
> modification_time=149579778; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14486-001.patch; 
> isDirectory=false; length=1322; replication=1; blocksize=4096; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop-main; 
> isDirectory=true; modification_time=149579712; access_time=0; owner=user; 
> group=group; permission=-; isSymlink=false}
> {code}
> In results above, {{FileStatus{path=ftp://test:123456@192.168.44.1/new; ……}} 
> is obviously the current Path, and  
> {{FileStatus{path=ftp://test:123456@192.168.44.1/;……}}  is obviously the 
> parent Path.
> So, if we want to walk the directory recursively, it will stuck.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14469) FTPFileSystem#listStatus get currentPath and parentPath at the same time, causing recursively list action stuck

2017-06-12 Thread Hongyuan Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14469?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongyuan Li updated HADOOP-14469:
-
Description: 
for some ftpsystems, liststatus method will return new Path(".") and new 
Path(".."), thus causing list op looping.for example, Serv-U
We can see the logic in code below:

{code}
  private FileStatus[] listStatus(FTPClient client, Path file)
  throws IOException {
……
FileStatus[] fileStats = new FileStatus[ftpFiles.length];
for (int i = 0; i < ftpFiles.length; i++) {
  fileStats[i] = getFileStatus(ftpFiles[i], absolute);
}
return fileStats;
  }
{code}


{code}
public void test() throws Exception{
FTPFileSystem ftpFileSystem = new FTPFileSystem();
ftpFileSystem.initialize(new 
Path("ftp://test:123456@192.168.44.1/;).toUri(),
new Configuration());
FileStatus[] fileStatus  = ftpFileSystem.listStatus(new Path("/new"));
for(FileStatus fileStatus1 : fileStatus)
  System.out.println(fileStatus1);
}
{code}
using test code below, the test results list below
{code}
FileStatus{path=ftp://test:123456@192.168.44.1/new; isDirectory=true; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/; isDirectory=true; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop; isDirectory=true; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14431-002.patch; 
isDirectory=false; length=2036; replication=1; blocksize=4096; 
modification_time=149579778; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14486-001.patch; 
isDirectory=false; length=1322; replication=1; blocksize=4096; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop-main; 
isDirectory=true; modification_time=149579712; access_time=0; owner=user; 
group=group; permission=-; isSymlink=false}
{code}
In results above, {{FileStatus{path=ftp://test:123456@192.168.44.1/new; ……}} is 
obviously the current Path, and  
{{FileStatus{path=ftp://test:123456@192.168.44.1/;……}}  is obviously the parent 
Path.
So, if we want to walk the directory recursively, it will stuck.

  was:
for some ftpsystems, liststatus method will return new Path(".") and new 
Path(".."), thus causing list op looping.for example, Serv-U
We can see the logic in code below:

{code}
  private FileStatus[] listStatus(FTPClient client, Path file)
  throws IOException {
……
FileStatus[] fileStats = new FileStatus[ftpFiles.length];
for (int i = 0; i < ftpFiles.length; i++) {
  fileStats[i] = getFileStatus(ftpFiles[i], absolute);
}
return fileStats;
  }
{code}


{code}
public void test() throws Exception{
FTPFileSystem ftpFileSystem = new FTPFileSystem();
ftpFileSystem.initialize(new 
Path("ftp://test:123456@192.168.44.1/;).toUri(),
new Configuration());
FileStatus[] fileStatus  = ftpFileSystem.listStatus(new Path("/new"));
for(FileStatus fileStatus1 : fileStatus)
  System.out.println(fileStatus1);
}
{code}
using test code below, the test results list below
{code}
FileStatus{path=ftp://test:123456@192.168.44.1/new; isDirectory=true; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/; isDirectory=true; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop; isDirectory=true; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14431-002.patch; 
isDirectory=false; length=2036; replication=1; blocksize=4096; 
modification_time=149579778; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14486-001.patch; 
isDirectory=false; length=1322; replication=1; blocksize=4096; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop-main; 
isDirectory=true; modification_time=149579712; access_time=0; owner=user; 
group=group; permission=-; isSymlink=false}
{code}
In results above, {{FileStatus{path=ftp://test:123456@192.168.44.1/new; ……}} is 
obviously the current Path, and  

[jira] [Updated] (HADOOP-14469) FTPFileSystem#listStatus get currentPath and parentPath at the same time, causing recursively list action stuck

2017-06-12 Thread Hongyuan Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14469?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongyuan Li updated HADOOP-14469:
-
Description: 
for some ftpsystems, liststatus method will return new Path(".") and new 
Path(".."), thus causing list op looping.for example, Serv-U
We can see the logic in code below:

{code}
  private FileStatus[] listStatus(FTPClient client, Path file)
  throws IOException {
……
FileStatus[] fileStats = new FileStatus[ftpFiles.length];
for (int i = 0; i < ftpFiles.length; i++) {
  fileStats[i] = getFileStatus(ftpFiles[i], absolute);
}
return fileStats;
  }
{code}


{code}
public void test() throws Exception{
FTPFileSystem ftpFileSystem = new FTPFileSystem();
ftpFileSystem.initialize(new 
Path("ftp://test:123456@192.168.44.1/;).toUri(),
new Configuration());
FileStatus[] fileStatus  = ftpFileSystem.listStatus(new Path("/new"));
for(FileStatus fileStatus1 : fileStatus)
  System.out.println(fileStatus1);
}
{code}
using test code below, the test results list below
{code}
FileStatus{path=ftp://test:123456@192.168.44.1/new; isDirectory=true; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/; isDirectory=true; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop; isDirectory=true; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14431-002.patch; 
isDirectory=false; length=2036; replication=1; blocksize=4096; 
modification_time=149579778; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14486-001.patch; 
isDirectory=false; length=1322; replication=1; blocksize=4096; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop-main; 
isDirectory=true; modification_time=149579712; access_time=0; owner=user; 
group=group; permission=-; isSymlink=false}
{code}
In results above, {{FileStatus{path=ftp://test:123456@192.168.44.1/new; ……}} is 
obviously the current Path, and  
{{FileStatus{path=ftp://test:123456@192.168.44.1/;…… }}  is obviously the 
parent Path.
So, if we want to walk the directory recursively, it will stuck.

  was:
for some ftpsystems, liststatus method will return new Path(".") and new 
Path(".."), thus causing list op looping.for example, Serv-U
We can see the logic in code below:

{code}
  private FileStatus[] listStatus(FTPClient client, Path file)
  throws IOException {
……
FileStatus[] fileStats = new FileStatus[ftpFiles.length];
for (int i = 0; i < ftpFiles.length; i++) {
  fileStats[i] = getFileStatus(ftpFiles[i], absolute);
}
return fileStats;
  }
{code}


{code}
public void test() throws Exception{
FTPFileSystem ftpFileSystem = new FTPFileSystem();
ftpFileSystem.initialize(new 
Path("ftp://test:123456@192.168.44.1/;).toUri(),
new Configuration());
FileStatus[] fileStatus  = ftpFileSystem.listStatus(new Path("/new"));
for(FileStatus fileStatus1 : fileStatus)
  System.out.println(fileStatus1);
}
{code}
using test code below, the test results list below
{code}
FileStatus{path=ftp://test:123456@192.168.44.1/new; isDirectory=true; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/; isDirectory=true; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop; isDirectory=true; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14431-002.patch; 
isDirectory=false; length=2036; replication=1; blocksize=4096; 
modification_time=149579778; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14486-001.patch; 
isDirectory=false; length=1322; replication=1; blocksize=4096; 
modification_time=149671698; access_time=0; owner=user; group=group; 
permission=-; isSymlink=false}
FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop-main; 
isDirectory=true; modification_time=149579712; access_time=0; owner=user; 
group=group; permission=-; isSymlink=false}
{code}
In results above, {{FileStatus{path=ftp://test:123456@192.168.44.1/new; ……}} is 
obviously the current Path, and 

[jira] [Updated] (HADOOP-14503) Make RollingAverages a mutable metric

2017-06-12 Thread Arpit Agarwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14503?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arpit Agarwal updated HADOOP-14503:
---
   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 3.0.0-alpha4
   Status: Resolved  (was: Patch Available)

I've committed this and fixed the checkstyle whitespace issue. Thank you for 
the contribution [~hanishakoneru]. 

I hit some merge conflicts while applying the patch to branch-2. If you want to 
post a branch-2 patch I'd be happy to commit that too.

> Make RollingAverages a mutable metric
> -
>
> Key: HADOOP-14503
> URL: https://issues.apache.org/jira/browse/HADOOP-14503
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Reporter: Hanisha Koneru
>Assignee: Hanisha Koneru
> Fix For: 3.0.0-alpha4
>
> Attachments: HADOOP-14503.001.patch, HADOOP-14503.002.patch, 
> HADOOP-14503.003.patch, HADOOP-14503.004.patch, HADOOP-14503.005.patch, 
> HADOOP-14503.006.patch, HADOOP-14503.007.patch
>
>
> RollingAverages metric extends on MutableRatesWithAggregation metric and 
> maintains a group of rolling average metrics. This class should be allowed to 
> register as a metric with the MetricSystem.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14521) KMS client needs retry logic

2017-06-12 Thread Xiao Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14521?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047382#comment-16047382
 ] 

Xiao Chen commented on HADOOP-14521:


Thanks for sticking with this Rushabh.

bq. From this comment, point 1,
Right... sorry I missed that one.

But what confuses me is that, what if {{maxNumRetries < providers.length}} ? If 
someone configures it to be 2, and we have 3 providers, the call will fail and 
print a message:
{code}
LOG.warn("Aborting since the Request has failed with all KMS"
  + " providers in the group or the exception is not recoverable.");
{code}
which isn't entirely accurate, because here it did not fail with all providers.

We can either update the message, or update the maxNumRetries to 
providers.length to make sure we try at least once on all providers. Either way 
feels okay to me. Slightly prefer the latter since by assumption all providers 
should work and we don't have retry delay for the whole sweep.

Besides that, I'm +1 on patch 7.

> KMS client needs retry logic
> 
>
> Key: HADOOP-14521
> URL: https://issues.apache.org/jira/browse/HADOOP-14521
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.6.0
>Reporter: Rushabh S Shah
>Assignee: Rushabh S Shah
> Attachments: HDFS-11804-trunk-1.patch, HDFS-11804-trunk-2.patch, 
> HDFS-11804-trunk-3.patch, HDFS-11804-trunk-4.patch, HDFS-11804-trunk-5.patch, 
> HDFS-11804-trunk-6.patch, HDFS-11804-trunk-7.patch, HDFS-11804-trunk.patch
>
>
> The kms client appears to have no retry logic – at all.  It's completely 
> decoupled from the ipc retry logic.  This has major impacts if the KMS is 
> unreachable for any reason, including but not limited to network connection 
> issues, timeouts, the +restart during an upgrade+.
> This has some major ramifications:
> # Jobs may fail to submit, although oozie resubmit logic should mask it
> # Non-oozie launchers may experience higher rates if they do not already have 
> retry logic.
> # Tasks reading EZ files will fail, probably be masked by framework reattempts
> # EZ file creation fails after creating a 0-length file – client receives 
> EDEK in the create response, then fails when decrypting the EDEK
> # Bulk hadoop fs copies, and maybe distcp, will prematurely fail



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14469) FTPFileSystem#listStatus get currentPath and parentPath at the same time, causing recursively list action stuck

2017-06-12 Thread Hongyuan Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14469?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongyuan Li updated HADOOP-14469:
-
Summary: FTPFileSystem#listStatus get currentPath and parentPath at the 
same time, causing recursively list action stuck  (was: 
FTPFileSystem#listStatus get currentPath and parentPath at the same time, which 
will cause recursively list action stuck)

> FTPFileSystem#listStatus get currentPath and parentPath at the same time, 
> causing recursively list action stuck
> ---
>
> Key: HADOOP-14469
> URL: https://issues.apache.org/jira/browse/HADOOP-14469
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs, tools/distcp
>Affects Versions: 3.0.0-alpha3
> Environment: ftp build by windows7 + Serv-U_64 12.1.0.8 
> code runs any os
>Reporter: Hongyuan Li
>Assignee: Hongyuan Li
>Priority: Critical
> Attachments: HADOOP-14469-001.patch, HADOOP-14469-002.patch, 
> HADOOP-14469-003.patch
>
>
> for some ftpsystems, liststatus method will return new Path(".") and new 
> Path(".."), thus causing list op looping.for example, Serv-U
> We can see the logic in code below:
> {code}
>   private FileStatus[] listStatus(FTPClient client, Path file)
>   throws IOException {
> ……
> FileStatus[] fileStats = new FileStatus[ftpFiles.length];
> for (int i = 0; i < ftpFiles.length; i++) {
>   fileStats[i] = getFileStatus(ftpFiles[i], absolute);
> }
> return fileStats;
>   }
> {code}
> {code}
> public void test() throws Exception{
> FTPFileSystem ftpFileSystem = new FTPFileSystem();
> ftpFileSystem.initialize(new 
> Path("ftp://test:123456@192.168.44.1/;).toUri(),
> new Configuration());
> FileStatus[] fileStatus  = ftpFileSystem.listStatus(new Path("/new"));
> for(FileStatus fileStatus1 : fileStatus)
>   System.out.println(fileStatus1);
> }
> {code}
> using test code below, the test results list below
> {code}
> FileStatus{path=ftp://test:123456@192.168.44.1/new; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14431-002.patch; 
> isDirectory=false; length=2036; replication=1; blocksize=4096; 
> modification_time=149579778; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14486-001.patch; 
> isDirectory=false; length=1322; replication=1; blocksize=4096; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop-main; 
> isDirectory=true; modification_time=149579712; access_time=0; owner=user; 
> group=group; permission=-; isSymlink=false}
> {code}
> In results above, {{FileStatus{path=ftp://test:123456@192.168.44.1/new; ……}} 
> is obviously the current Path, and 
> {{FileStatus{path=ftp://test:123456@192.168.44.1/;…… }} is obviously the 
> parent Path.
> So, if we want to walk the directory recursively, it will stuck.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14469) FTPFileSystem#listStatus get currentPath and parentPath at the same time, which will cause recursively list action stuck

2017-06-12 Thread Hongyuan Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14469?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongyuan Li updated HADOOP-14469:
-
Summary: FTPFileSystem#listStatus get currentPath and parentPath at the 
same time, which will cause recursively list action stuck  (was: 
FTPFileSystem#listStatus get currentPath and parentPath at the same time, which 
will cause recursively list stuck)

> FTPFileSystem#listStatus get currentPath and parentPath at the same time, 
> which will cause recursively list action stuck
> 
>
> Key: HADOOP-14469
> URL: https://issues.apache.org/jira/browse/HADOOP-14469
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs, tools/distcp
>Affects Versions: 3.0.0-alpha3
> Environment: ftp build by windows7 + Serv-U_64 12.1.0.8 
> code runs any os
>Reporter: Hongyuan Li
>Assignee: Hongyuan Li
>Priority: Critical
> Attachments: HADOOP-14469-001.patch, HADOOP-14469-002.patch, 
> HADOOP-14469-003.patch
>
>
> for some ftpsystems, liststatus method will return new Path(".") and new 
> Path(".."), thus causing list op looping.for example, Serv-U
> We can see the logic in code below:
> {code}
>   private FileStatus[] listStatus(FTPClient client, Path file)
>   throws IOException {
> ……
> FileStatus[] fileStats = new FileStatus[ftpFiles.length];
> for (int i = 0; i < ftpFiles.length; i++) {
>   fileStats[i] = getFileStatus(ftpFiles[i], absolute);
> }
> return fileStats;
>   }
> {code}
> {code}
> public void test() throws Exception{
> FTPFileSystem ftpFileSystem = new FTPFileSystem();
> ftpFileSystem.initialize(new 
> Path("ftp://test:123456@192.168.44.1/;).toUri(),
> new Configuration());
> FileStatus[] fileStatus  = ftpFileSystem.listStatus(new Path("/new"));
> for(FileStatus fileStatus1 : fileStatus)
>   System.out.println(fileStatus1);
> }
> {code}
> using test code below, the test results list below
> {code}
> FileStatus{path=ftp://test:123456@192.168.44.1/new; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14431-002.patch; 
> isDirectory=false; length=2036; replication=1; blocksize=4096; 
> modification_time=149579778; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14486-001.patch; 
> isDirectory=false; length=1322; replication=1; blocksize=4096; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop-main; 
> isDirectory=true; modification_time=149579712; access_time=0; owner=user; 
> group=group; permission=-; isSymlink=false}
> {code}
> In results above, {{FileStatus{path=ftp://test:123456@192.168.44.1/new; ……}} 
> is obviously the current Path, and 
> {{FileStatus{path=ftp://test:123456@192.168.44.1/;…… }} is obviously the 
> parent Path.
> So, if we want to walk the directory recursively, it will stuck.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14299) Hadoop Renew Thread for proxy users

2017-06-12 Thread Hongyuan Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14299?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047371#comment-16047371
 ] 

Hongyuan Li commented on HADOOP-14299:
--

[~Prabhu Joseph] any updates about this?

> Hadoop Renew Thread for proxy users
> ---
>
> Key: HADOOP-14299
> URL: https://issues.apache.org/jira/browse/HADOOP-14299
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.7.1
>Reporter: Prabhu Joseph
>
> Currently Hadoop Client has a separate renew thread which is created only for 
> Authentication type Kerberos and not for Proxy. So for proxy users, a yarn 
> client monitoring a long running job will fail after initial ticket lifetime 
> with GSS initiate failed unless there is a manual re-kinit. 
> https://github.com/apache/hadoop/blob/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/UserGroupInformation.java#L1030



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14503) Make RollingAverages a mutable metric

2017-06-12 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14503?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047368#comment-16047368
 ] 

Hadoop QA commented on HADOOP-14503:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
17s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 6 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
26s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 13m 
19s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 13m 
33s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
59s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m  
3s{color} | {color:green} trunk passed {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  1m 
28s{color} | {color:red} hadoop-common-project/hadoop-common in trunk has 19 
extant Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
40s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
17s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 10m 
16s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 10m 
16s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
2m  3s{color} | {color:orange} root: The patch generated 3 new + 595 unchanged 
- 3 fixed = 598 total (was 598) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m  
6s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 4 line(s) that end in whitespace. Use git 
apply --whitespace=fix <>. Refer https://git-scm.com/docs/git-apply 
{color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
33s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
43s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  8m 
57s{color} | {color:green} hadoop-common in the patch passed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 75m 13s{color} 
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
39s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}165m 43s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure |
| Timed out junit tests | 
org.apache.hadoop.hdfs.server.blockmanagement.TestBlockStatsMXBean |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:14b5c93 |
| JIRA Issue | HADOOP-14503 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12872783/HADOOP-14503.007.patch
 |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux 1cc0f3566a54 3.13.0-116-generic #163-Ubuntu SMP Fri Mar 31 
14:13:22 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / b3d3ede |
| Default Java | 1.8.0_131 |
| findbugs | v3.1.0-RC1 |
| findbugs | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12523/artifact/patchprocess/branch-findbugs-hadoop-common-project_hadoop-common-warnings.html
 |
| checkstyle | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12523/artifact/patchprocess/diff-checkstyle-root.txt
 |
| whitespace | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12523/artifact/patchprocess/whitespace-eol.txt
 |
| unit | 

[jira] [Comment Edited] (HADOOP-14469) FTPFileSystem#listStatus get currentPath and parentPath at the same time, which will cause recursively list stuck

2017-06-12 Thread Hongyuan Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14469?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16045372#comment-16045372
 ] 

Hongyuan Li edited comment on HADOOP-14469 at 6/13/17 3:12 AM:
---

[~ste...@apache.org] added test code and environment.

ftp built by serv-U x64  12.1.0.8, whose os is win7 x64



was (Author: hongyuan li):
[~ste...@apache.org] added test code and environment.

> FTPFileSystem#listStatus get currentPath and parentPath at the same time, 
> which will cause recursively list stuck
> -
>
> Key: HADOOP-14469
> URL: https://issues.apache.org/jira/browse/HADOOP-14469
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs, tools/distcp
>Affects Versions: 3.0.0-alpha3
> Environment: ftp build by windows7 + Serv-U_64 12.1.0.8 
> code runs any os
>Reporter: Hongyuan Li
>Assignee: Hongyuan Li
>Priority: Critical
> Attachments: HADOOP-14469-001.patch, HADOOP-14469-002.patch, 
> HADOOP-14469-003.patch
>
>
> for some ftpsystems, liststatus method will return new Path(".") and new 
> Path(".."), thus causing list op looping.for example, Serv-U
> We can see the logic in code below:
> {code}
>   private FileStatus[] listStatus(FTPClient client, Path file)
>   throws IOException {
> ……
> FileStatus[] fileStats = new FileStatus[ftpFiles.length];
> for (int i = 0; i < ftpFiles.length; i++) {
>   fileStats[i] = getFileStatus(ftpFiles[i], absolute);
> }
> return fileStats;
>   }
> {code}
> {code}
> public void test() throws Exception{
> FTPFileSystem ftpFileSystem = new FTPFileSystem();
> ftpFileSystem.initialize(new 
> Path("ftp://test:123456@192.168.44.1/;).toUri(),
> new Configuration());
> FileStatus[] fileStatus  = ftpFileSystem.listStatus(new Path("/new"));
> for(FileStatus fileStatus1 : fileStatus)
>   System.out.println(fileStatus1);
> }
> {code}
> using test code below, the test results list below
> {code}
> FileStatus{path=ftp://test:123456@192.168.44.1/new; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14431-002.patch; 
> isDirectory=false; length=2036; replication=1; blocksize=4096; 
> modification_time=149579778; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14486-001.patch; 
> isDirectory=false; length=1322; replication=1; blocksize=4096; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop-main; 
> isDirectory=true; modification_time=149579712; access_time=0; owner=user; 
> group=group; permission=-; isSymlink=false}
> {code}
> In results above, {{FileStatus{path=ftp://test:123456@192.168.44.1/new; ……}} 
> is obviously the current Path, and 
> {{FileStatus{path=ftp://test:123456@192.168.44.1/;…… }} is obviously the 
> parent Path.
> So, if we want to walk the directory recursively, it will stuck.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14429) FTPFileSystem#getFsAction always returns FsAction.NONE

2017-06-12 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047358#comment-16047358
 ] 

Hadoop QA commented on HADOOP-14429:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
19s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 15m 
18s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 15m 
54s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
39s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
13s{color} | {color:green} trunk passed {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  1m 
34s{color} | {color:red} hadoop-common-project/hadoop-common in trunk has 19 
extant Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
54s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
46s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 11m 
59s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 11m 
59s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
37s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m  
9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
42s{color} | {color:green} hadoop-common-project/hadoop-common generated 0 new 
+ 18 unchanged - 1 fixed = 18 total (was 19) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
56s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red}  8m 31s{color} 
| {color:red} hadoop-common in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
40s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 64m  3s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.fs.sftp.TestSFTPFileSystem |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:14b5c93 |
| JIRA Issue | HADOOP-14429 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12872789/HADOOP-14429-005.patch
 |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux 4bb0ab8eaae2 3.13.0-116-generic #163-Ubuntu SMP Fri Mar 31 
14:13:22 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / bec79ca |
| Default Java | 1.8.0_131 |
| findbugs | v3.1.0-RC1 |
| findbugs | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12524/artifact/patchprocess/branch-findbugs-hadoop-common-project_hadoop-common-warnings.html
 |
| unit | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12524/artifact/patchprocess/patch-unit-hadoop-common-project_hadoop-common.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12524/testReport/ |
| modules | C: hadoop-common-project/hadoop-common U: 
hadoop-common-project/hadoop-common |
| Console output | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12524/console |
| Powered by | Apache Yetus 0.5.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> FTPFileSystem#getFsAction  always returns FsAction.NONE
> ---
>
> Key: HADOOP-14429
> URL: https://issues.apache.org/jira/browse/HADOOP-14429
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 3.0.0-alpha2
> 

[jira] [Comment Edited] (HADOOP-14521) KMS client needs retry logic

2017-06-12 Thread Rushabh S Shah (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14521?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047355#comment-16047355
 ] 

Rushabh S Shah edited comment on HADOOP-14521 at 6/13/17 2:52 AM:
--

bq. For the retries, do you think maybe it's more intuitive to specify num of 
failovers for the whole provider set?
>From [this comment, point 
>1|https://issues.apache.org/jira/browse/HADOOP-14521?focusedCommentId=16036988=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-16036988],
> I think [~daryn] had a good reason explaining why have fix number of retries 
>is a good idea. I kind of agreed with him. So keeping the fix number of 
>retries even in patch #7.

bq. KMSCP: the first @Link seems redundant
Fixed in v7

bq. Good to have some validations on the config params (e.g. numRetries > 0 
etc.)
Added few pre conditions checks for all the config params.

bq.I'd revert the change that breaks instead of returning.
Thanks a lot for catching that. Fixed that in v7 of the patch.


was (Author: shahrs87):
bq. For the retries, do you think maybe it's more intuitive to specify num of 
failovers for the whole provider set?
>From [this comment, point 
>1|https://issues.apache.org/jira/browse/HADOOP-14521?focusedCommentId=16036988=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-16036988],
> I think [~daryn] had a good reason explaining why have fix number of retries 
>is a good idea. I kind of agreed with him. So keeping the fix number of 
>retries even in patch #7.

bq. KMSCP: the first @Link seems redundant
Fixed in v7

bq. Good to have some validations on the config params (e.g. numRetries > 0 
etc.)
Added few pre conditions checks for all the config params.


> KMS client needs retry logic
> 
>
> Key: HADOOP-14521
> URL: https://issues.apache.org/jira/browse/HADOOP-14521
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.6.0
>Reporter: Rushabh S Shah
>Assignee: Rushabh S Shah
> Attachments: HDFS-11804-trunk-1.patch, HDFS-11804-trunk-2.patch, 
> HDFS-11804-trunk-3.patch, HDFS-11804-trunk-4.patch, HDFS-11804-trunk-5.patch, 
> HDFS-11804-trunk-6.patch, HDFS-11804-trunk-7.patch, HDFS-11804-trunk.patch
>
>
> The kms client appears to have no retry logic – at all.  It's completely 
> decoupled from the ipc retry logic.  This has major impacts if the KMS is 
> unreachable for any reason, including but not limited to network connection 
> issues, timeouts, the +restart during an upgrade+.
> This has some major ramifications:
> # Jobs may fail to submit, although oozie resubmit logic should mask it
> # Non-oozie launchers may experience higher rates if they do not already have 
> retry logic.
> # Tasks reading EZ files will fail, probably be masked by framework reattempts
> # EZ file creation fails after creating a 0-length file – client receives 
> EDEK in the create response, then fails when decrypting the EDEK
> # Bulk hadoop fs copies, and maybe distcp, will prematurely fail



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14521) KMS client needs retry logic

2017-06-12 Thread Rushabh S Shah (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14521?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rushabh S Shah updated HADOOP-14521:

Attachment: HDFS-11804-trunk-7.patch

bq. For the retries, do you think maybe it's more intuitive to specify num of 
failovers for the whole provider set?
>From [this comment, point 
>1|https://issues.apache.org/jira/browse/HADOOP-14521?focusedCommentId=16036988=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-16036988],
> I think [~daryn] had a good reason explaining why have fix number of retries 
>is a good idea. I kind of agreed with him. So keeping the fix number of 
>retries even in patch #7.

bq. KMSCP: the first @Link seems redundant
Fixed in v7

bq. Good to have some validations on the config params (e.g. numRetries > 0 
etc.)
Added few pre conditions checks for all the config params.


> KMS client needs retry logic
> 
>
> Key: HADOOP-14521
> URL: https://issues.apache.org/jira/browse/HADOOP-14521
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.6.0
>Reporter: Rushabh S Shah
>Assignee: Rushabh S Shah
> Attachments: HDFS-11804-trunk-1.patch, HDFS-11804-trunk-2.patch, 
> HDFS-11804-trunk-3.patch, HDFS-11804-trunk-4.patch, HDFS-11804-trunk-5.patch, 
> HDFS-11804-trunk-6.patch, HDFS-11804-trunk-7.patch, HDFS-11804-trunk.patch
>
>
> The kms client appears to have no retry logic – at all.  It's completely 
> decoupled from the ipc retry logic.  This has major impacts if the KMS is 
> unreachable for any reason, including but not limited to network connection 
> issues, timeouts, the +restart during an upgrade+.
> This has some major ramifications:
> # Jobs may fail to submit, although oozie resubmit logic should mask it
> # Non-oozie launchers may experience higher rates if they do not already have 
> retry logic.
> # Tasks reading EZ files will fail, probably be masked by framework reattempts
> # EZ file creation fails after creating a 0-length file – client receives 
> EDEK in the create response, then fails when decrypting the EDEK
> # Bulk hadoop fs copies, and maybe distcp, will prematurely fail



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14521) KMS client needs retry logic

2017-06-12 Thread Rushabh S Shah (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14521?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rushabh S Shah updated HADOOP-14521:

Status: Patch Available  (was: Open)

> KMS client needs retry logic
> 
>
> Key: HADOOP-14521
> URL: https://issues.apache.org/jira/browse/HADOOP-14521
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.6.0
>Reporter: Rushabh S Shah
>Assignee: Rushabh S Shah
> Attachments: HDFS-11804-trunk-1.patch, HDFS-11804-trunk-2.patch, 
> HDFS-11804-trunk-3.patch, HDFS-11804-trunk-4.patch, HDFS-11804-trunk-5.patch, 
> HDFS-11804-trunk-6.patch, HDFS-11804-trunk-7.patch, HDFS-11804-trunk.patch
>
>
> The kms client appears to have no retry logic – at all.  It's completely 
> decoupled from the ipc retry logic.  This has major impacts if the KMS is 
> unreachable for any reason, including but not limited to network connection 
> issues, timeouts, the +restart during an upgrade+.
> This has some major ramifications:
> # Jobs may fail to submit, although oozie resubmit logic should mask it
> # Non-oozie launchers may experience higher rates if they do not already have 
> retry logic.
> # Tasks reading EZ files will fail, probably be masked by framework reattempts
> # EZ file creation fails after creating a 0-length file – client receives 
> EDEK in the create response, then fails when decrypting the EDEK
> # Bulk hadoop fs copies, and maybe distcp, will prematurely fail



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14521) KMS client needs retry logic

2017-06-12 Thread Rushabh S Shah (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14521?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rushabh S Shah updated HADOOP-14521:

Status: Open  (was: Patch Available)

Thanks [~daryn] [~xiaochen] for your valuable reviews.
Canceling the patch to address your comments.


> KMS client needs retry logic
> 
>
> Key: HADOOP-14521
> URL: https://issues.apache.org/jira/browse/HADOOP-14521
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.6.0
>Reporter: Rushabh S Shah
>Assignee: Rushabh S Shah
> Attachments: HDFS-11804-trunk-1.patch, HDFS-11804-trunk-2.patch, 
> HDFS-11804-trunk-3.patch, HDFS-11804-trunk-4.patch, HDFS-11804-trunk-5.patch, 
> HDFS-11804-trunk-6.patch, HDFS-11804-trunk.patch
>
>
> The kms client appears to have no retry logic – at all.  It's completely 
> decoupled from the ipc retry logic.  This has major impacts if the KMS is 
> unreachable for any reason, including but not limited to network connection 
> issues, timeouts, the +restart during an upgrade+.
> This has some major ramifications:
> # Jobs may fail to submit, although oozie resubmit logic should mask it
> # Non-oozie launchers may experience higher rates if they do not already have 
> retry logic.
> # Tasks reading EZ files will fail, probably be masked by framework reattempts
> # EZ file creation fails after creating a 0-length file – client receives 
> EDEK in the create response, then fails when decrypting the EDEK
> # Bulk hadoop fs copies, and maybe distcp, will prematurely fail



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14470) CommandWithDestination#create used redundant ternary operator

2017-06-12 Thread Hongyuan Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14470?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongyuan Li updated HADOOP-14470:
-
Component/s: fs

> CommandWithDestination#create used redundant ternary operator  
> ---
>
> Key: HADOOP-14470
> URL: https://issues.apache.org/jira/browse/HADOOP-14470
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common, fs
>Affects Versions: 3.0.0-alpha3
>Reporter: Hongyuan Li
>Assignee: Hongyuan Li
>Priority: Trivial
> Attachments: HADOOP-14470-001.patch
>
>
> in if statement,the lazyPersist  is always true, thus the ternary operator is 
> redundant,
> {{lazyPersist == true}} in if statment, so {{lazyPersist ? 1 : 
> getDefaultReplication(item.path)}} is redundant.
>   related code like below, which is in 
> {{org.apache.hadoop.fs.shell.CommandWithDestination}}  lineNumber : 504 :
> {code:java}
>FSDataOutputStream create(PathData item, boolean lazyPersist,
> boolean direct)
> throws IOException {
>   try {
> if (lazyPersist) { // in if stament, lazyPersist is always true
>   ……
>   return create(item.path,
> FsPermission.getFileDefault().applyUMask(
> FsPermission.getUMask(getConf())),
> createFlags,
> getConf().getInt(IO_FILE_BUFFER_SIZE_KEY,
> IO_FILE_BUFFER_SIZE_DEFAULT),
> lazyPersist ? 1 : getDefaultReplication(item.path), 
> // *this is redundant*
> getDefaultBlockSize(),
> null,
> null);
> } else {
>   return create(item.path, true);
> }
>   } finally { // might have been created but stream was interrupted
> if (!direct) {
>   deleteOnExit(item.path);
> }
>   }
> }
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14429) FTPFileSystem#getFsAction always returns FsAction.NONE

2017-06-12 Thread Hongyuan Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047318#comment-16047318
 ] 

Hongyuan Li commented on HADOOP-14429:
--

[~yzhangal] resubmit the patch 005 as you have said.Added empty line added 
between different sections in the unit test.

> FTPFileSystem#getFsAction  always returns FsAction.NONE
> ---
>
> Key: HADOOP-14429
> URL: https://issues.apache.org/jira/browse/HADOOP-14429
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 3.0.0-alpha2
>Reporter: Hongyuan Li
>Assignee: Hongyuan Li
> Attachments: HADOOP-14429-001.patch, HADOOP-14429-002.patch, 
> HADOOP-14429-003.patch, HADOOP-14429-004.patch, HADOOP-14429-005.patch
>
>
>   
> {code}
> private FsAction getFsAction(int accessGroup, FTPFile ftpFile) {
>   FsAction action = FsAction.NONE;
>   if (ftpFile.hasPermission(accessGroup, FTPFile.READ_PERMISSION)) {
>   action.or(FsAction.READ);
>   }
> if (ftpFile.hasPermission(accessGroup, FTPFile.WRITE_PERMISSION)) {
>   action.or(FsAction.WRITE);
> }
> if (ftpFile.hasPermission(accessGroup, FTPFile.EXECUTE_PERMISSION)) {
>   action.or(FsAction.EXECUTE);
> }
> return action;
>   }
> {code}
> from code above, we can see that the  getFsAction method doesnot modify the 
> action generated by FsAction action = FsAction.NONE,which means it return 
> FsAction.NONE all the time;



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14429) FTPFileSystem#getFsAction always returns FsAction.NONE

2017-06-12 Thread Hongyuan Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14429?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongyuan Li updated HADOOP-14429:
-
Attachment: HADOOP-14429-005.patch

> FTPFileSystem#getFsAction  always returns FsAction.NONE
> ---
>
> Key: HADOOP-14429
> URL: https://issues.apache.org/jira/browse/HADOOP-14429
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 3.0.0-alpha2
>Reporter: Hongyuan Li
>Assignee: Hongyuan Li
> Attachments: HADOOP-14429-001.patch, HADOOP-14429-002.patch, 
> HADOOP-14429-003.patch, HADOOP-14429-004.patch, HADOOP-14429-005.patch
>
>
>   
> {code}
> private FsAction getFsAction(int accessGroup, FTPFile ftpFile) {
>   FsAction action = FsAction.NONE;
>   if (ftpFile.hasPermission(accessGroup, FTPFile.READ_PERMISSION)) {
>   action.or(FsAction.READ);
>   }
> if (ftpFile.hasPermission(accessGroup, FTPFile.WRITE_PERMISSION)) {
>   action.or(FsAction.WRITE);
> }
> if (ftpFile.hasPermission(accessGroup, FTPFile.EXECUTE_PERMISSION)) {
>   action.or(FsAction.EXECUTE);
> }
> return action;
>   }
> {code}
> from code above, we can see that the  getFsAction method doesnot modify the 
> action generated by FsAction action = FsAction.NONE,which means it return 
> FsAction.NONE all the time;



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14486) TestSFTPFileSystem#testGetAccessTime test failure using openJDk 1.8.0

2017-06-12 Thread Hongyuan Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14486?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongyuan Li updated HADOOP-14486:
-
Component/s: fs

> TestSFTPFileSystem#testGetAccessTime test failure using openJDk 1.8.0  
> ---
>
> Key: HADOOP-14486
> URL: https://issues.apache.org/jira/browse/HADOOP-14486
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 3.0.0-alpha4
> Environment: Ubuntu 14.04 
> x86, ppc64le
> $ java -version
> openjdk version "1.8.0_111"
> OpenJDK Runtime Environment (build 1.8.0_111-8u111-b14-3~14.04.1-b14)
> OpenJDK 64-Bit Server VM (build 25.111-b14, mixed mode)
>Reporter: Sonia Garudi
>Assignee: Hongyuan Li
> Attachments: HADOOP-14486-001.patch
>
>
> The TestSFTPFileSystem#testGetAccessTime test fails consistently with the 
> error below:
> {code}
> java.lang.AssertionError: expected:<1496496040072> but was:<149649604>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.fs.sftp.TestSFTPFileSystem.testGetAccessTime(TestSFTPFileSystem.java:319)
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14469) FTPFileSystem#listStatus get currentPath and parentPath at the same time, which will cause recursively list stuck

2017-06-12 Thread Hongyuan Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14469?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongyuan Li updated HADOOP-14469:
-
Component/s: tools/distcp

> FTPFileSystem#listStatus get currentPath and parentPath at the same time, 
> which will cause recursively list stuck
> -
>
> Key: HADOOP-14469
> URL: https://issues.apache.org/jira/browse/HADOOP-14469
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs, tools/distcp
>Affects Versions: 3.0.0-alpha3
> Environment: ftp build by windows7 + Serv-U_64 12.1.0.8 
> code runs any os
>Reporter: Hongyuan Li
>Assignee: Hongyuan Li
>Priority: Critical
> Attachments: HADOOP-14469-001.patch, HADOOP-14469-002.patch, 
> HADOOP-14469-003.patch
>
>
> for some ftpsystems, liststatus method will return new Path(".") and new 
> Path(".."), thus causing list op looping.for example, Serv-U
> We can see the logic in code below:
> {code}
>   private FileStatus[] listStatus(FTPClient client, Path file)
>   throws IOException {
> ……
> FileStatus[] fileStats = new FileStatus[ftpFiles.length];
> for (int i = 0; i < ftpFiles.length; i++) {
>   fileStats[i] = getFileStatus(ftpFiles[i], absolute);
> }
> return fileStats;
>   }
> {code}
> {code}
> public void test() throws Exception{
> FTPFileSystem ftpFileSystem = new FTPFileSystem();
> ftpFileSystem.initialize(new 
> Path("ftp://test:123456@192.168.44.1/;).toUri(),
> new Configuration());
> FileStatus[] fileStatus  = ftpFileSystem.listStatus(new Path("/new"));
> for(FileStatus fileStatus1 : fileStatus)
>   System.out.println(fileStatus1);
> }
> {code}
> using test code below, the test results list below
> {code}
> FileStatus{path=ftp://test:123456@192.168.44.1/new; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14431-002.patch; 
> isDirectory=false; length=2036; replication=1; blocksize=4096; 
> modification_time=149579778; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14486-001.patch; 
> isDirectory=false; length=1322; replication=1; blocksize=4096; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop-main; 
> isDirectory=true; modification_time=149579712; access_time=0; owner=user; 
> group=group; permission=-; isSymlink=false}
> {code}
> In results above, {{FileStatus{path=ftp://test:123456@192.168.44.1/new; ……}} 
> is obviously the current Path, and 
> {{FileStatus{path=ftp://test:123456@192.168.44.1/;…… }} is obviously the 
> parent Path.
> So, if we want to walk the directory recursively, it will stuck.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14486) TestSFTPFileSystem#testGetAccessTime test failure using openJDK 1.8.0

2017-06-12 Thread Hongyuan Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14486?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongyuan Li updated HADOOP-14486:
-
Summary: TestSFTPFileSystem#testGetAccessTime test failure using openJDK 
1.8.0(was: TestSFTPFileSystem#testGetAccessTime test failure using openJDk 
1.8.0  )

> TestSFTPFileSystem#testGetAccessTime test failure using openJDK 1.8.0  
> ---
>
> Key: HADOOP-14486
> URL: https://issues.apache.org/jira/browse/HADOOP-14486
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 3.0.0-alpha4
> Environment: Ubuntu 14.04 
> x86, ppc64le
> $ java -version
> openjdk version "1.8.0_111"
> OpenJDK Runtime Environment (build 1.8.0_111-8u111-b14-3~14.04.1-b14)
> OpenJDK 64-Bit Server VM (build 25.111-b14, mixed mode)
>Reporter: Sonia Garudi
>Assignee: Hongyuan Li
> Attachments: HADOOP-14486-001.patch
>
>
> The TestSFTPFileSystem#testGetAccessTime test fails consistently with the 
> error below:
> {code}
> java.lang.AssertionError: expected:<1496496040072> but was:<149649604>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.fs.sftp.TestSFTPFileSystem.testGetAccessTime(TestSFTPFileSystem.java:319)
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14503) Make RollingAverages a mutable metric

2017-06-12 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14503?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047272#comment-16047272
 ] 

Arpit Agarwal commented on HADOOP-14503:


Thanks [~hanishakoneru]. +1 pending Jenkins.

> Make RollingAverages a mutable metric
> -
>
> Key: HADOOP-14503
> URL: https://issues.apache.org/jira/browse/HADOOP-14503
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Reporter: Hanisha Koneru
>Assignee: Hanisha Koneru
> Attachments: HADOOP-14503.001.patch, HADOOP-14503.002.patch, 
> HADOOP-14503.003.patch, HADOOP-14503.004.patch, HADOOP-14503.005.patch, 
> HADOOP-14503.006.patch, HADOOP-14503.007.patch
>
>
> RollingAverages metric extends on MutableRatesWithAggregation metric and 
> maintains a group of rolling average metrics. This class should be allowed to 
> register as a metric with the MetricSystem.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14503) Make RollingAverages a mutable metric

2017-06-12 Thread Hanisha Koneru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14503?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hanisha Koneru updated HADOOP-14503:

Attachment: HADOOP-14503.007.patch

Thanks [~arpitagarwal] for the review. 
I have addressed your comments in patch v07. 
The test failures are unrelated and pass locally.

> Make RollingAverages a mutable metric
> -
>
> Key: HADOOP-14503
> URL: https://issues.apache.org/jira/browse/HADOOP-14503
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Reporter: Hanisha Koneru
>Assignee: Hanisha Koneru
> Attachments: HADOOP-14503.001.patch, HADOOP-14503.002.patch, 
> HADOOP-14503.003.patch, HADOOP-14503.004.patch, HADOOP-14503.005.patch, 
> HADOOP-14503.006.patch, HADOOP-14503.007.patch
>
>
> RollingAverages metric extends on MutableRatesWithAggregation metric and 
> maintains a group of rolling average metrics. This class should be allowed to 
> register as a metric with the MetricSystem.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-14524) Make CryptoCodec Closeable so it can be cleaned up proactively

2017-06-12 Thread Xiao Chen (JIRA)
Xiao Chen created HADOOP-14524:
--

 Summary: Make CryptoCodec Closeable so it can be cleaned up 
proactively
 Key: HADOOP-14524
 URL: https://issues.apache.org/jira/browse/HADOOP-14524
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Xiao Chen
Assignee: Xiao Chen


See HADOOP-14523 for motivation. Credit to [~mi...@cloudera.com] for reporting 
initially on HADOOP-14523.

Basically, the CryptoCodec class is not a closeable, but the 
OpensslAesCtrCryptoCodec implementation of it contains a closeable member (the 
Random object). Currently it is left for {{finalize()}} to clean up, this would 
create problems if OpensslAesCtrCryptoCodec is used with OsSecureRandom, which 
could let OS run out of FDs on {{/dev/urandom}} if too many codecs created.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14524) Make CryptoCodec Closeable so it can be cleaned up proactively

2017-06-12 Thread Xiao Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14524?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiao Chen updated HADOOP-14524:
---
Description: 
See HADOOP-14523 for motivation. Credit to [~mi...@cloudera.com] for reporting 
initially there.

Basically, the CryptoCodec class is not a closeable, but the 
OpensslAesCtrCryptoCodec implementation of it contains a closeable member (the 
Random object). Currently it is left for {{finalize()}} to clean up, this would 
create problems if OpensslAesCtrCryptoCodec is used with OsSecureRandom, which 
could let OS run out of FDs on {{/dev/urandom}} if too many codecs created.

  was:
See HADOOP-14523 for motivation. Credit to [~mi...@cloudera.com] for reporting 
initially on HADOOP-14523.

Basically, the CryptoCodec class is not a closeable, but the 
OpensslAesCtrCryptoCodec implementation of it contains a closeable member (the 
Random object). Currently it is left for {{finalize()}} to clean up, this would 
create problems if OpensslAesCtrCryptoCodec is used with OsSecureRandom, which 
could let OS run out of FDs on {{/dev/urandom}} if too many codecs created.


> Make CryptoCodec Closeable so it can be cleaned up proactively
> --
>
> Key: HADOOP-14524
> URL: https://issues.apache.org/jira/browse/HADOOP-14524
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Xiao Chen
>Assignee: Xiao Chen
>
> See HADOOP-14523 for motivation. Credit to [~mi...@cloudera.com] for 
> reporting initially there.
> Basically, the CryptoCodec class is not a closeable, but the 
> OpensslAesCtrCryptoCodec implementation of it contains a closeable member 
> (the Random object). Currently it is left for {{finalize()}} to clean up, 
> this would create problems if OpensslAesCtrCryptoCodec is used with 
> OsSecureRandom, which could let OS run out of FDs on {{/dev/urandom}} if too 
> many codecs created.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Assigned] (HADOOP-14523) OpensslAesCtrCryptoCodec.finalize() holds excessive amounts of memory

2017-06-12 Thread Xiao Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14523?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiao Chen reassigned HADOOP-14523:
--

Assignee: Misha Dmitriev

> OpensslAesCtrCryptoCodec.finalize() holds excessive amounts of memory
> -
>
> Key: HADOOP-14523
> URL: https://issues.apache.org/jira/browse/HADOOP-14523
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Misha Dmitriev
>Assignee: Misha Dmitriev
>
> I recently analyzed JVM heap dumps from Hive running a big workload. Two 
> excerpts from the analysis done with jxray (www.jxray.com) are given below. 
> It turns out that nearly a half of live memory is taken by objects awaiting 
> finalization, and the biggest offender among them is class 
> OpensslAesCtrCryptoCodec:
> {code}
>   401,189K (39.7%) (1 of sun.misc.Cleaner)
>  <-- Java Static: sun.misc.Cleaner.first
>   400,572K (39.6%) (14001 of 
> org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec, 
> org.apache.hadoop.hive.ql.lockmgr.DummyTxnManager, java.util.jar.JarFile etc.)
>  <-- j.l.r.Finalizer.referent <-- j.l.r.Finalizer.{next} <-- 
> sun.misc.Cleaner.next <-- sun.misc.Cleaner.{next} <-- Java Static: 
> sun.misc.Cleaner.first
>   270,673K (26.8%) (2138 of org.apache.hadoop.mapred.JobConf)
>  <-- org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec.conf <-- 
> j.l.r.Finalizer.referent <-- j.l.r.Finalizer.{next} <-- sun.misc.Cleaner.next 
> <-- sun.misc.Cleaner.{next} <-- Java Static: sun.misc.Cleaner.first
> -
>   102,232K (10.1%) (1 of j.l.r.Finalizer)
>  <-- Java Static: java.lang.ref.Finalizer.unfinalized
>   101,676K (10.1%) (8613 of 
> org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec, 
> java.util.zip.ZipFile$ZipFileInflaterInputStream, 
> org.apache.hadoop.hive.ql.lockmgr.DummyTxnManager etc.)
>  <-- j.l.r.Finalizer.referent <-- j.l.r.Finalizer.{next} <-- Java Static: 
> java.lang.ref.Finalizer.unfinalized
> {code}
> This heap dump was taken using 'jmap -dump:live', which forces the JVM to run 
> full GC before dumping the heap. So we are already looking at the heap right 
> after GC, and yet all these unfinalized objects are there. I think this 
> happens because the JVM always runs only one finalization thread, and thus 
> the queue of objects that need finalization may get processed too slowly. My 
> understanding is that finalization works as follows:
> 1. When GC runs, it discovers that object x that overrides finalize() is 
> unreachable.
> 2. x is added to the finalization queue. So technically x is still reachable, 
> it occupies memory, and _all the objects that it references stay in memory as 
> well_.
> 3. The finalization thread processes objects from the finalization queue 
> serially, thus x may stay in memory for long time.
> 4. x.finalize() is invoked, then x is made unreachable. If x stayed in memory 
> for long time, it's now in Old Gen of the heap, so only full GC can clean it 
> up.
> 5. When full GC finally occurs, x gets cleaned up.
> So finalization is formally reliable, but in practice it's quite possible 
> that a lot of unreachable, but unfinalized objects flood the memory. I guess 
> we are seeing all these OpensslAesCtrCryptoCodec objects when they are in 
> phase 3 above. And the really bad thing is that these objects in turn keep in 
> memory a whole lot of other stuff, in particular JobConf objects. Such a 
> JobConf has nothing to do with finalization, yet the GC cannot release it 
> until the corresponding OpensslAesCtrCryptoCodec's is gone.
> Here is OpensslAesCtrCryptoCodec.finalize() method with my comments:
> {code}
> protected void finalize() throws Throwable {
>   try {
> Closeable r = (Closeable) this.random;
> r.close();  // Relevant only when (random instanceof OsSecureRandom == 
> true)
>   } catch (ClassCastException e) {
>   }
>   super.finalize();  // Not needed, no finalize() in superclasses
> }
> {code}
> So, finalize() in this class, that may keep in memory a whole tree of 
> objects, is relevant only when this codec is configured to use OsSecureRandom 
> class. The latter reads random bytes from the configured file, and needs 
> finalization to close the input stream associated with that file.
> The suggested fix is to remove finalize() from OpensslAesCtrCryptoCodec and 
> add it to the only class from this "family" that really needs it, 
> OsSecureRandom. That will ensure that only OsSecureRandom objects (if/when 
> they are used) stay in memory awaiting finalization, and no other, irrelevant 
> objects.
> Note that this solution means that streams are still closed lazily. This, in 
> principle, may cause its own problems. So the most reliable fix would be to 
> call OsSecureRandom.close() explicitly when it's not needed anymore. But the 
> above fix is a necessary 

[jira] [Created] (HADOOP-14523) OpensslAesCtrCryptoCodec.finalize() holds excessive amounts of memory

2017-06-12 Thread Misha Dmitriev (JIRA)
Misha Dmitriev created HADOOP-14523:
---

 Summary: OpensslAesCtrCryptoCodec.finalize() holds excessive 
amounts of memory
 Key: HADOOP-14523
 URL: https://issues.apache.org/jira/browse/HADOOP-14523
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Misha Dmitriev


I recently analyzed JVM heap dumps from Hive running a big workload. Two 
excerpts from the analysis done with jxray (www.jxray.com) are given below. It 
turns out that nearly a half of live memory is taken by objects awaiting 
finalization, and the biggest offender among them is class 
OpensslAesCtrCryptoCodec:

{code}
  401,189K (39.7%) (1 of sun.misc.Cleaner)
 <-- Java Static: sun.misc.Cleaner.first
  400,572K (39.6%) (14001 of org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec, 
org.apache.hadoop.hive.ql.lockmgr.DummyTxnManager, java.util.jar.JarFile etc.)
 <-- j.l.r.Finalizer.referent <-- j.l.r.Finalizer.{next} <-- 
sun.misc.Cleaner.next <-- sun.misc.Cleaner.{next} <-- Java Static: 
sun.misc.Cleaner.first
  270,673K (26.8%) (2138 of org.apache.hadoop.mapred.JobConf)
 <-- org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec.conf <-- 
j.l.r.Finalizer.referent <-- j.l.r.Finalizer.{next} <-- sun.misc.Cleaner.next 
<-- sun.misc.Cleaner.{next} <-- Java Static: sun.misc.Cleaner.first

-

  102,232K (10.1%) (1 of j.l.r.Finalizer)
 <-- Java Static: java.lang.ref.Finalizer.unfinalized
  101,676K (10.1%) (8613 of org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec, 
java.util.zip.ZipFile$ZipFileInflaterInputStream, 
org.apache.hadoop.hive.ql.lockmgr.DummyTxnManager etc.)
 <-- j.l.r.Finalizer.referent <-- j.l.r.Finalizer.{next} <-- Java Static: 
java.lang.ref.Finalizer.unfinalized
{code}

This heap dump was taken using 'jmap -dump:live', which forces the JVM to run 
full GC before dumping the heap. So we are already looking at the heap right 
after GC, and yet all these unfinalized objects are there. I think this happens 
because the JVM always runs only one finalization thread, and thus the queue of 
objects that need finalization may get processed too slowly. My understanding 
is that finalization works as follows:

1. When GC runs, it discovers that object x that overrides finalize() is 
unreachable.
2. x is added to the finalization queue. So technically x is still reachable, 
it occupies memory, and _all the objects that it references stay in memory as 
well_.
3. The finalization thread processes objects from the finalization queue 
serially, thus x may stay in memory for long time.
4. x.finalize() is invoked, then x is made unreachable. If x stayed in memory 
for long time, it's now in Old Gen of the heap, so only full GC can clean it up.
5. When full GC finally occurs, x gets cleaned up.

So finalization is formally reliable, but in practice it's quite possible that 
a lot of unreachable, but unfinalized objects flood the memory. I guess we are 
seeing all these OpensslAesCtrCryptoCodec objects when they are in phase 3 
above. And the really bad thing is that these objects in turn keep in memory a 
whole lot of other stuff, in particular JobConf objects. Such a JobConf has 
nothing to do with finalization, yet the GC cannot release it until the 
corresponding OpensslAesCtrCryptoCodec's is gone.

Here is OpensslAesCtrCryptoCodec.finalize() method with my comments:

{code}
protected void finalize() throws Throwable {
  try {
Closeable r = (Closeable) this.random;
r.close();  // Relevant only when (random instanceof OsSecureRandom == true)
  } catch (ClassCastException e) {
  }
  super.finalize();  // Not needed, no finalize() in superclasses
}
{code}

So, finalize() in this class, that may keep in memory a whole tree of objects, 
is relevant only when this codec is configured to use OsSecureRandom class. The 
latter reads random bytes from the configured file, and needs finalization to 
close the input stream associated with that file.

The suggested fix is to remove finalize() from OpensslAesCtrCryptoCodec and add 
it to the only class from this "family" that really needs it, OsSecureRandom. 
That will ensure that only OsSecureRandom objects (if/when they are used) stay 
in memory awaiting finalization, and no other, irrelevant objects.

Note that this solution means that streams are still closed lazily. This, in 
principle, may cause its own problems. So the most reliable fix would be to 
call OsSecureRandom.close() explicitly when it's not needed anymore. But the 
above fix is a necessary first step anyway, it will remove the most acute 
problem with memory and will not make any other things worse than they 
currently are.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: 

[jira] [Commented] (HADOOP-14521) KMS client needs retry logic

2017-06-12 Thread Xiao Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14521?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047207#comment-16047207
 ] 

Xiao Chen commented on HADOOP-14521:


Thanks for revving Rushabh. Looks pretty close to me.

Thanks for the explanations about {{AuthenticatedException}}, makes sense. I 
don't know about why the IOExceptions are wrapped either (except for guessing 
they're tricks with the method signature). Agree it's not related to this jira.

Patch 6 comments below:
For the retries, do you think maybe it's more intuitive to specify num of 
failovers for the whole provider set? (i.e. {{maxFailovers = 
conf.getInt(KMS_CLIENT_FAILOVER_MAX_RETRIES_KEY) * providers.length}}. (Then we 
can have the core-default.xml set to 1) Will leave the decision to you and 
Daryn.

Nits:
- KMSCP: the first {{@Link}} seems redundant in {{This will always create a 
@Link  ...}}
- Good to have some validations on the config params (e.g. numRetries > 0 etc.)

> KMS client needs retry logic
> 
>
> Key: HADOOP-14521
> URL: https://issues.apache.org/jira/browse/HADOOP-14521
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.6.0
>Reporter: Rushabh S Shah
>Assignee: Rushabh S Shah
> Attachments: HDFS-11804-trunk-1.patch, HDFS-11804-trunk-2.patch, 
> HDFS-11804-trunk-3.patch, HDFS-11804-trunk-4.patch, HDFS-11804-trunk-5.patch, 
> HDFS-11804-trunk-6.patch, HDFS-11804-trunk.patch
>
>
> The kms client appears to have no retry logic – at all.  It's completely 
> decoupled from the ipc retry logic.  This has major impacts if the KMS is 
> unreachable for any reason, including but not limited to network connection 
> issues, timeouts, the +restart during an upgrade+.
> This has some major ramifications:
> # Jobs may fail to submit, although oozie resubmit logic should mask it
> # Non-oozie launchers may experience higher rates if they do not already have 
> retry logic.
> # Tasks reading EZ files will fail, probably be masked by framework reattempts
> # EZ file creation fails after creating a 0-length file – client receives 
> EDEK in the create response, then fails when decrypting the EDEK
> # Bulk hadoop fs copies, and maybe distcp, will prematurely fail



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14521) KMS client needs retry logic

2017-06-12 Thread Daryn Sharp (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14521?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047200#comment-16047200
 ] 

Daryn Sharp commented on HADOOP-14521:
--

I'd revert the change that breaks instead of returning.  Otherwise this final 
condition: {{(ret == null && ex != null)}} can be faulty.  If the method was 
successfully invoked and it returned null, and it was a retry, the condition 
will erroneously throw the pre-retry exception.

> KMS client needs retry logic
> 
>
> Key: HADOOP-14521
> URL: https://issues.apache.org/jira/browse/HADOOP-14521
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.6.0
>Reporter: Rushabh S Shah
>Assignee: Rushabh S Shah
> Attachments: HDFS-11804-trunk-1.patch, HDFS-11804-trunk-2.patch, 
> HDFS-11804-trunk-3.patch, HDFS-11804-trunk-4.patch, HDFS-11804-trunk-5.patch, 
> HDFS-11804-trunk-6.patch, HDFS-11804-trunk.patch
>
>
> The kms client appears to have no retry logic – at all.  It's completely 
> decoupled from the ipc retry logic.  This has major impacts if the KMS is 
> unreachable for any reason, including but not limited to network connection 
> issues, timeouts, the +restart during an upgrade+.
> This has some major ramifications:
> # Jobs may fail to submit, although oozie resubmit logic should mask it
> # Non-oozie launchers may experience higher rates if they do not already have 
> retry logic.
> # Tasks reading EZ files will fail, probably be masked by framework reattempts
> # EZ file creation fails after creating a 0-length file – client receives 
> EDEK in the create response, then fails when decrypting the EDEK
> # Bulk hadoop fs copies, and maybe distcp, will prematurely fail



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14457) create() does not notify metadataStore of parent directories or ensure they're not existing files

2017-06-12 Thread Aaron Fabbri (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047186#comment-16047186
 ] 

Aaron Fabbri commented on HADOOP-14457:
---

Thanks for interesting discussion [~mackrorysd].

{quote}
A capabilities-based approach seems to me to be far more complicated
{quote}
I think capabilities APIs are good practice for any interfaces that have 
optional features.  They make things explicit and improve testability. IMO 
FileSystem should have something like this as well (i.e. HADOOP-9565, 
HDFS-11644 come to mind).

{quote}
 than just deciding whether it is or isn't a MetadataStore implementations job 
to track parent directories.
{quote}

By "parent directories", I assume you mean ancestor directories all the way to 
the root.  To be clear we have decided what MetadataStore's contract is.  
create() is the problem here.

We could change this, adding requirement that for any {{put(path)}}, the 
{{MetadataStore}} *must* account for all ancestor directories in the whole 
path.   Currently the contract is not a *must* but a *may*.  Reasons for this:

1. runtime complexity.  Consider an FS client doing this:

MetadataStore.put(/a/b/c/d)
.. put(/a/b/c)
.. put(/a/b)
.. put(/a)

if each put() is does work of writing ancestors, it is wasted (~2x round trips 
for any DB client, or depth^2 if it is dumb). I want to add more MetadataStore 
backends in the future, and I'm not sure they will have the same internal 
requirement that our Dynamo schema does (i.e. that all ancestors *must* be 
tracked anyways).

2. The FS client should be doing the ancestor enumeration anyways (see 
mkdirs()).  Why force all MetadataStore to repeat it?

MetadataStore is a trailing log of metadata modifications that have been made 
to a FileSystem.   We expect a FS to be able to tell us what is is doing, or it 
is broken (thus the bug here). I don't think putting a band-aid on create() is 
a compelling reason to add more requirements to any future MetadataStore 
implementations. Create is still broken here (i.e. if file is an ancestor) 
either way.

TL;DR we could add the requirement if it makes sense, but I don't think it buys 
us much here. I also feel like the implicit ancestor creation semantics of 
Hadoop FS was not the best decision and am not eager to have it spread to 
MetadataStore interface.



> create() does not notify metadataStore of parent directories or ensure 
> they're not existing files
> -
>
> Key: HADOOP-14457
> URL: https://issues.apache.org/jira/browse/HADOOP-14457
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Reporter: Sean Mackrory
>Assignee: Sean Mackrory
> Attachments: HADOOP-14457-HADOOP-13345.001.patch, 
> HADOOP-14457-HADOOP-13345.002.patch, HADOOP-14457-HADOOP-13345.003.patch, 
> HADOOP-14457-HADOOP-13345.004.patch, HADOOP-14457-HADOOP-13345.005.patch, 
> HADOOP-14457-HADOOP-13345.006.patch, HADOOP-14457-HADOOP-13345.007.patch, 
> HADOOP-14457-HADOOP-13345.008.patch, HADOOP-14457-HADOOP-13345.009.patch
>
>
> Not a great test yet, but it at least reliably demonstrates the issue. 
> LocalMetadataStore will sometimes erroneously report that a directory is 
> empty with isAuthoritative = true when it *definitely* has children the 
> metadatastore should know about. It doesn't appear to happen if the children 
> are just directory. The fact that it's returning an empty listing is 
> concerning, but the fact that it says it's authoritative *might* be a second 
> bug.
> {code}
> diff --git 
> a/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFileSystem.java
>  
> b/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFileSystem.java
> index 78b3970..1821d19 100644
> --- 
> a/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFileSystem.java
> +++ 
> b/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFileSystem.java
> @@ -965,7 +965,7 @@ public boolean hasMetadataStore() {
>}
>  
>@VisibleForTesting
> -  MetadataStore getMetadataStore() {
> +  public MetadataStore getMetadataStore() {
>  return metadataStore;
>}
>  
> diff --git 
> a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractRename.java
>  
> b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractRename.java
> index 4339649..881bdc9 100644
> --- 
> a/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractRename.java
> +++ 
> b/hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractRename.java
> @@ -23,6 +23,11 @@
>  import org.apache.hadoop.fs.contract.AbstractFSContract;
>  import org.apache.hadoop.fs.FileSystem;
>  import 

[jira] [Commented] (HADOOP-14503) Make RollingAverages a mutable metric

2017-06-12 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14503?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047161#comment-16047161
 ] 

Arpit Agarwal commented on HADOOP-14503:


Thanks [~hanishakoneru]. The patch lgtm. Can you please address the checkstyle 
issues? For the first checkstyle issue, you can add a private constructor to 
the MetricsTestHelper class and make it a final class.

Also there is a pre-existing typo in DataNodePeerMetrics 
(sendPacketDownstreamRollingAvgerages). Could you please fix that too?

> Make RollingAverages a mutable metric
> -
>
> Key: HADOOP-14503
> URL: https://issues.apache.org/jira/browse/HADOOP-14503
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Reporter: Hanisha Koneru
>Assignee: Hanisha Koneru
> Attachments: HADOOP-14503.001.patch, HADOOP-14503.002.patch, 
> HADOOP-14503.003.patch, HADOOP-14503.004.patch, HADOOP-14503.005.patch, 
> HADOOP-14503.006.patch
>
>
> RollingAverages metric extends on MutableRatesWithAggregation metric and 
> maintains a group of rolling average metrics. This class should be allowed to 
> register as a metric with the MetricSystem.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14501) Switch from aalto-xml to woodstox to handle odd XML features

2017-06-12 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14501?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047151#comment-16047151
 ] 

Hudson commented on HADOOP-14501:
-

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #11858 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/11858/])
HADOOP-14501. Switch from aalto-xml to woodstox to handle odd XML (jeagles: rev 
a81916ea89d59c1642b3462e3d7c6c1acb1e7109)
* (edit) 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/TestConfiguration.java
* (edit) 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java
* (edit) hadoop-project/pom.xml
* (edit) hadoop-common-project/hadoop-common/pom.xml


> Switch from aalto-xml to woodstox to handle odd XML features
> 
>
> Key: HADOOP-14501
> URL: https://issues.apache.org/jira/browse/HADOOP-14501
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: conf
>Affects Versions: 2.9.0, 3.0.0-alpha4
>Reporter: Andrew Wang
>Assignee: Jonathan Eagles
>Priority: Blocker
> Fix For: 2.9.0, 3.0.0-alpha4
>
> Attachments: HADOOP-14501.1.patch, HADOOP-14501.2.patch, 
> HADOOP-14501.3-branch-2.patch, HADOOP-14501.3.patch, 
> HADOOP-14501.4-branch-2.patch, HADOOP-14501.4.patch, HADOOP-14501.5.patch, 
> HADOOP-14501-branch-2.1.patch
>
>
> [~hgadre] tried testing solr with a Hadoop 3 client. He saw various test case 
> failures due to what look like functionality gaps in the new aalto-xml stax 
> implementation pulled in by HADOOP-14216:
> {noformat}
>[junit4]> Throwable #1: com.fasterxml.aalto.WFCException: Illegal XML 
> character ('ü' (code 252))
> 
>[junit4]> Caused by: com.fasterxml.aalto.WFCException: General entity 
> reference () encountered in entity expanding mode: operation not (yet) 
> implemented
> ...
>[junit4]> Throwable #1: org.apache.solr.common.SolrException: General 
> entity reference () encountered in entity expanding mode: operation 
> not (yet) implemented
> {noformat}
> These were from the following test case executions:
> {noformat}
> NOTE: reproduce with: ant test  -Dtestcase=DocumentAnalysisRequestHandlerTest 
> -Dtests.method=testCharsetOutsideDocument -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=und -Dtests.timezone=Atlantic/Faeroe 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=MBeansHandlerTest 
> -Dtests.method=testXMLDiffWithExternalEntity -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=en-US -Dtests.timezone=US/Aleutian 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=XmlUpdateRequestHandlerTest 
> -Dtests.method=testExternalEntities -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=hr -Dtests.timezone=America/Barbados 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=XmlUpdateRequestHandlerTest 
> -Dtests.method=testNamedEntity -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=hr -Dtests.timezone=America/Barbados 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14501) Switch from aalto-xml to woodstox to handle odd XML features

2017-06-12 Thread Jonathan Eagles (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14501?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047141#comment-16047141
 ] 

Jonathan Eagles commented on HADOOP-14501:
--

Thanks again, [~andrew.wang]!

> Switch from aalto-xml to woodstox to handle odd XML features
> 
>
> Key: HADOOP-14501
> URL: https://issues.apache.org/jira/browse/HADOOP-14501
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: conf
>Affects Versions: 2.9.0, 3.0.0-alpha4
>Reporter: Andrew Wang
>Assignee: Jonathan Eagles
>Priority: Blocker
> Fix For: 2.9.0, 3.0.0-alpha4
>
> Attachments: HADOOP-14501.1.patch, HADOOP-14501.2.patch, 
> HADOOP-14501.3-branch-2.patch, HADOOP-14501.3.patch, 
> HADOOP-14501.4-branch-2.patch, HADOOP-14501.4.patch, HADOOP-14501.5.patch, 
> HADOOP-14501-branch-2.1.patch
>
>
> [~hgadre] tried testing solr with a Hadoop 3 client. He saw various test case 
> failures due to what look like functionality gaps in the new aalto-xml stax 
> implementation pulled in by HADOOP-14216:
> {noformat}
>[junit4]> Throwable #1: com.fasterxml.aalto.WFCException: Illegal XML 
> character ('ü' (code 252))
> 
>[junit4]> Caused by: com.fasterxml.aalto.WFCException: General entity 
> reference () encountered in entity expanding mode: operation not (yet) 
> implemented
> ...
>[junit4]> Throwable #1: org.apache.solr.common.SolrException: General 
> entity reference () encountered in entity expanding mode: operation 
> not (yet) implemented
> {noformat}
> These were from the following test case executions:
> {noformat}
> NOTE: reproduce with: ant test  -Dtestcase=DocumentAnalysisRequestHandlerTest 
> -Dtests.method=testCharsetOutsideDocument -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=und -Dtests.timezone=Atlantic/Faeroe 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=MBeansHandlerTest 
> -Dtests.method=testXMLDiffWithExternalEntity -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=en-US -Dtests.timezone=US/Aleutian 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=XmlUpdateRequestHandlerTest 
> -Dtests.method=testExternalEntities -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=hr -Dtests.timezone=America/Barbados 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=XmlUpdateRequestHandlerTest 
> -Dtests.method=testNamedEntity -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=hr -Dtests.timezone=America/Barbados 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14501) Switch from aalto-xml to woodstox to handle odd XML features

2017-06-12 Thread Jonathan Eagles (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14501?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jonathan Eagles updated HADOOP-14501:
-
   Resolution: Fixed
Fix Version/s: 3.0.0-alpha4
   2.9.0
   Status: Resolved  (was: Patch Available)

> Switch from aalto-xml to woodstox to handle odd XML features
> 
>
> Key: HADOOP-14501
> URL: https://issues.apache.org/jira/browse/HADOOP-14501
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: conf
>Affects Versions: 2.9.0, 3.0.0-alpha4
>Reporter: Andrew Wang
>Assignee: Jonathan Eagles
>Priority: Blocker
> Fix For: 2.9.0, 3.0.0-alpha4
>
> Attachments: HADOOP-14501.1.patch, HADOOP-14501.2.patch, 
> HADOOP-14501.3-branch-2.patch, HADOOP-14501.3.patch, 
> HADOOP-14501.4-branch-2.patch, HADOOP-14501.4.patch, HADOOP-14501.5.patch, 
> HADOOP-14501-branch-2.1.patch
>
>
> [~hgadre] tried testing solr with a Hadoop 3 client. He saw various test case 
> failures due to what look like functionality gaps in the new aalto-xml stax 
> implementation pulled in by HADOOP-14216:
> {noformat}
>[junit4]> Throwable #1: com.fasterxml.aalto.WFCException: Illegal XML 
> character ('ü' (code 252))
> 
>[junit4]> Caused by: com.fasterxml.aalto.WFCException: General entity 
> reference () encountered in entity expanding mode: operation not (yet) 
> implemented
> ...
>[junit4]> Throwable #1: org.apache.solr.common.SolrException: General 
> entity reference () encountered in entity expanding mode: operation 
> not (yet) implemented
> {noformat}
> These were from the following test case executions:
> {noformat}
> NOTE: reproduce with: ant test  -Dtestcase=DocumentAnalysisRequestHandlerTest 
> -Dtests.method=testCharsetOutsideDocument -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=und -Dtests.timezone=Atlantic/Faeroe 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=MBeansHandlerTest 
> -Dtests.method=testXMLDiffWithExternalEntity -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=en-US -Dtests.timezone=US/Aleutian 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=XmlUpdateRequestHandlerTest 
> -Dtests.method=testExternalEntities -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=hr -Dtests.timezone=America/Barbados 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=XmlUpdateRequestHandlerTest 
> -Dtests.method=testNamedEntity -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=hr -Dtests.timezone=America/Barbados 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14522) Unable to see logs for Job end notification

2017-06-12 Thread Satish Subhashrao Saley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satish Subhashrao Saley updated HADOOP-14522:
-
Labels: applicationmaster mapreduce  (was: )

> Unable to see logs for Job end notification
> ---
>
> Key: HADOOP-14522
> URL: https://issues.apache.org/jira/browse/HADOOP-14522
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.8.0
>Reporter: Satish Subhashrao Saley
>  Labels: applicationmaster, mapreduce
>
> Hi, I have set {{mapreduce.job.end-notification.url}} and I am able to get 
> the notification correctly.  But I am unable to see log lines from class 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster regarding job end 
> notification. When I was trying out different notification urls (before 
> stabilizing my app), I didn't have any way to look at the logs and wasted a 
> lot of time figuring out actual issue. It would be good if these logs are 
> also included in syslog for container. 
> [relevant 
> code|https://github.com/apache/hadoop/blob/branch-2.8/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/MRAppMaster.java#L641-L650]
> {code}
> notifyIsLastAMRetry(isLastAMRetry);
>   // Stop all services
>   // This will also send the final report to the ResourceManager
>   LOG.info("Calling stop for all the services"); < Able to see this.
>   MRAppMaster.this.stop();
>   if (isLastAMRetry) {
> // Send job-end notification when it is safe to report termination to
> // users and it is the last AM retry
> if (getConfig().get(MRJobConfig.MR_JOB_END_NOTIFICATION_URL) != null) 
> {
>   try {
> LOG.info("Job end notification started for jobID : "
> + job.getReport().getJobId()); <--- unable to see this
> JobEndNotifier notifier = new JobEndNotifier();
> notifier.setConf(getConfig());
> JobReport report = job.getReport();
> {code}
> Is it because we are shutting down all services including 
> JobHistoryEventHandler in {{MRAppMaster.this.stop();}} ? 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14522) Unable to see logs for Job end notification

2017-06-12 Thread Satish Subhashrao Saley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satish Subhashrao Saley updated HADOOP-14522:
-
Affects Version/s: 2.8.0

> Unable to see logs for Job end notification
> ---
>
> Key: HADOOP-14522
> URL: https://issues.apache.org/jira/browse/HADOOP-14522
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.8.0
>Reporter: Satish Subhashrao Saley
>
> Hi, I have set {{mapreduce.job.end-notification.url}} and I am able to get 
> the notification correctly.  But I am unable to see log lines from class 
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster regarding job end 
> notification. When I was trying out different notification urls (before 
> stabilizing my app), I didn't have any way to look at the logs and wasted a 
> lot of time figuring out actual issue. It would be good if these logs are 
> also included in syslog for container. 
> [relevant 
> code|https://github.com/apache/hadoop/blob/branch-2.8/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/MRAppMaster.java#L641-L650]
> {code}
> notifyIsLastAMRetry(isLastAMRetry);
>   // Stop all services
>   // This will also send the final report to the ResourceManager
>   LOG.info("Calling stop for all the services"); < Able to see this.
>   MRAppMaster.this.stop();
>   if (isLastAMRetry) {
> // Send job-end notification when it is safe to report termination to
> // users and it is the last AM retry
> if (getConfig().get(MRJobConfig.MR_JOB_END_NOTIFICATION_URL) != null) 
> {
>   try {
> LOG.info("Job end notification started for jobID : "
> + job.getReport().getJobId()); <--- unable to see this
> JobEndNotifier notifier = new JobEndNotifier();
> notifier.setConf(getConfig());
> JobReport report = job.getReport();
> {code}
> Is it because we are shutting down all services including 
> JobHistoryEventHandler in {{MRAppMaster.this.stop();}} ? 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-14522) Unable to see logs for Job end notification

2017-06-12 Thread Satish Subhashrao Saley (JIRA)
Satish Subhashrao Saley created HADOOP-14522:


 Summary: Unable to see logs for Job end notification
 Key: HADOOP-14522
 URL: https://issues.apache.org/jira/browse/HADOOP-14522
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Satish Subhashrao Saley


Hi, I have set {{mapreduce.job.end-notification.url}} and I am able to get the 
notification correctly.  But I am unable to see log lines from class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster regarding job end notification. 
When I was trying out different notification urls (before stabilizing my app), 
I didn't have any way to look at the logs and wasted a lot of time figuring out 
actual issue. It would be good if these logs are also included in syslog for 
container. 

[relevant 
code|https://github.com/apache/hadoop/blob/branch-2.8/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/MRAppMaster.java#L641-L650]
{code}
notifyIsLastAMRetry(isLastAMRetry);
  // Stop all services
  // This will also send the final report to the ResourceManager
  LOG.info("Calling stop for all the services"); < Able to see this.
  MRAppMaster.this.stop();

  if (isLastAMRetry) {
// Send job-end notification when it is safe to report termination to
// users and it is the last AM retry
if (getConfig().get(MRJobConfig.MR_JOB_END_NOTIFICATION_URL) != null) {
  try {
LOG.info("Job end notification started for jobID : "
+ job.getReport().getJobId()); <--- unable to see this
JobEndNotifier notifier = new JobEndNotifier();
notifier.setConf(getConfig());
JobReport report = job.getReport();
{code}

Is it because we are shutting down all services including 
JobHistoryEventHandler in {{MRAppMaster.this.stop();}} ? 




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14520) Block compaction for WASB

2017-06-12 Thread Georgi Chalakov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Georgi Chalakov updated HADOOP-14520:
-
Status: Patch Available  (was: In Progress)

> Block compaction for WASB
> -
>
> Key: HADOOP-14520
> URL: https://issues.apache.org/jira/browse/HADOOP-14520
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/azure
>Affects Versions: 3.0.0-alpha3
>Reporter: Georgi Chalakov
>Assignee: Georgi Chalakov
> Attachments: HADOOP-14520-01.patch, HADOOP-14520-01-test.txt
>
>
> Block Compaction for WASB allows uploading new blocks for every hflush/hsync 
> call. When the number of blocks is above a predefined, configurable value, 
> next hflush/hsync triggers the block compaction process. Block compaction 
> replaces a sequence of blocks with one block. From all the sequences with 
> total length less than 4M, compaction chooses the longest one. It is a greedy 
> algorithm that preserve all potential candidates for the next round. Block 
> Compaction for WASB increases data durability and allows using block blobs 
> instead of page blobs. By default, block compaction is disabled. Similar to 
> the configuration for page blobs, the client needs to specify HDFS folders 
> where block compaction over block blobs is enabled. 
> Results for HADOOP-14520-01.patch
> Tests run: 704, Failures: 0, Errors: 0, Skipped: 119



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14520) Block compaction for WASB

2017-06-12 Thread Georgi Chalakov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Georgi Chalakov updated HADOOP-14520:
-
Status: In Progress  (was: Patch Available)

> Block compaction for WASB
> -
>
> Key: HADOOP-14520
> URL: https://issues.apache.org/jira/browse/HADOOP-14520
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/azure
>Affects Versions: 3.0.0-alpha3
>Reporter: Georgi Chalakov
>Assignee: Georgi Chalakov
> Attachments: HADOOP-14520-01.patch, HADOOP-14520-01-test.txt
>
>
> Block Compaction for WASB allows uploading new blocks for every hflush/hsync 
> call. When the number of blocks is above a predefined, configurable value, 
> next hflush/hsync triggers the block compaction process. Block compaction 
> replaces a sequence of blocks with one block. From all the sequences with 
> total length less than 4M, compaction chooses the longest one. It is a greedy 
> algorithm that preserve all potential candidates for the next round. Block 
> Compaction for WASB increases data durability and allows using block blobs 
> instead of page blobs. By default, block compaction is disabled. Similar to 
> the configuration for page blobs, the client needs to specify HDFS folders 
> where block compaction over block blobs is enabled. 
> Results for HADOOP-14520-01.patch
> Tests run: 704, Failures: 0, Errors: 0, Skipped: 119



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14520) Block compaction for WASB

2017-06-12 Thread Georgi Chalakov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Georgi Chalakov updated HADOOP-14520:
-
Status: Patch Available  (was: In Progress)

> Block compaction for WASB
> -
>
> Key: HADOOP-14520
> URL: https://issues.apache.org/jira/browse/HADOOP-14520
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/azure
>Affects Versions: 3.0.0-alpha3
>Reporter: Georgi Chalakov
>Assignee: Georgi Chalakov
> Attachments: HADOOP-14520-01.patch, HADOOP-14520-01-test.txt
>
>
> Block Compaction for WASB allows uploading new blocks for every hflush/hsync 
> call. When the number of blocks is above a predefined, configurable value, 
> next hflush/hsync triggers the block compaction process. Block compaction 
> replaces a sequence of blocks with one block. From all the sequences with 
> total length less than 4M, compaction chooses the longest one. It is a greedy 
> algorithm that preserve all potential candidates for the next round. Block 
> Compaction for WASB increases data durability and allows using block blobs 
> instead of page blobs. By default, block compaction is disabled. Similar to 
> the configuration for page blobs, the client needs to specify HDFS folders 
> where block compaction over block blobs is enabled. 
> Results for HADOOP-14520-01.patch
> Tests run: 704, Failures: 0, Errors: 0, Skipped: 119



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14501) Switch from aalto-xml to woodstox to handle odd XML features

2017-06-12 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14501?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047124#comment-16047124
 ] 

Andrew Wang commented on HADOOP-14501:
--

Yea SGTM, thanks Jonathan!

> Switch from aalto-xml to woodstox to handle odd XML features
> 
>
> Key: HADOOP-14501
> URL: https://issues.apache.org/jira/browse/HADOOP-14501
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: conf
>Affects Versions: 2.9.0, 3.0.0-alpha4
>Reporter: Andrew Wang
>Assignee: Jonathan Eagles
>Priority: Blocker
> Attachments: HADOOP-14501.1.patch, HADOOP-14501.2.patch, 
> HADOOP-14501.3-branch-2.patch, HADOOP-14501.3.patch, 
> HADOOP-14501.4-branch-2.patch, HADOOP-14501.4.patch, HADOOP-14501.5.patch, 
> HADOOP-14501-branch-2.1.patch
>
>
> [~hgadre] tried testing solr with a Hadoop 3 client. He saw various test case 
> failures due to what look like functionality gaps in the new aalto-xml stax 
> implementation pulled in by HADOOP-14216:
> {noformat}
>[junit4]> Throwable #1: com.fasterxml.aalto.WFCException: Illegal XML 
> character ('ü' (code 252))
> 
>[junit4]> Caused by: com.fasterxml.aalto.WFCException: General entity 
> reference () encountered in entity expanding mode: operation not (yet) 
> implemented
> ...
>[junit4]> Throwable #1: org.apache.solr.common.SolrException: General 
> entity reference () encountered in entity expanding mode: operation 
> not (yet) implemented
> {noformat}
> These were from the following test case executions:
> {noformat}
> NOTE: reproduce with: ant test  -Dtestcase=DocumentAnalysisRequestHandlerTest 
> -Dtests.method=testCharsetOutsideDocument -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=und -Dtests.timezone=Atlantic/Faeroe 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=MBeansHandlerTest 
> -Dtests.method=testXMLDiffWithExternalEntity -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=en-US -Dtests.timezone=US/Aleutian 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=XmlUpdateRequestHandlerTest 
> -Dtests.method=testExternalEntities -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=hr -Dtests.timezone=America/Barbados 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=XmlUpdateRequestHandlerTest 
> -Dtests.method=testNamedEntity -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=hr -Dtests.timezone=America/Barbados 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14501) Switch from aalto-xml to woodstox to handle odd XML features

2017-06-12 Thread Jonathan Eagles (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14501?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047121#comment-16047121
 ] 

Jonathan Eagles commented on HADOOP-14501:
--

[~andrew.wang], should I go ahead and check this into trunk and branch-2?

> Switch from aalto-xml to woodstox to handle odd XML features
> 
>
> Key: HADOOP-14501
> URL: https://issues.apache.org/jira/browse/HADOOP-14501
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: conf
>Affects Versions: 2.9.0, 3.0.0-alpha4
>Reporter: Andrew Wang
>Assignee: Jonathan Eagles
>Priority: Blocker
> Attachments: HADOOP-14501.1.patch, HADOOP-14501.2.patch, 
> HADOOP-14501.3-branch-2.patch, HADOOP-14501.3.patch, 
> HADOOP-14501.4-branch-2.patch, HADOOP-14501.4.patch, HADOOP-14501.5.patch, 
> HADOOP-14501-branch-2.1.patch
>
>
> [~hgadre] tried testing solr with a Hadoop 3 client. He saw various test case 
> failures due to what look like functionality gaps in the new aalto-xml stax 
> implementation pulled in by HADOOP-14216:
> {noformat}
>[junit4]> Throwable #1: com.fasterxml.aalto.WFCException: Illegal XML 
> character ('ü' (code 252))
> 
>[junit4]> Caused by: com.fasterxml.aalto.WFCException: General entity 
> reference () encountered in entity expanding mode: operation not (yet) 
> implemented
> ...
>[junit4]> Throwable #1: org.apache.solr.common.SolrException: General 
> entity reference () encountered in entity expanding mode: operation 
> not (yet) implemented
> {noformat}
> These were from the following test case executions:
> {noformat}
> NOTE: reproduce with: ant test  -Dtestcase=DocumentAnalysisRequestHandlerTest 
> -Dtests.method=testCharsetOutsideDocument -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=und -Dtests.timezone=Atlantic/Faeroe 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=MBeansHandlerTest 
> -Dtests.method=testXMLDiffWithExternalEntity -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=en-US -Dtests.timezone=US/Aleutian 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=XmlUpdateRequestHandlerTest 
> -Dtests.method=testExternalEntities -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=hr -Dtests.timezone=America/Barbados 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> NOTE: reproduce with: ant test  -Dtestcase=XmlUpdateRequestHandlerTest 
> -Dtests.method=testNamedEntity -Dtests.seed=2F739D88D9C723CA 
> -Dtests.slow=true -Dtests.locale=hr -Dtests.timezone=America/Barbados 
> -Dtests.asserts=true -Dtests.file.encoding=US-ASCII
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14503) Make RollingAverages a mutable metric

2017-06-12 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14503?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047103#comment-16047103
 ] 

Hadoop QA commented on HADOOP-14503:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
17s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 6 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
40s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 13m 
59s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 14m  
8s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
 3s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m 
11s{color} | {color:green} trunk passed {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  1m 
31s{color} | {color:red} hadoop-common-project/hadoop-common in trunk has 19 
extant Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
41s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
17s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
30s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 10m 
11s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 10m 
11s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
2m  4s{color} | {color:orange} root: The patch generated 4 new + 595 unchanged 
- 3 fixed = 599 total (was 598) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m  
8s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 4 line(s) that end in whitespace. Use git 
apply --whitespace=fix <>. Refer https://git-scm.com/docs/git-apply 
{color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
37s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
47s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  7m 
32s{color} | {color:green} hadoop-common in the patch passed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 70m 59s{color} 
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
44s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}161m 57s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure |
|   | hadoop.hdfs.server.datanode.TestDirectoryScanner |
|   | hadoop.hdfs.TestDFSStripedInputStreamWithRandomECPolicy |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:14b5c93 |
| JIRA Issue | HADOOP-14503 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12872763/HADOOP-14503.006.patch
 |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux ced59ec04b28 3.13.0-106-generic #153-Ubuntu SMP Tue Dec 6 
15:44:32 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 86368cc |
| Default Java | 1.8.0_131 |
| findbugs | v3.1.0-RC1 |
| findbugs | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12522/artifact/patchprocess/branch-findbugs-hadoop-common-project_hadoop-common-warnings.html
 |
| checkstyle | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12522/artifact/patchprocess/diff-checkstyle-root.txt
 |
| whitespace | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12522/artifact/patchprocess/whitespace-eol.txt
 |
| unit 

[jira] [Updated] (HADOOP-14520) Block compaction for WASB

2017-06-12 Thread Georgi Chalakov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Georgi Chalakov updated HADOOP-14520:
-
Status: In Progress  (was: Patch Available)

> Block compaction for WASB
> -
>
> Key: HADOOP-14520
> URL: https://issues.apache.org/jira/browse/HADOOP-14520
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/azure
>Affects Versions: 3.0.0-alpha3
>Reporter: Georgi Chalakov
>Assignee: Georgi Chalakov
> Attachments: HADOOP-14520-01.patch, HADOOP-14520-01-test.txt
>
>
> Block Compaction for WASB allows uploading new blocks for every hflush/hsync 
> call. When the number of blocks is above a predefined, configurable value, 
> next hflush/hsync triggers the block compaction process. Block compaction 
> replaces a sequence of blocks with one block. From all the sequences with 
> total length less than 4M, compaction chooses the longest one. It is a greedy 
> algorithm that preserve all potential candidates for the next round. Block 
> Compaction for WASB increases data durability and allows using block blobs 
> instead of page blobs. By default, block compaction is disabled. Similar to 
> the configuration for page blobs, the client needs to specify HDFS folders 
> where block compaction over block blobs is enabled. 
> Results for HADOOP-14520-01.patch
> Tests run: 704, Failures: 0, Errors: 0, Skipped: 119



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14518) Customize User-Agent header sent in HTTP/HTTPS requests by WASB.

2017-06-12 Thread Georgi Chalakov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Georgi Chalakov updated HADOOP-14518:
-
Target Version/s: 3.0.0-alpha3, 2.8.0  (was: 2.8.0, 3.0.0-alpha3)
  Status: In Progress  (was: Patch Available)

> Customize User-Agent header sent in HTTP/HTTPS requests by WASB.
> 
>
> Key: HADOOP-14518
> URL: https://issues.apache.org/jira/browse/HADOOP-14518
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/azure
>Affects Versions: 3.0.0-alpha3
>Reporter: Georgi Chalakov
>Assignee: Georgi Chalakov
>Priority: Minor
> Attachments: HADOOP-14518-01.patch, HADOOP-14518-01-test.txt, 
> HADOOP-14518-02.patch, HADOOP-14518-03.patch
>
>
> WASB passes a User-Agent header to the Azure back-end. Right now, it uses the 
> default value set by the Azure Client SDK, so Hadoop traffic doesn't appear 
> any different from general Blob traffic. If we customize the User-Agent 
> header, then it will enable better troubleshooting and analysis by Azure 
> service.
> The following configuration
>   
> fs.azure.user.agent.id
> MSFT
>   
> set the user agent to 
>  User-Agent: WASB/3.0.0-alpha4-SNAPSHOT (MSFT) Azure-Storage/4.2.0 
> (JavaJRE 1.8.0_131; WindowsServer2012R2 6.3)
> Test Results :
> Tests run: 703, Failures: 0, Errors: 0, Skipped: 119



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-14519) Client$Connection#waitForWork may suffer spurious wakeup

2017-06-12 Thread Chen Liang (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14519?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047047#comment-16047047
 ] 

Chen Liang edited comment on HADOOP-14519 at 6/12/17 9:10 PM:
--

Thanks [~jzhuge] for the catch! The patch LGTM. One minor thing though, I think 
the patch slightly changed the syntax of this function. (Please correct me if 
I'm wrong though).

Let's denote condition {{calls.isEmpty() && !shouldCloseConnection.get() && 
running.get()}} as condition A. The original logic works that, it only checks A 
before calling wait, meaning when it gets notified, it returns from wait and 
just proceed, potentially possible that A is still true at this point. While 
the patch changes it that, even when it gets notified, as long as A is true, it 
will keep looping. Could you please verify that this change to the syntax is 
correct? Namely, {{calls.isEmpty() && !shouldCloseConnection.get() && 
running.get()}} is indeed expected to be false always after the object gets 
notified. I'm inclined to believe it is all good based the fact of no test 
failures.



was (Author: vagarychen):
Thanks [~jzhuge] for the catch! The patch LGTM. One minor thing though, I think 
the patch slightly changed the syntax of this function. (Please correct me if 
I'm wrong though).

Let's denote condition {{calls.isEmpty() && !shouldCloseConnection.get() && 
running.get()}} as condition A. The original logic works that, it only checks A 
before calling wait, meaning when it gets notified, it returns from wait and 
just proceed, potentially possible that A is still true at this point. While 
the patch changes it that, even when it gets notified, as long as A is true, it 
will keep looping. Could you please verify that this change to the syntax is 
correct? I'm inclined to believe it is all good based the fact of no test 
failures.


> Client$Connection#waitForWork may suffer spurious wakeup
> 
>
> Key: HADOOP-14519
> URL: https://issues.apache.org/jira/browse/HADOOP-14519
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: ipc
>Affects Versions: 2.8.0
>Reporter: John Zhuge
>Assignee: John Zhuge
>Priority: Critical
> Attachments: HADOOP-14519.001.patch
>
>
> {{Client$Connection#waitForWork}} may suffer spurious wakeup because the 
> {{wait}} is not surrounded by a loop. See 
> [https://docs.oracle.com/javase/7/docs/api/java/lang/Object.html#wait()].
> {code:title=Client$Connection#waitForWork}
>   if (calls.isEmpty() && !shouldCloseConnection.get() && running.get())  {
> long timeout = maxIdleTime-
>   (Time.now()-lastActivity.get());
> if (timeout>0) {
>   try {
> wait(timeout);  << spurious wakeup
>   } catch (InterruptedException e) {}
> }
>   }
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14520) Block compaction for WASB

2017-06-12 Thread Georgi Chalakov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Georgi Chalakov updated HADOOP-14520:
-
Status: Patch Available  (was: Open)

> Block compaction for WASB
> -
>
> Key: HADOOP-14520
> URL: https://issues.apache.org/jira/browse/HADOOP-14520
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/azure
>Affects Versions: 3.0.0-alpha3
>Reporter: Georgi Chalakov
>Assignee: Georgi Chalakov
> Attachments: HADOOP-14520-01.patch, HADOOP-14520-01-test.txt
>
>
> Block Compaction for WASB allows uploading new blocks for every hflush/hsync 
> call. When the number of blocks is above a predefined, configurable value, 
> next hflush/hsync triggers the block compaction process. Block compaction 
> replaces a sequence of blocks with one block. From all the sequences with 
> total length less than 4M, compaction chooses the longest one. It is a greedy 
> algorithm that preserve all potential candidates for the next round. Block 
> Compaction for WASB increases data durability and allows using block blobs 
> instead of page blobs. By default, block compaction is disabled. Similar to 
> the configuration for page blobs, the client needs to specify HDFS folders 
> where block compaction over block blobs is enabled. 
> Results for HADOOP-14520-01.patch
> Tests run: 704, Failures: 0, Errors: 0, Skipped: 119



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14520) Block compaction for WASB

2017-06-12 Thread Georgi Chalakov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Georgi Chalakov updated HADOOP-14520:
-
Status: Open  (was: Patch Available)

> Block compaction for WASB
> -
>
> Key: HADOOP-14520
> URL: https://issues.apache.org/jira/browse/HADOOP-14520
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/azure
>Affects Versions: 3.0.0-alpha3
>Reporter: Georgi Chalakov
>Assignee: Georgi Chalakov
> Attachments: HADOOP-14520-01.patch, HADOOP-14520-01-test.txt
>
>
> Block Compaction for WASB allows uploading new blocks for every hflush/hsync 
> call. When the number of blocks is above a predefined, configurable value, 
> next hflush/hsync triggers the block compaction process. Block compaction 
> replaces a sequence of blocks with one block. From all the sequences with 
> total length less than 4M, compaction chooses the longest one. It is a greedy 
> algorithm that preserve all potential candidates for the next round. Block 
> Compaction for WASB increases data durability and allows using block blobs 
> instead of page blobs. By default, block compaction is disabled. Similar to 
> the configuration for page blobs, the client needs to specify HDFS folders 
> where block compaction over block blobs is enabled. 
> Results for HADOOP-14520-01.patch
> Tests run: 704, Failures: 0, Errors: 0, Skipped: 119



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14519) Client$Connection#waitForWork may suffer spurious wakeup

2017-06-12 Thread Chen Liang (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14519?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16047047#comment-16047047
 ] 

Chen Liang commented on HADOOP-14519:
-

Thanks [~jzhuge] for the catch! The patch LGTM. One minor thing though, I think 
the patch slightly changed the syntax of this function. (Please correct me if 
I'm wrong though).

Let's denote condition {{calls.isEmpty() && !shouldCloseConnection.get() && 
running.get()}} as condition A. The original logic works that, it only checks A 
before calling wait, meaning when it gets notified, it returns from wait and 
just proceed, potentially possible that A is still true at this point. While 
the patch changes it that, even when it gets notified, as long as A is true, it 
will keep looping. Could you please verify that this change to the syntax is 
correct? I'm inclined to believe it is all good based the fact of no test 
failures.


> Client$Connection#waitForWork may suffer spurious wakeup
> 
>
> Key: HADOOP-14519
> URL: https://issues.apache.org/jira/browse/HADOOP-14519
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: ipc
>Affects Versions: 2.8.0
>Reporter: John Zhuge
>Assignee: John Zhuge
>Priority: Critical
> Attachments: HADOOP-14519.001.patch
>
>
> {{Client$Connection#waitForWork}} may suffer spurious wakeup because the 
> {{wait}} is not surrounded by a loop. See 
> [https://docs.oracle.com/javase/7/docs/api/java/lang/Object.html#wait()].
> {code:title=Client$Connection#waitForWork}
>   if (calls.isEmpty() && !shouldCloseConnection.get() && running.get())  {
> long timeout = maxIdleTime-
>   (Time.now()-lastActivity.get());
> if (timeout>0) {
>   try {
> wait(timeout);  << spurious wakeup
>   } catch (InterruptedException e) {}
> }
>   }
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14521) KMS client needs retry logic

2017-06-12 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14521?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16046980#comment-16046980
 ] 

Hadoop QA commented on HADOOP-14521:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
16s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 2 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
15s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 13m 
 4s{color} | {color:green} trunk passed {color} |
| {color:red}-1{color} | {color:red} compile {color} | {color:red}  6m 
49s{color} | {color:red} root in trunk failed. {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
48s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
53s{color} | {color:green} trunk passed {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  1m 
18s{color} | {color:red} hadoop-common-project/hadoop-common in trunk has 19 
extant Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
29s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
17s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
31s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 13m  
4s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} javac {color} | {color:red} 13m  4s{color} 
| {color:red} root generated 501 new + 286 unchanged - 0 fixed = 787 total (was 
286) {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
1m 58s{color} | {color:orange} root: The patch generated 2 new + 137 unchanged 
- 0 fixed = 139 total (was 137) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m  
9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
1s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
46s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  7m 
59s{color} | {color:green} hadoop-common in the patch passed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 64m 10s{color} 
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
41s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}147m 18s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.hdfs.TestAclsEndToEnd |
|   | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure |
|   | hadoop.hdfs.TestDFSStripedInputStreamWithRandomECPolicy |
|   | hadoop.hdfs.server.blockmanagement.TestBlockTokenWithDFSStriped |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:14b5c93 |
| JIRA Issue | HADOOP-14521 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12872745/HDFS-11804-trunk-6.patch
 |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  xml  |
| uname | Linux b42f19aae496 3.13.0-116-generic #163-Ubuntu SMP Fri Mar 31 
14:13:22 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 86368cc |
| Default Java | 1.8.0_131 |
| compile | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12521/artifact/patchprocess/branch-compile-root.txt
 |
| findbugs | v3.1.0-RC1 |
| findbugs | 

[jira] [Commented] (HADOOP-13174) Add more debug logs for delegation tokens and authentication

2017-06-12 Thread Xiao Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-13174?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16046969#comment-16046969
 ] 

Xiao Chen commented on HADOOP-13174:


Clarified with [~yzhangal] offline about his question:
bq.  provide a set of debug configuration setting here so we can refer to when 
needed?
>From my experience, it's usually easier if we change general log levels as the 
>comment above. If that's not acceptable for the environment, what we need are 
>(from my experience):
For server:
- org.apache.hadoop.security.authentication.server.AuthenticationFilter
- org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler
- 
org.apache.hadoop.security.token.delegation.web.DelegationTokenAuthenticationFilter
- 
org.apache.hadoop.security.token.delegation.web.DelegationTokenAuthenticationHandler
- org.apache.hadoop.crypto.key.kms.server.*

For client:
- org.apache.hadoop.security.token.delegation.web.DelegationTokenAuthenticator
- 
org.apache.hadoop.security.token.delegation.web.DelegationTokenAuthenticatedURL
- org.apache.hadoop.security.authentication.client.KerberosAuthenticator
- org.apache.hadoop.security.authentication.client.AuthenticatedURL
- org.apache.hadoop.crypto.key.kms.*

> Add more debug logs for delegation tokens and authentication
> 
>
> Key: HADOOP-13174
> URL: https://issues.apache.org/jira/browse/HADOOP-13174
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: security
>Reporter: Xiao Chen
>Assignee: Xiao Chen
>Priority: Minor
> Fix For: 2.9.0, 3.0.0-alpha4
>
> Attachments: HADOOP-13174.01.patch, HADOOP-13174.02.patch, 
> HADOOP-13174.03.patch, HADOOP-13174.04.patch, HADOOP-13174.05.patch
>
>
> Recently I debugged several authentication related problems, and found that 
> the debug logs are not enough to identify a problem.
> This jira improves it by adding more debug/trace logs along the line.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14503) Make RollingAverages a mutable metric

2017-06-12 Thread Hanisha Koneru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14503?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hanisha Koneru updated HADOOP-14503:

Attachment: HADOOP-14503.006.patch

> Make RollingAverages a mutable metric
> -
>
> Key: HADOOP-14503
> URL: https://issues.apache.org/jira/browse/HADOOP-14503
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Reporter: Hanisha Koneru
>Assignee: Hanisha Koneru
> Attachments: HADOOP-14503.001.patch, HADOOP-14503.002.patch, 
> HADOOP-14503.003.patch, HADOOP-14503.004.patch, HADOOP-14503.005.patch, 
> HADOOP-14503.006.patch
>
>
> RollingAverages metric extends on MutableRatesWithAggregation metric and 
> maintains a group of rolling average metrics. This class should be allowed to 
> register as a metric with the MetricSystem.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14497) Logs for KMS delegation token lifecycle

2017-06-12 Thread Yongjun Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14497?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16046922#comment-16046922
 ] 

Yongjun Zhang commented on HADOOP-14497:


Thanks [~xiaochen]!


> Logs for KMS delegation token lifecycle
> ---
>
> Key: HADOOP-14497
> URL: https://issues.apache.org/jira/browse/HADOOP-14497
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Yongjun Zhang
>Assignee: Xiao Chen
> Fix For: 2.9.0, 3.0.0-alpha4
>
>
> We run into quite some customer cases about authentication failures related 
> to KMS delegation token. It would be nice to see a log for each stage of the 
> token:
> 1. creation
> 2. renewal
> 3. removal upon cancel
> 4. remove upon expiration
> So that when we correlate the logs for the same DT, we can have a good 
> picture about what's going on, and what could have caused the authentication 
> failure.
> The same is applicable to other delegation tokens.
> NOTE: When log info about delagation token, we don't want leak user's secret 
> info.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14429) FTPFileSystem#getFsAction always returns FsAction.NONE

2017-06-12 Thread Yongjun Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16046831#comment-16046831
 ] 

Yongjun Zhang commented on HADOOP-14429:


Thanks [~Hongyuan Li]. +1 after one nit: it would have better readability if we 
have a empty line added between different sections in the unit test. Sorry I 
did not look at this earlier.
 

> FTPFileSystem#getFsAction  always returns FsAction.NONE
> ---
>
> Key: HADOOP-14429
> URL: https://issues.apache.org/jira/browse/HADOOP-14429
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 3.0.0-alpha2
>Reporter: Hongyuan Li
>Assignee: Hongyuan Li
> Attachments: HADOOP-14429-001.patch, HADOOP-14429-002.patch, 
> HADOOP-14429-003.patch, HADOOP-14429-004.patch
>
>
>   
> {code}
> private FsAction getFsAction(int accessGroup, FTPFile ftpFile) {
>   FsAction action = FsAction.NONE;
>   if (ftpFile.hasPermission(accessGroup, FTPFile.READ_PERMISSION)) {
>   action.or(FsAction.READ);
>   }
> if (ftpFile.hasPermission(accessGroup, FTPFile.WRITE_PERMISSION)) {
>   action.or(FsAction.WRITE);
> }
> if (ftpFile.hasPermission(accessGroup, FTPFile.EXECUTE_PERMISSION)) {
>   action.or(FsAction.EXECUTE);
> }
> return action;
>   }
> {code}
> from code above, we can see that the  getFsAction method doesnot modify the 
> action generated by FsAction action = FsAction.NONE,which means it return 
> FsAction.NONE all the time;



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14310) RolloverSignerSecretProvider.LOG should be @VisibleForTesting

2017-06-12 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14310?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16046805#comment-16046805
 ] 

Hudson commented on HADOOP-14310:
-

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #11857 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/11857/])
HADOOP-14310. RolloverSignerSecretProvider.LOG should be (templedf: rev 
86368cc7667adfe224df77b039c2ffe8de7f889a)
* (edit) 
hadoop-common-project/hadoop-auth/src/main/java/org/apache/hadoop/security/authentication/util/RolloverSignerSecretProvider.java
* (edit) hadoop-common-project/hadoop-auth/pom.xml


> RolloverSignerSecretProvider.LOG should be @VisibleForTesting
> -
>
> Key: HADOOP-14310
> URL: https://issues.apache.org/jira/browse/HADOOP-14310
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: security
>Affects Versions: 2.8.0
>Reporter: Daniel Templeton
>Assignee: Arun Shanmugam Kumar
>Priority: Minor
>  Labels: newbie
> Fix For: 2.9.0, 3.0.0-alpha4
>
> Attachments: HADOOP-14310.001.patch, HADOOP-14310.002.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Moved] (HADOOP-14521) KMS client needs retry logic

2017-06-12 Thread Rushabh S Shah (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14521?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rushabh S Shah moved HDFS-11804 to HADOOP-14521:


Affects Version/s: (was: 2.6.0)
   2.6.0
 Target Version/s: 2.9.0, 3.0.0-alpha4, 2.8.2  (was: 3.0.0-alpha4, 2.8.2)
   Issue Type: Improvement  (was: Bug)
  Key: HADOOP-14521  (was: HDFS-11804)
  Project: Hadoop Common  (was: Hadoop HDFS)

> KMS client needs retry logic
> 
>
> Key: HADOOP-14521
> URL: https://issues.apache.org/jira/browse/HADOOP-14521
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.6.0
>Reporter: Rushabh S Shah
>Assignee: Rushabh S Shah
> Attachments: HDFS-11804-trunk-1.patch, HDFS-11804-trunk-2.patch, 
> HDFS-11804-trunk-3.patch, HDFS-11804-trunk-4.patch, HDFS-11804-trunk-5.patch, 
> HDFS-11804-trunk-6.patch, HDFS-11804-trunk.patch
>
>
> The kms client appears to have no retry logic – at all.  It's completely 
> decoupled from the ipc retry logic.  This has major impacts if the KMS is 
> unreachable for any reason, including but not limited to network connection 
> issues, timeouts, the +restart during an upgrade+.
> This has some major ramifications:
> # Jobs may fail to submit, although oozie resubmit logic should mask it
> # Non-oozie launchers may experience higher rates if they do not already have 
> retry logic.
> # Tasks reading EZ files will fail, probably be masked by framework reattempts
> # EZ file creation fails after creating a 0-length file – client receives 
> EDEK in the create response, then fails when decrypting the EDEK
> # Bulk hadoop fs copies, and maybe distcp, will prematurely fail



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14521) KMS client needs retry logic

2017-06-12 Thread Rushabh S Shah (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14521?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16046793#comment-16046793
 ] 

Rushabh S Shah commented on HADOOP-14521:
-

bq.  Should this JIRA be moved to project "Hadoop Common" since KMS belongs to 
common?
Moved to common. Thanks [~jzhuge].


> KMS client needs retry logic
> 
>
> Key: HADOOP-14521
> URL: https://issues.apache.org/jira/browse/HADOOP-14521
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.6.0
>Reporter: Rushabh S Shah
>Assignee: Rushabh S Shah
> Attachments: HDFS-11804-trunk-1.patch, HDFS-11804-trunk-2.patch, 
> HDFS-11804-trunk-3.patch, HDFS-11804-trunk-4.patch, HDFS-11804-trunk-5.patch, 
> HDFS-11804-trunk-6.patch, HDFS-11804-trunk.patch
>
>
> The kms client appears to have no retry logic – at all.  It's completely 
> decoupled from the ipc retry logic.  This has major impacts if the KMS is 
> unreachable for any reason, including but not limited to network connection 
> issues, timeouts, the +restart during an upgrade+.
> This has some major ramifications:
> # Jobs may fail to submit, although oozie resubmit logic should mask it
> # Non-oozie launchers may experience higher rates if they do not already have 
> retry logic.
> # Tasks reading EZ files will fail, probably be masked by framework reattempts
> # EZ file creation fails after creating a 0-length file – client receives 
> EDEK in the create response, then fails when decrypting the EDEK
> # Bulk hadoop fs copies, and maybe distcp, will prematurely fail



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14310) RolloverSignerSecretProvider.LOG should be @VisibleForTesting

2017-06-12 Thread Daniel Templeton (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14310?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton updated HADOOP-14310:
--
   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 3.0.0-alpha4
   2.9.0
   Status: Resolved  (was: Patch Available)

Thanks for the patch, [~arunsk].  Committed to trunk and branch-2.

> RolloverSignerSecretProvider.LOG should be @VisibleForTesting
> -
>
> Key: HADOOP-14310
> URL: https://issues.apache.org/jira/browse/HADOOP-14310
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: security
>Affects Versions: 2.8.0
>Reporter: Daniel Templeton
>Assignee: Arun Shanmugam Kumar
>Priority: Minor
>  Labels: newbie
> Fix For: 2.9.0, 3.0.0-alpha4
>
> Attachments: HADOOP-14310.001.patch, HADOOP-14310.002.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14440) Add metrics for connections dropped

2017-06-12 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14440?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HADOOP-14440:
---
Fix Version/s: 2.7.4

Thanks for the work [~ebadger]. I think this is a good improvement for 2.7.4; 
just backported to branch-2.7.

> Add metrics for connections dropped
> ---
>
> Key: HADOOP-14440
> URL: https://issues.apache.org/jira/browse/HADOOP-14440
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Eric Badger
>Assignee: Eric Badger
> Fix For: 2.9.0, 2.7.4, 3.0.0-alpha4, 2.8.2
>
> Attachments: HADOOP-14440.001.patch, HADOOP-14440.002.patch, 
> HADOOP-14440.003.patch
>
>
> Will be useful to figure out when the NN is getting overloaded with more 
> connections than it can handle



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14443) Azure: Add retry and client side failover for authorization, SASKey generation and delegation token generation requests to remote service

2017-06-12 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14443?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16046646#comment-16046646
 ] 

Hadoop QA commented on HADOOP-14443:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
16s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 2 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 13m 
19s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
18s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
14s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
20s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
29s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
15s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
16s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
16s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
11s{color} | {color:green} hadoop-tools/hadoop-azure: The patch generated 0 new 
+ 27 unchanged - 23 fixed = 27 total (was 50) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
33s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
11s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  1m 
18s{color} | {color:green} hadoop-azure in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
17s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 19m 45s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:14b5c93 |
| JIRA Issue | HADOOP-14443 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12872709/HADOOP-14443.4.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux c4ffdac51068 3.13.0-116-generic #163-Ubuntu SMP Fri Mar 31 
14:13:22 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / d64c842 |
| Default Java | 1.8.0_131 |
| findbugs | v3.1.0-RC1 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12520/testReport/ |
| modules | C: hadoop-tools/hadoop-azure U: hadoop-tools/hadoop-azure |
| Console output | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12520/console |
| Powered by | Apache Yetus 0.5.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> Azure: Add retry and client side failover for authorization, SASKey 
> generation and delegation token generation requests to remote service
> -
>
> Key: HADOOP-14443
> URL: https://issues.apache.org/jira/browse/HADOOP-14443
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/azure
>Affects Versions: 2.9.0
>Reporter: Santhosh G Nayak
>Assignee: Santhosh G Nayak
> Fix For: 2.9.0, 3.0.0-alpha4
>
> Attachments: HADOOP-14443.1.patch, HADOOP-14443.2.patch, 
> HADOOP-14443.3.patch, HADOOP-14443.4.patch
>
>
> Currently, {{WasRemoteCallHelper}} can be configured 

[jira] [Updated] (HADOOP-14443) Azure: Add retry and client side failover for authorization, SASKey generation and delegation token generation requests to remote service

2017-06-12 Thread Santhosh G Nayak (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14443?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Santhosh G Nayak updated HADOOP-14443:
--
Attachment: HADOOP-14443.4.patch

Attaching {{HADOOP-14443.4.patch}} to fix {{javac}}, {{Javadoc}} subsystem 
related failures.

> Azure: Add retry and client side failover for authorization, SASKey 
> generation and delegation token generation requests to remote service
> -
>
> Key: HADOOP-14443
> URL: https://issues.apache.org/jira/browse/HADOOP-14443
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/azure
>Affects Versions: 2.9.0
>Reporter: Santhosh G Nayak
>Assignee: Santhosh G Nayak
> Fix For: 2.9.0, 3.0.0-alpha4
>
> Attachments: HADOOP-14443.1.patch, HADOOP-14443.2.patch, 
> HADOOP-14443.3.patch, HADOOP-14443.4.patch
>
>
> Currently, {{WasRemoteCallHelper}} can be configured to talk to only one URL 
> for authorization, SASKey generation and delegation token generation. If for 
> some reason the service is down, all the requests will fail.
> So proposal is to,
> - Add support to configure multiple URLs, so that if communication to one URL 
> fails, client can retry on another instance of the service running on 
> different node for authorization, SASKey generation and delegation token 
> generation. 
> - Rename the configurations {{fs.azure.authorization.remote.service.url}} to 
> {{fs.azure.authorization.remote.service.urls}} and 
> {{fs.azure.cred.service.url}} to {{fs.azure.cred.service.urls}} to support 
> the comma separated list of URLs.
> Introduce a new configuration {{fs.azure.delegation.token.service.urls}} to 
> configure the comma separated list of service URLs to get the delegation 
> token.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14443) Azure: Add retry and client side failover for authorization, SASKey generation and delegation token generation requests to remote service

2017-06-12 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14443?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16046411#comment-16046411
 ] 

Hadoop QA commented on HADOOP-14443:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
15s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 2 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 13m 
54s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
22s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
13s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
22s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
31s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
13s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
21s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} javac {color} | {color:red}  0m 18s{color} 
| {color:red} hadoop-tools_hadoop-azure generated 1 new + 1 unchanged - 0 fixed 
= 2 total (was 1) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
13s{color} | {color:green} hadoop-tools/hadoop-azure: The patch generated 0 new 
+ 27 unchanged - 23 fixed = 27 total (was 50) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
22s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
38s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} javadoc {color} | {color:red}  0m 
12s{color} | {color:red} hadoop-azure in the patch failed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  1m 
27s{color} | {color:green} hadoop-azure in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
20s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 21m  1s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:14b5c93 |
| JIRA Issue | HADOOP-14443 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12872671/HADOOP-14443.3.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux d2e30b1c08b3 3.13.0-116-generic #163-Ubuntu SMP Fri Mar 31 
14:13:22 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / e86eef9 |
| Default Java | 1.8.0_131 |
| findbugs | v3.1.0-RC1 |
| javac | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12519/artifact/patchprocess/diff-compile-javac-hadoop-tools_hadoop-azure.txt
 |
| javadoc | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12519/artifact/patchprocess/patch-javadoc-hadoop-tools_hadoop-azure.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12519/testReport/ |
| modules | C: hadoop-tools/hadoop-azure U: hadoop-tools/hadoop-azure |
| Console output | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/12519/console |
| Powered by | Apache Yetus 0.5.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> Azure: Add retry and client side failover for authorization, SASKey 
> generation and delegation token generation requests to remote service
> -
>
> Key: HADOOP-14443
> URL: https://issues.apache.org/jira/browse/HADOOP-14443
> Project: Hadoop Common
>  Issue Type: Improvement
>   

[jira] [Updated] (HADOOP-14469) FTPFileSystem#listStatus get currentPath and parentPath at the same time, which will cause recursively list stuck

2017-06-12 Thread Hongyuan Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14469?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongyuan Li updated HADOOP-14469:
-
Summary: FTPFileSystem#listStatus get currentPath and parentPath at the 
same time, which will cause recursively list stuck  (was: 
FTPFileSystem#listStatus get currentPath and parentPath, which will cause 
recursively list stuck)

> FTPFileSystem#listStatus get currentPath and parentPath at the same time, 
> which will cause recursively list stuck
> -
>
> Key: HADOOP-14469
> URL: https://issues.apache.org/jira/browse/HADOOP-14469
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 3.0.0-alpha3
> Environment: ftp build by windows7 + Serv-U_64 12.1.0.8 
> code runs any os
>Reporter: Hongyuan Li
>Assignee: Hongyuan Li
>Priority: Critical
> Attachments: HADOOP-14469-001.patch, HADOOP-14469-002.patch, 
> HADOOP-14469-003.patch
>
>
> for some ftpsystems, liststatus method will return new Path(".") and new 
> Path(".."), thus causing list op looping.for example, Serv-U
> We can see the logic in code below:
> {code}
>   private FileStatus[] listStatus(FTPClient client, Path file)
>   throws IOException {
> ……
> FileStatus[] fileStats = new FileStatus[ftpFiles.length];
> for (int i = 0; i < ftpFiles.length; i++) {
>   fileStats[i] = getFileStatus(ftpFiles[i], absolute);
> }
> return fileStats;
>   }
> {code}
> {code}
> public void test() throws Exception{
> FTPFileSystem ftpFileSystem = new FTPFileSystem();
> ftpFileSystem.initialize(new 
> Path("ftp://test:123456@192.168.44.1/;).toUri(),
> new Configuration());
> FileStatus[] fileStatus  = ftpFileSystem.listStatus(new Path("/new"));
> for(FileStatus fileStatus1 : fileStatus)
>   System.out.println(fileStatus1);
> }
> {code}
> using test code below, the test results list below
> {code}
> FileStatus{path=ftp://test:123456@192.168.44.1/new; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14431-002.patch; 
> isDirectory=false; length=2036; replication=1; blocksize=4096; 
> modification_time=149579778; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14486-001.patch; 
> isDirectory=false; length=1322; replication=1; blocksize=4096; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop-main; 
> isDirectory=true; modification_time=149579712; access_time=0; owner=user; 
> group=group; permission=-; isSymlink=false}
> {code}
> In results above, {{FileStatus{path=ftp://test:123456@192.168.44.1/new; ……}} 
> is obviously the current Path, and 
> {{FileStatus{path=ftp://test:123456@192.168.44.1/;…… }} is obviously the 
> parent Path.
> So, if we want to walk the directory recursively, it will stuck.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14429) FTPFileSystem#getFsAction always returns FsAction.NONE

2017-06-12 Thread Hongyuan Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14429?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongyuan Li updated HADOOP-14429:
-
Priority: Major  (was: Trivial)

> FTPFileSystem#getFsAction  always returns FsAction.NONE
> ---
>
> Key: HADOOP-14429
> URL: https://issues.apache.org/jira/browse/HADOOP-14429
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 3.0.0-alpha2
>Reporter: Hongyuan Li
>Assignee: Hongyuan Li
> Attachments: HADOOP-14429-001.patch, HADOOP-14429-002.patch, 
> HADOOP-14429-003.patch, HADOOP-14429-004.patch
>
>
>   
> {code}
> private FsAction getFsAction(int accessGroup, FTPFile ftpFile) {
>   FsAction action = FsAction.NONE;
>   if (ftpFile.hasPermission(accessGroup, FTPFile.READ_PERMISSION)) {
>   action.or(FsAction.READ);
>   }
> if (ftpFile.hasPermission(accessGroup, FTPFile.WRITE_PERMISSION)) {
>   action.or(FsAction.WRITE);
> }
> if (ftpFile.hasPermission(accessGroup, FTPFile.EXECUTE_PERMISSION)) {
>   action.or(FsAction.EXECUTE);
> }
> return action;
>   }
> {code}
> from code above, we can see that the  getFsAction method doesnot modify the 
> action generated by FsAction action = FsAction.NONE,which means it return 
> FsAction.NONE all the time;



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14422) FTPFileSystem and distcp does not work when ecnountered a complex ftp password with char ':' or '@'

2017-06-12 Thread Hongyuan Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-14422?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16046398#comment-16046398
 ] 

Hongyuan Li commented on HADOOP-14422:
--

Ok, you can close the issue.

> FTPFileSystem and distcp does not work when ecnountered a complex ftp 
> password with char ':' or '@'
> ---
>
> Key: HADOOP-14422
> URL: https://issues.apache.org/jira/browse/HADOOP-14422
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs, tools/distcp
>Affects Versions: 3.0.0-alpha2
>Reporter: Hongyuan Li
>Assignee: Hongyuan Li
> Attachments: HADOOP-14422-002.patch, HADOOP-14422-003.patch, 
> HADOOP-14422-004.patch, HADOOP-14422-005.patch, HADOOP-14422.patch
>
>
> ftp username: test passwd:1@123
> hadoop distcp ftp://test:1@123@node2/home/test/  hdfs://piaopiao/test
> 17/05/15 15:24:56 INFO tools.DistCp: Input Options: 
> DistCpOptions{atomicCommit=false, syncFolder=false, deleteMissing=false, 
> ignoreFailures=false, maxMaps=20, sslConfigurationFile='null', 
> copyStrategy='uniformsize', sourceFileListing=null, 
> sourcePaths=[ftp://test:1@123@node2/home/test], targetPath=/test, 
> targetPathExists=true, preserveRawXattrs=false, filtersFile='null'}
> 17/05/15 15:24:57 INFO client.ConfiguredRMFailoverProxyProvider: Failing over 
> to rm2
> 17/05/15 15:24:59 ERROR tools.DistCp: Exception encountered 
> java.io.IOException: Login failed on server - 0.0.0.0, port - 21 as user 
> 'null'
>   at 
> org.apache.hadoop.fs.ftp.FTPFileSystem.connect(FTPFileSystem.java:154)
>   at 
> org.apache.hadoop.fs.ftp.FTPFileSystem.getFileStatus(FTPFileSystem.java:415)
>   at org.apache.hadoop.fs.Globber.getFileStatus(Globber.java:68)
>   at org.apache.hadoop.fs.Globber.glob(Globber.java:263)
>   at org.apache.hadoop.fs.FileSystem.globStatus(FileSystem.java:1630)
>   at 
> org.apache.hadoop.tools.GlobbedCopyListing.doBuildListing(GlobbedCopyListing.java:77)
>   at org.apache.hadoop.tools.CopyListing.buildListing(CopyListing.java:86)
>   at 
> org.apache.hadoop.tools.DistCp.createInputFileListing(DistCp.java:389)
>   at org.apache.hadoop.tools.DistCp.createAndSubmitJob(DistCp.java:187)
>   at org.apache.hadoop.tools.DistCp.execute(DistCp.java:153)
>   at org.apache.hadoop.tools.DistCp.run(DistCp.java:126)
>   at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>   at org.apache.hadoop.tools.DistCp.main(DistCp.java:453)



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14469) FTPFileSystem#listStatus get currentPath and parentPath, which will cause recursively list stuck

2017-06-12 Thread Hongyuan Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14469?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hongyuan Li updated HADOOP-14469:
-
Priority: Critical  (was: Major)

> FTPFileSystem#listStatus get currentPath and parentPath, which will cause 
> recursively list stuck
> 
>
> Key: HADOOP-14469
> URL: https://issues.apache.org/jira/browse/HADOOP-14469
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 3.0.0-alpha3
> Environment: ftp build by windows7 + Serv-U_64 12.1.0.8 
> code runs any os
>Reporter: Hongyuan Li
>Assignee: Hongyuan Li
>Priority: Critical
> Attachments: HADOOP-14469-001.patch, HADOOP-14469-002.patch, 
> HADOOP-14469-003.patch
>
>
> for some ftpsystems, liststatus method will return new Path(".") and new 
> Path(".."), thus causing list op looping.for example, Serv-U
> We can see the logic in code below:
> {code}
>   private FileStatus[] listStatus(FTPClient client, Path file)
>   throws IOException {
> ……
> FileStatus[] fileStats = new FileStatus[ftpFiles.length];
> for (int i = 0; i < ftpFiles.length; i++) {
>   fileStats[i] = getFileStatus(ftpFiles[i], absolute);
> }
> return fileStats;
>   }
> {code}
> {code}
> public void test() throws Exception{
> FTPFileSystem ftpFileSystem = new FTPFileSystem();
> ftpFileSystem.initialize(new 
> Path("ftp://test:123456@192.168.44.1/;).toUri(),
> new Configuration());
> FileStatus[] fileStatus  = ftpFileSystem.listStatus(new Path("/new"));
> for(FileStatus fileStatus1 : fileStatus)
>   System.out.println(fileStatus1);
> }
> {code}
> using test code below, the test results list below
> {code}
> FileStatus{path=ftp://test:123456@192.168.44.1/new; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop; isDirectory=true; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14431-002.patch; 
> isDirectory=false; length=2036; replication=1; blocksize=4096; 
> modification_time=149579778; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/HADOOP-14486-001.patch; 
> isDirectory=false; length=1322; replication=1; blocksize=4096; 
> modification_time=149671698; access_time=0; owner=user; group=group; 
> permission=-; isSymlink=false}
> FileStatus{path=ftp://test:123456@192.168.44.1/new/hadoop-main; 
> isDirectory=true; modification_time=149579712; access_time=0; owner=user; 
> group=group; permission=-; isSymlink=false}
> {code}
> In results above, {{FileStatus{path=ftp://test:123456@192.168.44.1/new; ……}} 
> is obviously the current Path, and 
> {{FileStatus{path=ftp://test:123456@192.168.44.1/;…… }} is obviously the 
> parent Path.
> So, if we want to walk the directory recursively, it will stuck.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-11886) Failed to run distcp against ftp server installed on Windows

2017-06-12 Thread Hongyuan Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11886?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16046394#comment-16046394
 ] 

Hongyuan Li commented on HADOOP-11886:
--

On windows os ,you could set fs.ftp.transfer.mode to FTP.STREAM_TRANSFER_MODE 
which is equal to  10.

> Failed to run distcp against ftp server installed on Windows
> 
>
> Key: HADOOP-11886
> URL: https://issues.apache.org/jira/browse/HADOOP-11886
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: tools/distcp
>Reporter: sam liu
>Assignee: sam liu
>Priority: Blocker
>
> Could run distcp against ftp server installed on Linux, but could NOT run 
> distcp against ftp server installed on Windows(such as IIS ftp service). 
> However, distcp works well for FileZilla ftp server installed on Windows 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-14443) Azure: Add retry and client side failover for authorization, SASKey generation and delegation token generation requests to remote service

2017-06-12 Thread Santhosh G Nayak (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14443?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Santhosh G Nayak updated HADOOP-14443:
--
Attachment: HADOOP-14443.3.patch

[~liuml07] Thanks for the comments. I have fixed few of those and attached a 
new patch,
1.  {{@Override}} annotation to new line.
2.  Marked the parameters such as {{op}}, {{renewer}}, {{service}} and 
{{token}} as constants? 
3. There is a default retry spec which will be used, if user wants to override, 
he can do so. If this is not recommended in Hadoop codebase, we can remove the 
configuration. 
4. Added {{Thread.currentThread().interrupt()}} in 
{{WasbRemoteCallHelper::shouldRetry()}} when {{InterruptedException}} occurs. 
5. This patch consolidates various http helpers which make common SPNEGO 
connection into {{WasbRemoteCallHelper}} and {{SecureWasbRemoteCallHelper}}. It 
makes easier to apply the same retry/failover logic to various remote HTTP 
calls. I am afraid separating it may be difficult.

> Azure: Add retry and client side failover for authorization, SASKey 
> generation and delegation token generation requests to remote service
> -
>
> Key: HADOOP-14443
> URL: https://issues.apache.org/jira/browse/HADOOP-14443
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/azure
>Affects Versions: 2.9.0
>Reporter: Santhosh G Nayak
>Assignee: Santhosh G Nayak
> Fix For: 2.9.0, 3.0.0-alpha4
>
> Attachments: HADOOP-14443.1.patch, HADOOP-14443.2.patch, 
> HADOOP-14443.3.patch
>
>
> Currently, {{WasRemoteCallHelper}} can be configured to talk to only one URL 
> for authorization, SASKey generation and delegation token generation. If for 
> some reason the service is down, all the requests will fail.
> So proposal is to,
> - Add support to configure multiple URLs, so that if communication to one URL 
> fails, client can retry on another instance of the service running on 
> different node for authorization, SASKey generation and delegation token 
> generation. 
> - Rename the configurations {{fs.azure.authorization.remote.service.url}} to 
> {{fs.azure.authorization.remote.service.urls}} and 
> {{fs.azure.cred.service.url}} to {{fs.azure.cred.service.urls}} to support 
> the comma separated list of URLs.
> Introduce a new configuration {{fs.azure.delegation.token.service.urls}} to 
> configure the comma separated list of service URLs to get the delegation 
> token.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org