[jira] [Updated] (HDFS-6776) distcp from insecure cluster (source) to secure cluster (destination) doesn't work via webhdfs

2014-08-28 Thread Yongjun Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6776?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yongjun Zhang updated HDFS-6776:


Attachment: HDFS-6776.010.patch

 distcp from insecure cluster (source) to secure cluster (destination) doesn't 
 work via webhdfs
 --

 Key: HDFS-6776
 URL: https://issues.apache.org/jira/browse/HDFS-6776
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.3.0, 2.5.0
Reporter: Yongjun Zhang
Assignee: Yongjun Zhang
 Attachments: HDFS-6776.001.patch, HDFS-6776.002.patch, 
 HDFS-6776.003.patch, HDFS-6776.004.patch, HDFS-6776.004.patch, 
 HDFS-6776.005.patch, HDFS-6776.006.NullToken.patch, 
 HDFS-6776.006.NullToken.patch, HDFS-6776.007.patch, HDFS-6776.008.patch, 
 HDFS-6776.009.patch, HDFS-6776.010.patch, dummy-token-proxy.js


 Issuing distcp command at the secure cluster side, trying to copy stuff from 
 insecure cluster to secure cluster, and see the following problem:
 {code}
 hadoopuser@yjc5u-1 ~]$ hadoop distcp webhdfs://insure-cluster:port/tmp 
 hdfs://sure-cluster:8020/tmp/tmptgt
 14/07/30 20:06:19 INFO tools.DistCp: Input Options: 
 DistCpOptions{atomicCommit=false, syncFolder=false, deleteMissing=false, 
 ignoreFailures=false, maxMaps=20, sslConfigurationFile='null', 
 copyStrategy='uniformsize', sourceFileListing=null, 
 sourcePaths=[webhdfs://insecure-cluster:port/tmp], 
 targetPath=hdfs://secure-cluster:8020/tmp/tmptgt, targetPathExists=true}
 14/07/30 20:06:19 INFO client.RMProxy: Connecting to ResourceManager at 
 secure-clister:8032
 14/07/30 20:06:20 WARN ssl.FileBasedKeyStoresFactory: The property 
 'ssl.client.truststore.location' has not been set, no TrustStore will be 
 loaded
 14/07/30 20:06:20 WARN security.UserGroupInformation: 
 PriviledgedActionException as:hadoopu...@xyz.com (auth:KERBEROS) 
 cause:java.io.IOException: Failed to get the token for hadoopuser, 
 user=hadoopuser
 14/07/30 20:06:20 WARN security.UserGroupInformation: 
 PriviledgedActionException as:hadoopu...@xyz.com (auth:KERBEROS) 
 cause:java.io.IOException: Failed to get the token for hadoopuser, 
 user=hadoopuser
 14/07/30 20:06:20 ERROR tools.DistCp: Exception encountered 
 java.io.IOException: Failed to get the token for hadoopuser, user=hadoopuser
   at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   at 
 sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
   at 
 sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
   at 
 org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
   at 
 org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem.toIOException(WebHdfsFileSystem.java:365)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem.access$600(WebHdfsFileSystem.java:84)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.shouldRetry(WebHdfsFileSystem.java:618)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.runWithRetry(WebHdfsFileSystem.java:584)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.access$100(WebHdfsFileSystem.java:438)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner$1.run(WebHdfsFileSystem.java:466)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:415)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1554)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.run(WebHdfsFileSystem.java:462)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getDelegationToken(WebHdfsFileSystem.java:1132)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getDelegationToken(WebHdfsFileSystem.java:218)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getAuthParameters(WebHdfsFileSystem.java:403)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem.toUrl(WebHdfsFileSystem.java:424)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractFsPathRunner.getUrl(WebHdfsFileSystem.java:640)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.runWithRetry(WebHdfsFileSystem.java:565)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.access$100(WebHdfsFileSystem.java:438)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner$1.run(WebHdfsFileSystem.java:466)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:415)
   at 
 

[jira] [Commented] (HDFS-6776) distcp from insecure cluster (source) to secure cluster (destination) doesn't work via webhdfs

2014-08-28 Thread Yongjun Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6776?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113387#comment-14113387
 ] 

Yongjun Zhang commented on HDFS-6776:
-

Hi [~atm], thanks a lot for the review and comments. I just uploaded rev 010 to 
address them.

I usually test against real clusters before I upload new rev, but my clusters 
are down now. I will test out tomorrow.

Thanks.

 distcp from insecure cluster (source) to secure cluster (destination) doesn't 
 work via webhdfs
 --

 Key: HDFS-6776
 URL: https://issues.apache.org/jira/browse/HDFS-6776
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.3.0, 2.5.0
Reporter: Yongjun Zhang
Assignee: Yongjun Zhang
 Attachments: HDFS-6776.001.patch, HDFS-6776.002.patch, 
 HDFS-6776.003.patch, HDFS-6776.004.patch, HDFS-6776.004.patch, 
 HDFS-6776.005.patch, HDFS-6776.006.NullToken.patch, 
 HDFS-6776.006.NullToken.patch, HDFS-6776.007.patch, HDFS-6776.008.patch, 
 HDFS-6776.009.patch, HDFS-6776.010.patch, dummy-token-proxy.js


 Issuing distcp command at the secure cluster side, trying to copy stuff from 
 insecure cluster to secure cluster, and see the following problem:
 {code}
 hadoopuser@yjc5u-1 ~]$ hadoop distcp webhdfs://insure-cluster:port/tmp 
 hdfs://sure-cluster:8020/tmp/tmptgt
 14/07/30 20:06:19 INFO tools.DistCp: Input Options: 
 DistCpOptions{atomicCommit=false, syncFolder=false, deleteMissing=false, 
 ignoreFailures=false, maxMaps=20, sslConfigurationFile='null', 
 copyStrategy='uniformsize', sourceFileListing=null, 
 sourcePaths=[webhdfs://insecure-cluster:port/tmp], 
 targetPath=hdfs://secure-cluster:8020/tmp/tmptgt, targetPathExists=true}
 14/07/30 20:06:19 INFO client.RMProxy: Connecting to ResourceManager at 
 secure-clister:8032
 14/07/30 20:06:20 WARN ssl.FileBasedKeyStoresFactory: The property 
 'ssl.client.truststore.location' has not been set, no TrustStore will be 
 loaded
 14/07/30 20:06:20 WARN security.UserGroupInformation: 
 PriviledgedActionException as:hadoopu...@xyz.com (auth:KERBEROS) 
 cause:java.io.IOException: Failed to get the token for hadoopuser, 
 user=hadoopuser
 14/07/30 20:06:20 WARN security.UserGroupInformation: 
 PriviledgedActionException as:hadoopu...@xyz.com (auth:KERBEROS) 
 cause:java.io.IOException: Failed to get the token for hadoopuser, 
 user=hadoopuser
 14/07/30 20:06:20 ERROR tools.DistCp: Exception encountered 
 java.io.IOException: Failed to get the token for hadoopuser, user=hadoopuser
   at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   at 
 sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
   at 
 sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
   at 
 org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
   at 
 org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem.toIOException(WebHdfsFileSystem.java:365)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem.access$600(WebHdfsFileSystem.java:84)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.shouldRetry(WebHdfsFileSystem.java:618)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.runWithRetry(WebHdfsFileSystem.java:584)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.access$100(WebHdfsFileSystem.java:438)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner$1.run(WebHdfsFileSystem.java:466)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:415)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1554)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.run(WebHdfsFileSystem.java:462)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getDelegationToken(WebHdfsFileSystem.java:1132)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getDelegationToken(WebHdfsFileSystem.java:218)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem.getAuthParameters(WebHdfsFileSystem.java:403)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem.toUrl(WebHdfsFileSystem.java:424)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractFsPathRunner.getUrl(WebHdfsFileSystem.java:640)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.runWithRetry(WebHdfsFileSystem.java:565)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$AbstractRunner.access$100(WebHdfsFileSystem.java:438)
   at 
 

[jira] [Commented] (HDFS-6914) Resolve huge memory consumption Issue with OIV processing PB-based fsimages

2014-08-28 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6914?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113415#comment-14113415
 ] 

Karthik Kambatla commented on HDFS-6914:


Do we really want to target this to 2.5.1? If not, I would suggest targetting 
2.6. 

 Resolve huge memory consumption Issue with OIV processing PB-based fsimages
 ---

 Key: HDFS-6914
 URL: https://issues.apache.org/jira/browse/HDFS-6914
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.4.1
Reporter: Hao Chen
  Labels: hdfs
 Fix For: 2.5.1

 Attachments: HDFS-6914.patch


 For better managing and supporting a lot of large hadoop clusters in 
 production, we internally need to automatically export fsimage to delimited 
 text files in LSR style and then analyse with hive or pig or build system 
 metrics for real-time analyzing. 
 However  due to the internal layout changes introduced by the protobuf-based 
 fsimage, OIV processing program consumes excessive amount of memory. For 
 example, in order to export the fsimage in size of 8GB, it should have taken 
 about 85GB memory which is really not reasonable and impacted performance of 
 other services badly in the same server.
 To resolve above problem, I submit this patch which will reduce memory 
 consumption of OIV LSR processing by 50%.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6831) Inconsistency between 'hdfs dfsadmin' and 'hdfs dfsadmin -help'

2014-08-28 Thread Xiaoyu Yao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6831?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao updated HDFS-6831:
-

Attachment: HDFS-6831.0.patch

 Inconsistency between 'hdfs dfsadmin' and 'hdfs dfsadmin -help'
 ---

 Key: HDFS-6831
 URL: https://issues.apache.org/jira/browse/HDFS-6831
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.4.0
Reporter: Akira AJISAKA
Priority: Minor
  Labels: newbie
 Attachments: HDFS-6831.0.patch


 There is an inconsistency between the console outputs of 'hdfs dfsadmin' 
 command and 'hdfs dfsadmin -help' command.
 {code}
 [root@trunk ~]# hdfs dfsadmin
 Usage: java DFSAdmin
 Note: Administrative commands can only be run as the HDFS superuser.
[-report]
[-safemode enter | leave | get | wait]
[-allowSnapshot snapshotDir]
[-disallowSnapshot snapshotDir]
[-saveNamespace]
[-rollEdits]
[-restoreFailedStorage true|false|check]
[-refreshNodes]
[-finalizeUpgrade]
[-rollingUpgrade [query|prepare|finalize]]
[-metasave filename]
[-refreshServiceAcl]
[-refreshUserToGroupsMappings]
[-refreshSuperUserGroupsConfiguration]
[-refreshCallQueue]
[-refresh]
[-printTopology]
[-refreshNamenodes datanodehost:port]
[-deleteBlockPool datanode-host:port blockpoolId [force]]
[-setQuota quota dirname...dirname]
[-clrQuota dirname...dirname]
[-setSpaceQuota quota dirname...dirname]
[-clrSpaceQuota dirname...dirname]
[-setBalancerBandwidth bandwidth in bytes per second]
[-fetchImage local directory]
[-shutdownDatanode datanode_host:ipc_port [upgrade]]
[-getDatanodeInfo datanode_host:ipc_port]
[-help [cmd]]
 {code}
 {code}
 [root@trunk ~]# hdfs dfsadmin -help
 hadoop dfsadmin performs DFS administrative commands.
 The full syntax is: 
 hadoop dfsadmin
   [-report [-live] [-dead] [-decommissioning]]
   [-safemode enter | leave | get | wait]
   [-saveNamespace]
   [-rollEdits]
   [-restoreFailedStorage true|false|check]
   [-refreshNodes]
   [-setQuota quota dirname...dirname]
   [-clrQuota dirname...dirname]
   [-setSpaceQuota quota dirname...dirname]
   [-clrSpaceQuota dirname...dirname]
   [-finalizeUpgrade]
   [-rollingUpgrade [query|prepare|finalize]]
   [-refreshServiceAcl]
   [-refreshUserToGroupsMappings]
   [-refreshSuperUserGroupsConfiguration]
   [-refreshCallQueue]
   [-refresh host:ipc_port key [arg1..argn]
   [-printTopology]
   [-refreshNamenodes datanodehost:port]
   [-deleteBlockPool datanodehost:port blockpoolId [force]]
   [-setBalancerBandwidth bandwidth]
   [-fetchImage local directory]
   [-allowSnapshot snapshotDir]
   [-disallowSnapshot snapshotDir]
   [-shutdownDatanode datanode_host:ipc_port [upgrade]]
   [-getDatanodeInfo datanode_host:ipc_port
   [-help [cmd]
 {code}
 These two outputs should be the same.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6634) inotify in HDFS

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113469#comment-14113469
 ] 

Hadoop QA commented on HDFS-6634:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12664819/HDFS-6634.7.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 5 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:red}-1 findbugs{color}.  The patch appears to introduce 1 new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs 
hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal:

  org.apache.hadoop.security.TestRefreshUserMappings
  
org.apache.hadoop.hdfs.server.namenode.ha.TestPipelinesFailover

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7803//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7803//artifact/trunk/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7803//console

This message is automatically generated.

 inotify in HDFS
 ---

 Key: HDFS-6634
 URL: https://issues.apache.org/jira/browse/HDFS-6634
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: hdfs-client, namenode, qjm
Reporter: James Thomas
Assignee: James Thomas
 Attachments: HDFS-6634.2.patch, HDFS-6634.3.patch, HDFS-6634.4.patch, 
 HDFS-6634.5.patch, HDFS-6634.6.patch, HDFS-6634.7.patch, HDFS-6634.patch, 
 inotify-design.2.pdf, inotify-design.3.pdf, inotify-design.4.pdf, 
 inotify-design.pdf, inotify-intro.2.pdf, inotify-intro.pdf


 Design a mechanism for applications like search engines to access the HDFS 
 edit stream.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6831) Inconsistency between 'hdfs dfsadmin' and 'hdfs dfsadmin -help'

2014-08-28 Thread Xiaoyu Yao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113475#comment-14113475
 ] 

Xiaoyu Yao commented on HDFS-6831:
--

Attach a patch that fixed the issue with a commonUsageSummary string for both 
commands and misc minor fixes.
* add the missing -metasave filename
* add the Note: Administrative commands can only be run with superuser 
permission.\n to summary of PrintHelp() as Jing suggested.
* change java DfsAdmin to hdfs DfsAdmin in PrintUsage()

Not sure why quota and rolling upgrade commands have different for display 
format with two + instead of  - and extra spaces. I just keep it as is and 
guess maybe it is to differentiate as new features. 

\t[ + SetQuotaCommand.USAGE + ]\n +

 Inconsistency between 'hdfs dfsadmin' and 'hdfs dfsadmin -help'
 ---

 Key: HDFS-6831
 URL: https://issues.apache.org/jira/browse/HDFS-6831
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.4.0
Reporter: Akira AJISAKA
Priority: Minor
  Labels: newbie
 Attachments: HDFS-6831.0.patch


 There is an inconsistency between the console outputs of 'hdfs dfsadmin' 
 command and 'hdfs dfsadmin -help' command.
 {code}
 [root@trunk ~]# hdfs dfsadmin
 Usage: java DFSAdmin
 Note: Administrative commands can only be run as the HDFS superuser.
[-report]
[-safemode enter | leave | get | wait]
[-allowSnapshot snapshotDir]
[-disallowSnapshot snapshotDir]
[-saveNamespace]
[-rollEdits]
[-restoreFailedStorage true|false|check]
[-refreshNodes]
[-finalizeUpgrade]
[-rollingUpgrade [query|prepare|finalize]]
[-metasave filename]
[-refreshServiceAcl]
[-refreshUserToGroupsMappings]
[-refreshSuperUserGroupsConfiguration]
[-refreshCallQueue]
[-refresh]
[-printTopology]
[-refreshNamenodes datanodehost:port]
[-deleteBlockPool datanode-host:port blockpoolId [force]]
[-setQuota quota dirname...dirname]
[-clrQuota dirname...dirname]
[-setSpaceQuota quota dirname...dirname]
[-clrSpaceQuota dirname...dirname]
[-setBalancerBandwidth bandwidth in bytes per second]
[-fetchImage local directory]
[-shutdownDatanode datanode_host:ipc_port [upgrade]]
[-getDatanodeInfo datanode_host:ipc_port]
[-help [cmd]]
 {code}
 {code}
 [root@trunk ~]# hdfs dfsadmin -help
 hadoop dfsadmin performs DFS administrative commands.
 The full syntax is: 
 hadoop dfsadmin
   [-report [-live] [-dead] [-decommissioning]]
   [-safemode enter | leave | get | wait]
   [-saveNamespace]
   [-rollEdits]
   [-restoreFailedStorage true|false|check]
   [-refreshNodes]
   [-setQuota quota dirname...dirname]
   [-clrQuota dirname...dirname]
   [-setSpaceQuota quota dirname...dirname]
   [-clrSpaceQuota dirname...dirname]
   [-finalizeUpgrade]
   [-rollingUpgrade [query|prepare|finalize]]
   [-refreshServiceAcl]
   [-refreshUserToGroupsMappings]
   [-refreshSuperUserGroupsConfiguration]
   [-refreshCallQueue]
   [-refresh host:ipc_port key [arg1..argn]
   [-printTopology]
   [-refreshNamenodes datanodehost:port]
   [-deleteBlockPool datanodehost:port blockpoolId [force]]
   [-setBalancerBandwidth bandwidth]
   [-fetchImage local directory]
   [-allowSnapshot snapshotDir]
   [-disallowSnapshot snapshotDir]
   [-shutdownDatanode datanode_host:ipc_port [upgrade]]
   [-getDatanodeInfo datanode_host:ipc_port
   [-help [cmd]
 {code}
 These two outputs should be the same.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6831) Inconsistency between 'hdfs dfsadmin' and 'hdfs dfsadmin -help'

2014-08-28 Thread Xiaoyu Yao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6831?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao updated HDFS-6831:
-

Status: Patch Available  (was: Open)

 Inconsistency between 'hdfs dfsadmin' and 'hdfs dfsadmin -help'
 ---

 Key: HDFS-6831
 URL: https://issues.apache.org/jira/browse/HDFS-6831
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.4.0
Reporter: Akira AJISAKA
Priority: Minor
  Labels: newbie
 Attachments: HDFS-6831.0.patch


 There is an inconsistency between the console outputs of 'hdfs dfsadmin' 
 command and 'hdfs dfsadmin -help' command.
 {code}
 [root@trunk ~]# hdfs dfsadmin
 Usage: java DFSAdmin
 Note: Administrative commands can only be run as the HDFS superuser.
[-report]
[-safemode enter | leave | get | wait]
[-allowSnapshot snapshotDir]
[-disallowSnapshot snapshotDir]
[-saveNamespace]
[-rollEdits]
[-restoreFailedStorage true|false|check]
[-refreshNodes]
[-finalizeUpgrade]
[-rollingUpgrade [query|prepare|finalize]]
[-metasave filename]
[-refreshServiceAcl]
[-refreshUserToGroupsMappings]
[-refreshSuperUserGroupsConfiguration]
[-refreshCallQueue]
[-refresh]
[-printTopology]
[-refreshNamenodes datanodehost:port]
[-deleteBlockPool datanode-host:port blockpoolId [force]]
[-setQuota quota dirname...dirname]
[-clrQuota dirname...dirname]
[-setSpaceQuota quota dirname...dirname]
[-clrSpaceQuota dirname...dirname]
[-setBalancerBandwidth bandwidth in bytes per second]
[-fetchImage local directory]
[-shutdownDatanode datanode_host:ipc_port [upgrade]]
[-getDatanodeInfo datanode_host:ipc_port]
[-help [cmd]]
 {code}
 {code}
 [root@trunk ~]# hdfs dfsadmin -help
 hadoop dfsadmin performs DFS administrative commands.
 The full syntax is: 
 hadoop dfsadmin
   [-report [-live] [-dead] [-decommissioning]]
   [-safemode enter | leave | get | wait]
   [-saveNamespace]
   [-rollEdits]
   [-restoreFailedStorage true|false|check]
   [-refreshNodes]
   [-setQuota quota dirname...dirname]
   [-clrQuota dirname...dirname]
   [-setSpaceQuota quota dirname...dirname]
   [-clrSpaceQuota dirname...dirname]
   [-finalizeUpgrade]
   [-rollingUpgrade [query|prepare|finalize]]
   [-refreshServiceAcl]
   [-refreshUserToGroupsMappings]
   [-refreshSuperUserGroupsConfiguration]
   [-refreshCallQueue]
   [-refresh host:ipc_port key [arg1..argn]
   [-printTopology]
   [-refreshNamenodes datanodehost:port]
   [-deleteBlockPool datanodehost:port blockpoolId [force]]
   [-setBalancerBandwidth bandwidth]
   [-fetchImage local directory]
   [-allowSnapshot snapshotDir]
   [-disallowSnapshot snapshotDir]
   [-shutdownDatanode datanode_host:ipc_port [upgrade]]
   [-getDatanodeInfo datanode_host:ipc_port
   [-help [cmd]
 {code}
 These two outputs should be the same.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6831) Inconsistency between 'hdfs dfsadmin' and 'hdfs dfsadmin -help'

2014-08-28 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113490#comment-14113490
 ] 

Akira AJISAKA commented on HDFS-6831:
-

Thanks [~xiaoyuyao] for the patch!
Would you change hdfs DFSAdmin to hdfs dfsadmin?

 Inconsistency between 'hdfs dfsadmin' and 'hdfs dfsadmin -help'
 ---

 Key: HDFS-6831
 URL: https://issues.apache.org/jira/browse/HDFS-6831
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.4.0
Reporter: Akira AJISAKA
Priority: Minor
  Labels: newbie
 Attachments: HDFS-6831.0.patch


 There is an inconsistency between the console outputs of 'hdfs dfsadmin' 
 command and 'hdfs dfsadmin -help' command.
 {code}
 [root@trunk ~]# hdfs dfsadmin
 Usage: java DFSAdmin
 Note: Administrative commands can only be run as the HDFS superuser.
[-report]
[-safemode enter | leave | get | wait]
[-allowSnapshot snapshotDir]
[-disallowSnapshot snapshotDir]
[-saveNamespace]
[-rollEdits]
[-restoreFailedStorage true|false|check]
[-refreshNodes]
[-finalizeUpgrade]
[-rollingUpgrade [query|prepare|finalize]]
[-metasave filename]
[-refreshServiceAcl]
[-refreshUserToGroupsMappings]
[-refreshSuperUserGroupsConfiguration]
[-refreshCallQueue]
[-refresh]
[-printTopology]
[-refreshNamenodes datanodehost:port]
[-deleteBlockPool datanode-host:port blockpoolId [force]]
[-setQuota quota dirname...dirname]
[-clrQuota dirname...dirname]
[-setSpaceQuota quota dirname...dirname]
[-clrSpaceQuota dirname...dirname]
[-setBalancerBandwidth bandwidth in bytes per second]
[-fetchImage local directory]
[-shutdownDatanode datanode_host:ipc_port [upgrade]]
[-getDatanodeInfo datanode_host:ipc_port]
[-help [cmd]]
 {code}
 {code}
 [root@trunk ~]# hdfs dfsadmin -help
 hadoop dfsadmin performs DFS administrative commands.
 The full syntax is: 
 hadoop dfsadmin
   [-report [-live] [-dead] [-decommissioning]]
   [-safemode enter | leave | get | wait]
   [-saveNamespace]
   [-rollEdits]
   [-restoreFailedStorage true|false|check]
   [-refreshNodes]
   [-setQuota quota dirname...dirname]
   [-clrQuota dirname...dirname]
   [-setSpaceQuota quota dirname...dirname]
   [-clrSpaceQuota dirname...dirname]
   [-finalizeUpgrade]
   [-rollingUpgrade [query|prepare|finalize]]
   [-refreshServiceAcl]
   [-refreshUserToGroupsMappings]
   [-refreshSuperUserGroupsConfiguration]
   [-refreshCallQueue]
   [-refresh host:ipc_port key [arg1..argn]
   [-printTopology]
   [-refreshNamenodes datanodehost:port]
   [-deleteBlockPool datanodehost:port blockpoolId [force]]
   [-setBalancerBandwidth bandwidth]
   [-fetchImage local directory]
   [-allowSnapshot snapshotDir]
   [-disallowSnapshot snapshotDir]
   [-shutdownDatanode datanode_host:ipc_port [upgrade]]
   [-getDatanodeInfo datanode_host:ipc_port
   [-help [cmd]
 {code}
 These two outputs should be the same.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6831) Inconsistency between 'hdfs dfsadmin' and 'hdfs dfsadmin -help'

2014-08-28 Thread Xiaoyu Yao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6831?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao updated HDFS-6831:
-

Attachment: HDFS-6831.1.patch

Update hdfs DFSAdmin to hdfs dfsadmin 

 Inconsistency between 'hdfs dfsadmin' and 'hdfs dfsadmin -help'
 ---

 Key: HDFS-6831
 URL: https://issues.apache.org/jira/browse/HDFS-6831
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.4.0
Reporter: Akira AJISAKA
Priority: Minor
  Labels: newbie
 Attachments: HDFS-6831.0.patch, HDFS-6831.1.patch


 There is an inconsistency between the console outputs of 'hdfs dfsadmin' 
 command and 'hdfs dfsadmin -help' command.
 {code}
 [root@trunk ~]# hdfs dfsadmin
 Usage: java DFSAdmin
 Note: Administrative commands can only be run as the HDFS superuser.
[-report]
[-safemode enter | leave | get | wait]
[-allowSnapshot snapshotDir]
[-disallowSnapshot snapshotDir]
[-saveNamespace]
[-rollEdits]
[-restoreFailedStorage true|false|check]
[-refreshNodes]
[-finalizeUpgrade]
[-rollingUpgrade [query|prepare|finalize]]
[-metasave filename]
[-refreshServiceAcl]
[-refreshUserToGroupsMappings]
[-refreshSuperUserGroupsConfiguration]
[-refreshCallQueue]
[-refresh]
[-printTopology]
[-refreshNamenodes datanodehost:port]
[-deleteBlockPool datanode-host:port blockpoolId [force]]
[-setQuota quota dirname...dirname]
[-clrQuota dirname...dirname]
[-setSpaceQuota quota dirname...dirname]
[-clrSpaceQuota dirname...dirname]
[-setBalancerBandwidth bandwidth in bytes per second]
[-fetchImage local directory]
[-shutdownDatanode datanode_host:ipc_port [upgrade]]
[-getDatanodeInfo datanode_host:ipc_port]
[-help [cmd]]
 {code}
 {code}
 [root@trunk ~]# hdfs dfsadmin -help
 hadoop dfsadmin performs DFS administrative commands.
 The full syntax is: 
 hadoop dfsadmin
   [-report [-live] [-dead] [-decommissioning]]
   [-safemode enter | leave | get | wait]
   [-saveNamespace]
   [-rollEdits]
   [-restoreFailedStorage true|false|check]
   [-refreshNodes]
   [-setQuota quota dirname...dirname]
   [-clrQuota dirname...dirname]
   [-setSpaceQuota quota dirname...dirname]
   [-clrSpaceQuota dirname...dirname]
   [-finalizeUpgrade]
   [-rollingUpgrade [query|prepare|finalize]]
   [-refreshServiceAcl]
   [-refreshUserToGroupsMappings]
   [-refreshSuperUserGroupsConfiguration]
   [-refreshCallQueue]
   [-refresh host:ipc_port key [arg1..argn]
   [-printTopology]
   [-refreshNamenodes datanodehost:port]
   [-deleteBlockPool datanodehost:port blockpoolId [force]]
   [-setBalancerBandwidth bandwidth]
   [-fetchImage local directory]
   [-allowSnapshot snapshotDir]
   [-disallowSnapshot snapshotDir]
   [-shutdownDatanode datanode_host:ipc_port [upgrade]]
   [-getDatanodeInfo datanode_host:ipc_port
   [-help [cmd]
 {code}
 These two outputs should be the same.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6840) Clients are always sent to the same datanode when read is off rack

2014-08-28 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-6840:
--

Status: Patch Available  (was: Open)

 Clients are always sent to the same datanode when read is off rack
 --

 Key: HDFS-6840
 URL: https://issues.apache.org/jira/browse/HDFS-6840
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Jason Lowe
Assignee: Andrew Wang
Priority: Critical
 Attachments: hdfs-6840.001.patch


 After HDFS-6268 the sorting order of block locations is deterministic for a 
 given block and locality level (e.g.: local, rack. off-rack), so off-rack 
 clients all see the same datanode for the same block.  This leads to very 
 poor behavior in distributed cache localization and other scenarios where 
 many clients all want the same block data at approximately the same time.  
 The one datanode is crushed by the load while the other replicas only handle 
 local and rack-local requests.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6840) Clients are always sent to the same datanode when read is off rack

2014-08-28 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-6840:
--

Attachment: hdfs-6840.001.patch

Here's a patch that removes the random seed stuff. Fortunately HDFS-6701 
hasn't been released yet, so we don't need to do any config key deprecation.

branch-2 is probably going to need a new patch like the others. This might be a 
good time to try and figure out why we have those extra test files.

 Clients are always sent to the same datanode when read is off rack
 --

 Key: HDFS-6840
 URL: https://issues.apache.org/jira/browse/HDFS-6840
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Jason Lowe
Assignee: Andrew Wang
Priority: Critical
 Attachments: hdfs-6840.001.patch


 After HDFS-6268 the sorting order of block locations is deterministic for a 
 given block and locality level (e.g.: local, rack. off-rack), so off-rack 
 clients all see the same datanode for the same block.  This leads to very 
 poor behavior in distributed cache localization and other scenarios where 
 many clients all want the same block data at approximately the same time.  
 The one datanode is crushed by the load while the other replicas only handle 
 local and rack-local requests.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-3529) Use direct buffers for data in write path

2014-08-28 Thread Yi Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113546#comment-14113546
 ] 

Yi Liu commented on HDFS-3529:
--

Hi [~scurrilous], could I take this JIRA and continue the work?

Currently HDFS cache is supported, and memory support is ongoing, so I think 
this will be a good improvement. 

Of course if use HDD, the benefit we get from saving bytes copy by using direct 
buffer is very small compared to the Disk IO.

 Use direct buffers for data in write path
 -

 Key: HDFS-3529
 URL: https://issues.apache.org/jira/browse/HDFS-3529
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode, performance
Affects Versions: 2.0.0-alpha
Reporter: Todd Lipcon
Assignee: Trevor Robinson
 Attachments: HDFS-3529.patch, dfsio-x86-trunk-vs-3529.png


 The write path currently makes several unnecessary data copies in order to go 
 to and from byte arrays. We can improve performance by using direct byte 
 buffers to avoid the copy. This is also a prerequisite for native checksum 
 calculation (HDFS-3528)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-3529) Use direct buffers for data in write path

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113551#comment-14113551
 ] 

Hadoop QA commented on HDFS-3529:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12552318/dfsio-x86-trunk-vs-3529.png
  against trunk revision .

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7808//console

This message is automatically generated.

 Use direct buffers for data in write path
 -

 Key: HDFS-3529
 URL: https://issues.apache.org/jira/browse/HDFS-3529
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode, performance
Affects Versions: 2.0.0-alpha
Reporter: Todd Lipcon
Assignee: Trevor Robinson
 Attachments: HDFS-3529.patch, dfsio-x86-trunk-vs-3529.png


 The write path currently makes several unnecessary data copies in order to go 
 to and from byte arrays. We can improve performance by using direct byte 
 buffers to avoid the copy. This is also a prerequisite for native checksum 
 calculation (HDFS-3528)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6840) Clients are always sent to the same datanode when read is off rack

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113557#comment-14113557
 ] 

Hadoop QA commented on HDFS-6840:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12664852/hdfs-6840.001.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:red}-1 javac{color:red}.  The patch appears to cause the build to 
fail.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7807//console

This message is automatically generated.

 Clients are always sent to the same datanode when read is off rack
 --

 Key: HDFS-6840
 URL: https://issues.apache.org/jira/browse/HDFS-6840
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Jason Lowe
Assignee: Andrew Wang
Priority: Critical
 Attachments: hdfs-6840.001.patch


 After HDFS-6268 the sorting order of block locations is deterministic for a 
 given block and locality level (e.g.: local, rack. off-rack), so off-rack 
 clients all see the same datanode for the same block.  This leads to very 
 poor behavior in distributed cache localization and other scenarios where 
 many clients all want the same block data at approximately the same time.  
 The one datanode is crushed by the load while the other replicas only handle 
 local and rack-local requests.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-3529) Use direct buffers for data in write path

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113583#comment-14113583
 ] 

Hadoop QA commented on HDFS-3529:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7810//console

This message is automatically generated.

 Use direct buffers for data in write path
 -

 Key: HDFS-3529
 URL: https://issues.apache.org/jira/browse/HDFS-3529
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode, performance
Affects Versions: 2.0.0-alpha
Reporter: Todd Lipcon
Assignee: Trevor Robinson
 Attachments: HDFS-3529.patch, dfsio-x86-trunk-vs-3529.png


 The write path currently makes several unnecessary data copies in order to go 
 to and from byte arrays. We can improve performance by using direct byte 
 buffers to avoid the copy. This is also a prerequisite for native checksum 
 calculation (HDFS-3528)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-3529) Use direct buffers for data in write path

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113613#comment-14113613
 ] 

Hadoop QA commented on HDFS-3529:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7813//console

This message is automatically generated.

 Use direct buffers for data in write path
 -

 Key: HDFS-3529
 URL: https://issues.apache.org/jira/browse/HDFS-3529
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode, performance
Affects Versions: 2.0.0-alpha
Reporter: Todd Lipcon
Assignee: Trevor Robinson
 Attachments: HDFS-3529.patch, dfsio-x86-trunk-vs-3529.png


 The write path currently makes several unnecessary data copies in order to go 
 to and from byte arrays. We can improve performance by using direct byte 
 buffers to avoid the copy. This is also a prerequisite for native checksum 
 calculation (HDFS-3528)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-3529) Use direct buffers for data in write path

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113624#comment-14113624
 ] 

Hadoop QA commented on HDFS-3529:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7814//console

This message is automatically generated.

 Use direct buffers for data in write path
 -

 Key: HDFS-3529
 URL: https://issues.apache.org/jira/browse/HDFS-3529
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode, performance
Affects Versions: 2.0.0-alpha
Reporter: Todd Lipcon
Assignee: Trevor Robinson
 Attachments: HDFS-3529.patch, dfsio-x86-trunk-vs-3529.png


 The write path currently makes several unnecessary data copies in order to go 
 to and from byte arrays. We can improve performance by using direct byte 
 buffers to avoid the copy. This is also a prerequisite for native checksum 
 calculation (HDFS-3528)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-3529) Use direct buffers for data in write path

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113626#comment-14113626
 ] 

Hadoop QA commented on HDFS-3529:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7815//console

This message is automatically generated.

 Use direct buffers for data in write path
 -

 Key: HDFS-3529
 URL: https://issues.apache.org/jira/browse/HDFS-3529
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode, performance
Affects Versions: 2.0.0-alpha
Reporter: Todd Lipcon
Assignee: Trevor Robinson
 Attachments: HDFS-3529.patch, dfsio-x86-trunk-vs-3529.png


 The write path currently makes several unnecessary data copies in order to go 
 to and from byte arrays. We can improve performance by using direct byte 
 buffers to avoid the copy. This is also a prerequisite for native checksum 
 calculation (HDFS-3528)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-3529) Use direct buffers for data in write path

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113627#comment-14113627
 ] 

Hadoop QA commented on HDFS-3529:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7818//console

This message is automatically generated.

 Use direct buffers for data in write path
 -

 Key: HDFS-3529
 URL: https://issues.apache.org/jira/browse/HDFS-3529
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode, performance
Affects Versions: 2.0.0-alpha
Reporter: Todd Lipcon
Assignee: Trevor Robinson
 Attachments: HDFS-3529.patch, dfsio-x86-trunk-vs-3529.png


 The write path currently makes several unnecessary data copies in order to go 
 to and from byte arrays. We can improve performance by using direct byte 
 buffers to avoid the copy. This is also a prerequisite for native checksum 
 calculation (HDFS-3528)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-3529) Use direct buffers for data in write path

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113631#comment-14113631
 ] 

Hadoop QA commented on HDFS-3529:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7819//console

This message is automatically generated.

 Use direct buffers for data in write path
 -

 Key: HDFS-3529
 URL: https://issues.apache.org/jira/browse/HDFS-3529
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode, performance
Affects Versions: 2.0.0-alpha
Reporter: Todd Lipcon
Assignee: Trevor Robinson
 Attachments: HDFS-3529.patch, dfsio-x86-trunk-vs-3529.png


 The write path currently makes several unnecessary data copies in order to go 
 to and from byte arrays. We can improve performance by using direct byte 
 buffers to avoid the copy. This is also a prerequisite for native checksum 
 calculation (HDFS-3528)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6705) Create an XAttr that disallows the HDFS admin from accessing a file

2014-08-28 Thread Charles Lamb (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6705?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113688#comment-14113688
 ] 

Charles Lamb commented on HDFS-6705:


[~hitliuyi],

{quote}
if the super user is the owner, then it can't access the file?
{quote}

Yes, that's the intended behavior.

{quote}
It is settable by any user which has hdfs access to that file.
It can only be set and never removed.

Then any user who has hdfs access can easily prevent HDFS admin to access file 
and the admin can't access that file any more. Could we find a better way?
{quote}

The xattr only prevents read access to the contents and is specifically 
intended to prevent the superuser from accessing the file. I realize that the 
current patch prevents read and write access to the file. Based on 
[~andrew.wang]'s comments, I'm going to change it to only restrict read access, 
not write access. The superuser will still have access to the meta data 
(xattrs, perms, etc.) so they can chown, chmod, etc.

It's unfortunate that the superuser could lock themselves out of a file, but it 
needs to be that way or else the root could just chown a users file to root and 
then read it.


 Create an XAttr that disallows the HDFS admin from accessing a file
 ---

 Key: HDFS-6705
 URL: https://issues.apache.org/jira/browse/HDFS-6705
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode, security
Affects Versions: 3.0.0
Reporter: Charles Lamb
Assignee: Charles Lamb
 Attachments: HDFS-6705.001.patch


 There needs to be an xattr that specifies that the HDFS admin can not access 
 a file. This is needed for m/r delegation tokens and data at rest encryption.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6705) Create an XAttr that disallows the HDFS admin from accessing a file

2014-08-28 Thread Charles Lamb (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6705?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113690#comment-14113690
 ] 

Charles Lamb commented on HDFS-6705:


bq. I think, restricting the file not accessible by the admin will create the 
problem for Fsck. Ideally a Super user means, will have access to everything 
right? Can you give more details about this requirement.?

Hi [~vinayrpet],

In general transparent encryption (HDFS-6134) is intended to ensure that 
encrypted files are not accessible to the superuser. For MR jobs that are 
accessing encrypted files, the DTs need to be made unreadable to the root. The 
intent is that DTs will have this xattr set for this purpose.



 Create an XAttr that disallows the HDFS admin from accessing a file
 ---

 Key: HDFS-6705
 URL: https://issues.apache.org/jira/browse/HDFS-6705
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode, security
Affects Versions: 3.0.0
Reporter: Charles Lamb
Assignee: Charles Lamb
 Attachments: HDFS-6705.001.patch


 There needs to be an xattr that specifies that the HDFS admin can not access 
 a file. This is needed for m/r delegation tokens and data at rest encryption.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6840) Clients are always sent to the same datanode when read is off rack

2014-08-28 Thread Yi Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113720#comment-14113720
 ] 

Yi Liu commented on HDFS-6840:
--

Thanks [~andrew.wang], the patch looks good to me.  Just one thing, we need 
also remove the configuration from hdfs-default.xml
{code}
property
  namedfs.namenode.randomize-block-locations-per-block/name
  valuefalse/value
  descriptionWhen fetching replica locations of a block, the replicas
   are sorted based on network distance. This configuration parameter
   determines whether the replicas at the same network distance are randomly
   shuffled. By default, this is false, such that repeated requests for a 
block's
   replicas always result in the same order. This potentially improves page 
cache
   behavior. However, for some network topologies, it is desirable to shuffle 
this
   order for better load balancing.
  /description
/property
{code}

 Clients are always sent to the same datanode when read is off rack
 --

 Key: HDFS-6840
 URL: https://issues.apache.org/jira/browse/HDFS-6840
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Jason Lowe
Assignee: Andrew Wang
Priority: Critical
 Attachments: hdfs-6840.001.patch


 After HDFS-6268 the sorting order of block locations is deterministic for a 
 given block and locality level (e.g.: local, rack. off-rack), so off-rack 
 clients all see the same datanode for the same block.  This leads to very 
 poor behavior in distributed cache localization and other scenarios where 
 many clients all want the same block data at approximately the same time.  
 The one datanode is crushed by the load while the other replicas only handle 
 local and rack-local requests.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6658) Namenode memory optimization - Block replicas list

2014-08-28 Thread Amir Langer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6658?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Amir Langer updated HDFS-6658:
--

Attachment: BlockListOptimizationComparison.xlsx

Added comparison of memory with and without CompressedOops in both the original 
and modified code.
(Memory in bytes collected using jmap).
Difference with compressed oops is marginal but significant without it.



 Namenode memory optimization - Block replicas list 
 ---

 Key: HDFS-6658
 URL: https://issues.apache.org/jira/browse/HDFS-6658
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Affects Versions: 2.4.1
Reporter: Amir Langer
Assignee: Amir Langer
 Attachments: BlockListOptimizationComparison.xlsx, Namenode Memory 
 Optimizations - Block replicas list.docx


 Part of the memory consumed by every BlockInfo object in the Namenode is a 
 linked list of block references for every DatanodeStorageInfo (called 
 triplets). 
 We propose to change the way we store the list in memory. 
 Using primitive integer indexes instead of object references will reduce the 
 memory needed for every block replica (when compressed oops is disabled) and 
 in our new design the list overhead will be per DatanodeStorageInfo and not 
 per block replica.
 see attached design doc. for details and evaluation results.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6843) Create FSDataInputStream FSDataOutputStream isEncrypted() method

2014-08-28 Thread Charles Lamb (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6843?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113765#comment-14113765
 ] 

Charles Lamb commented on HDFS-6843:


The test failure appears to be unrelated. I ran the test on my local machine 
and it passed.


 Create FSDataInputStream  FSDataOutputStream isEncrypted() method
 --

 Key: HDFS-6843
 URL: https://issues.apache.org/jira/browse/HDFS-6843
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode, security
Affects Versions: 3.0.0
Reporter: Charles Lamb
Assignee: Charles Lamb
 Attachments: HDFS-6843.001.patch, HDFS-6843.002.patch


 FileStatus should have a 'boolean isEncrypted()' method. (it was in the 
 context of discussing with AndreW about FileStatus being a Writable).
 Having this method would allow MR JobSubmitter do the following:
 -
 BOOLEAN intermediateEncryption = false
 IF jobconf.contains(mr.intermidate.encryption) THEN
   intermediateEncryption = jobConf.getBoolean(mr.intermidate.encryption)
 ELSE
   IF (I/O)Format INSTANCEOF File(I/O)Format THEN
 intermediateEncryption = ANY File(I/O)Format HAS a Path with status 
 isEncrypted()==TRUE
   FI
   jobConf.setBoolean(mr.intermidate.encryption, intermediateEncryption)
 FI



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6851) Flush EncryptionZoneWithId and add an id field to EncryptionZone

2014-08-28 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6851?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb updated HDFS-6851:
---

Status: Patch Available  (was: Open)

 Flush EncryptionZoneWithId and add an id field to EncryptionZone
 

 Key: HDFS-6851
 URL: https://issues.apache.org/jira/browse/HDFS-6851
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode, security
Affects Versions: 3.0.0
Reporter: Charles Lamb
Assignee: Charles Lamb
 Attachments: HDFS-6851.000.patch


 EncryptionZoneWithId can be flushed by moving the id field up to 
 EncryptionZone.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6851) Flush EncryptionZoneWithId and add an id field to EncryptionZone

2014-08-28 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6851?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb updated HDFS-6851:
---

Status: Open  (was: Patch Available)

Jenkins not running. Try to kick it in the head again.

 Flush EncryptionZoneWithId and add an id field to EncryptionZone
 

 Key: HDFS-6851
 URL: https://issues.apache.org/jira/browse/HDFS-6851
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode, security
Affects Versions: 3.0.0
Reporter: Charles Lamb
Assignee: Charles Lamb
 Attachments: HDFS-6851.000.patch


 EncryptionZoneWithId can be flushed by moving the id field up to 
 EncryptionZone.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6658) Namenode memory optimization - Block replicas list

2014-08-28 Thread Edward Bortnikov (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6658?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113773#comment-14113773
 ] 

Edward Bortnikov commented on HDFS-6658:


Looks convincing Amir! Seems that everybody is +1 for this short-term 
improvement. Can we move on towards committing this code, while proceeding with 
the long-term discussion on HDFS-5477 and HDFS-6709? 

 Namenode memory optimization - Block replicas list 
 ---

 Key: HDFS-6658
 URL: https://issues.apache.org/jira/browse/HDFS-6658
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Affects Versions: 2.4.1
Reporter: Amir Langer
Assignee: Amir Langer
 Attachments: BlockListOptimizationComparison.xlsx, Namenode Memory 
 Optimizations - Block replicas list.docx


 Part of the memory consumed by every BlockInfo object in the Namenode is a 
 linked list of block references for every DatanodeStorageInfo (called 
 triplets). 
 We propose to change the way we store the list in memory. 
 Using primitive integer indexes instead of object references will reduce the 
 memory needed for every block replica (when compressed oops is disabled) and 
 in our new design the list overhead will be per DatanodeStorageInfo and not 
 per block replica.
 see attached design doc. for details and evaluation results.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6888) Remove audit logging of getFIleInfo()

2014-08-28 Thread Chen He (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6888?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chen He updated HDFS-6888:
--

Attachment: HDFS-6888-3.patch

patch updated.

 Remove audit logging of getFIleInfo()
 -

 Key: HDFS-6888
 URL: https://issues.apache.org/jira/browse/HDFS-6888
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Kihwal Lee
Assignee: Chen He
  Labels: log
 Attachments: HDFS-6888-2.patch, HDFS-6888-3.patch, HDFS-6888.patch


 The audit logging of getFileInfo() was added in HDFS-3733.  Since this is a 
 one of the most called method, users have noticed that audit log is now 
 filled with this.  Since we now have HTTP request logging, this seems 
 unnecessary.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6888) Remove audit logging of getFIleInfo()

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113786#comment-14113786
 ] 

Hadoop QA commented on HDFS-6888:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7820//console

This message is automatically generated.

 Remove audit logging of getFIleInfo()
 -

 Key: HDFS-6888
 URL: https://issues.apache.org/jira/browse/HDFS-6888
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Kihwal Lee
Assignee: Chen He
  Labels: log
 Attachments: HDFS-6888-2.patch, HDFS-6888-3.patch, HDFS-6888.patch


 The audit logging of getFileInfo() was added in HDFS-3733.  Since this is a 
 one of the most called method, users have noticed that audit log is now 
 filled with this.  Since we now have HTTP request logging, this seems 
 unnecessary.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6888) Remove audit logging of getFIleInfo()

2014-08-28 Thread Chen He (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113795#comment-14113795
 ] 

Chen He commented on HDFS-6888:
---

It works on my mac. Is this because Hadoop repository move to git?

 Remove audit logging of getFIleInfo()
 -

 Key: HDFS-6888
 URL: https://issues.apache.org/jira/browse/HDFS-6888
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Kihwal Lee
Assignee: Chen He
  Labels: log
 Attachments: HDFS-6888-2.patch, HDFS-6888-3.patch, HDFS-6888.patch


 The audit logging of getFileInfo() was added in HDFS-3733.  Since this is a 
 one of the most called method, users have noticed that audit log is now 
 filled with this.  Since we now have HTTP request logging, this seems 
 unnecessary.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6888) Remove audit logging of getFIleInfo()

2014-08-28 Thread Chen He (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113797#comment-14113797
 ] 

Chen He commented on HDFS-6888:
---

HEAD is now at c4c9a78 HADOOP-11001. Fix test-patch to work with the git repo. 
(kasha)
Removing patchprocess/
Switched to branch 'trunk'
Your branch is up-to-date with 'origin/trunk'.
Current branch trunk is up to date.
/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS-Build/patchprocess/jira: 
No such file or directory
/bin/grep: 
/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS-Build/patchprocess/jira: 
No such file or directory
/bin/grep: 
/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS-Build/patchprocess/jira: 
No such file or directory
HDFS-6888 patch is being downloaded at Thu Aug 28 14:11:35 UTC 2014 from
http://issues.apache.org
/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS-Build/patchprocess/patch: 
No such file or directory
cp: cannot stat '/home/jenkins/buildSupport/lib/*': No such file or directory
/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS-Build/dev-support/smart-apply-patch.sh:
 line 44: 
/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS-Build/patchprocess/patch: 
No such file or directory
/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS-Build/dev-support/smart-apply-patch.sh:
 line 100: 
/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS-Build/patchprocess/patch: 
No such file or directory
/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS-Build/dev-support/smart-apply-patch.sh:
 line 102: 
/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS-Build/patchprocess/patch: 
No such file or directory
The patch does not appear to apply with p0 to p2
rm: cannot remove '/tmp/tmp.paths.7370': No such file or directory
PATCH APPLICATION FAILED

 Remove audit logging of getFIleInfo()
 -

 Key: HDFS-6888
 URL: https://issues.apache.org/jira/browse/HDFS-6888
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Kihwal Lee
Assignee: Chen He
  Labels: log
 Attachments: HDFS-6888-2.patch, HDFS-6888-3.patch, HDFS-6888.patch


 The audit logging of getFileInfo() was added in HDFS-3733.  Since this is a 
 one of the most called method, users have noticed that audit log is now 
 filled with this.  Since we now have HTTP request logging, this seems 
 unnecessary.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Work started] (HDFS-6954) With crypto, no native lib systems are too verbose

2014-08-28 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HDFS-6954 started by Charles Lamb.

 With crypto, no native lib systems are too verbose
 --

 Key: HDFS-6954
 URL: https://issues.apache.org/jira/browse/HDFS-6954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: encryption
Affects Versions: 3.0.0
Reporter: Allen Wittenauer
Assignee: Charles Lamb
 Attachments: HDFS-6954.001.patch


 Running commands on a machine without a native library results in:
 {code}
 $ bin/hdfs dfs -put /etc/hosts /tmp
 14/08/27 07:16:10 WARN util.NativeCodeLoader: Unable to load native-hadoop 
 library for your platform... using builtin-java classes where applicable
 14/08/27 07:16:11 WARN crypto.CryptoCodec: Crypto codec 
 org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec is not available.
 14/08/27 07:16:11 INFO hdfs.DFSClient: No KeyProvider found.
 {code}
 This is way too much.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6954) With crypto, no native lib systems are too verbose

2014-08-28 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb updated HDFS-6954:
---

 Target Version/s: 3.0.0, 2.6.0
Affects Version/s: 3.0.0
   Status: Patch Available  (was: In Progress)

 With crypto, no native lib systems are too verbose
 --

 Key: HDFS-6954
 URL: https://issues.apache.org/jira/browse/HDFS-6954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: encryption
Affects Versions: 3.0.0
Reporter: Allen Wittenauer
Assignee: Charles Lamb
 Attachments: HDFS-6954.001.patch


 Running commands on a machine without a native library results in:
 {code}
 $ bin/hdfs dfs -put /etc/hosts /tmp
 14/08/27 07:16:10 WARN util.NativeCodeLoader: Unable to load native-hadoop 
 library for your platform... using builtin-java classes where applicable
 14/08/27 07:16:11 WARN crypto.CryptoCodec: Crypto codec 
 org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec is not available.
 14/08/27 07:16:11 INFO hdfs.DFSClient: No KeyProvider found.
 {code}
 This is way too much.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6954) With crypto, no native lib systems are too verbose

2014-08-28 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb updated HDFS-6954:
---

Attachment: HDFS-6954.001.patch

Attached is a patch which changes the CryptoCodec and DFSClient messages from 
warn to debug. I considered changing the former to use the new 
PerformanceAdvisory log sink, but that still wouldn't suppress the messages per 
Allen's comment.

wrt distcp, I did not change the tools.DistCp or client.RMProxy messages as (I 
believe) they were not changed by the fs-encryption checkin. I can easily 
change those to debug also if that's desirable.

 With crypto, no native lib systems are too verbose
 --

 Key: HDFS-6954
 URL: https://issues.apache.org/jira/browse/HDFS-6954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: encryption
Affects Versions: 3.0.0
Reporter: Allen Wittenauer
Assignee: Charles Lamb
 Attachments: HDFS-6954.001.patch


 Running commands on a machine without a native library results in:
 {code}
 $ bin/hdfs dfs -put /etc/hosts /tmp
 14/08/27 07:16:10 WARN util.NativeCodeLoader: Unable to load native-hadoop 
 library for your platform... using builtin-java classes where applicable
 14/08/27 07:16:11 WARN crypto.CryptoCodec: Crypto codec 
 org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec is not available.
 14/08/27 07:16:11 INFO hdfs.DFSClient: No KeyProvider found.
 {code}
 This is way too much.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6954) With crypto, no native lib systems are too verbose

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113857#comment-14113857
 ] 

Hadoop QA commented on HDFS-6954:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7821//console

This message is automatically generated.

 With crypto, no native lib systems are too verbose
 --

 Key: HDFS-6954
 URL: https://issues.apache.org/jira/browse/HDFS-6954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: encryption
Affects Versions: 3.0.0
Reporter: Allen Wittenauer
Assignee: Charles Lamb
 Attachments: HDFS-6954.001.patch


 Running commands on a machine without a native library results in:
 {code}
 $ bin/hdfs dfs -put /etc/hosts /tmp
 14/08/27 07:16:10 WARN util.NativeCodeLoader: Unable to load native-hadoop 
 library for your platform... using builtin-java classes where applicable
 14/08/27 07:16:11 WARN crypto.CryptoCodec: Crypto codec 
 org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec is not available.
 14/08/27 07:16:11 INFO hdfs.DFSClient: No KeyProvider found.
 {code}
 This is way too much.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6694) TestPipelinesFailover.testPipelineRecoveryStress tests fail intermittently with various symptoms

2014-08-28 Thread Yongjun Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6694?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113870#comment-14113870
 ] 

Yongjun Zhang commented on HDFS-6694:
-

HI [~gkesavan] and [~ipv6guru],

Slave {{H9 (Build slave for Hadoop project builds : asf909.gq1.ygridcore.net)}} 
still has the problem, it's num open files is 1024. Would you please help to  
look into? Arpit suggested rebooting the host might solve the problem, which 
sounds good to me. 

Thanks a lot.


 TestPipelinesFailover.testPipelineRecoveryStress tests fail intermittently 
 with various symptoms
 

 Key: HDFS-6694
 URL: https://issues.apache.org/jira/browse/HDFS-6694
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 3.0.0
Reporter: Yongjun Zhang
Assignee: Yongjun Zhang
Priority: Critical
 Fix For: 2.6.0

 Attachments: HDFS-6694.001.dbg.patch, HDFS-6694.001.dbg.patch, 
 HDFS-6694.001.dbg.patch, HDFS-6694.002.dbg.patch, 
 org.apache.hadoop.hdfs.server.namenode.ha.TestPipelinesFailover-output.txt, 
 org.apache.hadoop.hdfs.server.namenode.ha.TestPipelinesFailover.txt


 TestPipelinesFailover.testPipelineRecoveryStress tests fail intermittently 
 with various symptoms. Typical failures are described in first comment.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6954) With crypto, no native lib systems are too verbose

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113874#comment-14113874
 ] 

Hadoop QA commented on HDFS-6954:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7824//console

This message is automatically generated.

 With crypto, no native lib systems are too verbose
 --

 Key: HDFS-6954
 URL: https://issues.apache.org/jira/browse/HDFS-6954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: encryption
Affects Versions: 3.0.0
Reporter: Allen Wittenauer
Assignee: Charles Lamb
 Attachments: HDFS-6954.001.patch


 Running commands on a machine without a native library results in:
 {code}
 $ bin/hdfs dfs -put /etc/hosts /tmp
 14/08/27 07:16:10 WARN util.NativeCodeLoader: Unable to load native-hadoop 
 library for your platform... using builtin-java classes where applicable
 14/08/27 07:16:11 WARN crypto.CryptoCodec: Crypto codec 
 org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec is not available.
 14/08/27 07:16:11 INFO hdfs.DFSClient: No KeyProvider found.
 {code}
 This is way too much.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6954) With crypto, no native lib systems are too verbose

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113882#comment-14113882
 ] 

Hadoop QA commented on HDFS-6954:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7825//console

This message is automatically generated.

 With crypto, no native lib systems are too verbose
 --

 Key: HDFS-6954
 URL: https://issues.apache.org/jira/browse/HDFS-6954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: encryption
Affects Versions: 3.0.0
Reporter: Allen Wittenauer
Assignee: Charles Lamb
 Attachments: HDFS-6954.001.patch


 Running commands on a machine without a native library results in:
 {code}
 $ bin/hdfs dfs -put /etc/hosts /tmp
 14/08/27 07:16:10 WARN util.NativeCodeLoader: Unable to load native-hadoop 
 library for your platform... using builtin-java classes where applicable
 14/08/27 07:16:11 WARN crypto.CryptoCodec: Crypto codec 
 org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec is not available.
 14/08/27 07:16:11 INFO hdfs.DFSClient: No KeyProvider found.
 {code}
 This is way too much.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-695) RaidNode should read in configuration from hdfs-site.xml

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113887#comment-14113887
 ] 

Hadoop QA commented on HDFS-695:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7826//console

This message is automatically generated.

 RaidNode should read in configuration from hdfs-site.xml
 

 Key: HDFS-695
 URL: https://issues.apache.org/jira/browse/HDFS-695
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: contrib/raid
Reporter: dhruba borthakur
Assignee: dhruba borthakur
 Fix For: 0.21.0

 Attachments: raidConfig.txt


 The RaidNode currently reads in the configuration from hadoop-*.xml. It 
 should read in its config parameters from hdfs-site.xml as well.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6954) With crypto, no native lib systems are too verbose

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113891#comment-14113891
 ] 

Hadoop QA commented on HDFS-6954:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7827//console

This message is automatically generated.

 With crypto, no native lib systems are too verbose
 --

 Key: HDFS-6954
 URL: https://issues.apache.org/jira/browse/HDFS-6954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: encryption
Affects Versions: 3.0.0
Reporter: Allen Wittenauer
Assignee: Charles Lamb
 Attachments: HDFS-6954.001.patch


 Running commands on a machine without a native library results in:
 {code}
 $ bin/hdfs dfs -put /etc/hosts /tmp
 14/08/27 07:16:10 WARN util.NativeCodeLoader: Unable to load native-hadoop 
 library for your platform... using builtin-java classes where applicable
 14/08/27 07:16:11 WARN crypto.CryptoCodec: Crypto codec 
 org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec is not available.
 14/08/27 07:16:11 INFO hdfs.DFSClient: No KeyProvider found.
 {code}
 This is way too much.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6888) Remove audit logging of getFIleInfo()

2014-08-28 Thread Chen He (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6888?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chen He updated HDFS-6888:
--

Attachment: HDFS-6888-3.patch

retrigger QA

 Remove audit logging of getFIleInfo()
 -

 Key: HDFS-6888
 URL: https://issues.apache.org/jira/browse/HDFS-6888
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Kihwal Lee
Assignee: Chen He
  Labels: log
 Attachments: HDFS-6888-2.patch, HDFS-6888-3.patch, HDFS-6888-3.patch, 
 HDFS-6888.patch


 The audit logging of getFileInfo() was added in HDFS-3733.  Since this is a 
 one of the most called method, users have noticed that audit log is now 
 filled with this.  Since we now have HTTP request logging, this seems 
 unnecessary.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-3529) Use direct buffers for data in write path

2014-08-28 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113917#comment-14113917
 ] 

Todd Lipcon commented on HDFS-3529:
---

Hey Yi. Please note that [~james.thomas] has been doing work recently so that 
native CRC32 can run against byte[] arrays instead of just against byte 
buffers. So, the major performance improvement enabled by this JIRA probably 
isn't as relevant anymore.

Using direct buffers might save us a memcpy, but for small cache-resident 
buffer sizes, my guess is that we won't see a big boost. I'd be interested to 
see some results, though, if you have a prototype of the patch against trunk.

 Use direct buffers for data in write path
 -

 Key: HDFS-3529
 URL: https://issues.apache.org/jira/browse/HDFS-3529
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode, performance
Affects Versions: 2.0.0-alpha
Reporter: Todd Lipcon
Assignee: Trevor Robinson
 Attachments: HDFS-3529.patch, dfsio-x86-trunk-vs-3529.png


 The write path currently makes several unnecessary data copies in order to go 
 to and from byte arrays. We can improve performance by using direct byte 
 buffers to avoid the copy. This is also a prerequisite for native checksum 
 calculation (HDFS-3528)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6962) ACLs inheritance conflict with umaskmode

2014-08-28 Thread LINTE (JIRA)
LINTE created HDFS-6962:
---

 Summary: ACLs inheritance conflict with umaskmode
 Key: HDFS-6962
 URL: https://issues.apache.org/jira/browse/HDFS-6962
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: security
Affects Versions: 2.4.1
 Environment: CentOS release 6.5 (Final)
Reporter: LINTE


In hdfs-site.xml 
property
namedfs.umaskmode/name
value027/value
/property

1/ Create a directory as superuser
bash# hdfs dfs -mkdir  /tmp/ACLS

2/ set default ACLs on this directory rwx access for group readwrite and user 
toto
bash# hdfs dfs -setfacl -m default:group:readwrite:rwx /tmp/ACLS
bash# hdfs dfs -setfacl -m default:user:toto:rwx /tmp/ACLS

3/ check ACLs /tmp/ACLS/
bash# hdfs dfs -getfacl /tmp/ACLS/
# file: /tmp/ACLS
# owner: hdfs
# group: hadoop
user::rwx
group::r-x
other::---
default:user::rwx
default:user:toto:rwx
default:group::r-x
default:group:readwrite:rwx
default:mask::rwx
default:other::---

user::rwx | group::r-x | other::--- matches with the umaskmode defined in 
hdfs-site.xml, everything ok !

default:group:readwrite:rwx allow readwrite group with rwx access for 
inhéritance.
default:user:toto:rwx allow toto user with rwx access for inhéritance.

default:mask::rwx inhéritance mask is rwx, so no mask

4/ Create a subdir to test inheritance of ACL
bash# hdfs dfs -mkdir  /tmp/ACLS/hdfs

5/ check ACLs /tmp/ACLS/hdfs
bash# hdfs dfs -getfacl /tmp/ACLS/hdfs
# file: /tmp/ACLS/hdfs
# owner: hdfs
# group: hadoop
user::rwx
user:toto:rwx   #effective:r-x
group::r-x
group:readwrite:rwx #effective:r-x
mask::r-x
other::---
default:user::rwx
default:user:toto:rwx
default:group::r-x
default:group:readwrite:rwx
default:mask::rwx
default:other::---

Here we can see that the readwrite group has rwx ACL bu only r-x is effective 
because the mask is r-x (mask::r-x) in spite of default mask for inheritance is 
set to default:mask::rwx on /tmp/ACLS/

6/ Modifiy hdfs-site.xml et restart namenode
property
namedfs.umaskmode/name
value010/value
/property

7/ Create a subdir to test inheritance of ACL with new parameter umaskmode
bash# hdfs dfs -mkdir  /tmp/ACLS/hdfs2

8/ Check ACL on /tmp/ACLS/hdfs2
bash# hdfs dfs -getfacl /tmp/ACLS/hdfs2
# file: /tmp/ACLS/hdfs2
# owner: hdfs
# group: hadoop
user::rwx
user:toto:rwx   #effective:rw-
group::r-x  #effective:r--
group:readwrite:rwx #effective:rw-
mask::rw-
other::---
default:user::rwx
default:user:toto:rwx
default:group::r-x
default:group:readwrite:rwx
default:mask::rwx
default:other::---

So HDFS masks the ACL value (user, group and other  -- exepted the POSIX owner 
-- ) with the group mask of dfs.umaskmode properties when creating directory 
with inherited ACL.






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6867) For DFSOutputStream, do pipeline recovery for a single block in the background

2014-08-28 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HDFS-6867:


Attachment: HDFS-6867-20140828-1.patch

Some test classes have RepliceNodeOnFailure disabled. This new patches skips 
recoveryWorker invocation if that's the case.

 For DFSOutputStream, do pipeline recovery for a single block in the background
 --

 Key: HDFS-6867
 URL: https://issues.apache.org/jira/browse/HDFS-6867
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Reporter: Colin Patrick McCabe
Assignee: Zhe Zhang
 Attachments: HDFS-6867-20140827-2.patch, HDFS-6867-20140827-3.patch, 
 HDFS-6867-20140827.patch, HDFS-6867-20140828-1.patch, 
 HDFS-6867-design-20140820.pdf, HDFS-6867-design-20140821.pdf, 
 HDFS-6867-design-20140822.pdf, HDFS-6867-design-20140827.pdf


 For DFSOutputStream, we should be able to do pipeline recovery in the 
 background, while the user is continuing to write to the file.  This is 
 especially useful for long-lived clients that write to an HDFS file slowly. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6958) Fix the condition of reporting failures in BlockPoolSlice

2014-08-28 Thread Lei (Eddy) Xu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6958?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14113961#comment-14113961
 ] 

Lei (Eddy) Xu commented on HDFS-6958:
-

[~atm] These two failures are all related to balancer, which is not related to 
this patch.

 Fix the condition of reporting failures in BlockPoolSlice
 -

 Key: HDFS-6958
 URL: https://issues.apache.org/jira/browse/HDFS-6958
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
Affects Versions: 2.5.0
Reporter: Lei (Eddy) Xu
Assignee: Lei (Eddy) Xu
Priority: Trivial
 Attachments: HDFS-6958.000.patch


 {{BlockPoolSlice#addToReplicsMap()}} should log the failure for the operation 
 that deletes restart metadata file. 
 The current code did the opposite, log the failures on succeed deletion:
 {code:java}
 if (restartMeta.delete()) {
 FsDatasetImpl.LOG.warn(Failed to delete restart meta file:  +
   restartMeta.getPath());
 }
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6963) Test partial failures when loading / removing volumes.

2014-08-28 Thread Lei (Eddy) Xu (JIRA)
Lei (Eddy) Xu created HDFS-6963:
---

 Summary: Test partial failures when loading / removing volumes.
 Key: HDFS-6963
 URL: https://issues.apache.org/jira/browse/HDFS-6963
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode
Affects Versions: 2.5.0
Reporter: Lei (Eddy) Xu
Assignee: Lei (Eddy) Xu


Test the cases that there are partial failures during loading new volumes. The 
expected behavior should be trying the best efforts to successfully load all 
good volumes, and then report the failures.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6964) NN fails to fix under replication leading to data loss

2014-08-28 Thread Daryn Sharp (JIRA)
Daryn Sharp created HDFS-6964:
-

 Summary: NN fails to fix under replication leading to data loss
 Key: HDFS-6964
 URL: https://issues.apache.org/jira/browse/HDFS-6964
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.0.0-alpha, 3.0.0
Reporter: Daryn Sharp
Priority: Blocker


We've encountered lost blocks due to node failure even when there is ample time 
to fix the under-replication.

2 nodes were lost.  The 3rd node with the last remaining replicas averaged 1 
copy block per heartbeat (3s) until ~7h later when that node was lost resulting 
in over 50 lost blocks.  When the node was restarted and sent its BR the NN 
immediately began fixing the replication.

In another data loss event, over 150 blocks were lost due to node failure but 
the timing of the node loss is not known so there may have been inadequate time 
to fix the under-replication unlike the first case.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6965) NN continues to issue block locations for DNs with full disks

2014-08-28 Thread Daryn Sharp (JIRA)
Daryn Sharp created HDFS-6965:
-

 Summary: NN continues to issue block locations for DNs with full 
disks
 Key: HDFS-6965
 URL: https://issues.apache.org/jira/browse/HDFS-6965
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.0.0-alpha, 3.0.0
Reporter: Daryn Sharp


Encountered issues where DNs have less space than a full block and reject 
incoming transfers.  The NN continues giving out locations for these nodes for 
some period of time.  It does not appear to be related to the DN's cached disk 
usage.

One impact is required replications are delayed when a full DN is chosen for 
the pipeline.  A DN cannot report a broken pipeline so the replication must 
timeout (5m) before new targets are chosen.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6965) NN continues to issue block locations for DNs with full disks

2014-08-28 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114065#comment-14114065
 ] 

Kihwal Lee commented on HDFS-6965:
--

I think there is a flaw in isGoodTarget().  When checking the available space, 
it checks against {{node.getRemaining()}}. It has too check at individual 
storage level.

 NN continues to issue block locations for DNs with full disks
 -

 Key: HDFS-6965
 URL: https://issues.apache.org/jira/browse/HDFS-6965
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.0.0-alpha, 3.0.0
Reporter: Daryn Sharp

 Encountered issues where DNs have less space than a full block and reject 
 incoming transfers.  The NN continues giving out locations for these nodes 
 for some period of time.  It does not appear to be related to the DN's cached 
 disk usage.
 One impact is required replications are delayed when a full DN is chosen for 
 the pipeline.  A DN cannot report a broken pipeline so the replication must 
 timeout (5m) before new targets are chosen.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6843) Create FSDataInputStream FSDataOutputStream isEncrypted() method

2014-08-28 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6843?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114075#comment-14114075
 ] 

Steve Loughran commented on HDFS-6843:
--

As this is extending the core FS APIs

# Can you update the specification of those FS APIs to add a new attribute to a 
file status
# the contract tests need to be extended too, we just need to verify that 
isEncrypted returns something ... that could be integrated with one of the 
exiting tests

This may need to be mentioned in a HADOOP- JIRA even if the patch is in here 
... this is broader than just HDFS


 Create FSDataInputStream  FSDataOutputStream isEncrypted() method
 --

 Key: HDFS-6843
 URL: https://issues.apache.org/jira/browse/HDFS-6843
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode, security
Affects Versions: 3.0.0
Reporter: Charles Lamb
Assignee: Charles Lamb
 Attachments: HDFS-6843.001.patch, HDFS-6843.002.patch


 FileStatus should have a 'boolean isEncrypted()' method. (it was in the 
 context of discussing with AndreW about FileStatus being a Writable).
 Having this method would allow MR JobSubmitter do the following:
 -
 BOOLEAN intermediateEncryption = false
 IF jobconf.contains(mr.intermidate.encryption) THEN
   intermediateEncryption = jobConf.getBoolean(mr.intermidate.encryption)
 ELSE
   IF (I/O)Format INSTANCEOF File(I/O)Format THEN
 intermediateEncryption = ANY File(I/O)Format HAS a Path with status 
 isEncrypted()==TRUE
   FI
   jobConf.setBoolean(mr.intermidate.encryption, intermediateEncryption)
 FI



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6966) Add additional unit tests for encryption zones

2014-08-28 Thread Stephen Chu (JIRA)
Stephen Chu created HDFS-6966:
-

 Summary: Add additional unit tests for encryption zones
 Key: HDFS-6966
 URL: https://issues.apache.org/jira/browse/HDFS-6966
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: encryption
Affects Versions: 3.0.0
Reporter: Stephen Chu
Assignee: Stephen Chu


There are some more unit tests that can be added for test encryption zones. For 
example, more encryption zone + snapshot tests, running fsck on encryption 
zones, and more.





--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6946) TestBalancerWithSaslDataTransfer fails in trunk

2014-08-28 Thread Stephen Chu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6946?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114095#comment-14114095
 ] 

Stephen Chu commented on HDFS-6946:
---

[~airbots], [~kihwal], any thoughts on this one? Checking because you worked on 
HDFS-5803.

 TestBalancerWithSaslDataTransfer fails in trunk
 ---

 Key: HDFS-6946
 URL: https://issues.apache.org/jira/browse/HDFS-6946
 Project: Hadoop HDFS
  Issue Type: Test
Reporter: Ted Yu
Assignee: Stephen Chu
Priority: Minor
 Attachments: HDFS-6946.1.patch


 From build #1849 :
 {code}
 REGRESSION:  
 org.apache.hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer.testBalancer0Integrity
 Error Message:
 Cluster failed to reached expected values of totalSpace (current: 750, 
 expected: 750), or usedSpace (current: 140, expected: 150), in more than 
 4 msec.
 Stack Trace:
 java.util.concurrent.TimeoutException: Cluster failed to reached expected 
 values of totalSpace (current: 750, expected: 750), or usedSpace (current: 
 140, expected: 150), in more than 4 msec.
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.waitForHeartBeat(TestBalancer.java:253)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.runBalancer(TestBalancer.java:578)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.doTest(TestBalancer.java:551)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.doTest(TestBalancer.java:437)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.oneNodeTest(TestBalancer.java:645)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.testBalancer0Internal(TestBalancer.java:759)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer.testBalancer0Integrity(TestBalancerWithSaslDataTransfer.java:34)
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6967) DNs may OOM under high webhdfs load

2014-08-28 Thread Daryn Sharp (JIRA)
Daryn Sharp created HDFS-6967:
-

 Summary: DNs may OOM under high webhdfs load
 Key: HDFS-6967
 URL: https://issues.apache.org/jira/browse/HDFS-6967
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode, webhdfs
Affects Versions: 2.0.0-alpha, 3.0.0
Reporter: Daryn Sharp


Webhdfs uses jetty.  The size of the request thread pool is limited, but jetty 
will accept and queue infinite connections.  Every queued connection is heavy 
with buffers, etc.  Unlike data streamer connections, thousands of webhdfs 
connections will quickly OOM a DN.  The accepted requests must be bounded and 
excess clients rejected so they retry on a new DN.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6888) Remove audit logging of getFIleInfo()

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114141#comment-14114141
 ] 

Hadoop QA commented on HDFS-6888:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12664998/HDFS-6888-3.patch
  against trunk revision c4c9a78.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 3 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  org.apache.hadoop.security.TestRefreshUserMappings

  The following test timeouts occurred in 
hadoop-hdfs-project/hadoop-hdfs:

org.apache.hadoop.hdfs.TestDecommission

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7829//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7829//console

This message is automatically generated.

 Remove audit logging of getFIleInfo()
 -

 Key: HDFS-6888
 URL: https://issues.apache.org/jira/browse/HDFS-6888
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Kihwal Lee
Assignee: Chen He
  Labels: log
 Attachments: HDFS-6888-2.patch, HDFS-6888-3.patch, HDFS-6888-3.patch, 
 HDFS-6888.patch


 The audit logging of getFileInfo() was added in HDFS-3733.  Since this is a 
 one of the most called method, users have noticed that audit log is now 
 filled with this.  Since we now have HTTP request logging, this seems 
 unnecessary.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6826) Plugin interface to enable delegation of HDFS authorization assertions

2014-08-28 Thread Jitendra Nath Pandey (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6826?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114137#comment-14114137
 ] 

Jitendra Nath Pandey commented on HDFS-6826:


I do prefer v7.4 approach. I believe plugins should have the flexibility to 
exclusively control the permission check.

 Plugin interface to enable delegation of HDFS authorization assertions
 --

 Key: HDFS-6826
 URL: https://issues.apache.org/jira/browse/HDFS-6826
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: security
Affects Versions: 2.4.1
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Attachments: HDFS-6826-idea.patch, HDFS-6826-idea2.patch, 
 HDFS-6826v3.patch, HDFS-6826v4.patch, HDFS-6826v5.patch, HDFS-6826v6.patch, 
 HDFS-6826v7.1.patch, HDFS-6826v7.2.patch, HDFS-6826v7.3.patch, 
 HDFS-6826v7.4.patch, HDFS-6826v7.5.patch, HDFS-6826v7.patch, 
 HDFS-6826v8.patch, HDFSPluggableAuthorizationProposal-v2.pdf, 
 HDFSPluggableAuthorizationProposal.pdf


 When Hbase data, HiveMetaStore data or Search data is accessed via services 
 (Hbase region servers, HiveServer2, Impala, Solr) the services can enforce 
 permissions on corresponding entities (databases, tables, views, columns, 
 search collections, documents). It is desirable, when the data is accessed 
 directly by users accessing the underlying data files (i.e. from a MapReduce 
 job), that the permission of the data files map to the permissions of the 
 corresponding data entity (i.e. table, column family or search collection).
 To enable this we need to have the necessary hooks in place in the NameNode 
 to delegate authorization to an external system that can map HDFS 
 files/directories to data entities and resolve their permissions based on the 
 data entities permissions.
 I’ll be posting a design proposal in the next few days.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6954) With crypto, no native lib systems are too verbose

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114140#comment-14114140
 ] 

Hadoop QA commented on HDFS-6954:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12664990/HDFS-6954.001.patch
  against trunk revision c4c9a78.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

  {color:red}-1 javac{color}.  The applied patch generated 1263 javac 
compiler warnings (more than the trunk's current 802 warnings).

{color:red}-1 javadoc{color}.  The javadoc tool appears to have generated 9 
warning messages.
See 
https://builds.apache.org/job/PreCommit-HDFS-Build/7828//artifact/trunk/patchprocess/diffJavadocWarnings.txt
 for details.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  
org.apache.hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer
  org.apache.hadoop.hdfs.TestDatanodeBlockScanner
  org.apache.hadoop.hdfs.web.TestWebHdfsFileSystemContract
  
org.apache.hadoop.hdfs.server.namenode.ha.TestPipelinesFailover

  The following test timeouts occurred in 
hadoop-hdfs-project/hadoop-hdfs:

org.apache.hadoop.hdfs.TestDecommission

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7828//testReport/
Javac warnings: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7828//artifact/trunk/patchprocess/diffJavacWarnings.txt
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7828//console

This message is automatically generated.

 With crypto, no native lib systems are too verbose
 --

 Key: HDFS-6954
 URL: https://issues.apache.org/jira/browse/HDFS-6954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: encryption
Affects Versions: 3.0.0
Reporter: Allen Wittenauer
Assignee: Charles Lamb
 Attachments: HDFS-6954.001.patch


 Running commands on a machine without a native library results in:
 {code}
 $ bin/hdfs dfs -put /etc/hosts /tmp
 14/08/27 07:16:10 WARN util.NativeCodeLoader: Unable to load native-hadoop 
 library for your platform... using builtin-java classes where applicable
 14/08/27 07:16:11 WARN crypto.CryptoCodec: Crypto codec 
 org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec is not available.
 14/08/27 07:16:11 INFO hdfs.DFSClient: No KeyProvider found.
 {code}
 This is way too much.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-4257) The ReplaceDatanodeOnFailure policies could have a forgiving option

2014-08-28 Thread Yongjun Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4257?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114150#comment-14114150
 ] 

Yongjun Zhang commented on HDFS-4257:
-

HI Nicholas, I wonder if we could log a message to see what policy is used at 
initialization stage and whenever the policy is changed, so when we check the 
log, we can see what policy is being used? thanks.


 The ReplaceDatanodeOnFailure policies could have a forgiving option
 ---

 Key: HDFS-4257
 URL: https://issues.apache.org/jira/browse/HDFS-4257
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: hdfs-client
Affects Versions: 2.0.2-alpha
Reporter: Harsh J
Assignee: Tsz Wo Nicholas Sze
Priority: Minor
 Attachments: h4257_20140325.patch, h4257_20140325b.patch, 
 h4257_20140326.patch, h4257_20140819.patch


 Similar question has previously come over HDFS-3091 and friends, but the 
 essential problem is: Why can't I write to my cluster of 3 nodes, when I 
 just have 1 node available at a point in time..
 The policies cover the 4 options, with {{Default}} being default:
 {{Disable}} - Disables the whole replacement concept by throwing out an 
 error (at the server) or acts as {{Never}} at the client.
 {{Never}} - Never replaces a DN upon pipeline failures (not too desirable in 
 many cases).
 {{Default}} - Replace based on a few conditions, but whose minimum never 
 touches 1. We always fail if only one DN remains and none others can be added.
 {{Always}} - Replace no matter what. Fail if can't replace.
 Would it not make sense to have an option similar to Always/Default, where 
 despite _trying_, if it isn't possible to have  1 DN in the pipeline, do not 
 fail. I think that is what the former write behavior was, and what fit with 
 the minimum replication factor allowed value.
 Why is it grossly wrong to pass a write from a client for a block with just 1 
 remaining replica in the pipeline (the minimum of 1 grows with the 
 replication factor demanded from the write), when replication is taken care 
 of immediately afterwards? How often have we seen missing blocks arise out of 
 allowing this + facing a big rack(s) failure or so?



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6928) 'hdfs put' command should accept lazyPersist flag for testing

2014-08-28 Thread Xiaoyu Yao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6928?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114176#comment-14114176
 ] 

Xiaoyu Yao commented on HDFS-6928:
--

Looks good. 
Can you update the comment of with lazyPersist flag on the stat Format 
sequences in Stat.java?

 'hdfs put' command should accept lazyPersist flag for testing
 -

 Key: HDFS-6928
 URL: https://issues.apache.org/jira/browse/HDFS-6928
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode
Affects Versions: HDFS-6581
Reporter: Tassapol Athiapinya
Assignee: Arpit Agarwal
 Attachments: HDFS-6928.01.patch, HDFS-6928.02.patch


 Add a '-l' flag to 'hdfs put' which creates the file with the LAZY_PERSIST 
 option.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6843) Create FSDataInputStream FSDataOutputStream isEncrypted() method

2014-08-28 Thread jay vyas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6843?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114192#comment-14114192
 ] 

jay vyas commented on HDFS-6843:


as per [~ste...@apache.org]'s comment, definetly important to  put this 
specification in HADOOP-... so that the other HCFS implementations can move it 
into the contract,  so we can have unit tests for it, etc, using .

 Create FSDataInputStream  FSDataOutputStream isEncrypted() method
 --

 Key: HDFS-6843
 URL: https://issues.apache.org/jira/browse/HDFS-6843
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode, security
Affects Versions: 3.0.0
Reporter: Charles Lamb
Assignee: Charles Lamb
 Attachments: HDFS-6843.001.patch, HDFS-6843.002.patch


 FileStatus should have a 'boolean isEncrypted()' method. (it was in the 
 context of discussing with AndreW about FileStatus being a Writable).
 Having this method would allow MR JobSubmitter do the following:
 -
 BOOLEAN intermediateEncryption = false
 IF jobconf.contains(mr.intermidate.encryption) THEN
   intermediateEncryption = jobConf.getBoolean(mr.intermidate.encryption)
 ELSE
   IF (I/O)Format INSTANCEOF File(I/O)Format THEN
 intermediateEncryption = ANY File(I/O)Format HAS a Path with status 
 isEncrypted()==TRUE
   FI
   jobConf.setBoolean(mr.intermidate.encryption, intermediateEncryption)
 FI



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6946) TestBalancerWithSaslDataTransfer fails in trunk

2014-08-28 Thread Chen He (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6946?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114194#comment-14114194
 ] 

Chen He commented on HDFS-6946:
---

[~schu] can you provide the full log if this error? Balancer test fails because 
of many reasons and some errors happen intermittently. 

 TestBalancerWithSaslDataTransfer fails in trunk
 ---

 Key: HDFS-6946
 URL: https://issues.apache.org/jira/browse/HDFS-6946
 Project: Hadoop HDFS
  Issue Type: Test
Reporter: Ted Yu
Assignee: Stephen Chu
Priority: Minor
 Attachments: HDFS-6946.1.patch


 From build #1849 :
 {code}
 REGRESSION:  
 org.apache.hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer.testBalancer0Integrity
 Error Message:
 Cluster failed to reached expected values of totalSpace (current: 750, 
 expected: 750), or usedSpace (current: 140, expected: 150), in more than 
 4 msec.
 Stack Trace:
 java.util.concurrent.TimeoutException: Cluster failed to reached expected 
 values of totalSpace (current: 750, expected: 750), or usedSpace (current: 
 140, expected: 150), in more than 4 msec.
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.waitForHeartBeat(TestBalancer.java:253)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.runBalancer(TestBalancer.java:578)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.doTest(TestBalancer.java:551)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.doTest(TestBalancer.java:437)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.oneNodeTest(TestBalancer.java:645)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.testBalancer0Internal(TestBalancer.java:759)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer.testBalancer0Integrity(TestBalancerWithSaslDataTransfer.java:34)
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6967) DNs may OOM under high webhdfs load

2014-08-28 Thread Daryn Sharp (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6967?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daryn Sharp updated HDFS-6967:
--

Assignee: Eric Payne

 DNs may OOM under high webhdfs load
 ---

 Key: HDFS-6967
 URL: https://issues.apache.org/jira/browse/HDFS-6967
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode, webhdfs
Affects Versions: 2.0.0-alpha, 3.0.0
Reporter: Daryn Sharp
Assignee: Eric Payne

 Webhdfs uses jetty.  The size of the request thread pool is limited, but 
 jetty will accept and queue infinite connections.  Every queued connection is 
 heavy with buffers, etc.  Unlike data streamer connections, thousands of 
 webhdfs connections will quickly OOM a DN.  The accepted requests must be 
 bounded and excess clients rejected so they retry on a new DN.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6888) Remove audit logging of getFIleInfo()

2014-08-28 Thread Chen He (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114200#comment-14114200
 ] 

Chen He commented on HDFS-6888:
---

Test error is because of HADOOP-11020

 Remove audit logging of getFIleInfo()
 -

 Key: HDFS-6888
 URL: https://issues.apache.org/jira/browse/HDFS-6888
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Kihwal Lee
Assignee: Chen He
  Labels: log
 Attachments: HDFS-6888-2.patch, HDFS-6888-3.patch, HDFS-6888-3.patch, 
 HDFS-6888.patch


 The audit logging of getFileInfo() was added in HDFS-3733.  Since this is a 
 one of the most called method, users have noticed that audit log is now 
 filled with this.  Since we now have HTTP request logging, this seems 
 unnecessary.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6946) TestBalancerWithSaslDataTransfer fails in trunk

2014-08-28 Thread Stephen Chu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6946?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephen Chu updated HDFS-6946:
--

Attachment: testBalancer0Integrity-failure.log

Attaching testBalancer0Integrity failure log from jenkins report. Will go 
through them again too.

 TestBalancerWithSaslDataTransfer fails in trunk
 ---

 Key: HDFS-6946
 URL: https://issues.apache.org/jira/browse/HDFS-6946
 Project: Hadoop HDFS
  Issue Type: Test
Reporter: Ted Yu
Assignee: Stephen Chu
Priority: Minor
 Attachments: HDFS-6946.1.patch, testBalancer0Integrity-failure.log


 From build #1849 :
 {code}
 REGRESSION:  
 org.apache.hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer.testBalancer0Integrity
 Error Message:
 Cluster failed to reached expected values of totalSpace (current: 750, 
 expected: 750), or usedSpace (current: 140, expected: 150), in more than 
 4 msec.
 Stack Trace:
 java.util.concurrent.TimeoutException: Cluster failed to reached expected 
 values of totalSpace (current: 750, expected: 750), or usedSpace (current: 
 140, expected: 150), in more than 4 msec.
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.waitForHeartBeat(TestBalancer.java:253)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.runBalancer(TestBalancer.java:578)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.doTest(TestBalancer.java:551)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.doTest(TestBalancer.java:437)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.oneNodeTest(TestBalancer.java:645)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.testBalancer0Internal(TestBalancer.java:759)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer.testBalancer0Integrity(TestBalancerWithSaslDataTransfer.java:34)
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6946) TestBalancerWithSaslDataTransfer fails in trunk

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6946?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114205#comment-14114205
 ] 

Hadoop QA commented on HDFS-6946:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12665044/testBalancer0Integrity-failure.log
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7831//console

This message is automatically generated.

 TestBalancerWithSaslDataTransfer fails in trunk
 ---

 Key: HDFS-6946
 URL: https://issues.apache.org/jira/browse/HDFS-6946
 Project: Hadoop HDFS
  Issue Type: Test
Reporter: Ted Yu
Assignee: Stephen Chu
Priority: Minor
 Attachments: HDFS-6946.1.patch, testBalancer0Integrity-failure.log


 From build #1849 :
 {code}
 REGRESSION:  
 org.apache.hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer.testBalancer0Integrity
 Error Message:
 Cluster failed to reached expected values of totalSpace (current: 750, 
 expected: 750), or usedSpace (current: 140, expected: 150), in more than 
 4 msec.
 Stack Trace:
 java.util.concurrent.TimeoutException: Cluster failed to reached expected 
 values of totalSpace (current: 750, expected: 750), or usedSpace (current: 
 140, expected: 150), in more than 4 msec.
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.waitForHeartBeat(TestBalancer.java:253)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.runBalancer(TestBalancer.java:578)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.doTest(TestBalancer.java:551)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.doTest(TestBalancer.java:437)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.oneNodeTest(TestBalancer.java:645)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.testBalancer0Internal(TestBalancer.java:759)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer.testBalancer0Integrity(TestBalancerWithSaslDataTransfer.java:34)
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6946) TestBalancerWithSaslDataTransfer fails in trunk

2014-08-28 Thread Chen He (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6946?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114210#comment-14114210
 ] 

Chen He commented on HDFS-6946:
---

TestRefreshUserMappings errors because of HADOOP-11020

 TestBalancerWithSaslDataTransfer fails in trunk
 ---

 Key: HDFS-6946
 URL: https://issues.apache.org/jira/browse/HDFS-6946
 Project: Hadoop HDFS
  Issue Type: Test
Reporter: Ted Yu
Assignee: Stephen Chu
Priority: Minor
 Attachments: HDFS-6946.1.patch, testBalancer0Integrity-failure.log


 From build #1849 :
 {code}
 REGRESSION:  
 org.apache.hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer.testBalancer0Integrity
 Error Message:
 Cluster failed to reached expected values of totalSpace (current: 750, 
 expected: 750), or usedSpace (current: 140, expected: 150), in more than 
 4 msec.
 Stack Trace:
 java.util.concurrent.TimeoutException: Cluster failed to reached expected 
 values of totalSpace (current: 750, expected: 750), or usedSpace (current: 
 140, expected: 150), in more than 4 msec.
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.waitForHeartBeat(TestBalancer.java:253)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.runBalancer(TestBalancer.java:578)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.doTest(TestBalancer.java:551)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.doTest(TestBalancer.java:437)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.oneNodeTest(TestBalancer.java:645)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.testBalancer0Internal(TestBalancer.java:759)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer.testBalancer0Integrity(TestBalancerWithSaslDataTransfer.java:34)
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6867) For DFSOutputStream, do pipeline recovery for a single block in the background

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6867?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114207#comment-14114207
 ] 

Hadoop QA commented on HDFS-6867:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12665006/HDFS-6867-20140828-1.patch
  against trunk revision c4c9a78.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  
org.apache.hadoop.hdfs.server.namenode.ha.TestPipelinesFailover
  org.apache.hadoop.hdfs.TestSafeMode
  org.apache.hadoop.hdfs.web.TestWebHdfsFileSystemContract

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7830//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7830//console

This message is automatically generated.

 For DFSOutputStream, do pipeline recovery for a single block in the background
 --

 Key: HDFS-6867
 URL: https://issues.apache.org/jira/browse/HDFS-6867
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Reporter: Colin Patrick McCabe
Assignee: Zhe Zhang
 Attachments: HDFS-6867-20140827-2.patch, HDFS-6867-20140827-3.patch, 
 HDFS-6867-20140827.patch, HDFS-6867-20140828-1.patch, 
 HDFS-6867-design-20140820.pdf, HDFS-6867-design-20140821.pdf, 
 HDFS-6867-design-20140822.pdf, HDFS-6867-design-20140827.pdf


 For DFSOutputStream, we should be able to do pipeline recovery in the 
 background, while the user is continuing to write to the file.  This is 
 especially useful for long-lived clients that write to an HDFS file slowly. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6940) Initial refactoring to allow ConsensusNode implementation

2014-08-28 Thread Konstantin Boudnik (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6940?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114213#comment-14114213
 ] 

Konstantin Boudnik commented on HDFS-6940:
--

bq. I think it'd be much better if you could somehow abstract out the behavior 
of the NN that the ConsensusNode needs to change into some sort of plugin 
interface, with a default implementation just being what the NN currently does, 
and then you could provide an alternate implementation that does what the 
ConsensusNode needs to do.
If I am reading this right, you'd be ok with a potentially huge refactoring of 
NN followed by one the two:
 # significant duplication of the NN code in the CNode
 # unnecessarily exposing the implementation of many intimate parts of NN
Or would it acceptable to add a dynamic dependency injection mechanism, perhaps?

 Initial refactoring to allow ConsensusNode implementation
 -

 Key: HDFS-6940
 URL: https://issues.apache.org/jira/browse/HDFS-6940
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Affects Versions: 3.0.0
Reporter: Konstantin Shvachko
Assignee: Konstantin Shvachko
 Attachments: HDFS-6940.patch


 Minor refactoring of FSNamesystem to open private methods that are needed for 
 CNode implementation.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6865) Byte array native checksumming on client side (HDFS changes)

2014-08-28 Thread stack (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6865?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114221#comment-14114221
 ] 

stack commented on HDFS-6865:
-

bq, Stack, by any chance did you run any HBase correctness or system tests with 
the patch enabled? 

Ran a short test.IntegrationTestBigLinkedList on a small cluster. 

{code}
2014-08-28 12:18:30,126 INFO  [main] test.IntegrationTestBigLinkedList$Loop: 
Verify finished with succees. Total nodes=200
{code}

I'd imagine it would fail if we were misreading.

 Byte array native checksumming on client side (HDFS changes)
 

 Key: HDFS-6865
 URL: https://issues.apache.org/jira/browse/HDFS-6865
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: hdfs-client, performance
Reporter: James Thomas
Assignee: James Thomas
 Attachments: HDFS-6865.2.patch, HDFS-6865.3.patch, HDFS-6865.4.patch, 
 HDFS-6865.5.patch, HDFS-6865.6.patch, HDFS-6865.7.patch, HDFS-6865.8.patch, 
 HDFS-6865.patch


 Refactor FSOutputSummer to buffer data and use the native checksum 
 calculation functionality introduced in HADOOP-10975.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6951) Saving namespace and restarting NameNode will remove existing encryption zones

2014-08-28 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6951?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb updated HDFS-6951:
---

Attachment: HDFS-6951.002.patch

[~andrew.wang], [~hitliuyi],

Thanks for your reviews and comments.  I've made the changes that you 
suggested. I've also run TestOfflineEditsViewer and it passes:

---
 T E S T S
---
Running org.apache.hadoop.hdfs.tools.offlineEditsViewer.TestOfflineEditsViewer
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 10.466 sec - in 
org.apache.hadoop.hdfs.tools.offlineEditsViewer.TestOfflineEditsViewer

Results :

Tests run: 3, Failures: 0, Errors: 0, Skipped: 0


 Saving namespace and restarting NameNode will remove existing encryption zones
 --

 Key: HDFS-6951
 URL: https://issues.apache.org/jira/browse/HDFS-6951
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: encryption
Affects Versions: 3.0.0
Reporter: Stephen Chu
Assignee: Charles Lamb
 Attachments: HDFS-6951-prelim.002.patch, HDFS-6951-testrepo.patch, 
 HDFS-6951.001.patch, HDFS-6951.002.patch, editsStored


 Currently, when users save namespace and restart the NameNode, pre-existing 
 encryption zones will be wiped out.
 I could reproduce this on a pseudo-distributed cluster:
 * Create an encryption zone
 * List encryption zones and verify the newly created zone is present
 * Save the namespace
 * Kill and restart the NameNode
 * List the encryption zones and you'll find the encryption zone is missing
 I've attached a test case for {{TestEncryptionZones}} that reproduces this as 
 well. Removing the saveNamespace call will get the test to pass.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HDFS-6965) NN continues to issue block locations for DNs with full disks

2014-08-28 Thread chang li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

chang li reassigned HDFS-6965:
--

Assignee: chang li

 NN continues to issue block locations for DNs with full disks
 -

 Key: HDFS-6965
 URL: https://issues.apache.org/jira/browse/HDFS-6965
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.0.0-alpha, 3.0.0
Reporter: Daryn Sharp
Assignee: chang li

 Encountered issues where DNs have less space than a full block and reject 
 incoming transfers.  The NN continues giving out locations for these nodes 
 for some period of time.  It does not appear to be related to the DN's cached 
 disk usage.
 One impact is required replications are delayed when a full DN is chosen for 
 the pipeline.  A DN cannot report a broken pipeline so the replication must 
 timeout (5m) before new targets are chosen.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6951) Saving namespace and restarting NameNode will remove existing encryption zones

2014-08-28 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6951?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb updated HDFS-6951:
---

Attachment: editsStored

The binary editsStored file is attached.

 Saving namespace and restarting NameNode will remove existing encryption zones
 --

 Key: HDFS-6951
 URL: https://issues.apache.org/jira/browse/HDFS-6951
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: encryption
Affects Versions: 3.0.0
Reporter: Stephen Chu
Assignee: Charles Lamb
 Attachments: HDFS-6951-prelim.002.patch, HDFS-6951-testrepo.patch, 
 HDFS-6951.001.patch, HDFS-6951.002.patch, editsStored


 Currently, when users save namespace and restart the NameNode, pre-existing 
 encryption zones will be wiped out.
 I could reproduce this on a pseudo-distributed cluster:
 * Create an encryption zone
 * List encryption zones and verify the newly created zone is present
 * Save the namespace
 * Kill and restart the NameNode
 * List the encryption zones and you'll find the encryption zone is missing
 I've attached a test case for {{TestEncryptionZones}} that reproduces this as 
 well. Removing the saveNamespace call will get the test to pass.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6951) Saving namespace and restarting NameNode will remove existing encryption zones

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6951?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114243#comment-14114243
 ] 

Hadoop QA commented on HDFS-6951:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12665047/editsStored
  against trunk revision c4c9a78.

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7832//console

This message is automatically generated.

 Saving namespace and restarting NameNode will remove existing encryption zones
 --

 Key: HDFS-6951
 URL: https://issues.apache.org/jira/browse/HDFS-6951
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: encryption
Affects Versions: 3.0.0
Reporter: Stephen Chu
Assignee: Charles Lamb
 Attachments: HDFS-6951-prelim.002.patch, HDFS-6951-testrepo.patch, 
 HDFS-6951.001.patch, HDFS-6951.002.patch, editsStored


 Currently, when users save namespace and restart the NameNode, pre-existing 
 encryption zones will be wiped out.
 I could reproduce this on a pseudo-distributed cluster:
 * Create an encryption zone
 * List encryption zones and verify the newly created zone is present
 * Save the namespace
 * Kill and restart the NameNode
 * List the encryption zones and you'll find the encryption zone is missing
 I've attached a test case for {{TestEncryptionZones}} that reproduces this as 
 well. Removing the saveNamespace call will get the test to pass.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6860) BlockStateChange logs are too noisy

2014-08-28 Thread chang li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6860?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

chang li updated HDFS-6860:
---

Attachment: HDFS6860.patch

have changed most of the blockstatechange log from info level to debug level.

 BlockStateChange logs are too noisy
 ---

 Key: HDFS-6860
 URL: https://issues.apache.org/jira/browse/HDFS-6860
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.5.0
Reporter: Arpit Agarwal
Assignee: chang li
  Labels: newbie
 Attachments: HDFS6860.patch


 Block State Change logs are too noisy at the default INFO level and affect NN 
 performance on busy clusters.
 Most of these state changes can be logged at debug level instead.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6800) Determine how Datanode layout changes should interact with rolling upgrade

2014-08-28 Thread James Thomas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6800?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

James Thomas updated HDFS-6800:
---

Attachment: HDFS-6800.7.patch

Uploaded patch that addresses case 2) above. [~arpitagarwal], thoughts?

 Determine how Datanode layout changes should interact with rolling upgrade
 --

 Key: HDFS-6800
 URL: https://issues.apache.org/jira/browse/HDFS-6800
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode
Affects Versions: 2.6.0
Reporter: Colin Patrick McCabe
Assignee: James Thomas
 Attachments: HDFS-6800.2.patch, HDFS-6800.3.patch, HDFS-6800.4.patch, 
 HDFS-6800.5.patch, HDFS-6800.6.patch, HDFS-6800.7.patch, HDFS-6800.patch


 We need to handle attempts to rolling-upgrade the DataNode to a new storage 
 directory layout.
 One approach is to disallow such upgrades.  If we choose this approach, we 
 should make sure that the system administrator gets a helpful error message 
 and a clean failure when trying to use rolling upgrade to a version that 
 doesn't support it.  Based on the compatibility guarantees described in 
 HDFS-5535, this would mean that *any* future DataNode layout changes would 
 require a major version upgrade.
 Another approach would be to support rolling upgrade from an old DN storage 
 layout to a new layout.  This approach requires us to change our 
 documentation to explain to users that they should supply the {{\-rollback}} 
 command on the command-line when re-starting the DataNodes during rolling 
 rollback.  Currently the documentation just says to restart the DataNode 
 normally.
 Another issue here is that the DataNode's usage message describes rollback 
 options that no longer exist.  The help text says that the DN supports 
 {{\-rollingupgrade rollback}}, but this option was removed by HDFS-6005.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6860) BlockStateChange logs are too noisy

2014-08-28 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6860?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114291#comment-14114291
 ] 

Arpit Agarwal commented on HDFS-6860:
-

Thanks for fixing this [~lichangleo]. Can you also wrap each call in a 
{{isDebugEnabled()}} call. That way we don't incur the overhead of generating 
the log strings unless debug logging is turned on. e.g.

{code}
if (NameNode.blockStateChangeLog.isDebugEnabled()) {
NameNode.blockStateChangeLog.debug(BLOCK* Removing stale replica 
 + from location:  + r.getExpectedStorageLocation());
}
{code}

 BlockStateChange logs are too noisy
 ---

 Key: HDFS-6860
 URL: https://issues.apache.org/jira/browse/HDFS-6860
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.5.0
Reporter: Arpit Agarwal
Assignee: chang li
  Labels: newbie
 Attachments: HDFS6860.patch


 Block State Change logs are too noisy at the default INFO level and affect NN 
 performance on busy clusters.
 Most of these state changes can be logged at debug level instead.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6940) Initial refactoring to allow ConsensusNode implementation

2014-08-28 Thread Aaron T. Myers (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6940?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114294#comment-14114294
 ] 

Aaron T. Myers commented on HDFS-6940:
--

{quote}
If I am reading this right, you'd be ok with a potentially huge refactoring of 
NN followed by one the two:
1. significant duplication of the NN code in the CNode
2. unnecessarily exposing the implementation of many intimate parts of NN
{quote}

I was not necessarily suggesting any of those things. It's not at all clear to 
me that said refactor would necessarily be huge, or that it would result in 
any duplication of NN code in the ConsensusNode. If it's designed well I don't 
see why either of those things would necessarily be the case. That very much 
remains to be seen, and just asserting that it would be so is not very 
constructive.

 Initial refactoring to allow ConsensusNode implementation
 -

 Key: HDFS-6940
 URL: https://issues.apache.org/jira/browse/HDFS-6940
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Affects Versions: 3.0.0
Reporter: Konstantin Shvachko
Assignee: Konstantin Shvachko
 Attachments: HDFS-6940.patch


 Minor refactoring of FSNamesystem to open private methods that are needed for 
 CNode implementation.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6867) For DFSOutputStream, do pipeline recovery for a single block in the background

2014-08-28 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HDFS-6867:


Attachment: HDFS-6867-20140828-2.patch

The new patch adds a type of policy (BACKGROUND) where background recovery 
worker will be invoked. 

 For DFSOutputStream, do pipeline recovery for a single block in the background
 --

 Key: HDFS-6867
 URL: https://issues.apache.org/jira/browse/HDFS-6867
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Reporter: Colin Patrick McCabe
Assignee: Zhe Zhang
 Attachments: HDFS-6867-20140827-2.patch, HDFS-6867-20140827-3.patch, 
 HDFS-6867-20140827.patch, HDFS-6867-20140828-1.patch, 
 HDFS-6867-20140828-2.patch, HDFS-6867-design-20140820.pdf, 
 HDFS-6867-design-20140821.pdf, HDFS-6867-design-20140822.pdf, 
 HDFS-6867-design-20140827.pdf


 For DFSOutputStream, we should be able to do pipeline recovery in the 
 background, while the user is continuing to write to the file.  This is 
 especially useful for long-lived clients that write to an HDFS file slowly. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6800) Determine how Datanode layout changes should interact with rolling upgrade

2014-08-28 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6800?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114306#comment-14114306
 ] 

Arpit Agarwal commented on HDFS-6800:
-

James, would it be simpler to just fix {{NameNodeRpcServer.blockReport}} to 
FinalizeCommand only if {{namesystem.isRollingUpgrade}} returns false or is 
there some other implication of leaving {{isUpgradeFinalized}} to false that I 
am not thinking of?

 Determine how Datanode layout changes should interact with rolling upgrade
 --

 Key: HDFS-6800
 URL: https://issues.apache.org/jira/browse/HDFS-6800
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode
Affects Versions: 2.6.0
Reporter: Colin Patrick McCabe
Assignee: James Thomas
 Attachments: HDFS-6800.2.patch, HDFS-6800.3.patch, HDFS-6800.4.patch, 
 HDFS-6800.5.patch, HDFS-6800.6.patch, HDFS-6800.7.patch, HDFS-6800.patch


 We need to handle attempts to rolling-upgrade the DataNode to a new storage 
 directory layout.
 One approach is to disallow such upgrades.  If we choose this approach, we 
 should make sure that the system administrator gets a helpful error message 
 and a clean failure when trying to use rolling upgrade to a version that 
 doesn't support it.  Based on the compatibility guarantees described in 
 HDFS-5535, this would mean that *any* future DataNode layout changes would 
 require a major version upgrade.
 Another approach would be to support rolling upgrade from an old DN storage 
 layout to a new layout.  This approach requires us to change our 
 documentation to explain to users that they should supply the {{\-rollback}} 
 command on the command-line when re-starting the DataNodes during rolling 
 rollback.  Currently the documentation just says to restart the DataNode 
 normally.
 Another issue here is that the DataNode's usage message describes rollback 
 options that no longer exist.  The help text says that the DN supports 
 {{\-rollingupgrade rollback}}, but this option was removed by HDFS-6005.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6860) BlockStateChange logs are too noisy

2014-08-28 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6860?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114304#comment-14114304
 ] 

Andrew Wang commented on HDFS-6860:
---

Is this a good time to also switch BlockStateChange over to use slf4j and to 
use {} templates? Then we could avoid the if-wrappers.

 BlockStateChange logs are too noisy
 ---

 Key: HDFS-6860
 URL: https://issues.apache.org/jira/browse/HDFS-6860
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.5.0
Reporter: Arpit Agarwal
Assignee: chang li
  Labels: newbie
 Attachments: HDFS6860.patch


 Block State Change logs are too noisy at the default INFO level and affect NN 
 performance on busy clusters.
 Most of these state changes can be logged at debug level instead.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Comment Edited] (HDFS-6800) Determine how Datanode layout changes should interact with rolling upgrade

2014-08-28 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6800?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114306#comment-14114306
 ] 

Arpit Agarwal edited comment on HDFS-6800 at 8/28/14 8:38 PM:
--

James, would it be simpler to just fix {{NameNodeRpcServer.blockReport}} to 
return FinalizeCommand only if {{namesystem.isRollingUpgrade}} returns false or 
is there some other implication of leaving {{isUpgradeFinalized}} to false that 
I am not thinking of?


was (Author: arpitagarwal):
James, would it be simpler to just fix {{NameNodeRpcServer.blockReport}} to 
FinalizeCommand only if {{namesystem.isRollingUpgrade}} returns false or is 
there some other implication of leaving {{isUpgradeFinalized}} to false that I 
am not thinking of?

 Determine how Datanode layout changes should interact with rolling upgrade
 --

 Key: HDFS-6800
 URL: https://issues.apache.org/jira/browse/HDFS-6800
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode
Affects Versions: 2.6.0
Reporter: Colin Patrick McCabe
Assignee: James Thomas
 Attachments: HDFS-6800.2.patch, HDFS-6800.3.patch, HDFS-6800.4.patch, 
 HDFS-6800.5.patch, HDFS-6800.6.patch, HDFS-6800.7.patch, HDFS-6800.patch


 We need to handle attempts to rolling-upgrade the DataNode to a new storage 
 directory layout.
 One approach is to disallow such upgrades.  If we choose this approach, we 
 should make sure that the system administrator gets a helpful error message 
 and a clean failure when trying to use rolling upgrade to a version that 
 doesn't support it.  Based on the compatibility guarantees described in 
 HDFS-5535, this would mean that *any* future DataNode layout changes would 
 require a major version upgrade.
 Another approach would be to support rolling upgrade from an old DN storage 
 layout to a new layout.  This approach requires us to change our 
 documentation to explain to users that they should supply the {{\-rollback}} 
 command on the command-line when re-starting the DataNodes during rolling 
 rollback.  Currently the documentation just says to restart the DataNode 
 normally.
 Another issue here is that the DataNode's usage message describes rollback 
 options that no longer exist.  The help text says that the DN supports 
 {{\-rollingupgrade rollback}}, but this option was removed by HDFS-6005.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6968) Configurable replication degree in the hadoop archive command

2014-08-28 Thread Zhe Zhang (JIRA)
Zhe Zhang created HDFS-6968:
---

 Summary: Configurable replication degree in the hadoop archive 
command
 Key: HDFS-6968
 URL: https://issues.apache.org/jira/browse/HDFS-6968
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Zhe Zhang
Assignee: Zhe Zhang
Priority: Minor


Due to be below hard-coded replication degree in {{HadoopArchives}}, the 
{{archive}} command will fail if HDFS maximum replication has already been 
configured to a number lower than 10. 

{code:java}
//increase the replication of src files
jobfs.setReplication(srcFiles, (short) 10);
{code}

This Jira will make the {{archive}} command configurable with desired 
replication degree.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6860) BlockStateChange logs are too noisy

2014-08-28 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6860?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114327#comment-14114327
 ] 

Arpit Agarwal commented on HDFS-6860:
-

[~lichangleo], see if you want to follow Andrew's suggestion and update 
BlockStateChange to use slf4j.

I'd be +1 either way - slf4j or existing log4j calls with {{isDebugEnabled}} 
checks.

 BlockStateChange logs are too noisy
 ---

 Key: HDFS-6860
 URL: https://issues.apache.org/jira/browse/HDFS-6860
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.5.0
Reporter: Arpit Agarwal
Assignee: chang li
  Labels: newbie
 Attachments: HDFS6860.patch


 Block State Change logs are too noisy at the default INFO level and affect NN 
 performance on busy clusters.
 Most of these state changes can be logged at debug level instead.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6951) Saving namespace and restarting NameNode will remove existing encryption zones

2014-08-28 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6951?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb updated HDFS-6951:
---

Attachment: HDFS-6951.002.patch

Resubmitting .002 again so that it is the latest attachment in the hopes that 
test-patch will pick this up instead of editsStored.

 Saving namespace and restarting NameNode will remove existing encryption zones
 --

 Key: HDFS-6951
 URL: https://issues.apache.org/jira/browse/HDFS-6951
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: encryption
Affects Versions: 3.0.0
Reporter: Stephen Chu
Assignee: Charles Lamb
 Attachments: HDFS-6951-prelim.002.patch, HDFS-6951-testrepo.patch, 
 HDFS-6951.001.patch, HDFS-6951.002.patch, editsStored


 Currently, when users save namespace and restart the NameNode, pre-existing 
 encryption zones will be wiped out.
 I could reproduce this on a pseudo-distributed cluster:
 * Create an encryption zone
 * List encryption zones and verify the newly created zone is present
 * Save the namespace
 * Kill and restart the NameNode
 * List the encryption zones and you'll find the encryption zone is missing
 I've attached a test case for {{TestEncryptionZones}} that reproduces this as 
 well. Removing the saveNamespace call will get the test to pass.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6951) Saving namespace and restarting NameNode will remove existing encryption zones

2014-08-28 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6951?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb updated HDFS-6951:
---

Attachment: (was: HDFS-6951.002.patch)

 Saving namespace and restarting NameNode will remove existing encryption zones
 --

 Key: HDFS-6951
 URL: https://issues.apache.org/jira/browse/HDFS-6951
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: encryption
Affects Versions: 3.0.0
Reporter: Stephen Chu
Assignee: Charles Lamb
 Attachments: HDFS-6951-prelim.002.patch, HDFS-6951-testrepo.patch, 
 HDFS-6951.001.patch, HDFS-6951.002.patch, editsStored


 Currently, when users save namespace and restart the NameNode, pre-existing 
 encryption zones will be wiped out.
 I could reproduce this on a pseudo-distributed cluster:
 * Create an encryption zone
 * List encryption zones and verify the newly created zone is present
 * Save the namespace
 * Kill and restart the NameNode
 * List the encryption zones and you'll find the encryption zone is missing
 I've attached a test case for {{TestEncryptionZones}} that reproduces this as 
 well. Removing the saveNamespace call will get the test to pass.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6800) Determine how Datanode layout changes should interact with rolling upgrade

2014-08-28 Thread James Thomas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6800?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

James Thomas updated HDFS-6800:
---

Attachment: HDFS-6800.8.patch

Good call. Updated patch accordingly.

 Determine how Datanode layout changes should interact with rolling upgrade
 --

 Key: HDFS-6800
 URL: https://issues.apache.org/jira/browse/HDFS-6800
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode
Affects Versions: 2.6.0
Reporter: Colin Patrick McCabe
Assignee: James Thomas
 Attachments: HDFS-6800.2.patch, HDFS-6800.3.patch, HDFS-6800.4.patch, 
 HDFS-6800.5.patch, HDFS-6800.6.patch, HDFS-6800.7.patch, HDFS-6800.8.patch, 
 HDFS-6800.patch


 We need to handle attempts to rolling-upgrade the DataNode to a new storage 
 directory layout.
 One approach is to disallow such upgrades.  If we choose this approach, we 
 should make sure that the system administrator gets a helpful error message 
 and a clean failure when trying to use rolling upgrade to a version that 
 doesn't support it.  Based on the compatibility guarantees described in 
 HDFS-5535, this would mean that *any* future DataNode layout changes would 
 require a major version upgrade.
 Another approach would be to support rolling upgrade from an old DN storage 
 layout to a new layout.  This approach requires us to change our 
 documentation to explain to users that they should supply the {{\-rollback}} 
 command on the command-line when re-starting the DataNodes during rolling 
 rollback.  Currently the documentation just says to restart the DataNode 
 normally.
 Another issue here is that the DataNode's usage message describes rollback 
 options that no longer exist.  The help text says that the DN supports 
 {{\-rollingupgrade rollback}}, but this option was removed by HDFS-6005.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6826) Plugin interface to enable delegation of HDFS authorization assertions

2014-08-28 Thread Alejandro Abdelnur (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6826?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alejandro Abdelnur updated HDFS-6826:
-

Attachment: HDFS-6826v7.6.patch

Uploading v7.6, which is based on v7.4 (based on Jitendra's preference).

This patch adds the {{isClientOp()}} method to the provider which indicates if 
the current invocation is done as part of a client FS operation or as part of a 
NameNode operation (fs image save, fs edit log, etc). I've had to create an 
interceptor for the {{ClientProtocol}}.

The testcases do positive and negative testing for this.

This addresses my action item #1 from my discussion with @daryn.

 Plugin interface to enable delegation of HDFS authorization assertions
 --

 Key: HDFS-6826
 URL: https://issues.apache.org/jira/browse/HDFS-6826
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: security
Affects Versions: 2.4.1
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Attachments: HDFS-6826-idea.patch, HDFS-6826-idea2.patch, 
 HDFS-6826v3.patch, HDFS-6826v4.patch, HDFS-6826v5.patch, HDFS-6826v6.patch, 
 HDFS-6826v7.1.patch, HDFS-6826v7.2.patch, HDFS-6826v7.3.patch, 
 HDFS-6826v7.4.patch, HDFS-6826v7.5.patch, HDFS-6826v7.6.patch, 
 HDFS-6826v7.patch, HDFS-6826v8.patch, 
 HDFSPluggableAuthorizationProposal-v2.pdf, 
 HDFSPluggableAuthorizationProposal.pdf


 When Hbase data, HiveMetaStore data or Search data is accessed via services 
 (Hbase region servers, HiveServer2, Impala, Solr) the services can enforce 
 permissions on corresponding entities (databases, tables, views, columns, 
 search collections, documents). It is desirable, when the data is accessed 
 directly by users accessing the underlying data files (i.e. from a MapReduce 
 job), that the permission of the data files map to the permissions of the 
 corresponding data entity (i.e. table, column family or search collection).
 To enable this we need to have the necessary hooks in place in the NameNode 
 to delegate authorization to an external system that can map HDFS 
 files/directories to data entities and resolve their permissions based on the 
 data entities permissions.
 I’ll be posting a design proposal in the next few days.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6800) Determine how Datanode layout changes should interact with rolling upgrade

2014-08-28 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6800?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114347#comment-14114347
 ] 

Arpit Agarwal commented on HDFS-6800:
-

Thanks for all the patch iterations and your patience James. Let me think about 
this some more today but I think your latest patch is ready to commit today.

Do you guys have any rolling upgrade testing planned with HDFS-6482 and 
HDFS-6800?

 Determine how Datanode layout changes should interact with rolling upgrade
 --

 Key: HDFS-6800
 URL: https://issues.apache.org/jira/browse/HDFS-6800
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode
Affects Versions: 2.6.0
Reporter: Colin Patrick McCabe
Assignee: James Thomas
 Attachments: HDFS-6800.2.patch, HDFS-6800.3.patch, HDFS-6800.4.patch, 
 HDFS-6800.5.patch, HDFS-6800.6.patch, HDFS-6800.7.patch, HDFS-6800.8.patch, 
 HDFS-6800.patch


 We need to handle attempts to rolling-upgrade the DataNode to a new storage 
 directory layout.
 One approach is to disallow such upgrades.  If we choose this approach, we 
 should make sure that the system administrator gets a helpful error message 
 and a clean failure when trying to use rolling upgrade to a version that 
 doesn't support it.  Based on the compatibility guarantees described in 
 HDFS-5535, this would mean that *any* future DataNode layout changes would 
 require a major version upgrade.
 Another approach would be to support rolling upgrade from an old DN storage 
 layout to a new layout.  This approach requires us to change our 
 documentation to explain to users that they should supply the {{\-rollback}} 
 command on the command-line when re-starting the DataNodes during rolling 
 rollback.  Currently the documentation just says to restart the DataNode 
 normally.
 Another issue here is that the DataNode's usage message describes rollback 
 options that no longer exist.  The help text says that the DN supports 
 {{\-rollingupgrade rollback}}, but this option was removed by HDFS-6005.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6968) Configurable replication degree in the hadoop archive command

2014-08-28 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6968?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HDFS-6968:


Attachment: HDFS-6968.patch

 Configurable replication degree in the hadoop archive command
 -

 Key: HDFS-6968
 URL: https://issues.apache.org/jira/browse/HDFS-6968
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Zhe Zhang
Assignee: Zhe Zhang
Priority: Minor
 Attachments: HDFS-6968.patch


 Due to be below hard-coded replication degree in {{HadoopArchives}}, the 
 {{archive}} command will fail if HDFS maximum replication has already been 
 configured to a number lower than 10. 
 {code:java}
 //increase the replication of src files
 jobfs.setReplication(srcFiles, (short) 10);
 {code}
 This Jira will make the {{archive}} command configurable with desired 
 replication degree.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6968) Configurable replication degree in the hadoop archive command

2014-08-28 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6968?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HDFS-6968:


Status: Patch Available  (was: Open)

 Configurable replication degree in the hadoop archive command
 -

 Key: HDFS-6968
 URL: https://issues.apache.org/jira/browse/HDFS-6968
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Zhe Zhang
Assignee: Zhe Zhang
Priority: Minor
 Attachments: HDFS-6968.patch


 Due to be below hard-coded replication degree in {{HadoopArchives}}, the 
 {{archive}} command will fail if HDFS maximum replication has already been 
 configured to a number lower than 10. 
 {code:java}
 //increase the replication of src files
 jobfs.setReplication(srcFiles, (short) 10);
 {code}
 This Jira will make the {{archive}} command configurable with desired 
 replication degree.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6800) Determine how Datanode layout changes should interact with rolling upgrade

2014-08-28 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6800?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114352#comment-14114352
 ] 

Arpit Agarwal commented on HDFS-6800:
-

The reason I asked about the testing is so we can start thinking about getting 
the changes to branch-2.

 Determine how Datanode layout changes should interact with rolling upgrade
 --

 Key: HDFS-6800
 URL: https://issues.apache.org/jira/browse/HDFS-6800
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode
Affects Versions: 2.6.0
Reporter: Colin Patrick McCabe
Assignee: James Thomas
 Attachments: HDFS-6800.2.patch, HDFS-6800.3.patch, HDFS-6800.4.patch, 
 HDFS-6800.5.patch, HDFS-6800.6.patch, HDFS-6800.7.patch, HDFS-6800.8.patch, 
 HDFS-6800.patch


 We need to handle attempts to rolling-upgrade the DataNode to a new storage 
 directory layout.
 One approach is to disallow such upgrades.  If we choose this approach, we 
 should make sure that the system administrator gets a helpful error message 
 and a clean failure when trying to use rolling upgrade to a version that 
 doesn't support it.  Based on the compatibility guarantees described in 
 HDFS-5535, this would mean that *any* future DataNode layout changes would 
 require a major version upgrade.
 Another approach would be to support rolling upgrade from an old DN storage 
 layout to a new layout.  This approach requires us to change our 
 documentation to explain to users that they should supply the {{\-rollback}} 
 command on the command-line when re-starting the DataNodes during rolling 
 rollback.  Currently the documentation just says to restart the DataNode 
 normally.
 Another issue here is that the DataNode's usage message describes rollback 
 options that no longer exist.  The help text says that the DN supports 
 {{\-rollingupgrade rollback}}, but this option was removed by HDFS-6005.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6800) Determine how Datanode layout changes should interact with rolling upgrade

2014-08-28 Thread James Thomas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6800?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114356#comment-14114356
 ] 

James Thomas commented on HDFS-6800:


Great, thanks for your help on this, Arpit. I will try to run some tests today 
or tomorrow.

 Determine how Datanode layout changes should interact with rolling upgrade
 --

 Key: HDFS-6800
 URL: https://issues.apache.org/jira/browse/HDFS-6800
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode
Affects Versions: 2.6.0
Reporter: Colin Patrick McCabe
Assignee: James Thomas
 Attachments: HDFS-6800.2.patch, HDFS-6800.3.patch, HDFS-6800.4.patch, 
 HDFS-6800.5.patch, HDFS-6800.6.patch, HDFS-6800.7.patch, HDFS-6800.8.patch, 
 HDFS-6800.patch


 We need to handle attempts to rolling-upgrade the DataNode to a new storage 
 directory layout.
 One approach is to disallow such upgrades.  If we choose this approach, we 
 should make sure that the system administrator gets a helpful error message 
 and a clean failure when trying to use rolling upgrade to a version that 
 doesn't support it.  Based on the compatibility guarantees described in 
 HDFS-5535, this would mean that *any* future DataNode layout changes would 
 require a major version upgrade.
 Another approach would be to support rolling upgrade from an old DN storage 
 layout to a new layout.  This approach requires us to change our 
 documentation to explain to users that they should supply the {{\-rollback}} 
 command on the command-line when re-starting the DataNodes during rolling 
 rollback.  Currently the documentation just says to restart the DataNode 
 normally.
 Another issue here is that the DataNode's usage message describes rollback 
 options that no longer exist.  The help text says that the DN supports 
 {{\-rollingupgrade rollback}}, but this option was removed by HDFS-6005.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6840) Clients are always sent to the same datanode when read is off rack

2014-08-28 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-6840:
--

Attachment: hdfs-6840.002.patch

Nice catch Yi, new patch that removes that param from the XML file.

 Clients are always sent to the same datanode when read is off rack
 --

 Key: HDFS-6840
 URL: https://issues.apache.org/jira/browse/HDFS-6840
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Jason Lowe
Assignee: Andrew Wang
Priority: Critical
 Attachments: hdfs-6840.001.patch, hdfs-6840.002.patch


 After HDFS-6268 the sorting order of block locations is deterministic for a 
 given block and locality level (e.g.: local, rack. off-rack), so off-rack 
 clients all see the same datanode for the same block.  This leads to very 
 poor behavior in distributed cache localization and other scenarios where 
 many clients all want the same block data at approximately the same time.  
 The one datanode is crushed by the load while the other replicas only handle 
 local and rack-local requests.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6928) 'hdfs put' command should accept lazyPersist flag for testing

2014-08-28 Thread Arpit Agarwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6928?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arpit Agarwal updated HDFS-6928:


Attachment: HDFS-6928.03.patch

Thanks for the review Xiaoyu! 

Fixed the comment. Also fixed {{TestCLI}} and added two test cases for 
{{TestHDFSCLI}}.

 'hdfs put' command should accept lazyPersist flag for testing
 -

 Key: HDFS-6928
 URL: https://issues.apache.org/jira/browse/HDFS-6928
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode
Affects Versions: HDFS-6581
Reporter: Tassapol Athiapinya
Assignee: Arpit Agarwal
 Attachments: HDFS-6928.01.patch, HDFS-6928.02.patch, 
 HDFS-6928.03.patch


 Add a '-l' flag to 'hdfs put' which creates the file with the LAZY_PERSIST 
 option.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6826) Plugin interface to enable delegation of HDFS authorization assertions

2014-08-28 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6826?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114385#comment-14114385
 ] 

Alejandro Abdelnur commented on HDFS-6826:
--

Regarding my action item #2, based on @daryn’s suggestion, I’ve move all data 
authz interception to the {{FsPermissionChecker}}, something like:

*Adding the following methods:*

{code}
  private String getUserName(INode[] nodes) {
return AuthorizationProvider.get().getUser(nodes);
  }

  private String getGroupName(INode[] nodes) {
// wee need to recreate INode[] using the snapshot version of the nodes, 
here or before calling
return AuthorizationProvider.get().getGroup(nodes);
  }

  private FsPermission getFsPermission(INode[] nodes) {
return AuthorizationProvider.get().getFsPermission(nodes);
  }
  
  private AclFeature getAclFeature(INode[] nodes) {
return AuthorizationProvider.get().getAclFeature(nodes);
  }
{code}

And then replacing all the calls user/group/permissions/acls getters of INodes 
calls within {{FsPermissionChecker}} to use the above methods, ie:

The intention was to be able to reuse the already calculated {{INode[]}} chain.

The issues I’m running with this are:

* The {{INode[]}} chain calculated on the entry point of checkPermission does 
not take into account snapshots, thus we will need to recalculate the 
{{INode[]}} to use the right snapshot for the full chain.
* All the logic in {{FSPermissionChecker}} has to be redone to pass the 
{{INode[]}} chain around. This will get tricky as in many places direct array 
access like this 'checkStickyBit(inodes[inodes.length - 2], last, snapshotId);' 
are being done. And this is not always done on the snapshot version of the 
{{INode}}.
* The {{INode[]}} may have null elements, complicating things on the plugin 
side.
* We’ll have to do the same in the {{FSDirectory}} to create file status.

Another issue is that doing this, the plugin is only intercepting getter calls, 
not setter calls.

Overall, it seems the plugin will have to be more complex than with the v7 
approach, it will have less functionality (no writes, no snapshots), and it 
will require some serious rewriting of the {{FsPermissionChecker}}.

 Plugin interface to enable delegation of HDFS authorization assertions
 --

 Key: HDFS-6826
 URL: https://issues.apache.org/jira/browse/HDFS-6826
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: security
Affects Versions: 2.4.1
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Attachments: HDFS-6826-idea.patch, HDFS-6826-idea2.patch, 
 HDFS-6826v3.patch, HDFS-6826v4.patch, HDFS-6826v5.patch, HDFS-6826v6.patch, 
 HDFS-6826v7.1.patch, HDFS-6826v7.2.patch, HDFS-6826v7.3.patch, 
 HDFS-6826v7.4.patch, HDFS-6826v7.5.patch, HDFS-6826v7.6.patch, 
 HDFS-6826v7.patch, HDFS-6826v8.patch, 
 HDFSPluggableAuthorizationProposal-v2.pdf, 
 HDFSPluggableAuthorizationProposal.pdf


 When Hbase data, HiveMetaStore data or Search data is accessed via services 
 (Hbase region servers, HiveServer2, Impala, Solr) the services can enforce 
 permissions on corresponding entities (databases, tables, views, columns, 
 search collections, documents). It is desirable, when the data is accessed 
 directly by users accessing the underlying data files (i.e. from a MapReduce 
 job), that the permission of the data files map to the permissions of the 
 corresponding data entity (i.e. table, column family or search collection).
 To enable this we need to have the necessary hooks in place in the NameNode 
 to delegate authorization to an external system that can map HDFS 
 files/directories to data entities and resolve their permissions based on the 
 data entities permissions.
 I’ll be posting a design proposal in the next few days.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6968) Configurable replication degree in the hadoop archive command

2014-08-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6968?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114395#comment-14114395
 ] 

Hadoop QA commented on HDFS-6968:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12665075/HDFS-6968.patch
  against trunk revision c4c9a78.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-tools/hadoop-archives.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7838//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7838//console

This message is automatically generated.

 Configurable replication degree in the hadoop archive command
 -

 Key: HDFS-6968
 URL: https://issues.apache.org/jira/browse/HDFS-6968
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Zhe Zhang
Assignee: Zhe Zhang
Priority: Minor
 Attachments: HDFS-6968.patch


 Due to be below hard-coded replication degree in {{HadoopArchives}}, the 
 {{archive}} command will fail if HDFS maximum replication has already been 
 configured to a number lower than 10. 
 {code:java}
 //increase the replication of src files
 jobfs.setReplication(srcFiles, (short) 10);
 {code}
 This Jira will make the {{archive}} command configurable with desired 
 replication degree.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6968) Configurable replication degree in the hadoop archive command

2014-08-28 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6968?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114401#comment-14114401
 ] 

Zhe Zhang commented on HDFS-6968:
-

[~andrew.wang] Could you review this? Thx..

 Configurable replication degree in the hadoop archive command
 -

 Key: HDFS-6968
 URL: https://issues.apache.org/jira/browse/HDFS-6968
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Zhe Zhang
Assignee: Zhe Zhang
Priority: Minor
 Attachments: HDFS-6968.patch


 Due to be below hard-coded replication degree in {{HadoopArchives}}, the 
 {{archive}} command will fail if HDFS maximum replication has already been 
 configured to a number lower than 10. 
 {code:java}
 //increase the replication of src files
 jobfs.setReplication(srcFiles, (short) 10);
 {code}
 This Jira will make the {{archive}} command configurable with desired 
 replication degree.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6960) Bugfix in LazyWriter, fix test case and some refactoring

2014-08-28 Thread Xiaoyu Yao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114451#comment-14114451
 ] 

Xiaoyu Yao commented on HDFS-6960:
--

In FsDataSetImpl.java Can you put  the following into debug log? 
{code}
saveNextReplica()
{...
  LOG.info(LazyWriter has no blocks to persist.  +
Thread going to sleep.); 
{code}



 Bugfix in LazyWriter, fix test case and some refactoring
 

 Key: HDFS-6960
 URL: https://issues.apache.org/jira/browse/HDFS-6960
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, test
Affects Versions: HDFS-6581
Reporter: Arpit Agarwal
Assignee: Arpit Agarwal
 Attachments: HDFS-6960.01.patch


 LazyWriter has a bug. While saving the replica to disk we would save it under 
 {{current/lazyPersist/}}. Instead it should be saved under the appropriate 
 subdirectory e.g. {{current/lazyPersist/subdir1/subdir0/}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6134) Transparent data at rest encryption

2014-08-28 Thread Alejandro Abdelnur (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6134?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alejandro Abdelnur updated HDFS-6134:
-

Fix Version/s: (was: 3.0.0)
   2.6.0

merged to branch-2.

 Transparent data at rest encryption
 ---

 Key: HDFS-6134
 URL: https://issues.apache.org/jira/browse/HDFS-6134
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: security
Affects Versions: 3.0.0, 2.3.0
Reporter: Alejandro Abdelnur
Assignee: Charles Lamb
 Fix For: 2.6.0

 Attachments: HDFS-6134.001.patch, HDFS-6134.002.patch, 
 HDFS-6134_test_plan.pdf, HDFSDataatRestEncryption.pdf, 
 HDFSDataatRestEncryptionProposal_obsolete.pdf, 
 HDFSEncryptionConceptualDesignProposal-2014-06-20.pdf, 
 fs-encryption.2014-08-18.patch, fs-encryption.2014-08-19.patch


 Because of privacy and security regulations, for many industries, sensitive 
 data at rest must be in encrypted form. For example: the health­care industry 
 (HIPAA regulations), the card payment industry (PCI DSS regulations) or the 
 US government (FISMA regulations).
 This JIRA aims to provide a mechanism to encrypt HDFS data at rest that can 
 be used transparently by any application accessing HDFS via Hadoop Filesystem 
 Java API, Hadoop libhdfs C library, or WebHDFS REST API.
 The resulting implementation should be able to be used in compliance with 
 different regulation requirements.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6960) Bugfix in LazyWriter, fix test case and some refactoring

2014-08-28 Thread Xiaoyu Yao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14114476#comment-14114476
 ] 

Xiaoyu Yao commented on HDFS-6960:
--

The change of return value from void to boolean for discardReplica() is not 
being used in the code. Is this necessary?

 Bugfix in LazyWriter, fix test case and some refactoring
 

 Key: HDFS-6960
 URL: https://issues.apache.org/jira/browse/HDFS-6960
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, test
Affects Versions: HDFS-6581
Reporter: Arpit Agarwal
Assignee: Arpit Agarwal
 Attachments: HDFS-6960.01.patch


 LazyWriter has a bug. While saving the replica to disk we would save it under 
 {{current/lazyPersist/}}. Instead it should be saved under the appropriate 
 subdirectory e.g. {{current/lazyPersist/subdir1/subdir0/}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


  1   2   >