[jira] [Commented] (HDFS-6507) Improve DFSAdmin to support HA cluster better

2014-06-18 Thread Zesheng Wu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14037044#comment-14037044
 ] 

Zesheng Wu commented on HDFS-6507:
--

You mean that we should revert the change to ProxyAndInfo and remove addresses 
in logs messages also?

> Improve DFSAdmin to support HA cluster better
> -
>
> Key: HDFS-6507
> URL: https://issues.apache.org/jira/browse/HDFS-6507
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: tools
>Affects Versions: 2.4.0
>Reporter: Zesheng Wu
>Assignee: Zesheng Wu
> Attachments: HDFS-6507.1.patch, HDFS-6507.2.patch, HDFS-6507.3.patch, 
> HDFS-6507.4-inprogress.patch, HDFS-6507.4.patch, HDFS-6507.5.patch, 
> HDFS-6507.6.patch, HDFS-6507.7.patch, HDFS-6507.7.patch, HDFS-6507.8.patch
>
>
> Currently, the commands supported in DFSAdmin can be classified into three 
> categories according to the protocol used:
> 1. ClientProtocol
> Commands in this category generally implement by calling the corresponding 
> function of the DFSClient class, and will call the corresponding remote 
> implementation function at the NN side finally. At the NN side, all these 
> operations are classified into five categories: UNCHECKED, READ, WRITE, 
> CHECKPOINT, JOURNAL. Active NN will allow all operations, and Standby NN only 
> allows UNCHECKED operations. In the current implementation of DFSClient, it 
> will connect one NN first, if the first NN is not Active and the operation is 
> not allowed, it will failover to the second NN. So here comes the problem, 
> some of the commands(setSafeMode, saveNameSpace, restoreFailedStorage, 
> refreshNodes, setBalancerBandwidth, metaSave) in DFSAdmin are classified as 
> UNCHECKED operations, and when executing these commands in the DFSAdmin 
> command line, they will be sent to a definite NN, no matter it is Active or 
> Standby. This may result in two problems: 
> a. If the first tried NN is standby, and the operation takes effect only on 
> Standby NN, which is not the expected result.
> b. If the operation needs to take effect on both NN, but it takes effect on 
> only one NN. In the future, when there is a NN failover, there may have 
> problems.
> Here I propose the following improvements:
> a. If the command can be classified as one of READ/WRITE/CHECKPOINT/JOURNAL 
> operations, we should classify it clearly.
> b. If the command can not be classified as one of the above four operations, 
> or if the command needs to take effect on both NN, we should send the request 
> to both Active and Standby NNs.
> 2. Refresh protocols: RefreshAuthorizationPolicyProtocol, 
> RefreshUserMappingsProtocol, RefreshUserMappingsProtocol, 
> RefreshCallQueueProtocol
> Commands in this category, including refreshServiceAcl, 
> refreshUserToGroupMapping, refreshSuperUserGroupsConfiguration and 
> refreshCallQueue, are implemented by creating a corresponding RPC proxy and 
> sending the request to remote NN. In the current implementation, these 
> requests will be sent to a definite NN, no matter it is Active or Standby. 
> Here I propose that we sent these requests to both NNs.
> 3. ClientDatanodeProtocol
> Commands in this category are handled correctly, no need to improve.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6507) Improve DFSAdmin to support HA cluster better

2014-06-18 Thread Vinayakumar B (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14037041#comment-14037041
 ] 

Vinayakumar B commented on HDFS-6507:
-

Hi [~wuzesheng],
I just realized one more thing
In {{NameNodeProxies.ProxyAndInfo}} we need to set the address only while 
creating the Non-HA proxies. 
because In case of HA, it will set to ":8020" which is not correct 
and its not used also.
This also can be mentioned somewhere in the comment.

So better not set the address for HA proxyinfo.

> Improve DFSAdmin to support HA cluster better
> -
>
> Key: HDFS-6507
> URL: https://issues.apache.org/jira/browse/HDFS-6507
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: tools
>Affects Versions: 2.4.0
>Reporter: Zesheng Wu
>Assignee: Zesheng Wu
> Attachments: HDFS-6507.1.patch, HDFS-6507.2.patch, HDFS-6507.3.patch, 
> HDFS-6507.4-inprogress.patch, HDFS-6507.4.patch, HDFS-6507.5.patch, 
> HDFS-6507.6.patch, HDFS-6507.7.patch, HDFS-6507.7.patch, HDFS-6507.8.patch
>
>
> Currently, the commands supported in DFSAdmin can be classified into three 
> categories according to the protocol used:
> 1. ClientProtocol
> Commands in this category generally implement by calling the corresponding 
> function of the DFSClient class, and will call the corresponding remote 
> implementation function at the NN side finally. At the NN side, all these 
> operations are classified into five categories: UNCHECKED, READ, WRITE, 
> CHECKPOINT, JOURNAL. Active NN will allow all operations, and Standby NN only 
> allows UNCHECKED operations. In the current implementation of DFSClient, it 
> will connect one NN first, if the first NN is not Active and the operation is 
> not allowed, it will failover to the second NN. So here comes the problem, 
> some of the commands(setSafeMode, saveNameSpace, restoreFailedStorage, 
> refreshNodes, setBalancerBandwidth, metaSave) in DFSAdmin are classified as 
> UNCHECKED operations, and when executing these commands in the DFSAdmin 
> command line, they will be sent to a definite NN, no matter it is Active or 
> Standby. This may result in two problems: 
> a. If the first tried NN is standby, and the operation takes effect only on 
> Standby NN, which is not the expected result.
> b. If the operation needs to take effect on both NN, but it takes effect on 
> only one NN. In the future, when there is a NN failover, there may have 
> problems.
> Here I propose the following improvements:
> a. If the command can be classified as one of READ/WRITE/CHECKPOINT/JOURNAL 
> operations, we should classify it clearly.
> b. If the command can not be classified as one of the above four operations, 
> or if the command needs to take effect on both NN, we should send the request 
> to both Active and Standby NNs.
> 2. Refresh protocols: RefreshAuthorizationPolicyProtocol, 
> RefreshUserMappingsProtocol, RefreshUserMappingsProtocol, 
> RefreshCallQueueProtocol
> Commands in this category, including refreshServiceAcl, 
> refreshUserToGroupMapping, refreshSuperUserGroupsConfiguration and 
> refreshCallQueue, are implemented by creating a corresponding RPC proxy and 
> sending the request to remote NN. In the current implementation, these 
> requests will be sent to a definite NN, no matter it is Active or Standby. 
> Here I propose that we sent these requests to both NNs.
> 3. ClientDatanodeProtocol
> Commands in this category are handled correctly, no need to improve.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6507) Improve DFSAdmin to support HA cluster better

2014-06-18 Thread Vinayakumar B (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14037035#comment-14037035
 ] 

Vinayakumar B commented on HDFS-6507:
-

Yes [~jingzhao], I also agree that metaSave and saveNamespace commands are very 
heavy. But as these are only DFSAdmin commands, we assume admin needs to know 
when to execute these commands.

> Improve DFSAdmin to support HA cluster better
> -
>
> Key: HDFS-6507
> URL: https://issues.apache.org/jira/browse/HDFS-6507
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: tools
>Affects Versions: 2.4.0
>Reporter: Zesheng Wu
>Assignee: Zesheng Wu
> Attachments: HDFS-6507.1.patch, HDFS-6507.2.patch, HDFS-6507.3.patch, 
> HDFS-6507.4-inprogress.patch, HDFS-6507.4.patch, HDFS-6507.5.patch, 
> HDFS-6507.6.patch, HDFS-6507.7.patch, HDFS-6507.7.patch, HDFS-6507.8.patch
>
>
> Currently, the commands supported in DFSAdmin can be classified into three 
> categories according to the protocol used:
> 1. ClientProtocol
> Commands in this category generally implement by calling the corresponding 
> function of the DFSClient class, and will call the corresponding remote 
> implementation function at the NN side finally. At the NN side, all these 
> operations are classified into five categories: UNCHECKED, READ, WRITE, 
> CHECKPOINT, JOURNAL. Active NN will allow all operations, and Standby NN only 
> allows UNCHECKED operations. In the current implementation of DFSClient, it 
> will connect one NN first, if the first NN is not Active and the operation is 
> not allowed, it will failover to the second NN. So here comes the problem, 
> some of the commands(setSafeMode, saveNameSpace, restoreFailedStorage, 
> refreshNodes, setBalancerBandwidth, metaSave) in DFSAdmin are classified as 
> UNCHECKED operations, and when executing these commands in the DFSAdmin 
> command line, they will be sent to a definite NN, no matter it is Active or 
> Standby. This may result in two problems: 
> a. If the first tried NN is standby, and the operation takes effect only on 
> Standby NN, which is not the expected result.
> b. If the operation needs to take effect on both NN, but it takes effect on 
> only one NN. In the future, when there is a NN failover, there may have 
> problems.
> Here I propose the following improvements:
> a. If the command can be classified as one of READ/WRITE/CHECKPOINT/JOURNAL 
> operations, we should classify it clearly.
> b. If the command can not be classified as one of the above four operations, 
> or if the command needs to take effect on both NN, we should send the request 
> to both Active and Standby NNs.
> 2. Refresh protocols: RefreshAuthorizationPolicyProtocol, 
> RefreshUserMappingsProtocol, RefreshUserMappingsProtocol, 
> RefreshCallQueueProtocol
> Commands in this category, including refreshServiceAcl, 
> refreshUserToGroupMapping, refreshSuperUserGroupsConfiguration and 
> refreshCallQueue, are implemented by creating a corresponding RPC proxy and 
> sending the request to remote NN. In the current implementation, these 
> requests will be sent to a definite NN, no matter it is Active or Standby. 
> Here I propose that we sent these requests to both NNs.
> 3. ClientDatanodeProtocol
> Commands in this category are handled correctly, no need to improve.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6430) HTTPFS - Implement XAttr support

2014-06-18 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6430?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036983#comment-14036983
 ] 

Alejandro Abdelnur commented on HDFS-6430:
--

Nice job Yi.

A couple of minor NITs:

* PBHelper.java, this seems like a false change
* HttpFSUtils.java, new createURL(), we should URLEncoder.encode() the name of 
the multivalue parameter as well.

Are testing the case of a FileSystem that does not support xAttrs?  

> HTTPFS - Implement XAttr support
> 
>
> Key: HDFS-6430
> URL: https://issues.apache.org/jira/browse/HDFS-6430
> Project: Hadoop HDFS
>  Issue Type: Task
>Affects Versions: 3.0.0
>Reporter: Yi Liu
>Assignee: Yi Liu
> Fix For: 3.0.0
>
> Attachments: HDFS-6430.1.patch, HDFS-6430.2.patch, HDFS-6430.3.patch, 
> HDFS-6430.patch
>
>
> Add xattr support to HttpFS.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6563) NameNode cannot save fsimage in certain circumstances when snapshots are in use

2014-06-18 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6563?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036982#comment-14036982
 ] 

Hudson commented on HDFS-6563:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5733 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5733/])
HDFS-6563. NameNode cannot save fsimage in certain circumstances when snapshots 
are in use. Contributed by Aaron T. Myers. (atm: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1603712)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormatPBINode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/snapshot/FileWithSnapshotFeature.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestSnapshotBlocksMap.java


> NameNode cannot save fsimage in certain circumstances when snapshots are in 
> use
> ---
>
> Key: HDFS-6563
> URL: https://issues.apache.org/jira/browse/HDFS-6563
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode, snapshots
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
>Priority: Critical
> Fix For: 2.5.0
>
> Attachments: HDFS-6563.patch
>
>
> Checkpoints will start to fail and the NameNode will not be able to manually 
> saveNamespace if the following set of steps occurs:
> # A zero-length file appears in a snapshot
> # That file is later lengthened to include at least one block
> # That file is subsequently deleted from the present file system but remains 
> in the snapshot
> More details in the first comment.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6549) Add support for accessing the NFS gateway from the AIX NFS client

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036977#comment-14036977
 ] 

Hadoop QA commented on HDFS-6549:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651329/HDFS-6549.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs hadoop-hdfs-project/hadoop-hdfs-nfs:

  org.apache.hadoop.hdfs.server.datanode.TestBPOfferService

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7173//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7173//console

This message is automatically generated.

> Add support for accessing the NFS gateway from the AIX NFS client
> -
>
> Key: HDFS-6549
> URL: https://issues.apache.org/jira/browse/HDFS-6549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Attachments: HDFS-6549.patch, HDFS-6549.patch
>
>
> We've identified two issues when trying to access the HDFS NFS Gateway from 
> an AIX NFS client:
> # In the case of COMMITs, the AIX NFS client will always send 4096, or a 
> multiple of the page size, for the offset to be committed, even if fewer 
> bytes than this have ever, or will ever, be written to the file. This will 
> cause a write to a file from the AIX NFS client to hang on close unless the 
> size of that file is a multiple of 4096.
> # In the case of READDIR and READDIRPLUS, the AIX NFS client will send the 
> same cookie verifier for a given directory seemingly forever after that 
> directory is first accessed over NFS, instead of getting a new cookie 
> verifier for every set of incremental readdir calls. This means that if a 
> directory's mtime ever changes, the FS must be unmounted/remounted before 
> readdir calls on that dir from AIX will ever succeed again.
> From my interpretation of RFC-1813, the NFS Gateway is in fact doing the 
> correct thing in both cases, but we can introduce simple changes on the NFS 
> Gateway side to be able to optionally work around these incompatibilities.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6563) NameNode cannot save fsimage in certain circumstances when snapshots are in use

2014-06-18 Thread Aaron T. Myers (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6563?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-6563:
-

   Resolution: Fixed
Fix Version/s: 2.5.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

Thanks again for the review, Jing.

I've just committed this to trunk and branch-2.

> NameNode cannot save fsimage in certain circumstances when snapshots are in 
> use
> ---
>
> Key: HDFS-6563
> URL: https://issues.apache.org/jira/browse/HDFS-6563
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode, snapshots
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
>Priority: Critical
> Fix For: 2.5.0
>
> Attachments: HDFS-6563.patch
>
>
> Checkpoints will start to fail and the NameNode will not be able to manually 
> saveNamespace if the following set of steps occurs:
> # A zero-length file appears in a snapshot
> # That file is later lengthened to include at least one block
> # That file is subsequently deleted from the present file system but remains 
> in the snapshot
> More details in the first comment.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6480) Move waitForReady() from FSDirectory to FSNamesystem

2014-06-18 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036966#comment-14036966
 ] 

Hudson commented on HDFS-6480:
--

FAILURE: Integrated in Hadoop-trunk-Commit #5732 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5732/])
HDFS-6480. Move waitForReady() from FSDirectory to FSNamesystem. Contributed by 
Haohui Mai. (wheat9: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1603705)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/Checkpointer.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SecondaryNameNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCommitBlockSynchronization.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSDirectory.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsLimits.java


> Move waitForReady() from FSDirectory to FSNamesystem
> 
>
> Key: HDFS-6480
> URL: https://issues.apache.org/jira/browse/HDFS-6480
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Fix For: 2.5.0
>
> Attachments: HDFS-6480.000.patch, HDFS-6480.001.patch, 
> HDFS-6480.002.patch, HDFS-6480.003.patch, HDFS-6480.004.patch, 
> HDFS-6480.005.patch
>
>
> Currently FSDirectory implements a barrier in {{waitForReady()}} / 
> {{setReady()}} so that it only serve requests once the FSImage is fully 
> loaded.
> As a part of the effort to evolve {{FSDirectory}} to a class which focuses on 
> implementing the data structure of the namespace, this jira proposes to move 
> the barrier one level higher to {{FSNamesystem}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6563) NameNode cannot save fsimage in certain circumstances when snapshots are in use

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6563?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036960#comment-14036960
 ] 

Hadoop QA commented on HDFS-6563:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651320/HDFS-6563.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7172//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7172//console

This message is automatically generated.

> NameNode cannot save fsimage in certain circumstances when snapshots are in 
> use
> ---
>
> Key: HDFS-6563
> URL: https://issues.apache.org/jira/browse/HDFS-6563
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode, snapshots
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
>Priority: Critical
> Attachments: HDFS-6563.patch
>
>
> Checkpoints will start to fail and the NameNode will not be able to manually 
> saveNamespace if the following set of steps occurs:
> # A zero-length file appears in a snapshot
> # That file is later lengthened to include at least one block
> # That file is subsequently deleted from the present file system but remains 
> in the snapshot
> More details in the first comment.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6312) WebHdfs HA failover is broken on secure clusters

2014-06-18 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6312?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036951#comment-14036951
 ] 

Alejandro Abdelnur commented on HDFS-6312:
--

planning to commit this tomorrow around noon PST.

> WebHdfs HA failover is broken on secure clusters
> 
>
> Key: HDFS-6312
> URL: https://issues.apache.org/jira/browse/HDFS-6312
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>Priority: Blocker
> Attachments: HDFS-6312.attempted.patch
>
>
> When webhdfs does a failover, it blanks out the delegation token.  This will 
> cause subsequent operations against the other NN to acquire a new token.  
> Tasks cannot acquire a token (no kerberos credentials) so jobs will fail.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6557) Move the reference of fsimage to FSNamesystem

2014-06-18 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6557?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-6557:
-

Attachment: HDFS-6557.000.patch

> Move the reference of fsimage to FSNamesystem
> -
>
> Key: HDFS-6557
> URL: https://issues.apache.org/jira/browse/HDFS-6557
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6557.000.patch
>
>
> Per the suggestion from HDFS-6480. {{FSDirectory}} becomes a in-memory data 
> structure so that the reference of fsimage should be moved to 
> {{FSNamesystem}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6557) Move the reference of fsimage to FSNamesystem

2014-06-18 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6557?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-6557:
-

Status: Patch Available  (was: Open)

> Move the reference of fsimage to FSNamesystem
> -
>
> Key: HDFS-6557
> URL: https://issues.apache.org/jira/browse/HDFS-6557
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6557.000.patch
>
>
> Per the suggestion from HDFS-6480. {{FSDirectory}} becomes a in-memory data 
> structure so that the reference of fsimage should be moved to 
> {{FSNamesystem}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6480) Move waitForReady() from FSDirectory to FSNamesystem

2014-06-18 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6480?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-6480:
-

   Resolution: Fixed
Fix Version/s: 2.5.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

I've committed the patch to trunk and branch-2. Thanks [~jingzhao] for the 
review.

> Move waitForReady() from FSDirectory to FSNamesystem
> 
>
> Key: HDFS-6480
> URL: https://issues.apache.org/jira/browse/HDFS-6480
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Fix For: 2.5.0
>
> Attachments: HDFS-6480.000.patch, HDFS-6480.001.patch, 
> HDFS-6480.002.patch, HDFS-6480.003.patch, HDFS-6480.004.patch, 
> HDFS-6480.005.patch
>
>
> Currently FSDirectory implements a barrier in {{waitForReady()}} / 
> {{setReady()}} so that it only serve requests once the FSImage is fully 
> loaded.
> As a part of the effort to evolve {{FSDirectory}} to a class which focuses on 
> implementing the data structure of the namespace, this jira proposes to move 
> the barrier one level higher to {{FSNamesystem}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6519) Document oiv_legacy command

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6519?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036909#comment-14036909
 ] 

Hadoop QA commented on HDFS-6519:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651296/HDFS-6519.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+0 tests included{color}.  The patch appears to be a 
documentation patch that doesn't require tests.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7171//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7171//console

This message is automatically generated.

> Document oiv_legacy command
> ---
>
> Key: HDFS-6519
> URL: https://issues.apache.org/jira/browse/HDFS-6519
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: documentation
>Affects Versions: 2.5.0
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
> Attachments: HDFS-6519.patch
>
>
> HDFS-6293 introduced oiv_legacy command.
> The usage of the command should be included in OfflineImageViewer.apt.vm.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6566) Separate public and private keys in DFSConfigKey

2014-06-18 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6566?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036899#comment-14036899
 ] 

Suresh Srinivas commented on HDFS-6566:
---

It may be good to segregate client and server side configs. 
[~stev...@iseran.com] Has made a case for annotating all the configs as public. 
We should explore that. 

> Separate public and private keys in DFSConfigKey
> 
>
> Key: HDFS-6566
> URL: https://issues.apache.org/jira/browse/HDFS-6566
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Haohui Mai
>
> DFSConfigKey contains both public and private keys. There is no clear 
> clarification on whether a key should be public and stable.
> This jira proposes to separate the public and stable keys in a separate class.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6403) Add metrics for log warnings reported by JVM pauses

2014-06-18 Thread Yongjun Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6403?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036898#comment-14036898
 ] 

Yongjun Zhang commented on HDFS-6403:
-

HI [~ajisakaa], thanks for the review, really appreciate it!


> Add metrics for log warnings reported by JVM pauses
> ---
>
> Key: HDFS-6403
> URL: https://issues.apache.org/jira/browse/HDFS-6403
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: datanode, namenode
>Affects Versions: 2.4.0
>Reporter: Yongjun Zhang
>Assignee: Yongjun Zhang
> Attachments: HDFS-6403.001.patch, HDFS-6403.002.patch, 
> HDFS-6403.003.patch, HDFS-6403.004.patch
>
>
> HADOOP-9618 logs warnings when there are long GC pauses. If this is exposed 
> as a metric, then they can be monitored.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6200) Create a separate jar for hdfs-client

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6200?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036888#comment-14036888
 ] 

Hadoop QA commented on HDFS-6200:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651333/HDFS-6200.007.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:red}-1 javac{color:red}.  The patch appears to cause the build to 
fail.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7174//console

This message is automatically generated.

> Create a separate jar for hdfs-client
> -
>
> Key: HDFS-6200
> URL: https://issues.apache.org/jira/browse/HDFS-6200
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6200.000.patch, HDFS-6200.001.patch, 
> HDFS-6200.002.patch, HDFS-6200.003.patch, HDFS-6200.004.patch, 
> HDFS-6200.005.patch, HDFS-6200.006.patch, HDFS-6200.007.patch
>
>
> Currently the hadoop-hdfs jar contain both the hdfs server and the hdfs 
> client. As discussed in the hdfs-dev mailing list 
> (http://mail-archives.apache.org/mod_mbox/hadoop-hdfs-dev/201404.mbox/browser),
>  downstream projects are forced to bring in additional dependency in order to 
> access hdfs. The additional dependency sometimes can be difficult to manage 
> for projects like Apache Falcon and Apache Oozie.
> This jira proposes to create a new project, hadoop-hdfs-cliient, which 
> contains the client side of the hdfs code. Downstream projects can use this 
> jar instead of the hadoop-hdfs to avoid unnecessary dependency.
> Note that it does not break the compatibility of downstream projects. This is 
> because old downstream projects implicitly depend on hadoop-hdfs-client 
> through the hadoop-hdfs jar.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6492) Support create-time xattrs and atomically setting multiple xattrs

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6492?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036825#comment-14036825
 ] 

Hadoop QA commented on HDFS-6492:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651295/HDFS-6492.003.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 3 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  
org.apache.hadoop.hdfs.tools.offlineEditsViewer.TestOfflineEditsViewer

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7170//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7170//console

This message is automatically generated.

> Support create-time xattrs and atomically setting multiple xattrs
> -
>
> Key: HDFS-6492
> URL: https://issues.apache.org/jira/browse/HDFS-6492
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Andrew Wang
>Assignee: Andrew Wang
> Attachments: HDFS-6492.001.patch, HDFS-6492.002.patch, 
> HDFS-6492.003.patch
>
>
> Ongoing work in HDFS-6134 requires being able to set system namespace 
> extended attributes at create and mkdir time, as well as being able to 
> atomically set multiple xattrs at once. There's currently no need to expose 
> this functionality in the client API, so let's not unless we have to.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6553) Add missing DeprecationDeltas for NFS Kerberos configurations

2014-06-18 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6553?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036816#comment-14036816
 ] 

Hudson commented on HDFS-6553:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5731 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5731/])
HDFS-6553. Add missing DeprecationDeltas for NFS Kerberos configurations. 
Contributed by Stephen Chu (brandonli: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1603677)
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/conf/NfsConfiguration.java
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


> Add missing DeprecationDeltas for NFS Kerberos configurations
> -
>
> Key: HDFS-6553
> URL: https://issues.apache.org/jira/browse/HDFS-6553
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 3.0.0, 2.5.0
>Reporter: Stephen Chu
>Assignee: Stephen Chu
> Fix For: 2.5.0
>
> Attachments: HDFS-6553.patch.001, HDFS-6553.patch.002
>
>
> HDFS-6056 (Clean up NFS config settings) improved NFS configuration naming 
> and added DeprecationDeltas to ensure that the old NFS configuration 
> properties could still be used.
> It's currently missing DeprecationDeltas for _dfs.nfs.keytab.file_ and 
> _dfs.nfs.kerberos.principal_.
> This patch adds those deltas so older configs for secure NFS can still work.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6562) Refactor rename() in FSDirectory

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6562?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036814#comment-14036814
 ] 

Hadoop QA commented on HDFS-6562:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651291/HDFS-6562.000.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:red}-1 findbugs{color}.  The patch appears to introduce 1 new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  
org.apache.hadoop.hdfs.server.namenode.snapshot.TestRenameWithSnapshots

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7169//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7169//artifact/trunk/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7169//console

This message is automatically generated.

> Refactor rename() in FSDirectory
> 
>
> Key: HDFS-6562
> URL: https://issues.apache.org/jira/browse/HDFS-6562
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6562.000.patch
>
>
> Currently there are two variants of {{rename()}} sitting in {{FSDirectory}}. 
> Both implementation shares quite a bit of common code.
> This jira proposes to clean up these two variants and extract the common code.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6480) Move waitForReady() from FSDirectory to FSNamesystem

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036806#comment-14036806
 ] 

Hadoop QA commented on HDFS-6480:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651290/HDFS-6480.005.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 4 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7168//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7168//console

This message is automatically generated.

> Move waitForReady() from FSDirectory to FSNamesystem
> 
>
> Key: HDFS-6480
> URL: https://issues.apache.org/jira/browse/HDFS-6480
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6480.000.patch, HDFS-6480.001.patch, 
> HDFS-6480.002.patch, HDFS-6480.003.patch, HDFS-6480.004.patch, 
> HDFS-6480.005.patch
>
>
> Currently FSDirectory implements a barrier in {{waitForReady()}} / 
> {{setReady()}} so that it only serve requests once the FSImage is fully 
> loaded.
> As a part of the effort to evolve {{FSDirectory}} to a class which focuses on 
> implementing the data structure of the namespace, this jira proposes to move 
> the barrier one level higher to {{FSNamesystem}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-2856) Fix block protocol so that Datanodes don't require root or jsvc

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2856?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036804#comment-14036804
 ] 

Hadoop QA commented on HDFS-2856:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651284/HDFS-2856.2.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 5 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  
org.apache.hadoop.hdfs.server.namenode.ha.TestPipelinesFailover

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7167//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7167//console

This message is automatically generated.

> Fix block protocol so that Datanodes don't require root or jsvc
> ---
>
> Key: HDFS-2856
> URL: https://issues.apache.org/jira/browse/HDFS-2856
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: datanode, security
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Owen O'Malley
>Assignee: Chris Nauroth
> Attachments: Datanode-Security-Design.pdf, 
> Datanode-Security-Design.pdf, Datanode-Security-Design.pdf, 
> HDFS-2856.1.patch, HDFS-2856.2.patch, HDFS-2856.prototype.patch
>
>
> Since we send the block tokens unencrypted to the datanode, we currently 
> start the datanode as root using jsvc and get a secure (< 1024) port.
> If we have the datanode generate a nonce and send it on the connection and 
> the sends an hmac of the nonce back instead of the block token it won't 
> reveal any secrets. Thus, we wouldn't require a secure port and would not 
> require root or jsvc.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6567) Clean up HdfsFileStatus

2014-06-18 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6567?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-6567:
-

Description: As suggested in HDFS-6200, the order of "public final" in 
{{HdfsFileStatus}} is reversed. This jira proposes to fix the order and to make 
the code more consistent.  (was: As suggested in HDFS-6200, the order of 
"public final" in {{HdfsFileStatus}} is reserved. This jira proposes to fix the 
order and to make the code more consistent.)

> Clean up HdfsFileStatus
> ---
>
> Key: HDFS-6567
> URL: https://issues.apache.org/jira/browse/HDFS-6567
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Haohui Mai
>
> As suggested in HDFS-6200, the order of "public final" in {{HdfsFileStatus}} 
> is reversed. This jira proposes to fix the order and to make the code more 
> consistent.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6200) Create a separate jar for hdfs-client

2014-06-18 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6200?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-6200:
-

Attachment: HDFS-6200.007.patch

> Create a separate jar for hdfs-client
> -
>
> Key: HDFS-6200
> URL: https://issues.apache.org/jira/browse/HDFS-6200
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6200.000.patch, HDFS-6200.001.patch, 
> HDFS-6200.002.patch, HDFS-6200.003.patch, HDFS-6200.004.patch, 
> HDFS-6200.005.patch, HDFS-6200.006.patch, HDFS-6200.007.patch
>
>
> Currently the hadoop-hdfs jar contain both the hdfs server and the hdfs 
> client. As discussed in the hdfs-dev mailing list 
> (http://mail-archives.apache.org/mod_mbox/hadoop-hdfs-dev/201404.mbox/browser),
>  downstream projects are forced to bring in additional dependency in order to 
> access hdfs. The additional dependency sometimes can be difficult to manage 
> for projects like Apache Falcon and Apache Oozie.
> This jira proposes to create a new project, hadoop-hdfs-cliient, which 
> contains the client side of the hdfs code. Downstream projects can use this 
> jar instead of the hadoop-hdfs to avoid unnecessary dependency.
> Note that it does not break the compatibility of downstream projects. This is 
> because old downstream projects implicitly depend on hadoop-hdfs-client 
> through the hadoop-hdfs jar.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6567) Clean up HdfsFileStatus

2014-06-18 Thread Haohui Mai (JIRA)
Haohui Mai created HDFS-6567:


 Summary: Clean up HdfsFileStatus
 Key: HDFS-6567
 URL: https://issues.apache.org/jira/browse/HDFS-6567
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai


As suggested in HDFS-6200, the order of "public final" in {{HdfsFileStatus}} is 
reserved. This jira proposes to fix the order and to make the code more 
consistent.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6549) Add support for accessing the NFS gateway from the AIX NFS client

2014-06-18 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036750#comment-14036750
 ] 

Brandon Li commented on HDFS-6549:
--

+1 to the latest patch.

> Add support for accessing the NFS gateway from the AIX NFS client
> -
>
> Key: HDFS-6549
> URL: https://issues.apache.org/jira/browse/HDFS-6549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Attachments: HDFS-6549.patch, HDFS-6549.patch
>
>
> We've identified two issues when trying to access the HDFS NFS Gateway from 
> an AIX NFS client:
> # In the case of COMMITs, the AIX NFS client will always send 4096, or a 
> multiple of the page size, for the offset to be committed, even if fewer 
> bytes than this have ever, or will ever, be written to the file. This will 
> cause a write to a file from the AIX NFS client to hang on close unless the 
> size of that file is a multiple of 4096.
> # In the case of READDIR and READDIRPLUS, the AIX NFS client will send the 
> same cookie verifier for a given directory seemingly forever after that 
> directory is first accessed over NFS, instead of getting a new cookie 
> verifier for every set of incremental readdir calls. This means that if a 
> directory's mtime ever changes, the FS must be unmounted/remounted before 
> readdir calls on that dir from AIX will ever succeed again.
> From my interpretation of RFC-1813, the NFS Gateway is in fact doing the 
> correct thing in both cases, but we can introduce simple changes on the NFS 
> Gateway side to be able to optionally work around these incompatibilities.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6200) Create a separate jar for hdfs-client

2014-06-18 Thread Haohui Mai (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6200?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036751#comment-14036751
 ] 

Haohui Mai commented on HDFS-6200:
--

{quote}
DFSConfigKeys should be public and contain those keys clients are expected to 
use
this is a good time to switch to SLF4J for the logging here, and drop 
commons-logging
HdfsFileStatus gets its "public final" declarations in the wrong order ... 
again, this is a good time to fix it.
JsonUtilClient uses org.mortbay.util.ajax.JSON to parse the json. Thus should 
be replaced by Jackson, so we don't need the mortbay libs on the classpath?
WebHdfsFileSystem also uses mortbay JSON for parsing
{quote}

It might be better to restrict this patch to moving the files only. I plan to 
address them in separate jiras. Filed HDFS-6564, HDFS-6565, HDFS-6566, and 
HDFS-6567 to track them.

> Create a separate jar for hdfs-client
> -
>
> Key: HDFS-6200
> URL: https://issues.apache.org/jira/browse/HDFS-6200
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6200.000.patch, HDFS-6200.001.patch, 
> HDFS-6200.002.patch, HDFS-6200.003.patch, HDFS-6200.004.patch, 
> HDFS-6200.005.patch, HDFS-6200.006.patch
>
>
> Currently the hadoop-hdfs jar contain both the hdfs server and the hdfs 
> client. As discussed in the hdfs-dev mailing list 
> (http://mail-archives.apache.org/mod_mbox/hadoop-hdfs-dev/201404.mbox/browser),
>  downstream projects are forced to bring in additional dependency in order to 
> access hdfs. The additional dependency sometimes can be difficult to manage 
> for projects like Apache Falcon and Apache Oozie.
> This jira proposes to create a new project, hadoop-hdfs-cliient, which 
> contains the client side of the hdfs code. Downstream projects can use this 
> jar instead of the hadoop-hdfs to avoid unnecessary dependency.
> Note that it does not break the compatibility of downstream projects. This is 
> because old downstream projects implicitly depend on hadoop-hdfs-client 
> through the hadoop-hdfs jar.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6566) Separate public and private keys in DFSConfigKey

2014-06-18 Thread Haohui Mai (JIRA)
Haohui Mai created HDFS-6566:


 Summary: Separate public and private keys in DFSConfigKey
 Key: HDFS-6566
 URL: https://issues.apache.org/jira/browse/HDFS-6566
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai


DFSConfigKey contains both public and private keys. There is no clear 
clarification on whether a key should be public and stable.

This jira proposes to separate the public and stable keys in a separate class.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6549) Add support for accessing the NFS gateway from the AIX NFS client

2014-06-18 Thread Aaron T. Myers (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-6549:
-

Attachment: HDFS-6549.patch

Great, thanks. Here's an updated patch which adds an explanation about why AIX 
compat mode should not be enabled in general.

> Add support for accessing the NFS gateway from the AIX NFS client
> -
>
> Key: HDFS-6549
> URL: https://issues.apache.org/jira/browse/HDFS-6549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Attachments: HDFS-6549.patch, HDFS-6549.patch
>
>
> We've identified two issues when trying to access the HDFS NFS Gateway from 
> an AIX NFS client:
> # In the case of COMMITs, the AIX NFS client will always send 4096, or a 
> multiple of the page size, for the offset to be committed, even if fewer 
> bytes than this have ever, or will ever, be written to the file. This will 
> cause a write to a file from the AIX NFS client to hang on close unless the 
> size of that file is a multiple of 4096.
> # In the case of READDIR and READDIRPLUS, the AIX NFS client will send the 
> same cookie verifier for a given directory seemingly forever after that 
> directory is first accessed over NFS, instead of getting a new cookie 
> verifier for every set of incremental readdir calls. This means that if a 
> directory's mtime ever changes, the FS must be unmounted/remounted before 
> readdir calls on that dir from AIX will ever succeed again.
> From my interpretation of RFC-1813, the NFS Gateway is in fact doing the 
> correct thing in both cases, but we can introduce simple changes on the NFS 
> Gateway side to be able to optionally work around these incompatibilities.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6565) Use jackson instead jetty json in hdfs-client

2014-06-18 Thread Haohui Mai (JIRA)
Haohui Mai created HDFS-6565:


 Summary: Use jackson instead jetty json in hdfs-client
 Key: HDFS-6565
 URL: https://issues.apache.org/jira/browse/HDFS-6565
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai


hdfs-client should use Jackson instead of jetty to parse JSON.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6564) Use slf4j instead of common-logging in hdfs-client

2014-06-18 Thread Haohui Mai (JIRA)
Haohui Mai created HDFS-6564:


 Summary: Use slf4j instead of common-logging in hdfs-client
 Key: HDFS-6564
 URL: https://issues.apache.org/jira/browse/HDFS-6564
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai


hdfs-client should depends on slf4j instead of common-logging.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6387) HDFS CLI admin tool for creating & deleting an encryption zone

2014-06-18 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6387?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb updated HDFS-6387:
---

Attachment: HDFS-6387.003.patch

.003 has all of the tests fixed to run correctly after HDFS-6386 (the backend 
of EZs) was checked in.

> HDFS CLI admin tool for creating & deleting an encryption zone
> --
>
> Key: HDFS-6387
> URL: https://issues.apache.org/jira/browse/HDFS-6387
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode, security
>Reporter: Alejandro Abdelnur
>Assignee: Charles Lamb
> Attachments: HDFS-6387.002.patch, HDFS-6387.003.patch, 
> HDFS-6387.1.patch
>
>
> CLI admin tool to create/delete an encryption zone in HDFS.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6563) NameNode cannot save fsimage in certain circumstances when snapshots are in use

2014-06-18 Thread Aaron T. Myers (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6563?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036725#comment-14036725
 ] 

Aaron T. Myers commented on HDFS-6563:
--

Thanks a lot for the quick review, Jing.

> NameNode cannot save fsimage in certain circumstances when snapshots are in 
> use
> ---
>
> Key: HDFS-6563
> URL: https://issues.apache.org/jira/browse/HDFS-6563
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode, snapshots
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
>Priority: Critical
> Attachments: HDFS-6563.patch
>
>
> Checkpoints will start to fail and the NameNode will not be able to manually 
> saveNamespace if the following set of steps occurs:
> # A zero-length file appears in a snapshot
> # That file is later lengthened to include at least one block
> # That file is subsequently deleted from the present file system but remains 
> in the snapshot
> More details in the first comment.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6507) Improve DFSAdmin to support HA cluster better

2014-06-18 Thread Zesheng Wu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036720#comment-14036720
 ] 

Zesheng Wu commented on HDFS-6507:
--

Thanks for feedback [~jingzhao]

I still think that we should send requests to all NNs. If people thinks some 
commands are heavy(just as you mentioned),  they can use the {{-fs}} option to 
specify a NN to do his operation. 
As the current default behavior of the commands(setSafeMode/metaSave, etc) 
doesn't meet our expectation(you also think so),  we should correct it. The way 
you suggested is just a compromise, and it's easier to be friendly for people, 
by not easier to other tools who use the DFSAdmin shell.

How do you think?

> Improve DFSAdmin to support HA cluster better
> -
>
> Key: HDFS-6507
> URL: https://issues.apache.org/jira/browse/HDFS-6507
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: tools
>Affects Versions: 2.4.0
>Reporter: Zesheng Wu
>Assignee: Zesheng Wu
> Attachments: HDFS-6507.1.patch, HDFS-6507.2.patch, HDFS-6507.3.patch, 
> HDFS-6507.4-inprogress.patch, HDFS-6507.4.patch, HDFS-6507.5.patch, 
> HDFS-6507.6.patch, HDFS-6507.7.patch, HDFS-6507.7.patch, HDFS-6507.8.patch
>
>
> Currently, the commands supported in DFSAdmin can be classified into three 
> categories according to the protocol used:
> 1. ClientProtocol
> Commands in this category generally implement by calling the corresponding 
> function of the DFSClient class, and will call the corresponding remote 
> implementation function at the NN side finally. At the NN side, all these 
> operations are classified into five categories: UNCHECKED, READ, WRITE, 
> CHECKPOINT, JOURNAL. Active NN will allow all operations, and Standby NN only 
> allows UNCHECKED operations. In the current implementation of DFSClient, it 
> will connect one NN first, if the first NN is not Active and the operation is 
> not allowed, it will failover to the second NN. So here comes the problem, 
> some of the commands(setSafeMode, saveNameSpace, restoreFailedStorage, 
> refreshNodes, setBalancerBandwidth, metaSave) in DFSAdmin are classified as 
> UNCHECKED operations, and when executing these commands in the DFSAdmin 
> command line, they will be sent to a definite NN, no matter it is Active or 
> Standby. This may result in two problems: 
> a. If the first tried NN is standby, and the operation takes effect only on 
> Standby NN, which is not the expected result.
> b. If the operation needs to take effect on both NN, but it takes effect on 
> only one NN. In the future, when there is a NN failover, there may have 
> problems.
> Here I propose the following improvements:
> a. If the command can be classified as one of READ/WRITE/CHECKPOINT/JOURNAL 
> operations, we should classify it clearly.
> b. If the command can not be classified as one of the above four operations, 
> or if the command needs to take effect on both NN, we should send the request 
> to both Active and Standby NNs.
> 2. Refresh protocols: RefreshAuthorizationPolicyProtocol, 
> RefreshUserMappingsProtocol, RefreshUserMappingsProtocol, 
> RefreshCallQueueProtocol
> Commands in this category, including refreshServiceAcl, 
> refreshUserToGroupMapping, refreshSuperUserGroupsConfiguration and 
> refreshCallQueue, are implemented by creating a corresponding RPC proxy and 
> sending the request to remote NN. In the current implementation, these 
> requests will be sent to a definite NN, no matter it is Active or Standby. 
> Here I propose that we sent these requests to both NNs.
> 3. ClientDatanodeProtocol
> Commands in this category are handled correctly, no need to improve.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6563) NameNode cannot save fsimage in certain circumstances when snapshots are in use

2014-06-18 Thread Aaron T. Myers (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6563?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-6563:
-

Status: Patch Available  (was: Open)

> NameNode cannot save fsimage in certain circumstances when snapshots are in 
> use
> ---
>
> Key: HDFS-6563
> URL: https://issues.apache.org/jira/browse/HDFS-6563
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode, snapshots
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
>Priority: Critical
> Attachments: HDFS-6563.patch
>
>
> Checkpoints will start to fail and the NameNode will not be able to manually 
> saveNamespace if the following set of steps occurs:
> # A zero-length file appears in a snapshot
> # That file is later lengthened to include at least one block
> # That file is subsequently deleted from the present file system but remains 
> in the snapshot
> More details in the first comment.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6563) NameNode cannot save fsimage in certain circumstances when snapshots are in use

2014-06-18 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6563?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036713#comment-14036713
 ] 

Jing Zhao commented on HDFS-6563:
-

Thanks for the fix, [~atm]! +1 pending Jenkins.

> NameNode cannot save fsimage in certain circumstances when snapshots are in 
> use
> ---
>
> Key: HDFS-6563
> URL: https://issues.apache.org/jira/browse/HDFS-6563
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode, snapshots
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
>Priority: Critical
> Attachments: HDFS-6563.patch
>
>
> Checkpoints will start to fail and the NameNode will not be able to manually 
> saveNamespace if the following set of steps occurs:
> # A zero-length file appears in a snapshot
> # That file is later lengthened to include at least one block
> # That file is subsequently deleted from the present file system but remains 
> in the snapshot
> More details in the first comment.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6563) NameNode cannot save fsimage in certain circumstances when snapshots are in use

2014-06-18 Thread Aaron T. Myers (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6563?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-6563:
-

Attachment: HDFS-6563.patch

Attached patch addresses the issue in two different ways:

# Don't set the blocks to {{null}} when cleaning up those blocks, but instead 
set them to {{BlockInfo.EMPTY_ARRAY}} as is done elsewhere in the code.
# Make the FSImage saving code be tolerant of {{null}} blocks on INodeFiles.

Either one of these by itself would be sufficient to address this issue, but I 
figure it's a good idea to do both so as to better future-proof this code 
against possible accidental goofs.

Please review.

> NameNode cannot save fsimage in certain circumstances when snapshots are in 
> use
> ---
>
> Key: HDFS-6563
> URL: https://issues.apache.org/jira/browse/HDFS-6563
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode, snapshots
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
>Priority: Critical
> Attachments: HDFS-6563.patch
>
>
> Checkpoints will start to fail and the NameNode will not be able to manually 
> saveNamespace if the following set of steps occurs:
> # A zero-length file appears in a snapshot
> # That file is later lengthened to include at least one block
> # That file is subsequently deleted from the present file system but remains 
> in the snapshot
> More details in the first comment.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6563) NameNode cannot save fsimage in certain circumstances when snapshots are in use

2014-06-18 Thread Aaron T. Myers (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6563?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036702#comment-14036702
 ] 

Aaron T. Myers commented on HDFS-6563:
--

I've filed this as critical for now, but if folks think this should be a 
blocker I'm fine raising the priority.

Though the issue is fairly critical, the bug is fairly straightforward. In 
{{FSImageFormatPBINode#save(OutputStream, INodeFile)}} we have the following 
code:

{code}
for (Block block : n.getBlocks()) {
  b.addBlocks(PBHelper.convert(block));
}
{code}

Perhaps not obviously, this assumes that {{n.getBlocks()}} will never return 
{{null}}. However, this is possible in the above-described scenario because of 
this code in {{FileWithSnapshotFeature#collectBlocksBeyondMax}}:

{code}
final BlockInfo[] newBlocks;
if (n == 0) {
  newBlocks = null;
} else {
  newBlocks = new BlockInfo[n];
  System.arraycopy(oldBlocks, 0, newBlocks, 0, n);
}

// set new blocks
file.setBlocks(newBlocks);
{code}

When attempting to save an fsimage after this code has been run, errors like 
the following will appear in the logs:

{noformat}
2014-06-18 16:55:11,295 ERROR namenode.FSImage (FSImage.java:run(988)) - Unable 
to save image for 
/home/atm/src/apache/hadoop.git/src/hadoop-hdfs-project/hadoop-hdfs/build/test/data/dfs/name1
java.lang.NullPointerException
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Saver.save(FSImageFormatPBINode.java:537)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Saver.save(FSImageFormatPBINode.java:518)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Saver.serializeINodeSection(FSImageFormatPBINode.java:491)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Saver.saveInodes(FSImageFormatProtobuf.java:412)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Saver.saveInternal(FSImageFormatProtobuf.java:457)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Saver.save(FSImageFormatProtobuf.java:393)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImage(FSImage.java:931)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage$FSImageSaver.run(FSImage.java:982)
at java.lang.Thread.run(Thread.java:724)
2014-06-18 16:55:11,295 ERROR namenode.FSImage (FSImage.java:run(988)) - Unable 
to save image for 
/home/atm/src/apache/hadoop.git/src/hadoop-hdfs-project/hadoop-hdfs/build/test/data/dfs/name2
java.lang.NullPointerException
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Saver.save(FSImageFormatPBINode.java:537)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Saver.save(FSImageFormatPBINode.java:518)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Saver.serializeINodeSection(FSImageFormatPBINode.java:491)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Saver.saveInodes(FSImageFormatProtobuf.java:412)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Saver.saveInternal(FSImageFormatProtobuf.java:457)
at 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatProtobuf$Saver.save(FSImageFormatProtobuf.java:393)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImage(FSImage.java:931)
at 
org.apache.hadoop.hdfs.server.namenode.FSImage$FSImageSaver.run(FSImage.java:982)
at java.lang.Thread.run(Thread.java:724)
2014-06-18 16:55:11,297 ERROR common.Storage 
(NNStorage.java:reportErrorsOnDirectory(808)) - Error reported on storage 
directory Storage Directory 
/home/atm/src/apache/hadoop.git/src/hadoop-hdfs-project/hadoop-hdfs/build/test/data/dfs/name1
2014-06-18 16:55:11,297 WARN  common.Storage 
(NNStorage.java:reportErrorsOnDirectory(813)) - About to remove corresponding 
storage: 
/home/atm/src/apache/hadoop.git/src/hadoop-hdfs-project/hadoop-hdfs/build/test/data/dfs/name1
2014-06-18 16:55:11,297 ERROR common.Storage 
(NNStorage.java:reportErrorsOnDirectory(808)) - Error reported on storage 
directory Storage Directory 
/home/atm/src/apache/hadoop.git/src/hadoop-hdfs-project/hadoop-hdfs/build/test/data/dfs/name2
2014-06-18 16:55:11,297 WARN  common.Storage 
(NNStorage.java:reportErrorsOnDirectory(813)) - About to remove corresponding 
storage: 
/home/atm/src/apache/hadoop.git/src/hadoop-hdfs-project/hadoop-hdfs/build/test/data/dfs/name2
{noformat}

> NameNode cannot save fsimage in certain circumstances when snapshots are in 
> use
> ---
>
> Key: HDFS-6563
> URL: https://issues.apache.org/jira/browse/HDFS-6563
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode, snapshots
>Affect

[jira] [Commented] (HDFS-6200) Create a separate jar for hdfs-client

2014-06-18 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6200?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036700#comment-14036700
 ] 

Steve Loughran commented on HDFS-6200:
--

# DFSConfigKeys should be public and contain those keys clients are expected to 
use
# this is a good time to switch to SLF4J for the logging here, and drop 
commons-logging
# HdfsFileStatus gets its "public final" declarations in the wrong order ... 
again, this is a good time to fix it.

# JsonUtilClient uses {{org.mortbay.util.ajax.JSON}} to parse the json. Thus 
should be replaced by Jackson, so we don't need the mortbay libs on the 
classpath?
# {{WebHdfsFileSystem}} also uses mortbay JSON for parsing

> Create a separate jar for hdfs-client
> -
>
> Key: HDFS-6200
> URL: https://issues.apache.org/jira/browse/HDFS-6200
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6200.000.patch, HDFS-6200.001.patch, 
> HDFS-6200.002.patch, HDFS-6200.003.patch, HDFS-6200.004.patch, 
> HDFS-6200.005.patch, HDFS-6200.006.patch
>
>
> Currently the hadoop-hdfs jar contain both the hdfs server and the hdfs 
> client. As discussed in the hdfs-dev mailing list 
> (http://mail-archives.apache.org/mod_mbox/hadoop-hdfs-dev/201404.mbox/browser),
>  downstream projects are forced to bring in additional dependency in order to 
> access hdfs. The additional dependency sometimes can be difficult to manage 
> for projects like Apache Falcon and Apache Oozie.
> This jira proposes to create a new project, hadoop-hdfs-cliient, which 
> contains the client side of the hdfs code. Downstream projects can use this 
> jar instead of the hadoop-hdfs to avoid unnecessary dependency.
> Note that it does not break the compatibility of downstream projects. This is 
> because old downstream projects implicitly depend on hadoop-hdfs-client 
> through the hadoop-hdfs jar.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6244) Make Trash Interval configurable for each of the namespaces

2014-06-18 Thread Siqi Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6244?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036698#comment-14036698
 ] 

Siqi Li commented on HDFS-6244:
---

can anyone give me some review for this

> Make Trash Interval configurable for each of the namespaces
> ---
>
> Key: HDFS-6244
> URL: https://issues.apache.org/jira/browse/HDFS-6244
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Affects Versions: 2.0.5-alpha
>Reporter: Siqi Li
>Assignee: Siqi Li
> Attachments: HDFS-6244.v1.patch, HDFS-6244.v2.patch, 
> HDFS-6244.v3.patch
>
>
> Somehow we need to avoid the cluster filling up.
> One solution is to have a different trash policy per namespace. However, if 
> we can simply make the property configurable per namespace, then the same 
> config can be rolled everywhere and we'd be done. This seems simple enough.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6563) NameNode cannot save fsimage in certain circumstances when snapshots are in use

2014-06-18 Thread Aaron T. Myers (JIRA)
Aaron T. Myers created HDFS-6563:


 Summary: NameNode cannot save fsimage in certain circumstances 
when snapshots are in use
 Key: HDFS-6563
 URL: https://issues.apache.org/jira/browse/HDFS-6563
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode, snapshots
Affects Versions: 2.4.0
Reporter: Aaron T. Myers
Assignee: Aaron T. Myers
Priority: Critical


Checkpoints will start to fail and the NameNode will not be able to manually 
saveNamespace if the following set of steps occurs:

# A zero-length file appears in a snapshot
# That file is later lengthened to include at least one block
# That file is subsequently deleted from the present file system but remains in 
the snapshot

More details in the first comment.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6553) Add missing DeprecationDeltas for NFS Kerberos configurations

2014-06-18 Thread Stephen Chu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6553?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036683#comment-14036683
 ] 

Stephen Chu commented on HDFS-6553:
---

Thanks, Brandon :)

> Add missing DeprecationDeltas for NFS Kerberos configurations
> -
>
> Key: HDFS-6553
> URL: https://issues.apache.org/jira/browse/HDFS-6553
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 3.0.0, 2.5.0
>Reporter: Stephen Chu
>Assignee: Stephen Chu
> Fix For: 2.5.0
>
> Attachments: HDFS-6553.patch.001, HDFS-6553.patch.002
>
>
> HDFS-6056 (Clean up NFS config settings) improved NFS configuration naming 
> and added DeprecationDeltas to ensure that the old NFS configuration 
> properties could still be used.
> It's currently missing DeprecationDeltas for _dfs.nfs.keytab.file_ and 
> _dfs.nfs.kerberos.principal_.
> This patch adds those deltas so older configs for secure NFS can still work.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6553) Add missing DeprecationDeltas for NFS Kerberos configurations

2014-06-18 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6553?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036681#comment-14036681
 ] 

Brandon Li commented on HDFS-6553:
--

I've committed the patch. Thanks! 

> Add missing DeprecationDeltas for NFS Kerberos configurations
> -
>
> Key: HDFS-6553
> URL: https://issues.apache.org/jira/browse/HDFS-6553
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 3.0.0, 2.5.0
>Reporter: Stephen Chu
>Assignee: Stephen Chu
> Fix For: 2.5.0
>
> Attachments: HDFS-6553.patch.001, HDFS-6553.patch.002
>
>
> HDFS-6056 (Clean up NFS config settings) improved NFS configuration naming 
> and added DeprecationDeltas to ensure that the old NFS configuration 
> properties could still be used.
> It's currently missing DeprecationDeltas for _dfs.nfs.keytab.file_ and 
> _dfs.nfs.kerberos.principal_.
> This patch adds those deltas so older configs for secure NFS can still work.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6553) Add missing DeprecationDeltas for NFS Kerberos configurations

2014-06-18 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6553?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-6553:
-

  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

> Add missing DeprecationDeltas for NFS Kerberos configurations
> -
>
> Key: HDFS-6553
> URL: https://issues.apache.org/jira/browse/HDFS-6553
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 3.0.0, 2.5.0
>Reporter: Stephen Chu
>Assignee: Stephen Chu
> Fix For: 2.5.0
>
> Attachments: HDFS-6553.patch.001, HDFS-6553.patch.002
>
>
> HDFS-6056 (Clean up NFS config settings) improved NFS configuration naming 
> and added DeprecationDeltas to ensure that the old NFS configuration 
> properties could still be used.
> It's currently missing DeprecationDeltas for _dfs.nfs.keytab.file_ and 
> _dfs.nfs.kerberos.principal_.
> This patch adds those deltas so older configs for secure NFS can still work.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6553) Add missing DeprecationDeltas for NFS Kerberos configurations

2014-06-18 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6553?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-6553:
-

Fix Version/s: 2.5.0

> Add missing DeprecationDeltas for NFS Kerberos configurations
> -
>
> Key: HDFS-6553
> URL: https://issues.apache.org/jira/browse/HDFS-6553
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 3.0.0, 2.5.0
>Reporter: Stephen Chu
>Assignee: Stephen Chu
> Fix For: 2.5.0
>
> Attachments: HDFS-6553.patch.001, HDFS-6553.patch.002
>
>
> HDFS-6056 (Clean up NFS config settings) improved NFS configuration naming 
> and added DeprecationDeltas to ensure that the old NFS configuration 
> properties could still be used.
> It's currently missing DeprecationDeltas for _dfs.nfs.keytab.file_ and 
> _dfs.nfs.kerberos.principal_.
> This patch adds those deltas so older configs for secure NFS can still work.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6553) Add missing DeprecationDeltas for NFS Kerberos configurations

2014-06-18 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6553?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036670#comment-14036670
 ] 

Brandon Li commented on HDFS-6553:
--

Sure.

> Add missing DeprecationDeltas for NFS Kerberos configurations
> -
>
> Key: HDFS-6553
> URL: https://issues.apache.org/jira/browse/HDFS-6553
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 3.0.0, 2.5.0
>Reporter: Stephen Chu
>Assignee: Stephen Chu
> Attachments: HDFS-6553.patch.001, HDFS-6553.patch.002
>
>
> HDFS-6056 (Clean up NFS config settings) improved NFS configuration naming 
> and added DeprecationDeltas to ensure that the old NFS configuration 
> properties could still be used.
> It's currently missing DeprecationDeltas for _dfs.nfs.keytab.file_ and 
> _dfs.nfs.kerberos.principal_.
> This patch adds those deltas so older configs for secure NFS can still work.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6549) Add support for accessing the NFS gateway from the AIX NFS client

2014-06-18 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036665#comment-14036665
 ] 

Brandon Li commented on HDFS-6549:
--

Yes. 
+1 once the doc is updated. 

> Add support for accessing the NFS gateway from the AIX NFS client
> -
>
> Key: HDFS-6549
> URL: https://issues.apache.org/jira/browse/HDFS-6549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Attachments: HDFS-6549.patch
>
>
> We've identified two issues when trying to access the HDFS NFS Gateway from 
> an AIX NFS client:
> # In the case of COMMITs, the AIX NFS client will always send 4096, or a 
> multiple of the page size, for the offset to be committed, even if fewer 
> bytes than this have ever, or will ever, be written to the file. This will 
> cause a write to a file from the AIX NFS client to hang on close unless the 
> size of that file is a multiple of 4096.
> # In the case of READDIR and READDIRPLUS, the AIX NFS client will send the 
> same cookie verifier for a given directory seemingly forever after that 
> directory is first accessed over NFS, instead of getting a new cookie 
> verifier for every set of incremental readdir calls. This means that if a 
> directory's mtime ever changes, the FS must be unmounted/remounted before 
> readdir calls on that dir from AIX will ever succeed again.
> From my interpretation of RFC-1813, the NFS Gateway is in fact doing the 
> correct thing in both cases, but we can introduce simple changes on the NFS 
> Gateway side to be able to optionally work around these incompatibilities.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6558) Missing '\n' in the description of dfsadmin -rollingUpgrade

2014-06-18 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6558?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036627#comment-14036627
 ] 

Akira AJISAKA commented on HDFS-6558:
-

The patch is just to fix the formatting, so new tests are not needed.

> Missing '\n' in the description of dfsadmin -rollingUpgrade
> ---
>
> Key: HDFS-6558
> URL: https://issues.apache.org/jira/browse/HDFS-6558
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Affects Versions: 2.4.0
>Reporter: Akira AJISAKA
>Assignee: Chen He
>Priority: Trivial
>  Labels: newbie
> Attachments: HDFS-6558.patch
>
>
> In DFSAdmin.java, '\n' should be added at the end of the line
> {code}
> + "   prepare: prepare a new rolling upgrade."
> {code}
> to clean up the following help message.
> {code}
> $ hdfs dfsadmin -help rollingUpgrade
> -rollingUpgrade []:
>  query: query the current rolling upgrade status.
>prepare: prepare a new rolling upgrade.  finalize: finalize the current 
> rolling upgrade.
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6558) Missing '\n' in the description of dfsadmin -rollingUpgrade

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6558?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036624#comment-14036624
 ] 

Hadoop QA commented on HDFS-6558:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651231/HDFS-6558.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7165//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7165//console

This message is automatically generated.

> Missing '\n' in the description of dfsadmin -rollingUpgrade
> ---
>
> Key: HDFS-6558
> URL: https://issues.apache.org/jira/browse/HDFS-6558
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Affects Versions: 2.4.0
>Reporter: Akira AJISAKA
>Assignee: Chen He
>Priority: Trivial
>  Labels: newbie
> Attachments: HDFS-6558.patch
>
>
> In DFSAdmin.java, '\n' should be added at the end of the line
> {code}
> + "   prepare: prepare a new rolling upgrade."
> {code}
> to clean up the following help message.
> {code}
> $ hdfs dfsadmin -help rollingUpgrade
> -rollingUpgrade []:
>  query: query the current rolling upgrade status.
>prepare: prepare a new rolling upgrade.  finalize: finalize the current 
> rolling upgrade.
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HDFS-6391) Get the Key/IV from the NameNode for encrypted files in DFSClient

2014-06-18 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang reassigned HDFS-6391:
-

Assignee: Andrew Wang  (was: Charles Lamb)

> Get the Key/IV from the NameNode for encrypted files in DFSClient
> -
>
> Key: HDFS-6391
> URL: https://issues.apache.org/jira/browse/HDFS-6391
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode, security
>Reporter: Alejandro Abdelnur
>Assignee: Andrew Wang
> Attachments: HDFS-6391.1.patch
>
>
> When creating/opening and encrypted file, the DFSClient should get the 
> encryption key material and the IV for the file in the create/open RPC call.
> HDFS admin users would never get key material/IV on encrypted files 
> create/open.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6134) Transparent data at rest encryption

2014-06-18 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036617#comment-14036617
 ] 

Alejandro Abdelnur commented on HDFS-6134:
--

[~lmccay], thanks for following up.

The proposed approach would be applicable using different cipher modes than CTR 
(ie CBC and XTR, even GCM if we handle the offset correction changes because of 
the GCM tag). In all cases, it would enable keeping the EZ keyVersion materials 
unknown to HDFS and clients and exposing the DEK for the files being accessed 
to the client accessing the file only. 

In the case of CTR, the proposed approach also helps avoiding the IV reuse 
scenario.

Trying to answer your questions:

bq. how does the keyprovider know what EZ key to use - is it the key that is 
referenced by the keyVersionName?

{code}
  // creates EDEK using specified version name
  KeyVersion generateEncryptedKey(String versionName, byte[] iv) 

  // receives EDEK returns DEK using specified version name
  KeyVersion decryptEncryptedKey(String versionName, byte[] iv, KeyVersion 
encryptedKey) 
{code}

The callers of both methods have the versionName at hand

bq. how do we key HDFS clients from asking for the EZ key - if it is stored by 
the passed in keyVersionName?

The file iNode will store (EZ-keyVersionName, IV, EDEK), that info is passed to 
the client on create()/open(). Using that info, the client can go to the 
KeyProvider to get the DEK for the file.

bq. will this require special access control protection for EZ keys?

KeyProviders could implement special control. For example KMS allows, via ACLs, 
to get KeyVersion without key material. This will effectively prevent EZ keys 
from leaving the KMS.

bq. would the unique DEK be stored in the provider as well or only in the 
extended attributes of the file? if stored in the provider what is the 
keyVersionName for it?

the unique EDEKs (encrypted with the EZ keyversion) are not stored by the 
KeyProvider but in the xAttr of the file.


> Transparent data at rest encryption
> ---
>
> Key: HDFS-6134
> URL: https://issues.apache.org/jira/browse/HDFS-6134
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: security
>Affects Versions: 2.3.0
>Reporter: Alejandro Abdelnur
>Assignee: Alejandro Abdelnur
> Attachments: HDFSDataAtRestEncryption.pdf
>
>
> Because of privacy and security regulations, for many industries, sensitive 
> data at rest must be in encrypted form. For example: the health­care industry 
> (HIPAA regulations), the card payment industry (PCI DSS regulations) or the 
> US government (FISMA regulations).
> This JIRA aims to provide a mechanism to encrypt HDFS data at rest that can 
> be used transparently by any application accessing HDFS via Hadoop Filesystem 
> Java API, Hadoop libhdfs C library, or WebHDFS REST API.
> The resulting implementation should be able to be used in compliance with 
> different regulation requirements.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6559) Fix wrong option "dfsadmin -rollingUpgrade start" in the document

2014-06-18 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036619#comment-14036619
 ] 

Akira AJISAKA commented on HDFS-6559:
-

Now Jenkins says +1, there's no need to revert. Thanks [~arpitagarwal] for the 
review!

> Fix wrong option "dfsadmin -rollingUpgrade start" in the document
> -
>
> Key: HDFS-6559
> URL: https://issues.apache.org/jira/browse/HDFS-6559
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.4.0
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>Priority: Minor
>  Labels: newbie
> Fix For: 3.0.0, 2.5.0
>
> Attachments: HDFS-6559.patch
>
>
> In HdfsRollingUpgrade.xml,
> {code}
>   hdfs dfsadmin -rollingUpgrade 
> {code}
> should be
> {code}
>   hdfs dfsadmin -rollingUpgrade 
> 
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6200) Create a separate jar for hdfs-client

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6200?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036611#comment-14036611
 ] 

Hadoop QA commented on HDFS-6200:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651236/HDFS-6200.006.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs hadoop-hdfs-project/hadoop-hdfs-client 
hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal:

  org.apache.hadoop.hdfs.web.TestWebHDFSAcl

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7164//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7164//console

This message is automatically generated.

> Create a separate jar for hdfs-client
> -
>
> Key: HDFS-6200
> URL: https://issues.apache.org/jira/browse/HDFS-6200
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6200.000.patch, HDFS-6200.001.patch, 
> HDFS-6200.002.patch, HDFS-6200.003.patch, HDFS-6200.004.patch, 
> HDFS-6200.005.patch, HDFS-6200.006.patch
>
>
> Currently the hadoop-hdfs jar contain both the hdfs server and the hdfs 
> client. As discussed in the hdfs-dev mailing list 
> (http://mail-archives.apache.org/mod_mbox/hadoop-hdfs-dev/201404.mbox/browser),
>  downstream projects are forced to bring in additional dependency in order to 
> access hdfs. The additional dependency sometimes can be difficult to manage 
> for projects like Apache Falcon and Apache Oozie.
> This jira proposes to create a new project, hadoop-hdfs-cliient, which 
> contains the client side of the hdfs code. Downstream projects can use this 
> jar instead of the hadoop-hdfs to avoid unnecessary dependency.
> Note that it does not break the compatibility of downstream projects. This is 
> because old downstream projects implicitly depend on hadoop-hdfs-client 
> through the hadoop-hdfs jar.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6519) Document oiv_legacy command

2014-06-18 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6519?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-6519:


Attachment: HDFS-6519.patch

* Add a document of oiv_legacy in HdfsImageViewer.apt.vm.
* Add a new property "dfs.namenode.legacy-oiv-image.dir" in hdfs-default.xml

> Document oiv_legacy command
> ---
>
> Key: HDFS-6519
> URL: https://issues.apache.org/jira/browse/HDFS-6519
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: documentation
>Affects Versions: 2.5.0
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
> Attachments: HDFS-6519.patch
>
>
> HDFS-6293 introduced oiv_legacy command.
> The usage of the command should be included in OfflineImageViewer.apt.vm.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6519) Document oiv_legacy command

2014-06-18 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6519?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-6519:


Target Version/s: 2.5.0
  Status: Patch Available  (was: Open)

> Document oiv_legacy command
> ---
>
> Key: HDFS-6519
> URL: https://issues.apache.org/jira/browse/HDFS-6519
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: documentation
>Affects Versions: 2.5.0
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
> Attachments: HDFS-6519.patch
>
>
> HDFS-6293 introduced oiv_legacy command.
> The usage of the command should be included in OfflineImageViewer.apt.vm.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6492) Support create-time xattrs and atomically setting multiple xattrs

2014-06-18 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6492?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-6492:
--

Attachment: HDFS-6492.003.patch

Thanks for the reviews Charles and Colin, this patch addresses your review 
comments.

> Support create-time xattrs and atomically setting multiple xattrs
> -
>
> Key: HDFS-6492
> URL: https://issues.apache.org/jira/browse/HDFS-6492
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Andrew Wang
>Assignee: Andrew Wang
> Attachments: HDFS-6492.001.patch, HDFS-6492.002.patch, 
> HDFS-6492.003.patch
>
>
> Ongoing work in HDFS-6134 requires being able to set system namespace 
> extended attributes at create and mkdir time, as well as being able to 
> atomically set multiple xattrs at once. There's currently no need to expose 
> this functionality in the client API, so let's not unless we have to.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6562) Refactor rename() in FSDirectory

2014-06-18 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6562?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-6562:
-

Attachment: HDFS-6562.000.patch

> Refactor rename() in FSDirectory
> 
>
> Key: HDFS-6562
> URL: https://issues.apache.org/jira/browse/HDFS-6562
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6562.000.patch
>
>
> Currently there are two variants of {{rename()}} sitting in {{FSDirectory}}. 
> Both implementation shares quite a bit of common code.
> This jira proposes to clean up these two variants and extract the common code.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6562) Refactor rename() in FSDirectory

2014-06-18 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6562?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-6562:
-

Status: Patch Available  (was: Open)

> Refactor rename() in FSDirectory
> 
>
> Key: HDFS-6562
> URL: https://issues.apache.org/jira/browse/HDFS-6562
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6562.000.patch
>
>
> Currently there are two variants of {{rename()}} sitting in {{FSDirectory}}. 
> Both implementation shares quite a bit of common code.
> This jira proposes to clean up these two variants and extract the common code.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6562) Refactor rename() in FSDirectory

2014-06-18 Thread Haohui Mai (JIRA)
Haohui Mai created HDFS-6562:


 Summary: Refactor rename() in FSDirectory
 Key: HDFS-6562
 URL: https://issues.apache.org/jira/browse/HDFS-6562
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Haohui Mai
Assignee: Haohui Mai
 Attachments: HDFS-6562.000.patch

Currently there are two variants of {{rename()}} sitting in {{FSDirectory}}. 
Both implementation shares quite a bit of common code.

This jira proposes to clean up these two variants and extract the common code.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6480) Move waitForReady() from FSDirectory to FSNamesystem

2014-06-18 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6480?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-6480:
-

Attachment: HDFS-6480.005.patch

> Move waitForReady() from FSDirectory to FSNamesystem
> 
>
> Key: HDFS-6480
> URL: https://issues.apache.org/jira/browse/HDFS-6480
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6480.000.patch, HDFS-6480.001.patch, 
> HDFS-6480.002.patch, HDFS-6480.003.patch, HDFS-6480.004.patch, 
> HDFS-6480.005.patch
>
>
> Currently FSDirectory implements a barrier in {{waitForReady()}} / 
> {{setReady()}} so that it only serve requests once the FSImage is fully 
> loaded.
> As a part of the effort to evolve {{FSDirectory}} to a class which focuses on 
> implementing the data structure of the namespace, this jira proposes to move 
> the barrier one level higher to {{FSNamesystem}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6549) Add support for accessing the NFS gateway from the AIX NFS client

2014-06-18 Thread Aaron T. Myers (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036593#comment-14036593
 ] 

Aaron T. Myers commented on HDFS-6549:
--

All makes sense. So are you good with this patch if I update the NFS user doc 
to explain why AIX compatibility mode is  not recommended in general?

> Add support for accessing the NFS gateway from the AIX NFS client
> -
>
> Key: HDFS-6549
> URL: https://issues.apache.org/jira/browse/HDFS-6549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Attachments: HDFS-6549.patch
>
>
> We've identified two issues when trying to access the HDFS NFS Gateway from 
> an AIX NFS client:
> # In the case of COMMITs, the AIX NFS client will always send 4096, or a 
> multiple of the page size, for the offset to be committed, even if fewer 
> bytes than this have ever, or will ever, be written to the file. This will 
> cause a write to a file from the AIX NFS client to hang on close unless the 
> size of that file is a multiple of 4096.
> # In the case of READDIR and READDIRPLUS, the AIX NFS client will send the 
> same cookie verifier for a given directory seemingly forever after that 
> directory is first accessed over NFS, instead of getting a new cookie 
> verifier for every set of incremental readdir calls. This means that if a 
> directory's mtime ever changes, the FS must be unmounted/remounted before 
> readdir calls on that dir from AIX will ever succeed again.
> From my interpretation of RFC-1813, the NFS Gateway is in fact doing the 
> correct thing in both cases, but we can introduce simple changes on the NFS 
> Gateway side to be able to optionally work around these incompatibilities.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6549) Add support for accessing the NFS gateway from the AIX NFS client

2014-06-18 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036589#comment-14036589
 ] 

Brandon Li commented on HDFS-6549:
--

Guess AIX NFS server doesn't put mtime in the verifier and uses it in a 
different way.  :-)

The patch should be able to fix the AIX NFS client problems. The only thing is 
that, the aix.compatibility.mode should not be recommended for non AIX client 
for the following reasons:
1. NFS client considers it's safe to clear the buffer cache as long as the data 
is committed. In aix.compatibility.mode, NFS could returns commit success even 
when the data is not received. 
2. Many NFS clients restart directory listing once it encounters verifier 
mismatch error. In aix.compatibility.mode, client may not get the most recent 
directory content.

> Add support for accessing the NFS gateway from the AIX NFS client
> -
>
> Key: HDFS-6549
> URL: https://issues.apache.org/jira/browse/HDFS-6549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Attachments: HDFS-6549.patch
>
>
> We've identified two issues when trying to access the HDFS NFS Gateway from 
> an AIX NFS client:
> # In the case of COMMITs, the AIX NFS client will always send 4096, or a 
> multiple of the page size, for the offset to be committed, even if fewer 
> bytes than this have ever, or will ever, be written to the file. This will 
> cause a write to a file from the AIX NFS client to hang on close unless the 
> size of that file is a multiple of 4096.
> # In the case of READDIR and READDIRPLUS, the AIX NFS client will send the 
> same cookie verifier for a given directory seemingly forever after that 
> directory is first accessed over NFS, instead of getting a new cookie 
> verifier for every set of incremental readdir calls. This means that if a 
> directory's mtime ever changes, the FS must be unmounted/remounted before 
> readdir calls on that dir from AIX will ever succeed again.
> From my interpretation of RFC-1813, the NFS Gateway is in fact doing the 
> correct thing in both cases, but we can introduce simple changes on the NFS 
> Gateway side to be able to optionally work around these incompatibilities.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6480) Move waitForReady() from FSDirectory to FSNamesystem

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036583#comment-14036583
 ] 

Hadoop QA commented on HDFS-6480:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651283/HDFS-6480.004.patch
  against trunk revision .

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7166//console

This message is automatically generated.

> Move waitForReady() from FSDirectory to FSNamesystem
> 
>
> Key: HDFS-6480
> URL: https://issues.apache.org/jira/browse/HDFS-6480
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6480.000.patch, HDFS-6480.001.patch, 
> HDFS-6480.002.patch, HDFS-6480.003.patch, HDFS-6480.004.patch
>
>
> Currently FSDirectory implements a barrier in {{waitForReady()}} / 
> {{setReady()}} so that it only serve requests once the FSImage is fully 
> loaded.
> As a part of the effort to evolve {{FSDirectory}} to a class which focuses on 
> implementing the data structure of the namespace, this jira proposes to move 
> the barrier one level higher to {{FSNamesystem}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-2856) Fix block protocol so that Datanodes don't require root or jsvc

2014-06-18 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2856?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-2856:


Attachment: HDFS-2856.2.patch

I'm attaching v2 of the patch.  This has the following changes since last time:
# Added new tests for the balancer with SASL on DataTransferProtocol.  This 
helped me find a bug with the balancer passing around the incorrect datanode ID 
(the source instead of the destination), so I fixed that.
# Removed TODO for inclusion of block pool ID in the SASL handshake.  I already 
include the token identifier, which contains the block pool ID as a component, 
so it's not necessary to add block pool ID again.
# Removed the client-generated timestamp from the SASL handshake.  The original 
intention of the timestamp was to make it harder for a man in the middle to 
replay the message.  The server side would have checked elapsed time since the 
timestamp and rejected the request if it was beyond a threshold.  However, the 
SASL DIGEST-MD5 handshake already protects against this, because the server 
initiates a random challenge at the start of any new connection.  It's highly 
likely that the challenge will be unique across different connection attempts, 
and thus a replayed message is highly likely to be rejected.  The timestamp 
wouldn't provide any additional benefit.
# Removed datanode ID from the SASL handshake.  This had been intended to 
protect against a man in the middle rerouting a message to a different 
datanode.  As described above, SASL DIGEST-MD5 already protects against this, 
because the server issues a different challenge on each connection attempt.  
The datanode ID wouldn't provide any additional benefit.
# Added code in {{DataNode#checkSecureConfig}} to check that when SASL is used 
on DataTransferProtocol, SSL must also be used on HTTP.  Plain HTTP wouldn't be 
safe, because the client could write a delegation token query parameter onto 
the socket without any authentication of the server.  By requiring SSL, we 
enforce that the server is authenticated before sending the delegation token.

> Fix block protocol so that Datanodes don't require root or jsvc
> ---
>
> Key: HDFS-2856
> URL: https://issues.apache.org/jira/browse/HDFS-2856
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: datanode, security
>Affects Versions: 3.0.0, 2.4.0
>Reporter: Owen O'Malley
>Assignee: Chris Nauroth
> Attachments: Datanode-Security-Design.pdf, 
> Datanode-Security-Design.pdf, Datanode-Security-Design.pdf, 
> HDFS-2856.1.patch, HDFS-2856.2.patch, HDFS-2856.prototype.patch
>
>
> Since we send the block tokens unencrypted to the datanode, we currently 
> start the datanode as root using jsvc and get a secure (< 1024) port.
> If we have the datanode generate a nonce and send it on the connection and 
> the sends an hmac of the nonce back instead of the block token it won't 
> reveal any secrets. Thus, we wouldn't require a secure port and would not 
> require root or jsvc.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Resolved] (HDFS-6386) HDFS Encryption Zones

2014-06-18 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6386?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb resolved HDFS-6386.


  Resolution: Fixed
Target Version/s: fs-encryption (HADOOP-10150 and HDFS-6134)

Committed to fs-encryption. Thanks Andrew.

> HDFS Encryption Zones
> -
>
> Key: HDFS-6386
> URL: https://issues.apache.org/jira/browse/HDFS-6386
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode, security
>Reporter: Alejandro Abdelnur
>Assignee: Charles Lamb
> Fix For: fs-encryption (HADOOP-10150 and HDFS-6134)
>
> Attachments: HDFS-6386.012.patch, HDFS-6386.013.patch, 
> HDFS-6386.014.patch, HDFS-6386.4.patch, HDFS-6386.5.patch, HDFS-6386.6.patch, 
> HDFS-6386.8.patch
>
>
> Define the required security xAttributes for directories and files within an 
> encryption zone and how they propagate to children. Implement the logic to 
> create/delete encryption zones.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6480) Move waitForReady() from FSDirectory to FSNamesystem

2014-06-18 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6480?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-6480:
-

Attachment: HDFS-6480.004.patch

> Move waitForReady() from FSDirectory to FSNamesystem
> 
>
> Key: HDFS-6480
> URL: https://issues.apache.org/jira/browse/HDFS-6480
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6480.000.patch, HDFS-6480.001.patch, 
> HDFS-6480.002.patch, HDFS-6480.003.patch, HDFS-6480.004.patch
>
>
> Currently FSDirectory implements a barrier in {{waitForReady()}} / 
> {{setReady()}} so that it only serve requests once the FSImage is fully 
> loaded.
> As a part of the effort to evolve {{FSDirectory}} to a class which focuses on 
> implementing the data structure of the namespace, this jira proposes to move 
> the barrier one level higher to {{FSNamesystem}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6134) Transparent data at rest encryption

2014-06-18 Thread Larry McCay (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036560#comment-14036560
 ] 

Larry McCay commented on HDFS-6134:
---

[~tucu00] - I realize that it is the first usecase - that doesn't make it the 
only one that we have in mind or in the works. The fact that you have run into 
an issue with the EZ key granularity while using the CTR mode is a problem with 
the usecase design not necessarily with the abstraction of key providers. The 
question is whether wrapped keys will be required by other usecases where 
either the key usage pattern or the encryption modes in use may not require 
them. 

Currently, the KeyProvider API doesn't do any encryption itself - I just want 
to make sure that adding the additional complexity and responsibility to this 
interface is really necessary.

Additional questions:

* how does the keyprovider know what EZ key to use - is it the key that is 
referenced by the keyVersionName?
* how do we key HDFS clients from asking for the EZ key - if it is stored by 
the passed in keyVersionName?
** will this require special access control protection for EZ keys?
* would the unique DEK be stored in the provider as well or only in the 
extended attributes of the file?
** if stored in the provider what is the keyVersionName for it?



> Transparent data at rest encryption
> ---
>
> Key: HDFS-6134
> URL: https://issues.apache.org/jira/browse/HDFS-6134
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: security
>Affects Versions: 2.3.0
>Reporter: Alejandro Abdelnur
>Assignee: Alejandro Abdelnur
> Attachments: HDFSDataAtRestEncryption.pdf
>
>
> Because of privacy and security regulations, for many industries, sensitive 
> data at rest must be in encrypted form. For example: the health­care industry 
> (HIPAA regulations), the card payment industry (PCI DSS regulations) or the 
> US government (FISMA regulations).
> This JIRA aims to provide a mechanism to encrypt HDFS data at rest that can 
> be used transparently by any application accessing HDFS via Hadoop Filesystem 
> Java API, Hadoop libhdfs C library, or WebHDFS REST API.
> The resulting implementation should be able to be used in compliance with 
> different regulation requirements.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-5711) Removing memory limitation of the Namenode by persisting Block - Block location mappings to disk.

2014-06-18 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-5711:
-

Assignee: (was: Haohui Mai)

> Removing memory limitation of the Namenode by persisting Block - Block 
> location mappings to disk.
> -
>
> Key: HDFS-5711
> URL: https://issues.apache.org/jira/browse/HDFS-5711
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Rohan Pasalkar
>Priority: Minor
>
> This jira is to track changes to be made to remove HDFS name-node memory 
> limitation to hold block - block location mappings.
> It is a known fact that the single Name-node architecture of HDFS has 
> scalability limits. The HDFS federation project alleviates this problem by 
> using horizontal scaling. This helps increase the throughput of metadata 
> operation and also the amount of data that can be stored in a Hadoop cluster.
> The Name-node stores all the filesystem metadata in memory (even in the 
> federated architecture), the
> Name-node design can be enhanced by persisting part of the metadata onto 
> secondary storage and retaining 
> the popular or recently accessed metadata information in main memory. This 
> design can benefit a HDFS deployment
> which doesn't use federation but needs to store a large number of files or 
> large number of blocks. Lin Xiao from Hortonworks attempted a similar
> project [1] in the Summer of 2013. They used LevelDB to persist the Namespace 
> information (i.e file and directory inode information).
> A patch with this change is yet to be submitted to code base. We also intend 
> to use LevelDB to persist metadata, and plan to 
> provide a complete solution, by not just persisting  the Namespace 
> information but also the Blocks Map onto secondary storage. 
> We did implement the basic prototype which stores the block-block location 
> mapping metadata to the persistent key-value store i.e. levelDB. Prototype 
> also maintains the in-memory cache of the recently used block-block location 
> mappings metadata. 
> References:
> [1] Lin Xiao, Hortonworks, Removing Name-node’s memory limitation, HDFS-5389, 
> http://www.slideshare.net/ydn/hadoop-meetup-hug-august-2013-removing-the-namenodes-memory-limitation.
>  



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6386) HDFS Encryption Zones

2014-06-18 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6386?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb updated HDFS-6386:
---

Attachment: HDFS-6386.014.patch

Thanks for the review Andrew. The .014 patch removes those args.

> HDFS Encryption Zones
> -
>
> Key: HDFS-6386
> URL: https://issues.apache.org/jira/browse/HDFS-6386
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode, security
>Reporter: Alejandro Abdelnur
>Assignee: Charles Lamb
> Fix For: fs-encryption (HADOOP-10150 and HDFS-6134)
>
> Attachments: HDFS-6386.012.patch, HDFS-6386.013.patch, 
> HDFS-6386.014.patch, HDFS-6386.4.patch, HDFS-6386.5.patch, HDFS-6386.6.patch, 
> HDFS-6386.8.patch
>
>
> Define the required security xAttributes for directories and files within an 
> encryption zone and how they propagate to children. Implement the logic to 
> create/delete encryption zones.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6553) Add missing DeprecationDeltas for NFS Kerberos configurations

2014-06-18 Thread Stephen Chu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6553?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036547#comment-14036547
 ] 

Stephen Chu commented on HDFS-6553:
---

Thanks, [~brandonli]! Would you have time to commit? (not a committer)

> Add missing DeprecationDeltas for NFS Kerberos configurations
> -
>
> Key: HDFS-6553
> URL: https://issues.apache.org/jira/browse/HDFS-6553
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 3.0.0, 2.5.0
>Reporter: Stephen Chu
>Assignee: Stephen Chu
> Attachments: HDFS-6553.patch.001, HDFS-6553.patch.002
>
>
> HDFS-6056 (Clean up NFS config settings) improved NFS configuration naming 
> and added DeprecationDeltas to ensure that the old NFS configuration 
> properties could still be used.
> It's currently missing DeprecationDeltas for _dfs.nfs.keytab.file_ and 
> _dfs.nfs.kerberos.principal_.
> This patch adds those deltas so older configs for secure NFS can still work.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HDFS-5389) A Namenode that keeps only a part of the namespace in memory

2014-06-18 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5389?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai reassigned HDFS-5389:


Assignee: Haohui Mai

> A Namenode that keeps only a part of the namespace in memory
> 
>
> Key: HDFS-5389
> URL: https://issues.apache.org/jira/browse/HDFS-5389
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: 0.23.1
>Reporter: Lin Xiao
>Assignee: Haohui Mai
>Priority: Minor
>
> *Background:*
> Currently, the NN Keeps all its namespace in memory. This has had the benefit 
> that the NN code is very simple and, more importantly, helps the NN scale to 
> over 4.5K machines with 60K  to 100K concurrently tasks.  HDFS namespace can 
> be scaled currently using more Ram on the NN and/or using Federation which 
> scales both namespace and performance. The current federation implementation 
> does not allow renames across volumes without data copying but there are 
> proposals to remove that limitation.
> *Motivation:*
>  Hadoop lets customers store huge amounts of data at very economical prices 
> and hence allows customers to store their data for several years. While most 
> customers perform analytics on recent  data (last hour, day, week, months, 
> quarter, year), the ability to have five year old data online for analytics 
> is very attractive for many businesses. Although one can use larger RAM in a 
> NN and/or use Federation, it not really necessary to store the entire 
> namespace in memory since only the recent data is typically heavily accessed. 
> *Proposed Solution:*
> Store a portion of the NN's namespace in memory- the "working set" of the 
> applications that are currently operating. LSM data structures are quite 
> appropriate for maintaining the full namespace in memory. One choice is 
> Google's LevelDB open-source implementation.
> *Benefits:*
>  *  Store larger namespaces without resorting to Federated namespace volumes.
>  * Complementary to NN Federated namespace volumes,  indeed will allow a 
> single NN to easily store multiple larger volumes.
>  *  Faster cold startup - the NN does not have read its full namespace before 
> responding to clients.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HDFS-5711) Removing memory limitation of the Namenode by persisting Block - Block location mappings to disk.

2014-06-18 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai reassigned HDFS-5711:


Assignee: Haohui Mai

> Removing memory limitation of the Namenode by persisting Block - Block 
> location mappings to disk.
> -
>
> Key: HDFS-5711
> URL: https://issues.apache.org/jira/browse/HDFS-5711
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Rohan Pasalkar
>Assignee: Haohui Mai
>Priority: Minor
>
> This jira is to track changes to be made to remove HDFS name-node memory 
> limitation to hold block - block location mappings.
> It is a known fact that the single Name-node architecture of HDFS has 
> scalability limits. The HDFS federation project alleviates this problem by 
> using horizontal scaling. This helps increase the throughput of metadata 
> operation and also the amount of data that can be stored in a Hadoop cluster.
> The Name-node stores all the filesystem metadata in memory (even in the 
> federated architecture), the
> Name-node design can be enhanced by persisting part of the metadata onto 
> secondary storage and retaining 
> the popular or recently accessed metadata information in main memory. This 
> design can benefit a HDFS deployment
> which doesn't use federation but needs to store a large number of files or 
> large number of blocks. Lin Xiao from Hortonworks attempted a similar
> project [1] in the Summer of 2013. They used LevelDB to persist the Namespace 
> information (i.e file and directory inode information).
> A patch with this change is yet to be submitted to code base. We also intend 
> to use LevelDB to persist metadata, and plan to 
> provide a complete solution, by not just persisting  the Namespace 
> information but also the Blocks Map onto secondary storage. 
> We did implement the basic prototype which stores the block-block location 
> mapping metadata to the persistent key-value store i.e. levelDB. Prototype 
> also maintains the in-memory cache of the recently used block-block location 
> mappings metadata. 
> References:
> [1] Lin Xiao, Hortonworks, Removing Name-node’s memory limitation, HDFS-5389, 
> http://www.slideshare.net/ydn/hadoop-meetup-hug-august-2013-removing-the-namenodes-memory-limitation.
>  



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HDFS-6519) Document oiv_legacy command

2014-06-18 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6519?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA reassigned HDFS-6519:
---

Assignee: Akira AJISAKA

> Document oiv_legacy command
> ---
>
> Key: HDFS-6519
> URL: https://issues.apache.org/jira/browse/HDFS-6519
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: documentation
>Affects Versions: 2.5.0
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>
> HDFS-6293 introduced oiv_legacy command.
> The usage of the command should be included in OfflineImageViewer.apt.vm.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6386) HDFS Encryption Zones

2014-06-18 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6386?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036510#comment-14036510
 ] 

Andrew Wang commented on HDFS-6386:
---

Hi Charles,

* logRetryCache is still being passed unnecessarily to FSDir methods

+1 pending that though, thanks

> HDFS Encryption Zones
> -
>
> Key: HDFS-6386
> URL: https://issues.apache.org/jira/browse/HDFS-6386
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode, security
>Reporter: Alejandro Abdelnur
>Assignee: Charles Lamb
> Fix For: fs-encryption (HADOOP-10150 and HDFS-6134)
>
> Attachments: HDFS-6386.012.patch, HDFS-6386.013.patch, 
> HDFS-6386.4.patch, HDFS-6386.5.patch, HDFS-6386.6.patch, HDFS-6386.8.patch
>
>
> Define the required security xAttributes for directories and files within an 
> encryption zone and how they propagate to children. Implement the logic to 
> create/delete encryption zones.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6492) Support create-time xattrs and atomically setting multiple xattrs

2014-06-18 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6492?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036504#comment-14036504
 ] 

Colin Patrick McCabe commented on HDFS-6492:


{code}
import java.util.*;
{code}

Shouldn't use wildcard includes (turn off this setting in your editor)

{code}
  @VisibleForTesting
  List filterINodeXAttrs(List existingXAttrs,
  List filterXAttrs, List filteredXAttrs) {
{code}

{{filterXAttrs}} sounds confusingly similar to {{filteredXAttrs}}.  Perhaps 
rename these to {{toFilter}} and {{removed}}?

{{FSNamesystem#removeXAttr}}: we don't log the name of the xattr that was 
removed?  Seems like we should do this.  Maybe in a separate JIRA, though.

+1 once these are addressed

> Support create-time xattrs and atomically setting multiple xattrs
> -
>
> Key: HDFS-6492
> URL: https://issues.apache.org/jira/browse/HDFS-6492
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Andrew Wang
>Assignee: Andrew Wang
> Attachments: HDFS-6492.001.patch, HDFS-6492.002.patch
>
>
> Ongoing work in HDFS-6134 requires being able to set system namespace 
> extended attributes at create and mkdir time, as well as being able to 
> atomically set multiple xattrs at once. There's currently no need to expose 
> this functionality in the client API, so let's not unless we have to.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6559) Fix wrong option "dfsadmin -rollingUpgrade start" in the document

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036474#comment-14036474
 ] 

Hadoop QA commented on HDFS-6559:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651221/HDFS-6559.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+0 tests included{color}.  The patch appears to be a 
documentation patch that doesn't require tests.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7162//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7162//console

This message is automatically generated.

> Fix wrong option "dfsadmin -rollingUpgrade start" in the document
> -
>
> Key: HDFS-6559
> URL: https://issues.apache.org/jira/browse/HDFS-6559
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.4.0
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>Priority: Minor
>  Labels: newbie
> Fix For: 3.0.0, 2.5.0
>
> Attachments: HDFS-6559.patch
>
>
> In HdfsRollingUpgrade.xml,
> {code}
>   hdfs dfsadmin -rollingUpgrade 
> {code}
> should be
> {code}
>   hdfs dfsadmin -rollingUpgrade 
> 
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6482) Use block ID-based block layout on datanodes

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6482?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036465#comment-14036465
 ] 

Hadoop QA commented on HDFS-6482:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651220/HDFS-6482.5.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 15 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:red}-1 release audit{color}.  The applied patch generated 1 
release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs:

  org.apache.hadoop.hdfs.TestDatanodeLayoutUpgrade

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7163//testReport/
Release audit warnings: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7163//artifact/trunk/patchprocess/patchReleaseAuditProblems.txt
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7163//console

This message is automatically generated.

> Use block ID-based block layout on datanodes
> 
>
> Key: HDFS-6482
> URL: https://issues.apache.org/jira/browse/HDFS-6482
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: datanode
>Affects Versions: 2.5.0
>Reporter: James Thomas
>Assignee: James Thomas
> Attachments: HDFS-6482.1.patch, HDFS-6482.2.patch, HDFS-6482.3.patch, 
> HDFS-6482.4.patch, HDFS-6482.5.patch, HDFS-6482.patch
>
>
> Right now blocks are placed into directories that are split into many 
> subdirectories when capacity is reached. Instead we can use a block's ID to 
> determine the path it should go in. This eliminates the need for the LDir 
> data structure that facilitates the splitting of directories when they reach 
> capacity as well as fields in ReplicaInfo that keep track of a replica's 
> location.
> An extension of the work in HDFS-3290.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6418) Regression: DFS_NAMENODE_USER_NAME_KEY missing in trunk

2014-06-18 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6418?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HDFS-6418:
-

Assignee: Tsz Wo Nicholas Sze

> Regression: DFS_NAMENODE_USER_NAME_KEY missing in trunk
> ---
>
> Key: HDFS-6418
> URL: https://issues.apache.org/jira/browse/HDFS-6418
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 3.0.0, 2.5.0
>Reporter: Steve Loughran
>Assignee: Tsz Wo Nicholas Sze
>
> Code i have that compiles against HADOOP 2.4 doesn't build against trunk as 
> someone took away {{DFSConfigKeys.DFS_NAMENODE_USER_NAME_KEY}} -apparently in 
> HDFS-6181.
> I know the name was obsolete, but anyone who has compiled code using that 
> reference -rather than cutting and pasting in the string- is going to find 
> their code doesn't work.
> More subtly: that will lead to a link exception trying to run that code on a 
> 2.5+  cluster.
> This is a regression: the old names need to go back in, even if they refer to 
> the new names and are marked as deprecated



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6549) Add support for accessing the NFS gateway from the AIX NFS client

2014-06-18 Thread Aaron T. Myers (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036446#comment-14036446
 ] 

Aaron T. Myers commented on HDFS-6549:
--

bq. Large directory may need multiple READDIR/READDIRPLUS requests. Many NFS 
client implementations send 0 as the verifier for the first READDIR/READDIRPLUS 
RPC request. In the AIX case, after it grabs a verifier from server it will 
always use the same verifier even in the initial READDIR/READIRPLUS request 
when it accesses the same directory again. Is my understanding correct?

Correct, this is exactly what I've observed.

> Add support for accessing the NFS gateway from the AIX NFS client
> -
>
> Key: HDFS-6549
> URL: https://issues.apache.org/jira/browse/HDFS-6549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Attachments: HDFS-6549.patch
>
>
> We've identified two issues when trying to access the HDFS NFS Gateway from 
> an AIX NFS client:
> # In the case of COMMITs, the AIX NFS client will always send 4096, or a 
> multiple of the page size, for the offset to be committed, even if fewer 
> bytes than this have ever, or will ever, be written to the file. This will 
> cause a write to a file from the AIX NFS client to hang on close unless the 
> size of that file is a multiple of 4096.
> # In the case of READDIR and READDIRPLUS, the AIX NFS client will send the 
> same cookie verifier for a given directory seemingly forever after that 
> directory is first accessed over NFS, instead of getting a new cookie 
> verifier for every set of incremental readdir calls. This means that if a 
> directory's mtime ever changes, the FS must be unmounted/remounted before 
> readdir calls on that dir from AIX will ever succeed again.
> From my interpretation of RFC-1813, the NFS Gateway is in fact doing the 
> correct thing in both cases, but we can introduce simple changes on the NFS 
> Gateway side to be able to optionally work around these incompatibilities.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6549) Add support for accessing the NFS gateway from the AIX NFS client

2014-06-18 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036443#comment-14036443
 ] 

Brandon Li commented on HDFS-6549:
--

This sometimes happens with some traditional NFS server/client implementations: 
when the NFS client does READDIR/READDIRPLUS with a huge directory which keeps 
changing, the READDIR/READDIRPLUS can never finish because the verifier changes 
and NFS client has to restart the listing request before it finishes. But based 
on what you described, the AIX client is a different case.

Large directory may need multiple READDIR/READDIRPLUS requests. Many NFS client 
implementations send 0 as the verifier for the first READDIR/READDIRPLUS RPC 
request. In the AIX case, after it grabs a verifier from server it will always 
use the same verifier even in the initial READDIR/READIRPLUS request when it 
accesses the same directory again. Is my understanding correct?

> Add support for accessing the NFS gateway from the AIX NFS client
> -
>
> Key: HDFS-6549
> URL: https://issues.apache.org/jira/browse/HDFS-6549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.4.0
>Reporter: Aaron T. Myers
>Assignee: Aaron T. Myers
> Attachments: HDFS-6549.patch
>
>
> We've identified two issues when trying to access the HDFS NFS Gateway from 
> an AIX NFS client:
> # In the case of COMMITs, the AIX NFS client will always send 4096, or a 
> multiple of the page size, for the offset to be committed, even if fewer 
> bytes than this have ever, or will ever, be written to the file. This will 
> cause a write to a file from the AIX NFS client to hang on close unless the 
> size of that file is a multiple of 4096.
> # In the case of READDIR and READDIRPLUS, the AIX NFS client will send the 
> same cookie verifier for a given directory seemingly forever after that 
> directory is first accessed over NFS, instead of getting a new cookie 
> verifier for every set of incremental readdir calls. This means that if a 
> directory's mtime ever changes, the FS must be unmounted/remounted before 
> readdir calls on that dir from AIX will ever succeed again.
> From my interpretation of RFC-1813, the NFS Gateway is in fact doing the 
> correct thing in both cases, but we can introduce simple changes on the NFS 
> Gateway side to be able to optionally work around these incompatibilities.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6559) Fix wrong option "dfsadmin -rollingUpgrade start" in the document

2014-06-18 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036437#comment-14036437
 ] 

Hudson commented on HDFS-6559:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5729 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5729/])
HDFS-6559. Fix wrong option "dfsadmin -rollingUpgrade start" in the document. 
(Contributed by Akira Ajisaka) (arp: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1603638)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/site/xdoc/HdfsRollingUpgrade.xml


> Fix wrong option "dfsadmin -rollingUpgrade start" in the document
> -
>
> Key: HDFS-6559
> URL: https://issues.apache.org/jira/browse/HDFS-6559
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.4.0
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>Priority: Minor
>  Labels: newbie
> Fix For: 3.0.0, 2.5.0
>
> Attachments: HDFS-6559.patch
>
>
> In HdfsRollingUpgrade.xml,
> {code}
>   hdfs dfsadmin -rollingUpgrade 
> {code}
> should be
> {code}
>   hdfs dfsadmin -rollingUpgrade 
> 
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6559) Fix wrong option "dfsadmin -rollingUpgrade start" in the document

2014-06-18 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036434#comment-14036434
 ] 

Arpit Agarwal commented on HDFS-6559:
-

My bad, I missed waiting for the Jenkins +1.

Since this is a trivial documentation-only patch I won't revert it for now. In 
the unlikely case there is a problem I'll back it out.

> Fix wrong option "dfsadmin -rollingUpgrade start" in the document
> -
>
> Key: HDFS-6559
> URL: https://issues.apache.org/jira/browse/HDFS-6559
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.4.0
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>Priority: Minor
>  Labels: newbie
> Fix For: 3.0.0, 2.5.0
>
> Attachments: HDFS-6559.patch
>
>
> In HdfsRollingUpgrade.xml,
> {code}
>   hdfs dfsadmin -rollingUpgrade 
> {code}
> should be
> {code}
>   hdfs dfsadmin -rollingUpgrade 
> 
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6559) Fix wrong option "dfsadmin -rollingUpgrade start" in the document

2014-06-18 Thread Arpit Agarwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6559?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arpit Agarwal updated HDFS-6559:


   Resolution: Fixed
Fix Version/s: 2.5.0
   3.0.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

I committed this to trunk and branch-2.

Thanks for the contribution [~ajisakaa].

> Fix wrong option "dfsadmin -rollingUpgrade start" in the document
> -
>
> Key: HDFS-6559
> URL: https://issues.apache.org/jira/browse/HDFS-6559
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.4.0
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>Priority: Minor
>  Labels: newbie
> Fix For: 3.0.0, 2.5.0
>
> Attachments: HDFS-6559.patch
>
>
> In HdfsRollingUpgrade.xml,
> {code}
>   hdfs dfsadmin -rollingUpgrade 
> {code}
> should be
> {code}
>   hdfs dfsadmin -rollingUpgrade 
> 
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6559) Fix wrong option "dfsadmin -rollingUpgrade start" in the document

2014-06-18 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036419#comment-14036419
 ] 

Arpit Agarwal commented on HDFS-6559:
-

+1

I will commit it shortly.

> Fix wrong option "dfsadmin -rollingUpgrade start" in the document
> -
>
> Key: HDFS-6559
> URL: https://issues.apache.org/jira/browse/HDFS-6559
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.4.0
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>Priority: Minor
>  Labels: newbie
> Attachments: HDFS-6559.patch
>
>
> In HdfsRollingUpgrade.xml,
> {code}
>   hdfs dfsadmin -rollingUpgrade 
> {code}
> should be
> {code}
>   hdfs dfsadmin -rollingUpgrade 
> 
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6480) Move waitForReady() from FSDirectory to FSNamesystem

2014-06-18 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036417#comment-14036417
 ] 

Jing Zhao commented on HDFS-6480:
-

The 003 patch looks pretty good to me. Two minors:
# Currently in FSNamesystem#renameTo, we call waitForLoadingFSImage right 
before we call FSDirectory#renameTo. In that way, we may invoke 
FSDirectory.resolvePath before we can guarantee that the fsdir has been fully 
loaded.
# We need to update the following comment in several places:
{code}
-  // check if we are ready to initialize replication queues
+  // check if we are imageLoaded to initialize replication queues
{code}

+1 after addressing the comments.

> Move waitForReady() from FSDirectory to FSNamesystem
> 
>
> Key: HDFS-6480
> URL: https://issues.apache.org/jira/browse/HDFS-6480
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6480.000.patch, HDFS-6480.001.patch, 
> HDFS-6480.002.patch, HDFS-6480.003.patch
>
>
> Currently FSDirectory implements a barrier in {{waitForReady()}} / 
> {{setReady()}} so that it only serve requests once the FSImage is fully 
> loaded.
> As a part of the effort to evolve {{FSDirectory}} to a class which focuses on 
> implementing the data structure of the namespace, this jira proposes to move 
> the barrier one level higher to {{FSNamesystem}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HDFS-6561) Byte array native checksumming on client side

2014-06-18 Thread James Thomas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6561?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

James Thomas reassigned HDFS-6561:
--

Assignee: James Thomas

> Byte array native checksumming on client side
> -
>
> Key: HDFS-6561
> URL: https://issues.apache.org/jira/browse/HDFS-6561
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: datanode, hdfs-client, performance
>Reporter: James Thomas
>Assignee: James Thomas
>




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HDFS-6560) Byte array native checksumming on DN side

2014-06-18 Thread James Thomas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6560?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

James Thomas reassigned HDFS-6560:
--

Assignee: James Thomas

> Byte array native checksumming on DN side
> -
>
> Key: HDFS-6560
> URL: https://issues.apache.org/jira/browse/HDFS-6560
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: datanode, hdfs-client, performance
>Reporter: James Thomas
>Assignee: James Thomas
>




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6560) Byte array native checksumming on DN side

2014-06-18 Thread James Thomas (JIRA)
James Thomas created HDFS-6560:
--

 Summary: Byte array native checksumming on DN side
 Key: HDFS-6560
 URL: https://issues.apache.org/jira/browse/HDFS-6560
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: James Thomas






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6561) Byte array native checksumming on client side

2014-06-18 Thread James Thomas (JIRA)
James Thomas created HDFS-6561:
--

 Summary: Byte array native checksumming on client side
 Key: HDFS-6561
 URL: https://issues.apache.org/jira/browse/HDFS-6561
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: James Thomas






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6439) NFS should not reject NFS requests to the NULL procedure whether port monitoring is enabled or not

2014-06-18 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6439?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036370#comment-14036370
 ] 

Hudson commented on HDFS-6439:
--

FAILURE: Integrated in Hadoop-trunk-Commit #5728 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5728/])
HDFS-6439. NFS should not reject NFS requests to the NULL procedure whether 
port monitoring is enabled or not. Contributed by Brandon Li (brandonli: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1603622)
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/nfs/nfs3/Nfs3Interface.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/RpcProgram.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-nfs/src/test/java/org/apache/hadoop/oncrpc/TestFrameDecoder.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/conf/NfsConfigKeys.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/conf/NfsConfiguration.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/mount/RpcProgramMountd.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/Nfs3.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/RpcProgramNfs3.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/TestReaddir.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/nfs3/TestReaddir.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/nfs3/TestWrites.java
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/site/apt/HdfsNfsGateway.apt.vm


> NFS should not reject NFS requests to the NULL procedure whether port 
> monitoring is enabled or not
> --
>
> Key: HDFS-6439
> URL: https://issues.apache.org/jira/browse/HDFS-6439
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.5.0
>Reporter: Brandon Li
>Assignee: Aaron T. Myers
> Fix For: 2.5.0
>
> Attachments: HDFS-6439.003.patch, HDFS-6439.004.patch, 
> HDFS-6439.005.patch, HDFS-6439.patch, HDFS-6439.patch, 
> linux-nfs-disallow-request-from-nonsecure-port.pcapng, 
> mount-nfs-requests.pcapng
>
>
> As discussed in HDFS-6406, this JIRA is to track the follow update:
> 1. Port monitoring is the feature name with traditional NFS server and we may 
> want to make the config property (along with related variable 
> allowInsecurePorts) something as dfs.nfs.port.monitoring. 
> 2 . According to RFC2623 (http://www.rfc-editor.org/rfc/rfc2623.txt):
> {quote}Whether port monitoring is enabled or not, NFS servers SHOULD NOT 
> reject NFS requests to the NULL procedure (procedure number 0). See 
> subsection 2.3.1, "NULL procedure" for a complete explanation. {quote}
> I do notice that NFS clients (most time) send mount NULL and nfs NULL from 
> non-privileged port. If we deny NULL call in mountd or nfs server, the client 
> can't mount the export even as user root.
> 3. it would be nice to have the user guide updated for the port monitoring 
> feature.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6480) Move waitForReady() from FSDirectory to FSNamesystem

2014-06-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036363#comment-14036363
 ] 

Hadoop QA commented on HDFS-6480:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12651198/HDFS-6480.003.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 4 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/7161//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/7161//console

This message is automatically generated.

> Move waitForReady() from FSDirectory to FSNamesystem
> 
>
> Key: HDFS-6480
> URL: https://issues.apache.org/jira/browse/HDFS-6480
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-6480.000.patch, HDFS-6480.001.patch, 
> HDFS-6480.002.patch, HDFS-6480.003.patch
>
>
> Currently FSDirectory implements a barrier in {{waitForReady()}} / 
> {{setReady()}} so that it only serve requests once the FSImage is fully 
> loaded.
> As a part of the effort to evolve {{FSDirectory}} to a class which focuses on 
> implementing the data structure of the namespace, this jira proposes to move 
> the barrier one level higher to {{FSNamesystem}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6439) NFS should not reject NFS requests to the NULL procedure whether port monitoring is enabled or not

2014-06-18 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6439?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-6439:
-

Affects Version/s: (was: 2.4.0)
   2.5.0

> NFS should not reject NFS requests to the NULL procedure whether port 
> monitoring is enabled or not
> --
>
> Key: HDFS-6439
> URL: https://issues.apache.org/jira/browse/HDFS-6439
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.5.0
>Reporter: Brandon Li
>Assignee: Aaron T. Myers
> Fix For: 2.5.0
>
> Attachments: HDFS-6439.003.patch, HDFS-6439.004.patch, 
> HDFS-6439.005.patch, HDFS-6439.patch, HDFS-6439.patch, 
> linux-nfs-disallow-request-from-nonsecure-port.pcapng, 
> mount-nfs-requests.pcapng
>
>
> As discussed in HDFS-6406, this JIRA is to track the follow update:
> 1. Port monitoring is the feature name with traditional NFS server and we may 
> want to make the config property (along with related variable 
> allowInsecurePorts) something as dfs.nfs.port.monitoring. 
> 2 . According to RFC2623 (http://www.rfc-editor.org/rfc/rfc2623.txt):
> {quote}Whether port monitoring is enabled or not, NFS servers SHOULD NOT 
> reject NFS requests to the NULL procedure (procedure number 0). See 
> subsection 2.3.1, "NULL procedure" for a complete explanation. {quote}
> I do notice that NFS clients (most time) send mount NULL and nfs NULL from 
> non-privileged port. If we deny NULL call in mountd or nfs server, the client 
> can't mount the export even as user root.
> 3. it would be nice to have the user guide updated for the port monitoring 
> feature.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6439) NFS should not reject NFS requests to the NULL procedure whether port monitoring is enabled or not

2014-06-18 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6439?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-6439:
-

Fix Version/s: 2.5.0

> NFS should not reject NFS requests to the NULL procedure whether port 
> monitoring is enabled or not
> --
>
> Key: HDFS-6439
> URL: https://issues.apache.org/jira/browse/HDFS-6439
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.5.0
>Reporter: Brandon Li
>Assignee: Aaron T. Myers
> Fix For: 2.5.0
>
> Attachments: HDFS-6439.003.patch, HDFS-6439.004.patch, 
> HDFS-6439.005.patch, HDFS-6439.patch, HDFS-6439.patch, 
> linux-nfs-disallow-request-from-nonsecure-port.pcapng, 
> mount-nfs-requests.pcapng
>
>
> As discussed in HDFS-6406, this JIRA is to track the follow update:
> 1. Port monitoring is the feature name with traditional NFS server and we may 
> want to make the config property (along with related variable 
> allowInsecurePorts) something as dfs.nfs.port.monitoring. 
> 2 . According to RFC2623 (http://www.rfc-editor.org/rfc/rfc2623.txt):
> {quote}Whether port monitoring is enabled or not, NFS servers SHOULD NOT 
> reject NFS requests to the NULL procedure (procedure number 0). See 
> subsection 2.3.1, "NULL procedure" for a complete explanation. {quote}
> I do notice that NFS clients (most time) send mount NULL and nfs NULL from 
> non-privileged port. If we deny NULL call in mountd or nfs server, the client 
> can't mount the export even as user root.
> 3. it would be nice to have the user guide updated for the port monitoring 
> feature.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6439) NFS should not reject NFS requests to the NULL procedure whether port monitoring is enabled or not

2014-06-18 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6439?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036356#comment-14036356
 ] 

Brandon Li commented on HDFS-6439:
--

I've committed the patch. Thanks again, Aaron.

> NFS should not reject NFS requests to the NULL procedure whether port 
> monitoring is enabled or not
> --
>
> Key: HDFS-6439
> URL: https://issues.apache.org/jira/browse/HDFS-6439
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.5.0
>Reporter: Brandon Li
>Assignee: Aaron T. Myers
> Fix For: 2.5.0
>
> Attachments: HDFS-6439.003.patch, HDFS-6439.004.patch, 
> HDFS-6439.005.patch, HDFS-6439.patch, HDFS-6439.patch, 
> linux-nfs-disallow-request-from-nonsecure-port.pcapng, 
> mount-nfs-requests.pcapng
>
>
> As discussed in HDFS-6406, this JIRA is to track the follow update:
> 1. Port monitoring is the feature name with traditional NFS server and we may 
> want to make the config property (along with related variable 
> allowInsecurePorts) something as dfs.nfs.port.monitoring. 
> 2 . According to RFC2623 (http://www.rfc-editor.org/rfc/rfc2623.txt):
> {quote}Whether port monitoring is enabled or not, NFS servers SHOULD NOT 
> reject NFS requests to the NULL procedure (procedure number 0). See 
> subsection 2.3.1, "NULL procedure" for a complete explanation. {quote}
> I do notice that NFS clients (most time) send mount NULL and nfs NULL from 
> non-privileged port. If we deny NULL call in mountd or nfs server, the client 
> can't mount the export even as user root.
> 3. it would be nice to have the user guide updated for the port monitoring 
> feature.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6439) NFS should not reject NFS requests to the NULL procedure whether port monitoring is enabled or not

2014-06-18 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6439?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-6439:
-

  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

> NFS should not reject NFS requests to the NULL procedure whether port 
> monitoring is enabled or not
> --
>
> Key: HDFS-6439
> URL: https://issues.apache.org/jira/browse/HDFS-6439
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.5.0
>Reporter: Brandon Li
>Assignee: Aaron T. Myers
> Fix For: 2.5.0
>
> Attachments: HDFS-6439.003.patch, HDFS-6439.004.patch, 
> HDFS-6439.005.patch, HDFS-6439.patch, HDFS-6439.patch, 
> linux-nfs-disallow-request-from-nonsecure-port.pcapng, 
> mount-nfs-requests.pcapng
>
>
> As discussed in HDFS-6406, this JIRA is to track the follow update:
> 1. Port monitoring is the feature name with traditional NFS server and we may 
> want to make the config property (along with related variable 
> allowInsecurePorts) something as dfs.nfs.port.monitoring. 
> 2 . According to RFC2623 (http://www.rfc-editor.org/rfc/rfc2623.txt):
> {quote}Whether port monitoring is enabled or not, NFS servers SHOULD NOT 
> reject NFS requests to the NULL procedure (procedure number 0). See 
> subsection 2.3.1, "NULL procedure" for a complete explanation. {quote}
> I do notice that NFS clients (most time) send mount NULL and nfs NULL from 
> non-privileged port. If we deny NULL call in mountd or nfs server, the client 
> can't mount the export even as user root.
> 3. it would be nice to have the user guide updated for the port monitoring 
> feature.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6134) Transparent data at rest encryption

2014-06-18 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036344#comment-14036344
 ] 

Alejandro Abdelnur commented on HDFS-6134:
--

[~lmccay], it is great that you agree that the proposed changes makes things 
more secure.

Answering your bullets:

#1. this is the first use case in Hadoop
#2. no, the client does not have acces to the EZ keyVersion
#3. the KeyProvider own the decryption of the DEK ,the client does not have the 
EZ key

What is being proposed is the equivalent of KMIP's key-wrapping functionality. 
What makes the system more secure is the fact that the EZ keyVersion is never 
exposed to HDFS, nor HDFS clients. And the DEK for the file is never exposed to 
HDFS NN. 


> Transparent data at rest encryption
> ---
>
> Key: HDFS-6134
> URL: https://issues.apache.org/jira/browse/HDFS-6134
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: security
>Affects Versions: 2.3.0
>Reporter: Alejandro Abdelnur
>Assignee: Alejandro Abdelnur
> Attachments: HDFSDataAtRestEncryption.pdf
>
>
> Because of privacy and security regulations, for many industries, sensitive 
> data at rest must be in encrypted form. For example: the health­care industry 
> (HIPAA regulations), the card payment industry (PCI DSS regulations) or the 
> US government (FISMA regulations).
> This JIRA aims to provide a mechanism to encrypt HDFS data at rest that can 
> be used transparently by any application accessing HDFS via Hadoop Filesystem 
> Java API, Hadoop libhdfs C library, or WebHDFS REST API.
> The resulting implementation should be able to be used in compliance with 
> different regulation requirements.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6553) Add missing DeprecationDeltas for NFS Kerberos configurations

2014-06-18 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6553?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036332#comment-14036332
 ] 

Brandon Li commented on HDFS-6553:
--

+1

> Add missing DeprecationDeltas for NFS Kerberos configurations
> -
>
> Key: HDFS-6553
> URL: https://issues.apache.org/jira/browse/HDFS-6553
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 3.0.0, 2.5.0
>Reporter: Stephen Chu
>Assignee: Stephen Chu
> Attachments: HDFS-6553.patch.001, HDFS-6553.patch.002
>
>
> HDFS-6056 (Clean up NFS config settings) improved NFS configuration naming 
> and added DeprecationDeltas to ensure that the old NFS configuration 
> properties could still be used.
> It's currently missing DeprecationDeltas for _dfs.nfs.keytab.file_ and 
> _dfs.nfs.kerberos.principal_.
> This patch adds those deltas so older configs for secure NFS can still work.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6439) NFS should not reject NFS requests to the NULL procedure whether port monitoring is enabled or not

2014-06-18 Thread Aaron T. Myers (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6439?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036310#comment-14036310
 ] 

Aaron T. Myers commented on HDFS-6439:
--

Agreed. +1

> NFS should not reject NFS requests to the NULL procedure whether port 
> monitoring is enabled or not
> --
>
> Key: HDFS-6439
> URL: https://issues.apache.org/jira/browse/HDFS-6439
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.4.0
>Reporter: Brandon Li
>Assignee: Aaron T. Myers
> Attachments: HDFS-6439.003.patch, HDFS-6439.004.patch, 
> HDFS-6439.005.patch, HDFS-6439.patch, HDFS-6439.patch, 
> linux-nfs-disallow-request-from-nonsecure-port.pcapng, 
> mount-nfs-requests.pcapng
>
>
> As discussed in HDFS-6406, this JIRA is to track the follow update:
> 1. Port monitoring is the feature name with traditional NFS server and we may 
> want to make the config property (along with related variable 
> allowInsecurePorts) something as dfs.nfs.port.monitoring. 
> 2 . According to RFC2623 (http://www.rfc-editor.org/rfc/rfc2623.txt):
> {quote}Whether port monitoring is enabled or not, NFS servers SHOULD NOT 
> reject NFS requests to the NULL procedure (procedure number 0). See 
> subsection 2.3.1, "NULL procedure" for a complete explanation. {quote}
> I do notice that NFS clients (most time) send mount NULL and nfs NULL from 
> non-privileged port. If we deny NULL call in mountd or nfs server, the client 
> can't mount the export even as user root.
> 3. it would be nice to have the user guide updated for the port monitoring 
> feature.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HDFS-6439) NFS should not reject NFS requests to the NULL procedure whether port monitoring is enabled or not

2014-06-18 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6439?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14036303#comment-14036303
 ] 

Brandon Li commented on HDFS-6439:
--

The test failures are not introduced by this patch.

> NFS should not reject NFS requests to the NULL procedure whether port 
> monitoring is enabled or not
> --
>
> Key: HDFS-6439
> URL: https://issues.apache.org/jira/browse/HDFS-6439
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Affects Versions: 2.4.0
>Reporter: Brandon Li
>Assignee: Aaron T. Myers
> Attachments: HDFS-6439.003.patch, HDFS-6439.004.patch, 
> HDFS-6439.005.patch, HDFS-6439.patch, HDFS-6439.patch, 
> linux-nfs-disallow-request-from-nonsecure-port.pcapng, 
> mount-nfs-requests.pcapng
>
>
> As discussed in HDFS-6406, this JIRA is to track the follow update:
> 1. Port monitoring is the feature name with traditional NFS server and we may 
> want to make the config property (along with related variable 
> allowInsecurePorts) something as dfs.nfs.port.monitoring. 
> 2 . According to RFC2623 (http://www.rfc-editor.org/rfc/rfc2623.txt):
> {quote}Whether port monitoring is enabled or not, NFS servers SHOULD NOT 
> reject NFS requests to the NULL procedure (procedure number 0). See 
> subsection 2.3.1, "NULL procedure" for a complete explanation. {quote}
> I do notice that NFS clients (most time) send mount NULL and nfs NULL from 
> non-privileged port. If we deny NULL call in mountd or nfs server, the client 
> can't mount the export even as user root.
> 3. it would be nice to have the user guide updated for the port monitoring 
> feature.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HDFS-6558) Missing '\n' in the description of dfsadmin -rollingUpgrade

2014-06-18 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-6558:


Assignee: Chen He
Target Version/s: 2.5.0
  Status: Patch Available  (was: Open)

> Missing '\n' in the description of dfsadmin -rollingUpgrade
> ---
>
> Key: HDFS-6558
> URL: https://issues.apache.org/jira/browse/HDFS-6558
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Affects Versions: 2.4.0
>Reporter: Akira AJISAKA
>Assignee: Chen He
>Priority: Trivial
>  Labels: newbie
> Attachments: HDFS-6558.patch
>
>
> In DFSAdmin.java, '\n' should be added at the end of the line
> {code}
> + "   prepare: prepare a new rolling upgrade."
> {code}
> to clean up the following help message.
> {code}
> $ hdfs dfsadmin -help rollingUpgrade
> -rollingUpgrade []:
>  query: query the current rolling upgrade status.
>prepare: prepare a new rolling upgrade.  finalize: finalize the current 
> rolling upgrade.
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


  1   2   >