[jira] [Commented] (HDFS-5228) The RemoteIterator returned by DistributedFileSystem.listFiles(..) may throw NPE

2013-09-23 Thread Tsz Wo (Nicholas), SZE (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5228?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774354#comment-13774354
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-5228:
--

> We have extensive symlink tests checked in as part of HADOOP-8040, ...

Which part of HADOOP-8040?  Where are the new tests?

> The RemoteIterator returned by DistributedFileSystem.listFiles(..) may throw 
> NPE
> 
>
> Key: HDFS-5228
> URL: https://issues.apache.org/jira/browse/HDFS-5228
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.1.0-beta
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
>Priority: Blocker
> Attachments: h5228_20130919.patch, h5228_20130919_test.patch, 
> HDFS-5228.2.patch
>
>
> Get a RemoteIterator from DistributedFileSystem.listFiles(..) with a relative 
> path.  Then, it will result a NullPointerException when calling hasNext() 
> from the RemoteIterator.
> This bug was discovered by Arnaud:
> http://hortonworks.com/community/forums/topic/new-bug-in-hdfs-listfiles-method/

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5228) The RemoteIterator returned by DistributedFileSystem.listFiles(..) may throw NPE

2013-09-23 Thread Tsz Wo (Nicholas), SZE (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5228?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774366#comment-13774366
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-5228:
--

Chris, thanks for posting a patch.  I do think that the best solution is to 
revert HADOOP-9418, which seemed not well tested.  HADOOP-9418 first broke 
HADOOP-9761 and then this.  As Andrew said, "... I think some number of bugs 
like this are unavoidable as symlinks was a pretty big change that required 
touching basically all the client API code. ..."  We should not let HADOOP-9418 
destabilize 2.1.1.

> The RemoteIterator returned by DistributedFileSystem.listFiles(..) may throw 
> NPE
> 
>
> Key: HDFS-5228
> URL: https://issues.apache.org/jira/browse/HDFS-5228
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.1.0-beta
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
>Priority: Blocker
> Attachments: h5228_20130919.patch, h5228_20130919_test.patch, 
> HDFS-5228.2.patch
>
>
> Get a RemoteIterator from DistributedFileSystem.listFiles(..) with a relative 
> path.  Then, it will result a NullPointerException when calling hasNext() 
> from the RemoteIterator.
> This bug was discovered by Arnaud:
> http://hortonworks.com/community/forums/topic/new-bug-in-hdfs-listfiles-method/

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5228) The RemoteIterator returned by DistributedFileSystem.listFiles(..) may throw NPE

2013-09-23 Thread Tsz Wo (Nicholas), SZE (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5228?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774368#comment-13774368
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-5228:
--

Andrew, you have not answered my following question.  I hope that you are not 
avoiding it.

bq. How exactly to audit DFS do you suggest so that it could avoid further bugs?

> The RemoteIterator returned by DistributedFileSystem.listFiles(..) may throw 
> NPE
> 
>
> Key: HDFS-5228
> URL: https://issues.apache.org/jira/browse/HDFS-5228
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.1.0-beta
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
>Priority: Blocker
> Attachments: h5228_20130919.patch, h5228_20130919_test.patch, 
> HDFS-5228.2.patch
>
>
> Get a RemoteIterator from DistributedFileSystem.listFiles(..) with a relative 
> path.  Then, it will result a NullPointerException when calling hasNext() 
> from the RemoteIterator.
> This bug was discovered by Arnaud:
> http://hortonworks.com/community/forums/topic/new-bug-in-hdfs-listfiles-method/

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5216) NumDecomDeadDataNodes not returning correct number of dead decommissioned nodes

2013-09-23 Thread Trevor Lorimer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5216?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Trevor Lorimer updated HDFS-5216:
-

Status: Patch Available  (was: Open)

> NumDecomDeadDataNodes not returning correct number of dead decommissioned 
> nodes 
> 
>
> Key: HDFS-5216
> URL: https://issues.apache.org/jira/browse/HDFS-5216
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.0-beta
>Reporter: Trevor Lorimer
>Assignee: Trevor Lorimer
> Attachments: HDFS-5216.diff, HDFS-5216.diff
>
>
> For HDFS-4860 I essentially copied the process in 
> NamenodejspHelper.generateHealthReport(), so it would be in sync with the 
> original dfsHealth.jsp.
> However looking at this now there may be a bug? in 
> getNumDecomDeadDataNodes(), where:
> getBlockManager().getDatanodeManager().fetchDatanodes(dead, null, true);
> 
> Where the parameter true indicates that decommissioned nodes should be 
> removed from the list.
> If the flag is true fetchDatanodes calls removeDecomNodeFromList, which will 
> remove a node if an existing datanode does not appear in both include or 
> exclude lists and it has been decommissioned.
> If I am looking to return the Number of Dead Decommissioned Nodes, should I 
> change the remove decommissioned nodes flag to False? i.e.:
> getBlockManager().getDatanodeManager().fetchDatanodes(null, dead, false);

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5216) NumDecomDeadDataNodes not returning correct number of dead decommissioned nodes

2013-09-23 Thread Trevor Lorimer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5216?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Trevor Lorimer updated HDFS-5216:
-

Attachment: HDFS-5216.diff

> NumDecomDeadDataNodes not returning correct number of dead decommissioned 
> nodes 
> 
>
> Key: HDFS-5216
> URL: https://issues.apache.org/jira/browse/HDFS-5216
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.0-beta
>Reporter: Trevor Lorimer
>Assignee: Trevor Lorimer
> Attachments: HDFS-5216.diff, HDFS-5216.diff
>
>
> For HDFS-4860 I essentially copied the process in 
> NamenodejspHelper.generateHealthReport(), so it would be in sync with the 
> original dfsHealth.jsp.
> However looking at this now there may be a bug? in 
> getNumDecomDeadDataNodes(), where:
> getBlockManager().getDatanodeManager().fetchDatanodes(dead, null, true);
> 
> Where the parameter true indicates that decommissioned nodes should be 
> removed from the list.
> If the flag is true fetchDatanodes calls removeDecomNodeFromList, which will 
> remove a node if an existing datanode does not appear in both include or 
> exclude lists and it has been decommissioned.
> If I am looking to return the Number of Dead Decommissioned Nodes, should I 
> change the remove decommissioned nodes flag to False? i.e.:
> getBlockManager().getDatanodeManager().fetchDatanodes(null, dead, false);

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5216) NumDecomDeadDataNodes not returning correct number of dead decommissioned nodes

2013-09-23 Thread Trevor Lorimer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5216?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Trevor Lorimer updated HDFS-5216:
-

Assignee: Trevor Lorimer
  Status: Open  (was: Patch Available)

> NumDecomDeadDataNodes not returning correct number of dead decommissioned 
> nodes 
> 
>
> Key: HDFS-5216
> URL: https://issues.apache.org/jira/browse/HDFS-5216
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.0-beta
>Reporter: Trevor Lorimer
>Assignee: Trevor Lorimer
> Attachments: HDFS-5216.diff, HDFS-5216.diff
>
>
> For HDFS-4860 I essentially copied the process in 
> NamenodejspHelper.generateHealthReport(), so it would be in sync with the 
> original dfsHealth.jsp.
> However looking at this now there may be a bug? in 
> getNumDecomDeadDataNodes(), where:
> getBlockManager().getDatanodeManager().fetchDatanodes(dead, null, true);
> 
> Where the parameter true indicates that decommissioned nodes should be 
> removed from the list.
> If the flag is true fetchDatanodes calls removeDecomNodeFromList, which will 
> remove a node if an existing datanode does not appear in both include or 
> exclude lists and it has been decommissioned.
> If I am looking to return the Number of Dead Decommissioned Nodes, should I 
> change the remove decommissioned nodes flag to False? i.e.:
> getBlockManager().getDatanodeManager().fetchDatanodes(null, dead, false);

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1213) Implement an Apache Commons VFS Driver for HDFS

2013-09-23 Thread Jakub Narloch (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1213?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774520#comment-13774520
 ] 

Jakub Narloch commented on HDFS-1213:
-

Hi,

What is the status of this issue? Is there any change I would be released 
someday?
Actually we could use such level of abstraction and I think the general idea is 
great.

> Implement an Apache Commons VFS Driver for HDFS
> ---
>
> Key: HDFS-1213
> URL: https://issues.apache.org/jira/browse/HDFS-1213
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: hdfs-client
>Reporter: Michael D'Amour
> Attachments: HADOOP-HDFS-Apache-VFS.patch, 
> pentaho-hdfs-vfs-TRUNK-SNAPSHOT.jar, 
> pentaho-hdfs-vfs-TRUNK-SNAPSHOT-sources.tar.gz
>
>
> We have an open source ETL tool (Kettle) which uses VFS for many input/output 
> steps/jobs.  We would like to be able to read/write HDFS from Kettle using 
> VFS.  
>  
> I haven't been able to find anything out there other than "it would be nice."
>  
> I had some time a few weeks ago to begin writing a VFS driver for HDFS and we 
> (Pentaho) would like to be able to contribute this driver.  I believe it 
> supports all the major file/folder operations and I have written unit tests 
> for all of these operations.  The code is currently checked into an open 
> Pentaho SVN repository under the Apache 2.0 license.  There are some current 
> limitations, such as a lack of authentication (kerberos), which appears to be 
> coming in 0.22.0, however, the driver supports username/password, but I just 
> can't use them yet.
> I will be attaching the code for the driver once the case is created.  The 
> project does not modify existing hadoop/hdfs source.
> Our JIRA case can be found at http://jira.pentaho.com/browse/PDI-4146

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1213) Implement an Apache Commons VFS Driver for HDFS

2013-09-23 Thread Matt Casters (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1213?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774553#comment-13774553
 ] 

Matt Casters commented on HDFS-1213:


Jakub, until the VFS driver can be incorporated into the HDFS or VFS Apache 
projects, Pentaho maintains a separate source tree.  It's over here: 
http://source.pentaho.org/viewvc/svnroot/pentaho-commons/pentaho-hdfs-vfs/trunk/

> Implement an Apache Commons VFS Driver for HDFS
> ---
>
> Key: HDFS-1213
> URL: https://issues.apache.org/jira/browse/HDFS-1213
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: hdfs-client
>Reporter: Michael D'Amour
> Attachments: HADOOP-HDFS-Apache-VFS.patch, 
> pentaho-hdfs-vfs-TRUNK-SNAPSHOT.jar, 
> pentaho-hdfs-vfs-TRUNK-SNAPSHOT-sources.tar.gz
>
>
> We have an open source ETL tool (Kettle) which uses VFS for many input/output 
> steps/jobs.  We would like to be able to read/write HDFS from Kettle using 
> VFS.  
>  
> I haven't been able to find anything out there other than "it would be nice."
>  
> I had some time a few weeks ago to begin writing a VFS driver for HDFS and we 
> (Pentaho) would like to be able to contribute this driver.  I believe it 
> supports all the major file/folder operations and I have written unit tests 
> for all of these operations.  The code is currently checked into an open 
> Pentaho SVN repository under the Apache 2.0 license.  There are some current 
> limitations, such as a lack of authentication (kerberos), which appears to be 
> coming in 0.22.0, however, the driver supports username/password, but I just 
> can't use them yet.
> I will be attaching the code for the driver once the case is created.  The 
> project does not modify existing hadoop/hdfs source.
> Our JIRA case can be found at http://jira.pentaho.com/browse/PDI-4146

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5239) Allow FSNamesystem lock fairness to be configurable

2013-09-23 Thread Daryn Sharp (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5239?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daryn Sharp updated HDFS-5239:
--

   Resolution: Fixed
Fix Version/s: 2.3.0
   3.0.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

Thanks Todd!  I've committed to trunk and branch-2.

> Allow FSNamesystem lock fairness to be configurable
> ---
>
> Key: HDFS-5239
> URL: https://issues.apache.org/jira/browse/HDFS-5239
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
> Fix For: 3.0.0, 2.3.0
>
> Attachments: HDFS-5239.patch
>
>
> The fairness of the {{FSNamesystem#fsLock}} is hardcoded to {{true}}.  Using 
> an unfair lock provides a negligible increase to throughput.  However this is 
> due to bottlenecks elsewhere in the system. In combination with other 
> changes, such as RPC layer and audit logging, preliminary tests show up to a 
> 5X improvement for a read heavy workload.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5240) Separate formatting from logging in the audit logger API

2013-09-23 Thread Daryn Sharp (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5240?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daryn Sharp updated HDFS-5240:
--

   Resolution: Fixed
Fix Version/s: 2.3.0
   3.0.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

Thanks Todd!  I've committed to trunk and branch-2.

> Separate formatting from logging in the audit logger API
> 
>
> Key: HDFS-5240
> URL: https://issues.apache.org/jira/browse/HDFS-5240
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
> Fix For: 3.0.0, 2.3.0
>
> Attachments: HDFS-5240.patch
>
>
> The audit logger API should be extended (in a compatible manner) to separate 
> the formatting of the log message from the actual logging of the message.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5240) Separate formatting from logging in the audit logger API

2013-09-23 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5240?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774654#comment-13774654
 ] 

Hudson commented on HDFS-5240:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #4453 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/4453/])
HDFS-5240. Separate formatting from logging in the audit logger API (daryn) 
(daryn: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1525626)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java


> Separate formatting from logging in the audit logger API
> 
>
> Key: HDFS-5240
> URL: https://issues.apache.org/jira/browse/HDFS-5240
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
> Fix For: 3.0.0, 2.3.0
>
> Attachments: HDFS-5240.patch
>
>
> The audit logger API should be extended (in a compatible manner) to separate 
> the formatting of the log message from the actual logging of the message.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5242) Reduce contention on DatanodeInfo instances

2013-09-23 Thread Daryn Sharp (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5242?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daryn Sharp updated HDFS-5242:
--

Status: Patch Available  (was: Open)

No test because there's no way to squeeze in a barrier.

> Reduce contention on DatanodeInfo instances
> ---
>
> Key: HDFS-5242
> URL: https://issues.apache.org/jira/browse/HDFS-5242
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
> Attachments: HDFS-5242.patch
>
>
> Synchronization in {{DatanodeInfo}} instances causes unnecessary contention 
> between call handlers.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5242) Reduce contention on DatanodeInfo instances

2013-09-23 Thread Daryn Sharp (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5242?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daryn Sharp updated HDFS-5242:
--

Attachment: HDFS-5242.patch

Remove synchronized keyword from {{get/setNetworkLocation}}. HDFS-1985 added 
the synch with no clear reason and it's the only synch'ed mutable field in the 
class.  A profiler identified this unnecessary synch after peeling away other 
issues.  Unnecessary synchs cause thread contention and which harms performance 
due to context switches.

The synch doesn't really protect against any race conditions.  Further, the 
network location is only set when registering a datanode so it's not changing 
during normal operation.

> Reduce contention on DatanodeInfo instances
> ---
>
> Key: HDFS-5242
> URL: https://issues.apache.org/jira/browse/HDFS-5242
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
> Attachments: HDFS-5242.patch
>
>
> Synchronization in {{DatanodeInfo}} instances causes unnecessary contention 
> between call handlers.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5139) Remove redundant -R option from setrep

2013-09-23 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5139?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774799#comment-13774799
 ] 

Hudson commented on HDFS-5139:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #4455 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/4455/])
HDFS-5139. Remove redundant -R option from setrep. (arp: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1525659)
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/SetReplication.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/site/apt/FileSystemShell.apt.vm
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/resources/testConf.xml
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSShell.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/resources/testHDFSConf.xml


> Remove redundant -R option from setrep
> --
>
> Key: HDFS-5139
> URL: https://issues.apache.org/jira/browse/HDFS-5139
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: tools
>Affects Versions: 3.0.0, 1.3.0
>Reporter: Arpit Agarwal
>Assignee: Arpit Agarwal
> Attachments: HDFS-5139.01.patch, HDFS-5139.02.patch, 
> HDFS-5139.03.patch, HDFS-5139.04.patch
>
>
> The -R option to setrep is redundant because it is required for directory 
> targets and ignored for file targets.
> We can just remove the option and make -R the default for directories.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5139) Remove redundant -R option from setrep

2013-09-23 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5139?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774775#comment-13774775
 ] 

Suresh Srinivas commented on HDFS-5139:
---

+1 for the change. 

> Remove redundant -R option from setrep
> --
>
> Key: HDFS-5139
> URL: https://issues.apache.org/jira/browse/HDFS-5139
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: tools
>Affects Versions: 3.0.0, 1.3.0
>Reporter: Arpit Agarwal
>Assignee: Arpit Agarwal
> Attachments: HDFS-5139.01.patch, HDFS-5139.02.patch, 
> HDFS-5139.03.patch, HDFS-5139.04.patch
>
>
> The -R option to setrep is redundant because it is required for directory 
> targets and ignored for file targets.
> We can just remove the option and make -R the default for directories.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-5248) Checkpointing fails in uploading the Image to the Namenode if the merge requires more than 10 hours in secure cluster

2013-09-23 Thread Benoy Antony (JIRA)
Benoy Antony created HDFS-5248:
--

 Summary: Checkpointing fails in uploading the Image to the 
Namenode if the merge requires more than 10 hours in secure cluster
 Key: HDFS-5248
 URL: https://issues.apache.org/jira/browse/HDFS-5248
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 1.2.1, 2.1.0-beta
Reporter: Benoy Antony
Assignee: Benoy Antony


If the merge takes up more than 10 hours, the TGT of the SNN expires and SNN 
will fail to upload the image to the Namenode



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5248) Checkpointing fails in uploading the Image to the Namenode if the merge requires more than 10 hours in secure cluster

2013-09-23 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-5248:
---

Attachment: HDFS-5248-1.patch

Attaching a patch for branch-1.2

Will upload the patch for trunk soon.

> Checkpointing fails in uploading the Image to the Namenode if the merge 
> requires more than 10 hours in secure cluster
> -
>
> Key: HDFS-5248
> URL: https://issues.apache.org/jira/browse/HDFS-5248
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.0-beta, 1.2.1
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-5248-1.patch
>
>
> If the merge takes up more than 10 hours, the TGT of the SNN expires and SNN 
> will fail to upload the image to the Namenode

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5248) Checkpointing fails in uploading the Image to the Namenode if the merge requires more than 10 hours in secure cluster

2013-09-23 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-5248:
---

Attachment: (was: HDFS-5248-1.patch)

> Checkpointing fails in uploading the Image to the Namenode if the merge 
> requires more than 10 hours in secure cluster
> -
>
> Key: HDFS-5248
> URL: https://issues.apache.org/jira/browse/HDFS-5248
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.0-beta, 1.2.1
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-5248-1.patch
>
>
> If the merge takes up more than 10 hours, the TGT of the SNN expires and SNN 
> will fail to upload the image to the Namenode

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5139) Remove redundant -R option from setrep

2013-09-23 Thread Arpit Agarwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5139?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arpit Agarwal updated HDFS-5139:


   Resolution: Fixed
Fix Version/s: 2.2.0
   Status: Resolved  (was: Patch Available)

Thanks Suresh. I have committed this to trunk, branch-2 and branch-2.1-beta. 
Fix version has been set to 2.2.0.

> Remove redundant -R option from setrep
> --
>
> Key: HDFS-5139
> URL: https://issues.apache.org/jira/browse/HDFS-5139
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: tools
>Affects Versions: 3.0.0, 1.3.0
>Reporter: Arpit Agarwal
>Assignee: Arpit Agarwal
> Fix For: 2.2.0
>
> Attachments: HDFS-5139.01.patch, HDFS-5139.02.patch, 
> HDFS-5139.03.patch, HDFS-5139.04.patch
>
>
> The -R option to setrep is redundant because it is required for directory 
> targets and ignored for file targets.
> We can just remove the option and make -R the default for directories.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5248) Checkpointing fails in uploading the Image to the Namenode if the merge requires more than 10 hours in secure cluster

2013-09-23 Thread Kai Zheng (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774813#comment-13774813
 ] 

Kai Zheng commented on HDFS-5248:
-

I guess there might be a place calling ugi.doAs or like that in the callers of 
putFSImage(). If so, perhaps there adding the checkTGTAndReloginFromKeytab() 
call might work with minimized change. HADOOP-9797 is working on something that 
does such job automatically. For ugi logined from keytab or ticket cache, it 
has choice to be automatically renewed.

> Checkpointing fails in uploading the Image to the Namenode if the merge 
> requires more than 10 hours in secure cluster
> -
>
> Key: HDFS-5248
> URL: https://issues.apache.org/jira/browse/HDFS-5248
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.0-beta, 1.2.1
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-5248-1.patch
>
>
> If the merge takes up more than 10 hours, the TGT of the SNN expires and SNN 
> will fail to upload the image to the Namenode

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5248) Checkpointing fails in uploading the Image to the Namenode if the merge requires more than 10 hours in secure cluster

2013-09-23 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-5248:
---

Attachment: HDFS-5248-1.patch

> Checkpointing fails in uploading the Image to the Namenode if the merge 
> requires more than 10 hours in secure cluster
> -
>
> Key: HDFS-5248
> URL: https://issues.apache.org/jira/browse/HDFS-5248
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.0-beta, 1.2.1
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-5248-1.patch
>
>
> If the merge takes up more than 10 hours, the TGT of the SNN expires and SNN 
> will fail to upload the image to the Namenode

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5216) NumDecomDeadDataNodes not returning correct number of dead decommissioned nodes

2013-09-23 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5216?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775551#comment-13775551
 ] 

Hadoop QA commented on HDFS-5216:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12604543/HDFS-5216.diff
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/5018//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5018//console

This message is automatically generated.

> NumDecomDeadDataNodes not returning correct number of dead decommissioned 
> nodes 
> 
>
> Key: HDFS-5216
> URL: https://issues.apache.org/jira/browse/HDFS-5216
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.0-beta
>Reporter: Trevor Lorimer
>Assignee: Trevor Lorimer
> Attachments: HDFS-5216.diff, HDFS-5216.diff
>
>
> For HDFS-4860 I essentially copied the process in 
> NamenodejspHelper.generateHealthReport(), so it would be in sync with the 
> original dfsHealth.jsp.
> However looking at this now there may be a bug? in 
> getNumDecomDeadDataNodes(), where:
> getBlockManager().getDatanodeManager().fetchDatanodes(dead, null, true);
> 
> Where the parameter true indicates that decommissioned nodes should be 
> removed from the list.
> If the flag is true fetchDatanodes calls removeDecomNodeFromList, which will 
> remove a node if an existing datanode does not appear in both include or 
> exclude lists and it has been decommissioned.
> If I am looking to return the Number of Dead Decommissioned Nodes, should I 
> change the remove decommissioned nodes flag to False? i.e.:
> getBlockManager().getDatanodeManager().fetchDatanodes(null, dead, false);

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5246) Hadoop nfs server binds to port 2049 which is the same as Linux nfs server

2013-09-23 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5246?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-5246:
-

Target Version/s: 2.1.0-beta, 3.0.0, 2.1.1-beta  (was: 3.0.0, 2.1.0-beta, 
2.1.1-beta)
  Status: Patch Available  (was: Open)

> Hadoop nfs server binds to port 2049 which is the same as Linux nfs server
> --
>
> Key: HDFS-5246
> URL: https://issues.apache.org/jira/browse/HDFS-5246
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM Java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: HDFS-5246.patch
>
>
> Hadoop nfs binds the nfs server to port 2049,
> which is also the default port that Linux nfs uses. If Linux nfs is already 
> running on the machine then Hadoop nfs will not be albe to start.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5246) Hadoop nfs server binds to port 2049 which is the same as Linux nfs server

2013-09-23 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774792#comment-13774792
 ] 

Brandon Li commented on HDFS-5246:
--

{quote}"...2049, which is also the default port that Linux nfs uses. If Linux 
nfs is already running on the machine then Hadoop nfs will not be albe to 
start."{quote}

There can be only one NFSv3 server running on the same host no matter which 
port is configured. right?

Regardless, this patch is a good idea. In the patch, you set the default port 
to 2079 for tests, so the NFS test can run when an NFS server is using the 
default port 2049 on the same host. This makes the testing more convenient.


 




> Hadoop nfs server binds to port 2049 which is the same as Linux nfs server
> --
>
> Key: HDFS-5246
> URL: https://issues.apache.org/jira/browse/HDFS-5246
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM Java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: HDFS-5246.patch
>
>
> Hadoop nfs binds the nfs server to port 2049,
> which is also the default port that Linux nfs uses. If Linux nfs is already 
> running on the machine then Hadoop nfs will not be albe to start.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4971) Move IO operations out of locking in OpenFileCtx

2013-09-23 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4971?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775568#comment-13775568
 ] 

Jing Zhao commented on HDFS-4971:
-

Thanks again Brandon! I will commit the patch soon.

> Move IO operations out of locking in OpenFileCtx
> 
>
> Key: HDFS-4971
> URL: https://issues.apache.org/jira/browse/HDFS-4971
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Jing Zhao
>Assignee: Jing Zhao
> Attachments: HDFS-4971.000.patch, HDFS-4971.001.patch, 
> HDFS-4971.002.patch, HDFS-4971.003.patch, HDFS-4971.004.patch, 
> HDFS-4971.005.patch, HDFS-4971.006.patch, HDFS-4971.007.patch
>
>
> Currently some IO operations (such as writing data to HDFS and dumping to 
> local disk) in OpenFileCtx may hold a lock which can block processing 
> incoming writing requests. This jira aims to optimize OpenFileCtx and move 
> the IO operations out of the locking.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5246) Hadoop nfs server binds to port 2049 which is the same as Linux nfs server

2013-09-23 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775574#comment-13775574
 ] 

Brandon Li commented on HDFS-5246:
--

The patch looks good to me. Some minor comments:
1. "nfsPort" can be final in Nfs3Base
2. nfs3.server.port could be a better name than nfs.server.port 

> Hadoop nfs server binds to port 2049 which is the same as Linux nfs server
> --
>
> Key: HDFS-5246
> URL: https://issues.apache.org/jira/browse/HDFS-5246
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM Java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: HDFS-5246.patch
>
>
> Hadoop nfs binds the nfs server to port 2049,
> which is also the default port that Linux nfs uses. If Linux nfs is already 
> running on the machine then Hadoop nfs will not be albe to start.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5191) revisit zero-copy API in FSDataInputStream to make it more intuitive

2013-09-23 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5191?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775611#comment-13775611
 ] 

Chris Nauroth commented on HDFS-5191:
-

It looks like there are still some import-only changes in {{DFSUtil}}.

{code}
if (buffer == null) {
  throw new UnsupportedOperationException("zero-copy reads " +
  "were not available, and the ByteBufferPool did not provide " +
  "us with a " + (useDirect ? " direct" : "indirect") +
  "buffer.");
}
{code}

Minor nit: this exception message would have an extraneous space in the direct 
case (i.e. "did not provide us with a  direct"), and no space between the last 
two words (i.e. "did not provide us with a indirectbuffer").

{code}
  while (true) {
countingVisitor.reset();
mmapManager.visitEvictable(countingVisitor);
if (0 == countingVisitor.count) {
  break;
}
  }
{code}

This test would cause an infinite loop if a bug was introduced that left mmaps 
lingering in evictable state, which could be hard to diagnose.  Should we use 
{{GenericTestUtils#waitFor}}, so that there is a timeout?

Lastly, can you help clarify something about the data structure behind 
{{IdentityHashStore}} for me?  In particular, I'm wondering about the math in 
{{realloc}}:

{code}
  private void realloc(int newCapacity) {
Preconditions.checkArgument(newCapacity > 0);
Object prevBuffer[] = buffer;
this.capacity = newCapacity;
int numElements = 1 + (2 * newCapacity);
this.buffer = new Object[2 * numElements];
this.numInserted = 0;
if (prevBuffer != null) {
  for (int i = 0; i < prevBuffer.length; i += 2) {
if (prevBuffer[i] != null) {
  putInternal(prevBuffer[i], prevBuffer[i + 1]);
}
  }
}
  }
{code}

{{put}} will call {{realloc}} to double capacity when needed.  {{numElements}} 
is double of the new capacity plus one extra.  Then, the size is doubled again 
for allocating the new array.  Therefore the growth pattern looks like:

capacity == 2, buffer.length == 10
capacity == 4, buffer.length == 18
capacity == 8, buffer.length == 34

It looks like this causes a fair amount of extra unused slack in the array, 
because only 2 * {{numInserted}} elements are used at a time.  Is the slack 
required, or should {{realloc}} only double the size once instead of twice?  
Also, I wasn't sure why we needed 1 extra in the calculation of 
{{numElements}}.  Is that a defensive thing so that the loop in {{putInternal}} 
always terminates?  I would expect the check of capacity in {{put}} to be 
sufficient to prevent that without needing an extra sentinel element.

On to the libhdfs changes...


> revisit zero-copy API in FSDataInputStream to make it more intuitive
> 
>
> Key: HDFS-5191
> URL: https://issues.apache.org/jira/browse/HDFS-5191
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: hdfs-client, libhdfs
>Affects Versions: HDFS-4949
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
> Attachments: HDFS-5191-caching.001.patch, 
> HDFS-5191-caching.003.patch, HDFS-5191-caching.006.patch, 
> HDFS-5191-caching.007.patch, HDFS-5191-caching.008.patch
>
>
> As per the discussion on HDFS-4953, we should revisit the zero-copy API to 
> make it more intuitive for new users.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5230) Introduce RpcInfo to decouple XDR classes from the RPC API

2013-09-23 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775576#comment-13775576
 ] 

Jing Zhao commented on HDFS-5230:
-

The patch looks great to me. Some comments and thoughts:

# RpcInfo's constructor, we do not need to call super() explicitly.
# It will be helpful to add more javadoc for RpcInfo and RpcResponse.
# Similarly, we may want more javadoc for RpcxxxStage classes in RpcUtil to 
clearly state the next stage for each stage (i.e., the whole pipeline).
# Creating Verifier#VERIFIER_RPC_GSS may be unnecessary here since we need to 
create different GSS Verifier objects each time. This part we can leave it as 
it is and update it in HDFS-5086.
# Currently we may hide ChannelBuffer creation into the RpcResponse. I.e., by 
introducing a new constructor or static method in RpcResponse which takes 
XDR/(Rpc Header + Rpc Response Body) as argument(s), we can avoid calling
{code}
ChannelBuffer buf = 
ChannelBuffers.wrappedBuffer(reply.asReadOnlyWrap().buffer());
{code}
outside of RpcResponse. 
# Currently we call Nfs3Response#writeHeaderAndResponse to generate an XDR 
containing complete RPC response. We should put this logic into the new 
RpcResponse class so that NFS3Response and its subclasses only need to handle 
response information returned from NFS handlers.

> Introduce RpcInfo to decouple XDR classes from the RPC API
> --
>
> Key: HDFS-5230
> URL: https://issues.apache.org/jira/browse/HDFS-5230
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-5230.002.patch
>
>
> The XDR class is one fundamental aspect in the current implementation of NFS 
> server. While the client might potentially have a higher level APIs, it also 
> requires redundant copying since the upstream clients have insufficient 
> information.
> This JIRA introduces a new class, RpcInfo, which (1) decouples XDR from the 
> APIs, turning it into a utility class, and (2) exposes ChannelBuffer directly 
> to the client in order to open the opportunity for avoid copying.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5139) Remove redundant -R option from setrep

2013-09-23 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5139?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774820#comment-13774820
 ] 

Hudson commented on HDFS-5139:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #4456 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/4456/])
HDFS-5139. Remove redundant -R option from setrep (update CHANGES.txt). (arp: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1525665)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


> Remove redundant -R option from setrep
> --
>
> Key: HDFS-5139
> URL: https://issues.apache.org/jira/browse/HDFS-5139
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: tools
>Affects Versions: 3.0.0, 1.3.0
>Reporter: Arpit Agarwal
>Assignee: Arpit Agarwal
> Fix For: 2.2.0
>
> Attachments: HDFS-5139.01.patch, HDFS-5139.02.patch, 
> HDFS-5139.03.patch, HDFS-5139.04.patch
>
>
> The -R option to setrep is redundant because it is required for directory 
> targets and ignored for file targets.
> We can just remove the option and make -R the default for directories.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4971) Move IO operations out of locking in OpenFileCtx

2013-09-23 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4971?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775617#comment-13775617
 ] 

Hudson commented on HDFS-4971:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #4458 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/4458/])
HDFS-4971. Move IO operations out of locking in OpenFileCtx. Contributed by 
Jing Zhao and Brandon Li. (jing9: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1525681)
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/AsyncDataService.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OffsetRange.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/WriteCtx.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/WriteManager.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/test/java/org/apache/hadoop/hdfs/nfs/nfs3/TestOffsetRange.java
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


> Move IO operations out of locking in OpenFileCtx
> 
>
> Key: HDFS-4971
> URL: https://issues.apache.org/jira/browse/HDFS-4971
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Jing Zhao
>Assignee: Jing Zhao
> Attachments: HDFS-4971.000.patch, HDFS-4971.001.patch, 
> HDFS-4971.002.patch, HDFS-4971.003.patch, HDFS-4971.004.patch, 
> HDFS-4971.005.patch, HDFS-4971.006.patch, HDFS-4971.007.patch
>
>
> Currently some IO operations (such as writing data to HDFS and dumping to 
> local disk) in OpenFileCtx may hold a lock which can block processing 
> incoming writing requests. This jira aims to optimize OpenFileCtx and move 
> the IO operations out of the locking.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4971) Move IO operations out of locking in OpenFileCtx

2013-09-23 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4971?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774750#comment-13774750
 ] 

Hadoop QA commented on HDFS-4971:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12604472/HDFS-4971.007.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs-nfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/5019//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5019//console

This message is automatically generated.

> Move IO operations out of locking in OpenFileCtx
> 
>
> Key: HDFS-4971
> URL: https://issues.apache.org/jira/browse/HDFS-4971
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Jing Zhao
>Assignee: Jing Zhao
> Attachments: HDFS-4971.000.patch, HDFS-4971.001.patch, 
> HDFS-4971.002.patch, HDFS-4971.003.patch, HDFS-4971.004.patch, 
> HDFS-4971.005.patch, HDFS-4971.006.patch, HDFS-4971.007.patch
>
>
> Currently some IO operations (such as writing data to HDFS and dumping to 
> local disk) in OpenFileCtx may hold a lock which can block processing 
> incoming writing requests. This jira aims to optimize OpenFileCtx and move 
> the IO operations out of the locking.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5191) revisit zero-copy API in FSDataInputStream to make it more intuitive

2013-09-23 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5191?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774757#comment-13774757
 ] 

Colin Patrick McCabe commented on HDFS-5191:


I decided to test {{TestByteBufferUtil#fallbackRead}} inside 
{{TestEnhancedByteBufferAccess}}.  The other file is not needed-- I will remove.

Removed extra imports in DFSUtil.

> revisit zero-copy API in FSDataInputStream to make it more intuitive
> 
>
> Key: HDFS-5191
> URL: https://issues.apache.org/jira/browse/HDFS-5191
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: hdfs-client, libhdfs
>Affects Versions: HDFS-4949
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
> Attachments: HDFS-5191-caching.001.patch, 
> HDFS-5191-caching.003.patch, HDFS-5191-caching.006.patch, 
> HDFS-5191-caching.007.patch, HDFS-5191-caching.008.patch
>
>
> As per the discussion on HDFS-4953, we should revisit the zero-copy API to 
> make it more intuitive for new users.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5191) revisit zero-copy API in FSDataInputStream to make it more intuitive

2013-09-23 Thread Colin Patrick McCabe (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5191?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-5191:
---

Attachment: HDFS-5191-caching.008.patch

> revisit zero-copy API in FSDataInputStream to make it more intuitive
> 
>
> Key: HDFS-5191
> URL: https://issues.apache.org/jira/browse/HDFS-5191
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: hdfs-client, libhdfs
>Affects Versions: HDFS-4949
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
> Attachments: HDFS-5191-caching.001.patch, 
> HDFS-5191-caching.003.patch, HDFS-5191-caching.006.patch, 
> HDFS-5191-caching.007.patch, HDFS-5191-caching.008.patch
>
>
> As per the discussion on HDFS-4953, we should revisit the zero-copy API to 
> make it more intuitive for new users.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4971) Move IO operations out of locking in OpenFileCtx

2013-09-23 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4971?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao updated HDFS-4971:


   Resolution: Fixed
Fix Version/s: 2.1.1-beta
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

I've committed this to trunk, branch-2 and branch-2.1-beta.

> Move IO operations out of locking in OpenFileCtx
> 
>
> Key: HDFS-4971
> URL: https://issues.apache.org/jira/browse/HDFS-4971
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Jing Zhao
>Assignee: Jing Zhao
> Fix For: 2.1.1-beta
>
> Attachments: HDFS-4971.000.patch, HDFS-4971.001.patch, 
> HDFS-4971.002.patch, HDFS-4971.003.patch, HDFS-4971.004.patch, 
> HDFS-4971.005.patch, HDFS-4971.006.patch, HDFS-4971.007.patch
>
>
> Currently some IO operations (such as writing data to HDFS and dumping to 
> local disk) in OpenFileCtx may hold a lock which can block processing 
> incoming writing requests. This jira aims to optimize OpenFileCtx and move 
> the IO operations out of the locking.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5245) Retry policies should return false rather than rethrow an exception in branch-1

2013-09-23 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5245?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775559#comment-13775559
 ] 

Suresh Srinivas commented on HDFS-5245:
---

[~wheat9], Jenkins only runs the submitted patch against trunk. There is no 
need to submit a patch, if it is for non-trunk.

> Retry policies should return false rather than rethrow an exception in 
> branch-1
> ---
>
> Key: HDFS-5245
> URL: https://issues.apache.org/jira/browse/HDFS-5245
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 1.2.0, 1.2.1
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-5245.000.patch
>
>
> A RetryPolicy inspects the exception generated when performing work and 
> decides to try the work again.
> In branch-1 a RetryPolicy either returns false or rethrow the exception in 
> the shouldRetry() function to indicate it decides not to retry. However, 
> rethrowing an exception confuses the upstream code and causes it to print out 
> excessive warning messages.
> This bug is fixed in trunk and branch-2.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5228) The RemoteIterator returned by DistributedFileSystem.listFiles(..) may throw NPE

2013-09-23 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5228?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775561#comment-13775561
 ] 

Andrew Wang commented on HDFS-5228:
---

Chris, I'm +1 for your v2 patch, thanks for the help.

Hey Nicholas:

bq. Which part of HADOOP-8040? Where are the new tests?

HADOOP-9370 and HADOOP-9355 extended the FileContext symlink tests to 
FileSystem. These are both linked on HADOOP-8040. You can also see this if you 
look for tests in the HADOOP-9418 patch.

bq. Andrew, you have not answered my following question. I hope that you are 
not avoiding it.

I'm going to quote myself to answer the question:

bq.  Basically all DFS methods are supposed to pass Paths through 
fixRelativePart and then getPathName, it just looks like we missed this one. It 
might be worth auditing DFS to make sure we didn't miss anything else.

Chris already did this audit, perhaps you can ask him if you need additional 
in-person clarification.

> The RemoteIterator returned by DistributedFileSystem.listFiles(..) may throw 
> NPE
> 
>
> Key: HDFS-5228
> URL: https://issues.apache.org/jira/browse/HDFS-5228
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.1.0-beta
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
>Priority: Blocker
> Attachments: h5228_20130919.patch, h5228_20130919_test.patch, 
> HDFS-5228.2.patch
>
>
> Get a RemoteIterator from DistributedFileSystem.listFiles(..) with a relative 
> path.  Then, it will result a NullPointerException when calling hasNext() 
> from the RemoteIterator.
> This bug was discovered by Arnaud:
> http://hortonworks.com/community/forums/topic/new-bug-in-hdfs-listfiles-method/

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5239) Allow FSNamesystem lock fairness to be configurable

2013-09-23 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5239?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13774645#comment-13774645
 ] 

Hudson commented on HDFS-5239:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #4452 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/4452/])
HDFS-5239.  Allow FSNamesystem lock fairness to be configurable (daryn) (daryn: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1525624)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSNamesystem.java


> Allow FSNamesystem lock fairness to be configurable
> ---
>
> Key: HDFS-5239
> URL: https://issues.apache.org/jira/browse/HDFS-5239
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
> Fix For: 3.0.0, 2.3.0
>
> Attachments: HDFS-5239.patch
>
>
> The fairness of the {{FSNamesystem#fsLock}} is hardcoded to {{true}}.  Using 
> an unfair lock provides a negligible increase to throughput.  However this is 
> due to bottlenecks elsewhere in the system. In combination with other 
> changes, such as RPC layer and audit logging, preliminary tests show up to a 
> 5X improvement for a read heavy workload.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5246) Hadoop nfs server binds to port 2049 which is the same as Linux nfs server

2013-09-23 Thread Jinghui Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775693#comment-13775693
 ] 

Jinghui Wang commented on HDFS-5246:


"default port to 2079 for tests, so the NFS test can run when an NFS server is 
using the default port 2049 on the same host. This makes the testing more 
convenient"
That's exactly the reason why we want to make the change as we encountered the 
failure during a UT job when NFS was running in the testing env.

Attach new patch HDFS-5246-2.patch here with changes for the comments.

> Hadoop nfs server binds to port 2049 which is the same as Linux nfs server
> --
>
> Key: HDFS-5246
> URL: https://issues.apache.org/jira/browse/HDFS-5246
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM Java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: HDFS-5246-2.patch, HDFS-5246.patch
>
>
> Hadoop nfs binds the nfs server to port 2049,
> which is also the default port that Linux nfs uses. If Linux nfs is already 
> running on the machine then Hadoop nfs will not be albe to start.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5246) Hadoop nfs server binds to port 2049 which is the same as Linux nfs server

2013-09-23 Thread Jinghui Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5246?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jinghui Wang updated HDFS-5246:
---

Attachment: HDFS-5246-2.patch

> Hadoop nfs server binds to port 2049 which is the same as Linux nfs server
> --
>
> Key: HDFS-5246
> URL: https://issues.apache.org/jira/browse/HDFS-5246
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM Java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: HDFS-5246-2.patch, HDFS-5246.patch
>
>
> Hadoop nfs binds the nfs server to port 2049,
> which is also the default port that Linux nfs uses. If Linux nfs is already 
> running on the machine then Hadoop nfs will not be albe to start.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5137) Improve WriteManager for processing stable write requests and commit requests

2013-09-23 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5137?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775701#comment-13775701
 ] 

Brandon Li commented on HDFS-5137:
--

Hi Jing, you may want to rebase the patch since HDFS-4971 has been resolved.

> Improve WriteManager for processing stable write requests and commit requests
> -
>
> Key: HDFS-5137
> URL: https://issues.apache.org/jira/browse/HDFS-5137
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Affects Versions: 3.0.0
>Reporter: Jing Zhao
>Assignee: Jing Zhao
>Priority: Minor
> Attachments: HDFS-5137.patch
>
>
> WriteManager#handleCommit needs to block until all the data before the 
> commitOffset has been received. This jira plans to improve the current 
> concurrency implementation for this blocking call.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5246) Hadoop nfs server binds to port 2049 which is the same as Linux nfs server

2013-09-23 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775708#comment-13775708
 ] 

Brandon Li commented on HDFS-5246:
--

{quote}...we encountered the failure during a UT job when NFS was running in 
the testing env.{quote}
I suspect you might have same port problem with Mount daemon. 

> Hadoop nfs server binds to port 2049 which is the same as Linux nfs server
> --
>
> Key: HDFS-5246
> URL: https://issues.apache.org/jira/browse/HDFS-5246
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM Java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: HDFS-5246-2.patch, HDFS-5246.patch
>
>
> Hadoop nfs binds the nfs server to port 2049,
> which is also the default port that Linux nfs uses. If Linux nfs is already 
> running on the machine then Hadoop nfs will not be albe to start.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5230) Introduce RpcInfo to decouple XDR classes from the RPC API

2013-09-23 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5230?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-5230:
-

Attachment: HDFS-5230.003.patch

Address Jing's comments. Add JavaDoc to document how the pipeline works.

I agree with Jing that more refactoring can be addressed in HDFS-5086.

> Introduce RpcInfo to decouple XDR classes from the RPC API
> --
>
> Key: HDFS-5230
> URL: https://issues.apache.org/jira/browse/HDFS-5230
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-5230.002.patch, HDFS-5230.003.patch
>
>
> The XDR class is one fundamental aspect in the current implementation of NFS 
> server. While the client might potentially have a higher level APIs, it also 
> requires redundant copying since the upstream clients have insufficient 
> information.
> This JIRA introduces a new class, RpcInfo, which (1) decouples XDR from the 
> APIs, turning it into a utility class, and (2) exposes ChannelBuffer directly 
> to the client in order to open the opportunity for avoid copying.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5246) Hadoop nfs server binds to port 2049 which is the same as Linux nfs server

2013-09-23 Thread Jinghui Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775716#comment-13775716
 ] 

Jinghui Wang commented on HDFS-5246:


Right. The test was not able to start the Mount daemon since NFS was using port 
2049.

> Hadoop nfs server binds to port 2049 which is the same as Linux nfs server
> --
>
> Key: HDFS-5246
> URL: https://issues.apache.org/jira/browse/HDFS-5246
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM Java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: HDFS-5246-2.patch, HDFS-5246.patch
>
>
> Hadoop nfs binds the nfs server to port 2049,
> which is also the default port that Linux nfs uses. If Linux nfs is already 
> running on the machine then Hadoop nfs will not be albe to start.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-5249) Fix dumper thread which may die silently

2013-09-23 Thread Brandon Li (JIRA)
Brandon Li created HDFS-5249:


 Summary: Fix dumper thread which may die silently
 Key: HDFS-5249
 URL: https://issues.apache.org/jira/browse/HDFS-5249
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: nfs
Reporter: Brandon Li
Assignee: Brandon Li


Dumper thread can get an NPE when the WriteCtx it's about to work on is just 
deleted by write back thread. 

A dead dumper thread could cause out-of-memory error when too many pending 
writes accumulated for one opened file.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5249) Fix dumper thread which may die silently

2013-09-23 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-5249:
-

Status: Patch Available  (was: Open)

> Fix dumper thread which may die silently
> 
>
> Key: HDFS-5249
> URL: https://issues.apache.org/jira/browse/HDFS-5249
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Brandon Li
>Assignee: Brandon Li
> Attachments: HDFS-5249.patch
>
>
> Dumper thread can get an NPE when the WriteCtx it's about to work on is just 
> deleted by write back thread. 
> A dead dumper thread could cause out-of-memory error when too many pending 
> writes accumulated for one opened file.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5249) Fix dumper thread which may die silently

2013-09-23 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-5249:
-

Attachment: HDFS-5249.patch

> Fix dumper thread which may die silently
> 
>
> Key: HDFS-5249
> URL: https://issues.apache.org/jira/browse/HDFS-5249
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Brandon Li
>Assignee: Brandon Li
> Attachments: HDFS-5249.patch
>
>
> Dumper thread can get an NPE when the WriteCtx it's about to work on is just 
> deleted by write back thread. 
> A dead dumper thread could cause out-of-memory error when too many pending 
> writes accumulated for one opened file.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5246) Hadoop nfs server binds to port 2049 which is the same as Linux nfs server

2013-09-23 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775736#comment-13775736
 ] 

Brandon Li commented on HDFS-5246:
--

Mount daemon(implemented inside HDFS-NFS server) uses a different port, default 
as 4242, as in class RpcProgramMountd.
That port is also need to be made configurable as you did for NFS server, so 
that the Mount daemon(in the HDFS-NFS server) in tests will not use the same 
default port which is already used by the native Mountd on Linux.

I think either way is ok: make mount port configurable here or in a different 
JIRA.

> Hadoop nfs server binds to port 2049 which is the same as Linux nfs server
> --
>
> Key: HDFS-5246
> URL: https://issues.apache.org/jira/browse/HDFS-5246
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM Java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: HDFS-5246-2.patch, HDFS-5246.patch
>
>
> Hadoop nfs binds the nfs server to port 2049,
> which is also the default port that Linux nfs uses. If Linux nfs is already 
> running on the machine then Hadoop nfs will not be albe to start.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-5250) Enable one url to browse the entire federated hdfs file system on dfsclusterhealth summary page

2013-09-23 Thread Vrushali C (JIRA)
Vrushali C created HDFS-5250:


 Summary: Enable one url to browse the entire federated hdfs file 
system on dfsclusterhealth summary page
 Key: HDFS-5250
 URL: https://issues.apache.org/jira/browse/HDFS-5250
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Vrushali C




Ideally, we should have one url to browse the entire federated file system on 
the main dfsclusterhealth summary page along with the list of namenodes in the 
cluster


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5246) Hadoop nfs server binds to port 2049 which is the same as Linux nfs server

2013-09-23 Thread Haohui Mai (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775751#comment-13775751
 ] 

Haohui Mai commented on HDFS-5246:
--

In my opinion, Portmap is designed to take care of the exact issue -- any 
servers that register themselves in portmap server should use port 0, and 
portmap should take care of the rest.

> Hadoop nfs server binds to port 2049 which is the same as Linux nfs server
> --
>
> Key: HDFS-5246
> URL: https://issues.apache.org/jira/browse/HDFS-5246
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM Java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: HDFS-5246-2.patch, HDFS-5246.patch
>
>
> Hadoop nfs binds the nfs server to port 2049,
> which is also the default port that Linux nfs uses. If Linux nfs is already 
> running on the machine then Hadoop nfs will not be albe to start.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5191) revisit zero-copy API in FSDataInputStream to make it more intuitive

2013-09-23 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5191?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775753#comment-13775753
 ] 

Colin Patrick McCabe commented on HDFS-5191:


re: test timouts, I like to add them at the test level using  
\@Test(timeout=...), rather than looking at every part of the test.

re: hash table sizes.  Any time you insert something into IdentityHashStore, 
you use two array slots... one for the key, and another for the value.  The 
other 2x factor is because the hash table should stay half empty, to avoid the 
risk of collisions.  In other words, we have a load factor of 0.50.  This is 
especially important since every element takes up space (we don't use separate 
chaining).

After thinking about it, I don't think we need the extra +1 element.  It seems 
that System#identityHashCode provides a well-distributed enough hash that 
dividing by a power of two size works well.

IdentityHashStore was necessary because ByteBuffer#hash and ByteBuffer#equals 
were just very unsuitable for what I was trying to do.  And there's no way to 
parameterize HashTable to use different functions.  Another nice advantage of 
IdentityHashStore is that it does zero allocations, unless you are growing the 
hash table.  This might seem like a trivial point, but given how frequently 
read is called, it's nice to avoid generating garbage.  We learned that when 
dealing with the edit log code...

> revisit zero-copy API in FSDataInputStream to make it more intuitive
> 
>
> Key: HDFS-5191
> URL: https://issues.apache.org/jira/browse/HDFS-5191
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: hdfs-client, libhdfs
>Affects Versions: HDFS-4949
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
> Attachments: HDFS-5191-caching.001.patch, 
> HDFS-5191-caching.003.patch, HDFS-5191-caching.006.patch, 
> HDFS-5191-caching.007.patch, HDFS-5191-caching.008.patch
>
>
> As per the discussion on HDFS-4953, we should revisit the zero-copy API to 
> make it more intuitive for new users.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5249) Fix dumper thread which may die silently

2013-09-23 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775755#comment-13775755
 ] 

Jing Zhao commented on HDFS-5249:
-

Patch looks pretty good. One comment is that for the dumper thread, we can put 
the try-catch under the while loop:
{code}
  while (activeState && enabledDump) {
try {
  if (nonSequentialWriteInMemory.get() >= DUMP_WRITE_WATER_MARK) {
dump();
  }
  synchronized (OpenFileCtx.this) {
if (nonSequentialWriteInMemory.get() < DUMP_WRITE_WATER_MARK) {
  try {
OpenFileCtx.this.wait();
  } catch (InterruptedException e) {
  }
}
  }
} catch (Throwable t) {
  // print and process
}
  }
{code}

> Fix dumper thread which may die silently
> 
>
> Key: HDFS-5249
> URL: https://issues.apache.org/jira/browse/HDFS-5249
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Brandon Li
>Assignee: Brandon Li
> Attachments: HDFS-5249.patch
>
>
> Dumper thread can get an NPE when the WriteCtx it's about to work on is just 
> deleted by write back thread. 
> A dead dumper thread could cause out-of-memory error when too many pending 
> writes accumulated for one opened file.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-5251) Race between the initialization of NameNode and the http server

2013-09-23 Thread Haohui Mai (JIRA)
Haohui Mai created HDFS-5251:


 Summary: Race between the initialization of NameNode and the http 
server
 Key: HDFS-5251
 URL: https://issues.apache.org/jira/browse/HDFS-5251
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Haohui Mai


The constructor of NameNode starts a HTTP server before the FSNameSystem is 
initialized. Currently there is a race where the HTTP server can access the 
uninitialized namesystem variable, throwing a NullPointerException.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5191) revisit zero-copy API in FSDataInputStream to make it more intuitive

2013-09-23 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5191?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775765#comment-13775765
 ] 

Chris Nauroth commented on HDFS-5191:
-

bq. In other words, we have a load factor of 0.50.

That clears it right up for me.  Thanks!  Do you mind dropping a quick comment 
on the second 2x multiplication?

> revisit zero-copy API in FSDataInputStream to make it more intuitive
> 
>
> Key: HDFS-5191
> URL: https://issues.apache.org/jira/browse/HDFS-5191
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: hdfs-client, libhdfs
>Affects Versions: HDFS-4949
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
> Attachments: HDFS-5191-caching.001.patch, 
> HDFS-5191-caching.003.patch, HDFS-5191-caching.006.patch, 
> HDFS-5191-caching.007.patch, HDFS-5191-caching.008.patch
>
>
> As per the discussion on HDFS-4953, we should revisit the zero-copy API to 
> make it more intuitive for new users.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5249) Fix dumper thread which may die silently

2013-09-23 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775815#comment-13775815
 ] 

Hadoop QA commented on HDFS-5249:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12604683/HDFS-5249.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs-nfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/5020//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5020//console

This message is automatically generated.

> Fix dumper thread which may die silently
> 
>
> Key: HDFS-5249
> URL: https://issues.apache.org/jira/browse/HDFS-5249
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Brandon Li
>Assignee: Brandon Li
> Attachments: HDFS-5249.patch
>
>
> Dumper thread can get an NPE when the WriteCtx it's about to work on is just 
> deleted by write back thread. 
> A dead dumper thread could cause out-of-memory error when too many pending 
> writes accumulated for one opened file.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5137) Improve WriteManager for processing stable write requests and commit requests

2013-09-23 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5137?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao updated HDFS-5137:


Attachment: HDFS-5137.001.patch

Rebase the patch.

> Improve WriteManager for processing stable write requests and commit requests
> -
>
> Key: HDFS-5137
> URL: https://issues.apache.org/jira/browse/HDFS-5137
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Affects Versions: 3.0.0
>Reporter: Jing Zhao
>Assignee: Jing Zhao
>Priority: Minor
> Attachments: HDFS-5137.001.patch, HDFS-5137.patch
>
>
> WriteManager#handleCommit needs to block until all the data before the 
> commitOffset has been received. This jira plans to improve the current 
> concurrency implementation for this blocking call.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5137) Improve WriteManager for processing stable write requests and commit requests

2013-09-23 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5137?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775825#comment-13775825
 ] 

Jing Zhao commented on HDFS-5137:
-

Currently the commit request will block the processing thread. In general we 
should convert it to a non-blocking implementation. Maybe we can do it later in 
a separate jira.

> Improve WriteManager for processing stable write requests and commit requests
> -
>
> Key: HDFS-5137
> URL: https://issues.apache.org/jira/browse/HDFS-5137
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Affects Versions: 3.0.0
>Reporter: Jing Zhao
>Assignee: Jing Zhao
>Priority: Minor
> Attachments: HDFS-5137.001.patch, HDFS-5137.patch
>
>
> WriteManager#handleCommit needs to block until all the data before the 
> commitOffset has been received. This jira plans to improve the current 
> concurrency implementation for this blocking call.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5230) Introduce RpcInfo to decouple XDR classes from the RPC API

2013-09-23 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775830#comment-13775830
 ] 

Hadoop QA commented on HDFS-5230:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12604679/HDFS-5230.003.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-common-project/hadoop-nfs hadoop-hdfs-project/hadoop-hdfs-nfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/5022//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5022//console

This message is automatically generated.

> Introduce RpcInfo to decouple XDR classes from the RPC API
> --
>
> Key: HDFS-5230
> URL: https://issues.apache.org/jira/browse/HDFS-5230
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-5230.002.patch, HDFS-5230.003.patch
>
>
> The XDR class is one fundamental aspect in the current implementation of NFS 
> server. While the client might potentially have a higher level APIs, it also 
> requires redundant copying since the upstream clients have insufficient 
> information.
> This JIRA introduces a new class, RpcInfo, which (1) decouples XDR from the 
> APIs, turning it into a utility class, and (2) exposes ChannelBuffer directly 
> to the client in order to open the opportunity for avoid copying.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5222) Move block schedule information from DatanodeDescriptor to DatanodeStorageInfo

2013-09-23 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5222?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775837#comment-13775837
 ] 

Arpit Agarwal commented on HDFS-5222:
-

Hi Nicholas, couple of spurious newlines. +1 otherwise. Please feel free to 
commit after removing these.

{code}
   private State state;
+
   private long capacity;
   private long dfsUsed;
   private long remaining;
+
   private volatile BlockInfo blockList = null;
{code}

> Move block schedule information from DatanodeDescriptor to DatanodeStorageInfo
> --
>
> Key: HDFS-5222
> URL: https://issues.apache.org/jira/browse/HDFS-5222
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Attachments: h5222_20130819.patch
>
>
> In HDFS-4990, the block placement target type was changed from 
> DatanodeDescriptor to DatanodeStorageInfo.  The block schedule information, 
> such as the number of blocks scheduled for replication (i.e. 
> getBlocksScheduled()), should be moved from DatanodeDescriptor to 
> DatanodeStorageInfo.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5246) Hadoop nfs server binds to port 2049 which is the same as Linux nfs server

2013-09-23 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775841#comment-13775841
 ] 

Brandon Li commented on HDFS-5246:
--

As Haohui and I discussed offline, in theory using port 0 should work, but more 
people are used to the default port 2049 for NFS. Therefore, let's go with the 
approach in the current patch: make the port configurable, but use 2049 as 
default port for NFS.
 

> Hadoop nfs server binds to port 2049 which is the same as Linux nfs server
> --
>
> Key: HDFS-5246
> URL: https://issues.apache.org/jira/browse/HDFS-5246
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM Java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: HDFS-5246-2.patch, HDFS-5246.patch
>
>
> Hadoop nfs binds the nfs server to port 2049,
> which is also the default port that Linux nfs uses. If Linux nfs is already 
> running on the machine then Hadoop nfs will not be albe to start.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5246) Hadoop nfs server binds to port 2049 which is the same as Linux nfs server

2013-09-23 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775840#comment-13775840
 ] 

Hadoop QA commented on HDFS-5246:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12604673/HDFS-5246-2.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-common-project/hadoop-common hadoop-common-project/hadoop-nfs 
hadoop-hdfs-project/hadoop-hdfs-nfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/5021//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5021//console

This message is automatically generated.

> Hadoop nfs server binds to port 2049 which is the same as Linux nfs server
> --
>
> Key: HDFS-5246
> URL: https://issues.apache.org/jira/browse/HDFS-5246
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM Java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: HDFS-5246-2.patch, HDFS-5246.patch
>
>
> Hadoop nfs binds the nfs server to port 2049,
> which is also the default port that Linux nfs uses. If Linux nfs is already 
> running on the machine then Hadoop nfs will not be albe to start.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5246) Hadoop nfs server binds to port 2049 which is the same as Linux nfs server

2013-09-23 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775842#comment-13775842
 ] 

Brandon Li commented on HDFS-5246:
--

[~jwang302] Please let us know if you want to address the same issue for 
RpcProgramMountd in this JIRA.

> Hadoop nfs server binds to port 2049 which is the same as Linux nfs server
> --
>
> Key: HDFS-5246
> URL: https://issues.apache.org/jira/browse/HDFS-5246
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: nfs
>Affects Versions: 2.1.0-beta
> Environment: Red Hat Enterprise 6 with Sun Java 1.7 and IBM Java 1.6
>Reporter: Jinghui Wang
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: HDFS-5246-2.patch, HDFS-5246.patch
>
>
> Hadoop nfs binds the nfs server to port 2049,
> which is also the default port that Linux nfs uses. If Linux nfs is already 
> running on the machine then Hadoop nfs will not be albe to start.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5230) Introduce RpcInfo to decouple XDR classes from the RPC API

2013-09-23 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775851#comment-13775851
 ] 

Jing Zhao commented on HDFS-5230:
-

[~wheat9], I guess you want to move 5&6 to HDFS-5086? Looks fine to me. 
Could you post how you test the patch and the result of your test?

> Introduce RpcInfo to decouple XDR classes from the RPC API
> --
>
> Key: HDFS-5230
> URL: https://issues.apache.org/jira/browse/HDFS-5230
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-5230.002.patch, HDFS-5230.003.patch
>
>
> The XDR class is one fundamental aspect in the current implementation of NFS 
> server. While the client might potentially have a higher level APIs, it also 
> requires redundant copying since the upstream clients have insufficient 
> information.
> This JIRA introduces a new class, RpcInfo, which (1) decouples XDR from the 
> APIs, turning it into a utility class, and (2) exposes ChannelBuffer directly 
> to the client in order to open the opportunity for avoid copying.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5230) Introduce RpcInfo to decouple XDR classes from the RPC API

2013-09-23 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775854#comment-13775854
 ] 

Brandon Li commented on HDFS-5230:
--

Please give me a few hours. I will also take a look of the patch tomorrow if I 
can't get to it tonight.

> Introduce RpcInfo to decouple XDR classes from the RPC API
> --
>
> Key: HDFS-5230
> URL: https://issues.apache.org/jira/browse/HDFS-5230
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-5230.002.patch, HDFS-5230.003.patch
>
>
> The XDR class is one fundamental aspect in the current implementation of NFS 
> server. While the client might potentially have a higher level APIs, it also 
> requires redundant copying since the upstream clients have insufficient 
> information.
> This JIRA introduces a new class, RpcInfo, which (1) decouples XDR from the 
> APIs, turning it into a utility class, and (2) exposes ChannelBuffer directly 
> to the client in order to open the opportunity for avoid copying.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5249) Fix dumper thread which may die silently

2013-09-23 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-5249:
-

Attachment: HDFS-5249.02.patch

Upload a new patch to address Jing's comment.

> Fix dumper thread which may die silently
> 
>
> Key: HDFS-5249
> URL: https://issues.apache.org/jira/browse/HDFS-5249
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Brandon Li
>Assignee: Brandon Li
> Attachments: HDFS-5249.02.patch, HDFS-5249.patch
>
>
> Dumper thread can get an NPE when the WriteCtx it's about to work on is just 
> deleted by write back thread. 
> A dead dumper thread could cause out-of-memory error when too many pending 
> writes accumulated for one opened file.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5249) Fix dumper thread which may die silently

2013-09-23 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775859#comment-13775859
 ] 

Brandon Li commented on HDFS-5249:
--

No unit test included since it's difficult to time the race condition in a unit 
test. I manually tested file loading on a Linux cluster. 

> Fix dumper thread which may die silently
> 
>
> Key: HDFS-5249
> URL: https://issues.apache.org/jira/browse/HDFS-5249
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Brandon Li
>Assignee: Brandon Li
> Attachments: HDFS-5249.02.patch, HDFS-5249.patch
>
>
> Dumper thread can get an NPE when the WriteCtx it's about to work on is just 
> deleted by write back thread. 
> A dead dumper thread could cause out-of-memory error when too many pending 
> writes accumulated for one opened file.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5249) Fix dumper thread which may die silently

2013-09-23 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775860#comment-13775860
 ] 

Jing Zhao commented on HDFS-5249:
-

+1 for the patch.

> Fix dumper thread which may die silently
> 
>
> Key: HDFS-5249
> URL: https://issues.apache.org/jira/browse/HDFS-5249
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Brandon Li
>Assignee: Brandon Li
> Attachments: HDFS-5249.02.patch, HDFS-5249.patch
>
>
> Dumper thread can get an NPE when the WriteCtx it's about to work on is just 
> deleted by write back thread. 
> A dead dumper thread could cause out-of-memory error when too many pending 
> writes accumulated for one opened file.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5250) Enable one url to browse the entire federated hdfs file system on dfsclusterhealth summary page

2013-09-23 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5250?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775861#comment-13775861
 ] 

Andrew Wang commented on HDFS-5250:
---

Hi Vrushali,

ViewFileSystem is a purely client-side abstraction, each client can have their 
own viewfs mount points, and the federated namenodes are unaware of where they 
are mounted as well as the list of other namenodes/mount points.

I think something like this is only really possible with a single sharded 
namespace, not the client-side mount table that ViewFileSystem provides. Maybe 
I missed something, but we should probably close this JIRA as wontfix.

> Enable one url to browse the entire federated hdfs file system on 
> dfsclusterhealth summary page
> ---
>
> Key: HDFS-5250
> URL: https://issues.apache.org/jira/browse/HDFS-5250
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Reporter: Vrushali C
>
> Ideally, we should have one url to browse the entire federated file system on 
> the main dfsclusterhealth summary page along with the list of namenodes in 
> the cluster

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-5252) Do unstable write only when stable write can't be honored

2013-09-23 Thread Brandon Li (JIRA)
Brandon Li created HDFS-5252:


 Summary: Do unstable write only when stable write can't be honored
 Key: HDFS-5252
 URL: https://issues.apache.org/jira/browse/HDFS-5252
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: nfs
Reporter: Brandon Li


When the client asks for a stable write but the prerequisite writes are not 
transferred to NFS gateway, the stableness can't be honored. NFS gateway has to 
treat the write as unstable write.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5249) Fix dumper thread which may die silently

2013-09-23 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775890#comment-13775890
 ] 

Hadoop QA commented on HDFS-5249:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12604706/HDFS-5249.02.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs-nfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/5025//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5025//console

This message is automatically generated.

> Fix dumper thread which may die silently
> 
>
> Key: HDFS-5249
> URL: https://issues.apache.org/jira/browse/HDFS-5249
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Brandon Li
>Assignee: Brandon Li
> Attachments: HDFS-5249.02.patch, HDFS-5249.patch
>
>
> Dumper thread can get an NPE when the WriteCtx it's about to work on is just 
> deleted by write back thread. 
> A dead dumper thread could cause out-of-memory error when too many pending 
> writes accumulated for one opened file.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5251) Race between the initialization of NameNode and the http server

2013-09-23 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5251?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-5251:
-

Attachment: HDFS-5251.000.patch

> Race between the initialization of NameNode and the http server
> ---
>
> Key: HDFS-5251
> URL: https://issues.apache.org/jira/browse/HDFS-5251
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-5251.000.patch
>
>
> The constructor of NameNode starts a HTTP server before the FSNameSystem is 
> initialized. Currently there is a race where the HTTP server can access the 
> uninitialized namesystem variable, throwing a NullPointerException.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5251) Race between the initialization of NameNode and the http server

2013-09-23 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5251?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-5251:
-

Status: Patch Available  (was: Open)

> Race between the initialization of NameNode and the http server
> ---
>
> Key: HDFS-5251
> URL: https://issues.apache.org/jira/browse/HDFS-5251
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-5251.000.patch
>
>
> The constructor of NameNode starts a HTTP server before the FSNameSystem is 
> initialized. Currently there is a race where the HTTP server can access the 
> uninitialized namesystem variable, throwing a NullPointerException.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5237) Get rid of nodes' registration names

2013-09-23 Thread Eli Collins (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5237?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775898#comment-13775898
 ] 

Eli Collins commented on HDFS-5237:
---

Agree it's reasonable to validate dfs.datanode.hostname, but Junping's patch 
removes this config option which will break some configurations that use 
HDFS-3150 and need to set the reported hostname to something other than the 
default (for example if getDefaultHost returns a hostname on network A and the 
clients want to talk to the DN via network B).  Shall we re-purpose this to add 
dfs.datanode.hostname validation?

> Get rid of nodes' registration names
> 
>
> Key: HDFS-5237
> URL: https://issues.apache.org/jira/browse/HDFS-5237
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Junping Du
> Fix For: 3.0.0
>
> Attachments: HDFS-5237.patch
>
>
> Per discussion in HDFS-5208 and may be some other discussions before, Node's 
> registration name is pretty confusing and shouldn't be used in production 
> environment as topology resolving issues. So we remove related configuration 
> dfs.datanode.hostname or its old one "slave.host.name".

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5225) datanode keeps logging the same 'is no longer in the dataset' message over and over again

2013-09-23 Thread Tsuyoshi OZAWA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsuyoshi OZAWA updated HDFS-5225:
-

Attachment: HDFS-5225-reproduce.1.txt

A patch for reproducing this problem.

The analysis of the code is as follows:

1. DataBlockScanner#run() tries to scan blocks via 
BlockPoolSliceScanner#scanBlockPoolSlice in the main loop.
2. At the same time, the other theads(BlockReceiver's constructor) issue 
blockScanner.deleteBlock.
3. BlockPoolSliceScanner#deleteBlock() tries to remove BlockScanInfo from 
blockInfoSet. However, this is not visible from 
BlockPoolSliceScanner#scanBlockPoolSlice's thread, because there is no memory 
barrier. This may be the critical section we've faced.
4. Other threads issue FsDatasetImpl#unfinalizeBlock, FsDataImpl, 
checkAndUpdate, and updates BlockPoolSliceScanner#dataset, and remove the block 
entry from dataset. This is because the log 'is no longer in the dataset' is 
dumped repeatedly.

Note taht this may be not complete: I've faced the log 'is no longer in the 
dataset' dumped repeatedly, but sometimes with my code.

> datanode keeps logging the same 'is no longer in the dataset' message over 
> and over again
> -
>
> Key: HDFS-5225
> URL: https://issues.apache.org/jira/browse/HDFS-5225
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 2.1.1-beta
>Reporter: Roman Shaposhnik
> Attachments: HDFS-5225.1.patch, HDFS-5225-reproduce.1.txt
>
>
> I was running the usual Bigtop testing on 2.1.1-beta RC1 with the following 
> configuration: 4 nodes fully distributed cluster with security on.
> All of a sudden my DN ate up all of the space in /var/log logging the 
> following message repeatedly:
> {noformat}
> 2013-09-18 20:51:12,046 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: 
> BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1369 is no longer 
> in the dataset
> {noformat}
> It wouldn't answer to a jstack and jstack -F ended up being useless.
> Here's what I was able to find in the NameNode logs regarding this block ID:
> {noformat}
> fgrep -rI 'blk_1073742189' hadoop-hdfs-namenode-ip-10-224-158-152.log
> 2013-09-18 18:03:16,972 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
> allocateBlock: 
> /user/jenkins/testAppendInputWedSep18180222UTC2013/test4.fileWedSep18180222UTC2013._COPYING_.
>  BP-1884637155-10.224.158.152-1379524544853 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]}
> 2013-09-18 18:03:17,222 INFO BlockStateChange: BLOCK* addStoredBlock: 
> blockMap updated: 10.224.158.152:1004 is added to 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]} size 0
> 2013-09-18 18:03:17,222 INFO BlockStateChange: BLOCK* addStoredBlock: 
> blockMap updated: 10.34.74.206:1004 is added to 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]} size 0
> 2013-09-18 18:03:17,224 INFO BlockStateChange: BLOCK* addStoredBlock: 
> blockMap updated: 10.83.107.80:1004 is added to 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]} size 0
> 2013-09-18 18:03:17,899 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(block=BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1369,
>  newGenerationStamp=1370, newLength=1048576, newNodes=[10.83.107.80:1004, 
> 10.34.74.206:1004, 10.224.158.152:1004], 
> clientName=DFSClient_NONMAPREDUCE_-450304083_1)
> 2013-09-18 18:03:17,904 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1369)
>  successfully to 
> BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1370
> 2013-09-18 18:03:18,540 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(block=BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1370,
>  newGenerationStamp=1371, newLength=2097152, newNodes=[10.83.107.80:1004, 
> 10.34.74.206:1004, 10.224.158.152:1004], 
> clientName=DFSClient_NONMAPREDUCE_-450304083_1)
> 2013-09-18 18:03

[jira] [Created] (HDFS-5253) Add requesting user's name to PathBasedCacheEntry

2013-09-23 Thread Andrew Wang (JIRA)
Andrew Wang created HDFS-5253:
-

 Summary: Add requesting user's name to PathBasedCacheEntry
 Key: HDFS-5253
 URL: https://issues.apache.org/jira/browse/HDFS-5253
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: HDFS-4949
Reporter: Andrew Wang
Assignee: Andrew Wang


It'll be useful to have the requesting user's name in {{PathBasedCacheEntry}} 
for tracking per-user statistics (e.g. amount of data cached by a user).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5225) datanode keeps logging the same 'is no longer in the dataset' message over and over again

2013-09-23 Thread Tsuyoshi OZAWA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5225?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775924#comment-13775924
 ] 

Tsuyoshi OZAWA commented on HDFS-5225:
--

{quote}
Note that this may be not complete: I've faced the log 'is no longer in the 
dataset' dumped repeatedly, but sometimes with my code.
{quote}

I mean the log messages itself are dumped every time, but the "logging the same 
'is no longer in the dataset' message over and over again" can happen sometimes.

> datanode keeps logging the same 'is no longer in the dataset' message over 
> and over again
> -
>
> Key: HDFS-5225
> URL: https://issues.apache.org/jira/browse/HDFS-5225
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 2.1.1-beta
>Reporter: Roman Shaposhnik
> Attachments: HDFS-5225.1.patch, HDFS-5225-reproduce.1.txt
>
>
> I was running the usual Bigtop testing on 2.1.1-beta RC1 with the following 
> configuration: 4 nodes fully distributed cluster with security on.
> All of a sudden my DN ate up all of the space in /var/log logging the 
> following message repeatedly:
> {noformat}
> 2013-09-18 20:51:12,046 INFO 
> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: 
> BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1369 is no longer 
> in the dataset
> {noformat}
> It wouldn't answer to a jstack and jstack -F ended up being useless.
> Here's what I was able to find in the NameNode logs regarding this block ID:
> {noformat}
> fgrep -rI 'blk_1073742189' hadoop-hdfs-namenode-ip-10-224-158-152.log
> 2013-09-18 18:03:16,972 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
> allocateBlock: 
> /user/jenkins/testAppendInputWedSep18180222UTC2013/test4.fileWedSep18180222UTC2013._COPYING_.
>  BP-1884637155-10.224.158.152-1379524544853 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]}
> 2013-09-18 18:03:17,222 INFO BlockStateChange: BLOCK* addStoredBlock: 
> blockMap updated: 10.224.158.152:1004 is added to 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]} size 0
> 2013-09-18 18:03:17,222 INFO BlockStateChange: BLOCK* addStoredBlock: 
> blockMap updated: 10.34.74.206:1004 is added to 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]} size 0
> 2013-09-18 18:03:17,224 INFO BlockStateChange: BLOCK* addStoredBlock: 
> blockMap updated: 10.83.107.80:1004 is added to 
> blk_1073742189_1369{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, 
> replicas=[ReplicaUnderConstruction[10.83.107.80:1004|RBW], 
> ReplicaUnderConstruction[10.34.74.206:1004|RBW], 
> ReplicaUnderConstruction[10.224.158.152:1004|RBW]]} size 0
> 2013-09-18 18:03:17,899 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(block=BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1369,
>  newGenerationStamp=1370, newLength=1048576, newNodes=[10.83.107.80:1004, 
> 10.34.74.206:1004, 10.224.158.152:1004], 
> clientName=DFSClient_NONMAPREDUCE_-450304083_1)
> 2013-09-18 18:03:17,904 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1369)
>  successfully to 
> BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1370
> 2013-09-18 18:03:18,540 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(block=BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1370,
>  newGenerationStamp=1371, newLength=2097152, newNodes=[10.83.107.80:1004, 
> 10.34.74.206:1004, 10.224.158.152:1004], 
> clientName=DFSClient_NONMAPREDUCE_-450304083_1)
> 2013-09-18 18:03:18,548 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
> updatePipeline(BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1370)
>  successfully to 
> BP-1884637155-10.224.158.152-1379524544853:blk_1073742189_1371
> 2013-09-18 18:03:26,150 INFO BlockStateChange: BLOCK* addToInvalidates: 
> blk_1073742189_1371 10.83.107.80:1004 10.34.74.206:1004 10.224.158.152:1004 
> 2013-09-18 18:03:26,847 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
> InvalidateBlocks: ask 10.34.74.206:1004 to delete [blk_1073742178_1359, 
> blk_1073742183_1362, blk_1073742184_1363, blk_1073742186_136

[jira] [Commented] (HDFS-5251) Race between the initialization of NameNode and the http server

2013-09-23 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5251?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775926#comment-13775926
 ] 

Suresh Srinivas commented on HDFS-5251:
---

+1 for the change. I will commit it shortly.

> Race between the initialization of NameNode and the http server
> ---
>
> Key: HDFS-5251
> URL: https://issues.apache.org/jira/browse/HDFS-5251
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-5251.000.patch
>
>
> The constructor of NameNode starts a HTTP server before the FSNameSystem is 
> initialized. Currently there is a race where the HTTP server can access the 
> uninitialized namesystem variable, throwing a NullPointerException.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5242) Reduce contention on DatanodeInfo instances

2013-09-23 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5242?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775927#comment-13775927
 ] 

Hadoop QA commented on HDFS-5242:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12604625/HDFS-5242.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/5024//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5024//console

This message is automatically generated.

> Reduce contention on DatanodeInfo instances
> ---
>
> Key: HDFS-5242
> URL: https://issues.apache.org/jira/browse/HDFS-5242
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
> Attachments: HDFS-5242.patch
>
>
> Synchronization in {{DatanodeInfo}} instances causes unnecessary contention 
> between call handlers.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5216) NumDecomDeadDataNodes not returning correct number of dead decommissioned nodes

2013-09-23 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5216?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775928#comment-13775928
 ] 

Hadoop QA commented on HDFS-5216:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12604543/HDFS-5216.diff
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/5023//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5023//console

This message is automatically generated.

> NumDecomDeadDataNodes not returning correct number of dead decommissioned 
> nodes 
> 
>
> Key: HDFS-5216
> URL: https://issues.apache.org/jira/browse/HDFS-5216
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.1.0-beta
>Reporter: Trevor Lorimer
>Assignee: Trevor Lorimer
> Attachments: HDFS-5216.diff, HDFS-5216.diff
>
>
> For HDFS-4860 I essentially copied the process in 
> NamenodejspHelper.generateHealthReport(), so it would be in sync with the 
> original dfsHealth.jsp.
> However looking at this now there may be a bug? in 
> getNumDecomDeadDataNodes(), where:
> getBlockManager().getDatanodeManager().fetchDatanodes(dead, null, true);
> 
> Where the parameter true indicates that decommissioned nodes should be 
> removed from the list.
> If the flag is true fetchDatanodes calls removeDecomNodeFromList, which will 
> remove a node if an existing datanode does not appear in both include or 
> exclude lists and it has been decommissioned.
> If I am looking to return the Number of Dead Decommissioned Nodes, should I 
> change the remove decommissioned nodes flag to False? i.e.:
> getBlockManager().getDatanodeManager().fetchDatanodes(null, dead, false);

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5250) Enable one url to browse the entire federated hdfs file system on dfsclusterhealth summary page

2013-09-23 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5250?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775932#comment-13775932
 ] 

Suresh Srinivas commented on HDFS-5250:
---

[~vrushalic], dfscluster health page was a way to show summary of all the 
namenodes in the federated cluster. Currently you can go to each namenode from 
that page and browse the file system supported on each of them.

I agree, this should be closed as wontfix.

> Enable one url to browse the entire federated hdfs file system on 
> dfsclusterhealth summary page
> ---
>
> Key: HDFS-5250
> URL: https://issues.apache.org/jira/browse/HDFS-5250
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Reporter: Vrushali C
>
> Ideally, we should have one url to browse the entire federated file system on 
> the main dfsclusterhealth summary page along with the list of namenodes in 
> the cluster

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5249) Fix dumper thread which may die silently

2013-09-23 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775977#comment-13775977
 ] 

Brandon Li commented on HDFS-5249:
--

Thank you Jing for the review. I've committed the patch.

> Fix dumper thread which may die silently
> 
>
> Key: HDFS-5249
> URL: https://issues.apache.org/jira/browse/HDFS-5249
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Brandon Li
>Assignee: Brandon Li
> Attachments: HDFS-5249.02.patch, HDFS-5249.patch
>
>
> Dumper thread can get an NPE when the WriteCtx it's about to work on is just 
> deleted by write back thread. 
> A dead dumper thread could cause out-of-memory error when too many pending 
> writes accumulated for one opened file.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5249) Fix dumper thread which may die silently

2013-09-23 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-5249:
-

Fix Version/s: 2.1.1-beta

> Fix dumper thread which may die silently
> 
>
> Key: HDFS-5249
> URL: https://issues.apache.org/jira/browse/HDFS-5249
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Brandon Li
>Assignee: Brandon Li
> Fix For: 2.1.1-beta
>
> Attachments: HDFS-5249.02.patch, HDFS-5249.patch
>
>
> Dumper thread can get an NPE when the WriteCtx it's about to work on is just 
> deleted by write back thread. 
> A dead dumper thread could cause out-of-memory error when too many pending 
> writes accumulated for one opened file.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5249) Fix dumper thread which may die silently

2013-09-23 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HDFS-5249:
-

Resolution: Fixed
Status: Resolved  (was: Patch Available)

> Fix dumper thread which may die silently
> 
>
> Key: HDFS-5249
> URL: https://issues.apache.org/jira/browse/HDFS-5249
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Brandon Li
>Assignee: Brandon Li
> Fix For: 2.1.1-beta
>
> Attachments: HDFS-5249.02.patch, HDFS-5249.patch
>
>
> Dumper thread can get an NPE when the WriteCtx it's about to work on is just 
> deleted by write back thread. 
> A dead dumper thread could cause out-of-memory error when too many pending 
> writes accumulated for one opened file.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5251) Race between the initialization of NameNode and the http server

2013-09-23 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5251?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775980#comment-13775980
 ] 

Hadoop QA commented on HDFS-5251:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12604720/HDFS-5251.000.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/5026//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/5026//console

This message is automatically generated.

> Race between the initialization of NameNode and the http server
> ---
>
> Key: HDFS-5251
> URL: https://issues.apache.org/jira/browse/HDFS-5251
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Attachments: HDFS-5251.000.patch
>
>
> The constructor of NameNode starts a HTTP server before the FSNameSystem is 
> initialized. Currently there is a race where the HTTP server can access the 
> uninitialized namesystem variable, throwing a NullPointerException.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5249) Fix dumper thread which may die silently

2013-09-23 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13775985#comment-13775985
 ] 

Hudson commented on HDFS-5249:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #4459 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/4459/])
HDFS-5249. Fix dumper thread which may die silently. Contributed by Brandon Li 
(brandonli: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1525770)
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/java/org/apache/hadoop/hdfs/nfs/nfs3/OpenFileCtx.java
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


> Fix dumper thread which may die silently
> 
>
> Key: HDFS-5249
> URL: https://issues.apache.org/jira/browse/HDFS-5249
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: nfs
>Reporter: Brandon Li
>Assignee: Brandon Li
> Fix For: 2.1.1-beta
>
> Attachments: HDFS-5249.02.patch, HDFS-5249.patch
>
>
> Dumper thread can get an NPE when the WriteCtx it's about to work on is just 
> deleted by write back thread. 
> A dead dumper thread could cause out-of-memory error when too many pending 
> writes accumulated for one opened file.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-5251) Race between the initialization of NameNode and the http server

2013-09-23 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5251?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-5251:
--

   Resolution: Fixed
Fix Version/s: 2.2.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

I committed the patch to trunk, branch-2 and branch-2.1-beta. Thank you Haohui!

> Race between the initialization of NameNode and the http server
> ---
>
> Key: HDFS-5251
> URL: https://issues.apache.org/jira/browse/HDFS-5251
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Fix For: 2.2.0
>
> Attachments: HDFS-5251.000.patch
>
>
> The constructor of NameNode starts a HTTP server before the FSNameSystem is 
> initialized. Currently there is a race where the HTTP server can access the 
> uninitialized namesystem variable, throwing a NullPointerException.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-5251) Race between the initialization of NameNode and the http server

2013-09-23 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5251?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13776015#comment-13776015
 ] 

Hudson commented on HDFS-5251:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #4460 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/4460/])
HDFS-5251. Race between the initialization of NameNode and the http server. 
Contributed by Haohui Mai. (suresh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1525787)
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NamenodeJspHelper.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeJspHelper.java


> Race between the initialization of NameNode and the http server
> ---
>
> Key: HDFS-5251
> URL: https://issues.apache.org/jira/browse/HDFS-5251
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Haohui Mai
>Assignee: Haohui Mai
> Fix For: 2.2.0
>
> Attachments: HDFS-5251.000.patch
>
>
> The constructor of NameNode starts a HTTP server before the FSNameSystem is 
> initialized. Currently there is a race where the HTTP server can access the 
> uninitialized namesystem variable, throwing a NullPointerException.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira