[jira] [Commented] (HDFS-1964) Incorrect HTML unescaping in DatanodeJspHelper.java

2011-05-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1964?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13038974#comment-13038974
 ] 

Hudson commented on HDFS-1964:
--

Integrated in Hadoop-Hdfs-trunk-Commit #683 (See 
[https://builds.apache.org/hudson/job/Hadoop-Hdfs-trunk-Commit/683/])
HDFS-1964. Fix incorrect HTML unescaping in DatanodeJspHelper. Contributed 
by Aaron T. Myers.

todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1127390
Files : 
* 
/hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/DatanodeJspHelper.java
* 
/hadoop/hdfs/trunk/src/test/hdfs/org/apache/hadoop/hdfs/server/datanode/TestDatanodeJsp.java
* /hadoop/hdfs/trunk/CHANGES.txt


 Incorrect HTML unescaping in DatanodeJspHelper.java
 ---

 Key: HDFS-1964
 URL: https://issues.apache.org/jira/browse/HDFS-1964
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.22.0, 0.23.0
Reporter: Aaron T. Myers
Assignee: Aaron T. Myers
 Fix For: 0.22.0, 0.23.0

 Attachments: hdfs-1964-trunk.0.patch, hdfs-1964-trunk.1.patch


 HDFS-1575 introduced some HTML unescaping of parameters so that viewing a 
 file would work for paths containing HTML-escaped characters, but in two of 
 the places did the unescaping either too early or too late.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1623) High Availability Framework for HDFS NN

2011-05-25 Thread Sanjay Radia (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1623?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sanjay Radia updated HDFS-1623:
---

Attachment: NameNode HA_v2.pdf
NameNode HA_v2.pdf

Attached version 2 of Namenode HA document. Added a significant amount of 
design details.

 High Availability Framework for HDFS NN
 ---

 Key: HDFS-1623
 URL: https://issues.apache.org/jira/browse/HDFS-1623
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Sanjay Radia
Assignee: Sanjay Radia
 Attachments: HDFS-High-Availability.pdf, NameNode HA_v2.pdf, Namenode 
 HA Framework.pdf




--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1623) High Availability Framework for HDFS NN

2011-05-25 Thread Sanjay Radia (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1623?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sanjay Radia updated HDFS-1623:
---

Attachment: (was: NameNode HA_v2.pdf)

 High Availability Framework for HDFS NN
 ---

 Key: HDFS-1623
 URL: https://issues.apache.org/jira/browse/HDFS-1623
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Sanjay Radia
Assignee: Sanjay Radia
 Attachments: HDFS-High-Availability.pdf, NameNode HA_v2.pdf, Namenode 
 HA Framework.pdf




--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1623) High Availability Framework for HDFS NN

2011-05-25 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1623?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039019#comment-13039019
 ] 

Sanjay Radia commented on HDFS-1623:


 How does NN fail-over impact federation? Eg does viewfs have any special 
 requirements as a client?
In case of federation, each NN needs its own warm or hot failover. Cannot do 
N+k failover because of the large memory requirements for a NN unless one wants 
to limit it to cold failover.

Viewfs is separate layer and not required for federation; one could choose to 
use symbolic links or one could choose to not provide a global namespace. But 
if one were to use viewfs, I am not aware of any *additional* failover issues.
Viewfs has bindings like: /user -hdfs://nnOfUser/  
The failover issues are the same as if hdfs://nnofUser was your default file 
system.

 High Availability Framework for HDFS NN
 ---

 Key: HDFS-1623
 URL: https://issues.apache.org/jira/browse/HDFS-1623
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Sanjay Radia
Assignee: Sanjay Radia
 Attachments: HDFS-High-Availability.pdf, NameNode HA_v2.pdf, Namenode 
 HA Framework.pdf




--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1623) High Availability Framework for HDFS NN

2011-05-25 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1623?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039022#comment-13039022
 ] 

Sanjay Radia commented on HDFS-1623:


Wrt requirement #2, ... fail over to new version of HDFS ...  is out of scope 
for this framework. 

For failover during upgrades, the DNs need to be able to communicate to both 
versions of the NN. This is true for dot releases and will only be true if we 
support rolling upgrades. I will clarify this on the next version of the 
document.

 High Availability Framework for HDFS NN
 ---

 Key: HDFS-1623
 URL: https://issues.apache.org/jira/browse/HDFS-1623
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Sanjay Radia
Assignee: Sanjay Radia
 Attachments: HDFS-High-Availability.pdf, NameNode HA_v2.pdf, Namenode 
 HA Framework.pdf




--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1941) Remove -genclusterid from NameNode startup options

2011-05-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1941?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039090#comment-13039090
 ] 

Hudson commented on HDFS-1941:
--

Integrated in Hadoop-Hdfs-trunk #677 (See 
[https://builds.apache.org/hudson/job/Hadoop-Hdfs-trunk/677/])
Reverting the change r1125031 - HDFS-1941. Remove -genclusterid option from 
namenode command.

suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1127311
Files : 
* 
/hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java
* 
/hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/common/HdfsConstants.java
* /hadoop/hdfs/trunk/CHANGES.txt


 Remove -genclusterid from NameNode startup options
 --

 Key: HDFS-1941
 URL: https://issues.apache.org/jira/browse/HDFS-1941
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: name-node
Affects Versions: 0.23.0
Reporter: Bharath Mundlapudi
Assignee: Bharath Mundlapudi
Priority: Minor
 Fix For: 0.23.0

 Attachments: HDFS-1941-1.patch


 Currently, namenode -genclusterid is a helper utility to generate unique 
 clusterid. This option is useless once namenode -format automatically 
 generates the clusterid.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1964) Incorrect HTML unescaping in DatanodeJspHelper.java

2011-05-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1964?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039091#comment-13039091
 ] 

Hudson commented on HDFS-1964:
--

Integrated in Hadoop-Hdfs-trunk #677 (See 
[https://builds.apache.org/hudson/job/Hadoop-Hdfs-trunk/677/])
HDFS-1964. Fix incorrect HTML unescaping in DatanodeJspHelper. Contributed 
by Aaron T. Myers.

todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1127390
Files : 
* 
/hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/DatanodeJspHelper.java
* 
/hadoop/hdfs/trunk/src/test/hdfs/org/apache/hadoop/hdfs/server/datanode/TestDatanodeJsp.java
* /hadoop/hdfs/trunk/CHANGES.txt


 Incorrect HTML unescaping in DatanodeJspHelper.java
 ---

 Key: HDFS-1964
 URL: https://issues.apache.org/jira/browse/HDFS-1964
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.22.0, 0.23.0
Reporter: Aaron T. Myers
Assignee: Aaron T. Myers
 Fix For: 0.22.0, 0.23.0

 Attachments: hdfs-1964-trunk.0.patch, hdfs-1964-trunk.1.patch


 HDFS-1575 introduced some HTML unescaping of parameters so that viewing a 
 file would work for paths containing HTML-escaped characters, but in two of 
 the places did the unescaping either too early or too late.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1169) Can't read binary data off HDFS via thrift API

2011-05-25 Thread Johnny Boy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1169?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Johnny Boy updated HDFS-1169:
-

Attachment: thriftfs.jar

HadoopThriftServer with changed encoding for reading and writing

 Can't read binary data off HDFS via thrift API
 --

 Key: HDFS-1169
 URL: https://issues.apache.org/jira/browse/HDFS-1169
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.20.2
Reporter: Erik Forsberg
 Attachments: HadoopThriftServer.java, hadoopfs.thrift, thriftfs.jar


 Trying to access binary data stored in HDFS (in my case, TypedByte files 
 generated by Dumbo) via thrift talking to 
 org.apache.hadoop.thriftfs.HadoopThriftServer, the data I get back is 
 mangled. For example, when I read a file which contains the value 0xa2, it's 
 coming back as 0xef 0xbf 0xbd, also known as the Unicode replacement 
 character.
 I think this is because the read method in HadoopThriftServer.java is trying 
 to convert the data read from HDFS into UTF-8 via the String() constructor. 
 This essentially makes the HDFS thrift API useless for me :-(.
 Not being an expert on Thrift, but would it be possible to modify the API so 
 that it uses the binary type listed on 
 http://wiki.apache.org/thrift/ThriftTypes?

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1169) Can't read binary data off HDFS via thrift API

2011-05-25 Thread Johnny Boy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1169?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Johnny Boy updated HDFS-1169:
-

Attachment: thriftfs.jar

Modified version of HadoopThriftServer

 Can't read binary data off HDFS via thrift API
 --

 Key: HDFS-1169
 URL: https://issues.apache.org/jira/browse/HDFS-1169
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.20.2
Reporter: Erik Forsberg
 Attachments: HadoopThriftServer.java, hadoopfs.thrift, thriftfs.jar, 
 thriftfs.jar


 Trying to access binary data stored in HDFS (in my case, TypedByte files 
 generated by Dumbo) via thrift talking to 
 org.apache.hadoop.thriftfs.HadoopThriftServer, the data I get back is 
 mangled. For example, when I read a file which contains the value 0xa2, it's 
 coming back as 0xef 0xbf 0xbd, also known as the Unicode replacement 
 character.
 I think this is because the read method in HadoopThriftServer.java is trying 
 to convert the data read from HDFS into UTF-8 via the String() constructor. 
 This essentially makes the HDFS thrift API useless for me :-(.
 Not being an expert on Thrift, but would it be possible to modify the API so 
 that it uses the binary type listed on 
 http://wiki.apache.org/thrift/ThriftTypes?

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1169) Can't read binary data off HDFS via thrift API

2011-05-25 Thread Johnny Boy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1169?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039124#comment-13039124
 ] 

Johnny Boy commented on HDFS-1169:
--

Bleh double post and I don't know how to remove it :)
The comment should be that it was enough to change utf-8 into latin1 for me to 
get the binary data to work, see attached file above

 Can't read binary data off HDFS via thrift API
 --

 Key: HDFS-1169
 URL: https://issues.apache.org/jira/browse/HDFS-1169
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.20.2
Reporter: Erik Forsberg
 Attachments: HadoopThriftServer.java, hadoopfs.thrift, thriftfs.jar, 
 thriftfs.jar


 Trying to access binary data stored in HDFS (in my case, TypedByte files 
 generated by Dumbo) via thrift talking to 
 org.apache.hadoop.thriftfs.HadoopThriftServer, the data I get back is 
 mangled. For example, when I read a file which contains the value 0xa2, it's 
 coming back as 0xef 0xbf 0xbd, also known as the Unicode replacement 
 character.
 I think this is because the read method in HadoopThriftServer.java is trying 
 to convert the data read from HDFS into UTF-8 via the String() constructor. 
 This essentially makes the HDFS thrift API useless for me :-(.
 Not being an expert on Thrift, but would it be possible to modify the API so 
 that it uses the binary type listed on 
 http://wiki.apache.org/thrift/ThriftTypes?

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-977) DataNode.createInterDataNodeProtocolProxy() guards a log at the wrong level

2011-05-25 Thread Harsh J Chouraria (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-977?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J Chouraria updated HDFS-977:
---

Fix Version/s: 0.23.0
 Release Note: DataNode.createInterDataNodeProtocolProxy() guarded a log at 
the wrong level
   Status: Patch Available  (was: Open)

 DataNode.createInterDataNodeProtocolProxy() guards a log at the wrong level
 ---

 Key: HDFS-977
 URL: https://issues.apache.org/jira/browse/HDFS-977
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: data-node
Affects Versions: 0.22.0
Reporter: Steve Loughran
Assignee: Harsh J Chouraria
Priority: Trivial
 Fix For: 0.23.0

 Attachments: HDFS-977.r1.diff


 My IDE tells me that this code snippet in 
 {{DataNode.createInterDataNodeProtocolProxy()}} is guarding the info log 
 entry at debug level, and it should be reviewed
 {code}
 if (InterDatanodeProtocol.LOG.isDebugEnabled()) {
   InterDatanodeProtocol.LOG.info(InterDatanodeProtocol addr= + addr);
 }
 {code}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-977) DataNode.createInterDataNodeProtocolProxy() guards a log at the wrong level

2011-05-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-977?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039132#comment-13039132
 ] 

Hadoop QA commented on HDFS-977:


-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12478501/HDFS-977.r1.diff
  against trunk revision 1127390.

+1 @author.  The patch does not contain any @author tags.

-1 tests included.  The patch doesn't appear to include any new or modified 
tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

-1 patch.  The patch command could not apply the patch.

Console output: 
https://builds.apache.org/hudson/job/PreCommit-HDFS-Build/622//console

This message is automatically generated.

 DataNode.createInterDataNodeProtocolProxy() guards a log at the wrong level
 ---

 Key: HDFS-977
 URL: https://issues.apache.org/jira/browse/HDFS-977
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: data-node
Affects Versions: 0.22.0
Reporter: Steve Loughran
Assignee: Harsh J Chouraria
Priority: Trivial
 Fix For: 0.23.0

 Attachments: HDFS-977.r1.diff


 My IDE tells me that this code snippet in 
 {{DataNode.createInterDataNodeProtocolProxy()}} is guarding the info log 
 entry at debug level, and it should be reviewed
 {code}
 if (InterDatanodeProtocol.LOG.isDebugEnabled()) {
   InterDatanodeProtocol.LOG.info(InterDatanodeProtocol addr= + addr);
 }
 {code}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1636) If dfs.name.dir points to an empty dir, namenode format shouldn't require confirmation

2011-05-25 Thread Harsh J Chouraria (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1636?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J Chouraria updated HDFS-1636:


Attachment: HDFS-1636.r2.diff

Added sort of a test case for this.

 If dfs.name.dir points to an empty dir, namenode format shouldn't require 
 confirmation
 --

 Key: HDFS-1636
 URL: https://issues.apache.org/jira/browse/HDFS-1636
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Affects Versions: 0.23.0
Reporter: Todd Lipcon
Assignee: Harsh J Chouraria
Priority: Minor
 Fix For: 0.23.0

 Attachments: HDFS-1636.r1.diff, HDFS-1636.r2.diff


 Right now, running namenode -format when dfs.name.dir is configured to a dir 
 which exists but is empty still asks for confirmation. This is unnecessary 
 since it isn't blowing away any real data.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1636) If dfs.name.dir points to an empty dir, namenode format shouldn't require confirmation

2011-05-25 Thread Harsh J Chouraria (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1636?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J Chouraria updated HDFS-1636:


Fix Version/s: 0.23.0
 Release Note: If dfs.name.dir points to an empty dir, namenode format 
shouldn't require confirmation.
   Status: Patch Available  (was: Open)

 If dfs.name.dir points to an empty dir, namenode format shouldn't require 
 confirmation
 --

 Key: HDFS-1636
 URL: https://issues.apache.org/jira/browse/HDFS-1636
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Affects Versions: 0.23.0
Reporter: Todd Lipcon
Assignee: Harsh J Chouraria
Priority: Minor
 Fix For: 0.23.0

 Attachments: HDFS-1636.r1.diff, HDFS-1636.r2.diff


 Right now, running namenode -format when dfs.name.dir is configured to a dir 
 which exists but is empty still asks for confirmation. This is unnecessary 
 since it isn't blowing away any real data.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1620) FSConstants vs FsConstants

2011-05-25 Thread Harsh J Chouraria (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1620?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039145#comment-13039145
 ] 

Harsh J Chouraria commented on HDFS-1620:
-

The failing test: 
org.apache.hadoop.hdfs.TestDFSStorageStateRecovery.testBlockPoolStorageStates 
is unrelated to this patch (age 93 as per hudson).

 FSConstants vs FsConstants 
 ---

 Key: HDFS-1620
 URL: https://issues.apache.org/jira/browse/HDFS-1620
 Project: Hadoop HDFS
  Issue Type: Improvement
Affects Versions: 0.22.0
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Harsh J Chouraria
Priority: Minor
 Fix For: 0.23.0

 Attachments: HDFS-1620.r1.diff


 We have {{org.apache.hadoop.hdfs.protocol.*FSConstants*}} and 
 {{org.apache.hadoop.fs.*FsConstants*}}.  Elegant or confused?

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1640) Transmission of large files in compressed format will save the network traffic and can improve the data transfer time.

2011-05-25 Thread Harsh J Chouraria (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039159#comment-13039159
 ] 

Harsh J Chouraria commented on HDFS-1640:
-

@Option 2 -- This can be done at the writer level itself. Much too expensive 
maintaining a per-file property for compression and implementing the same to be 
guaranteed as well.

Besides, compressing on the fly for tx would still invoke the same amount of 
IO. No savings, right?

 Transmission of large files in compressed format will save the network 
 traffic and can improve the data transfer time.
 --

 Key: HDFS-1640
 URL: https://issues.apache.org/jira/browse/HDFS-1640
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: data-node, hdfs client
Reporter: Uma Maheswara Rao G

 *In Write scenario:*
 DFSClient can Compress the data when transmitting it over the network, Data 
 Nodes can forward the same compressed data to other Data Nodes in pipeline 
 and as well as it can decompress that data and write on to their local disks. 
 *In Read Scenario:*
  Data Node can compress the data when transmitting it over the network. 
 DFSClient can decompress it and write on to the local stream.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1636) If dfs.name.dir points to an empty dir, namenode format shouldn't require confirmation

2011-05-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1636?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039160#comment-13039160
 ] 

Hadoop QA commented on HDFS-1636:
-

-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12480416/HDFS-1636.r2.diff
  against trunk revision 1127390.

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 3 new or modified tests.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

-1 core tests.  The patch failed these core unit tests:
  org.apache.hadoop.hdfs.TestDFSStorageStateRecovery

+1 contrib tests.  The patch passed contrib unit tests.

+1 system test framework.  The patch passed system test framework compile.

Test results: 
https://builds.apache.org/hudson/job/PreCommit-HDFS-Build/623//testReport/
Findbugs warnings: 
https://builds.apache.org/hudson/job/PreCommit-HDFS-Build/623//artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Console output: 
https://builds.apache.org/hudson/job/PreCommit-HDFS-Build/623//console

This message is automatically generated.

 If dfs.name.dir points to an empty dir, namenode format shouldn't require 
 confirmation
 --

 Key: HDFS-1636
 URL: https://issues.apache.org/jira/browse/HDFS-1636
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Affects Versions: 0.23.0
Reporter: Todd Lipcon
Assignee: Harsh J Chouraria
Priority: Minor
 Fix For: 0.23.0

 Attachments: HDFS-1636.r1.diff, HDFS-1636.r2.diff


 Right now, running namenode -format when dfs.name.dir is configured to a dir 
 which exists but is empty still asks for confirmation. This is unnecessary 
 since it isn't blowing away any real data.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1636) If dfs.name.dir points to an empty dir, namenode format shouldn't require confirmation

2011-05-25 Thread Harsh J Chouraria (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1636?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039178#comment-13039178
 ] 

Harsh J Chouraria commented on HDFS-1636:
-

Failing test looks unrelated.

 If dfs.name.dir points to an empty dir, namenode format shouldn't require 
 confirmation
 --

 Key: HDFS-1636
 URL: https://issues.apache.org/jira/browse/HDFS-1636
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Affects Versions: 0.23.0
Reporter: Todd Lipcon
Assignee: Harsh J Chouraria
Priority: Minor
 Fix For: 0.23.0

 Attachments: HDFS-1636.r1.diff, HDFS-1636.r2.diff


 Right now, running namenode -format when dfs.name.dir is configured to a dir 
 which exists but is empty still asks for confirmation. This is unnecessary 
 since it isn't blowing away any real data.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1636) If dfs.name.dir points to an empty dir, namenode format shouldn't require confirmation

2011-05-25 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1636?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039195#comment-13039195
 ] 

Todd Lipcon commented on HDFS-1636:
---

Hey Harsh. There's now an issue where the NN will throw an NPE if one of the 
directories exists but isn't accessible. eg I did chmod 000 /my/namedir and 
run namenode -format and got:

11/05/25 09:43:32 ERROR namenode.NameNode: java.lang.NullPointerException
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1487)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1671)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1730)

Consider looking at HADOOP-7322 for a utility method on its way into trunk to 
deal with this.

 If dfs.name.dir points to an empty dir, namenode format shouldn't require 
 confirmation
 --

 Key: HDFS-1636
 URL: https://issues.apache.org/jira/browse/HDFS-1636
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Affects Versions: 0.23.0
Reporter: Todd Lipcon
Assignee: Harsh J Chouraria
Priority: Minor
 Fix For: 0.23.0

 Attachments: HDFS-1636.r1.diff, HDFS-1636.r2.diff


 Right now, running namenode -format when dfs.name.dir is configured to a dir 
 which exists but is empty still asks for confirmation. This is unnecessary 
 since it isn't blowing away any real data.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-977) DataNode.createInterDataNodeProtocolProxy() guards a log at the wrong level

2011-05-25 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-977?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039198#comment-13039198
 ] 

Todd Lipcon commented on HDFS-977:
--

Can you re-upload as a -p0 patch?

 DataNode.createInterDataNodeProtocolProxy() guards a log at the wrong level
 ---

 Key: HDFS-977
 URL: https://issues.apache.org/jira/browse/HDFS-977
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: data-node
Affects Versions: 0.22.0
Reporter: Steve Loughran
Assignee: Harsh J Chouraria
Priority: Trivial
 Fix For: 0.23.0

 Attachments: HDFS-977.r1.diff


 My IDE tells me that this code snippet in 
 {{DataNode.createInterDataNodeProtocolProxy()}} is guarding the info log 
 entry at debug level, and it should be reviewed
 {code}
 if (InterDatanodeProtocol.LOG.isDebugEnabled()) {
   InterDatanodeProtocol.LOG.info(InterDatanodeProtocol addr= + addr);
 }
 {code}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-977) DataNode.createInterDataNodeProtocolProxy() guards a log at the wrong level

2011-05-25 Thread Harsh J Chouraria (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-977?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J Chouraria updated HDFS-977:
---

Attachment: HDFS-977.r2.diff

Sorry about that!

 DataNode.createInterDataNodeProtocolProxy() guards a log at the wrong level
 ---

 Key: HDFS-977
 URL: https://issues.apache.org/jira/browse/HDFS-977
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: data-node
Affects Versions: 0.22.0
Reporter: Steve Loughran
Assignee: Harsh J Chouraria
Priority: Trivial
 Fix For: 0.23.0

 Attachments: HDFS-977.r1.diff, HDFS-977.r2.diff


 My IDE tells me that this code snippet in 
 {{DataNode.createInterDataNodeProtocolProxy()}} is guarding the info log 
 entry at debug level, and it should be reviewed
 {code}
 if (InterDatanodeProtocol.LOG.isDebugEnabled()) {
   InterDatanodeProtocol.LOG.info(InterDatanodeProtocol addr= + addr);
 }
 {code}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1727) fsck command can display command usage if user passes any illegal argument

2011-05-25 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1727?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039208#comment-13039208
 ] 

Todd Lipcon commented on HDFS-1727:
---

A few comments:
- I think the {{null == dir}} check needs an {{else}} clause that prints an 
error that only one path may be passed. That is to say, if you run hadoop fsck 
/foo /bar, it should print an error rather than ignoring /bar.
- To make the tests run faster, I think it would be better for the two test 
cases to be combined and share the same MiniDFSCluster.
- The test case would be easier to follow if you used an argument like 
-asdfjisadfijs or -thisIsNotAValidFlag instead of -listcorruptfileblocks.
- The error message could be formatted nicer. Instead of Invalid arguments 
passed 'foo' I would use the same formatting as the FsShell errors - ie fsck: 
Illegal option 'foo'



 fsck command can display command usage if user passes any illegal argument
 --

 Key: HDFS-1727
 URL: https://issues.apache.org/jira/browse/HDFS-1727
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.20.1, 0.23.0
Reporter: Uma Maheswara Rao G
Priority: Minor
 Attachments: HDFS-1727.patch


 In fsck command if user passes the arguments like
 ./hadoop fsck -test -files -blocks -racks
 In this case it will take / and will display whole DFS information regarding 
 to files,blocks,racks.
 But here, we are hiding the user mistake. Instead of this, we can display the 
 command usage if user passes any invalid argument like above.
 If user passes illegal optional arguments like
 ./hadoop fsck /test -listcorruptfileblocks instead of
 ./hadoop fsck /test -list-corruptfileblocks also we can display the proper 
 command usage

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1727) fsck command can display command usage if user passes any illegal argument

2011-05-25 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1727?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039210#comment-13039210
 ] 

Todd Lipcon commented on HDFS-1727:
---

One more comment: please remove the @throws Exception from the test javadoc. 
It doesn't give any extra information, so just clutters the code.

 fsck command can display command usage if user passes any illegal argument
 --

 Key: HDFS-1727
 URL: https://issues.apache.org/jira/browse/HDFS-1727
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.20.1, 0.23.0
Reporter: Uma Maheswara Rao G
Priority: Minor
 Attachments: HDFS-1727.patch


 In fsck command if user passes the arguments like
 ./hadoop fsck -test -files -blocks -racks
 In this case it will take / and will display whole DFS information regarding 
 to files,blocks,racks.
 But here, we are hiding the user mistake. Instead of this, we can display the 
 command usage if user passes any invalid argument like above.
 If user passes illegal optional arguments like
 ./hadoop fsck /test -listcorruptfileblocks instead of
 ./hadoop fsck /test -list-corruptfileblocks also we can display the proper 
 command usage

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1592) Datanode startup doesn't honor volumes.tolerated

2011-05-25 Thread Jitendra Nath Pandey (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1592?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039240#comment-13039240
 ] 

Jitendra Nath Pandey commented on HDFS-1592:


+1 for the patch.

 Datanode startup doesn't honor volumes.tolerated 
 -

 Key: HDFS-1592
 URL: https://issues.apache.org/jira/browse/HDFS-1592
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.20.204.0
Reporter: Bharath Mundlapudi
Assignee: Bharath Mundlapudi
 Fix For: 0.20.204.0, 0.23.0

 Attachments: HDFS-1592-1.patch, HDFS-1592-2.patch, HDFS-1592-3.patch, 
 HDFS-1592-4.patch, HDFS-1592-5.patch, HDFS-1592-rel20.patch


 Datanode startup doesn't honor volumes.tolerated for hadoop 20 version.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-977) DataNode.createInterDataNodeProtocolProxy() guards a log at the wrong level

2011-05-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-977?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039251#comment-13039251
 ] 

Hadoop QA commented on HDFS-977:


-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12480430/HDFS-977.r2.diff
  against trunk revision 1127390.

+1 @author.  The patch does not contain any @author tags.

-1 tests included.  The patch doesn't appear to include any new or modified 
tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

-1 core tests.  The patch failed these core unit tests:
  org.apache.hadoop.hdfs.TestDFSStorageStateRecovery

+1 contrib tests.  The patch passed contrib unit tests.

+1 system test framework.  The patch passed system test framework compile.

Test results: 
https://builds.apache.org/hudson/job/PreCommit-HDFS-Build/624//testReport/
Findbugs warnings: 
https://builds.apache.org/hudson/job/PreCommit-HDFS-Build/624//artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Console output: 
https://builds.apache.org/hudson/job/PreCommit-HDFS-Build/624//console

This message is automatically generated.

 DataNode.createInterDataNodeProtocolProxy() guards a log at the wrong level
 ---

 Key: HDFS-977
 URL: https://issues.apache.org/jira/browse/HDFS-977
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: data-node
Affects Versions: 0.22.0
Reporter: Steve Loughran
Assignee: Harsh J Chouraria
Priority: Trivial
 Fix For: 0.23.0

 Attachments: HDFS-977.r1.diff, HDFS-977.r2.diff


 My IDE tells me that this code snippet in 
 {{DataNode.createInterDataNodeProtocolProxy()}} is guarding the info log 
 entry at debug level, and it should be reviewed
 {code}
 if (InterDatanodeProtocol.LOG.isDebugEnabled()) {
   InterDatanodeProtocol.LOG.info(InterDatanodeProtocol addr= + addr);
 }
 {code}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1884) Recurring failures in TestDFSStorageStateRecovery due to HADOOP-7146

2011-05-25 Thread Matt Foley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1884?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt Foley updated HDFS-1884:
-

Description: (was: Changing summary to make it more informative in the 
HDFS-1852 roll-up.)

 Recurring failures in TestDFSStorageStateRecovery due to HADOOP-7146
 

 Key: HDFS-1884
 URL: https://issues.apache.org/jira/browse/HDFS-1884
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: test
Reporter: Matt Foley
Assignee: Aaron T. Myers
 Attachments: hdfs-1884.0.patch




--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1983) Fix path display for copy rm

2011-05-25 Thread Daryn Sharp (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1983?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daryn Sharp updated HDFS-1983:
--

Status: Patch Available  (was: Open)

 Fix path display for copy  rm
 --

 Key: HDFS-1983
 URL: https://issues.apache.org/jira/browse/HDFS-1983
 Project: Hadoop HDFS
  Issue Type: Test
  Components: test
Affects Versions: 0.23.0
Reporter: Daryn Sharp
Assignee: Daryn Sharp
 Attachments: HDFS-1983.patch


 This will also fix a few misc broken tests.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1983) Fix path display for copy rm

2011-05-25 Thread Daryn Sharp (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1983?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daryn Sharp updated HDFS-1983:
--

Attachment: HDFS-1983.patch

Fixes all remaining issues in DFSShell and HDFSCLI tests.

 Fix path display for copy  rm
 --

 Key: HDFS-1983
 URL: https://issues.apache.org/jira/browse/HDFS-1983
 Project: Hadoop HDFS
  Issue Type: Test
  Components: test
Affects Versions: 0.23.0
Reporter: Daryn Sharp
Assignee: Daryn Sharp
 Attachments: HDFS-1983.patch


 This will also fix a few misc broken tests.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1943) fail to start datanode while start-dfs.sh is executed by root user

2011-05-25 Thread Bharath Mundlapudi (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1943?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039320#comment-13039320
 ] 

Bharath Mundlapudi commented on HDFS-1943:
--

+1 to this patch. 

 fail to start datanode while start-dfs.sh is executed by root user
 --

 Key: HDFS-1943
 URL: https://issues.apache.org/jira/browse/HDFS-1943
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: scripts
Affects Versions: 0.23.0
Reporter: Wei Yongjun
Priority: Blocker
 Fix For: 0.23.0

 Attachments: HDFS-1943.patch


 When start-dfs.sh is run by root user, we got the following error message:
 # start-dfs.sh
 Starting namenodes on [localhost ]
 localhost: namenode running as process 2556. Stop it first.
 localhost: starting datanode, logging to 
 /usr/hadoop/hadoop-common-0.23.0-SNAPSHOT/bin/../logs/hadoop-root-datanode-cspf01.out
 localhost: Unrecognized option: -jvm
 localhost: Could not create the Java virtual machine.
 The -jvm options should be passed to jsvc when we starting a secure
 datanode, but it still passed to java when start-dfs.sh is run by root
 while secure datanode is disabled. This is a bug of bin/hdfs.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1983) Fix path display for copy rm

2011-05-25 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1983?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039331#comment-13039331
 ] 

Todd Lipcon commented on HDFS-1983:
---

- looks like some hard tabs snuck in (in runCmd())
- can you use Joiner.on( ).join(args) instead of using the StringBuilder in 
the same function?

Otherwise, looks good

 Fix path display for copy  rm
 --

 Key: HDFS-1983
 URL: https://issues.apache.org/jira/browse/HDFS-1983
 Project: Hadoop HDFS
  Issue Type: Test
  Components: test
Affects Versions: 0.23.0
Reporter: Daryn Sharp
Assignee: Daryn Sharp
 Attachments: HDFS-1983.patch


 This will also fix a few misc broken tests.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-1131) each DFSClient instance uses a daemon thread for lease checking, there should be a singleton daemon for all DFSClient instances

2011-05-25 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1131?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE resolved HDFS-1131.
--

Resolution: Duplicate

 each DFSClient instance uses a daemon thread for lease checking, there should 
 be a singleton daemon for all DFSClient instances
 ---

 Key: HDFS-1131
 URL: https://issues.apache.org/jira/browse/HDFS-1131
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Alejandro Abdelnur
Priority: Critical

 When accessing HDFS from a server application that acts on behalf of several 
 users each user has its own file system handle (DFSClient), this creates one 
 daemon thread per file system instance.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1964) Incorrect HTML unescaping in DatanodeJspHelper.java

2011-05-25 Thread Aaron T. Myers (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1964?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-1964:
-

Attachment: hdfs-1964-22.0.patch

Sure, Todd. Here's a patch for branch-0.22.

 Incorrect HTML unescaping in DatanodeJspHelper.java
 ---

 Key: HDFS-1964
 URL: https://issues.apache.org/jira/browse/HDFS-1964
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.22.0, 0.23.0
Reporter: Aaron T. Myers
Assignee: Aaron T. Myers
 Fix For: 0.22.0, 0.23.0

 Attachments: hdfs-1964-22.0.patch, hdfs-1964-trunk.0.patch, 
 hdfs-1964-trunk.1.patch


 HDFS-1575 introduced some HTML unescaping of parameters so that viewing a 
 file would work for paths containing HTML-escaped characters, but in two of 
 the places did the unescaping either too early or too late.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1983) Fix path display for copy rm

2011-05-25 Thread Daryn Sharp (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1983?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daryn Sharp updated HDFS-1983:
--

Attachment: HDFS-1983-2.patch

Removed the tabs.  Time to change my vimrc...  We don't appear to currently 
have a dep on google's common base, so I didn't change to using a Joiner.  Let 
me know if that's not ok.

 Fix path display for copy  rm
 --

 Key: HDFS-1983
 URL: https://issues.apache.org/jira/browse/HDFS-1983
 Project: Hadoop HDFS
  Issue Type: Test
  Components: test
Affects Versions: 0.23.0
Reporter: Daryn Sharp
Assignee: Daryn Sharp
 Attachments: HDFS-1983-2.patch, HDFS-1983.patch


 This will also fix a few misc broken tests.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1943) fail to start datanode while start-dfs.sh is executed by root user

2011-05-25 Thread Jakob Homan (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1943?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039367#comment-13039367
 ] 

Jakob Homan commented on HDFS-1943:
---

+1

 fail to start datanode while start-dfs.sh is executed by root user
 --

 Key: HDFS-1943
 URL: https://issues.apache.org/jira/browse/HDFS-1943
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: scripts
Affects Versions: 0.23.0
Reporter: Wei Yongjun
Priority: Blocker
 Fix For: 0.23.0

 Attachments: HDFS-1943.patch


 When start-dfs.sh is run by root user, we got the following error message:
 # start-dfs.sh
 Starting namenodes on [localhost ]
 localhost: namenode running as process 2556. Stop it first.
 localhost: starting datanode, logging to 
 /usr/hadoop/hadoop-common-0.23.0-SNAPSHOT/bin/../logs/hadoop-root-datanode-cspf01.out
 localhost: Unrecognized option: -jvm
 localhost: Could not create the Java virtual machine.
 The -jvm options should be passed to jsvc when we starting a secure
 datanode, but it still passed to java when start-dfs.sh is run by root
 while secure datanode is disabled. This is a bug of bin/hdfs.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1943) fail to start datanode while start-dfs.sh is executed by root user

2011-05-25 Thread Jakob Homan (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1943?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039369#comment-13039369
 ] 

Jakob Homan commented on HDFS-1943:
---

It's not a good idea to run the dn as root, but this shouldn't be the way it's 
prevented.  Unit test failures can't be related.  Committing.

 fail to start datanode while start-dfs.sh is executed by root user
 --

 Key: HDFS-1943
 URL: https://issues.apache.org/jira/browse/HDFS-1943
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: scripts
Affects Versions: 0.23.0
Reporter: Wei Yongjun
Priority: Blocker
 Fix For: 0.23.0

 Attachments: HDFS-1943.patch


 When start-dfs.sh is run by root user, we got the following error message:
 # start-dfs.sh
 Starting namenodes on [localhost ]
 localhost: namenode running as process 2556. Stop it first.
 localhost: starting datanode, logging to 
 /usr/hadoop/hadoop-common-0.23.0-SNAPSHOT/bin/../logs/hadoop-root-datanode-cspf01.out
 localhost: Unrecognized option: -jvm
 localhost: Could not create the Java virtual machine.
 The -jvm options should be passed to jsvc when we starting a secure
 datanode, but it still passed to java when start-dfs.sh is run by root
 while secure datanode is disabled. This is a bug of bin/hdfs.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1943) fail to start datanode while start-dfs.sh is executed by root user

2011-05-25 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1943?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan updated HDFS-1943:
--

   Resolution: Fixed
Fix Version/s: (was: 0.23.0)
   Edit log branch (HDFS-1073)
 Assignee: Wei Yongjun
 Hadoop Flags: [Reviewed]
   Status: Resolved  (was: Patch Available)

Committed to trunk.  Resolving.  Thanks, Wei!

 fail to start datanode while start-dfs.sh is executed by root user
 --

 Key: HDFS-1943
 URL: https://issues.apache.org/jira/browse/HDFS-1943
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: scripts
Affects Versions: 0.23.0
Reporter: Wei Yongjun
Assignee: Wei Yongjun
Priority: Blocker
 Fix For: Edit log branch (HDFS-1073)

 Attachments: HDFS-1943.patch


 When start-dfs.sh is run by root user, we got the following error message:
 # start-dfs.sh
 Starting namenodes on [localhost ]
 localhost: namenode running as process 2556. Stop it first.
 localhost: starting datanode, logging to 
 /usr/hadoop/hadoop-common-0.23.0-SNAPSHOT/bin/../logs/hadoop-root-datanode-cspf01.out
 localhost: Unrecognized option: -jvm
 localhost: Could not create the Java virtual machine.
 The -jvm options should be passed to jsvc when we starting a secure
 datanode, but it still passed to java when start-dfs.sh is run by root
 while secure datanode is disabled. This is a bug of bin/hdfs.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1943) fail to start datanode while start-dfs.sh is executed by root user

2011-05-25 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1943?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan updated HDFS-1943:
--

Fix Version/s: (was: Edit log branch (HDFS-1073))
   0.23.0

 fail to start datanode while start-dfs.sh is executed by root user
 --

 Key: HDFS-1943
 URL: https://issues.apache.org/jira/browse/HDFS-1943
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: scripts
Affects Versions: 0.23.0
Reporter: Wei Yongjun
Assignee: Wei Yongjun
Priority: Blocker
 Fix For: 0.23.0

 Attachments: HDFS-1943.patch


 When start-dfs.sh is run by root user, we got the following error message:
 # start-dfs.sh
 Starting namenodes on [localhost ]
 localhost: namenode running as process 2556. Stop it first.
 localhost: starting datanode, logging to 
 /usr/hadoop/hadoop-common-0.23.0-SNAPSHOT/bin/../logs/hadoop-root-datanode-cspf01.out
 localhost: Unrecognized option: -jvm
 localhost: Could not create the Java virtual machine.
 The -jvm options should be passed to jsvc when we starting a secure
 datanode, but it still passed to java when start-dfs.sh is run by root
 while secure datanode is disabled. This is a bug of bin/hdfs.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1983) Fix path display for copy rm

2011-05-25 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1983?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039395#comment-13039395
 ] 

Todd Lipcon commented on HDFS-1983:
---

We should be picking up guava via common, which now depends on it. No big deal, 
though. +1 on this patch. Unfortunately Hudson seems stalled, but since this is 
just a test change, I will manually run the modified tests before commit.

 Fix path display for copy  rm
 --

 Key: HDFS-1983
 URL: https://issues.apache.org/jira/browse/HDFS-1983
 Project: Hadoop HDFS
  Issue Type: Test
  Components: test
Affects Versions: 0.23.0
Reporter: Daryn Sharp
Assignee: Daryn Sharp
 Attachments: HDFS-1983-2.patch, HDFS-1983.patch


 This will also fix a few misc broken tests.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1983) Fix path display for copy rm

2011-05-25 Thread Todd Lipcon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1983?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-1983:
--

   Resolution: Fixed
Fix Version/s: 0.23.0
 Hadoop Flags: [Reviewed]
   Status: Resolved  (was: Patch Available)

Related passed manually. Committed to trunk, thanks Daryn!

 Fix path display for copy  rm
 --

 Key: HDFS-1983
 URL: https://issues.apache.org/jira/browse/HDFS-1983
 Project: Hadoop HDFS
  Issue Type: Test
  Components: test
Affects Versions: 0.23.0
Reporter: Daryn Sharp
Assignee: Daryn Sharp
 Fix For: 0.23.0

 Attachments: HDFS-1983-2.patch, HDFS-1983.patch


 This will also fix a few misc broken tests.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1936) Updating the layout version from HDFS-1822 causes upgrade problems.

2011-05-25 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1936?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039398#comment-13039398
 ] 

Suresh Srinivas commented on HDFS-1936:
---

 imgVersion  -23
This is a bug in the existing code. Delegation token was added in -24. So the 
check should be  -24.

 Removed // added in -23
I would revert this. The comments are replete with when a particular field is 
added. I prefer not adding these details to comments, we can keep that 
documentation really up to date. 

 DataNode.getUpgradeManagerDatanode, which treats the DN like a singleton...
This is very similar to what is in 22. 22 uses a static method as well. I will 
look into this. If this is not closely related to this change, will address it 
in a separate bug.

 Updating the layout version from HDFS-1822 causes upgrade problems.
 ---

 Key: HDFS-1936
 URL: https://issues.apache.org/jira/browse/HDFS-1936
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: name-node
Affects Versions: 0.22.0, 0.23.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
Priority: Blocker
 Fix For: 0.22.0, 0.23.0

 Attachments: HDFS-1936.22.patch, HDFS-1936.trunk.patch, 
 hadoop-22-dfs-dir.tgz, hdfs-1936-with-testcase.txt


 In HDFS-1822 and HDFS-1842, the layout versions for 203, 204, 22 and trunk 
 were changed. Some of the namenode logic that depends on layout version is 
 broken because of this. Read the comment for more description.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-1996) ivy: hdfs test jar should be independent to common test jar

2011-05-25 Thread Tsz Wo (Nicholas), SZE (JIRA)
ivy: hdfs test jar should be independent to common test jar
---

 Key: HDFS-1996
 URL: https://issues.apache.org/jira/browse/HDFS-1996
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: build
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Eric Yang


hdfs tests and common tests may require different libraries, e.g. common tests 
need ftpserver for testing {{FTPFileSystem}} but hdfs does not.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-1997) Image transfer process misreports client side exceptions

2011-05-25 Thread Todd Lipcon (JIRA)
Image transfer process misreports client side exceptions


 Key: HDFS-1997
 URL: https://issues.apache.org/jira/browse/HDFS-1997
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Todd Lipcon
Assignee: Todd Lipcon




--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1997) Image transfer process misreports client side exceptions

2011-05-25 Thread Todd Lipcon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1997?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-1997:
--

  Component/s: name-node
  Description: If the image transfer process receives a client-side 
error when transferring edit logs, it will throw an exception before it has 
completely read all of the input from the server-side servlet. Then, the 
finally clause will throw a new error, since the received length is less than 
the length given in the header. This masks the client-side exception and makes 
it look like a network error or a server-side problem.
Affects Version/s: Edit log branch (HDFS-1073)
Fix Version/s: Edit log branch (HDFS-1073)

 Image transfer process misreports client side exceptions
 

 Key: HDFS-1997
 URL: https://issues.apache.org/jira/browse/HDFS-1997
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Affects Versions: Edit log branch (HDFS-1073)
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: Edit log branch (HDFS-1073)


 If the image transfer process receives a client-side error when transferring 
 edit logs, it will throw an exception before it has completely read all of 
 the input from the server-side servlet. Then, the finally clause will throw a 
 new error, since the received length is less than the length given in the 
 header. This masks the client-side exception and makes it look like a network 
 error or a server-side problem.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-1998) make refresh-namodenodes.sh refreshing all namenodes

2011-05-25 Thread Tanping Wang (JIRA)
make refresh-namodenodes.sh refreshing all namenodes


 Key: HDFS-1998
 URL: https://issues.apache.org/jira/browse/HDFS-1998
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: scripts
Affects Versions: 0.23.0
Reporter: Tanping Wang
Assignee: Tanping Wang
Priority: Minor
 Fix For: 0.23.0


refresh-namenodes.sh is used to refresh name nodes in the cluster to check for 
updates of include/exclude list.  It is used when decommissioning or adding a 
data node.  Currently it only refreshes the name node who serves the defaultFs, 
if there is defaultFs defined.  Fix it by refreshing all the name nodes in the 
cluster.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1983) Fix path display for copy rm

2011-05-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1983?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039414#comment-13039414
 ] 

Hudson commented on HDFS-1983:
--

Integrated in Hadoop-Hdfs-trunk-Commit #685 (See 
[https://builds.apache.org/hudson/job/Hadoop-Hdfs-trunk-Commit/685/])
HDFS-1983. Fix path display for copy and rm commands in TestHDFSCLI and 
TestDFSShell. Contributed by Daryn Sharp.

todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1127712
Files : 
* /hadoop/hdfs/trunk/src/test/hdfs/org/apache/hadoop/cli/testHDFSConf.xml
* /hadoop/hdfs/trunk/src/test/hdfs/org/apache/hadoop/hdfs/TestDFSShell.java
* /hadoop/hdfs/trunk/CHANGES.txt


 Fix path display for copy  rm
 --

 Key: HDFS-1983
 URL: https://issues.apache.org/jira/browse/HDFS-1983
 Project: Hadoop HDFS
  Issue Type: Test
  Components: test
Affects Versions: 0.23.0
Reporter: Daryn Sharp
Assignee: Daryn Sharp
 Fix For: 0.23.0

 Attachments: HDFS-1983-2.patch, HDFS-1983.patch


 This will also fix a few misc broken tests.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1943) fail to start datanode while start-dfs.sh is executed by root user

2011-05-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1943?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039413#comment-13039413
 ] 

Hudson commented on HDFS-1943:
--

Integrated in Hadoop-Hdfs-trunk-Commit #685 (See 
[https://builds.apache.org/hudson/job/Hadoop-Hdfs-trunk-Commit/685/])


 fail to start datanode while start-dfs.sh is executed by root user
 --

 Key: HDFS-1943
 URL: https://issues.apache.org/jira/browse/HDFS-1943
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: scripts
Affects Versions: 0.23.0
Reporter: Wei Yongjun
Assignee: Wei Yongjun
Priority: Blocker
 Fix For: 0.23.0

 Attachments: HDFS-1943.patch


 When start-dfs.sh is run by root user, we got the following error message:
 # start-dfs.sh
 Starting namenodes on [localhost ]
 localhost: namenode running as process 2556. Stop it first.
 localhost: starting datanode, logging to 
 /usr/hadoop/hadoop-common-0.23.0-SNAPSHOT/bin/../logs/hadoop-root-datanode-cspf01.out
 localhost: Unrecognized option: -jvm
 localhost: Could not create the Java virtual machine.
 The -jvm options should be passed to jsvc when we starting a secure
 datanode, but it still passed to java when start-dfs.sh is run by root
 while secure datanode is disabled. This is a bug of bin/hdfs.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1997) Image transfer process misreports client side exceptions

2011-05-25 Thread Todd Lipcon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1997?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-1997:
--

Attachment: hdfs-1997.txt

 Image transfer process misreports client side exceptions
 

 Key: HDFS-1997
 URL: https://issues.apache.org/jira/browse/HDFS-1997
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Affects Versions: 0.22.0, Edit log branch (HDFS-1073), 0.23.0
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: 0.22.0, Edit log branch (HDFS-1073), 0.23.0

 Attachments: hdfs-1997.txt


 If the image transfer process receives a client-side error when transferring 
 edit logs, it will throw an exception before it has completely read all of 
 the input from the server-side servlet. Then, the finally clause will throw a 
 new error, since the received length is less than the length given in the 
 header. This masks the client-side exception and makes it look like a network 
 error or a server-side problem.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1997) Image transfer process misreports client side exceptions

2011-05-25 Thread Todd Lipcon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1997?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-1997:
--

Affects Version/s: 0.23.0
   0.22.0
Fix Version/s: 0.23.0
   0.22.0

This affects trunk as well, so marking as such.

 Image transfer process misreports client side exceptions
 

 Key: HDFS-1997
 URL: https://issues.apache.org/jira/browse/HDFS-1997
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Affects Versions: 0.22.0, Edit log branch (HDFS-1073), 0.23.0
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: 0.22.0, Edit log branch (HDFS-1073), 0.23.0

 Attachments: hdfs-1997.txt


 If the image transfer process receives a client-side error when transferring 
 edit logs, it will throw an exception before it has completely read all of 
 the input from the server-side servlet. Then, the finally clause will throw a 
 new error, since the received length is less than the length given in the 
 header. This masks the client-side exception and makes it look like a network 
 error or a server-side problem.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1997) Image transfer process misreports client side exceptions

2011-05-25 Thread Todd Lipcon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1997?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-1997:
--

Status: Patch Available  (was: Open)

 Image transfer process misreports client side exceptions
 

 Key: HDFS-1997
 URL: https://issues.apache.org/jira/browse/HDFS-1997
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Affects Versions: 0.22.0, Edit log branch (HDFS-1073), 0.23.0
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: 0.22.0, Edit log branch (HDFS-1073), 0.23.0

 Attachments: hdfs-1997.txt


 If the image transfer process receives a client-side error when transferring 
 edit logs, it will throw an exception before it has completely read all of 
 the input from the server-side servlet. Then, the finally clause will throw a 
 new error, since the received length is less than the length given in the 
 header. This masks the client-side exception and makes it look like a network 
 error or a server-side problem.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-1999) Tests use deprecated configs

2011-05-25 Thread Aaron T. Myers (JIRA)
Tests use deprecated configs


 Key: HDFS-1999
 URL: https://issues.apache.org/jira/browse/HDFS-1999
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 0.22.0, 0.23.0
Reporter: Aaron T. Myers
Assignee: Aaron T. Myers
 Fix For: 0.22.0, 0.23.0


A few of the HDFS tests (not intended to test deprecation) use config keys 
which are deprecated.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-2000) Missing deprecation for io.bytes.per.checksum

2011-05-25 Thread Aaron T. Myers (JIRA)
Missing deprecation for io.bytes.per.checksum
-

 Key: HDFS-2000
 URL: https://issues.apache.org/jira/browse/HDFS-2000
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.22.0, 0.23.0
Reporter: Aaron T. Myers
Assignee: Aaron T. Myers
 Fix For: 0.22.0, 0.23.0


Hadoop long ago deprecated the configuration io.bytes.per.checksum in favor 
of dfs.bytes-per-checksum, but when the programmatic deprecation support was 
added, we didn't add an entry for this pair.

This is causing some tests to fail on branch-0.22 since the inclusion of 
HADOOP-7287, since some tests which were inadvertently using default config 
values are now having their settings actually picked up.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1994) Fix race conditions when running two rapidly checkpointing 2NNs

2011-05-25 Thread Todd Lipcon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1994?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-1994:
--

Attachment: hdfs-1994.txt

Updated patch that does solution a above. Also includes a new test case to 
trigger the interleaving and make sure only one is successful.

If we decide we also want to do b above, it could be done in a followup. But 
it's a very rare race that only really shows up in this kind of stress test, 
and it doesn't cause any corruption, just a missed checkpoint.

 Fix race conditions when running two rapidly checkpointing 2NNs
 ---

 Key: HDFS-1994
 URL: https://issues.apache.org/jira/browse/HDFS-1994
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Affects Versions: Edit log branch (HDFS-1073)
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: Edit log branch (HDFS-1073)

 Attachments: hdfs-1994.txt, hdfs-1994.txt


 HDFS-1984 added the ability to run two secondary namenodes at the same time. 
 However, there were two races I found when stress testing this (by running 
 two NNs each checkpointing in a tight loop with no sleep):
 1) the writing of the seen_txid file was not atomic, so it was at some points 
 reading an empty file
 2) it was possible for two checkpointers to try to take a checkpoint at the 
 same transaction ID, which would cause the two image downloads to collide and 
 fail

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-2000) Missing deprecation for io.bytes.per.checksum

2011-05-25 Thread Aaron T. Myers (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2000?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-2000:
-

Attachment: hdfs-2000-22.0.patch

Here's a patch for 0.22, which should also apply cleanly to trunk. No tests are 
included, but I can confirm that the presence of this patch fixes the 
{{TestFiRename}} test which had been broken on branch-0.22 since the commit of 
HADOOP-7287.

 Missing deprecation for io.bytes.per.checksum
 -

 Key: HDFS-2000
 URL: https://issues.apache.org/jira/browse/HDFS-2000
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.22.0, 0.23.0
Reporter: Aaron T. Myers
Assignee: Aaron T. Myers
 Fix For: 0.22.0, 0.23.0

 Attachments: hdfs-2000-22.0.patch


 Hadoop long ago deprecated the configuration io.bytes.per.checksum in favor 
 of dfs.bytes-per-checksum, but when the programmatic deprecation support 
 was added, we didn't add an entry for this pair.
 This is causing some tests to fail on branch-0.22 since the inclusion of 
 HADOOP-7287, since some tests which were inadvertently using default config 
 values are now having their settings actually picked up.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-2001) HDFS-1073: Kill previous.checkpoint, lastcheckpoint.tmp directories

2011-05-25 Thread Todd Lipcon (JIRA)
HDFS-1073: Kill previous.checkpoint, lastcheckpoint.tmp directories
---

 Key: HDFS-2001
 URL: https://issues.apache.org/jira/browse/HDFS-2001
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Todd Lipcon
Assignee: Todd Lipcon




--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-2000) Missing deprecation for io.bytes.per.checksum

2011-05-25 Thread Aaron T. Myers (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2000?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-2000:
-

Status: Patch Available  (was: Open)

 Missing deprecation for io.bytes.per.checksum
 -

 Key: HDFS-2000
 URL: https://issues.apache.org/jira/browse/HDFS-2000
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.22.0, 0.23.0
Reporter: Aaron T. Myers
Assignee: Aaron T. Myers
 Fix For: 0.22.0, 0.23.0

 Attachments: hdfs-2000-22.0.patch


 Hadoop long ago deprecated the configuration io.bytes.per.checksum in favor 
 of dfs.bytes-per-checksum, but when the programmatic deprecation support 
 was added, we didn't add an entry for this pair.
 This is causing some tests to fail on branch-0.22 since the inclusion of 
 HADOOP-7287, since some tests which were inadvertently using default config 
 values are now having their settings actually picked up.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1963) HDFS rpm integration project

2011-05-25 Thread Eric Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Yang updated HDFS-1963:


Attachment: HDFS-1963-4.patch

Make sure webapps directory is setup in 
$HADOOP_PREFIX/share/hadoop/hdfs/webapps, and hdfs script is setup accordingly.

 HDFS rpm integration project
 

 Key: HDFS-1963
 URL: https://issues.apache.org/jira/browse/HDFS-1963
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: build
 Environment: Java 6, RHEL 5.5
Reporter: Eric Yang
Assignee: Eric Yang
 Attachments: HDFS-1963-1.patch, HDFS-1963-2.patch, HDFS-1963-3.patch, 
 HDFS-1963-4.patch, HDFS-1963.patch


 This jira is corresponding to HADOOP-6255 and associated directory layout 
 change.  The patch for creating HDFS rpm packaging should be posted here for 
 patch test build to verify against hdfs svn trunk.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-2001) HDFS-1073: Kill previous.checkpoint, lastcheckpoint.tmp directories

2011-05-25 Thread Todd Lipcon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2001?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-2001:
--

  Component/s: name-node
  Description: 
In the old design for edits/image storage, the secondary namenode does a 
complicated dance of moving current/ to lastcheckpoint.tmp, checkpointing 
into current/, then moving lastcheckpoint.tmp back to 
previous.checkpoint. The idea here was so that there would always be one 
directory with a valid set of storage files.

In the HDFS-1073 design, this complicated dance isn't necessary. If a 
checkpoint fails, we can just rm that single fsimage_N.ckpt file and still be 
left with a valid storage directory.

So, we can just let the 2NN keep a single current/ dir around for all 
checkpoints and eliminate the complexity of the dance.
Affects Version/s: Edit log branch (HDFS-1073)
Fix Version/s: Edit log branch (HDFS-1073)

 HDFS-1073: Kill previous.checkpoint, lastcheckpoint.tmp directories
 ---

 Key: HDFS-2001
 URL: https://issues.apache.org/jira/browse/HDFS-2001
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Affects Versions: Edit log branch (HDFS-1073)
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: Edit log branch (HDFS-1073)


 In the old design for edits/image storage, the secondary namenode does a 
 complicated dance of moving current/ to lastcheckpoint.tmp, checkpointing 
 into current/, then moving lastcheckpoint.tmp back to 
 previous.checkpoint. The idea here was so that there would always be one 
 directory with a valid set of storage files.
 In the HDFS-1073 design, this complicated dance isn't necessary. If a 
 checkpoint fails, we can just rm that single fsimage_N.ckpt file and still be 
 left with a valid storage directory.
 So, we can just let the 2NN keep a single current/ dir around for all 
 checkpoints and eliminate the complexity of the dance.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1996) ivy: hdfs test jar should be independent to common test jar

2011-05-25 Thread Eric Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Yang updated HDFS-1996:


Attachment: HDFS-1996.patch

Make sure hdfs test profile contains junit jar file because it is a direct 
dependency.

 ivy: hdfs test jar should be independent to common test jar
 ---

 Key: HDFS-1996
 URL: https://issues.apache.org/jira/browse/HDFS-1996
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: build
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Eric Yang
 Attachments: HDFS-1996.patch


 hdfs tests and common tests may require different libraries, e.g. common 
 tests need ftpserver for testing {{FTPFileSystem}} but hdfs does not.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1996) ivy: hdfs test jar should be independent to common test jar

2011-05-25 Thread Eric Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Yang updated HDFS-1996:


Status: Patch Available  (was: Open)

Declare junit as direct dependency for hdfs test ivy profile.

 ivy: hdfs test jar should be independent to common test jar
 ---

 Key: HDFS-1996
 URL: https://issues.apache.org/jira/browse/HDFS-1996
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: build
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Eric Yang
 Attachments: HDFS-1996.patch


 hdfs tests and common tests may require different libraries, e.g. common 
 tests need ftpserver for testing {{FTPFileSystem}} but hdfs does not.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1996) ivy: hdfs test jar should be independent to common test jar

2011-05-25 Thread Tsz Wo (Nicholas), SZE (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1996?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039434#comment-13039434
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-1996:
--

Could you also change mockito from common-master to test-master?

 ivy: hdfs test jar should be independent to common test jar
 ---

 Key: HDFS-1996
 URL: https://issues.apache.org/jira/browse/HDFS-1996
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: build
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Eric Yang
 Attachments: HDFS-1996.patch


 hdfs tests and common tests may require different libraries, e.g. common 
 tests need ftpserver for testing {{FTPFileSystem}} but hdfs does not.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-2002) Incorrect computation of needed blocks in getTurnOffTip()

2011-05-25 Thread Konstantin Shvachko (JIRA)
Incorrect computation of needed blocks in getTurnOffTip()
-

 Key: HDFS-2002
 URL: https://issues.apache.org/jira/browse/HDFS-2002
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: name-node
Affects Versions: 0.22.0
Reporter: Konstantin Shvachko
 Fix For: 0.22.0


{{SafeModeInfo.getTurnOffTip()}} under-reports the number of blocks needed to 
reach the safemode threshold.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-2002) Incorrect computation of needed blocks in getTurnOffTip()

2011-05-25 Thread Konstantin Shvachko (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2002?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039447#comment-13039447
 ] 

Konstantin Shvachko commented on HDFS-2002:
---

In case there are 3 blocks in the system with the default threshold 0.9990, and 
with no replicas reported yet the safemode tip says:
_*{{Safe mode is ON. The reported blocks 0 needs additional 2 blocks to reach 
the threshold 0.9990 of total blocks 3. Safe mode will be turned off 
automatically.}}*_
While it should report one more _*{{needs additional 3 blocks}}*_.

And when the threshold is set larger than one, say 1.5, the numbers don't make 
any sense at all.

 Incorrect computation of needed blocks in getTurnOffTip()
 -

 Key: HDFS-2002
 URL: https://issues.apache.org/jira/browse/HDFS-2002
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: name-node
Affects Versions: 0.22.0
Reporter: Konstantin Shvachko
 Fix For: 0.22.0


 {{SafeModeInfo.getTurnOffTip()}} under-reports the number of blocks needed to 
 reach the safemode threshold.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1996) ivy: hdfs test jar should be independent to common test jar

2011-05-25 Thread Eric Yang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Yang updated HDFS-1996:


Attachment: HDFS-1996-1.patch

Change mockit to test-master.  Thanks Nicholas. :)

 ivy: hdfs test jar should be independent to common test jar
 ---

 Key: HDFS-1996
 URL: https://issues.apache.org/jira/browse/HDFS-1996
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: build
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Eric Yang
 Attachments: HDFS-1996-1.patch, HDFS-1996.patch


 hdfs tests and common tests may require different libraries, e.g. common 
 tests need ftpserver for testing {{FTPFileSystem}} but hdfs does not.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1875) MiniDFSCluster hard-codes dfs.datanode.address to localhost

2011-05-25 Thread Matt Foley (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1875?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039455#comment-13039455
 ] 

Matt Foley commented on HDFS-1875:
--

Hi Eric, looks generally good.  A few comments:

# MiniDFSCluster
** startDataNodes() - the @param checkDataNodeAddrConfig comment has been 
added to the javadocs for the stubbed-out original method signature.  It needs 
to be associated with the new longer-args-list method signature.

# TestDFSAddressConfig
** Configuration.unset() method is only available in trunk, not yahoo-merge, at 
this time.  This was added to trunk in HADOOP-7001 on Nov 24, 2010, but seems 
to not be in yahoo-merge.  Have you asked Suresh to merge HADOOP-7001 to 
yahoo-merge?

# DataNodeCluster
** In the class javadocs and/or the USAGE string, should document the change in 
semantics - that the datanode addresses specified in config will be used unless 
not set.

Also, with this setup, if you then run the Client on the same server as the 
DataNodeCluster, can the Client communicate successfully?  Or does it try to 
use 127.0.0.1 and fail to connect?  Have you searched for other unit test 
classes that use a Client and a DataNodeCluster?  Do they all still work?

Thanks.

 MiniDFSCluster hard-codes dfs.datanode.address to localhost
 ---

 Key: HDFS-1875
 URL: https://issues.apache.org/jira/browse/HDFS-1875
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 0.22.0
Reporter: Eric Payne
Assignee: Eric Payne
 Fix For: 0.23.0

 Attachments: HDFS-1875.patch


 When creating RPC addresses that represent the communication sockets for each 
 simulated DataNode, the MiniDFSCluster class hard-codes the address of the 
 dfs.datanode.address port to be 127.0.0.1:0
 The DataNodeCluster test tool uses the MiniDFSCluster class to create a 
 selected number of simulated datanodes on a single host. In the 
 DataNodeCluster setup, the NameNode is not simulated but is started as a 
 separate daemon.
 The problem is that if the write requrests into the simulated datanodes are 
 originated on a host that is not the same host running the simulated 
 datanodes, the connections are refused. This is because the RPC sockets that 
 are started by MiniDFSCluster are for localhost (127.0.0.1) and are not 
 accessible from outside that same machine.
 It is proposed that the MiniDFSCluster.setupDatanodeAddress() method be 
 overloaded in order to accommodate an environment where the NameNode is on 
 one host, the client is on another host, and the simulated DataNodes are on 
 yet another host (or even multiple hosts simulating multiple DataNodes each).
 The overloaded API would add a parameter that would be used as the basis for 
 creating the RPS sockets. By default, it would remain 127.0.0.1

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1996) ivy: hdfs test jar should be independent to common test jar

2011-05-25 Thread Tsz Wo (Nicholas), SZE (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1996?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039459#comment-13039459
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-1996:
--

+1 patch looks good.

 ivy: hdfs test jar should be independent to common test jar
 ---

 Key: HDFS-1996
 URL: https://issues.apache.org/jira/browse/HDFS-1996
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: build
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Eric Yang
 Attachments: HDFS-1996-1.patch, HDFS-1996.patch


 hdfs tests and common tests may require different libraries, e.g. common 
 tests need ftpserver for testing {{FTPFileSystem}} but hdfs does not.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1623) High Availability Framework for HDFS NN

2011-05-25 Thread Sanjay Radia (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1623?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sanjay Radia updated HDFS-1623:
---

Attachment: NameNode HA_v2_1.pdf

Update doc: v_2_1
Some minor updates I missed. Added section numbers

 High Availability Framework for HDFS NN
 ---

 Key: HDFS-1623
 URL: https://issues.apache.org/jira/browse/HDFS-1623
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Sanjay Radia
Assignee: Sanjay Radia
 Attachments: HDFS-High-Availability.pdf, NameNode HA_v2.pdf, NameNode 
 HA_v2_1.pdf, Namenode HA Framework.pdf




--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1934) Fix NullPointerException when certain File APIs return null

2011-05-25 Thread Bharath Mundlapudi (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1934?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bharath Mundlapudi updated HDFS-1934:
-

Affects Version/s: (was: 0.20.205.0)
Fix Version/s: (was: 0.20.205.0)

 Fix NullPointerException when certain File APIs return null
 ---

 Key: HDFS-1934
 URL: https://issues.apache.org/jira/browse/HDFS-1934
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.23.0
Reporter: Bharath Mundlapudi
Assignee: Bharath Mundlapudi
 Fix For: 0.23.0


 While testing Disk Fail Inplace, We encountered the NPE from this part of the 
 code. 
 File[] files = dir.listFiles();
 for (File f : files) {
 ...
 }
 This is kinda of an API issue. When a disk is bad (or name is not a 
 directory), this API (listFiles, list) return null rather than throwing an 
 exception. This 'for loop' throws a NPE exception. And same applies to 
 dir.list() API.
 Fix all the places where null condition was not checked.
  

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1875) MiniDFSCluster hard-codes dfs.datanode.address to localhost

2011-05-25 Thread Matt Foley (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1875?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039467#comment-13039467
 ] 

Matt Foley commented on HDFS-1875:
--

Assumed context for the last question above is a box where datanode config 
addresses are specified in the environment, e.g. by hdfs-site.xml.

 MiniDFSCluster hard-codes dfs.datanode.address to localhost
 ---

 Key: HDFS-1875
 URL: https://issues.apache.org/jira/browse/HDFS-1875
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 0.22.0
Reporter: Eric Payne
Assignee: Eric Payne
 Fix For: 0.23.0

 Attachments: HDFS-1875.patch


 When creating RPC addresses that represent the communication sockets for each 
 simulated DataNode, the MiniDFSCluster class hard-codes the address of the 
 dfs.datanode.address port to be 127.0.0.1:0
 The DataNodeCluster test tool uses the MiniDFSCluster class to create a 
 selected number of simulated datanodes on a single host. In the 
 DataNodeCluster setup, the NameNode is not simulated but is started as a 
 separate daemon.
 The problem is that if the write requrests into the simulated datanodes are 
 originated on a host that is not the same host running the simulated 
 datanodes, the connections are refused. This is because the RPC sockets that 
 are started by MiniDFSCluster are for localhost (127.0.0.1) and are not 
 accessible from outside that same machine.
 It is proposed that the MiniDFSCluster.setupDatanodeAddress() method be 
 overloaded in order to accommodate an environment where the NameNode is on 
 one host, the client is on another host, and the simulated DataNodes are on 
 yet another host (or even multiple hosts simulating multiple DataNodes each).
 The overloaded API would add a parameter that would be used as the basis for 
 creating the RPS sockets. By default, it would remain 127.0.0.1

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-2001) HDFS-1073: Kill previous.checkpoint, lastcheckpoint.tmp directories

2011-05-25 Thread Todd Lipcon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2001?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HDFS-2001:
--

Attachment: hdfs-2001.txt

This patch goes on top of HDFS-1991 and HDFS-1994.

It removes the previouscheckpoint/lastcheckpoint directories and updates 
TestDFSStorageStateRecovery to match that.

I currently commented out a couple of the test cases in TestSaveNamespace which 
need to still be updated and improved on this branch.

I also added a new test case for the scenario when the 2NN fails two 
checkpoints in a row. Previous to this patch, there was a bug with that 
situation, but this patch addresses it.

With this patch, I'm able to run a stress test with two 2NNs checkpointing as 
fast as they can for a while without issues. I'll let it loop for a few hours 
tonight.

 HDFS-1073: Kill previous.checkpoint, lastcheckpoint.tmp directories
 ---

 Key: HDFS-2001
 URL: https://issues.apache.org/jira/browse/HDFS-2001
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Affects Versions: Edit log branch (HDFS-1073)
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: Edit log branch (HDFS-1073)

 Attachments: hdfs-2001.txt


 In the old design for edits/image storage, the secondary namenode does a 
 complicated dance of moving current/ to lastcheckpoint.tmp, checkpointing 
 into current/, then moving lastcheckpoint.tmp back to 
 previous.checkpoint. The idea here was so that there would always be one 
 directory with a valid set of storage files.
 In the HDFS-1073 design, this complicated dance isn't necessary. If a 
 checkpoint fails, we can just rm that single fsimage_N.ckpt file and still be 
 left with a valid storage directory.
 So, we can just let the 2NN keep a single current/ dir around for all 
 checkpoints and eliminate the complexity of the dance.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1592) Datanode startup doesn't honor volumes.tolerated

2011-05-25 Thread Jitendra Nath Pandey (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1592?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039473#comment-13039473
 ] 

Jitendra Nath Pandey commented on HDFS-1592:


Eli,
   Do you have any more concerns? I intend to commit this patch by tomorrow.

 Datanode startup doesn't honor volumes.tolerated 
 -

 Key: HDFS-1592
 URL: https://issues.apache.org/jira/browse/HDFS-1592
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.20.204.0
Reporter: Bharath Mundlapudi
Assignee: Bharath Mundlapudi
 Fix For: 0.20.204.0, 0.23.0

 Attachments: HDFS-1592-1.patch, HDFS-1592-2.patch, HDFS-1592-3.patch, 
 HDFS-1592-4.patch, HDFS-1592-5.patch, HDFS-1592-rel20.patch


 Datanode startup doesn't honor volumes.tolerated for hadoop 20 version.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-420) Fuse-dfs should cache fs handles

2011-05-25 Thread Eli Collins (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins updated HDFS-420:
-

Fix Version/s: (was: 0.20.3)
   0.23.0
   Issue Type: Improvement  (was: Bug)
  Summary: Fuse-dfs should cache fs handles  (was: fuse_dfs is unable 
to connect to the dfs after a copying a large number of files into the dfs over 
fuse)

Latest patch looks good. I agree we should cache fs handles in fuse-dfs. In the 
updated patch attached I refactored out functions for creating the fs handle 
table, finding and inserting references into it. I also implemented 
doDisconnect, it works (I bet the segv you saw was from trying to free an entry 
in the table when the ref-count drops to zero) however the performance is much 
worse. I did some measurements comparing hdfsConnectAsUserNewInstance vs 
hdfsConnectAsUser and they perform similarly. The initial call is always slow 
because it creates a JVM, but after that it looks like the bulk of the time is 
spent loading configuration, the FileSystem cache in the Java client doesn't 
make a difference (even when the client and NN are a hop away). So it turns out 
the big performance gain comes from caching the fs handle in fuse-dfs vs 
getting a handle from libHdfs. Since eg an ls can result in many fuse-dfs 
operations it really pays to cache the handle.

The updated patch also contains the following additional changes:
* Improved the fix for HDFS-1757.
* Modifed doConnectAsUser to return the fs rather than use a IN/OUT arg.
* Made some the ERRORs in fuse_options.c INFOs since they're not error cases
* Removed the dead code in fuse_impls_access.c. This is HDFS-428.
* Removed the non-PERMS code since this no longer compiles, is not used.
* Removed the auto-connect in dfs_init since connection testing is already done 
in main.
* Minor update fuse_dfs_wrapper.sh to reflect the new ivy dir names.
* Removed doConnect, it's only used once and we don't want to encourage more 
uses.

I verified TestFuseDFS passes. Also did some stress testing of a fuse-dfs mount 
using multiple clients/users on a pseudo-cluster running from a build of trunk.

Mind taking a look at the latest code?

 Fuse-dfs should cache fs handles
 

 Key: HDFS-420
 URL: https://issues.apache.org/jira/browse/HDFS-420
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: contrib/fuse-dfs
Affects Versions: 0.20.2
 Environment: Fedora core 10, x86_64, 2.6.27.7-134.fc10.x86_64 #1 SMP 
 (AMD 64), gcc 4.3.2, java 1.6.0 (IcedTea6 1.4 (fedora-7.b12.fc10-x86_64) 
 Runtime Environment (build 1.6.0_0-b12) OpenJDK 64-Bit Server VM (build 
 10.0-b19, mixed mode)
Reporter: Dima Brodsky
Assignee: Brian Bockelman
 Fix For: 0.23.0

 Attachments: fuse_dfs_020_memleaks.patch, 
 fuse_dfs_020_memleaks_v3.patch, fuse_dfs_020_memleaks_v8.patch, 
 hdfs-420-1.patch


 I run the following test:
 1.  Run hadoop DFS in single node mode
 2.  start up fuse_dfs
 3.  copy my source tree, about 250 megs, into the DFS
  cp -av * /mnt/hdfs/
 in /var/log/messages I keep seeing:
 Dec 22 09:02:08 bodum fuse_dfs: ERROR: hdfs trying to utime 
 /bar/backend-trunk2/src/machinery/hadoop/output/2008/11/19 to 
 1229385138/1229963739
 and then eventually
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1333
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1333
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1333
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1333
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1209
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1209
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not 

[jira] [Updated] (HDFS-420) Fuse-dfs should cache fs handles

2011-05-25 Thread Eli Collins (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins updated HDFS-420:
-

Attachment: hdfs-420-1.patch

 Fuse-dfs should cache fs handles
 

 Key: HDFS-420
 URL: https://issues.apache.org/jira/browse/HDFS-420
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: contrib/fuse-dfs
Affects Versions: 0.20.2
 Environment: Fedora core 10, x86_64, 2.6.27.7-134.fc10.x86_64 #1 SMP 
 (AMD 64), gcc 4.3.2, java 1.6.0 (IcedTea6 1.4 (fedora-7.b12.fc10-x86_64) 
 Runtime Environment (build 1.6.0_0-b12) OpenJDK 64-Bit Server VM (build 
 10.0-b19, mixed mode)
Reporter: Dima Brodsky
Assignee: Brian Bockelman
 Fix For: 0.23.0

 Attachments: fuse_dfs_020_memleaks.patch, 
 fuse_dfs_020_memleaks_v3.patch, fuse_dfs_020_memleaks_v8.patch, 
 hdfs-420-1.patch


 I run the following test:
 1.  Run hadoop DFS in single node mode
 2.  start up fuse_dfs
 3.  copy my source tree, about 250 megs, into the DFS
  cp -av * /mnt/hdfs/
 in /var/log/messages I keep seeing:
 Dec 22 09:02:08 bodum fuse_dfs: ERROR: hdfs trying to utime 
 /bar/backend-trunk2/src/machinery/hadoop/output/2008/11/19 to 
 1229385138/1229963739
 and then eventually
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1333
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1333
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1333
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1333
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1209
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1209
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1333
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1209
 Dec 22 09:03:49 bodum fuse_dfs: ERROR: could not connect to dfs 
 fuse_dfs.c:1037
 and the file system hangs.  hadoop is still running and I don't see any 
 errors in it's logs.  I have to unmount the dfs and restart fuse_dfs and then 
 everything is fine again.  At some point I see the following messages in the 
 /var/log/messages:
 ERROR: dfs problem - could not close file_handle(139677114350528) for 
 /bar/backend-trunk2/src/machinery/hadoop/input/2008/12/14/actionrecordlog-8339-93825052368848-1229278807.log
  fuse_dfs.c:1464
 Dec 22 09:04:49 bodum fuse_dfs: ERROR: dfs problem - could not close 
 file_handle(139676770220176) for 
 /bar/backend-trunk2/src/machinery/hadoop/input/2008/12/14/actionrecordlog-8140-93825025883216-1229278759.log
  fuse_dfs.c:1464
 Dec 22 09:05:13 bodum fuse_dfs: ERROR: dfs problem - could not close 
 file_handle(139677114812832) for 
 /bar/backend-trunk2/src/machinery/hadoop/input/2008/12/14/actionrecordlog-8138-93825070138960-1229251587.log
  fuse_dfs.c:1464
 Is this a known issue?  Am I just flooding the system too much.  All of this 
 is being performed on a single, dual core, machine.
 Thanks!
 ttyl
 Dima

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1580) Add interface for generic Write Ahead Logging mechanisms

2011-05-25 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1580?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039478#comment-13039478
 ] 

Sanjay Radia commented on HDFS-1580:


Please add some text doc to motivate the 
Journal#getNumberOfTransactions(long 
sinceTxnId)
.
From what I understand this method gives the # of transactions since tX in a 
particular
storage dir. 
Would it be better instead to ask getHighestTxid()?




 Add interface for generic Write Ahead Logging mechanisms
 

 Key: HDFS-1580
 URL: https://issues.apache.org/jira/browse/HDFS-1580
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Ivan Kelly
 Fix For: Edit log branch (HDFS-1073)

 Attachments: EditlogInterface.1.pdf, EditlogInterface.2.pdf, 
 HDFS-1580+1521.diff, HDFS-1580.diff, HDFS-1580.diff, HDFS-1580.diff, 
 generic_wal_iface.pdf, generic_wal_iface.pdf, generic_wal_iface.pdf, 
 generic_wal_iface.txt




--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1996) ivy: hdfs test jar should be independent to common test jar

2011-05-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1996?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039484#comment-13039484
 ] 

Hadoop QA commented on HDFS-1996:
-

-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12480486/HDFS-1996-1.patch
  against trunk revision 1127712.

+1 @author.  The patch does not contain any @author tags.

-1 tests included.  The patch doesn't appear to include any new or modified 
tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

-1 core tests.  The patch failed these core unit tests:
  org.apache.hadoop.hdfs.TestDFSStorageStateRecovery

+1 contrib tests.  The patch passed contrib unit tests.

+1 system test framework.  The patch passed system test framework compile.

Test results: 
https://builds.apache.org/hudson/job/PreCommit-HDFS-Build/625//testReport/
Findbugs warnings: 
https://builds.apache.org/hudson/job/PreCommit-HDFS-Build/625//artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Console output: 
https://builds.apache.org/hudson/job/PreCommit-HDFS-Build/625//console

This message is automatically generated.

 ivy: hdfs test jar should be independent to common test jar
 ---

 Key: HDFS-1996
 URL: https://issues.apache.org/jira/browse/HDFS-1996
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: build
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Eric Yang
 Attachments: HDFS-1996-1.patch, HDFS-1996.patch


 hdfs tests and common tests may require different libraries, e.g. common 
 tests need ftpserver for testing {{FTPFileSystem}} but hdfs does not.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-1996) ivy: hdfs test jar should be independent to common test jar

2011-05-25 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-1996:
-

   Resolution: Fixed
Fix Version/s: 0.23.0
 Hadoop Flags: [Reviewed]
   Status: Resolved  (was: Patch Available)

I have committed this.  Thanks, Eric!

 ivy: hdfs test jar should be independent to common test jar
 ---

 Key: HDFS-1996
 URL: https://issues.apache.org/jira/browse/HDFS-1996
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: build
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Eric Yang
 Fix For: 0.23.0

 Attachments: HDFS-1996-1.patch, HDFS-1996.patch


 hdfs tests and common tests may require different libraries, e.g. common 
 tests need ftpserver for testing {{FTPFileSystem}} but hdfs does not.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1996) ivy: hdfs test jar should be independent to common test jar

2011-05-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1996?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039490#comment-13039490
 ] 

Hudson commented on HDFS-1996:
--

Integrated in Hadoop-Hdfs-trunk-Commit #686 (See 
[https://builds.apache.org/hudson/job/Hadoop-Hdfs-trunk-Commit/686/])
HDFS-1996.  ivy: hdfs test jar should be independent to common test jar.  
Contributed by Eric Yang

szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1127759
Files : 
* /hadoop/hdfs/trunk/CHANGES.txt
* /hadoop/hdfs/trunk/ivy.xml


 ivy: hdfs test jar should be independent to common test jar
 ---

 Key: HDFS-1996
 URL: https://issues.apache.org/jira/browse/HDFS-1996
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: build
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Eric Yang
 Fix For: 0.23.0

 Attachments: HDFS-1996-1.patch, HDFS-1996.patch


 hdfs tests and common tests may require different libraries, e.g. common 
 tests need ftpserver for testing {{FTPFileSystem}} but hdfs does not.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1592) Datanode startup doesn't honor volumes.tolerated

2011-05-25 Thread Eli Collins (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1592?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039509#comment-13039509
 ] 

Eli Collins commented on HDFS-1592:
---

+1 lgtm. Thanks Bharath

Nit: due an exception should due to exception

 Datanode startup doesn't honor volumes.tolerated 
 -

 Key: HDFS-1592
 URL: https://issues.apache.org/jira/browse/HDFS-1592
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.20.204.0
Reporter: Bharath Mundlapudi
Assignee: Bharath Mundlapudi
 Fix For: 0.20.204.0, 0.23.0

 Attachments: HDFS-1592-1.patch, HDFS-1592-2.patch, HDFS-1592-3.patch, 
 HDFS-1592-4.patch, HDFS-1592-5.patch, HDFS-1592-rel20.patch


 Datanode startup doesn't honor volumes.tolerated for hadoop 20 version.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-2000) Missing deprecation for io.bytes.per.checksum

2011-05-25 Thread Eli Collins (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2000?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins updated HDFS-2000:
--

Fix Version/s: (was: 0.23.0)
 Hadoop Flags: [Reviewed]

+1  pending Hudson

 Missing deprecation for io.bytes.per.checksum
 -

 Key: HDFS-2000
 URL: https://issues.apache.org/jira/browse/HDFS-2000
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.22.0, 0.23.0
Reporter: Aaron T. Myers
Assignee: Aaron T. Myers
 Fix For: 0.22.0

 Attachments: hdfs-2000-22.0.patch


 Hadoop long ago deprecated the configuration io.bytes.per.checksum in favor 
 of dfs.bytes-per-checksum, but when the programmatic deprecation support 
 was added, we didn't add an entry for this pair.
 This is causing some tests to fail on branch-0.22 since the inclusion of 
 HADOOP-7287, since some tests which were inadvertently using default config 
 values are now having their settings actually picked up.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-1991) HDFS-1073: Some refactoring of 2NN to easier share code with BN and CN

2011-05-25 Thread Eli Collins (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039519#comment-13039519
 ] 

Eli Collins commented on HDFS-1991:
---

+1  lgtm

 HDFS-1073: Some refactoring of 2NN to easier share code with BN and CN
 --

 Key: HDFS-1991
 URL: https://issues.apache.org/jira/browse/HDFS-1991
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: name-node
Affects Versions: Edit log branch (HDFS-1073)
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: Edit log branch (HDFS-1073)

 Attachments: hdfs-1991.txt


 Currently the Checkpointer class shares a lot of very similar code with the 
 secondary namenode.
 This JIRA is to do a little cleanup in SecondaryNameNode that will make it 
 easier to share code with the BackupNode and CheckpointNode.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira