[jira] [Commented] (HDFS-2197) Refactor RPC call implementations out of NameNode class

2011-09-08 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2197?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100114#comment-13100114
 ] 

Todd Lipcon commented on HDFS-2197:
---

This has been merged into the HA branch, but still pending merge into 0.23.

 Refactor RPC call implementations out of NameNode class
 ---

 Key: HDFS-2197
 URL: https://issues.apache.org/jira/browse/HDFS-2197
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Affects Versions: 0.23.0
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Attachments: hdfs-2197-2.txt, hdfs-2197-3.txt, hdfs-2197-4.txt, 
 hdfs-2197-trunk-final.txt, hdfs-2197.txt


 For HA, the NameNode will gain a bit of a state machine, to be able to 
 transition between standby and active states. This would be cleaner in the 
 code if the {{NameNode}} class were just a container for various services, as 
 discussed in HDFS-1974. It's also nice for testing, where it would become 
 easier to construct just the RPC handlers around a mock NameSystem, with no 
 HTTP server, for example.
 This JIRA is to move all of the protocol implementations out of {{NameNode}} 
 into a separate {{NameNodeRPCServer}} class.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2316) webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP

2011-09-08 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100120#comment-13100120
 ] 

Sanjay Radia commented on HDFS-2316:


HDFS-2178 (HOOP) is HDFS Proxy using http protocol - a replacement for hdfs 
proxy v2, but providing rw access. It runs as separate daemons *typically* on 
an array of servers sitting next to an HDFS cluster (like HDFS proxy v2).

HDFS-2316 is http rw access that replaces hftp but is *built into* the hdfs 
system and provides bandwidth 
scaling by redirecting from the NN to the datanode that contains the block. It 
will use spnego and delegation token. It does not requires a notion of trust. 
HDFS-2178 (HOOP) is run as a separate daemons that are trusted by hdfs. 
hdfs-2178 (HOOP)  can provide additional features like bandwidth management and 
user authentication mapping.

There is an overlap but a need for both.

 webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP
 --

 Key: HDFS-2316
 URL: https://issues.apache.org/jira/browse/HDFS-2316
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE

 We current have hftp for accessing HDFS over HTTP.  However, hftp is a 
 read-only FileSystem and does not provide write accesses.
 In HDFS-2284, we propose to have webhdfs for providing a complete FileSystem 
 implementation for accessing HDFS over HTTP.  The is the umbrella JIRA for 
 the tasks.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2284) Write Http access to HDFS

2011-09-08 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100124#comment-13100124
 ] 

Sanjay Radia commented on HDFS-2284:



@Sanjay  Hoop as a proxy is still valuable: bandwidth management, 
authentication translation, etc.

I has assumed that folks had understood that HDFS-2178 (HOOP) is still needed 
in addition to this this jira-- but I guess not. I have answered in more detail 
in HDFS-2316. 
Sorry I should have caught on earlier by Alejandro and ELi's comments:
@Alejandro And yes, Hoop supports all read/write/metadata FS operations.
@eli  Hoop is rw already btw.
Lets continue the discussion in the parent jira (2316)

 Write Http access to HDFS
 -

 Key: HDFS-2284
 URL: https://issues.apache.org/jira/browse/HDFS-2284
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Sanjay Radia
Assignee: Tsz Wo (Nicholas), SZE
 Attachments: h2284_20110902b.patch, h2284_20110904_0.20s.patch, 
 h2284_20110904b_0.20s.patch, h2284_20110905_0.20s.patch, 
 h2284_20110906_0.20s.patch, h2284_20110906b.patch, 
 h2284_20110906b_0.20s.patch, h2284_20110906c.patch, 
 h2284_20110906c_0.20s.patch, h2284_20110906d.patch, h2284_20110907.patch, 
 h2284_20110907_0.20s.patch


 HFTP allows on read access to HDFS via HTTP. Add write HTTP access to HDFS.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2298) TestDfsOverAvroRpc is failing on trunk

2011-09-08 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2298?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100128#comment-13100128
 ] 

Sanjay Radia commented on HDFS-2298:


I will comment in a couple of days after reading this in more detail.

 TestDfsOverAvroRpc is failing on trunk
 --

 Key: HDFS-2298
 URL: https://issues.apache.org/jira/browse/HDFS-2298
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Reporter: Aaron T. Myers
Assignee: Doug Cutting
 Fix For: 0.23.0

 Attachments: HDFS-2298.patch, HDFS-2298.patch, HDFS-2298.patch


 The relevant bit of the error:
 {noformat}
 ---
 Test set: org.apache.hadoop.hdfs.TestDfsOverAvroRpc
 ---
 Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 1.486 sec  
 FAILURE!
 testWorkingDirectory(org.apache.hadoop.hdfs.TestDfsOverAvroRpc)  Time 
 elapsed: 1.424 sec   ERROR!
 org.apache.avro.AvroTypeException: Two methods with same name: delete
 {noformat}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2251) Namenode does not recognize incorrectly sized blocks

2011-09-08 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2251?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100254#comment-13100254
 ] 

Uma Maheswara Rao G commented on HDFS-2251:
---

Hi Brian,

 In which version of Hadoop you faced this?

 Did you observe any recovery (or) write failures?


Thanks
Uma

 Namenode does not recognize incorrectly sized blocks
 

 Key: HDFS-2251
 URL: https://issues.apache.org/jira/browse/HDFS-2251
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Brian Bockelman

 We had a lot of file system corruption resulting in incorrectly sized blocks 
 (on disk, they're truncated to 192KB when they should be 64MB).
 However, I cannot make Hadoop realize that these blocks are incorrectly 
 sized.  When I try to drain off the node, I get the following messages:
 2008-10-29 18:46:51,293 WARN org.apache.hadoop.fs.FSNamesystem: Inconsistent 
 size for block blk_-4403534125663454855_9937 reported from 172.16.1.150:50010 
 current size is 67108864 reported size is 196608
 Here 172.16.1.150 is not the node which has the problematic block, but the 
 destination of the file transfer.  I propose that Hadoop should either:
 a) Upon startup, make sure that all blocks are properly sized (pro: rather 
 cheap check; con: doesn't catch any truncations which happen while on disk)
 b) Upon detecting the incorrectly sized copy, Hadoop should ask the source of 
 the block to perform a block verification.
 Thanks,
 Brian

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2232) TestHDFSCLI fails on 0.22 branch

2011-09-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2232?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100275#comment-13100275
 ] 

Hudson commented on HDFS-2232:
--

Integrated in Hadoop-Hdfs-trunk #787 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/787/])
HDFS-2232. Generalize regular expressions in TestHDFSCLI. Contributed by 
Plamen Jeliazkov.

shv : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1166470
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/resources/testHDFSConf.xml


 TestHDFSCLI fails on 0.22 branch
 

 Key: HDFS-2232
 URL: https://issues.apache.org/jira/browse/HDFS-2232
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 0.22.0
Reporter: Konstantin Shvachko
Assignee: Plamen Jeliazkov
Priority: Blocker
 Fix For: 0.22.0, 0.23.0, 0.24.0

 Attachments: HDFS-2232-TRUNK.patch, HDFS-2232-TRUNK.patch, 
 HDFS-2232-TRUNK.patch, HDFS-2232-TRUNK.patch, HDFS-2232.patch, 
 HDFS-2232.patch, HDFS-2232.patch, HDFS-2232.patch, 
 TEST-org.apache.hadoop.cli.TestHDFSCLI.txt, 
 TEST-org.apache.hadoop.cli.TestHDFSCLI.txt, TestHDFSCLI-output.txt


 Several HDFS CLI tests fail on 0.22 branch. I can see 2 reasons:
 # Not generic enough regular expression for host names and paths. Similar to 
 MAPREDUCE-2304.
 # Some command outputs have new-line in the end.
 # And some seem to produce [much] more output than expected.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2223) Untangle depencencies between NN components

2011-09-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100276#comment-13100276
 ] 

Hudson commented on HDFS-2223:
--

Integrated in Hadoop-Hdfs-trunk #787 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/787/])
HDFS-2223. Untangle depencencies between NN components. Contributed by Todd 
Lipcon.

todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1166466
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupImage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/Checkpointer.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormat.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNodeResourceChecker.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SecondaryNameNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/common/TestDistributedUpgrade.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCheckpoint.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestClusterId.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogRace.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestSaveNamespace.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/unit/org/apache/hadoop/hdfs/server/namenode/TestNNLeaseRecovery.java


 Untangle depencencies between NN components
 ---

 Key: HDFS-2223
 URL: https://issues.apache.org/jira/browse/HDFS-2223
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Affects Versions: 0.23.0
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: 0.24.0

 Attachments: hdfs-2223-1.txt, hdfs-2223-10.txt, hdfs-2223-2.txt, 
 hdfs-2223-3.txt, hdfs-2223-4.txt, hdfs-2223-5.txt, hdfs-2223-6.txt, 
 hdfs-2223-7.txt, hdfs-2223-8.txt, hdfs-2223-9.txt


 Working in the NN a lot for HA (HDFS-1623) I've come across a number of 
 situations where the tangled dependencies between NN components has been 
 problematic for adding new features and for testability. It would be good to 
 untangle some of these and clarify what the distinction is between the 
 different components: NameNode, FSNamesystem, FSDirectory, FSImage, 
 NNStorage, and FSEditLog

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2321) HDFS Build failure due to unresolved commons.daemon.os.name commons.daemon.os.arch properties in Windows 7

2011-09-08 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2321?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100280#comment-13100280
 ] 

Uma Maheswara Rao G commented on HDFS-2321:
---

I am proceeding with below conditions in my local environment as a workaround.

  condition property=commons.daemon.os.name value=linux
   os family=windows/
  /condition
  condition property=commons.daemon.os.arch value=i686
os family=windows/
  /condition

Here our building env is in windows, because of that we are getting this error 
for every updation. 

Alejandro, Do you have any better suggestion here to handle this?
 it will be problem if target os is solaris and building env is windows.

Thanks
Uma


 HDFS Build failure due to unresolved commons.daemon.os.name  
 commons.daemon.os.arch properties in Windows 7
 

 Key: HDFS-2321
 URL: https://issues.apache.org/jira/browse/HDFS-2321
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 0.24.0
 Environment: Windows 7, Windows XP
Reporter: Uma Maheswara Rao G

 This is the script for resolving the commons-deamons***tar.gz in hdfs pom.xml.
 target
 condition property=commons.daemon.os.name value=darwin
 os name=Mac OS X/
 /condition
 condition property=commons.daemon.os.arch 
 value=universal
 os name=Mac OS X/
 /condition
 condition property=commons.daemon.os.name value=linux
 os name=Linux /
 /condition
 !-- Set commons.daemon.os.arch to either i686 or x86_64 for 
 GNU/Linux --
 condition property=commons.daemon.os.arch value=x86_64
 os name=Linux arch=amd64/
 /condition
 condition property=commons.daemon.os.arch value=i686
 os name=Linux / !-- This is a guess --
 /condition
 property name=commons.daemon.tar.name
   
 value=commons-daemon-${commons-daemon.version}-bin-${commons.daemon.os.name}-${commons.daemon.os.arch}.tar.gz/
 Here we are not handling for windows.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2223) Untangle depencencies between NN components

2011-09-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100295#comment-13100295
 ] 

Hudson commented on HDFS-2223:
--

Integrated in Hadoop-Mapreduce-trunk #810 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/810/])
HDFS-2223. Untangle depencencies between NN components. Contributed by Todd 
Lipcon.

todd : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1166466
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupImage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/Checkpointer.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormat.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNodeResourceChecker.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SecondaryNameNode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/common/TestDistributedUpgrade.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCheckpoint.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestClusterId.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogRace.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestSaveNamespace.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/unit/org/apache/hadoop/hdfs/server/namenode/TestNNLeaseRecovery.java


 Untangle depencencies between NN components
 ---

 Key: HDFS-2223
 URL: https://issues.apache.org/jira/browse/HDFS-2223
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: name-node
Affects Versions: 0.23.0
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: 0.24.0

 Attachments: hdfs-2223-1.txt, hdfs-2223-10.txt, hdfs-2223-2.txt, 
 hdfs-2223-3.txt, hdfs-2223-4.txt, hdfs-2223-5.txt, hdfs-2223-6.txt, 
 hdfs-2223-7.txt, hdfs-2223-8.txt, hdfs-2223-9.txt


 Working in the NN a lot for HA (HDFS-1623) I've come across a number of 
 situations where the tangled dependencies between NN components has been 
 problematic for adding new features and for testability. It would be good to 
 untangle some of these and clarify what the distinction is between the 
 different components: NameNode, FSNamesystem, FSDirectory, FSImage, 
 NNStorage, and FSEditLog

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2232) TestHDFSCLI fails on 0.22 branch

2011-09-08 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2232?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100294#comment-13100294
 ] 

Hudson commented on HDFS-2232:
--

Integrated in Hadoop-Mapreduce-trunk #810 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/810/])
HDFS-2232. Generalize regular expressions in TestHDFSCLI. Contributed by 
Plamen Jeliazkov.

shv : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1166470
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/resources/testHDFSConf.xml


 TestHDFSCLI fails on 0.22 branch
 

 Key: HDFS-2232
 URL: https://issues.apache.org/jira/browse/HDFS-2232
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 0.22.0
Reporter: Konstantin Shvachko
Assignee: Plamen Jeliazkov
Priority: Blocker
 Fix For: 0.22.0, 0.23.0, 0.24.0

 Attachments: HDFS-2232-TRUNK.patch, HDFS-2232-TRUNK.patch, 
 HDFS-2232-TRUNK.patch, HDFS-2232-TRUNK.patch, HDFS-2232.patch, 
 HDFS-2232.patch, HDFS-2232.patch, HDFS-2232.patch, 
 TEST-org.apache.hadoop.cli.TestHDFSCLI.txt, 
 TEST-org.apache.hadoop.cli.TestHDFSCLI.txt, TestHDFSCLI-output.txt


 Several HDFS CLI tests fail on 0.22 branch. I can see 2 reasons:
 # Not generic enough regular expression for host names and paths. Similar to 
 MAPREDUCE-2304.
 # Some command outputs have new-line in the end.
 # And some seem to produce [much] more output than expected.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2321) HDFS Build failure due to unresolved commons.daemon.os.name commons.daemon.os.arch properties in Windows 7

2011-09-08 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2321?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100299#comment-13100299
 ] 

Uma Maheswara Rao G commented on HDFS-2321:
---

-D options also would work fine.
E:\Trunk-Commonmvn clean install -DskipTests -Dcommons.daemon.os.arch=i686 -Dco
mmons.daemon.os.name=linux

 HDFS Build failure due to unresolved commons.daemon.os.name  
 commons.daemon.os.arch properties in Windows 7
 

 Key: HDFS-2321
 URL: https://issues.apache.org/jira/browse/HDFS-2321
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 0.24.0
 Environment: Windows 7, Windows XP
Reporter: Uma Maheswara Rao G

 This is the script for resolving the commons-deamons***tar.gz in hdfs pom.xml.
 target
 condition property=commons.daemon.os.name value=darwin
 os name=Mac OS X/
 /condition
 condition property=commons.daemon.os.arch 
 value=universal
 os name=Mac OS X/
 /condition
 condition property=commons.daemon.os.name value=linux
 os name=Linux /
 /condition
 !-- Set commons.daemon.os.arch to either i686 or x86_64 for 
 GNU/Linux --
 condition property=commons.daemon.os.arch value=x86_64
 os name=Linux arch=amd64/
 /condition
 condition property=commons.daemon.os.arch value=i686
 os name=Linux / !-- This is a guess --
 /condition
 property name=commons.daemon.tar.name
   
 value=commons-daemon-${commons-daemon.version}-bin-${commons.daemon.os.name}-${commons.daemon.os.arch}.tar.gz/
 Here we are not handling for windows.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Assigned] (HDFS-197) du fails on Cygwin

2011-09-08 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-197?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J reassigned HDFS-197:


Assignee: Harsh J  (was: Kohsuke Kawaguchi)

 du fails on Cygwin
 

 Key: HDFS-197
 URL: https://issues.apache.org/jira/browse/HDFS-197
 Project: Hadoop HDFS
  Issue Type: Bug
 Environment: Windows + Cygwin
Reporter: Kohsuke Kawaguchi
Assignee: Harsh J
 Attachments: HADOOP-5486


 When I try to run a datanode on Windows, I get the following exception:
 {noformat}
 java.io.IOException: Expecting a line not the end of stream
   at org.apache.hadoop.fs.DU.parseExecResult(DU.java:181)
   at org.apache.hadoop.util.Shell.runCommand(Shell.java:179)
   at org.apache.hadoop.util.Shell.run(Shell.java:134)
   at org.apache.hadoop.fs.DU.init(DU.java:53)
   at org.apache.hadoop.fs.DU.init(DU.java:63)
   at 
 org.apache.hadoop.hdfs.server.datanode.FSDataset$FSVolume.init(FSDataset.java:325)
   at 
 org.apache.hadoop.hdfs.server.datanode.FSDataset.init(FSDataset.java:681)
   at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:291)
   at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:205)
   at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1238)
   at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1193)
 {noformat}
 This is because Hadoop execs du -sk C:\tmp\hadoop-SYSTEM\dfs\data with a 
 Windows path representation, which cygwin du doesn't understand.
 {noformat}
 C:\hudsondu -sk C:\tmp\hadoop-SYSTEM\dfs\data
 du -sk C:\tmp\hadoop-SYSTEM\dfs\data
 du: cannot access `C:\\tmp\\hadoop-SYSTEM\\dfs\\data': No such file or 
 directory
 {noformat}
 For this to work correctly, Hadoop would have to run cygpath first to get a 
 Unix path representation, then to call DU.
 Also, I had to use the debugger to get this information. Shell.runCommand 
 should catch IOException from parseExecResult and add the buffered stderr to 
 simplify the error diagnostics.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-197) du fails on Cygwin

2011-09-08 Thread Harsh J (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-197?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100325#comment-13100325
 ] 

Harsh J commented on HDFS-197:
--

(Assigning to self, will try a test case and update patch accordingly.)

 du fails on Cygwin
 

 Key: HDFS-197
 URL: https://issues.apache.org/jira/browse/HDFS-197
 Project: Hadoop HDFS
  Issue Type: Bug
 Environment: Windows + Cygwin
Reporter: Kohsuke Kawaguchi
Assignee: Harsh J
 Attachments: HADOOP-5486


 When I try to run a datanode on Windows, I get the following exception:
 {noformat}
 java.io.IOException: Expecting a line not the end of stream
   at org.apache.hadoop.fs.DU.parseExecResult(DU.java:181)
   at org.apache.hadoop.util.Shell.runCommand(Shell.java:179)
   at org.apache.hadoop.util.Shell.run(Shell.java:134)
   at org.apache.hadoop.fs.DU.init(DU.java:53)
   at org.apache.hadoop.fs.DU.init(DU.java:63)
   at 
 org.apache.hadoop.hdfs.server.datanode.FSDataset$FSVolume.init(FSDataset.java:325)
   at 
 org.apache.hadoop.hdfs.server.datanode.FSDataset.init(FSDataset.java:681)
   at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:291)
   at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:205)
   at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1238)
   at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1193)
 {noformat}
 This is because Hadoop execs du -sk C:\tmp\hadoop-SYSTEM\dfs\data with a 
 Windows path representation, which cygwin du doesn't understand.
 {noformat}
 C:\hudsondu -sk C:\tmp\hadoop-SYSTEM\dfs\data
 du -sk C:\tmp\hadoop-SYSTEM\dfs\data
 du: cannot access `C:\\tmp\\hadoop-SYSTEM\\dfs\\data': No such file or 
 directory
 {noformat}
 For this to work correctly, Hadoop would have to run cygpath first to get a 
 Unix path representation, then to call DU.
 Also, I had to use the debugger to get this information. Shell.runCommand 
 should catch IOException from parseExecResult and add the buffered stderr to 
 simplify the error diagnostics.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2271) startJournalSpool should invoke ProcessIOError with failed storage directories if createEditLogFile throws any exception.

2011-09-08 Thread Uma Maheswara Rao G (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uma Maheswara Rao G updated HDFS-2271:
--

Component/s: name-node

 startJournalSpool should invoke ProcessIOError with failed storage 
 directories if createEditLogFile throws any exception.  
 ---

 Key: HDFS-2271
 URL: https://issues.apache.org/jira/browse/HDFS-2271
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: name-node
Affects Versions: 0.22.0
Reporter: Uma Maheswara Rao G
Assignee: Uma Maheswara Rao G
 Attachments: HDFS-2271.patch


 Even If createEditsLogFile failes in startJournalSpool of BackUpStorage, 
 BackUPNode will proceed with exceptions. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2060) DFS client RPCs using protobufs

2011-09-08 Thread Doug Cutting (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2060?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100466#comment-13100466
 ] 

Doug Cutting commented on HDFS-2060:


I agree that the MR2-style protocol wrappers are a good pattern, but I'm not 
convinced they're required before this could be committed.  The current 
Writable-based RPC implementation is not wrapped, nor does it provide version 
interoperability.  This patch addresses version-interoperability but not 
wrapping.  Wrapping seems like it might be handled as an independent issue, 
especially if version-interoperability is a priority in the near term.

 DFS client RPCs using protobufs
 ---

 Key: HDFS-2060
 URL: https://issues.apache.org/jira/browse/HDFS-2060
 Project: Hadoop HDFS
  Issue Type: New Feature
Affects Versions: 0.23.0
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Attachments: hdfs-2060-getblocklocations.txt


 The most important place for wire-compatibility in DFS is between clients and 
 the cluster, since lockstep upgrade is very difficult and a single client may 
 want to talk to multiple server versions. So, I'd like to focus this JIRA on 
 making the RPCs between the DFS client and the NN/DNs wire-compatible using 
 protocol buffer based serialization.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-1242) 0.20 append: Add test for appendFile() race solved in HDFS-142

2011-09-08 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1242?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HDFS-1242:
---

Fix Version/s: 0.20.205.0

 0.20 append: Add test for appendFile() race solved in HDFS-142
 --

 Key: HDFS-1242
 URL: https://issues.apache.org/jira/browse/HDFS-1242
 Project: Hadoop HDFS
  Issue Type: Test
Affects Versions: 0.20-append
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: 0.20-append, 0.20.205.0

 Attachments: hdfs-1242.txt


 This is a unit test that didn't make it into branch-0.20-append, but worth 
 having in TestFileAppend4.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-1242) 0.20 append: Add test for appendFile() race solved in HDFS-142

2011-09-08 Thread Jitendra Nath Pandey (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1242?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100577#comment-13100577
 ] 

Jitendra Nath Pandey commented on HDFS-1242:


+1. It is a useful test.

 0.20 append: Add test for appendFile() race solved in HDFS-142
 --

 Key: HDFS-1242
 URL: https://issues.apache.org/jira/browse/HDFS-1242
 Project: Hadoop HDFS
  Issue Type: Test
Affects Versions: 0.20-append
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: 0.20-append, 0.20.205.0

 Attachments: hdfs-1242.txt


 This is a unit test that didn't make it into branch-0.20-append, but worth 
 having in TestFileAppend4.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-2322) the build fails in Windows because commons-daemon TAR cannot be fetched

2011-09-08 Thread Alejandro Abdelnur (JIRA)
the build fails in Windows because commons-daemon TAR cannot be fetched
---

 Key: HDFS-2322
 URL: https://issues.apache.org/jira/browse/HDFS-2322
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 0.23.0, 0.24.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Fix For: 0.23.0, 0.24.0


For windows there is no commons-daemon TAR but a ZIP, plus the name follows a 
different convention. 



--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2322) the build fails in Windows because commons-daemon TAR cannot be fetched

2011-09-08 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2322?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100579#comment-13100579
 ] 

Alejandro Abdelnur commented on HDFS-2322:
--

As native stuff is not supported on windows, we should just skip the 
downloading of the commons-daemons TAR

 the build fails in Windows because commons-daemon TAR cannot be fetched
 ---

 Key: HDFS-2322
 URL: https://issues.apache.org/jira/browse/HDFS-2322
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 0.23.0, 0.24.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Fix For: 0.23.0, 0.24.0


 For windows there is no commons-daemon TAR but a ZIP, plus the name follows a 
 different convention. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2322) the build fails in Windows because commons-daemon TAR cannot be fetched

2011-09-08 Thread Alejandro Abdelnur (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2322?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alejandro Abdelnur updated HDFS-2322:
-

Status: Patch Available  (was: Open)

 the build fails in Windows because commons-daemon TAR cannot be fetched
 ---

 Key: HDFS-2322
 URL: https://issues.apache.org/jira/browse/HDFS-2322
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 0.23.0, 0.24.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Fix For: 0.23.0, 0.24.0

 Attachments: HDFS-2322v1.patch


 For windows there is no commons-daemon TAR but a ZIP, plus the name follows a 
 different convention. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2322) the build fails in Windows because commons-daemon TAR cannot be fetched

2011-09-08 Thread Alejandro Abdelnur (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2322?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alejandro Abdelnur updated HDFS-2322:
-

Attachment: HDFS-2322v1.patch

 the build fails in Windows because commons-daemon TAR cannot be fetched
 ---

 Key: HDFS-2322
 URL: https://issues.apache.org/jira/browse/HDFS-2322
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 0.23.0, 0.24.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Fix For: 0.23.0, 0.24.0

 Attachments: HDFS-2322v1.patch


 For windows there is no commons-daemon TAR but a ZIP, plus the name follows a 
 different convention. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2322) the build fails in Windows because commons-daemon TAR cannot be fetched

2011-09-08 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2322?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100594#comment-13100594
 ] 

Uma Maheswara Rao G commented on HDFS-2322:
---

Duplicate of HDFS-2321

 the build fails in Windows because commons-daemon TAR cannot be fetched
 ---

 Key: HDFS-2322
 URL: https://issues.apache.org/jira/browse/HDFS-2322
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 0.23.0, 0.24.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Fix For: 0.23.0, 0.24.0

 Attachments: HDFS-2322v1.patch


 For windows there is no commons-daemon TAR but a ZIP, plus the name follows a 
 different convention. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-1242) 0.20 append: Add test for appendFile() race solved in HDFS-142

2011-09-08 Thread Jitendra Nath Pandey (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1242?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100596#comment-13100596
 ] 

Jitendra Nath Pandey commented on HDFS-1242:


I have committed to 20-security. Thanks to Todd.

 0.20 append: Add test for appendFile() race solved in HDFS-142
 --

 Key: HDFS-1242
 URL: https://issues.apache.org/jira/browse/HDFS-1242
 Project: Hadoop HDFS
  Issue Type: Test
Affects Versions: 0.20-append
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: 0.20-append, 0.20.205.0

 Attachments: hdfs-1242.txt


 This is a unit test that didn't make it into branch-0.20-append, but worth 
 having in TestFileAppend4.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2321) HDFS Build failure due to unresolved commons.daemon.os.name commons.daemon.os.arch properties in Windows 7

2011-09-08 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2321?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100625#comment-13100625
 ] 

Alejandro Abdelnur commented on HDFS-2321:
--

Uma, I wouldn't force the os.arch/os.name to linux, instead I'd skip that 
download as per patch HDFS-2322

 HDFS Build failure due to unresolved commons.daemon.os.name  
 commons.daemon.os.arch properties in Windows 7
 

 Key: HDFS-2321
 URL: https://issues.apache.org/jira/browse/HDFS-2321
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 0.24.0
 Environment: Windows 7, Windows XP
Reporter: Uma Maheswara Rao G

 This is the script for resolving the commons-deamons***tar.gz in hdfs pom.xml.
 target
 condition property=commons.daemon.os.name value=darwin
 os name=Mac OS X/
 /condition
 condition property=commons.daemon.os.arch 
 value=universal
 os name=Mac OS X/
 /condition
 condition property=commons.daemon.os.name value=linux
 os name=Linux /
 /condition
 !-- Set commons.daemon.os.arch to either i686 or x86_64 for 
 GNU/Linux --
 condition property=commons.daemon.os.arch value=x86_64
 os name=Linux arch=amd64/
 /condition
 condition property=commons.daemon.os.arch value=i686
 os name=Linux / !-- This is a guess --
 /condition
 property name=commons.daemon.tar.name
   
 value=commons-daemon-${commons-daemon.version}-bin-${commons.daemon.os.name}-${commons.daemon.os.arch}.tar.gz/
 Here we are not handling for windows.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2322) the build fails in Windows because commons-daemon TAR cannot be fetched

2011-09-08 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2322?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100647#comment-13100647
 ] 

Hadoop QA commented on HDFS-2322:
-

-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12493640/HDFS-2322v1.patch
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

-1 tests included.  The patch doesn't appear to include any new or modified 
tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

-1 core tests.  The patch failed these unit tests:
  org.apache.hadoop.hdfs.TestDfsOverAvroRpc
  
org.apache.hadoop.hdfs.server.blockmanagement.TestHost2NodesMap
  org.apache.hadoop.hdfs.server.datanode.TestReplicasMap

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/1224//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HDFS-Build/1224//artifact/trunk/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/1224//console

This message is automatically generated.

 the build fails in Windows because commons-daemon TAR cannot be fetched
 ---

 Key: HDFS-2322
 URL: https://issues.apache.org/jira/browse/HDFS-2322
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 0.23.0, 0.24.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Fix For: 0.23.0, 0.24.0

 Attachments: HDFS-2322v1.patch


 For windows there is no commons-daemon TAR but a ZIP, plus the name follows a 
 different convention. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-1122) client block verification may result in blocks in DataBlockScanner prematurely

2011-09-08 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HDFS-1122:
---

Attachment: HDFS-1122-20-security.2.patch

The attached patch is Sam's original patch with following changes
1) Rebased against 20-security.
2) Changed testcase to junit4.
3) Addressed Todd's comments.

 client block verification may result in blocks in DataBlockScanner prematurely
 --

 Key: HDFS-1122
 URL: https://issues.apache.org/jira/browse/HDFS-1122
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: 0.20-append
Reporter: sam rash
Assignee: sam rash
 Attachments: HDFS-1122-20-security.2.patch, hdfs-1122-for-0.20.txt


 found that when the DN uses client verification of a block that is open for 
 writing, it will add it to the DataBlockScanner prematurely. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-1122) client block verification may result in blocks in DataBlockScanner prematurely

2011-09-08 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HDFS-1122:
---

Affects Version/s: 0.20.205.0
Fix Version/s: 0.20.205.0
   0.20-append

 client block verification may result in blocks in DataBlockScanner prematurely
 --

 Key: HDFS-1122
 URL: https://issues.apache.org/jira/browse/HDFS-1122
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: 0.20-append, 0.20.205.0
Reporter: sam rash
Assignee: sam rash
 Fix For: 0.20-append, 0.20.205.0

 Attachments: HDFS-1122-20-security.2.patch, hdfs-1122-for-0.20.txt


 found that when the DN uses client verification of a block that is open for 
 writing, it will add it to the DataBlockScanner prematurely. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-1122) client block verification may result in blocks in DataBlockScanner prematurely

2011-09-08 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100671#comment-13100671
 ] 

Suresh Srinivas commented on HDFS-1122:
---

+1 for the patch.

 client block verification may result in blocks in DataBlockScanner prematurely
 --

 Key: HDFS-1122
 URL: https://issues.apache.org/jira/browse/HDFS-1122
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: 0.20-append, 0.20.205.0
Reporter: sam rash
Assignee: sam rash
 Fix For: 0.20-append, 0.20.205.0

 Attachments: HDFS-1122-20-security.2.patch, hdfs-1122-for-0.20.txt


 found that when the DN uses client verification of a block that is open for 
 writing, it will add it to the DataBlockScanner prematurely. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2246) Shortcut a local client reads to a Datanodes files directly

2011-09-08 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100675#comment-13100675
 ] 

Sanjay Radia commented on HDFS-2246:


getBlockPathInfo(block) - add  block access token parameter which is verified 
on the data node.

 Shortcut a local client reads to a Datanodes files directly
 ---

 Key: HDFS-2246
 URL: https://issues.apache.org/jira/browse/HDFS-2246
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Sanjay Radia
 Attachments: 0001-HDFS-347.-Local-reads.patch




--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2320) Make merged protocol changes from 0.20-append to 0.20-security compatible with previous releases.

2011-09-08 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2320?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-2320:
--

Attachment: HDFS-2320.patch

Updated patch to remove an unnecessary file change.

I also tested this, killing datanodes while running TestDFSIO

 Make merged protocol changes from 0.20-append to 0.20-security compatible 
 with previous releases.
 -

 Key: HDFS-2320
 URL: https://issues.apache.org/jira/browse/HDFS-2320
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: data-node, hdfs client, name-node
Affects Versions: 0.20.205.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
 Fix For: 0.20.205.0

 Attachments: HDFS-2320.patch, HDFS-2320.patch


 0.20-append changes have been merged to 0.20-security. The merge has changes 
 to version numbers in several protocols. This jira makes the protocol changes 
 compatible with older release, allowing clients running older version to talk 
 to server running 205 version and clients running 205 version talk to older 
 servers running 203, 204.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HDFS-2323) start-dfs.sh script fails for tarball install

2011-09-08 Thread Tom White (JIRA)
start-dfs.sh script fails for tarball install
-

 Key: HDFS-2323
 URL: https://issues.apache.org/jira/browse/HDFS-2323
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Tom White
Assignee: Tom White


I build Common and HDFS tarballs from trunk then tried to start a cluster with 
start-dfs.sh, but I got the following error:

{noformat}
Starting namenodes on [localhost ]
sbin/start-dfs.sh: line 55: 
/Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
directory
sbin/start-dfs.sh: line 68: 
/Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
directory
Starting secondary namenodes [0.0.0.0 ]
sbin/start-dfs.sh: line 88: 
/Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
directory
{noformat}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2323) start-dfs.sh script fails for tarball install

2011-09-08 Thread Tom White (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2323?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tom White updated HDFS-2323:


Attachment: HDFS-2323.patch

There are a couple of problems:

1. The scripts refer to bin/hadoop-daemons.sh not sbin/hadoop-daemons.sh.
2. The secondary namenode should not be started at all since it's not 
configured to, but the script tries to start it since the test for 0.0.0.0 
fails due to a trailing space.

This patch fixes both of these.

 start-dfs.sh script fails for tarball install
 -

 Key: HDFS-2323
 URL: https://issues.apache.org/jira/browse/HDFS-2323
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Tom White
Assignee: Tom White
 Attachments: HDFS-2323.patch


 I build Common and HDFS tarballs from trunk then tried to start a cluster 
 with start-dfs.sh, but I got the following error:
 {noformat}
 Starting namenodes on [localhost ]
 sbin/start-dfs.sh: line 55: 
 /Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
 directory
 sbin/start-dfs.sh: line 68: 
 /Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
 directory
 Starting secondary namenodes [0.0.0.0 ]
 sbin/start-dfs.sh: line 88: 
 /Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
 directory
 {noformat}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2323) start-dfs.sh script fails for tarball install

2011-09-08 Thread Tom White (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2323?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100782#comment-13100782
 ] 

Tom White commented on HDFS-2323:
-

To ran a pseudo distributed HDFS cluster I ran 

{noformat}
mvn clean package -Pdist -Dtar -DskipTests -P-cbuild
{noformat}

from the top level, then from another directory

{noformat}
mkdir hadoop
cd hadoop
tar zxf 
~/workspace/hadoop-trunk/hadoop-hdfs-project/hadoop-hdfs/target/hadoop-hdfs-0.24.0-SNAPSHOT.tar.gz
 --strip 1
tar zxf 
~/workspace/hadoop-trunk/hadoop-common-project/hadoop-common/target/hadoop-common-0.24.0-SNAPSHOT.tar.gz
 --strip 1
export HADOOP_COMMON_HOME=$(pwd)
export HADOOP_HDFS_HOME=$(pwd)

cat  $HADOOP_COMMON_HOME/etc/hadoop/core-site.xml   EOF
?xml version=1.0?
!-- core-site.xml --
configuration
  property
namefs.default.name/name
valuehdfs://localhost//value
  /property
/configuration


EOF

sbin/start-dfs.sh 
{noformat}

 start-dfs.sh script fails for tarball install
 -

 Key: HDFS-2323
 URL: https://issues.apache.org/jira/browse/HDFS-2323
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Tom White
Assignee: Tom White
 Attachments: HDFS-2323.patch


 I build Common and HDFS tarballs from trunk then tried to start a cluster 
 with start-dfs.sh, but I got the following error:
 {noformat}
 Starting namenodes on [localhost ]
 sbin/start-dfs.sh: line 55: 
 /Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
 directory
 sbin/start-dfs.sh: line 68: 
 /Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
 directory
 Starting secondary namenodes [0.0.0.0 ]
 sbin/start-dfs.sh: line 88: 
 /Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
 directory
 {noformat}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HDFS-2323) start-dfs.sh script fails for tarball install

2011-09-08 Thread Tom White (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2323?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tom White updated HDFS-2323:


Status: Patch Available  (was: Open)

 start-dfs.sh script fails for tarball install
 -

 Key: HDFS-2323
 URL: https://issues.apache.org/jira/browse/HDFS-2323
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Tom White
Assignee: Tom White
 Attachments: HDFS-2323.patch


 I build Common and HDFS tarballs from trunk then tried to start a cluster 
 with start-dfs.sh, but I got the following error:
 {noformat}
 Starting namenodes on [localhost ]
 sbin/start-dfs.sh: line 55: 
 /Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
 directory
 sbin/start-dfs.sh: line 68: 
 /Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
 directory
 Starting secondary namenodes [0.0.0.0 ]
 sbin/start-dfs.sh: line 88: 
 /Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
 directory
 {noformat}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2316) webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP

2011-09-08 Thread Eli Collins (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100797#comment-13100797
 ] 

Eli Collins commented on HDFS-2316:
---

I definitely agree with these distinct use cases, I'm just wondering if we need 
to have two separate FileSystem over HTTP implementations vs one client that 
may or may not use a proxy server (there's no reason http or FileSystem clients 
need to care whether they're being proxied). Sounds like we were duplicating 
code w/o understanding what could be shared. You and Alejandro have looked at 
the specifics more than I have so I trust your judgement.


 webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP
 --

 Key: HDFS-2316
 URL: https://issues.apache.org/jira/browse/HDFS-2316
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE

 We current have hftp for accessing HDFS over HTTP.  However, hftp is a 
 read-only FileSystem and does not provide write accesses.
 In HDFS-2284, we propose to have webhdfs for providing a complete FileSystem 
 implementation for accessing HDFS over HTTP.  The is the umbrella JIRA for 
 the tasks.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2320) Make merged protocol changes from 0.20-append to 0.20-security compatible with previous releases.

2011-09-08 Thread Jitendra Nath Pandey (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2320?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100801#comment-13100801
 ] 

Jitendra Nath Pandey commented on HDFS-2320:


+1. The patch looks good to me.

 Make merged protocol changes from 0.20-append to 0.20-security compatible 
 with previous releases.
 -

 Key: HDFS-2320
 URL: https://issues.apache.org/jira/browse/HDFS-2320
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: data-node, hdfs client, name-node
Affects Versions: 0.20.205.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
 Fix For: 0.20.205.0

 Attachments: HDFS-2320.patch, HDFS-2320.patch


 0.20-append changes have been merged to 0.20-security. The merge has changes 
 to version numbers in several protocols. This jira makes the protocol changes 
 compatible with older release, allowing clients running older version to talk 
 to server running 205 version and clients running 205 version talk to older 
 servers running 203, 204.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-1122) client block verification may result in blocks in DataBlockScanner prematurely

2011-09-08 Thread Jitendra Nath Pandey (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100804#comment-13100804
 ] 

Jitendra Nath Pandey commented on HDFS-1122:


Committed to branch-0.20-security. Thanks to Sam.

 client block verification may result in blocks in DataBlockScanner prematurely
 --

 Key: HDFS-1122
 URL: https://issues.apache.org/jira/browse/HDFS-1122
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: 0.20-append, 0.20.205.0
Reporter: sam rash
Assignee: sam rash
 Fix For: 0.20-append, 0.20.205.0

 Attachments: HDFS-1122-20-security.2.patch, hdfs-1122-for-0.20.txt


 found that when the DN uses client verification of a block that is open for 
 writing, it will add it to the DataBlockScanner prematurely. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HDFS-2320) Make merged protocol changes from 0.20-append to 0.20-security compatible with previous releases.

2011-09-08 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2320?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-2320.
---

  Resolution: Fixed
Hadoop Flags: [Reviewed]

I committed this patch. This patch is only relevant to 0.20.205, due to merging 
0.20-append changes and is not relevant to trunk which has a new implementation 
of append.

 Make merged protocol changes from 0.20-append to 0.20-security compatible 
 with previous releases.
 -

 Key: HDFS-2320
 URL: https://issues.apache.org/jira/browse/HDFS-2320
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: data-node, hdfs client, name-node
Affects Versions: 0.20.205.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
 Fix For: 0.20.205.0

 Attachments: HDFS-2320.patch, HDFS-2320.patch


 0.20-append changes have been merged to 0.20-security. The merge has changes 
 to version numbers in several protocols. This jira makes the protocol changes 
 compatible with older release, allowing clients running older version to talk 
 to server running 205 version and clients running 205 version talk to older 
 servers running 203, 204.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2316) webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP

2011-09-08 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100816#comment-13100816
 ] 

Sanjay Radia commented on HDFS-2316:


One issue: we have to co exist with hftp which uses a prefixes like /data. Two 
proposal on the table:
# (by Nicholas) use a different fixed prefix like /webhdfs/path since it does 
not conflict with /data.
# (by Alejandro) always use an operation in the url so that one can send /data 
with an operation to webhdfs and /data without operation to hftp. 

1) is is simpler to implement since urls within the context of /webhdfs can be 
sent to the webhdfs servlet. 2)has a nicer url since the path is the pathname 
being refereced. 

 webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP
 --

 Key: HDFS-2316
 URL: https://issues.apache.org/jira/browse/HDFS-2316
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE

 We current have hftp for accessing HDFS over HTTP.  However, hftp is a 
 read-only FileSystem and does not provide write accesses.
 In HDFS-2284, we propose to have webhdfs for providing a complete FileSystem 
 implementation for accessing HDFS over HTTP.  The is the umbrella JIRA for 
 the tasks.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2323) start-dfs.sh script fails for tarball install

2011-09-08 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2323?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100836#comment-13100836
 ] 

Hadoop QA commented on HDFS-2323:
-

-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12493689/HDFS-2323.patch
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

-1 tests included.  The patch doesn't appear to include any new or modified 
tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

-1 core tests.  The patch failed these unit tests:
  org.apache.hadoop.hdfs.TestDfsOverAvroRpc
  
org.apache.hadoop.hdfs.server.blockmanagement.TestHost2NodesMap

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/1225//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HDFS-Build/1225//artifact/trunk/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/1225//console

This message is automatically generated.

 start-dfs.sh script fails for tarball install
 -

 Key: HDFS-2323
 URL: https://issues.apache.org/jira/browse/HDFS-2323
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Tom White
Assignee: Tom White
 Attachments: HDFS-2323.patch


 I build Common and HDFS tarballs from trunk then tried to start a cluster 
 with start-dfs.sh, but I got the following error:
 {noformat}
 Starting namenodes on [localhost ]
 sbin/start-dfs.sh: line 55: 
 /Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
 directory
 sbin/start-dfs.sh: line 68: 
 /Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
 directory
 Starting secondary namenodes [0.0.0.0 ]
 sbin/start-dfs.sh: line 88: 
 /Users/tom/tmp/hadoop/libexec/../bin/hadoop-daemons.sh: No such file or 
 directory
 {noformat}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2316) webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP

2011-09-08 Thread Tsz Wo (Nicholas), SZE (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100862#comment-13100862
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-2316:
--

2) may be confusion since the same http url path represents two different file 
system paths.  E.g.
- http://namenode:port/data/foo/bar?... means reading /foo/bar
- http://namenode:port/data/foo/bar?opGet=read;... means reading /data/foo/bar


 webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP
 --

 Key: HDFS-2316
 URL: https://issues.apache.org/jira/browse/HDFS-2316
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE

 We current have hftp for accessing HDFS over HTTP.  However, hftp is a 
 read-only FileSystem and does not provide write accesses.
 In HDFS-2284, we propose to have webhdfs for providing a complete FileSystem 
 implementation for accessing HDFS over HTTP.  The is the umbrella JIRA for 
 the tasks.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2316) webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP

2011-09-08 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100891#comment-13100891
 ] 

Alejandro Abdelnur commented on HDFS-2316:
--

IMO, the nice thing about #2 is that the file path of *HDFS:* and a *HTTP:* 
URIs will be exactly the same, and in the case of using the NN/DD deployment of 
HOOP it will be even the same host. 

In addition is it intuitive without any caveat, a given path will just work by 
replacing the SCHEME://HOST:PORT part of it. 

Finally, and IMO this is very important from the Usability perspective, user 
applications that take are designed to take the URI of the FS as parameter and 
operate via HDFS: or HTTP: will be otherwise difficult to code. Hadoop's 
*Path(String parent, String child)* uses the *URI.resolve(...)* that uses a 
well defined logic to resolve URIs based on other URIs[ 
http://download.oracle.com/javase/6/docs/api/java/net/URI.html#resolve(java.net.URI)
 ]. If we use a prefix for HTTP URIs then it will become difficult and error 
prone to compose HDFS: URIs from HTTP: URIs and viceversa. (And I believe the 
same is true for libraries in other languages)

Finally, I have not seen HDFS files under */data* as a common practice, thus 
the name collision won't be that common.


 webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP
 --

 Key: HDFS-2316
 URL: https://issues.apache.org/jira/browse/HDFS-2316
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE

 We current have hftp for accessing HDFS over HTTP.  However, hftp is a 
 read-only FileSystem and does not provide write accesses.
 In HDFS-2284, we propose to have webhdfs for providing a complete FileSystem 
 implementation for accessing HDFS over HTTP.  The is the umbrella JIRA for 
 the tasks.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2316) webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP

2011-09-08 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100917#comment-13100917
 ] 

Todd Lipcon commented on HDFS-2316:
---

Maybe I'm not following completely: is the idea that webhdfs would run on the 
same port as the NN web UI?

It seems crazy to me that http://nn:50030/jmx would be the JMX servlet but 
http://nn:50030/jmx?opGet=read would be a file at path /jmx... hopefully I'm 
misunderstanding :)

 webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP
 --

 Key: HDFS-2316
 URL: https://issues.apache.org/jira/browse/HDFS-2316
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE

 We current have hftp for accessing HDFS over HTTP.  However, hftp is a 
 read-only FileSystem and does not provide write accesses.
 In HDFS-2284, we propose to have webhdfs for providing a complete FileSystem 
 implementation for accessing HDFS over HTTP.  The is the umbrella JIRA for 
 the tasks.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2316) webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP

2011-09-08 Thread Tsz Wo (Nicholas), SZE (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100926#comment-13100926
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-2316:
--

@Todd, we are using the same port.  If we implement (2), you are right that we 
will have such problem.

 webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP
 --

 Key: HDFS-2316
 URL: https://issues.apache.org/jira/browse/HDFS-2316
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE

 We current have hftp for accessing HDFS over HTTP.  However, hftp is a 
 read-only FileSystem and does not provide write accesses.
 In HDFS-2284, we propose to have webhdfs for providing a complete FileSystem 
 implementation for accessing HDFS over HTTP.  The is the umbrella JIRA for 
 the tasks.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2316) webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP

2011-09-08 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100934#comment-13100934
 ] 

Todd Lipcon commented on HDFS-2316:
---

Yea, that seems crazy. I'm for option #1 or for opening yet-another-port.

 webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP
 --

 Key: HDFS-2316
 URL: https://issues.apache.org/jira/browse/HDFS-2316
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE

 We current have hftp for accessing HDFS over HTTP.  However, hftp is a 
 read-only FileSystem and does not provide write accesses.
 In HDFS-2284, we propose to have webhdfs for providing a complete FileSystem 
 implementation for accessing HDFS over HTTP.  The is the umbrella JIRA for 
 the tasks.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2316) webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP

2011-09-08 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100960#comment-13100960
 ] 

Sanjay Radia commented on HDFS-2316:


Operations like to minimize the number of ports. So we would use the same port.

 webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP
 --

 Key: HDFS-2316
 URL: https://issues.apache.org/jira/browse/HDFS-2316
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE

 We current have hftp for accessing HDFS over HTTP.  However, hftp is a 
 read-only FileSystem and does not provide write accesses.
 In HDFS-2284, we propose to have webhdfs for providing a complete FileSystem 
 implementation for accessing HDFS over HTTP.  The is the umbrella JIRA for 
 the tasks.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2316) webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP

2011-09-08 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100965#comment-13100965
 ] 

Sanjay Radia commented on HDFS-2316:


Like hftp, Operations are processed  at the NN if they involve no data 
transfer. If the operation involves data transfer (r or w) then the request is 
redirected to the DN. This allows bandwidth scaling and load distribution.
Alejandro has pointed out that when you redirect a put or post operation, the 
initial part of the payload has been sent to the NN. I believe this is true. 
Hence for writes we could consider a two request mode - getTheWrite handle 
using a get and then do a put or post to the data node.

 webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP
 --

 Key: HDFS-2316
 URL: https://issues.apache.org/jira/browse/HDFS-2316
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE

 We current have hftp for accessing HDFS over HTTP.  However, hftp is a 
 read-only FileSystem and does not provide write accesses.
 In HDFS-2284, we propose to have webhdfs for providing a complete FileSystem 
 implementation for accessing HDFS over HTTP.  The is the umbrella JIRA for 
 the tasks.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2316) webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP

2011-09-08 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100978#comment-13100978
 ] 

Alejandro Abdelnur commented on HDFS-2316:
--

@Todd, well the problem here is that we are overloading the use of a PORT for a 
functionality that requires the whole domain of the namespace (in HFDS we don't 
have special dirs like /dev). I'd be OK with a different port then.

@Sanjay, the in-direction for writes it would mean that we moving away from a 
the REST protocol, which is a well understood and known way of interacting with 
resources. I think there a big value in having a full REST API like Hoop has 
today.


 webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP
 --

 Key: HDFS-2316
 URL: https://issues.apache.org/jira/browse/HDFS-2316
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE

 We current have hftp for accessing HDFS over HTTP.  However, hftp is a 
 read-only FileSystem and does not provide write accesses.
 In HDFS-2284, we propose to have webhdfs for providing a complete FileSystem 
 implementation for accessing HDFS over HTTP.  The is the umbrella JIRA for 
 the tasks.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2316) webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP

2011-09-08 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13100993#comment-13100993
 ] 

Alejandro Abdelnur commented on HDFS-2316:
--

If we use a webhdfs://HOST:PORT from the FS client impl and internally we 
replace to it to http://HOST:PORT/webhdfs, I'd live with that one. After all 
(as Todd pointed offline) it is not fully REST.

But the write/append handle thingy, any other option?


 webhdfs: a complete FileSystem implementation for accessing HDFS over HTTP
 --

 Key: HDFS-2316
 URL: https://issues.apache.org/jira/browse/HDFS-2316
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE

 We current have hftp for accessing HDFS over HTTP.  However, hftp is a 
 read-only FileSystem and does not provide write accesses.
 In HDFS-2284, we propose to have webhdfs for providing a complete FileSystem 
 implementation for accessing HDFS over HTTP.  The is the umbrella JIRA for 
 the tasks.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-2178) Contributing Hoop to HDFS, replacement for HDFS proxy with read/write capabilities

2011-09-08 Thread Tsz Wo (Nicholas), SZE (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13101001#comment-13101001
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-2178:
--

Nicholas, thank you very much for taking the time to go through Hoop's code 
and thanks for your positive feedback on it.

Regarding create  mkdir being POST operation. 

...

Hi Alejandro, I still think that we should use PUT for mkdir and create.  Below 
is quoted from Section 9.6 PUT in 
http://www.w3.org/Protocols/rfc2616/rfc2616-sec9.html
{quote}
The fundamental difference between the POST and PUT requests is reflected in 
the different meaning of the Request-URI. The URI in a POST request identifies 
the resource that will handle the enclosed entity. That resource might be a 
data-accepting process, a gateway to some other protocol, or a separate entity 
that accepts annotations. In contrast, the URI in a PUT request identifies the 
entity enclosed with the request -- the user agent knows what URI is intended 
and the server MUST NOT attempt to apply the request to some other resource. ...
{quote}

 Contributing Hoop to HDFS, replacement for HDFS proxy with read/write 
 capabilities
 --

 Key: HDFS-2178
 URL: https://issues.apache.org/jira/browse/HDFS-2178
 Project: Hadoop HDFS
  Issue Type: Improvement
Affects Versions: 0.23.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Fix For: 0.23.0


 We'd like to contribute Hoop to Hadoop HDFS as a replacement (an improvement) 
 for HDFS Proxy.
 Hoop provides access to all Hadoop Distributed File System (HDFS) operations 
 (read and write) over HTTP/S.
 The Hoop server component is a REST HTTP gateway to HDFS supporting all file 
 system operations. It can be accessed using standard HTTP tools (i.e. curl 
 and wget), HTTP libraries from different programing languages (i.e. Perl, 
 Java Script) as well as using the Hoop client. The Hoop server component is a 
 standard Java web-application and it has been implemented using Jersey 
 (JAX-RS).
 The Hoop client component is an implementation of Hadoop FileSystem client 
 that allows using the familiar Hadoop filesystem API to access HDFS data 
 through a Hoop server.
   Repo: https://github.com/cloudera/hoop
   Docs: http://cloudera.github.com/hoop
   Blog: http://www.cloudera.com/blog/2011/07/hoop-hadoop-hdfs-over-http/
 Hoop is a Maven based project that depends on Hadoop HDFS and Alfredo (for 
 Kerberos HTTP SPNEGO authentication). 
 To make the integration easy, HDFS Mavenization (HDFS-2096) would have to be 
 done first, as well as the Alfredo contribution (HADOOP-7119).

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HDFS-1779) After NameNode restart , Clients can not read partial files even after client invokes Sync.

2011-09-08 Thread Jitendra Nath Pandey (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1779?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13101005#comment-13101005
 ] 

Jitendra Nath Pandey commented on HDFS-1779:


+1 for the patch. This needs to be ported to 20-security branch as well.

 After NameNode restart , Clients can not read partial files even after client 
 invokes Sync.
 ---

 Key: HDFS-1779
 URL: https://issues.apache.org/jira/browse/HDFS-1779
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: data-node, name-node
Affects Versions: 0.20-append
 Environment: Linux
Reporter: Uma Maheswara Rao G
Assignee: Uma Maheswara Rao G
 Fix For: 0.20-append, 0.20.205.0

 Attachments: HDFS-1779.1.patch, HDFS-1779.2.patch, 
 HDFS-1779.2a.patch, HDFS-1779.patch, bbwReportAppend.patch


 In Append HDFS-200 issue,
 If file has 10 blocks and after writing 5 blocks if client invokes sync 
 method then NN will persist the blocks information in edits. 
 After this if we restart the NN, All the DataNodes will reregister with NN. 
 But DataNodes are not sending the blocks being written information to NN. DNs 
 are sending the blocksBeingWritten information in DN startup. So, here 
 NameNode can not find that the 5 persisted blocks belongs to which datanodes. 
 This information can build based on block reports from DN. Otherwise we will 
 loose this 5 blocks information even NN persisted that block information in 
 edits. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira