[jira] [Updated] (HDFS-4581) DataNode#checkDiskError should not be called on network errors

2013-04-18 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J updated HDFS-4581:
--

Fix Version/s: (was: 3.0.0)
   1.3.0

Applying Fix Version 1.3.0 for the branch-1 commit done here; seems to have 
been missed post commit.

> DataNode#checkDiskError should not be called on network errors
> --
>
> Key: HDFS-4581
> URL: https://issues.apache.org/jira/browse/HDFS-4581
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 1.1.1, 3.0.0, 0.23.7, 2.0.4-alpha
>Reporter: Rohit Kochar
>Assignee: Rohit Kochar
> Fix For: 0.23.7, 2.0.4-alpha, 1.3.0
>
> Attachments: HDFS-4581-branch1.patch, HDFS-4581-branch1.patch, 
> HDFS-4581-branch1.patch, HDFS-4581-branch1.patch, HDFS-4581.patch, 
> HDFS-4581.patch, HDFS-4581.patch, HDFS-4581.patch, HDFS-4581.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4581) DataNode#checkDiskError should not be called on network errors

2013-04-18 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J updated HDFS-4581:
--

Target Version/s:   (was: 1.2.0, 3.0.0)

> DataNode#checkDiskError should not be called on network errors
> --
>
> Key: HDFS-4581
> URL: https://issues.apache.org/jira/browse/HDFS-4581
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 1.1.1, 3.0.0, 0.23.7, 2.0.4-alpha
>Reporter: Rohit Kochar
>Assignee: Rohit Kochar
> Fix For: 0.23.7, 2.0.4-alpha, 1.3.0
>
> Attachments: HDFS-4581-branch1.patch, HDFS-4581-branch1.patch, 
> HDFS-4581-branch1.patch, HDFS-4581-branch1.patch, HDFS-4581.patch, 
> HDFS-4581.patch, HDFS-4581.patch, HDFS-4581.patch, HDFS-4581.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4622) Remove redundant synchronized from FSNamesystem#rollEditLog in branch-1

2013-04-18 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4622?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J updated HDFS-4622:
--

Fix Version/s: (was: 1.2.1)
   1.3.0

> Remove redundant synchronized from FSNamesystem#rollEditLog in branch-1
> ---
>
> Key: HDFS-4622
> URL: https://issues.apache.org/jira/browse/HDFS-4622
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 1.2.0
>Reporter: Jing Zhao
>Assignee: Jing Zhao
>Priority: Trivial
> Fix For: 1.3.0
>
> Attachments: HDFS-4622.b1.001.patch
>
>
> HDFS-4222 moves checkSuperuserPrivilege out of the namespace lock scope but 
> missed the rollEditLog method.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4622) Remove redundant synchronized from FSNamesystem#rollEditLog in branch-1

2013-04-18 Thread Harsh J (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4622?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13634978#comment-13634978
 ] 

Harsh J commented on HDFS-4622:
---

Reset the Fix Version to point to the right one. 1.2.1 was not a valid point, 
as 1.2.0 itself is unreleased and this patch didn't go to the 1.2 branch.

If it did, please reset to 1.2.0.

> Remove redundant synchronized from FSNamesystem#rollEditLog in branch-1
> ---
>
> Key: HDFS-4622
> URL: https://issues.apache.org/jira/browse/HDFS-4622
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 1.2.0
>Reporter: Jing Zhao
>Assignee: Jing Zhao
>Priority: Trivial
> Fix For: 1.3.0
>
> Attachments: HDFS-4622.b1.001.patch
>
>
> HDFS-4222 moves checkSuperuserPrivilege out of the namespace lock scope but 
> missed the rollEditLog method.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4710) Turning off HDFS short-circuit checksums unexpectedly slows down Hive

2013-04-18 Thread Gopal V (JIRA)
Gopal V created HDFS-4710:
-

 Summary: Turning off HDFS short-circuit checksums unexpectedly 
slows down Hive
 Key: HDFS-4710
 URL: https://issues.apache.org/jira/browse/HDFS-4710
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 2.0.4-alpha
 Environment: Centos (EC2) + short-circuit reads on
Reporter: Gopal V
Priority: Minor


When short-circuit reads are on, HDFS client slows down when checksums are 
turned off.

With checksums on, the query takes 45.341 seconds and with it turned off, it 
takes 56.345 seconds. This is slower than the speeds observed when 
short-circuiting is turned off.

The issue seems to be that FSDataInputStream.readByte() calls are directly 
transferred to the disk fd when the checksums are turned off.

Even though all the columns are integers, the data being read will be read via 
DataInputStream which does

{code}
public final int readInt() throws IOException {
int ch1 = in.read();
int ch2 = in.read();
int ch3 = in.read();
int ch4 = in.read();
{code}

To confirm, an strace of the Yarn container shows

{code}
26690 read(154, "B", 1) = 1
26690 read(154, "\250", 1)  = 1
26690 read(154, ".", 1) = 1
26690 read(154, "\24", 1)   = 1
{code}

To emulate this without the entirety of Hive code, I have written a simpler 
test app 

https://github.com/t3rmin4t0r/shortcircuit-reader

The jar will read a file in -bs  sized buffers. Running it with 1 byte 
blocks gives similar results to the Hive test run.


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4686) Fix quota computation for rename with snapshots

2013-04-18 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4686?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao updated HDFS-4686:


Attachment: HDFS-4686.001.patch

Initial patch for review. Still need to fix quota computation for (snapshot 
deletion + rename).

> Fix quota computation for rename with snapshots
> ---
>
> Key: HDFS-4686
> URL: https://issues.apache.org/jira/browse/HDFS-4686
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: datanode, namenode
>Reporter: Jing Zhao
>Assignee: Jing Zhao
> Attachments: HDFS-4686.001.patch
>
>
> Currently after a rename operation within/from a snapshottable directory, a 
> reference node is created in both src and dst subtree, pointing to the 
> original renamed inode. With this change the original quota computation may 
> count the quota usage of the renamed subtree multiple times. This jira tries 
> to fix this issue.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4695) TestEditLog leaks open file handles between tests

2013-04-18 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635081#comment-13635081
 ] 

Hudson commented on HDFS-4695:
--

Integrated in Hadoop-Yarn-trunk #187 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/187/])
HDFS-4695. TestEditLog leaks open file handles between tests. Contributed 
by Ivan Mitic. (Revision 1469015)

 Result = FAILURE
suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1469015
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java


> TestEditLog leaks open file handles between tests
> -
>
> Key: HDFS-4695
> URL: https://issues.apache.org/jira/browse/HDFS-4695
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.0.0-alpha
>Reporter: Ivan Mitic
>Assignee: Ivan Mitic
> Fix For: 2.0.5-beta
>
> Attachments: HDFS-4695.2.patch, HDFS-4695.patch
>
>
> The test leaks open file handles causing subsequent test cases to fail on 
> Windows (common cross-platform issue we've seen multiple times so far).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4695) TestEditLog leaks open file handles between tests

2013-04-18 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635137#comment-13635137
 ] 

Hudson commented on HDFS-4695:
--

Integrated in Hadoop-Hdfs-trunk #1376 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1376/])
HDFS-4695. TestEditLog leaks open file handles between tests. Contributed 
by Ivan Mitic. (Revision 1469015)

 Result = FAILURE
suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1469015
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java


> TestEditLog leaks open file handles between tests
> -
>
> Key: HDFS-4695
> URL: https://issues.apache.org/jira/browse/HDFS-4695
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.0.0-alpha
>Reporter: Ivan Mitic
>Assignee: Ivan Mitic
> Fix For: 2.0.5-beta
>
> Attachments: HDFS-4695.2.patch, HDFS-4695.patch
>
>
> The test leaks open file handles causing subsequent test cases to fail on 
> Windows (common cross-platform issue we've seen multiple times so far).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4707) Fix FilterFileSystem and findbugs warning in Snapshot branch

2013-04-18 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635139#comment-13635139
 ] 

Hudson commented on HDFS-4707:
--

Integrated in Hadoop-Hdfs-Snapshots-Branch-build #161 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-Snapshots-Branch-build/161/])
HDFS-4707. Add snapshot methods to FilterFileSystem and fix findbugs 
warnings. (Revision 1469119)

 Result = FAILURE
szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1469119
Files : 
* 
/hadoop/common/branches/HDFS-2802/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FilterFileSystem.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestFilterFileSystem.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/CHANGES.HDFS-2802.txt
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormat.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageSerialization.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/snapshot/Snapshot.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/snapshot/SnapshotFSImageFormat.java


> Fix FilterFileSystem and findbugs warning in Snapshot branch
> 
>
> Key: HDFS-4707
> URL: https://issues.apache.org/jira/browse/HDFS-4707
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: datanode, namenode
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: Snapshot (HDFS-2802)
>
> Attachments: h4707_20130417.patch
>
>
> The snapshot methods are not added to FilterFileSystem.
> Findbugs warnings:
> - SIC Should snapshot.Snapshot$Root be a _static_ inner class?
> - WMI Method FSImageFormat$Saver.saveImage(ByteBuffer, 
> INodeDirectory, DataOutputStream, Snapshot, boolean) makes inefficient use of 
> keySet iterator instead of entrySet iterator
> - BC  Unchecked/unconfirmed cast from 
> FSImageSerialization.writeINodeFile(INodeFile, DataOutput, boolean)
> - BC  Unchecked/unconfirmed cast from 
> snapshot.SnapshotFSImageFormat.loadDirectoryDiffList(INodeDirectory, 
> DataInput, FSImageFormat$Loader)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4706) disallowSnapshot does not work for root

2013-04-18 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4706?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635140#comment-13635140
 ] 

Hudson commented on HDFS-4706:
--

Integrated in Hadoop-Hdfs-Snapshots-Branch-build #161 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-Snapshots-Branch-build/161/])
HDFS-4706. Do not replace root inode for disallowSnapshot. (Revision 
1469122)

 Result = FAILURE
szetszwo : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1469122
Files : 
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/CHANGES.HDFS-2802.txt
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSUtil.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/protocol/SnapshottableDirectoryStatus.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INode.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/snapshot/INodeDirectorySnapshottable.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/snapshot/SnapshotManager.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/snapshot/SnapshotStats.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestNestedSnapshots.java
* 
/hadoop/common/branches/HDFS-2802/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestSnapshotMetrics.java


> disallowSnapshot does not work for root
> ---
>
> Key: HDFS-4706
> URL: https://issues.apache.org/jira/browse/HDFS-4706
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: Snapshot (HDFS-2802)
>
> Attachments: h4706_20130417.patch
>
>
> disallowSnapshot replaces a snapshottable directory to a normal directory.  
> However, we cannot replace root.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4711) Can not change replication factor of file during moving or deleting.

2013-04-18 Thread Vladimir Barinov (JIRA)
Vladimir Barinov created HDFS-4711:
--

 Summary: Can not change replication factor of file during moving 
or deleting.
 Key: HDFS-4711
 URL: https://issues.apache.org/jira/browse/HDFS-4711
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Vladimir Barinov
Priority: Minor


I don't know is it a feature or a bug. 
According to hdfs dfs -help we can use key -D to set specific options for 
action;
When we copying or uploading file to hdfs we can override replication factor 
with -D dfs.replication=N. That's works well.
But it doesn't work for moving or removing(to trash) file.
Step to reproduce:
Uploading file
hdfs dfs -put somefile /tmp/somefile
Copying with changing replication:
hdfs dfs -D dfs.replication=1 -mv /tmp/somefile /tmp/somefile2

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4711) Can not change replication factor of file during moving or deleting.

2013-04-18 Thread Vladimir Barinov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vladimir Barinov updated HDFS-4711:
---

Description: 
I don't know is it a feature or a bug. 
According to hdfs dfs -help we can use key -D to set specific options for 
action;
When we copying or uploading file to hdfs we can override replication factor 
with -D dfs.replication=N. That's works well.
But it doesn't work for moving or removing(to trash) file.
Steps to reproduce:
Uploading file
hdfs dfs -put somefile /tmp/somefile
Copying with changing replication:
hdfs dfs -D dfs.replication=1 -mv /tmp/somefile /tmp/somefile2

  was:
I don't know is it a feature or a bug. 
According to hdfs dfs -help we can use key -D to set specific options for 
action;
When we copying or uploading file to hdfs we can override replication factor 
with -D dfs.replication=N. That's works well.
But it doesn't work for moving or removing(to trash) file.
Step to reproduce:
Uploading file
hdfs dfs -put somefile /tmp/somefile
Copying with changing replication:
hdfs dfs -D dfs.replication=1 -mv /tmp/somefile /tmp/somefile2


> Can not change replication factor of file during moving or deleting.
> 
>
> Key: HDFS-4711
> URL: https://issues.apache.org/jira/browse/HDFS-4711
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Vladimir Barinov
>Priority: Minor
>
> I don't know is it a feature or a bug. 
> According to hdfs dfs -help we can use key -D to set specific options for 
> action;
> When we copying or uploading file to hdfs we can override replication factor 
> with -D dfs.replication=N. That's works well.
> But it doesn't work for moving or removing(to trash) file.
> Steps to reproduce:
> Uploading file
> hdfs dfs -put somefile /tmp/somefile
> Copying with changing replication:
> hdfs dfs -D dfs.replication=1 -mv /tmp/somefile /tmp/somefile2

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4711) Can not change replication factor of file during moving or deleting.

2013-04-18 Thread Vladimir Barinov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vladimir Barinov updated HDFS-4711:
---

Affects Version/s: 2.0.0-alpha

> Can not change replication factor of file during moving or deleting.
> 
>
> Key: HDFS-4711
> URL: https://issues.apache.org/jira/browse/HDFS-4711
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.0-alpha
>Reporter: Vladimir Barinov
>Priority: Minor
>
> I don't know is it a feature or a bug. 
> According to hdfs dfs -help we can use key -D to set specific options for 
> action;
> When we copying or uploading file to hdfs we can override replication factor 
> with -D dfs.replication=N. That's works well.
> But it doesn't work for moving or removing(to trash) file.
> Steps to reproduce:
> Uploading file
> hdfs dfs -put somefile /tmp/somefile
> Copying with changing replication:
> hdfs dfs -D dfs.replication=1 -mv /tmp/somefile /tmp/somefile2

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4711) Can not change replication factor of file during moving or deleting.

2013-04-18 Thread Vladimir Barinov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vladimir Barinov updated HDFS-4711:
---

Description: 
I don't know is it a feature or a bug. 
According to hdfs dfs -help we can use key -D to set specific options for 
action;
When we copying or uploading file to hdfs we can override replication factor 
with -D dfs.replication=N. That's works well.
But it doesn't work for moving or removing(to trash) file.
Steps to reproduce:
Uploading file
hdfs dfs -put somefile /tmp/somefile
Copying with changing replication:
hdfs dfs -D dfs.replication=1 -mv /tmp/somefile /tmp/somefile2

hadoop version:
Hadoop 2.0.0-cdh4.1.2


  was:
I don't know is it a feature or a bug. 
According to hdfs dfs -help we can use key -D to set specific options for 
action;
When we copying or uploading file to hdfs we can override replication factor 
with -D dfs.replication=N. That's works well.
But it doesn't work for moving or removing(to trash) file.
Steps to reproduce:
Uploading file
hdfs dfs -put somefile /tmp/somefile
Copying with changing replication:
hdfs dfs -D dfs.replication=1 -mv /tmp/somefile /tmp/somefile2


> Can not change replication factor of file during moving or deleting.
> 
>
> Key: HDFS-4711
> URL: https://issues.apache.org/jira/browse/HDFS-4711
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.0-alpha
>Reporter: Vladimir Barinov
>Priority: Minor
>
> I don't know is it a feature or a bug. 
> According to hdfs dfs -help we can use key -D to set specific options for 
> action;
> When we copying or uploading file to hdfs we can override replication factor 
> with -D dfs.replication=N. That's works well.
> But it doesn't work for moving or removing(to trash) file.
> Steps to reproduce:
> Uploading file
> hdfs dfs -put somefile /tmp/somefile
> Copying with changing replication:
> hdfs dfs -D dfs.replication=1 -mv /tmp/somefile /tmp/somefile2
> hadoop version:
> Hadoop 2.0.0-cdh4.1.2

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4712) New libhdfs method hdfsGetDataNodes

2013-04-18 Thread andrea manzi (JIRA)
andrea manzi created HDFS-4712:
--

 Summary: New libhdfs method hdfsGetDataNodes
 Key: HDFS-4712
 URL: https://issues.apache.org/jira/browse/HDFS-4712
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: libhdfs
Reporter: andrea manzi


we have implemented a possible extension to libhdfs to retrieve information 
about the available datanodes ( there was a mail on the hadoop-hdsf-dev mailing 
list initially abut this :
http://mail-archives.apache.org/mod_mbox/hadoop-hdfs-dev/201204.mbox/%3CCANhO-
s0mvororrxpjnjbql6brkj4c7l+u816xkdc+2r0whj...@mail.gmail.com%3E)

i would like to know how to proceed to create a patch, cause on the wiki 
http://wiki.apache.org/hadoop/HowToContribute i can see info about JAVA patches 
but nothing related to extensions in C.





--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4695) TestEditLog leaks open file handles between tests

2013-04-18 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635180#comment-13635180
 ] 

Hudson commented on HDFS-4695:
--

Integrated in Hadoop-Mapreduce-trunk #1403 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1403/])
HDFS-4695. TestEditLog leaks open file handles between tests. Contributed 
by Ivan Mitic. (Revision 1469015)

 Result = SUCCESS
suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1469015
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java


> TestEditLog leaks open file handles between tests
> -
>
> Key: HDFS-4695
> URL: https://issues.apache.org/jira/browse/HDFS-4695
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.0.0-alpha
>Reporter: Ivan Mitic
>Assignee: Ivan Mitic
> Fix For: 2.0.5-beta
>
> Attachments: HDFS-4695.2.patch, HDFS-4695.patch
>
>
> The test leaks open file handles causing subsequent test cases to fail on 
> Windows (common cross-platform issue we've seen multiple times so far).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4709) TestDFSClientRetries#testGetFileChecksum

2013-04-18 Thread Tian Hong Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tian Hong Wang updated HDFS-4709:
-

Summary: TestDFSClientRetries#testGetFileChecksum  (was: 
TestDFSClientRetries#testGetFileChecksum fails using IBM java 6)

> TestDFSClientRetries#testGetFileChecksum
> 
>
> Key: HDFS-4709
> URL: https://issues.apache.org/jira/browse/HDFS-4709
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.0.3-alpha
>Reporter: Tian Hong Wang
>Assignee: Tian Hong Wang
>  Labels: patch
> Fix For: 2.0.3-alpha
>
> Attachments: HDFS-4709.patch
>
>
> testGetFileChecksum(org.apache.hadoop.hdfs.TestDFSClientRetries)  Time 
> elapsed: 3993 sec  <<< ERROR!
> org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
> /testGetFileChecksum could only be replicated to 0 nodes instead of 
> minReplication (=1).  There are 3 datanode(s) running and 3 node(s) are 
> excluded in this operation.
> at 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1339)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2186)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:491)
> at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:351)
> at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:40744)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:454)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1014)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1735)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1731)
> at 
> java.security.AccessController.doPrivileged(AccessController.java:310)
> at javax.security.auth.Subject.doAs(Subject.java:573)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1441)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1729)
> at org.apache.hadoop.ipc.Client.call(Client.java:1235)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202)
> at $Proxy10.addBlock(Unknown Source)
> at sun.reflect.GeneratedMethodAccessor23.invoke(Unknown Source)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:37)
> at java.lang.reflect.Method.invoke(Method.java:611)
> at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164)
> at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83)
> at $Proxy10.addBlock(Unknown Source)
> at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:311)
> at 
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1156)
> at 
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1009)
> at 
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:464)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4709) TestDFSClientRetries#testGetFileChecksum fails

2013-04-18 Thread Tian Hong Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tian Hong Wang updated HDFS-4709:
-

Summary: TestDFSClientRetries#testGetFileChecksum fails  (was: 
TestDFSClientRetries#testGetFileChecksum)

> TestDFSClientRetries#testGetFileChecksum fails
> --
>
> Key: HDFS-4709
> URL: https://issues.apache.org/jira/browse/HDFS-4709
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 2.0.3-alpha
>Reporter: Tian Hong Wang
>Assignee: Tian Hong Wang
>  Labels: patch
> Fix For: 2.0.3-alpha
>
> Attachments: HDFS-4709.patch
>
>
> testGetFileChecksum(org.apache.hadoop.hdfs.TestDFSClientRetries)  Time 
> elapsed: 3993 sec  <<< ERROR!
> org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
> /testGetFileChecksum could only be replicated to 0 nodes instead of 
> minReplication (=1).  There are 3 datanode(s) running and 3 node(s) are 
> excluded in this operation.
> at 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1339)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2186)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:491)
> at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:351)
> at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:40744)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:454)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1014)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1735)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1731)
> at 
> java.security.AccessController.doPrivileged(AccessController.java:310)
> at javax.security.auth.Subject.doAs(Subject.java:573)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1441)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1729)
> at org.apache.hadoop.ipc.Client.call(Client.java:1235)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202)
> at $Proxy10.addBlock(Unknown Source)
> at sun.reflect.GeneratedMethodAccessor23.invoke(Unknown Source)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:37)
> at java.lang.reflect.Method.invoke(Method.java:611)
> at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164)
> at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83)
> at $Proxy10.addBlock(Unknown Source)
> at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:311)
> at 
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1156)
> at 
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1009)
> at 
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:464)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4699) TestPipelinesFailover#testPipelineRecoveryStress fails sporadically

2013-04-18 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-4699:


Status: Open  (was: Patch Available)

> TestPipelinesFailover#testPipelineRecoveryStress fails sporadically
> ---
>
> Key: HDFS-4699
> URL: https://issues.apache.org/jira/browse/HDFS-4699
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 3.0.0
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-4699.1.patch
>
>
> I have seen {{TestPipelinesFailover#testPipelineRecoveryStress}} fail 
> sporadically due to timeout during {{loopRecoverLease}}, which waits for up 
> to 30 seconds before timing out.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4699) TestPipelinesFailover#testPipelineRecoveryStress fails sporadically

2013-04-18 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-4699:


 Target Version/s: 3.0.0, 0.23.7, 2.0.4-alpha, 1.3.0  (was: 3.0.0)
Affects Version/s: 1.3.0
   2.0.4-alpha
   0.23.7

> TestPipelinesFailover#testPipelineRecoveryStress fails sporadically
> ---
>
> Key: HDFS-4699
> URL: https://issues.apache.org/jira/browse/HDFS-4699
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 3.0.0, 0.23.7, 2.0.4-alpha, 1.3.0
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-4699.1.patch
>
>
> I have seen {{TestPipelinesFailover#testPipelineRecoveryStress}} fail 
> sporadically due to timeout during {{loopRecoverLease}}, which waits for up 
> to 30 seconds before timing out.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4699) TestPipelinesFailover#testPipelineRecoveryStress fails sporadically

2013-04-18 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-4699:


Attachment: HDFS-4699.branch-1.1.patch
HDFS-4699.branch-0.23.1.patch

> TestPipelinesFailover#testPipelineRecoveryStress fails sporadically
> ---
>
> Key: HDFS-4699
> URL: https://issues.apache.org/jira/browse/HDFS-4699
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 3.0.0, 0.23.7, 2.0.4-alpha, 1.3.0
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-4699.1.patch, HDFS-4699.branch-0.23.1.patch, 
> HDFS-4699.branch-1.1.patch
>
>
> I have seen {{TestPipelinesFailover#testPipelineRecoveryStress}} fail 
> sporadically due to timeout during {{loopRecoverLease}}, which waits for up 
> to 30 seconds before timing out.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4699) TestPipelinesFailover#testPipelineRecoveryStress fails sporadically

2013-04-18 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-4699:


Attachment: (was: HDFS-4699.1.patch)

> TestPipelinesFailover#testPipelineRecoveryStress fails sporadically
> ---
>
> Key: HDFS-4699
> URL: https://issues.apache.org/jira/browse/HDFS-4699
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 3.0.0, 0.23.7, 2.0.4-alpha, 1.3.0
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-4699.1.patch, HDFS-4699.branch-0.23.1.patch, 
> HDFS-4699.branch-1.1.patch
>
>
> I have seen {{TestPipelinesFailover#testPipelineRecoveryStress}} fail 
> sporadically due to timeout during {{loopRecoverLease}}, which waits for up 
> to 30 seconds before timing out.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4699) TestPipelinesFailover#testPipelineRecoveryStress fails sporadically

2013-04-18 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-4699:


Attachment: HDFS-4699.1.patch

> TestPipelinesFailover#testPipelineRecoveryStress fails sporadically
> ---
>
> Key: HDFS-4699
> URL: https://issues.apache.org/jira/browse/HDFS-4699
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 3.0.0, 0.23.7, 2.0.4-alpha, 1.3.0
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-4699.1.patch, HDFS-4699.branch-0.23.1.patch, 
> HDFS-4699.branch-1.1.patch
>
>
> I have seen {{TestPipelinesFailover#testPipelineRecoveryStress}} fail 
> sporadically due to timeout during {{loopRecoverLease}}, which waits for up 
> to 30 seconds before timing out.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4699) TestPipelinesFailover#testPipelineRecoveryStress fails sporadically

2013-04-18 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-4699:


Status: Patch Available  (was: Open)

> TestPipelinesFailover#testPipelineRecoveryStress fails sporadically
> ---
>
> Key: HDFS-4699
> URL: https://issues.apache.org/jira/browse/HDFS-4699
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 3.0.0, 0.23.7, 2.0.4-alpha, 1.3.0
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-4699.1.patch, HDFS-4699.branch-0.23.1.patch, 
> HDFS-4699.branch-1.1.patch
>
>
> I have seen {{TestPipelinesFailover#testPipelineRecoveryStress}} fail 
> sporadically due to timeout during {{loopRecoverLease}}, which waits for up 
> to 30 seconds before timing out.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4699) TestPipelinesFailover#testPipelineRecoveryStress fails sporadically

2013-04-18 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635290#comment-13635290
 ] 

Chris Nauroth commented on HDFS-4699:
-

I noticed that HDFS-4581 also put the network error filtering logic into 
branch-1 and branch-0.23, so I've added patches for those branches too.  I 
uploaded the trunk patch again, just so that Jenkins sees it as the newest file.

> TestPipelinesFailover#testPipelineRecoveryStress fails sporadically
> ---
>
> Key: HDFS-4699
> URL: https://issues.apache.org/jira/browse/HDFS-4699
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 3.0.0, 0.23.7, 2.0.4-alpha, 1.3.0
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-4699.1.patch, HDFS-4699.branch-0.23.1.patch, 
> HDFS-4699.branch-1.1.patch
>
>
> I have seen {{TestPipelinesFailover#testPipelineRecoveryStress}} fail 
> sporadically due to timeout during {{loopRecoverLease}}, which waits for up 
> to 30 seconds before timing out.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4581) DataNode#checkDiskError should not be called on network errors

2013-04-18 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4581?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635292#comment-13635292
 ] 

Chris Nauroth commented on HDFS-4581:
-

In HDFS-4699, I have added more cases to the network error filtering logic, 
based on reviewing logs from tests of rapid NN failovers.

> DataNode#checkDiskError should not be called on network errors
> --
>
> Key: HDFS-4581
> URL: https://issues.apache.org/jira/browse/HDFS-4581
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 1.1.1, 3.0.0, 0.23.7, 2.0.4-alpha
>Reporter: Rohit Kochar
>Assignee: Rohit Kochar
> Fix For: 0.23.7, 2.0.4-alpha, 1.3.0
>
> Attachments: HDFS-4581-branch1.patch, HDFS-4581-branch1.patch, 
> HDFS-4581-branch1.patch, HDFS-4581-branch1.patch, HDFS-4581.patch, 
> HDFS-4581.patch, HDFS-4581.patch, HDFS-4581.patch, HDFS-4581.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4713) Wrong server principal is used for rpc calls to namenode if HA is enabled

2013-04-18 Thread Kihwal Lee (JIRA)
Kihwal Lee created HDFS-4713:


 Summary: Wrong server principal is used for rpc calls to namenode 
if HA is enabled
 Key: HDFS-4713
 URL: https://issues.apache.org/jira/browse/HDFS-4713
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: ha, namenode
Affects Versions: 2.0.4-alpha
Reporter: Kihwal Lee
Priority: Blocker


When various components are connecting to a namenode in a HA-enabled 
environment, a wrong server principal may be picked up.  This result in SASL 
failure, since the client-side used a wrong service ticket for the connection.


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4713) Wrong server principal is used for rpc calls to namenode if HA is enabled

2013-04-18 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4713?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635435#comment-13635435
 ] 

Kihwal Lee commented on HDFS-4713:
--

For standby bootstrapping, HDFS-3284 made it work but HDFS-3438 broke it again. 
In general the proxy object for a protocol should be created with the config 
that has the correct server principle set.  For example, standby namenode talks 
to active namenode via NamenodeProtocol, where the server principal key is 
defined as DFS_NAMENODE_USER_NAME_KEY, "dfs.namenode.kerberos.principal".  The 
standby bootstrapping and checkpointing fail because the namenode proxy object 
has a conf with DFS_NAMENODE_USER_NAME_KEY set to itself. The RPC address is 
correctly set, but the wrong server principle is used.

When I modified the code to create the proxy with "other NN config", everything 
worked. 

I haven't checked thoroughly, but ConfiguredFailoverProxyProvider may have a 
similar issue.

> Wrong server principal is used for rpc calls to namenode if HA is enabled
> -
>
> Key: HDFS-4713
> URL: https://issues.apache.org/jira/browse/HDFS-4713
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha, namenode
>Affects Versions: 2.0.4-alpha
>Reporter: Kihwal Lee
>Priority: Blocker
>
> When various components are connecting to a namenode in a HA-enabled 
> environment, a wrong server principal may be picked up.  This result in SASL 
> failure, since the client-side used a wrong service ticket for the connection.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4713) Wrong server principal is used for rpc calls to namenode if HA is enabled

2013-04-18 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4713?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee updated HDFS-4713:
-

Affects Version/s: (was: 2.0.4-alpha)
   2.0.5-beta

> Wrong server principal is used for rpc calls to namenode if HA is enabled
> -
>
> Key: HDFS-4713
> URL: https://issues.apache.org/jira/browse/HDFS-4713
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha, namenode
>Affects Versions: 2.0.5-beta
>Reporter: Kihwal Lee
>Priority: Blocker
>
> When various components are connecting to a namenode in a HA-enabled 
> environment, a wrong server principal may be picked up.  This result in SASL 
> failure, since the client-side used a wrong service ticket for the connection.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4699) TestPipelinesFailover#testPipelineRecoveryStress fails sporadically

2013-04-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635450#comment-13635450
 ] 

Hadoop QA commented on HDFS-4699:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12579356/HDFS-4699.1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/4274//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4274//console

This message is automatically generated.

> TestPipelinesFailover#testPipelineRecoveryStress fails sporadically
> ---
>
> Key: HDFS-4699
> URL: https://issues.apache.org/jira/browse/HDFS-4699
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Affects Versions: 3.0.0, 0.23.7, 2.0.4-alpha, 1.3.0
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-4699.1.patch, HDFS-4699.branch-0.23.1.patch, 
> HDFS-4699.branch-1.1.patch
>
>
> I have seen {{TestPipelinesFailover#testPipelineRecoveryStress}} fail 
> sporadically due to timeout during {{loopRecoverLease}}, which waits for up 
> to 30 seconds before timing out.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4710) Turning off HDFS short-circuit checksums unexpectedly slows down Hive

2013-04-18 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4710?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635476#comment-13635476
 ] 

Colin Patrick McCabe commented on HDFS-4710:


Hi Gopal,

Good find.

I suppose the fix here is to change {{BlockReaderLocal}} and 
{{BlockReaderLocalLegacy}} to honor 
{{dfs.client.read.shortcircuit.buffer.size}} when 
{{dfs.client.read.shortcircuit.skip.checksum}} is turned on.

> Turning off HDFS short-circuit checksums unexpectedly slows down Hive
> -
>
> Key: HDFS-4710
> URL: https://issues.apache.org/jira/browse/HDFS-4710
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.0.4-alpha
> Environment: Centos (EC2) + short-circuit reads on
>Reporter: Gopal V
>Priority: Minor
>  Labels: perfomance
>
> When short-circuit reads are on, HDFS client slows down when checksums are 
> turned off.
> With checksums on, the query takes 45.341 seconds and with it turned off, it 
> takes 56.345 seconds. This is slower than the speeds observed when 
> short-circuiting is turned off.
> The issue seems to be that FSDataInputStream.readByte() calls are directly 
> transferred to the disk fd when the checksums are turned off.
> Even though all the columns are integers, the data being read will be read 
> via DataInputStream which does
> {code}
> public final int readInt() throws IOException {
> int ch1 = in.read();
> int ch2 = in.read();
> int ch3 = in.read();
> int ch4 = in.read();
> {code}
> To confirm, an strace of the Yarn container shows
> {code}
> 26690 read(154, "B", 1) = 1
> 26690 read(154, "\250", 1)  = 1
> 26690 read(154, ".", 1) = 1
> 26690 read(154, "\24", 1)   = 1
> {code}
> To emulate this without the entirety of Hive code, I have written a simpler 
> test app 
> https://github.com/t3rmin4t0r/shortcircuit-reader
> The jar will read a file in -bs  sized buffers. Running it with 1 byte 
> blocks gives similar results to the Hive test run.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4714) Support logging short messages in Namenode IPC server for configurable list of exception classes.

2013-04-18 Thread Kihwal Lee (JIRA)
Kihwal Lee created HDFS-4714:


 Summary: Support logging short messages in Namenode IPC server for 
configurable list of exception classes.
 Key: HDFS-4714
 URL: https://issues.apache.org/jira/browse/HDFS-4714
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 3.0.0, 0.23.7, 2.0.5-beta
Reporter: Kihwal Lee


Namenode can slow down significantly if a rogue client/job issues massive 
number of requests that will fail. E.g. permission denied, quota overage, etc.  
The major contributing factor in slow down is the long namenode log message, 
which includes full stack trace.  

Previously similar issues involving safe mode and standby node have been 
addressed and we can extend it to suppress logging stack traces for configured 
list of exception classes.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4710) Turning off HDFS short-circuit checksums unexpectedly slows down Hive

2013-04-18 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4710?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635501#comment-13635501
 ] 

Gopal V commented on HDFS-4710:
---

It is not really a one-size-fits-all fix, I guess.

For large reads, a direct read would be the right option to avoid having 
another copy.

But for small repeated reads, it does make sense to bounce buffer.size chunks 
through a buffer to avoid syscall overhead.

> Turning off HDFS short-circuit checksums unexpectedly slows down Hive
> -
>
> Key: HDFS-4710
> URL: https://issues.apache.org/jira/browse/HDFS-4710
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.0.4-alpha
> Environment: Centos (EC2) + short-circuit reads on
>Reporter: Gopal V
>Priority: Minor
>  Labels: perfomance
>
> When short-circuit reads are on, HDFS client slows down when checksums are 
> turned off.
> With checksums on, the query takes 45.341 seconds and with it turned off, it 
> takes 56.345 seconds. This is slower than the speeds observed when 
> short-circuiting is turned off.
> The issue seems to be that FSDataInputStream.readByte() calls are directly 
> transferred to the disk fd when the checksums are turned off.
> Even though all the columns are integers, the data being read will be read 
> via DataInputStream which does
> {code}
> public final int readInt() throws IOException {
> int ch1 = in.read();
> int ch2 = in.read();
> int ch3 = in.read();
> int ch4 = in.read();
> {code}
> To confirm, an strace of the Yarn container shows
> {code}
> 26690 read(154, "B", 1) = 1
> 26690 read(154, "\250", 1)  = 1
> 26690 read(154, ".", 1) = 1
> 26690 read(154, "\24", 1)   = 1
> {code}
> To emulate this without the entirety of Hive code, I have written a simpler 
> test app 
> https://github.com/t3rmin4t0r/shortcircuit-reader
> The jar will read a file in -bs  sized buffers. Running it with 1 byte 
> blocks gives similar results to the Hive test run.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4712) New libhdfs method hdfsGetDataNodes

2013-04-18 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4712?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635506#comment-13635506
 ] 

Colin Patrick McCabe commented on HDFS-4712:


Hi Andrea,

The procedure for creating C patches is much the same as that for Java.  Post 
your patch here in the form of a diff, get Jenkins results and feedback, and 
then get a +1 from a committer.

It sounds like what you want is some way to expose 
{{NameNodeRpcServer#getDatanodeReport}}, which ultimately calls 
{{FSNamesystem#datanodeReport}}.

If you return dynamically allocated memory, please be sure to also offer a 
function in libhdfs to free that memory.  You should not assume that the 
application and the library use the same memory allocator.

Also, as M. C. Srivas commented, please be aware that the list of DataNodes can 
change at any point.  It would help if you would describe what you're trying to 
do.

> New libhdfs method hdfsGetDataNodes
> ---
>
> Key: HDFS-4712
> URL: https://issues.apache.org/jira/browse/HDFS-4712
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: libhdfs
>Reporter: andrea manzi
>
> we have implemented a possible extension to libhdfs to retrieve information 
> about the available datanodes ( there was a mail on the hadoop-hdsf-dev 
> mailing list initially abut this :
> http://mail-archives.apache.org/mod_mbox/hadoop-hdfs-dev/201204.mbox/%3CCANhO-
> s0mvororrxpjnjbql6brkj4c7l+u816xkdc+2r0whj...@mail.gmail.com%3E)
> i would like to know how to proceed to create a patch, cause on the wiki 
> http://wiki.apache.org/hadoop/HowToContribute i can see info about JAVA 
> patches but nothing related to extensions in C.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4711) Can not change replication factor of file during moving or deleting.

2013-04-18 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635513#comment-13635513
 ] 

Colin Patrick McCabe commented on HDFS-4711:


What's wrong with using {{hadoop fs -setrep}}?

The move command is not supposed to alter replication status...

> Can not change replication factor of file during moving or deleting.
> 
>
> Key: HDFS-4711
> URL: https://issues.apache.org/jira/browse/HDFS-4711
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.0-alpha
>Reporter: Vladimir Barinov
>Priority: Minor
>
> I don't know is it a feature or a bug. 
> According to hdfs dfs -help we can use key -D to set specific options for 
> action;
> When we copying or uploading file to hdfs we can override replication factor 
> with -D dfs.replication=N. That's works well.
> But it doesn't work for moving or removing(to trash) file.
> Steps to reproduce:
> Uploading file
> hdfs dfs -put somefile /tmp/somefile
> Copying with changing replication:
> hdfs dfs -D dfs.replication=1 -mv /tmp/somefile /tmp/somefile2
> hadoop version:
> Hadoop 2.0.0-cdh4.1.2

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4710) Turning off HDFS short-circuit checksums unexpectedly slows down Hive

2013-04-18 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4710?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635514#comment-13635514
 ] 

Colin Patrick McCabe commented on HDFS-4710:


If we honor {{dfs.client.read.shortcircuit.buffer.size}}, the client can decide 
whether to use a bounce buffer or not, and if so, how big.

> Turning off HDFS short-circuit checksums unexpectedly slows down Hive
> -
>
> Key: HDFS-4710
> URL: https://issues.apache.org/jira/browse/HDFS-4710
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs-client
>Affects Versions: 2.0.4-alpha
> Environment: Centos (EC2) + short-circuit reads on
>Reporter: Gopal V
>Priority: Minor
>  Labels: perfomance
>
> When short-circuit reads are on, HDFS client slows down when checksums are 
> turned off.
> With checksums on, the query takes 45.341 seconds and with it turned off, it 
> takes 56.345 seconds. This is slower than the speeds observed when 
> short-circuiting is turned off.
> The issue seems to be that FSDataInputStream.readByte() calls are directly 
> transferred to the disk fd when the checksums are turned off.
> Even though all the columns are integers, the data being read will be read 
> via DataInputStream which does
> {code}
> public final int readInt() throws IOException {
> int ch1 = in.read();
> int ch2 = in.read();
> int ch3 = in.read();
> int ch4 = in.read();
> {code}
> To confirm, an strace of the Yarn container shows
> {code}
> 26690 read(154, "B", 1) = 1
> 26690 read(154, "\250", 1)  = 1
> 26690 read(154, ".", 1) = 1
> 26690 read(154, "\24", 1)   = 1
> {code}
> To emulate this without the entirety of Hive code, I have written a simpler 
> test app 
> https://github.com/t3rmin4t0r/shortcircuit-reader
> The jar will read a file in -bs  sized buffers. Running it with 1 byte 
> blocks gives similar results to the Hive test run.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4711) Can not change replication factor of file during moving or deleting.

2013-04-18 Thread Vladimir Barinov (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635530#comment-13635530
 ] 

Vladimir Barinov commented on HDFS-4711:


-setrep works well but it was confusing when move does not supported it and 
usage tells nothing about it. Probably I should change ticket to "new feature" 
or "improvement"?

> Can not change replication factor of file during moving or deleting.
> 
>
> Key: HDFS-4711
> URL: https://issues.apache.org/jira/browse/HDFS-4711
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.0-alpha
>Reporter: Vladimir Barinov
>Priority: Minor
>
> I don't know is it a feature or a bug. 
> According to hdfs dfs -help we can use key -D to set specific options for 
> action;
> When we copying or uploading file to hdfs we can override replication factor 
> with -D dfs.replication=N. That's works well.
> But it doesn't work for moving or removing(to trash) file.
> Steps to reproduce:
> Uploading file
> hdfs dfs -put somefile /tmp/somefile
> Copying with changing replication:
> hdfs dfs -D dfs.replication=1 -mv /tmp/somefile /tmp/somefile2
> hadoop version:
> Hadoop 2.0.0-cdh4.1.2

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4549) WebHDFS hits a Jetty performance issue

2013-04-18 Thread Mark Wagner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Wagner updated HDFS-4549:
--

Attachment: HDFS-4549.2.patch

I've backported HDFS-3577, HDFS-3318, and HDFS-3788 as best as possible. 
However, I'm seeing performance problems above 2GB. It looks like when the 
MessageBodyWriter reports a length greater than 2GB, Jersey is reverting to 
chunked transfer, causing the previously seen performance problems with Jetty.

I tested this on a 2.0.3 deployment and got the same results (chunked transfer 
above a certain size).

[~szetszwo], can you expand on the motivation for using MessageBodyWriter and 
OpenEntity instead of manually setting the header values? Would it be 
problematic to make all branches use the method of my first patch? It seems 
that Jersey is taking the MessageBodyWriter information only as a suggestion.

> WebHDFS hits a Jetty performance issue
> --
>
> Key: HDFS-4549
> URL: https://issues.apache.org/jira/browse/HDFS-4549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 1.1.2
>Reporter: Mark Wagner
>Assignee: Mark Wagner
> Attachments: example4549.txt, HDFS-4549.1.patch, HDFS-4549.2.patch
>
>
> WebHDFS on branch-1 is hitting a Jetty issue for me when it does chunked 
> transfers. This is the same Jetty issue as MAPREDUCE-4399. I have not 
> observed this on trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4549) WebHDFS hits a Jetty performance issue

2013-04-18 Thread Mark Wagner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Wagner updated HDFS-4549:
--

Attachment: example4549.txt

> WebHDFS hits a Jetty performance issue
> --
>
> Key: HDFS-4549
> URL: https://issues.apache.org/jira/browse/HDFS-4549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 1.1.2
>Reporter: Mark Wagner
>Assignee: Mark Wagner
> Attachments: example4549.txt, HDFS-4549.1.patch, HDFS-4549.2.patch
>
>
> WebHDFS on branch-1 is hitting a Jetty issue for me when it does chunked 
> transfers. This is the same Jetty issue as MAPREDUCE-4399. I have not 
> observed this on trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Assigned] (HDFS-4711) Can not change replication factor of file during moving or deleting.

2013-04-18 Thread Colin Patrick McCabe (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe reassigned HDFS-4711:
--

Assignee: Colin Patrick McCabe

> Can not change replication factor of file during moving or deleting.
> 
>
> Key: HDFS-4711
> URL: https://issues.apache.org/jira/browse/HDFS-4711
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.0-alpha
>Reporter: Vladimir Barinov
>Assignee: Colin Patrick McCabe
>Priority: Minor
>
> I don't know is it a feature or a bug. 
> According to hdfs dfs -help we can use key -D to set specific options for 
> action;
> When we copying or uploading file to hdfs we can override replication factor 
> with -D dfs.replication=N. That's works well.
> But it doesn't work for moving or removing(to trash) file.
> Steps to reproduce:
> Uploading file
> hdfs dfs -put somefile /tmp/somefile
> Copying with changing replication:
> hdfs dfs -D dfs.replication=1 -mv /tmp/somefile /tmp/somefile2
> hadoop version:
> Hadoop 2.0.0-cdh4.1.2

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4549) WebHDFS hits a Jetty performance issue

2013-04-18 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635618#comment-13635618
 ] 

Alejandro Abdelnur commented on HDFS-4549:
--

I think we should leave Jersey to decide what to use, chunk encoding or not. If 
we always force the content-length, we can run into issues in clients (I 
believe Java HTTP does -or used to do this-) that try to allocate a buffer in 
memory for the full content-length. By using chunk encoding, you are forcing 
the client to fallback on partial-caching/flushing.

Is the Jetty version use by Hadoop 1 that has issues with chunk encoding?

Also, if we don't see the problem in trunk/branch-2, why change it there? not 
broken, don't fix it. 


> WebHDFS hits a Jetty performance issue
> --
>
> Key: HDFS-4549
> URL: https://issues.apache.org/jira/browse/HDFS-4549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 1.1.2
>Reporter: Mark Wagner
>Assignee: Mark Wagner
> Attachments: example4549.txt, HDFS-4549.1.patch, HDFS-4549.2.patch
>
>
> WebHDFS on branch-1 is hitting a Jetty issue for me when it does chunked 
> transfers. This is the same Jetty issue as MAPREDUCE-4399. I have not 
> observed this on trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4711) Can not change replication factor of file during moving or deleting.

2013-04-18 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635619#comment-13635619
 ] 

Colin Patrick McCabe commented on HDFS-4711:


I don't think this is a good idea.  Consider moving a large directory with 
hundreds of thousands of files.  Are you going to check each file to see if its 
replication is equal to the current dfs.replication setting?  And if it's not, 
are you going to do setrep on it?  This will also create a lot of problems for 
current users, since they'll find that mv unexpectedly changes their 
replication settings.  Remember that dfs.replication is usually set to a value 
due to the configuration XML files, even if the user didn't set it explicitly.

> Can not change replication factor of file during moving or deleting.
> 
>
> Key: HDFS-4711
> URL: https://issues.apache.org/jira/browse/HDFS-4711
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.0-alpha
>Reporter: Vladimir Barinov
>Assignee: Colin Patrick McCabe
>Priority: Minor
>
> I don't know is it a feature or a bug. 
> According to hdfs dfs -help we can use key -D to set specific options for 
> action;
> When we copying or uploading file to hdfs we can override replication factor 
> with -D dfs.replication=N. That's works well.
> But it doesn't work for moving or removing(to trash) file.
> Steps to reproduce:
> Uploading file
> hdfs dfs -put somefile /tmp/somefile
> Copying with changing replication:
> hdfs dfs -D dfs.replication=1 -mv /tmp/somefile /tmp/somefile2
> hadoop version:
> Hadoop 2.0.0-cdh4.1.2

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4549) WebHDFS hits a Jetty performance issue

2013-04-18 Thread Tsz Wo (Nicholas), SZE (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635637#comment-13635637
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-4549:
--

@Mark, the motivation for using MessageBodyWriter is that manually setting the 
Content-Length header may cause problems such as Jersey may add the another 
Content-Length header.  It seems that MessageBodyWriter is a standard way in 
Jersey.

BTW, do you see any performance problem in trunk/2.0.3?

> WebHDFS hits a Jetty performance issue
> --
>
> Key: HDFS-4549
> URL: https://issues.apache.org/jira/browse/HDFS-4549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 1.1.2
>Reporter: Mark Wagner
>Assignee: Mark Wagner
> Attachments: example4549.txt, HDFS-4549.1.patch, HDFS-4549.2.patch
>
>
> WebHDFS on branch-1 is hitting a Jetty issue for me when it does chunked 
> transfers. This is the same Jetty issue as MAPREDUCE-4399. I have not 
> observed this on trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4711) Can not change replication factor of file during moving or deleting.

2013-04-18 Thread Vladimir Barinov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vladimir Barinov resolved HDFS-4711.


Resolution: Won't Fix

> Can not change replication factor of file during moving or deleting.
> 
>
> Key: HDFS-4711
> URL: https://issues.apache.org/jira/browse/HDFS-4711
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.0.0-alpha
>Reporter: Vladimir Barinov
>Assignee: Colin Patrick McCabe
>Priority: Minor
>
> I don't know is it a feature or a bug. 
> According to hdfs dfs -help we can use key -D to set specific options for 
> action;
> When we copying or uploading file to hdfs we can override replication factor 
> with -D dfs.replication=N. That's works well.
> But it doesn't work for moving or removing(to trash) file.
> Steps to reproduce:
> Uploading file
> hdfs dfs -put somefile /tmp/somefile
> Copying with changing replication:
> hdfs dfs -D dfs.replication=1 -mv /tmp/somefile /tmp/somefile2
> hadoop version:
> Hadoop 2.0.0-cdh4.1.2

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4504) DFSOutputStream#close doesn't always release resources (such as leases)

2013-04-18 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635672#comment-13635672
 ] 

Colin Patrick McCabe commented on HDFS-4504:


bq. RPC retry should be done in RPC level. Please don't add retry to DFSClient.

OK.  I will remove this in the next patch.

I suppose lease recovery will take care of un-completed files eventually.

> DFSOutputStream#close doesn't always release resources (such as leases)
> ---
>
> Key: HDFS-4504
> URL: https://issues.apache.org/jira/browse/HDFS-4504
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
> Attachments: HDFS-4504.001.patch
>
>
> {{DFSOutputStream#close}} can throw an {{IOException}} in some cases.  One 
> example is if there is a pipeline error and then pipeline recovery fails.  
> Unfortunately, in this case, some of the resources used by the 
> {{DFSOutputStream}} are leaked.  One particularly important resource is file 
> leases.
> So it's possible for a long-lived HDFS client, such as Flume, to write many 
> blocks to a file, but then fail to close it.  Unfortunately, the 
> {{LeaseRenewerThread}} inside the client will continue to renew the lease for 
> the "undead" file.  Future attempts to close the file will just rethrow the 
> previous exception, and no progress can be made by the client.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4549) WebHDFS hits a Jetty performance issue

2013-04-18 Thread Mark Wagner (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635687#comment-13635687
 ] 

Mark Wagner commented on HDFS-4549:
---

bq. I think we should leave Jersey to decide what to use, chunk encoding or not.

I agree that that would usually be best, however this is causing significant 
performance problems, so I think it needs to be handled somehow. I just took a 
look at the memory usage when pulling a 1GB file (which does use 
content-length) and didn't see any memory usage increase. I'll look into this 
further though.

{quote}
Is the Jetty version use by Hadoop 1 that has issues with chunk encoding?
Also, if we don't see the problem in trunk/branch-2, why change it there? not 
broken, don't fix it.
{quote}

The issue has been observed in 6.1.26, which is used by branch-1, branch-2, and 
trunk. This problem was first seen (chunked encoding being used, leading to 
performance loss) in branch-1 (1.0.4 to be precise) for files of any size. I 
didn't see it on branch-2/trunk because I didn't test larger files. When I was 
backporting though, I noticed that it was still slow above 2GB, and it turns 
out the same is true for branch-2/trunk (I've attached logs for a demonstration 
of this issue using 2.0.3-alpha).

> WebHDFS hits a Jetty performance issue
> --
>
> Key: HDFS-4549
> URL: https://issues.apache.org/jira/browse/HDFS-4549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 1.1.2
>Reporter: Mark Wagner
>Assignee: Mark Wagner
> Attachments: example4549.txt, HDFS-4549.1.patch, HDFS-4549.2.patch
>
>
> WebHDFS on branch-1 is hitting a Jetty issue for me when it does chunked 
> transfers. This is the same Jetty issue as MAPREDUCE-4399. I have not 
> observed this on trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4549) WebHDFS hits a Jetty performance issue

2013-04-18 Thread Mark Wagner (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635695#comment-13635695
 ] 

Mark Wagner commented on HDFS-4549:
---

[~szetszwo], that makes sense. Yes, I see performance issues above 2GB on 2.0.3.

> WebHDFS hits a Jetty performance issue
> --
>
> Key: HDFS-4549
> URL: https://issues.apache.org/jira/browse/HDFS-4549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 1.1.2
>Reporter: Mark Wagner
>Assignee: Mark Wagner
> Attachments: example4549.txt, HDFS-4549.1.patch, HDFS-4549.2.patch
>
>
> WebHDFS on branch-1 is hitting a Jetty issue for me when it does chunked 
> transfers. This is the same Jetty issue as MAPREDUCE-4399. I have not 
> observed this on trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4549) WebHDFS hits a Jetty performance issue

2013-04-18 Thread Tsz Wo (Nicholas), SZE (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635710#comment-13635710
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-4549:
--

@Mark, Since there are also performance problem in trunk, let's first backport 
the patch to branch-1 and then fix all the branch the same way.  If manually 
setting content-length works well, we may set it when the file size is >= 2GB.

Let me create another JIRA for backporting and continue the performance 
improvement here.

> WebHDFS hits a Jetty performance issue
> --
>
> Key: HDFS-4549
> URL: https://issues.apache.org/jira/browse/HDFS-4549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 1.1.2
>Reporter: Mark Wagner
>Assignee: Mark Wagner
> Attachments: example4549.txt, HDFS-4549.1.patch, HDFS-4549.2.patch
>
>
> WebHDFS on branch-1 is hitting a Jetty issue for me when it does chunked 
> transfers. This is the same Jetty issue as MAPREDUCE-4399. I have not 
> observed this on trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4715) Backport HDFS-3577 and other related WebHDFS to branch-1

2013-04-18 Thread Tsz Wo (Nicholas), SZE (JIRA)
Tsz Wo (Nicholas), SZE created HDFS-4715:


 Summary: Backport HDFS-3577 and other related WebHDFS to branch-1
 Key: HDFS-4715
 URL: https://issues.apache.org/jira/browse/HDFS-4715
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: webhdfs
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Mark Wagner


The related JIRAs are HDFS-3577, HDFS-3318, and HDFS-3788.  Backporting them 
can fix some WebHDFS performance issues in branch-1.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4715) Backport HDFS-3577 and other related WebHDFS issue to branch-1

2013-04-18 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4715?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-4715:
-

Summary: Backport HDFS-3577 and other related WebHDFS issue to branch-1  
(was: Backport HDFS-3577 and other related WebHDFS to branch-1)

> Backport HDFS-3577 and other related WebHDFS issue to branch-1
> --
>
> Key: HDFS-4715
> URL: https://issues.apache.org/jira/browse/HDFS-4715
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Mark Wagner
>
> The related JIRAs are HDFS-3577, HDFS-3318, and HDFS-3788.  Backporting them 
> can fix some WebHDFS performance issues in branch-1.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4549) WebHDFS hits a Jetty performance issue

2013-04-18 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635725#comment-13635725
 ] 

Alejandro Abdelnur commented on HDFS-4549:
--

Before changing to remove chunk encoding, can we try the following?

http://stackoverflow.com/questions/9031311/slow-transfers-in-jetty-with-chunked-transfer-encoding-at-certain-buffer-size


> WebHDFS hits a Jetty performance issue
> --
>
> Key: HDFS-4549
> URL: https://issues.apache.org/jira/browse/HDFS-4549
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 1.1.2
>Reporter: Mark Wagner
>Assignee: Mark Wagner
> Attachments: example4549.txt, HDFS-4549.1.patch, HDFS-4549.2.patch
>
>
> WebHDFS on branch-1 is hitting a Jetty issue for me when it does chunked 
> transfers. This is the same Jetty issue as MAPREDUCE-4399. I have not 
> observed this on trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4504) DFSOutputStream#close doesn't always release resources (such as leases)

2013-04-18 Thread Colin Patrick McCabe (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4504?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-4504:
---

Attachment: HDFS-4504.002.patch

* remove manual retry (allow RPC retry to do this job)

* remove debug messages

* TestBlockRecovery: set lower RPC timeout so that we give up on close prior to 
the test timeout.

* If both {{completeFile}} and the block data flush fail, throw a 
{{MultipleIOException}} with information about both failures.

> DFSOutputStream#close doesn't always release resources (such as leases)
> ---
>
> Key: HDFS-4504
> URL: https://issues.apache.org/jira/browse/HDFS-4504
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
> Attachments: HDFS-4504.001.patch, HDFS-4504.002.patch
>
>
> {{DFSOutputStream#close}} can throw an {{IOException}} in some cases.  One 
> example is if there is a pipeline error and then pipeline recovery fails.  
> Unfortunately, in this case, some of the resources used by the 
> {{DFSOutputStream}} are leaked.  One particularly important resource is file 
> leases.
> So it's possible for a long-lived HDFS client, such as Flume, to write many 
> blocks to a file, but then fail to close it.  Unfortunately, the 
> {{LeaseRenewerThread}} inside the client will continue to renew the lease for 
> the "undead" file.  Future attempts to close the file will just rethrow the 
> previous exception, and no progress can be made by the client.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4715) Backport HDFS-3577 and other related WebHDFS issue to branch-1

2013-04-18 Thread Tsz Wo (Nicholas), SZE (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4715?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635733#comment-13635733
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-4715:
--

The patch (HDFS-4549.2.patch) looks different from the current trunk code.  For 
backporting, we would try to backport the code exactly so that it is easier to 
maintain the code.  Could you make the backport patch the same as the original 
patches?  Some other comments:

- The import org.apache.log4j.helpers.BoundedFIFO is not used.

- the parameter in update(..) should be the read length but not the read value.
{code}
+return update(getInputStream().read())
{code}

- the filelength in udpate(..) may be null

> Backport HDFS-3577 and other related WebHDFS issue to branch-1
> --
>
> Key: HDFS-4715
> URL: https://issues.apache.org/jira/browse/HDFS-4715
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Mark Wagner
>
> The related JIRAs are HDFS-3577, HDFS-3318, and HDFS-3788.  Backporting them 
> can fix some WebHDFS performance issues in branch-1.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4716) TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs fails on Windows

2013-04-18 Thread Chris Nauroth (JIRA)
Chris Nauroth created HDFS-4716:
---

 Summary: 
TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs fails on Windows
 Key: HDFS-4716
 URL: https://issues.apache.org/jira/browse/HDFS-4716
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode, test
Affects Versions: 3.0.0
Reporter: Chris Nauroth
Assignee: Chris Nauroth


{{TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs}} fails on 
Windows due to an incorrectly initialized name dir in the {{Configuration}} 
used by the test.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4717) Change the parameter type of the snapshot methods in HdfsAdmin to Path

2013-04-18 Thread Tsz Wo (Nicholas), SZE (JIRA)
Tsz Wo (Nicholas), SZE created HDFS-4717:


 Summary: Change the parameter type of the snapshot methods in 
HdfsAdmin to Path
 Key: HDFS-4717
 URL: https://issues.apache.org/jira/browse/HDFS-4717
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: hdfs-client
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Tsz Wo (Nicholas), SZE


In HdfsAdmin, the path parameter type in allowSnapshot(String path) and 
allowSnapshot(String path) should be Path but not String.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4717) Change the parameter type of the snapshot methods in HdfsAdmin to Path

2013-04-18 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4717?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-4717:
-

Description: In HdfsAdmin, the path parameter type in allowSnapshot(String 
path) and disallowSnapshot(String path) should be Path but not String.  (was: 
In HdfsAdmin, the path parameter type in allowSnapshot(String path) and 
allowSnapshot(String path) should be Path but not String.)

> Change the parameter type of the snapshot methods in HdfsAdmin to Path
> --
>
> Key: HDFS-4717
> URL: https://issues.apache.org/jira/browse/HDFS-4717
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: hdfs-client
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
>
> In HdfsAdmin, the path parameter type in allowSnapshot(String path) and 
> disallowSnapshot(String path) should be Path but not String.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4716) TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs fails on Windows

2013-04-18 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4716?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-4716:


Attachment: HDFS-4716.1.patch

The test was not setting dfs.namenode.name.dir.  It was using the default.  On 
Windows, this would end up being an invalid URI containing '\' characters after 
configuration performed property substitution, i.e. file://C:\foo\bar/dfs/name.

This patch is a one-line fix to initialize dfs.namenode.name.dir to a correct 
value under test.build.data.

> TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs fails on Windows
> -
>
> Key: HDFS-4716
> URL: https://issues.apache.org/jira/browse/HDFS-4716
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode, test
>Affects Versions: 3.0.0
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-4716.1.patch
>
>
> {{TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs}} fails on 
> Windows due to an incorrectly initialized name dir in the {{Configuration}} 
> used by the test.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4716) TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs fails on Windows

2013-04-18 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4716?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-4716:


Status: Patch Available  (was: Open)

> TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs fails on Windows
> -
>
> Key: HDFS-4716
> URL: https://issues.apache.org/jira/browse/HDFS-4716
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode, test
>Affects Versions: 3.0.0
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-4716.1.patch
>
>
> {{TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs}} fails on 
> Windows due to an incorrectly initialized name dir in the {{Configuration}} 
> used by the test.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4716) TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs fails on Windows

2013-04-18 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4716?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-4716:


Status: Open  (was: Patch Available)

> TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs fails on Windows
> -
>
> Key: HDFS-4716
> URL: https://issues.apache.org/jira/browse/HDFS-4716
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode, test
>Affects Versions: 3.0.0
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-4716.1.patch
>
>
> {{TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs}} fails on 
> Windows due to an incorrectly initialized name dir in the {{Configuration}} 
> used by the test.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4705) TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir

2013-04-18 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4705?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-4705:


Attachment: HDFS-4705.1.patch

Hi, Ivan.  I had a patch in progress for this one.  I accidentally filed 
HDFS-4716, but I just resolved it as duplicate.  I'm attaching my current 
patch, which is a one-liner fix for just {{TestAllowFormat}} (not the other 
ones that you mentioned in the comments).

> TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir
> -
>
> Key: HDFS-4705
> URL: https://issues.apache.org/jira/browse/HDFS-4705
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Ivan Mitic
>Assignee: Ivan Mitic
>Priority: Minor
> Attachments: HDFS-4705.1.patch
>
>
> Test fails on Windows with the below exception:
> {code}
> testFormatShouldBeIgnoredForNonFileBasedDirs(org.apache.hadoop.hdfs.server.namenode.TestAllowFormat)
>   Time elapsed: 49 sec  <<< ERROR!
> java.io.IOException: No image directories available!
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:905)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:151)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:758)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:259)
>   at 
> org.apache.hadoop.hdfs.server.namenode.TestAllowFormat.testFormatShouldBeIgnoredForNonFileBasedDirs(TestAllowFormat.java:181)
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4705) TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir

2013-04-18 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4705?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-4705:


Status: Patch Available  (was: Open)

> TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir
> -
>
> Key: HDFS-4705
> URL: https://issues.apache.org/jira/browse/HDFS-4705
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Ivan Mitic
>Assignee: Ivan Mitic
>Priority: Minor
> Attachments: HDFS-4705.1.patch
>
>
> Test fails on Windows with the below exception:
> {code}
> testFormatShouldBeIgnoredForNonFileBasedDirs(org.apache.hadoop.hdfs.server.namenode.TestAllowFormat)
>   Time elapsed: 49 sec  <<< ERROR!
> java.io.IOException: No image directories available!
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:905)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:151)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:758)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:259)
>   at 
> org.apache.hadoop.hdfs.server.namenode.TestAllowFormat.testFormatShouldBeIgnoredForNonFileBasedDirs(TestAllowFormat.java:181)
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4713) Wrong server principal is used for rpc calls to namenode if HA is enabled

2013-04-18 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4713?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635767#comment-13635767
 ] 

Kihwal Lee commented on HDFS-4713:
--

It seems client side is okay.

> Wrong server principal is used for rpc calls to namenode if HA is enabled
> -
>
> Key: HDFS-4713
> URL: https://issues.apache.org/jira/browse/HDFS-4713
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: ha, namenode
>Affects Versions: 2.0.5-beta
>Reporter: Kihwal Lee
>Priority: Blocker
>
> When various components are connecting to a namenode in a HA-enabled 
> environment, a wrong server principal may be picked up.  This result in SASL 
> failure, since the client-side used a wrong service ticket for the connection.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4716) TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs fails on Windows

2013-04-18 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4716?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth resolved HDFS-4716.
-

Resolution: Duplicate

> TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs fails on Windows
> -
>
> Key: HDFS-4716
> URL: https://issues.apache.org/jira/browse/HDFS-4716
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode, test
>Affects Versions: 3.0.0
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-4716.1.patch
>
>
> {{TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs}} fails on 
> Windows due to an incorrectly initialized name dir in the {{Configuration}} 
> used by the test.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4705) TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir

2013-04-18 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4705?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635778#comment-13635778
 ] 

Chris Nauroth commented on HDFS-4705:
-

I can see 2 different approaches to fixing these tests:

# Fix each individual test to initialize dfs.namenode.name.dir properly.
# Add special case logic in {{Configuration#substituteVars}} saying that if the 
value starts with "file:", then pass any substitution values through 
{{File#toURI}} and {{URI#getPath}}.  On Windows, this would have the effect of 
converting '\' to '/' and yield a valid URI.

My initial patch takes the first approach.  I expect I can quickly prep a patch 
for the second approach too so we can compare and contrast.

BTW, I apologize if I stepped on your toes here.  Somehow, I missed that you 
had already filed this jira.


> TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir
> -
>
> Key: HDFS-4705
> URL: https://issues.apache.org/jira/browse/HDFS-4705
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Ivan Mitic
>Assignee: Ivan Mitic
>Priority: Minor
> Attachments: HDFS-4705.1.patch
>
>
> Test fails on Windows with the below exception:
> {code}
> testFormatShouldBeIgnoredForNonFileBasedDirs(org.apache.hadoop.hdfs.server.namenode.TestAllowFormat)
>   Time elapsed: 49 sec  <<< ERROR!
> java.io.IOException: No image directories available!
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:905)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:151)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:758)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:259)
>   at 
> org.apache.hadoop.hdfs.server.namenode.TestAllowFormat.testFormatShouldBeIgnoredForNonFileBasedDirs(TestAllowFormat.java:181)
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4715) Backport HDFS-3577 and other related WebHDFS issue to branch-1

2013-04-18 Thread Mark Wagner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4715?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Wagner updated HDFS-4715:
--

Attachment: HDFS-4751.1.patch

Addressed your points, and got things as close to trunk as possible.

> Backport HDFS-3577 and other related WebHDFS issue to branch-1
> --
>
> Key: HDFS-4715
> URL: https://issues.apache.org/jira/browse/HDFS-4715
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Mark Wagner
> Attachments: HDFS-4751.1.patch
>
>
> The related JIRAs are HDFS-3577, HDFS-3318, and HDFS-3788.  Backporting them 
> can fix some WebHDFS performance issues in branch-1.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4715) Backport HDFS-3577 and other related WebHDFS issue to branch-1

2013-04-18 Thread Mark Wagner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4715?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Wagner updated HDFS-4715:
--

Status: Patch Available  (was: Open)

> Backport HDFS-3577 and other related WebHDFS issue to branch-1
> --
>
> Key: HDFS-4715
> URL: https://issues.apache.org/jira/browse/HDFS-4715
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Mark Wagner
> Attachments: HDFS-4751.1.patch
>
>
> The related JIRAs are HDFS-3577, HDFS-3318, and HDFS-3788.  Backporting them 
> can fix some WebHDFS performance issues in branch-1.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4650) Add unit tests for rename with snapshots

2013-04-18 Thread Arpit Agarwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4650?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arpit Agarwal updated HDFS-4650:


Assignee: Jing Zhao  (was: Arpit Agarwal)

> Add unit tests for rename with snapshots
> 
>
> Key: HDFS-4650
> URL: https://issues.apache.org/jira/browse/HDFS-4650
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: datanode, namenode
>Reporter: Jing Zhao
>Assignee: Jing Zhao
>
> Add more unit tests and update current unit tests to cover different cases 
> for rename with existence of snapshottable directories and snapshots.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4686) Fix quota computation for rename with snapshots

2013-04-18 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4686?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao updated HDFS-4686:


Attachment: HDFS-4686.001.patch

Update the patch. Now the patch can pass all the unit tests with quota check 
enabled. Things to do next:
1. Fix INode#computeContentSummary
2. Add javadoc and clean the code

> Fix quota computation for rename with snapshots
> ---
>
> Key: HDFS-4686
> URL: https://issues.apache.org/jira/browse/HDFS-4686
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: datanode, namenode
>Reporter: Jing Zhao
>Assignee: Jing Zhao
> Attachments: HDFS-4686.001.patch, HDFS-4686.001.patch
>
>
> Currently after a rename operation within/from a snapshottable directory, a 
> reference node is created in both src and dst subtree, pointing to the 
> original renamed inode. With this change the original quota computation may 
> count the quota usage of the renamed subtree multiple times. This jira tries 
> to fix this issue.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4504) DFSOutputStream#close doesn't always release resources (such as leases)

2013-04-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635834#comment-13635834
 ] 

Hadoop QA commented on HDFS-4504:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12579431/HDFS-4504.002.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/4275//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4275//console

This message is automatically generated.

> DFSOutputStream#close doesn't always release resources (such as leases)
> ---
>
> Key: HDFS-4504
> URL: https://issues.apache.org/jira/browse/HDFS-4504
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
> Attachments: HDFS-4504.001.patch, HDFS-4504.002.patch
>
>
> {{DFSOutputStream#close}} can throw an {{IOException}} in some cases.  One 
> example is if there is a pipeline error and then pipeline recovery fails.  
> Unfortunately, in this case, some of the resources used by the 
> {{DFSOutputStream}} are leaked.  One particularly important resource is file 
> leases.
> So it's possible for a long-lived HDFS client, such as Flume, to write many 
> blocks to a file, but then fail to close it.  Unfortunately, the 
> {{LeaseRenewerThread}} inside the client will continue to renew the lease for 
> the "undead" file.  Future attempts to close the file will just rethrow the 
> previous exception, and no progress can be made by the client.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4715) Backport HDFS-3577 and other related WebHDFS issue to branch-1

2013-04-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4715?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635839#comment-13635839
 ] 

Hadoop QA commented on HDFS-4715:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12579452/HDFS-4751.1.patch
  against trunk revision .

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4278//console

This message is automatically generated.

> Backport HDFS-3577 and other related WebHDFS issue to branch-1
> --
>
> Key: HDFS-4715
> URL: https://issues.apache.org/jira/browse/HDFS-4715
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Mark Wagner
> Attachments: HDFS-4751.1.patch
>
>
> The related JIRAs are HDFS-3577, HDFS-3318, and HDFS-3788.  Backporting them 
> can fix some WebHDFS performance issues in branch-1.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4705) TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir

2013-04-18 Thread Ivan Mitic (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4705?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635857#comment-13635857
 ] 

Ivan Mitic commented on HDFS-4705:
--

Hey Chris, no worries at all, and thanks for attaching the patch!

I started with a similar approach to yours and then saw the other test 
failures, so I thought it is worth to spent some time seeing if it makes sense 
to fix them all at ones.

I was thinking along the lines of changing {{Util#fileAsURI}} such that it 
converts the given File to Path and then from Path to the URI. However, on top 
of this we'd also have to change the default value for 
{{dfs.namenode.name.dir}} as "file://$hadoop.tmp.dir/dfs/name" is actually not 
a valid local URI (there should 3 forward slashes after "file:", IOW: 
"file:///$hadoop.tmp.dir/dfs/name"). I haven't tested this out, so it might be 
that there are some problems here.

You're welcome to take this up if you want :)

> TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir
> -
>
> Key: HDFS-4705
> URL: https://issues.apache.org/jira/browse/HDFS-4705
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Ivan Mitic
>Assignee: Ivan Mitic
>Priority: Minor
> Attachments: HDFS-4705.1.patch
>
>
> Test fails on Windows with the below exception:
> {code}
> testFormatShouldBeIgnoredForNonFileBasedDirs(org.apache.hadoop.hdfs.server.namenode.TestAllowFormat)
>   Time elapsed: 49 sec  <<< ERROR!
> java.io.IOException: No image directories available!
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:905)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:151)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:758)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:259)
>   at 
> org.apache.hadoop.hdfs.server.namenode.TestAllowFormat.testFormatShouldBeIgnoredForNonFileBasedDirs(TestAllowFormat.java:181)
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4705) TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir

2013-04-18 Thread Ivan Mitic (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4705?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635859#comment-13635859
 ] 

Ivan Mitic commented on HDFS-4705:
--

PS. I think the approach from the current patch is actually a fine way to 
address the problem, I just thought it is worth to check if there are better 
ways.

> TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir
> -
>
> Key: HDFS-4705
> URL: https://issues.apache.org/jira/browse/HDFS-4705
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Ivan Mitic
>Assignee: Ivan Mitic
>Priority: Minor
> Attachments: HDFS-4705.1.patch
>
>
> Test fails on Windows with the below exception:
> {code}
> testFormatShouldBeIgnoredForNonFileBasedDirs(org.apache.hadoop.hdfs.server.namenode.TestAllowFormat)
>   Time elapsed: 49 sec  <<< ERROR!
> java.io.IOException: No image directories available!
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:905)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:151)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:758)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:259)
>   at 
> org.apache.hadoop.hdfs.server.namenode.TestAllowFormat.testFormatShouldBeIgnoredForNonFileBasedDirs(TestAllowFormat.java:181)
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4705) TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir

2013-04-18 Thread Ivan Mitic (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4705?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635865#comment-13635865
 ] 

Ivan Mitic commented on HDFS-4705:
--

bq. However, on top of this we'd also have to change the default value for 
dfs.namenode.name.dir as "file://$hadoop.tmp.dir/dfs/name" is actually not a 
valid local URI (there should 3 forward slashes after "file:", IOW: 
"file:///$hadoop.tmp.dir/dfs/name").
Correcting myself here. The above URI is actually valid since $hadoop.tmp.dir 
contains one forward slash at the beginning. The rest of what I said below 
should still make sense.

> TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir
> -
>
> Key: HDFS-4705
> URL: https://issues.apache.org/jira/browse/HDFS-4705
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Ivan Mitic
>Assignee: Ivan Mitic
>Priority: Minor
> Attachments: HDFS-4705.1.patch
>
>
> Test fails on Windows with the below exception:
> {code}
> testFormatShouldBeIgnoredForNonFileBasedDirs(org.apache.hadoop.hdfs.server.namenode.TestAllowFormat)
>   Time elapsed: 49 sec  <<< ERROR!
> java.io.IOException: No image directories available!
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:905)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:151)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:758)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:259)
>   at 
> org.apache.hadoop.hdfs.server.namenode.TestAllowFormat.testFormatShouldBeIgnoredForNonFileBasedDirs(TestAllowFormat.java:181)
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4716) TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs fails on Windows

2013-04-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4716?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635869#comment-13635869
 ] 

Hadoop QA commented on HDFS-4716:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12579438/HDFS-4716.1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/4276//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4276//console

This message is automatically generated.

> TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs fails on Windows
> -
>
> Key: HDFS-4716
> URL: https://issues.apache.org/jira/browse/HDFS-4716
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode, test
>Affects Versions: 3.0.0
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
> Attachments: HDFS-4716.1.patch
>
>
> {{TestAllowFormat#testFormatShouldBeIgnoredForNonFileBasedDirs}} fails on 
> Windows due to an incorrectly initialized name dir in the {{Configuration}} 
> used by the test.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4705) TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir

2013-04-18 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4705?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635876#comment-13635876
 ] 

Hadoop QA commented on HDFS-4705:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12579441/HDFS-4705.1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/4277//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4277//console

This message is automatically generated.

> TestAllowFormat fails on Windows because of invalid dfs.namenode.name.dir
> -
>
> Key: HDFS-4705
> URL: https://issues.apache.org/jira/browse/HDFS-4705
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Ivan Mitic
>Assignee: Ivan Mitic
>Priority: Minor
> Attachments: HDFS-4705.1.patch
>
>
> Test fails on Windows with the below exception:
> {code}
> testFormatShouldBeIgnoredForNonFileBasedDirs(org.apache.hadoop.hdfs.server.namenode.TestAllowFormat)
>   Time elapsed: 49 sec  <<< ERROR!
> java.io.IOException: No image directories available!
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:912)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:905)
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:151)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:758)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:259)
>   at 
> org.apache.hadoop.hdfs.server.namenode.TestAllowFormat.testFormatShouldBeIgnoredForNonFileBasedDirs(TestAllowFormat.java:181)
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4434) Provide a mapping from INodeId to INode

2013-04-18 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4434?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635882#comment-13635882
 ] 

Hudson commented on HDFS-4434:
--

Integrated in Hadoop-trunk-Commit #3632 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/3632/])
HDFS-4434. Provide a mapping from INodeId to INode. Contributed by Suresh 
Srinivas. (Revision 1469644)

 Result = SUCCESS
suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1469644
Files : 
* /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlocksMap.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormat.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INode.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeDirectory.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeId.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestINodeFile.java


> Provide a mapping from INodeId to INode
> ---
>
> Key: HDFS-4434
> URL: https://issues.apache.org/jira/browse/HDFS-4434
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: 3.0.0
>Reporter: Brandon Li
>Assignee: Suresh Srinivas
> Attachments: HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch
>
>
> This JIRA is to provide a way to access the INode via its id. The proposed 
> solution is to have an in-memory mapping from INodeId to INode. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4434) Provide a mapping from INodeId to INode

2013-04-18 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4434?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-4434:
--

   Resolution: Fixed
Fix Version/s: 3.0.0
 Release Note: This change adds support for referencing files and 
directories based on fileID/inodeID using a path /.reserved/.inodes/. 
With this change creating a file or directory /.reserved is not longer allowed. 
Before upgrading to a release with this change, files /.reserved needs to be 
renamed to another name.
 Hadoop Flags: Incompatible change,Reviewed
   Status: Resolved  (was: Patch Available)

> Provide a mapping from INodeId to INode
> ---
>
> Key: HDFS-4434
> URL: https://issues.apache.org/jira/browse/HDFS-4434
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Affects Versions: 3.0.0
>Reporter: Brandon Li
>Assignee: Suresh Srinivas
> Fix For: 3.0.0
>
> Attachments: HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
> HDFS-4434.patch
>
>
> This JIRA is to provide a way to access the INode via its id. The proposed 
> solution is to have an in-memory mapping from INodeId to INode. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4489) Use InodeID as as an identifier of a file in HDFS protocols and APIs

2013-04-18 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li resolved HDFS-4489.
--

Resolution: Fixed

Close this JIRA since all its sub-issues have been resolved. 

> Use InodeID as as an identifier of a file in HDFS protocols and APIs
> 
>
> Key: HDFS-4489
> URL: https://issues.apache.org/jira/browse/HDFS-4489
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Reporter: Brandon Li
>Assignee: Brandon Li
>
> The benefit of using InodeID to uniquely identify a file can be multiple 
> folds. Here are a few of them:
> 1. uniquely identify a file cross rename, related JIRAs include HDFS-4258, 
> HDFS-4437.
> 2. modification checks in tools like distcp. Since a file could have been 
> replaced or renamed to, the file name and size combination is no t reliable, 
> but the combination of file id and size is unique.
> 3. id based protocol support (e.g., NFS)
> 4. to make the pluggable block placement policy use fileid instead of 
> filename (HDFS-385).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HDFS-4717) Change the parameter type of the snapshot methods in HdfsAdmin to Path

2013-04-18 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4717?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-4717:
-

Attachment: h4717_20130418.patch

h4717_20130418.patch: change the type.

> Change the parameter type of the snapshot methods in HdfsAdmin to Path
> --
>
> Key: HDFS-4717
> URL: https://issues.apache.org/jira/browse/HDFS-4717
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: hdfs-client
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Attachments: h4717_20130418.patch
>
>
> In HdfsAdmin, the path parameter type in allowSnapshot(String path) and 
> disallowSnapshot(String path) should be Path but not String.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HDFS-4717) Change the parameter type of the snapshot methods in HdfsAdmin to Path

2013-04-18 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4717?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13635933#comment-13635933
 ] 

Jing Zhao commented on HDFS-4717:
-

The patch looks good to me. +1 for the patch.

> Change the parameter type of the snapshot methods in HdfsAdmin to Path
> --
>
> Key: HDFS-4717
> URL: https://issues.apache.org/jira/browse/HDFS-4717
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: hdfs-client
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Attachments: h4717_20130418.patch
>
>
> In HdfsAdmin, the path parameter type in allowSnapshot(String path) and 
> disallowSnapshot(String path) should be Path but not String.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4717) Change the parameter type of the snapshot methods in HdfsAdmin to Path

2013-04-18 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4717?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE resolved HDFS-4717.
--

   Resolution: Fixed
Fix Version/s: Snapshot (HDFS-2802)
 Hadoop Flags: Reviewed

Thanks Jing for reviewing the patch.

I have committed this.

> Change the parameter type of the snapshot methods in HdfsAdmin to Path
> --
>
> Key: HDFS-4717
> URL: https://issues.apache.org/jira/browse/HDFS-4717
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: hdfs-client
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: Snapshot (HDFS-2802)
>
> Attachments: h4717_20130418.patch
>
>
> In HdfsAdmin, the path parameter type in allowSnapshot(String path) and 
> disallowSnapshot(String path) should be Path but not String.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira