[jira] [Commented] (HDFS-7827) Erasure Coding: support striped blocks in non-protobuf fsimage

2015-03-16 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7827?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364100#comment-14364100
 ] 

Jing Zhao commented on HDFS-7827:
-

Currently we still have non-protobuf image for offline processing. Thus this 
change can still be helpful especially for storage usage analysis.

Some comments on the current patch:
# After updating {{writeINodeUnderConstruction}}, we also need to update 
{{readINodeUnderConstruction}} accordingly.
# We also need a unit test to cover the under construction file with striped 
blocks.
# Instead of the following code, we can simply use 
out.writeBoolean(cons.isWithStripedBlocks()). Whether we keep using 
EC_STORAGE_POLICY_ID is still under construction.
{code}
+if (cons.isStriped()){
+  out.writeByte(HdfsConstants.EC_STORAGE_POLICY_ID);
+}else{
+  out.writeByte(0);
+}
{code}
# Please correct the coding format according to the project conventions.

Besides, as [~wheat9] pointed out, we also need to update WebImageViewer for 
file size calculation. Please open a separate jira to do this.

 Erasure Coding: support striped blocks in non-protobuf fsimage
 --

 Key: HDFS-7827
 URL: https://issues.apache.org/jira/browse/HDFS-7827
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jing Zhao
Assignee: Hui Zheng
 Attachments: HDFS-7827.000.patch, HDFS-7827.001.patch, 
 HDFS-7827.002.patch


 HDFS-7749 only adds code to persist striped blocks to protobuf-based fsimage. 
 We should also add this support to the non-protobuf fsimage since it is still 
 used for use cases like offline image processing.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-7937) Erasure Coding: INodeFile quota computation unit tests

2015-03-16 Thread Kai Sasaki (JIRA)
Kai Sasaki created HDFS-7937:


 Summary: Erasure Coding: INodeFile quota computation unit tests
 Key: HDFS-7937
 URL: https://issues.apache.org/jira/browse/HDFS-7937
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Sasaki
Assignee: Kai Sasaki
Priority: Minor


Unit test for [HDFS-7826|https://issues.apache.org/jira/browse/HDFS-7826]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7854) Separate class DataStreamer out of DFSOutputStream

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7854?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364386#comment-14364386
 ] 

Hadoop QA commented on HDFS-7854:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12704951/HDFS-7854-004-duplicate3.patch
  against trunk revision 5608520.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:red}-1 eclipse:eclipse{color}.  The patch failed to build with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in .

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/9908//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9908//console

This message is automatically generated.

 Separate class DataStreamer out of DFSOutputStream
 --

 Key: HDFS-7854
 URL: https://issues.apache.org/jira/browse/HDFS-7854
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Li Bo
Assignee: Li Bo
 Attachments: HDFS-7854-001.patch, HDFS-7854-002.patch, 
 HDFS-7854-003.patch, HDFS-7854-004-duplicate.patch, 
 HDFS-7854-004-duplicate2.patch, HDFS-7854-004-duplicate3.patch, 
 HDFS-7854-004.patch


 This sub task separate DataStreamer from DFSOutputStream. New DataStreamer 
 will accept packets and write them to remote datanodes.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-7938) OpensslSecureRandom.c pthread_threadid_np usage signature is wrong on 32-bit Mac

2015-03-16 Thread Colin Patrick McCabe (JIRA)
Colin Patrick McCabe created HDFS-7938:
--

 Summary: OpensslSecureRandom.c pthread_threadid_np usage signature 
is wrong on 32-bit Mac
 Key: HDFS-7938
 URL: https://issues.apache.org/jira/browse/HDFS-7938
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.7.0
Reporter: Colin Patrick McCabe
Priority: Critical


In OpensslSecureRandom.c, pthread_threadid_np is being used with an unsigned 
long, but the type signature requires a uint64_t.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7935) Support multi-homed networks when Kerberos security is enabled

2015-03-16 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363895#comment-14363895
 ] 

Kihwal Lee commented on HDFS-7935:
--

Secure HA NN does come up with dfs.namenode.http-address.logical_name.nn_id 
set to 0.0.0.0:port. The clients should of course have a vaild address. The 
kerberos principal used by the two name nodes comes from 
dfs.web.authentication.kerberos.principal, but SPNEGO will work with any SPN 
found in its keytab after HADOOP-10322.  So I guess HDFS-4448 is fixed?

 Support multi-homed networks when Kerberos security is enabled
 --

 Key: HDFS-7935
 URL: https://issues.apache.org/jira/browse/HDFS-7935
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Arun Suresh
Assignee: Arun Suresh

 Currently, during SASL negotiation stage between ipc Client and Server, The 
 server sends only a single serviceId (curresponding to a single principal) to 
 the client. This is the principal the the server process is logged in as 
 during startup.
 It is possible that in a multi-homed network, the server might be associated 
 with more than one principal, and thus severs must provide the clients all 
 possible principals it can use to connect to.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-2743) Streamline usage of bookkeeper journal manager

2015-03-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2743?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HDFS-2743:
---
Fix Version/s: (was: 2.0.0-alpha)
   3.0.0

 Streamline usage of bookkeeper journal manager
 --

 Key: HDFS-2743
 URL: https://issues.apache.org/jira/browse/HDFS-2743
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Ivan Kelly
Assignee: Ivan Kelly
 Fix For: 3.0.0

 Attachments: HDFS-2743.diff, HDFS-2743.diff


 The current method of installing bkjournal manager involves generating a 
 tarball, and extracting it with special flags over the hdfs distribution. 
 This is cumbersome and prone to being broken by other changes (see 
 https://svn.apache.org/repos/asf/hadoop/common/trunk@1220940). I think a 
 cleaner way to doing this is to generate a single jar that can be placed in 
 the lib dir of hdfs.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7936) Erasure coding: resolving conflicts when merging with HDFS-7903 and HDFS-7435

2015-03-16 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7936?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HDFS-7936:

Attachment: HDFS-7936-002.patch

001 patch had unnecessary changes to import

 Erasure coding: resolving conflicts when merging with HDFS-7903 and HDFS-7435
 -

 Key: HDFS-7936
 URL: https://issues.apache.org/jira/browse/HDFS-7936
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Zhe Zhang
Assignee: Jing Zhao
 Attachments: HDFS-7936-001.patch, HDFS-7936-002.patch


 A few non-trivial conflicts were found when merging trunk changes. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7936) Erasure coding: resolving conflicts when merging with HDFS-7903 and HDFS-7435

2015-03-16 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7936?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363961#comment-14363961
 ] 

Zhe Zhang commented on HDFS-7936:
-

[~jingzhao] Thanks! I actually have a patch (guess JIRA has a delay in showing 
it). Let me know if the 002 patch looks correct to you.

 Erasure coding: resolving conflicts when merging with HDFS-7903 and HDFS-7435
 -

 Key: HDFS-7936
 URL: https://issues.apache.org/jira/browse/HDFS-7936
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Zhe Zhang
Assignee: Jing Zhao
 Attachments: HDFS-7936-001.patch, HDFS-7936-002.patch


 A few non-trivial conflicts were found when merging trunk changes. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7935) Support multi-homed networks when Kerberos security is enabled

2015-03-16 Thread Arun Suresh (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363979#comment-14363979
 ] 

Arun Suresh commented on HDFS-7935:
---

Hit enter by mistake... 

Continuing..
So looks like we still need the HDFS-4448 fix (I guess it needs a minor 
rebasing though..)

 Support multi-homed networks when Kerberos security is enabled
 --

 Key: HDFS-7935
 URL: https://issues.apache.org/jira/browse/HDFS-7935
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Arun Suresh
Assignee: Arun Suresh

 Currently, during SASL negotiation stage between ipc Client and Server, The 
 server sends only a single serviceId (curresponding to a single principal) to 
 the client. This is the principal the the server process is logged in as 
 during startup.
 It is possible that in a multi-homed network, the server might be associated 
 with more than one principal, and thus severs must provide the clients all 
 possible principals it can use to connect to.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-3265) PowerPc Build error.

2015-03-16 Thread Allen Wittenauer (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3265?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364042#comment-14364042
 ] 

Allen Wittenauer commented on HDFS-3265:


For those spelunking:  this was never committed to any other branches other 
than 1.x and trunk.

 PowerPc Build error.
 

 Key: HDFS-3265
 URL: https://issues.apache.org/jira/browse/HDFS-3265
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 1.0.2, 1.0.3, 2.0.0-alpha
 Environment: Linux RHEL 6.1 PowerPC + IBM JVM 6.0 SR10
Reporter: Kumar Ravi
Assignee: Kumar Ravi
  Labels: patch
 Fix For: 1.0.3, 3.0.0

 Attachments: HADOOP-8271.patch, HADOOP-8271.patch

   Original Estimate: 168h
  Remaining Estimate: 168h

 When attempting to build branch-1, the following error is seen and ant exits.
 [exec] configure: error: Unsupported CPU architecture powerpc64
 The following command was used to build hadoop-common
 ant -Dlibhdfs=true -Dcompile.native=true -Dfusedfs=true -Dcompile.c++=true 
 -Dforrest.home=$FORREST_HOME compile-core-native compile-c++ 
 compile-c++-examples task-controller tar record-parser compile-hdfs-classes 
 package -Djava5.home=/opt/ibm/ibm-java2-ppc64-50/ 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HDFS-7933) fsck should also report decommissioning replicas.

2015-03-16 Thread Xiaoyu Yao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7933?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao reassigned HDFS-7933:


Assignee: Xiaoyu Yao

 fsck should also report decommissioning replicas. 
 --

 Key: HDFS-7933
 URL: https://issues.apache.org/jira/browse/HDFS-7933
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Reporter: Jitendra Nath Pandey
Assignee: Xiaoyu Yao

 Fsck doesn't count replicas that are on decommissioning nodes. If a block has 
 all replicas on the decommissioning nodes, it will be marked as missing, 
 which is alarming for the admins, although the system will replicate them 
 before nodes are decommissioned.
 Fsck output should also show decommissioning replicas along with the live 
 replicas.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-3549) dist tar build fails in hadoop-hdfs-raid project

2015-03-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HDFS-3549:
---
Fix Version/s: (was: 2.0.0-alpha)
   3.0.0

 dist tar build fails in hadoop-hdfs-raid project
 

 Key: HDFS-3549
 URL: https://issues.apache.org/jira/browse/HDFS-3549
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 3.0.0
Reporter: Jason Lowe
Assignee: Jason Lowe
Priority: Critical
 Fix For: 3.0.0

 Attachments: HDFS-3549.patch, HDFS-3549.patch, HDFS-3549.patch, 
 HDFS-3549.patch


 Trying to build the distribution tarball in a clean tree via {{mvn install 
 -Pdist -Dtar -DskipTests -Dmaven.javadoc.skip}} fails with this error:
 {noformat}
 main:
  [exec] tar: hadoop-hdfs-raid-3.0.0-SNAPSHOT: Cannot stat: No such file 
 or directory
  [exec] tar: Exiting with failure status due to previous errors
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7878) API - expose an unique file identifier

2015-03-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7878?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364104#comment-14364104
 ] 

Sergey Shelukhin commented on HDFS-7878:


Note: confusingly, HdfsFileStatus is not FileStatus, that's an internal class

[~cmccabe] can you clarify what you mean by two RPCs? The API makes one RPC.
Two RPCs will only happen if the user calls both get ID and get status, which 
is not necessary.


 API - expose an unique file identifier
 --

 Key: HDFS-7878
 URL: https://issues.apache.org/jira/browse/HDFS-7878
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HDFS-7878.01.patch, HDFS-7878.02.patch, HDFS-7878.patch


 See HDFS-487.
 Even though that is resolved as duplicate, the ID is actually not exposed by 
 the JIRA it supposedly duplicates.
 INode ID for the file should be easy to expose; alternatively ID could be 
 derived from block IDs, to account for appends...
 This is useful e.g. for cache key by file, to make sure cache stays correct 
 when file is overwritten.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7826) Erasure Coding: Update INodeFile quota computation for striped blocks

2015-03-16 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7826?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364218#comment-14364218
 ] 

Jing Zhao commented on HDFS-7826:
-

+1 for the .5 patch. I will commit it shortly.

 Erasure Coding: Update INodeFile quota computation for striped blocks
 -

 Key: HDFS-7826
 URL: https://issues.apache.org/jira/browse/HDFS-7826
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jing Zhao
Assignee: Kai Sasaki
 Attachments: HDFS-7826.1.patch, HDFS-7826.2.patch, HDFS-7826.3.patch, 
 HDFS-7826.4.patch, HDFS-7826.5.patch


 Currently INodeFile's quota computation only considers contiguous blocks 
 (i.e., {{INodeFile#blocks}}). We need to update it to support striped blocks.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7054) Make DFSOutputStream tracing more fine-grained

2015-03-16 Thread Colin Patrick McCabe (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7054?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-7054:
---
Attachment: HDFS-7054.001.patch

 Make DFSOutputStream tracing more fine-grained
 --

 Key: HDFS-7054
 URL: https://issues.apache.org/jira/browse/HDFS-7054
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Affects Versions: 2.6.0
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
 Attachments: HDFS-7054.001.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7054) Make DFSOutputStream tracing more fine-grained

2015-03-16 Thread Colin Patrick McCabe (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7054?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-7054:
---
Target Version/s: 2.7.0  (was: 2.6.0)
  Status: Patch Available  (was: Open)

 Make DFSOutputStream tracing more fine-grained
 --

 Key: HDFS-7054
 URL: https://issues.apache.org/jira/browse/HDFS-7054
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Affects Versions: 2.6.0
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
 Attachments: HDFS-7054.001.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7854) Separate class DataStreamer out of DFSOutputStream

2015-03-16 Thread Li Bo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7854?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Li Bo updated HDFS-7854:

Attachment: HDFS-7854-004-duplicate3.patch

The findbugsWarnings's url cannot be opened, trigger Jenkins again.

 Separate class DataStreamer out of DFSOutputStream
 --

 Key: HDFS-7854
 URL: https://issues.apache.org/jira/browse/HDFS-7854
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Li Bo
Assignee: Li Bo
 Attachments: HDFS-7854-001.patch, HDFS-7854-002.patch, 
 HDFS-7854-003.patch, HDFS-7854-004-duplicate.patch, 
 HDFS-7854-004-duplicate2.patch, HDFS-7854-004-duplicate3.patch, 
 HDFS-7854-004.patch


 This sub task separate DataStreamer from DFSOutputStream. New DataStreamer 
 will accept packets and write them to remote datanodes.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7912) Erasure Coding: track BlockInfo instead of Block in UnderReplicatedBlocks and PendingReplicationBlocks

2015-03-16 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7912?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364346#comment-14364346
 ] 

Zhe Zhang commented on HDFS-7912:
-

Thanks Jing for clarifying about NN overhead.

About the second minor point, isn't {{storedBlock}} always equal to {{lBlk}} 
earlier in the code?
{code}
for (LocatedBlock lBlk : blocks.getLocatedBlocks()) {
  ExtendedBlock block = lBlk.getBlock();
{code}

 Erasure Coding: track BlockInfo instead of Block in UnderReplicatedBlocks and 
 PendingReplicationBlocks
 --

 Key: HDFS-7912
 URL: https://issues.apache.org/jira/browse/HDFS-7912
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jing Zhao
Assignee: Jing Zhao
 Attachments: HDFS-7912.000.patch


 Now with striped blocks and the design that uses a single BlockInfoStriped 
 object to track all the corresponding blocks, we need to clearly distinguish 
 the type Block and BlockInfo in BlockManager. Specifically, data structures 
 like {{UnderReplicatedBlocks}} and {{PendingReplicationBlocks}} should track 
 BlockInfo instead of Block in order to support striped block recovery.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7878) API - expose an unique file identifier

2015-03-16 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7878?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364389#comment-14364389
 ] 

Colin Patrick McCabe commented on HDFS-7878:


Hi [~jingzhao],

I understand that the current patch only makes one RPC.  My concern is that in 
order to get both the file ID and the other information about the file, two 
RPCs will be needed when using this API.  It's just a bad API from an 
efficiency and concurrency point of view.  It would be like if you needed 5 
separate RPCs to get the length, access time, modification time, group, and 
user.  There is a reason why FileStatus combines together all these fields into 
one class, and that same logic should lead us to including a way of getting 
file ID out of FileStatus.

I understand that HdfsFileStatus already has {{HdfsFileStatus#getFileId}}.  But 
this functionality should not be HDFS-specific.  There are other filesystems 
that have file IDs.  (Plus, if casting to HdfsFileStatus was adequate, there 
would be no need for this JIRA at all, since this cast is already possible.)

Why not just add an accessor function to {{FileStatus}} with a default 
implementation like I suggested earlier?  Adding a new function to a stable 
interface is a compatible change (and I note, that adding a new function to 
FileSystem is also adding a new function to a stable class).

 API - expose an unique file identifier
 --

 Key: HDFS-7878
 URL: https://issues.apache.org/jira/browse/HDFS-7878
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HDFS-7878.01.patch, HDFS-7878.02.patch, HDFS-7878.patch


 See HDFS-487.
 Even though that is resolved as duplicate, the ID is actually not exposed by 
 the JIRA it supposedly duplicates.
 INode ID for the file should be easy to expose; alternatively ID could be 
 derived from block IDs, to account for appends...
 This is useful e.g. for cache key by file, to make sure cache stays correct 
 when file is overwritten.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7937) Erasure Coding: INodeFile quota computation unit tests

2015-03-16 Thread Kai Sasaki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7937?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kai Sasaki updated HDFS-7937:
-
Attachment: (was: HDFS-7937.1.patch)

 Erasure Coding: INodeFile quota computation unit tests
 --

 Key: HDFS-7937
 URL: https://issues.apache.org/jira/browse/HDFS-7937
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Sasaki
Assignee: Kai Sasaki
Priority: Minor

 Unit test for [HDFS-7826|https://issues.apache.org/jira/browse/HDFS-7826]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7902) Expose truncate API for libwebhdfs

2015-03-16 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7902?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364457#comment-14364457
 ] 

Colin Patrick McCabe commented on HDFS-7902:


{code}
130 int createUrlForNnTRUNCATE(const char *host, int nnPort, const char 
*path,
131size_t newlength, const char *user, char 
**url)
...
139   strlength = snprintf(newlengthStr, LONG_STR_LEN, %lu, 
newlength);
{code}

newlength should be a {{tOffset}} like it is in the {{hdfs.h}} API.  It 
should be {{%llu}} because on 32-bit systems, long is not a 64-bit type.

{code}
1143ret = createUrlForNnTRUNCATE(fs-nn, fs-port, absPath,
1144 newlength, fs-userName, url);
1145free(absPath);
1146if (ret) {
1147errno = ret;
1148return -1;
1149}
{code}
url isn't freed on the error case here.  Similarly with resp.

 Expose truncate API for libwebhdfs
 --

 Key: HDFS-7902
 URL: https://issues.apache.org/jira/browse/HDFS-7902
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: native, webhdfs
Affects Versions: 2.7.0
Reporter: Yi Liu
Assignee: Yi Liu
 Attachments: HDFS-7902.001.patch, HDFS-7902.002.patch


 As Colin suggested in HDFS-7838, we will add truncate support for libwebhdfs.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7938) OpensslSecureRandom.c pthread_threadid_np usage signature is wrong on 32-bit Mac

2015-03-16 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7938?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364488#comment-14364488
 ] 

Chris Nauroth commented on HDFS-7938:
-

Hello, [~kiranmr].  Are you interested in taking this as a follow-up?

 OpensslSecureRandom.c pthread_threadid_np usage signature is wrong on 32-bit 
 Mac
 

 Key: HDFS-7938
 URL: https://issues.apache.org/jira/browse/HDFS-7938
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.7.0
Reporter: Colin Patrick McCabe
Priority: Critical

 In OpensslSecureRandom.c, pthread_threadid_np is being used with an unsigned 
 long, but the type signature requires a uint64_t.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7937) Erasure Coding: INodeFile quota computation unit tests

2015-03-16 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7937?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364250#comment-14364250
 ] 

Jing Zhao commented on HDFS-7937:
-

Thanks for opening the jira, Kai.

BTW, there is some code cleanup work we can do for the change brought by 
HDFS-7826. Specifically, {{computeQuotaUsageWithStriped}} does not use the 
parameter bsps. Also we need to fix its javadoc. You can include the cleanup in 
this jira.

 Erasure Coding: INodeFile quota computation unit tests
 --

 Key: HDFS-7937
 URL: https://issues.apache.org/jira/browse/HDFS-7937
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Sasaki
Assignee: Kai Sasaki
Priority: Minor

 Unit test for [HDFS-7826|https://issues.apache.org/jira/browse/HDFS-7826]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-2360) Ugly stacktrace when quota exceeds

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364326#comment-14364326
 ] 

Hadoop QA commented on HDFS-2360:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12704900/HDFS-2360.patch
  against trunk revision 2681ed9.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The following test timeouts occurred in 
hadoop-hdfs-project/hadoop-hdfs:

org.apache.hadoop.hdfs.TestFileCreation

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/9906//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9906//console

This message is automatically generated.

 Ugly stacktrace when quota exceeds
 --

 Key: HDFS-2360
 URL: https://issues.apache.org/jira/browse/HDFS-2360
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 0.23.0
Reporter: Rajit Saha
Assignee: Harsh J
Priority: Minor
 Attachments: HDFS-2360.patch, HDFS-2360.patch


 Will it be better to catch the exception and throw a small reasonable messege 
 to user when they exceed quota?
 $hdfs  dfs -mkdir testDir
 $hdfs  dfsadmin -setSpaceQuota 191M  testDir
 $hdfs dfs -count -q testDir
 none inf   200278016   2002780161 
0  0
 hdfs://NN hostname:port/user/hdfsqa/testDir
 $hdfs dfs -put /etc/passwd /user/hadoopqa/testDir 
 11/09/19 08:08:15 WARN hdfs.DFSClient: DataStreamer Exception
 org.apache.hadoop.hdfs.protocol.DSQuotaExceededException: The DiskSpace quota 
 of /user/hdfsqa/testDir is exceeded:
 quota=191.0m diskspace consumed=768.0m
 at 
 org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota.verifyQuota(INodeDirectoryWithQuota.java:159)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.verifyQuota(FSDirectory.java:1609)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.updateCount(FSDirectory.java:1383)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.addBlock(FSDirectory.java:370)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allocateBlock(FSNamesystem.java:1681)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1476)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:389)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at 
 org.apache.hadoop.ipc.WritableRpcEngine$Server.call(WritableRpcEngine.java:365)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1496)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1492)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1135)
 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1490)
 at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
 Method)
 at 
 sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
 at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
 at 
 org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:90)
 at 
 org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:57)
 at 
 

[jira] [Commented] (HDFS-7054) Make DFSOutputStream tracing more fine-grained

2015-03-16 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7054?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364418#comment-14364418
 ] 

Colin Patrick McCabe commented on HDFS-7054:


v2 fixes a missing import

 Make DFSOutputStream tracing more fine-grained
 --

 Key: HDFS-7054
 URL: https://issues.apache.org/jira/browse/HDFS-7054
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Affects Versions: 2.6.0
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
 Attachments: HDFS-7054.001.patch, HDFS-7054.002.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-5523) Support multiple subdirectory exports in HDFS NFS gateway

2015-03-16 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364259#comment-14364259
 ] 

Brandon Li commented on HDFS-5523:
--

{quote}We should probably disallow nesting in this phase.{quote}
Sounds ok to me. Otherwise, it would be hard to decide the access mode, for 
example: export /a is read-only but /a/b is read-write, when user traverse from 
/a to /a/b, it's tricky to decide which access the user should have. 

{quote}I think we should allow mounting a non-top-level directory.{quote}
In this case, we can have root as the only mount point and the users can mount 
any sub-directory. This might be the easiest way to implement, but provide only 
marginal security provision. Users can mount any sub-directories  as long as he 
has the right to do mount operation. Usually a root user is needed to do mount, 
so if customers' environment has control over who can do mount, then this 
option is viable solution. Of course, to add a bit more security control, the 
user can choose to export multiple sub-directories instead of only root.


 Support multiple subdirectory exports in HDFS NFS gateway 
 --

 Key: HDFS-5523
 URL: https://issues.apache.org/jira/browse/HDFS-5523
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: nfs
Reporter: Brandon Li

 Currently, the HDFS NFS Gateway only supports configuring a single 
 subdirectory export via the  {{dfs.nfs3.export.point}} configuration setting. 
 Supporting multiple subdirectory exports can make data and security 
 management easier when using the HDFS NFS Gateway.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7864) Erasure Coding: Update safemode calculation for striped blocks

2015-03-16 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7864?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364289#comment-14364289
 ] 

Jing Zhao commented on HDFS-7864:
-

Thanks for updating the patch, Rui! Some comments for the latest patch:
# Actually we do not need {{DFS_NAMENODE_STRIPE_MIN_DEFAULT}}. For a striped
  block, the minimal safe number should depend on the EC schema of the
  individual block (i.e., the number of data blocks of a striped block group).
  Therefore we also need to update the logic of {{incrementSafeBlockCount}}.
# There are still multiple places not following the current project code 
convention. E.g., we need to put a space after each comma in the function 
parameter list.

 Erasure Coding: Update safemode calculation for striped blocks
 --

 Key: HDFS-7864
 URL: https://issues.apache.org/jira/browse/HDFS-7864
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jing Zhao
Assignee: GAO Rui
 Attachments: HDFS-7864.1.patch, HDFS-7864.2.patch


 We need to update the safemode calculation for striped blocks. Specifically, 
 each striped block now consists of multiple data/parity blocks stored in 
 corresponding DataNodes. The current code's calculation is thus inconsistent: 
 each striped block is only counted as 1 expected block, while each of its 
 member block may increase the number of received blocks by 1.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7054) Make DFSOutputStream tracing more fine-grained

2015-03-16 Thread Colin Patrick McCabe (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7054?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HDFS-7054:
---
Attachment: HDFS-7054.002.patch

 Make DFSOutputStream tracing more fine-grained
 --

 Key: HDFS-7054
 URL: https://issues.apache.org/jira/browse/HDFS-7054
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Affects Versions: 2.6.0
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
 Attachments: HDFS-7054.001.patch, HDFS-7054.002.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7933) fsck should also report decommissioning replicas.

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7933?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364388#comment-14364388
 ] 

Hadoop QA commented on HDFS-7933:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12704903/HDFS-7933.00.patch
  against trunk revision 2681ed9.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  org.apache.hadoop.hdfs.TestClientReportBadBlock

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/9905//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9905//console

This message is automatically generated.

 fsck should also report decommissioning replicas. 
 --

 Key: HDFS-7933
 URL: https://issues.apache.org/jira/browse/HDFS-7933
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Reporter: Jitendra Nath Pandey
Assignee: Xiaoyu Yao
 Attachments: HDFS-7933.00.patch


 Fsck doesn't count replicas that are on decommissioning nodes. If a block has 
 all replicas on the decommissioning nodes, it will be marked as missing, 
 which is alarming for the admins, although the system will replicate them 
 before nodes are decommissioned.
 Fsck output should also show decommissioning replicas along with the live 
 replicas.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7937) Erasure Coding: INodeFile quota computation unit tests

2015-03-16 Thread Kai Sasaki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7937?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kai Sasaki updated HDFS-7937:
-
Status: Patch Available  (was: Open)

[~jingzhao] I submitted an initial patch including clean up which is mentioned 
above. Can you review it? Thank you.

 Erasure Coding: INodeFile quota computation unit tests
 --

 Key: HDFS-7937
 URL: https://issues.apache.org/jira/browse/HDFS-7937
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Sasaki
Assignee: Kai Sasaki
Priority: Minor
 Attachments: HDFS-7937.1.patch


 Unit test for [HDFS-7826|https://issues.apache.org/jira/browse/HDFS-7826]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7937) Erasure Coding: INodeFile quota computation unit tests

2015-03-16 Thread Kai Sasaki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7937?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kai Sasaki updated HDFS-7937:
-
Attachment: HDFS-7937.1.patch

 Erasure Coding: INodeFile quota computation unit tests
 --

 Key: HDFS-7937
 URL: https://issues.apache.org/jira/browse/HDFS-7937
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Sasaki
Assignee: Kai Sasaki
Priority: Minor
 Attachments: HDFS-7937.1.patch


 Unit test for [HDFS-7826|https://issues.apache.org/jira/browse/HDFS-7826]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7054) Make DFSOutputStream tracing more fine-grained

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7054?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364448#comment-14364448
 ] 

Hadoop QA commented on HDFS-7054:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12704958/HDFS-7054.002.patch
  against trunk revision 5608520.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

  {color:red}-1 javac{color}.  The applied patch generated 1155 javac 
compiler warnings (more than the trunk's current 1152 warnings).

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:red}-1 findbugs{color}.  The patch appears to introduce 5 new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager:

  
org.apache.hadoop.yarn.server.resourcemanager.TestRMAdminService
  
org.apache.hadoop.yarn.server.resourcemanager.TestClientRMTokens

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/9909//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HDFS-Build/9909//artifact/patchprocess/newPatchFindbugsWarningshadoop-yarn-server-resourcemanager.html
Javac warnings: 
https://builds.apache.org/job/PreCommit-HDFS-Build/9909//artifact/patchprocess/diffJavacWarnings.txt
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9909//console

This message is automatically generated.

 Make DFSOutputStream tracing more fine-grained
 --

 Key: HDFS-7054
 URL: https://issues.apache.org/jira/browse/HDFS-7054
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Affects Versions: 2.6.0
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
 Attachments: HDFS-7054.001.patch, HDFS-7054.002.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7616) Change FSImage to support BlockGroup

2015-03-16 Thread Takuya Fukudome (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7616?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364496#comment-14364496
 ] 

Takuya Fukudome commented on HDFS-7616:
---

Hi [~szetszwo],

I am interested in working on this task. Could you assign it to me? Thank you.

 Change FSImage to support BlockGroup
 

 Key: HDFS-7616
 URL: https://issues.apache.org/jira/browse/HDFS-7616
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Reporter: Tsz Wo Nicholas Sze
Assignee: Tsz Wo Nicholas Sze

 We need to change FSImage for support BlockGroup and other new structures 
 introduced for EC.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7932) Speed up the shutdown of datanode during rolling upgrade

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7932?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364236#comment-14364236
 ] 

Hadoop QA commented on HDFS-7932:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12704863/HDFS-7932.patch
  against trunk revision ce5de93.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-hdfs-project/hadoop-hdfs.

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/9904//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9904//console

This message is automatically generated.

 Speed up the shutdown of datanode during rolling upgrade
 

 Key: HDFS-7932
 URL: https://issues.apache.org/jira/browse/HDFS-7932
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Kihwal Lee
Assignee: Kihwal Lee
 Attachments: HDFS-7932.patch, HDFS-7932.patch


 Datanode normally exits in 3 seconds after receiving {{shutdownDatanode}} 
 command. However, sometimes it doesn't, especially when the IO is busy. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7054) Make DFSOutputStream tracing more fine-grained

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7054?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364300#comment-14364300
 ] 

Hadoop QA commented on HDFS-7054:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12704936/HDFS-7054.001.patch
  against trunk revision 5608520.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:red}-1 javac{color:red}.  The patch appears to cause the build to 
fail.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9907//console

This message is automatically generated.

 Make DFSOutputStream tracing more fine-grained
 --

 Key: HDFS-7054
 URL: https://issues.apache.org/jira/browse/HDFS-7054
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Affects Versions: 2.6.0
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
 Attachments: HDFS-7054.001.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7891) A block placement policy with best fault tolerance

2015-03-16 Thread Kai Zheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7891?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kai Zheng updated HDFS-7891:

Summary: A block placement policy with best fault tolerance  (was: a block 
placement policy with best fault tolerance)

 A block placement policy with best fault tolerance
 --

 Key: HDFS-7891
 URL: https://issues.apache.org/jira/browse/HDFS-7891
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su
 Attachments: HDFS-7891.002.patch, HDFS-7891.patch


 a block placement policy tries its best to place replicas to most racks.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7912) Erasure Coding: track BlockInfo instead of Block in UnderReplicatedBlocks and PendingReplicationBlocks

2015-03-16 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7912?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364334#comment-14364334
 ] 

Jing Zhao commented on HDFS-7912:
-

Thanks for the review, Zhe!

bq. My only concern is that it adds some getStoredBlock calls

The patch adds {{getStoredBlock}} call in mainly two places: 
{{removeStoredBlock}} and {{addBlock}}. The {{addBlock}} function is currently 
called to handle a RECEIVED_BLOCK msg in incremental block report, and it is 
not happening when handling a full block report. Thus I guess here we will not 
bring too much extra overhead to the NN.

bq. do we need storedBlock in the following code

Here my thought is that we should use BlockInfo for {{countNodes}}, since the 
function tries to retrieve the storages from the blocksMap for the given block. 
And with this we need to make the change in NameNodeFsck.

 Erasure Coding: track BlockInfo instead of Block in UnderReplicatedBlocks and 
 PendingReplicationBlocks
 --

 Key: HDFS-7912
 URL: https://issues.apache.org/jira/browse/HDFS-7912
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jing Zhao
Assignee: Jing Zhao
 Attachments: HDFS-7912.000.patch


 Now with striped blocks and the design that uses a single BlockInfoStriped 
 object to track all the corresponding blocks, we need to clearly distinguish 
 the type Block and BlockInfo in BlockManager. Specifically, data structures 
 like {{UnderReplicatedBlocks}} and {{PendingReplicationBlocks}} should track 
 BlockInfo instead of Block in order to support striped block recovery.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7937) Erasure Coding: INodeFile quota computation unit tests

2015-03-16 Thread Kai Sasaki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7937?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kai Sasaki updated HDFS-7937:
-
Attachment: HDFS-7937.1.patch

 Erasure Coding: INodeFile quota computation unit tests
 --

 Key: HDFS-7937
 URL: https://issues.apache.org/jira/browse/HDFS-7937
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Sasaki
Assignee: Kai Sasaki
Priority: Minor
 Attachments: HDFS-7937.1.patch


 Unit test for [HDFS-7826|https://issues.apache.org/jira/browse/HDFS-7826]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7936) Erasure coding: resolving conflicts when merging with HDFS-7903 and HDFS-7435

2015-03-16 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7936?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364239#comment-14364239
 ] 

Jing Zhao commented on HDFS-7936:
-

Yeah, but we do not have the class BlockInfo in trunk currently.. 

 Erasure coding: resolving conflicts when merging with HDFS-7903 and HDFS-7435
 -

 Key: HDFS-7936
 URL: https://issues.apache.org/jira/browse/HDFS-7936
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Zhe Zhang
Assignee: Zhe Zhang
 Fix For: HDFS-7285

 Attachments: HDFS-7936-001.patch, HDFS-7936-002.patch


 A few non-trivial conflicts were found when merging trunk changes. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7915) The DataNode can sometimes allocate a ShortCircuitShm slot and fail to tell the DFSClient about it because of a network error

2015-03-16 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364392#comment-14364392
 ] 

Colin Patrick McCabe commented on HDFS-7915:


Good catch.  I committed to branch-2 but not branch-2.7, which was incorrect.  
Fixed.

 The DataNode can sometimes allocate a ShortCircuitShm slot and fail to tell 
 the DFSClient about it because of a network error
 -

 Key: HDFS-7915
 URL: https://issues.apache.org/jira/browse/HDFS-7915
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.7.0
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
 Fix For: 2.7.0

 Attachments: HDFS-7915.001.patch, HDFS-7915.002.patch, 
 HDFS-7915.004.patch, HDFS-7915.005.patch, HDFS-7915.006.patch


 The DataNode can sometimes allocate a ShortCircuitShm slot and fail to tell 
 the DFSClient about it because of a network error.  In 
 {{DataXceiver#requestShortCircuitFds}}, the DataNode can succeed at the first 
 part (mark the slot as used) and fail at the second part (tell the DFSClient 
 what it did). The try block for unregistering the slot only covers a 
 failure in the first part, not the second part. In this way, a divergence can 
 form between the views of which slots are allocated on DFSClient and on 
 server.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7929) inotify unable fetch pre-upgrade edit log segments once upgrade starts

2015-03-16 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364411#comment-14364411
 ] 

Colin Patrick McCabe commented on HDFS-7929:


The approach looks reasonable here.

{code}
131 File[] files = tmpDir.listFiles();
132 if (files != null) {
{code}

We should use {{IOUtils#listDirectory}} here.  That function will throw an 
exception if there is an IO error; in contrast, {{File#listFiles}} will 
silently hide the error and just return null.

I am +1 pending that (and getting a good jenkins run)

We might want to file a follow-up JIRA for looking into whether we need any 
changes in {{BackupJournalManager}} or {{BookKeeperJournalManager}}.  Of 
course, I think in most cases people configure local file journal managers to 
go along with those journal managers, so it probably won't be an issue for most 
systems.

 inotify unable fetch pre-upgrade edit log segments once upgrade starts
 --

 Key: HDFS-7929
 URL: https://issues.apache.org/jira/browse/HDFS-7929
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Zhe Zhang
Assignee: Zhe Zhang
 Attachments: HDFS-7929-000.patch, HDFS-7929-001.patch


 inotify is often used to periodically poll HDFS events. However, once an HDFS 
 upgrade has started, edit logs are moved to /previous on the NN, which is not 
 accessible. Moreover, once the upgrade is finalized /previous is currently 
 lost forever.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7838) Expose truncate API for libhdfs

2015-03-16 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364455#comment-14364455
 ] 

Colin Patrick McCabe commented on HDFS-7838:


+1.  Thanks, Yi.

 Expose truncate API for libhdfs
 ---

 Key: HDFS-7838
 URL: https://issues.apache.org/jira/browse/HDFS-7838
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Affects Versions: 2.7.0
Reporter: Yi Liu
Assignee: Yi Liu
 Attachments: HDFS-7838.001.patch, HDFS-7838.002.patch, 
 HDFS-7838.003.patch, HDFS-7838.004.patch


 It's good to expose truncate in libhdfs.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7837) Erasure Coding: allocate and persist striped blocks in NameNode

2015-03-16 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7837?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363837#comment-14363837
 ] 

Zhe Zhang commented on HDFS-7837:
-

[~jingzhao] I rebased trunk changes today and manually fixed some conflicts. I 
also squashed the 2 HDFS-7837 commits. Please take a look at the [commit | 
https://git-wip-us.apache.org/repos/asf?p=hadoop.git;a=commitdiff;h=bc962947a19ed0a9ee130f62e819b65a64a8fb18]
 and see if it looks OK. Thanks!

 Erasure Coding: allocate and persist striped blocks in NameNode
 ---

 Key: HDFS-7837
 URL: https://issues.apache.org/jira/browse/HDFS-7837
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jing Zhao
Assignee: Jing Zhao
 Fix For: HDFS-7285

 Attachments: HDFS-7837.000.patch, HDFS-7837.001.patch


 Try to finish the remaining work from HDFS-7339 (except the 
 ClientProtocol/DFSClient part):
 # Allow FSNamesystem#getAdditionalBlock to create striped blocks and persist 
 striped blocks to editlog
 # Update FSImage for max allocated striped block ID
 # Update the block commit/complete logic in BlockManager



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7932) Speed up the shutdown of datanode during rolling upgrade

2015-03-16 Thread Kihwal Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7932?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kihwal Lee updated HDFS-7932:
-
Attachment: HDFS-7932.patch

Making it wait shorter for the dataXfer thread group to exit. (From 2.5 seconds 
to 1 second)

 Speed up the shutdown of datanode during rolling upgrade
 

 Key: HDFS-7932
 URL: https://issues.apache.org/jira/browse/HDFS-7932
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Kihwal Lee
Assignee: Kihwal Lee
 Attachments: HDFS-7932.patch, HDFS-7932.patch


 Datanode normally exits in 3 seconds after receiving {{shutdownDatanode}} 
 command. However, sometimes it doesn't, especially when the IO is busy. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7935) Support multi-homed networks when Kerberos security is enabled

2015-03-16 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363175#comment-14363175
 ] 

Kihwal Lee commented on HDFS-7935:
--

Have you tried HADOOP-9789?  For example, you can configure 
dfs.namenode.kerberos.principal.pattern on the client side to make it accept 
a SPN that is different from the connecting namenode address.

 Support multi-homed networks when Kerberos security is enabled
 --

 Key: HDFS-7935
 URL: https://issues.apache.org/jira/browse/HDFS-7935
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Arun Suresh
Assignee: Arun Suresh

 Currently, during SASL negotiation stage between ipc Client and Server, The 
 server sends only a single serviceId (curresponding to a single principal) to 
 the client. This is the principal the the server process is logged in as 
 during startup.
 It is possible that in a multi-homed network, the server might be associated 
 with more than one principal, and thus severs must provide the clients all 
 possible principals it can use to connect to.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7060) Contentions of the monitor of FsDatasetImpl block DN's heartbeat

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7060?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363215#comment-14363215
 ] 

Hadoop QA commented on HDFS-7060:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12704754/HDFS-7060.001.patch
  against trunk revision 3ff1ba2.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  
org.apache.hadoop.hdfs.server.namenode.snapshot.TestUpdatePipelineWithSnapshots

  The following test timeouts occurred in 
hadoop-hdfs-project/hadoop-hdfs:

org.apache.hadoop.hdfs.server.blockmanagement.TestDatanodeManager

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/9896//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9896//console

This message is automatically generated.

 Contentions of the monitor of FsDatasetImpl block DN's heartbeat
 

 Key: HDFS-7060
 URL: https://issues.apache.org/jira/browse/HDFS-7060
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Xinwei Qin 
 Attachments: HDFS-7060.000.patch, HDFS-7060.001.patch


 We're seeing the heartbeat is blocked by the monitor of {{FsDatasetImpl}} 
 when the DN is under heavy load of writes:
 {noformat}
java.lang.Thread.State: BLOCKED (on object monitor)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsVolumeImpl.getDfsUsed(FsVolumeImpl.java:115)
 - waiting to lock 0x000780304fb8 (a 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.getStorageReports(FsDatasetImpl.java:91)
 - locked 0x000780612fd8 (a java.lang.Object)
 at 
 org.apache.hadoop.hdfs.server.datanode.BPServiceActor.sendHeartBeat(BPServiceActor.java:563)
 at 
 org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:668)
 at 
 org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:827)
 at java.lang.Thread.run(Thread.java:744)
java.lang.Thread.State: BLOCKED (on object monitor)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:743)
 - waiting to lock 0x000780304fb8 (a 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:60)
 at 
 org.apache.hadoop.hdfs.server.datanode.BlockReceiver.init(BlockReceiver.java:169)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:621)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:124)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:71)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:232)
 at java.lang.Thread.run(Thread.java:744)
java.lang.Thread.State: RUNNABLE
 at java.io.UnixFileSystem.createFileExclusively(Native Method)
 at java.io.File.createNewFile(File.java:1006)
 at 
 org.apache.hadoop.hdfs.server.datanode.DatanodeUtil.createTmpFile(DatanodeUtil.java:59)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice.createRbwFile(BlockPoolSlice.java:244)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsVolumeImpl.createRbwFile(FsVolumeImpl.java:195)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:753)
 - locked 0x000780304fb8 (a 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl)
  

[jira] [Commented] (HDFS-7854) Separate class DataStreamer out of DFSOutputStream

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7854?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363169#comment-14363169
 ] 

Hadoop QA commented on HDFS-7854:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12704748/HDFS-7854-004-duplicate2.patch
  against trunk revision 3ff1ba2.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:red}-1 findbugs{color}.  The patch appears to introduce 1 new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  org.apache.hadoop.hdfs.server.namenode.ha.TestRetryCacheWithHA

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/9895//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-HDFS-Build/9895//artifact/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9895//console

This message is automatically generated.

 Separate class DataStreamer out of DFSOutputStream
 --

 Key: HDFS-7854
 URL: https://issues.apache.org/jira/browse/HDFS-7854
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Li Bo
Assignee: Li Bo
 Attachments: HDFS-7854-001.patch, HDFS-7854-002.patch, 
 HDFS-7854-003.patch, HDFS-7854-004-duplicate.patch, 
 HDFS-7854-004-duplicate2.patch, HDFS-7854-004.patch


 This sub task separate DataStreamer from DFSOutputStream. New DataStreamer 
 will accept packets and write them to remote datanodes.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7847) Modify NNThroughputBenchmark to be able to operate on a remote NameNode

2015-03-16 Thread Charles Lamb (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7847?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Lamb updated HDFS-7847:
---
Attachment: HDFS-7847.002.patch

@cmccabe, @stack, thanks for the review!

bq. DFSClient.java: this change adds three new fields to DFSClient. But they 
only seem to be used by unit tests. It seems like we should just put these 
inside the unit test(s) that are using these-- if necessary, by adding a helper 
method. There's no reason to add more fields to DFSClient. Also remember that 
when using FileContext, we create new DFSClients all the time.

Good point. I've left the existing {code}ClientProtocol namenode{code} field 
alone. The other 3 proxies are created on-demand by their getters. That means 
no change in DFSClient instance size.

bq. It seems kind of odd to have NameNodeProxies#createProxy create a proxy to 
the datanode.

It's actually a proxy to the NN for the DatanodeProtocol. That's the same 
protocol that the DN uses to speak with the NN when it's sending (among other 
things) block reports. But with some ideas from @stack, I got rid of the 
changes to NameNodeProxies.

bq. Of course the NameNode may or may not be remote here. It seems like --nnuri 
or just --namenode or something like that would be more descriptive.

Yeah, I agree. I changed it to -namenode.

bq. Instead of this boilerplate, just use StringUtils#popOptionWithArgument.

Changed. I was just trying to match the existing code, but the using 
StringUtils is for the better. 

{code}
-  replication, BLOCK_SIZE, null);
+  replication, BLOCK_SIZE, CryptoProtocolVersion.supported());
{code}

bq. This fix is a little bit separate, right? I suppose we can do it in this 
JIRA, though.

Without this, the relevant PBHelper.convert code throws NPE on the 
supportVersions arg.


 Modify NNThroughputBenchmark to be able to operate on a remote NameNode
 ---

 Key: HDFS-7847
 URL: https://issues.apache.org/jira/browse/HDFS-7847
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: HDFS-7836
Reporter: Colin Patrick McCabe
Assignee: Charles Lamb
 Attachments: HDFS-7847.000.patch, HDFS-7847.001.patch, 
 HDFS-7847.002.patch, make_blocks.tar.gz


 Modify NNThroughputBenchmark to be able to operate on a NN that is not in 
 process. A followon Jira will modify it some more to allow quantifying native 
 and java heap sizes, and some latency numbers.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-5356) MiniDFSCluster shoud close all open FileSystems when shutdown()

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5356?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363045#comment-14363045
 ] 

Hadoop QA commented on HDFS-5356:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12704667/HDFS-5356-8.patch
  against trunk revision 3ff1ba2.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 3 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  org.apache.hadoop.hdfs.server.namenode.ha.TestRetryCacheWithHA

  The following test timeouts occurred in 
hadoop-hdfs-project/hadoop-hdfs:

org.apache.hadoop.hdfs.server.blockmanagement.TestDatanodeManager

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/9892//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9892//console

This message is automatically generated.

 MiniDFSCluster shoud close all open FileSystems when shutdown()
 ---

 Key: HDFS-5356
 URL: https://issues.apache.org/jira/browse/HDFS-5356
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 3.0.0, 2.2.0
Reporter: haosdent
Assignee: Rakesh R
Priority: Critical
 Attachments: HDFS-5356-1.patch, HDFS-5356-2.patch, HDFS-5356-3.patch, 
 HDFS-5356-4.patch, HDFS-5356-5.patch, HDFS-5356-6.patch, HDFS-5356-7.patch, 
 HDFS-5356-8.patch, HDFS-5356.patch


 After add some metrics functions to DFSClient, I found that some unit tests 
 relates to metrics are failed. Because MiniDFSCluster are never close open 
 FileSystems, DFSClients are alive after MiniDFSCluster shutdown(). The 
 metrics of DFSClients in DefaultMetricsSystem are still exist and this make 
 other unit tests failed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-3325) When configuring dfs.namenode.safemode.threshold-pct to a value greater or equal to 1 there is mismatch in the UI report

2015-03-16 Thread J.Andreina (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3325?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

J.Andreina updated HDFS-3325:
-
Attachment: HDFS-3325.2.patch

Hi Vinayakumar B ,

Attached a patch with updated testcases. 
Please review. 

 When configuring dfs.namenode.safemode.threshold-pct to a value greater or 
 equal to 1 there is mismatch in the UI report
 --

 Key: HDFS-3325
 URL: https://issues.apache.org/jira/browse/HDFS-3325
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.0.0-alpha
Reporter: J.Andreina
Assignee: J.Andreina
Priority: Minor
 Attachments: HDFS-3325.1.patch, HDFS-3325.2.patch


 When dfs.namenode.safemode.threshold-pct is configured to n
 Namenode will be in safemode until n percentage of blocks that should satisfy 
 the minimal replication requirement defined by 
 dfs.namenode.replication.min is reported to namenode
 But in UI it displays that n percentage of total blocks + 1 blocks  are 
 additionally needed
 to come out of the safemode
 Scenario 1:
 
 Configurations:
 dfs.namenode.safemode.threshold-pct = 2
 dfs.replication = 2
 dfs.namenode.replication.min =2
 Step 1: Start NN,DN1,DN2
 Step 2: Write a file a.txt which has got 167 blocks
 step 3: Stop NN,DN1,DN2
 Step 4: start NN
 In UI report the Number of blocks needed to come out of safemode and number 
 of blocks actually present is different.
 {noformat}
 Cluster Summary
 Security is OFF 
 Safe mode is ON. The reported blocks 0 needs additional 335 blocks to reach 
 the threshold 2. of total blocks 167. Safe mode will be turned off 
 automatically.
 2 files and directories, 167 blocks = 169 total.
 Heap Memory used 57.05 MB is 2% of Commited Heap Memory 2 GB. Max Heap Memory 
 is 2 GB. 
 Non Heap Memory used 23.37 MB is 17% of Commited Non Heap Memory 130.44 MB. 
 Max Non Heap Memory is 176 MB.{noformat}
 Scenario 2:
 ===
 Configurations:
 dfs.namenode.safemode.threshold-pct = 1
 dfs.replication = 2
 dfs.namenode.replication.min =2
 Step 1: Start NN,DN1,DN2
 Step 2: Write a file a.txt which has got 167 blocks
 step 3: Stop NN,DN1,DN2
 Step 4: start NN
 In UI report the Number of blocks needed to come out of safemode and number 
 of blocks actually present is different
 {noformat}
 Cluster Summary
 Security is OFF 
 Safe mode is ON. The reported blocks 0 needs additional 168 blocks to reach 
 the threshold 1. of total blocks 167. Safe mode will be turned off 
 automatically.
 2 files and directories, 167 blocks = 169 total.
 Heap Memory used 56.2 MB is 2% of Commited Heap Memory 2 GB. Max Heap Memory 
 is 2 GB. 
 Non Heap Memory used 23.37 MB is 17% of Commited Non Heap Memory 130.44 MB. 
 Max Non Heap Memory is 176 MB.{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7838) Expose truncate API for libhdfs

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363126#comment-14363126
 ] 

Hadoop QA commented on HDFS-7838:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12704742/HDFS-7838.004.patch
  against trunk revision 3ff1ba2.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  
org.apache.hadoop.hdfs.server.blockmanagement.TestDatanodeManager
  org.apache.hadoop.hdfs.server.balancer.TestBalancer

  The following test timeouts occurred in 
hadoop-hdfs-project/hadoop-hdfs:

org.apache.hadoop.hdfs.server.namenode.TestNameNodeXAttr

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/9894//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9894//console

This message is automatically generated.

 Expose truncate API for libhdfs
 ---

 Key: HDFS-7838
 URL: https://issues.apache.org/jira/browse/HDFS-7838
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Affects Versions: 2.7.0
Reporter: Yi Liu
Assignee: Yi Liu
 Attachments: HDFS-7838.001.patch, HDFS-7838.002.patch, 
 HDFS-7838.003.patch, HDFS-7838.004.patch


 It's good to expose truncate in libhdfs.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-3325) When configuring dfs.namenode.safemode.threshold-pct to a value greater or equal to 1 there is mismatch in the UI report

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3325?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363065#comment-14363065
 ] 

Hadoop QA commented on HDFS-3325:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12704758/HDFS-3325.2.patch
  against trunk revision 3ff1ba2.

{color:red}-1 patch{color}.  Trunk compilation may be broken.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9897//console

This message is automatically generated.

 When configuring dfs.namenode.safemode.threshold-pct to a value greater or 
 equal to 1 there is mismatch in the UI report
 --

 Key: HDFS-3325
 URL: https://issues.apache.org/jira/browse/HDFS-3325
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.0.0-alpha
Reporter: J.Andreina
Assignee: J.Andreina
Priority: Minor
 Attachments: HDFS-3325.1.patch, HDFS-3325.2.patch


 When dfs.namenode.safemode.threshold-pct is configured to n
 Namenode will be in safemode until n percentage of blocks that should satisfy 
 the minimal replication requirement defined by 
 dfs.namenode.replication.min is reported to namenode
 But in UI it displays that n percentage of total blocks + 1 blocks  are 
 additionally needed
 to come out of the safemode
 Scenario 1:
 
 Configurations:
 dfs.namenode.safemode.threshold-pct = 2
 dfs.replication = 2
 dfs.namenode.replication.min =2
 Step 1: Start NN,DN1,DN2
 Step 2: Write a file a.txt which has got 167 blocks
 step 3: Stop NN,DN1,DN2
 Step 4: start NN
 In UI report the Number of blocks needed to come out of safemode and number 
 of blocks actually present is different.
 {noformat}
 Cluster Summary
 Security is OFF 
 Safe mode is ON. The reported blocks 0 needs additional 335 blocks to reach 
 the threshold 2. of total blocks 167. Safe mode will be turned off 
 automatically.
 2 files and directories, 167 blocks = 169 total.
 Heap Memory used 57.05 MB is 2% of Commited Heap Memory 2 GB. Max Heap Memory 
 is 2 GB. 
 Non Heap Memory used 23.37 MB is 17% of Commited Non Heap Memory 130.44 MB. 
 Max Non Heap Memory is 176 MB.{noformat}
 Scenario 2:
 ===
 Configurations:
 dfs.namenode.safemode.threshold-pct = 1
 dfs.replication = 2
 dfs.namenode.replication.min =2
 Step 1: Start NN,DN1,DN2
 Step 2: Write a file a.txt which has got 167 blocks
 step 3: Stop NN,DN1,DN2
 Step 4: start NN
 In UI report the Number of blocks needed to come out of safemode and number 
 of blocks actually present is different
 {noformat}
 Cluster Summary
 Security is OFF 
 Safe mode is ON. The reported blocks 0 needs additional 168 blocks to reach 
 the threshold 1. of total blocks 167. Safe mode will be turned off 
 automatically.
 2 files and directories, 167 blocks = 169 total.
 Heap Memory used 56.2 MB is 2% of Commited Heap Memory 2 GB. Max Heap Memory 
 is 2 GB. 
 Non Heap Memory used 23.37 MB is 17% of Commited Non Heap Memory 130.44 MB. 
 Max Non Heap Memory is 176 MB.{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7060) Contentions of the monitor of FsDatasetImpl block DN's heartbeat

2015-03-16 Thread Xinwei Qin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7060?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xinwei Qin  updated HDFS-7060:
--
Attachment: HDFS-7060.001.patch

 Contentions of the monitor of FsDatasetImpl block DN's heartbeat
 

 Key: HDFS-7060
 URL: https://issues.apache.org/jira/browse/HDFS-7060
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Xinwei Qin 
 Attachments: HDFS-7060.000.patch, HDFS-7060.001.patch


 We're seeing the heartbeat is blocked by the monitor of {{FsDatasetImpl}} 
 when the DN is under heavy load of writes:
 {noformat}
java.lang.Thread.State: BLOCKED (on object monitor)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsVolumeImpl.getDfsUsed(FsVolumeImpl.java:115)
 - waiting to lock 0x000780304fb8 (a 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.getStorageReports(FsDatasetImpl.java:91)
 - locked 0x000780612fd8 (a java.lang.Object)
 at 
 org.apache.hadoop.hdfs.server.datanode.BPServiceActor.sendHeartBeat(BPServiceActor.java:563)
 at 
 org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:668)
 at 
 org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:827)
 at java.lang.Thread.run(Thread.java:744)
java.lang.Thread.State: BLOCKED (on object monitor)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:743)
 - waiting to lock 0x000780304fb8 (a 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:60)
 at 
 org.apache.hadoop.hdfs.server.datanode.BlockReceiver.init(BlockReceiver.java:169)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:621)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:124)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:71)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:232)
 at java.lang.Thread.run(Thread.java:744)
java.lang.Thread.State: RUNNABLE
 at java.io.UnixFileSystem.createFileExclusively(Native Method)
 at java.io.File.createNewFile(File.java:1006)
 at 
 org.apache.hadoop.hdfs.server.datanode.DatanodeUtil.createTmpFile(DatanodeUtil.java:59)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice.createRbwFile(BlockPoolSlice.java:244)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsVolumeImpl.createRbwFile(FsVolumeImpl.java:195)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:753)
 - locked 0x000780304fb8 (a 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl)
 at 
 org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.createRbw(FsDatasetImpl.java:60)
 at 
 org.apache.hadoop.hdfs.server.datanode.BlockReceiver.init(BlockReceiver.java:169)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:621)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:124)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:71)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:232)
 at java.lang.Thread.run(Thread.java:744)
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7267) TestBalancer#testUnknownDatanode occasionally fails in trunk

2015-03-16 Thread Walter Su (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7267?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363070#comment-14363070
 ] 

Walter Su commented on HDFS-7267:
-

set block report to 1s is not a decent solution. I can't think up a better 
solution.

 TestBalancer#testUnknownDatanode occasionally fails in trunk
 

 Key: HDFS-7267
 URL: https://issues.apache.org/jira/browse/HDFS-7267
 Project: Hadoop HDFS
  Issue Type: Test
Reporter: Ted Yu
Priority: Minor
 Attachments: testUnknownDatanode-failed-log.html


 In build #1907 (https://builds.apache.org/job/Hadoop-Hdfs-trunk/1907/):
 {code}
 REGRESSION:  
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.testUnknownDatanode
 Error Message:
 expected:0 but was:-3
 Stack Trace:
 java.lang.AssertionError: expected:0 but was:-3
 at org.junit.Assert.fail(Assert.java:88)
 at org.junit.Assert.failNotEquals(Assert.java:743)
 at org.junit.Assert.assertEquals(Assert.java:118)
 at org.junit.Assert.assertEquals(Assert.java:555)
 at org.junit.Assert.assertEquals(Assert.java:542)
 at 
 org.apache.hadoop.hdfs.server.balancer.TestBalancer.testUnknownDatanode(TestBalancer.java:737)
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7934) During Rolling upgrade rollback ,standby namenode startup fails.

2015-03-16 Thread Vinayakumar B (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14362978#comment-14362978
 ] 

Vinayakumar B commented on HDFS-7934:
-

I think the problematic area is below code block in Fsimage#loadFsImage(..)
{code}  // For rollback in rolling upgrade, we need to set the 
toAtLeastTxId to
  // the txid right before the upgrade marker.  
  long toAtLeastTxId = editLog.isOpenForWrite() ? inspector
  .getMaxSeenTxId() : 0;
  if (rollingRollback) {
// note that the first image in imageFiles is the special checkpoint
// for the rolling upgrade
toAtLeastTxId = imageFiles.get(0).getCheckpointTxId() + 2;
  }{code}
In Case of rollingRollback, there is nothing read from edits streams. So 
setting {{toAtLeastTxId = imageFiles.get(0).getCheckpointTxId() + 2;}} is not 
required. removing this line will solve the problem IMO.

Any thoughts?

 During Rolling upgrade rollback ,standby namenode startup fails.
 

 Key: HDFS-7934
 URL: https://issues.apache.org/jira/browse/HDFS-7934
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: J.Andreina
Assignee: J.Andreina
Priority: Critical

 During Rolling upgrade rollback , standby namenode startup fails , while 
 loading edits and when  there is no local copy of edits created after upgrade 
 ( which is already been removed  by Active Namenode from journal manager and 
 from Active's local). 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7930) commitBlockSynchronization() does not remove locations

2015-03-16 Thread Yi Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7930?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yi Liu updated HDFS-7930:
-
Status: Patch Available  (was: Open)

 commitBlockSynchronization() does not remove locations
 --

 Key: HDFS-7930
 URL: https://issues.apache.org/jira/browse/HDFS-7930
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.7.0
Reporter: Konstantin Shvachko
Assignee: Yi Liu
Priority: Blocker
 Attachments: HDFS-7930.001.patch


 When {{commitBlockSynchronization()}} has less {{newTargets}} than in the 
 original block it does not remove unconfirmed locations. This results in that 
 the the block stores locations of different lengths or genStamp (corrupt).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-3325) When configuring dfs.namenode.safemode.threshold-pct to a value greater or equal to 1 there is mismatch in the UI report

2015-03-16 Thread J.Andreina (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3325?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363069#comment-14363069
 ] 

J.Andreina commented on HDFS-3325:
--

Failures are not related to this patch

 When configuring dfs.namenode.safemode.threshold-pct to a value greater or 
 equal to 1 there is mismatch in the UI report
 --

 Key: HDFS-3325
 URL: https://issues.apache.org/jira/browse/HDFS-3325
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.0.0-alpha
Reporter: J.Andreina
Assignee: J.Andreina
Priority: Minor
 Attachments: HDFS-3325.1.patch, HDFS-3325.2.patch


 When dfs.namenode.safemode.threshold-pct is configured to n
 Namenode will be in safemode until n percentage of blocks that should satisfy 
 the minimal replication requirement defined by 
 dfs.namenode.replication.min is reported to namenode
 But in UI it displays that n percentage of total blocks + 1 blocks  are 
 additionally needed
 to come out of the safemode
 Scenario 1:
 
 Configurations:
 dfs.namenode.safemode.threshold-pct = 2
 dfs.replication = 2
 dfs.namenode.replication.min =2
 Step 1: Start NN,DN1,DN2
 Step 2: Write a file a.txt which has got 167 blocks
 step 3: Stop NN,DN1,DN2
 Step 4: start NN
 In UI report the Number of blocks needed to come out of safemode and number 
 of blocks actually present is different.
 {noformat}
 Cluster Summary
 Security is OFF 
 Safe mode is ON. The reported blocks 0 needs additional 335 blocks to reach 
 the threshold 2. of total blocks 167. Safe mode will be turned off 
 automatically.
 2 files and directories, 167 blocks = 169 total.
 Heap Memory used 57.05 MB is 2% of Commited Heap Memory 2 GB. Max Heap Memory 
 is 2 GB. 
 Non Heap Memory used 23.37 MB is 17% of Commited Non Heap Memory 130.44 MB. 
 Max Non Heap Memory is 176 MB.{noformat}
 Scenario 2:
 ===
 Configurations:
 dfs.namenode.safemode.threshold-pct = 1
 dfs.replication = 2
 dfs.namenode.replication.min =2
 Step 1: Start NN,DN1,DN2
 Step 2: Write a file a.txt which has got 167 blocks
 step 3: Stop NN,DN1,DN2
 Step 4: start NN
 In UI report the Number of blocks needed to come out of safemode and number 
 of blocks actually present is different
 {noformat}
 Cluster Summary
 Security is OFF 
 Safe mode is ON. The reported blocks 0 needs additional 168 blocks to reach 
 the threshold 1. of total blocks 167. Safe mode will be turned off 
 automatically.
 2 files and directories, 167 blocks = 169 total.
 Heap Memory used 56.2 MB is 2% of Commited Heap Memory 2 GB. Max Heap Memory 
 is 2 GB. 
 Non Heap Memory used 23.37 MB is 17% of Commited Non Heap Memory 130.44 MB. 
 Max Non Heap Memory is 176 MB.{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-3745) fsck prints that it's using KSSL even when it's in fact using SPNEGO for authentication

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3745?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J updated HDFS-3745:
--
Target Version/s:   (was: 2.0.2-alpha)

 fsck prints that it's using KSSL even when it's in fact using SPNEGO for 
 authentication
 ---

 Key: HDFS-3745
 URL: https://issues.apache.org/jira/browse/HDFS-3745
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client, security
Affects Versions: 1.2.0, 2.0.0-alpha
Reporter: Aaron T. Myers
Priority: Trivial
  Labels: newbie

 In branch-2 (which exclusively uses SPNEGO for HTTP authentication) and in 
 branch-1 (which can optionally use SPNEGO for HTTP authentication), running 
 fsck will print the following, which isn't quite right:
 {quote}
 FSCK started by hdfs (auth:KERBEROS_SSL) from...
 {quote}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-3745) fsck prints that it's using KSSL even when it's in fact using SPNEGO for authentication

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3745?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J updated HDFS-3745:
--
Affects Version/s: (was: 1.2.0)

 fsck prints that it's using KSSL even when it's in fact using SPNEGO for 
 authentication
 ---

 Key: HDFS-3745
 URL: https://issues.apache.org/jira/browse/HDFS-3745
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client, security
Affects Versions: 2.0.0-alpha
Reporter: Aaron T. Myers
Priority: Trivial
  Labels: newbie

 In branch-2 (which exclusively uses SPNEGO for HTTP authentication) and in 
 branch-1 (which can optionally use SPNEGO for HTTP authentication), running 
 fsck will print the following, which isn't quite right:
 {quote}
 FSCK started by hdfs (auth:KERBEROS_SSL) from...
 {quote}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-4829) Make the fs-shell tail command behave more like Unix tail command

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4829?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J updated HDFS-4829:
--
Labels: newbie  (was: )

 Make the fs-shell tail command behave more like Unix tail command
 -

 Key: HDFS-4829
 URL: https://issues.apache.org/jira/browse/HDFS-4829
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.0.0-alpha
Reporter: Todd Grayson
Priority: Minor
  Labels: newbie

 Strange behavior of the hadoop fs -tail command - its default for output 
 seems to be 9 lines of output vs 10 lines of output in the OS version of the 
 command (minor issue).  The strange thing (bug behavior?) appears to drop the 
 initial octect from an IP address when examining a file over HDFS.  
 [training@localhost hands-on]$ hadoop fs -tail weblog/access_log
 .190.174.142 - - [03/Dec/2011:13:28:08 -0800] GET 
 /assets/js/javascript_combined.js HTTP/1.1 200 20404
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /assets/img/home-logo.png HTTP/1.1 200 3892
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/019.jpg HTTP/1.1 200 74446
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/g_still_04.jpg HTTP/1.1 200 761555
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/07082218.jpg HTTP/1.1 200 154609
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000163.jpg HTTP/1.1 200 60117
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/Chacha.jpg HTTP/1.1 200 109379
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000159.jpg HTTP/1.1 200 161657
 *When looking at the original log data outside of HDFS with the os version of 
 the tail command we see the following*
 [training@localhost hands-on]$ hadoop fs -get weblog/access_log ./
 [training@localhost hands-on]$ tail access_log 
 10.190.174.142 - - [03/Dec/2011:13:28:06 -0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 10.190.174.142 - - [03/Dec/2011:13:28:08 -0800] GET 
 /assets/js/javascript_combined.js HTTP/1.1 200 20404
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /assets/img/home-logo.png HTTP/1.1 200 3892
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/019.jpg HTTP/1.1 200 74446
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/g_still_04.jpg HTTP/1.1 200 761555
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/07082218.jpg HTTP/1.1 200 154609
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000163.jpg HTTP/1.1 200 60117
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/Chacha.jpg HTTP/1.1 200 109379
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000159.jpg HTTP/1.1 200 161657
 When using non ip data seperated by periods, it gets even worse and even more 
 data is masked? (same data subtituting names for IP octects).  Note we loose 
 the first line well into the URI string? *
 [training@localhost hands-on]$ hadoop fs -tail weblog/test_log
 s/javascript_combined.js HTTP/1.1 200 20404
 larry.billy.will.amy - - [03/Dec/2011:13:28:09 -0800] GET 
 /assets/img/home-logo.png HTTP/1.1 200 3892
 larry.billy.will.amy - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/019.jpg HTTP/1.1 200 74446
 larry.billy.will.amy - - [03/Dec/2011:13:28:larry.-0800] GET 
 /images/filmmediablock/360/g_still_04.jpg HTTP/1.1 200 761555
 larry.billy.will.amy - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/07082218.jpg HTTP/1.1 200 154609
 larry.billy.will.amy - - [03/Dec/2011:13:28:larry.-0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 larry.billy.will.amy - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000163.jpg HTTP/1.1 200 60117
 larry.billy.will.amy - - [03/Dec/2011:13:28:larry.-0800] GET 
 /images/filmmediablock/360/Chacha.jpg HTTP/1.1 200 larry.379
 larry.billy.will.amy - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000159.jpg HTTP/1.1 200 161657
 * and verifying what we are looking at in normal tail matches - note the 
 first line is not represented in the hadoop fs -tail as its only grabbing 9 
 lines instead of 10... as I mentioned before. Align the two text 

[jira] [Commented] (HDFS-7935) Support multi-homed networks when Kerberos security is enabled

2015-03-16 Thread Arun Suresh (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363477#comment-14363477
 ] 

Arun Suresh commented on HDFS-7935:
---

[~kihwal], Yup looks like that would work. I guess then.. that the only thing 
remaining is a minor fix for https://issues.apache.org/jira/browse/HDFS-4448 
which would allow one to deploy NN in HA with security in a multi-homed env.

 Support multi-homed networks when Kerberos security is enabled
 --

 Key: HDFS-7935
 URL: https://issues.apache.org/jira/browse/HDFS-7935
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Arun Suresh
Assignee: Arun Suresh

 Currently, during SASL negotiation stage between ipc Client and Server, The 
 server sends only a single serviceId (curresponding to a single principal) to 
 the client. This is the principal the the server process is logged in as 
 during startup.
 It is possible that in a multi-homed network, the server might be associated 
 with more than one principal, and thus severs must provide the clients all 
 possible principals it can use to connect to.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HDFS-2360) Ugly stacktrace when quota exceeds

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2360?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HDFS-2360.
---
Resolution: Not a Problem

The last line of the command (excluding the log and its stack trace via the 
WARN) does today print the base message reason that should catch the eye 
clearly:

{code}
put: The DiskSpace quota of /testDir is exceeded: quota = 1024 B = 1 KB but 
diskspace consumed = 402653184 B = 384 MB
{code}

Resolving this as it should be clear enough. To get rid of the WARN, the client 
logger can be nullified, but the catch layer is rather generic today to 
specifically turn it off without causing other impact (for other use-cases and 
troubles) I think.

As always though, feel free to reopen with any counter-point.

 Ugly stacktrace when quota exceeds
 --

 Key: HDFS-2360
 URL: https://issues.apache.org/jira/browse/HDFS-2360
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 0.23.0
Reporter: Rajit Saha
Priority: Minor

 Will it be better to catch the exception and throw a small reasonable messege 
 to user when they exceed quota?
 $hdfs  dfs -mkdir testDir
 $hdfs  dfsadmin -setSpaceQuota 191M  testDir
 $hdfs dfs -count -q testDir
 none inf   200278016   2002780161 
0  0
 hdfs://NN hostname:port/user/hdfsqa/testDir
 $hdfs dfs -put /etc/passwd /user/hadoopqa/testDir 
 11/09/19 08:08:15 WARN hdfs.DFSClient: DataStreamer Exception
 org.apache.hadoop.hdfs.protocol.DSQuotaExceededException: The DiskSpace quota 
 of /user/hdfsqa/testDir is exceeded:
 quota=191.0m diskspace consumed=768.0m
 at 
 org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota.verifyQuota(INodeDirectoryWithQuota.java:159)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.verifyQuota(FSDirectory.java:1609)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.updateCount(FSDirectory.java:1383)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.addBlock(FSDirectory.java:370)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allocateBlock(FSNamesystem.java:1681)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1476)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:389)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at 
 org.apache.hadoop.ipc.WritableRpcEngine$Server.call(WritableRpcEngine.java:365)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1496)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1492)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1135)
 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1490)
 at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
 Method)
 at 
 sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
 at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
 at 
 org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:90)
 at 
 org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:57)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1100)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:972)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:454)
 Caused by: org.apache.hadoop.hdfs.protocol.DSQuotaExceededException: The 
 DiskSpace quota of /user/hdfsqa/testDir is
 exceeded: quota=191.0m diskspace consumed=768.0m
 at 
 org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota.verifyQuota(INodeDirectoryWithQuota.java:159)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.verifyQuota(FSDirectory.java:1609)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.updateCount(FSDirectory.java:1383)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.addBlock(FSDirectory.java:370)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allocateBlock(FSNamesystem.java:1681)
 at 
 

[jira] [Commented] (HDFS-5356) MiniDFSCluster shoud close all open FileSystems when shutdown()

2015-03-16 Thread Rakesh R (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5356?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363428#comment-14363428
 ] 

Rakesh R commented on HDFS-5356:


Thanks [~vinayrpet] for triggering the build. It looks like the testcase 
failure is unrelated to this patch. Please have a look at the latest patch, I 
addressed the review comments.

 MiniDFSCluster shoud close all open FileSystems when shutdown()
 ---

 Key: HDFS-5356
 URL: https://issues.apache.org/jira/browse/HDFS-5356
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 3.0.0, 2.2.0
Reporter: haosdent
Assignee: Rakesh R
Priority: Critical
 Attachments: HDFS-5356-1.patch, HDFS-5356-2.patch, HDFS-5356-3.patch, 
 HDFS-5356-4.patch, HDFS-5356-5.patch, HDFS-5356-6.patch, HDFS-5356-7.patch, 
 HDFS-5356-8.patch, HDFS-5356.patch


 After add some metrics functions to DFSClient, I found that some unit tests 
 relates to metrics are failed. Because MiniDFSCluster are never close open 
 FileSystems, DFSClients are alive after MiniDFSCluster shutdown(). The 
 metrics of DFSClients in DefaultMetricsSystem are still exist and this make 
 other unit tests failed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-5640) Add snapshot methods to FileContext.

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5640?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363432#comment-14363432
 ] 

Hadoop QA commented on HDFS-5640:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12704411/HDFS-5640-001.patch
  against trunk revision d1eebd9.

{color:red}-1 patch{color}.  Trunk compilation may be broken.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9899//console

This message is automatically generated.

 Add snapshot methods to FileContext.
 

 Key: HDFS-5640
 URL: https://issues.apache.org/jira/browse/HDFS-5640
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client, snapshots
Affects Versions: 3.0.0, 2.2.0
Reporter: Chris Nauroth
Assignee: Rakesh R
 Attachments: HDFS-5640-001.patch


 Currently, methods related to HDFS snapshots are defined on {{FileSystem}}.  
 For feature parity, these methods need to be added to {{FileContext}}.  This 
 would also require updating {{AbstractFileSystem}} and the {{Hdfs}} subclass.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-4829) Make the fs-shell tail command behave more like Unix tail command

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4829?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J updated HDFS-4829:
--
Summary: Make the fs-shell tail command behave more like Unix tail command  
(was: Strange loss of data displayed in hadoop fs -tail command)

 Make the fs-shell tail command behave more like Unix tail command
 -

 Key: HDFS-4829
 URL: https://issues.apache.org/jira/browse/HDFS-4829
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.0.0-alpha
 Environment: OS Centos 6.3 (on Intel Core2 Duo, VMware Player VM 
 running under windows 7)
 Testing on both 2.0.0-cdh4.1.1 and 2.0.0-cdh4.1.2
Reporter: Todd Grayson
Priority: Minor

 Strange behavior of the hadoop fs -tail command - its default for output 
 seems to be 9 lines of output vs 10 lines of output in the OS version of the 
 command (minor issue).  The strange thing (bug behavior?) appears to drop the 
 initial octect from an IP address when examining a file over HDFS.  
 [training@localhost hands-on]$ hadoop fs -tail weblog/access_log
 .190.174.142 - - [03/Dec/2011:13:28:08 -0800] GET 
 /assets/js/javascript_combined.js HTTP/1.1 200 20404
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /assets/img/home-logo.png HTTP/1.1 200 3892
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/019.jpg HTTP/1.1 200 74446
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/g_still_04.jpg HTTP/1.1 200 761555
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/07082218.jpg HTTP/1.1 200 154609
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000163.jpg HTTP/1.1 200 60117
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/Chacha.jpg HTTP/1.1 200 109379
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000159.jpg HTTP/1.1 200 161657
 *When looking at the original log data outside of HDFS with the os version of 
 the tail command we see the following*
 [training@localhost hands-on]$ hadoop fs -get weblog/access_log ./
 [training@localhost hands-on]$ tail access_log 
 10.190.174.142 - - [03/Dec/2011:13:28:06 -0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 10.190.174.142 - - [03/Dec/2011:13:28:08 -0800] GET 
 /assets/js/javascript_combined.js HTTP/1.1 200 20404
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /assets/img/home-logo.png HTTP/1.1 200 3892
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/019.jpg HTTP/1.1 200 74446
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/g_still_04.jpg HTTP/1.1 200 761555
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/07082218.jpg HTTP/1.1 200 154609
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000163.jpg HTTP/1.1 200 60117
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/Chacha.jpg HTTP/1.1 200 109379
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000159.jpg HTTP/1.1 200 161657
 When using non ip data seperated by periods, it gets even worse and even more 
 data is masked? (same data subtituting names for IP octects).  Note we loose 
 the first line well into the URI string? *
 [training@localhost hands-on]$ hadoop fs -tail weblog/test_log
 s/javascript_combined.js HTTP/1.1 200 20404
 larry.billy.will.amy - - [03/Dec/2011:13:28:09 -0800] GET 
 /assets/img/home-logo.png HTTP/1.1 200 3892
 larry.billy.will.amy - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/019.jpg HTTP/1.1 200 74446
 larry.billy.will.amy - - [03/Dec/2011:13:28:larry.-0800] GET 
 /images/filmmediablock/360/g_still_04.jpg HTTP/1.1 200 761555
 larry.billy.will.amy - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/07082218.jpg HTTP/1.1 200 154609
 larry.billy.will.amy - - [03/Dec/2011:13:28:larry.-0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 larry.billy.will.amy - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000163.jpg HTTP/1.1 200 60117
 larry.billy.will.amy - - [03/Dec/2011:13:28:larry.-0800] GET 
 /images/filmmediablock/360/Chacha.jpg HTTP/1.1 200 larry.379
 larry.billy.will.amy - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000159.jpg 

[jira] [Commented] (HDFS-3745) fsck prints that it's using KSSL even when it's in fact using SPNEGO for authentication

2015-03-16 Thread Harsh J (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363409#comment-14363409
 ] 

Harsh J commented on HDFS-3745:
---

I think it should not be harmful to keep the sub-mechanism info around. We 
could just rename the KERBEROS_SSL into KERBEROS_SPNEGO on trunk I think, since 
we do not use/support KERBEROS_SSL (KSSL) anymore in trunk and branch-2?

 fsck prints that it's using KSSL even when it's in fact using SPNEGO for 
 authentication
 ---

 Key: HDFS-3745
 URL: https://issues.apache.org/jira/browse/HDFS-3745
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client, security
Affects Versions: 1.2.0, 2.0.0-alpha
Reporter: Aaron T. Myers
Priority: Trivial
  Labels: newbie

 In branch-2 (which exclusively uses SPNEGO for HTTP authentication) and in 
 branch-1 (which can optionally use SPNEGO for HTTP authentication), running 
 fsck will print the following, which isn't quite right:
 {quote}
 FSCK started by hdfs (auth:KERBEROS_SSL) from...
 {quote}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7930) commitBlockSynchronization() does not remove locations

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7930?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363443#comment-14363443
 ] 

Hadoop QA commented on HDFS-7930:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12704734/HDFS-7930.001.patch
  against trunk revision 3ff1ba2.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-hdfs-project/hadoop-hdfs:

  org.apache.hadoop.hdfs.server.namenode.ha.TestRetryCacheWithHA
  
org.apache.hadoop.hdfs.server.blockmanagement.TestDatanodeManager

Test results: 
https://builds.apache.org/job/PreCommit-HDFS-Build/9898//testReport/
Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9898//console

This message is automatically generated.

 commitBlockSynchronization() does not remove locations
 --

 Key: HDFS-7930
 URL: https://issues.apache.org/jira/browse/HDFS-7930
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.7.0
Reporter: Konstantin Shvachko
Assignee: Yi Liu
Priority: Blocker
 Attachments: HDFS-7930.001.patch


 When {{commitBlockSynchronization()}} has less {{newTargets}} than in the 
 original block it does not remove unconfirmed locations. This results in that 
 the the block stores locations of different lengths or genStamp (corrupt).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HDFS-3349) DFSAdmin fetchImage command should initialize security credentials

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3349?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HDFS-3349.
---
  Resolution: Cannot Reproduce
Target Version/s:   (was: 2.0.0-alpha)

Trying with lack of credentials throws the proper response back (No tgt). I 
think this is stale given Aaron's comment as well, marking as resolved.

 DFSAdmin fetchImage command should initialize security credentials
 --

 Key: HDFS-3349
 URL: https://issues.apache.org/jira/browse/HDFS-3349
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 2.0.0-alpha
Reporter: Aaron T. Myers
Priority: Minor

 The `hdfs dfsadmin -fetchImage' command should fetch the fsimage using the 
 appropriate credentials if security is enabled.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-4829) Make the fs-shell tail command behave more like Unix tail command

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4829?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J updated HDFS-4829:
--
Environment: (was: OS Centos 6.3 (on Intel Core2 Duo, VMware Player VM 
running under windows 7)
Testing on both 2.0.0-cdh4.1.1 and 2.0.0-cdh4.1.2)

 Make the fs-shell tail command behave more like Unix tail command
 -

 Key: HDFS-4829
 URL: https://issues.apache.org/jira/browse/HDFS-4829
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.0.0-alpha
Reporter: Todd Grayson
Priority: Minor
  Labels: newbie

 Strange behavior of the hadoop fs -tail command - its default for output 
 seems to be 9 lines of output vs 10 lines of output in the OS version of the 
 command (minor issue).  The strange thing (bug behavior?) appears to drop the 
 initial octect from an IP address when examining a file over HDFS.  
 [training@localhost hands-on]$ hadoop fs -tail weblog/access_log
 .190.174.142 - - [03/Dec/2011:13:28:08 -0800] GET 
 /assets/js/javascript_combined.js HTTP/1.1 200 20404
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /assets/img/home-logo.png HTTP/1.1 200 3892
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/019.jpg HTTP/1.1 200 74446
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/g_still_04.jpg HTTP/1.1 200 761555
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/07082218.jpg HTTP/1.1 200 154609
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000163.jpg HTTP/1.1 200 60117
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/Chacha.jpg HTTP/1.1 200 109379
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000159.jpg HTTP/1.1 200 161657
 *When looking at the original log data outside of HDFS with the os version of 
 the tail command we see the following*
 [training@localhost hands-on]$ hadoop fs -get weblog/access_log ./
 [training@localhost hands-on]$ tail access_log 
 10.190.174.142 - - [03/Dec/2011:13:28:06 -0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 10.190.174.142 - - [03/Dec/2011:13:28:08 -0800] GET 
 /assets/js/javascript_combined.js HTTP/1.1 200 20404
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /assets/img/home-logo.png HTTP/1.1 200 3892
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/019.jpg HTTP/1.1 200 74446
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/g_still_04.jpg HTTP/1.1 200 761555
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/07082218.jpg HTTP/1.1 200 154609
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000163.jpg HTTP/1.1 200 60117
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/Chacha.jpg HTTP/1.1 200 109379
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000159.jpg HTTP/1.1 200 161657
 When using non ip data seperated by periods, it gets even worse and even more 
 data is masked? (same data subtituting names for IP octects).  Note we loose 
 the first line well into the URI string? *
 [training@localhost hands-on]$ hadoop fs -tail weblog/test_log
 s/javascript_combined.js HTTP/1.1 200 20404
 larry.billy.will.amy - - [03/Dec/2011:13:28:09 -0800] GET 
 /assets/img/home-logo.png HTTP/1.1 200 3892
 larry.billy.will.amy - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/019.jpg HTTP/1.1 200 74446
 larry.billy.will.amy - - [03/Dec/2011:13:28:larry.-0800] GET 
 /images/filmmediablock/360/g_still_04.jpg HTTP/1.1 200 761555
 larry.billy.will.amy - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/07082218.jpg HTTP/1.1 200 154609
 larry.billy.will.amy - - [03/Dec/2011:13:28:larry.-0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 larry.billy.will.amy - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000163.jpg HTTP/1.1 200 60117
 larry.billy.will.amy - - [03/Dec/2011:13:28:larry.-0800] GET 
 /images/filmmediablock/360/Chacha.jpg HTTP/1.1 200 larry.379
 larry.billy.will.amy - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000159.jpg HTTP/1.1 200 161657
 * and verifying what we are looking at in normal tail matches - note the 
 first line is not 

[jira] [Updated] (HDFS-4829) Strange loss of data displayed in hadoop fs -tail command

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4829?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J updated HDFS-4829:
--
Issue Type: Improvement  (was: Bug)

 Strange loss of data displayed in hadoop fs -tail command
 -

 Key: HDFS-4829
 URL: https://issues.apache.org/jira/browse/HDFS-4829
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.0.0-alpha
 Environment: OS Centos 6.3 (on Intel Core2 Duo, VMware Player VM 
 running under windows 7)
 Testing on both 2.0.0-cdh4.1.1 and 2.0.0-cdh4.1.2
Reporter: Todd Grayson
Priority: Minor

 Strange behavior of the hadoop fs -tail command - its default for output 
 seems to be 9 lines of output vs 10 lines of output in the OS version of the 
 command (minor issue).  The strange thing (bug behavior?) appears to drop the 
 initial octect from an IP address when examining a file over HDFS.  
 [training@localhost hands-on]$ hadoop fs -tail weblog/access_log
 .190.174.142 - - [03/Dec/2011:13:28:08 -0800] GET 
 /assets/js/javascript_combined.js HTTP/1.1 200 20404
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /assets/img/home-logo.png HTTP/1.1 200 3892
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/019.jpg HTTP/1.1 200 74446
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/g_still_04.jpg HTTP/1.1 200 761555
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/07082218.jpg HTTP/1.1 200 154609
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000163.jpg HTTP/1.1 200 60117
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/Chacha.jpg HTTP/1.1 200 109379
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000159.jpg HTTP/1.1 200 161657
 *When looking at the original log data outside of HDFS with the os version of 
 the tail command we see the following*
 [training@localhost hands-on]$ hadoop fs -get weblog/access_log ./
 [training@localhost hands-on]$ tail access_log 
 10.190.174.142 - - [03/Dec/2011:13:28:06 -0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 10.190.174.142 - - [03/Dec/2011:13:28:08 -0800] GET 
 /assets/js/javascript_combined.js HTTP/1.1 200 20404
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /assets/img/home-logo.png HTTP/1.1 200 3892
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/019.jpg HTTP/1.1 200 74446
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/g_still_04.jpg HTTP/1.1 200 761555
 10.190.174.142 - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/07082218.jpg HTTP/1.1 200 154609
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000163.jpg HTTP/1.1 200 60117
 10.190.174.142 - - [03/Dec/2011:13:28:10 -0800] GET 
 /images/filmmediablock/360/Chacha.jpg HTTP/1.1 200 109379
 10.190.174.142 - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000159.jpg HTTP/1.1 200 161657
 When using non ip data seperated by periods, it gets even worse and even more 
 data is masked? (same data subtituting names for IP octects).  Note we loose 
 the first line well into the URI string? *
 [training@localhost hands-on]$ hadoop fs -tail weblog/test_log
 s/javascript_combined.js HTTP/1.1 200 20404
 larry.billy.will.amy - - [03/Dec/2011:13:28:09 -0800] GET 
 /assets/img/home-logo.png HTTP/1.1 200 3892
 larry.billy.will.amy - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/019.jpg HTTP/1.1 200 74446
 larry.billy.will.amy - - [03/Dec/2011:13:28:larry.-0800] GET 
 /images/filmmediablock/360/g_still_04.jpg HTTP/1.1 200 761555
 larry.billy.will.amy - - [03/Dec/2011:13:28:09 -0800] GET 
 /images/filmmediablock/360/07082218.jpg HTTP/1.1 200 154609
 larry.billy.will.amy - - [03/Dec/2011:13:28:larry.-0800] GET 
 /images/filmpics//2229/GOEMON-NUKI-000163.jpg HTTP/1.1 200 184976
 larry.billy.will.amy - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000163.jpg HTTP/1.1 200 60117
 larry.billy.will.amy - - [03/Dec/2011:13:28:larry.-0800] GET 
 /images/filmmediablock/360/Chacha.jpg HTTP/1.1 200 larry.379
 larry.billy.will.amy - - [03/Dec/2011:13:28:11 -0800] GET 
 /images/filmmediablock/360/GOEMON-NUKI-000159.jpg HTTP/1.1 200 161657
 * and verifying what we are looking at in normal tail matches - note the 
 first line is not 

[jira] [Resolved] (HDFS-5740) getmerge file system shell command needs error message for user error

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5740?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HDFS-5740.
---
Resolution: Not a Problem

This is no longer an issue on branch-2 and trunk today. The command accepts a 
collection of files now, and prepares the output accordingly.

 getmerge file system shell command needs error message for user error
 -

 Key: HDFS-5740
 URL: https://issues.apache.org/jira/browse/HDFS-5740
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 1.1.2
 Environment: {noformat}[jpfuntner@h58 tmp]$ cat /etc/redhat-release
 Red Hat Enterprise Linux Server release 6.0 (Santiago)
 [jpfuntner@h58 tmp]$ hadoop version
 Hadoop 1.1.2.21
 Subversion  -r 
 Compiled by jenkins on Thu Jan 10 03:38:39 PST 2013
 From source with checksum ce0aa0de785f572347f1afee69c73861{noformat}
Reporter: John Pfuntner
Priority: Minor

 I naively tried a {{getmerge}} operation but it didn't seem to do anything 
 and there was no error message:
 {noformat}[jpfuntner@h58 tmp]$ hadoop fs -mkdir /user/jpfuntner/tmp
 [jpfuntner@h58 tmp]$ num=0; while [ $num -lt 5 ]; do echo file$num | hadoop 
 fs -put - /user/jpfuntner/tmp/file$num; let num=num+1; done
 [jpfuntner@h58 tmp]$ ls -A
 [jpfuntner@h58 tmp]$ hadoop fs -getmerge /user/jpfuntner/tmp/file* files.txt
 [jpfuntner@h58 tmp]$ ls -A
 [jpfuntner@h58 tmp]$ hadoop fs -ls /user/jpfuntner/tmp
 Found 5 items
 -rw---   3 jpfuntner hdfs  6 2014-01-08 17:37 
 /user/jpfuntner/tmp/file0
 -rw---   3 jpfuntner hdfs  6 2014-01-08 17:37 
 /user/jpfuntner/tmp/file1
 -rw---   3 jpfuntner hdfs  6 2014-01-08 17:37 
 /user/jpfuntner/tmp/file2
 -rw---   3 jpfuntner hdfs  6 2014-01-08 17:37 
 /user/jpfuntner/tmp/file3
 -rw---   3 jpfuntner hdfs  6 2014-01-08 17:37 
 /user/jpfuntner/tmp/file4
 [jpfuntner@h58 tmp]$ {noformat}
 It was pointed out to me that I made a mistake and my source should have been 
 a directory not a set of regular files.  It works if I use the directory:
 {noformat}[jpfuntner@h58 tmp]$ hadoop fs -getmerge /user/jpfuntner/tmp/ 
 files.txt
 [jpfuntner@h58 tmp]$ ls -A
 files.txt  .files.txt.crc
 [jpfuntner@h58 tmp]$ cat files.txt
 file0
 file1
 file2
 file3
 file4
 [jpfuntner@h58 tmp]$ {noformat}
 I think the {{getmerge}} command should issue an error message to let the 
 user know they made a mistake.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Reopened] (HDFS-2360) Ugly stacktrace when quota exceeds

2015-03-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2360?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer reopened HDFS-2360:


OK, then let me re-open it.

Having oodles of useless stack trace here is *incredibly* user-unfriendly.  
Users do miss this message very very often because, believe it or not, they 
aren't Java programmers who are used to reading these things.

 Ugly stacktrace when quota exceeds
 --

 Key: HDFS-2360
 URL: https://issues.apache.org/jira/browse/HDFS-2360
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 0.23.0
Reporter: Rajit Saha
Priority: Minor

 Will it be better to catch the exception and throw a small reasonable messege 
 to user when they exceed quota?
 $hdfs  dfs -mkdir testDir
 $hdfs  dfsadmin -setSpaceQuota 191M  testDir
 $hdfs dfs -count -q testDir
 none inf   200278016   2002780161 
0  0
 hdfs://NN hostname:port/user/hdfsqa/testDir
 $hdfs dfs -put /etc/passwd /user/hadoopqa/testDir 
 11/09/19 08:08:15 WARN hdfs.DFSClient: DataStreamer Exception
 org.apache.hadoop.hdfs.protocol.DSQuotaExceededException: The DiskSpace quota 
 of /user/hdfsqa/testDir is exceeded:
 quota=191.0m diskspace consumed=768.0m
 at 
 org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota.verifyQuota(INodeDirectoryWithQuota.java:159)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.verifyQuota(FSDirectory.java:1609)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.updateCount(FSDirectory.java:1383)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.addBlock(FSDirectory.java:370)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allocateBlock(FSNamesystem.java:1681)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1476)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:389)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at 
 org.apache.hadoop.ipc.WritableRpcEngine$Server.call(WritableRpcEngine.java:365)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1496)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1492)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1135)
 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1490)
 at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
 Method)
 at 
 sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
 at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
 at 
 org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:90)
 at 
 org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:57)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1100)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:972)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:454)
 Caused by: org.apache.hadoop.hdfs.protocol.DSQuotaExceededException: The 
 DiskSpace quota of /user/hdfsqa/testDir is
 exceeded: quota=191.0m diskspace consumed=768.0m
 at 
 org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota.verifyQuota(INodeDirectoryWithQuota.java:159)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.verifyQuota(FSDirectory.java:1609)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.updateCount(FSDirectory.java:1383)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.addBlock(FSDirectory.java:370)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allocateBlock(FSNamesystem.java:1681)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1476)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:389)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at 

[jira] [Resolved] (HDFS-147) Stack trace on spaceQuota excced .

2015-03-16 Thread Xiaoyu Yao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-147?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao resolved HDFS-147.
-
Resolution: Duplicate

 Stack trace on spaceQuota excced .
 --

 Key: HDFS-147
 URL: https://issues.apache.org/jira/browse/HDFS-147
 Project: Hadoop HDFS
  Issue Type: Bug
 Environment: All
Reporter: Ravi Phulari
Assignee: Xiaoyu Yao
  Labels: newbie

 Currently disk space quota exceed exception spits out stack trace . It's 
 better to show error message instead of stack trace .
 {code}
 somehost:Hadoop guesti$ bin/hdfs dfsadmin -setSpaceQuota 2 2344
 somehost:Hadoop guest$ bin/hadoop fs -put conf 2344
 09/06/19 16:44:30 WARN hdfs.DFSClient: DataStreamer Exception: 
 org.apache.hadoop.hdfs.protocol.QuotaExceededException: 
 org.apache.hadoop.hdfs.protocol.QuotaExceededException: The quota of 
 /user/guest/2344 is exceeded: namespace quota=-1 file count=4, diskspace 
 quota=2 diskspace=67108864
   at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   at 
 sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
   at 
 sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
 ..
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-4494) Confusing exception for unresolvable hdfs host with security enabled

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4494?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J updated HDFS-4494:
--
Fix Version/s: 2.6.0

 Confusing exception for unresolvable hdfs host with security enabled
 

 Key: HDFS-4494
 URL: https://issues.apache.org/jira/browse/HDFS-4494
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0
Reporter: Daryn Sharp
Priority: Minor
 Fix For: 2.6.0


 {noformat}
 $ hadoop fs -ls hdfs://unresolvable-host
 ls: Can't replace _HOST pattern since client address is null
 {noformat}
 It's misleading because it's not even related to the client's address.  It'd 
 be a bit more informative to see something like {{UnknownHostException: 
 unresolvable-host}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-1330) Make RPCs to DataNodes timeout

2015-03-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1330?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HDFS-1330:
---
Fix Version/s: (was: 2.0.0-alpha)

 Make RPCs to DataNodes timeout
 --

 Key: HDFS-1330
 URL: https://issues.apache.org/jira/browse/HDFS-1330
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: datanode
Affects Versions: 0.22.0, 0.23.0
Reporter: Hairong Kuang
Assignee: John George
 Fix For: 0.22.0, 0.23.0

 Attachments: HADOOP-6889-fortrunk-2.patch, hdfsRpcTimeout.patch


 This jira aims to make client/datanode or datanode/datanode RPC to have a 
 timeout of DataNode#socketTimeout.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-2333) HDFS-2284 introduced 2 findbugs warnings on trunk

2015-03-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2333?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HDFS-2333:
---
Fix Version/s: (was: 2.0.0-alpha)

 HDFS-2284 introduced 2 findbugs warnings on trunk
 -

 Key: HDFS-2333
 URL: https://issues.apache.org/jira/browse/HDFS-2333
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Ivan Kelly
Assignee: Tsz Wo Nicholas Sze
 Fix For: 0.20.205.0, 0.23.0

 Attachments: HDFS-2333.diff, h2333_20110914.patch, 
 h2333_20110914b.patch, h2333_20110915.patch, h2333_20110915_0.20s.patch


 When HDFS-2284 was submitted it made DFSOutputStream public which triggered 
 two SC_START_IN_CTOR findbug warnings.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-2360) Ugly stacktrace when quota exceeds

2015-03-16 Thread Allen Wittenauer (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363619#comment-14363619
 ] 

Allen Wittenauer commented on HDFS-2360:


I'm re-opening to get rid of the stack trace as well.  I see that someone else 
has also duped that request to this issue as well. :)

 Ugly stacktrace when quota exceeds
 --

 Key: HDFS-2360
 URL: https://issues.apache.org/jira/browse/HDFS-2360
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 0.23.0
Reporter: Rajit Saha
Priority: Minor

 Will it be better to catch the exception and throw a small reasonable messege 
 to user when they exceed quota?
 $hdfs  dfs -mkdir testDir
 $hdfs  dfsadmin -setSpaceQuota 191M  testDir
 $hdfs dfs -count -q testDir
 none inf   200278016   2002780161 
0  0
 hdfs://NN hostname:port/user/hdfsqa/testDir
 $hdfs dfs -put /etc/passwd /user/hadoopqa/testDir 
 11/09/19 08:08:15 WARN hdfs.DFSClient: DataStreamer Exception
 org.apache.hadoop.hdfs.protocol.DSQuotaExceededException: The DiskSpace quota 
 of /user/hdfsqa/testDir is exceeded:
 quota=191.0m diskspace consumed=768.0m
 at 
 org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota.verifyQuota(INodeDirectoryWithQuota.java:159)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.verifyQuota(FSDirectory.java:1609)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.updateCount(FSDirectory.java:1383)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.addBlock(FSDirectory.java:370)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allocateBlock(FSNamesystem.java:1681)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1476)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:389)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at 
 org.apache.hadoop.ipc.WritableRpcEngine$Server.call(WritableRpcEngine.java:365)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1496)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1492)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1135)
 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1490)
 at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
 Method)
 at 
 sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
 at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
 at 
 org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:90)
 at 
 org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:57)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1100)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:972)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:454)
 Caused by: org.apache.hadoop.hdfs.protocol.DSQuotaExceededException: The 
 DiskSpace quota of /user/hdfsqa/testDir is
 exceeded: quota=191.0m diskspace consumed=768.0m
 at 
 org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota.verifyQuota(INodeDirectoryWithQuota.java:159)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.verifyQuota(FSDirectory.java:1609)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.updateCount(FSDirectory.java:1383)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.addBlock(FSDirectory.java:370)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allocateBlock(FSNamesystem.java:1681)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1476)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:389)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at 
 

[jira] [Commented] (HDFS-7929) inotify unable fetch pre-upgrade edit log segments once upgrade starts

2015-03-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363646#comment-14363646
 ] 

Hadoop QA commented on HDFS-7929:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12704832/HDFS-7929-001.patch
  against trunk revision ed4e72a.

{color:red}-1 patch{color}.  Trunk compilation may be broken.

Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/9900//console

This message is automatically generated.

 inotify unable fetch pre-upgrade edit log segments once upgrade starts
 --

 Key: HDFS-7929
 URL: https://issues.apache.org/jira/browse/HDFS-7929
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Zhe Zhang
Assignee: Zhe Zhang
 Attachments: HDFS-7929-000.patch, HDFS-7929-001.patch


 inotify is often used to periodically poll HDFS events. However, once an HDFS 
 upgrade has started, edit logs are moved to /previous on the NN, which is not 
 accessible. Moreover, once the upgrade is finalized /previous is currently 
 lost forever.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-4494) Confusing exception for unresolvable hdfs host with security enabled

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4494?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J updated HDFS-4494:
--
Target Version/s:   (was: 3.0.0, 2.1.0-beta)

 Confusing exception for unresolvable hdfs host with security enabled
 

 Key: HDFS-4494
 URL: https://issues.apache.org/jira/browse/HDFS-4494
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0
Reporter: Daryn Sharp
Priority: Minor
 Fix For: 2.6.0


 {noformat}
 $ hadoop fs -ls hdfs://unresolvable-host
 ls: Can't replace _HOST pattern since client address is null
 {noformat}
 It's misleading because it's not even related to the client's address.  It'd 
 be a bit more informative to see something like {{UnknownHostException: 
 unresolvable-host}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-2290) Block with corrupt replica is not getting replicated

2015-03-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2290?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HDFS-2290:
---
Fix Version/s: (was: 2.0.0-alpha)

 Block with corrupt replica is not getting replicated
 

 Key: HDFS-2290
 URL: https://issues.apache.org/jira/browse/HDFS-2290
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 0.22.0
Reporter: Konstantin Shvachko
Assignee: Benoy Antony
  Labels: blockmanagement, replication
 Fix For: 0.22.0, 0.23.0

 Attachments: HDFS-2290_0.22.patch, HDFS-2290_022.patch, 
 HDFS-2290_022.patch, HDFS-2290_trunk.patch, HDFS-2290_trunk.patch


 A block has one replica marked as corrupt and two good ones. countNodes() 
 correctly detects that there are only 2 live replicas, and fsck reports the 
 block as under-replicated. But ReplicationMonitor never schedules replication 
 of good replicas.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-2360) Ugly stacktrace when quota exceeds

2015-03-16 Thread Harsh J (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14363612#comment-14363612
 ] 

Harsh J commented on HDFS-2360:
---

[~aw] - I agree with that sentiment, but the previous state was the lack of 
message. Now we do have the message, like my comment example indicates. Is that 
still unclear, or is the reopen just to get rid of the stack trace WARN as well 
on clients?

 Ugly stacktrace when quota exceeds
 --

 Key: HDFS-2360
 URL: https://issues.apache.org/jira/browse/HDFS-2360
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 0.23.0
Reporter: Rajit Saha
Priority: Minor

 Will it be better to catch the exception and throw a small reasonable messege 
 to user when they exceed quota?
 $hdfs  dfs -mkdir testDir
 $hdfs  dfsadmin -setSpaceQuota 191M  testDir
 $hdfs dfs -count -q testDir
 none inf   200278016   2002780161 
0  0
 hdfs://NN hostname:port/user/hdfsqa/testDir
 $hdfs dfs -put /etc/passwd /user/hadoopqa/testDir 
 11/09/19 08:08:15 WARN hdfs.DFSClient: DataStreamer Exception
 org.apache.hadoop.hdfs.protocol.DSQuotaExceededException: The DiskSpace quota 
 of /user/hdfsqa/testDir is exceeded:
 quota=191.0m diskspace consumed=768.0m
 at 
 org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota.verifyQuota(INodeDirectoryWithQuota.java:159)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.verifyQuota(FSDirectory.java:1609)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.updateCount(FSDirectory.java:1383)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.addBlock(FSDirectory.java:370)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allocateBlock(FSNamesystem.java:1681)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1476)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:389)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at 
 org.apache.hadoop.ipc.WritableRpcEngine$Server.call(WritableRpcEngine.java:365)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1496)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1492)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1135)
 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1490)
 at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
 Method)
 at 
 sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
 at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
 at 
 org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:90)
 at 
 org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:57)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1100)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:972)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:454)
 Caused by: org.apache.hadoop.hdfs.protocol.DSQuotaExceededException: The 
 DiskSpace quota of /user/hdfsqa/testDir is
 exceeded: quota=191.0m diskspace consumed=768.0m
 at 
 org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota.verifyQuota(INodeDirectoryWithQuota.java:159)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.verifyQuota(FSDirectory.java:1609)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.updateCount(FSDirectory.java:1383)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.addBlock(FSDirectory.java:370)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allocateBlock(FSNamesystem.java:1681)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1476)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:389)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
  

[jira] [Updated] (HDFS-2197) Refactor RPC call implementations out of NameNode class

2015-03-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2197?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HDFS-2197:
---
Fix Version/s: (was: 2.0.0-alpha)

 Refactor RPC call implementations out of NameNode class
 ---

 Key: HDFS-2197
 URL: https://issues.apache.org/jira/browse/HDFS-2197
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Affects Versions: 0.23.0
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: HA branch (HDFS-1623), 0.23.0

 Attachments: hdfs-2197-2.txt, hdfs-2197-3.txt, hdfs-2197-4.txt, 
 hdfs-2197-trunk-final.txt, hdfs-2197.txt


 For HA, the NameNode will gain a bit of a state machine, to be able to 
 transition between standby and active states. This would be cleaner in the 
 code if the {{NameNode}} class were just a container for various services, as 
 discussed in HDFS-1974. It's also nice for testing, where it would become 
 easier to construct just the RPC handlers around a mock NameSystem, with no 
 HTTP server, for example.
 This JIRA is to move all of the protocol implementations out of {{NameNode}} 
 into a separate {{NameNodeRPCServer}} class.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7929) inotify unable fetch pre-upgrade edit log segments once upgrade starts

2015-03-16 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7929?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HDFS-7929:

Attachment: HDFS-7929-001.patch

Some pre-created edits log files in {{TestDFSUpgradeFromImage}} do not have 
version numbers. 001 patch uses {{EditLogFileInputStream#nextValidOp}} to skip 
those.

 inotify unable fetch pre-upgrade edit log segments once upgrade starts
 --

 Key: HDFS-7929
 URL: https://issues.apache.org/jira/browse/HDFS-7929
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Zhe Zhang
Assignee: Zhe Zhang
 Attachments: HDFS-7929-000.patch, HDFS-7929-001.patch


 inotify is often used to periodically poll HDFS events. However, once an HDFS 
 upgrade has started, edit logs are moved to /previous on the NN, which is not 
 accessible. Moreover, once the upgrade is finalized /previous is currently 
 lost forever.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HDFS-4494) Confusing exception for unresolvable hdfs host with security enabled

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4494?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HDFS-4494.
---
  Resolution: Done
Target Version/s: 2.1.0-beta, 3.0.0  (was: 3.0.0, 2.1.0-beta)

This seems resolved now (as of 2.6.0):

{code}
[root@host ~]# hdfs getconf -confKey hadoop.security.authentication
kerberos
[root@host ~]# hadoop fs -ls hdfs://asdfsdfsdf/
-ls: java.net.UnknownHostException: asdfsdfsdf
Usage: hadoop fs [generic options] -ls [-d] [-h] [-R] [path ...]
{code}

Marking as Done.

 Confusing exception for unresolvable hdfs host with security enabled
 

 Key: HDFS-4494
 URL: https://issues.apache.org/jira/browse/HDFS-4494
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0
Reporter: Daryn Sharp
Priority: Minor

 {noformat}
 $ hadoop fs -ls hdfs://unresolvable-host
 ls: Can't replace _HOST pattern since client address is null
 {noformat}
 It's misleading because it's not even related to the client's address.  It'd 
 be a bit more informative to see something like {{UnknownHostException: 
 unresolvable-host}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-4494) Confusing exception for unresolvable hdfs host with security enabled

2015-03-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4494?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HDFS-4494:
---
Fix Version/s: (was: 2.6.0)

 Confusing exception for unresolvable hdfs host with security enabled
 

 Key: HDFS-4494
 URL: https://issues.apache.org/jira/browse/HDFS-4494
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0
Reporter: Daryn Sharp
Priority: Minor

 {noformat}
 $ hadoop fs -ls hdfs://unresolvable-host
 ls: Can't replace _HOST pattern since client address is null
 {noformat}
 It's misleading because it's not even related to the client's address.  It'd 
 be a bit more informative to see something like {{UnknownHostException: 
 unresolvable-host}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HDFS-4290) Expose an event listener interface in DFSOutputStreams for block write pipeline status changes

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4290?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved HDFS-4290.
---
Resolution: Later

Specific problems/use-cases driving this need haven't been bought up in the 
past years. Resolving as Later for now.

 Expose an event listener interface in DFSOutputStreams for block write 
 pipeline status changes
 --

 Key: HDFS-4290
 URL: https://issues.apache.org/jira/browse/HDFS-4290
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: hdfs-client
Affects Versions: 2.0.0-alpha
Reporter: Harsh J
Priority: Minor

 I've noticed HBase periodically polls the current status of block replicas 
 for its HLog files via the API presented by HDFS-826.
 It would perhaps be better for such clients if they could register a listener 
 instead. The listener(s) can be sent an event in case things change in the 
 last open block (due to DN fall but no replacement found, etc. cases). This 
 would avoid having a periodic, parallel looped check in such clients and be 
 more efficient.
 Just a thought :)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-2232) TestHDFSCLI fails on 0.22 branch

2015-03-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2232?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HDFS-2232:
---
Fix Version/s: (was: 2.0.0-alpha)

 TestHDFSCLI fails on 0.22 branch
 

 Key: HDFS-2232
 URL: https://issues.apache.org/jira/browse/HDFS-2232
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 0.22.0
Reporter: Konstantin Shvachko
Assignee: Plamen Jeliazkov
Priority: Blocker
 Fix For: 0.22.0, 0.23.0

 Attachments: HDFS-2232-TRUNK.patch, HDFS-2232-TRUNK.patch, 
 HDFS-2232-TRUNK.patch, HDFS-2232-TRUNK.patch, HDFS-2232.patch, 
 HDFS-2232.patch, HDFS-2232.patch, HDFS-2232.patch, 
 TEST-org.apache.hadoop.cli.TestHDFSCLI.txt, 
 TEST-org.apache.hadoop.cli.TestHDFSCLI.txt, TestHDFSCLI-output.txt


 Several HDFS CLI tests fail on 0.22 branch. I can see 2 reasons:
 # Not generic enough regular expression for host names and paths. Similar to 
 MAPREDUCE-2304.
 # Some command outputs have new-line in the end.
 # And some seem to produce [much] more output than expected.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-2331) Hdfs compilation fails

2015-03-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2331?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HDFS-2331:
---
Fix Version/s: (was: 2.0.0-alpha)

 Hdfs compilation fails
 --

 Key: HDFS-2331
 URL: https://issues.apache.org/jira/browse/HDFS-2331
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 0.20.205.0, 2.0.0-alpha
Reporter: Abhijit Suresh Shingate
Assignee: Abhijit Suresh Shingate
 Fix For: 0.20.205.0, 0.23.0

 Attachments: HDFS-2331-1.patch, HDFS-2331.patch

   Original Estimate: 1h
  Remaining Estimate: 1h

 I am trying to perform complete build from trunk folder but the compilation 
 fails.
 *Commandline:*
 mvn clean install  
 *Error Message:*
 [ERROR] Failed to execute goal 
 org.apache.maven.plugins:maven-compiler-plugin:2.
 3.2:compile (default-compile) on project hadoop-hdfs: Compilation failure
 [ERROR] \Hadoop\SVN\trunk\hadoop-hdfs-project\hadoop-hdfs\src\main\java\org
 \apache\hadoop\hdfs\web\WebHdfsFileSystem.java:[209,21] type parameters of 
 TT
 cannot be determined; no unique maximal instance exists for type variable T 
 with
  upper bounds T,java.lang.Object
 [ERROR] - [Help 1]
 [ERROR]
 [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
 swit
 ch.
 [ERROR] Re-run Maven using the -X switch to enable full debug logging.
 [ERROR]
 [ERROR] For more information about the errors and possible solutions, please 
 rea
 d the following articles:
 [ERROR] [Help 1] 
 http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureExc
 eption
 [ERROR]
 [ERROR] After correcting the problems, you can resume the build with the 
 command
 This is because of known reason
 [http://bugs.sun.com/bugdatabase/view_bug.do?bug_id=6302954]
 {code:title=WebHdfsFileSystem.java|borderStyle=solid}
 Following is the code snippet with issue.
   private T T run(final HttpOpParam.Op op, final Path fspath,
   final Param?,?... parameters) throws IOException {
 final HttpURLConnection conn = httpConnect(op, fspath, parameters);
 validateResponse(op, conn);
 try {
   return jsonParse(conn.getInputStream());
 } finally {
   conn.disconnect();
 }
   }
 {code} 
 Fix to the issue
 {code:title=WebHdfsFileSystem.java|borderStyle=solid}
 Following is the fix to the issue.
   private T T run(final HttpOpParam.Op op, final Path fspath,
   final Param?,?... parameters) throws IOException {
 final HttpURLConnection conn = httpConnect(op, fspath, parameters);
 validateResponse(op, conn);
 try {
   return WebHdfsFileSystem.TjsonParse(conn.getInputStream());
 } finally {
   conn.disconnect();
 }
   }
 {code} 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-3101) cannot read empty file using webhdfs

2015-03-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3101?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HDFS-3101:
---
Fix Version/s: (was: 2.0.0-alpha)

 cannot read empty file using webhdfs
 

 Key: HDFS-3101
 URL: https://issues.apache.org/jira/browse/HDFS-3101
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: webhdfs
Affects Versions: 0.23.1
Reporter: Zhanwei Wang
Assignee: Tsz Wo Nicholas Sze
 Fix For: 1.0.2, 0.23.2

 Attachments: h3101_20120315.patch, h3101_20120315_branch-1.patch


 STEP:
 1, create a new EMPTY file
 2, read it using webhdfs.
 RESULT:
 expected: get a empty file
 I got: 
 {RemoteException:{exception:IOException,javaClassName:java.io.IOException,message:Offset=0
  out of the range [0, 0); OPEN, path=/testFile}}
 First of all, [0, 0) is not a valid range, and I think read a empty file 
 should be OK.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-2002) Incorrect computation of needed blocks in getTurnOffTip()

2015-03-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2002?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HDFS-2002:
---
Fix Version/s: (was: 2.0.0-alpha)

 Incorrect computation of needed blocks in getTurnOffTip()
 -

 Key: HDFS-2002
 URL: https://issues.apache.org/jira/browse/HDFS-2002
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 0.22.0
Reporter: Konstantin Shvachko
Assignee: Plamen Jeliazkov
  Labels: newbie
 Fix For: 0.22.0, 0.23.0

 Attachments: HADOOP-2002_TRUNK.patch, hdfs-2002.patch, 
 testsafemode.patch, testsafemode.patch


 {{SafeModeInfo.getTurnOffTip()}} under-reports the number of blocks needed to 
 reach the safemode threshold.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-2012) Recurring failure of TestBalancer due to incorrect treatment of nodes whose utilization equals avgUtilization.

2015-03-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2012?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HDFS-2012:
---
Fix Version/s: (was: 2.0.0-alpha)

 Recurring failure of TestBalancer due to incorrect treatment of nodes whose 
 utilization equals avgUtilization.
 --

 Key: HDFS-2012
 URL: https://issues.apache.org/jira/browse/HDFS-2012
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: balancer  mover, test
Affects Versions: 0.22.0
Reporter: Aaron T. Myers
Assignee: Uma Maheswara Rao G
Priority: Blocker
 Fix For: 0.22.0

 Attachments: HDFS-2012-0.22Branch.patch, HDFS-2012-Trunk.patch, 
 HDFS-2012.patch, TestBalancerLog.html


 This has been failing on Hudson for the last two builds and fails on my local 
 box as well.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-2360) Ugly stacktrace when quota exceeds

2015-03-16 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2360?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J updated HDFS-2360:
--
   Resolution: Fixed
Fix Version/s: 2.8.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

Thanks again Arpit (and Allen).

Failing test was unrelated again, so I went ahead and committed this to 
branch-2 and trunk.

 Ugly stacktrace when quota exceeds
 --

 Key: HDFS-2360
 URL: https://issues.apache.org/jira/browse/HDFS-2360
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 0.23.0
Reporter: Rajit Saha
Assignee: Harsh J
Priority: Minor
 Fix For: 2.8.0

 Attachments: HDFS-2360.patch, HDFS-2360.patch


 Will it be better to catch the exception and throw a small reasonable messege 
 to user when they exceed quota?
 $hdfs  dfs -mkdir testDir
 $hdfs  dfsadmin -setSpaceQuota 191M  testDir
 $hdfs dfs -count -q testDir
 none inf   200278016   2002780161 
0  0
 hdfs://NN hostname:port/user/hdfsqa/testDir
 $hdfs dfs -put /etc/passwd /user/hadoopqa/testDir 
 11/09/19 08:08:15 WARN hdfs.DFSClient: DataStreamer Exception
 org.apache.hadoop.hdfs.protocol.DSQuotaExceededException: The DiskSpace quota 
 of /user/hdfsqa/testDir is exceeded:
 quota=191.0m diskspace consumed=768.0m
 at 
 org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota.verifyQuota(INodeDirectoryWithQuota.java:159)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.verifyQuota(FSDirectory.java:1609)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.updateCount(FSDirectory.java:1383)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.addBlock(FSDirectory.java:370)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allocateBlock(FSNamesystem.java:1681)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1476)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:389)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at 
 org.apache.hadoop.ipc.WritableRpcEngine$Server.call(WritableRpcEngine.java:365)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1496)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1492)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1135)
 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1490)
 at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
 Method)
 at 
 sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
 at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
 at 
 org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:90)
 at 
 org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:57)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1100)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:972)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:454)
 Caused by: org.apache.hadoop.hdfs.protocol.DSQuotaExceededException: The 
 DiskSpace quota of /user/hdfsqa/testDir is
 exceeded: quota=191.0m diskspace consumed=768.0m
 at 
 org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota.verifyQuota(INodeDirectoryWithQuota.java:159)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.verifyQuota(FSDirectory.java:1609)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.updateCount(FSDirectory.java:1383)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.addBlock(FSDirectory.java:370)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allocateBlock(FSNamesystem.java:1681)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1476)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:389)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 

[jira] [Assigned] (HDFS-7616) Change FSImage to support BlockGroup

2015-03-16 Thread Tsz Wo Nicholas Sze (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7616?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo Nicholas Sze reassigned HDFS-7616:
-

Assignee: Takuya Fukudome  (was: Tsz Wo Nicholas Sze)

Sure, assigned to you.  Thanks.

 Change FSImage to support BlockGroup
 

 Key: HDFS-7616
 URL: https://issues.apache.org/jira/browse/HDFS-7616
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Reporter: Tsz Wo Nicholas Sze
Assignee: Takuya Fukudome

 We need to change FSImage for support BlockGroup and other new structures 
 introduced for EC.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-2360) Ugly stacktrace when quota exceeds

2015-03-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-2360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364570#comment-14364570
 ] 

Hudson commented on HDFS-2360:
--

FAILURE: Integrated in Hadoop-trunk-Commit #7340 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/7340/])
HDFS-2360. Ugly stacktrce when quota exceeds. (harsh) (harsh: rev 
046521cd6511b7fc6d9478cb2bed90d8e75fca20)
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSOutputStream.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


 Ugly stacktrace when quota exceeds
 --

 Key: HDFS-2360
 URL: https://issues.apache.org/jira/browse/HDFS-2360
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 0.23.0
Reporter: Rajit Saha
Assignee: Harsh J
Priority: Minor
 Fix For: 2.8.0

 Attachments: HDFS-2360.patch, HDFS-2360.patch


 Will it be better to catch the exception and throw a small reasonable messege 
 to user when they exceed quota?
 $hdfs  dfs -mkdir testDir
 $hdfs  dfsadmin -setSpaceQuota 191M  testDir
 $hdfs dfs -count -q testDir
 none inf   200278016   2002780161 
0  0
 hdfs://NN hostname:port/user/hdfsqa/testDir
 $hdfs dfs -put /etc/passwd /user/hadoopqa/testDir 
 11/09/19 08:08:15 WARN hdfs.DFSClient: DataStreamer Exception
 org.apache.hadoop.hdfs.protocol.DSQuotaExceededException: The DiskSpace quota 
 of /user/hdfsqa/testDir is exceeded:
 quota=191.0m diskspace consumed=768.0m
 at 
 org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota.verifyQuota(INodeDirectoryWithQuota.java:159)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.verifyQuota(FSDirectory.java:1609)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.updateCount(FSDirectory.java:1383)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.addBlock(FSDirectory.java:370)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allocateBlock(FSNamesystem.java:1681)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1476)
 at 
 org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:389)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at 
 org.apache.hadoop.ipc.WritableRpcEngine$Server.call(WritableRpcEngine.java:365)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1496)
 at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1492)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1135)
 at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1490)
 at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
 Method)
 at 
 sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
 at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
 at 
 org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:90)
 at 
 org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:57)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1100)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:972)
 at 
 org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:454)
 Caused by: org.apache.hadoop.hdfs.protocol.DSQuotaExceededException: The 
 DiskSpace quota of /user/hdfsqa/testDir is
 exceeded: quota=191.0m diskspace consumed=768.0m
 at 
 org.apache.hadoop.hdfs.server.namenode.INodeDirectoryWithQuota.verifyQuota(INodeDirectoryWithQuota.java:159)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.verifyQuota(FSDirectory.java:1609)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.updateCount(FSDirectory.java:1383)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSDirectory.addBlock(FSDirectory.java:370)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.allocateBlock(FSNamesystem.java:1681)
 at 
 org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1476)
 at 
 

[jira] [Commented] (HDFS-5522) Datanode disk error check may be incorrectly skipped

2015-03-16 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5522?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364604#comment-14364604
 ] 

Todd Lipcon commented on HDFS-5522:
---

I know this has been closed for a while, but wanted to get some clarification:

{quote}
If I/O hangs for a long time, network read/write may timeout first and the peer 
may close the connection. Although the error was caused by a faulty local disk, 
disk check is not being carried out in this case
{quote}

It seems like this JIRA applied a rather heavy hammer for a specific case that 
could be better identified in another way. After applying this patch, it seems 
that DNs will run checkDiskError when any other node experiences an issue. So, 
if one node is down (eg due to a rolling restart or a crash) all of the other 
nodes are very soon running checkDiskError for no particularly good reason. 
Coupled with HDFS-7489, this failure can also cascade.

Can you describe in more detail the scenario you were facing that inspired this 
JIRA? Would it not make more sense to actually look for the underlying symptom 
(by adding a timer around the I/O, perhaps?) and running checkDiskError in the 
specific scenarios we're looking for, rather than all network errors?

 Datanode disk error check may be incorrectly skipped
 

 Key: HDFS-5522
 URL: https://issues.apache.org/jira/browse/HDFS-5522
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.23.9, 2.2.0
Reporter: Kihwal Lee
Assignee: Rushabh S Shah
 Fix For: 2.5.0

 Attachments: HDFS-5522-v2.patch, HDFS-5522-v3.patch, HDFS-5522.patch


 After HDFS-4581 and HDFS-4699, {{checkDiskError()}} is not called when 
 network errors occur during processing data node requests.  This appears to 
 create problems when a disk is having problems, but not failing I/O soon. 
 If I/O hangs for a long time, network read/write may timeout first and the 
 peer may close the connection. Although the error was caused by a faulty 
 local disk, disk check is not being carried out in this case. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7930) commitBlockSynchronization() does not remove locations

2015-03-16 Thread Konstantin Shvachko (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7930?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364621#comment-14364621
 ] 

Konstantin Shvachko commented on HDFS-7930:
---

Looks like the right fix. Few suggestions:
# Merge test cases {{testTruncateWithDataNodesRestart()}} and 
{{testTruncateWithDataNodesRestart1()}} into one. The latter seems to be a more 
detailed case of the former.
# Also you should use {{restartDataNode()}} with {{expireOnNN == true}} now 
that HDFS-7886 is in.
# Remove misleading comment in {{FSN.commitBlockSync()}}
{{// There should be no locations in the blockManager till now }}
# Few improvements to {{markBlockReplicasAsCorrupt()}}
#- {{excludedStorages}} should be {{newStorages}} or {{committedStorages}}
#- Better keep the parameter declaration on a single line
#- {{toMark}} could be {{isCorrupt}}
#- Just {{else return;}} when the genStamp and the length are the same as the 
old ones.
#- Redundant check {{excludedStorages.length  0}}

 commitBlockSynchronization() does not remove locations
 --

 Key: HDFS-7930
 URL: https://issues.apache.org/jira/browse/HDFS-7930
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.7.0
Reporter: Konstantin Shvachko
Assignee: Yi Liu
Priority: Blocker
 Attachments: HDFS-7930.001.patch


 When {{commitBlockSynchronization()}} has less {{newTargets}} than in the 
 original block it does not remove unconfirmed locations. This results in that 
 the the block stores locations of different lengths or genStamp (corrupt).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7717) Erasure Coding: provide a tool for convert files between replication and erasure coding

2015-03-16 Thread Kai Sasaki (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7717?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364641#comment-14364641
 ] 

Kai Sasaki commented on HDFS-7717:
--

[~jingzhao] Hello, Jing. I'd like to work on this issue. Could you assign it to 
me? Thank you.

 Erasure Coding: provide a tool for convert files between replication and 
 erasure coding
 ---

 Key: HDFS-7717
 URL: https://issues.apache.org/jira/browse/HDFS-7717
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jing Zhao
Assignee: Jing Zhao

 We need a tool to do offline conversion between replication and erasure 
 coding. The tool itself can either utilize MR just like the current distcp, 
 or act like the balancer/mover. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7864) Erasure Coding: Update safemode calculation for striped blocks

2015-03-16 Thread GAO Rui (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7864?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364536#comment-14364536
 ] 

GAO Rui commented on HDFS-7864:
---

[~jingzhao]

Thanks for your comment, Jing! 
  I have updated a new patch which used {{getDataBlockNum()}} to obtain the 
minimal safe number for striped block. And I also modified code style according 
to code convention as possible as I can.  Please let me know if there are any 
future problems.
  Please review the patch when you are free, thank you very much!
  And can you recommend another Jira for me to work on? Thank you!



 Erasure Coding: Update safemode calculation for striped blocks
 --

 Key: HDFS-7864
 URL: https://issues.apache.org/jira/browse/HDFS-7864
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jing Zhao
Assignee: GAO Rui
 Attachments: HDFS-7864.1.patch, HDFS-7864.2.patch, HDFS-7864.3.patch


 We need to update the safemode calculation for striped blocks. Specifically, 
 each striped block now consists of multiple data/parity blocks stored in 
 corresponding DataNodes. The current code's calculation is thus inconsistent: 
 each striped block is only counted as 1 expected block, while each of its 
 member block may increase the number of received blocks by 1.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7829) Code clean up for LocatedBlock

2015-03-16 Thread Takanobu Asanuma (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7829?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Takanobu Asanuma updated HDFS-7829:
---
Attachment: HDFS-7829.4.patch

Thank you for your review, Jing. I thought and changed my patch:

1. I removed some constructors:{{LocatedBlock(ExtendedBlock, DatanodeInfo[], 
long, boolean)}},{{LocatedBlock(ExtendedBlock, DatanodeStorageInfo[])}} and 
changed callers.
And I kept {{LocatedBlock(ExtendedBlock b, DatanodeInfo[], String[], 
StorageType[])}}.

2. Since {{setStartOffset}} and {{setCorrupt}} are used by tests out of the 
package, I set public to them.

3. I set final to {{storageIDs}} and {{storageTypes}}.

How does that look?

 Code clean up for LocatedBlock
 --

 Key: HDFS-7829
 URL: https://issues.apache.org/jira/browse/HDFS-7829
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Jing Zhao
Assignee: Takanobu Asanuma
Priority: Minor
 Attachments: HDFS-7829.1.patch, HDFS-7829.2.patch, HDFS-7829.3.patch, 
 HDFS-7829.4.patch


 We can do some code cleanup for {{LocatedBlock}}, including:
 # Using a simple Builder pattern to avoid multiple constructors
 # Setting data fields like {{corrupt}} and {{offset}} to final



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7864) Erasure Coding: Update safemode calculation for striped blocks

2015-03-16 Thread GAO Rui (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7864?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

GAO Rui updated HDFS-7864:
--
Attachment: HDFS-7864.3.patch

 Erasure Coding: Update safemode calculation for striped blocks
 --

 Key: HDFS-7864
 URL: https://issues.apache.org/jira/browse/HDFS-7864
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jing Zhao
Assignee: GAO Rui
 Attachments: HDFS-7864.1.patch, HDFS-7864.2.patch, HDFS-7864.3.patch


 We need to update the safemode calculation for striped blocks. Specifically, 
 each striped block now consists of multiple data/parity blocks stored in 
 corresponding DataNodes. The current code's calculation is thus inconsistent: 
 each striped block is only counted as 1 expected block, while each of its 
 member block may increase the number of received blocks by 1.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


  1   2   >