[jira] [Commented] (HDFS-11183) Document XOR policy XOR-2-1-64k support in HDFS

2017-03-06 Thread SammiChen (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11183?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898856#comment-15898856
 ] 

SammiChen commented on HDFS-11183:
--

Hi, [~andrew.wang], I see you have mentioned XOR-2-1-64K policy in 
HDFSErasureCoding.md in last check-in for HDFS-11427. So I'm going to close 
this JIRA.   Agree? 

> Document XOR policy XOR-2-1-64k support in HDFS
> ---
>
> Key: HDFS-11183
> URL: https://issues.apache.org/jira/browse/HDFS-11183
> Project: Hadoop HDFS
>  Issue Type: Task
>Affects Versions: 3.0.0-alpha2
>Reporter: SammiChen
>Assignee: SammiChen
>  Labels: hdfs-ec-3.0-must-do
>
> A new EC policy XOR-2-1-64k is added in HDFS. This task is to update HDFS 
> related documents about how to use this policy



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-9705) Refine the behaviour of getFileChecksum when length = 0

2017-03-06 Thread SammiChen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9705?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

SammiChen updated HDFS-9705:

Attachment: HDFS-9705-v7.patch

The previous comment is kind of misleading. When the crcType is "NULL" which 
means no checksum for files, the execution will go through the "default" branch 
of {makeFinalResult} function.

> Refine the behaviour of getFileChecksum when length = 0
> ---
>
> Key: HDFS-9705
> URL: https://issues.apache.org/jira/browse/HDFS-9705
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Kai Zheng
>Assignee: Kai Zheng
>Priority: Minor
> Attachments: HDFS-9705-v1.patch, HDFS-9705-v2.patch, 
> HDFS-9705-v3.patch, HDFS-9705-v4.patch, HDFS-9705-v5.patch, 
> HDFS-9705-v6.patch, HDFS-9705-v7.patch
>
>
> {{FileSystem#getFileChecksum}} may accept {{length}} parameter and 0 is a 
> valid value. Currently it will return {{null}} when length is 0, in the 
> following code block:
> {code}
> //compute file MD5
> final MD5Hash fileMD5 = MD5Hash.digest(md5out.getData());
> switch (crcType) {
> case CRC32:
>   return new MD5MD5CRC32GzipFileChecksum(bytesPerCRC,
>   crcPerBlock, fileMD5);
> case CRC32C:
>   return new MD5MD5CRC32CastagnoliFileChecksum(bytesPerCRC,
>   crcPerBlock, fileMD5);
> default:
>   // If there is no block allocated for the file,
>   // return one with the magic entry that matches what previous
>   // hdfs versions return.
>   if (locatedblocks.size() == 0) {
> return new MD5MD5CRC32GzipFileChecksum(0, 0, fileMD5);
>   }
>   // we should never get here since the validity was checked
>   // when getCrcType() was called above.
>   return null;
> }
> {code}
> The comment says "we should never get here since the validity was checked" 
> but it does. As we're using the MD5-MD5-X approach, and {{EMPTY--CONTENT}} 
> actually is a valid case in which the MD5 value is 
> {{d41d8cd98f00b204e9800998ecf8427e}}, so suggest we return a reasonable value 
> other than null. At least some useful information in the returned value can 
> be seen, like values from block checksum header.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HDFS-9705) Refine the behaviour of getFileChecksum when length = 0

2017-03-06 Thread SammiChen (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9705?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898838#comment-15898838
 ] 

SammiChen edited comment on HDFS-9705 at 3/7/17 6:45 AM:
-

The previous comment is kind of misleading. When the crcType is "NULL" which 
means no checksum for files, the execution will go through the "default" branch 
of {{makeFinalResult}} function.


was (Author: sammi):
The previous comment is kind of misleading. When the crcType is "NULL" which 
means no checksum for files, the execution will go through the "default" branch 
of {makeFinalResult} function.

> Refine the behaviour of getFileChecksum when length = 0
> ---
>
> Key: HDFS-9705
> URL: https://issues.apache.org/jira/browse/HDFS-9705
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Kai Zheng
>Assignee: Kai Zheng
>Priority: Minor
> Attachments: HDFS-9705-v1.patch, HDFS-9705-v2.patch, 
> HDFS-9705-v3.patch, HDFS-9705-v4.patch, HDFS-9705-v5.patch, 
> HDFS-9705-v6.patch, HDFS-9705-v7.patch
>
>
> {{FileSystem#getFileChecksum}} may accept {{length}} parameter and 0 is a 
> valid value. Currently it will return {{null}} when length is 0, in the 
> following code block:
> {code}
> //compute file MD5
> final MD5Hash fileMD5 = MD5Hash.digest(md5out.getData());
> switch (crcType) {
> case CRC32:
>   return new MD5MD5CRC32GzipFileChecksum(bytesPerCRC,
>   crcPerBlock, fileMD5);
> case CRC32C:
>   return new MD5MD5CRC32CastagnoliFileChecksum(bytesPerCRC,
>   crcPerBlock, fileMD5);
> default:
>   // If there is no block allocated for the file,
>   // return one with the magic entry that matches what previous
>   // hdfs versions return.
>   if (locatedblocks.size() == 0) {
> return new MD5MD5CRC32GzipFileChecksum(0, 0, fileMD5);
>   }
>   // we should never get here since the validity was checked
>   // when getCrcType() was called above.
>   return null;
> }
> {code}
> The comment says "we should never get here since the validity was checked" 
> but it does. As we're using the MD5-MD5-X approach, and {{EMPTY--CONTENT}} 
> actually is a valid case in which the MD5 value is 
> {{d41d8cd98f00b204e9800998ecf8427e}}, so suggest we return a reasonable value 
> other than null. At least some useful information in the returned value can 
> be seen, like values from block checksum header.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11158) Ozone: TestDatanodeStateMachine is failing intermittently

2017-03-06 Thread Anu Engineer (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11158?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898797#comment-15898797
 ] 

Anu Engineer commented on HDFS-11158:
-

+1, I will add another test case to make sure we can launch the state machine 
in the thread. It is more important to get a consistent working test case 
without race conditions.

> Ozone: TestDatanodeStateMachine is failing intermittently
> -
>
> Key: HDFS-11158
> URL: https://issues.apache.org/jira/browse/HDFS-11158
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>Assignee: Weiwei Yang
> Attachments: HDFS-11158-HDFS-7240.001.patch
>
>
> Failed in HDFS-11138 patch v6 run



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11451) Ozone: Add protobuf definitions for container reports

2017-03-06 Thread Anu Engineer (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898788#comment-15898788
 ] 

Anu Engineer commented on HDFS-11451:
-

[~xyao] verbally approved checking the patch v4 since jenkins is not working 
and all tests pass locally. Committing based on that.


> Ozone: Add protobuf definitions for container reports
> -
>
> Key: HDFS-11451
> URL: https://issues.apache.org/jira/browse/HDFS-11451
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>Assignee: Anu Engineer
> Fix For: HDFS-7240
>
> Attachments: HDFS-11451-HDFS-7240.001.patch, 
> HDFS-11451-HDFS-7240.002.patch, HDFS-11451-HDFS-7240.003.patch, 
> HDFS-11451-HDFS-7240.004.patch
>
>
> Container reports are send from datanodes describing the state of the 
> container. Since a full container report can be as large as 2 MB, the 
> container reports are send by datanode only when SCM approves it.
> In this change, datanode informs the SCM that it has a container report 
> ready. SCM based on load will send a command to datanode to send the actual 
> report. The protobuf classes and plumbing required for that change is part of 
> this patch. This whole container reports handling will be broken into 
> multiple JIRAs to make it easy to code review.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Assigned] (HDFS-11158) Ozone: TestDatanodeStateMachine is failing intermittently

2017-03-06 Thread Anu Engineer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11158?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer reassigned HDFS-11158:
---

Assignee: Weiwei Yang  (was: Anu Engineer)

> Ozone: TestDatanodeStateMachine is failing intermittently
> -
>
> Key: HDFS-11158
> URL: https://issues.apache.org/jira/browse/HDFS-11158
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>Assignee: Weiwei Yang
> Attachments: HDFS-11158-HDFS-7240.001.patch
>
>
> Failed in HDFS-11138 patch v6 run



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11508) Fix bind failure in SimpleTCPServer & Portmap where bind fails because socket is in TIME_WAIT state

2017-03-06 Thread Mukul Kumar Singh (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mukul Kumar Singh updated HDFS-11508:
-
Status: Patch Available  (was: Open)

> Fix bind failure in SimpleTCPServer & Portmap where bind fails because socket 
> is in TIME_WAIT state
> ---
>
> Key: HDFS-11508
> URL: https://issues.apache.org/jira/browse/HDFS-11508
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Reporter: Mukul Kumar Singh
>Assignee: Mukul Kumar Singh
> Attachments: HDFS-11508.001.patch
>
>
> Bind can fail in SimpleTCPServer & Portmap because socket is in TIME_WAIT 
> state.
> Socket options should be changed here to use the setReuseAddress option.
> {noformat}
> 2017-03-05 08:37:13,283 INFO  oncrpc.SimpleUdpServer 
> (SimpleUdpServer.java:run(73)) - Started listening to UDP requests at port 
> 4242 for Rpc program: mountd at localhost:4242 with workerCount 1
> 2017-03-05 08:37:13,298 FATAL mount.MountdBase 
> (MountdBase.java:startTCPServer(85)) - Failed to start the TCP server.
> org.jboss.netty.channel.ChannelException: Failed to bind to: 
> 0.0.0.0/0.0.0.0:4242
>   at 
> org.jboss.netty.bootstrap.ServerBootstrap.bind(ServerBootstrap.java:272)
>   at org.apache.hadoop.oncrpc.SimpleTcpServer.run(SimpleTcpServer.java:86)
>   at org.apache.hadoop.mount.MountdBase.startTCPServer(MountdBase.java:83)
>   at org.apache.hadoop.mount.MountdBase.start(MountdBase.java:98)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startServiceInternal(Nfs3.java:56)
>   at org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startService(Nfs3.java:69)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.PrivilegedNfsGatewayStarter.start(PrivilegedNfsGatewayStarter.java:71)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.apache.commons.daemon.support.DaemonLoader.start(DaemonLoader.java:243)
> Caused by: java.net.BindException: Address already in use
>   at sun.nio.ch.Net.bind0(Native Method)
>   at sun.nio.ch.Net.bind(Net.java:433)
>   at sun.nio.ch.Net.bind(Net.java:425)
>   at 
> sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:223)
>   at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
>   at 
> org.jboss.netty.channel.socket.nio.NioServerBoss$RegisterTask.run(NioServerBoss.java:193)
>   at 
> org.jboss.netty.channel.socket.nio.AbstractNioSelector.processTaskQueue(AbstractNioSelector.java:366)
>   at 
> org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:290)
>   at 
> org.jboss.netty.channel.socket.nio.NioServerBoss.run(NioServerBoss.java:42)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)
> 2017-03-05 08:37:13,304 INFO  util.ExitUtil (ExitUtil.java:terminate(124)) - 
> Exiting with status 1
> 2017-03-05 08:37:13,308 INFO  nfs3.Nfs3Base (LogAdapter.java:info(45)) - 
> SHUTDOWN_MSG: 
> /
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11508) Fix bind failure in SimpleTCPServer & Portmap where bind fails because socket is in TIME_WAIT state

2017-03-06 Thread Mukul Kumar Singh (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mukul Kumar Singh updated HDFS-11508:
-
Attachment: HDFS-11508.001.patch

Tested this patch manually by starting NFS gateway and portmap multiple times

> Fix bind failure in SimpleTCPServer & Portmap where bind fails because socket 
> is in TIME_WAIT state
> ---
>
> Key: HDFS-11508
> URL: https://issues.apache.org/jira/browse/HDFS-11508
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: nfs
>Reporter: Mukul Kumar Singh
>Assignee: Mukul Kumar Singh
> Attachments: HDFS-11508.001.patch
>
>
> Bind can fail in SimpleTCPServer & Portmap because socket is in TIME_WAIT 
> state.
> Socket options should be changed here to use the setReuseAddress option.
> {noformat}
> 2017-03-05 08:37:13,283 INFO  oncrpc.SimpleUdpServer 
> (SimpleUdpServer.java:run(73)) - Started listening to UDP requests at port 
> 4242 for Rpc program: mountd at localhost:4242 with workerCount 1
> 2017-03-05 08:37:13,298 FATAL mount.MountdBase 
> (MountdBase.java:startTCPServer(85)) - Failed to start the TCP server.
> org.jboss.netty.channel.ChannelException: Failed to bind to: 
> 0.0.0.0/0.0.0.0:4242
>   at 
> org.jboss.netty.bootstrap.ServerBootstrap.bind(ServerBootstrap.java:272)
>   at org.apache.hadoop.oncrpc.SimpleTcpServer.run(SimpleTcpServer.java:86)
>   at org.apache.hadoop.mount.MountdBase.startTCPServer(MountdBase.java:83)
>   at org.apache.hadoop.mount.MountdBase.start(MountdBase.java:98)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startServiceInternal(Nfs3.java:56)
>   at org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startService(Nfs3.java:69)
>   at 
> org.apache.hadoop.hdfs.nfs.nfs3.PrivilegedNfsGatewayStarter.start(PrivilegedNfsGatewayStarter.java:71)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.apache.commons.daemon.support.DaemonLoader.start(DaemonLoader.java:243)
> Caused by: java.net.BindException: Address already in use
>   at sun.nio.ch.Net.bind0(Native Method)
>   at sun.nio.ch.Net.bind(Net.java:433)
>   at sun.nio.ch.Net.bind(Net.java:425)
>   at 
> sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:223)
>   at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
>   at 
> org.jboss.netty.channel.socket.nio.NioServerBoss$RegisterTask.run(NioServerBoss.java:193)
>   at 
> org.jboss.netty.channel.socket.nio.AbstractNioSelector.processTaskQueue(AbstractNioSelector.java:366)
>   at 
> org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:290)
>   at 
> org.jboss.netty.channel.socket.nio.NioServerBoss.run(NioServerBoss.java:42)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)
> 2017-03-05 08:37:13,304 INFO  util.ExitUtil (ExitUtil.java:terminate(124)) - 
> Exiting with status 1
> 2017-03-05 08:37:13,308 INFO  nfs3.Nfs3Base (LogAdapter.java:info(45)) - 
> SHUTDOWN_MSG: 
> /
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11465) Rename BlockType#CONTIGUOUS to BlockType#REPLICATED

2017-03-06 Thread Yuanbo Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11465?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yuanbo Liu updated HDFS-11465:
--
Status: Patch Available  (was: Open)

> Rename BlockType#CONTIGUOUS to BlockType#REPLICATED
> ---
>
> Key: HDFS-11465
> URL: https://issues.apache.org/jira/browse/HDFS-11465
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: 3.0.0-alpha2
>Reporter: Andrew Wang
>Assignee: Yuanbo Liu
> Attachments: HDFS-11465.001.patch
>
>
> HDFS-10759 introduced the a BlockType enum to INodeFile, with possible values 
> CONTIGUOUS or STRIPED.
> Since HDFS-8030 wants to implement "contiguous EC", CONTIGUOUS isn't an 
> appropriate name. I propose we rename CONTIGUOUS to REPLICATED for clarity.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11465) Rename BlockType#CONTIGUOUS to BlockType#REPLICATED

2017-03-06 Thread Yuanbo Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11465?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yuanbo Liu updated HDFS-11465:
--
Attachment: HDFS-11465.001.patch

upload v1 patch for this JIRA, basically there are two changes here:
* rename BlockType#CONTIGUOUS to BlockType#REPLICATED
* rename BlockInfoContiguous.java to BlockInfoReplicated.java

> Rename BlockType#CONTIGUOUS to BlockType#REPLICATED
> ---
>
> Key: HDFS-11465
> URL: https://issues.apache.org/jira/browse/HDFS-11465
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: 3.0.0-alpha2
>Reporter: Andrew Wang
>Assignee: Yuanbo Liu
> Attachments: HDFS-11465.001.patch
>
>
> HDFS-10759 introduced the a BlockType enum to INodeFile, with possible values 
> CONTIGUOUS or STRIPED.
> Since HDFS-8030 wants to implement "contiguous EC", CONTIGUOUS isn't an 
> appropriate name. I propose we rename CONTIGUOUS to REPLICATED for clarity.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11497) Ozone: Fix Container test regression.

2017-03-06 Thread Anu Engineer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11497?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer updated HDFS-11497:

  Resolution: Fixed
Hadoop Flags: Reviewed
   Fix Version/s: HDFS-7240
Target Version/s: HDFS-7240
  Status: Resolved  (was: Patch Available)

[~xyao] Thanks for the reviews. I have committed this to the feature branch.

> Ozone: Fix Container test regression. 
> --
>
> Key: HDFS-11497
> URL: https://issues.apache.org/jira/browse/HDFS-11497
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>Assignee: Anu Engineer
> Fix For: HDFS-7240
>
> Attachments: HDFS-11497-HDFS-7240.001.patch
>
>
> One the recent commits caused a regression in the container HB path. This 
> JIRA fixes it. Thanks to [~vagarychen] for bringing it to my attention.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Masatake Iwasaki (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898769#comment-15898769
 ] 

Masatake Iwasaki commented on HDFS-11499:
-

Thanks for working on this, [~lukmajercak] and [~manojg].

While testing the 03 patch, the added 
testDecommissionWithOpenFileAndDatanodeFailing intermittently timeouts waiting 
for decomission. I'm looking into the cause.


> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.02.patch, HDFS-11499.03.patch, 
> HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-10838) Last full block report received time for each DN should be easily discoverable

2017-03-06 Thread Surendra Singh Lilhore (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898770#comment-15898770
 ] 

Surendra Singh Lilhore commented on HDFS-10838:
---

Thanks [~arpitagarwal] for review and commit ...

> Last full block report received time for each DN should be easily discoverable
> --
>
> Key: HDFS-10838
> URL: https://issues.apache.org/jira/browse/HDFS-10838
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: ui
>Reporter: Arpit Agarwal
>Assignee: Surendra Singh Lilhore
> Fix For: 2.9.0, 3.0.0-alpha3
>
> Attachments: DFSAdmin-Report.png, HDFS-10838-001.patch, 
> HDFS-10838.002.patch, HDFS-10838.003.patch, HDFS-10838.004.patch, 
> HDFS-10838.005.patch, HDFS-10838.006.patch, NN_UI.png, 
> NN_UI_relative_time_minutes.png, NN_UI_relative_time.png
>
>
> It should be easy for administrators to discover the time of last full block 
> report from each DataNode.
> We can show it in the NameNode web UI or in the output of {{hdfs dfsadmin 
> -report}}, or both.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11497) Ozone: Fix Container test regression.

2017-03-06 Thread Anu Engineer (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11497?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898763#comment-15898763
 ] 

Anu Engineer commented on HDFS-11497:
-

None of these failures are reproducible  on local machine, and it looks like 
Jenkins machines ran out of memory. So I am going to commit this fix.


> Ozone: Fix Container test regression. 
> --
>
> Key: HDFS-11497
> URL: https://issues.apache.org/jira/browse/HDFS-11497
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>Assignee: Anu Engineer
> Attachments: HDFS-11497-HDFS-7240.001.patch
>
>
> One the recent commits caused a regression in the container HB path. This 
> JIRA fixes it. Thanks to [~vagarychen] for bringing it to my attention.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Lukas Majercak (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lukas Majercak updated HDFS-11499:
--
Hadoop Flags: Reviewed
  Status: Patch Available  (was: Open)

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 3.0.0-alpha2, 2.7.3, 2.7.2, 2.7.1
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.02.patch, HDFS-11499.03.patch, 
> HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Lukas Majercak (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lukas Majercak updated HDFS-11499:
--
Status: Open  (was: Patch Available)

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 3.0.0-alpha2, 2.7.3, 2.7.2, 2.7.1
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.02.patch, HDFS-11499.03.patch, 
> HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Created] (HDFS-11508) Fix bind failure in SimpleTCPServer & Portmap where bind fails because socket is in TIME_WAIT state

2017-03-06 Thread Mukul Kumar Singh (JIRA)
Mukul Kumar Singh created HDFS-11508:


 Summary: Fix bind failure in SimpleTCPServer & Portmap where bind 
fails because socket is in TIME_WAIT state
 Key: HDFS-11508
 URL: https://issues.apache.org/jira/browse/HDFS-11508
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: nfs
Reporter: Mukul Kumar Singh
Assignee: Mukul Kumar Singh


Bind can fail in SimpleTCPServer & Portmap because socket is in TIME_WAIT state.

Socket options should be changed here to use the setReuseAddress option.

{noformat}
2017-03-05 08:37:13,283 INFO  oncrpc.SimpleUdpServer 
(SimpleUdpServer.java:run(73)) - Started listening to UDP requests at port 4242 
for Rpc program: mountd at localhost:4242 with workerCount 1
2017-03-05 08:37:13,298 FATAL mount.MountdBase 
(MountdBase.java:startTCPServer(85)) - Failed to start the TCP server.
org.jboss.netty.channel.ChannelException: Failed to bind to: 
0.0.0.0/0.0.0.0:4242
at 
org.jboss.netty.bootstrap.ServerBootstrap.bind(ServerBootstrap.java:272)
at org.apache.hadoop.oncrpc.SimpleTcpServer.run(SimpleTcpServer.java:86)
at org.apache.hadoop.mount.MountdBase.startTCPServer(MountdBase.java:83)
at org.apache.hadoop.mount.MountdBase.start(MountdBase.java:98)
at 
org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startServiceInternal(Nfs3.java:56)
at org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startService(Nfs3.java:69)
at 
org.apache.hadoop.hdfs.nfs.nfs3.PrivilegedNfsGatewayStarter.start(PrivilegedNfsGatewayStarter.java:71)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.commons.daemon.support.DaemonLoader.start(DaemonLoader.java:243)
Caused by: java.net.BindException: Address already in use
at sun.nio.ch.Net.bind0(Native Method)
at sun.nio.ch.Net.bind(Net.java:433)
at sun.nio.ch.Net.bind(Net.java:425)
at 
sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:223)
at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
at 
org.jboss.netty.channel.socket.nio.NioServerBoss$RegisterTask.run(NioServerBoss.java:193)
at 
org.jboss.netty.channel.socket.nio.AbstractNioSelector.processTaskQueue(AbstractNioSelector.java:366)
at 
org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:290)
at 
org.jboss.netty.channel.socket.nio.NioServerBoss.run(NioServerBoss.java:42)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
2017-03-05 08:37:13,304 INFO  util.ExitUtil (ExitUtil.java:terminate(124)) - 
Exiting with status 1
2017-03-05 08:37:13,308 INFO  nfs3.Nfs3Base (LogAdapter.java:info(45)) - 
SHUTDOWN_MSG: 
/
{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11502) dn.js set datanode UI to window.location.hostname, it should use jmx "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get tag.hostname

2017-03-06 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11502?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898730#comment-15898730
 ] 

Hadoop QA commented on HDFS-11502:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
16s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 1 line(s) that end in whitespace. Use git 
apply --whitespace=fix <>. Refer https://git-scm.com/docs/git-apply 
{color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
19s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}  0m 50s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | HDFS-11502 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12856430/HDFS-11502.001.patch |
| Optional Tests |  asflicense  |
| uname | Linux d535ef6d5eb9 3.13.0-107-generic #154-Ubuntu SMP Tue Dec 20 
09:57:27 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 1441398 |
| whitespace | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18610/artifact/patchprocess/whitespace-eol.txt
 |
| modules | C: hadoop-hdfs-project/hadoop-hdfs U: 
hadoop-hdfs-project/hadoop-hdfs |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18610/console |
| Powered by | Apache Yetus 0.5.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> dn.js set datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname
> --
>
> Key: HDFS-11502
> URL: https://issues.apache.org/jira/browse/HDFS-11502
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.7.2, 2.7.3
> Environment: all
>Reporter: Jeffrey E  Rodriguez
>Assignee: Jeffrey E  Rodriguez
> Fix For: 2.8.0, 2.7.3, 3.0.0-alpha1, 3.0.0-alpha2
>
> Attachments: HDFS-11502.001.patch
>
>
> Datanode UI calls dn.js which loads properties for datanode..
> dn.js sets datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname.
> The issue is that if we proxy datanode UI we would get the proxy as the 
> hostname because UI code uses window.location.hostname instead of the actual 
> datanode hostname.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11502) dn.js set datanode UI to window.location.hostname, it should use jmx "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get tag.hostname

2017-03-06 Thread Jeffrey E Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11502?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jeffrey E  Rodriguez updated HDFS-11502:

Fix Version/s: 2.7.3
   3.0.0-alpha1
   3.0.0-alpha2
   2.8.0
   Status: Patch Available  (was: In Progress)

> dn.js set datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname
> --
>
> Key: HDFS-11502
> URL: https://issues.apache.org/jira/browse/HDFS-11502
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.7.3, 2.7.2
> Environment: all
>Reporter: Jeffrey E  Rodriguez
>Assignee: Jeffrey E  Rodriguez
> Fix For: 2.8.0, 3.0.0-alpha2, 3.0.0-alpha1, 2.7.3
>
> Attachments: HDFS-11502.001.patch
>
>
> Datanode UI calls dn.js which loads properties for datanode..
> dn.js sets datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname.
> The issue is that if we proxy datanode UI we would get the proxy as the 
> hostname because UI code uses window.location.hostname instead of the actual 
> datanode hostname.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11502) dn.js set datanode UI to window.location.hostname, it should use jmx "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get tag.hostname

2017-03-06 Thread Jeffrey E Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11502?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jeffrey E  Rodriguez updated HDFS-11502:

Attachment: HDFS-11502.001.patch

Patch fixes Datanode html UI so we use datahostname tag.Hostname from 
JvmMetrics jmx instead of relying on window.location.hostname. 
window.location.hostname which in a case where we proxy datanode UI would 
return host of proxy instead.
Since we want the real hostname of datanode it may be preferable to use actual 
datanode jmx data. 

> dn.js set datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname
> --
>
> Key: HDFS-11502
> URL: https://issues.apache.org/jira/browse/HDFS-11502
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.7.2, 2.7.3
> Environment: all
>Reporter: Jeffrey E  Rodriguez
>Assignee: Jeffrey E  Rodriguez
> Attachments: HDFS-11502.001.patch
>
>
> Datanode UI calls dn.js which loads properties for datanode..
> dn.js sets datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname.
> The issue is that if we proxy datanode UI we would get the proxy as the 
> hostname because UI code uses window.location.hostname instead of the actual 
> datanode hostname.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work started] (HDFS-11502) dn.js set datanode UI to window.location.hostname, it should use jmx "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get tag.hostname

2017-03-06 Thread Jeffrey E Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11502?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HDFS-11502 started by Jeffrey E  Rodriguez.
---
> dn.js set datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname
> --
>
> Key: HDFS-11502
> URL: https://issues.apache.org/jira/browse/HDFS-11502
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.7.2, 2.7.3
> Environment: all
>Reporter: Jeffrey E  Rodriguez
>Assignee: Jeffrey E  Rodriguez
>
> Datanode UI calls dn.js which loads properties for datanode..
> dn.js sets datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname.
> The issue is that if we proxy datanode UI we would get the proxy as the 
> hostname because UI code uses window.location.hostname instead of the actual 
> datanode hostname.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Yiqun Lin (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898700#comment-15898700
 ] 

Yiqun Lin commented on HDFS-11499:
--

You are right, I misreading for this. +1 for the patch. Pending Jenkins.

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.02.patch, HDFS-11499.03.patch, 
> HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-10983) OIV tool should make an EC file explicit

2017-03-06 Thread Manoj Govindassamy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10983?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Manoj Govindassamy updated HDFS-10983:
--
Attachment: HDFS-10983.03.patch

[~andrew.wang],

Removed the explicit replication field write for Striped files in FSImage at 
{{FSImageFormatPBINode#buildPermissionStatus}}. And extended the  
{{FSImageFormatPBINode#loadINodeFile}} with an assert to verify that FSImage 
doesn't have replication field set for Striped files. With this new 
verification {{TestFSImage#testSupportBlockGroup}} fails at FSImage load for EC 
Files when the replication fields are additionally set.  

{code}
   @@ -332,6 +332,7 @@ private INodeFile loadINodeFile(INodeSection.INode n) {
  assert n.getType() == INodeSection.INode.Type.FILE;
..
  BlockType blockType = PBHelperClient.convert(f.getBlockType());
  LoaderContext state = parent.getLoaderContext();
  boolean isStriped = f.hasErasureCodingPolicyID();
   + assert ((!isStriped) || (isStriped && !f.hasReplication()));
 Short replication = (!isStriped ? (short) f.getReplication() : null);
{code}

Currently, FSImage Loaders and Savers are not exposed properly for testing and 
having an external test for verifying the fsimage inode_section image contents 
seems not feasible. So, extending the {{loadINodeFile()}} with assert like few 
other asserts in the same file.  Please let me know if this is ok ?

> OIV tool should make an EC file explicit
> 
>
> Key: HDFS-10983
> URL: https://issues.apache.org/jira/browse/HDFS-10983
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 3.0.0-alpha1
>Reporter: Wei-Chiu Chuang
>Assignee: Manoj Govindassamy
>  Labels: hdfs-ec-3.0-nice-to-have
> Attachments: HDFS-10983.01.patch, HDFS-10983.02.patch, 
> HDFS-10983.03.patch
>
>
> The OIV tool's webhdfs interface does not print if a file is striped or not.
> Also, it prints the file's EC policy ID as replication factor, which is 
> inconsistent to the output of a typical webhdfs call to the cluster, which 
> always shows replication factor of 0 for EC files.
> Not just webhdfs, but delimiter output does not print if a file is stripped 
> or not either.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Assigned] (HDFS-11502) dn.js set datanode UI to window.location.hostname, it should use jmx "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get tag.hostname

2017-03-06 Thread Brahma Reddy Battula (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11502?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brahma Reddy Battula reassigned HDFS-11502:
---

Assignee: Jeffrey E  Rodriguez

> dn.js set datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname
> --
>
> Key: HDFS-11502
> URL: https://issues.apache.org/jira/browse/HDFS-11502
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.7.2, 2.7.3
> Environment: all
>Reporter: Jeffrey E  Rodriguez
>Assignee: Jeffrey E  Rodriguez
>
> Datanode UI calls dn.js which loads properties for datanode..
> dn.js sets datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname.
> The issue is that if we proxy datanode UI we would get the proxy as the 
> hostname because UI code uses window.location.hostname instead of the actual 
> datanode hostname.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Manoj Govindassamy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898651#comment-15898651
 ] 

Manoj Govindassamy commented on HDFS-11499:
---

[~linyiqun], 

Thanks for the review. I was using the below {{writeFile}} method version, 
where the last param is the file replication factor and not block count. The 
method in turn creates a total of 2 blocks for the file with the provided repl 
factor. Is the comment still wrong ? please let me know.

{code}
  static protected void writeFile(FileSystem fileSys, Path name, int repl)
  throws IOException {
writeFile(fileSys, name, repl, 2);
  }
{code}

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.02.patch, HDFS-11499.03.patch, 
> HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11071) Ozone: SCM: Move SCM config keys to ScmConfig

2017-03-06 Thread Weiwei Yang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11071?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898634#comment-15898634
 ] 

Weiwei Yang commented on HDFS-11071:


There seems to be some issue on Jenkins

{quote}
FATAL: command execution failed
java.io.IOException: Backing channel 'H0' is disconnected
{quote}

I've run {{test-patch}} locally that I did not see so many issues. Will trigger 
another jenkins job if necessary.

> Ozone: SCM: Move SCM config keys to ScmConfig
> -
>
> Key: HDFS-11071
> URL: https://issues.apache.org/jira/browse/HDFS-11071
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>Assignee: Weiwei Yang
>Priority: Minor
> Fix For: HDFS-7240
>
> Attachments: HDFS-11071-HDFS-7240.001.patch
>
>
> Move SCM keys from OzoneConfigKeys.Ozone to SCMConfigKeys.SCM * 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11446) TestMaintenanceState#testWithNNAndDNRestart fails intermittently

2017-03-06 Thread Yiqun Lin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11446?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yiqun Lin updated HDFS-11446:
-
Attachment: HDFS-11446-branch-2.patch
HDFS-11446.003.patch

Reupload the patch to trigger the Jenkins.

> TestMaintenanceState#testWithNNAndDNRestart fails intermittently
> 
>
> Key: HDFS-11446
> URL: https://issues.apache.org/jira/browse/HDFS-11446
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 3.0.0-alpha2
>Reporter: Yiqun Lin
>Assignee: Yiqun Lin
> Attachments: HDFS-11446.001.patch, HDFS-11446.002.patch, 
> HDFS-11446.003.patch, HDFS-11446-branch-2.patch
>
>
> The test {{TestMaintenanceState#testWithNNAndDNRestart}} fails in trunk. The 
> stack info( 
> https://builds.apache.org/job/PreCommit-HDFS-Build/18423/testReport/ ):
> {code}
> java.lang.AssertionError: expected null, but was: for block BP-1367163238-172.17.0.2-1487836532907:blk_1073741825_1001: 
> expected 3, got 2 
> ,DatanodeInfoWithStorage[127.0.0.1:42649,DS-c499e6ef-ce14-428b-baef-8cf2a122b248,DISK],DatanodeInfoWithStorage[127.0.0.1:40774,DS-cc484c09-6e32-4804-a337-2871f37b62e1,DISK],pending
>  block # 1 ,under replicated # 0 ,>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotNull(Assert.java:664)
>   at org.junit.Assert.assertNull(Assert.java:646)
>   at org.junit.Assert.assertNull(Assert.java:656)
>   at 
> org.apache.hadoop.hdfs.TestMaintenanceState.testWithNNAndDNRestart(TestMaintenanceState.java:731)
> {code}
> The failure seems due to pending block has not been replicated. We can bump 
> the retry times since sometimes the cluster would be busy. Also we can use 
> {{GenericTestUtils#waitFor}} to simplified the current compared logic.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11446) TestMaintenanceState#testWithNNAndDNRestart fails intermittently

2017-03-06 Thread Yiqun Lin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11446?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yiqun Lin updated HDFS-11446:
-
Attachment: (was: HDFS-11446-branch-2.patch)

> TestMaintenanceState#testWithNNAndDNRestart fails intermittently
> 
>
> Key: HDFS-11446
> URL: https://issues.apache.org/jira/browse/HDFS-11446
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 3.0.0-alpha2
>Reporter: Yiqun Lin
>Assignee: Yiqun Lin
> Attachments: HDFS-11446.001.patch, HDFS-11446.002.patch
>
>
> The test {{TestMaintenanceState#testWithNNAndDNRestart}} fails in trunk. The 
> stack info( 
> https://builds.apache.org/job/PreCommit-HDFS-Build/18423/testReport/ ):
> {code}
> java.lang.AssertionError: expected null, but was: for block BP-1367163238-172.17.0.2-1487836532907:blk_1073741825_1001: 
> expected 3, got 2 
> ,DatanodeInfoWithStorage[127.0.0.1:42649,DS-c499e6ef-ce14-428b-baef-8cf2a122b248,DISK],DatanodeInfoWithStorage[127.0.0.1:40774,DS-cc484c09-6e32-4804-a337-2871f37b62e1,DISK],pending
>  block # 1 ,under replicated # 0 ,>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotNull(Assert.java:664)
>   at org.junit.Assert.assertNull(Assert.java:646)
>   at org.junit.Assert.assertNull(Assert.java:656)
>   at 
> org.apache.hadoop.hdfs.TestMaintenanceState.testWithNNAndDNRestart(TestMaintenanceState.java:731)
> {code}
> The failure seems due to pending block has not been replicated. We can bump 
> the retry times since sometimes the cluster would be busy. Also we can use 
> {{GenericTestUtils#waitFor}} to simplified the current compared logic.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11446) TestMaintenanceState#testWithNNAndDNRestart fails intermittently

2017-03-06 Thread Yiqun Lin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11446?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yiqun Lin updated HDFS-11446:
-
Attachment: (was: HDFS-11446.003.patch)

> TestMaintenanceState#testWithNNAndDNRestart fails intermittently
> 
>
> Key: HDFS-11446
> URL: https://issues.apache.org/jira/browse/HDFS-11446
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 3.0.0-alpha2
>Reporter: Yiqun Lin
>Assignee: Yiqun Lin
> Attachments: HDFS-11446.001.patch, HDFS-11446.002.patch
>
>
> The test {{TestMaintenanceState#testWithNNAndDNRestart}} fails in trunk. The 
> stack info( 
> https://builds.apache.org/job/PreCommit-HDFS-Build/18423/testReport/ ):
> {code}
> java.lang.AssertionError: expected null, but was: for block BP-1367163238-172.17.0.2-1487836532907:blk_1073741825_1001: 
> expected 3, got 2 
> ,DatanodeInfoWithStorage[127.0.0.1:42649,DS-c499e6ef-ce14-428b-baef-8cf2a122b248,DISK],DatanodeInfoWithStorage[127.0.0.1:40774,DS-cc484c09-6e32-4804-a337-2871f37b62e1,DISK],pending
>  block # 1 ,under replicated # 0 ,>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotNull(Assert.java:664)
>   at org.junit.Assert.assertNull(Assert.java:646)
>   at org.junit.Assert.assertNull(Assert.java:656)
>   at 
> org.apache.hadoop.hdfs.TestMaintenanceState.testWithNNAndDNRestart(TestMaintenanceState.java:731)
> {code}
> The failure seems due to pending block has not been replicated. We can bump 
> the retry times since sometimes the cluster would be busy. Also we can use 
> {{GenericTestUtils#waitFor}} to simplified the current compared logic.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Yiqun Lin (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898628#comment-15898628
 ] 

Yiqun Lin commented on HDFS-11499:
--

Great work, everyone!
Only one nit for the v03 patch:
{code}
+
+Path openFile = new Path("/testClosingFileInMaintenance.dat");
+// Lets write 2 blocks of data to the openFile
+writeFile(getCluster().getFileSystem(), openFile, (short) 3);
+
{code}
The comment seems not accurate, here we write three block replica to openFile , 
right? [~manojg]. Or this should be replaced by "two more blocks".

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.02.patch, HDFS-11499.03.patch, 
> HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-10983) OIV tool should make an EC file explicit

2017-03-06 Thread Manoj Govindassamy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10983?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898592#comment-15898592
 ] 

Manoj Govindassamy commented on HDFS-10983:
---

Thanks for the review [~andrew.wang].

If we don't set replication explicitly for Striped blocks, the field is 0 by 
default and anyone reading this replication field from the FSImage would be 
surprised. So, am explicitly setting it to DEFAULT_REPL_FOR_STRIPED_BLOCKS. 
Example, in {{PBImageXMLWriter#dumpINodeFile}}, the previous code was just read 
the replication field and dump in xml as is, which could be wrong for Striped 
blocks. So, I later changed to read the file type and dump replication param 
accordingly. Though this was fixed, i was worried about similar readers not 
caring about striped files and just reading replication param. 

{{FSImageFormatPBInode}}
{code}
public static INodeSection.INodeFile.Builder buildINodeFile(
INodeFileAttributes file, final SaverContext state) {
  INodeSection.INodeFile.Builder b = INodeSection.INodeFile.newBuilder()
  .setAccessTime ...

  if (file.isStriped()) {
b.setReplication(INodeFile.DEFAULT_REPL_FOR_STRIPED_BLOCKS);
b.setErasureCodingPolicyID(file.getErasureCodingPolicyID());
  } else {
b.setReplication(file.getFileReplication());
  }
{code}

> OIV tool should make an EC file explicit
> 
>
> Key: HDFS-10983
> URL: https://issues.apache.org/jira/browse/HDFS-10983
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 3.0.0-alpha1
>Reporter: Wei-Chiu Chuang
>Assignee: Manoj Govindassamy
>  Labels: hdfs-ec-3.0-nice-to-have
> Attachments: HDFS-10983.01.patch, HDFS-10983.02.patch
>
>
> The OIV tool's webhdfs interface does not print if a file is striped or not.
> Also, it prints the file's EC policy ID as replication factor, which is 
> inconsistent to the output of a typical webhdfs call to the cluster, which 
> always shows replication factor of 0 for EC files.
> Not just webhdfs, but delimiter output does not print if a file is stripped 
> or not either.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11314) Enforce set of enabled EC policies on the NameNode

2017-03-06 Thread Kai Zheng (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898588#comment-15898588
 ] 

Kai Zheng commented on HDFS-11314:
--

Thanks for your clarifying and consideration, Andrew. Yes renaming the getters 
should work. 

> Enforce set of enabled EC policies on the NameNode
> --
>
> Key: HDFS-11314
> URL: https://issues.apache.org/jira/browse/HDFS-11314
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: erasure-coding
>Affects Versions: 3.0.0-alpha1
>Reporter: Andrew Wang
>Assignee: Andrew Wang
>Priority: Blocker
>  Labels: hdfs-ec-3.0-must-do
> Attachments: HDFS-11314.001.patch, HDFS-11314.002.patch, 
> HDFS-11314.003.patch, HDFS-11314.004.patch, HDFS-11314.005.patch, 
> HDFS-11314.006.patch
>
>
> Filing based on discussion in HDFS-8095. A user might specify a policy that 
> is not appropriate for the cluster, e.g. a RS (10,4) policy when the cluster 
> only has 10 nodes. The NN should only allow the client to choose from a 
> pre-approved list determined by the cluster administrator.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11502) dn.js set datanode UI to window.location.hostname, it should use jmx "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get tag.hostname

2017-03-06 Thread Jeffrey E Rodriguez (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11502?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898587#comment-15898587
 ] 

Jeffrey E  Rodriguez commented on HDFS-11502:
-

Thanks I would do so

> dn.js set datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname
> --
>
> Key: HDFS-11502
> URL: https://issues.apache.org/jira/browse/HDFS-11502
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.7.2, 2.7.3
> Environment: all
>Reporter: Jeffrey E  Rodriguez
>
> Datanode UI calls dn.js which loads properties for datanode..
> dn.js sets datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname.
> The issue is that if we proxy datanode UI we would get the proxy as the 
> hostname because UI code uses window.location.hostname instead of the actual 
> datanode hostname.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-9705) Refine the behaviour of getFileChecksum when length = 0

2017-03-06 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9705?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898584#comment-15898584
 ] 

Hadoop QA commented on HDFS-9705:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
24s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
37s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 15m 
 3s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
46s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
45s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
37s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
27s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
26s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
1s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m  
7s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
16s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
19s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
19s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
38s{color} | {color:green} hadoop-hdfs-project: The patch generated 0 new + 89 
unchanged - 1 fixed = 89 total (was 90) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
21s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
20s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
54s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  0m 
55s{color} | {color:green} hadoop-hdfs-client in the patch passed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 70m 55s{color} 
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
24s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}108m 16s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.hdfs.TestFSOutputSummer |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | HDFS-9705 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12856226/HDFS-9705-v6.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux 0d53ef231060 3.13.0-106-generic #153-Ubuntu SMP Tue Dec 6 
15:44:32 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 5e74196 |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| unit | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18601/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18601/testReport/ |
| modules | C: hadoop-hdfs-project/hadoop-hdfs-client 
hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18601/console

[jira] [Commented] (HDFS-11505) Do not enable any erasure coding policies by default

2017-03-06 Thread Kai Zheng (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898583#comment-15898583
 ] 

Kai Zheng commented on HDFS-11505:
--

Thanks Andrew for your thought and it makes sense. The client-side config 
similar to dfs.replication to specify a default policy is a great idea and I 
guess we can do it later given some real user requirement that incurs much need 
of setting policy per file. 

> Do not enable any erasure coding policies by default
> 
>
> Key: HDFS-11505
> URL: https://issues.apache.org/jira/browse/HDFS-11505
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: erasure-coding
>Affects Versions: 3.0.0-alpha3
>Reporter: Andrew Wang
>Assignee: Manoj Govindassamy
>  Labels: hdfs-ec-3.0-must-do
>
> As discussed on HDFS-11314, administrators need to choose the correct set of 
> EC policies based on cluster size and desired fault-tolerance properties.
> This means we should not enable any EC policies by default, since any default 
> value could be incorrect.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Manoj Govindassamy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Manoj Govindassamy updated HDFS-11499:
--
Attachment: HDFS-11499.03.patch

[~lukmajercak],
Attached v03 patch to have the same test timeout as the other tests and also 
fixed checkstyle issues. Please take a look.

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.02.patch, HDFS-11499.03.patch, 
> HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11502) dn.js set datanode UI to window.location.hostname, it should use jmx "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get tag.hostname

2017-03-06 Thread Yuanbo Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11502?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898579#comment-15898579
 ] 

Yuanbo Liu commented on HDFS-11502:
---

[~jeffreyr97] If you want to work on it, please send a email to hadoop-dev and 
apply for "assign to myself" permission.
Currently we don't have the privilege to assign JIRAs.

> dn.js set datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname
> --
>
> Key: HDFS-11502
> URL: https://issues.apache.org/jira/browse/HDFS-11502
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 2.7.2, 2.7.3
> Environment: all
>Reporter: Jeffrey E  Rodriguez
>
> Datanode UI calls dn.js which loads properties for datanode..
> dn.js sets datanode UI to window.location.hostname, it should use jmx 
> "qry=Hadoop:service=DataNode,name=RpcDetailedActivityForPort8010" to get 
> tag.hostname.
> The issue is that if we proxy datanode UI we would get the proxy as the 
> hostname because UI code uses window.location.hostname instead of the actual 
> datanode hostname.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11314) Enforce set of enabled EC policies on the NameNode

2017-03-06 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11314?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-11314:
---
Attachment: HDFS-11314.006.patch

Thanks for reviewing Kai! Regarding the naming, I was trying to differentiate 
between:

# The set of all possible policies
# The set of enabled policies

I reused "system" to refer to the first set, while using "enabled" to refer to 
the second set. I forgot that "system" specifically refers to the hardcoded 
policies, so let's make the names more general as you say.

I shortened your proposal a little bit if that's okay, named the getters 
{{getPolicyByName}} and {{getPolicyByID}}.

> Enforce set of enabled EC policies on the NameNode
> --
>
> Key: HDFS-11314
> URL: https://issues.apache.org/jira/browse/HDFS-11314
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: erasure-coding
>Affects Versions: 3.0.0-alpha1
>Reporter: Andrew Wang
>Assignee: Andrew Wang
>Priority: Blocker
>  Labels: hdfs-ec-3.0-must-do
> Attachments: HDFS-11314.001.patch, HDFS-11314.002.patch, 
> HDFS-11314.003.patch, HDFS-11314.004.patch, HDFS-11314.005.patch, 
> HDFS-11314.006.patch
>
>
> Filing based on discussion in HDFS-8095. A user might specify a policy that 
> is not appropriate for the cluster, e.g. a RS (10,4) policy when the cluster 
> only has 10 nodes. The NN should only allow the client to choose from a 
> pre-approved list determined by the cluster administrator.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Lukas Majercak (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898571#comment-15898571
 ] 

Lukas Majercak commented on HDFS-11499:
---

[~manojg], yes 
TestDecommission#testDecommissionWithOpenFileAndDatanodeFailing(). It will wait 
for all three DNs to be decommissioned right, and the log you showed is just 1 
of them. I would suggest increasing it to 360sec yes. It finishes in 30seconds~ 
on my machine, similar to 
TestDecommission#testDeadNodeCountAfterNamenodeRestart, which has 360sec 
timeout.

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.02.patch, HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11314) Enforce set of enabled EC policies on the NameNode

2017-03-06 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11314?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-11314:
---
Attachment: HDFS-11314.005.patch

New patch to address checkstyle errors.

> Enforce set of enabled EC policies on the NameNode
> --
>
> Key: HDFS-11314
> URL: https://issues.apache.org/jira/browse/HDFS-11314
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: erasure-coding
>Affects Versions: 3.0.0-alpha1
>Reporter: Andrew Wang
>Assignee: Andrew Wang
>Priority: Blocker
>  Labels: hdfs-ec-3.0-must-do
> Attachments: HDFS-11314.001.patch, HDFS-11314.002.patch, 
> HDFS-11314.003.patch, HDFS-11314.004.patch, HDFS-11314.005.patch
>
>
> Filing based on discussion in HDFS-8095. A user might specify a policy that 
> is not appropriate for the cluster, e.g. a RS (10,4) policy when the cluster 
> only has 10 nodes. The NN should only allow the client to choose from a 
> pre-approved list determined by the cluster administrator.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11314) Enforce set of enabled EC policies on the NameNode

2017-03-06 Thread Kai Zheng (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898546#comment-15898546
 ] 

Kai Zheng commented on HDFS-11314:
--

I agree it's much safer to have admin configure and ensure the valid policies. 
The list should only affect the choices that users would choose for other 
files/folders, and incur no impact for existing erasure coded files/folders. 
Thanks [~andrew.wang] for the great work!

One question and minor, noticed pretty many changes like follows and wondered 
why. If we would allow user defined policies in future, due to whatever 
considerations like new codec, new option even for existing codec and etc., I 
guess the more general {{getPolicyByPolicyID}} name would be better.
{code}
-return ErasureCodingPolicyManager.getPolicyByPolicyID(
+return ErasureCodingPolicyManager.getSystemPolicyByID(
{code}

> Enforce set of enabled EC policies on the NameNode
> --
>
> Key: HDFS-11314
> URL: https://issues.apache.org/jira/browse/HDFS-11314
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: erasure-coding
>Affects Versions: 3.0.0-alpha1
>Reporter: Andrew Wang
>Assignee: Andrew Wang
>Priority: Blocker
>  Labels: hdfs-ec-3.0-must-do
> Attachments: HDFS-11314.001.patch, HDFS-11314.002.patch, 
> HDFS-11314.003.patch, HDFS-11314.004.patch
>
>
> Filing based on discussion in HDFS-8095. A user might specify a policy that 
> is not appropriate for the cluster, e.g. a RS (10,4) policy when the cluster 
> only has 10 nodes. The NN should only allow the client to choose from a 
> pre-approved list determined by the cluster administrator.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11505) Do not enable any erasure coding policies by default

2017-03-06 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898544#comment-15898544
 ] 

Andrew Wang commented on HDFS-11505:


Hi Kai,

Your understanding is correct. However, I think the vast majority of EC writes 
will happen by inheriting the policy from a parent directory at create time, so 
usage will be transparent. It's likely that only command-line users will be 
calling the setPolicy API, so they'll (hopefully) get nice error messages about 
what's an invalid or disabled policy and how to enable policies. As a reference 
point, I don't think we've seen a similar need for the storage policy APIs. 
Apps like HBase or Hive have often depended on admins setting up 
directory-level policies (e.g. storage policies, encryption).

However, if you think having a default policy is important, we could add a 
client-side config similar to {{dfs.replication}} to specify a default policy. 
Alternatively, this could also be set in NN configuration and queried via 
getServerDefaults. As per above though, I'd hope that dir-level policies are 
the best fit for users who want a simple way of using EC.

> Do not enable any erasure coding policies by default
> 
>
> Key: HDFS-11505
> URL: https://issues.apache.org/jira/browse/HDFS-11505
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: erasure-coding
>Affects Versions: 3.0.0-alpha3
>Reporter: Andrew Wang
>Assignee: Manoj Govindassamy
>  Labels: hdfs-ec-3.0-must-do
>
> As discussed on HDFS-11314, administrators need to choose the correct set of 
> EC policies based on cluster size and desired fault-tolerance properties.
> This means we should not enable any EC policies by default, since any default 
> value could be incorrect.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11314) Enforce set of enabled EC policies on the NameNode

2017-03-06 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898541#comment-15898541
 ] 

Hadoop QA commented on HDFS-11314:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
12s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 17 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 11m 
55s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
51s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
38s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
48s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
11s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
38s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
37s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
44s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
39s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
39s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
0m 36s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch 
generated 8 new + 771 unchanged - 6 fixed = 779 total (was 777) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
47s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
 9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
3s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
44s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
35s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 70m  1s{color} 
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 93m 34s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | 
hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting |
|   | hadoop.hdfs.web.TestWebHdfsTimeouts |
| Timed out junit tests | org.apache.hadoop.hdfs.TestLeaseRecovery2 |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | HDFS-11314 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12856372/HDFS-11314.004.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  xml  |
| uname | Linux 0568dc0418e4 4.4.0-43-generic #63-Ubuntu SMP Wed Oct 12 
13:48:03 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 5e74196 |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| checkstyle | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18599/artifact/patchprocess/diff-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt
 |
| unit | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18599/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18599/testReport/ |
| modules | C: hadoop-hdfs-project/hadoop-hdfs U: 
hadoop-hdfs-project/hadoop-hdfs |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18599/co

[jira] [Commented] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Manoj Govindassamy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898537#comment-15898537
 ] 

Manoj Govindassamy commented on HDFS-11499:
---

[~lukmajercak],
Are you referring to the timeout in 
TestDecommission#testDecommissionWithOpenFileAndDatanodeFailing() which was 
part of the patch v01 ? In the patch v02 I added Maintenance State related 
test. Not sure, if extending the timeout for the failed test is going to solve 
the problem. Because, the nodes didn't move to DECOMMISSIONED state as the test 
is expecting .

{noformat}

2017-03-06 23:33:49,462 [Thread-782] INFO  hdfs.AdminStatesBaseTest 
(AdminStatesBaseTest.java:waitNodeState(342)) - Waiting for node 
127.0.0.1:33069 to change state to Decommissioned current state: Decommission 
In Progress
2017-03-06 23:33:49,462 [Thread-782] INFO  hdfs.AdminStatesBaseTest 
(AdminStatesBaseTest.java:waitNodeState(342)) - Waiting for node 
127.0.0.1:33069 to change state to Decommissioned current state: Decommission 
In Progress

[test timeout]

2017-03-06 23:33:49,486 [main] INFO  hdfs.MiniDFSCluster 
(MiniDFSCluster.java:shutdown(1951)) - Shutting down the Mini HDFS Cluster
{noformat}

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.02.patch, HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11482) Add storage type demand to into DFSNetworkTopology#chooseRandom

2017-03-06 Thread Chen Liang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chen Liang updated HDFS-11482:
--
Attachment: HDFS-11482.003.patch

v001 and v002 patch were missing the change to one more file. Fixed in v003 
patch.

Also, there is potential race condition that can cause the code to loop forever 
when choosing a random node. This issue also exists in current NetworkTopology, 
tracked in HDFS-11507

> Add storage type demand to into DFSNetworkTopology#chooseRandom
> ---
>
> Key: HDFS-11482
> URL: https://issues.apache.org/jira/browse/HDFS-11482
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Chen Liang
>Assignee: Chen Liang
> Attachments: HDFS-11482.001.patch, HDFS-11482.002.patch, 
> HDFS-11482.003.patch
>
>
> HDFS-11450 adds storage type info into network topology, with on this info we 
> may change chooseRandom to take storage type requirement as parameter, only 
> checking subtrees required storage type available. This way we avoid blindly 
> picking up nodes that are not applicable.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11505) Do not enable any erasure coding policies by default

2017-03-06 Thread Kai Zheng (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898529#comment-15898529
 ] 

Kai Zheng commented on HDFS-11505:
--

So assume some policies are enabled by the configuration, will users have to 
specify some one when create a file/folder or assign/set policy to it later? 
I'm asking this because, I'm wondering, for most applications/users they might 
not want to bother to query/check the enabled policies first to pick up some to 
specify. If users don't have to or the policy parameter is optional, I guess we 
could use the first one in the enabled list. Thoughts?

> Do not enable any erasure coding policies by default
> 
>
> Key: HDFS-11505
> URL: https://issues.apache.org/jira/browse/HDFS-11505
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: erasure-coding
>Affects Versions: 3.0.0-alpha3
>Reporter: Andrew Wang
>Assignee: Manoj Govindassamy
>  Labels: hdfs-ec-3.0-must-do
>
> As discussed on HDFS-11314, administrators need to choose the correct set of 
> EC policies based on cluster size and desired fault-tolerance properties.
> This means we should not enable any EC policies by default, since any default 
> value could be incorrect.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Lukas Majercak (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898524#comment-15898524
 ] 

Lukas Majercak commented on HDFS-11499:
---

Looks like the test timed out, [~manojg], do you mind increasing the timeout to 
360sec (based on the other tests in TestDecommission) and submitting a patch? 
Thanks!

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.02.patch, HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Created] (HDFS-11507) NetworkTopology#chooseRandom may run into a dead loop due to race condition

2017-03-06 Thread Chen Liang (JIRA)
Chen Liang created HDFS-11507:
-

 Summary: NetworkTopology#chooseRandom may run into a dead loop due 
to race condition
 Key: HDFS-11507
 URL: https://issues.apache.org/jira/browse/HDFS-11507
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Reporter: Chen Liang
Assignee: Chen Liang


{{NetworkTopology#chooseRandom()}} works as:
1. counts the number of available nodes as {{availableNodes}},
2. checks how many nodes are excluded, deduct from {{availableNodes}}
3. if {{availableNodes}} still > 0, then there are nodes available.
4. keep looping to find that node

But now imagine, in the meantime, the actually available nodes got removed in 
step 3 or step 4, and all remaining nodes are excluded nodes. Then, although 
there are no more nodes actually available, the code would still run as 
{{availableNodes}} > 0, and then it would keep getting excluded node and loop 
forever, as 
{{if (excludedNodes == null || !excludedNodes.contains(ret))}} 
will always be false.

We may fix this by expanding the while loop to also include the 
{{availableNodes}} calculation. Such that we re-calculate {{availableNodes}} 
every time it fails to find an available node.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11071) Ozone: SCM: Move SCM config keys to ScmConfig

2017-03-06 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11071?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898519#comment-15898519
 ] 

Hadoop QA commented on HDFS-11071:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
33s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 4 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
10s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 18m 
51s{color} | {color:green} HDFS-7240 passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
49s{color} | {color:green} HDFS-7240 passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
48s{color} | {color:green} HDFS-7240 passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
58s{color} | {color:green} HDFS-7240 passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
33s{color} | {color:green} HDFS-7240 passed {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  2m  
5s{color} | {color:red} hadoop-hdfs-project/hadoop-hdfs-client in HDFS-7240 has 
98 extant Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
40s{color} | {color:green} HDFS-7240 passed {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m  
8s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
56s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m  
0s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  2m  
0s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
0m 48s{color} | {color:orange} hadoop-hdfs-project: The patch generated 7 new + 
5 unchanged - 0 fixed = 12 total (was 5) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
53s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
21s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  1m 
11s{color} | {color:green} hadoop-hdfs-client in the patch passed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red}142m 47s{color} 
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
49s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}190m  2s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.hdfs.TestReadStripedFileWithDecoding |
|   | hadoop.hdfs.TestFileStatus |
|   | hadoop.hdfs.server.balancer.TestBalancerWithHANameNodes |
|   | hadoop.hdfs.server.datanode.fsdataset.impl.TestDatanodeRestart |
|   | hadoop.hdfs.TestFileCreationDelete |
|   | hadoop.hdfs.TestEncryptionZonesWithHA |
|   | hadoop.hdfs.TestDFSShell |
|   | hadoop.hdfs.server.namenode.TestNameNodeXAttr |
|   | hadoop.hdfs.shortcircuit.TestShortCircuitCache |
|   | hadoop.hdfs.TestDecommission |
|   | hadoop.hdfs.server.namenode.TestFSEditLogLoader |
|   | hadoop.hdfs.TestDFSStripedOutputStreamWithFailure180 |
|   | hadoop.ozone.web.TestOzoneWebAccess |
|   | hadoop.hdfs.server.balancer.TestBalancerWithMultipleNameNodes |
|   | hadoop.hdfs.server.blockmanagement.TestNameNodePrunesMissingStorages |
|   | hadoop.ozone.web.client.TestKeys |
|   | hadoop.hdfs.server.datanode.TestCachingStrategy |
|   | hadoop.hdfs.server.datanode.TestDirectoryScanner |
|   | hadoop.cli.TestXAttrCLI |
|   | hadoop.ozone

[jira] [Commented] (HDFS-10838) Last full block report received time for each DN should be easily discoverable

2017-03-06 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898518#comment-15898518
 ] 

Hudson commented on HDFS-10838:
---

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #11358 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/11358/])
HDFS-10838. Last full block report received time for each DN should be (arp: 
rev b5adc5c3011f111f86d232cb33ec522547f68a95)
* (edit) 
hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/protocol/DatanodeInfo.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/protocolPB/PBHelperClient.java
* (edit) hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/hdfs/dfshealth.js
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeMXBean.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java
* (edit) hadoop-hdfs-project/hadoop-hdfs-client/src/main/proto/hdfs.proto
* (edit) hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/hdfs/dfshealth.html
* (edit) 
hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/web/JsonUtilClient.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/JsonUtil.java


> Last full block report received time for each DN should be easily discoverable
> --
>
> Key: HDFS-10838
> URL: https://issues.apache.org/jira/browse/HDFS-10838
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: ui
>Reporter: Arpit Agarwal
>Assignee: Surendra Singh Lilhore
> Fix For: 2.9.0, 3.0.0-alpha3
>
> Attachments: DFSAdmin-Report.png, HDFS-10838-001.patch, 
> HDFS-10838.002.patch, HDFS-10838.003.patch, HDFS-10838.004.patch, 
> HDFS-10838.005.patch, HDFS-10838.006.patch, NN_UI.png, 
> NN_UI_relative_time_minutes.png, NN_UI_relative_time.png
>
>
> It should be easy for administrators to discover the time of last full block 
> report from each DataNode.
> We can show it in the NameNode web UI or in the output of {{hdfs dfsadmin 
> -report}}, or both.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-6984) In Hadoop 3, make FileStatus serialize itself via protobuf

2017-03-06 Thread Chris Douglas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6984?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Douglas updated HDFS-6984:

Attachment: HDFS-6984.004.patch

> In Hadoop 3, make FileStatus serialize itself via protobuf
> --
>
> Key: HDFS-6984
> URL: https://issues.apache.org/jira/browse/HDFS-6984
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Affects Versions: 3.0.0-alpha1
>Reporter: Colin P. McCabe
>Assignee: Colin P. McCabe
>  Labels: BB2015-05-TBR
> Attachments: HDFS-6984.001.patch, HDFS-6984.002.patch, 
> HDFS-6984.003.patch, HDFS-6984.004.patch, HDFS-6984.nowritable.patch
>
>
> FileStatus was a Writable in Hadoop 2 and earlier.  Originally, we used this 
> to serialize it and send it over the wire.  But in Hadoop 2 and later, we 
> have the protobuf {{HdfsFileStatusProto}} which serves to serialize this 
> information.  The protobuf form is preferable, since it allows us to add new 
> fields in a backwards-compatible way.  Another issue is that already a lot of 
> subclasses of FileStatus don't override the Writable methods of the 
> superclass, breaking the interface contract that read(status.write) should be 
> equal to the original status.
> In Hadoop 3, we should just make FileStatus serialize itself via protobuf so 
> that we don't have to deal with these issues.  It's probably too late to do 
> this in Hadoop 2, since user code may be relying on the existing FileStatus 
> serialization there.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11482) Add storage type demand to into DFSNetworkTopology#chooseRandom

2017-03-06 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11482?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898493#comment-15898493
 ] 

Hadoop QA commented on HDFS-11482:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
18s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 3 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
50s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 14m 
24s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 12m 
33s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
14s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m 
34s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
49s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m  
4s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
45s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
13s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:red}-1{color} | {color:red} mvninstall {color} | {color:red}  0m 
21s{color} | {color:red} hadoop-common in the patch failed. {color} |
| {color:red}-1{color} | {color:red} mvninstall {color} | {color:red}  0m 
42s{color} | {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:red}-1{color} | {color:red} compile {color} | {color:red}  0m 
34s{color} | {color:red} root in the patch failed. {color} |
| {color:red}-1{color} | {color:red} javac {color} | {color:red}  0m 34s{color} 
| {color:red} root in the patch failed. {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
1m 47s{color} | {color:orange} root: The patch generated 2 new + 51 unchanged - 
0 fixed = 53 total (was 51) {color} |
| {color:red}-1{color} | {color:red} mvnsite {color} | {color:red}  0m 
23s{color} | {color:red} hadoop-common in the patch failed. {color} |
| {color:red}-1{color} | {color:red} mvnsite {color} | {color:red}  0m 
42s{color} | {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
20s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  0m 
15s{color} | {color:red} hadoop-common in the patch failed. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  0m 
23s{color} | {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
17s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red}  0m 22s{color} 
| {color:red} hadoop-common in the patch failed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red}  0m 41s{color} 
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
18s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 49m 44s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | HDFS-11482 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12856373/HDFS-11482.002.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux e08333756f58 3.13.0-106-generic #153-Ubuntu SMP Tue Dec 6 
15:44:32 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 5e74196 |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| mvninstall | 
https://builds.apache.org/job/PreCommit-HDFS-Build/1860

[jira] [Commented] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898491#comment-15898491
 ] 

Hadoop QA commented on HDFS-11499:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
33s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 19m 
58s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
18s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
58s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
28s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
23s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m 
25s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
0s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
11s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
0m 43s{color} | {color:orange} hadoop-hdfs-project/hadoop-hdfs: The patch 
generated 4 new + 146 unchanged - 0 fixed = 150 total (was 146) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
14s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m 
16s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
50s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 96m 49s{color} 
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  1m 
58s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}136m 19s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | 
hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting |
|   | hadoop.hdfs.TestFileAppend3 |
|   | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure |
|   | hadoop.hdfs.TestDecommission |
|   | hadoop.hdfs.server.datanode.TestDirectoryScanner |
|   | hadoop.hdfs.server.datanode.TestDataNodeUUID |
| Timed out junit tests | org.apache.hadoop.hdfs.server.namenode.TestEditLog |
|   | org.apache.hadoop.hdfs.server.namenode.TestQuotaByStorageType |
|   | org.apache.hadoop.hdfs.server.blockmanagement.TestBlockStatsMXBean |
|   | org.apache.hadoop.hdfs.TestDFSStripedOutputStreamWithFailure030 |
|   | org.apache.hadoop.hdfs.server.namenode.TestEditLogAutoroll |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | HDFS-11499 |
| GITHUB PR | https://github.com/apache/hadoop/pull/199 |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux 278985d535e8 3.13.0-92-generic #139-Ubuntu SMP Tue Jun 28 
20:42:26 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / d9dc444 |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| checkstyle | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18598/artifact/patchprocess/diff-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt
 |
| unit | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18598/artifact/patchprocess/patch-

[jira] [Commented] (HDFS-11477) Simplify file IO profiling configuration

2017-03-06 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11477?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898478#comment-15898478
 ] 

Hadoop QA commented on HDFS-11477:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
19s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 2 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
13s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 12m 
38s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 10m 
28s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
 5s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m 
12s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
48s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
21s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
45s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
14s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
22s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 10m 
38s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 10m 
38s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
 5s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m  
5s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
48s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
44s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
48s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red}  9m 11s{color} 
| {color:red} hadoop-common in the patch failed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 72m 40s{color} 
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
43s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}140m 32s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.security.TestKDiag |
|   | hadoop.hdfs.tools.TestDFSZKFailoverController |
|   | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure |
|   | hadoop.hdfs.server.datanode.TestDataNodeErasureCodingMetrics |
|   | hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistReplicaRecovery 
|
|   | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | HDFS-11477 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12855960/HDFS-11477.003.patch |
| Optional Tests |  asflicense  mvnsite  compile  javac  javadoc  mvninstall  
unit  findbugs  checkstyle  |
| uname | Linux 4b91f7864c20 3.13.0-106-generic #153-Ubuntu SMP Tue Dec 6 
15:44:32 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / d9dc444 |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| unit | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18597/artifact/patchprocess/patch-unit-hadoop-common-project_hadoop-common.txt
 |
| unit | 
https://builds.apache

[jira] [Commented] (HDFS-11482) Add storage type demand to into DFSNetworkTopology#chooseRandom

2017-03-06 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11482?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898476#comment-15898476
 ] 

Hadoop QA commented on HDFS-11482:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
22s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 3 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
55s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 13m 
19s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 10m 
34s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
 0s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m 
18s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
50s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
24s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
45s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
14s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:red}-1{color} | {color:red} mvninstall {color} | {color:red}  0m 
21s{color} | {color:red} hadoop-common in the patch failed. {color} |
| {color:red}-1{color} | {color:red} mvninstall {color} | {color:red}  0m 
42s{color} | {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:red}-1{color} | {color:red} compile {color} | {color:red}  0m 
40s{color} | {color:red} root in the patch failed. {color} |
| {color:red}-1{color} | {color:red} javac {color} | {color:red}  0m 40s{color} 
| {color:red} root in the patch failed. {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
1m 47s{color} | {color:orange} root: The patch generated 2 new + 51 unchanged - 
0 fixed = 53 total (was 51) {color} |
| {color:red}-1{color} | {color:red} mvnsite {color} | {color:red}  0m 
25s{color} | {color:red} hadoop-common in the patch failed. {color} |
| {color:red}-1{color} | {color:red} mvnsite {color} | {color:red}  0m 
48s{color} | {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  0m 
16s{color} | {color:red} hadoop-common in the patch failed. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  0m 
26s{color} | {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
29s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red}  0m 22s{color} 
| {color:red} hadoop-common in the patch failed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red}  0m 52s{color} 
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
20s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 46m 39s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | HDFS-11482 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12856373/HDFS-11482.002.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux 73ebee8cdb84 3.13.0-106-generic #153-Ubuntu SMP Tue Dec 6 
15:44:32 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 5e74196 |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| mvninstall | 
https://builds.apache.org/job/PreCommit-HDFS-Build/1860

[jira] [Updated] (HDFS-10838) Last full block report received time for each DN should be easily discoverable

2017-03-06 Thread Arpit Agarwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10838?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arpit Agarwal updated HDFS-10838:
-
   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 3.0.0-alpha3
   2.9.0
   Status: Resolved  (was: Patch Available)

Committed this to trunk and branch-2. Thanks for the contribution 
[~surendrasingh]!

> Last full block report received time for each DN should be easily discoverable
> --
>
> Key: HDFS-10838
> URL: https://issues.apache.org/jira/browse/HDFS-10838
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: ui
>Reporter: Arpit Agarwal
>Assignee: Surendra Singh Lilhore
> Fix For: 2.9.0, 3.0.0-alpha3
>
> Attachments: DFSAdmin-Report.png, HDFS-10838-001.patch, 
> HDFS-10838.002.patch, HDFS-10838.003.patch, HDFS-10838.004.patch, 
> HDFS-10838.005.patch, HDFS-10838.006.patch, NN_UI.png, 
> NN_UI_relative_time_minutes.png, NN_UI_relative_time.png
>
>
> It should be easy for administrators to discover the time of last full block 
> report from each DataNode.
> We can show it in the NameNode web UI or in the output of {{hdfs dfsadmin 
> -report}}, or both.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HDFS-10838) Last full block report received time for each DN should be easily discoverable

2017-03-06 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898471#comment-15898471
 ] 

Arpit Agarwal edited comment on HDFS-10838 at 3/7/17 12:44 AM:
---

Committed this to trunk and branch-2. Thanks for the contribution 
[~surendrasingh]!

(verified that the TestBalancer failure did not repro locally).


was (Author: arpitagarwal):
Committed this to trunk and branch-2. Thanks for the contribution 
[~surendrasingh]!

> Last full block report received time for each DN should be easily discoverable
> --
>
> Key: HDFS-10838
> URL: https://issues.apache.org/jira/browse/HDFS-10838
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: ui
>Reporter: Arpit Agarwal
>Assignee: Surendra Singh Lilhore
> Fix For: 2.9.0, 3.0.0-alpha3
>
> Attachments: DFSAdmin-Report.png, HDFS-10838-001.patch, 
> HDFS-10838.002.patch, HDFS-10838.003.patch, HDFS-10838.004.patch, 
> HDFS-10838.005.patch, HDFS-10838.006.patch, NN_UI.png, 
> NN_UI_relative_time_minutes.png, NN_UI_relative_time.png
>
>
> It should be easy for administrators to discover the time of last full block 
> report from each DataNode.
> We can show it in the NameNode web UI or in the output of {{hdfs dfsadmin 
> -report}}, or both.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-10838) Last full block report received time for each DN should be easily discoverable

2017-03-06 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898443#comment-15898443
 ] 

Hadoop QA commented on HDFS-10838:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
17s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
33s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 16m 
21s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
51s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
56s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
42s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
29s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
58s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
23s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
10s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
55s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
51s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} cc {color} | {color:green}  1m 
51s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
51s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
1m  1s{color} | {color:orange} hadoop-hdfs-project: The patch generated 2 new + 
430 unchanged - 2 fixed = 432 total (was 432) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
56s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
28s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  1m 
14s{color} | {color:green} hadoop-hdfs-client in the patch passed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 68m 36s{color} 
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
23s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}112m 41s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.hdfs.server.balancer.TestBalancer |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | HDFS-10838 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12856072/HDFS-10838.006.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  cc  |
| uname | Linux 78709fe36bcb 3.13.0-107-generic #154-Ubuntu SMP Tue Dec 20 
09:57:27 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / d9dc444 |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| checkstyle | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18596/artifact/patchprocess/diff-checkstyle-hadoop-hdfs-project.txt
 |
| unit | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18596/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.

[jira] [Commented] (HDFS-10983) OIV tool should make an EC file explicit

2017-03-06 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10983?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898441#comment-15898441
 ] 

Andrew Wang commented on HDFS-10983:


Looks good overall, thanks for revving Manoj! One question though, in 
FSImageFormatPBInode, is it necessary to set the replication factor for an EC 
file? I commented this line out and ran a few EC unit tests successfully. It'd 
be nice to unit test this to make sure we aren't writing extraneous fields to 
the FSImage.

> OIV tool should make an EC file explicit
> 
>
> Key: HDFS-10983
> URL: https://issues.apache.org/jira/browse/HDFS-10983
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 3.0.0-alpha1
>Reporter: Wei-Chiu Chuang
>Assignee: Manoj Govindassamy
>  Labels: hdfs-ec-3.0-nice-to-have
> Attachments: HDFS-10983.01.patch, HDFS-10983.02.patch
>
>
> The OIV tool's webhdfs interface does not print if a file is striped or not.
> Also, it prints the file's EC policy ID as replication factor, which is 
> inconsistent to the output of a typical webhdfs call to the cluster, which 
> always shows replication factor of 0 for EC files.
> Not just webhdfs, but delimiter output does not print if a file is stripped 
> or not either.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11498) Make RestCsrfPreventionHandler and WebHdfsHandler compatible with Netty 4.0

2017-03-06 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11498?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898416#comment-15898416
 ] 

Hudson commented on HDFS-11498:
---

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #11357 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/11357/])
HDFS-11498. Make RestCsrfPreventionHandler and WebHdfsHandler compatible (wang: 
rev 5e74196ede9bfc20eb6d6fe3aa6a0e5c47a40fdd)
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/web/webhdfs/WebHdfsHandler.java
* (edit) 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/web/RestCsrfPreventionFilterHandler.java
* (edit) hadoop-project/pom.xml


> Make RestCsrfPreventionHandler and WebHdfsHandler compatible with Netty 4.0
> ---
>
> Key: HDFS-11498
> URL: https://issues.apache.org/jira/browse/HDFS-11498
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.8.0
>Reporter: Andrew Wang
>Assignee: Andrew Wang
> Fix For: 2.8.0, 3.0.0-alpha3
>
> Attachments: HDFS-11498.001.patch, HDFS-11498.branch-2.001.patch
>
>
> Per discussion in HADOOP-13866, it looks like we can change 2.8.0 back to 
> exposing Netty 4.0, but still be ABI compatible with Netty 4.1 for users like 
> HBase that want to swap out the version.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Issue Comment Deleted] (HDFS-11482) Add storage type demand to into DFSNetworkTopology#chooseRandom

2017-03-06 Thread Chen Liang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chen Liang updated HDFS-11482:
--
Comment: was deleted

(was: I was looking into what would happen when chooseRandom and add/remove are 
being called simultaneously. I'm inclined to believe that as long as the 
{{DFSNetworkTopology}} object is synchronized by the caller when calling 
chooseRandom and add/remove, there should be no problem. But any comments are 
more than welcome.)

> Add storage type demand to into DFSNetworkTopology#chooseRandom
> ---
>
> Key: HDFS-11482
> URL: https://issues.apache.org/jira/browse/HDFS-11482
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Chen Liang
>Assignee: Chen Liang
> Attachments: HDFS-11482.001.patch, HDFS-11482.002.patch
>
>
> HDFS-11450 adds storage type info into network topology, with on this info we 
> may change chooseRandom to take storage type requirement as parameter, only 
> checking subtrees required storage type available. This way we avoid blindly 
> picking up nodes that are not applicable.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11482) Add storage type demand to into DFSNetworkTopology#chooseRandom

2017-03-06 Thread Chen Liang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11482?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898412#comment-15898412
 ] 

Chen Liang commented on HDFS-11482:
---

I was looking into what would happen when chooseRandom and add/remove are being 
called simultaneously. I'm inclined to believe that as long as the 
{{DFSNetworkTopology}} object is synchronized by the caller when calling 
chooseRandom and add/remove, there should be no problem. But any comments are 
more than welcome.

> Add storage type demand to into DFSNetworkTopology#chooseRandom
> ---
>
> Key: HDFS-11482
> URL: https://issues.apache.org/jira/browse/HDFS-11482
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Chen Liang
>Assignee: Chen Liang
> Attachments: HDFS-11482.001.patch, HDFS-11482.002.patch
>
>
> HDFS-11450 adds storage type info into network topology, with on this info we 
> may change chooseRandom to take storage type requirement as parameter, only 
> checking subtrees required storage type available. This way we avoid blindly 
> picking up nodes that are not applicable.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11497) Ozone: Fix Container test regression.

2017-03-06 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11497?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898404#comment-15898404
 ] 

Hadoop QA commented on HDFS-11497:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
20s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red}  0m  
0s{color} | {color:red} The patch doesn't appear to include any new or modified 
tests. Please justify why no new tests are needed for this patch. Also please 
list what manual steps were performed to verify this patch. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 15m 
45s{color} | {color:green} HDFS-7240 passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
2s{color} | {color:green} HDFS-7240 passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
44s{color} | {color:green} HDFS-7240 passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m  
4s{color} | {color:green} HDFS-7240 passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
14s{color} | {color:green} HDFS-7240 passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m  
0s{color} | {color:green} HDFS-7240 passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
49s{color} | {color:green} HDFS-7240 passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
53s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
47s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
47s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
34s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
51s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m  
7s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
46s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 73m 43s{color} 
| {color:red} hadoop-hdfs in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
26s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}103m 34s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.hdfs.server.namenode.metrics.TestNameNodeMetrics |
|   | hadoop.hdfs.TestBlocksScheduledCounter |
|   | hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistReplicaRecovery 
|
|   | hadoop.hdfs.server.datanode.TestDataNodeMXBean |
|   | hadoop.hdfs.server.diskbalancer.TestDiskBalancer |
|   | hadoop.hdfs.server.namenode.ha.TestDNFencingWithReplication |
|   | hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureToleration |
|   | hadoop.hdfs.server.namenode.TestFSImage |
|   | hadoop.hdfs.server.namenode.TestAuditLogs |
|   | hadoop.hdfs.TestDataTransferKeepalive |
|   | hadoop.hdfs.server.blockmanagement.TestBlockTokenWithDFSStriped |
|   | hadoop.hdfs.TestLeaseRecovery |
|   | hadoop.hdfs.server.namenode.ha.TestQuotasWithHA |
|   | hadoop.hdfs.server.datanode.TestDataNodeErasureCodingMetrics |
|   | hadoop.hdfs.server.namenode.TestAddBlockRetry |
|   | hadoop.security.TestRefreshUserMappings |
|   | hadoop.tracing.TestTraceAdmin |
|   | hadoop.hdfs.TestHFlush |
|   | hadoop.hdfs.TestMiniDFSCluster |
|   | hadoop.TestGenericRefresh |
|   | hadoop.hdfs.server.namenode.ha.TestBootstrapStandbyWithQJM |
|   | hadoop.hdfs.TestWriteConfigurationToDFS |
|   | hadoop.hdfs.server.blockmanagement.TestSequentialBlockId |
|   | hadoop.hdfs.server.datanode.TestIncrementalBrVariations |
|   | hadoop.hdfs.server.datanode.TestDataNodeLifeline |
|   | hadoop.hdfs.TestDFSStartupVersions |
|   | hadoop.hdfs.server.namenode.TestDefau

[jira] [Updated] (HDFS-11482) Add storage type demand to into DFSNetworkTopology#chooseRandom

2017-03-06 Thread Chen Liang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chen Liang updated HDFS-11482:
--
Attachment: HDFS-11482.002.patch

> Add storage type demand to into DFSNetworkTopology#chooseRandom
> ---
>
> Key: HDFS-11482
> URL: https://issues.apache.org/jira/browse/HDFS-11482
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Chen Liang
>Assignee: Chen Liang
> Attachments: HDFS-11482.001.patch, HDFS-11482.002.patch
>
>
> HDFS-11450 adds storage type info into network topology, with on this info we 
> may change chooseRandom to take storage type requirement as parameter, only 
> checking subtrees required storage type available. This way we avoid blindly 
> picking up nodes that are not applicable.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11314) Enforce set of enabled EC policies on the NameNode

2017-03-06 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11314?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-11314:
---
Attachment: HDFS-11314.004.patch

Here's a patch to fix the TestOIV failure, had to enable an EC policy.

> Enforce set of enabled EC policies on the NameNode
> --
>
> Key: HDFS-11314
> URL: https://issues.apache.org/jira/browse/HDFS-11314
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: erasure-coding
>Affects Versions: 3.0.0-alpha1
>Reporter: Andrew Wang
>Assignee: Andrew Wang
>Priority: Blocker
>  Labels: hdfs-ec-3.0-must-do
> Attachments: HDFS-11314.001.patch, HDFS-11314.002.patch, 
> HDFS-11314.003.patch, HDFS-11314.004.patch
>
>
> Filing based on discussion in HDFS-8095. A user might specify a policy that 
> is not appropriate for the cluster, e.g. a RS (10,4) policy when the cluster 
> only has 10 nodes. The NN should only allow the client to choose from a 
> pre-approved list determined by the cluster administrator.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11482) Add storage type demand to into DFSNetworkTopology#chooseRandom

2017-03-06 Thread Chen Liang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chen Liang updated HDFS-11482:
--
Status: Patch Available  (was: Open)

> Add storage type demand to into DFSNetworkTopology#chooseRandom
> ---
>
> Key: HDFS-11482
> URL: https://issues.apache.org/jira/browse/HDFS-11482
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Chen Liang
>Assignee: Chen Liang
> Attachments: HDFS-11482.001.patch
>
>
> HDFS-11450 adds storage type info into network topology, with on this info we 
> may change chooseRandom to take storage type requirement as parameter, only 
> checking subtrees required storage type available. This way we avoid blindly 
> picking up nodes that are not applicable.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11482) Add storage type demand to into DFSNetworkTopology#chooseRandom

2017-03-06 Thread Chen Liang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chen Liang updated HDFS-11482:
--
Attachment: HDFS-11482.001.patch

> Add storage type demand to into DFSNetworkTopology#chooseRandom
> ---
>
> Key: HDFS-11482
> URL: https://issues.apache.org/jira/browse/HDFS-11482
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode
>Reporter: Chen Liang
>Assignee: Chen Liang
> Attachments: HDFS-11482.001.patch
>
>
> HDFS-11450 adds storage type info into network topology, with on this info we 
> may change chooseRandom to take storage type requirement as parameter, only 
> checking subtrees required storage type available. This way we avoid blindly 
> picking up nodes that are not applicable.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11498) Make RestCsrfPreventionHandler and WebHdfsHandler compatible with Netty 4.0

2017-03-06 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11498?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-11498:
---
Release Note: This JIRA sets the Netty 4 dependency to 4.0.23. This is an 
incompatible change for the 3.0 release line, as 3.0.0-alpha1 and 3.0.0-alpha2 
depended on Netty 4.1.0.Beta5.

> Make RestCsrfPreventionHandler and WebHdfsHandler compatible with Netty 4.0
> ---
>
> Key: HDFS-11498
> URL: https://issues.apache.org/jira/browse/HDFS-11498
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.8.0
>Reporter: Andrew Wang
>Assignee: Andrew Wang
> Fix For: 2.8.0, 3.0.0-alpha3
>
> Attachments: HDFS-11498.001.patch, HDFS-11498.branch-2.001.patch
>
>
> Per discussion in HADOOP-13866, it looks like we can change 2.8.0 back to 
> exposing Netty 4.0, but still be ABI compatible with Netty 4.1 for users like 
> HBase that want to swap out the version.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11498) Make RestCsrfPreventionHandler and WebHdfsHandler compatible with Netty 4.0

2017-03-06 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11498?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-11498:
---
   Resolution: Fixed
Fix Version/s: 3.0.0-alpha3
   2.8.0
   Status: Resolved  (was: Patch Available)

Thanks again for the reviews Akira and Jason, after backporting the revert of 
HDFS-8377, the patch here applied cleanly. I've committed to all the branches 
back through branch-2.8.0, we can resume discussion on the Netty 4.1 upgrade 
over at HADOOP-13866.

> Make RestCsrfPreventionHandler and WebHdfsHandler compatible with Netty 4.0
> ---
>
> Key: HDFS-11498
> URL: https://issues.apache.org/jira/browse/HDFS-11498
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.8.0
>Reporter: Andrew Wang
>Assignee: Andrew Wang
> Fix For: 2.8.0, 3.0.0-alpha3
>
> Attachments: HDFS-11498.001.patch, HDFS-11498.branch-2.001.patch
>
>
> Per discussion in HADOOP-13866, it looks like we can change 2.8.0 back to 
> exposing Netty 4.0, but still be ABI compatible with Netty 4.1 for users like 
> HBase that want to swap out the version.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11376) Revert HDFS-8377 Support HTTP/2 in datanode

2017-03-06 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11376?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HDFS-11376:
---
Affects Version/s: (was: 2.9.0)
   2.8.0
 Target Version/s: 3.0.0-alpha3  (was: 2.9.0, 3.0.0-alpha3)
Fix Version/s: (was: 2.9.0)
   2.8.0

> Revert HDFS-8377 Support HTTP/2 in datanode
> ---
>
> Key: HDFS-11376
> URL: https://issues.apache.org/jira/browse/HDFS-11376
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 2.8.0, 3.0.0-alpha3
>Reporter: Andrew Wang
>Assignee: Xiao Chen
> Fix For: 2.8.0, 3.0.0-alpha3
>
>
> Tracking JIRA so this revert shows up in the release notes. See discussion on 
> HADOOP-13866.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11376) Revert HDFS-8377 Support HTTP/2 in datanode

2017-03-06 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11376?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898346#comment-15898346
 ] 

Andrew Wang commented on HDFS-11376:


Per discussion on HDFS-11498, I've backported this to branch-2.8 and 
branch-2.8.0 for inclusion in 2.8.0.

> Revert HDFS-8377 Support HTTP/2 in datanode
> ---
>
> Key: HDFS-11376
> URL: https://issues.apache.org/jira/browse/HDFS-11376
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 2.8.0, 3.0.0-alpha3
>Reporter: Andrew Wang
>Assignee: Xiao Chen
> Fix For: 2.8.0, 3.0.0-alpha3
>
>
> Tracking JIRA so this revert shows up in the release notes. See discussion on 
> HADOOP-13866.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11498) Make RestCsrfPreventionHandler and WebHdfsHandler compatible with Netty 4.0

2017-03-06 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11498?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898323#comment-15898323
 ] 

Andrew Wang commented on HDFS-11498:


Yea, sorry for only testing on branch-2, I assumed it'd be a smooth backport to 
2.8 as well.

+1 on Jason's recommendation. I'll work on backporting HDFS-11376, then I'll 
apply this JIRA (HDFS-11498) everywhere. Given that there are two +1s on this 
pretty straightforward change, I'll go ahead and commit everything if I can do 
it with minimal cherry-pick conflicts.



> Make RestCsrfPreventionHandler and WebHdfsHandler compatible with Netty 4.0
> ---
>
> Key: HDFS-11498
> URL: https://issues.apache.org/jira/browse/HDFS-11498
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.8.0
>Reporter: Andrew Wang
>Assignee: Andrew Wang
> Attachments: HDFS-11498.001.patch, HDFS-11498.branch-2.001.patch
>
>
> Per discussion in HADOOP-13866, it looks like we can change 2.8.0 back to 
> exposing Netty 4.0, but still be ABI compatible with Netty 4.1 for users like 
> HBase that want to swap out the version.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Lukas Majercak (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898290#comment-15898290
 ] 

Lukas Majercak commented on HDFS-11499:
---

LGTM

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.02.patch, HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Assigned] (HDFS-4934) add symlink support to WebHDFS server side

2017-03-06 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4934?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang reassigned HDFS-4934:
-

Assignee: (was: Andrew Wang)

> add symlink support to WebHDFS server side
> --
>
> Key: HDFS-4934
> URL: https://issues.apache.org/jira/browse/HDFS-4934
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.3.0
> Environment: followup on HADOOP-8040
>Reporter: Alejandro Abdelnur
>
> follow up on HADOOP-8040



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Assigned] (HDFS-4933) add symlink support to WebHDFS to HTTP REST API & client filesystem

2017-03-06 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4933?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang reassigned HDFS-4933:
-

Assignee: (was: Andrew Wang)

> add symlink support to WebHDFS to HTTP REST API & client filesystem
> ---
>
> Key: HDFS-4933
> URL: https://issues.apache.org/jira/browse/HDFS-4933
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.3.0
>Reporter: Alejandro Abdelnur
>
> follow up on HADOOP-8040



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Assigned] (HDFS-4935) add symlink support to HttpFS server side

2017-03-06 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang reassigned HDFS-4935:
-

Assignee: (was: Andrew Wang)

> add symlink support to HttpFS server side
> -
>
> Key: HDFS-4935
> URL: https://issues.apache.org/jira/browse/HDFS-4935
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 2.3.0
> Environment: followup on HADOOP-8040
>Reporter: Alejandro Abdelnur
>
> follow up on HADOOP-8040



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Manoj Govindassamy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Manoj Govindassamy updated HDFS-11499:
--
Attachment: HDFS-11499.02.patch

[~lukmajercak], [~jojochuang], [~linyiqun],
Attached v02 patch to address the following. Can you please take a look at the 
patch.
* {{BlockManager#commitOrCompleteLastBlock()}} to consider entering_maintenance 
replicase along with decommissioning replicas for usable replicas.
* Added unit test testFileCloseAfterEnteringMaintenance to 
{{TestMaintenanceState}} based on the test given by Yiqun. Without fix, the 
test fails at file close.

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.02.patch, HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-11487) Ozone: Exclude container protobuf files from findbugs check

2017-03-06 Thread Anu Engineer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11487?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer updated HDFS-11487:

  Resolution: Fixed
Hadoop Flags: Reviewed
Target Version/s: HDFS-7240
  Status: Resolved  (was: Patch Available)

[~yuanbo] Thank you for the contribution. I have committed this patch to the 
feature branch.

> Ozone: Exclude container protobuf files from findbugs check
> ---
>
> Key: HDFS-11487
> URL: https://issues.apache.org/jira/browse/HDFS-11487
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>Assignee: Yuanbo Liu
>  Labels: newbie
> Fix For: HDFS-7240
>
> Attachments: HDFS-11487-HDFS-7240.001.patch
>
>
> We moved the container protobuf definitions from hdfs project to hdfs-client. 
> This is to track that we need to add an exclusion to hdfs-client findbugs xml 
> to ignore those files.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11487) Ozone: Exclude container protobuf files from findbugs check

2017-03-06 Thread Anu Engineer (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898259#comment-15898259
 ] 

Anu Engineer commented on HDFS-11487:
-

+1, i will commit this shortly.

> Ozone: Exclude container protobuf files from findbugs check
> ---
>
> Key: HDFS-11487
> URL: https://issues.apache.org/jira/browse/HDFS-11487
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>Assignee: Yuanbo Liu
>  Labels: newbie
> Fix For: HDFS-7240
>
> Attachments: HDFS-11487-HDFS-7240.001.patch
>
>
> We moved the container protobuf definitions from hdfs project to hdfs-client. 
> This is to track that we need to add an exclusion to hdfs-client findbugs xml 
> to ignore those files.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Assigned] (HDFS-11505) Do not enable any erasure coding policies by default

2017-03-06 Thread Manoj Govindassamy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11505?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Manoj Govindassamy reassigned HDFS-11505:
-

Assignee: Manoj Govindassamy  (was: Andrew Wang)

> Do not enable any erasure coding policies by default
> 
>
> Key: HDFS-11505
> URL: https://issues.apache.org/jira/browse/HDFS-11505
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: erasure-coding
>Affects Versions: 3.0.0-alpha3
>Reporter: Andrew Wang
>Assignee: Manoj Govindassamy
>  Labels: hdfs-ec-3.0-must-do
>
> As discussed on HDFS-11314, administrators need to choose the correct set of 
> EC policies based on cluster size and desired fault-tolerance properties.
> This means we should not enable any EC policies by default, since any default 
> value could be incorrect.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Assigned] (HDFS-11506) Move ErasureCodingPolicyManager#getSystemDefaultPolicy to test code

2017-03-06 Thread Manoj Govindassamy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11506?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Manoj Govindassamy reassigned HDFS-11506:
-

Assignee: Manoj Govindassamy  (was: Andrew Wang)

> Move ErasureCodingPolicyManager#getSystemDefaultPolicy to test code
> ---
>
> Key: HDFS-11506
> URL: https://issues.apache.org/jira/browse/HDFS-11506
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: erasure-coding
>Affects Versions: 3.0.0-alpha3
>Reporter: Andrew Wang
>Assignee: Manoj Govindassamy
>  Labels: hdfs-ec-3.0-nice-to-have
>
> In HDFS-11416, we removed usages of getSystemDefaultPolicy in production 
> code. Let's move it to a test package to make sure no production code calls 
> it in the future.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-10838) Last full block report received time for each DN should be easily discoverable

2017-03-06 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898251#comment-15898251
 ] 

Arpit Agarwal commented on HDFS-10838:
--

Thanks [~surendrasingh]. +1 pending Jenkins.

HDFS Pre-commit runs appear to be busted right now due to a bad node, we should 
retrigger the Jenkins runs when the node is fixed.

> Last full block report received time for each DN should be easily discoverable
> --
>
> Key: HDFS-10838
> URL: https://issues.apache.org/jira/browse/HDFS-10838
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: ui
>Reporter: Arpit Agarwal
>Assignee: Surendra Singh Lilhore
> Attachments: DFSAdmin-Report.png, HDFS-10838-001.patch, 
> HDFS-10838.002.patch, HDFS-10838.003.patch, HDFS-10838.004.patch, 
> HDFS-10838.005.patch, HDFS-10838.006.patch, NN_UI.png, 
> NN_UI_relative_time_minutes.png, NN_UI_relative_time.png
>
>
> It should be easy for administrators to discover the time of last full block 
> report from each DataNode.
> We can show it in the NameNode web UI or in the output of {{hdfs dfsadmin 
> -report}}, or both.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Lukas Majercak (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898250#comment-15898250
 ] 

Lukas Majercak commented on HDFS-11499:
---

[~manojg], I don't really mind, let's go for the second option? Also, + 
[~mingma], do you want to check this? Seems like it is related to your 
HDFS-9390.

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Created] (HDFS-11506) Move ErasureCodingPolicyManager#getSystemDefaultPolicy to test code

2017-03-06 Thread Andrew Wang (JIRA)
Andrew Wang created HDFS-11506:
--

 Summary: Move ErasureCodingPolicyManager#getSystemDefaultPolicy to 
test code
 Key: HDFS-11506
 URL: https://issues.apache.org/jira/browse/HDFS-11506
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: erasure-coding
Affects Versions: 3.0.0-alpha3
Reporter: Andrew Wang
Assignee: Andrew Wang


In HDFS-11416, we removed usages of getSystemDefaultPolicy in production code. 
Let's move it to a test package to make sure no production code calls it in the 
future.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11314) Enforce set of enabled EC policies on the NameNode

2017-03-06 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898246#comment-15898246
 ] 

Andrew Wang commented on HDFS-11314:


Hi Rakesh, thanks for reviewing!

bq. I'd also prefer keep the list empty and lets users configure the right 
values based on their topology.

Great, I filed HDFS-11505. I'll handle this after we get this one in.

bq.  Could you please point me the way Namenode handles this during restarts 
and read from edit log. It would be good to add test case if not added.

I extended TestErasureCodingPolicies#testBasicSetECPolicy, little snippets here:

{code}
+// Verify that policies are successfully loaded even when policies
+// are disabled
+cluster.getConfiguration(0).set(
+DFSConfigKeys.DFS_NAMENODE_EC_POLICIES_ENABLED_KEY, "");
+cluster.restartNameNodes();
+cluster.waitActive();
{code}

{code}
+// Also check loading disabled EC policies from fsimage
+fs.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_ENTER);
+fs.saveNamespace();
+fs.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_LEAVE);
+cluster.restartNameNodes();
{code}

bq. Whats your opinion to provide a facility to dynamically/reconfigure 
enabling the policies with out needing to restart Namenode?

I think changing the set of EC policies is a pretty rare operation, so it's not 
that important.

One related nice work item would be doing this for the replication throttles, 
since that's an issue even now.

> Enforce set of enabled EC policies on the NameNode
> --
>
> Key: HDFS-11314
> URL: https://issues.apache.org/jira/browse/HDFS-11314
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: erasure-coding
>Affects Versions: 3.0.0-alpha1
>Reporter: Andrew Wang
>Assignee: Andrew Wang
>Priority: Blocker
>  Labels: hdfs-ec-3.0-must-do
> Attachments: HDFS-11314.001.patch, HDFS-11314.002.patch, 
> HDFS-11314.003.patch
>
>
> Filing based on discussion in HDFS-8095. A user might specify a policy that 
> is not appropriate for the cluster, e.g. a RS (10,4) policy when the cluster 
> only has 10 nodes. The NN should only allow the client to choose from a 
> pre-approved list determined by the cluster administrator.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11487) Ozone: Exclude container protobuf files from findbugs check

2017-03-06 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898239#comment-15898239
 ] 

Hadoop QA commented on HDFS-11487:
--

| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
31s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
2s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
19s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}  1m  7s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | HDFS-11487 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12856240/HDFS-11487-HDFS-7240.001.patch
 |
| Optional Tests |  asflicense  xml  |
| uname | Linux 2d1bf10b2882 3.13.0-107-generic #154-Ubuntu SMP Tue Dec 20 
09:57:27 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | HDFS-7240 / 0951726 |
| modules | C: hadoop-hdfs-project/hadoop-hdfs-client U: 
hadoop-hdfs-project/hadoop-hdfs-client |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18592/console |
| Powered by | Apache Yetus 0.5.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> Ozone: Exclude container protobuf files from findbugs check
> ---
>
> Key: HDFS-11487
> URL: https://issues.apache.org/jira/browse/HDFS-11487
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>Assignee: Yuanbo Liu
>  Labels: newbie
> Fix For: HDFS-7240
>
> Attachments: HDFS-11487-HDFS-7240.001.patch
>
>
> We moved the container protobuf definitions from hdfs project to hdfs-client. 
> This is to track that we need to add an exclusion to hdfs-client findbugs xml 
> to ignore those files.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11445) FSCK shows overall health stauts as corrupt even one replica is corrupt

2017-03-06 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11445?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898236#comment-15898236
 ] 

Jing Zhao commented on HDFS-11445:
--

Thanks for working on this, [~brahmareddy]!

I think you just find a scenario that the following inconsistency happens:
{code:title=BlockManager#createLocatedBlock}
NumberReplicas numReplicas = countNodes(blk);
final int numCorruptNodes = numReplicas.corruptReplicas();
final int numCorruptReplicas = corruptReplicas.numCorruptReplicas(blk);
if (numCorruptNodes != numCorruptReplicas) {
  LOG.warn("Inconsistent number of corrupt replicas for "
  + blk + " blockMap has " + numCorruptNodes
  + " but corrupt replicas map has " + numCorruptReplicas);
}
{code}

I also did some debugging using your unit test. Looks like the root cause for 
this inconsistency is: {{BlockInfo#setGenerationStampAndVerifyReplicas}} may 
remove a datanode storage from the block's storage list, but still leave the 
storage in the CorruptReplicasMap.

This inconsistency later can be fixed automatically, e.g., by a full block 
report. But maybe we should consider using 
{{BlockManager#removeStoredBlock(BlockInfo, DatanodeDescriptor)}} to remove all 
the records related to the block-dn pair.

> FSCK shows overall health stauts as corrupt even one replica is corrupt
> ---
>
> Key: HDFS-11445
> URL: https://issues.apache.org/jira/browse/HDFS-11445
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Brahma Reddy Battula
>Assignee: Brahma Reddy Battula
> Attachments: HDFS-11445.patch
>
>
> In the following scenario,FSCK shows overall health status as corrupt even 
> it's has one good replica.
> 1. Create file with 2 RF.
> 2. Shutdown one DN
> 3. Append to file again. 
> 4. Restart the DN
> 5. After block report, check Fsck



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Created] (HDFS-11505) Do not enable any erasure coding policies by default

2017-03-06 Thread Andrew Wang (JIRA)
Andrew Wang created HDFS-11505:
--

 Summary: Do not enable any erasure coding policies by default
 Key: HDFS-11505
 URL: https://issues.apache.org/jira/browse/HDFS-11505
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: erasure-coding
Affects Versions: 3.0.0-alpha3
Reporter: Andrew Wang
Assignee: Andrew Wang


As discussed on HDFS-11314, administrators need to choose the correct set of EC 
policies based on cluster size and desired fault-tolerance properties.

This means we should not enable any EC policies by default, since any default 
value could be incorrect.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11196) Ozone: Improve logging and error handling in the container layer

2017-03-06 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11196?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898231#comment-15898231
 ] 

Hadoop QA commented on HDFS-11196:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m  
0s{color} | {color:blue} Docker mode activated. {color} |
| {color:red}-1{color} | {color:red} patch {color} | {color:red}  0m  8s{color} 
| {color:red} HDFS-11196 does not apply to HDFS-7240. Rebase required? Wrong 
Branch? See https://wiki.apache.org/hadoop/HowToContribute for help. {color} |
\\
\\
|| Subsystem || Report/Notes ||
| JIRA Issue | HDFS-11196 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12856024/HDFS-11196-HDFS-7240.003.patch
 |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18589/console |
| Powered by | Apache Yetus 0.5.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> Ozone: Improve logging and error handling in the container layer
> 
>
> Key: HDFS-11196
> URL: https://issues.apache.org/jira/browse/HDFS-11196
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>Assignee: Anu Engineer
> Fix For: HDFS-7240
>
> Attachments: HDFS-11196-HDFS-7240.001.patch, 
> HDFS-11196-HDFS-7240.002.patch, HDFS-11196-HDFS-7240.003.patch
>
>
> Improve logging and error handling in container layer.
>  * With this change Storage Containers return StorageContainerException.
>  * Precondition checks fail with a human readable error.
>  * All failed requests are logged with traceID in the dispatcher.
>  * Returns proper error codes for corresponding failures.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11196) Ozone: Improve logging and error handling in the container layer

2017-03-06 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11196?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898232#comment-15898232
 ] 

Hadoop QA commented on HDFS-11196:
--

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m  
0s{color} | {color:blue} Docker mode activated. {color} |
| {color:red}-1{color} | {color:red} patch {color} | {color:red}  0m  8s{color} 
| {color:red} HDFS-11196 does not apply to HDFS-7240. Rebase required? Wrong 
Branch? See https://wiki.apache.org/hadoop/HowToContribute for help. {color} |
\\
\\
|| Subsystem || Report/Notes ||
| JIRA Issue | HDFS-11196 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12856024/HDFS-11196-HDFS-7240.003.patch
 |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/18590/console |
| Powered by | Apache Yetus 0.5.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> Ozone: Improve logging and error handling in the container layer
> 
>
> Key: HDFS-11196
> URL: https://issues.apache.org/jira/browse/HDFS-11196
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>Assignee: Anu Engineer
> Fix For: HDFS-7240
>
> Attachments: HDFS-11196-HDFS-7240.001.patch, 
> HDFS-11196-HDFS-7240.002.patch, HDFS-11196-HDFS-7240.003.patch
>
>
> Improve logging and error handling in container layer.
>  * With this change Storage Containers return StorageContainerException.
>  * Precondition checks fail with a human readable error.
>  * All failed requests are logged with traceID in the dispatcher.
>  * Returns proper error codes for corresponding failures.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Manoj Govindassamy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898222#comment-15898222
 ] 

Manoj Govindassamy commented on HDFS-11499:
---

yes [~lukmajercak], we need to additionally count in the ENTERING MAINTENANCE 
nodes as well. I am adding a new test based on the one given by [~linyiqun] in 
HDFS-11486 in TestMaintenanceState to cover this case. 

[~lukmajercak]/[~jojochuang], Shall I merge all the fix and test patches along 
with mine and post a complete patch covering both HDFS-11499 and HDFS-11486 ? 
Or shall I submit v02 patch for this HDFS-11499 alone with maintenance state 
included. Your suggestion please ?

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Lukas Majercak (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898211#comment-15898211
 ] 

Lukas Majercak commented on HDFS-11499:
---

Hi [~jojochuang], [~manojg]

So shall we count in the replicas on maintenance nodes as well? Can we need to 
add a test case to cover this/modify the one in the patch?

Thanks

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HDFS-11500) Java Unlimited Cryptography Extension doesn't work

2017-03-06 Thread Jeff W (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898191#comment-15898191
 ] 

Jeff W edited comment on HDFS-11500 at 3/6/17 9:42 PM:
---

Run command:

curl -k -v -u nn/hdfs-nn1.cloud.local@CLOUD.LOCAL:password --negotiate 
https://jn1.cloud.local:8481/getJournal?jid=hadoop1&segmentTxId=1&storageInfo=-63%3A1598677718%3A0%3ACID-2eecd392-dae7-480b-a867-5e0295c78648

Will got the same error information, it must be the Jetty server configuration 
issue.


was (Author: jeff...@yahoo.com):
Run command:

curl -k -v --negotiate 
https://jn1.cloud.local:8481/getJournal?jid=hadoop1&segmentTxId=1&storageInfo=-63%3A1598677718%3A0%3ACID-2eecd392-dae7-480b-a867-5e0295c78648

Will got the same error information, it must be the Jetty server configuration 
issue.

> Java Unlimited Cryptography Extension doesn't work
> --
>
> Key: HDFS-11500
> URL: https://issues.apache.org/jira/browse/HDFS-11500
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: journal-node, namenode
>Affects Versions: 2.7.3
> Environment: Centos 7, Java 1.8, HDFS HA
>Reporter: Jeff W
>  Labels: security
> Attachments: log_from_journal_node.txt, log_from_name_node.txt
>
>
> I'm setting up my HDFS cluster with JQM service, and enabled Kerberos 
> authentication. replaced JCE to unlimited extension, but I still got the 
> following error information from the Journal node:
> -
> 2017-03-04 15:31:39,566 INFO SecurityLogger.org.apache.hadoop.ipc.Server: 
> Auth successful for nn/hdfs-nn1.cloud.local@CLOUD.LOCAL (auth:KERBEROS)
> 2017-03-04 15:31:39,585 INFO 
> SecurityLogger.org.apache.hadoop.security.authorize.ServiceAuthorizationManager:
>  Authorization successful for nn/hdfs-nn1.cloud.local@CLOUD.LOCAL 
> (auth:KERBEROS) for protocol=interface 
> org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocol
> ...
> 2017-03-04 15:31:40,345 WARN 
> org.apache.hadoop.security.authentication.server.AuthenticationFilter: 
> Authentication exception: GSSException: Failure unspecified at GSS-API level 
> (Mechanism level: Invalid argument (400) - Cannot find key of appropriate 
> type to decrypt AP REP - AES256 CTS mode with HMAC SHA1-96)
> 
>  
> IPC authentication for Namenode looks good, but following with 
> AuthenticationFilter exception came from HTTP authentication, as I'm able to 
> see the same error message appears in Namenode's log:
> -
> 2017-03-04 15:31:40,402 ERROR 
> org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: caught exception 
> initializing 
> https://jn1.cloud.local:8481/getJournal?jid=hadoop1&segmentTxId=1&storageInfo=-63%3A1598677718%3A0%3ACID-2eecd392-dae7-480b-a867-5e0295c78648
> java.io.IOException: 
> org.apache.hadoop.security.authentication.client.AuthenticationException: 
> Authentication failed, status: 403, message: GSSException: Failure 
> unspecified at GSS-API level (Mechanism level: Invalid argument (400) - 
> Cannot find key of appropriate type to decrypt AP REP - AES256 CTS mode with 
> HMAC SHA1-96)
> at 
> org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:464)
> at 
> org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:456)
> at java.security.AccessController.doPrivileged(Native Method)
> ...
> Caused by: 
> org.apache.hadoop.security.authentication.client.AuthenticationException: 
> Authentication failed, status: 403, message: GSSException: Failure 
> unspecified at GSS-API level (Mechanism level: Invalid argument (400) - 
> Cannot find key of appropriate type to decrypt AP REP - AES256 CTS mode with 
> HMAC SHA1-96)
> at 
> org.apache.hadoop.security.authentication.client.AuthenticatedURL.extractToken(AuthenticatedURL.java:274)
> at 
> org.apache.hadoop.security.authentication.client.PseudoAuthenticator.authenticate(PseudoAuthenticator.java:77)
> at 
> org.apache.hadoop.security.authentication.client.KerberosAuthenticator.authenticate(KerberosAuthenticator.java:214)
> at 
> org.apache.hadoop.security.authentication.client.AuthenticatedURL.openConnection(AuthenticatedURL.java:215)
> at 
> org.apache.hadoop.hdfs.web.URLConnectionFactory.openConnection(URLConnectionFactory.java:161)
> at 
> org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:461)
> ... 30 more
> 2017-03-04 20:06:51,612 ERROR 
> org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Got error reading 
> edit log input stream 
> https://jn1.cloud.local:8481/getJournal?jid=hadoop1&segmentTxId=1&storageInfo=-63%3A1598677718%3A0%3ACID-2e

[jira] [Comment Edited] (HDFS-11500) Java Unlimited Cryptography Extension doesn't work

2017-03-06 Thread Jeff W (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898191#comment-15898191
 ] 

Jeff W edited comment on HDFS-11500 at 3/6/17 9:41 PM:
---

Run command:

curl -k -v --negotiate 
https://jn1.cloud.local:8481/getJournal?jid=hadoop1&segmentTxId=1&storageInfo=-63%3A1598677718%3A0%3ACID-2eecd392-dae7-480b-a867-5e0295c78648

Will got the same error information, it must be the Jetty server configuration 
issue.


was (Author: jeff...@yahoo.com):
Run command:

curl -k -v --negotiate 
https://jn1.cloud.local:8481/getJournal?jid=hadoop1&segmentTxId=1&storageInfo=-63%3A1598677718%3A0%3ACID-2eecd392-dae7-480b-a867-5e0295c78648

Will got same error information, it must be the Jetty server configuration 
issue.

> Java Unlimited Cryptography Extension doesn't work
> --
>
> Key: HDFS-11500
> URL: https://issues.apache.org/jira/browse/HDFS-11500
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: journal-node, namenode
>Affects Versions: 2.7.3
> Environment: Centos 7, Java 1.8, HDFS HA
>Reporter: Jeff W
>  Labels: security
> Attachments: log_from_journal_node.txt, log_from_name_node.txt
>
>
> I'm setting up my HDFS cluster with JQM service, and enabled Kerberos 
> authentication. replaced JCE to unlimited extension, but I still got the 
> following error information from the Journal node:
> -
> 2017-03-04 15:31:39,566 INFO SecurityLogger.org.apache.hadoop.ipc.Server: 
> Auth successful for nn/hdfs-nn1.cloud.local@CLOUD.LOCAL (auth:KERBEROS)
> 2017-03-04 15:31:39,585 INFO 
> SecurityLogger.org.apache.hadoop.security.authorize.ServiceAuthorizationManager:
>  Authorization successful for nn/hdfs-nn1.cloud.local@CLOUD.LOCAL 
> (auth:KERBEROS) for protocol=interface 
> org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocol
> ...
> 2017-03-04 15:31:40,345 WARN 
> org.apache.hadoop.security.authentication.server.AuthenticationFilter: 
> Authentication exception: GSSException: Failure unspecified at GSS-API level 
> (Mechanism level: Invalid argument (400) - Cannot find key of appropriate 
> type to decrypt AP REP - AES256 CTS mode with HMAC SHA1-96)
> 
>  
> IPC authentication for Namenode looks good, but following with 
> AuthenticationFilter exception came from HTTP authentication, as I'm able to 
> see the same error message appears in Namenode's log:
> -
> 2017-03-04 15:31:40,402 ERROR 
> org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: caught exception 
> initializing 
> https://jn1.cloud.local:8481/getJournal?jid=hadoop1&segmentTxId=1&storageInfo=-63%3A1598677718%3A0%3ACID-2eecd392-dae7-480b-a867-5e0295c78648
> java.io.IOException: 
> org.apache.hadoop.security.authentication.client.AuthenticationException: 
> Authentication failed, status: 403, message: GSSException: Failure 
> unspecified at GSS-API level (Mechanism level: Invalid argument (400) - 
> Cannot find key of appropriate type to decrypt AP REP - AES256 CTS mode with 
> HMAC SHA1-96)
> at 
> org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:464)
> at 
> org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:456)
> at java.security.AccessController.doPrivileged(Native Method)
> ...
> Caused by: 
> org.apache.hadoop.security.authentication.client.AuthenticationException: 
> Authentication failed, status: 403, message: GSSException: Failure 
> unspecified at GSS-API level (Mechanism level: Invalid argument (400) - 
> Cannot find key of appropriate type to decrypt AP REP - AES256 CTS mode with 
> HMAC SHA1-96)
> at 
> org.apache.hadoop.security.authentication.client.AuthenticatedURL.extractToken(AuthenticatedURL.java:274)
> at 
> org.apache.hadoop.security.authentication.client.PseudoAuthenticator.authenticate(PseudoAuthenticator.java:77)
> at 
> org.apache.hadoop.security.authentication.client.KerberosAuthenticator.authenticate(KerberosAuthenticator.java:214)
> at 
> org.apache.hadoop.security.authentication.client.AuthenticatedURL.openConnection(AuthenticatedURL.java:215)
> at 
> org.apache.hadoop.hdfs.web.URLConnectionFactory.openConnection(URLConnectionFactory.java:161)
> at 
> org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:461)
> ... 30 more
> 2017-03-04 20:06:51,612 ERROR 
> org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Got error reading 
> edit log input stream 
> https://jn1.cloud.local:8481/getJournal?jid=hadoop1&segmentTxId=1&storageInfo=-63%3A1598677718%3A0%3ACID-2eecd392-dae7-480b-a867-5e0295c78648;
>  failing over 

[jira] [Commented] (HDFS-11500) Java Unlimited Cryptography Extension doesn't work

2017-03-06 Thread Jeff W (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898191#comment-15898191
 ] 

Jeff W commented on HDFS-11500:
---

Run command:

curl -k -v --negotiate 
https://jn1.cloud.local:8481/getJournal?jid=hadoop1&segmentTxId=1&storageInfo=-63%3A1598677718%3A0%3ACID-2eecd392-dae7-480b-a867-5e0295c78648

Will got same error information, it must be the Jetty server configuration 
issue.

> Java Unlimited Cryptography Extension doesn't work
> --
>
> Key: HDFS-11500
> URL: https://issues.apache.org/jira/browse/HDFS-11500
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: journal-node, namenode
>Affects Versions: 2.7.3
> Environment: Centos 7, Java 1.8, HDFS HA
>Reporter: Jeff W
>  Labels: security
> Attachments: log_from_journal_node.txt, log_from_name_node.txt
>
>
> I'm setting up my HDFS cluster with JQM service, and enabled Kerberos 
> authentication. replaced JCE to unlimited extension, but I still got the 
> following error information from the Journal node:
> -
> 2017-03-04 15:31:39,566 INFO SecurityLogger.org.apache.hadoop.ipc.Server: 
> Auth successful for nn/hdfs-nn1.cloud.local@CLOUD.LOCAL (auth:KERBEROS)
> 2017-03-04 15:31:39,585 INFO 
> SecurityLogger.org.apache.hadoop.security.authorize.ServiceAuthorizationManager:
>  Authorization successful for nn/hdfs-nn1.cloud.local@CLOUD.LOCAL 
> (auth:KERBEROS) for protocol=interface 
> org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocol
> ...
> 2017-03-04 15:31:40,345 WARN 
> org.apache.hadoop.security.authentication.server.AuthenticationFilter: 
> Authentication exception: GSSException: Failure unspecified at GSS-API level 
> (Mechanism level: Invalid argument (400) - Cannot find key of appropriate 
> type to decrypt AP REP - AES256 CTS mode with HMAC SHA1-96)
> 
>  
> IPC authentication for Namenode looks good, but following with 
> AuthenticationFilter exception came from HTTP authentication, as I'm able to 
> see the same error message appears in Namenode's log:
> -
> 2017-03-04 15:31:40,402 ERROR 
> org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: caught exception 
> initializing 
> https://jn1.cloud.local:8481/getJournal?jid=hadoop1&segmentTxId=1&storageInfo=-63%3A1598677718%3A0%3ACID-2eecd392-dae7-480b-a867-5e0295c78648
> java.io.IOException: 
> org.apache.hadoop.security.authentication.client.AuthenticationException: 
> Authentication failed, status: 403, message: GSSException: Failure 
> unspecified at GSS-API level (Mechanism level: Invalid argument (400) - 
> Cannot find key of appropriate type to decrypt AP REP - AES256 CTS mode with 
> HMAC SHA1-96)
> at 
> org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:464)
> at 
> org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:456)
> at java.security.AccessController.doPrivileged(Native Method)
> ...
> Caused by: 
> org.apache.hadoop.security.authentication.client.AuthenticationException: 
> Authentication failed, status: 403, message: GSSException: Failure 
> unspecified at GSS-API level (Mechanism level: Invalid argument (400) - 
> Cannot find key of appropriate type to decrypt AP REP - AES256 CTS mode with 
> HMAC SHA1-96)
> at 
> org.apache.hadoop.security.authentication.client.AuthenticatedURL.extractToken(AuthenticatedURL.java:274)
> at 
> org.apache.hadoop.security.authentication.client.PseudoAuthenticator.authenticate(PseudoAuthenticator.java:77)
> at 
> org.apache.hadoop.security.authentication.client.KerberosAuthenticator.authenticate(KerberosAuthenticator.java:214)
> at 
> org.apache.hadoop.security.authentication.client.AuthenticatedURL.openConnection(AuthenticatedURL.java:215)
> at 
> org.apache.hadoop.hdfs.web.URLConnectionFactory.openConnection(URLConnectionFactory.java:161)
> at 
> org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:461)
> ... 30 more
> 2017-03-04 20:06:51,612 ERROR 
> org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Got error reading 
> edit log input stream 
> https://jn1.cloud.local:8481/getJournal?jid=hadoop1&segmentTxId=1&storageInfo=-63%3A1598677718%3A0%3ACID-2eecd392-dae7-480b-a867-5e0295c78648;
>  failing over to edit log 
> https://jn2.cloud.local:8481/getJournal?jid=hadoop1&segmentTxId=1&storageInfo=-63%3A1598677718%3A0%3ACID-2eecd392-dae7-480b-a867-5e0295c78648
> 
> org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream$PrematureEOFException:
>  got premature end-of-file at txid 0; expected file to go up to 2
> at 
> org.apache

[jira] [Commented] (HDFS-11496) Ozone: Merge with trunk needed a ignore duplicate entry in pom file due to shading

2017-03-06 Thread Anu Engineer (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11496?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898138#comment-15898138
 ] 

Anu Engineer commented on HDFS-11496:
-

[~busbey] That makes sense, I think we did add netty dependency. Thanks for 
your suggestions,  I will fix this before merging the way you are suggesting.

> Ozone: Merge with trunk needed a ignore duplicate entry in pom file due to 
> shading
> --
>
> Key: HDFS-11496
> URL: https://issues.apache.org/jira/browse/HDFS-11496
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>
> The trunk merge needed a hack in the 
> hadoop-client-modules/hadoop-client-check-test-invariants/pom.xml to ignore 
> netty files. Not sure that is the right thing to do, so tracking here if we 
> need to revert this change before we merge.
> {code}
> ~/a/hadoop> git diff 
> hadoop-client-modules/hadoop-client-check-test-invariants/pom.xml
> ─
> modified: hadoop-client-modules/hadoop-client-check-test-invariants/pom.xml
> ─
> @ pom.xml:106 @
> *
>   
> 
> 
>   
>   
>   
> 
>   
> 
>   
> io.netty
> netty
> *
>   
> 
>   
> 
>   
> {code}
> [~andrew.wang] or [~busbey] Would one of you care to comment if this is the 
> right thing to do ? Thanks in advance.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11496) Ozone: Merge with trunk needed a ignore duplicate entry in pom file due to shading

2017-03-06 Thread Sean Busbey (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11496?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898103#comment-15898103
 ] 

Sean Busbey commented on HDFS-11496:


I would guess that the Ozone branch added netty as a dependency for something 
that is transitively included in minicluster. using maven's dependency tree 
should show you where the dependency comes in.

> Ozone: Merge with trunk needed a ignore duplicate entry in pom file due to 
> shading
> --
>
> Key: HDFS-11496
> URL: https://issues.apache.org/jira/browse/HDFS-11496
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>
> The trunk merge needed a hack in the 
> hadoop-client-modules/hadoop-client-check-test-invariants/pom.xml to ignore 
> netty files. Not sure that is the right thing to do, so tracking here if we 
> need to revert this change before we merge.
> {code}
> ~/a/hadoop> git diff 
> hadoop-client-modules/hadoop-client-check-test-invariants/pom.xml
> ─
> modified: hadoop-client-modules/hadoop-client-check-test-invariants/pom.xml
> ─
> @ pom.xml:106 @
> *
>   
> 
> 
>   
>   
>   
> 
>   
> 
>   
> io.netty
> netty
> *
>   
> 
>   
> 
>   
> {code}
> [~andrew.wang] or [~busbey] Would one of you care to comment if this is the 
> right thing to do ? Thanks in advance.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11499) Decommissioning stuck because of failing recovery

2017-03-06 Thread Manoj Govindassamy (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898100#comment-15898100
 ] 

Manoj Govindassamy commented on HDFS-11499:
---

Sure [~jojochuang]. I can give you a patch for the same, soon.

> Decommissioning stuck because of failing recovery
> -
>
> Key: HDFS-11499
> URL: https://issues.apache.org/jira/browse/HDFS-11499
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs, namenode
>Affects Versions: 2.7.1, 2.7.2, 2.7.3, 3.0.0-alpha2
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>  Labels: blockmanagement, decommission, recovery
> Fix For: 3.0.0-alpha3
>
> Attachments: HDFS-11499.patch
>
>
> Block recovery will fail to finalize the file if the locations of the last, 
> incomplete block are being decommissioned. Vice versa, the decommissioning 
> will be stuck, waiting for the last block to be completed.
> {code:xml}
> org.apache.hadoop.ipc.RemoteException(java.lang.IllegalStateException): 
> Failed to finalize INodeFile testRecoveryFile since blocks[255] is 
> non-complete, where blocks=[blk_1073741825_1001, blk_1073741826_1002...
> {code}
> The fix is to count replicas on decommissioning nodes when completing last 
> block in BlockManager.commitOrCompleteLastBlock, as we know that the 
> DecommissionManager will not decommission a node that has UC blocks.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11163) Mover should move the file blocks to default storage once policy is unset

2017-03-06 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898092#comment-15898092
 ] 

Chris Nauroth commented on HDFS-11163:
--

The {{FsServerDefaults}} class is annotated {{Public}}, so let's maintain the 
existing constructor signature and add a new constructor that supports passing 
default storage policy ID.  The old constructor can delegate to the new 
constructor with default storage policy ID of 0.  That also would remove the 
need to change {{FtpConfigKeys}} and {{LocalConfigKeys}} in this patch.

The logic is looking good to me, but I'd still like a second opinion review 
before committing anything.

> Mover should move the file blocks to default storage once policy is unset
> -
>
> Key: HDFS-11163
> URL: https://issues.apache.org/jira/browse/HDFS-11163
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: balancer & mover
>Affects Versions: 2.8.0
>Reporter: Surendra Singh Lilhore
>Assignee: Surendra Singh Lilhore
> Attachments: HDFS-11163-001.patch, HDFS-11163-002.patch, 
> HDFS-11163-003.patch, HDFS-11163-004.patch, HDFS-11163-005.patch, 
> temp-YARN-6278.HDFS-11163.patch
>
>
> HDFS-9534 added new API in FileSystem to unset the storage policy. Once 
> policy is unset blocks should move back to the default storage policy.
> Currently mover is not moving file blocks which have zero storage ID
> {code}
>   // currently we ignore files with unspecified storage policy
>   if (policyId == HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED) {
> return;
>   }
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11496) Ozone: Merge with trunk needed a ignore duplicate entry in pom file due to shading

2017-03-06 Thread Anu Engineer (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11496?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15898035#comment-15898035
 ] 

Anu Engineer commented on HDFS-11496:
-

[~busbey] Thanks for your comment. I agree that changing invariants is a bad 
idea. That is why I filed this JIRA to get your expert comments. I will try to 
fix it the way you are suggesting. Before I do that, I wanted to ask you why 
this is not happening in trunk and why this change is needed in Ozone branch, 
any guesses to where the root issue might be ?

> Ozone: Merge with trunk needed a ignore duplicate entry in pom file due to 
> shading
> --
>
> Key: HDFS-11496
> URL: https://issues.apache.org/jira/browse/HDFS-11496
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: ozone
>Affects Versions: HDFS-7240
>Reporter: Anu Engineer
>
> The trunk merge needed a hack in the 
> hadoop-client-modules/hadoop-client-check-test-invariants/pom.xml to ignore 
> netty files. Not sure that is the right thing to do, so tracking here if we 
> need to revert this change before we merge.
> {code}
> ~/a/hadoop> git diff 
> hadoop-client-modules/hadoop-client-check-test-invariants/pom.xml
> ─
> modified: hadoop-client-modules/hadoop-client-check-test-invariants/pom.xml
> ─
> @ pom.xml:106 @
> *
>   
> 
> 
>   
>   
>   
> 
>   
> 
>   
> io.netty
> netty
> *
>   
> 
>   
> 
>   
> {code}
> [~andrew.wang] or [~busbey] Would one of you care to comment if this is the 
> right thing to do ? Thanks in advance.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



  1   2   >