[jira] [Commented] (HDFS-13541) NameNode Port based selective encryption

2018-05-25 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-13541?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16491147#comment-16491147
 ] 

Benoy Antony commented on HDFS-13541:
-

{quote}
 in our environment, cross data center traffic actually compose a small 
fraction of all traffic, having additional DataXceiverServer thread sitting and 
listening on every single data node, but being idle most of time does not seem 
to be ideal. 
{quote}

If its a small fraction, then one option would be to use (1)  swebhdfs  or (2) 
HTTPFS gateway which uses RPC. They work seamlessly for clients via the 
FileSystem implementation and swebhdfs scheme.

(1) introduces HTTP protocol overhead and lack of QOS on namenode.  
(2) introduces HTTP protocol overhead and an additional hop. 

But these are the common solutions to this problem for small fractions of data 
which needs to be treated differently.  But your use case may be different 
which warrants secure and non secure  RPC and data transfer. 

> NameNode Port based selective encryption
> 
>
> Key: HDFS-13541
> URL: https://issues.apache.org/jira/browse/HDFS-13541
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: datanode, namenode, security
>Reporter: Chen Liang
>Assignee: Chen Liang
>Priority: Major
> Attachments: NameNode Port based selective encryption-v1.pdf
>
>
> Here at LinkedIn, one issue we face is that we need to enforce different 
> security requirement based on the location of client and the cluster. 
> Specifically, for clients from outside of the data center, it is required by 
> regulation that all traffic must be encrypted. But for clients within the 
> same data center, unencrypted connections are more desired to avoid the high 
> encryption overhead. 
> HADOOP-10221 introduced pluggable SASL resolver, based on which HADOOP-10335 
> introduced WhitelistBasedResolver which solves the same problem. However we 
> found it difficult to fit into our environment for several reasons. In this 
> JIRA, on top of pluggable SASL resolver, *we propose a different approach of 
> running RPC two ports on NameNode, and the two ports will be enforcing 
> encrypted and unencrypted connections respectively, and the following 
> DataNode access will simply follow the same behaviour of 
> encryption/unencryption*. Then by blocking unencrypted port on datacenter 
> firewall, we can completely block unencrypted external access.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-13541) NameNode Port based selective encryption

2018-05-10 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-13541?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16470684#comment-16470684
 ] 

Benoy Antony commented on HDFS-13541:
-

Thanks [~vagarychen]. I have reviewed the document. Overall, I believe, this 
approach makes the administration of selective encryption simpler. The 
selection can be done using firewall rules.

I have a few comments based on my understanding of the design.
 # Currently the input to _SaslPropertiesResolver_ methods is ip address. This 
feature adds ingress port as an input. I believe, it should be made generic so 
that we do not have to change it in future. We could also think of passing 
additional parameters based on the connection even though they may not be used 
in the current implementations of _SaslPropertiesResolver._
 # _The_  _selective data transfer protection_  design needs more scrutiny as 
it's protection is derived from _selective RPC protection_. Based on what I 
understood, the protection is dictated by the value of _encrypted message_ sent 
by the client when it handshakes with datanode just before the data transfer. 
The _encrypted message is either_ _auth_ or _auth_conf_ encrypted by the secret 
shared between namenode and the datanode. If so, what prevents the external  
client from replaying the _encrypted message_ from a different connection 
between an internal client and datanode ?
 # Another side effect of derived QOP for data transfer protection is that one 
cannot enable RPC protection alone with this approach.
 # As mentioned in the document, Encrypting the entire data pipeline is not 
necessary. I believe, it should be optimized.
 # All things equal, I prefer the approach where datanode also listens on two 
ports, as it makes the entire approach easy to understand. It will also solve 
issues specified in #2  and #3 above. Those issues are the result of QOP of 
data transfer operation becoming a derivative of RPC operation. 

 

> NameNode Port based selective encryption
> 
>
> Key: HDFS-13541
> URL: https://issues.apache.org/jira/browse/HDFS-13541
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: datanode, namenode, security
>Reporter: Chen Liang
>Assignee: Chen Liang
>Priority: Major
> Attachments: NameNode Port based selective encryption-v1.pdf
>
>
> Here at LinkedIn, one issue we face is that we need to enforce different 
> security requirement based on the location of client and the cluster. 
> Specifically, for clients from outside of the data center, it is required by 
> regulation that all traffic must be encrypted. But for clients within the 
> same data center, unencrypted connections are more desired to avoid the high 
> encryption overhead. 
> HADOOP-10221 introduced pluggable SASL resolver, based on which HADOOP-10335 
> introduced WhitelistBasedResolver which solves the same problem. However we 
> found it difficult to fit into our environment for several reasons. In this 
> JIRA, on top of pluggable SASL resolver, *we propose a different approach of 
> running RPC two ports on NameNode, and the two ports will be enforcing 
> encrypted and unencrypted connections respectively, and the following 
> DataNode access will simply follow the same behaviour of 
> encryption/unencryption*. Then by blocking unencrypted port on datacenter 
> firewall, we can completely block unencrypted external access.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-7466) Allow different values for dfs.datanode.balance.max.concurrent.moves per datanode

2017-10-16 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7466?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16206818#comment-16206818
 ] 

Benoy Antony commented on HDFS-7466:


I think [~dan...@cloudera.com]'s suggestion makes a  lot of sense. I got some 
time now to implement this scheme. I'll first submit a brief design spec on 
this.

> Allow different values for dfs.datanode.balance.max.concurrent.moves per 
> datanode
> -
>
> Key: HDFS-7466
> URL: https://issues.apache.org/jira/browse/HDFS-7466
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>
> It is possible to configure different values for  
> _dfs.datanode.balance.max.concurrent.moves_ per datanode.  But the value will 
> be used by balancer/mover which obtains the value from its own configuration. 
> The correct approach will be to obtain the value from the datanode itself.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Work started] (HDFS-7466) Allow different values for dfs.datanode.balance.max.concurrent.moves per datanode

2017-10-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HDFS-7466 started by Benoy Antony.
--
> Allow different values for dfs.datanode.balance.max.concurrent.moves per 
> datanode
> -
>
> Key: HDFS-7466
> URL: https://issues.apache.org/jira/browse/HDFS-7466
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>
> It is possible to configure different values for  
> _dfs.datanode.balance.max.concurrent.moves_ per datanode.  But the value will 
> be used by balancer/mover which obtains the value from its own configuration. 
> The correct approach will be to obtain the value from the datanode itself.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-5290) Port Encryption Support in DataTransfer protocol to Hadoop 1

2017-10-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5290?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony resolved HDFS-5290.

Resolution: Won't Do

Cleaning up jiras which is not relevant anymore.

> Port Encryption Support in DataTransfer protocol to Hadoop 1
> 
>
> Key: HDFS-5290
> URL: https://issues.apache.org/jira/browse/HDFS-5290
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: security
>Affects Versions: 1.2.1
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 1.3.0
>
> Attachments: HDFS-5290.patch
>
>
> Currently all HDFS RPCs performed by NNs/DNs/clients can be optionally 
> encrypted. However, actual data read or written between DNs and clients (or 
> DNs to DNs) is sent in the clear. When processing sensitive data on a shared 
> cluster, confidentiality of the data read/written from/to HDFS may be desired.
> This change is already done in Hadoop 2 and trunk. 
> The task is to port this change to Hadoop 1.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-4823) Inode.toString () should return the full path

2017-10-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4823?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony resolved HDFS-4823.

Resolution: Won't Do

Cleaning up jiras which is not relevant anymore.

> Inode.toString ()  should return the full path 
> ---
>
> Key: HDFS-4823
> URL: https://issues.apache.org/jira/browse/HDFS-4823
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Affects Versions: 1.1.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>Priority: Minor
> Attachments: HDFS-4823.patch
>
>
> Indoe.ToString() is used in many error messages. This gives the name of the 
> file / directory, but not the fullpath.
> org.apache.hadoop.security.AccessControlException 
> org.apache.hadoop.security.AccessControlException: Permission denied: 
> user=user1, access=WRITE, inode="warehouse":user2:supergroup:rwxrwxr-x)
> The fix is to provide the full path n line with Hadoop 2.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11384) Add option for balancer to disperse getBlocks calls to avoid NameNode's rpc.CallQueueLength spike

2017-03-09 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11384?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15903651#comment-15903651
 ] 

Benoy Antony commented on HDFS-11384:
-

 [~zhaoyunjiong], Could you please check to see if sleeping outside the sync 
block with a potentially longer delay is sufficient to avoid spike ?



> Add option for balancer to disperse getBlocks calls to avoid NameNode's 
> rpc.CallQueueLength spike
> -
>
> Key: HDFS-11384
> URL: https://issues.apache.org/jira/browse/HDFS-11384
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer & mover
>Affects Versions: 2.7.3
>Reporter: yunjiong zhao
>Assignee: yunjiong zhao
> Attachments: balancer.day.png, balancer.week.png, HDFS-11384.001.patch
>
>
> When running balancer on hadoop cluster which have more than 3000 Datanodes 
> will cause NameNode's rpc.CallQueueLength spike. We observed this situation 
> could cause Hbase cluster failure due to RegionServer's WAL timeout.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11384) Add option for balancer to disperse getBlocks calls to avoid NameNode's rpc.CallQueueLength spike

2017-03-01 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11384?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15891587#comment-15891587
 ] 

Benoy Antony commented on HDFS-11384:
-

Thanks for the explanation, [~zhaoyunjiong].
The patch looks good.  I will commit this tomorrow if there are no other 
comments.

> Add option for balancer to disperse getBlocks calls to avoid NameNode's 
> rpc.CallQueueLength spike
> -
>
> Key: HDFS-11384
> URL: https://issues.apache.org/jira/browse/HDFS-11384
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer & mover
>Affects Versions: 2.7.3
>Reporter: yunjiong zhao
>Assignee: yunjiong zhao
> Attachments: balancer.day.png, balancer.week.png, HDFS-11384.001.patch
>
>
> When running balancer on hadoop cluster which have more than 3000 Datanodes 
> will cause NameNode's rpc.CallQueueLength spike. We observed this situation 
> could cause Hbase cluster failure due to RegionServer's WAL timeout.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11384) Add option for balancer to disperse getBlocks calls to avoid NameNode's rpc.CallQueueLength spike

2017-03-01 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11384?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15890957#comment-15890957
 ] 

Benoy Antony commented on HDFS-11384:
-

Sleeping inside the *Synchronized* block should be avoided as it will lock 
prevent other threads from obtaining the lock while the thread is sleeping. 
One tradeoff in sleeping fixed vs variable time is that code gets complicated. 
Since by default, the delay is not applied, it is okay to sleep for a fixed 
interval after getBlocks(). 

> Add option for balancer to disperse getBlocks calls to avoid NameNode's 
> rpc.CallQueueLength spike
> -
>
> Key: HDFS-11384
> URL: https://issues.apache.org/jira/browse/HDFS-11384
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer & mover
>Affects Versions: 2.7.3
>Reporter: yunjiong zhao
>Assignee: yunjiong zhao
> Attachments: balancer.day.png, balancer.week.png, HDFS-11384.001.patch
>
>
> When running balancer on hadoop cluster which have more than 3000 Datanodes 
> will cause NameNode's rpc.CallQueueLength spike. We observed this situation 
> could cause Hbase cluster failure due to RegionServer's WAL timeout.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HDFS-11384) Add option for balancer to disperse getBlocks calls to avoid NameNode's rpc.CallQueueLength spike

2017-03-01 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11384?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15890957#comment-15890957
 ] 

Benoy Antony edited comment on HDFS-11384 at 3/1/17 8:17 PM:
-

Sleeping inside the *Synchronized* block should be avoided as it will prevent 
other threads from obtaining the lock while the thread is sleeping. 
One tradeoff in sleeping fixed vs variable time is that code gets complicated. 
Since by default, the delay is not applied, it is okay to sleep for a fixed 
interval after getBlocks(). 


was (Author: benoyantony):
Sleeping inside the *Synchronized* block should be avoided as it will lock 
prevent other threads from obtaining the lock while the thread is sleeping. 
One tradeoff in sleeping fixed vs variable time is that code gets complicated. 
Since by default, the delay is not applied, it is okay to sleep for a fixed 
interval after getBlocks(). 

> Add option for balancer to disperse getBlocks calls to avoid NameNode's 
> rpc.CallQueueLength spike
> -
>
> Key: HDFS-11384
> URL: https://issues.apache.org/jira/browse/HDFS-11384
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer & mover
>Affects Versions: 2.7.3
>Reporter: yunjiong zhao
>Assignee: yunjiong zhao
> Attachments: balancer.day.png, balancer.week.png, HDFS-11384.001.patch
>
>
> When running balancer on hadoop cluster which have more than 3000 Datanodes 
> will cause NameNode's rpc.CallQueueLength spike. We observed this situation 
> could cause Hbase cluster failure due to RegionServer's WAL timeout.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-11384) Add option for balancer to disperse getBlocks calls to avoid NameNode's rpc.CallQueueLength spike

2017-02-28 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-11384?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15888766#comment-15888766
 ] 

Benoy Antony commented on HDFS-11384:
-

[~zhaoyunjiong],
 If there are blocks to balance, then there will be sufficient delays between 
successive getBlocks. 
In such cases, we do not have to sleep. 
It will be better to keep track of the interval between successive getBlocks 
and sleep only for the required time.
Can you also write a unit test to cover this change ?


> Add option for balancer to disperse getBlocks calls to avoid NameNode's 
> rpc.CallQueueLength spike
> -
>
> Key: HDFS-11384
> URL: https://issues.apache.org/jira/browse/HDFS-11384
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer & mover
>Affects Versions: 2.7.3
>Reporter: yunjiong zhao
>Assignee: yunjiong zhao
> Attachments: balancer.day.png, balancer.week.png, HDFS-11384.001.patch
>
>
> When running balancer on hadoop cluster which have more than 3000 Datanodes 
> will cause NameNode's rpc.CallQueueLength spike. We observed this situation 
> could cause Hbase cluster failure due to RegionServer's WAL timeout.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-10735) Distcp using webhdfs on secure HA clusters fails with StandbyException

2016-10-14 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-10735:

  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

Committed to trunk , branch-2  and branch-2.8

> Distcp using webhdfs on secure HA clusters fails with StandbyException
> --
>
> Key: HDFS-10735
> URL: https://issues.apache.org/jira/browse/HDFS-10735
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.7.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.8.0
>
> Attachments: HDFS-10735-001.patch, exception.txt
>
>
> A distcp running on HADOOP 2.7.2 cluster , copying from HADOOP 2.4.1 cluster 
> fails.
> Both are secure clusters.
> The logs indicates that task contacts the StandBy name node on the remote 
> cluster i first and gets a standByException. 
> Instead of retrying with the active name node on the remote cluster, the 
> execution errors out throwing InvalidToken
> The exception is attached



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-10735) Distcp using webhdfs on secure HA clusters fails with StandbyException

2016-10-13 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-10735:

Description: 
A distcp running on HADOOP 2.7.2 cluster , copying from HADOOP 2.4.1 cluster 
fails.
Both are secure clusters.
The logs indicates that task contacts the StandBy name node on the remote 
cluster i first and gets a standByException. 
Instead of retrying with the active name node on the remote cluster, the 
execution errors out throwing InvalidToken

The exception is attached

  was:
A distcp running on HADOOP 2.7.2 cluster , copying from HADOOP 2.4.1 cluster 
fails.
Both are secure clusters.
The logs indicates that task contacts the StandBy name node on the remote 
cluster is contacted first and gets a standByException. 
Instead of retrying with the active name node on the remote cluster, the 
execution errors out throwing InvalidToken

The exception is attached


> Distcp using webhdfs on secure HA clusters fails with StandbyException
> --
>
> Key: HDFS-10735
> URL: https://issues.apache.org/jira/browse/HDFS-10735
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.7.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.8.0
>
> Attachments: HDFS-10735-001.patch, exception.txt
>
>
> A distcp running on HADOOP 2.7.2 cluster , copying from HADOOP 2.4.1 cluster 
> fails.
> Both are secure clusters.
> The logs indicates that task contacts the StandBy name node on the remote 
> cluster i first and gets a standByException. 
> Instead of retrying with the active name node on the remote cluster, the 
> execution errors out throwing InvalidToken
> The exception is attached



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-10735) Distcp using webhdfs on secure HA clusters fails with StandbyException

2016-10-12 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10735?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15570939#comment-15570939
 ] 

Benoy Antony commented on HDFS-10735:
-

[~arpitagarwal], Thanks for the review. Since the issue happens when 
interacting with an older version of HDFS, it will be difficult to test.
If there are no further comments, I'll commit this in a couple of days.

> Distcp using webhdfs on secure HA clusters fails with StandbyException
> --
>
> Key: HDFS-10735
> URL: https://issues.apache.org/jira/browse/HDFS-10735
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.7.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.8.0
>
> Attachments: HDFS-10735-001.patch, exception.txt
>
>
> A distcp running on HADOOP 2.7.2 cluster , copying from HADOOP 2.4.1 cluster 
> fails.
> Both are secure clusters.
> The logs indicates that task contacts the StandBy name node on the remote 
> cluster is contacted first and gets a standByException. 
> Instead of retrying with the active name node on the remote cluster, the 
> execution errors out throwing InvalidToken
> The exception is attached



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-10735) Distcp using webhdfs on secure HA clusters fails with StandbyException

2016-08-08 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-10735:

Attachment: HDFS-10735-001.patch

> Distcp using webhdfs on secure HA clusters fails with StandbyException
> --
>
> Key: HDFS-10735
> URL: https://issues.apache.org/jira/browse/HDFS-10735
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.7.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.8.0
>
> Attachments: HDFS-10735-001.patch, exception.txt
>
>
> A distcp running on HADOOP 2.7.2 cluster , copying from HADOOP 2.4.1 cluster 
> fails.
> Both are secure clusters.
> The logs indicates that task contacts the StandBy name node on the remote 
> cluster is contacted first and gets a standByException. 
> Instead of retrying with the active name node on the remote cluster, the 
> execution errors out throwing InvalidToken
> The exception is attached



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-10735) Distcp using webhdfs on secure HA clusters fails with StandbyException

2016-08-08 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-10735:

Attachment: (was: HDFS-10735-001.patch)

> Distcp using webhdfs on secure HA clusters fails with StandbyException
> --
>
> Key: HDFS-10735
> URL: https://issues.apache.org/jira/browse/HDFS-10735
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.7.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.8.0
>
> Attachments: exception.txt
>
>
> A distcp running on HADOOP 2.7.2 cluster , copying from HADOOP 2.4.1 cluster 
> fails.
> Both are secure clusters.
> The logs indicates that task contacts the StandBy name node on the remote 
> cluster is contacted first and gets a standByException. 
> Instead of retrying with the active name node on the remote cluster, the 
> execution errors out throwing InvalidToken
> The exception is attached



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Issue Comment Deleted] (HDFS-10735) Distcp using webhdfs on secure HA clusters fails with StandbyException

2016-08-08 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-10735:

Comment: was deleted

(was: | (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m  
0s{color} | {color:blue} Docker mode activated. {color} |
| {color:red}-1{color} | {color:red} patch {color} | {color:red}  0m  4s{color} 
| {color:red} HDFS-10735 does not apply to trunk. Rebase required? Wrong 
Branch? See https://wiki.apache.org/hadoop/HowToContribute for help. {color} |
\\
\\
|| Subsystem || Report/Notes ||
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12822708/HDFS-10735-001.patch |
| JIRA Issue | HDFS-10735 |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/16349/console |
| Powered by | Apache Yetus 0.4.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.

)

> Distcp using webhdfs on secure HA clusters fails with StandbyException
> --
>
> Key: HDFS-10735
> URL: https://issues.apache.org/jira/browse/HDFS-10735
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.7.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.8.0
>
> Attachments: exception.txt
>
>
> A distcp running on HADOOP 2.7.2 cluster , copying from HADOOP 2.4.1 cluster 
> fails.
> Both are secure clusters.
> The logs indicates that task contacts the StandBy name node on the remote 
> cluster is contacted first and gets a standByException. 
> Instead of retrying with the active name node on the remote cluster, the 
> execution errors out throwing InvalidToken
> The exception is attached



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-10735) Distcp using webhdfs on secure HA clusters fails with StandbyException

2016-08-08 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-10735:

Status: Patch Available  (was: Open)

> Distcp using webhdfs on secure HA clusters fails with StandbyException
> --
>
> Key: HDFS-10735
> URL: https://issues.apache.org/jira/browse/HDFS-10735
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.7.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.8.0
>
> Attachments: HDFS-10735-001.patch, exception.txt
>
>
> A distcp running on HADOOP 2.7.2 cluster , copying from HADOOP 2.4.1 cluster 
> fails.
> Both are secure clusters.
> The logs indicates that task contacts the StandBy name node on the remote 
> cluster is contacted first and gets a standByException. 
> Instead of retrying with the active name node on the remote cluster, the 
> execution errors out throwing InvalidToken
> The exception is attached



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-10735) Distcp using webhdfs on secure HA clusters fails with StandbyException

2016-08-08 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-10735:

Attachment: HDFS-10735-001.patch

> Distcp using webhdfs on secure HA clusters fails with StandbyException
> --
>
> Key: HDFS-10735
> URL: https://issues.apache.org/jira/browse/HDFS-10735
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.7.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.8.0
>
> Attachments: HDFS-10735-001.patch, exception.txt
>
>
> A distcp running on HADOOP 2.7.2 cluster , copying from HADOOP 2.4.1 cluster 
> fails.
> Both are secure clusters.
> The logs indicates that task contacts the StandBy name node on the remote 
> cluster is contacted first and gets a standByException. 
> Instead of retrying with the active name node on the remote cluster, the 
> execution errors out throwing InvalidToken
> The exception is attached



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-10735) Distcp using webhdfs on secure HA clusters fails with StandbyException

2016-08-08 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10735?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15412749#comment-15412749
 ] 

Benoy Antony commented on HDFS-10735:
-

Attaching the patch.
The patch is tested on our clusters.


> Distcp using webhdfs on secure HA clusters fails with StandbyException
> --
>
> Key: HDFS-10735
> URL: https://issues.apache.org/jira/browse/HDFS-10735
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.7.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.8.0
>
> Attachments: exception.txt
>
>
> A distcp running on HADOOP 2.7.2 cluster , copying from HADOOP 2.4.1 cluster 
> fails.
> Both are secure clusters.
> The logs indicates that task contacts the StandBy name node on the remote 
> cluster is contacted first and gets a standByException. 
> Instead of retrying with the active name node on the remote cluster, the 
> execution errors out throwing InvalidToken
> The exception is attached



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Updated] (HDFS-10735) Distcp using webhdfs on secure HA clusters fails with StandbyException

2016-08-08 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-10735:

Attachment: exception.txt

> Distcp using webhdfs on secure HA clusters fails with StandbyException
> --
>
> Key: HDFS-10735
> URL: https://issues.apache.org/jira/browse/HDFS-10735
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: webhdfs
>Affects Versions: 2.7.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.8.0
>
> Attachments: exception.txt
>
>
> A distcp running on HADOOP 2.7.2 cluster , copying from HADOOP 2.4.1 cluster 
> fails.
> Both are secure clusters.
> The logs indicates that task contacts the StandBy name node on the remote 
> cluster is contacted first and gets a standByException. 
> Instead of retrying with the active name node on the remote cluster, the 
> execution errors out throwing InvalidToken
> The exception is attached



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Created] (HDFS-10735) Distcp using webhdfs on secure HA clusters fails with StandbyException

2016-08-08 Thread Benoy Antony (JIRA)
Benoy Antony created HDFS-10735:
---

 Summary: Distcp using webhdfs on secure HA clusters fails with 
StandbyException
 Key: HDFS-10735
 URL: https://issues.apache.org/jira/browse/HDFS-10735
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: webhdfs
Affects Versions: 2.7.2
Reporter: Benoy Antony
Assignee: Benoy Antony
 Fix For: 2.8.0


A distcp running on HADOOP 2.7.2 cluster , copying from HADOOP 2.4.1 cluster 
fails.
Both are secure clusters.
The logs indicates that task contacts the StandBy name node on the remote 
cluster is contacted first and gets a standByException. 
Instead of retrying with the active name node on the remote cluster, the 
execution errors out throwing InvalidToken

The exception is attached



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org



[jira] [Commented] (HDFS-10477) Stop decommission a rack of DataNodes caused NameNode fail over to standby

2016-08-02 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10477?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15404918#comment-15404918
 ] 

Benoy Antony commented on HDFS-10477:
-

Didn't see Arpit's previous comments. If the recommended approach is set to 
dfs.namenode.fslock.fair  to false, then it may be good to sleep for a 
millisecond between lock release and re-acquisition. 

The patch looks good to me , +1

> Stop decommission a rack of DataNodes caused NameNode fail over to standby
> --
>
> Key: HDFS-10477
> URL: https://issues.apache.org/jira/browse/HDFS-10477
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.7.2
>Reporter: yunjiong zhao
>Assignee: yunjiong zhao
> Attachments: HDFS-10477.002.patch, HDFS-10477.003.patch, 
> HDFS-10477.004.patch, HDFS-10477.005.patch, HDFS-10477.patch
>
>
> In our cluster, when we stop decommissioning a rack which have 46 DataNodes, 
> it locked Namesystem for about 7 minutes as below log shows:
> {code}
> 2016-05-26 20:11:41,697 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.27:1004
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 285258 over-replicated blocks on 10.142.27.27:1004 during recommissioning
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.118:1004
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 279923 over-replicated blocks on 10.142.27.118:1004 during recommissioning
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.113:1004
> 2016-05-26 20:12:09,007 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 294307 over-replicated blocks on 10.142.27.113:1004 during recommissioning
> 2016-05-26 20:12:09,008 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.117:1004
> 2016-05-26 20:12:18,055 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 314381 over-replicated blocks on 10.142.27.117:1004 during recommissioning
> 2016-05-26 20:12:18,056 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.130:1004
> 2016-05-26 20:12:25,938 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 272779 over-replicated blocks on 10.142.27.130:1004 during recommissioning
> 2016-05-26 20:12:25,939 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.121:1004
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 287248 over-replicated blocks on 10.142.27.121:1004 during recommissioning
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.33:1004
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 299868 over-replicated blocks on 10.142.27.33:1004 during recommissioning
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.137:1004
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 303914 over-replicated blocks on 10.142.27.137:1004 during recommissioning
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.51:1004
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 281175 over-replicated blocks on 10.142.27.51:1004 during recommissioning
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.12:1004
> 2016-05-26 20:13:08,756 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 274880 over-replicated blocks on 10.142.27.12:1004 during recommissioning
> 2016-05-26 20:13:08,757 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.15:1004
> 2016-05-26 20:13:17,185 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 286334 over-replicated blocks on 10.142.27.15:1004 during recommissioning
> 2016-05-26 20:13:17,185 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.14:1004
> 2016-05-26 20:13:25,369 INFO 
> org.apache.hadoop.hdfs.server.blockma

[jira] [Commented] (HDFS-10477) Stop decommission a rack of DataNodes caused NameNode fail over to standby

2016-08-02 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10477?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15404912#comment-15404912
 ] 

Benoy Antony commented on HDFS-10477:
-

The patch looks good. 
There is no need to sleep as the admin can control the lock fairness via 
dfs.namenode.fslock.fair.  The default value is true which means that the 
earlier threads will acquire the lock.

> Stop decommission a rack of DataNodes caused NameNode fail over to standby
> --
>
> Key: HDFS-10477
> URL: https://issues.apache.org/jira/browse/HDFS-10477
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.7.2
>Reporter: yunjiong zhao
>Assignee: yunjiong zhao
> Attachments: HDFS-10477.002.patch, HDFS-10477.003.patch, 
> HDFS-10477.004.patch, HDFS-10477.005.patch, HDFS-10477.patch
>
>
> In our cluster, when we stop decommissioning a rack which have 46 DataNodes, 
> it locked Namesystem for about 7 minutes as below log shows:
> {code}
> 2016-05-26 20:11:41,697 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.27:1004
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 285258 over-replicated blocks on 10.142.27.27:1004 during recommissioning
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.118:1004
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 279923 over-replicated blocks on 10.142.27.118:1004 during recommissioning
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.113:1004
> 2016-05-26 20:12:09,007 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 294307 over-replicated blocks on 10.142.27.113:1004 during recommissioning
> 2016-05-26 20:12:09,008 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.117:1004
> 2016-05-26 20:12:18,055 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 314381 over-replicated blocks on 10.142.27.117:1004 during recommissioning
> 2016-05-26 20:12:18,056 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.130:1004
> 2016-05-26 20:12:25,938 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 272779 over-replicated blocks on 10.142.27.130:1004 during recommissioning
> 2016-05-26 20:12:25,939 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.121:1004
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 287248 over-replicated blocks on 10.142.27.121:1004 during recommissioning
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.33:1004
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 299868 over-replicated blocks on 10.142.27.33:1004 during recommissioning
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.137:1004
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 303914 over-replicated blocks on 10.142.27.137:1004 during recommissioning
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.51:1004
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 281175 over-replicated blocks on 10.142.27.51:1004 during recommissioning
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.12:1004
> 2016-05-26 20:13:08,756 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 274880 over-replicated blocks on 10.142.27.12:1004 during recommissioning
> 2016-05-26 20:13:08,757 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.15:1004
> 2016-05-26 20:13:17,185 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 286334 over-replicated blocks on 10.142.27.15:1004 during recommissioning
> 2016-05-26 20:13:17,185 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.14:1004
> 2016-05-26 20:13:25,369 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: In

[jira] [Commented] (HDFS-10477) Stop decommission a rack of DataNodes caused NameNode fail over to standby

2016-07-06 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10477?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15364835#comment-15364835
 ] 

Benoy Antony commented on HDFS-10477:
-

Thanks for the info, Akira.

> Stop decommission a rack of DataNodes caused NameNode fail over to standby
> --
>
> Key: HDFS-10477
> URL: https://issues.apache.org/jira/browse/HDFS-10477
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.7.2
>Reporter: yunjiong zhao
>Assignee: yunjiong zhao
> Attachments: HDFS-10477.002.patch, HDFS-10477.003.patch, 
> HDFS-10477.patch
>
>
> In our cluster, when we stop decommissioning a rack which have 46 DataNodes, 
> it locked Namesystem for about 7 minutes as below log shows:
> {code}
> 2016-05-26 20:11:41,697 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.27:1004
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 285258 over-replicated blocks on 10.142.27.27:1004 during recommissioning
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.118:1004
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 279923 over-replicated blocks on 10.142.27.118:1004 during recommissioning
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.113:1004
> 2016-05-26 20:12:09,007 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 294307 over-replicated blocks on 10.142.27.113:1004 during recommissioning
> 2016-05-26 20:12:09,008 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.117:1004
> 2016-05-26 20:12:18,055 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 314381 over-replicated blocks on 10.142.27.117:1004 during recommissioning
> 2016-05-26 20:12:18,056 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.130:1004
> 2016-05-26 20:12:25,938 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 272779 over-replicated blocks on 10.142.27.130:1004 during recommissioning
> 2016-05-26 20:12:25,939 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.121:1004
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 287248 over-replicated blocks on 10.142.27.121:1004 during recommissioning
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.33:1004
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 299868 over-replicated blocks on 10.142.27.33:1004 during recommissioning
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.137:1004
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 303914 over-replicated blocks on 10.142.27.137:1004 during recommissioning
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.51:1004
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 281175 over-replicated blocks on 10.142.27.51:1004 during recommissioning
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.12:1004
> 2016-05-26 20:13:08,756 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 274880 over-replicated blocks on 10.142.27.12:1004 during recommissioning
> 2016-05-26 20:13:08,757 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.15:1004
> 2016-05-26 20:13:17,185 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 286334 over-replicated blocks on 10.142.27.15:1004 during recommissioning
> 2016-05-26 20:13:17,185 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.14:1004
> 2016-05-26 20:13:25,369 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 280219 over-replicated blocks on 10.142.27.14:1004 during recommissioning
> 2016-05-26 20:13:25,370 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.28:1004
>

[jira] [Commented] (HDFS-10477) Stop decommission a rack of DataNodes caused NameNode fail over to standby

2016-07-06 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10477?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15364805#comment-15364805
 ] 

Benoy Antony commented on HDFS-10477:
-

Looks good. Could you please re-trigger the build to make sure the tests pass ?


> Stop decommission a rack of DataNodes caused NameNode fail over to standby
> --
>
> Key: HDFS-10477
> URL: https://issues.apache.org/jira/browse/HDFS-10477
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.7.2
>Reporter: yunjiong zhao
>Assignee: yunjiong zhao
> Attachments: HDFS-10477.002.patch, HDFS-10477.003.patch, 
> HDFS-10477.patch
>
>
> In our cluster, when we stop decommissioning a rack which have 46 DataNodes, 
> it locked Namesystem for about 7 minutes as below log shows:
> {code}
> 2016-05-26 20:11:41,697 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.27:1004
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 285258 over-replicated blocks on 10.142.27.27:1004 during recommissioning
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.118:1004
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 279923 over-replicated blocks on 10.142.27.118:1004 during recommissioning
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.113:1004
> 2016-05-26 20:12:09,007 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 294307 over-replicated blocks on 10.142.27.113:1004 during recommissioning
> 2016-05-26 20:12:09,008 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.117:1004
> 2016-05-26 20:12:18,055 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 314381 over-replicated blocks on 10.142.27.117:1004 during recommissioning
> 2016-05-26 20:12:18,056 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.130:1004
> 2016-05-26 20:12:25,938 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 272779 over-replicated blocks on 10.142.27.130:1004 during recommissioning
> 2016-05-26 20:12:25,939 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.121:1004
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 287248 over-replicated blocks on 10.142.27.121:1004 during recommissioning
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.33:1004
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 299868 over-replicated blocks on 10.142.27.33:1004 during recommissioning
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.137:1004
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 303914 over-replicated blocks on 10.142.27.137:1004 during recommissioning
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.51:1004
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 281175 over-replicated blocks on 10.142.27.51:1004 during recommissioning
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.12:1004
> 2016-05-26 20:13:08,756 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 274880 over-replicated blocks on 10.142.27.12:1004 during recommissioning
> 2016-05-26 20:13:08,757 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.15:1004
> 2016-05-26 20:13:17,185 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 286334 over-replicated blocks on 10.142.27.15:1004 during recommissioning
> 2016-05-26 20:13:17,185 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.14:1004
> 2016-05-26 20:13:25,369 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 280219 over-replicated blocks on 10.142.27.14:1004 during recommissioning
> 2016-05-26 20:13:25,370 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.Datanod

[jira] [Commented] (HDFS-10477) Stop decommission a rack of DataNodes caused NameNode fail over to standby

2016-06-10 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10477?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15325033#comment-15325033
 ] 

Benoy Antony commented on HDFS-10477:
-

[~kihwal],  Your comments regarding starvation makes sense. 
[~zhaoyunjiong], I think its a good idea to combine these two patches. 
 





> Stop decommission a rack of DataNodes caused NameNode fail over to standby
> --
>
> Key: HDFS-10477
> URL: https://issues.apache.org/jira/browse/HDFS-10477
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.7.2
>Reporter: yunjiong zhao
>Assignee: yunjiong zhao
> Attachments: HDFS-10477.002.patch, HDFS-10477.patch
>
>
> In our cluster, when we stop decommissioning a rack which have 46 DataNodes, 
> it locked Namesystem for about 7 minutes as below log shows:
> {code}
> 2016-05-26 20:11:41,697 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.27:1004
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 285258 over-replicated blocks on 10.142.27.27:1004 during recommissioning
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.118:1004
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 279923 over-replicated blocks on 10.142.27.118:1004 during recommissioning
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.113:1004
> 2016-05-26 20:12:09,007 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 294307 over-replicated blocks on 10.142.27.113:1004 during recommissioning
> 2016-05-26 20:12:09,008 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.117:1004
> 2016-05-26 20:12:18,055 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 314381 over-replicated blocks on 10.142.27.117:1004 during recommissioning
> 2016-05-26 20:12:18,056 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.130:1004
> 2016-05-26 20:12:25,938 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 272779 over-replicated blocks on 10.142.27.130:1004 during recommissioning
> 2016-05-26 20:12:25,939 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.121:1004
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 287248 over-replicated blocks on 10.142.27.121:1004 during recommissioning
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.33:1004
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 299868 over-replicated blocks on 10.142.27.33:1004 during recommissioning
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.137:1004
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 303914 over-replicated blocks on 10.142.27.137:1004 during recommissioning
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.51:1004
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 281175 over-replicated blocks on 10.142.27.51:1004 during recommissioning
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.12:1004
> 2016-05-26 20:13:08,756 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 274880 over-replicated blocks on 10.142.27.12:1004 during recommissioning
> 2016-05-26 20:13:08,757 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.15:1004
> 2016-05-26 20:13:17,185 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 286334 over-replicated blocks on 10.142.27.15:1004 during recommissioning
> 2016-05-26 20:13:17,185 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.14:1004
> 2016-05-26 20:13:25,369 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 280219 over-replicated blocks on 10.142.27.14:1004 during recommissioning
> 2016-05-26 20:13:25,370 INFO 
> org.apache.hadoop.hd

[jira] [Commented] (HDFS-10477) Stop decommission a rack of DataNodes caused NameNode fail over to standby

2016-06-02 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10477?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15313282#comment-15313282
 ] 

Benoy Antony commented on HDFS-10477:
-

If its possible to release lock per storage, then that's better. 
If not , I prefer the first version which does releases the lock per each 
datanode without the additional processing. 
The logs show that  the each node is processed in around 10 seconds. 

> Stop decommission a rack of DataNodes caused NameNode fail over to standby
> --
>
> Key: HDFS-10477
> URL: https://issues.apache.org/jira/browse/HDFS-10477
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.7.2
>Reporter: yunjiong zhao
>Assignee: yunjiong zhao
> Attachments: HDFS-10477.002.patch, HDFS-10477.patch
>
>
> In our cluster, when we stop decommissioning a rack which have 46 DataNodes, 
> it locked Namesystem for about 7 minutes as below log shows:
> {code}
> 2016-05-26 20:11:41,697 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.27:1004
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 285258 over-replicated blocks on 10.142.27.27:1004 during recommissioning
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.118:1004
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 279923 over-replicated blocks on 10.142.27.118:1004 during recommissioning
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.113:1004
> 2016-05-26 20:12:09,007 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 294307 over-replicated blocks on 10.142.27.113:1004 during recommissioning
> 2016-05-26 20:12:09,008 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.117:1004
> 2016-05-26 20:12:18,055 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 314381 over-replicated blocks on 10.142.27.117:1004 during recommissioning
> 2016-05-26 20:12:18,056 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.130:1004
> 2016-05-26 20:12:25,938 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 272779 over-replicated blocks on 10.142.27.130:1004 during recommissioning
> 2016-05-26 20:12:25,939 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.121:1004
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 287248 over-replicated blocks on 10.142.27.121:1004 during recommissioning
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.33:1004
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 299868 over-replicated blocks on 10.142.27.33:1004 during recommissioning
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.137:1004
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 303914 over-replicated blocks on 10.142.27.137:1004 during recommissioning
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.51:1004
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 281175 over-replicated blocks on 10.142.27.51:1004 during recommissioning
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.12:1004
> 2016-05-26 20:13:08,756 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 274880 over-replicated blocks on 10.142.27.12:1004 during recommissioning
> 2016-05-26 20:13:08,757 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.15:1004
> 2016-05-26 20:13:17,185 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 286334 over-replicated blocks on 10.142.27.15:1004 during recommissioning
> 2016-05-26 20:13:17,185 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.14:1004
> 2016-05-26 20:13:25,369 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 280219 over

[jira] [Comment Edited] (HDFS-10477) Stop decommission a rack of DataNodes caused NameNode fail over to standby

2016-06-02 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-10477?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15313282#comment-15313282
 ] 

Benoy Antony edited comment on HDFS-10477 at 6/2/16 11:21 PM:
--

If its possible to release lock per storage, then that's better. 
If not , I prefer the first version which releases the lock per each datanode 
without the additional processing. 
The logs show that  the each node is processed in around 10 seconds. 


was (Author: benoyantony):
If its possible to release lock per storage, then that's better. 
If not , I prefer the first version which does releases the lock per each 
datanode without the additional processing. 
The logs show that  the each node is processed in around 10 seconds. 

> Stop decommission a rack of DataNodes caused NameNode fail over to standby
> --
>
> Key: HDFS-10477
> URL: https://issues.apache.org/jira/browse/HDFS-10477
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.7.2
>Reporter: yunjiong zhao
>Assignee: yunjiong zhao
> Attachments: HDFS-10477.002.patch, HDFS-10477.patch
>
>
> In our cluster, when we stop decommissioning a rack which have 46 DataNodes, 
> it locked Namesystem for about 7 minutes as below log shows:
> {code}
> 2016-05-26 20:11:41,697 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.27:1004
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 285258 over-replicated blocks on 10.142.27.27:1004 during recommissioning
> 2016-05-26 20:11:51,171 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.118:1004
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 279923 over-replicated blocks on 10.142.27.118:1004 during recommissioning
> 2016-05-26 20:11:59,972 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.113:1004
> 2016-05-26 20:12:09,007 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 294307 over-replicated blocks on 10.142.27.113:1004 during recommissioning
> 2016-05-26 20:12:09,008 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.117:1004
> 2016-05-26 20:12:18,055 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 314381 over-replicated blocks on 10.142.27.117:1004 during recommissioning
> 2016-05-26 20:12:18,056 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.130:1004
> 2016-05-26 20:12:25,938 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 272779 over-replicated blocks on 10.142.27.130:1004 during recommissioning
> 2016-05-26 20:12:25,939 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.121:1004
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 287248 over-replicated blocks on 10.142.27.121:1004 during recommissioning
> 2016-05-26 20:12:34,134 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.33:1004
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 299868 over-replicated blocks on 10.142.27.33:1004 during recommissioning
> 2016-05-26 20:12:43,020 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.137:1004
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 303914 over-replicated blocks on 10.142.27.137:1004 during recommissioning
> 2016-05-26 20:12:52,220 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.51:1004
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 281175 over-replicated blocks on 10.142.27.51:1004 during recommissioning
> 2016-05-26 20:13:00,362 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.12:1004
> 2016-05-26 20:13:08,756 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 274880 over-replicated blocks on 10.142.27.12:1004 during recommissioning
> 2016-05-26 20:13:08,757 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Stop 
> Decommissioning 10.142.27.15:1004
> 2016-05-26 20:13:17,185 INFO 
> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Invalidated 
> 2863

[jira] [Commented] (HDFS-9597) BaseReplicationPolicyTest should update data node stats after adding a data node

2016-01-14 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9597?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15099063#comment-15099063
 ] 

Benoy Antony commented on HDFS-9597:


committed to branch-2.8 also.

> BaseReplicationPolicyTest should update data node stats after adding a data 
> node
> 
>
> Key: HDFS-9597
> URL: https://issues.apache.org/jira/browse/HDFS-9597
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 3.0.0, 2.8.0
>Reporter: Wei-Chiu Chuang
>Assignee: Wei-Chiu Chuang
>Priority: Blocker
> Fix For: 2.8.0
>
> Attachments: HDFS-9597.001.patch
>
>
> Looks like HDFS-9034 broke 
> TestReplicationPolicyConsiderLoad#testChooseTargetWithDecomNodes.
> This test has been failing since yesterday.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9034) "StorageTypeStats" Metric should not count failed storage.

2016-01-14 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9034?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15099064#comment-15099064
 ] 

Benoy Antony commented on HDFS-9034:


committed to branch-2.8.

> "StorageTypeStats" Metric should not count failed storage.
> --
>
> Key: HDFS-9034
> URL: https://issues.apache.org/jira/browse/HDFS-9034
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Reporter: Archana T
>Assignee: Surendra Singh Lilhore
> Attachments: HDFS-9034.01.patch, HDFS-9034.02.patch, 
> HDFS-9034.03.patch, HDFS-9034.04.patch, dfsStorage_NN_UI2.png
>
>
> When we remove one storage type from all the DNs, still NN UI shows entry of 
> those storage type --
> Ex:for ARCHIVE
> Steps--
> 1. ARCHIVE Storage type was added for all DNs
> 2. Stop DNs
> 3. Removed ARCHIVE Storages from all DNs
> 4. Restarted DNs
> NN UI shows below --
> DFS Storage Types
> Storage Type Configured Capacity Capacity Used Capacity Remaining 
> ARCHIVE   57.18 GB64 KB (0%)  39.82 GB (69.64%)   64 KB   
> 1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9597) BaseReplicationPolicyTest should update data node stats after adding a data node

2015-12-25 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9597?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-9597:
---
   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 2.8.0
   Status: Resolved  (was: Patch Available)

Committed in trunk and branch-2. Thanks [~jojochuang]  for the fix and 
[~yangzhe1991]  for the review. 



> BaseReplicationPolicyTest should update data node stats after adding a data 
> node
> 
>
> Key: HDFS-9597
> URL: https://issues.apache.org/jira/browse/HDFS-9597
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 3.0.0, 2.8.0
>Reporter: Wei-Chiu Chuang
>Assignee: Wei-Chiu Chuang
> Fix For: 2.8.0
>
> Attachments: HDFS-9597.001.patch
>
>
> Looks like HDFS-9034 broke 
> TestReplicationPolicyConsiderLoad#testChooseTargetWithDecomNodes.
> This test has been failing since yesterday.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9597) BaseReplicationPolicyTest should update data node stats after adding a data node

2015-12-25 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9597?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-9597:
---
Priority: Blocker  (was: Major)

> BaseReplicationPolicyTest should update data node stats after adding a data 
> node
> 
>
> Key: HDFS-9597
> URL: https://issues.apache.org/jira/browse/HDFS-9597
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 3.0.0, 2.8.0
>Reporter: Wei-Chiu Chuang
>Assignee: Wei-Chiu Chuang
>Priority: Blocker
> Fix For: 2.8.0
>
> Attachments: HDFS-9597.001.patch
>
>
> Looks like HDFS-9034 broke 
> TestReplicationPolicyConsiderLoad#testChooseTargetWithDecomNodes.
> This test has been failing since yesterday.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9597) BaseReplicationPolicyTest should update data node stats after adding a data node

2015-12-25 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9597?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15071664#comment-15071664
 ] 

Benoy Antony commented on HDFS-9597:


+1. The fix looks good to me. Thanks for taking care of it.
Will commit the patch after the tests. 

> BaseReplicationPolicyTest should update data node stats after adding a data 
> node
> 
>
> Key: HDFS-9597
> URL: https://issues.apache.org/jira/browse/HDFS-9597
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: datanode
>Affects Versions: 3.0.0, 2.8.0
>Reporter: Wei-Chiu Chuang
>Assignee: Wei-Chiu Chuang
> Attachments: HDFS-9597.001.patch
>
>
> Looks like HDFS-9034 broke 
> TestReplicationPolicyConsiderLoad#testChooseTargetWithDecomNodes.
> This test has been failing since yesterday.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-9034) "StorageTypeStats" Metric should not count failed storage.

2015-12-22 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9034?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-9034:
---
  Resolution: Fixed
Hadoop Flags: Reviewed
Target Version/s: 2.8.0
  Status: Resolved  (was: Patch Available)

Committed in trunk and branch-2. 
Thanks [~surendrasingh] for the contribution. 
Thanks [~archanat] for reporting the issue.

> "StorageTypeStats" Metric should not count failed storage.
> --
>
> Key: HDFS-9034
> URL: https://issues.apache.org/jira/browse/HDFS-9034
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Reporter: Archana T
>Assignee: Surendra Singh Lilhore
> Attachments: HDFS-9034.01.patch, HDFS-9034.02.patch, 
> HDFS-9034.03.patch, HDFS-9034.04.patch, dfsStorage_NN_UI2.png
>
>
> When we remove one storage type from all the DNs, still NN UI shows entry of 
> those storage type --
> Ex:for ARCHIVE
> Steps--
> 1. ARCHIVE Storage type was added for all DNs
> 2. Stop DNs
> 3. Removed ARCHIVE Storages from all DNs
> 4. Restarted DNs
> NN UI shows below --
> DFS Storage Types
> Storage Type Configured Capacity Capacity Used Capacity Remaining 
> ARCHIVE   57.18 GB64 KB (0%)  39.82 GB (69.64%)   64 KB   
> 1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9034) "StorageTypeStats" Metric should not count failed storage.

2015-12-21 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9034?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15066868#comment-15066868
 ] 

Benoy Antony commented on HDFS-9034:


Thanks for the clarification, Surendra. The patch looks good . 
+1.
If there are no other comments, I'll commit this, tomorrow.



> "StorageTypeStats" Metric should not count failed storage.
> --
>
> Key: HDFS-9034
> URL: https://issues.apache.org/jira/browse/HDFS-9034
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Reporter: Archana T
>Assignee: Surendra Singh Lilhore
> Attachments: HDFS-9034.01.patch, HDFS-9034.02.patch, 
> HDFS-9034.03.patch, HDFS-9034.04.patch, dfsStorage_NN_UI2.png
>
>
> When we remove one storage type from all the DNs, still NN UI shows entry of 
> those storage type --
> Ex:for ARCHIVE
> Steps--
> 1. ARCHIVE Storage type was added for all DNs
> 2. Stop DNs
> 3. Removed ARCHIVE Storages from all DNs
> 4. Restarted DNs
> NN UI shows below --
> DFS Storage Types
> Storage Type Configured Capacity Capacity Used Capacity Remaining 
> ARCHIVE   57.18 GB64 KB (0%)  39.82 GB (69.64%)   64 KB   
> 1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9034) "StorageTypeStats" Metric should not count failed storage.

2015-12-18 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9034?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15064719#comment-15064719
 ] 

Benoy Antony commented on HDFS-9034:


Let say, node registers with  10 Storages of 2 TB. Assume none of these failed. 
 Assume 9 are DISK and 1 is ARCHIVE
stats.add(d) is called via updateDnStat .
StorageTypes will be recorded as DISK and ARCHIVE.

For the next hearBeat , the single ARCHIVE is reported as FAILED. 
stats.subtract(node) will be called, but it will not remove the FAILED storage.
StorageTypes will remain as DISK and ARCHIVE.

stats.add(node) will be called, but will skip the FAILED storage.
StorageTypes will remain as DISK and ARCHIVE.

Wouldn't this happen ?

> "StorageTypeStats" Metric should not count failed storage.
> --
>
> Key: HDFS-9034
> URL: https://issues.apache.org/jira/browse/HDFS-9034
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Reporter: Archana T
>Assignee: Surendra Singh Lilhore
> Attachments: HDFS-9034.01.patch, HDFS-9034.02.patch, 
> HDFS-9034.03.patch, HDFS-9034.04.patch, dfsStorage_NN_UI2.png
>
>
> When we remove one storage type from all the DNs, still NN UI shows entry of 
> those storage type --
> Ex:for ARCHIVE
> Steps--
> 1. ARCHIVE Storage type was added for all DNs
> 2. Stop DNs
> 3. Removed ARCHIVE Storages from all DNs
> 4. Restarted DNs
> NN UI shows below --
> DFS Storage Types
> Storage Type Configured Capacity Capacity Used Capacity Remaining 
> ARCHIVE   57.18 GB64 KB (0%)  39.82 GB (69.64%)   64 KB   
> 1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9034) "StorageTypeStats" Metric should not count failed storage.

2015-12-17 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9034?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15063140#comment-15063140
 ] 

Benoy Antony commented on HDFS-9034:


Reviewed the patch. Sorry for the delay.
Patch looks good. I have following comment:

1. In DatanodeStats.substract,  we subtract the storage capacity only if it is 
NOT FAILED. But suppose it was good when we added it, then not subtracting will 
be a mistake. But if it was bad during add, we should not subtract it. So we 
need some additional information to correctly determine whether we should 
subtract it or not. 
We also need test cases to make sure that these cases are handled correctly.


> "StorageTypeStats" Metric should not count failed storage.
> --
>
> Key: HDFS-9034
> URL: https://issues.apache.org/jira/browse/HDFS-9034
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Reporter: Archana T
>Assignee: Surendra Singh Lilhore
> Attachments: HDFS-9034.01.patch, HDFS-9034.02.patch, 
> HDFS-9034.03.patch, HDFS-9034.04.patch, dfsStorage_NN_UI2.png
>
>
> When we remove one storage type from all the DNs, still NN UI shows entry of 
> those storage type --
> Ex:for ARCHIVE
> Steps--
> 1. ARCHIVE Storage type was added for all DNs
> 2. Stop DNs
> 3. Removed ARCHIVE Storages from all DNs
> 4. Restarted DNs
> NN UI shows below --
> DFS Storage Types
> Storage Type Configured Capacity Capacity Used Capacity Remaining 
> ARCHIVE   57.18 GB64 KB (0%)  39.82 GB (69.64%)   64 KB   
> 1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9034) "StorageTypeStats" Metric should not count failed storage.

2015-12-09 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9034?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15049781#comment-15049781
 ] 

Benoy Antony commented on HDFS-9034:


I'll review this tomorrow.

> "StorageTypeStats" Metric should not count failed storage.
> --
>
> Key: HDFS-9034
> URL: https://issues.apache.org/jira/browse/HDFS-9034
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Reporter: Archana T
>Assignee: Surendra Singh Lilhore
> Attachments: HDFS-9034.01.patch, HDFS-9034.02.patch, 
> HDFS-9034.03.patch, HDFS-9034.04.patch, dfsStorage_NN_UI2.png
>
>
> When we remove one storage type from all the DNs, still NN UI shows entry of 
> those storage type --
> Ex:for ARCHIVE
> Steps--
> 1. ARCHIVE Storage type was added for all DNs
> 2. Stop DNs
> 3. Removed ARCHIVE Storages from all DNs
> 4. Restarted DNs
> NN UI shows below --
> DFS Storage Types
> Storage Type Configured Capacity Capacity Used Capacity Remaining 
> ARCHIVE   57.18 GB64 KB (0%)  39.82 GB (69.64%)   64 KB   
> 1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-6363) Improve concurrency while checking inclusion and exclusion of datanodes

2015-11-30 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-6363:
---
Attachment: HDFS-6363-004.patch

> Improve concurrency while checking inclusion and exclusion of datanodes
> ---
>
> Key: HDFS-6363
> URL: https://issues.apache.org/jira/browse/HDFS-6363
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Attachments: HDFS-6363-002.patch, HDFS-6363-003.patch, 
> HDFS-6363-004.patch, HDFS-6363.patch
>
>
> HostFileManager holds two effectively immutable objects - includes and 
> excludes. These two objects can be safely published together using a volatile 
> container instead of synchronizing for all mutators and accessors.
> This improves the concurrency while using HostFileManager.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-6363) Improve concurrency while checking inclusion and exclusion of datanodes

2015-11-30 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-6363:
---
Attachment: HDFS-6363-003.patch

Attaching the patch after fixing checkstyle issues. 

> Improve concurrency while checking inclusion and exclusion of datanodes
> ---
>
> Key: HDFS-6363
> URL: https://issues.apache.org/jira/browse/HDFS-6363
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Attachments: HDFS-6363-002.patch, HDFS-6363-003.patch, HDFS-6363.patch
>
>
> HostFileManager holds two effectively immutable objects - includes and 
> excludes. These two objects can be safely published together using a volatile 
> container instead of synchronizing for all mutators and accessors.
> This improves the concurrency while using HostFileManager.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-6363) Improve concurrency while checking inclusion and exclusion of datanodes

2015-11-25 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-6363:
---
Attachment: HDFS-6363-002.patch

Thanks for the review , [~wheat9].
Attaching a patch using _AtomicReference_.


> Improve concurrency while checking inclusion and exclusion of datanodes
> ---
>
> Key: HDFS-6363
> URL: https://issues.apache.org/jira/browse/HDFS-6363
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Attachments: HDFS-6363-002.patch, HDFS-6363.patch
>
>
> HostFileManager holds two effectively immutable objects - includes and 
> excludes. These two objects can be safely published together using a volatile 
> container instead of synchronizing for all mutators and accessors.
> This improves the concurrency while using HostFileManager.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9034) "StorageTypeStats" Metric should not count failed storage.

2015-11-25 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9034?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15027483#comment-15027483
 ] 

Benoy Antony commented on HDFS-9034:


Patch generally looks good. I cannot apply it to the current code base. Could 
you please rebase the patch ?

> "StorageTypeStats" Metric should not count failed storage.
> --
>
> Key: HDFS-9034
> URL: https://issues.apache.org/jira/browse/HDFS-9034
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Reporter: Archana T
>Assignee: Surendra Singh Lilhore
> Attachments: HDFS-9034.01.patch, HDFS-9034.02.patch, 
> dfsStorage_NN_UI2.png
>
>
> When we remove one storage type from all the DNs, still NN UI shows entry of 
> those storage type --
> Ex:for ARCHIVE
> Steps--
> 1. ARCHIVE Storage type was added for all DNs
> 2. Stop DNs
> 3. Removed ARCHIVE Storages from all DNs
> 4. Restarted DNs
> NN UI shows below --
> DFS Storage Types
> Storage Type Configured Capacity Capacity Used Capacity Remaining 
> ARCHIVE   57.18 GB64 KB (0%)  39.82 GB (69.64%)   64 KB   
> 1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-9034) "StorageTypeStats" Metric should not count failed storage.

2015-11-24 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-9034?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15026044#comment-15026044
 ] 

Benoy Antony commented on HDFS-9034:


Sure, I will review tomorrow.

> "StorageTypeStats" Metric should not count failed storage.
> --
>
> Key: HDFS-9034
> URL: https://issues.apache.org/jira/browse/HDFS-9034
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Reporter: Archana T
>Assignee: Surendra Singh Lilhore
> Attachments: HDFS-9034.01.patch, HDFS-9034.02.patch, 
> dfsStorage_NN_UI2.png
>
>
> When we remove one storage type from all the DNs, still NN UI shows entry of 
> those storage type --
> Ex:for ARCHIVE
> Steps--
> 1. ARCHIVE Storage type was added for all DNs
> 2. Stop DNs
> 3. Removed ARCHIVE Storages from all DNs
> 4. Restarted DNs
> NN UI shows below --
> DFS Storage Types
> Storage Type Configured Capacity Capacity Used Capacity Remaining 
> ARCHIVE   57.18 GB64 KB (0%)  39.82 GB (69.64%)   64 KB   
> 1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-6407) new namenode UI, lost ability to sort columns in datanode tab

2015-08-17 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6407?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14699893#comment-14699893
 ] 

Benoy Antony commented on HDFS-6407:


It will be good to specify the version information of the datatables component. 
This will help in maintaining this functionality.
For other Js components, the version information is included in the file name. 

> new namenode UI, lost ability to sort columns in datanode tab
> -
>
> Key: HDFS-6407
> URL: https://issues.apache.org/jira/browse/HDFS-6407
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Nathan Roberts
>Assignee: Haohui Mai
>Priority: Critical
>  Labels: BB2015-05-TBR
> Attachments: 002-datanodes-sorted-capacityUsed.png, 
> 002-datanodes.png, 002-filebrowser.png, 002-snapshots.png, 
> HDFS-6407-002.patch, HDFS-6407-003.patch, HDFS-6407.008.patch, 
> HDFS-6407.009.patch, HDFS-6407.010.patch, HDFS-6407.011.patch, 
> HDFS-6407.4.patch, HDFS-6407.5.patch, HDFS-6407.6.patch, HDFS-6407.7.patch, 
> HDFS-6407.patch, browse_directory.png, datanodes.png, snapshots.png, sorting 
> 2.png, sorting table.png
>
>
> old ui supported clicking on column header to sort on that column. The new ui 
> seems to have dropped this very useful feature.
> There are a few tables in the Namenode UI to display  datanodes information, 
> directory listings and snapshots.
> When there are many items in the tables, it is useful to have ability to sort 
> on the different columns.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7483) Display information per tier on the Namenode UI

2015-07-20 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14634243#comment-14634243
 ] 

Benoy Antony commented on HDFS-7483:


Nice. Thanks for the snippet. I like this approach. 
+1 pending jenkins. 


> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7483-001.patch, HDFS-7483-002.patch, 
> HDFS-7483.003.patch, overview.png, storagetypes.png, 
> storagetypes_withnostorage.png, withOneStorageType.png, withTwoStorageType.png
>
>
> If cluster has different types of storage, it is useful to display the 
> storage information per type. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7483) Display information per tier on the Namenode UI

2015-07-16 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14630563#comment-14630563
 ] 

Benoy Antony commented on HDFS-7483:


In the existing use cases of fmt_percentage, the percentage is already 
generated on the server side. I do not understand the approach that you are 
suggesting. Could you please provide me with a working snippet  ?


> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7483-001.patch, HDFS-7483-002.patch, overview.png, 
> storagetypes.png, storagetypes_withnostorage.png, withOneStorageType.png, 
> withTwoStorageType.png
>
>
> If cluster has different types of storage, it is useful to display the 
> storage information per type. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7483) Display information per tier on the Namenode UI

2015-07-15 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14629045#comment-14629045
 ] 

Benoy Antony commented on HDFS-7483:


[~wheat9], As I mentioned, there is no good way of displaying percentage using 
math helper and fmt_percentage filter. If you have no further comments, I'll 
commit this patch by end of day tomorrow.

> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7483-001.patch, HDFS-7483-002.patch, overview.png, 
> storagetypes.png, storagetypes_withnostorage.png, withOneStorageType.png, 
> withTwoStorageType.png
>
>
> If cluster has different types of storage, it is useful to display the 
> storage information per type. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7483) Display information per tier on the Namenode UI

2015-07-15 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14628406#comment-14628406
 ] 

Benoy Antony commented on HDFS-7483:


To add on, I had tried the approach while working on the patch. _math_ is a 
helper whereas fmt_percentage is a filter. We cannot do something like   
"helper | filter" . Some helpers support a "filters" attribute. But math helper 
does not support "filters" attribute. So I could not reuse math helper and 
fmt_percentage filter. That's why I wrote a a new percentage helper. 


> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7483-001.patch, HDFS-7483-002.patch, overview.png, 
> storagetypes.png, storagetypes_withnostorage.png, withOneStorageType.png, 
> withTwoStorageType.png
>
>
> If cluster has different types of storage, it is useful to display the 
> storage information per type. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7483) Display information per tier on the Namenode UI

2015-07-14 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14627342#comment-14627342
 ] 

Benoy Antony commented on HDFS-7483:


I don't think the suggested approach will work. But I could be wrong. 
[~wheat9], Could you please provide a working code snippet for this approach ?

> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7483-001.patch, HDFS-7483-002.patch, overview.png, 
> storagetypes.png, storagetypes_withnostorage.png, withOneStorageType.png, 
> withTwoStorageType.png
>
>
> If cluster has different types of storage, it is useful to display the 
> storage information per type. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7483) Display information per tier on the Namenode UI

2015-07-13 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7483:
---
Attachment: HDFS-7483-002.patch

> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7483-001.patch, HDFS-7483-002.patch, overview.png, 
> storagetypes.png, storagetypes_withnostorage.png, withOneStorageType.png, 
> withTwoStorageType.png
>
>
> If cluster has different types of storage, it is useful to display the 
> storage information per type. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7483) Display information per tier on the Namenode UI

2015-07-13 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7483:
---
Attachment: (was: HDFS-7483-002.patch)

> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7483-001.patch, HDFS-7483-002.patch, overview.png, 
> storagetypes.png, storagetypes_withnostorage.png, withOneStorageType.png, 
> withTwoStorageType.png
>
>
> If cluster has different types of storage, it is useful to display the 
> storage information per type. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7483) Display information per tier on the Namenode UI

2015-07-11 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7483:
---
Status: Patch Available  (was: In Progress)

> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7483-001.patch, HDFS-7483-002.patch, overview.png, 
> storagetypes.png, storagetypes_withnostorage.png, withOneStorageType.png, 
> withTwoStorageType.png
>
>
> If cluster has different types of storage, it is useful to display the 
> storage information per type. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7483) Display information per tier on the Namenode UI

2015-07-10 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7483:
---
Attachment: withTwoStorageType.png
withOneStorageType.png
HDFS-7483-002.patch

Thanks for the review and comments , [~wheat9] and [~arpitagarwal].

Updated the patch based on the comments.

# Displays Storage type information on the overview screen. The table is 
displayed only if there are more than one storage types. Following screenshots 
are attached:
#* Overview page when there is only one storage type.
#* Overview page when there are two storage types.
# Added dust.js helper to calculate percentage on the client side. 
# Removed an unnecessary  _System.out.println_ added by HDFS-7390.



> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7483-001.patch, HDFS-7483-002.patch, overview.png, 
> storagetypes.png, storagetypes_withnostorage.png, withOneStorageType.png, 
> withTwoStorageType.png
>
>
> If cluster has different types of storage, it is useful to display the 
> storage information per type. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7483) Display information per tier on the Namenode UI

2015-07-01 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7483:
---
Attachment: (was: HDFS-7483-001.patch)

> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7483-001.patch, overview.png, storagetypes.png, 
> storagetypes_withnostorage.png
>
>
> If cluster has different types of storage, it is useful to display the 
> storage information per type. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7483) Display information per tier on the Namenode UI

2015-07-01 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7483:
---
Attachment: storagetypes.png
storagetypes_withnostorage.png
HDFS-7483-001.patch
overview.png

> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7483-001.patch, overview.png, storagetypes.png, 
> storagetypes_withnostorage.png
>
>
> If cluster has different types of storage, it is useful to display the 
> storage information per type. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7483) Display information per tier on the Namenode UI

2015-07-01 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7483:
---
Attachment: HDFS-7483-001.patch

Attaching a patch which displays storage stats by storage type on the name node 
ui. The screen shots are attached.
The patch has the following changes.

1. The jmx for Blocktstats  is enhanced by adding PercentRemaining and 
PercentUsed. 
2. Refactored header part of name node ui to header.html
3. Added a page for storage types as shown in the screen shots and tested on a 
local cluster.


> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7483-001.patch
>
>
> If cluster has different types of storage, it is useful to display the 
> storage information per type. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7483) Display information per tier on the Namenode UI

2015-06-29 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7483:
---
Description: 
If cluster has different types of storage, it is useful to display the storage 
information per type. 

The information will be available via JMX (HDFS-7390)

  was:
If cluster has different types of storage, it is useful to display the storage 
information per tier. 

The information will be available via JMX (HDFS-7390)


> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>
> If cluster has different types of storage, it is useful to display the 
> storage information per type. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Work started] (HDFS-7483) Display information per tier on the Namenode UI

2015-06-29 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HDFS-7483 started by Benoy Antony.
--
> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>
> If cluster has different tiers of storage, it is useful to display the 
> storage information per tier. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7483) Display information per tier on the Namenode UI

2015-06-29 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7483:
---
Description: 
If cluster has different types of storage, it is useful to display the storage 
information per tier. 

The information will be available via JMX (HDFS-7390)

  was:
If cluster has different tiers of storage, it is useful to display the storage 
information per tier. 

The information will be available via JMX (HDFS-7390)


> Display information per tier on the Namenode UI
> ---
>
> Key: HDFS-7483
> URL: https://issues.apache.org/jira/browse/HDFS-7483
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>
> If cluster has different types of storage, it is useful to display the 
> storage information per tier. 
> The information will be available via JMX (HDFS-7390)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7390) Provide JMX metrics per storage type

2015-06-29 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7390:
---
   Resolution: Fixed
Fix Version/s: 2.8.0
   Status: Resolved  (was: Patch Available)

> Provide JMX metrics per storage type
> 
>
> Key: HDFS-7390
> URL: https://issues.apache.org/jira/browse/HDFS-7390
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.5.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Fix For: 2.8.0
>
> Attachments: HDFS-7390-003.patch, HDFS-7390-004.patch, 
> HDFS-7390-005.patch, HDFS-7390-006.patch, HDFS-7390-007.patch, 
> HDFS-7390-008.patch, HDFS-7390.patch, HDFS-7390.patch
>
>
> HDFS-2832  added heterogeneous support. In a cluster with different storage 
> types, it is useful to have metrics per storage type. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7390) Provide JMX metrics per storage type

2015-06-29 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7390?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14606020#comment-14606020
 ] 

Benoy Antony commented on HDFS-7390:


committed to trunk and branch-2. 

> Provide JMX metrics per storage type
> 
>
> Key: HDFS-7390
> URL: https://issues.apache.org/jira/browse/HDFS-7390
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.5.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Fix For: 2.8.0
>
> Attachments: HDFS-7390-003.patch, HDFS-7390-004.patch, 
> HDFS-7390-005.patch, HDFS-7390-006.patch, HDFS-7390-007.patch, 
> HDFS-7390-008.patch, HDFS-7390.patch, HDFS-7390.patch
>
>
> HDFS-2832  added heterogeneous support. In a cluster with different storage 
> types, it is useful to have metrics per storage type. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7390) Provide JMX metrics per storage type

2015-06-29 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7390?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14605987#comment-14605987
 ] 

Benoy Antony commented on HDFS-7390:


The test failure on _ TestWebHDFS_ is unrelated. Thanks for the review, 
[~arpitagarwal]. I'll commit this shortly.

> Provide JMX metrics per storage type
> 
>
> Key: HDFS-7390
> URL: https://issues.apache.org/jira/browse/HDFS-7390
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.5.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Attachments: HDFS-7390-003.patch, HDFS-7390-004.patch, 
> HDFS-7390-005.patch, HDFS-7390-006.patch, HDFS-7390-007.patch, 
> HDFS-7390-008.patch, HDFS-7390.patch, HDFS-7390.patch
>
>
> HDFS-2832  added heterogeneous support. In a cluster with different storage 
> types, it is useful to have metrics per storage type. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7390) Provide JMX metrics per storage type

2015-06-26 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7390:
---
Attachment: HDFS-7390-008.patch

Attaching the patch which addresses [~arpitagarwal]'s comments.

> Provide JMX metrics per storage type
> 
>
> Key: HDFS-7390
> URL: https://issues.apache.org/jira/browse/HDFS-7390
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.5.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Attachments: HDFS-7390-003.patch, HDFS-7390-004.patch, 
> HDFS-7390-005.patch, HDFS-7390-006.patch, HDFS-7390-007.patch, 
> HDFS-7390-008.patch, HDFS-7390.patch, HDFS-7390.patch
>
>
> HDFS-2832  added heterogeneous support. In a cluster with different storage 
> types, it is useful to have metrics per storage type. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7390) Provide JMX metrics per storage type

2015-06-25 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7390?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14601772#comment-14601772
 ] 

Benoy Antony commented on HDFS-7390:


The checkstyle issue is due to the length of _BlockManager_  which was already 
above 2000 lines. Not some that was introduced by the patch

> Provide JMX metrics per storage type
> 
>
> Key: HDFS-7390
> URL: https://issues.apache.org/jira/browse/HDFS-7390
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.5.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Attachments: HDFS-7390-003.patch, HDFS-7390-004.patch, 
> HDFS-7390-005.patch, HDFS-7390-006.patch, HDFS-7390-007.patch, 
> HDFS-7390.patch, HDFS-7390.patch
>
>
> HDFS-2832  added heterogeneous support. In a cluster with different storage 
> types, it is useful to have metrics per storage type. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7390) Provide JMX metrics per storage type

2015-06-25 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7390:
---
Attachment: HDFS-7390-007.patch

Attaching a patch after fixing the compilation issues.

> Provide JMX metrics per storage type
> 
>
> Key: HDFS-7390
> URL: https://issues.apache.org/jira/browse/HDFS-7390
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.5.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Attachments: HDFS-7390-003.patch, HDFS-7390-004.patch, 
> HDFS-7390-005.patch, HDFS-7390-006.patch, HDFS-7390-007.patch, 
> HDFS-7390.patch, HDFS-7390.patch
>
>
> HDFS-2832  added heterogeneous support. In a cluster with different storage 
> types, it is useful to have metrics per storage type. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7390) Provide JMX metrics per storage type

2015-06-24 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7390:
---
Attachment: HDFS-7390-006.patch

Thanks for the review [~arpitagarwal]
Attaching a patch which addresses the comments.

> Provide JMX metrics per storage type
> 
>
> Key: HDFS-7390
> URL: https://issues.apache.org/jira/browse/HDFS-7390
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.5.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Attachments: HDFS-7390-003.patch, HDFS-7390-004.patch, 
> HDFS-7390-005.patch, HDFS-7390-006.patch, HDFS-7390.patch, HDFS-7390.patch
>
>
> HDFS-2832  added heterogeneous support. In a cluster with different storage 
> types, it is useful to have metrics per storage type. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7390) Provide JMX metrics per storage type

2015-06-22 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7390:
---
Attachment: (was: HDFS-7390-005.patch)

> Provide JMX metrics per storage type
> 
>
> Key: HDFS-7390
> URL: https://issues.apache.org/jira/browse/HDFS-7390
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.5.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Attachments: HDFS-7390-003.patch, HDFS-7390-004.patch, 
> HDFS-7390-005.patch, HDFS-7390.patch, HDFS-7390.patch
>
>
> HDFS-2832  added heterogeneous support. In a cluster with different storage 
> types, it is useful to have metrics per storage type. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7390) Provide JMX metrics per storage type

2015-06-22 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7390:
---
Attachment: HDFS-7390-005.patch

> Provide JMX metrics per storage type
> 
>
> Key: HDFS-7390
> URL: https://issues.apache.org/jira/browse/HDFS-7390
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.5.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Attachments: HDFS-7390-003.patch, HDFS-7390-004.patch, 
> HDFS-7390-005.patch, HDFS-7390-005.patch, HDFS-7390.patch, HDFS-7390.patch
>
>
> HDFS-2832  added heterogeneous support. In a cluster with different storage 
> types, it is useful to have metrics per storage type. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7390) Provide JMX metrics per storage type

2015-06-18 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7390:
---
Attachment: HDFS-7390-005.patch

fixing checkstyle warning

> Provide JMX metrics per storage type
> 
>
> Key: HDFS-7390
> URL: https://issues.apache.org/jira/browse/HDFS-7390
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.5.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Attachments: HDFS-7390-003.patch, HDFS-7390-004.patch, 
> HDFS-7390-005.patch, HDFS-7390.patch, HDFS-7390.patch
>
>
> HDFS-2832  added heterogeneous support. In a cluster with different storage 
> types, it is useful to have metrics per storage type. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7390) Provide JMX metrics per storage type

2015-06-17 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7390:
---
Attachment: HDFS-7390-004.patch

Attaching a patch after fixing checkstyle warnings.

> Provide JMX metrics per storage type
> 
>
> Key: HDFS-7390
> URL: https://issues.apache.org/jira/browse/HDFS-7390
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.5.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Attachments: HDFS-7390-003.patch, HDFS-7390-004.patch, 
> HDFS-7390.patch, HDFS-7390.patch
>
>
> HDFS-2832  added heterogeneous support. In a cluster with different storage 
> types, it is useful to have metrics per storage type. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7390) Provide JMX metrics per storage type

2015-06-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7390?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7390:
---
Attachment: HDFS-7390-003.patch

Updated the patch. A new MXBean _BlockStats_ is defined and is implemented by 
_BlockManager_.
Used standard MXBean approach (based on CompositeDataSupport) to output Json.
Testcases are added.
Dependency on HADOOP-11308 is no longer needed and it is closed.
The sample Json output is as below:
{code}
{
"name" : "Hadoop:service=NameNode,name=BlockStats",
"modelerType" : 
"org.apache.hadoop.hdfs.server.blockmanagement.BlockManager",
"TieredStorageStats" : [ {
  "key" : "RAM_DISK",
  "value" : {
"blockPoolUsed" : 24576,
"capacityRemaining" : 1634963226624,
"capacityTotal" : 2994494914560,
"capacityUsed" : 24576,
"nodesInService" : 6
  }
}, {
  "key" : "ARCHIVE",
  "value" : {
"blockPoolUsed" : 12288,
"capacityRemaining" : 817481576448,
"capacityTotal" : 1497247457280,
"capacityUsed" : 12288,
"nodesInService" : 3
  }
}, {
  "key" : "DISK",
  "value" : {
"blockPoolUsed" : 12288,
"capacityRemaining" : 817481650176,
"capacityTotal" : 1497247457280,
"capacityUsed" : 12288,
"nodesInService" : 3
  }
} ]
  }
{code}

> Provide JMX metrics per storage type
> 
>
> Key: HDFS-7390
> URL: https://issues.apache.org/jira/browse/HDFS-7390
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>Affects Versions: 2.5.2
>Reporter: Benoy Antony
>Assignee: Benoy Antony
>  Labels: BB2015-05-TBR
> Attachments: HDFS-7390-003.patch, HDFS-7390.patch, HDFS-7390.patch
>
>
> HDFS-2832  added heterogeneous support. In a cluster with different storage 
> types, it is useful to have metrics per storage type. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8291) Modify NN WebUI to display correct unit

2015-05-05 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8291?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14529751#comment-14529751
 ] 

Benoy Antony commented on HDFS-8291:


Terabyte (TB) is more widely used than Tebibyte(TiB). I suggest fixing the 
calculation than changing from TB to TiB.

> Modify NN WebUI to display correct unit 
> 
>
> Key: HDFS-8291
> URL: https://issues.apache.org/jira/browse/HDFS-8291
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Zhongyi Xie
>Assignee: Zhongyi Xie
>Priority: Minor
> Attachments: HDFS-8291.001.patch, HDFS-8291.002.patch
>
>
> NN Web UI displays its capacity and usage in TB, but it is actually TiB. We 
> should either change the unit name or the calculation to ensure it follows 
> standards.
> http://en.wikipedia.org/wiki/Tebibyte
> http://en.wikipedia.org/wiki/Terabyte



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-6407) new namenode UI, lost ability to sort columns in datanode tab

2015-05-04 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-6407:
---
Attachment: HDFS-6407-003.patch

Adjusted the page size to a lower number (instead of All) to ensure that pages 
get painted fast when there are lots of items. 
This is an existing problem when there are thousands of items to display.

Sorting thousands of items has no noticeable delay to require any optimization. 
Tested with 8500 files in one directory. 



> new namenode UI, lost ability to sort columns in datanode tab
> -
>
> Key: HDFS-6407
> URL: https://issues.apache.org/jira/browse/HDFS-6407
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Nathan Roberts
>Assignee: Benoy Antony
>Priority: Minor
> Attachments: 002-datanodes-sorted-capacityUsed.png, 
> 002-datanodes.png, 002-filebrowser.png, 002-snapshots.png, 
> HDFS-6407-002.patch, HDFS-6407-003.patch, HDFS-6407.patch, 
> browse_directory.png, datanodes.png, snapshots.png
>
>
> old ui supported clicking on column header to sort on that column. The new ui 
> seems to have dropped this very useful feature.
> There are a few tables in the Namenode UI to display  datanodes information, 
> directory listings and snapshots.
> When there are many items in the tables, it is useful to have ability to sort 
> on the different columns.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-6407) new namenode UI, lost ability to sort columns in datanode tab

2015-05-01 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6407?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14524415#comment-14524415
 ] 

Benoy Antony commented on HDFS-6407:


The sorting seems to be correct for all the fields. Please let me know what 
makes you think otherwise and I can fix it.

The purpose of this jira is to put back sorting on data nodes tab. Since it was 
easy to enable sorting on other tables using the plugin, sorting and pagination 
was added to those tables as well. 

The processing is done completely on the client side and so there is no impact 
on the server side. At this point, the process of enabling sorting pagination 
for a table is simple and uniform.  To give up this simplicity, there should be 
some noticeable delay  with the current approach.

Tested with 6,000 files.  It took time to render 6,000 items irrespective of 
sorting is enabled or not.  
If pagination is set , then items are painted instantly.

In any case, no noticeable delay is caused by sorting. So I am not sure whether 
there is any point in optimizing this client side logic. 

One enhancement which enables the page to render fast , will be to set the 
default pagination to reasonably higher number like 200. In cases where there 
are 10,000+ items to paint, the pages gets rendered fast if page size is 200 
items. If the user wants all items, then he can select "All" display all files. 
Or User can search for the file in other pages.I will update the patch with 
page size set to 200 items for all tables.







> new namenode UI, lost ability to sort columns in datanode tab
> -
>
> Key: HDFS-6407
> URL: https://issues.apache.org/jira/browse/HDFS-6407
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Nathan Roberts
>Assignee: Benoy Antony
>Priority: Minor
> Attachments: 002-datanodes-sorted-capacityUsed.png, 
> 002-datanodes.png, 002-filebrowser.png, 002-snapshots.png, 
> HDFS-6407-002.patch, HDFS-6407.patch, browse_directory.png, datanodes.png, 
> snapshots.png
>
>
> old ui supported clicking on column header to sort on that column. The new ui 
> seems to have dropped this very useful feature.
> There are a few tables in the Namenode UI to display  datanodes information, 
> directory listings and snapshots.
> When there are many items in the tables, it is useful to have ability to sort 
> on the different columns.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-6407) new namenode UI, lost ability to sort columns in datanode tab

2015-05-01 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6407?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14523563#comment-14523563
 ] 

Benoy Antony commented on HDFS-6407:


That will be a good optimization. Can we please do it as an enhancement in a 
separate jira ? 


> new namenode UI, lost ability to sort columns in datanode tab
> -
>
> Key: HDFS-6407
> URL: https://issues.apache.org/jira/browse/HDFS-6407
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Nathan Roberts
>Assignee: Benoy Antony
>Priority: Minor
> Attachments: 002-datanodes-sorted-capacityUsed.png, 
> 002-datanodes.png, 002-filebrowser.png, 002-snapshots.png, 
> HDFS-6407-002.patch, HDFS-6407.patch, browse_directory.png, datanodes.png, 
> snapshots.png
>
>
> old ui supported clicking on column header to sort on that column. The new ui 
> seems to have dropped this very useful feature.
> There are a few tables in the Namenode UI to display  datanodes information, 
> directory listings and snapshots.
> When there are many items in the tables, it is useful to have ability to sort 
> on the different columns.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HDFS-6407) new namenode UI, lost ability to sort columns in datanode tab

2015-05-01 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6407?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14523438#comment-14523438
 ] 

Benoy Antony edited comment on HDFS-6407 at 5/1/15 4:52 PM:


The patch is updated to use the data tables plugin 
(http://www.datatables.net/). The same plugin is used in the Resource Manager 
UI also. Please check the attached screenshots to see the sorting in action.

The patch takes care of sorting based on file sizes. Also supports pagination 
on the client side.

Notes about the sorting and pagination features added to the different tables 
on Namenode UI:

# Uses data tables plugin (http://www.datatables.net/).
# Uses latest version 1.10.7 (Yarn uses 1.9.4  and Yarn can potentially migrate 
to the latest version)
# Added logic to detect columns showing file sizes and sort based on file 
sizes. 
# Added custom logic detect columns showing file size and percentage and sort 
based on file sizes. This is used to detect and sort “BlockPool Used” column 
which displays value like 7.74 MB (10%).
# The default pagination is to display all the rows to match the existing 
behavior. The user can further select smaller page sizes  using the drop down.
# The following tables are set up to be sortable and pageable : datanodes, 
decommissioning nodes, datanode volume failures, snapshot table directories, 
snapshotted directories and file/dir listing.
# For file/dir listings , default sort happens on the name column (last 
column). For snapshotted directories, the default sort happens on the snapshot 
directory (second column). For all other tables, the sort happens on the first 
column.



was (Author: benoyantony):
The patch is updated to use the data tables plugin 
(http://www.datatables.net/). The same plugin is used in the Resource Manager 
UI also.

The patch takes care of sorting based on file sizes. Also supports pagination 
on the client side.

Notes about the sorting and pagination features added to the different tables 
on Namenode UI:

# Uses data tables plugin (http://www.datatables.net/).
# Uses latest version 1.10.7 (Yarn uses 1.9.4  and Yarn can potentially migrate 
to the latest version)
# Added logic to detect columns showing file sizes and sort based on file 
sizes. 
# Added custom logic detect columns showing file size and percentage and sort 
based on file sizes. This is used to detect and sort “BlockPool Used” column 
which displays value like 7.74 MB (10%).
# The default pagination is to display all the rows to match the existing 
behavior. The user can further select smaller page sizes  using the drop down.
# The following tables are set up to be sortable and pageable : datanodes, 
decommissioning nodes, datanode volume failures, snapshot table directories, 
snapshotted directories and file/dir listing.
# For file/dir listings , default sort happens on the name column (last 
column). For snapshotted directories, the default sort happens on the snapshot 
directory (second column). For all other tables, the sort happens on the first 
column.


> new namenode UI, lost ability to sort columns in datanode tab
> -
>
> Key: HDFS-6407
> URL: https://issues.apache.org/jira/browse/HDFS-6407
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Nathan Roberts
>Assignee: Benoy Antony
>Priority: Minor
> Attachments: 002-datanodes-sorted-capacityUsed.png, 
> 002-datanodes.png, 002-filebrowser.png, 002-snapshots.png, 
> HDFS-6407-002.patch, HDFS-6407.patch, browse_directory.png, datanodes.png, 
> snapshots.png
>
>
> old ui supported clicking on column header to sort on that column. The new ui 
> seems to have dropped this very useful feature.
> There are a few tables in the Namenode UI to display  datanodes information, 
> directory listings and snapshots.
> When there are many items in the tables, it is useful to have ability to sort 
> on the different columns.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HDFS-6407) new namenode UI, lost ability to sort columns in datanode tab

2015-05-01 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6407?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14523438#comment-14523438
 ] 

Benoy Antony edited comment on HDFS-6407 at 5/1/15 4:51 PM:


The patch is updated to use the data tables plugin 
(http://www.datatables.net/). The same plugin is used in the Resource Manager 
UI also.

The patch takes care of sorting based on file sizes. Also supports pagination 
on the client side.

Notes about the sorting and pagination features added to the different tables 
on Namenode UI:

# Uses data tables plugin (http://www.datatables.net/).
# Uses latest version 1.10.7 (Yarn uses 1.9.4  and Yarn can potentially migrate 
to the latest version)
# Added logic to detect columns showing file sizes and sort based on file 
sizes. 
# Added custom logic detect columns showing file size and percentage and sort 
based on file sizes. This is used to detect and sort “BlockPool Used” column 
which displays value like 7.74 MB (10%).
# The default pagination is to display all the rows to match the existing 
behavior. The user can further select smaller page sizes  using the drop down.
# The following tables are set up to be sortable and pageable : datanodes, 
decommissioning nodes, datanode volume failures, snapshot table directories, 
snapshotted directories and file/dir listing.
# For file/dir listings , default sort happens on the name column (last 
column). For snapshotted directories, the default sort happens on the snapshot 
directory (second column). For all other tables, the sort happens on the first 
column.



was (Author: benoyantony):
The patch is update using the data tables plugin (http://www.datatables.net/). 
The same plugin is used in the Resource Manager UI also.

The patch takes care of sorting based on file sizes. Also supports pagination 
on the client side.

Notes about the sorting and pagination features added to the different tables 
on Namenode UI:

# Uses data tables plugin (http://www.datatables.net/).
# Uses latest version 1.10.7 (Yarn uses 1.9.4  and Yarn can potentially migrate 
to the latest version)
# Added logic to detect columns showing file sizes and sort based on file 
sizes. 
# Added custom logic detect columns showing file size and percentage and sort 
based on file sizes. This is used to detect and sort “BlockPool Used” column 
which displays value like 7.74 MB (10%).
# The default pagination is to display all the rows to match the existing 
behavior. The user can further select smaller page sizes  using the drop down.
# The following tables are set up to be sortable and pageable : datanodes, 
decommissioning nodes, datanode volume failures, snapshot table directories, 
snapshotted directories and file/dir listing.
# For file/dir listings , default sort happens on the name column (last 
column). For snapshotted directories, the default sort happens on the snapshot 
directory (second column). For all other tables, the sort happens on the first 
column.


> new namenode UI, lost ability to sort columns in datanode tab
> -
>
> Key: HDFS-6407
> URL: https://issues.apache.org/jira/browse/HDFS-6407
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Nathan Roberts
>Assignee: Benoy Antony
>Priority: Minor
> Attachments: 002-datanodes-sorted-capacityUsed.png, 
> 002-datanodes.png, 002-filebrowser.png, 002-snapshots.png, 
> HDFS-6407-002.patch, HDFS-6407.patch, browse_directory.png, datanodes.png, 
> snapshots.png
>
>
> old ui supported clicking on column header to sort on that column. The new ui 
> seems to have dropped this very useful feature.
> There are a few tables in the Namenode UI to display  datanodes information, 
> directory listings and snapshots.
> When there are many items in the tables, it is useful to have ability to sort 
> on the different columns.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-6407) new namenode UI, lost ability to sort columns in datanode tab

2015-05-01 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-6407:
---
Attachment: 002-snapshots.png
002-filebrowser.png
002-datanodes-sorted-capacityUsed.png
002-datanodes.png

> new namenode UI, lost ability to sort columns in datanode tab
> -
>
> Key: HDFS-6407
> URL: https://issues.apache.org/jira/browse/HDFS-6407
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Nathan Roberts
>Assignee: Benoy Antony
>Priority: Minor
> Attachments: 002-datanodes-sorted-capacityUsed.png, 
> 002-datanodes.png, 002-filebrowser.png, 002-snapshots.png, HDFS-6407.patch, 
> browse_directory.png, datanodes.png, snapshots.png
>
>
> old ui supported clicking on column header to sort on that column. The new ui 
> seems to have dropped this very useful feature.
> There are a few tables in the Namenode UI to display  datanodes information, 
> directory listings and snapshots.
> When there are many items in the tables, it is useful to have ability to sort 
> on the different columns.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-6407) new namenode UI, lost ability to sort columns in datanode tab

2015-05-01 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-6407:
---
Attachment: HDFS-6407-002.patch

The patch is update using the data tables plugin (http://www.datatables.net/). 
The same plugin is used in the Resource Manager UI also.

The patch takes care of sorting based on file sizes. Also supports pagination 
on the client side.

Notes about the sorting and pagination features added to the different tables 
on Namenode UI:

# Uses data tables plugin (http://www.datatables.net/).
# Uses latest version 1.10.7 (Yarn uses 1.9.4  and Yarn can potentially migrate 
to the latest version)
# Added logic to detect columns showing file sizes and sort based on file 
sizes. 
# Added custom logic detect columns showing file size and percentage and sort 
based on file sizes. This is used to detect and sort “BlockPool Used” column 
which displays value like 7.74 MB (10%).
# The default pagination is to display all the rows to match the existing 
behavior. The user can further select smaller page sizes  using the drop down.
# The following tables are set up to be sortable and pageable : datanodes, 
decommissioning nodes, datanode volume failures, snapshot table directories, 
snapshotted directories and file/dir listing.
# For file/dir listings , default sort happens on the name column (last 
column). For snapshotted directories, the default sort happens on the snapshot 
directory (second column). For all other tables, the sort happens on the first 
column.


> new namenode UI, lost ability to sort columns in datanode tab
> -
>
> Key: HDFS-6407
> URL: https://issues.apache.org/jira/browse/HDFS-6407
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: namenode
>Affects Versions: 2.4.0
>Reporter: Nathan Roberts
>Assignee: Benoy Antony
>Priority: Minor
> Attachments: 002-datanodes-sorted-capacityUsed.png, 
> 002-datanodes.png, 002-filebrowser.png, 002-snapshots.png, 
> HDFS-6407-002.patch, HDFS-6407.patch, browse_directory.png, datanodes.png, 
> snapshots.png
>
>
> old ui supported clicking on column header to sort on that column. The new ui 
> seems to have dropped this very useful feature.
> There are a few tables in the Namenode UI to display  datanodes information, 
> directory listings and snapshots.
> When there are many items in the tables, it is useful to have ability to sort 
> on the different columns.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-5823) Document async audit logging

2015-03-11 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5823?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14357338#comment-14357338
 ] 

Benoy Antony commented on HDFS-5823:


[~daryn]

Can't I achieve the same by configuring an AsyncAppender as below :
Do you see any downside in doing this ?

{code}

 
 
 




{code}

> Document async audit logging
> 
>
> Key: HDFS-5823
> URL: https://issues.apache.org/jira/browse/HDFS-5823
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>
> HDFS-5241 added an option for async log4j audit logging.  The option is 
> considered semi-experimental and should be documented in hdfs-defaults.xml 
> after it's stability under stress is proven.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-5823) Document async audit logging

2015-03-09 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5823?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14353731#comment-14353731
 ] 

Benoy Antony commented on HDFS-5823:


[~daryn]
Is the async audit log feature feature stable?

> Document async audit logging
> 
>
> Key: HDFS-5823
> URL: https://issues.apache.org/jira/browse/HDFS-5823
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Daryn Sharp
>Assignee: Daryn Sharp
>
> HDFS-5241 added an option for async log4j audit logging.  The option is 
> considered semi-experimental and should be documented in hdfs-defaults.xml 
> after it's stability under stress is proven.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-02-25 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7467:
---
Fix Version/s: 2.7.0

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.7.0
>
> Attachments: HDFS-7467-002.patch, HDFS-7467-003.patch, 
> HDFS-7467-004.patch, HDFS-7467.patch, storagepolicydisplay.pdf
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-02-25 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14337588#comment-14337588
 ] 

Benoy Antony commented on HDFS-7467:


committed to trunk and branch-2.

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.7.0
>
> Attachments: HDFS-7467-002.patch, HDFS-7467-003.patch, 
> HDFS-7467-004.patch, HDFS-7467.patch, storagepolicydisplay.pdf
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-02-25 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7467:
---
Resolution: Fixed
Status: Resolved  (was: Patch Available)

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467-002.patch, HDFS-7467-003.patch, 
> HDFS-7467-004.patch, HDFS-7467.patch, storagepolicydisplay.pdf
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-02-24 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14335606#comment-14335606
 ] 

Benoy Antony commented on HDFS-7467:


Thanks for the review [~szetszwo].
If there are no further comments, I'll commit the patch tomorrow.



> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467-002.patch, HDFS-7467-003.patch, 
> HDFS-7467-004.patch, HDFS-7467.patch, storagepolicydisplay.pdf
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-02-23 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7467:
---
Attachment: HDFS-7467-004.patch

Attaching the patch which takes care of the recommended changes. 
Removed the method StoragePolicySummary.process()  as the statements in that 
method could be done via other methods.

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467-002.patch, HDFS-7467-003.patch, 
> HDFS-7467-004.patch, HDFS-7467.patch, storagepolicydisplay.pdf
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-02-09 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14312484#comment-14312484
 ] 

Benoy Antony commented on HDFS-7467:


The findbugs warning and test failure are unrelated.

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467-002.patch, HDFS-7467-003.patch, 
> HDFS-7467.patch, storagepolicydisplay.pdf
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-02-08 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14311609#comment-14311609
 ] 

Benoy Antony edited comment on HDFS-7467 at 2/8/15 10:57 PM:
-

Attaching the patch which fixes the findbugs


was (Author: benoyantony):
Fixing the patch which fixes the findbugs

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467-002.patch, HDFS-7467-003.patch, 
> HDFS-7467.patch, storagepolicydisplay.pdf
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-02-08 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7467:
---
Attachment: HDFS-7467-003.patch

Fixing the patch which fixes the findbugs

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467-002.patch, HDFS-7467-003.patch, 
> HDFS-7467.patch, storagepolicydisplay.pdf
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-02-07 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7467:
---
Status: Patch Available  (was: In Progress)

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467-002.patch, HDFS-7467.patch, 
> storagepolicydisplay.pdf
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-02-07 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7467:
---
Attachment: HDFS-7467-002.patch

Attaching the patch which provides the _storagepolicy_ information via _fsck_.
The above comments are considered.
Sample outputs are shown the pdf attached.

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467-002.patch, HDFS-7467.patch, 
> storagepolicydisplay.pdf
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-02-07 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HDFS-7467:
---
Attachment: storagepolicydisplay.pdf

Attaching the document containing sample fsck outputs when invoked with 
_storagepolicies_ option

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467.patch, storagepolicydisplay.pdf
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-6133) Make Balancer support exclude specified path

2015-02-06 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6133?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14310564#comment-14310564
 ] 

Benoy Antony commented on HDFS-6133:


I know that there are different ways to accomplish the desired functionality of 
pinning the blocks under a directory to a set of nodes.
I want to suggest yet another approach using storage policies.

This is based on the fact that the set of favored nodes can be considered as a 
tier. 

Let’s say blocks under /hbase needs to be pinned to a set of nodes.

1. Mark the data.dirs on those set of datanodes as a special storage type , say 
HBASE_TIER. [HBASE_TIER]/hadoop01/data
2. Define a storage policy - HBASE_POLICY which insists that all blocks 
following the policy, reside in HBASE_TIER
3. Set storage policy of /hbase to HBASE_POLICY 

Since the balancer(and mover) considers the storage policy while 
moving/balancing, balancer will not move these blocks out of the favored 
nodes(HBASE_TIER)

Now if we want to keep other blocks ( non hbase blocks) on these datanodes as 
well(may not be a great idea), then we can model it by having multiple storage 
types on a single directory like [HBASE_TIER,DISK]/hadoop01/data. This make 
sense since one want to have blocks having multiple storage characteristics on 
the same node.
This will need a (minor) code modification.


> Make Balancer support exclude specified path
> 
>
> Key: HDFS-6133
> URL: https://issues.apache.org/jira/browse/HDFS-6133
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer & mover, namenode
>Reporter: zhaoyunjiong
>Assignee: zhaoyunjiong
> Attachments: HDFS-6133-1.patch, HDFS-6133-2.patch, HDFS-6133-3.patch, 
> HDFS-6133-4.patch, HDFS-6133-5.patch, HDFS-6133-6.patch, HDFS-6133-7.patch, 
> HDFS-6133-8.patch, HDFS-6133.patch
>
>
> Currently, run Balancer will destroying Regionserver's data locality.
> If getBlocks could exclude blocks belongs to files which have specific path 
> prefix, like "/hbase", then we can run Balancer without destroying 
> Regionserver's data locality.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-01-07 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14268598#comment-14268598
 ] 

Benoy Antony commented on HDFS-7467:


{quote}
If a file does not satisfies the specified policy, fsck should show such 
information.
{quote}

That's good for a file. What do we do for directories which can potentially 
have children with different policies ? 

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467.patch
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-01-06 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14266871#comment-14266871
 ] 

Benoy Antony edited comment on HDFS-7467 at 1/6/15 10:22 PM:
-

1. 
{quote}
Are all storage policies in fallback storage equivalent to other storage 
policies that this output can always be fully described by the percentages that 
Tsz has suggested?
{quote}

There is a possibility that some storage tier combination may not belong to a 
storage policy. 
My recommendation is to display the policy along with the combination if 
possible. If not, display the combination. Lowercase for policy name is 
intentional.
{code}
Storage Policy   # of blocks   % of blocks
cold(DISK:1,ARCHIVE:2)  340730  97.7393%
frozen(ARCHIVE:3)   39281.1268%
DISK:2,ARCHIVE:2   31220.8956%
warm(DISK:2,ARCHIVE:1)   7480.2146%
DISK:1,ARCHIVE:3440.0126%
DISK:3,ARCHIVE:2300.0086%
DISK:3,ARCHIVE:1  90.0026%
{code}

2.
{quote}
There should also be some warning messages as well in fsck for all files that 
are unable to meet the requested ideal for their storage policy and are using 
fallback storage, perhaps with a switch since that could become overly volumous 
output.
{quote}

This is a nice feature. Will look into that .


was (Author: benoyantony):
1. 
{quote}
Are all storage policies in fallback storage equivalent to other storage 
policies that this output can always be fully described by the percentages that 
Tsz has suggested?
{quote}

There is a possibility that some storage tier combination may not belong to a 
storage policy. 
My recommendation is to display the policy along with the combination if 
possible. If not, display the combination. Lowercase for policy name is 
intentional.
{code}
Storage Policy  # of blocks   % of blocks
cold(DISK:1,ARCHIVE:2) 340730  97.7393%
frozen(ARCHIVE:3)  39281.1268%
DISK:2,ARCHIVE:2  31220.8956%
warm(DISK:2,ARCHIVE:1) 7480.2146%
DISK:1,ARCHIVE:3  440.0126%
DISK:3,ARCHIVE:2  300.0086%
DISK:3,ARCHIVE:190.0026%
 
{code}

2.
{quote}
There should also be some warning messages as well in fsck for all files that 
are unable to meet the requested ideal for their storage policy and are using 
fallback storage, perhaps with a switch since that could become overly volumous 
output.
{quote}

This is a nice feature. Will look into that .

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467.patch
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-01-06 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14266871#comment-14266871
 ] 

Benoy Antony edited comment on HDFS-7467 at 1/6/15 10:18 PM:
-

1. 
{quote}
Are all storage policies in fallback storage equivalent to other storage 
policies that this output can always be fully described by the percentages that 
Tsz has suggested?
{quote}

There is a possibility that some storage tier combination may not belong to a 
storage policy. 
My recommendation is to display the policy along with the combination if 
possible. If not, display the combination. Lowercase for policy name is 
intentional.
{code}
Storage Policy  # of blocks   % of blocks
cold(DISK:1,ARCHIVE:2) 340730  97.7393%
frozen(ARCHIVE:3)  39281.1268%
DISK:2,ARCHIVE:2  31220.8956%
warm(DISK:2,ARCHIVE:1) 7480.2146%
DISK:1,ARCHIVE:3  440.0126%
DISK:3,ARCHIVE:2  300.0086%
DISK:3,ARCHIVE:190.0026%
 
{code}

2.
{quote}
There should also be some warning messages as well in fsck for all files that 
are unable to meet the requested ideal for their storage policy and are using 
fallback storage, perhaps with a switch since that could become overly volumous 
output.
{quote}

This is a nice feature. Will look into that .


was (Author: benoyantony):
1. 
{quote}
Are all storage policies in fallback storage equivalent to other storage 
policies that this output can always be fully described by the percentages that 
Tsz has suggested?
{quote}

There is a possibility that some storage tier combination may not belong to a 
storage policy. 
My recommendation is to display the policy along with the combination if 
possible. If not, display the combination. Lowercase for policy name is 
intentional.
{code}
Storage Policy  # of blocks   % of blocks
cold(DISK:1,ARCHIVE:2) 340730   97.7393%
frozen(ARCHIVE:3)  39281.1268%
DISK:2,ARCHIVE:2  31220.8956%
warm(DISK:2,ARCHIVE:1) 7480.2146%
DISK:1,ARCHIVE:3  440.0126%
DISK:3,ARCHIVE:2  300.0086%
DISK:3,ARCHIVE:190.0026%
 
{code}

2.
{quote}
There should also be some warning messages as well in fsck for all files that 
are unable to meet the requested ideal for their storage policy and are using 
fallback storage, perhaps with a switch since that could become overly volumous 
output.
{quote}

This is a nice feature. Will look into that .

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467.patch
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-01-06 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14266871#comment-14266871
 ] 

Benoy Antony edited comment on HDFS-7467 at 1/6/15 10:17 PM:
-

1. 
{quote}
Are all storage policies in fallback storage equivalent to other storage 
policies that this output can always be fully described by the percentages that 
Tsz has suggested?
{quote}

There is a possibility that some storage tier combination may not belong to a 
storage policy. 
My recommendation is to display the policy along with the combination if 
possible. If not, display the combination. Lowercase for policy name is 
intentional.
{code}
Storage Policy  # of blocks   % of blocks
cold(DISK:1,ARCHIVE:2) 340730   97.7393%
frozen(ARCHIVE:3)  39281.1268%
DISK:2,ARCHIVE:2  31220.8956%
warm(DISK:2,ARCHIVE:1) 7480.2146%
DISK:1,ARCHIVE:3  440.0126%
DISK:3,ARCHIVE:2  300.0086%
DISK:3,ARCHIVE:190.0026%
 
{code}

2.
{quote}
There should also be some warning messages as well in fsck for all files that 
are unable to meet the requested ideal for their storage policy and are using 
fallback storage, perhaps with a switch since that could become overly volumous 
output.
{quote}

This is a nice feature. Will look into that .


was (Author: benoyantony):
1. 
{quote}
Are all storage policies in fallback storage equivalent to other storage 
policies that this output can always be fully described by the percentages that 
Tsz has suggested?
{quote}

There is a possibility that some storage tier combination may not belong to a 
storage policy. 
My recommendation is to display the policy along with the combination if 
possible. If not, display the combination. Lowercase for policy name is 
intentional.
{code}
Storage Policy  # of blocks   % of blocks
cold (DISK:1,ARCHIVE:2)340730   97.7393%
frozen (ARCHIVE:3) 39281.1268%
DISK:2,ARCHIVE:2  31220.8956%
warm (DISK:2,ARCHIVE:1)7480.2146%
DISK:1,ARCHIVE:3  440.0126%
DISK:3,ARCHIVE:2  300.0086%
DISK:3,ARCHIVE:190.0026%
 
{code}

2.
{quote}
There should also be some warning messages as well in fsck for all files that 
are unable to meet the requested ideal for their storage policy and are using 
fallback storage, perhaps with a switch since that could become overly volumous 
output.
{quote}

This is a nice feature. Will look into that .

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467.patch
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-01-06 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14266871#comment-14266871
 ] 

Benoy Antony edited comment on HDFS-7467 at 1/6/15 10:16 PM:
-

1. 
{quote}
Are all storage policies in fallback storage equivalent to other storage 
policies that this output can always be fully described by the percentages that 
Tsz has suggested?
{quote}

There is a possibility that some storage tier combination may not belong to a 
storage policy. 
My recommendation is to display the policy along with the combination if 
possible. If not, display the combination. Lowercase for policy name is 
intentional.
{code}
Storage Policy  # of blocks   % of blocks
cold (DISK:1,ARCHIVE:2)340730   97.7393%
frozen (ARCHIVE:3) 39281.1268%
DISK:2,ARCHIVE:2  31220.8956%
warm (DISK:2,ARCHIVE:1)7480.2146%
DISK:1,ARCHIVE:3  440.0126%
DISK:3,ARCHIVE:2  300.0086%
DISK:3,ARCHIVE:190.0026%
 
{code}

2.
{quote}
There should also be some warning messages as well in fsck for all files that 
are unable to meet the requested ideal for their storage policy and are using 
fallback storage, perhaps with a switch since that could become overly volumous 
output.
{quote}

This is a nice feature. Will look into that .


was (Author: benoyantony):
1. 
{quote}
Are all storage policies in fallback storage equivalent to other storage 
policies that this output can always be fully described by the percentages that 
Tsz has suggested?
{quote}

There is a possibility that some storage tier combination may not belong to a 
storage policy. 
My recommendation is to display the policy along with the combination if 
possible. If not, display the combination. Lowercase for policy name is 
intentional.
{code}
Storage Policy  # of blocks   % of blocks
cold (DISK:1,ARCHIVE:2)  340730   97.7393%
frozen (ARCHIVE:3)   39281.1268%
DISK:2,ARCHIVE:231220.8956%
warm (DISK:2,ARCHIVE:1)   7480.2146%
DISK:1,ARCHIVE:3 440.0126%
DISK:3,ARCHIVE:2 300.0086%
DISK:3,ARCHIVE:1   90.0026% 

{code}

2.
{quote}
There should also be some warning messages as well in fsck for all files that 
are unable to meet the requested ideal for their storage policy and are using 
fallback storage, perhaps with a switch since that could become overly volumous 
output.
{quote}

This is a nice feature. Will look into that .

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467.patch
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7467) Provide storage tier information for a directory via fsck

2015-01-06 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14266871#comment-14266871
 ] 

Benoy Antony commented on HDFS-7467:


1. 
{quote}
Are all storage policies in fallback storage equivalent to other storage 
policies that this output can always be fully described by the percentages that 
Tsz has suggested?
{quote}

There is a possibility that some storage tier combination may not belong to a 
storage policy. 
My recommendation is to display the policy along with the combination if 
possible. If not, display the combination. Lowercase for policy name is 
intentional.
{code}
Storage Policy  # of blocks   % of blocks
cold (DISK:1,ARCHIVE:2)  340730   97.7393%
frozen (ARCHIVE:3)   39281.1268%
DISK:2,ARCHIVE:231220.8956%
warm (DISK:2,ARCHIVE:1)   7480.2146%
DISK:1,ARCHIVE:3 440.0126%
DISK:3,ARCHIVE:2 300.0086%
DISK:3,ARCHIVE:1   90.0026% 

{code}

2.
{quote}
There should also be some warning messages as well in fsck for all files that 
are unable to meet the requested ideal for their storage policy and are using 
fallback storage, perhaps with a switch since that could become overly volumous 
output.
{quote}

This is a nice feature. Will look into that .

> Provide storage tier information for a directory via fsck
> -
>
> Key: HDFS-7467
> URL: https://issues.apache.org/jira/browse/HDFS-7467
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: balancer & mover
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Attachments: HDFS-7467.patch
>
>
> Currently _fsck_  provides information regarding blocks for a directory.
> It should be augmented to provide storage tier information (optionally). 
> The sample report could be as follows :
> {code}
> Storage Tier Combination# of blocks   % of blocks
> DISK:1,ARCHIVE:2  340730   97.7393%
>  
> ARCHIVE:3   39281.1268%
>  
> DISK:2,ARCHIVE:231220.8956%
>  
> DISK:2,ARCHIVE:1 7480.2146%
>  
> DISK:1,ARCHIVE:3  440.0126%
>  
> DISK:3,ARCHIVE:2  300.0086%
>  
> DISK:3,ARCHIVE:1   90.0026%
> {code}
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


  1   2   3   4   5   >