[ 
https://issues.apache.org/jira/browse/HDFS-5517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13823344#comment-13823344
 ] 

Uma Maheswara Rao G commented on HDFS-5517:
-------------------------------------------

+1 for lowering the value. Thanks Aaron for proposing this change. Recently we 
also faced this issue. When file having 2L blocks itself NN is behaving 
something abnormal.
For every block addition we do update blocks, so this big array will be 
persisted every time incrementally I think. This will be overhead really, if 
allow to write more and more blocks for same file with the current code But in 
practice I am not sure what will be the scenario to configure less block sizes 
and more blocks for same file. But we saw this in testing cluster.

> Lower the default maximum number of blocks per file
> ---------------------------------------------------
>
>                 Key: HDFS-5517
>                 URL: https://issues.apache.org/jira/browse/HDFS-5517
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: namenode
>    Affects Versions: 2.2.0
>            Reporter: Aaron T. Myers
>            Assignee: Aaron T. Myers
>
> We introduced the maximum number of blocks per file in HDFS-4305, but we set 
> the default to 1MM. In practice this limit is so high as to never be hit, 
> whereas we know that an individual file with 10s of thousands of blocks can 
> cause problems. We should lower the default value, in my opinion to 10k.



--
This message was sent by Atlassian JIRA
(v6.1#6144)

Reply via email to