[ 
https://issues.apache.org/jira/browse/HDFS-16270?focusedWorklogId=669879&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-669879
 ]

ASF GitHub Bot logged work on HDFS-16270:
-----------------------------------------

                Author: ASF GitHub Bot
            Created on: 26/Oct/21 02:23
            Start Date: 26/Oct/21 02:23
    Worklog Time Spent: 10m 
      Work Description: jianghuazhu closed pull request #3547:
URL: https://github.com/apache/hadoop/pull/3547


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
-------------------

    Worklog Id:     (was: 669879)
    Time Spent: 1h 20m  (was: 1h 10m)

> Improve NNThroughputBenchmark#printUsage() related to block size
> ----------------------------------------------------------------
>
>                 Key: HDFS-16270
>                 URL: https://issues.apache.org/jira/browse/HDFS-16270
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: benchmarks, namenode
>            Reporter: JiangHua Zhu
>            Assignee: JiangHua Zhu
>            Priority: Major
>              Labels: pull-request-available
>          Time Spent: 1h 20m
>  Remaining Estimate: 0h
>
> When using the NNThroughputBenchmark test, if the usage is not correct, we 
> will get some prompt messages.
> E.g:
> '
> If connecting to a remote NameNode with -fs option, 
> dfs.namenode.fs-limits.min-block-size should be set to 16.
> 21/10/13 11:55:32 INFO util.ExitUtil: Exiting with status -1: ExitException
> '
> Yes, this way is good.
> However, the setting of'dfs.blocksize' has been completed before execution, 
> for example:
> conf.setInt(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, 16);
> We will still get the above prompt, which is wrong.
> At the same time, it should also be explained. The hint here should not be 
> for'dfs.namenode.fs-limits.min-block-size', but should be'dfs.blocksize'.
> Because in the NNThroughputBenchmark construction, 
> the'dfs.namenode.fs-limits.min-block-size' has been set to 0 in advance.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org

Reply via email to