[ https://issues.apache.org/jira/browse/HDFS-1877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13036320#comment-13036320 ]
Tsz Wo (Nicholas), SZE commented on HDFS-1877: ---------------------------------------------- - The variables, {{inJunitMode}}, {{BLOCK_SIZE}}, {{dfs}}, are not actually used. Please remove them. - How about the default {{filenameOption}} equals {{ROOT_DIR}}? - You may simply have {{static private Log LOG = LogFactory.getLog(TestWriteRead.class);}} {code} + static private Log LOG; + + @Before + public void initJunitModeTest() throws Exception { + LOG = LogFactory.getLog(TestWriteRead.class); {code} - Please remove the following. The default is already INFO. {code} + ((Log4JLogger) FSNamesystem.LOG).getLogger().setLevel(Level.INFO); + ((Log4JLogger) DFSClient.LOG).getLogger().setLevel(Level.INFO); {code} - Most public methods should be package private. - Please add comments to tell how to use the command options and the default values. > Create a functional test for file read/write > -------------------------------------------- > > Key: HDFS-1877 > URL: https://issues.apache.org/jira/browse/HDFS-1877 > Project: Hadoop HDFS > Issue Type: Test > Components: test > Affects Versions: 0.22.0 > Reporter: CW Chung > Priority: Minor > Attachments: TestWriteRead.java, TestWriteRead.patch > > > It would be a great to have a tool, running on a real grid, to perform > function test (and stress tests to certain extent) for the file operations. > The tool would be written in Java and makes HDFS API calls to read, write, > append, hflush hadoop files. The tool would be usable standalone, or as a > building block for other regression or stress test suites (written in shell, > perl, python, etc). -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira