[ https://issues.apache.org/jira/browse/HDFS-7630?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14294898#comment-14294898 ]
sam liu commented on HDFS-7630: ------------------------------- Arpit, Not all such hard-coding will cause failure and the patch is mainly to remove the hard-coding. But sometimes the hard-coding could cause failure. For example, without patch HDFS-7585, test TestEnhancedByteBufferAccess will fail on power platform. In the tests, the size of BLOCK_SIZE is usually set as 4096 and it just equals to the default page size of x86 Linux operating system, but on power Linux operating system the default page size is 65536. As HDFS is based on operating system, it might be better the unit tests could consider the differences of operating systems. Thanks! > TestConnCache hardcode block size without considering native OS > --------------------------------------------------------------- > > Key: HDFS-7630 > URL: https://issues.apache.org/jira/browse/HDFS-7630 > Project: Hadoop HDFS > Issue Type: Bug > Components: test > Reporter: sam liu > Assignee: sam liu > Attachments: HDFS-7630.001.patch, HDFS-7630.002.patch > > > TestConnCache hardcode block size with 'BLOCK_SIZE = 4096', however it's > incorrect on some platforms. For example, on power platform, the correct > value is 65536. -- This message was sent by Atlassian JIRA (v6.3.4#6332)