[ https://issues.apache.org/jira/browse/HDDS-2259?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Anu Engineer updated HDDS-2259: ------------------------------- Fix Version/s: 0.5.0 Resolution: Fixed Status: Resolved (was: Patch Available) Thank you for the contribution. I have committed this patch to the trunk. > Container Data Scrubber computes wrong checksum > ----------------------------------------------- > > Key: HDDS-2259 > URL: https://issues.apache.org/jira/browse/HDDS-2259 > Project: Hadoop Distributed Data Store > Issue Type: Sub-task > Components: Ozone Datanode > Affects Versions: 0.5.0 > Reporter: Attila Doroszlai > Assignee: Attila Doroszlai > Priority: Critical > Labels: pull-request-available > Fix For: 0.5.0 > > Time Spent: 50m > Remaining Estimate: 0h > > Chunk checksum verification fails for (almost) any file. This is caused by > computing checksum for the entire buffer, regardless of the actual size of > the chunk. > {code:title=https://github.com/apache/hadoop/blob/55c5436f39120da0d7dabf43d7e5e6404307123b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/KeyValueContainerCheck.java#L259-L273} > byte[] buffer = new byte[cData.getBytesPerChecksum()]; > ... > v = fs.read(buffer); > ... > bytesRead += v; > ... > ByteString actual = cal.computeChecksum(buffer) > .getChecksums().get(0); > {code} > This results in marking all closed containers as unhealthy. -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org