[ 
https://issues.apache.org/jira/browse/HDDS-284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16692490#comment-16692490
 ] 

Hanisha Koneru commented on HDDS-284:
-------------------------------------

Thank you for the review [~shashikant].
{quote} 2. With the patch it aways seems to be computing the checksum in 
writeChunkToContainerCall. With HTTP headers, if the checksum is already 
available in a Rest call, we might not require to recompute again. Are we going 
to address such cases later?{quote}
We can add support for this later. If the checksum is already provided in the 
HTTP header, we can use that and skip the computation. Let's open a new Jira to 
track this?
{quote} 3. ChunkManagerImpl#writeChunk:-> while handling the overWrites of a 
chunkFile we can just verify the checksum if its already present and return 
accordingly without actually doing I/O ( addressed as TODO in the code). We can 
also add the checksum verification here, though these can be addressed in a 
separate patch as well.{quote}
Yes, let's address this also in a separate patch.
{quote} 4. ChunkInputStream.java : L213-215 : why is this change specifically 
required? Is it just for making the tests added to work?{quote}
Added this to propagate the actual exception to the client. Otherwise, we just 
get "Unexpected OzoneException" without specifying the actual reason for the 
failure.

> CRC for ChunksData
> ------------------
>
>                 Key: HDDS-284
>                 URL: https://issues.apache.org/jira/browse/HDDS-284
>             Project: Hadoop Distributed Data Store
>          Issue Type: Improvement
>            Reporter: Bharat Viswanadham
>            Assignee: Hanisha Koneru
>            Priority: Major
>         Attachments: CRC and Error Detection for Containers.pdf, 
> HDDS-284.00.patch, HDDS-284.005.patch, HDDS-284.01.patch, HDDS-284.02.patch, 
> HDDS-284.03.patch, HDDS-284.04.patch, Interleaving CRC and Error Detection 
> for Containers.pdf
>
>
> This Jira is to add CRC for chunks data.
> Right now a Chunk Info structure looks like this:
> _message ChunkInfo {_
>  _required string chunkName =_ _1__;_
> _required uint64 offset =_ _2__;_
> _required uint64 len =_ _3__;_
> _optional string checksum =_ _4__;_
> _repeated KeyValue metadata =_ _5__;_
> _}_
> _Proposal is to change ChunkInfo structure as below:_
> _message ChunkInfo {_
>  _required string chunkName =_ _1__;_
>  _required uint64 offset =_ _2__;_
>  _required uint64 len =_ _3__;_
>  _optional bytes checksum =_ _4__;_
>  _optional CRCType checksumType =_ _5__;_
>  _optional string legacyMetadata =_ _6__;_
>  _optional string legacyData =_ _7__;_
>  _repeated KeyValue metadata =_ _8__;_
> _}_
> _Instead of changing disk format, we put the checksum, checksumtype and 
> legacy data fields in to chunkInfo._



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org

Reply via email to