[ 
https://issues.apache.org/jira/browse/HDFS-7337?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14359825#comment-14359825
 ] 

Kai Zheng commented on HDFS-7337:
---------------------------------

Thanks [~zhz] for the review and thoughts.
bq.ErasureCodec is like a factory or an utility class, which creates 
ErasureCoder and BlockGrouper based on ECSchema
{{ErasureCodec}} would be the high level construct in the framework that covers 
all the potential erasure code specific aspects, including but might not be 
limited to {{ErasureCoder}} and {{BlockGrouper}}, which allows to be 
implemented and deployed as a whole for a new code. All the underlying code 
specific logic can be hooked via codec and can only be accessible thru codec. I 
understand there will be something more to think about, it's generally one of 
the major goal for the framework.
bq.I think we can leverage the pattern of BlockStoragePolicySuite
It's a good pattern. {{ErasureCodec}} follows another good pattern, 
{{CompressionCodec}}. 
bq. Something like:...your illustration codes...
I understand we need to hard-code a default schema for the system. What we have 
discussed and been doing is we allow to predefine EC schemas in an external 
file (XML currently used as we regularly do in the project). For easy 
reference, unique schema name (string) and codec name (string) are used. Do you 
have any concern for this way ?
bq.Then NN can just pass around the schema ID when communicating with DN and 
client, which is much smaller than an ErasureCodec object.
Yes similarly it's to pass around the schema NAME between any pair among NN, 
DN, client. It's not meaning to pass ErasureCodec object. Is there confusing 
sentence I need to clarify in the doc ? All the {{ErasureCodec}}s are loaded 
thru core-site configuration or service locators, and kept in map with codec 
name as the key. Providing the codec name, a codec will be fetched from the 
map. Codec object isn't needed to be passed around, codec name is. I guess 
you're meaning schema object. In the f2f meetup discussion with [~jingzhao], we 
mentioned it may need to pass around schema object. If we don't want to 
hard-code all the schemas, then we need to pass schema object I guess.

> Configurable and pluggable Erasure Codec and schema
> ---------------------------------------------------
>
>                 Key: HDFS-7337
>                 URL: https://issues.apache.org/jira/browse/HDFS-7337
>             Project: Hadoop HDFS
>          Issue Type: Sub-task
>            Reporter: Zhe Zhang
>            Assignee: Kai Zheng
>         Attachments: HDFS-7337-prototype-v1.patch, 
> HDFS-7337-prototype-v2.zip, HDFS-7337-prototype-v3.zip, 
> PluggableErasureCodec-v2.pdf, PluggableErasureCodec.pdf
>
>
> According to HDFS-7285 and the design, this considers to support multiple 
> Erasure Codecs via pluggable approach. It allows to define and configure 
> multiple codec schemas with different coding algorithms and parameters. The 
> resultant codec schemas can be utilized and specified via command tool for 
> different file folders. While design and implement such pluggable framework, 
> it’s also to implement a concrete codec by default (Reed Solomon) to prove 
> the framework is useful and workable. Separate JIRA could be opened for the 
> RS codec implementation.
> Note HDFS-7353 will focus on the very low level codec API and implementation 
> to make concrete vendor libraries transparent to the upper layer. This JIRA 
> focuses on high level stuffs that interact with configuration, schema and etc.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to