Hello Zelin,

  I pinged you on this PR: https://github.com/apache/paimon/pull/2070.
Checking in to see what release of paimon you plan to pick this item:
"support debezium avro format with schema registry".

Thanks,
Umesh

On Mon, Jan 22, 2024 at 8:55 AM umesh dangat <[email protected]> wrote:

> Thanks for the update Zelin.
>
> Currently, the intermediate records from Kafka source are string type. But
> for debezium-avro, the intermediate records are avro objects,
> This is indeed the case for nested avro records containing arrays, maps,
> avro records etc. There is already a TODO comment here
> <https://github.com/apache/incubator-paimon/blob/master/paimon-flink/paimon-flink-cdc/src/main/java/org/apache/paimon/flink/sink/cdc/CdcRecordUtils.java#L102>
>  that
> mentions that we need to either extend TypeUtils to handle such types or
> change CdcRecord.fields Map to not have String as values.  My branch in [2]
> took the former approach. Ofc I also needed to change the
> DebeziumAvroParser to handle such types (rather than convert them to
> String).
>
>  I will continue on Debezium-avro format in 0.8.0
> Thanks for working on this. I am fine with the debezium avro being
> available in 0.8. One thing that would be nice is if you can rebase branch
> [1] on master, then I can continue working off it in the meanwhile as the
> current branch [2] is based on [1] and it's quite diverted from master.
>
> Thanks,
> Umesh
>
>
>
>
> On Sun, Jan 21, 2024 at 8:43 PM yu zelin <[email protected]> wrote:
>
>> Hi Umesh,
>>
>> Recently I’m working on support Confluent debezium avro format
>> in Kafka cdc based on [1]. But the Paimon community is planning
>> to cut 0.7.0 release branch at Jan. 25th. And I think there is not enough
>> time for me to complete the job before the deadline for some reasons:
>>
>> 1. I have to modify the current CDC framework. Currently, the
>> intermediate
>> records from Kafka source are string type. But for debezium-avro, the
>> intermediate records are avro objects, so we have to adjust the framework.
>> It needs some time.
>>
>> 2. I noticed that you want to support some complex type in [2] which
>> made some changes to TypeUtils. Since this util is used by many
>> features, we should do some tests to see if the changes are compatible
>> with other features. I think if we implement a simple version in this
>> release
>> which doesn’t support those complex types, this release cannot meet
>> your situation. So I suggest that you continue to use the jar buit by
>> yourself.
>>
>> Recently I’m also woking on preparing to release 0.7.0. I will continue
>> on
>> Debezium-avro format in 0.8.0. If you have any problems with [1], welcome
>> to discuss with us in mailing list.
>>
>> Best,
>> Zelin Yu
>>
>> [1] https://github.com/apache/incubator-paimon/pull/2070
>> [2] https://github.com/harveyyue/incubator-paimon/pull/1
>>
>> 2024年1月10日 01:21,umesh dangat <[email protected]> 写道:
>>
>> Hello,
>>
>> I am a software engineer at Yelp Inc and lead the data infrastructure
>> group at Yelp. We have a complex real time streaming ecosystem comprising
>> flink, kafka and our custom schema registry service. I am trying to
>> evaluate Apache Paimon as a potential replacement for a lot of our data
>> pipelines, involving streaming reads, joins and aggregations to help
>> minimize our growing operational complexity and cost. Also paimon seems to
>> solve the schema evolution problem better than flink sqlclient? (which we
>> use currently)
>>
>> One issue with integrating paimon in our ecosystem seems to be that it
>> does
>> not support debezium avro format. Although Jingsong Li pointed me to this
>> <https://github.com/apache/incubator-paimon/pull/2070> branch that does
>> seem to add support for debezium avro format using confluent schema
>> registry. This would allow us to ingest our data from kafka into paimon
>> and
>> then evaluate it.
>>
>> I wanted to know if we have plans to push this branch to master soonish. I
>> can help with reviewing, since I plan to consume data written using this
>> format for some of our production workflows.
>>
>> Thanks,
>> Umesh
>>
>>
>>

Reply via email to