[ https://issues.apache.org/jira/browse/FLINK-34440?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17817076#comment-17817076 ]
Kevin Lam commented on FLINK-34440: ----------------------------------- I just added this Jira issue, following the [Contribution Guide.|https://flink.apache.org/how-to-contribute/contribute-code/] If this ticket warrants a dev@ discussion, I'm happy to open one. I am happy to work on contributing the code to complete this issue. Looking forward to hearing others' thoughts! > Support Debezium Protobuf Confluent Format > ------------------------------------------ > > Key: FLINK-34440 > URL: https://issues.apache.org/jira/browse/FLINK-34440 > Project: Flink > Issue Type: New Feature > Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile) > Affects Versions: 1.19.0, 1.18.1 > Reporter: Kevin Lam > Priority: Minor > > *Motivation* > Debezium and the Confluent Schema registry can be used to emit Protobuf > Encoded messages to Kafka, but Flink does not easily support consuming these > messages through a connector. > *Definition of Done* > Add a format `debezium-protobuf-confluent` provided by > DebeziumProtobufFormatFactory that supports Debezium messages encoded using > Protocol Buffer and the Confluent Schema Registry. > To consider > * Mirror the implementation of the `debezium-avro-confluent` format. First > implement a `protobuf-confluent` format similar to the existing [Confluent > Avro|https://nightlies.apache.org/flink/flink-docs-release-1.18/docs/connectors/table/formats/avro-confluent/] > format that's provided today, which allows reading/writing protobuf using > the Confluent Schema Registry -- This message was sent by Atlassian Jira (v8.20.10#820010)