[ https://issues.apache.org/jira/browse/FLINK-14108?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16939901#comment-16939901 ]
Lasse Nedergaard commented on FLINK-14108: ------------------------------------------- Sure. I'm on the road next week and the week after to Flink Forward. I will try to find the time for it. > Support for Confluent Kafka schema registry for Avro serialisation > ------------------------------------------------------------------- > > Key: FLINK-14108 > URL: https://issues.apache.org/jira/browse/FLINK-14108 > Project: Flink > Issue Type: New Feature > Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile) > Affects Versions: 1.10.0 > Reporter: Lasse Nedergaard > Assignee: Lasse Nedergaard > Priority: Minor > > The current implementation in flink-avro-confluent-registry support > deserialization with schema lookup in Confluent Kafka schema registry. > I would like support for serialization as well, following the same structure > as deserialization. With the feature it would be possible to use Confluent > schema registry in Sink writing Avro to Kafka and at the same time register > the schema used. > The test in TestAvroConsumerConfluent need to be updated together with the > comment as it indicate it use Confluent schema registry for write, but the > example code use SimpleStringSchema. > We have a running version, that we would like to give back to the community. > -- This message was sent by Atlassian Jira (v8.3.4#803005)