Re: Protobuf support with Flink SQL and Kafka Connector
Hi Shipeng, Matthias is correct. FLINK-18202 should address this topic. There is already a pull request there which is in good shape. You can also download the PR and build the format jar yourself, and then it should work with Flink 1.12. Best, Jark On Mon, 3 May 2021 at 21:41, Matthias Pohl wrote: > Hi Shipeng, > it looks like there is an open Jira issue FLINK-18202 [1] addressing this > topic. You might want to follow up on that one. I'm adding Timo and Jark to > this thread. They might have more insights. > > Best, > Matthias > > [1] https://issues.apache.org/jira/browse/FLINK-18202 > > On Sat, May 1, 2021 at 2:00 AM Fuyao Li wrote: > >> Hello Shipeng, >> >> >> >> I am not an expert in Flink, just want to share some of my thoughts. >> Maybe others can give you better ideas. >> >> I think there is no directly available Protobuf support for Flink SQL. >> However, you can write a user-defined format to support it [1]. >> >> If you use DataStream API, you can leverage Kryo Serializer to serialize >> and deserialize with Protobuf format. [2]. There is an out-of-box >> integration for Protobuf here. You will need to convert it to Flink SQL >> after data ingestion. >> >> >> >> [1] >> https://ci.apache.org/projects/flink/flink-docs-release-1.12/dev/table/sourceSinks.html#user-defined-sources-sinks >> >> [2] >> https://ci.apache.org/projects/flink/flink-docs-stable/dev/custom_serializers.html >> >> >> >> Best, >> >> Fuyao >> >> >> >> >> >> *From: *Shipeng Xie >> *Date: *Friday, April 30, 2021 at 14:58 >> *To: *user@flink.apache.org >> *Subject: *[External] : Protobuf support with Flink SQL and Kafka >> Connector >> >> Hi, >> >> >> >> In >> https://ci.apache.org/projects/flink/flink-docs-release-1.12/dev/table/connectors/formats/ >> <https://urldefense.com/v3/__https:/ci.apache.org/projects/flink/flink-docs-release-1.12/dev/table/connectors/formats/__;!!GqivPVa7Brio!P98PmXGjSsCo9MVOyHDZg39FQfaYhAKXwkO6xMz5STZfl95lJFWQDT7IOH2pa9w$>, >> it does not mention protobuf format. Does Flink SQL support protobuf >> format? If not, is there any plan to support it in the near future? >> >> Thanks! >> >
Re: Protobuf support with Flink SQL and Kafka Connector
Hi Shipeng, it looks like there is an open Jira issue FLINK-18202 [1] addressing this topic. You might want to follow up on that one. I'm adding Timo and Jark to this thread. They might have more insights. Best, Matthias [1] https://issues.apache.org/jira/browse/FLINK-18202 On Sat, May 1, 2021 at 2:00 AM Fuyao Li wrote: > Hello Shipeng, > > > > I am not an expert in Flink, just want to share some of my thoughts. Maybe > others can give you better ideas. > > I think there is no directly available Protobuf support for Flink SQL. > However, you can write a user-defined format to support it [1]. > > If you use DataStream API, you can leverage Kryo Serializer to serialize > and deserialize with Protobuf format. [2]. There is an out-of-box > integration for Protobuf here. You will need to convert it to Flink SQL > after data ingestion. > > > > [1] > https://ci.apache.org/projects/flink/flink-docs-release-1.12/dev/table/sourceSinks.html#user-defined-sources-sinks > > [2] > https://ci.apache.org/projects/flink/flink-docs-stable/dev/custom_serializers.html > > > > Best, > > Fuyao > > > > > > *From: *Shipeng Xie > *Date: *Friday, April 30, 2021 at 14:58 > *To: *user@flink.apache.org > *Subject: *[External] : Protobuf support with Flink SQL and Kafka > Connector > > Hi, > > > > In > https://ci.apache.org/projects/flink/flink-docs-release-1.12/dev/table/connectors/formats/ > <https://urldefense.com/v3/__https:/ci.apache.org/projects/flink/flink-docs-release-1.12/dev/table/connectors/formats/__;!!GqivPVa7Brio!P98PmXGjSsCo9MVOyHDZg39FQfaYhAKXwkO6xMz5STZfl95lJFWQDT7IOH2pa9w$>, > it does not mention protobuf format. Does Flink SQL support protobuf > format? If not, is there any plan to support it in the near future? > > Thanks! >
Re: Protobuf support with Flink SQL and Kafka Connector
Hello Shipeng, I am not an expert in Flink, just want to share some of my thoughts. Maybe others can give you better ideas. I think there is no directly available Protobuf support for Flink SQL. However, you can write a user-defined format to support it [1]. If you use DataStream API, you can leverage Kryo Serializer to serialize and deserialize with Protobuf format. [2]. There is an out-of-box integration for Protobuf here. You will need to convert it to Flink SQL after data ingestion. [1] https://ci.apache.org/projects/flink/flink-docs-release-1.12/dev/table/sourceSinks.html#user-defined-sources-sinks [2] https://ci.apache.org/projects/flink/flink-docs-stable/dev/custom_serializers.html Best, Fuyao From: Shipeng Xie Date: Friday, April 30, 2021 at 14:58 To: user@flink.apache.org Subject: [External] : Protobuf support with Flink SQL and Kafka Connector Hi, In https://ci.apache.org/projects/flink/flink-docs-release-1.12/dev/table/connectors/formats/<https://urldefense.com/v3/__https:/ci.apache.org/projects/flink/flink-docs-release-1.12/dev/table/connectors/formats/__;!!GqivPVa7Brio!P98PmXGjSsCo9MVOyHDZg39FQfaYhAKXwkO6xMz5STZfl95lJFWQDT7IOH2pa9w$>, it does not mention protobuf format. Does Flink SQL support protobuf format? If not, is there any plan to support it in the near future? Thanks!
Protobuf support with Flink SQL and Kafka Connector
Hi, In https://ci.apache.org/projects/flink/flink-docs-release-1.12/dev/table/connectors/formats/, it does not mention protobuf format. Does Flink SQL support protobuf format? If not, is there any plan to support it in the near future? Thanks!