Just to add, we target that for 1.14.

However, it's also not too complicated to add a new TableFactory that uses
the new sources (or your source).

On Thu, Jun 3, 2021 at 10:04 AM Chesnay Schepler <ches...@apache.org> wrote:

> The FLIP-27 were primarily aimed at the DataStream API; the integration
> into the SQL/Table APIs will happen at a later date.
>
> On 6/1/2021 5:59 PM, deepthi Sridharan wrote:
>
> Thank you, Roman. I should have said our own flavor of Kafka and not
> version. Thanks for the reference of the new source and sink
> interfaces, though, as it seems like the interfaces we should be
> implementing to use our custom Kafka connector.
>
> I did notice however that the FLIP does not cover table interfaces. The
> KafkaDynamicTableFactory for example is still creating a FlinkKafkaConsumer
> instance. Is that something that will change in the future or are the table
> interfaces somehow exceptions to the advantages of the new interface?
>
> --
> Regards,
> Deepthi
>
> On Thu, May 20, 2021 at 12:23 PM Roman Khachatryan <ro...@apache.org>
> wrote:
>
>> Hi,
>>
>> Those classes will likely be deprecated in the future in favor of
>> FLIP-27 [1][2] source and FLIP-143 [3] sink implementations and
>> eventually removed (though it won't happen soon).
>> You probably should take a look at the above new APIs.
>>
>> Either way, there is no such a recommendation AFAIK. Copied connector
>> classes will have to be updated if something in Flink changes. Maybe a
>> better way would be to build your own kafka client and use it to build
>> flink-kafka connector (by overriding ${kafka.version} for example).
>>
>> [1]
>>
>> https://cwiki.apache.org/confluence/display/FLINK/FLIP-27%3A+Refactor+Source+Interface
>> [2]
>> https://issues.apache.org/jira/browse/FLINK-18323
>> [3]
>>
>> https://cwiki.apache.org/confluence/display/FLINK/FLIP-143%3A+Unified+Sink+API
>>
>> Regards,
>> Roman
>>
>> On Thu, May 20, 2021 at 7:45 PM deepthi Sridharan
>> <deepthi.sridha...@gmail.com> wrote:
>> >
>> > Hi,
>> >
>> > We have an internal version of Open source Kafka consumer and producer
>> that we use and are working on adding that as a source and sink for flink.
>> >
>> > It seems like the easiest way to add the consumer as source would be to
>> override the FlinkKafkaConsumer class's createFetcher method to provide our
>> own derived class of KafkaFetcher class which can hookup its own version of
>> the consumerThread. But the fetcher classes are annotated as Internal and
>> seems like it is not meant to be used this way. (And the changes for
>> Producer would be on similar lines).
>> >
>> > Is there a recommendation for how to add new flavors of Kafka
>> Consumer/Producer from the community? Would it be recommended to maintain a
>> copy of all the connector classes so we don't have to deal with changes to
>> classes tagged as internal?
>> >
>> > --
>> > Thanks & Regards
>> >
>>
>
>

Reply via email to