Prasanna,
You might want to check the kafka broker configs where
'auto.create.topics.enable' helps with creating a new topic whenever a new
message with non existent topic is published.
https://kafka.apache.org/documentation/#brokerconfigs

I am not too sure about pitfalls if any.

On Mon, Jun 1, 2020 at 3:20 PM Leonard Xu <xbjt...@gmail.com> wrote:

> Hi, kumar
>
> Sorry for missed the original question, I think we can not create topic
> dynamically current, creating topic should belong to control flow rather a
> data flow, and user may has some custom configurations of the topic from my
> understanding. Maybe you need implement the logic of check/create/manage
> topic in your custom SinkFunction so that the topic can create dynamically
> in runtime.
>
> Best,
> Leonard Xu
>
> 在 2020年6月1日,17:02,Prasanna kumar <prasannakumarram...@gmail.com> 写道:
>
> Leaonard,
>
> Thanks for the reply and would look into those options.
> But as for the original question, could we create a topic dynamically when
> required .
>
> Prasanna.
>
> On Mon, Jun 1, 2020 at 2:18 PM Leonard Xu <xbjt...@gmail.com> wrote:
>
>> Hi, kumar
>>
>> Flink support consume/produce from/to multiple kafka topics[1], in your
>> case you can implement KeyedSerializationSchema(legacy interface) or
>> KafkaSerializationSchema[2] to make one producer instance support send data
>> to multiple topics. There is an ITCase you can reference[3].
>>
>>
>> Best,
>> Leonard Xu
>>
>> [1]
>> https://ci.apache.org/projects/flink/flink-docs-master/dev/connectors/kafka.html#kafka-producer
>> [2]
>> https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-kafka-base/src/main/java/org/apache/flink/streaming/connectors/kafka/KafkaSerializationSchema.java
>>
>> [3]
>> https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-kafka-0.11/src/test/java/org/apache/flink/streaming/connectors/kafka/Kafka011ITCase.java#L126
>>
>>
>> 在 2020年6月1日,15:35,Prasanna kumar <prasannakumarram...@gmail.com> 写道:
>>
>> Hi,
>>
>> I have Use Case where i read events from a Single kafka Stream comprising
>> of JSON messages.
>>
>> Requirement is to split the stream into multiple output streams based on
>> some criteria say based on Type of Event or Based on Type and Customer
>> associated with the event.
>>
>> We could achieve the splitting of stream using Side outputs as i have
>> seen in the documentation.
>>
>> Our business environment is such that there could be new event types
>> flowing in and would the Flink Kafka producer create the topics dynamically
>> based on the inflowing events. I did not see any documentation saying
>> that it could create.
>>
>> Or should it be always pre created by running a script separately. (Not a
>> good scalable practice in our case)
>>
>> Thanks,
>> Prasanna.
>>
>>
>>
>

Reply via email to