Leaonard,

Thanks for the reply and would look into those options.
But as for the original question, could we create a topic dynamically when
required .

Prasanna.

On Mon, Jun 1, 2020 at 2:18 PM Leonard Xu <xbjt...@gmail.com> wrote:

> Hi, kumar
>
> Flink support consume/produce from/to multiple kafka topics[1], in your
> case you can implement KeyedSerializationSchema(legacy interface) or
> KafkaSerializationSchema[2] to make one producer instance support send data
> to multiple topics. There is an ITCase you can reference[3].
>
>
> Best,
> Leonard Xu
>
> [1]
> https://ci.apache.org/projects/flink/flink-docs-master/dev/connectors/kafka.html#kafka-producer
> [2]
> https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-kafka-base/src/main/java/org/apache/flink/streaming/connectors/kafka/KafkaSerializationSchema.java
>
> [3]
> https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-kafka-0.11/src/test/java/org/apache/flink/streaming/connectors/kafka/Kafka011ITCase.java#L126
>
>
> 在 2020年6月1日,15:35,Prasanna kumar <prasannakumarram...@gmail.com> 写道:
>
> Hi,
>
> I have Use Case where i read events from a Single kafka Stream comprising
> of JSON messages.
>
> Requirement is to split the stream into multiple output streams based on
> some criteria say based on Type of Event or Based on Type and Customer
> associated with the event.
>
> We could achieve the splitting of stream using Side outputs as i have seen
> in the documentation.
>
> Our business environment is such that there could be new event types
> flowing in and would the Flink Kafka producer create the topics dynamically
> based on the inflowing events. I did not see any documentation saying
> that it could create.
>
> Or should it be always pre created by running a script separately. (Not a
> good scalable practice in our case)
>
> Thanks,
> Prasanna.
>
>
>

Reply via email to