Hello,

One of the big attractions of Kafka for me was the ability to write new
consumers of topics that would then be able to connect to a topic and
replay all the previous events.

However, most of the time, Kafka appears to be used with a retention period
- presumably in such cases, the events have been warehoused into HDFS
or something similar.

So my question is - how do people typically approach the scenario where a
new piece of code needs to process all events in a topic from "day one",
but has to source some of them from e.g HDFS and then connect to the
real-time Kafka topic?  Are there any wrinkles with such an approach?

Thanks,
Ted

Reply via email to