@Jay, I currently use reactive-kaka for my Kafka sources and sinks in my stream
processing apps. I was interested to see if this new stream API would make that
setup easier/simpler/better in the future when it becomes available.
How does the Streams API handle the commit offsets? Since you are processing
"1-at-a-time”, is it auto magic on commit handling at the beginning/end of the
processing or can we specify where in the processing an offset commit happens?
Thanks,
DaVe.
David Buschman
[email protected]
> On Mar 11, 2016, at 7:21 AM, Dick Davies <[email protected]> wrote:
>
> Nice - I've read topics on the idea of a database as the 'now' view of a
> stream
> of updates, it's a very powerful concept.
>
> Reminds me of Rich Hickeys talk on DAtomic, if anyone's seen that.
>
>
>
> On 10 March 2016 at 21:26, Jay Kreps <[email protected]> wrote:
>> Hey all,
>>
>> Lot's of people have probably seen the ongoing work on Kafka Streams
>> happening. There is no real way to design a system like this in a vacuum,
>> so we put up a blog, some snapshot docs, and something you can download and
>> use easily to get feedback:
>>
>> http://www.confluent.io/blog/introducing-kafka-streams-stream-processing-made-simple
>>
>> We'd love comments or thoughts from anyone...
>>
>> -Jay