@Jay, I currently use reactive-kaka for my Kafka sources and sinks in my stream processing apps. I was interested to see if this new stream API would make that setup easier/simpler/better in the future when it becomes available.
How does the Streams API handle the commit offsets? Since you are processing "1-at-a-time”, is it auto magic on commit handling at the beginning/end of the processing or can we specify where in the processing an offset commit happens? Thanks, DaVe. David Buschman d...@timeli.io > On Mar 11, 2016, at 7:21 AM, Dick Davies <d...@hellooperator.net> wrote: > > Nice - I've read topics on the idea of a database as the 'now' view of a > stream > of updates, it's a very powerful concept. > > Reminds me of Rich Hickeys talk on DAtomic, if anyone's seen that. > > > > On 10 March 2016 at 21:26, Jay Kreps <j...@confluent.io> wrote: >> Hey all, >> >> Lot's of people have probably seen the ongoing work on Kafka Streams >> happening. There is no real way to design a system like this in a vacuum, >> so we put up a blog, some snapshot docs, and something you can download and >> use easily to get feedback: >> >> http://www.confluent.io/blog/introducing-kafka-streams-stream-processing-made-simple >> >> We'd love comments or thoughts from anyone... >> >> -Jay