Kafka has quite a few tricks up its sleeve that can help implementing event
sourced systems ..

   - application state - one of the things that u may want to do in an
   event sourced system is manage and query the state of the application. If
   you use Kafka Streams, you get the full functionality of state management
   through local states backed up by persistent Kafka log-compacted topics. So
   you get full statefulness which is resilient as well.
   - for database like functionalities you can use Kafka Connect to store
   events in a real database. And enjoy queries from it.
   - and with 0.11 you get exactly-once processing with transactions as
   well. Though usually in Kafka you don't do updates, you add to the log and
   the state gets updated in the local store and in the backed up topic.

regards.

On Fri, Jul 21, 2017 at 7:21 PM, Chris Richardson <ch...@chrisrichardson.net
> wrote:

> Hi,
>
> I like Kafka but I don't understand the claim that it can be used for Event
> Sourcing (here <http://microservices.io/patterns/data/event-sourcing.html>
> and here <https://martinfowler.com/eaaDev/EventSourcing.html>)
>
> One part of the event sourcing is the ability to subscribe to events
> published by aggregates and clearly Kafka works well there.
>
> But the other part of Event Sourcing is "database" like functionality,
> which includes
>
>    - findEventsByPrimaryKey() - needed to be able to reconstruct an
>    aggregate from its events - the essence of event sourcing
>    - Atomic updates -  for updating aggregates  - findEventsByPrimaryKey()
>    - business logic - insertNewEvents()) in order to handle this kind of
>    scenario.
>
> The approach we have taken is to implement event sourcing using a database
> and Kafka.
> For instance: see
> https://blog.eventuate.io/2016/10/06/eventuate-local-
> event-sourcing-and-cqrs-with-spring-boot-apache-kafka-and-mysql/
>
> Chris
>
> --
> Learn microservices - http://learnmicroservices.io
> Microservices application platform http://eventuate.io
>
>
> On Fri, Jul 21, 2017 at 12:25 AM, José Antonio Iñigo <
> joseantonio.in...@gmail.com> wrote:
>
> > Hi everybody,
> >
> > I have been struggling with this problem for quite a while now, resorting
> > to stackoverflow
> > <https://stackoverflow.com/questions/45144429/event-
> > sourcing-apache-kafka-kafka-streams-how-to-assure-atomicity-transa>
> > for some help with no success. I am hoping to that here I'll get a more
> > authoritative answer from experienced Kafka users.
> >
> > This is the summary of my problem:
> >
> > - I am developing an application based on Spring Boot Microservices for a
> > shopping domain.
> > - I want to use Event Sourcing, having Kafka to register the events and
> > Kafka Streams API stores to materialize the views.
> > - To simplify the scenario we'll consider only two domains: Orders and
> > Products.
> > - The conflicting part:
> >    1) OrderService publishes an OrderPlaced event indicating a productId
> > and the quantity.
> >    2) ProductService consumes the event and queries (with an interactive
> > query) its Kafka Streams Store (ProductsStore) to check the availability
> of
> > the product. If there is availabilty it publishes a ProductReserved event
> > with productId and quantity:
> >
> > if("OrderPlaced".equals(event.get("eventType"))){
> >
> >     Order order = new Order();
> >     order.setId((String)event.get("orderId"));
> >     order.setProductId((Integer)(event.get("productId")));
> >     order.setUid(event.get("uid").toString());
> >
> >     // QUERY PRODUCTSTOCK TO CHECK AVAILABILITY
> >     Integer productStock = getProductStock(order.getProductId());
> >
> >     if(productStock > 0) {
> >         Map<String, Object> event = new HashMap<>();
> >         event.put("name", "ProductReserved");
> >         event.put("orderId", order.getId());
> >         event.put("productId", order.getProductId());
> >
> >         // WRITES A PRODUCT RESERVED EVENT TO orders topic
> >
> > orderProcessor.output().send(MessageBuilder.withPayload(event).build(),
> > 500);
> >     }else{
> >         //XXX CANCEL ORDER
> >     }
> > }
> >
> >    Then ProductService consumes its own event in a Kafka Streams
> processor
> > to update the stock of the product in the ProductsStore.
> >
> > KStream<Integer, JsonNode> stream = kStreamBuilder.stream(integerSerde,
> > jsonSerde, "orders");
> > stream.filter(...).groupByKey().reduce((...) -> {...}, "ProductsStock");
> >
> >    3.1) Suppose that in 1) two orders were placed simultaneously for the
> > same product and there is only stock for one of them
> >    3.2) ProductService would process the first one, the stock is ok and
> > would publish the ProductReserved event.
> >    3.3) We can't assure that ProductService will always process in the
> > Kafka Streams processor the order1 ProductReserved event to update
> > ProductsStore before the order2 OrderCreated is processed. Then in cases
> > ProductService will generate a ProductReserved for order2 incorrectly,
> > generating an inconsistency.
> >
> > IMPORTANT: You can find the detailed description, with code and the
> events
> > that are published and consumed in the previously referenced
> stackoverflow
> > question.
> >
> > After so much thinking and looking up online I haven't found a single
> place
> > where I can get a clear way to deal with Event Sourcing with Kafka+Kafka
> > Streams solving the problem of atomicity.
> >
> > I'd really appreciate if someone could propose a solution for this.
> >
> > Regards
> > Jose
> >
>



-- 
Debasish Ghosh
http://manning.com/ghosh2
http://manning.com/ghosh

Twttr: @debasishg
Blog: http://debasishg.blogspot.com
Code: http://github.com/debasishg

Reply via email to