GenericRecord.toString produces invalid JSON for logical types

2018-08-23 Thread kant kodali
Hi All, I have a serialized avro binary data represented by byte[] where one of the field is long with a logical type of TimeStamp. Timestamp tsp = new Timestamp(1530228588182l); Schema schema = SchemaBuilder.builder() .record("hello") .fields()

Re: Messages are repeating in kafka

2017-05-24 Thread kant kodali
ask whether this should also be tuned or not. > > > > > > Regards, > > Abhimanyu > > > > > > > > > > On Mon, May 22, 2017 at 2:24 PM, kant kodali <kanth...@gmail.com> wrote: > > > >> @Abhimanyu Why do you think you need to set

Re: Messages are repeating in kafka

2017-05-22 Thread kant kodali
017 03:25, "Abhimanyu Nagrath" <abhimanyunagr...@gmail.com> > > wrote: > > > >> Hi Kant, > >> > >> Thanks for the suggestion. > >> > >> > >> Regards, > >> Abhimanyu > >> > >> On Sun, May 21, 2017 a

Re: Messages are repeating in kafka

2017-05-21 Thread kant kodali
@Abhimanyu You can try setting offset.retention = 30 (log.retention). At most, you will have a storage overhead of 5 million msgs per day * 30 (days) * 8 bytes (for each offset) = 1.2GB (not that much since you have a TB of hard disk) On Sun, May 21, 2017 at 3:05 AM, kant kodali <ka

Re: Messages are repeating in kafka

2017-05-21 Thread kant kodali
at 2:19 AM, kant kodali <kanth...@gmail.com> wrote: > What is your average message size and network speed? > > On Sun, May 21, 2017 at 2:04 AM, Abhimanyu Nagrath < > abhimanyunagr...@gmail.com> wrote: > >> Hi Girish, >> >> I did not set any value for

Re: Messages are repeating in kafka

2017-05-21 Thread kant kodali
What is your average message size and network speed? On Sun, May 21, 2017 at 2:04 AM, Abhimanyu Nagrath < abhimanyunagr...@gmail.com> wrote: > Hi Girish, > > I did not set any value for offsets.retention.minutes so therefore what I > think is picking its default value i.e 1440 minutes so what do

Re: is anyone able to create 1M or 10M or 100M or 1B partitions in a topic?

2017-05-16 Thread kant kodali
gt; https://www.confluent.io/blog/how-to-choose-the-number-of- > topicspartitions-in-a-kafka-cluster/ > > -Sameer. > > On Tue, May 16, 2017 at 2:40 PM, kant kodali <kanth...@gmail.com> wrote: > > > Forgot to mention: The question in this thread is for one node which has

Re: is anyone able to create 1M or 10M or 100M or 1B partitions in a topic?

2017-05-16 Thread kant kodali
Forgot to mention: The question in this thread is for one node which has 8 CPU's 16GB RAM & 500GB hard disk space. On Tue, May 16, 2017 at 2:06 AM, kant kodali <kanth...@gmail.com> wrote: > Hi All, > > 1. I was wondering if anyone has seen or heard or able to create 1M or

is anyone able to create 1M or 10M or 100M or 1B partitions in a topic?

2017-05-16 Thread kant kodali
Hi All, 1. I was wondering if anyone has seen or heard or able to create 1M or 10M or 100M or 1B partitions in a topic? I understand lot of this depends on filesystem limitations (we are using ext4) and the OS limitations but I just would like to know what is the scale one had seen in production?

Kafka Streams vs Flink (or any other stream processing framework)

2017-04-12 Thread kant kodali
Hi All, I have read enough blogs from Confluent and others and also books that tried to talk about the differences between the two and while it is great to know those differences I hardly find them any useful when it comes to decision making process of which one to pick since I don't see the

Re: Can multiple Kafka consumers read from the same partition of same topic by default?

2017-03-30 Thread kant kodali
, Matthias J. Sax <matth...@confluent.io> wrote: > Yes, you can do that. > > -Matthias > > > > On 3/30/17 6:09 PM, kant kodali wrote: > > Hi All, > > > > Can multiple Kafka consumers read from the same partition of same topic > by > > default? B

Can multiple Kafka consumers read from the same partition of same topic by default?

2017-03-30 Thread kant kodali
Hi All, Can multiple Kafka consumers read from the same partition of same topic by default? By default, I mean since group.id is not mandatory I am wondering if I spawn multiple kafka consumers without specifying any group.id and give them the same topic and partition name will they be able to

Stack Overflow Questions Jobs Documentation Tags Users Badges Ask Question Does the number of App instances and Zookeeper servers should be the same?

2017-01-25 Thread kant kodali
Does the number of App instances and Zookeeper servers should be the same? I understand the requirement of 2F+1 to tolerate F failures but this is to tolerate failures of Zookeeper instances itself. But how about the number of App instances ? For example say I have 3 zookeeper servers and I have 2

Re: How does Kafka emulate exactly once processing?

2016-12-21 Thread kant kodali
es. That way after a > crash, they can check the external store (HSFS, JSDC, etc) retrieve the > last committed offset and seek the the next message and continue processing > with no duplicates and exactly once semantics. > > -hans > > > > > > On Dec 21, 2016, at 4:

How does Kafka emulate exactly once processing?

2016-12-21 Thread kant kodali
How does Kafka emulate exactly once processing currently? Does it require the producer to send at least once and consumer to de dupe? I did do my research but I feel like I am going all over the place so a simple short answer would be great! Thanks!

Re: is there a way to make sure two consumers receive the same message from the broker?

2016-11-08 Thread kant kodali
ecord with higher writetime will win. > > Using es's document _version field is just one option, you can use > something from you domain or kafka's offset or machine timestamp (not > recommended at all). > > I hope it could help > > From: ka

Re: is there a way to make sure two consumers receive the same message from the broker?

2016-11-07 Thread kant kodali
And there is this https://github.com/vroyer/elassandra which is still under active development and not sure how they plan to keep up with Apache Cassandra moving forward. On Mon, Nov 7, 2016 at 9:36 AM, kant kodali <kanth...@gmail.com> wrote: > Fixing typo's > > Hi Tauzell, >

Re: is there a way to make sure two consumers receive the same message from the broker?

2016-11-07 Thread kant kodali
Fixing typo's Hi Tauzell, Yeah our users want to query, do aggregations on Elastic Search directly and we cannot have inconsistent data because say the writes didn't make it into Cassandra but made it to Elastic search then a simple aggregations like count will lead to a wrong answer but again

Re: is there a way to make sure two consumers receive the same message from the broker?

2016-11-07 Thread kant kodali
; same set of messages from Kafka ... I don't think this is possible. > > -Dave > > -Original Message- > From: kant kodali [mailto:kanth...@gmail.com] > Sent: Monday, November 7, 2016 10:48 AM > To: users@kafka.apache.org > Subject: Re: is there a way to make sure two con

Re: is there a way to make sure two consumers receive the same message from the broker?

2016-11-07 Thread kant kodali
e elasticsearch _version field as cassandra's > writetime?(_version is strictly increasing, cassandra uses writetime for > applying LWW, so last write in elasticsearch will always win) > > It needs no transaction and makes databases convergent. > > > _______

Re: is there a way to make sure two consumers receive the same message from the broker?

2016-11-07 Thread kant kodali
ngineer, Confluent Inc. > * h...@confluent.io (650)924-2670 > */ > > On Sun, Nov 6, 2016 at 7:08 PM, kant kodali <kanth...@gmail.com> wrote: > > > Hi Hans, > > > > The two storages we use are Cassandra and Elastic search and they are on > > the same datacenter f

Re: is there a way to make sure two consumers receive the same message from the broker?

2016-11-06 Thread kant kodali
Hi Hans, The two storages we use are Cassandra and Elastic search and they are on the same datacenter for now. The Programming Language we use is Java and OS would be Ubuntu or CentOS. We get messages in JSON format so we insert into Elastic Search directly and for Cassandra we transform JSON

Re: is there a way to make sure two consumers receive the same message from the broker?

2016-11-06 Thread kant kodali
Hi! Thanks. any pointers on how to do that? On Sun, Nov 6, 2016 at 2:32 PM, Tauzell, Dave <dave.tauz...@surescripts.com> wrote: > You should have one consumer pull the message and submit the data to each > storage using an XA transaction. > > > On Nov 5, 2016, at 19:4

Re: is there a way to make sure two consumers receive the same message from the broker?

2016-11-05 Thread kant kodali
here may be a better way to get you the functionality you want? > > -hans > > > > > > On Nov 5, 2016, at 4:31 PM, kant kodali <kanth...@gmail.com> wrote: > > > > I am new to Kafka and reading this statement "write consumer 1 and > consumer > > 2 to

Re: is there a way to make sure two consumers receive the same message from the broker?

2016-11-05 Thread kant kodali
2016 at 4:15 PM, kant kodali <kanth...@gmail.com> wrote: > Hi Hans, > > What do you mean by "write consumer 1 and consumer 2 to share a common > external offset storage" ? can you please elaborate a bit more. > > Thanks! > > On Sat, Nov 5, 2016 at

Re: is there a way to make sure two consumers receive the same message from the broker?

2016-11-05 Thread kant kodali
his in Apache Kafka but if you can > write consumer 1 and consumer 2 to share a common external offset storage > then you may be able to build the functionality you seek. > > -hans > > > > > On Nov 5, 2016, at 3:55 PM, kant kodali <kanth...@gmail.com> wrote:

Re: is there a way to make sure two consumers receive the same message from the broker?

2016-11-05 Thread kant kodali
Sorry there is a typo. here is a restatement. Is there a way to make sure two consumers receive the same message from the kafka broker in a atomic way? such that if consumer 1 gets a message consumer 2 should also get that message and if consumer 1 fails for whatever reason consumer 2 should

is there a way to make sure two consumers receive the same message from the broker?

2016-11-05 Thread kant kodali
is there a way to make sure two consumers receive the same message from the kafka broker in a atomic way? such that if consumer 1 gets a message consumer 2 should also get that message and in case one of the consumer fails for whatever reason consumer 2 should also rollback to previous offset or

Re: producer can't push msg sometimes with 1 broker recoved

2016-09-23 Thread kant kodali
@Fei Just curious why you guys are interested in using Kafka. I thought alcatel-lucent usually create their own software no? On Fri, Sep 23, 2016 10:36 PM, Kamal C kamaltar...@gmail.com wrote: Reduce the metadata refresh interval 'metadata.max.age.ms' from 5 min to your desired time

Re: why did Kafka choose pull instead of push for a consumer ?

2016-09-23 Thread kant kodali
to limit latency, and that may also be enough to get a low enough latency. It would certainly be interesting to compare the two, with the same hardware, and with high load. On Thu, Sep 22, 2016 at 6:01 PM kant kodali <kanth...@gmail.com> wrote: > @Gerard Thanks for this. It looks

Does Kafka Sync/persist every message from a publisher by default?

2016-09-22 Thread kant kodali
Does Kafka Sync/persist every message from a publisher by default? If not, What settings should I change so I Sync every message?

Re: why did Kafka choose pull instead of push for a consumer ?

2016-09-22 Thread kant kodali
real-time-datas-unifying > > > > – > > Best regards, > > Radek Gruchalski > > ra...@gruchalski.com > > > > > > On September 17, 2016 at 9:49:43 PM, kant kodali (kanth...@gmail.com) > wrote: > > > > Still it should be possible to

Re: any update on this?

2016-09-19 Thread kant kodali
independent of Zookeeper alltogether. Maybe not worth spending time on migrating to Consul in that case. Cheers, Jens On Sat, Sep 17, 2016 at 10:38 PM Jennifer Fountain <jfount...@meetme.com> wrote: > +2 watching. > > On Sat, Sep 17, 2016 at 2:45 AM, kant kodali <kanth..

Re: Kafka usecase

2016-09-18 Thread kant kodali
Why does comcast needs to do better than 1-2 seconds? On Sun, Sep 18, 2016 8:08 PM, Ghosh, Achintya (Contractor) achintya_gh...@comcast.com wrote: Hi there, We have an usecase where we do a lot of business logic to process each message and sometime it takes 1-2 sec, so will be Kafka

Re: why did Kafka choose pull instead of push for a consumer ?

2016-09-17 Thread kant kodali
your thoughts. On Sat, Sep 17, 2016 12:39 PM, Radoslaw Gruchalski ra...@gruchalski.com wrote: Kafka is not a queue. It’s a distributed commit log. – Best regards, Radek Gruchalski ra...@gruchalski.com On September 17, 2016 at 9:23:09 PM, kant kodali (kanth...@gmail.com) wrote

Re: why did Kafka choose pull instead of push for a consumer ?

2016-09-17 Thread kant kodali
regarding if this is the reason: Pull is much more sensible when a lot of data is pushed through. It allows consumers consuming at their own pace, slow consumers do not slow the complete system down. -- Best regards, Rad On Sat, Sep 17, 2016 at 11:18 AM +0200, "kant kodali&qu

why did Kafka choose pull instead of push for a consumer ?

2016-09-17 Thread kant kodali
why did Kafka choose pull instead of push for a consumer? push sounds like it is more realtime to me than poll and also wouldn't poll just keeps polling even when they are no messages in the broker causing more traffic? please enlighten me

Re: can one topic be registered in multiple brokers?

2016-09-17 Thread kant kodali
ker will be the leader of 1 of the 3 partitions. P.S how did the benchmarking go? On Sat, Sep 17, 2016 at 1:36 PM, kant kodali <kanth...@gmail.com> wrote: > can one topic be registered in multiple brokers? if so, which component of > kafka decides which broker should get the message for that particular > topic? > Thanks!

can one topic be registered in multiple brokers?

2016-09-17 Thread kant kodali
can one topic be registered in multiple brokers? if so, which component of kafka decides which broker should get the message for that particular topic? Thanks!

any update on this?

2016-09-17 Thread kant kodali
https://issues.apache.org/jira/browse/KAFKA-1793 It would be great to use Consul instead of Zookeeper for Kafka and I think it would benefit Kafka a lot from the exponentially growing consul community.

Re: which port should I use 9091 or 9092 or 2181 to send messages through kafka when using a client Library?

2016-09-15 Thread kant kodali
. On Thu, 15 Sep 2016 at 17:03 kant kodali <kanth...@gmail.com> wrote: > I haven't changed anything from > https://github.com/apache/kafka/blob/trunk/config/server.properties > and it looks like it is pointing to zookeeper. > Question: > Does producer client need to po

Re: which port should I use 9091 or 9092 or 2181 to send messages through kafka when using a client Library?

2016-09-15 Thread kant kodali
? On Thu, Sep 15, 2016 4:24 AM, Ali Akhtar ali.rac...@gmail.com wrote: Examine server.properties and see which port you're using in there On Thu, Sep 15, 2016 at 3:52 PM, kant kodali <kanth...@gmail.com> wrote: which port should I use 9091 or 9092 or 2181 to send messages through kafka

which port should I use 9091 or 9092 or 2181 to send messages through kafka when using a client Library?

2016-09-15 Thread kant kodali
which port should I use 9091 or 9092 or 2181 to send messages through kafka when using a client Library? I start kafka as follows: sudo bin/zookeeper-server-start.sh config/zookeeper.propertiessudo ./bin/kafka-server-start.sh config/server.properties and I dont see any process running on 9091 or

Re: What is the fair setup of Kafka to be comparable with NATS or NSQ?

2016-09-15 Thread kant kodali
that a go..but can you please confirm that one partition on one broker should be able to handle 300K messages of 1KB data size for each message? Thanks,kant On Thu, Sep 15, 2016 2:28 AM, kant kodali kanth...@gmail.com wrote: Hi Ben, I can give that a try but can you tell me the suspicion

Re: What is the fair setup of Kafka to be comparable with NATS or NSQ?

2016-09-15 Thread kant kodali
:16 AM, Ben Davison ben.davi...@7digital.com wrote: Hi Kant, I was following the other thread, can you try using a different benchmarking client for a test. https://grey-boundary.io/load-testing-apache-kafka-on-aws/ Ben On Thursday, 15 September 2016, kant kodali <kanth...@gmail.

What is the fair setup of Kafka to be comparable with NATS or NSQ?

2016-09-15 Thread kant kodali
with Kafka I tried it with 10 messages with single broker and only one partiton that looked instantaneous and ~5K messages/sec for the data size of 1KB I tried it with 1000 messages that looked instantaneous as well ~5K messages/sec for the data size of 1KBI tried it with 10K messages with single

Re: hi

2016-09-15 Thread kant kodali
. Here's a case study that should help: https://engineering.linkedin.com/kafka/benchmarking-apache-kafka-2-million-writes-second-three-cheap-machines Good luck, let us know how it goes On Thu, Sep 15, 2016 at 1:42 PM, kant kodali <kanth...@gmail.com> wrote: yeah.. I

Re: hi

2016-09-15 Thread kant kodali
com wrote: Lower the workload gradually, start from 10 messages, increase to 100, then 1000, and so on. See if it slows down as the workload increases. If so, you need more brokers + partitions to handle the workload. On Thu, Sep 15, 2016 at 12:42 PM, kant kodali <kanth...@gmail.com>

Re: hi

2016-09-15 Thread kant kodali
m4.xlarge On Thu, Sep 15, 2016 12:33 AM, Ali Akhtar ali.rac...@gmail.com wrote: What's the instance size that you're using? With 300k messages your single broker might not be able to handle it. On Thu, Sep 15, 2016 at 12:30 PM, kant kodali <kanth...@gmail.com> wrote: M

Re: hi

2016-09-15 Thread kant kodali
My goal is to test the throughput (#messages per second) given my setup and with a data size of 1KB. if you guys already have some idea on these numbers that would be helpful as well. On Thu, Sep 15, 2016 12:24 AM, kant kodali kanth...@gmail.com wrote: 172.* is all private ip's for my

Re: hi

2016-09-15 Thread kant kodali
-producer.sh instead of the Node code? On Thu, Sep 15, 2016 at 12:01 PM, kant kodali <kanth...@gmail.com> wrote: > They are hosted on AWS and I dont think there are any network issues > because I > tried testing other Queuing systems with no issues however I am using a >

Re: hi

2016-09-15 Thread kant kodali
tar ali.rac...@gmail.com wrote: It sounds like a network issue. Where are the 3 servers located / hosted? On Thu, Sep 15, 2016 at 11:51 AM, kant kodali <kanth...@gmail.com> wrote: Hi, I have the following setup. Single Kafka broker and Zookeeper on Machine 1single Kafka

hi

2016-09-15 Thread kant kodali
Hi, I have the following setup. Single Kafka broker and Zookeeper on Machine 1single Kafka producer on Machine 2 Single Kafka Consumer on Machine 3 When a producer client sends a message to the Kafka broker by pointing at the Zookeeper Server the consumer doesn't seem to get the message right

Consumer stops after reaching an offset of 1644

2016-09-14 Thread kant kodali
Hi All, I am trying to do a simple benchmark test  for Kafka using single broker, producer and consumer however my consumer doesn't seem to receive all the messages produced by the producer so not sure what is going on any help? Here is the full description of the problem.