Yeah, I've seen the same issue.  It appears that the storm-kafka-client in
versions < 1.1 has significant throughput problems.  We saw a 10x speedup
in moving to the 1.1 version.  There is a PR out for this currently:
https://github.com/apache/metron/pull/584

Casey

On Tue, May 16, 2017 at 4:26 AM, Ali Nazemian <alinazem...@gmail.com> wrote:

> I am still facing this issue and couldn't manage to fix it. I would be
> really grateful If somebody can help me.
>
> Thanks,
> Ali
>
> On Sun, May 14, 2017 at 1:58 PM, Ali Nazemian <alinazem...@gmail.com>
> wrote:
>
>> I was wrong. I think I couldn't increase the timeout value for Kafka
>> spout properly. Therefore, I was wondering how I can increase the timeout
>> value for Kafka spout? What is the right "-esc" property name I need to set
>> in this case? Also, what has changed in the newer version since I didn't
>> have this issue with the previous version?
>>
>>
>>
>> On Sun, May 14, 2017 at 3:00 AM, Ali Nazemian <alinazem...@gmail.com>
>> wrote:
>>
>>> Hi,
>>>
>>> I have installed the new version of HCP recently. I can see that the
>>> following error has appeared in Storm UI at Kafka spout section related to
>>> Parser topologies:
>>>
>>> org.apache.kafka.clients.consumer.CommitFailedException: Commit cannot
>>> be completed since the group has already rebalanced and assigned the
>>> partitions to another member. This means that the time between subsequent
>>> calls to poll() was longer than the configured session.timeout.ms,
>>> which typically implies that the poll loop is spending too much time
>>> message processing. You can address this either by increasing the session
>>> timeout or by reducing the maximum size of batches returned in poll() with
>>> max.poll.records. at org.apache.kafka.clients.consu
>>> mer.internals.ConsumerCoordinator$OffsetCommitResponseHandle
>>> r.handle(ConsumerCoordinator.java:600) at org.apache.kafka.clients.consu
>>> mer.internals.ConsumerCoordinator$OffsetCommitResponseHandle
>>> r.handle(ConsumerCoordinator.java:541) at org.apache.kafka.clients.consu
>>> mer.internals.AbstractCoordinator$CoordinatorResponseHandler
>>> .onSuccess(AbstractCoordinator.java:679) at
>>> org.apache.kafka.clients.consumer.internals.AbstractCoordina
>>> tor$CoordinatorResponseHandler.onSuccess(AbstractCoordinator.java:658)
>>> at org.apache.kafka.clients.consumer.internals.RequestFuture$1.
>>> onSuccess(RequestFuture.java:167) at org.apache.kafka.clients.consu
>>> mer.internals.RequestFuture.fireSuccess(RequestFuture.java:133) at
>>> org.apache.kafka.clients.consumer.internals.RequestFuture.complete(RequestFuture.java:107)
>>> at org.apache.kafka.clients.consumer.internals.ConsumerNetworkC
>>> lient$RequestFutureCompletionHandler.onComplete(ConsumerNetworkClient.java:426)
>>> at org.apache.kafka.clients.NetworkClient.poll(NetworkClient.java:278)
>>> at org.apache.kafka.clients.consumer.internals.ConsumerNetworkC
>>> lient.clientPoll(ConsumerNetworkClient.java:360) at
>>> org.apache.kafka.clients.consumer.internals.ConsumerNetworkC
>>> lient.poll(ConsumerNetworkClient.java:224) at
>>> org.apache.kafka.clients.consumer.internals.ConsumerNetworkC
>>> lient.poll(ConsumerNetworkClient.java:192) at
>>> org.apache.kafka.clients.consumer.internals.ConsumerNetworkC
>>> lient.poll(ConsumerNetworkClient.java:163) at
>>> org.apache.kafka.clients.consumer.internals.ConsumerCoordina
>>> tor.commitOffsetsSync(ConsumerCoordinator.java:426) at
>>> org.apache.kafka.clients.consumer.KafkaConsumer.commitSync(KafkaConsumer.java:1059)
>>> at 
>>> org.apache.storm.kafka.spout.KafkaSpout.commitOffsetsForAckedTuples(KafkaSpout.java:302)
>>> at org.apache.storm.kafka.spout.KafkaSpout.nextTuple(KafkaSpout.java:204)
>>> at 
>>> org.apache.storm.daemon.executor$fn__6505$fn__6520$fn__6551.invoke(executor.clj:651)
>>> at org.apache.storm.util$async_loop$fn__554.invoke(util.clj:484) at
>>> clojure.lang.AFn.run(AFn.java:22) at java.lang.Thread.run(Thread.ja
>>> va:748)
>>>
>>>
>>> This error has affected the Parsers throughput significantly!
>>>
>>> I have tried to increase the session timeout, but It didn't affect my
>>> situation. I would be grateful if you can help me to find the source of
>>> this issue. Please be advised that I haven't had this issue with the
>>> previous version of Metron (0.3.1).
>>>
>>> Regards,
>>> Ali
>>>
>>>
>>
>>
>> --
>> A.Nazemian
>>
>
>
>
> --
> A.Nazemian
>

Reply via email to