The RC candidate build is here: 
http://home.apache.org/~gwenshap/kafka-0.10.2.1-rc1/ 
<http://home.apache.org/~gwenshap/kafka-0.10.2.1-rc1/>

Eno
> On 17 Apr 2017, at 17:20, Mahendra Kariya <mahendra.kar...@go-jek.com> wrote:
> 
> Thanks!
> 
> In the meantime, is the jar published somewhere on github or as a part of
> build pipeline?
> 
> On Mon, Apr 17, 2017 at 9:18 PM, Eno Thereska <eno.there...@gmail.com>
> wrote:
> 
>> Not yet, but as soon as 0.10.2 is voted it should be. Hopefully this week.
>> 
>> Eno
>>> On 17 Apr 2017, at 13:25, Mahendra Kariya <mahendra.kar...@go-jek.com>
>> wrote:
>>> 
>>> Are the bug fix releases published to Maven central repo?
>>> 
>>> On Sat, Apr 1, 2017 at 12:26 PM, Eno Thereska <eno.there...@gmail.com>
>>> wrote:
>>> 
>>>> Hi Sachin,
>>>> 
>>>> In the bug fix release for 0.10.2 (and in trunk) we have now set
>>>> max.poll.interval to infinite since from our experience with streams
>> this
>>>> should not be something that users set: https://github.com/apache/
>>>> kafka/pull/2770/files <https://github.com/apache/kafka/pull/2770/files
>>> .
>>>> 
>>>> We're in the process of documenting that change. For now you can
>> increase
>>>> the request timeout without worrying about max.poll.interval anymore. In
>>>> fact I'd suggest you also increase max.poll.interval as we've done it
>> above.
>>>> 
>>>> Thanks
>>>> Eno
>>>> 
>>>>> On 1 Apr 2017, at 03:28, Sachin Mittal <sjmit...@gmail.com> wrote:
>>>>> 
>>>>> Should this timeout be less than max poll interval value? if yes than
>>>>> generally speaking what should be the ratio between two or range for
>> this
>>>>> timeout value .
>>>>> 
>>>>> Thanks
>>>>> Sachin
>>>>> 
>>>>> On 1 Apr 2017 04:57, "Matthias J. Sax" <matth...@confluent.io> wrote:
>>>>> 
>>>>> Yes, you can increase ProducerConfig.REQUEST_TIMEOUT_MS_CONFIG
>>>>> 
>>>>> 
>>>>> -Matthias
>>>>> 
>>>>> 
>>>>> On 3/31/17 11:32 AM, Sachin Mittal wrote:
>>>>>> Hi,
>>>>>> So I have added the config ProducerConfig.RETRIES_CONFIG,
>>>>> Integer.MAX_VALUE
>>>>>> and the NotLeaderForPartitionException is gone.
>>>>>> 
>>>>>> However we see a new exception especially under heavy load:
>>>>>> org.apache.kafka.streams.errors.StreamsException: task [0_1]
>> exception
>>>>>> caught when producing
>>>>>> at
>>>>>> org.apache.kafka.streams.processor.internals.RecordCollectorImpl.
>>>>> checkForException(RecordCollectorImpl.java:119)
>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>> at
>>>>>> org.apache.kafka.streams.processor.internals.
>> RecordCollectorImpl.flush(
>>>>> RecordCollectorImpl.java:127)
>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
>>>>>> org.apache.kafka.streams.processor.internals.
>>>> StreamTask$1.run(StreamTask.
>>>>> java:76)
>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>> at
>>>>>> org.apache.kafka.streams.processor.internals.StreamsMetricsImpl.
>>>>> measureLatencyNs(StreamsMetricsImpl.java:188)
>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>> at
>>>>>> org.apache.kafka.streams.processor.internals.
>>>> StreamTask.commit(StreamTask.
>>>>> java:280)
>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
>>>>>> org.apache.kafka.streams.processor.internals.StreamThread.commitOne(
>>>>> StreamThread.java:787)
>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>> at
>>>>>> org.apache.kafka.streams.processor.internals.StreamThread.commitAll(
>>>>> StreamThread.java:774)
>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
>>>>>> org.apache.kafka.streams.processor.internals.
>> StreamThread.maybeCommit(
>>>>> StreamThread.java:749)
>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>> at
>>>>>> org.apache.kafka.streams.processor.internals.StreamThread.runLoop(
>>>>> StreamThread.java:671)
>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
>>>>>> org.apache.kafka.streams.processor.internals.
>>>>> StreamThread.run(StreamThread.java:378)
>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>> org.apache.kafka.common.errors.TimeoutException: Expiring 1 record(s)
>>>> for
>>>>>> new-part-advice-key-table-changelog-1: 30001 ms has passed since last
>>>>> append
>>>>>> 
>>>>>> So any idea as why TimeoutException is happening.
>>>>>> Is this controlled by
>>>>>> ProducerConfig.REQUEST_TIMEOUT_MS_CONFIG
>>>>>> 
>>>>>> If yes
>>>>>> What should the value be set in this given that out consumer
>>>>>> max.poll.interval.ms is defaul 5 minutes.
>>>>>> 
>>>>>> Is there any other setting that we should try to avoid such errors
>> which
>>>>>> causes stream thread to die.
>>>>>> 
>>>>>> Thanks
>>>>>> Sachin
>>>>>> 
>>>>>> 
>>>>>> On Sun, Mar 26, 2017 at 1:39 AM, Eno Thereska <eno.there...@gmail.com
>>> 
>>>>>> wrote:
>>>>>> 
>>>>>>> Hi Sachin,
>>>>>>> 
>>>>>>> Not in this case.
>>>>>>> 
>>>>>>> Thanks
>>>>>>> Eno
>>>>>>> 
>>>>>>>> On Mar 25, 2017, at 6:19 PM, Sachin Mittal <sjmit...@gmail.com>
>>>> wrote:
>>>>>>>> 
>>>>>>>> OK.
>>>>>>>> I will try this out.
>>>>>>>> 
>>>>>>>> Do I need to change anything for
>>>>>>>> max.in.flight.requests.per.connection
>>>>>>>> 
>>>>>>>> Thanks
>>>>>>>> Sachin
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On Sat, Mar 25, 2017 at 10:59 PM, Eno Thereska <
>>>> eno.there...@gmail.com>
>>>>>>>> wrote:
>>>>>>>> 
>>>>>>>>> Hi Sachin,
>>>>>>>>> 
>>>>>>>>> For this particular error, “org.apache.kafka.common.errors.
>>>>>>>>> NotLeaderForPartitionException: This server is not the leader for
>>>> that
>>>>>>>>> topic-partition.”, could you try setting the number of retries to
>>>>>>> something
>>>>>>>>> large like this:
>>>>>>>>> 
>>>>>>>>> Properties props = new Properties();
>>>>>>>>> props.put(StreamsConfig.APPLICATION_ID_CONFIG, applicationId);
>>>>>>>>> ...
>>>>>>>>> props.put(ProducerConfig.RETRIES_CONFIG, Integer.MAX_VALUE);
>>>>>>>>> 
>>>>>>>>> This will retry the produce requests and should hopefully solve
>> your
>>>>>>>>> immediate problem.
>>>>>>>>> 
>>>>>>>>> Thanks
>>>>>>>>> Eno
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> On 25/03/2017, 08:35, "Sachin Mittal" <sjmit...@gmail.com> wrote:
>>>>>>>>> 
>>>>>>>>> Hi,
>>>>>>>>> We have encountered another case of series of errors which I would
>>>>>>> need
>>>>>>>>> more help in understanding.
>>>>>>>>> 
>>>>>>>>> In logs we see message like this:
>>>>>>>>> ERROR 2017-03-25 03:41:40,001 [kafka-producer-network-thread |
>>>>>>>>> 85-StreamThread-3-producer]:
>>>>>>>>> org.apache.kafka.streams.processor.internals.RecordCollectorImpl
>> -
>>>>>>>>> task
>>>>>>>>> [0_1] Error sending record to topic new-part-advice-key-table-
>>>>>>> changelog.
>>>>>>>>> No
>>>>>>>>> more offsets will be recorded for this task and the exception will
>>>>>>>>> eventually be thrown
>>>>>>>>> 
>>>>>>>>> then some millisecond later
>>>>>>>>> ERROR 2017-03-25 03:41:40,149 [StreamThread-3]:
>>>>>>>>> org.apache.kafka.streams.processor.internals.StreamThread -
>>>>>>>>> stream-thread
>>>>>>>>> [StreamThread-3] Failed while executing StreamTask 0_1 due to
>> flush
>>>>>>>>> state:
>>>>>>>>> org.apache.kafka.streams.errors.StreamsException: task [0_1]
>>>>>>> exception
>>>>>>>>> caught when producing
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.RecordCollectorImpl.
>>>>>>>>> checkForException(RecordCollectorImpl.java:119)
>>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.
>>>>>>>>> RecordCollectorImpl.flush(RecordCollectorImpl.java:127)
>>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.
>>>> StreamTask.flushState(
>>>>>>>>> StreamTask.java:422)
>>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.
>> StreamThread$4.apply(
>>>>>>>>> StreamThread.java:555)
>>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.StreamThread.
>>>>>>>>> performOnAllTasks(StreamThread.java:513)
>>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.
>>>>>>>>> StreamThread.flushAllState(StreamThread.java:551)
>>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.StreamThread.
>>>>>>>>> shutdownTasksAndState(StreamThread.java:463)
>>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.
>>>> StreamThread.shutdown(
>>>>>>>>> StreamThread.java:408)
>>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.
>>>>>>>>> StreamThread.run(StreamThread.java:389)
>>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>> org.apache.kafka.common.errors.NotLeaderForPartitionException:
>>>> This
>>>>>>>>> server
>>>>>>>>> is not the leader for that topic-partition.
>>>>>>>>> 
>>>>>>>>> finally we get this
>>>>>>>>> ERROR 2017-03-25 03:41:45,724 [StreamThread-3]:
>>>>>>>>> com.advice.TestKafkaAdvice
>>>>>>>>> - Uncaught exception:
>>>>>>>>> org.apache.kafka.streams.errors.StreamsException: Exception
>> caught
>>>>>>> in
>>>>>>>>> process. taskId=0_1, processor=KSTREAM-SOURCE-0000000000,
>>>>>>>>> topic=advice-stream, partition=1, offset=48062286
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.
>>>>>>>>> StreamTask.process(StreamTask.java:216)
>>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.
>> StreamThread.runLoop(
>>>>>>>>> StreamThread.java:651)
>>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.
>>>>>>>>> StreamThread.run(StreamThread.java:378)
>>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>> Caused by: org.apache.kafka.streams.errors.StreamsException: task
>>>>>>>>> [0_1]
>>>>>>>>> exception caught when producing
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.RecordCollectorImpl.
>>>>>>>>> checkForException(RecordCollectorImpl.java:119)
>>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.
>>>>>>> RecordCollectorImpl.send(
>>>>>>>>> RecordCollectorImpl.java:76)
>>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>>     at
>>>>>>>>> org.apache.kafka.streams.processor.internals.
>>>>>>> RecordCollectorImpl.send(
>>>>>>>>> RecordCollectorImpl.java:64)
>>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> Again it is not clear why in this case we need to shut down the
>>>>>>> steams
>>>>>>>>> thread and eventually the application. Shouldn't we capture this
>>>>>>> error
>>>>>>>>> too?
>>>>>>>>> 
>>>>>>>>> Thanks
>>>>>>>>> Sachin
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>> 
>>>> 
>>>> 
>> 
>> 

Reply via email to