Are the bug fix releases published to Maven central repo?

On Sat, Apr 1, 2017 at 12:26 PM, Eno Thereska <eno.there...@gmail.com>
wrote:

> Hi Sachin,
>
> In the bug fix release for 0.10.2 (and in trunk) we have now set
> max.poll.interval to infinite since from our experience with streams this
> should not be something that users set: https://github.com/apache/
> kafka/pull/2770/files <https://github.com/apache/kafka/pull/2770/files>.
>
> We're in the process of documenting that change. For now you can increase
> the request timeout without worrying about max.poll.interval anymore. In
> fact I'd suggest you also increase max.poll.interval as we've done it above.
>
> Thanks
> Eno
>
> > On 1 Apr 2017, at 03:28, Sachin Mittal <sjmit...@gmail.com> wrote:
> >
> > Should this timeout be less than max poll interval value? if yes than
> > generally speaking what should be the ratio between two or range for this
> > timeout value .
> >
> > Thanks
> > Sachin
> >
> > On 1 Apr 2017 04:57, "Matthias J. Sax" <matth...@confluent.io> wrote:
> >
> > Yes, you can increase ProducerConfig.REQUEST_TIMEOUT_MS_CONFIG
> >
> >
> > -Matthias
> >
> >
> > On 3/31/17 11:32 AM, Sachin Mittal wrote:
> >> Hi,
> >> So I have added the config ProducerConfig.RETRIES_CONFIG,
> > Integer.MAX_VALUE
> >> and the NotLeaderForPartitionException is gone.
> >>
> >> However we see a new exception especially under heavy load:
> >> org.apache.kafka.streams.errors.StreamsException: task [0_1] exception
> >> caught when producing
> >>  at
> >> org.apache.kafka.streams.processor.internals.RecordCollectorImpl.
> > checkForException(RecordCollectorImpl.java:119)
> >> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>  at
> >> org.apache.kafka.streams.processor.internals.RecordCollectorImpl.flush(
> > RecordCollectorImpl.java:127)
> >> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
> >> org.apache.kafka.streams.processor.internals.
> StreamTask$1.run(StreamTask.
> > java:76)
> >> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>  at
> >> org.apache.kafka.streams.processor.internals.StreamsMetricsImpl.
> > measureLatencyNs(StreamsMetricsImpl.java:188)
> >> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>  at
> >> org.apache.kafka.streams.processor.internals.
> StreamTask.commit(StreamTask.
> > java:280)
> >> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
> >> org.apache.kafka.streams.processor.internals.StreamThread.commitOne(
> > StreamThread.java:787)
> >> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>  at
> >> org.apache.kafka.streams.processor.internals.StreamThread.commitAll(
> > StreamThread.java:774)
> >> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
> >> org.apache.kafka.streams.processor.internals.StreamThread.maybeCommit(
> > StreamThread.java:749)
> >> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>  at
> >> org.apache.kafka.streams.processor.internals.StreamThread.runLoop(
> > StreamThread.java:671)
> >> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
> >> org.apache.kafka.streams.processor.internals.
> > StreamThread.run(StreamThread.java:378)
> >> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >> org.apache.kafka.common.errors.TimeoutException: Expiring 1 record(s)
> for
> >> new-part-advice-key-table-changelog-1: 30001 ms has passed since last
> > append
> >>
> >> So any idea as why TimeoutException is happening.
> >> Is this controlled by
> >> ProducerConfig.REQUEST_TIMEOUT_MS_CONFIG
> >>
> >> If yes
> >> What should the value be set in this given that out consumer
> >> max.poll.interval.ms is defaul 5 minutes.
> >>
> >> Is there any other setting that we should try to avoid such errors which
> >> causes stream thread to die.
> >>
> >> Thanks
> >> Sachin
> >>
> >>
> >> On Sun, Mar 26, 2017 at 1:39 AM, Eno Thereska <eno.there...@gmail.com>
> >> wrote:
> >>
> >>> Hi Sachin,
> >>>
> >>> Not in this case.
> >>>
> >>> Thanks
> >>> Eno
> >>>
> >>>> On Mar 25, 2017, at 6:19 PM, Sachin Mittal <sjmit...@gmail.com>
> wrote:
> >>>>
> >>>> OK.
> >>>> I will try this out.
> >>>>
> >>>> Do I need to change anything for
> >>>> max.in.flight.requests.per.connection
> >>>>
> >>>> Thanks
> >>>> Sachin
> >>>>
> >>>>
> >>>> On Sat, Mar 25, 2017 at 10:59 PM, Eno Thereska <
> eno.there...@gmail.com>
> >>>> wrote:
> >>>>
> >>>>> Hi Sachin,
> >>>>>
> >>>>> For this particular error, “org.apache.kafka.common.errors.
> >>>>> NotLeaderForPartitionException: This server is not the leader for
> that
> >>>>> topic-partition.”, could you try setting the number of retries to
> >>> something
> >>>>> large like this:
> >>>>>
> >>>>> Properties props = new Properties();
> >>>>> props.put(StreamsConfig.APPLICATION_ID_CONFIG, applicationId);
> >>>>> ...
> >>>>> props.put(ProducerConfig.RETRIES_CONFIG, Integer.MAX_VALUE);
> >>>>>
> >>>>> This will retry the produce requests and should hopefully solve your
> >>>>> immediate problem.
> >>>>>
> >>>>> Thanks
> >>>>> Eno
> >>>>>
> >>>>>
> >>>>> On 25/03/2017, 08:35, "Sachin Mittal" <sjmit...@gmail.com> wrote:
> >>>>>
> >>>>>   Hi,
> >>>>>   We have encountered another case of series of errors which I would
> >>> need
> >>>>>   more help in understanding.
> >>>>>
> >>>>>   In logs we see message like this:
> >>>>>   ERROR 2017-03-25 03:41:40,001 [kafka-producer-network-thread |
> >>>>>   85-StreamThread-3-producer]:
> >>>>>   org.apache.kafka.streams.processor.internals.RecordCollectorImpl -
> >>>>> task
> >>>>>   [0_1] Error sending record to topic new-part-advice-key-table-
> >>> changelog.
> >>>>> No
> >>>>>   more offsets will be recorded for this task and the exception will
> >>>>>   eventually be thrown
> >>>>>
> >>>>>   then some millisecond later
> >>>>>   ERROR 2017-03-25 03:41:40,149 [StreamThread-3]:
> >>>>>   org.apache.kafka.streams.processor.internals.StreamThread -
> >>>>> stream-thread
> >>>>>   [StreamThread-3] Failed while executing StreamTask 0_1 due to flush
> >>>>> state:
> >>>>>   org.apache.kafka.streams.errors.StreamsException: task [0_1]
> >>> exception
> >>>>>   caught when producing
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.RecordCollectorImpl.
> >>>>> checkForException(RecordCollectorImpl.java:119)
> >>>>>   ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.
> >>>>> RecordCollectorImpl.flush(RecordCollectorImpl.java:127)
> >>>>>   ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.
> StreamTask.flushState(
> >>>>> StreamTask.java:422)
> >>>>>   ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.StreamThread$4.apply(
> >>>>> StreamThread.java:555)
> >>>>>   ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.StreamThread.
> >>>>> performOnAllTasks(StreamThread.java:513)
> >>>>>   [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.
> >>>>> StreamThread.flushAllState(StreamThread.java:551)
> >>>>>   [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.StreamThread.
> >>>>> shutdownTasksAndState(StreamThread.java:463)
> >>>>>   [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.
> StreamThread.shutdown(
> >>>>> StreamThread.java:408)
> >>>>>   [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.
> >>>>> StreamThread.run(StreamThread.java:389)
> >>>>>   [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>   org.apache.kafka.common.errors.NotLeaderForPartitionException:
> This
> >>>>> server
> >>>>>   is not the leader for that topic-partition.
> >>>>>
> >>>>>   finally we get this
> >>>>>   ERROR 2017-03-25 03:41:45,724 [StreamThread-3]:
> >>>>> com.advice.TestKafkaAdvice
> >>>>>   - Uncaught exception:
> >>>>>   org.apache.kafka.streams.errors.StreamsException: Exception caught
> >>> in
> >>>>>   process. taskId=0_1, processor=KSTREAM-SOURCE-0000000000,
> >>>>>   topic=advice-stream, partition=1, offset=48062286
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.
> >>>>> StreamTask.process(StreamTask.java:216)
> >>>>>   ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.StreamThread.runLoop(
> >>>>> StreamThread.java:651)
> >>>>>   ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.
> >>>>> StreamThread.run(StreamThread.java:378)
> >>>>>   ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>   Caused by: org.apache.kafka.streams.errors.StreamsException: task
> >>>>> [0_1]
> >>>>>   exception caught when producing
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.RecordCollectorImpl.
> >>>>> checkForException(RecordCollectorImpl.java:119)
> >>>>>   ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.
> >>> RecordCollectorImpl.send(
> >>>>> RecordCollectorImpl.java:76)
> >>>>>   ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>       at
> >>>>>   org.apache.kafka.streams.processor.internals.
> >>> RecordCollectorImpl.send(
> >>>>> RecordCollectorImpl.java:64)
> >>>>>   ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>
> >>>>>
> >>>>>   Again it is not clear why in this case we need to shut down the
> >>> steams
> >>>>>   thread and eventually the application. Shouldn't we capture this
> >>> error
> >>>>> too?
> >>>>>
> >>>>>   Thanks
> >>>>>   Sachin
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>
> >>>
> >>
>
>

Reply via email to