Hi Gabor,

I guess you are looking at Kafka 2.1 but Guillermo mentioned initially that
they are working with Kafka 1.0

Akshay Bhardwaj
+91-97111-33849


On Wed, Feb 27, 2019 at 5:41 PM Gabor Somogyi <gabor.g.somo...@gmail.com>
wrote:

> Where exactly? In Kafka broker configuration section here it's 10080:
> https://kafka.apache.org/documentation/
>
> offsets.retention.minutes After a consumer group loses all its consumers
> (i.e. becomes empty) its offsets will be kept for this retention period
> before getting discarded. For standalone consumers (using manual
> assignment), offsets will be expired after the time of last commit plus
> this retention period. int 10080 [1,...] high read-only
>
> On Wed, Feb 27, 2019 at 1:04 PM Guillermo Ortiz <konstt2...@gmail.com>
> wrote:
>
>> I'm going to check the value, but I didn't change it., normally, the
>> process is always running but sometimes I have to restarted to apply some
>> changes. Sometimes it starts from the beginning and others continue for the
>> last offset.
>>
>> El mié., 27 feb. 2019 a las 12:25, Akshay Bhardwaj (<
>> akshay.bhardwaj1...@gmail.com>) escribió:
>>
>>> Hi Gabor,
>>>
>>> I am talking about offset.retention.minutes which is set default as
>>> 1440 (or 24 hours)
>>>
>>> Akshay Bhardwaj
>>> +91-97111-33849
>>>
>>>
>>> On Wed, Feb 27, 2019 at 4:47 PM Gabor Somogyi <gabor.g.somo...@gmail.com>
>>> wrote:
>>>
>>>> Hi Akshay,
>>>>
>>>> The feature what you've mentioned has a default value of 7 days...
>>>>
>>>> BR,
>>>> G
>>>>
>>>>
>>>> On Wed, Feb 27, 2019 at 7:38 AM Akshay Bhardwaj <
>>>> akshay.bhardwaj1...@gmail.com> wrote:
>>>>
>>>>> Hi Guillermo,
>>>>>
>>>>> What was the interval in between restarting the spark job? As a
>>>>> feature in Kafka, a broker deleted offsets for a consumer group after
>>>>> inactivity of 24 hours.
>>>>> In such a case, the newly started spark streaming job will read
>>>>> offsets from beginning for the same groupId.
>>>>>
>>>>> Akshay Bhardwaj
>>>>> +91-97111-33849
>>>>>
>>>>>
>>>>> On Thu, Feb 21, 2019 at 9:08 PM Gabor Somogyi <
>>>>> gabor.g.somo...@gmail.com> wrote:
>>>>>
>>>>>> From the info you've provided not much to say.
>>>>>> Maybe you could collect sample app, logs etc, open a jira and we can
>>>>>> take a deeper look at it...
>>>>>>
>>>>>> BR,
>>>>>> G
>>>>>>
>>>>>>
>>>>>> On Thu, Feb 21, 2019 at 4:14 PM Guillermo Ortiz <konstt2...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> I' working with Spark Streaming 2.0.2 and Kafka 1.0.0 using Direct
>>>>>>> Stream as connector. I consume data from Kafka and autosave the offsets.
>>>>>>> I can see Spark doing commits in the logs of the last offsets
>>>>>>> processed, Sometimes I have restarted spark and it starts from the
>>>>>>> beginning, when I'm using the same groupId.
>>>>>>>
>>>>>>> Why could it happen? it only happen rarely.
>>>>>>>
>>>>>>

Reply via email to