Good to hear :)

On Wed, Dec 9, 2020 at 7:44 PM Eric Beabes <mailinglist...@gmail.com> wrote:

> Gabor,
>
> I waited to revert for a long time to ensure that this is working as
> expected. I am VERY HAPPY to tell you that this configuration change has
> fixed this issue! Not a single task has failed for over 2 weeks!
>
> THANKS once again. Hopefully, at some point we can switch to Spark 3.0.
>
>
> On Fri, Nov 20, 2020 at 7:30 AM Gabor Somogyi <gabor.g.somo...@gmail.com>
> wrote:
>
>> Happy that saved some time for you :)
>> We've invested quite an effort in the latest releases into streaming and
>> hope there will be less and less headaches like this.
>>
>> On Thu, Nov 19, 2020 at 5:55 PM Eric Beabes <mailinglist...@gmail.com>
>> wrote:
>>
>>> THANK YOU SO MUCH! Will try it out & revert.
>>>
>>> On Thu, Nov 19, 2020 at 8:18 AM Gabor Somogyi <gabor.g.somo...@gmail.com>
>>> wrote:
>>>
>>>> "spark.kafka.producer.cache.timeout" is available since 2.2.1 which can
>>>> be increased as a temporary workaround.
>>>> This is not super elegant but works which gives enough time to migrate
>>>> to Spark 3.
>>>>
>>>>
>>>> On Wed, Nov 18, 2020 at 11:12 PM Eric Beabes <mailinglist...@gmail.com>
>>>> wrote:
>>>>
>>>>> I must say.. *Spark has let me down in this case*. I am surprised an
>>>>> important issue like this hasn't been fixed in Spark 2.4.
>>>>>
>>>>> I am fighting a battle of 'Spark Structured Streaming' Vs 'Flink' at
>>>>> work & now because Spark 2.4 can't handle this *I've been asked to
>>>>> rewrite the code in Flink*.
>>>>>
>>>>> Moving to Spark 3.0 is not an easy option 'cause Cloudera 6.2 doesn't
>>>>> have a Spark 3.0 parcel!!!! So we can't upgrade to 3.0.
>>>>>
>>>>> So sad. Let me ask one more time. *Is there no way to fix this in
>>>>> Spark 2.4?*
>>>>>
>>>>>
>>>>> On Tue, Nov 10, 2020 at 11:33 AM Eric Beabes <mailinglist...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> BTW, we are seeing this message as well: 
>>>>>> *"org.apache.kafka.common.KafkaException:
>>>>>> Producer** closed while send in progress"*. I am assuming this
>>>>>> happens because of the previous issue.."producer has been closed", right?
>>>>>> Or are they unrelated? Please advise. Thanks.
>>>>>>
>>>>>> On Tue, Nov 10, 2020 at 11:17 AM Eric Beabes <
>>>>>> mailinglist...@gmail.com> wrote:
>>>>>>
>>>>>>> Thanks for the reply. We are on Spark 2.4. Is there no way to get
>>>>>>> this fixed in Spark 2.4?
>>>>>>>
>>>>>>> On Mon, Nov 2, 2020 at 8:32 PM Jungtaek Lim <
>>>>>>> kabhwan.opensou...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Which Spark version do you use? There's a known issue on Kafka
>>>>>>>> producer pool in Spark 2.x which was fixed in Spark 3.0, so you'd like 
>>>>>>>> to
>>>>>>>> check whether your case is bound to the known issue or not.
>>>>>>>>
>>>>>>>> https://issues.apache.org/jira/browse/SPARK-21869
>>>>>>>>
>>>>>>>>
>>>>>>>> On Tue, Nov 3, 2020 at 1:53 AM Eric Beabes <
>>>>>>>> mailinglist...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> I know this is related to Kafka but it happens during the Spark
>>>>>>>>> Structured Streaming job that's why I am asking on this mailing list.
>>>>>>>>>
>>>>>>>>> How would you debug this or get around this in Spark Structured
>>>>>>>>> Streaming? Any tips would be appreciated. Thanks.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> java.lang.IllegalStateException: Cannot perform operation after
>>>>>>>>> producer has been closed at
>>>>>>>>> org.apache.kafka.clients.producer.KafkaProducer.throwIfProducerClosed(KafkaProducer.java:853)
>>>>>>>>> at
>>>>>>>>> org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:862)
>>>>>>>>> at
>>>>>>>>> org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:846)
>>>>>>>>> at
>>>>>>>>> org.apache.spark.sql.kafka010.KafkaRowWriter.sendRow(KafkaWriteTask.scala:92)
>>>>>>>>> at
>>>>>>>>> org.apache.spark.sql.kafka010.KafkaStreamDataWriter.write(KafkaStreamWriter.scala:95)
>>>>>>>>>
>>>>>>>>

Reply via email to