Hi Keith,

I don't think that we keep such references.
But we do experience exceptions during the job execution that we catch and
retry (timeouts/network issues from different data sources).
Can they affect RDD cleanup?

Thanks,
Alex

On Sun, Jul 21, 2019 at 10:49 PM Keith Chapman <keithgchap...@gmail.com>
wrote:

> Hi Alex,
>
> Shuffle files in spark are deleted when the object holding a reference to
> the shuffle file on disk goes out of scope (is garbage collected by the
> JVM).  Could it be the case that you are keeping these objects alive?
>
> Regards,
> Keith.
>
> http://keith-chapman.com
>
>
> On Sun, Jul 21, 2019 at 12:19 AM Alex Landa <metalo...@gmail.com> wrote:
>
>> Thanks,
>> I looked into these options, the cleaner periodic interval is set to 30
>> min by default.
>> The block option for shuffle -
>> *spark.cleaner.referenceTracking.blocking.shuffle* - is set to false by
>> default.
>> What are the implications of setting it to true?
>> Will it make the driver slower?
>>
>> Thanks,
>> Alex
>>
>> On Sun, Jul 21, 2019 at 9:06 AM Prathmesh Ranaut Gmail <
>> prathmesh.ran...@gmail.com> wrote:
>>
>>> This is the job of ContextCleaner. There are few a property that you can
>>> tweak to see if that helps:
>>> spark.cleaner.periodicGC.interval
>>> spark.cleaner.referenceTracking
>>> spark.cleaner.referenceTracking.blocking.shuffle
>>>
>>> Regards
>>> Prathmesh Ranaut
>>>
>>> On Jul 21, 2019, at 11:31 AM, Alex Landa <metalo...@gmail.com> wrote:
>>>
>>> Hi,
>>>
>>> We are running a long running Spark application ( which executes lots of
>>> quick jobs using our scheduler ) on Spark stand-alone cluster 2.4.0.
>>> We see that old shuffle files ( a week old for example ) are not deleted
>>> during the execution of the application, which leads to out of disk space
>>> errors on the executor.
>>> If we re-deploy the application, the Spark cluster take care of the
>>> cleaning
>>> and deletes the old shuffle data (since we have
>>> /-Dspark.worker.cleanup.enabled=true/ in the worker config).
>>> I don't want to re-deploy our app every week or two, but to be able to
>>> configure spark to clean old shuffle data (as it should).
>>>
>>> How can I configure Spark to delete old shuffle data during the life
>>> time of
>>> the application (not after)?
>>>
>>>
>>> Thanks,
>>> Alex
>>>
>>>

Reply via email to