Hi,
wanted to get some advice regarding tunning spark application
I see for some of the tasks many log entries like this
Executor task launch worker-38 ExternalAppendOnlyMap: Thread 239 spilling
in-memory map of 5.1 MB to disk (272 times so far)
(especially when inputs are considerable)
I
the tasks many log entries like this
> Executor task launch worker-38 ExternalAppendOnlyMap: Thread 239 spilling
> in-memory map of 5.1 MB to disk (272 times so far)
> (especially when inputs are considerable)
> I understand that this is connected to shuffle and joins, so that data is
>