Hi, I have a Flink application sourcing from a topic in Kafka (400 partitions)
and sinking to S3 using bucketingsink and using RocksDb for checkpointing every
2 mins. The Flink app runs with parallelism 400 so that each worker handles a
partition. This is using Flink 1.5.2. The Flink cluster use
Hi,
It looks that some of your slots were freed during the job execution (possibly
due to idle for too long). AFAIK the exception was thrown when a pending Slot
request was removed. You can try increase the “Slot.idle.timeout” to mitigate
this issue (default is 5, try 360 or higher).
R
Thanks for the suggestion Qi. I tried increasing slot.idle.timeout to 360
but it seems to still have encountered the issue. Does this mean if a slot or
"flink worker" has not processed items for 1 hour, that it will be removed?
Would any other flink configuration properties help for this?
s
In addition, after the Flink job has failed from the above exception, the Flink
job is unable to recover from previous checkpoint. Is this the expected
behavior? How can the job be recovered successfully from this?
‐‐‐ Original Message ‐‐‐
On Monday, November 26, 2018 12:30 AM, Flink Dev
This is weird. Could you paste your entire exception trace here?
> On Nov 26, 2018, at 4:37 PM, Flink Developer
> wrote:
>
> In addition, after the Flink job has failed from the above exception, the
> Flink job is unable to recover from previous checkpoint. Is this the expected
> behavior? Ho
Hi
It can be also that some task managers fail, then pending checkpoints are
declined and slots are removed.
Can you also have a look into task managers logs?
Best,
Andrey
> On 26 Nov 2018, at 12:19, qi luo wrote:
>
> This is weird. Could you paste your entire exception trace here?
>
>> On N