Is there a way in Spark to automatically terminate laggard "stage's", ones that appear to be hanging? In other words, is there a timeout for processing of a given RDD?
In the Spark GUI, I see the "kill" function for a given Stage under 'Details for Job <...>". Is there something in Spark that would identify and kill laggards proactively? Thanks.