Github user squito commented on the issue: https://github.com/apache/spark/pull/17208 hmm I think you're right @zsxwing that we should be updating `isZombie` before `sched.dagScheduler.taskEnded` and `sched.dagScheduler.taskSetFailed` is called, just to keep state consistent. I don't think you'll actually hit the bug described here, as (a) if it was from a fetch failure, `isZombie` is already set first or if (b) its just a regular task failure, and it leads to the stage getting aborted, then there aren't any more retries of the stage anyway.
--- --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org