Github user squito commented on the issue:

    https://github.com/apache/spark/pull/17208
  
    hmm I think you're right @zsxwing that we should be updating `isZombie` 
before `sched.dagScheduler.taskEnded` and `sched.dagScheduler.taskSetFailed` is 
called, just to keep state consistent.  I don't think you'll actually hit the 
bug described here, as (a) if it was from a fetch failure, `isZombie` is 
already set first or if (b) its just a regular task failure, and it leads to 
the stage getting aborted, then there aren't any more retries of the stage 
anyway.  


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to