[ https://issues.apache.org/jira/browse/SPARK-29950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Marcelo Masiero Vanzin reassigned SPARK-29950: ---------------------------------------------- Assignee: Marcelo Masiero Vanzin > Deleted excess executors can connect back to driver in K8S with dyn alloc on > ---------------------------------------------------------------------------- > > Key: SPARK-29950 > URL: https://issues.apache.org/jira/browse/SPARK-29950 > Project: Spark > Issue Type: Improvement > Components: Kubernetes > Affects Versions: 3.0.0 > Reporter: Marcelo Masiero Vanzin > Assignee: Marcelo Masiero Vanzin > Priority: Minor > > {{ExecutorPodsAllocator}} currently has code to delete excess pods that the > K8S server hasn't started yet, and aren't needed anymore due to downscaling. > The problem is that there is a race between K8S starting the pod and the > Spark code deleting it. This may cause the pod to connect back to Spark and > do a lot of initialization, sometimes even being considered for task > allocation, just to be killed almost immediately. > This doesn't cause any problems that I could detect in my tests, but wastes > resources, and causes logs to contains misleading messages about the executor > being killed. It would be nice to avoid that. -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org