wankunde commented on code in PR #37533: URL: https://github.com/apache/spark/pull/37533#discussion_r955858394
########## core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala: ########## @@ -2259,37 +2259,51 @@ private[spark] class DAGScheduler( } override def onShuffleMergeFailure(e: Throwable): Unit = { + if (e.isInstanceOf[IOException]) { + logInfo(s"Failed to connect external shuffle service " + + s"${shuffleServiceLoc.hostPort}") + blockManagerMaster.removeShufflePushMergerLocation(shuffleServiceLoc.host) + } } }) } } }, 0, TimeUnit.SECONDS) } else { - stage.shuffleDep.getMergerLocs.zipWithIndex.foreach { - case (shuffleServiceLoc, index) => - // Sends async request to shuffle service to finalize shuffle merge on that host - // TODO: SPARK-35536: Cancel finalizeShuffleMerge if the stage is cancelled - // TODO: during shuffleMergeFinalizeWaitSec - shuffleClient.finalizeShuffleMerge(shuffleServiceLoc.host, - shuffleServiceLoc.port, shuffleId, shuffleMergeId, - new MergeFinalizerListener { - override def onShuffleMergeSuccess(statuses: MergeStatuses): Unit = { - assert(shuffleId == statuses.shuffleId) - eventProcessLoop.post(RegisterMergeStatuses(stage, MergeStatus. - convertMergeStatusesToMergeStatusArr(statuses, shuffleServiceLoc))) - results(index).set(true) - } + shuffleMergeFinalizeScheduler.schedule(new Runnable { Review Comment: For a long-running spark application, if no finalize partition RPC is received, ess will not close open data files, meta files, and index files until the application is complete. Too many opening files can be a potential risk. Or should we add a parameter to control whether to continue sending finalize RPC to the remaining nodes ? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org