AngersZhuuuu commented on code in PR #43936: URL: https://github.com/apache/spark/pull/43936#discussion_r1401447419
########## core/src/main/scala/org/apache/spark/SparkContext.scala: ########## @@ -1822,7 +1822,7 @@ class SparkContext(config: SparkConf) extends Logging { logInfo(s"Added file $path at $key with timestamp $timestamp") // Fetch the file locally so that closures which are run on the driver can still use the // SparkFiles API to access files. - Utils.fetchFile(uri.toString, root, conf, hadoopConfiguration, timestamp, useCache = false) + Utils.fetchFile(uri.toString, root, conf, hadoopConfiguration, timestamp, useCache = true) Review Comment: > So are you saying that it doesn't fetch the file to `root` directory? didn't copy as same filename -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org