Github user HyukjinKwon commented on a diff in the pull request: https://github.com/apache/spark/pull/15131#discussion_r79735513 --- Diff: core/src/main/scala/org/apache/spark/SparkContext.scala --- @@ -1426,7 +1426,7 @@ class SparkContext(config: SparkConf) extends Logging with ExecutorAllocationCli * supported for Hadoop-supported filesystems. */ def addFile(path: String, recursive: Boolean): Unit = { - val uri = new URI(path) + val uri = new Path(path).toUri --- End diff -- Meanwhile, we could alternatively use `Utils.resolveURI` as well which we are already testing within Spark. However, this util seems not hadling `C:/a/b/c` case (not `C:\a\b\c` which we should fix). So I suggested `Path (...).toUri` instead but if you feel strongly about this, we could use that. I will try to find and share doc and tests for `Path` as I got in my home though.
--- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. --- --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org