[ https://issues.apache.org/jira/browse/SPARK-18922?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15823150#comment-15823150 ]
Apache Spark commented on SPARK-18922: -------------------------------------- User 'HyukjinKwon' has created a pull request for this issue: https://github.com/apache/spark/pull/16586 > Fix more resource-closing-related and path-related test failures in > identified ones on Windows > ---------------------------------------------------------------------------------------------- > > Key: SPARK-18922 > URL: https://issues.apache.org/jira/browse/SPARK-18922 > Project: Spark > Issue Type: Sub-task > Components: Tests > Reporter: Hyukjin Kwon > Assignee: Hyukjin Kwon > Priority: Minor > Fix For: 2.2.0 > > > There are more instances that are failed on Windows as below: > - {{LauncherBackendSuite}}: > {code} > - local: launcher handle *** FAILED *** (30 seconds, 120 milliseconds) > The code passed to eventually never returned normally. Attempted 283 times > over 30.0960053 seconds. Last failure message: The reference was null. > (LauncherBackendSuite.scala:56) > org.scalatest.exceptions.TestFailedDueToTimeoutException: > at > org.scalatest.concurrent.Eventually$class.tryTryAgain$1(Eventually.scala:420) > at > org.scalatest.concurrent.Eventually$class.eventually(Eventually.scala:438) > - standalone/client: launcher handle *** FAILED *** (30 seconds, 47 > milliseconds) > The code passed to eventually never returned normally. Attempted 282 times > over 30.037987100000002 seconds. Last failure message: The reference was > null. (LauncherBackendSuite.scala:56) > org.scalatest.exceptions.TestFailedDueToTimeoutException: > at > org.scalatest.concurrent.Eventually$class.tryTryAgain$1(Eventually.scala:420) > at > org.scalatest.concurrent.Eventually$class.eventually(Eventually.scala:438) > {code} > - {{SQLQuerySuite}}: > {code} > - specifying database name for a temporary table is not allowed *** FAILED > *** (125 milliseconds) > org.apache.spark.sql.AnalysisException: Path does not exist: > file:/C:projectsspark arget mpspark-1f4471ab-aac0-4239-ae35-833d54b37e52; > at > org.apache.spark.sql.execution.datasources.DataSource$$anonfun$14.apply(DataSource.scala:382) > at > org.apache.spark.sql.execution.datasources.DataSource$$anonfun$14.apply(DataSource.scala:370) > {code} > - {{JsonSuite}}: > {code} > - Loading a JSON dataset from a text file with SQL *** FAILED *** (94 > milliseconds) > org.apache.spark.sql.AnalysisException: Path does not exist: > file:/C:projectsspark arget mpspark-c918a8b7-fc09-433c-b9d0-36c0f78ae918; > at > org.apache.spark.sql.execution.datasources.DataSource$$anonfun$14.apply(DataSource.scala:382) > at > org.apache.spark.sql.execution.datasources.DataSource$$anonfun$14.apply(DataSource.scala:370) > {code} > - {{StateStoreSuite}}: > {code} > - SPARK-18342: commit fails when rename fails *** FAILED *** (16 milliseconds) > java.lang.IllegalArgumentException: java.net.URISyntaxException: Relative > path in absolute URI: > StateStoreSuite29777261fs://C:%5Cprojects%5Cspark%5Ctarget%5Ctmp%5Cspark-ef349862-7281-4963-aaf3-add0d670a4ad%5C?????-2218c2f8-2cf6-4f80-9cdf-96354e8246a77685899733421033312/0 > at org.apache.hadoop.fs.Path.initialize(Path.java:206) > at org.apache.hadoop.fs.Path.<init>(Path.java:116) > at org.apache.hadoop.fs.Path.<init>(Path.java:89) > ... > Cause: java.net.URISyntaxException: Relative path in absolute URI: > StateStoreSuite29777261fs://C:%5Cprojects%5Cspark%5Ctarget%5Ctmp%5Cspark-ef349862-7281-4963-aaf3-add0d670a4ad%5C?????-2218c2f8-2cf6-4f80-9cdf-96354e8246a77685899733421033312/0 > at java.net.URI.checkPath(URI.java:1823) > at java.net.URI.<init>(URI.java:745) > at org.apache.hadoop.fs.Path.initialize(Path.java:203) > {code} > - {{HDFSMetadataLogSuite}}: > {code} > - FileManager: FileContextManager *** FAILED *** (94 milliseconds) > java.io.IOException: Failed to delete: > C:\projects\spark\target\tmp\spark-415bb0bd-396b-444d-be82-04599e025f21 > at org.apache.spark.util.Utils$.deleteRecursively(Utils.scala:1010) > at > org.apache.spark.sql.test.SQLTestUtils$class.withTempDir(SQLTestUtils.scala:127) > at > org.apache.spark.sql.execution.streaming.HDFSMetadataLogSuite.withTempDir(HDFSMetadataLogSuite.scala:38) > - FileManager: FileSystemManager *** FAILED *** (78 milliseconds) > java.io.IOException: Failed to delete: > C:\projects\spark\target\tmp\spark-ef8222cd-85aa-47c0-a396-bc7979e15088 > at org.apache.spark.util.Utils$.deleteRecursively(Utils.scala:1010) > at > org.apache.spark.sql.test.SQLTestUtils$class.withTempDir(SQLTestUtils.scala:127) > at > org.apache.spark.sql.execution.streaming.HDFSMetadataLogSuite.withTempDir(HDFSMetadataLogSuite.scala:38) > {code} > Please refer, for full logs, > https://ci.appveyor.com/project/spark-test/spark/build/283-tmp-test-base -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org