[jira] [Updated] (SPARK-32991) RESET can clear StaticSQLConfs
[ https://issues.apache.org/jira/browse/SPARK-32991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiao Li updated SPARK-32991: Target Version/s: 3.1.0 > RESET can clear StaticSQLConfs > -- > > Key: SPARK-32991 > URL: https://issues.apache.org/jira/browse/SPARK-32991 > Project: Spark > Issue Type: Improvement > Components: SQL >Affects Versions: 3.0.1 >Reporter: Herman van Hövell >Assignee: Kent Yao >Priority: Blocker > Fix For: 3.1.0 > > > The RESET command can clear a sessions' static SQL configurations, when that > static SQL configuration was set on a SparkSession that uses a pre-existing > SparkContext. Here is repro: > {code:java} > // Blow away any pre-existing session thread locals > org.apache.spark.sql.SparkSession.clearDefaultSession() > org.apache.spark.sql.SparkSession.clearActiveSession() > // Create new session and explicitly set a spark context > val newSession = org.apache.spark.sql.SparkSession.builder > .sparkContext(sc) > .config("spark.sql.globalTempDatabase", "bob") > .getOrCreate() > assert(newSession.conf.get("spark.sql.globalTempDatabase") == "bob") > newSession.sql("reset") > assert(newSession.conf.get("spark.sql.globalTempDatabase") == "bob") // Boom! > {code} > The problem is that RESET assumes it can use the SparkContext's > configurations as its default. > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-32991) RESET can clear StaticSQLConfs
[ https://issues.apache.org/jira/browse/SPARK-32991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiao Li updated SPARK-32991: Priority: Blocker (was: Major) > RESET can clear StaticSQLConfs > -- > > Key: SPARK-32991 > URL: https://issues.apache.org/jira/browse/SPARK-32991 > Project: Spark > Issue Type: Improvement > Components: SQL >Affects Versions: 3.0.1 >Reporter: Herman van Hövell >Assignee: Kent Yao >Priority: Blocker > Fix For: 3.1.0 > > > The RESET command can clear a sessions' static SQL configurations, when that > static SQL configuration was set on a SparkSession that uses a pre-existing > SparkContext. Here is repro: > {code:java} > // Blow away any pre-existing session thread locals > org.apache.spark.sql.SparkSession.clearDefaultSession() > org.apache.spark.sql.SparkSession.clearActiveSession() > // Create new session and explicitly set a spark context > val newSession = org.apache.spark.sql.SparkSession.builder > .sparkContext(sc) > .config("spark.sql.globalTempDatabase", "bob") > .getOrCreate() > assert(newSession.conf.get("spark.sql.globalTempDatabase") == "bob") > newSession.sql("reset") > assert(newSession.conf.get("spark.sql.globalTempDatabase") == "bob") // Boom! > {code} > The problem is that RESET assumes it can use the SparkContext's > configurations as its default. > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-32991) RESET can clear StaticSQLConfs
[ https://issues.apache.org/jira/browse/SPARK-32991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244675#comment-17244675 ] Xiao Li commented on SPARK-32991: - [~Qin Yao] I reopened the Jira and set it to Blocker, because the follow-up PR is needed before the 3.1 release. > RESET can clear StaticSQLConfs > -- > > Key: SPARK-32991 > URL: https://issues.apache.org/jira/browse/SPARK-32991 > Project: Spark > Issue Type: Improvement > Components: SQL >Affects Versions: 3.0.1 >Reporter: Herman van Hövell >Assignee: Kent Yao >Priority: Blocker > Fix For: 3.1.0 > > > The RESET command can clear a sessions' static SQL configurations, when that > static SQL configuration was set on a SparkSession that uses a pre-existing > SparkContext. Here is repro: > {code:java} > // Blow away any pre-existing session thread locals > org.apache.spark.sql.SparkSession.clearDefaultSession() > org.apache.spark.sql.SparkSession.clearActiveSession() > // Create new session and explicitly set a spark context > val newSession = org.apache.spark.sql.SparkSession.builder > .sparkContext(sc) > .config("spark.sql.globalTempDatabase", "bob") > .getOrCreate() > assert(newSession.conf.get("spark.sql.globalTempDatabase") == "bob") > newSession.sql("reset") > assert(newSession.conf.get("spark.sql.globalTempDatabase") == "bob") // Boom! > {code} > The problem is that RESET assumes it can use the SparkContext's > configurations as its default. > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Reopened] (SPARK-32991) RESET can clear StaticSQLConfs
[ https://issues.apache.org/jira/browse/SPARK-32991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiao Li reopened SPARK-32991: - > RESET can clear StaticSQLConfs > -- > > Key: SPARK-32991 > URL: https://issues.apache.org/jira/browse/SPARK-32991 > Project: Spark > Issue Type: Improvement > Components: SQL >Affects Versions: 3.0.1 >Reporter: Herman van Hövell >Assignee: Kent Yao >Priority: Major > Fix For: 3.1.0 > > > The RESET command can clear a sessions' static SQL configurations, when that > static SQL configuration was set on a SparkSession that uses a pre-existing > SparkContext. Here is repro: > {code:java} > // Blow away any pre-existing session thread locals > org.apache.spark.sql.SparkSession.clearDefaultSession() > org.apache.spark.sql.SparkSession.clearActiveSession() > // Create new session and explicitly set a spark context > val newSession = org.apache.spark.sql.SparkSession.builder > .sparkContext(sc) > .config("spark.sql.globalTempDatabase", "bob") > .getOrCreate() > assert(newSession.conf.get("spark.sql.globalTempDatabase") == "bob") > newSession.sql("reset") > assert(newSession.conf.get("spark.sql.globalTempDatabase") == "bob") // Boom! > {code} > The problem is that RESET assumes it can use the SparkContext's > configurations as its default. > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33044) Add a Jenkins build and test job for Scala 2.13
[ https://issues.apache.org/jira/browse/SPARK-33044?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244672#comment-17244672 ] Dongjoon Hyun commented on SPARK-33044: --- I don't think so, [~srowen]. And, new Jenkins servers rejects my login. Is it still accessible to you? > Add a Jenkins build and test job for Scala 2.13 > --- > > Key: SPARK-33044 > URL: https://issues.apache.org/jira/browse/SPARK-33044 > Project: Spark > Issue Type: Sub-task > Components: jenkins >Affects Versions: 3.1.0 >Reporter: Yang Jie >Priority: Major > > {{Master}} branch seems to be almost ready for Scala 2.13 now, we need a > Jenkins test job to verify current work results and CI. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-33668) Fix flaky test "Verify logging configuration is picked from the provided SPARK_CONF_DIR/log4j.properties."
[ https://issues.apache.org/jira/browse/SPARK-33668?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dongjoon Hyun reassigned SPARK-33668: - Assignee: Prashant Sharma > Fix flaky test "Verify logging configuration is picked from the provided > SPARK_CONF_DIR/log4j.properties." > -- > > Key: SPARK-33668 > URL: https://issues.apache.org/jira/browse/SPARK-33668 > Project: Spark > Issue Type: Bug > Components: Kubernetes, Tests >Affects Versions: 3.1.0 >Reporter: Prashant Sharma >Assignee: Prashant Sharma >Priority: Major > > The test is flaking, with multiple flaked instances - the reason for the > failure has been similar to: > {code:java} > The code passed to eventually never returned normally. Attempted 109 times > over 3.007988241397 minutes. Last failure message: Failure executing: GET > at: > https://192.168.39.167:8443/api/v1/namespaces/b37fc72a991b49baa68a2eaaa1516463/pods/spark-pi-97a9bc76308e7fe3-exec-1/log?pretty=false. > Message: pods "spark-pi-97a9bc76308e7fe3-exec-1" not found. Received status: > Status(apiVersion=v1, code=404, details=StatusDetails(causes=[], group=null, > kind=pods, name=spark-pi-97a9bc76308e7fe3-exec-1, retryAfterSeconds=null, > uid=null, additionalProperties={}), kind=Status, message=pods > "spark-pi-97a9bc76308e7fe3-exec-1" not found, > metadata=ListMeta(_continue=null, remainingItemCount=null, > resourceVersion=null, selfLink=null, additionalProperties={}), > reason=NotFound, status=Failure, additionalProperties={}).. > (KubernetesSuite.scala:402) > {code} > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36854/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36852/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36850/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36848/console > From the above failures, it seems, that executor finishes too quickly and is > removed by spark before the test can complete. > So, in order to mitigate this situation, one way is to turn on the flag > {code} >"spark.kubernetes.executor.deleteOnTermination" > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-33668) Fix flaky test "Verify logging configuration is picked from the provided SPARK_CONF_DIR/log4j.properties."
[ https://issues.apache.org/jira/browse/SPARK-33668?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dongjoon Hyun resolved SPARK-33668. --- Fix Version/s: 3.2.0 Resolution: Fixed Issue resolved by pull request 30616 [https://github.com/apache/spark/pull/30616] > Fix flaky test "Verify logging configuration is picked from the provided > SPARK_CONF_DIR/log4j.properties." > -- > > Key: SPARK-33668 > URL: https://issues.apache.org/jira/browse/SPARK-33668 > Project: Spark > Issue Type: Bug > Components: Kubernetes, Tests >Affects Versions: 3.1.0 >Reporter: Prashant Sharma >Assignee: Prashant Sharma >Priority: Major > Fix For: 3.2.0 > > > The test is flaking, with multiple flaked instances - the reason for the > failure has been similar to: > {code:java} > The code passed to eventually never returned normally. Attempted 109 times > over 3.007988241397 minutes. Last failure message: Failure executing: GET > at: > https://192.168.39.167:8443/api/v1/namespaces/b37fc72a991b49baa68a2eaaa1516463/pods/spark-pi-97a9bc76308e7fe3-exec-1/log?pretty=false. > Message: pods "spark-pi-97a9bc76308e7fe3-exec-1" not found. Received status: > Status(apiVersion=v1, code=404, details=StatusDetails(causes=[], group=null, > kind=pods, name=spark-pi-97a9bc76308e7fe3-exec-1, retryAfterSeconds=null, > uid=null, additionalProperties={}), kind=Status, message=pods > "spark-pi-97a9bc76308e7fe3-exec-1" not found, > metadata=ListMeta(_continue=null, remainingItemCount=null, > resourceVersion=null, selfLink=null, additionalProperties={}), > reason=NotFound, status=Failure, additionalProperties={}).. > (KubernetesSuite.scala:402) > {code} > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36854/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36852/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36850/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36848/console > From the above failures, it seems, that executor finishes too quickly and is > removed by spark before the test can complete. > So, in order to mitigate this situation, one way is to turn on the flag > {code} >"spark.kubernetes.executor.deleteOnTermination" > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33256) Update contribution guide about NumPy documentation style
[ https://issues.apache.org/jira/browse/SPARK-33256?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244649#comment-17244649 ] Apache Spark commented on SPARK-33256: -- User 'HyukjinKwon' has created a pull request for this issue: https://github.com/apache/spark/pull/30622 > Update contribution guide about NumPy documentation style > - > > Key: SPARK-33256 > URL: https://issues.apache.org/jira/browse/SPARK-33256 > Project: Spark > Issue Type: Sub-task > Components: Documentation, PySpark >Affects Versions: 3.1.0 >Reporter: Hyukjin Kwon >Priority: Major > > We should document that PySpark uses NumPy documentation style. > See also https://github.com/apache/spark/pull/30181#discussion_r517314341 -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-33256) Update contribution guide about NumPy documentation style
[ https://issues.apache.org/jira/browse/SPARK-33256?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-33256: Assignee: (was: Apache Spark) > Update contribution guide about NumPy documentation style > - > > Key: SPARK-33256 > URL: https://issues.apache.org/jira/browse/SPARK-33256 > Project: Spark > Issue Type: Sub-task > Components: Documentation, PySpark >Affects Versions: 3.1.0 >Reporter: Hyukjin Kwon >Priority: Major > > We should document that PySpark uses NumPy documentation style. > See also https://github.com/apache/spark/pull/30181#discussion_r517314341 -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-33256) Update contribution guide about NumPy documentation style
[ https://issues.apache.org/jira/browse/SPARK-33256?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-33256: Assignee: Apache Spark > Update contribution guide about NumPy documentation style > - > > Key: SPARK-33256 > URL: https://issues.apache.org/jira/browse/SPARK-33256 > Project: Spark > Issue Type: Sub-task > Components: Documentation, PySpark >Affects Versions: 3.1.0 >Reporter: Hyukjin Kwon >Assignee: Apache Spark >Priority: Major > > We should document that PySpark uses NumPy documentation style. > See also https://github.com/apache/spark/pull/30181#discussion_r517314341 -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33256) Update contribution guide about NumPy documentation style
[ https://issues.apache.org/jira/browse/SPARK-33256?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244648#comment-17244648 ] Apache Spark commented on SPARK-33256: -- User 'HyukjinKwon' has created a pull request for this issue: https://github.com/apache/spark/pull/30622 > Update contribution guide about NumPy documentation style > - > > Key: SPARK-33256 > URL: https://issues.apache.org/jira/browse/SPARK-33256 > Project: Spark > Issue Type: Sub-task > Components: Documentation, PySpark >Affects Versions: 3.1.0 >Reporter: Hyukjin Kwon >Priority: Major > > We should document that PySpark uses NumPy documentation style. > See also https://github.com/apache/spark/pull/30181#discussion_r517314341 -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33674) Show Slowpoke notifications in SBT tests
[ https://issues.apache.org/jira/browse/SPARK-33674?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244647#comment-17244647 ] Apache Spark commented on SPARK-33674: -- User 'gatorsmile' has created a pull request for this issue: https://github.com/apache/spark/pull/30621 > Show Slowpoke notifications in SBT tests > > > Key: SPARK-33674 > URL: https://issues.apache.org/jira/browse/SPARK-33674 > Project: Spark > Issue Type: Test > Components: Tests >Affects Versions: 3.0.1 >Reporter: Xiao Li >Assignee: Xiao Li >Priority: Major > > When the tests/code has bug and enters the infinite loop, it is hard to tell > which test cases hit some issues from the log, especially when we are running > the tests in parallel. It would be nice to show the Slowpoke notifications. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33674) Show Slowpoke notifications in SBT tests
[ https://issues.apache.org/jira/browse/SPARK-33674?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244646#comment-17244646 ] Apache Spark commented on SPARK-33674: -- User 'gatorsmile' has created a pull request for this issue: https://github.com/apache/spark/pull/30621 > Show Slowpoke notifications in SBT tests > > > Key: SPARK-33674 > URL: https://issues.apache.org/jira/browse/SPARK-33674 > Project: Spark > Issue Type: Test > Components: Tests >Affects Versions: 3.0.1 >Reporter: Xiao Li >Assignee: Xiao Li >Priority: Major > > When the tests/code has bug and enters the infinite loop, it is hard to tell > which test cases hit some issues from the log, especially when we are running > the tests in parallel. It would be nice to show the Slowpoke notifications. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-33674) Show Slowpoke notifications in SBT tests
[ https://issues.apache.org/jira/browse/SPARK-33674?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-33674: Assignee: Xiao Li (was: Apache Spark) > Show Slowpoke notifications in SBT tests > > > Key: SPARK-33674 > URL: https://issues.apache.org/jira/browse/SPARK-33674 > Project: Spark > Issue Type: Test > Components: Tests >Affects Versions: 3.0.1 >Reporter: Xiao Li >Assignee: Xiao Li >Priority: Major > > When the tests/code has bug and enters the infinite loop, it is hard to tell > which test cases hit some issues from the log, especially when we are running > the tests in parallel. It would be nice to show the Slowpoke notifications. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-33674) Show Slowpoke notifications in SBT tests
[ https://issues.apache.org/jira/browse/SPARK-33674?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-33674: Assignee: Apache Spark (was: Xiao Li) > Show Slowpoke notifications in SBT tests > > > Key: SPARK-33674 > URL: https://issues.apache.org/jira/browse/SPARK-33674 > Project: Spark > Issue Type: Test > Components: Tests >Affects Versions: 3.0.1 >Reporter: Xiao Li >Assignee: Apache Spark >Priority: Major > > When the tests/code has bug and enters the infinite loop, it is hard to tell > which test cases hit some issues from the log, especially when we are running > the tests in parallel. It would be nice to show the Slowpoke notifications. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33256) Update contribution guide about NumPy documentation style
[ https://issues.apache.org/jira/browse/SPARK-33256?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244645#comment-17244645 ] Hyukjin Kwon commented on SPARK-33256: -- We should have a page like https://pandas.pydata.org/docs/development/contributing_docstring.html but I will leave it to the future work. > Update contribution guide about NumPy documentation style > - > > Key: SPARK-33256 > URL: https://issues.apache.org/jira/browse/SPARK-33256 > Project: Spark > Issue Type: Sub-task > Components: Documentation, PySpark >Affects Versions: 3.1.0 >Reporter: Hyukjin Kwon >Priority: Major > > We should document that PySpark uses NumPy documentation style. > See also https://github.com/apache/spark/pull/30181#discussion_r517314341 -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Created] (SPARK-33674) Show Slowpoke notifications in SBT tests
Xiao Li created SPARK-33674: --- Summary: Show Slowpoke notifications in SBT tests Key: SPARK-33674 URL: https://issues.apache.org/jira/browse/SPARK-33674 Project: Spark Issue Type: Test Components: Tests Affects Versions: 3.0.1 Reporter: Xiao Li Assignee: Xiao Li When the tests/code has bug and enters the infinite loop, it is hard to tell which test cases hit some issues from the log, especially when we are running the tests in parallel. It would be nice to show the Slowpoke notifications. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-33637) Spark sql drop non-existent table will not report an error after failure
[ https://issues.apache.org/jira/browse/SPARK-33637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hyukjin Kwon resolved SPARK-33637. -- Resolution: Invalid No feedback from the reporter. > Spark sql drop non-existent table will not report an error after failure > > > Key: SPARK-33637 > URL: https://issues.apache.org/jira/browse/SPARK-33637 > Project: Spark > Issue Type: Bug > Components: SQL >Affects Versions: 2.4.0 >Reporter: guihuawen >Priority: Major > > In the migration of spark sql to hive sql, in order to reduce user changes, > spark sql will report an error when executing a table where drop does not > exist, but hive sql will not. In order to ensure that the grammar can be > executed normally, no error is reported when changing to a table that does > not exist in drop. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33639) The external table does not specify a location
[ https://issues.apache.org/jira/browse/SPARK-33639?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244640#comment-17244640 ] Hyukjin Kwon commented on SPARK-33639: -- [~guihuawen] what is the actual result and expected result? Can you share a self-contained reproducer with outputs? > The external table does not specify a location > --- > > Key: SPARK-33639 > URL: https://issues.apache.org/jira/browse/SPARK-33639 > Project: Spark > Issue Type: Bug > Components: SQL >Affects Versions: 2.4.0 >Reporter: guihuawen >Priority: Major > > If the location of the external table is not specified, an error will be > reported. If the partition is not specified, the default is the same as the > specified partition of the internal table. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-33648) Moving file stage failed cause dulpicated data
[ https://issues.apache.org/jira/browse/SPARK-33648?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hyukjin Kwon resolved SPARK-33648. -- Resolution: Invalid [~blackpig] please fill the JIRA description. > Moving file stage failed cause dulpicated data > --- > > Key: SPARK-33648 > URL: https://issues.apache.org/jira/browse/SPARK-33648 > Project: Spark > Issue Type: Bug > Components: SQL >Affects Versions: 3.0.1 >Reporter: Tongwei >Priority: Major > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33661) Unable to load RandomForestClassificationModel trained in Spark 2.x
[ https://issues.apache.org/jira/browse/SPARK-33661?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244638#comment-17244638 ] Hyukjin Kwon commented on SPARK-33661: -- There's a major version bumped up from 2 -> 3. Breaking change is legitimate here. > Unable to load RandomForestClassificationModel trained in Spark 2.x > --- > > Key: SPARK-33661 > URL: https://issues.apache.org/jira/browse/SPARK-33661 > Project: Spark > Issue Type: Bug > Components: ML >Affects Versions: 3.0.1 >Reporter: Marcus Levine >Priority: Major > > When attempting to load a RandomForestClassificationModel that was trained in > Spark 2.x using Spark 3.x, an exception is raised: > {code:python} > ... > RandomForestClassificationModel.load('/path/to/my/model') > File "/usr/spark/python/lib/pyspark.zip/pyspark/ml/util.py", line 330, in > load > File "/usr/spark/python/lib/pyspark.zip/pyspark/ml/pipeline.py", line 291, > in load > File "/usr/spark/python/lib/pyspark.zip/pyspark/ml/util.py", line 280, in > load > File "/usr/spark/python/lib/py4j-0.10.9-src.zip/py4j/java_gateway.py", line > 1305, in __call__ > File "/usr/spark/python/lib/pyspark.zip/pyspark/sql/utils.py", line 134, in > deco > File "", line 3, in raise_from > pyspark.sql.utils.AnalysisException: No such struct field rawCount in id, > prediction, impurity, impurityStats, gain, leftChild, rightChild, split; > {code} > There seems to be a schema incompatibility between the trained model data > saved by Spark 2.x and the expected data for a model trained in Spark 3.x > If this issue is not resolved, users will be forced to retrain any existing > random forest models they trained in Spark 2.x using Spark 3.x before they > can upgrade -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-33661) Unable to load RandomForestClassificationModel trained in Spark 2.x
[ https://issues.apache.org/jira/browse/SPARK-33661?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hyukjin Kwon resolved SPARK-33661. -- Resolution: Not A Problem > Unable to load RandomForestClassificationModel trained in Spark 2.x > --- > > Key: SPARK-33661 > URL: https://issues.apache.org/jira/browse/SPARK-33661 > Project: Spark > Issue Type: Bug > Components: ML >Affects Versions: 3.0.1 >Reporter: Marcus Levine >Priority: Major > > When attempting to load a RandomForestClassificationModel that was trained in > Spark 2.x using Spark 3.x, an exception is raised: > {code:python} > ... > RandomForestClassificationModel.load('/path/to/my/model') > File "/usr/spark/python/lib/pyspark.zip/pyspark/ml/util.py", line 330, in > load > File "/usr/spark/python/lib/pyspark.zip/pyspark/ml/pipeline.py", line 291, > in load > File "/usr/spark/python/lib/pyspark.zip/pyspark/ml/util.py", line 280, in > load > File "/usr/spark/python/lib/py4j-0.10.9-src.zip/py4j/java_gateway.py", line > 1305, in __call__ > File "/usr/spark/python/lib/pyspark.zip/pyspark/sql/utils.py", line 134, in > deco > File "", line 3, in raise_from > pyspark.sql.utils.AnalysisException: No such struct field rawCount in id, > prediction, impurity, impurityStats, gain, leftChild, rightChild, split; > {code} > There seems to be a schema incompatibility between the trained model data > saved by Spark 2.x and the expected data for a model trained in Spark 3.x > If this issue is not resolved, users will be forced to retrain any existing > random forest models they trained in Spark 2.x using Spark 3.x before they > can upgrade -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33673) Do not push down partition filters to ParquetScan for DataSourceV2
[ https://issues.apache.org/jira/browse/SPARK-33673?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244624#comment-17244624 ] Yuming Wang commented on SPARK-33673: - Yes, you can verify it by: {code:sh} mvn -Dtest=none -DwildcardSuites=org.apache.spark.sql.execution.datasources.parquet.ParquetV2SchemaPruningSuite -Dparquet.version=1.11.1 test {code} > Do not push down partition filters to ParquetScan for DataSourceV2 > -- > > Key: SPARK-33673 > URL: https://issues.apache.org/jira/browse/SPARK-33673 > Project: Spark > Issue Type: Sub-task > Components: SQL >Affects Versions: 3.2.0 >Reporter: Yuming Wang >Priority: Major > > {noformat} > - Spark vectorized reader - with partition data column - select a single > complex field array and its parent struct array *** FAILED *** > - Non-vectorized reader - with partition data column - select a single > complex field array and its parent struct array *** FAILED *** > - Spark vectorized reader - with partition data column - select a single > complex field from a map entry and its parent map entry *** FAILED *** > - Non-vectorized reader - with partition data column - select a single > complex field from a map entry and its parent map entry *** FAILED *** > - Spark vectorized reader - with partition data column - partial schema > intersection - select missing subfield *** FAILED *** > - Non-vectorized reader - with partition data column - partial schema > intersection - select missing subfield *** FAILED *** > - Spark vectorized reader - with partition data column - no unnecessary > schema pruning *** FAILED *** > - Non-vectorized reader - with partition data column - no unnecessary schema > pruning *** FAILED *** > - Spark vectorized reader - with partition data column - empty schema > intersection *** FAILED *** > - Non-vectorized reader - with partition data column - empty schema > intersection *** FAILED *** > - Spark vectorized reader - with partition data column - select a single > complex field and in where clause *** FAILED *** > - Non-vectorized reader - with partition data column - select a single > complex field and in where clause *** FAILED *** > - Spark vectorized reader - with partition data column - select nullable > complex field and having is not null predicate *** FAILED *** > - Non-vectorized reader - with partition data column - select nullable > complex field and having is not null predicate *** FAILED *** > {noformat} > These test will fail, this is because Parquet will return empty results for > non-existent column since PARQUET-1765. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33673) Do not push down partition filters to ParquetScan for DataSourceV2
[ https://issues.apache.org/jira/browse/SPARK-33673?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244615#comment-17244615 ] L. C. Hsieh commented on SPARK-33673: - You mean these test failed once upgrading to Parquet 1.11.1? > Do not push down partition filters to ParquetScan for DataSourceV2 > -- > > Key: SPARK-33673 > URL: https://issues.apache.org/jira/browse/SPARK-33673 > Project: Spark > Issue Type: Sub-task > Components: SQL >Affects Versions: 3.2.0 >Reporter: Yuming Wang >Priority: Major > > {noformat} > - Spark vectorized reader - with partition data column - select a single > complex field array and its parent struct array *** FAILED *** > - Non-vectorized reader - with partition data column - select a single > complex field array and its parent struct array *** FAILED *** > - Spark vectorized reader - with partition data column - select a single > complex field from a map entry and its parent map entry *** FAILED *** > - Non-vectorized reader - with partition data column - select a single > complex field from a map entry and its parent map entry *** FAILED *** > - Spark vectorized reader - with partition data column - partial schema > intersection - select missing subfield *** FAILED *** > - Non-vectorized reader - with partition data column - partial schema > intersection - select missing subfield *** FAILED *** > - Spark vectorized reader - with partition data column - no unnecessary > schema pruning *** FAILED *** > - Non-vectorized reader - with partition data column - no unnecessary schema > pruning *** FAILED *** > - Spark vectorized reader - with partition data column - empty schema > intersection *** FAILED *** > - Non-vectorized reader - with partition data column - empty schema > intersection *** FAILED *** > - Spark vectorized reader - with partition data column - select a single > complex field and in where clause *** FAILED *** > - Non-vectorized reader - with partition data column - select a single > complex field and in where clause *** FAILED *** > - Spark vectorized reader - with partition data column - select nullable > complex field and having is not null predicate *** FAILED *** > - Non-vectorized reader - with partition data column - select nullable > complex field and having is not null predicate *** FAILED *** > {noformat} > These test will fail, this is because Parquet will return empty results for > non-existent column since PARQUET-1765. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33673) Do not push down partition filters to ParquetScan for DataSourceV2
[ https://issues.apache.org/jira/browse/SPARK-33673?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244612#comment-17244612 ] Yuming Wang commented on SPARK-33673: - [~cloud_fan] [~rdblue] [~viirya] > Do not push down partition filters to ParquetScan for DataSourceV2 > -- > > Key: SPARK-33673 > URL: https://issues.apache.org/jira/browse/SPARK-33673 > Project: Spark > Issue Type: Sub-task > Components: SQL >Affects Versions: 3.2.0 >Reporter: Yuming Wang >Priority: Major > > {noformat} > - Spark vectorized reader - with partition data column - select a single > complex field array and its parent struct array *** FAILED *** > - Non-vectorized reader - with partition data column - select a single > complex field array and its parent struct array *** FAILED *** > - Spark vectorized reader - with partition data column - select a single > complex field from a map entry and its parent map entry *** FAILED *** > - Non-vectorized reader - with partition data column - select a single > complex field from a map entry and its parent map entry *** FAILED *** > - Spark vectorized reader - with partition data column - partial schema > intersection - select missing subfield *** FAILED *** > - Non-vectorized reader - with partition data column - partial schema > intersection - select missing subfield *** FAILED *** > - Spark vectorized reader - with partition data column - no unnecessary > schema pruning *** FAILED *** > - Non-vectorized reader - with partition data column - no unnecessary schema > pruning *** FAILED *** > - Spark vectorized reader - with partition data column - empty schema > intersection *** FAILED *** > - Non-vectorized reader - with partition data column - empty schema > intersection *** FAILED *** > - Spark vectorized reader - with partition data column - select a single > complex field and in where clause *** FAILED *** > - Non-vectorized reader - with partition data column - select a single > complex field and in where clause *** FAILED *** > - Spark vectorized reader - with partition data column - select nullable > complex field and having is not null predicate *** FAILED *** > - Non-vectorized reader - with partition data column - select nullable > complex field and having is not null predicate *** FAILED *** > {noformat} > These test will fail, this is because Parquet will return empty results for > non-existent column since PARQUET-1765. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Created] (SPARK-33673) Do not push down partition filters to ParquetScan for DataSourceV2
Yuming Wang created SPARK-33673: --- Summary: Do not push down partition filters to ParquetScan for DataSourceV2 Key: SPARK-33673 URL: https://issues.apache.org/jira/browse/SPARK-33673 Project: Spark Issue Type: Sub-task Components: SQL Affects Versions: 3.2.0 Reporter: Yuming Wang {noformat} - Spark vectorized reader - with partition data column - select a single complex field array and its parent struct array *** FAILED *** - Non-vectorized reader - with partition data column - select a single complex field array and its parent struct array *** FAILED *** - Spark vectorized reader - with partition data column - select a single complex field from a map entry and its parent map entry *** FAILED *** - Non-vectorized reader - with partition data column - select a single complex field from a map entry and its parent map entry *** FAILED *** - Spark vectorized reader - with partition data column - partial schema intersection - select missing subfield *** FAILED *** - Non-vectorized reader - with partition data column - partial schema intersection - select missing subfield *** FAILED *** - Spark vectorized reader - with partition data column - no unnecessary schema pruning *** FAILED *** - Non-vectorized reader - with partition data column - no unnecessary schema pruning *** FAILED *** - Spark vectorized reader - with partition data column - empty schema intersection *** FAILED *** - Non-vectorized reader - with partition data column - empty schema intersection *** FAILED *** - Spark vectorized reader - with partition data column - select a single complex field and in where clause *** FAILED *** - Non-vectorized reader - with partition data column - select a single complex field and in where clause *** FAILED *** - Spark vectorized reader - with partition data column - select nullable complex field and having is not null predicate *** FAILED *** - Non-vectorized reader - with partition data column - select nullable complex field and having is not null predicate *** FAILED *** {noformat} These test will fail, this is because Parquet will return empty results for non-existent column since PARQUET-1765. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-33499) Enable `-Wunused:imports` in Scala 2.13 SBT
[ https://issues.apache.org/jira/browse/SPARK-33499?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen updated SPARK-33499: - Parent: (was: SPARK-25075) Issue Type: Improvement (was: Sub-task) > Enable `-Wunused:imports` in Scala 2.13 SBT > --- > > Key: SPARK-33499 > URL: https://issues.apache.org/jira/browse/SPARK-33499 > Project: Spark > Issue Type: Improvement > Components: Build >Affects Versions: 3.1.0 > Environment: !image-2020-11-20-18-49-16-384.png! > Scala 2.13 think this is a {{unused import}} but Scala 2.12 compile failed > without this import, so comments {{-Wunused:imports}} in Scala 2.13 and left > a TODO in this pr > >Reporter: Yang Jie >Priority: Minor > Attachments: image-2020-11-20-18-50-21-567.png > > > As the image, Scala 2.13 think `scala.language.higherKinds` is a {{unused > import, }}but Scala 2.12 compile failed without this import, so comments > {{-Wunused:imports}} in Scala 2.13 and left a TODO, > we should enabled `-Wunused:imports` when Scala 2.12 is no longer supported. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-33499) Enable `-Wunused:imports` in Scala 2.13 SBT
[ https://issues.apache.org/jira/browse/SPARK-33499?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen resolved SPARK-33499. -- Resolution: Not A Problem OK, so we don't remove it. I don't see the issue. > Enable `-Wunused:imports` in Scala 2.13 SBT > --- > > Key: SPARK-33499 > URL: https://issues.apache.org/jira/browse/SPARK-33499 > Project: Spark > Issue Type: Sub-task > Components: Build >Affects Versions: 3.1.0 > Environment: !image-2020-11-20-18-49-16-384.png! > Scala 2.13 think this is a {{unused import}} but Scala 2.12 compile failed > without this import, so comments {{-Wunused:imports}} in Scala 2.13 and left > a TODO in this pr > >Reporter: Yang Jie >Priority: Minor > Attachments: image-2020-11-20-18-50-21-567.png > > > As the image, Scala 2.13 think `scala.language.higherKinds` is a {{unused > import, }}but Scala 2.12 compile failed without this import, so comments > {{-Wunused:imports}} in Scala 2.13 and left a TODO, > we should enabled `-Wunused:imports` when Scala 2.12 is no longer supported. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-33348) Use scala.jdk.CollectionConverters replace scala.collection.JavaConverters
[ https://issues.apache.org/jira/browse/SPARK-33348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen updated SPARK-33348: - Parent: (was: SPARK-25075) Issue Type: Improvement (was: Sub-task) > Use scala.jdk.CollectionConverters replace scala.collection.JavaConverters > -- > > Key: SPARK-33348 > URL: https://issues.apache.org/jira/browse/SPARK-33348 > Project: Spark > Issue Type: Improvement > Components: Build >Affects Versions: 3.1.0 >Reporter: Yang Jie >Priority: Minor > > `scala.collection.JavaConverters` is deprecated in Scala 2.13, there are many > compilation warnings about this, should use `scala.jdk.CollectionConverters` > replace it. > But `scala.jdk.CollectionConverters` only available in Scala 2.13. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-33345) Batch fix compilation warnings about "Widening conversion from XXX to XXX is deprecated"
[ https://issues.apache.org/jira/browse/SPARK-33345?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen updated SPARK-33345: - Parent: (was: SPARK-25075) Issue Type: Improvement (was: Sub-task) > Batch fix compilation warnings about "Widening conversion from XXX to XXX is > deprecated" > - > > Key: SPARK-33345 > URL: https://issues.apache.org/jira/browse/SPARK-33345 > Project: Spark > Issue Type: Improvement > Components: Build >Affects Versions: 3.1.0 >Reporter: Yang Jie >Priority: Minor > > There is a batch of compilation warnings in Scala 2.13 as follows: > {code:java} > [WARNING] [Warn] > /spark/core/src/main/scala/org/apache/spark/input/FixedLengthBinaryInputFormat.scala:77: > Widening conversion from Long to Double is deprecated because it loses > precision. Write `.toDouble` instead. > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-33355) Batch fix compilation warnings about "method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is deprecated (since 2.13.0)"
[ https://issues.apache.org/jira/browse/SPARK-33355?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen updated SPARK-33355: - Parent: (was: SPARK-25075) Issue Type: Improvement (was: Sub-task) > Batch fix compilation warnings about "method copyArrayToImmutableIndexedSeq > in class LowPriorityImplicits2 is deprecated (since 2.13.0)" > > > Key: SPARK-33355 > URL: https://issues.apache.org/jira/browse/SPARK-33355 > Project: Spark > Issue Type: Improvement > Components: Build >Affects Versions: 3.1.0 >Reporter: Yang Jie >Priority: Minor > > compilation warnings as follow: > {code:java} > [WARNING] [Warn] > /spark/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/Cast.scala:1711: > method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is > deprecated (since 2.13.0): Implicit conversions from Array to > immutable.IndexedSeq are implemented by copying; Use the more efficient > non-copying ArraySeq.unsafeWrapArray or an explicit toIndexedSeq call > {code} > this Jira is a placeholder now, wait until Scala 2.12 is no longer supported -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33348) Use scala.jdk.CollectionConverters replace scala.collection.JavaConverters
[ https://issues.apache.org/jira/browse/SPARK-33348?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244607#comment-17244607 ] Sean R. Owen commented on SPARK-33348: -- Yes, this isn't required for 2.13, and can't be done for 2.12, right? > Use scala.jdk.CollectionConverters replace scala.collection.JavaConverters > -- > > Key: SPARK-33348 > URL: https://issues.apache.org/jira/browse/SPARK-33348 > Project: Spark > Issue Type: Sub-task > Components: Build >Affects Versions: 3.1.0 >Reporter: Yang Jie >Priority: Minor > > `scala.collection.JavaConverters` is deprecated in Scala 2.13, there are many > compilation warnings about this, should use `scala.jdk.CollectionConverters` > replace it. > But `scala.jdk.CollectionConverters` only available in Scala 2.13. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-33344) Fix Compilation warnings of "multiarg infix syntax looks like a tuple and will be deprecated" in Scala 2.13
[ https://issues.apache.org/jira/browse/SPARK-33344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen updated SPARK-33344: - Summary: Fix Compilation warnings of "multiarg infix syntax looks like a tuple and will be deprecated" in Scala 2.13 (was: Fix Compilation warings of "multiarg infix syntax looks like a tuple and will be deprecated" in Scala 2.13) Because this isn't required for 2.13, I moved it out of the umbrella > Fix Compilation warnings of "multiarg infix syntax looks like a tuple and > will be deprecated" in Scala 2.13 > --- > > Key: SPARK-33344 > URL: https://issues.apache.org/jira/browse/SPARK-33344 > Project: Spark > Issue Type: Improvement > Components: Build >Affects Versions: 3.1.0 >Reporter: Yang Jie >Priority: Minor > > There is a batch of compilation warnings in Scala 2.13 as follow: > {code:java} > [WARNING] [Warn] > /spark/core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala:656: > multiarg infix syntax looks like a tuple and will be deprecated > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-33344) Fix Compilation warings of "multiarg infix syntax looks like a tuple and will be deprecated" in Scala 2.13
[ https://issues.apache.org/jira/browse/SPARK-33344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen updated SPARK-33344: - Parent: (was: SPARK-25075) Issue Type: Improvement (was: Sub-task) > Fix Compilation warings of "multiarg infix syntax looks like a tuple and will > be deprecated" in Scala 2.13 > -- > > Key: SPARK-33344 > URL: https://issues.apache.org/jira/browse/SPARK-33344 > Project: Spark > Issue Type: Improvement > Components: Build >Affects Versions: 3.1.0 >Reporter: Yang Jie >Priority: Minor > > There is a batch of compilation warnings in Scala 2.13 as follow: > {code:java} > [WARNING] [Warn] > /spark/core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala:656: > multiarg infix syntax looks like a tuple and will be deprecated > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33285) Too many "Auto-application to `()` is deprecated." related compilation warnings
[ https://issues.apache.org/jira/browse/SPARK-33285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244605#comment-17244605 ] Sean R. Owen commented on SPARK-33285: -- I moved this out from the 2.13 umbrella as it is not required for 2.13 > Too many "Auto-application to `()` is deprecated." related compilation > warnings > > > Key: SPARK-33285 > URL: https://issues.apache.org/jira/browse/SPARK-33285 > Project: Spark > Issue Type: Improvement > Components: Build >Affects Versions: 3.1.0 >Reporter: Yang Jie >Priority: Minor > > There are too many "Auto-application to `()` is deprecated." related > compilation warnings when compile with Scala 2.13 like > {code:java} > [WARNING] [Warn] > /spark-src/core/src/test/scala/org/apache/spark/PartitioningSuite.scala:246: > Auto-application to `()` is deprecated. Supply the empty argument list `()` > explicitly to invoke method stdev, > or remove the empty argument list from its definition (Java-defined methods > are exempt). > In Scala 3, an unapplied method like this will be eta-expanded into a > function. > {code} > A lot of them, but it's easy to fix. > If there is a definition as follows: > {code:java} > Class Foo { >def bar(): Unit = {} > } > val foo = new Foo{code} > Should be > {code:java} > foo.bar() > {code} > not > {code:java} > foo.bar {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-33285) Too many "Auto-application to `()` is deprecated." related compilation warnings
[ https://issues.apache.org/jira/browse/SPARK-33285?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen updated SPARK-33285: - Parent: (was: SPARK-25075) Issue Type: Improvement (was: Sub-task) > Too many "Auto-application to `()` is deprecated." related compilation > warnings > > > Key: SPARK-33285 > URL: https://issues.apache.org/jira/browse/SPARK-33285 > Project: Spark > Issue Type: Improvement > Components: Build >Affects Versions: 3.1.0 >Reporter: Yang Jie >Priority: Minor > > There are too many "Auto-application to `()` is deprecated." related > compilation warnings when compile with Scala 2.13 like > {code:java} > [WARNING] [Warn] > /spark-src/core/src/test/scala/org/apache/spark/PartitioningSuite.scala:246: > Auto-application to `()` is deprecated. Supply the empty argument list `()` > explicitly to invoke method stdev, > or remove the empty argument list from its definition (Java-defined methods > are exempt). > In Scala 3, an unapplied method like this will be eta-expanded into a > function. > {code} > A lot of them, but it's easy to fix. > If there is a definition as follows: > {code:java} > Class Foo { >def bar(): Unit = {} > } > val foo = new Foo{code} > Should be > {code:java} > foo.bar() > {code} > not > {code:java} > foo.bar {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33044) Add a Jenkins build and test job for Scala 2.13
[ https://issues.apache.org/jira/browse/SPARK-33044?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244604#comment-17244604 ] Sean R. Owen commented on SPARK-33044: -- [~dongjoon] This is done now right? > Add a Jenkins build and test job for Scala 2.13 > --- > > Key: SPARK-33044 > URL: https://issues.apache.org/jira/browse/SPARK-33044 > Project: Spark > Issue Type: Sub-task > Components: jenkins >Affects Versions: 3.1.0 >Reporter: Yang Jie >Priority: Major > > {{Master}} branch seems to be almost ready for Scala 2.13 now, we need a > Jenkins test job to verify current work results and CI. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-32348) Get tests working for Scala 2.13 build
[ https://issues.apache.org/jira/browse/SPARK-32348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen resolved SPARK-32348. -- Fix Version/s: 3.1.0 Resolution: Fixed This should be largely done now, or at least, superseded by a few much more specific issues > Get tests working for Scala 2.13 build > -- > > Key: SPARK-32348 > URL: https://issues.apache.org/jira/browse/SPARK-32348 > Project: Spark > Issue Type: Sub-task > Components: ML, Spark Core, SQL, Structured Streaming >Affects Versions: 3.0.0 >Reporter: Sean R. Owen >Assignee: Sean R. Owen >Priority: Major > Fix For: 3.1.0 > > > This is a placeholder for the general task of getting the tests to pass in > the Scala 2.13 build, after it compiles. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33672) Check SQLContext.tables() for V2 session catalog
[ https://issues.apache.org/jira/browse/SPARK-33672?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244542#comment-17244542 ] Maxim Gekk commented on SPARK-33672: [~cloud_fan] FYI > Check SQLContext.tables() for V2 session catalog > > > Key: SPARK-33672 > URL: https://issues.apache.org/jira/browse/SPARK-33672 > Project: Spark > Issue Type: Sub-task > Components: SQL >Affects Versions: 3.1.0 >Reporter: Maxim Gekk >Priority: Major > > V1 ShowTablesCommand is hard coded in SQLContext: > https://github.com/apache/spark/blob/a088a801ed8c17171545c196a3f26ce415de0cd1/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala#L671 > The ticket aims to checks tables() behavior for V2 session catalog. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Created] (SPARK-33672) Check SQLContext.tables() for V2 session catalog
Maxim Gekk created SPARK-33672: -- Summary: Check SQLContext.tables() for V2 session catalog Key: SPARK-33672 URL: https://issues.apache.org/jira/browse/SPARK-33672 Project: Spark Issue Type: Sub-task Components: SQL Affects Versions: 3.1.0 Reporter: Maxim Gekk V1 ShowTablesCommand is hard coded in SQLContext: https://github.com/apache/spark/blob/a088a801ed8c17171545c196a3f26ce415de0cd1/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala#L671 The ticket aims to checks tables() behavior for V2 session catalog. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33671) Remove VIEW checks from V1 table commands
[ https://issues.apache.org/jira/browse/SPARK-33671?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244538#comment-17244538 ] Apache Spark commented on SPARK-33671: -- User 'MaxGekk' has created a pull request for this issue: https://github.com/apache/spark/pull/30620 > Remove VIEW checks from V1 table commands > - > > Key: SPARK-33671 > URL: https://issues.apache.org/jira/browse/SPARK-33671 > Project: Spark > Issue Type: Sub-task > Components: SQL >Affects Versions: 3.0.2, 3.1.0 >Reporter: Maxim Gekk >Priority: Major > > Checking of VIEWs is performed earlier, see > https://github.com/apache/spark/pull/30461 . So, the checks can be removed > from some V1 commands. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33671) Remove VIEW checks from V1 table commands
[ https://issues.apache.org/jira/browse/SPARK-33671?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244536#comment-17244536 ] Apache Spark commented on SPARK-33671: -- User 'MaxGekk' has created a pull request for this issue: https://github.com/apache/spark/pull/30620 > Remove VIEW checks from V1 table commands > - > > Key: SPARK-33671 > URL: https://issues.apache.org/jira/browse/SPARK-33671 > Project: Spark > Issue Type: Sub-task > Components: SQL >Affects Versions: 3.0.2, 3.1.0 >Reporter: Maxim Gekk >Priority: Major > > Checking of VIEWs is performed earlier, see > https://github.com/apache/spark/pull/30461 . So, the checks can be removed > from some V1 commands. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-33671) Remove VIEW checks from V1 table commands
[ https://issues.apache.org/jira/browse/SPARK-33671?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-33671: Assignee: (was: Apache Spark) > Remove VIEW checks from V1 table commands > - > > Key: SPARK-33671 > URL: https://issues.apache.org/jira/browse/SPARK-33671 > Project: Spark > Issue Type: Sub-task > Components: SQL >Affects Versions: 3.0.2, 3.1.0 >Reporter: Maxim Gekk >Priority: Major > > Checking of VIEWs is performed earlier, see > https://github.com/apache/spark/pull/30461 . So, the checks can be removed > from some V1 commands. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-33671) Remove VIEW checks from V1 table commands
[ https://issues.apache.org/jira/browse/SPARK-33671?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-33671: Assignee: Apache Spark > Remove VIEW checks from V1 table commands > - > > Key: SPARK-33671 > URL: https://issues.apache.org/jira/browse/SPARK-33671 > Project: Spark > Issue Type: Sub-task > Components: SQL >Affects Versions: 3.0.2, 3.1.0 >Reporter: Maxim Gekk >Assignee: Apache Spark >Priority: Major > > Checking of VIEWs is performed earlier, see > https://github.com/apache/spark/pull/30461 . So, the checks can be removed > from some V1 commands. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Created] (SPARK-33671) Remove VIEW checks from V1 table commands
Maxim Gekk created SPARK-33671: -- Summary: Remove VIEW checks from V1 table commands Key: SPARK-33671 URL: https://issues.apache.org/jira/browse/SPARK-33671 Project: Spark Issue Type: Sub-task Components: SQL Affects Versions: 3.0.2, 3.1.0 Reporter: Maxim Gekk Checking of VIEWs is performed earlier, see https://github.com/apache/spark/pull/30461 . So, the checks can be removed from some V1 commands. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-31711) Register the executor source with the metrics system when running in local mode.
[ https://issues.apache.org/jira/browse/SPARK-31711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244497#comment-17244497 ] Apache Spark commented on SPARK-31711: -- User 'LucaCanali' has created a pull request for this issue: https://github.com/apache/spark/pull/30619 > Register the executor source with the metrics system when running in local > mode. > > > Key: SPARK-31711 > URL: https://issues.apache.org/jira/browse/SPARK-31711 > Project: Spark > Issue Type: Improvement > Components: Spark Core >Affects Versions: 3.0.0 >Reporter: Luca Canali >Assignee: Luca Canali >Priority: Minor > Fix For: 3.1.0 > > > The Apache Spark metrics system provides many useful insights on the Spark > workload. In particular, the executor source metrics > (https://github.com/apache/spark/blob/master/docs/monitoring.md#component-instance--executor) > provide detailed info, including the number of active tasks, some I/O > metrics, and task metrics details. Executor source metrics, contrary to other > sources (for example ExecutorMetrics source), are not yet available when > running in local mode. > This JIRA proposes to register the executor source with the Spark metrics > system when running in local mode, as this can be very useful when testing > and troubleshooting Spark workloads. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-31711) Register the executor source with the metrics system when running in local mode.
[ https://issues.apache.org/jira/browse/SPARK-31711?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244496#comment-17244496 ] Apache Spark commented on SPARK-31711: -- User 'LucaCanali' has created a pull request for this issue: https://github.com/apache/spark/pull/30619 > Register the executor source with the metrics system when running in local > mode. > > > Key: SPARK-31711 > URL: https://issues.apache.org/jira/browse/SPARK-31711 > Project: Spark > Issue Type: Improvement > Components: Spark Core >Affects Versions: 3.0.0 >Reporter: Luca Canali >Assignee: Luca Canali >Priority: Minor > Fix For: 3.1.0 > > > The Apache Spark metrics system provides many useful insights on the Spark > workload. In particular, the executor source metrics > (https://github.com/apache/spark/blob/master/docs/monitoring.md#component-instance--executor) > provide detailed info, including the number of active tasks, some I/O > metrics, and task metrics details. Executor source metrics, contrary to other > sources (for example ExecutorMetrics source), are not yet available when > running in local mode. > This JIRA proposes to register the executor source with the Spark metrics > system when running in local mode, as this can be very useful when testing > and troubleshooting Spark workloads. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33670) Verify the partition provider is Hive in v1 SHOW TABLE EXTENDED
[ https://issues.apache.org/jira/browse/SPARK-33670?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244486#comment-17244486 ] Apache Spark commented on SPARK-33670: -- User 'MaxGekk' has created a pull request for this issue: https://github.com/apache/spark/pull/30618 > Verify the partition provider is Hive in v1 SHOW TABLE EXTENDED > --- > > Key: SPARK-33670 > URL: https://issues.apache.org/jira/browse/SPARK-33670 > Project: Spark > Issue Type: Sub-task > Components: SQL >Affects Versions: 3.0.2, 3.1.0 >Reporter: Maxim Gekk >Priority: Major > > Invoke the check verifyPartitionProviderIsHive() from v1 implementation of > SHOW TABLE EXTENDED. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-33670) Verify the partition provider is Hive in v1 SHOW TABLE EXTENDED
[ https://issues.apache.org/jira/browse/SPARK-33670?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-33670: Assignee: (was: Apache Spark) > Verify the partition provider is Hive in v1 SHOW TABLE EXTENDED > --- > > Key: SPARK-33670 > URL: https://issues.apache.org/jira/browse/SPARK-33670 > Project: Spark > Issue Type: Sub-task > Components: SQL >Affects Versions: 3.0.2, 3.1.0 >Reporter: Maxim Gekk >Priority: Major > > Invoke the check verifyPartitionProviderIsHive() from v1 implementation of > SHOW TABLE EXTENDED. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-33670) Verify the partition provider is Hive in v1 SHOW TABLE EXTENDED
[ https://issues.apache.org/jira/browse/SPARK-33670?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-33670: Assignee: Apache Spark > Verify the partition provider is Hive in v1 SHOW TABLE EXTENDED > --- > > Key: SPARK-33670 > URL: https://issues.apache.org/jira/browse/SPARK-33670 > Project: Spark > Issue Type: Sub-task > Components: SQL >Affects Versions: 3.0.2, 3.1.0 >Reporter: Maxim Gekk >Assignee: Apache Spark >Priority: Major > > Invoke the check verifyPartitionProviderIsHive() from v1 implementation of > SHOW TABLE EXTENDED. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-33670) Verify the partition provider is Hive in v1 SHOW TABLE EXTENDED
[ https://issues.apache.org/jira/browse/SPARK-33670?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-33670: Assignee: Apache Spark > Verify the partition provider is Hive in v1 SHOW TABLE EXTENDED > --- > > Key: SPARK-33670 > URL: https://issues.apache.org/jira/browse/SPARK-33670 > Project: Spark > Issue Type: Sub-task > Components: SQL >Affects Versions: 3.0.2, 3.1.0 >Reporter: Maxim Gekk >Assignee: Apache Spark >Priority: Major > > Invoke the check verifyPartitionProviderIsHive() from v1 implementation of > SHOW TABLE EXTENDED. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Created] (SPARK-33670) Verify the partition provider is Hive in v1 SHOW TABLE EXTENDED
Maxim Gekk created SPARK-33670: -- Summary: Verify the partition provider is Hive in v1 SHOW TABLE EXTENDED Key: SPARK-33670 URL: https://issues.apache.org/jira/browse/SPARK-33670 Project: Spark Issue Type: Sub-task Components: SQL Affects Versions: 3.0.2, 3.1.0 Reporter: Maxim Gekk Invoke the check verifyPartitionProviderIsHive() from v1 implementation of SHOW TABLE EXTENDED. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33669) Wrong error message from YARN application state monitor when sc.stop in yarn client mode
[ https://issues.apache.org/jira/browse/SPARK-33669?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244456#comment-17244456 ] Apache Spark commented on SPARK-33669: -- User 'sqlwindspeaker' has created a pull request for this issue: https://github.com/apache/spark/pull/30617 > Wrong error message from YARN application state monitor when sc.stop in yarn > client mode > > > Key: SPARK-33669 > URL: https://issues.apache.org/jira/browse/SPARK-33669 > Project: Spark > Issue Type: Bug > Components: YARN >Affects Versions: 2.4.3, 3.0.1 >Reporter: Su Qilong >Priority: Minor > > For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries > to interrupt Yarn application monitor thread. In MonitorThread.run() it > catches InterruptedException to gracefully response to stopping request. > But client.monitorApplication method also throws InterruptedIOException when > the hadoop rpc call is calling. In this case, MonitorThread will not know it > is interrupted, a Yarn App failed is returned with "Failed to contact YARN > for application x; YARN application has exited unexpectedly with state > x" is logged with error level. which confuse user a lot. > We Should take considerate InterruptedIOException here to make it the same > behavior with InterruptedException. > {code:java} > private class MonitorThread extends Thread { > private var allowInterrupt = true > override def run() { > try { > val YarnAppReport(_, state, diags) = > client.monitorApplication(appId.get, logApplicationReport = false) > logError(s"YARN application has exited unexpectedly with state $state! > " + > "Check the YARN application logs for more details.") > diags.foreach { err => > logError(s"Diagnostics message: $err") > } > allowInterrupt = false > sc.stop() > } catch { > case e: InterruptedException => logInfo("Interrupting monitor thread") > } > } > > {code} > {code:java} > // wrong error message > 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: > org.apache.spark.deploy.yarn.Client(91) - Failed to contact YARN for > application application_1605868815011_1154961. > java.io.InterruptedIOException: Call interrupted > at org.apache.hadoop.ipc.Client.call(Client.java:1466) > at org.apache.hadoop.ipc.Client.call(Client.java:1409) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230) > at com.sun.proxy.$Proxy38.getApplicationReport(Unknown Source) > at > org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getApplicationReport(ApplicationClientProtocolPBClientImpl.java:187) > at sun.reflect.GeneratedMethodAccessor22.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:256) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104) > at com.sun.proxy.$Proxy39.getApplicationReport(Unknown Source) > at > org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getApplicationReport(YarnClientImpl.java:408) > at > org.apache.spark.deploy.yarn.Client.getApplicationReport(Client.scala:327) > at > org.apache.spark.deploy.yarn.Client.monitorApplication(Client.scala:1039) > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:116) > 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - YARN > application has exited unexpectedly with state FAILED! Check the YARN > application logs for more details. > 2020-12-05 03:06:58,001 ERROR [YARN application state monitor]: > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - > Diagnostics message: Failed to contact YARN for application > application_1605868815011_1154961. > {code} > > {code:java} > // hadoop ipc code > public Writable call(RPC.RpcKind rpcKind, Writable rpcRequest, > ConnectionId remoteId, int serviceClass, > AtomicBoolean fallbackToSimpleAuth) throws IOException { > final Call call = createCall(rpcKind, rpcRequest); > Connection connection = getConnection(remoteId, call, serviceClass, > fallbackToSimpleAuth); > try { > connection.sendRpcRequest(call); // send the rpc request > } catch (RejectedExecutionException e) { > throw new IOException("connection has been
[jira] [Commented] (SPARK-33669) Wrong error message from YARN application state monitor when sc.stop in yarn client mode
[ https://issues.apache.org/jira/browse/SPARK-33669?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244455#comment-17244455 ] Apache Spark commented on SPARK-33669: -- User 'sqlwindspeaker' has created a pull request for this issue: https://github.com/apache/spark/pull/30617 > Wrong error message from YARN application state monitor when sc.stop in yarn > client mode > > > Key: SPARK-33669 > URL: https://issues.apache.org/jira/browse/SPARK-33669 > Project: Spark > Issue Type: Bug > Components: YARN >Affects Versions: 2.4.3, 3.0.1 >Reporter: Su Qilong >Priority: Minor > > For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries > to interrupt Yarn application monitor thread. In MonitorThread.run() it > catches InterruptedException to gracefully response to stopping request. > But client.monitorApplication method also throws InterruptedIOException when > the hadoop rpc call is calling. In this case, MonitorThread will not know it > is interrupted, a Yarn App failed is returned with "Failed to contact YARN > for application x; YARN application has exited unexpectedly with state > x" is logged with error level. which confuse user a lot. > We Should take considerate InterruptedIOException here to make it the same > behavior with InterruptedException. > {code:java} > private class MonitorThread extends Thread { > private var allowInterrupt = true > override def run() { > try { > val YarnAppReport(_, state, diags) = > client.monitorApplication(appId.get, logApplicationReport = false) > logError(s"YARN application has exited unexpectedly with state $state! > " + > "Check the YARN application logs for more details.") > diags.foreach { err => > logError(s"Diagnostics message: $err") > } > allowInterrupt = false > sc.stop() > } catch { > case e: InterruptedException => logInfo("Interrupting monitor thread") > } > } > > {code} > {code:java} > // wrong error message > 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: > org.apache.spark.deploy.yarn.Client(91) - Failed to contact YARN for > application application_1605868815011_1154961. > java.io.InterruptedIOException: Call interrupted > at org.apache.hadoop.ipc.Client.call(Client.java:1466) > at org.apache.hadoop.ipc.Client.call(Client.java:1409) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230) > at com.sun.proxy.$Proxy38.getApplicationReport(Unknown Source) > at > org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getApplicationReport(ApplicationClientProtocolPBClientImpl.java:187) > at sun.reflect.GeneratedMethodAccessor22.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:256) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104) > at com.sun.proxy.$Proxy39.getApplicationReport(Unknown Source) > at > org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getApplicationReport(YarnClientImpl.java:408) > at > org.apache.spark.deploy.yarn.Client.getApplicationReport(Client.scala:327) > at > org.apache.spark.deploy.yarn.Client.monitorApplication(Client.scala:1039) > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:116) > 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - YARN > application has exited unexpectedly with state FAILED! Check the YARN > application logs for more details. > 2020-12-05 03:06:58,001 ERROR [YARN application state monitor]: > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - > Diagnostics message: Failed to contact YARN for application > application_1605868815011_1154961. > {code} > > {code:java} > // hadoop ipc code > public Writable call(RPC.RpcKind rpcKind, Writable rpcRequest, > ConnectionId remoteId, int serviceClass, > AtomicBoolean fallbackToSimpleAuth) throws IOException { > final Call call = createCall(rpcKind, rpcRequest); > Connection connection = getConnection(remoteId, call, serviceClass, > fallbackToSimpleAuth); > try { > connection.sendRpcRequest(call); // send the rpc request > } catch (RejectedExecutionException e) { > throw new IOException("connection has been
[jira] [Assigned] (SPARK-33669) Wrong error message from YARN application state monitor when sc.stop in yarn client mode
[ https://issues.apache.org/jira/browse/SPARK-33669?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-33669: Assignee: (was: Apache Spark) > Wrong error message from YARN application state monitor when sc.stop in yarn > client mode > > > Key: SPARK-33669 > URL: https://issues.apache.org/jira/browse/SPARK-33669 > Project: Spark > Issue Type: Bug > Components: YARN >Affects Versions: 2.4.3, 3.0.1 >Reporter: Su Qilong >Priority: Minor > > For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries > to interrupt Yarn application monitor thread. In MonitorThread.run() it > catches InterruptedException to gracefully response to stopping request. > But client.monitorApplication method also throws InterruptedIOException when > the hadoop rpc call is calling. In this case, MonitorThread will not know it > is interrupted, a Yarn App failed is returned with "Failed to contact YARN > for application x; YARN application has exited unexpectedly with state > x" is logged with error level. which confuse user a lot. > We Should take considerate InterruptedIOException here to make it the same > behavior with InterruptedException. > {code:java} > private class MonitorThread extends Thread { > private var allowInterrupt = true > override def run() { > try { > val YarnAppReport(_, state, diags) = > client.monitorApplication(appId.get, logApplicationReport = false) > logError(s"YARN application has exited unexpectedly with state $state! > " + > "Check the YARN application logs for more details.") > diags.foreach { err => > logError(s"Diagnostics message: $err") > } > allowInterrupt = false > sc.stop() > } catch { > case e: InterruptedException => logInfo("Interrupting monitor thread") > } > } > > {code} > {code:java} > // wrong error message > 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: > org.apache.spark.deploy.yarn.Client(91) - Failed to contact YARN for > application application_1605868815011_1154961. > java.io.InterruptedIOException: Call interrupted > at org.apache.hadoop.ipc.Client.call(Client.java:1466) > at org.apache.hadoop.ipc.Client.call(Client.java:1409) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230) > at com.sun.proxy.$Proxy38.getApplicationReport(Unknown Source) > at > org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getApplicationReport(ApplicationClientProtocolPBClientImpl.java:187) > at sun.reflect.GeneratedMethodAccessor22.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:256) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104) > at com.sun.proxy.$Proxy39.getApplicationReport(Unknown Source) > at > org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getApplicationReport(YarnClientImpl.java:408) > at > org.apache.spark.deploy.yarn.Client.getApplicationReport(Client.scala:327) > at > org.apache.spark.deploy.yarn.Client.monitorApplication(Client.scala:1039) > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:116) > 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - YARN > application has exited unexpectedly with state FAILED! Check the YARN > application logs for more details. > 2020-12-05 03:06:58,001 ERROR [YARN application state monitor]: > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - > Diagnostics message: Failed to contact YARN for application > application_1605868815011_1154961. > {code} > > {code:java} > // hadoop ipc code > public Writable call(RPC.RpcKind rpcKind, Writable rpcRequest, > ConnectionId remoteId, int serviceClass, > AtomicBoolean fallbackToSimpleAuth) throws IOException { > final Call call = createCall(rpcKind, rpcRequest); > Connection connection = getConnection(remoteId, call, serviceClass, > fallbackToSimpleAuth); > try { > connection.sendRpcRequest(call); // send the rpc request > } catch (RejectedExecutionException e) { > throw new IOException("connection has been closed", e); > } catch (InterruptedException e) { > Thread.currentThread().interrupt(); > LOG.warn("interrupte
[jira] [Assigned] (SPARK-33669) Wrong error message from YARN application state monitor when sc.stop in yarn client mode
[ https://issues.apache.org/jira/browse/SPARK-33669?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-33669: Assignee: Apache Spark > Wrong error message from YARN application state monitor when sc.stop in yarn > client mode > > > Key: SPARK-33669 > URL: https://issues.apache.org/jira/browse/SPARK-33669 > Project: Spark > Issue Type: Bug > Components: YARN >Affects Versions: 2.4.3, 3.0.1 >Reporter: Su Qilong >Assignee: Apache Spark >Priority: Minor > > For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries > to interrupt Yarn application monitor thread. In MonitorThread.run() it > catches InterruptedException to gracefully response to stopping request. > But client.monitorApplication method also throws InterruptedIOException when > the hadoop rpc call is calling. In this case, MonitorThread will not know it > is interrupted, a Yarn App failed is returned with "Failed to contact YARN > for application x; YARN application has exited unexpectedly with state > x" is logged with error level. which confuse user a lot. > We Should take considerate InterruptedIOException here to make it the same > behavior with InterruptedException. > {code:java} > private class MonitorThread extends Thread { > private var allowInterrupt = true > override def run() { > try { > val YarnAppReport(_, state, diags) = > client.monitorApplication(appId.get, logApplicationReport = false) > logError(s"YARN application has exited unexpectedly with state $state! > " + > "Check the YARN application logs for more details.") > diags.foreach { err => > logError(s"Diagnostics message: $err") > } > allowInterrupt = false > sc.stop() > } catch { > case e: InterruptedException => logInfo("Interrupting monitor thread") > } > } > > {code} > {code:java} > // wrong error message > 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: > org.apache.spark.deploy.yarn.Client(91) - Failed to contact YARN for > application application_1605868815011_1154961. > java.io.InterruptedIOException: Call interrupted > at org.apache.hadoop.ipc.Client.call(Client.java:1466) > at org.apache.hadoop.ipc.Client.call(Client.java:1409) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230) > at com.sun.proxy.$Proxy38.getApplicationReport(Unknown Source) > at > org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getApplicationReport(ApplicationClientProtocolPBClientImpl.java:187) > at sun.reflect.GeneratedMethodAccessor22.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:256) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104) > at com.sun.proxy.$Proxy39.getApplicationReport(Unknown Source) > at > org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getApplicationReport(YarnClientImpl.java:408) > at > org.apache.spark.deploy.yarn.Client.getApplicationReport(Client.scala:327) > at > org.apache.spark.deploy.yarn.Client.monitorApplication(Client.scala:1039) > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:116) > 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - YARN > application has exited unexpectedly with state FAILED! Check the YARN > application logs for more details. > 2020-12-05 03:06:58,001 ERROR [YARN application state monitor]: > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - > Diagnostics message: Failed to contact YARN for application > application_1605868815011_1154961. > {code} > > {code:java} > // hadoop ipc code > public Writable call(RPC.RpcKind rpcKind, Writable rpcRequest, > ConnectionId remoteId, int serviceClass, > AtomicBoolean fallbackToSimpleAuth) throws IOException { > final Call call = createCall(rpcKind, rpcRequest); > Connection connection = getConnection(remoteId, call, serviceClass, > fallbackToSimpleAuth); > try { > connection.sendRpcRequest(call); // send the rpc request > } catch (RejectedExecutionException e) { > throw new IOException("connection has been closed", e); > } catch (InterruptedException e) { > Thread.currentThread().interrupt(); >
[jira] [Updated] (SPARK-33669) Wrong error message from YARN application state monitor when sc.stop in yarn client mode
[ https://issues.apache.org/jira/browse/SPARK-33669?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Su Qilong updated SPARK-33669: -- Description: For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries to interrupt Yarn application monitor thread. In MonitorThread.run() it catches InterruptedException to gracefully response to stopping request. But client.monitorApplication method also throws InterruptedIOException when the hadoop rpc call is calling. In this case, MonitorThread will not know it is interrupted, a Yarn App failed is returned with "Failed to contact YARN for application x; YARN application has exited unexpectedly with state x" is logged with error level. which confuse user a lot. We Should take considerate InterruptedIOException here to make it the same behavior with InterruptedException. {code:java} private class MonitorThread extends Thread { private var allowInterrupt = true override def run() { try { val YarnAppReport(_, state, diags) = client.monitorApplication(appId.get, logApplicationReport = false) logError(s"YARN application has exited unexpectedly with state $state! " + "Check the YARN application logs for more details.") diags.foreach { err => logError(s"Diagnostics message: $err") } allowInterrupt = false sc.stop() } catch { case e: InterruptedException => logInfo("Interrupting monitor thread") } } {code} {code:java} // wrong error message 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: org.apache.spark.deploy.yarn.Client(91) - Failed to contact YARN for application application_1605868815011_1154961. java.io.InterruptedIOException: Call interrupted at org.apache.hadoop.ipc.Client.call(Client.java:1466) at org.apache.hadoop.ipc.Client.call(Client.java:1409) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230) at com.sun.proxy.$Proxy38.getApplicationReport(Unknown Source) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getApplicationReport(ApplicationClientProtocolPBClientImpl.java:187) at sun.reflect.GeneratedMethodAccessor22.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:256) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104) at com.sun.proxy.$Proxy39.getApplicationReport(Unknown Source) at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getApplicationReport(YarnClientImpl.java:408) at org.apache.spark.deploy.yarn.Client.getApplicationReport(Client.scala:327) at org.apache.spark.deploy.yarn.Client.monitorApplication(Client.scala:1039) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:116) 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - YARN application has exited unexpectedly with state FAILED! Check the YARN application logs for more details. 2020-12-05 03:06:58,001 ERROR [YARN application state monitor]: org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - Diagnostics message: Failed to contact YARN for application application_1605868815011_1154961. {code} {code:java} // hadoop ipc code public Writable call(RPC.RpcKind rpcKind, Writable rpcRequest, ConnectionId remoteId, int serviceClass, AtomicBoolean fallbackToSimpleAuth) throws IOException { final Call call = createCall(rpcKind, rpcRequest); Connection connection = getConnection(remoteId, call, serviceClass, fallbackToSimpleAuth); try { connection.sendRpcRequest(call); // send the rpc request } catch (RejectedExecutionException e) { throw new IOException("connection has been closed", e); } catch (InterruptedException e) { Thread.currentThread().interrupt(); LOG.warn("interrupted waiting to send rpc request to server", e); throw new IOException(e); } synchronized (call) { while (!call.done) { try { call.wait(); // wait for the result } catch (InterruptedException ie) { Thread.currentThread().interrupt(); throw new InterruptedIOException("Call interrupted"); } } {code} was: For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries to interrupt Yarn application monitor thread. In MonitorThread.run() it catches InterruptedException to gracefully response to stopping request. But client.monitorApplication method also throws InterruptedIOE
[jira] [Updated] (SPARK-33669) Wrong error message from YARN application state monitor when sc.stop in yarn client mode
[ https://issues.apache.org/jira/browse/SPARK-33669?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Su Qilong updated SPARK-33669: -- Description: For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries to interrupt Yarn application monitor thread. In MonitorThread.run() it catches InterruptedException to gracefully response to stopping request. But client.monitorApplication method also throws InterruptedIOException when the hadoop rpc call is calling. In this case, MonitorThread will not know it is interrupted, a Yarn App failed is returned with "Failed to contact YARN for application x; YARN application has exited unexpectedly with state x" is logged with error level. which confuse user a lot. We Should take considerate InterruptedIOException here to make it the same behavior with InterruptedException. {code:java} private class MonitorThread extends Thread { private var allowInterrupt = true override def run() { try { val YarnAppReport(_, state, diags) = client.monitorApplication(appId.get, logApplicationReport = false) logError(s"YARN application has exited unexpectedly with state $state! " + "Check the YARN application logs for more details.") diags.foreach { err => logError(s"Diagnostics message: $err") } allowInterrupt = false sc.stop() } catch { case e: InterruptedException => logInfo("Interrupting monitor thread") } } {code} {code:java} 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: org.apache.spark.deploy.yarn.Client(91) - Failed to contact YARN for application application_1605868815011_1154961. java.io.InterruptedIOException: Call interrupted at org.apache.hadoop.ipc.Client.call(Client.java:1466) at org.apache.hadoop.ipc.Client.call(Client.java:1409) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230) at com.sun.proxy.$Proxy38.getApplicationReport(Unknown Source) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getApplicationReport(ApplicationClientProtocolPBClientImpl.java:187) at sun.reflect.GeneratedMethodAccessor22.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:256) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104) at com.sun.proxy.$Proxy39.getApplicationReport(Unknown Source) at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getApplicationReport(YarnClientImpl.java:408) at org.apache.spark.deploy.yarn.Client.getApplicationReport(Client.scala:327) at org.apache.spark.deploy.yarn.Client.monitorApplication(Client.scala:1039) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:116) 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - YARN application has exited unexpectedly with state FAILED! Check the YARN application logs for more details. 2020-12-05 03:06:58,001 ERROR [YARN application state monitor]: org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - Diagnostics message: Failed to contact YARN for application application_1605868815011_1154961. {code} {code:java} // hadoop ipc code public Writable call(RPC.RpcKind rpcKind, Writable rpcRequest, ConnectionId remoteId, int serviceClass, AtomicBoolean fallbackToSimpleAuth) throws IOException { final Call call = createCall(rpcKind, rpcRequest); Connection connection = getConnection(remoteId, call, serviceClass, fallbackToSimpleAuth); try { connection.sendRpcRequest(call); // send the rpc request } catch (RejectedExecutionException e) { throw new IOException("connection has been closed", e); } catch (InterruptedException e) { Thread.currentThread().interrupt(); LOG.warn("interrupted waiting to send rpc request to server", e); throw new IOException(e); } synchronized (call) { while (!call.done) { try { call.wait(); // wait for the result } catch (InterruptedException ie) { Thread.currentThread().interrupt(); throw new InterruptedIOException("Call interrupted"); } } {code} was: For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries to interrupt Yarn application monitor thread. In MonitorThread.run() it catches InterruptedException to gracefully response to stopping request. But client.monitorApplication method also throws InterruptedIOException when the hado
[jira] [Updated] (SPARK-33669) Wrong error message from YARN application state monitor when sc.stop in yarn client mode
[ https://issues.apache.org/jira/browse/SPARK-33669?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Su Qilong updated SPARK-33669: -- Description: For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries to interrupt Yarn application monitor thread. In MonitorThread.run() it catches InterruptedException to gracefully response to stopping request. But client.monitorApplication method also throws InterruptedIOException when the hadoop rpc call is calling. In this case, MonitorThread will not know it is interrupted, a Yarn App failed is returned with "Failed to contact YARN for application x; YARN application has exited unexpectedly with state x" is logged with error level. which confuse user a lot. We Should take considerate InterruptedIOException here to make it the same behavior with InterruptedException. {code:java} private class MonitorThread extends Thread { private var allowInterrupt = true override def run() { try { val YarnAppReport(_, state, diags) = client.monitorApplication(appId.get, logApplicationReport = false) logError(s"YARN application has exited unexpectedly with state $state! " + "Check the YARN application logs for more details.") diags.foreach { err => logError(s"Diagnostics message: $err") } allowInterrupt = false sc.stop() } catch { case e: InterruptedException => logInfo("Interrupting monitor thread") } } {code} {code:java} 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: org.apache.spark.deploy.yarn.Client(91) - Failed to contact YARN for application application_1605868815011_1154961. - sessionId[99e46a14-7995-41da-ba0a-c4c7387728a4] java.io.InterruptedIOException: Call interrupted at org.apache.hadoop.ipc.Client.call(Client.java:1466) at org.apache.hadoop.ipc.Client.call(Client.java:1409) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230) at com.sun.proxy.$Proxy38.getApplicationReport(Unknown Source) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getApplicationReport(ApplicationClientProtocolPBClientImpl.java:187) at sun.reflect.GeneratedMethodAccessor22.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:256) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104) at com.sun.proxy.$Proxy39.getApplicationReport(Unknown Source) at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getApplicationReport(YarnClientImpl.java:408) at org.apache.spark.deploy.yarn.Client.getApplicationReport(Client.scala:327) at org.apache.spark.deploy.yarn.Client.monitorApplication(Client.scala:1039) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:116) 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - YARN application has exited unexpectedly with state FAILED! Check the YARN application logs for more details. - sessionId[99e46a14-7995-41da-ba0a-c4c7387728a4] 2020-12-05 03:06:58,000 INFO [FrontendService-Handler-Pool: Thread-6560]: org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(54) - Shutting down all executors - sessionId[99e46a14-7995-41da-ba0a-c4c7387728a4] 2020-12-05 03:06:58,001 ERROR [YARN application state monitor]: org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - Diagnostics message: Failed to contact YARN for application application_1605868815011_1154961. {code} was: For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries to interrupt Yarn application monitor thread. In MonitorThread.run() it catches InterruptedException to gracefully response to stopping request. But client.monitorApplication method also throws InterruptedIOException when the hadoop rpc call is calling. In this case, MonitorThread will not know it is interrupted, a Yarn App failed is returned with "Failed to contact YARN for application x; YARN application has exited unexpectedly with state x" is logged with error level. which confuse user a lot. We Should take considerate InterruptedIOException here to make it the same behavior with InterruptedException. {code:java} private class MonitorThread extends Thread { private var allowInterrupt = true override def run() { try { val YarnAppReport(_, state, diags) = client.monitorApplication(appId.get, logApplicationReport = false) logError(s"YARN application has exited unexpectedly with state $state! " + "Check the YARN application logs for more details.") diags.foreach { err =>
[jira] [Updated] (SPARK-33669) Wrong error message from YARN application state monitor when sc.stop in yarn client mode
[ https://issues.apache.org/jira/browse/SPARK-33669?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Su Qilong updated SPARK-33669: -- Description: For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries to interrupt Yarn application monitor thread. In MonitorThread.run() it catches InterruptedException to gracefully response to stopping request. But client.monitorApplication method also throws InterruptedIOException when the hadoop rpc call is calling. In this case, MonitorThread will not know it is interrupted, a Yarn App failed is returned with "Failed to contact YARN for application x; YARN application has exited unexpectedly with state x" is logged with error level. which confuse user a lot. We Should take considerate InterruptedIOException here to make it the same behavior with InterruptedException. {code:java} private class MonitorThread extends Thread { private var allowInterrupt = true override def run() { try { val YarnAppReport(_, state, diags) = client.monitorApplication(appId.get, logApplicationReport = false) logError(s"YARN application has exited unexpectedly with state $state! " + "Check the YARN application logs for more details.") diags.foreach { err => logError(s"Diagnostics message: $err") } allowInterrupt = false sc.stop() } catch { case e: InterruptedException => logInfo("Interrupting monitor thread") } } {code} {noformat} // error message 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: org.apache.spark.deploy.yarn.Client(91) - Failed to contact YARN for application application_1605868815011_1154961. java.io.InterruptedIOException: Call interrupted at org.apache.hadoop.ipc.Client.call(Client.java:1466) at org.apache.hadoop.ipc.Client.call(Client.java:1409) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230) at com.sun.proxy.$Proxy38.getApplicationReport(Unknown Source) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getApplicationReport(ApplicationClientProtocolPBClientImpl.java:187) at sun.reflect.GeneratedMethodAccessor22.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:256) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104) at com.sun.proxy.$Proxy39.getApplicationReport(Unknown Source) at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getApplicationReport(YarnClientImpl.java:408) at org.apache.spark.deploy.yarn.Client.getApplicationReport(Client.scala:327) at org.apache.spark.deploy.yarn.Client.monitorApplication(Client.scala:1039) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:116) 2020-12-05 03:06:58,000 ERROR [YARN application state monitor]: org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - YARN application has exited unexpectedly with state FAILED! Check the YARN application logs for more details. 2020-12-05 03:06:58,001 ERROR [YARN application state monitor]: org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend(70) - Diagnostics message: Failed to contact YARN for application application_1605868815011_1154961. {noformat} was: For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries to interrupt Yarn application monitor thread. In MonitorThread.run() it catches InterruptedException to gracefully response to stopping request. But client.monitorApplication method also throws InterruptedIOException when the hadoop rpc call is calling. In this case, MonitorThread will not know it is interrupted, a Yarn App failed is returned with "Failed to contact YARN for application x; YARN application has exited unexpectedly with state x" is logged with error level. which confuse user a lot. We Should take considerate InterruptedIOException here to make it the same behavior with InterruptedException. {code:java} // code placeholder private class MonitorThread extends Thread { private var allowInterrupt = true override def run() { try { val YarnAppReport(_, state, diags) = client.monitorApplication(appId.get, logApplicationReport = false) logError(s"YARN application has exited unexpectedly with state $state! " + "Check the YARN application logs for more details.") diags.foreach { err => logError(s"Diagnostics message: $err") } allowInterrupt = false sc.stop() } catch { case e: InterruptedException => logInfo("Interrupting monitor thread") } } {code} > Wrong error message from YARN application state monitor when sc.stop in y
[jira] [Updated] (SPARK-33669) Wrong error message from YARN application state monitor when sc.stop in yarn client mode
[ https://issues.apache.org/jira/browse/SPARK-33669?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Su Qilong updated SPARK-33669: -- Description: For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries to interrupt Yarn application monitor thread. In MonitorThread.run() it catches InterruptedException to gracefully response to stopping request. But client.monitorApplication method also throws InterruptedIOException when the hadoop rpc call is calling. In this case, MonitorThread will not know it is interrupted, a Yarn App failed is returned with "Failed to contact YARN for application x; YARN application has exited unexpectedly with state x" is logged with error level. which confuse user a lot. We Should take considerate InterruptedIOException here to make it the same behavior with InterruptedException. {code:java} // code placeholder private class MonitorThread extends Thread { private var allowInterrupt = true override def run() { try { val YarnAppReport(_, state, diags) = client.monitorApplication(appId.get, logApplicationReport = false) logError(s"YARN application has exited unexpectedly with state $state! " + "Check the YARN application logs for more details.") diags.foreach { err => logError(s"Diagnostics message: $err") } allowInterrupt = false sc.stop() } catch { case e: InterruptedException => logInfo("Interrupting monitor thread") } } {code} was: For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries to interrupt Yarn application monitor thread. In MonitorThread.run() it catches InterruptedException to gracefully response to stopping request. But client.monitorApplication method also throws InterruptedIOException when the hadoop rpc call is calling. In this case, MonitorThread will not know it is interrupted, a Yarn App failed is returned with "Failed to contact YARN for application x; YARN application has exited unexpectedly with state x" is logged with error level. which confuse user a lot. We Should take considerate InterruptedIOException here to make it the same behavior with InterruptedException. > Wrong error message from YARN application state monitor when sc.stop in yarn > client mode > > > Key: SPARK-33669 > URL: https://issues.apache.org/jira/browse/SPARK-33669 > Project: Spark > Issue Type: Bug > Components: YARN >Affects Versions: 2.4.3, 3.0.1 >Reporter: Su Qilong >Priority: Minor > > For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries > to interrupt Yarn application monitor thread. In MonitorThread.run() it > catches InterruptedException to gracefully response to stopping request. > But client.monitorApplication method also throws InterruptedIOException when > the hadoop rpc call is calling. In this case, MonitorThread will not know it > is interrupted, a Yarn App failed is returned with "Failed to contact YARN > for application x; YARN application has exited unexpectedly with state > x" is logged with error level. which confuse user a lot. > We Should take considerate InterruptedIOException here to make it the same > behavior with InterruptedException. > {code:java} > // code placeholder > private class MonitorThread extends Thread { > private var allowInterrupt = true > override def run() { > try { > val YarnAppReport(_, state, diags) = > client.monitorApplication(appId.get, logApplicationReport = false) > logError(s"YARN application has exited unexpectedly with state $state! > " + > "Check the YARN application logs for more details.") > diags.foreach { err => > logError(s"Diagnostics message: $err") > } > allowInterrupt = false > sc.stop() > } catch { > case e: InterruptedException => logInfo("Interrupting monitor thread") > } > } > {code} > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Created] (SPARK-33669) Wrong error message from YARN application state monitor when sc.stop in yarn client mode
Su Qilong created SPARK-33669: - Summary: Wrong error message from YARN application state monitor when sc.stop in yarn client mode Key: SPARK-33669 URL: https://issues.apache.org/jira/browse/SPARK-33669 Project: Spark Issue Type: Bug Components: YARN Affects Versions: 3.0.1, 2.4.3 Reporter: Su Qilong For YarnClient mode, when stopping YarnClientSchedulerBackend, it first tries to interrupt Yarn application monitor thread. In MonitorThread.run() it catches InterruptedException to gracefully response to stopping request. But client.monitorApplication method also throws InterruptedIOException when the hadoop rpc call is calling. In this case, MonitorThread will not know it is interrupted, a Yarn App failed is returned with "Failed to contact YARN for application x; YARN application has exited unexpectedly with state x" is logged with error level. which confuse user a lot. We Should take considerate InterruptedIOException here to make it the same behavior with InterruptedException. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-33668) Fix flaky test "Verify logging configuration is picked from the provided SPARK_CONF_DIR/log4j.properties."
[ https://issues.apache.org/jira/browse/SPARK-33668?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-33668: Assignee: (was: Apache Spark) > Fix flaky test "Verify logging configuration is picked from the provided > SPARK_CONF_DIR/log4j.properties." > -- > > Key: SPARK-33668 > URL: https://issues.apache.org/jira/browse/SPARK-33668 > Project: Spark > Issue Type: Bug > Components: Kubernetes, Tests >Affects Versions: 3.1.0 >Reporter: Prashant Sharma >Priority: Major > > The test is flaking, with multiple flaked instances - the reason for the > failure has been similar to: > {code:java} > The code passed to eventually never returned normally. Attempted 109 times > over 3.007988241397 minutes. Last failure message: Failure executing: GET > at: > https://192.168.39.167:8443/api/v1/namespaces/b37fc72a991b49baa68a2eaaa1516463/pods/spark-pi-97a9bc76308e7fe3-exec-1/log?pretty=false. > Message: pods "spark-pi-97a9bc76308e7fe3-exec-1" not found. Received status: > Status(apiVersion=v1, code=404, details=StatusDetails(causes=[], group=null, > kind=pods, name=spark-pi-97a9bc76308e7fe3-exec-1, retryAfterSeconds=null, > uid=null, additionalProperties={}), kind=Status, message=pods > "spark-pi-97a9bc76308e7fe3-exec-1" not found, > metadata=ListMeta(_continue=null, remainingItemCount=null, > resourceVersion=null, selfLink=null, additionalProperties={}), > reason=NotFound, status=Failure, additionalProperties={}).. > (KubernetesSuite.scala:402) > {code} > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36854/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36852/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36850/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36848/console > From the above failures, it seems, that executor finishes too quickly and is > removed by spark before the test can complete. > So, in order to mitigate this situation, one way is to turn on the flag > {code} >"spark.kubernetes.executor.deleteOnTermination" > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-33668) Fix flaky test "Verify logging configuration is picked from the provided SPARK_CONF_DIR/log4j.properties."
[ https://issues.apache.org/jira/browse/SPARK-33668?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Apache Spark reassigned SPARK-33668: Assignee: Apache Spark > Fix flaky test "Verify logging configuration is picked from the provided > SPARK_CONF_DIR/log4j.properties." > -- > > Key: SPARK-33668 > URL: https://issues.apache.org/jira/browse/SPARK-33668 > Project: Spark > Issue Type: Bug > Components: Kubernetes, Tests >Affects Versions: 3.1.0 >Reporter: Prashant Sharma >Assignee: Apache Spark >Priority: Major > > The test is flaking, with multiple flaked instances - the reason for the > failure has been similar to: > {code:java} > The code passed to eventually never returned normally. Attempted 109 times > over 3.007988241397 minutes. Last failure message: Failure executing: GET > at: > https://192.168.39.167:8443/api/v1/namespaces/b37fc72a991b49baa68a2eaaa1516463/pods/spark-pi-97a9bc76308e7fe3-exec-1/log?pretty=false. > Message: pods "spark-pi-97a9bc76308e7fe3-exec-1" not found. Received status: > Status(apiVersion=v1, code=404, details=StatusDetails(causes=[], group=null, > kind=pods, name=spark-pi-97a9bc76308e7fe3-exec-1, retryAfterSeconds=null, > uid=null, additionalProperties={}), kind=Status, message=pods > "spark-pi-97a9bc76308e7fe3-exec-1" not found, > metadata=ListMeta(_continue=null, remainingItemCount=null, > resourceVersion=null, selfLink=null, additionalProperties={}), > reason=NotFound, status=Failure, additionalProperties={}).. > (KubernetesSuite.scala:402) > {code} > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36854/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36852/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36850/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36848/console > From the above failures, it seems, that executor finishes too quickly and is > removed by spark before the test can complete. > So, in order to mitigate this situation, one way is to turn on the flag > {code} >"spark.kubernetes.executor.deleteOnTermination" > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33668) Fix flaky test "Verify logging configuration is picked from the provided SPARK_CONF_DIR/log4j.properties."
[ https://issues.apache.org/jira/browse/SPARK-33668?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244434#comment-17244434 ] Apache Spark commented on SPARK-33668: -- User 'ScrapCodes' has created a pull request for this issue: https://github.com/apache/spark/pull/30616 > Fix flaky test "Verify logging configuration is picked from the provided > SPARK_CONF_DIR/log4j.properties." > -- > > Key: SPARK-33668 > URL: https://issues.apache.org/jira/browse/SPARK-33668 > Project: Spark > Issue Type: Bug > Components: Kubernetes, Tests >Affects Versions: 3.1.0 >Reporter: Prashant Sharma >Priority: Major > > The test is flaking, with multiple flaked instances - the reason for the > failure has been similar to: > {code:java} > The code passed to eventually never returned normally. Attempted 109 times > over 3.007988241397 minutes. Last failure message: Failure executing: GET > at: > https://192.168.39.167:8443/api/v1/namespaces/b37fc72a991b49baa68a2eaaa1516463/pods/spark-pi-97a9bc76308e7fe3-exec-1/log?pretty=false. > Message: pods "spark-pi-97a9bc76308e7fe3-exec-1" not found. Received status: > Status(apiVersion=v1, code=404, details=StatusDetails(causes=[], group=null, > kind=pods, name=spark-pi-97a9bc76308e7fe3-exec-1, retryAfterSeconds=null, > uid=null, additionalProperties={}), kind=Status, message=pods > "spark-pi-97a9bc76308e7fe3-exec-1" not found, > metadata=ListMeta(_continue=null, remainingItemCount=null, > resourceVersion=null, selfLink=null, additionalProperties={}), > reason=NotFound, status=Failure, additionalProperties={}).. > (KubernetesSuite.scala:402) > {code} > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36854/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36852/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36850/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36848/console > From the above failures, it seems, that executor finishes too quickly and is > removed by spark before the test can complete. > So, in order to mitigate this situation, one way is to turn on the flag > {code} >"spark.kubernetes.executor.deleteOnTermination" > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-33668) Fix flaky test "Verify logging configuration is picked from the provided SPARK_CONF_DIR/log4j.properties."
[ https://issues.apache.org/jira/browse/SPARK-33668?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17244433#comment-17244433 ] Apache Spark commented on SPARK-33668: -- User 'ScrapCodes' has created a pull request for this issue: https://github.com/apache/spark/pull/30616 > Fix flaky test "Verify logging configuration is picked from the provided > SPARK_CONF_DIR/log4j.properties." > -- > > Key: SPARK-33668 > URL: https://issues.apache.org/jira/browse/SPARK-33668 > Project: Spark > Issue Type: Bug > Components: Kubernetes, Tests >Affects Versions: 3.1.0 >Reporter: Prashant Sharma >Priority: Major > > The test is flaking, with multiple flaked instances - the reason for the > failure has been similar to: > {code:java} > The code passed to eventually never returned normally. Attempted 109 times > over 3.007988241397 minutes. Last failure message: Failure executing: GET > at: > https://192.168.39.167:8443/api/v1/namespaces/b37fc72a991b49baa68a2eaaa1516463/pods/spark-pi-97a9bc76308e7fe3-exec-1/log?pretty=false. > Message: pods "spark-pi-97a9bc76308e7fe3-exec-1" not found. Received status: > Status(apiVersion=v1, code=404, details=StatusDetails(causes=[], group=null, > kind=pods, name=spark-pi-97a9bc76308e7fe3-exec-1, retryAfterSeconds=null, > uid=null, additionalProperties={}), kind=Status, message=pods > "spark-pi-97a9bc76308e7fe3-exec-1" not found, > metadata=ListMeta(_continue=null, remainingItemCount=null, > resourceVersion=null, selfLink=null, additionalProperties={}), > reason=NotFound, status=Failure, additionalProperties={}).. > (KubernetesSuite.scala:402) > {code} > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36854/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36852/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36850/console > https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder-K8s/36848/console > From the above failures, it seems, that executor finishes too quickly and is > removed by spark before the test can complete. > So, in order to mitigate this situation, one way is to turn on the flag > {code} >"spark.kubernetes.executor.deleteOnTermination" > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org