[jira] [Assigned] (SPARK-26321) Split a SQL in a correct way
[ https://issues.apache.org/jira/browse/SPARK-26321?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yuming Wang reassigned SPARK-26321: --- Assignee: Yuming Wang > Split a SQL in a correct way > > > Key: SPARK-26321 > URL: https://issues.apache.org/jira/browse/SPARK-26321 > Project: Spark > Issue Type: Improvement > Components: SQL >Affects Versions: 3.0.0 >Reporter: Darcy Shen >Assignee: Yuming Wang >Priority: Major > Fix For: 3.0.0 > > > First: > ./build/mvn -Phive-thriftserver -DskipTests package > > Then: > $ bin/spark-sql > > 18/12/10 19:35:02 INFO SparkSQLCLIDriver: Time taken: 4.483 seconds, Fetched > 1 row(s) > spark-sql> select "1;2"; > Error in query: > no viable alternative at input 'select "'(line 1, pos 7) > == SQL == > select "1 > ---^^^ -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-26321) Split a SQL in a correct way
[ https://issues.apache.org/jira/browse/SPARK-26321?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yuming Wang resolved SPARK-26321. - Fix Version/s: 3.0.0 Resolution: Fixed Issue resolved by pull request 25018 [https://github.com/apache/spark/pull/25018] > Split a SQL in a correct way > > > Key: SPARK-26321 > URL: https://issues.apache.org/jira/browse/SPARK-26321 > Project: Spark > Issue Type: Improvement > Components: SQL >Affects Versions: 3.0.0 >Reporter: Darcy Shen >Priority: Major > Fix For: 3.0.0 > > > First: > ./build/mvn -Phive-thriftserver -DskipTests package > > Then: > $ bin/spark-sql > > 18/12/10 19:35:02 INFO SparkSQLCLIDriver: Time taken: 4.483 seconds, Fetched > 1 row(s) > spark-sql> select "1;2"; > Error in query: > no viable alternative at input 'select "'(line 1, pos 7) > == SQL == > select "1 > ---^^^ -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-29359) Better exception handling in SQLQueryTestSuite and ThriftServerQueryTestSuite
[ https://issues.apache.org/jira/browse/SPARK-29359?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yuming Wang resolved SPARK-29359. - Fix Version/s: 3.0.0 Resolution: Fixed Issue resolved by pull request 26028 [https://github.com/apache/spark/pull/26028] > Better exception handling in SQLQueryTestSuite and ThriftServerQueryTestSuite > - > > Key: SPARK-29359 > URL: https://issues.apache.org/jira/browse/SPARK-29359 > Project: Spark > Issue Type: Bug > Components: Tests >Affects Versions: 3.0.0 >Reporter: Peter Toth >Assignee: Peter Toth >Priority: Minor > Fix For: 3.0.0 > > > SQLQueryTestSuite and ThriftServerQueryTestSuite should have the same > exception handling to avoid issues like this: > {noformat} > Expected "[Recursion level limit 100 reached but query has not exhausted, > try increasing spark.sql.cte.recursion.level.limit > org.apache.spark.SparkException]", but got "[org.apache.spark.SparkException > Recursion level limit 100 reached but query has not exhausted, try > increasing spark.sql.cte.recursion.level.limit]" > {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-29359) Better exception handling in SQLQueryTestSuite and ThriftServerQueryTestSuite
[ https://issues.apache.org/jira/browse/SPARK-29359?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yuming Wang reassigned SPARK-29359: --- Assignee: Peter Toth > Better exception handling in SQLQueryTestSuite and ThriftServerQueryTestSuite > - > > Key: SPARK-29359 > URL: https://issues.apache.org/jira/browse/SPARK-29359 > Project: Spark > Issue Type: Bug > Components: Tests >Affects Versions: 3.0.0 >Reporter: Peter Toth >Assignee: Peter Toth >Priority: Minor > > SQLQueryTestSuite and ThriftServerQueryTestSuite should have the same > exception handling to avoid issues like this: > {noformat} > Expected "[Recursion level limit 100 reached but query has not exhausted, > try increasing spark.sql.cte.recursion.level.limit > org.apache.spark.SparkException]", but got "[org.apache.spark.SparkException > Recursion level limit 100 reached but query has not exhausted, try > increasing spark.sql.cte.recursion.level.limit]" > {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-29435) Spark 3 doesnt work with older shuffle service
[ https://issues.apache.org/jira/browse/SPARK-29435?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16950212#comment-16950212 ] Sandeep Katta commented on SPARK-29435: --- cc [~cloud_fan] [~XuanYuan] this patched is tested by [~koert] please help to review patch > Spark 3 doesnt work with older shuffle service > -- > > Key: SPARK-29435 > URL: https://issues.apache.org/jira/browse/SPARK-29435 > Project: Spark > Issue Type: Bug > Components: Shuffle >Affects Versions: 3.0.0 > Environment: Spark 3 from Sept 26, commit > 8beb736a00b004f97de7fcdf9ff09388d80fc548 > Spark 2.4.1 shuffle service in yarn >Reporter: koert kuipers >Priority: Major > > SPARK-27665 introduced a change to the shuffle protocol. It also introduced a > setting spark.shuffle.useOldFetchProtocol which would allow spark 3 to run > with old shuffle service. > However i have not gotten that to work. I have been testing with Spark 3 > master (from Sept 26) and shuffle service from Spark 2.4.1 in yarn. > The errors i see are for example on EMR: > {code} > Error occurred while fetching local blocks > java.nio.file.NoSuchFileException: > /mnt1/yarn/usercache/hadoop/appcache/application_1570697024032_0058/blockmgr-d1d009b1-1c95-4e2a-9a71-0ff20078b9a8/38/shuffle_0_0_0.index > {code} > And on CDH5: > {code} > org.apache.spark.shuffle.FetchFailedException: > /data/9/hadoop/nm/usercache/koert/appcache/application_1568061697664_8250/blockmgr-57f28014-cdf2-431e-8e11-447ba5c2b2f2/0b/shuffle_0_0_0.index > at > org.apache.spark.storage.ShuffleBlockFetcherIterator.throwFetchFailedException(ShuffleBlockFetcherIterator.scala:596) > at > org.apache.spark.storage.ShuffleBlockFetcherIterator.next(ShuffleBlockFetcherIterator.scala:511) > at > org.apache.spark.storage.ShuffleBlockFetcherIterator.next(ShuffleBlockFetcherIterator.scala:67) > at > org.apache.spark.util.CompletionIterator.next(CompletionIterator.scala:29) > at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:484) > at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:490) > at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458) > at > org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:31) > at > org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:37) > at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458) > at scala.collection.Iterator$SliceIterator.hasNext(Iterator.scala:266) > at > org.apache.spark.sql.execution.SparkPlan.$anonfun$getByteArrayRdd$1(SparkPlan.scala:337) > at > org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2(RDD.scala:850) > at > org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2$adapted(RDD.scala:850) > at > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:327) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:291) > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) > at org.apache.spark.scheduler.Task.run(Task.scala:127) > at > org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:455) > at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1377) > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:458) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) > Caused by: java.nio.file.NoSuchFileException: > /data/9/hadoop/nm/usercache/koert/appcache/application_1568061697664_8250/blockmgr-57f28014-cdf2-431e-8e11-447ba5c2b2f2/0b/shuffle_0_0_0.index > at > sun.nio.fs.UnixException.translateToIOException(UnixException.java:86) > at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:102) > at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:107) > at > sun.nio.fs.UnixFileSystemProvider.newByteChannel(UnixFileSystemProvider.java:214) > at java.nio.file.Files.newByteChannel(Files.java:361) > at java.nio.file.Files.newByteChannel(Files.java:407) > at > org.apache.spark.shuffle.IndexShuffleBlockResolver.getBlockData(IndexShuffleBlockResolver.scala:204) > at > org.apache.spark.storage.BlockManager.getBlockData(BlockManager.scala:551) > at > org.apache.spark.storage.ShuffleBlockFetcherIterator.fetchLocalBlocks(ShuffleBlockFetcherIterator.scala:349) > at > org.apache.spark.storage.ShuffleBlockFetcherIterator.initialize(ShuffleBlockFetcherIterator.scala:391) > at >
[jira] [Resolved] (SPARK-29368) Port interval.sql
[ https://issues.apache.org/jira/browse/SPARK-29368?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dongjoon Hyun resolved SPARK-29368. --- Fix Version/s: 3.0.0 Resolution: Fixed Issue resolved by pull request 26055 [https://github.com/apache/spark/pull/26055] > Port interval.sql > - > > Key: SPARK-29368 > URL: https://issues.apache.org/jira/browse/SPARK-29368 > Project: Spark > Issue Type: Sub-task > Components: SQL, Tests >Affects Versions: 3.0.0 >Reporter: Maxim Gekk >Assignee: Maxim Gekk >Priority: Major > Fix For: 3.0.0 > > > Here is interval.sql: > [https://raw.githubusercontent.com/postgres/postgres/REL_12_STABLE/src/test/regress/sql/interval.sql] > Results: > https://github.com/postgres/postgres/blob/REL_12_STABLE/src/test/regress/expected/interval.out -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-29368) Port interval.sql
[ https://issues.apache.org/jira/browse/SPARK-29368?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dongjoon Hyun reassigned SPARK-29368: - Assignee: Maxim Gekk > Port interval.sql > - > > Key: SPARK-29368 > URL: https://issues.apache.org/jira/browse/SPARK-29368 > Project: Spark > Issue Type: Sub-task > Components: SQL, Tests >Affects Versions: 3.0.0 >Reporter: Maxim Gekk >Assignee: Maxim Gekk >Priority: Major > > Here is interval.sql: > [https://raw.githubusercontent.com/postgres/postgres/REL_12_STABLE/src/test/regress/sql/interval.sql] > Results: > https://github.com/postgres/postgres/blob/REL_12_STABLE/src/test/regress/expected/interval.out -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Comment Edited] (SPARK-29435) Spark 3 doesnt work with older shuffle service
[ https://issues.apache.org/jira/browse/SPARK-29435?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16950122#comment-16950122 ] koert kuipers edited comment on SPARK-29435 at 10/12/19 11:50 PM: -- i checked the patch and it works with dynamic execution using spark 3 shuffle service and using spark 2 shuffle service was (Author: koert): i checked and it works with dynamic execution using spark 3 shuffle service and using spark 2 shuffle service > Spark 3 doesnt work with older shuffle service > -- > > Key: SPARK-29435 > URL: https://issues.apache.org/jira/browse/SPARK-29435 > Project: Spark > Issue Type: Bug > Components: Shuffle >Affects Versions: 3.0.0 > Environment: Spark 3 from Sept 26, commit > 8beb736a00b004f97de7fcdf9ff09388d80fc548 > Spark 2.4.1 shuffle service in yarn >Reporter: koert kuipers >Priority: Major > > SPARK-27665 introduced a change to the shuffle protocol. It also introduced a > setting spark.shuffle.useOldFetchProtocol which would allow spark 3 to run > with old shuffle service. > However i have not gotten that to work. I have been testing with Spark 3 > master (from Sept 26) and shuffle service from Spark 2.4.1 in yarn. > The errors i see are for example on EMR: > {code} > Error occurred while fetching local blocks > java.nio.file.NoSuchFileException: > /mnt1/yarn/usercache/hadoop/appcache/application_1570697024032_0058/blockmgr-d1d009b1-1c95-4e2a-9a71-0ff20078b9a8/38/shuffle_0_0_0.index > {code} > And on CDH5: > {code} > org.apache.spark.shuffle.FetchFailedException: > /data/9/hadoop/nm/usercache/koert/appcache/application_1568061697664_8250/blockmgr-57f28014-cdf2-431e-8e11-447ba5c2b2f2/0b/shuffle_0_0_0.index > at > org.apache.spark.storage.ShuffleBlockFetcherIterator.throwFetchFailedException(ShuffleBlockFetcherIterator.scala:596) > at > org.apache.spark.storage.ShuffleBlockFetcherIterator.next(ShuffleBlockFetcherIterator.scala:511) > at > org.apache.spark.storage.ShuffleBlockFetcherIterator.next(ShuffleBlockFetcherIterator.scala:67) > at > org.apache.spark.util.CompletionIterator.next(CompletionIterator.scala:29) > at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:484) > at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:490) > at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458) > at > org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:31) > at > org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:37) > at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458) > at scala.collection.Iterator$SliceIterator.hasNext(Iterator.scala:266) > at > org.apache.spark.sql.execution.SparkPlan.$anonfun$getByteArrayRdd$1(SparkPlan.scala:337) > at > org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2(RDD.scala:850) > at > org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2$adapted(RDD.scala:850) > at > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:327) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:291) > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) > at org.apache.spark.scheduler.Task.run(Task.scala:127) > at > org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:455) > at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1377) > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:458) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) > Caused by: java.nio.file.NoSuchFileException: > /data/9/hadoop/nm/usercache/koert/appcache/application_1568061697664_8250/blockmgr-57f28014-cdf2-431e-8e11-447ba5c2b2f2/0b/shuffle_0_0_0.index > at > sun.nio.fs.UnixException.translateToIOException(UnixException.java:86) > at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:102) > at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:107) > at > sun.nio.fs.UnixFileSystemProvider.newByteChannel(UnixFileSystemProvider.java:214) > at java.nio.file.Files.newByteChannel(Files.java:361) > at java.nio.file.Files.newByteChannel(Files.java:407) > at > org.apache.spark.shuffle.IndexShuffleBlockResolver.getBlockData(IndexShuffleBlockResolver.scala:204) > at > org.apache.spark.storage.BlockManager.getBlockData(BlockManager.scala:551) > at >
[jira] [Commented] (SPARK-29435) Spark 3 doesnt work with older shuffle service
[ https://issues.apache.org/jira/browse/SPARK-29435?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16950122#comment-16950122 ] koert kuipers commented on SPARK-29435: --- i checked and it works with dynamic execution using spark 3 shuffle service and using spark 2 shuffle service > Spark 3 doesnt work with older shuffle service > -- > > Key: SPARK-29435 > URL: https://issues.apache.org/jira/browse/SPARK-29435 > Project: Spark > Issue Type: Bug > Components: Shuffle >Affects Versions: 3.0.0 > Environment: Spark 3 from Sept 26, commit > 8beb736a00b004f97de7fcdf9ff09388d80fc548 > Spark 2.4.1 shuffle service in yarn >Reporter: koert kuipers >Priority: Major > > SPARK-27665 introduced a change to the shuffle protocol. It also introduced a > setting spark.shuffle.useOldFetchProtocol which would allow spark 3 to run > with old shuffle service. > However i have not gotten that to work. I have been testing with Spark 3 > master (from Sept 26) and shuffle service from Spark 2.4.1 in yarn. > The errors i see are for example on EMR: > {code} > Error occurred while fetching local blocks > java.nio.file.NoSuchFileException: > /mnt1/yarn/usercache/hadoop/appcache/application_1570697024032_0058/blockmgr-d1d009b1-1c95-4e2a-9a71-0ff20078b9a8/38/shuffle_0_0_0.index > {code} > And on CDH5: > {code} > org.apache.spark.shuffle.FetchFailedException: > /data/9/hadoop/nm/usercache/koert/appcache/application_1568061697664_8250/blockmgr-57f28014-cdf2-431e-8e11-447ba5c2b2f2/0b/shuffle_0_0_0.index > at > org.apache.spark.storage.ShuffleBlockFetcherIterator.throwFetchFailedException(ShuffleBlockFetcherIterator.scala:596) > at > org.apache.spark.storage.ShuffleBlockFetcherIterator.next(ShuffleBlockFetcherIterator.scala:511) > at > org.apache.spark.storage.ShuffleBlockFetcherIterator.next(ShuffleBlockFetcherIterator.scala:67) > at > org.apache.spark.util.CompletionIterator.next(CompletionIterator.scala:29) > at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:484) > at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:490) > at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458) > at > org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:31) > at > org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:37) > at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458) > at scala.collection.Iterator$SliceIterator.hasNext(Iterator.scala:266) > at > org.apache.spark.sql.execution.SparkPlan.$anonfun$getByteArrayRdd$1(SparkPlan.scala:337) > at > org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2(RDD.scala:850) > at > org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2$adapted(RDD.scala:850) > at > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) > at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:327) > at org.apache.spark.rdd.RDD.iterator(RDD.scala:291) > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) > at org.apache.spark.scheduler.Task.run(Task.scala:127) > at > org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:455) > at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1377) > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:458) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) > Caused by: java.nio.file.NoSuchFileException: > /data/9/hadoop/nm/usercache/koert/appcache/application_1568061697664_8250/blockmgr-57f28014-cdf2-431e-8e11-447ba5c2b2f2/0b/shuffle_0_0_0.index > at > sun.nio.fs.UnixException.translateToIOException(UnixException.java:86) > at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:102) > at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:107) > at > sun.nio.fs.UnixFileSystemProvider.newByteChannel(UnixFileSystemProvider.java:214) > at java.nio.file.Files.newByteChannel(Files.java:361) > at java.nio.file.Files.newByteChannel(Files.java:407) > at > org.apache.spark.shuffle.IndexShuffleBlockResolver.getBlockData(IndexShuffleBlockResolver.scala:204) > at > org.apache.spark.storage.BlockManager.getBlockData(BlockManager.scala:551) > at > org.apache.spark.storage.ShuffleBlockFetcherIterator.fetchLocalBlocks(ShuffleBlockFetcherIterator.scala:349) > at > org.apache.spark.storage.ShuffleBlockFetcherIterator.initialize(ShuffleBlockFetcherIterator.scala:391) > at >
[jira] [Resolved] (SPARK-29446) Upgrade netty-all to 4.1.42 and fix vulnerabilities.
[ https://issues.apache.org/jira/browse/SPARK-29446?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yuming Wang resolved SPARK-29446. - Resolution: Duplicate > Upgrade netty-all to 4.1.42 and fix vulnerabilities. > > > Key: SPARK-29446 > URL: https://issues.apache.org/jira/browse/SPARK-29446 > Project: Spark > Issue Type: Dependency upgrade > Components: Build >Affects Versions: 3.0.0 >Reporter: jiaan.geng >Priority: Major > > The current code uses io.netty:netty-all:jar:4.1.17 and it will cause a > security vulnerabilities. We could get some security info from > [https://www.tenable.com/cve/CVE-2019-16869]. > This reference remind to upgrate the version of netty-all to 4.1.42 or later. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-29449) Add tooltip to Spark WebUI
[ https://issues.apache.org/jira/browse/SPARK-29449?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiao Li updated SPARK-29449: Description: The initial effort was made in https://issues.apache.org/jira/browse/SPARK-2384. This umbrella Jira is to track the progress of adding tooltip to all the WebUI for better usability. was:This umbrella Jira is to track the progress of adding tooltip to all the WebUI for better usability. > Add tooltip to Spark WebUI > -- > > Key: SPARK-29449 > URL: https://issues.apache.org/jira/browse/SPARK-29449 > Project: Spark > Issue Type: Umbrella > Components: Web UI >Affects Versions: 3.0.0 >Reporter: Xiao Li >Priority: Major > > The initial effort was made in > https://issues.apache.org/jira/browse/SPARK-2384. This umbrella Jira is to > track the progress of adding tooltip to all the WebUI for better usability. > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-29019) Improve tooltip information in JDBC/ODBC Server tab
[ https://issues.apache.org/jira/browse/SPARK-29019?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiao Li updated SPARK-29019: Parent: SPARK-29449 Issue Type: Sub-task (was: Improvement) > Improve tooltip information in JDBC/ODBC Server tab > --- > > Key: SPARK-29019 > URL: https://issues.apache.org/jira/browse/SPARK-29019 > Project: Spark > Issue Type: Sub-task > Components: Web UI >Affects Versions: 3.0.0 >Reporter: Pablo Langa Blanco >Assignee: Pablo Langa Blanco >Priority: Trivial > Fix For: 3.0.0 > > > Some of the columns of JDBC/ODBC server tab in Web UI are hard to understand. > We have documented it at SPARK-28373 but I think it is better to have some > tooltips in the SQL statistics table to explain the columns > More information at the pull request -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-29323) Add tooltip for The Executors Tab's column names in the Spark history server Page
[ https://issues.apache.org/jira/browse/SPARK-29323?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiao Li updated SPARK-29323: Parent: SPARK-29449 Issue Type: Sub-task (was: Improvement) > Add tooltip for The Executors Tab's column names in the Spark history server > Page > - > > Key: SPARK-29323 > URL: https://issues.apache.org/jira/browse/SPARK-29323 > Project: Spark > Issue Type: Sub-task > Components: Web UI >Affects Versions: 3.0.0 >Reporter: liucht-inspur >Assignee: liucht-inspur >Priority: Trivial > Fix For: 3.0.0 > > Attachments: image-2019-10-04-09-42-14-174.png > > > the spark Executors of history Tab page, the Summary part shows the line in > the list of title, but format is irregular. > Some column names have tooltip, such as Storage Memory, Task Time(GC Time), > Input, Shuffle Read,Shuffle Write and Blacklisted, but there are still some > list names that have not tooltip. They are RDD Blocks, Disk Used,Cores, > Activity Tasks, Failed Tasks , Complete Tasks and Total Tasks. oddly, > Executors section below,All the column names Contains the column names above > have tooltip . -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Created] (SPARK-29449) Add tooltip to Spark WebUI
Xiao Li created SPARK-29449: --- Summary: Add tooltip to Spark WebUI Key: SPARK-29449 URL: https://issues.apache.org/jira/browse/SPARK-29449 Project: Spark Issue Type: Umbrella Components: Web UI Affects Versions: 3.0.0 Reporter: Xiao Li This umbrella Jira is to track the progress of adding tooltip to all the WebUI for better usability. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-29433) Web UI Stages table tooltip correction
[ https://issues.apache.org/jira/browse/SPARK-29433?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen resolved SPARK-29433. -- Fix Version/s: 3.0.0 Resolution: Fixed Issue resolved by pull request 26084 [https://github.com/apache/spark/pull/26084] > Web UI Stages table tooltip correction > -- > > Key: SPARK-29433 > URL: https://issues.apache.org/jira/browse/SPARK-29433 > Project: Spark > Issue Type: Bug > Components: Web UI >Affects Versions: 3.0.0 >Reporter: Pablo Langa Blanco >Assignee: Pablo Langa Blanco >Priority: Trivial > Fix For: 3.0.0 > > > In the Web UI, Stages table, the tool tip of Input and output column are not > corrrect. > Actual tooltip messages: > * Bytes and records read from Hadoop or from Spark storage. > * Bytes and records written to Hadoop. > In this column we are only showing bytes, not records > More information at the pull request -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-29433) Web UI Stages table tooltip correction
[ https://issues.apache.org/jira/browse/SPARK-29433?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen reassigned SPARK-29433: Assignee: Pablo Langa Blanco > Web UI Stages table tooltip correction > -- > > Key: SPARK-29433 > URL: https://issues.apache.org/jira/browse/SPARK-29433 > Project: Spark > Issue Type: Bug > Components: Web UI >Affects Versions: 3.0.0 >Reporter: Pablo Langa Blanco >Assignee: Pablo Langa Blanco >Priority: Trivial > > In the Web UI, Stages table, the tool tip of Input and output column are not > corrrect. > Actual tooltip messages: > * Bytes and records read from Hadoop or from Spark storage. > * Bytes and records written to Hadoop. > In this column we are only showing bytes, not records > More information at the pull request -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-29445) Bump netty-all from 4.1.39.Final to 4.1.42.Final
[ https://issues.apache.org/jira/browse/SPARK-29445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen reassigned SPARK-29445: Assignee: Fokko Driesprong > Bump netty-all from 4.1.39.Final to 4.1.42.Final > > > Key: SPARK-29445 > URL: https://issues.apache.org/jira/browse/SPARK-29445 > Project: Spark > Issue Type: Task > Components: Spark Core >Affects Versions: 2.4.4 >Reporter: Fokko Driesprong >Assignee: Fokko Driesprong >Priority: Major > > https://www.cvedetails.com/cve/CVE-2019-16869/ -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-29445) Bump netty-all from 4.1.39.Final to 4.1.42.Final
[ https://issues.apache.org/jira/browse/SPARK-29445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen resolved SPARK-29445. -- Fix Version/s: 3.0.0 2.4.5 Resolution: Fixed Issue resolved by pull request 26099 [https://github.com/apache/spark/pull/26099] > Bump netty-all from 4.1.39.Final to 4.1.42.Final > > > Key: SPARK-29445 > URL: https://issues.apache.org/jira/browse/SPARK-29445 > Project: Spark > Issue Type: Task > Components: Spark Core >Affects Versions: 2.4.4 >Reporter: Fokko Driesprong >Assignee: Fokko Driesprong >Priority: Major > Fix For: 2.4.5, 3.0.0 > > > https://www.cvedetails.com/cve/CVE-2019-16869/ -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-29323) Add tooltip for The Executors Tab's column names in the Spark history server Page
[ https://issues.apache.org/jira/browse/SPARK-29323?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen reassigned SPARK-29323: Assignee: liucht-inspur > Add tooltip for The Executors Tab's column names in the Spark history server > Page > - > > Key: SPARK-29323 > URL: https://issues.apache.org/jira/browse/SPARK-29323 > Project: Spark > Issue Type: Improvement > Components: Web UI >Affects Versions: 3.0.0 >Reporter: liucht-inspur >Assignee: liucht-inspur >Priority: Trivial > Attachments: image-2019-10-04-09-42-14-174.png > > > the spark Executors of history Tab page, the Summary part shows the line in > the list of title, but format is irregular. > Some column names have tooltip, such as Storage Memory, Task Time(GC Time), > Input, Shuffle Read,Shuffle Write and Blacklisted, but there are still some > list names that have not tooltip. They are RDD Blocks, Disk Used,Cores, > Activity Tasks, Failed Tasks , Complete Tasks and Total Tasks. oddly, > Executors section below,All the column names Contains the column names above > have tooltip . -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-29323) Add tooltip for The Executors Tab's column names in the Spark history server Page
[ https://issues.apache.org/jira/browse/SPARK-29323?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen resolved SPARK-29323. -- Fix Version/s: 3.0.0 Resolution: Fixed Issue resolved by pull request 25994 [https://github.com/apache/spark/pull/25994] > Add tooltip for The Executors Tab's column names in the Spark history server > Page > - > > Key: SPARK-29323 > URL: https://issues.apache.org/jira/browse/SPARK-29323 > Project: Spark > Issue Type: Improvement > Components: Web UI >Affects Versions: 3.0.0 >Reporter: liucht-inspur >Assignee: liucht-inspur >Priority: Trivial > Fix For: 3.0.0 > > Attachments: image-2019-10-04-09-42-14-174.png > > > the spark Executors of history Tab page, the Summary part shows the line in > the list of title, but format is irregular. > Some column names have tooltip, such as Storage Memory, Task Time(GC Time), > Input, Shuffle Read,Shuffle Write and Blacklisted, but there are still some > list names that have not tooltip. They are RDD Blocks, Disk Used,Cores, > Activity Tasks, Failed Tasks , Complete Tasks and Total Tasks. oddly, > Executors section below,All the column names Contains the column names above > have tooltip . -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-29404) Add some explanations about the color of execution bars on Web UI.
[ https://issues.apache.org/jira/browse/SPARK-29404?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen resolved SPARK-29404. -- Resolution: Won't Fix > Add some explanations about the color of execution bars on Web UI. > -- > > Key: SPARK-29404 > URL: https://issues.apache.org/jira/browse/SPARK-29404 > Project: Spark > Issue Type: Improvement > Components: Documentation >Affects Versions: 3.0.0 >Reporter: Tomoko Komiyama >Priority: Trivial > > Documentation doesn't explain enough about why the executor bar's color > changes with click on it on Web UI. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-29328) Incorrect calculation mean seconds per month
[ https://issues.apache.org/jira/browse/SPARK-29328?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen updated SPARK-29328: - Labels: (was: correctness) > Incorrect calculation mean seconds per month > > > Key: SPARK-29328 > URL: https://issues.apache.org/jira/browse/SPARK-29328 > Project: Spark > Issue Type: Improvement > Components: SQL >Affects Versions: 2.1.3, 2.2.3, 2.3.4, 2.4.4 >Reporter: Maxim Gekk >Assignee: Maxim Gekk >Priority: Minor > > Existing implementation assumes 31 days per month or 372 days per year which > is far away from the correct number. Spark uses the proleptic Gregorian > calendar by default SPARK-26651 in which the average year is 365.2425 days > long: https://en.wikipedia.org/wiki/Gregorian_calendar . Need to fix > calculation in 3 places at least: > - GroupStateImpl.scala:167:val millisPerMonth = > TimeUnit.MICROSECONDS.toMillis(CalendarInterval.MICROS_PER_DAY) * 31 > - EventTimeWatermark.scala:32:val millisPerMonth = > TimeUnit.MICROSECONDS.toMillis(CalendarInterval.MICROS_PER_DAY) * 31 -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-29328) Incorrect calculation mean seconds per month
[ https://issues.apache.org/jira/browse/SPARK-29328?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen resolved SPARK-29328. -- Fix Version/s: (was: 3.0.0) Resolution: Won't Fix > Incorrect calculation mean seconds per month > > > Key: SPARK-29328 > URL: https://issues.apache.org/jira/browse/SPARK-29328 > Project: Spark > Issue Type: Improvement > Components: SQL >Affects Versions: 2.1.3, 2.2.3, 2.3.4, 2.4.4 >Reporter: Maxim Gekk >Assignee: Maxim Gekk >Priority: Minor > Labels: correctness > > Existing implementation assumes 31 days per month or 372 days per year which > is far away from the correct number. Spark uses the proleptic Gregorian > calendar by default SPARK-26651 in which the average year is 365.2425 days > long: https://en.wikipedia.org/wiki/Gregorian_calendar . Need to fix > calculation in 3 places at least: > - GroupStateImpl.scala:167:val millisPerMonth = > TimeUnit.MICROSECONDS.toMillis(CalendarInterval.MICROS_PER_DAY) * 31 > - EventTimeWatermark.scala:32:val millisPerMonth = > TimeUnit.MICROSECONDS.toMillis(CalendarInterval.MICROS_PER_DAY) * 31 -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Reopened] (SPARK-29328) Incorrect calculation mean seconds per month
[ https://issues.apache.org/jira/browse/SPARK-29328?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen reopened SPARK-29328: -- Reverted in https://github.com/apache/spark/pull/26101 > Incorrect calculation mean seconds per month > > > Key: SPARK-29328 > URL: https://issues.apache.org/jira/browse/SPARK-29328 > Project: Spark > Issue Type: Improvement > Components: SQL >Affects Versions: 2.1.3, 2.2.3, 2.3.4, 2.4.4 >Reporter: Maxim Gekk >Assignee: Maxim Gekk >Priority: Minor > Labels: correctness > Fix For: 3.0.0 > > > Existing implementation assumes 31 days per month or 372 days per year which > is far away from the correct number. Spark uses the proleptic Gregorian > calendar by default SPARK-26651 in which the average year is 365.2425 days > long: https://en.wikipedia.org/wiki/Gregorian_calendar . Need to fix > calculation in 3 places at least: > - GroupStateImpl.scala:167:val millisPerMonth = > TimeUnit.MICROSECONDS.toMillis(CalendarInterval.MICROS_PER_DAY) * 31 > - EventTimeWatermark.scala:32:val millisPerMonth = > TimeUnit.MICROSECONDS.toMillis(CalendarInterval.MICROS_PER_DAY) * 31 -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-29410) Update Commons BeanUtils to 1.9.4
[ https://issues.apache.org/jira/browse/SPARK-29410?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen updated SPARK-29410: - Fix Version/s: 2.4.5 > Update Commons BeanUtils to 1.9.4 > - > > Key: SPARK-29410 > URL: https://issues.apache.org/jira/browse/SPARK-29410 > Project: Spark > Issue Type: Improvement > Components: Spark Core >Affects Versions: 3.0.0 >Reporter: Peter Toth >Assignee: Peter Toth >Priority: Minor > Fix For: 2.4.5, 3.0.0 > > > Update commons-beanutils to 1.9.4 to fix CVE: > [http://commons.apache.org/proper/commons-beanutils/javadocs/v1.9.4/RELEASE-NOTES.txt] -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-29380) RFormula avoid repeated 'first' jobs to get vector size
[ https://issues.apache.org/jira/browse/SPARK-29380?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] zhengruifeng reassigned SPARK-29380: Assignee: zhengruifeng > RFormula avoid repeated 'first' jobs to get vector size > --- > > Key: SPARK-29380 > URL: https://issues.apache.org/jira/browse/SPARK-29380 > Project: Spark > Issue Type: Improvement > Components: ML >Affects Versions: 3.0.0 >Reporter: zhengruifeng >Assignee: zhengruifeng >Priority: Minor > > In current impl, {{RFormula}} will trigger one {{first}} job to get the > vector size, if the size can not be obtained from {{AttributeGroup.}} > {{This can be optimized by get the first row lazily, and reuse it for each > vector column.}} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-29380) RFormula avoid repeated 'first' jobs to get vector size
[ https://issues.apache.org/jira/browse/SPARK-29380?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] zhengruifeng resolved SPARK-29380. -- Fix Version/s: 3.0.0 Resolution: Fixed Issue resolved by pull request 26052 [https://github.com/apache/spark/pull/26052] > RFormula avoid repeated 'first' jobs to get vector size > --- > > Key: SPARK-29380 > URL: https://issues.apache.org/jira/browse/SPARK-29380 > Project: Spark > Issue Type: Improvement > Components: ML >Affects Versions: 3.0.0 >Reporter: zhengruifeng >Assignee: zhengruifeng >Priority: Minor > Fix For: 3.0.0 > > > In current impl, {{RFormula}} will trigger one {{first}} job to get the > vector size, if the size can not be obtained from {{AttributeGroup.}} > {{This can be optimized by get the first row lazily, and reuse it for each > vector column.}} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-29410) Update Commons BeanUtils to 1.9.4
[ https://issues.apache.org/jira/browse/SPARK-29410?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen reassigned SPARK-29410: Assignee: Peter Toth > Update Commons BeanUtils to 1.9.4 > - > > Key: SPARK-29410 > URL: https://issues.apache.org/jira/browse/SPARK-29410 > Project: Spark > Issue Type: Improvement > Components: Spark Core >Affects Versions: 3.0.0 >Reporter: Peter Toth >Assignee: Peter Toth >Priority: Minor > > Update commons-beanutils to 1.9.4 to fix CVE: > [http://commons.apache.org/proper/commons-beanutils/javadocs/v1.9.4/RELEASE-NOTES.txt] -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-29410) Update Commons BeanUtils to 1.9.4
[ https://issues.apache.org/jira/browse/SPARK-29410?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen resolved SPARK-29410. -- Fix Version/s: 3.0.0 Resolution: Fixed Issue resolved by pull request 26069 [https://github.com/apache/spark/pull/26069] > Update Commons BeanUtils to 1.9.4 > - > > Key: SPARK-29410 > URL: https://issues.apache.org/jira/browse/SPARK-29410 > Project: Spark > Issue Type: Improvement > Components: Spark Core >Affects Versions: 3.0.0 >Reporter: Peter Toth >Assignee: Peter Toth >Priority: Minor > Fix For: 3.0.0 > > > Update commons-beanutils to 1.9.4 to fix CVE: > [http://commons.apache.org/proper/commons-beanutils/javadocs/v1.9.4/RELEASE-NOTES.txt] -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-28810) Document SHOW TABLES in SQL Reference.
[ https://issues.apache.org/jira/browse/SPARK-28810?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen resolved SPARK-28810. -- Fix Version/s: 3.0.0 Resolution: Fixed Issue resolved by pull request 25561 [https://github.com/apache/spark/pull/25561] > Document SHOW TABLES in SQL Reference. > -- > > Key: SPARK-28810 > URL: https://issues.apache.org/jira/browse/SPARK-28810 > Project: Spark > Issue Type: Sub-task > Components: Documentation, SQL >Affects Versions: 2.4.3 >Reporter: Dilip Biswal >Assignee: Shivu Sondur >Priority: Major > Fix For: 3.0.0 > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-28810) Document SHOW TABLES in SQL Reference.
[ https://issues.apache.org/jira/browse/SPARK-28810?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sean R. Owen reassigned SPARK-28810: Assignee: Shivu Sondur > Document SHOW TABLES in SQL Reference. > -- > > Key: SPARK-28810 > URL: https://issues.apache.org/jira/browse/SPARK-28810 > Project: Spark > Issue Type: Sub-task > Components: Documentation, SQL >Affects Versions: 2.4.3 >Reporter: Dilip Biswal >Assignee: Shivu Sondur >Priority: Major > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-29116) Refactor py classes related to DecisionTree
[ https://issues.apache.org/jira/browse/SPARK-29116?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] zhengruifeng resolved SPARK-29116. -- Fix Version/s: 3.0.0 Resolution: Fixed Issue resolved by pull request 25929 [https://github.com/apache/spark/pull/25929] > Refactor py classes related to DecisionTree > --- > > Key: SPARK-29116 > URL: https://issues.apache.org/jira/browse/SPARK-29116 > Project: Spark > Issue Type: Sub-task > Components: ML, PySpark >Affects Versions: 3.0.0 >Reporter: zhengruifeng >Assignee: Huaxin Gao >Priority: Minor > Fix For: 3.0.0 > > > 1, Like the scala side, move related classes to a seperate file 'tree.py' > 2, add method 'predictLeaf' in 'DecisionTreeModel' & 'TreeEnsembleModel' -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Assigned] (SPARK-29116) Refactor py classes related to DecisionTree
[ https://issues.apache.org/jira/browse/SPARK-29116?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] zhengruifeng reassigned SPARK-29116: Assignee: Huaxin Gao > Refactor py classes related to DecisionTree > --- > > Key: SPARK-29116 > URL: https://issues.apache.org/jira/browse/SPARK-29116 > Project: Spark > Issue Type: Sub-task > Components: ML, PySpark >Affects Versions: 3.0.0 >Reporter: zhengruifeng >Assignee: Huaxin Gao >Priority: Minor > > 1, Like the scala side, move related classes to a seperate file 'tree.py' > 2, add method 'predictLeaf' in 'DecisionTreeModel' & 'TreeEnsembleModel' -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Created] (SPARK-29448) Support the `INTERVAL` type by Parquet datasource
Maxim Gekk created SPARK-29448: -- Summary: Support the `INTERVAL` type by Parquet datasource Key: SPARK-29448 URL: https://issues.apache.org/jira/browse/SPARK-29448 Project: Spark Issue Type: New Feature Components: SQL Affects Versions: 2.4.4 Reporter: Maxim Gekk Parquet format allows to store intervals as triple of (milliseconds, days, months) see https://github.com/apache/parquet-format/blob/master/LogicalTypes.md#interval . The `INTERVAL` logical type is used for an interval of time. _It must annotate a fixed_len_byte_array of length 12. This array stores three little-endian unsigned integers that represent durations at different granularities of time. The first stores a number in months, the second stores a number in days, and the third stores a number in milliseconds. This representation is independent of any particular timezone or date._ Need to support writing and reading values of Catalyst's CalendarIntervalType. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Closed] (SPARK-29447) Allow users to update the name of a column
[ https://issues.apache.org/jira/browse/SPARK-29447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lantao Jin closed SPARK-29447. -- > Allow users to update the name of a column > -- > > Key: SPARK-29447 > URL: https://issues.apache.org/jira/browse/SPARK-29447 > Project: Spark > Issue Type: New Feature > Components: SQL >Affects Versions: 3.0.0 >Reporter: Lantao Jin >Priority: Major > > Right now, {{ALTER TABLE CHANGE COLUMN}} command only supports changing the > comment of a non-partition column for now > (https://issues.apache.org/jira/browse/SPARK-17910). > Support changing column's name is very useful. I will file a PR to address it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Resolved] (SPARK-29447) Allow users to update the name of a column
[ https://issues.apache.org/jira/browse/SPARK-29447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lantao Jin resolved SPARK-29447. Resolution: Invalid Close since we have had {{ALTER TABLE table1 RENAME COLUMN a.b.c TO x}} > Allow users to update the name of a column > -- > > Key: SPARK-29447 > URL: https://issues.apache.org/jira/browse/SPARK-29447 > Project: Spark > Issue Type: New Feature > Components: SQL >Affects Versions: 3.0.0 >Reporter: Lantao Jin >Priority: Major > > Right now, {{ALTER TABLE CHANGE COLUMN}} command only supports changing the > comment of a non-partition column for now > (https://issues.apache.org/jira/browse/SPARK-17910). > Support changing column's name is very useful. I will file a PR to address it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-29428) Can't persist/set None-valued param
[ https://issues.apache.org/jira/browse/SPARK-29428?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16949968#comment-16949968 ] Borys Biletskyy commented on SPARK-29428: - Thanks for your inputs. Maybe makes sense to mention it in the python docs, where I really missed conventions regarding None params. For me it was a way to workaround https://issues.apache.org/jira/browse/SPARK-29414. >From the following Params method I've got an impression that None params are >acceptable. {code:java} def _set(self, **kwargs): """ Sets user-supplied params. """ for param, value in kwargs.items(): p = getattr(self, param) if value is not None: try: value = p.typeConverter(value) except TypeError as e: raise TypeError('Invalid param value given for param "%s". %s' % (p.name, e)) self._paramMap[p] = value return self {code} Which is not the case here, where None params are not acceptable. {code:java} def set(self, param, value): """ Sets a parameter in the embedded param map. """ self._shouldOwn(param) try: value = param.typeConverter(value) except ValueError as e: raise ValueError('Invalid param value given for param "%s". %s' % (param.name, e)) self._paramMap[param] = value {code} > Can't persist/set None-valued param > > > Key: SPARK-29428 > URL: https://issues.apache.org/jira/browse/SPARK-29428 > Project: Spark > Issue Type: Bug > Components: ML, PySpark >Affects Versions: 2.3.2 >Reporter: Borys Biletskyy >Priority: Major > > {code:java} > import pytest > from pyspark import keyword_only > from pyspark.ml import Model > from pyspark.sql import DataFrame > from pyspark.ml.util import DefaultParamsReadable, DefaultParamsWritable > from pyspark.ml.param.shared import HasInputCol > from pyspark.sql.functions import * > class NoneParamTester(Model, > HasInputCol, > DefaultParamsReadable, > DefaultParamsWritable > ): > @keyword_only > def __init__(self, inputCol: str = None): > super(NoneParamTester, self).__init__() > kwargs = self._input_kwargs > self.setParams(**kwargs) > @keyword_only > def setParams(self, inputCol: str = None): > kwargs = self._input_kwargs > self._set(**kwargs) > return self > def _transform(self, data: DataFrame) -> DataFrame: > return data > class TestNoneParam(object): > def test_persist_none(self, spark, temp_dir): > path = temp_dir + '/test_model' > model = NoneParamTester(inputCol=None) > assert model.isDefined(model.inputCol) > assert model.isSet(model.inputCol) > assert model.getInputCol() is None > model.write().overwrite().save(path) > NoneParamTester.load(path) # TypeError: Could not convert 'NoneType'> to string type > def test_set_none(self, spark): > model = NoneParamTester(inputCol=None) > assert model.isDefined(model.inputCol) > assert model.isSet(model.inputCol) > assert model.getInputCol() is None > model.set(model.inputCol, None) # TypeError: Could not convert > to string type > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-29447) Allow users to update the name of a column
[ https://issues.apache.org/jira/browse/SPARK-29447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lantao Jin updated SPARK-29447: --- Description: Right now, {{ALTER TABLE CHANGE COLUMN}} command only supports changing the comment of a non-partition column for now (https://issues.apache.org/jira/browse/SPARK-17910). Support changing column's name is very useful. I will file a PR to address it. was: Right now, {{ALTER TABLE CHANGE COLUMN}} command only supports changing the comment of a non-partition column for now (https://issues.apache.org/jira/browse/SPARK-17910). Support changing column's name is very useful. > Allow users to update the name of a column > -- > > Key: SPARK-29447 > URL: https://issues.apache.org/jira/browse/SPARK-29447 > Project: Spark > Issue Type: New Feature > Components: SQL >Affects Versions: 3.0.0 >Reporter: Lantao Jin >Priority: Major > > Right now, {{ALTER TABLE CHANGE COLUMN}} command only supports changing the > comment of a non-partition column for now > (https://issues.apache.org/jira/browse/SPARK-17910). > Support changing column's name is very useful. I will file a PR to address it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Created] (SPARK-29447) Allow users to update the name of a column
Lantao Jin created SPARK-29447: -- Summary: Allow users to update the name of a column Key: SPARK-29447 URL: https://issues.apache.org/jira/browse/SPARK-29447 Project: Spark Issue Type: New Feature Components: SQL Affects Versions: 3.0.0 Reporter: Lantao Jin Right now, {{ALTER TABLE CHANGE COLUMN}} command only supports changing the comment of a non-partition column for now (https://issues.apache.org/jira/browse/SPARK-17910). Support changing column's name is very useful. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-29441) Unable to Alter table column type in spark.
[ https://issues.apache.org/jira/browse/SPARK-29441?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16949961#comment-16949961 ] Lantao Jin commented on SPARK-29441: ALTER TABLE CHANGE COLUMN command only supports changing the comment of a non-partition column for now. > Unable to Alter table column type in spark. > --- > > Key: SPARK-29441 > URL: https://issues.apache.org/jira/browse/SPARK-29441 > Project: Spark > Issue Type: Bug > Components: Spark Shell >Affects Versions: 2.3.1 > Environment: spark -2.3 > hadoop -2.4 >Reporter: prudhviraj >Priority: Major > > Unable to alter table column type in spark. > scala> spark.sql("""alter table tablename change col1 col1 string""") > org.apache.spark.sql.AnalysisException: ALTER TABLE CHANGE COLUMN is not > supported for changing column 'col1' with type 'LongType' to 'col1' with type > 'StringType'; -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-29446) Upgrade netty-all to 4.1.42 and fix vulnerabilities.
[ https://issues.apache.org/jira/browse/SPARK-29446?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16949958#comment-16949958 ] Iskender Unlu commented on SPARK-29446: --- I will have a look around this issue. > Upgrade netty-all to 4.1.42 and fix vulnerabilities. > > > Key: SPARK-29446 > URL: https://issues.apache.org/jira/browse/SPARK-29446 > Project: Spark > Issue Type: Dependency upgrade > Components: Build >Affects Versions: 3.0.0 >Reporter: jiaan.geng >Priority: Major > > The current code uses io.netty:netty-all:jar:4.1.17 and it will cause a > security vulnerabilities. We could get some security info from > [https://www.tenable.com/cve/CVE-2019-16869]. > This reference remind to upgrate the version of netty-all to 4.1.42 or later. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Created] (SPARK-29446) Upgrade netty-all to 4.1.42 and fix vulnerabilities.
jiaan.geng created SPARK-29446: -- Summary: Upgrade netty-all to 4.1.42 and fix vulnerabilities. Key: SPARK-29446 URL: https://issues.apache.org/jira/browse/SPARK-29446 Project: Spark Issue Type: Dependency upgrade Components: Build Affects Versions: 3.0.0 Reporter: jiaan.geng The current code uses io.netty:netty-all:jar:4.1.17 and it will cause a security vulnerabilities. We could get some security info from [https://www.tenable.com/cve/CVE-2019-16869]. This reference remind to upgrate the version of netty-all to 4.1.42 or later. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org