HyukjinKwon commented on code in PR #36773:
URL: https://github.com/apache/spark/pull/36773#discussion_r916607535


##########
sql/core/src/test/scala/org/apache/spark/sql/jdbc/JDBCV2Suite.scala:
##########
@@ -1685,6 +1709,42 @@ class JDBCV2Suite extends QueryTest with 
SharedSparkSession with ExplainSuiteHel
     checkAnswer(df2, Seq(Row(1d), Row(1d), Row(null)))
   }
 
+  test("scan with aggregate push-down: linear regression functions with filter 
and group by") {

Review Comment:
   This also seems to be failing with ANSI mode on:
   
   ```
   2022-07-08T01:56:48.3914077Z [info] - scan with 
aggregate push-down: linear regression functions with filter and group by *** 
FAILED *** (350 milliseconds)
   2022-07-08T01:56:48.3915454Z [info]   
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in 
stage 416.0 failed 1 times, most recent failure: Lost task 0.0 in stage 416.0 
(TID 379) (localhost executor driver): 
org.apache.spark.SparkArithmeticException: [DIVIDE_BY_ZERO] Division by zero. 
Use `try_divide` to tolerate divisor being 0 and return NULL instead. If 
necessary set "spark.sql.ansi.enabled" to "false" (except for ANSI interval 
type) to bypass this error.
   2022-07-08T01:56:48.3916839Z [info]     at 
org.apache.spark.sql.errors.QueryExecutionErrors$.divideByZeroError(QueryExecutionErrors.scala:184)
   2022-07-08T01:56:48.3917602Z [info]     at 
org.apache.spark.sql.errors.QueryExecutionErrors.divideByZeroError(QueryExecutionErrors.scala)
   2022-07-08T01:56:48.3918487Z [info]     at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage3.hashAgg_doAggregateWithKeysOutput_0$(Unknown
 Source)
   2022-07-08T01:56:48.3919355Z [info]     at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage3.processNext(Unknown
 Source)
   2022-07-08T01:56:48.3920199Z [info]     at 
org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
   2022-07-08T01:56:48.3921079Z [info]     at 
org.apache.spark.sql.execution.WholeStageCodegenExec$$anon$1.hasNext(WholeStageCodegenExec.scala:760)
   2022-07-08T01:56:48.3921693Z [info]     at 
scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
   2022-07-08T01:56:48.3922208Z [info]     at 
scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
   2022-07-08T01:56:48.3922715Z [info]     at 
org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1917)
   2022-07-08T01:56:48.3923231Z [info]     at 
org.apache.spark.rdd.RDD.$anonfun$count$1(RDD.scala:1268)
   2022-07-08T01:56:48.3923746Z [info]     at 
org.apache.spark.rdd.RDD.$anonfun$count$1$adapted(RDD.scala:1268)
   2022-07-08T01:56:48.3924290Z [info]     at 
org.apache.spark.SparkContext.$anonfun$runJob$5(SparkContext.scala:2262)
   2022-07-08T01:56:48.3924862Z [info]     at 
org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:92)
   2022-07-08T01:56:48.3925414Z [info]     at 
org.apache.spark.scheduler.Task.run(Task.scala:139)
   2022-07-08T01:56:48.3925982Z [info]     at 
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:548)
   2022-07-08T01:56:48.3926545Z [info]     at 
org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1490)
   2022-07-08T01:56:48.3927090Z [info]     at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:551)
   2022-07-08T01:56:48.3927701Z [info]     at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
   2022-07-08T01:56:48.3928327Z [info]     at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
   2022-07-08T01:56:48.3928833Z [info]     at 
java.lang.Thread.run(Thread.java:750)
   2022-07-08T01:56:48.3929196Z [info] 
   2022-07-08T01:56:48.3929548Z [info] Driver 
stacktrace:
   2022-07-08T01:56:48.3930152Z [info]   at 
org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2706)
   2022-07-08T01:56:48.3930839Z [info]   at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2642)
   2022-07-08T01:56:48.3931492Z [info]   at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2641)
   2022-07-08T01:56:48.3932111Z [info]   at 
scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
   2022-07-08T01:56:48.3932717Z [info]   at 
scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
   2022-07-08T01:56:48.3933306Z [info]   at 
scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
   2022-07-08T01:56:48.3933969Z [info]   at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2641)
   2022-07-08T01:56:48.3937642Z [info]   at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1189)
   2022-07-08T01:56:48.3941965Z [info]   at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1189)
   2022-07-08T01:56:48.3943119Z [info]   at 
scala.Option.foreach(Option.scala:407)
   2022-07-08T01:56:48.3943813Z [info]   at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1189)
   2022-07-08T01:56:48.3944644Z [info]   at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2897)
   2022-07-08T01:56:48.3945466Z [info]   at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2836)
   2022-07-08T01:56:48.3946265Z [info]   at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2825)
   2022-07-08T01:56:48.3947177Z [info]   at 
org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
   2022-07-08T01:56:48.3947776Z [info]   at 
org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:952)
   2022-07-08T01:56:48.4330080Z [info]   at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:2222)
   2022-07-08T01:56:48.4330873Z [info]   at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:2243)
   2022-07-08T01:56:48.4331499Z [info]   at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:2262)
   2022-07-08T01:56:48.4332552Z [info]   at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:2287)
   2022-07-08T01:56:48.4333184Z [info]   at 
org.apache.spark.rdd.RDD.count(RDD.scala:1268)
   2022-07-08T01:56:48.4333923Z [info]   at 
org.apache.spark.sql.QueryTest$.$anonfun$getErrorMessageInCheckAnswer$1(QueryTest.scala:265)
   2022-07-08T01:56:48.4334537Z [info]   at 
scala.runtime.java8.JFunction0$mcJ$sp.apply(JFunction0$mcJ$sp.java:23)
   2022-07-08T01:56:48.4335504Z [info]   at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:171)
   2022-07-08T01:56:48.4336158Z [info]   at 
org.apache.spark.sql.QueryTest$.getErrorMessageInCheckAnswer(QueryTest.scala:265)
   2022-07-08T01:56:48.4336802Z [info]   at 
org.apache.spark.sql.QueryTest$.checkAnswer(QueryTest.scala:242)
   2022-07-08T01:56:48.4337377Z [info]   at 
org.apache.spark.sql.QueryTest.checkAnswer(QueryTest.scala:150)
   2022-07-08T01:56:48.4337959Z [info]   at 
org.apache.spark.sql.jdbc.JDBCV2Suite.$anonfun$new$204(JDBCV2Suite.scala:1745)
   2022-07-08T01:56:48.4338601Z [info]   at 
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
   2022-07-08T01:56:48.4339126Z [info]   at 
org.scalatest.OutcomeOf.outcomeOf(OutcomeOf.scala:85)
   2022-07-08T01:56:48.4339645Z [info]   at 
org.scalatest.OutcomeOf.outcomeOf$(OutcomeOf.scala:83)
   2022-07-08T01:56:48.4340150Z [info]   at 
org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
   2022-07-08T01:56:48.4340653Z [info]   at 
org.scalatest.Transformer.apply(Transformer.scala:22)
   2022-07-08T01:56:48.4341312Z [info]   at 
org.scalatest.Transformer.apply(Transformer.scala:20)
   2022-07-08T01:56:48.4369891Z [info]   at 
org.scalatest.funsuite.AnyFunSuiteLike$$anon$1.apply(AnyFunSuiteLike.scala:190)
   2022-07-08T01:56:48.4370698Z [info]   at 
org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:204)
   2022-07-08T01:56:48.4371447Z [info]   at 
org.scalatest.funsuite.AnyFunSuiteLike.invokeWithFixture$1(AnyFunSuiteLike.scala:188)
   2022-07-08T01:56:48.4372349Z [info]   at 
org.scalatest.funsuite.AnyFunSuiteLike.$anonfun$runTest$1(AnyFunSuiteLike.scala:200)
   2022-07-08T01:56:48.4373008Z [info]   at 
org.scalatest.SuperEngine.runTestImpl(Engine.scala:306)
   2022-07-08T01:56:48.4373668Z [info]   at 
org.scalatest.funsuite.AnyFunSuiteLike.runTest(AnyFunSuiteLike.scala:200)
   2022-07-08T01:56:48.4374361Z [info]   at 
org.scalatest.funsuite.AnyFunSuiteLike.runTest$(AnyFunSuiteLike.scala:182)
   2022-07-08T01:56:48.4375096Z [info]   at 
org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterEach$$super$runTest(SparkFunSuite.scala:65)
   2022-07-08T01:56:48.4375800Z [info]   at 
org.scalatest.BeforeAndAfterEach.runTest(BeforeAndAfterEach.scala:234)
   2022-07-08T01:56:48.4376475Z [info]   at 
org.scalatest.BeforeAndAfterEach.runTest$(BeforeAndAfterEach.scala:227)
   2022-07-08T01:56:48.4377124Z [info]   at 
org.apache.spark.SparkFunSuite.runTest(SparkFunSuite.scala:65)
   2022-07-08T01:56:48.4377802Z [info]   at 
org.scalatest.funsuite.AnyFunSuiteLike.$anonfun$runTests$1(AnyFunSuiteLike.scala:233)
   2022-07-08T01:56:48.4378466Z [info]   at 
org.scalatest.SuperEngine.$anonfun$runTestsInBranch$1(Engine.scala:413)
   2022-07-08T01:56:48.4379077Z [info]   at 
scala.collection.immutable.List.foreach(List.scala:431)
   2022-07-08T01:56:48.4379698Z [info]   at 
org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401)
   2022-07-08T01:56:48.4380329Z [info]   at 
org.scalatest.SuperEngine.runTestsInBranch(Engine.scala:396)
   2022-07-08T01:56:48.4380955Z [info]   at 
org.scalatest.SuperEngine.runTestsImpl(Engine.scala:475)
   2022-07-08T01:56:48.4381607Z [info]   at 
org.scalatest.funsuite.AnyFunSuiteLike.runTests(AnyFunSuiteLike.scala:233)
   2022-07-08T01:56:48.4382292Z [info]   at 
org.scalatest.funsuite.AnyFunSuiteLike.runTests$(AnyFunSuiteLike.scala:232)
   2022-07-08T01:56:48.4382973Z [info]   at 
org.scalatest.funsuite.AnyFunSuite.runTests(AnyFunSuite.scala:1563)
   2022-07-08T01:56:48.4383559Z [info]   at 
org.scalatest.Suite.run(Suite.scala:1112)
   2022-07-08T01:56:48.4384087Z [info]   at 
org.scalatest.Suite.run$(Suite.scala:1094)
   2022-07-08T01:56:48.4384760Z [info]   at 
org.scalatest.funsuite.AnyFunSuite.org$scalatest$funsuite$AnyFunSuiteLike$$super$run(AnyFunSuite.scala:1563)
   2022-07-08T01:56:48.4385468Z [info]   at 
org.scalatest.funsuite.AnyFunSuiteLike.$anonfun$run$1(AnyFunSuiteLike.scala:237)
   2022-07-08T01:56:48.4386094Z [info]   at 
org.scalatest.SuperEngine.runImpl(Engine.scala:535)
   2022-07-08T01:56:48.4386725Z [info]   at 
org.scalatest.funsuite.AnyFunSuiteLike.run(AnyFunSuiteLike.scala:237)
   2022-07-08T01:56:48.4387381Z [info]   at 
org.scalatest.funsuite.AnyFunSuiteLike.run$(AnyFunSuiteLike.scala:236)
   2022-07-08T01:56:48.4388169Z [info]   at 
org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterAll$$super$run(SparkFunSuite.scala:65)
   2022-07-08T01:56:48.4388882Z [info]   at 
org.scalatest.BeforeAndAfterAll.liftedTree1$1(BeforeAndAfterAll.scala:213)
   2022-07-08T01:56:48.4389533Z [info]   at 
org.scalatest.BeforeAndAfterAll.run(BeforeAndAfterAll.scala:210)
   2022-07-08T01:56:48.4390219Z [info]   at 
org.scalatest.BeforeAndAfterAll.run$(BeforeAndAfterAll.scala:208)
   2022-07-08T01:56:48.4390838Z [info]   at 
org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:65)
   2022-07-08T01:56:48.4391511Z [info]   at 
org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:318)
   2022-07-08T01:56:48.4392296Z [info]   at 
org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:513)
   2022-07-08T01:56:48.4433669Z [info]   at 
sbt.ForkMain$Run.lambda$runTest$1(ForkMain.java:413)
   2022-07-08T01:56:48.4434367Z [info]   at 
java.util.concurrent.FutureTask.run(FutureTask.java:266)
   2022-07-08T01:56:48.4435084Z [info]   at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
   2022-07-08T01:56:48.4435808Z [info]   at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
   2022-07-08T01:56:48.4436787Z [info]   at 
java.lang.Thread.run(Thread.java:750)
   2022-07-08T01:56:48.4437780Z [info]   Cause: 
org.apache.spark.SparkArithmeticException: [DIVIDE_BY_ZERO] Division by zero. 
Use `try_divide` to tolerate divisor being 0 and return NULL instead. If 
necessary set "spark.sql.ansi.enabled" to "false" (except for ANSI interval 
type) to bypass this error.
   2022-07-08T01:56:48.4438954Z [info]   at 
org.apache.spark.sql.errors.QueryExecutionErrors$.divideByZeroError(QueryExecutionErrors.scala:184)
   2022-07-08T01:56:48.4439793Z [info]   at 
org.apache.spark.sql.errors.QueryExecutionErrors.divideByZeroError(QueryExecutionErrors.scala)
   2022-07-08T01:56:48.4441245Z [info]   at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage3.hashAgg_doAggregateWithKeysOutput_0$(Unknown
 Source)
   2022-07-08T01:56:48.4442235Z [info]   at 
org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage3.processNext(Unknown
 Source)
   2022-07-08T01:56:48.4443261Z [info]   at 
org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
   2022-07-08T01:56:48.4444051Z [info]   at 
org.apache.spark.sql.execution.WholeStageCodegenExec$$anon$1.hasNext(WholeStageCodegenExec.scala:760)
   2022-07-08T01:56:48.4444879Z [info]   at 
scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
   2022-07-08T01:56:48.4445474Z [info]   at 
scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
   2022-07-08T01:56:48.4446193Z [info]   at 
org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1917)
   2022-07-08T01:56:48.4447254Z [info]   at 
org.apache.spark.rdd.RDD.$anonfun$count$1(RDD.scala:1268)
   2022-07-08T01:56:48.4448107Z [info]   at 
org.apache.spark.rdd.RDD.$anonfun$count$1$adapted(RDD.scala:1268)
   2022-07-08T01:56:48.4449171Z [info]   at 
org.apache.spark.SparkContext.$anonfun$runJob$5(SparkContext.scala:2262)
   2022-07-08T01:56:48.4449891Z [info]   at 
org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:92)
   2022-07-08T01:56:48.4450452Z [info]   at 
org.apache.spark.scheduler.Task.run(Task.scala:139)
   2022-07-08T01:56:48.4451018Z [info]   at 
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:548)
   2022-07-08T01:56:48.4451581Z [info]   at 
org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1490)
   2022-07-08T01:56:48.4452202Z [info]   at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:551)
   2022-07-08T01:56:48.4452818Z [info]   at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
   2022-07-08T01:56:48.4453447Z [info]   at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
   2022-07-08T01:56:48.4453968Z [info]   at 
java.lang.Thread.run(Thread.java:750)
   ```
   
   https://github.com/apache/spark/runs/7244240118?check_suite_focus=true



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to