What version of Spark SQL?

On Sat, Nov 15, 2014 at 10:25 PM, Eric Zhen <zhpeng...@gmail.com> wrote:

> Hi all,
>
> We run SparkSQL on TPCDS benchmark Q19 with  spark.sql.codegen=true, we
> got exceptions as below, has anyone else saw these before?
>
> java.lang.ExceptionInInitializerError
>         at
> org.apache.spark.sql.execution.SparkPlan.newProjection(SparkPlan.scala:92)
>         at
> org.apache.spark.sql.execution.Exchange$$anonfun$execute$1$$anonfun$1.apply(Exchange.scala:51)
>         at
> org.apache.spark.sql.execution.Exchange$$anonfun$execute$1$$anonfun$1.apply(Exchange.scala:48)
>         at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:596)
>         at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:596)
>         at
> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
>         at
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
>         at
> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
>         at org.apache.spark.scheduler.Task.run(Task.scala:54)
>         at
> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:178)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:744)
> Caused by: java.lang.NullPointerException
>         at
> scala.reflect.internal.Types$TypeRef.computeHashCode(Types.scala:2358)
>         at scala.reflect.internal.Types$UniqueType.<init>(Types.scala:1304)
>         at scala.reflect.internal.Types$TypeRef.<init>(Types.scala:2341)
>         at
> scala.reflect.internal.Types$NoArgsTypeRef.<init>(Types.scala:2137)
>         at
> scala.reflect.internal.Types$TypeRef$$anon$6.<init>(Types.scala:2544)
>         at scala.reflect.internal.Types$TypeRef$.apply(Types.scala:2544)
>         at scala.reflect.internal.Types$class.typeRef(Types.scala:3615)
>         at scala.reflect.internal.SymbolTable.typeRef(SymbolTable.scala:13)
>         at
> scala.reflect.internal.Symbols$TypeSymbol.newTypeRef(Symbols.scala:2752)
>         at
> scala.reflect.internal.Symbols$TypeSymbol.typeConstructor(Symbols.scala:2806)
>         at
> scala.reflect.internal.Symbols$SymbolContextApiImpl.toTypeConstructor(Symbols.scala:103)
>         at
> scala.reflect.internal.Symbols$TypeSymbol.toTypeConstructor(Symbols.scala:2698)
>         at
> org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$$typecreator1$1.apply(CodeGenerator.scala:46)
>         at
> scala.reflect.api.TypeTags$WeakTypeTagImpl.tpe$lzycompute(TypeTags.scala:231)
>         at
> scala.reflect.api.TypeTags$WeakTypeTagImpl.tpe(TypeTags.scala:231)
>         at scala.reflect.api.TypeTags$class.typeOf(TypeTags.scala:335)
>         at scala.reflect.api.Universe.typeOf(Universe.scala:59)
>         at
> org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator.<init>(CodeGenerator.scala:46)
>         at
> org.apache.spark.sql.catalyst.expressions.codegen.GenerateProjection$.<init>(GenerateProjection.scala:29)
>         at
> org.apache.spark.sql.catalyst.expressions.codegen.GenerateProjection$.<clinit>(GenerateProjection.scala)
>         ... 15 more
> --
> Best Regards
>

Reply via email to