Okay, thank you Micheal.

On Wed, Nov 19, 2014 at 3:45 AM, Michael Armbrust <mich...@databricks.com>
wrote:

> Those are probably related.  It looks like we are somehow not being thread
> safe when initializing various parts of the scala compiler.
>
> Since code gen is pretty experimental we probably won't have the resources
> to investigate backporting a fix.  However, if you can reproduce the
> problem in Spark 1.2 then please file a JIRA.
>
> On Mon, Nov 17, 2014 at 9:37 PM, Eric Zhen <zhpeng...@gmail.com> wrote:
>
>> Yes, it's always appears on a part of the whole tasks in a stage(i.e. 100/100
>> (65 failed)), and sometimes cause the stage to fail.
>>
>> And there is another error that I'm not sure if there is a correlation.
>>
>> java.lang.NoClassDefFoundError: Could not initialize class
>> org.apache.spark.sql.catalyst.expressions.codegen.GeneratePredicate$
>>         at
>> org.apache.spark.sql.execution.SparkPlan.newPredicate(SparkPlan.scala:114)
>>         at
>> org.apache.spark.sql.execution.Filter.conditionEvaluator$lzycompute(basicOperators.scala:55)
>>         at
>> org.apache.spark.sql.execution.Filter.conditionEvaluator(basicOperators.scala:55)
>>         at
>> org.apache.spark.sql.execution.Filter$$anonfun$2.apply(basicOperators.scala:58)
>>         at
>> org.apache.spark.sql.execution.Filter$$anonfun$2.apply(basicOperators.scala:57)
>>         at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:596)
>>         at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:596)
>>         at
>> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
>>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
>>         at
>> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
>>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
>>         at
>> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
>>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
>>         at
>> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
>>         at
>> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
>>         at org.apache.spark.scheduler.Task.run(Task.scala:54)
>>         at
>> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:178)
>>         at
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>         at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>         at java.lang.Thread.run(Thread.java:744)
>>
>> On Tue, Nov 18, 2014 at 11:41 AM, Michael Armbrust <
>> mich...@databricks.com> wrote:
>>
>>> Interesting, I believe we have run that query with version 1.1.0 with
>>> codegen turned on and not much has changed there.  Is the error
>>> deterministic?
>>>
>>> On Mon, Nov 17, 2014 at 7:04 PM, Eric Zhen <zhpeng...@gmail.com> wrote:
>>>
>>>> Hi Michael,
>>>>
>>>> We use Spark v1.1.1-rc1 with jdk 1.7.0_51 and scala 2.10.4.
>>>>
>>>> On Tue, Nov 18, 2014 at 7:09 AM, Michael Armbrust <
>>>> mich...@databricks.com> wrote:
>>>>
>>>>> What version of Spark SQL?
>>>>>
>>>>> On Sat, Nov 15, 2014 at 10:25 PM, Eric Zhen <zhpeng...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi all,
>>>>>>
>>>>>> We run SparkSQL on TPCDS benchmark Q19 with  spark.sql.codegen=true,
>>>>>> we got exceptions as below, has anyone else saw these before?
>>>>>>
>>>>>> java.lang.ExceptionInInitializerError
>>>>>>         at
>>>>>> org.apache.spark.sql.execution.SparkPlan.newProjection(SparkPlan.scala:92)
>>>>>>         at
>>>>>> org.apache.spark.sql.execution.Exchange$$anonfun$execute$1$$anonfun$1.apply(Exchange.scala:51)
>>>>>>         at
>>>>>> org.apache.spark.sql.execution.Exchange$$anonfun$execute$1$$anonfun$1.apply(Exchange.scala:48)
>>>>>>         at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:596)
>>>>>>         at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:596)
>>>>>>         at
>>>>>> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
>>>>>>         at
>>>>>> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
>>>>>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
>>>>>>         at
>>>>>> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
>>>>>>         at
>>>>>> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
>>>>>>         at org.apache.spark.scheduler.Task.run(Task.scala:54)
>>>>>>         at
>>>>>> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:178)
>>>>>>         at
>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>>>>         at
>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>>>>         at java.lang.Thread.run(Thread.java:744)
>>>>>> Caused by: java.lang.NullPointerException
>>>>>>         at
>>>>>> scala.reflect.internal.Types$TypeRef.computeHashCode(Types.scala:2358)
>>>>>>         at
>>>>>> scala.reflect.internal.Types$UniqueType.<init>(Types.scala:1304)
>>>>>>         at
>>>>>> scala.reflect.internal.Types$TypeRef.<init>(Types.scala:2341)
>>>>>>         at
>>>>>> scala.reflect.internal.Types$NoArgsTypeRef.<init>(Types.scala:2137)
>>>>>>         at
>>>>>> scala.reflect.internal.Types$TypeRef$$anon$6.<init>(Types.scala:2544)
>>>>>>         at
>>>>>> scala.reflect.internal.Types$TypeRef$.apply(Types.scala:2544)
>>>>>>         at
>>>>>> scala.reflect.internal.Types$class.typeRef(Types.scala:3615)
>>>>>>         at
>>>>>> scala.reflect.internal.SymbolTable.typeRef(SymbolTable.scala:13)
>>>>>>         at
>>>>>> scala.reflect.internal.Symbols$TypeSymbol.newTypeRef(Symbols.scala:2752)
>>>>>>         at
>>>>>> scala.reflect.internal.Symbols$TypeSymbol.typeConstructor(Symbols.scala:2806)
>>>>>>         at
>>>>>> scala.reflect.internal.Symbols$SymbolContextApiImpl.toTypeConstructor(Symbols.scala:103)
>>>>>>         at
>>>>>> scala.reflect.internal.Symbols$TypeSymbol.toTypeConstructor(Symbols.scala:2698)
>>>>>>         at
>>>>>> org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$$typecreator1$1.apply(CodeGenerator.scala:46)
>>>>>>         at
>>>>>> scala.reflect.api.TypeTags$WeakTypeTagImpl.tpe$lzycompute(TypeTags.scala:231)
>>>>>>         at
>>>>>> scala.reflect.api.TypeTags$WeakTypeTagImpl.tpe(TypeTags.scala:231)
>>>>>>         at scala.reflect.api.TypeTags$class.typeOf(TypeTags.scala:335)
>>>>>>         at scala.reflect.api.Universe.typeOf(Universe.scala:59)
>>>>>>         at
>>>>>> org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator.<init>(CodeGenerator.scala:46)
>>>>>>         at
>>>>>> org.apache.spark.sql.catalyst.expressions.codegen.GenerateProjection$.<init>(GenerateProjection.scala:29)
>>>>>>         at
>>>>>> org.apache.spark.sql.catalyst.expressions.codegen.GenerateProjection$.<clinit>(GenerateProjection.scala)
>>>>>>         ... 15 more
>>>>>> --
>>>>>> Best Regards
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> Best Regards
>>>>
>>>
>>>
>>
>>
>> --
>> Best Regards
>>
>
>


-- 
Best Regards

Reply via email to