Any idea?

Actually anyone using zeppelin 0.9+ on k8s, with spark interpreter scope per 
note ? 


On 2021/04/24 10:46:06, Sylvain Gibier <sylv...@apefactory.com> wrote: 
> Hi,
> 
> we have an issue with our current deployment of zeppelin on k8s, and more
> precisely with spark interpreter.
> 
> For reference - the spark context is: scala 2.12.10 / spark 2.4.7
> 
> We have a weird behaviour, running the spark interpreter in per note, scoped
> 
> To reproduce currently - we restart the spark interpreter in scoped per
> note, and create two notebooks (A & B) with the same following code:
> 
> %spark
> > import spark.implicits._
> >
> > List(1, 2, 3).toDS.map(_ + 1).show
> >
> 
> 1- we run notebook A successfully
> 2 - we run notebook B  - it fails with class cast exception
> 
> org.apache.spark.SparkException: Job aborted due to stage failure: Task 0
> > in stage 24.0 failed 4 times, most recent failure: Lost task 0.3 in stage
> > 24.0 (TID 161, 10.11.18.133, executor 2): java.lang.ClassCastException:
> > cannot assign instance of java.lang.invoke.SerializedLambda to field
> > org.apache.spark.rdd.MapPartitionsRDD.f of type scala.Function3 in instance
> > of org.apache.spark.rdd.MapPartitionsRDD at
> > java.io.ObjectStreamClass$FieldReflector.setObjFieldValues(ObjectStreamClass.java:2287)
> > at java.io.ObjectStreamClass.setObjFieldValues(ObjectStreamClass.java:1417)
> > at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2293)
> >
> 
> Anyone having a working zeppelin deployment with k8s / spark 2.4 - scala
> 2.12 ?
> 
> or let anyone interested to make some $$$ to help us fix the issue?
> 
> cheers
> 

Reply via email to