Any idea? Actually anyone using zeppelin 0.9+ on k8s, with spark interpreter scope per note ?
On 2021/04/24 10:46:06, Sylvain Gibier <sylv...@apefactory.com> wrote: > Hi, > > we have an issue with our current deployment of zeppelin on k8s, and more > precisely with spark interpreter. > > For reference - the spark context is: scala 2.12.10 / spark 2.4.7 > > We have a weird behaviour, running the spark interpreter in per note, scoped > > To reproduce currently - we restart the spark interpreter in scoped per > note, and create two notebooks (A & B) with the same following code: > > %spark > > import spark.implicits._ > > > > List(1, 2, 3).toDS.map(_ + 1).show > > > > 1- we run notebook A successfully > 2 - we run notebook B - it fails with class cast exception > > org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 > > in stage 24.0 failed 4 times, most recent failure: Lost task 0.3 in stage > > 24.0 (TID 161, 10.11.18.133, executor 2): java.lang.ClassCastException: > > cannot assign instance of java.lang.invoke.SerializedLambda to field > > org.apache.spark.rdd.MapPartitionsRDD.f of type scala.Function3 in instance > > of org.apache.spark.rdd.MapPartitionsRDD at > > java.io.ObjectStreamClass$FieldReflector.setObjFieldValues(ObjectStreamClass.java:2287) > > at java.io.ObjectStreamClass.setObjFieldValues(ObjectStreamClass.java:1417) > > at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2293) > > > > Anyone having a working zeppelin deployment with k8s / spark 2.4 - scala > 2.12 ? > > or let anyone interested to make some $$$ to help us fix the issue? > > cheers >