helium repository down ?
Hi, we noticed that we are no longer able to download/ access to the helium registry ? It seems https://s3.amazonaws.com/helium-package/helium.json is no longer available. Known issue ? or where can I find the original helium registry to run it on our own infrastructure? Cheers, Sylvain *Sylvain Gibie*r cell: +49 172 666 2435 ape factory GmbH Rumfordstr. 21a, 80469 München https://apefactory.com' Geschäftsführer: Sylvain Gibier Sitz: München; Registergericht: München, HRB 221698
Re: Zeppelin 0.9 / Kubernetes / Spark interpreter
Hi, Cf. ZEPPELIN-5337. Switching to isolated mode is not really an option - as it means one spark interpreter per note. per user -- which consumes a lot of resources, as there is no mechanism to clean k8s pods created afterwards. The scope mode allows us to share the spark interpreter along with our 100+ analysts. On Fri, Apr 30, 2021 at 5:05 PM moon soo Lee wrote: > Hi, > > Thanks for sharing the issue. > > I tried zeppelin 0.9+ on k8s with per note scoped, scala 2.12, spark 3.0+. > And I could reproduce the problem. But isolated mode works without problem. > Does isolated mode work for your use case? > > Best, > moon > > > > On Tue, Apr 27, 2021 at 12:39 PM Sylvain Gibier > wrote: > >> Any idea? >> >> Actually anyone using zeppelin 0.9+ on k8s, with spark interpreter scope >> per note ? >> >> >> On 2021/04/24 10:46:06, Sylvain Gibier wrote: >> > Hi, >> > >> > we have an issue with our current deployment of zeppelin on k8s, and >> more >> > precisely with spark interpreter. >> > >> > For reference - the spark context is: scala 2.12.10 / spark 2.4.7 >> > >> > We have a weird behaviour, running the spark interpreter in per note, >> scoped >> > >> > To reproduce currently - we restart the spark interpreter in scoped per >> > note, and create two notebooks (A & B) with the same following code: >> > >> > %spark >> > > import spark.implicits._ >> > > >> > > List(1, 2, 3).toDS.map(_ + 1).show >> > > >> > >> > 1- we run notebook A successfully >> > 2 - we run notebook B - it fails with class cast exception >> > >> > org.apache.spark.SparkException: Job aborted due to stage failure: Task >> 0 >> > > in stage 24.0 failed 4 times, most recent failure: Lost task 0.3 in >> stage >> > > 24.0 (TID 161, 10.11.18.133, executor 2): >> java.lang.ClassCastException: >> > > cannot assign instance of java.lang.invoke.SerializedLambda to field >> > > org.apache.spark.rdd.MapPartitionsRDD.f of type scala.Function3 in >> instance >> > > of org.apache.spark.rdd.MapPartitionsRDD at >> > > >> java.io.ObjectStreamClass$FieldReflector.setObjFieldValues(ObjectStreamClass.java:2287) >> > > at >> java.io.ObjectStreamClass.setObjFieldValues(ObjectStreamClass.java:1417) >> > > at >> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2293) >> > > >> > >> > Anyone having a working zeppelin deployment with k8s / spark 2.4 - scala >> > 2.12 ? >> > >> > or let anyone interested to make some $$$ to help us fix the issue? >> > >> > cheers >> > >> >
Re: Zeppelin 0.9 / Kubernetes / Spark interpreter
Any idea? Actually anyone using zeppelin 0.9+ on k8s, with spark interpreter scope per note ? On 2021/04/24 10:46:06, Sylvain Gibier wrote: > Hi, > > we have an issue with our current deployment of zeppelin on k8s, and more > precisely with spark interpreter. > > For reference - the spark context is: scala 2.12.10 / spark 2.4.7 > > We have a weird behaviour, running the spark interpreter in per note, scoped > > To reproduce currently - we restart the spark interpreter in scoped per > note, and create two notebooks (A & B) with the same following code: > > %spark > > import spark.implicits._ > > > > List(1, 2, 3).toDS.map(_ + 1).show > > > > 1- we run notebook A successfully > 2 - we run notebook B - it fails with class cast exception > > org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 > > in stage 24.0 failed 4 times, most recent failure: Lost task 0.3 in stage > > 24.0 (TID 161, 10.11.18.133, executor 2): java.lang.ClassCastException: > > cannot assign instance of java.lang.invoke.SerializedLambda to field > > org.apache.spark.rdd.MapPartitionsRDD.f of type scala.Function3 in instance > > of org.apache.spark.rdd.MapPartitionsRDD at > > java.io.ObjectStreamClass$FieldReflector.setObjFieldValues(ObjectStreamClass.java:2287) > > at java.io.ObjectStreamClass.setObjFieldValues(ObjectStreamClass.java:1417) > > at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2293) > > > > Anyone having a working zeppelin deployment with k8s / spark 2.4 - scala > 2.12 ? > > or let anyone interested to make some $$$ to help us fix the issue? > > cheers >
Zeppelin 0.9 / Kubernetes / Spark interpreter
Hi, we have an issue with our current deployment of zeppelin on k8s, and more precisely with spark interpreter. For reference - the spark context is: scala 2.12.10 / spark 2.4.7 We have a weird behaviour, running the spark interpreter in per note, scoped To reproduce currently - we restart the spark interpreter in scoped per note, and create two notebooks (A & B) with the same following code: %spark > import spark.implicits._ > > List(1, 2, 3).toDS.map(_ + 1).show > 1- we run notebook A successfully 2 - we run notebook B - it fails with class cast exception org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 > in stage 24.0 failed 4 times, most recent failure: Lost task 0.3 in stage > 24.0 (TID 161, 10.11.18.133, executor 2): java.lang.ClassCastException: > cannot assign instance of java.lang.invoke.SerializedLambda to field > org.apache.spark.rdd.MapPartitionsRDD.f of type scala.Function3 in instance > of org.apache.spark.rdd.MapPartitionsRDD at > java.io.ObjectStreamClass$FieldReflector.setObjFieldValues(ObjectStreamClass.java:2287) > at java.io.ObjectStreamClass.setObjFieldValues(ObjectStreamClass.java:1417) > at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2293) > Anyone having a working zeppelin deployment with k8s / spark 2.4 - scala 2.12 ? or let anyone interested to make some $$$ to help us fix the issue? cheers
slack invitation
Hi, Could you invite me to the zeppelin slack workspace ? Thanks sylvain