browse/SPARK-3447
>
>
>
> -
> Thanks & Regards,
> Mohan
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/Kryo-fails-with-avro-having-Arrays-and-unions-but-succeeds-with-simple-avro-tp14549p14557.html
> Sent from
Added some more info on this issue in the tracker Spark-3447
https://issues.apache.org/jira/browse/SPARK-3447
-
Thanks & Regards,
Mohan
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Kryo-fails-with-avro-having-Arrays-and-unions-but-succeeds-
*I am facing similar issue to Spark-3447 with spark streaming Api, Kryo
Serializer, Avro messages.
If avro message is simple, its fine. but if the avro message has
Union/Arrays its failing with the exception Below:*
ERROR scheduler.JobScheduler: Error running job streaming job 1411043845000
ms.0
Hi,I was able to set this parameter in my application to resolve this issue:
set("spark.kryoserializer.buffer.mb", "256")
Please let me know if this helps.
Date: Mon, 18 Aug 2014 21:50:02 +0800
From: dujinh...@hzduozhun.com
To: user@spark.apache.org
Subject: spark kryo se
hi all,
In RDD map , i invoke an object that is *Serialized* by java standard ,
and exception ::
com.esotericsoftware.kryo.KryoException: Buffer overflow. Available: 0,
required: 13
at com.esotericsoftware.kryo.io.Output.require(Output.java:138)
at com.esotericsoftware.kryo.io.Output.writeAscii_
Hi All,
I was doing a groupBy and apparently some keys were very frequent making
the serializer fail with buffer overflow exception. I did not need a
groupBy so I switched to combineByKey in this case but would like to know
how to increase the kryo buffer sizes to avoid this error. I hope there is
any pointers to this issue.
Thanks
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Issue-using-kryo-serilization-tp11129p11191.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
No,it doesn't implement serializable..It's third party class
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Issue-using-kryo-serilization-tp11129p11136.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
Does the class your serializing implement serializable?
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Issue-using-kryo-serilization-tp11129p11134.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
Yes,I did enable that
conf.set("spark.serializer",
"org.apache.spark.serializer.KryoSerializer")
conf.set("spark.kryo.registrator", "com.bigdata.MyRegistrator")
--
View this message in context:
http://apache-spark-user-list.1001560.n3.
Did you enable Kryo and have it use your registrator
using spark.serializer=org.apache.spark.serializer.KryoSerializer
and spark.kryo.registrator=mypackage.MyRegistrator ? It looks like the
serializer being used is the default Java one
http://spark.apache.org/docs/latest/tuning.html#data
I'm new to spark programming and here I'm trying to use third party class in
map with kryo serializer
val deviceApi = new DeviceApi()
deviceApi.loadDataFromStream(this.getClass.getClassLoader.getResourceAsStream("20140730.json"))
val properties = uaRDD1.map(line =>
g loaded.
On Fri, Jul 25, 2014 at 2:27 PM, Gary Malouf wrote:
> After upgrading to Spark 1.0.1 from 0.9.1 everything seemed to be going
> well. Looking at the Mesos slave logs, I noticed:
>
> ERROR KryoSerializer: Failed to run spark.kryo.registrator
> java.lang.Cla
After upgrading to Spark 1.0.1 from 0.9.1 everything seemed to be going
well. Looking at the Mesos slave logs, I noticed:
ERROR KryoSerializer: Failed to run spark.kryo.registrator
java.lang.ClassNotFoundException:
com/mediacrossing/verrazano/kryo/MxDataRegistrator
My spark-env.sh has the
Hi there,
I've been sucessfully using the precompiled Spark 1.0.0 Java api on a small
cluster in standalone mode. However, when I try to use Kryo serializer by
adding
conf.set("spark.serializer","org.apache.spark.serializer.KryoSerializer");
as suggested, Spark crash
Hi
My setup is to use localMode standalone, Sprak 1.0.0 release version, scala
2.10.4
I made a job that receive serialized object from Kafka broker. The objects
are serialized using kryo.
The code :
val sparkConf = new
SparkConf().setMaster("local[4]").setAppName("SparkTe
ginal Message-
From: wxhsdp [mailto:wxh...@gmail.com]
Sent: Wednesday, July 09, 2014 5:47 PM
To: u...@spark.incubator.apache.org
Subject: Re: Kryo is slower, and the size saving is minimal
i'am not familiar with kryo and my opinion may be not right. in my case,
kryo only saves about 5% of th
i'am not familiar with kryo and my opinion may be not right. in my case, kryo
only saves about 5% of the original size when dealing with primitive types
such as Arrays. i'am not sure whether it is the common case.
--
View this message in context:
http://apache-spark-user-list.
Hi,
For my test case, using Kryo serializer does not help.
It is slower than default Java serializer, and the size saving is minimal.
I've registered almost all classes to the Kryo registrator.
What is happening to my test case?
Have Anyone experienced a case like this?
Is this supposed to be supported? It doesn't work, at least in mesos fine
grained mode. First it fails a bunch of times because it can't find my
registrator class because my assembly jar hasn't been fetched like so:
java.lang.ClassNotFoundException: pickles.kryo.PicklesRegistrator
at java.
ay 2, 2014 at 3:35 PM, Soren Macbeth wrote:
>
>> Hallo,
>>
>> I've getting this rather crazy kryo exception trying to run my spark job:
>>
>> Exception in thread "main" org.apache.spark.SparkException: Job aborted:
>> E
so
it seems that it dying while trying to fetch results from my tasks to
return back to the driver.
Am I close?
On Fri, May 2, 2014 at 3:35 PM, Soren Macbeth wrote:
> Hallo,
>
> I've getting this rather crazy kryo exception trying to run my spark job:
>
> Ex
Hallo,
I've getting this rather crazy kryo exception trying to run my spark job:
Exception in thread "main" org.apache.spark.SparkException: Job aborted:
Exception while deserializing and fetching task:
com.esotericsoftware.kryo.KryoException:
java.lang.IllegalArgumentExcepti
For me, PageRank fails when I use Kryo (works fine if I don't). I
found the same problem reported here:
https://groups.google.com/forum/#!topic/spark-users/unngi3JdRk8 .
Has this been resolved?
I'm not launching code from spark-shell. I tried registering
GraphKryoRegistrator (instead
Kryo won’t make a major impact on PySpark because it just stores data as byte[]
objects, which are fast to serialize even with Java. But it may be worth a try
— you would just set spark.serializer and not try to register any classes. What
might make more impact is storing data as
I'm looking at the Tuning Guide suggestion to use Kryo instead of default
serialization. My questions:
Does pyspark use Java serialization by default, as Scala spark does? If
so, then...
can I use Kryo with pyspark instead? The instructions say I should
register my classes with the
Hi Patrick,
Thanks for your reply.
I am guessing even an array type will be registered automatically. Is this
correct?
Thanks,
Pradeep
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Kryo-serialization-does-not-compress-tp2042p2400.html
Sent from the
other question, it's possible that serializing doesn't
provide a big space savings for your objects, especially if you are
serializing mostly primitive types. It depends a bit what the type of
the object it is. One thing is, it would be good to register all of
the object types you plan to ser
We are trying to use kryo serialization, but with kryo serialization ON the
memory consumption does not change. We have tried this on multiple sets of
data.
We have also checked the logs of Kryo serialization and have confirmed that
Kryo is being used.
Can somebody please help us with this?
The
Has no one ever registered generic classes in scala? Is it possible?
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Kryo-Registration-class-is-not-registered-but-Log-TRACE-says-otherwise-tp2077p2182.html
Sent from the Apache Spark User List mailing list
(511.5 MB)
which uses Kryo serialization.
Both consumes almost equivalent storage (519.1 MB vs 511.5 MB respectively).
Is this behavior expected?
Because we were under the impression that kryo serialization is efficient
and were expecting it to compress further.
Also,we have noticed that when we
201 - 231 of 231 matches
Mail list logo