I tried with 1.6.2 and saw the same behavior. -Joey
On Tue, Oct 11, 2016 at 5:18 PM, Shixiong(Ryan) Zhu <shixi...@databricks.com> wrote: > There are some known issues in 1.6.0, e.g., > https://issues.apache.org/jira/browse/SPARK-12591 > > Could you try 1.6.1? > > On Tue, Oct 11, 2016 at 9:55 AM, Joey Echeverria <j...@rocana.com> wrote: >> >> I tried wrapping my Tuple class (which is generated by Avro) in a >> class that implements Serializable, but now I'm getting a >> ClassNotFoundException in my Spark application. The exception is >> thrown while trying to deserialize checkpoint state: >> >> https://gist.github.com/joey/7b374a2d483e25f15e20c0c4cb81b5cf >> >> I set some flags[1] on the JVM and I can see the class get loaded in the >> logs. >> >> Does anyone have any suggestions/recommendations for debugging class >> loading issues during checkpoint deserialization? >> >> I also looked into switching to byte[] for the state keys, but byte[] >> doesn't implement value-based equals() or hashCode(). I can't use >> ByteBuffer because it doesn't implement Serializable. Spark has a >> SerializableBuffer class that wraps ByteBuffer, but it also doesn't >> have value-based equals() or hashCode(). >> >> -Joey >> >> [1] -verbose:class -Dsun.misc.URLClassPath.debug >> >> On Mon, Oct 10, 2016 at 11:28 AM, Joey Echeverria <j...@rocana.com> wrote: >> > I do, I get the stack trace in this gist: >> > >> > https://gist.github.com/joey/d3bf040af31e854b3be374e2c016d7e1 >> > >> > The class it references, com.rocana.data.Tuple, is registered with >> > Kryo. Also, this is with 1.6.0 so if this behavior changed/got fixed >> > in a later release let me know. >> > >> > -Joey >> > >> > On Mon, Oct 10, 2016 at 9:54 AM, Shixiong(Ryan) Zhu >> > <shixi...@databricks.com> wrote: >> >> That's enough. Did you see any error? >> >> >> >> On Mon, Oct 10, 2016 at 5:08 AM, Joey Echeverria <j...@rocana.com> >> >> wrote: >> >>> >> >>> Hi Ryan! >> >>> >> >>> Do you know where I need to configure Kryo for this? I already have >> >>> spark.serializer=org.apache.spark.serializer.KryoSerializer in my >> >>> SparkConf and I registered the class. Is there a different >> >>> configuration setting for the state map keys? >> >>> >> >>> Thanks! >> >>> >> >>> -Joey >> >>> >> >>> On Sun, Oct 9, 2016 at 10:58 PM, Shixiong(Ryan) Zhu >> >>> <shixi...@databricks.com> wrote: >> >>> > You can use Kryo. It also implements KryoSerializable which is >> >>> > supported >> >>> > by >> >>> > Kryo. >> >>> > >> >>> > On Fri, Oct 7, 2016 at 11:39 AM, Joey Echeverria <j...@rocana.com> >> >>> > wrote: >> >>> >> >> >>> >> Looking at the source code for StateMap[1], which is used by >> >>> >> JavaPairDStream#mapWithState(), it looks like state keys are >> >>> >> serialized using an ObjectOutputStream. I couldn't find a reference >> >>> >> to >> >>> >> this restriction in the documentation. Did I miss that? >> >>> >> >> >>> >> Unless I'm mistaken, I'm guessing there isn't a way to use Kryo for >> >>> >> this serialization? >> >>> >> >> >>> >> Thanks! >> >>> >> >> >>> >> -Joey >> >>> >> >> >>> >> [1] >> >>> >> >> >>> >> >> >>> >> https://github.com/apache/spark/blob/master/streaming/src/main/scala/org/apache/spark/streaming/util/StateMap.scala#L251 >> >>> >> >> >>> >> >> >>> >> --------------------------------------------------------------------- >> >>> >> To unsubscribe e-mail: user-unsubscr...@spark.apache.org >> >>> >> >> >>> > >> >>> >> >>> >> >>> >> >>> -- >> >>> -Joey >> >> >> >> >> > >> > >> > >> > -- >> > -Joey >> >> >> >> -- >> -Joey > > -- -Joey --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org