After watching Ryan Weald's talk about integrating Spark Streaming with
Algebird and Storehaus, I decided that I had to give this a try! :-) But I'm
having some rather basic problems.
My code looks a lot like the example Ryan gives. Here's the basic structure:
dstream.foreach(rdd => if(rdd.count != 0){ val data =
rdd.collect.toMap; store.multiMerge(data)})
where dstream is my Spark DStream and store is a Storehaus MySqlLongStore.
The problem is that the Spark Streaming CheckpointWriter wants to serialize the
store object and it can't (I end up getting a NotSerializableException).
Anybody have an example of working code? Is this problem specific to
MySqlLongStores, or is this generic?
Thanks,
Jim Donahue
Adobe