This worked with java serialization.I am using 1.2.0 you are right if I use
1.2.1 or 1.3.0 this issue will not occur
I will test this and let you know

On 15 April 2015 at 19:48, Imran Rashid <iras...@cloudera.com> wrote:

> oh interesting.  The suggested workaround is to wrap the result from
> collectAsMap into another hashmap, you should try that:
>
> Map<Long, MatcherReleventData> matchData =RddForMarch.collectAsMap();
> Map<String, String> tmp = new HashMap<String, String>(matchData);
> final Broadcast<Map<Long, MatcherReleventData>> dataMatchGlobal =
> jsc.broadcast(tmp);
>
> Can you please clarify:
> * Does it work w/ java serialization in the end?  Or is this kryo only?
> * which Spark version you are using? (one of the relevant bugs was fixed
> in 1.2.1 and 1.3.0)
>
>
>
> On Wed, Apr 15, 2015 at 9:06 AM, Jeetendra Gangele <gangele...@gmail.com>
> wrote:
>
>> This looks like known issue? check this out
>>
>> http://apache-spark-user-list.1001560.n3.nabble.com/java-io-InvalidClassException-org-apache-spark-api-java-JavaUtils-SerializableMapWrapper-no-valid-cor-td20034.html
>>
>> Can you please suggest any work around I am broad casting HashMap return
>> from RDD.collectasMap().
>>
>> On 15 April 2015 at 19:33, Imran Rashid <iras...@cloudera.com> wrote:
>>
>>> this is a really strange exception ... I'm especially surprised that it
>>> doesn't work w/ java serialization.  Do you think you could try to boil it
>>> down to a minimal example?
>>>
>>> On Wed, Apr 15, 2015 at 8:58 AM, Jeetendra Gangele <gangele...@gmail.com
>>> > wrote:
>>>
>>>> Yes Without Kryo it did work out.when I remove kryo registration it did
>>>> worked out
>>>>
>>>> On 15 April 2015 at 19:24, Jeetendra Gangele <gangele...@gmail.com>
>>>> wrote:
>>>>
>>>>> its not working with the combination of Broadcast.
>>>>> Without Kyro also not working.
>>>>>
>>>>>
>>>>> On 15 April 2015 at 19:20, Akhil Das <ak...@sigmoidanalytics.com>
>>>>> wrote:
>>>>>
>>>>>> Is it working without kryo?
>>>>>>
>>>>>> Thanks
>>>>>> Best Regards
>>>>>>
>>>>>> On Wed, Apr 15, 2015 at 6:38 PM, Jeetendra Gangele <
>>>>>> gangele...@gmail.com> wrote:
>>>>>>
>>>>>>> Hi All I am getting below exception while using Kyro serializable
>>>>>>> with broadcast variable. I am broadcating a hasmap with below line.
>>>>>>>
>>>>>>> Map<Long, MatcherReleventData> matchData =RddForMarch.collectAsMap();
>>>>>>> final Broadcast<Map<Long, MatcherReleventData>> dataMatchGlobal =
>>>>>>> jsc.broadcast(matchData);
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> 15/04/15 12:58:51 ERROR executor.Executor: Exception in task 0.3 in
>>>>>>> stage 4.0 (TID 7)
>>>>>>> java.io.IOException: java.lang.UnsupportedOperationException
>>>>>>>         at
>>>>>>> org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1003)
>>>>>>>         at
>>>>>>> org.apache.spark.broadcast.TorrentBroadcast.readBroadcastBlock(TorrentBroadcast.scala:164)
>>>>>>>         at
>>>>>>> org.apache.spark.broadcast.TorrentBroadcast._value$lzycompute(TorrentBroadcast.scala:64)
>>>>>>>         at
>>>>>>> org.apache.spark.broadcast.TorrentBroadcast._value(TorrentBroadcast.scala:64)
>>>>>>>         at
>>>>>>> org.apache.spark.broadcast.TorrentBroadcast.getValue(TorrentBroadcast.scala:87)
>>>>>>>         at
>>>>>>> com.insideview.yam.CompanyMatcher$4.call(CompanyMatcher.java:103)
>>>>>>>         at
>>>>>>> com.insideview.yam.CompanyMatcher$4.call(CompanyMatcher.java:1)
>>>>>>>         at
>>>>>>> org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1002)
>>>>>>>         at
>>>>>>> org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1002)
>>>>>>>         at
>>>>>>> scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
>>>>>>>         at
>>>>>>> org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:204)
>>>>>>>         at
>>>>>>> org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:58)
>>>>>>>         at
>>>>>>> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
>>>>>>>         at
>>>>>>> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
>>>>>>>         at org.apache.spark.scheduler.Task.run(Task.scala:56)
>>>>>>>         at
>>>>>>> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
>>>>>>>         at
>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>>>>>         at
>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>>> Caused by: java.lang.UnsupportedOperationException
>>>>>>>         at java.util.AbstractMap.put(AbstractMap.java:203)
>>>>>>>         at
>>>>>>> com.esotericsoftware.kryo.serializers.MapSerializer.read(MapSerializer.java:135)
>>>>>>>         at
>>>>>>> com.esotericsoftware.kryo.serializers.MapSerializer.read(MapSerializer.java:17)
>>>>>>>         at
>>>>>>> com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:729)
>>>>>>>         at
>>>>>>> org.apache.spark.serializer.KryoDeserializationStream.readObject(KryoSerializer.scala:142)
>>>>>>>         at
>>>>>>> org.apache.spark.broadcast.TorrentBroadcast$.unBlockifyObject(TorrentBroadcast.scala:216)
>>>>>>>         at
>>>>>>> org.apache.spark.broadcast.TorrentBroadcast$$anonfun$readBroadcastBlock$1.apply(TorrentBroadcast.scala:177)
>>>>>>>         at
>>>>>>> org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1000)
>>>>>>>         ... 18 more
>>>>>>> 15/04/15 12:58:51 INFO executor.CoarseGrainedExecutorBackend: Driver
>>>>>>> commanded a shutdown
>>>>>>> 15/04/15 12:58:51 INFO storage.MemoryStore: MemoryStore cleared
>>>>>>> 15/04/15 12:58:51 INFO storage.BlockManager: BlockManager stopped
>>>>>>> 15/04/15 12:58:51 INFO
>>>>>>> remote.RemoteActorRefProvider$RemotingTerminator: Shutting down remote
>>>>>>> daemon.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>>
>>>>
>>>
>>
>>
>>
>>
>

Reply via email to