Finally got it working by increasing the spark version in maven to 1.2.1

On Thu, Apr 9, 2015 at 12:30 PM, Mohit Anchlia <mohitanch...@gmail.com>
wrote:

> I changed the JDK to Oracle but I still get this error. Not sure what it
> means by "Stream class is incompatible with local class". I am using the
> following build on the server "spark-1.2.1-bin-hadoop2.4"
>
> 15/04/09 15:26:24 ERROR JobScheduler: Error running job streaming job
> 1428607584000 ms.0
> org.apache.spark.SparkException: Job aborted due to stage failure: Task 0
> in stage 15.0 failed 4 times, most recent failure: Lost task 0.3 in stage
> 15.0 (TID 66, ip-10-241-251-232.us-west-2.compute.internal):
> java.io.InvalidClassException: org.apache.spark.Aggregator; local class
> incompatible: stream classdesc serialVersionUID = 5032037208639381169,
> local class serialVersionUID = -9085606473104903453
>         at
> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617)
>
> On Wed, Apr 8, 2015 at 3:40 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>
>> bq. one is Oracle and the other is OpenJDK
>>
>> I don't have experience with mixed JDK's.
>> Can you try with using single JDK ?
>>
>> Cheers
>>
>> On Wed, Apr 8, 2015 at 3:26 PM, Mohit Anchlia <mohitanch...@gmail.com>
>> wrote:
>>
>>> For the build I am using java version "1.7.0_65" which seems to be the
>>> same as the one on the spark host. However one is Oracle and the other is
>>> OpenJDK. Does that make any difference?
>>>
>>> On Wed, Apr 8, 2015 at 1:24 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>>
>>>> What version of Java do you use to build ?
>>>>
>>>> Cheers
>>>>
>>>> On Wed, Apr 8, 2015 at 12:43 PM, Mohit Anchlia <mohitanch...@gmail.com>
>>>> wrote:
>>>>
>>>>> I am seeing the following, is this because of my maven version?
>>>>>
>>>>> 15/04/08 15:42:22 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID
>>>>> 0, ip-10-241-251-232.us-west-2.compute.internal):
>>>>> java.io.InvalidClassException: org.apache.spark.Aggregator; local class
>>>>> incompatible: stream classdesc serialVersionUID = 5032037208639381169,
>>>>> local class serialVersionUID = -9085606473104903453
>>>>>         at
>>>>> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617)
>>>>>         at
>>>>> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622)
>>>>>         at
>>>>> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
>>>>>         at
>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
>>>>>         at
>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>>         at
>>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>>>>>         at
>>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>>>>>         at
>>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>>>>         at
>>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>>
>>>>>
>>>>> <dependency>
>>>>>
>>>>> <groupId>org.apache.spark</groupId>
>>>>>
>>>>> <artifactId>spark-streaming_2.10</artifactId>
>>>>>
>>>>> <version>1.2.0</version>
>>>>>
>>>>> </dependency>
>>>>>
>>>>> <dependency>
>>>>>
>>>>> <groupId>org.apache.spark</groupId>
>>>>>
>>>>> <artifactId>spark-core_2.10</artifactId>
>>>>>
>>>>> <version>1.2.0</version>
>>>>>
>>>>> </dependency>
>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to