I changed the JDK to Oracle but I still get this error. Not sure what it
means by "Stream class is incompatible with local class". I am using the
following build on the server "spark-1.2.1-bin-hadoop2.4"

15/04/09 15:26:24 ERROR JobScheduler: Error running job streaming job
1428607584000 ms.0
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0
in stage 15.0 failed 4 times, most recent failure: Lost task 0.3 in stage
15.0 (TID 66, ip-10-241-251-232.us-west-2.compute.internal):
java.io.InvalidClassException: org.apache.spark.Aggregator; local class
incompatible: stream classdesc serialVersionUID = 5032037208639381169,
local class serialVersionUID = -9085606473104903453
        at
java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617)

On Wed, Apr 8, 2015 at 3:40 PM, Ted Yu <yuzhih...@gmail.com> wrote:

> bq. one is Oracle and the other is OpenJDK
>
> I don't have experience with mixed JDK's.
> Can you try with using single JDK ?
>
> Cheers
>
> On Wed, Apr 8, 2015 at 3:26 PM, Mohit Anchlia <mohitanch...@gmail.com>
> wrote:
>
>> For the build I am using java version "1.7.0_65" which seems to be the
>> same as the one on the spark host. However one is Oracle and the other is
>> OpenJDK. Does that make any difference?
>>
>> On Wed, Apr 8, 2015 at 1:24 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>
>>> What version of Java do you use to build ?
>>>
>>> Cheers
>>>
>>> On Wed, Apr 8, 2015 at 12:43 PM, Mohit Anchlia <mohitanch...@gmail.com>
>>> wrote:
>>>
>>>> I am seeing the following, is this because of my maven version?
>>>>
>>>> 15/04/08 15:42:22 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID
>>>> 0, ip-10-241-251-232.us-west-2.compute.internal):
>>>> java.io.InvalidClassException: org.apache.spark.Aggregator; local class
>>>> incompatible: stream classdesc serialVersionUID = 5032037208639381169,
>>>> local class serialVersionUID = -9085606473104903453
>>>>         at
>>>> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617)
>>>>         at
>>>> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622)
>>>>         at
>>>> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
>>>>         at
>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
>>>>         at
>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>         at
>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>>>>         at
>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>>>>         at
>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>>>         at
>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>>
>>>>
>>>> <dependency>
>>>>
>>>> <groupId>org.apache.spark</groupId>
>>>>
>>>> <artifactId>spark-streaming_2.10</artifactId>
>>>>
>>>> <version>1.2.0</version>
>>>>
>>>> </dependency>
>>>>
>>>> <dependency>
>>>>
>>>> <groupId>org.apache.spark</groupId>
>>>>
>>>> <artifactId>spark-core_2.10</artifactId>
>>>>
>>>> <version>1.2.0</version>
>>>>
>>>> </dependency>
>>>>
>>>>
>>>
>>
>

Reply via email to