yes, it works after checkout branch-0.8.

thanks.


On Tue, Oct 29, 2013 at 12:51 PM, Patrick Wendell <pwend...@gmail.com>wrote:

> If you just add the "extends Serializable" changes from here it should
> work.
>
> On Tue, Oct 29, 2013 at 9:36 AM, Patrick Wendell <pwend...@gmail.com>
> wrote:
> > This was fixed on 0.8 branch and master:
> > https://github.com/apache/incubator-spark/pull/63/files
> >
> > - Patrick
> >
> > On Tue, Oct 29, 2013 at 9:17 AM, Thunder Stumpges
> > <thunder.stump...@gmail.com> wrote:
> >> I vaguely remember running into this same error. It says there
> >> "java.io.NotSerializableException:
> >> org.apache.spark.streaming.examples.clickstream.PageView"... can you
> >> check the PageView class in the examples and make sure it has the
> >> @serializable directive? I seem to remember having to add it.
> >>
> >> good luck,
> >> Thunder
> >>
> >>
> >> On Tue, Oct 29, 2013 at 6:54 AM, dachuan <hdc1...@gmail.com> wrote:
> >>> Hi,
> >>>
> >>> I have tried the clickstream example, it runs into an exception,
> anybody met
> >>> this before?
> >>>
> >>> Since the program mentioned "local[2]", so I run it in my local
> machine.
> >>>
> >>> thanks in advance,
> >>> dachuan.
> >>>
> >>> Log Snippet 1:
> >>>
> >>> 13/10/29 08:50:25 INFO scheduler.DAGScheduler: Submitting 46 missing
> tasks
> >>> from Stage 12 (MapPartitionsRDD[63] at combineByKey at
> >>> ShuffledDStream.scala:41)
> >>> 13/10/29 08:50:25 INFO local.LocalTaskSetManager: Size of task 75 is
> 4230
> >>> bytes
> >>> 13/10/29 08:50:25 INFO local.LocalScheduler: Running 75
> >>> 13/10/29 08:50:25 INFO spark.CacheManager: Cache key is rdd_9_0
> >>> 13/10/29 08:50:25 INFO spark.CacheManager: Computing partition
> >>> org.apache.spark.rdd.BlockRDDPartition@0
> >>> 13/10/29 08:50:25 WARN storage.BlockManager: Putting block rdd_9_0
> failed
> >>> 13/10/29 08:50:25 INFO local.LocalTaskSetManager: Loss was due to
> >>> java.io.NotSerializableException
> >>> java.io.NotSerializableException:
> >>> org.apache.spark.streaming.examples.clickstream.PageView
> >>>
> >>> Log Snippet 2:
> >>> org.apache.spark.SparkException: Job failed: Task 12.0:0 failed more
> than 4
> >>> times; aborting job java.io.NotSerializableException:
> >>> org.apache.spark.streaming.examples.clickstream.PageView
> >>>         at
> >>>
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:760)
> >>>         at
> >>>
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:758)
> >>>         at
> >>>
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:60)
> >>>         at
> >>> scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
> >>>         at
> >>>
> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:758)
> >>>         at
> >>>
> org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:379)
> >>>         at
> >>> org.apache.spark.scheduler.DAGScheduler.org
> $apache$spark$scheduler$DAGScheduler$$run(DAGScheduler.scala:441)
> >>>         at
> >>>
> org.apache.spark.scheduler.DAGScheduler$$anon$1.run(DAGScheduler.scala:149)
> >>>
> >>> Two commands that run this app:
> >>> ./run-example
> >>> org.apache.spark.streaming.exampl.clickstream.PageViewGenerator 44444
> 10
> >>> ./run-example
> org.apache.spark.streaming.examples.clickstream.PageViewStream
> >>> errorRatePerZipCode localhost 44444
> >>>
>



-- 
Dachuan Huang
Cellphone: 614-390-7234
2015 Neil Avenue
Ohio State University
Columbus, Ohio
U.S.A.
43210

Reply via email to