Are you expecting something like this?

val data = ssc.textFileStream("hdfs://akhldz:9000/input/")


     val rdd = ssc.sparkContext.parallelize(Seq("foo", "bar"))

     val sample = data.foreachRDD(x=> {
       val new_rdd = x.union(rdd)
       new_rdd.saveAsTextFile("hdfs://akhldz:9000/output/")
     })

Thanks
Best Regards

On Fri, Oct 31, 2014 at 10:46 AM, Tobias Pfeiffer <t...@preferred.jp> wrote:

> Harold,
>
> just mentioning it in case you run into it: If you are in a separate
> thread, there are apparently stricter limits to what you can and cannot
> serialize:
>
> val someVal
> future {
>   // be very careful with defining RDD operations using someVal here
>   val myLocalVal = someVal
>   // use myLocalVal instead
> }
>
> On Thu, Oct 30, 2014 at 4:55 PM, Harold Nguyen <har...@nexgate.com> wrote:
>
>> In Spark Streaming, when I do "foreachRDD" on my DStreams, I get a
>> NonSerializable exception when I try to do something like:
>>
>> DStream.foreachRDD( rdd => {
>>   var sc.parallelize(Seq(("test", "blah")))
>> })
>>
>
> Is this the code you are actually using? "var sc.parallelize(...)" doesn't
> really look like valid Scala to me.
>
> Tobias
>
>
>

Reply via email to