Iterative Programming by keeping data across micro-batches in spark-streaming?

2015-06-17 Thread Nipun Arora
Hi, Is there anyway in spark streaming to keep data across multiple micro-batches? Like in a HashMap or something? Can anyone make suggestions on how to keep data across iterations where each iteration is an RDD being processed in JavaDStream? This is especially the case when I am trying to

Re: Iterative Programming by keeping data across micro-batches in spark-streaming?

2015-06-17 Thread Silvio Fiorito
Hi, just answered in your other thread as well... Depending on your requirements, you can look at the updateStateByKey API From: Nipun Arora Date: Wednesday, June 17, 2015 at 10:51 PM To: user@spark.apache.orgmailto:user@spark.apache.org Subject: Iterative Programming by keeping data across

Re: Iterative Programming by keeping data across micro-batches in spark-streaming?

2015-06-17 Thread Nipun Arora
, 2015 at 10:51 PM To: user@spark.apache.org Subject: Iterative Programming by keeping data across micro-batches in spark-streaming? Hi, Is there anyway in spark streaming to keep data across multiple micro-batches? Like in a HashMap or something? Can anyone make suggestions on how to keep