As of Spark 1.2 you can do Streaming k-means, see examples here:
http://spark.apache.org/docs/latest/mllib-clustering.html#examples-1
Best,
Reza

On Fri, Dec 26, 2014 at 1:36 AM, vishnu <johnfedrickena...@gmail.com> wrote:

> Hi,
>
> Say I have created a clustering model using KMeans for 100million
> transactions at time t1. I am using streaming and say for every 1 hour i
> need to update my existing model. How do I do it. Should it include every
> time all the data or can it be incrementally updated.
>
> If I can do an incrementally updating , how do i do it.
>
> Thanks,
> Vishnu
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/how-to-do-incremental-model-updates-using-spark-streaming-and-mllib-tp20862.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to