Hi to all,

I'm building a recommendation system to my application.
I have a set of logs (that contains the user info, the hour, the button
that was clicked ect...) that arrive to my Flink by kafka, then I save
every log in a HDFS (HADOOP), but know I have a problem, I want to apply ML
to (all) my data.

I think in 2 scenarios:
First : Transform my DataStream in a DataSet and perform the ML task. It is
possible?
Second : Preform a task in flink that get the data from Hadoop and perform
the ML task.

What is the best way to do it?

I already check the IncrementalLearningSkeleton but I didn't understand how
to apply that to an actual real case. Is there some complex example that I
could look?
(
https://github.com/apache/flink/tree/master/flink-examples/flink-examples-streaming/src/main/java/org/apache/flink/streaming/examples/ml
)

Another thing that I would like to ask is how to perform the second
scenario, where I need to perform this task every hour, what it is the best
way to do it?

Thanks,
Fábio Dias.

Reply via email to