Hi, I’m new on Spark and I’m trying to understand if it can fit my use case.
I have the following scenario. I have a file (it can be a log file, .txt, .csv, .xml or .json, I can produce the data in whatever format I prefer) with some data, e.g.: *Event “X”, City “Y”, Zone “Z”* with different events, cities and zones. This data can be represented by string (like the one I wrote) in a .txt, or by XML , CSV, or JSON, as I wish. I can also send this data through TCP Socket, if I need it. What I really want to do is to *correlate each single entry with other similar entries by declaring rules*. For example, I want to declare some rules on the data flow: if I received event X1 and event X2 in same city and same zone, I’ll want to do something (execute a .bat script, write a log file, etc). Same thing if I received the same string multiple times, or whatever rule I want to produce with these data strings. I’m trying to understand if Apache Spark can fit my use case. The only input data will be these strings from this file. Can I correlate these events and how? Is there a GUI to do it? Any hints and advices will be appreciated. Best regards, Simone -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org