Another way to do it is to extract your filters as SQL code and load it in a transform – which allows you to change the filters at runtime.
Inside the transform you could apply the filters by goind RDD -> DF -> SQL -> RDD. Lastly, depending on how complex your filters are, you could skip SQL and create your own mini-DSL that runs inside transform. I’d definitely start here if the filter predicates are simple enough… -adrian From: Stefano Baghino Date: Wednesday, November 4, 2015 at 10:15 AM To: Cassa L Cc: user Subject: Re: Rule Engine for Spark Hi LCassa, unfortunately I don't have actual experience on this matter, however for a similar use case I have briefly evaluated Decision<https://github.com/Stratio/Decision> (then called literally Streaming CEP Engine) and it looked interesting. I hope it may help. On Wed, Nov 4, 2015 at 1:42 AM, Cassa L <lcas...@gmail.com<mailto:lcas...@gmail.com>> wrote: Hi, Has anyone used rule engine with spark streaming? I have a case where data is streaming from Kafka and I need to apply some rules on it (instead of hard coding in a code). Thanks, LCassa -- BR, Stefano Baghino Software Engineer @ Radicalbit