Thanks for reply. How about DROOLs. Does it worj with Spark?
LCassa On Wed, Nov 4, 2015 at 3:02 AM, Adrian Tanase <atan...@adobe.com> wrote: > Another way to do it is to extract your filters as SQL code and load it in > a transform – which allows you to change the filters at runtime. > > Inside the transform you could apply the filters by goind RDD -> DF -> SQL > -> RDD. > > Lastly, depending on how complex your filters are, you could skip SQL and > create your own mini-DSL that runs inside transform. I’d definitely start > here if the filter predicates are simple enough… > > -adrian > > From: Stefano Baghino > Date: Wednesday, November 4, 2015 at 10:15 AM > To: Cassa L > Cc: user > Subject: Re: Rule Engine for Spark > > Hi LCassa, > unfortunately I don't have actual experience on this matter, however for a > similar use case I have briefly evaluated Decision > <https://github.com/Stratio/Decision> (then called literally Streaming > CEP Engine) and it looked interesting. I hope it may help. > > On Wed, Nov 4, 2015 at 1:42 AM, Cassa L <lcas...@gmail.com> wrote: > >> Hi, >> Has anyone used rule engine with spark streaming? I have a case where >> data is streaming from Kafka and I need to apply some rules on it (instead >> of hard coding in a code). >> >> Thanks, >> LCassa >> > > > > -- > BR, > Stefano Baghino > > Software Engineer @ Radicalbit >