Re: Anything like this been implemented yet?
More specifically: I know you can do this functionality by chaining other processors or using a bunch of generateflowfiles. What I'm looking for is a one stop way of associating dozens of small queries to one processor and have it batch send them on its own with no backend dependency like a database. On Wed, Mar 28, 2018 at 5:02 PM, Mike Thomsen wrote: > I have a client that would benefit from being able to run certain queries > periodically. Like half a dozen or more. Is there any processor where you > can associate a bunch of strings (like JSON) and send them out individually > as flowfiles? > > Thanks, > > Mike >
Re: Anything like this been implemented yet?
Since it sounds like each query is a JSON document, can you create a JSON array of all your queries and put that as the Custom Text of a GenerateFlowFile processor? Then follow it with SplitJson to split the array into a query per flow file, assuming that is what you want. Could also use ExecuteScript and enter all the queries as user defined properties, then write a small script that loops of the properties and produces a flow file for each dynamic property, where the content is the value of the property which would be the query string. On Wed, Mar 28, 2018 at 5:03 PM, Mike Thomsen wrote: > More specifically: I know you can do this functionality by chaining other > processors or using a bunch of generateflowfiles. What I'm looking for is a > one stop way of associating dozens of small queries to one processor and > have it batch send them on its own with no backend dependency like a > database. > > On Wed, Mar 28, 2018 at 5:02 PM, Mike Thomsen > wrote: > >> I have a client that would benefit from being able to run certain queries >> periodically. Like half a dozen or more. Is there any processor where you >> can associate a bunch of strings (like JSON) and send them out individually >> as flowfiles? >> >> Thanks, >> >> Mike >>
Re: Anything like this been implemented yet?
I would also suggest if you want to set a static flow and not have to update data directly in NiFi, put the JSON strings in a file and use GetFile to read it in, SplitText to split it into individual lines (flowfiles), and process that way. Andy LoPresto alopre...@apache.org alopresto.apa...@gmail.com PGP Fingerprint: 70EC B3E5 98A6 5A3F D3C4 BACE 3C6E F65B 2F7D EF69 > On Mar 28, 2018, at 2:13 PM, Bryan Bende wrote: > > Since it sounds like each query is a JSON document, can you create a > JSON array of all your queries and put that as the Custom Text of a > GenerateFlowFile processor? > > Then follow it with SplitJson to split the array into a query per flow > file, assuming that is what you want. > > Could also use ExecuteScript and enter all the queries as user defined > properties, then write a small script that loops of the properties and > produces a flow file for each dynamic property, where the content is > the value of the property which would be the query string. > > On Wed, Mar 28, 2018 at 5:03 PM, Mike Thomsen wrote: >> More specifically: I know you can do this functionality by chaining other >> processors or using a bunch of generateflowfiles. What I'm looking for is a >> one stop way of associating dozens of small queries to one processor and >> have it batch send them on its own with no backend dependency like a >> database. >> >> On Wed, Mar 28, 2018 at 5:02 PM, Mike Thomsen >> wrote: >> >>> I have a client that would benefit from being able to run certain queries >>> periodically. Like half a dozen or more. Is there any processor where you >>> can associate a bunch of strings (like JSON) and send them out individually >>> as flowfiles? >>> >>> Thanks, >>> >>> Mike >>> signature.asc Description: Message signed with OpenPGP using GPGMail