Question about NiFi and bulk database inserts - is PutSQL the only out of the box option?

2018-08-09 Thread Kuhfahl, Bob
I’m trying to get bulk inserts going using PutSQL processor but it’s starting to get ugly so I need to reach out and see if any of you have been down this path. If you have, here’s some info. If not, thanks for reading this far ☺ Background: Legacy database migration ETL task. Extract from on

Need a sample JSON input file for InferAvroSchema

2018-08-13 Thread Kuhfahl, Bob
Trying to develop a sample input file of json data to feed into InferAvroSchema so I can feed that into PutDatabaseRecord. Need a hello world example ☺ But, to get started, I’d be happy to get InferAvroSchema working. I’m “trial and error”-ing the input file hoping to get lucky, but.. No log m

Re: Need a sample JSON input file for InferAvroSchema -> PutDatabaseRecord

2018-08-14 Thread Kuhfahl, Bob
; Subject: Re: Need a sample JSON input file for InferAvroSchema Bob, InferAvroSchema can infer types like boolean, integer, long, float, double, and I believe for JSON can correctly descend into arrays and nested maps/structs/objects. Here is an example record from NiFi provenance data that has mo

ReplaceText regex question

2018-08-16 Thread Kuhfahl, Bob
I have to massage some JSON to pass into PutDatabaseRecord so I can insert the data into postgresql. Postgres wants the identifers to have double quotes and it seems the values need to have single quotes (for the strings). Sample JSON input: "producer_userid": "fdolomite", Search Value: (?<

OPS : ReplaceText regex question

2018-08-16 Thread Kuhfahl, Bob
Failed to notice the JSON is “minified” (compressed so it has no spaces) in the flow… From: Bob Kuhfahl Reply-To: "users@nifi.apache.org" Date: Thursday, August 16, 2018 at 9:44 AM To: "users@nifi.apache.org" Subject: ReplaceText regex question I have to massage some JSON to pass into PutDat

Design pattern advice needed

2018-08-17 Thread Kuhfahl, Bob
Problem: * Source database with over 200 tables. * Current Nifi ‘system’ we are developing can extract data from those 200 tables into NiFi flows of JSON-formatted data, essentially separate flows for each table with an attribute that indicates the tablename and other useful attributes