Re: PutKafka use with large quantity of data?

2019-04-04 Thread Bryan Bende
Each queue has back-pressure settings on it which default to 10k flow files or 1GB size of flow files. When one of these thresholds is exceeded, the preceeding processor will not execute until the queue goes back below the threshold. Most likely if GenerateFlowFile has a Run Schedule of 0

Re: PutKafka use with large quantity of data?

2019-04-04 Thread l vic
Actually, it's not Kafka topic but Nifi queue between "generateFlow" and "PutKafka" gets overflown On Thu, Apr 4, 2019 at 10:58 AM Joe Witt wrote: > Can you share screenshots, logs, and a more detailed description of what > you're doing, observing with nifi and the system and what you expect it

Re: PutKafka use with large quantity of data?

2019-04-04 Thread Bryan Bende
We need to define what "NiFi stops responding" means... Are there tons of flow files queued up before before PublishKafka? Are there back-pressure indicators on any of the queues? Do the kafka related processors show active threads in the top right corners of the processors? Does NiFI crash? On

Re: PutKafka use with large quantity of data?

2019-04-04 Thread Andrew Grande
What's the concurrency for these processors? What's a global NiFi thread pool size? I wonder if you might be running out of available threads while they are waiting for external system i/o under load. Andrew On Thu, Apr 4, 2019, 8:24 AM l vic wrote: > What's this particular processing group

Re: PutKafka use with large quantity of data?

2019-04-04 Thread l vic
What's this particular processing group does: writes large dataset to Kafka topic, one consumer reads from topic and saves data to Hbase/PQS table, another consumer writes to ES index On Thu, Apr 4, 2019 at 10:58 AM Joe Witt wrote: > Can you share screenshots, logs, and a more detailed

Re: PutKafka use with large quantity of data?

2019-04-04 Thread Joe Witt
Can you share screenshots, logs, and a more detailed description of what you're doing, observing with nifi and the system and what you expect it to be doing. Thanks On Thu, Apr 4, 2019 at 10:56 AM l vic wrote: > No, actually what happens is - NiFi stops responding ( if I use it without > rate

Re: PutKafka use with large quantity of data?

2019-04-04 Thread l vic
No, actually what happens is - NiFi stops responding ( if I use it without rate contol) On Thu, Apr 4, 2019 at 10:42 AM Joe Witt wrote: > Hello > > There isn't really a feedback mechanism based on load on the Kafka topic. > When you say overrunning the topic do you mean that you don't want

Re: PutKafka use with large quantity of data?

2019-04-04 Thread Joe Witt
Hello There isn't really a feedback mechanism based on load on the Kafka topic. When you say overrunning the topic do you mean that you don't want there to be a large lag between consumers and their current offset and if that grows you want NiFi to slow down? I dont believe there is anything

PutKafka use with large quantity of data?

2019-04-04 Thread l vic
I have to ingest large (200,000 messages) data set into Kafka topic as quickly as possible without overrunning topic... Right now I just use rate limiter to do it but can be there some better "adaptive" way to do it? Thank you... -V

Re: Reusing same flow for different database connections

2019-04-04 Thread Bryan Bende
Hello, Take a look at the DBCP lookup service, it allows you to register one or more connection pool services and then select one at runtime based on an incoming flow file having an attribute called database.name. Thanks, Bryan On Thu, Apr 4, 2019 at 8:47 AM Max wrote: > Hello! > > We are

Reusing same flow for different database connections

2019-04-04 Thread Max
Hello! We are working on a project that requires importing data from tables across different database servers (as in, different db connection pools.) The data flow itself is the same across maybe 40-50 tables and around 10 connections. I tried to create an abstract flow that can be parameterized

GetHbase state

2019-04-04 Thread Dwane Hall
Hey fellow NiFi fans, I was recently loading data into into Solr via HBase (around 700G ~60,000,000 db rows) using NiFi and noticed some inconsistent behaviour with the GetHbase processor and I'm wondering if anyone else has noticed similar behaviour when using it. Here's our environment and

Re: NiFi Registry Not Auditing Denied Errors

2019-04-04 Thread Shawn Weeks
It looks like it will do this if you don’t grant the host access to /buckets which is a valid resource. Sent from my iPhone > On Apr 4, 2019, at 1:45 AM, Koji Kawamura wrote: > > Hi Shawn, > > The 'No applicable policies could be found.' message can be logged > when a request is made

Re: NiFi Registry Not Auditing Denied Errors

2019-04-04 Thread Koji Kawamura
Hi Shawn, The 'No applicable policies could be found.' message can be logged when a request is made against a resource which doesn't exist.