Yea ListenTCP also doesn't handle the back-pressure with the client
the way it really should.

Regarding the load balancing, I believe traditional s2s does factor in
the load of each node when deciding how to load balance, but I don't
know if this is part of load balanced connections or not. Mark P would
know for sure.

On Wed, Mar 6, 2019 at 4:47 PM James Srinivasan
<james.sriniva...@gmail.com> wrote:
>
> Yup, but because of the unfortunate way the source (outside NiFi)
> works, it doesn't buffer for long when the connection doesn't pull or
> drops. It behaves far more like a 5 Mbps UDP stream really :-(
>
> On Wed, 6 Mar 2019 at 21:44, Bryan Bende <bbe...@gmail.com> wrote:
> >
> > James, just curious, what was your source processor in this case? ListenTCP?
> >
> > On Wed, Mar 6, 2019 at 4:26 PM Jon Logan <jmlo...@buffalo.edu> wrote:
> > >
> > > What really would resolve some of these issues is backpressure on CPU -- 
> > > ie. let Nifi throttle itself down to not choke the machine until it dies 
> > > if constrained on CPU. Easier said than done unfortunately.
> > >
> > > On Wed, Mar 6, 2019 at 4:23 PM James Srinivasan 
> > > <james.sriniva...@gmail.com> wrote:
> > >>
> > >> In our case, backpressure applied all the way up to the TCP network
> > >> source which meant we lost data. AIUI, current load balancing is round
> > >> robin (and two other options prob not relevant). Would actual load
> > >> balancing (e.g. send to node with lowest OS load, or number of active
> > >> threads) be a reasonable request?
> > >>
> > >> On Wed, 6 Mar 2019 at 20:51, Joe Witt <joe.w...@gmail.com> wrote:
> > >> >
> > >> > This is generally workable (heterogenous node capabilities) in NiFi 
> > >> > clustering.  But you do want to leverage back-pressure and load 
> > >> > balanced connections so that faster nodes will have an opportunity to 
> > >> > take on the workload for slower nodes.
> > >> >
> > >> > Thanks
> > >> >
> > >> > On Wed, Mar 6, 2019 at 3:48 PM James Srinivasan 
> > >> > <james.sriniva...@gmail.com> wrote:
> > >> >>
> > >> >> Yes, we hit this with the new load balanced queues (which, to be 
> > >> >> fair, we also had with remote process groups previously). Two "old" 
> > >> >> nodes got saturated and their queues filled while three "new" nodes 
> > >> >> were fine.
> > >> >>
> > >> >> My "solution" was to move everything to new hardware which we had 
> > >> >> inbound anyway.
> > >> >>
> > >> >> On Wed, 6 Mar 2019, 20:40 Jon Logan, <jmlo...@buffalo.edu> wrote:
> > >> >>>
> > >> >>> You may run into issues with different processing power, as some 
> > >> >>> machines may be overwhelmed in order to saturate other machines.
> > >> >>>
> > >> >>> On Wed, Mar 6, 2019 at 3:34 PM Mark Payne <marka...@hotmail.com> 
> > >> >>> wrote:
> > >> >>>>
> > >> >>>> Chad,
> > >> >>>>
> > >> >>>> This should not be a problem, given that all nodes have enough 
> > >> >>>> storage available to handle the influx of data.
> > >> >>>>
> > >> >>>> Thanks
> > >> >>>> -Mark
> > >> >>>>
> > >> >>>>
> > >> >>>> > On Mar 6, 2019, at 1:44 PM, Chad Woodhead 
> > >> >>>> > <chadwoodh...@gmail.com> wrote:
> > >> >>>> >
> > >> >>>> > Are there any negative effects of having filesystem mounts 
> > >> >>>> > (dedicated mounts for each repo) used by the different NiFi 
> > >> >>>> > repositories differ in size on NiFi nodes within the same 
> > >> >>>> > cluster? For instance, if some nodes have a content_repo mount of 
> > >> >>>> > 130 GB and other nodes have a content_repo mount of 125 GB, could 
> > >> >>>> > that cause any problems or cause one node to be used more since 
> > >> >>>> > it has more space? What about if the difference was larger, by 
> > >> >>>> > say a 100 GB difference?
> > >> >>>> >
> > >> >>>> > Trying to repurpose old nodes and add them as NiFi nodes, but 
> > >> >>>> > their mount sizes are different than my current cluster’s nodes 
> > >> >>>> > and I’ve noticed I can’t set the max size limit to use of a 
> > >> >>>> > particular mount for a repo.
> > >> >>>> >
> > >> >>>> > -Chad
> > >> >>>>

Reply via email to