Thanks for getting back to me Mark. Unfortunately it is running on an intranet so I can’t get logs and so forth off the system. Is there anything in particular I can look out for?
I am running zookeeper as a separate service (not embedded) on three nodes. The nifi cluster currently has two nodes. Cheers, Phil On Thu, 3 Jun 2021 at 10:17, Mark Payne <[email protected]> wrote: > Hey Phil, > > Can you grab a diagnostics dump from one of the nodes (preferably the > cluster coordinator)? Ideally grab 3 of them, with about 5 mins in between. > > To do that, run: > > bin/nifi.sh diagnostics <filename> > > So run something like: > > bin/nifi.sh diagnostics diagnostics1.txt > <wait 3-5 mins> > bin/nifi.sh diagnostics diagnostics2.txt > <wait 3-5 mins> > bin/nifi.sh diagnostics diagnostics3.txt > > And then upload those diagnostics text files? > They should not contain any sensitive information, aside from maybe file > paths (which most don’t consider sensitive but you may). But recommend you > glance through it to make sure that you leave out any sensitive information. > > Those dumps should help in understanding the problem, or at least zeroing > in on it. > > Also, is NiFi using its own dedicated zookeeper or is it shared with other > services? How many nodes is the zookeeper? > > > > > On Jun 2, 2021, at 7:54 PM, Phil H <[email protected]> wrote: > > > > Hi there, > > > > I am getting a lot of these both in the web interface to my servers, and > in > > the cluster communication between the nodes. All other aspects of the > > servers are fine. TCP connections to NiFi, as well as SSH connections to > > the servers are stable (running for days at a time). I’m lucky if I go 5 > > minutes without the web UI dropping out or a cluster re-election due to a > > heartbeat being missed. > > > > Running 1.13.2, recently upgraded from 1.9.2. I was getting the same > issues > > with the old version, but they seem to be much more frequent now. > > > > Help! > >
