Okay, found the offload, but the data is still stuck on the “offloaded” node, in a “single node” queue (I am bringing the data to a single node to deduplicate multiple parallel inputs).
If I refresh the UI, I can see the missing items numbered in the queue, but can’t open the queue because the other node is “currently offloaded”. I’m sure I’m just missing something here?? On Thu, 2 Sep 2021 at 08:20, Shawn Weeks <[email protected]> wrote: > On newer versions there is an option in the UI to Offload the data if you > have NiFi's cluster load balancing setup. Then you'd disconnect the node > and shut it down. > > Thanks > Shawn > > -----Original Message----- > From: Phil H <[email protected]> > Sent: Wednesday, September 1, 2021 4:36 PM > To: [email protected] > Subject: Primary node vs shutdown > > Hi there, > > I am noticing a number of situations where shutting down one node in a > cluster is leaving data stranded in the flows on that shut down server. > > Is there any way to tell NiFi to ship data off to other cluster members > before it shuts down? Note I am restarting via the nifi.sh script, not > just killing the process/host with no notice > > Thanks, > Phil >
