Hi all,

I’m investigating an issue where the content disk of one of our nifi 1.12 
clusters is filling up while the active usage is very low (10-12 GB active data 
per node while the cluster view shows 98% full of 600 GB per node).
My first idea was that archiving wasn’t being run often enough or cleaning up 
enough data, but from what I read, I should see global backpressure being 
applied then.

For reference, these are the settings:
# Content Repository
nifi.content.claim.max.appendable.size=1 MB
nifi.content.repository.archive.max.retention.period=12 hours
nifi.content.repository.archive.max.usage.percentage=80%
nifi.content.repository.archive.enabled=true
nifi.content.repository.always.sync=false

Now my next suspect is a queue full of old flowfiles (50+ days in queue) in a 
disabled flow that keep content claims active amounting to much more diskspace 
than needed.
I know I can look up the Content Claim identifier in the content repository 
directory, but while I can see the size of the total file, I have no idea 
whether those are all active flowfiles or mostly archived.

Is there a more systematic way of inspecting active/inactive content in content 
claims?

Regards,

Isha

Reply via email to