[ 
https://issues.apache.org/jira/browse/NIFI-6202?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17188966#comment-17188966
 ] 

Nadeem commented on NIFI-6202:
------------------------------

Are you using MergeContent/MergeRecord Processor in your DataFlow? This could 
possibly happen when MergeX processors are waiting for more flowfiles based on 
Defragment strategy. Flowfile repository gets huge due to antipattern store of 
data payloads in flowfile attributes which makes NiFi difficult to serialize 
off memory when the rate of flowfiles arriving gets higher. Eventually, leading 
to canvas not accessible as the system run OOM quickly. I have myself seen 
flowfile repository getting 100+ GB in our customer dataflow. You should 
probably fix filthy dataflow. That should pretty much solve the problem. 

> NiFI flow states one file struck in process and not getting cleared 
> --------------------------------------------------------------------
>
>                 Key: NIFI-6202
>                 URL: https://issues.apache.org/jira/browse/NIFI-6202
>             Project: Apache NiFi
>          Issue Type: Bug
>          Components: Core UI
>    Affects Versions: 1.8.0
>            Reporter: sayee
>            Priority: Blocker
>         Attachments: defect image.PNG
>
>
> Processing large files through NiFi:
> 1. Running a large number of data files through NiFi processor causes 1 file 
> struck in Q and not getting cleared.
> 2. only way to clean is delete flow files from repository
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to