Hi,
I studied the nifi website a bit and if I missed a key part, forgive me for asking this question. But I am wondering if or how nifi can accommodate processing large data sets with possibly compute intensive operations. For example, if we have say 2 million documents, how does nifi make processing these documents efficient? I understand the visual workflow and its nice. How is that parallelized across a data set?

Do we submit all the documents to a cluster of flows (how many?) that execute some number of documents simultaneously?
Does nifi support batch processing? Is it elastic?

Thanks.

Reply via email to