Hi,
I studied the nifi website a bit and if I missed a key part, forgive
me for asking this question.
But I am wondering if or how nifi can accommodate processing large data
sets with possibly compute intensive operations.
For example, if we have say 2 million documents, how does nifi make
processing these documents efficient?
I understand the visual workflow and its nice. How is that parallelized
across a data set?
Do we submit all the documents to a cluster of flows (how many?) that
execute some number of documents simultaneously?
Does nifi support batch processing? Is it elastic?
Thanks.