Hi Pierre

Atlas indeed comes into my attention in the beginning, but my production
environment does not support hdfs which is the fundamental of Atlas.

The provenance is a out of box feature in NiFi, and we also use NiFi to do
simple transformations, therefore I am searching some provenance solution
that can be leveraged in NiFi, that would be perfect.

Cheers

Pierre Villard <[email protected]>于2021年1月11日 周一下午6:42写道:

> Hi,
>
> Just want to also mention that the integration that NiFi provides with
> Apache Atlas (through a dedicated reporting task) is supposed to provide
> this kind of end-to-end lineage across multiple components.
>
> Thanks,
> Pierre
>
> Le lun. 11 janv. 2021 à 18:58, Chris Sampson <[email protected]>
> a écrit :
>
>> Might be worth taking a look at the "alternate.identifier" (see
>> https://nifi.apache.org/docs/nifi-docs/html/nifi-in-depth.html#associating-disparate-data
>> ).
>>
>>
>> ---
>> *Chris Sampson*
>> IT Consultant
>> [email protected]
>> <https://www.naimuri.com/>
>>
>>
>> On Mon, 11 Jan 2021 at 14:53, Yi Wang <[email protected]> wrote:
>>
>>> Hi Nifi team and experts,
>>>
>>> In such an example scenario:
>>>
>>> Data —> Nifi —> Kafka —> external system (spark, flink etc) —> Kafka —>
>>> Nifi —> s3
>>>
>>> How do I fill in the ‘gap’ of data lineage?
>>> Once the data leave Nifi, its provenance life is ended (as far as I
>>> know), even when the same data is sent back to Nifi later, Nifi treats them
>>> as different data. So how can I handle this in order to get a complete data
>>> lineage graph?
>>>
>>> Any idea or suggestions? Thanks in advance.
>>>
>>>
>>> Cheers
>>>
>>>

Reply via email to