Thank you so much for your guidance.  I was able to get data flowing into
Prometheus fairly easily once all the pieces were understood.

Now, I just need to dig into Prometheus queries and make some Grafana
dashboards.

On Tue, Mar 3, 2020 at 2:54 PM Yolanda Davis <[email protected]>
wrote:

> Sure not a problem!  Hopefully below thoughts can help you get started:
>
> As you may know the PrometheusReportingTask is a bit different from other
> tasks in that it actually exposes an endpoint for Prometheus to scrape (vs.
> pushing data directly to Prometheus).  When the task is started the
> endpoint is created on the port you designate under “/metrics”; so just
> ensure that you don’t have anything already on the port you select. If you
> want to ensure that you have a secured endpoint for Prometheus to connect,
> be sure to use a SSL Context Service (a controller service that will allow
> the reporting task to use the appropriate key/trust stores for TLS). Also
> you'll want to consider the levels at which you are reporting (Root Group,
> Process Group or All Components), especially in terms of the amount of data
> you are looking to send back.  Jvm metrics can be sent as well flow
> specific metrics. Finally consider how often metrics should be refreshed by
> adjusting the Scheduling Strategy in the settings tab for the task.
>
> When starting the task you should be able to go directly to the endpoint
> (without Prometheus) to confirm it’s output (e.g.
> http://locahost:9092/metrics ).  You should see a format similar to what
> Prometheus supports for it’s scraping jobs (see example
> https://prometheus.io/docs/instrumenting/exposition_formats/#text-format-example
> )
>
> On the Prometheus side you’ll want to follow their instructions on how to
> setup a scrape configuration that  will point to the newly created metrics
> endpoint . I’d recommend checking out the first steps for help (
> https://prometheus.io/docs/introduction/first_steps/#configuring-prometheus)
> and then when you need to provide more advanced settings take a look here
> https://prometheus.io/docs/prometheus/latest/configuration/configuration/#scrape_config.
> The key is you’ll want to define a new scrape job that looks at the NiFi
> endpoint for scraping.  To start you may want to refer to the cluster
> directly but later add the security credentials or use another method for
> discovering the endpoint.
>
> Once these configurations are in place, and Prometheus is started (or
> restarted) after a few seconds you should begin to see metrics landing when
> querying in Grafana.
>
> I hope this helps!  Please let me know if you have any further questions.
>
> -yolanda
>
> On Tue, Mar 3, 2020 at 2:10 PM Eric Ladner <[email protected]> wrote:
>
>> Yes, exactly!   Reporting Task -> Prometheus -> Grafana for keeping an
>> eye on things running in NiFi.
>>
>> If you have any hints/tips on getting things working, I'd be grateful.
>>
>> On Tue, Mar 3, 2020 at 12:35 PM Yolanda Davis <[email protected]>
>> wrote:
>>
>>> Hi Eric,
>>>
>>> Were you looking to use the Prometheus Reporting Task for making metrics
>>> available for Prometheus scraping? I don't believe any documentation
>>> outside of what is in NiFi exists just yet, but I'm happy to help answer
>>> questions you may have (I've used this task recently).
>>>
>>> -yolanda
>>>
>>> On Tue, Mar 3, 2020 at 10:51 AM Eric Ladner <[email protected]>
>>> wrote:
>>>
>>>> Is there a guide to setting up Nifi and Prometheus anywhere?  The nar
>>>> docs are a little vague.
>>>>
>>>> Thanks,
>>>>
>>>> Eric Ladner
>>>>
>>>
>>>
>>> --
>>> --
>>> [email protected]
>>> @YolandaMDavis
>>>
>>>
>>
>> --
>> Eric Ladner
>>
>
>
> --
> --
> [email protected]
> @YolandaMDavis
>
>

-- 
Eric Ladner

Reply via email to