We do the same thing at Twitter, we have a local agent that polls the the
metrics endpoint and sends them to our internal timeseries database.

Cheers,
David

On Tue, Aug 22, 2017 at 4:05 PM, Derek Slager <[email protected]> wrote:

> Our approach is similar. We poll /vars.json on an interval and send a
> selection of those metrics to Riemann. We configure alerts there, and also
> pass these metrics through to InfluxDB for historical reporting (mostly via
> Grafana dashboards). This has worked well for us.
>
> --
> Derek Slager
> CTO
> Amperity
>
> On Tue, Aug 22, 2017 at 3:23 PM, De, Bipra <[email protected]> wrote:
>
>> Hello Friends,
>>
>> Greetings!!
>>
>> We are currently using *Aurora 0.17.0* and have a use-case wherein we
>> want to continuously monitor the below SLA metrics for our clusters to
>> detect any anomalies :
>>
>>    - Median Time To Assigned (MTTA
>>    
>> <http://aurora.apache.org/documentation/latest/features/sla-metrics/#median-time-to-assigned-(mtta)>
>>    )
>>    - Median Time To Starting (MTTS
>>    
>> <http://aurora.apache.org/documentation/latest/features/sla-metrics/#median-time-to-starting-(mtts)>
>>    )
>>    - Median Time To Running (MTTR
>>    
>> <http://aurora.apache.org/documentation/latest/features/sla-metrics/#median-time-to-running-(mttr)>
>>    )
>>
>> Currently, the *sla_stat_refresh_interval* for us is set to default *1
>> min*.
>>
>> Now, while using the */vars* api endpoint to fetch the SLA metrics,
>> aurora samples the data for metrics calculation of the above metrics only
>> for the last one min at every 1 minute interval. It won’t give us the
>> historical data for these metrics.
>>
>> Does aurora expose any api endpoint to provide the historical data for
>> these metrics over some configurable period of time? Is there any metric in 
>> */graphview
>> *endpoint for this?
>>
>> Also, it will be great if anyone can suggest some ideas for monitoring
>> around these metrics. I am , at present,  planning to keep polling the
>> /vars endpoint regularly for data collection and use ELK stack for graphing
>> and alerting.
>>
>> Thanks for your time in advance !!
>>
>> Regards,
>>
>> Bipra.
>>
>
>

Reply via email to