Hi Sean,

Wow, that sounds like a good guess!

Unfortunately, when I list the tables in influxdb (via 'show
measurements'), I can't even find an entry for the StatsD metrics. I'll
explain: I assume that for a StatsD metric like the one below, I would find
a 'measurement' (i.e. table) with name
'statsd_swift-object-replicator_partition_update_timing', correct?

```
swift-object-replicator.partition.update.timing:6.03103637695|ms
```

Unfortunately I don't find any measurements with a 'statsd_*' prefix. I can
only find measurements for metrics I already know about and are being
correctly exported by telegraf.

Best,
Damião

On Thu, Aug 18, 2016 at 10:41 PM, Sean Beckett <[email protected]> wrote:

> It seems that StatsD is sending timestamps in milliseconds, but the
> Telegraf writes are at nanosecond precision ("precision=ns" in the query
> string.)
>
> I suspect your data is in InfluxDB, it's just tightly clustered a few
> seconds after midnight, Jan 1 1970, which is what a millisecond timestamp
> looks like when converted to nanoseconds since epoch.
>
> Can you query close to epoch 0 and see if you have results?
>
> On Tue, Aug 16, 2016 at 3:45 AM, Damião Rodrigues <[email protected]>
> wrote:
>
>> Hi Sean,
>>
>> Thanks for the answer. I'll answer your queries by parts.
>>
>> 1) telegraf can write to influxdb:
>>
>>    - Other metrics other than StatsD metrics are being written (e.g. I'm
>>    using the Docker and exec input plugins)
>>    - I get the following influxdb log entries at regular intervals (the
>>    IP address of the telegraf server is indeed 10.42.75.237):
>>
>> ```
>>
>> [httpd] 2016/08/16 09:01:11 10.42.75.237 - influxdb [16/Aug/2016:09:01:11
>> +0000] POST /write?consistency=&db=metrics&precision=ns&rp=default
>> HTTP/1.1 204 0 - InfluxDBClient fa9967c1-638f-11e6-98c8-000000000000
>> 100.970492ms
>> [httpd] 2016/08/16 09:01:16 10.42.75.237 - influxdb [16/Aug/2016:09:01:16
>> +0000] POST /write?consistency=&db=metrics&precision=ns&rp=default
>> HTTP/1.1 204 0 - InfluxDBClient fd928252-638f-11e6-98ce-000000000000
>> 104.447255ms
>> [httpd] 2016/08/16 09:01:17 10.42.75.237 - influxdb [16/Aug/2016:09:01:17
>> +0000] POST /write?consistency=&db=metrics&precision=ns&rp=default
>> HTTP/1.1 204 0 - InfluxDBClient fdcff4c3-638f-11e6-98cf-000000000000
>> 60.178059ms
>>
>> ```
>>
>> 2) Output configuration
>>
>> ```
>>
>> [[outputs.influxdb]]
>>   urls = ["http://influxdb:8086";]
>>   database = "metrics" # required
>>   retention_policy = "default"
>>   username = <my-user>
>>   password = <my-pass>
>>
>> ```
>>
>> 3) Running SHOW RETENTION POLICIES ON 'metrics'
>>
>> ```
>>
>> > SHOW RETENTION POLICIES ON metrics
>> name    duration        shardGroupDuration      replicaN        default
>> default 0               168h0m0s                1               true
>>
>> ```
>>
>> Best,
>>
>> Damiao
>>
>> On Thu, Aug 11, 2016 at 10:21 PM, Sean Beckett <[email protected]> wrote:
>>
>>> What do the InfluxDB logs show? Are there writes coming from the
>>> Telegraf server? The Telegraf logs indicate it is writing to InfluxDB.
>>>
>>> Can you include the output configuration
>>> <https://docs.influxdata.com/telegraf/v0.13/administration/configuration/#output-configuration>?
>>> Also the results of running "SHOW RETENTION POLICIES ON <db>" against
>>> InfluxDB, where <db> is replaced by the destination database.
>>>
>>> On Mon, Aug 8, 2016 at 7:27 AM, <[email protected]> wrote:
>>>
>>>> Hi all,
>>>>
>>>> I followed this tutorial (https://influxdata.com/blog/g
>>>> etting-started-with-sending-statsd-metrics-to-telegraf-influxdb/) to
>>>> get a statsd > telegraf > influxdb setup working. However, I've noticed
>>>> that while the statsd metrics reach telegraf, they are not relayed to
>>>> influxdb.
>>>>
>>>> Do you have an idea of what might be the problem? Find more information
>>>> below.
>>>>
>>>> 1) telegraf version 0.13.1 / influxdb 0.13.0
>>>>
>>>> 2) telegraf seems to get some statsd metrics, even though the
>>>> processing time seems suspiciously short:
>>>>
>>>> ```
>>>> [docker] gathered metrics, (5s interval) in 2.270456355s
>>>> 2016/08/08 13:16:18 Output [influxdb] buffer fullness: 900 / 20000
>>>> metrics. Total gathered metrics: 261096. Total dropped metrics: 0.
>>>> 2016/08/08 13:16:18 Output [influxdb] wrote batch of 900 metrics in
>>>> 104.723614ms
>>>>
>>>> 2016/08/08 13:16:20 Input [statsd] gathered metrics, (5s interval) in
>>>> 91.964µs
>>>>
>>>> 2016/08/08 13:16:20 Input [memcached] gathered metrics, (5s interval)
>>>> in 3.099963ms
>>>> ```
>>>>
>>>> 3) here's the relevant configuration for telegraf.conf:
>>>>
>>>> ```
>>>>
>>>> [[inputs.statsd]]
>>>>   ## Address and port to host UDP listener on
>>>>   service_address = ":8125"
>>>>   ## Delete gauges every interval (default=false)
>>>>   delete_gauges = false
>>>>   ## Delete counters every interval (default=false)
>>>>   delete_counters = false
>>>>   ## Delete sets every interval (default=false)
>>>>   delete_sets = false
>>>>   ## Delete timings & histograms every interval (default=true)
>>>>   delete_timings = true
>>>>   ## Percentiles to calculate for timing & histogram stats
>>>>   percentiles = [90]
>>>>
>>>>   ## separator to use between elements of a statsd metric
>>>>   metric_separator = "_"
>>>>
>>>>   ## Parses tags in the datadog statsd format
>>>>   ## http://docs.datadoghq.com/guides/dogstatsd/
>>>>   parse_data_dog_tags = false
>>>>
>>>>   ## Statsd data translation templates, more info can be read here:
>>>>   ## https://github.com/influxdata/telegraf/blob/master/docs/DATA
>>>> _FORMATS_INPUT.md#graphite
>>>>   # templates = [
>>>>   #     "cpu.* measurement*"
>>>>   # ]
>>>>   ## Number of UDP messages allowed to queue up, once filled,
>>>>   ## the statsd server will start dropping packets
>>>>   allowed_pending_messages = 10000
>>>>
>>>>   ## Number of timing/histogram values to track per-measurement in the
>>>>   ## calculation of percentiles. Raising this limit increases the
>>>> accuracy
>>>>   ## of percentiles but also increases the memory usage and cpu time.
>>>>   percentile_limit = 1000
>>>>
>>>> ```
>>>>
>>>> 4) here's an example of the statsd metrics being received at telegraf's
>>>> host (obtained via `tcpdump`):
>>>>
>>>> ```
>>>> 13:20:49.431549 IP (tos 0x0, ttl 64, id 2976, offset 0, flags [DF],
>>>> proto UDP (17), length 92)
>>>>     big43.local.60178 > macmini6.local.8125: [udp sum ok] UDP, length 64
>>>>         0x0000:  4500 005c 0ba0 4000 4011 9d51 c0a8 082b
>>>> E..\..@[email protected]...+
>>>>         0x0010:  c0a8 0824 eb12 1fbd 0048 d989 7377 6966
>>>> ...$.....H..swif
>>>>         0x0020:  742d 6f62 6a65 6374 2d72 6570 6c69 6361
>>>> t-object-replica
>>>>         0x0030:  746f 722e 7061 7274 6974 696f 6e2e 7570
>>>> tor.partition.up
>>>>         0x0040:  6461 7465 2e74 696d 696e 673a 362e 3033
>>>> date.timing:6.03
>>>>         0x0050:  3130 3336 3337 3639 357c 6d73            103637695|ms
>>>> ```
>>>>
>>>> note the 
>>>> `swiftt-object-replicator.partition.update.timing:6.04103637695|ms`.
>>>> this is therefore of the statsd 'timing' type, which is supposedly
>>>> supported by telegraf.
>>>>
>>>> 5) i've noticed that the measurements don't show up in influxdb. in
>>>> fact, i don't even spot any outgoing messages to influxdb exporting the
>>>> statsd metrics!
>>>>
>>>> Thanks!
>>>>
>>>> Damião
>>>>
>>>> --
>>>> Remember to include the InfluxDB version number with all issue reports
>>>> ---
>>>> You received this message because you are subscribed to the Google
>>>> Groups "InfluxDB" group.
>>>> To unsubscribe from this group and stop receiving emails from it, send
>>>> an email to [email protected].
>>>> To post to this group, send email to [email protected].
>>>> Visit this group at https://groups.google.com/group/influxdb.
>>>> To view this discussion on the web visit https://groups.google.com/d/ms
>>>> gid/influxdb/6ba76909-d4a5-4a58-b0c5-55e33f1462f9%40googlegroups.com.
>>>> For more options, visit https://groups.google.com/d/optout.
>>>>
>>>
>>>
>>>
>>> --
>>> Sean Beckett
>>> Director of Support and Professional Services
>>> InfluxDB
>>>
>>> --
>>> Remember to include the InfluxDB version number with all issue reports
>>> ---
>>> You received this message because you are subscribed to a topic in the
>>> Google Groups "InfluxDB" group.
>>> To unsubscribe from this topic, visit https://groups.google.com/d/to
>>> pic/influxdb/4gYkiSu4PUk/unsubscribe.
>>> To unsubscribe from this group and all its topics, send an email to
>>> [email protected].
>>> To post to this group, send email to [email protected].
>>> Visit this group at https://groups.google.com/group/influxdb.
>>> To view this discussion on the web visit https://groups.google.com/d/ms
>>> gid/influxdb/CALGqCvOJ9cjd_UkeaY0J2YO07ekg5kQ68QehKnjnK2MBQ0
>>> vn0A%40mail.gmail.com
>>> <https://groups.google.com/d/msgid/influxdb/CALGqCvOJ9cjd_UkeaY0J2YO07ekg5kQ68QehKnjnK2MBQ0vn0A%40mail.gmail.com?utm_medium=email&utm_source=footer>
>>> .
>>>
>>> For more options, visit https://groups.google.com/d/optout.
>>>
>>
>>
>
>
> --
> Sean Beckett
> Director of Support and Professional Services
> InfluxDB
>

-- 
Remember to include the InfluxDB version number with all issue reports
--- 
You received this message because you are subscribed to the Google Groups 
"InfluxDB" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at https://groups.google.com/group/influxdb.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/influxdb/CAF8xhAvxrdzCaaRknkd8O2Vck_WYOb_PwqvpmfefpVE-5p%2BbUw%40mail.gmail.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to