What about on the /query endpoint. I tried something similar, but it didn't
seem to work.My measurement names, and retention policies, are just about
identically named, but with the retention policy having "RP_" in the front
of it, and the measurement having "Metric_" in front of it.
This was my URL. It gave me an empty result ({"results":[{}]}) , but have a
200 response code.
http://localhost:8086/query?db=jefferyk&epoch=ms&rp=RP_jefferyk_Windows+System&q=show+field+keys+FROM+%22Metric_jefferyk_Windows+System%22
In the database, I can see it's selecting from the autogen retention
policy, despite me providing the rp in the query end point.
database console output:
[query] 2016/12/22 10:27:00 SELECT fieldKey, fieldType FROM
jefferyk.autogen._fieldKeys WHERE _name = 'Metric_NEWPETRONAS_Windows
System'
[httpd] 127.0.0.1 - - [22/Dec/2016:10:27:00 -0500] "GET
/query?db=jefferyk&epoch=ms&q=show+field+keys+FROM+%22Metric_NEWPETRONAS_Windows+System%22&rp=RP_NEWPETRONAS_Windows+System
HTTP/1.1" 200 17 "-" "Java/1.8.0_73" 14f22859-c85b-11e6-9276-000000000000
1001
On Wednesday, December 21, 2016 at 1:59:00 PM UTC-5, Mark Rushakoff wrote:
>
> Specifying a retention policy on the /write endpoint requires the rp query
> parameter. The space between the RP and measurement name only works in the
> CLI as a convenience.
>
> It's difficult to precisely quantify these performance differences.
> Ultimately, the best answer will come from profiling your data on your
> operating system and your hardware.
>
> That being said, there's small CPU and memory overhead for any database.
> We have users with hundreds of databases in production, but in general,
> prefer the smallest number of databases that fit your needs.
>
> You're more likely to run into disk contention with a larger number of
> databases because there are more WAL and TSM files being managed. TSM files
> in a single database and retention policy can be compacted together to
> minimize disk reads; if they're spread out, you lose some compaction
> benefits. Likewise, you'll be appending to a single WAL file per database.
> You can batch writes to multiple measurements in a single request to
> /write, but if your measurements are in separate databases and retention
> policies, they must necessarily be multiple requests to /write.
>
> On Wed, Dec 21, 2016 at 10:11 AM, Jeffery K <
> [email protected] <javascript:>> wrote:
>
>> I think i found the issue.
>> I am using a http client and in the line protocol, specifying the insert
>> as "RPNAME"."measurementName", but i actually need a space between the RP
>> and measurement name. I thought that since the query required
>> RP.measurement, then the insert was the same. My mistake.
>> Do you know what the difference in the performance profile would be
>> splitting out these measurements into more shards like this?
>>
>>
>> On Wednesday, December 21, 2016 at 12:02:42 PM UTC-5, Mark Rushakoff
>> wrote:
>>>
>>> It looks like you probably are writing without providing the rp query
>>> parameter [1] and so writes are going into the default RP (that is, the one
>>> with default=true, which also happens to be named "default").
>>>
>>> TSM and WAL files are stored under
>>> $INFLUXDB_DIR/{data,wal}/<database_name>/<rp_name>. If you're intending to
>>> use a unique retention policy per measurement, you're raising the disk
>>> usage and shard management overhead as compared to multiple measurements in
>>> a single retention policy and database.
>>>
>>> You are correct that the replication factor has no effect in open source
>>> InfluxDB.
>>>
>>> [1] https://docs.influxdata.com/influxdb/v1.1/tools/api/#write
>>>
>>> On Tuesday, December 20, 2016 at 10:05:13 PM UTC-8, Jeffery K wrote:
>>>>
>>>> In looking at other diagnostics in the github issue list, here is the
>>>> output of a show shards. I'm confused why they all have the "default" RP.
>>>> I
>>>> have a retention policy for every measurement in my database.
>>>>
>>>>
>>>> name: SL
>>>> id database retention_policy shard_group
>>>> start_time end_time expiry_time
>>>> owners
>>>> -- -------- ---------------- -----------
>>>> ---------- -------- -----------
>>>> ------
>>>> 110 SL default 110
>>>> 2015-10-26T00:00:00Z 2015-11-02T00:00:00Z 2015-11-02T00:00:00Z
>>>> 86 SL default 86
>>>> 2016-11-14T00:00:00Z 2016-11-21T00:00:00Z 2016-11-21T00:00:00Z
>>>> 88 SL default 88
>>>> 2016-11-21T00:00:00Z 2016-11-28T00:00:00Z 2016-11-28T00:00:00Z
>>>> 95 SL default 95
>>>> 2016-11-28T00:00:00Z 2016-12-05T00:00:00Z 2016-12-05T00:00:00Z
>>>> 104 SL default 104
>>>> 2016-12-05T00:00:00Z 2016-12-12T00:00:00Z 2016-12-12T00:00:00Z
>>>> 107 SL default 107
>>>> 2016-12-12T00:00:00Z 2016-12-19T00:00:00Z 2016-12-19T00:00:00Z
>>>> 115 SL default 115
>>>> 2016-12-19T00:00:00Z 2016-12-26T00:00:00Z 2016-12-26T00:00:00Z
>>>>
>>>>
>>>>
>>>> On Tuesday, December 20, 2016 at 6:35:35 PM UTC-5, Jeffery K wrote:
>>>>>
>>>>> I'm having an issue with the influx 1.1 release. I just implemented
>>>>> retention policies, but they don't seen to be deleting the data. I
>>>>> believe
>>>>> I understand that they won't be deleted up until the value of the Shard
>>>>> Group, which can vary bu the duration (if not specified) but even using
>>>>> small durations, my data is still not being deleted.
>>>>>
>>>>> I have the following retention policies:
>>>>> even the 10 hour retention policy, still has all the data i've added
>>>>> to it, which is 5 days (collected live over the last 5 days). Infact all
>>>>> of
>>>>> these retention policies still have all the data that was inserted into
>>>>> them. It seems like no deletes are running.
>>>>> Does the replica number affect that? I set 2, just so that if this is
>>>>> used in a cluster, it already has the retention policy setup correctly
>>>>> for
>>>>> 2, but it is currently not used in a cluster. My understanding is the
>>>>> replica number has no effect unless in a cluster, right?
>>>>>
>>>>> > show retention policies
>>>>> name
>>>>> duration shardGroupDuration replicaN default
>>>>> ----
>>>>> -------- ------------------ -------- -------
>>>>> default
>>>>> 0s 168h0m0s 1 true
>>>>> RP_debir_Test10hours
>>>>> 10h0m0s 1h0m0s 2 false
>>>>> RP_debir_Test1day
>>>>> 24h0m0s 1h0m0s 2 false
>>>>> RP_debir_Test2days
>>>>> 48h0m0s 24h0m0s 2 false
>>>>> RP_debir_Test3days
>>>>> 72h0m0s 24h0m0s 2 false
>>>>> RP_debir_Test1000datablocks
>>>>> 8333h20m0s 168h0m0s 2 false
>>>>> RP_it-edmTest1000datablocks
>>>>> 8333h20m0s 168h0m0s 2 false
>>>>> RP_negtest_Millies
>>>>> 336h0m0s 24h0m0s 2 false
>>>>>
>>>> --
>> Remember to include the version number!
>> ---
>> You received this message because you are subscribed to the Google Groups
>> "InfluxData" group.
>> To unsubscribe from this group and stop receiving emails from it, send an
>> email to [email protected] <javascript:>.
>> To post to this group, send email to [email protected]
>> <javascript:>.
>> Visit this group at https://groups.google.com/group/influxdb.
>> To view this discussion on the web visit
>> https://groups.google.com/d/msgid/influxdb/c5bddcba-ed9c-40eb-b884-364680e0a22b%40googlegroups.com
>>
>> <https://groups.google.com/d/msgid/influxdb/c5bddcba-ed9c-40eb-b884-364680e0a22b%40googlegroups.com?utm_medium=email&utm_source=footer>
>> .
>>
>> For more options, visit https://groups.google.com/d/optout.
>>
>
>
--
Remember to include the version number!
---
You received this message because you are subscribed to the Google Groups
"InfluxData" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To post to this group, send email to [email protected].
Visit this group at https://groups.google.com/group/influxdb.
To view this discussion on the web visit
https://groups.google.com/d/msgid/influxdb/1c7656ce-65f8-498d-93ec-b5ecfcf3ab15%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.