There have been several WAL corruption fixes since that version.

On Thu, Sep 23, 2021 at 5:23 PM Brandon Duffany <[email protected]>
wrote:

> Yep --
>
> Prometheus: 2.22.1 -- revision 00f16d1ac3a4c94561e5133b821d8e4d9ef78ec2
>
> Filesystem: ext4
>
>
> On Thursday, September 23, 2021 at 12:18:21 AM UTC-4 Julien Pivotto wrote:
>
>> Can we know the filesystem you use and your Prometheus version?
>>
>> Le jeu. 23 sept. 2021 à 06:06, Brandon Duffany <[email protected]> a
>> écrit :
>>
>>>
>>> One detail I forgot to mention which might be relevant: on Fri (Sep 10)
>>> we had an issue where Prometheus server was failing to write new metrics to
>>> disk because it was hitting disk capacity limits. We were seeing this in
>>> the logs:
>>>
>>> level=error ts=2021-09-10T19:16:01.392Z caller=scrape.go:1085
>>> component="scrape manager" scrape_pool=kubernetes-service-endpoints target=
>>> http://10.138.0.102:19100/metrics msg="Scrape commit failed" err="write
>>> to WAL: log samples: write /data/wal/00007614: no space left on device"
>>>
>>> We resolved this by increasing the size of the PersistentVolumeClaim
>>> from 32GB to 64GB. We ran into the issue again on Mon, Sep 20 and resolved
>>> it once again by doubling the storage capacity from 64GB to 128GB.
>>>
>>> Is Prometheus designed to be resilient to running out of disk space? Or
>>> could that be part of what caused the corruption?
>>> On Wednesday, September 22, 2021 at 10:53:59 PM UTC-4 Brandon Duffany
>>> wrote:
>>>
>>>> After our Prometheus server restarted today (Sep 22), we noticed that
>>>> 11 days worth of data somehow got deleted from TSDB, from Sep 10 through
>>>> Sep 21. (We are running Prometheus on Kubernetes using a persistent volume
>>>> for the TSDB data directory.)
>>>>
>>>> I think the data was actually deleted by Prometheus itself, because we
>>>> saw the disk usage at 50GB just after Prometheus server started, but then
>>>> dropped to around 8GB shortly after.
>>>>
>>>> Furthermore, we saw the following in Prometheus server logs:
>>>>
>>>> level=warn ts=2021-09-22T16:24:09.931Z caller=db.go:662 component=tsdb
>>>> msg="Encountered WAL read error, attempting repair" err="read records:
>>>> corruption in segment /data/wal/00007611 at 25123899: unexpected checksum
>>>> a70d7089, expected 30cb982b"
>>>>
>>>> level=warn ts=2021-09-22T16:24:09.931Z caller=wal.go:354 component=tsdb
>>>> msg="Starting corruption repair" segment=7611 offset=25123899
>>>>
>>>> level=warn ts=2021-09-22T16:24:09.933Z caller=wal.go:362 component=tsdb
>>>> msg="Deleting all segments newer than corrupted segment" segment=7611
>>>>
>>>>
>>>> And if I look at the TSDB data volume (/data) I see that there are a
>>>> bunch of data directories which look like they got deleted:
>>>>
>>>> /data $ stat -c '%y %n' * | sort
>>>> 2020-11-24 18:17:06.000000000 lost+found
>>>> 2020-11-24 18:17:14.000000000 lock
>>>> 2021-09-08 09:02:11.000000000 01FF2A7WB036B9VEN19TXFN6K6
>>>> 2021-09-09 03:03:33.000000000 01FF47ZTKY1SJBPKT11VED354K
>>>> 2021-09-09 21:06:33.000000000 01FF65WM7VS00ZQ98C1W0GAJ99
>>>> 2021-09-10 20:52:19.000000000 01FF8QK0AF514208XDW7W576QD
>>>> 2021-09-22 17:01:29.000000000 01FG776WNKM6W4ZPADSG20GK6H
>>>> 2021-09-22 21:00:36.000000000 01FG7MXRH65JXVR46PD0SWE951
>>>> 2021-09-22 21:01:30.000000000 01FG7MZF9EC1HMW57K9GR866WY
>>>> 2021-09-22 23:00:30.000000000 01FG7VSG0FBFFXWS1QJ68208HW
>>>> 2021-09-23 01:00:01.000000000 01FG82N78GAX6C31MMWR14K8BH
>>>> 2021-09-23 01:00:01.000000000 chunks_head
>>>> 2021-09-23 01:00:01.000000000 wal
>>>> 2021-09-23 02:42:48.000000000 queries.active
>>>>
>>>>
>>>> Has anyone run into similar issues before or know why this data
>>>> corruption might be happening?
>>>>
>>>> Or, is there anywhere we can look for hints as to why TSDB thought the
>>>> data was corrupted and removed so many of the chunks (over 40GB of data in
>>>> our case?)
>>>>
>>>> Any help would be greatly appreciated. Thanks!
>>>>
>>>> --
>>> You received this message because you are subscribed to the Google
>>> Groups "Prometheus Users" group.
>>> To unsubscribe from this group and stop receiving emails from it, send
>>> an email to [email protected].
>>> To view this discussion on the web visit
>>> https://groups.google.com/d/msgid/prometheus-users/19423814-230d-42a0-ade4-c90abc272cc0n%40googlegroups.com
>>> <https://groups.google.com/d/msgid/prometheus-users/19423814-230d-42a0-ade4-c90abc272cc0n%40googlegroups.com?utm_medium=email&utm_source=footer>
>>> .
>>>
>> --
> You received this message because you are subscribed to the Google Groups
> "Prometheus Users" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/prometheus-users/40aa0db9-7297-4ede-bcb1-37cf29c5f6d1n%40googlegroups.com
> <https://groups.google.com/d/msgid/prometheus-users/40aa0db9-7297-4ede-bcb1-37cf29c5f6d1n%40googlegroups.com?utm_medium=email&utm_source=footer>
> .
>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/CABbyFmqFbMZ-Py%2BZPnSfRL5_z0C528UjEVWDeVhXazESNZtkmg%40mail.gmail.com.

Reply via email to