Hi. We have AKS cluster on Azure, which is monitored by in-cluster prometheus. Prometheus pod has configured persistent storage which uses azure disk as a storage backend (it is discrbed here <https://docs.microsoft.com/en-us/azure/aks/azure-disks-dynamic-pv>).
Some time age we decided to use azure file as storage backend (which basic description can be found here <https://docs.microsoft.com/en-us/azure/aks/azure-files-dynamic-pv>). After some initial tests I have decided to run it, and migrate data from previous instance (by coping old directories with data). So I have started new instance, copied some old data to /data dir, restarted prometheus and everything was fine. Newly scraped data are present on dashboards along with old migrated. I have started to migrate another batch of data. Everything is working fine, after ~1,5h hour from start I noticed that server was creating a one folder with data each minute: drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:00 01E20SAAAK4CC27H07QH3ZXN3C drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:00 01E20SAQWA1KVBCNDBRZ5KTVQ1 drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:00 01E20SAYDGXC12DSXA7066NJ03 drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:00 01E20SB6G921Y6T0ZFJTKNWDNX drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:00 01E20SBDYENAF1NE7YRDHHY34K drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:00 01E20SBP0VW47HWK00EREXW15A drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:00 01E20SBY3EN5JS0K32X9WTNXC2 drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:01 01E20SC70QMKGF0QSSN25PESF0 drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:01 01E20SCHR65MQRMPG58XG8JMT9 drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:01 01E20SCVD861GEFYHR5695EQNH drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:01 01E20SD8C78SG806AN5X22DPHS drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:01 01E20SDK67WV2X82K8RNWAM1FD drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:01 01E20SDY6R4XVK6RJQA2PZH16H drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:02 01E20SE9X5ZG2WPV6E9RZ88JPK drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:02 01E20SEQG0T38AZK4G8SNZEAEG drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:02 01E20SF9AR8TJZE372ZRP06JYS drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:03 01E10T77TYT8YBJ7RH3R1CE88Z drwxr-xr-x 2 nobody nogroup 0 Feb 26 13:03 01E20SG2NKH38JM55SS8Y3N0M5 And in logs i saw something like this: level=error ts=2020-02-26T13:01:58.895708548Z caller=db.go:341 component=tsdb msg="compaction failed" err="reload blocks: invalid block sequence: block time ranges overlap: [mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s, blocks: 12]: <ulid: 01E20SBDYENAF1NE7YRDHHY34K, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SAQWA1KVBCNDBRZ5KTVQ1, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SAYDGXC12DSXA7066NJ03, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SB6G921Y6T0ZFJTKNWDNX, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SAAAK4CC27H07QH3ZXN3C, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SBP0VW47HWK00EREXW15A, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SBY3EN5JS0K32X9WTNXC2, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SC70QMKGF0QSSN25PESF0, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SCHR65MQRMPG58XG8JMT9, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SCVD861GEFYHR5695EQNH, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SD8C78SG806AN5X22DPHS, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SDK67WV2X82K8RNWAM1FD, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>" level=info ts=2020-02-26T13:02:02.134855312Z caller=compact.go:443 component=tsdb msg="write block" mint=1582711200000 maxt=1582718400000 ulid=01E20SDY6R4XVK6RJQA2PZH16H ... (shortened for readability) And after some time this log entry is repeated multiple times but with higher block count: level=info ts=2020-02-26T14:02:24.079677532Z caller=compact.go:443 component=tsdb msg="write block" mint=1582711200000 maxt=1582718400000 ulid=01E20WWEN4Y1FW535M5VN1WZCZ level=error ts=2020-02-26T14:03:40.144058489Z caller=db.go:341 component=tsdb msg="compaction failed" err="reload blocks: invalid block sequence: block time ranges overlap: [mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s, blocks: 95]: <ulid: 01E20TSHHXVCCEQZ7RMR1F0N8H, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SAQWA1KVBCNDBRZ5KTVQ1, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SAYDGXC12DSXA7066NJ03, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SB6G921Y6 T0ZFJTKNWDNX, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SBDYENAF1NE7YRDHHY34K, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SBP0VW47HWK00EREXW15A, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20TYZV0Y8FMM3WT1DY6FRKB, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SC70QMKGF0QSSN25PESF0, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SCHR65MQRMPG58XG8JMT9, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SCVD861GEFYHR5695EQNH, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SD8C78SG806AN5X22DPHS, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SDK67WV2X82K8RNWAM1FD, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SDY6R4XVK6RJQA2PZH16H, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SE9X5ZG2WPV6E9RZ88JPK, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SEQG0T38AZK4G8SNZEAEG, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SF9AR8TJZE372ZRP06JYS, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SG2NKH38JM55SS8Y3N0M5, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SGDREP1SQNRH738259AC5, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SGR6W3VCM0MH8QB6VTCPS, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SH2D4WAD4FM2DC6ZH0TAM, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SHEBKHMV1P7KP7B89FRYM, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SHWT3C6WA3TG09SB80H45, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SJ9NXN1M0AXS18R6X5CK2, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SJQ3Z3KD60YN6WQ805DNH, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SK3F800SK1MKN19PGGETZ, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SKM5AM6FB00248HPRZ1W4, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SM1Z00J5S07ZMCP5V3JR1, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SMF3F3PEGJ8244CBBG784, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SMXWGCF7J2ZW510TE5QVW, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SND70FW39613K843ZVH4Z, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SNVH2ACCPASRWT8H5RZH5, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SPARSEFP8E9EX9461NE4M, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SPT29R98NC852BMQTNZ1Y, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SQAKQE8T41XS5Q70049FG, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SQV17F1F4PBJYSJ9TXVJZ, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SRCYKKG5G210SX9RC4RGA, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SAAAK4CC27H07QH3ZXN3C, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SSG76YSCXA7K9Q75CWDJA, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20ST49KD8029JGAH7Y090EX, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20STV260VAPK05DY5DES9PV, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SVGFTJPGA2YENCGW75QTF, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SX2EEP791A6JJMFMSFRDG, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20SZFX1AG9NYBNHV22XAVQP, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T0683AYR4J7PSC8YTMBFR, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T0TA05FFDNYMJ0W33PFGV, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T1DQGFZ0SA3FW60H6TY2D, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T231MT7CTDQRHR7KR3A4Z, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T2VBWQR4D2T7C11V4HCGB, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T3H789XEW22454B17KHBF, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T46199V8E0WRWQ0GN3J69, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T4X0PXDM79TCW22A1ANJN, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T5K9DGF3TS09EYW5B3E9Q, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T6A0PFCT3JJS86MBQNNYS, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T71QNK9WBGT5QFDWMGYY2, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T7TEM7Y1TNKQYV8VPJW9H, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T8H90FC39T0G7JZG8E86Y, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20T99XASQ8MQTNVQRRPG9XZ, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20TA5JRJZZVP5AXJWTKCYHD, mint: 1582711200000, maxt: 1582718400000, range: 2h0m0s>, <ulid: 01E20TCS5H585B02X2SZCRFN9W, mint: 1582711200000, maxt: ... (shortened for readability, but this is much longer than the first entry) When I run prometheus server and I don't copy data everything is ok. Copied data don't overlap, for sure, they are from beginning of this month. Timestamps and increasing number of blocks suggests that compacting loop creates additional folder with data, and don't delete old one, and creates another folder with compacted data... and so one. What can be cause of this problem? -- You received this message because you are subscribed to the Google Groups "Prometheus Users" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/prometheus-users/d5d5f178-fb31-4737-99f8-efdd0b8a4ea2%40googlegroups.com.

