Am 29.08.19 um 15:11 schrieb Qu Wenruo:
>
>
> On 2019/8/29 下午8:46, Oliver Freyermuth wrote:
>> Am 27.08.19 um 14:40 schrieb Hans van Kranenburg:
>>> On 8/27/19 11:14 AM, Swâmi Petaramesh wrote:
>>>> On 8/27/19 8:52 AM, Qu Wenruo wrote:
>>>>>> or to use the V2 space
>>>>>> cache generally speaking, on any machine that I use (I had understood it
>>>>>> was useful only on multi-TB filesystems...)
>>>>> 10GiB is enough to create large enough block groups to utilize free
>>>>> space cache.
>>>>> So you can't really escape from free space cache.
>>>>
>>>> I meant that I had understood that the V2 space cache was preferable to
>>>> V1 only for multi-TB filesystems.
>>>>
>>>> So would you advise to use V2 space cache also for filesystems < 1 TB ?
>>>
>>> Yes.
>>>
>>
>> This makes me wonder if it should be the default?
>
> It will be.
>
> Just a spoiler, I believe features like no-holes and v2 space cache will
> be default in not so far future.
>
>>
>> This thread made me check on my various BTRFS volumes and for almost all of
>> them (in different machines), I find cases of
>> failed to load free space cache for block group XXXX, rebuilding it now
>> at several points during the last months in my syslogs - and that's for
>> machines without broken memory, for disks for which FUA should be working
>> fine,
>> without any unsafe shutdowns over their lifetime, and with histories as
>> short as only having seen 5.x kernels.
>
> That's interesting. In theory that shouldn't happen, especially without
> unsafe shutdown.
I also forgot to add that in addition on the machines there is no mdraid / dm /
LUKS in between (i.e. purely btrfs on the drives).
The messages _seem_ to be more prominent for spinning disks, but after all, my
statistics is just 5 devices in total.
So it really "feels" like a bug crawling somewhere. However, the machines seem
to not have not seen any actual corruption as consequence.
I'm playing with "btrfs check --readonly" now to see if there's really
everything still fine, but I'm already running kernel 5.2 with the new checks
without issues.
> But please also be aware that, there is no concrete proof that corrupted
> v1 space cache is causing all the problems.
> What I said is just, corrupted v1 space cache may cause problem, I need
> to at least craft an image to proof my assumption.
I see - that might be useful in any case to hopefully track down the issue.
>
>>
>> So if this may cause harmful side effects, happens without clear origin, and
>> v2 is safer due to being CoW,
>> I guess I should switch all my nodes to v2 (or this should become the
>> default in a future kernel?).
>
> At least, your experience would definitely help the btrfs community.
Ok, then I will slowly switch the nodes one by one - in case I do not come and
cry on the list, this means all is well (but I'm only a small datapoint with 5
disks in three machines) ;-).
Cheers,
Oliver
>
> Thanks,
> Qu
>
>>
>> Cheers,
>> Oliver
>>