On 2018年01月29日 23:08, ^m'e wrote: > Lowmem check on my backup pool reports dozens of 'backref lost' on > extents... Excerpt: > ---------------------------------------------------------- > # ./btrfsck.static check --mode=lowmem /dev/sda1 > checking extents > ERROR: data extent[33866182656 4096] backref lost > ERROR: data extent[37102219264 114688] backref lost > ERROR: data extent[37090353152 45056] backref lost > ERROR: data extent[37193342976 114688] backref lost > ERROR: data extent[50151686144 53248] backref lost > ERROR: data extent[49782943744 126976] backref lost > ERROR: data extent[49718861824 77824] backref lost > ERROR: data extent[33853538304 4096] backref lost > ERROR: data extent[41170333696 692224] backref lost > ERROR: data extent[37550239744 4096] backref lost > ERROR: data extent[33866182656 4096] backref lost > ERROR: data extent[37102219264 114688] backref lost > ... > > ERROR: data extent[37193342976 114688] backref lost > ERROR: data extent[50151686144 53248] backref lost > ERROR: data extent[49782943744 126976] backref lost > ERROR: data extent[49718861824 77824] backref lost > ERROR: data extent[33853538304 4096] backref lost > ERROR: data extent[41170333696 692224] backref lost > ERROR: data extent[37550239744 4096] backref lost > ERROR: data extent[44122832896 8192] backref lost > ERROR: data extent[45874237440 40960] backref lost > ERROR: errors found in extent allocation tree or chunk allocation > checking free space cache > block group 112772251648 has wrong amount of free space > failed to load free space cache for block group 112772251648 > Wanted offset 127268339712, found 127268323328 > Wanted offset 127268339712, found 127268323328 > cache appears valid but isn't 127267766272 > block group 133173346304 has wrong amount of free space > failed to load free space cache for block group 133173346304 > Wanted offset 142837039104, found 142837022720 > Wanted offset 142837039104, found 142837022720 > cache appears valid but isn't 142837022720 > ERROR: errors found in free space cache > Checking filesystem on /dev/sda1 > UUID: 854e1bf5-7a98-4bcb-b971-0d9f2ac9452a > found 200684883968 bytes used, error(s) found > total csum bytes: 186712500 > total tree bytes: 40191000576 > total fs tree bytes: 39574110208 > total extent tree bytes: 359038976 > btree space waste bytes: 7605597010 > file data blocks allocated: 946099404800 > referenced 1091731312640 > ---------------------------------------------------------- > > Is that related to flawed snapshots of the ailing rootfs vol?
For that specific problem, lowmem is not that reliable. Use original mode to make sure. Thanks, Qu > Is it safe to try to rollback from it? > > Cheers, > > M. > > > On Mon, Jan 29, 2018 at 2:49 PM, ^m'e <marc...@gmail.com> wrote: >> On Mon, Jan 29, 2018 at 2:04 PM, Qu Wenruo <quwenruo.bt...@gmx.com> wrote: >>> >>> >>> On 2018年01月29日 21:58, ^m'e wrote: >>>> Thanks for the advice, Qu! >>>> >>>> I used the system for a while, did some package upgrades -- writing in >>>> the suspect corrupted area. Then tried a btrfs-send to my backup vol, >>>> and it failed miserably with a nice kernel oops. >>>> >>>> So I went for a lowmem repair: >>>> ---------------------------------------------------------------------------------------- >>>> # ./btrfsck.static check --repair --mode=lowmem /dev/sdb3 2>&1 | tee >>>> /mnt/custom/rescue/btrfs-recovery/btrfs-repair.BTR-POOL.1.log >>>> WARNING: low-memory mode repair support is only partial >>>> Fixed 0 roots. >>>> checking extents >>>> checking free space cache >>>> checking fs roots >>>> ERROR: failed to add inode 28891726 as orphan item root 257 >>>> ERROR: root 257 INODE[28891726] is orphan item >>> >>> At least I need dig the kernel code further to determine if the orphan >>> inode handling in btrfs-progs is correct or not. >>> >>> So there won't be more dirty fix soon. >>> >>> Hopefully you could get some good backup and restore the system. >>> >>> At least the problem is limited to a very small range, and it's >>> something we could handle easily. >>> >>> Thanks for all your report, >>> Qu >>> >>> >> >> Right. >> >> Meanwhile, could you please suggest the best course of action? btrfs >> rescue or restore? >> I have snapshots of my two subvols (rootfs, home -- now fs-checking >> them just in case...) >> >> Cheers, >> >> Marco > > >
signature.asc
Description: OpenPGP digital signature