Re: bad tree block start, want 705757184 have 82362368
Am Sonntag, 18. November 2018, 14:31:36 CET schrieb Anand Jain: > On 11/18/2018 03:56 PM, Stephan Olbrich wrote: > > Am Sonntag, 18. November 2018, 01:30:14 CET schrieb Qu Wenruo: > > Late on I got the same errors for my /home partition (on the same > > drive) > > as well. I have snapshots of all partitions on another drive made by > > btrbk. To get a working system, I made new (rw) snapshots of the most > > recent backup and setup grub and fstab, so my system would boot from > > the > > other drive. Unfortunately now I got the "bad tree block start" error > > again at least once in dmesg but I didn't save it and it's not in > > syslog > > > > :-( What I remember is, that it was followed by other btrfs error > > > > messages saying something about correcting something. And the > > filesystem > > was still read/write this time. > > At the moment I can't reproduce it. > > > > Today it happend again (sdb is my backup drive, which is my main drive at > > the moment): [ 286.325857] BTRFS error (device sdb1): bad tree block > > start, want 787719208960 have 11268016545161247416 [ 286.363245] BTRFS > > info (device sdb1): read error corrected: ino 0 off 787719208960 (dev > > /dev/sdb1 sector 1243815072) [ 286.364087] BTRFS info (device sdb1): > > read error corrected: ino 0 off 787719213056 (dev /dev/sdb1 sector > > 1243815080) [ 286.425946] BTRFS info (device sdb1): read error > > corrected: ino 0 off 787719217152 (dev /dev/sdb1 sector 1243815088) [ > > 286.427530] BTRFS info (device sdb1): read error corrected: ino 0 off > > 787719221248 (dev /dev/sdb1 sector 1243815096) > Was there any hardware issues? How about the following data from the > system.. > >btrfs fi df >btrfs dev stat I didn't have any hardware issues (that I know of). $ btrfs fi df / Data, single: total=841.00GiB, used=791.92GiB System, DUP: total=8.00MiB, used=112.00KiB System, single: total=4.00MiB, used=0.00B Metadata, DUP: total=10.50GiB, used=8.44GiB Metadata, single: total=8.00MiB, used=0.00B GlobalReserve, single: total=512.00MiB, used=0.00B $ btrfs dev stat / [/dev/sdb1].write_io_errs0 [/dev/sdb1].read_io_errs 0 [/dev/sdb1].flush_io_errs0 [/dev/sdb1].corruption_errs 0 [/dev/sdb1].generation_errs 0 Also the numbers in the error messages keep changing: [ 1577.302959] BTRFS error (device sdb1): bad tree block start, want 788086226944 have 411602213549769407 [ 1577.369083] BTRFS info (device sdb1): read error corrected: ino 0 off 788086226944 (dev /dev/sdb1 sector 1244531904) [ 1577.428139] BTRFS info (device sdb1): read error corrected: ino 0 off 788086231040 (dev /dev/sdb1 sector 1244531912) [ 1577.429091] BTRFS info (device sdb1): read error corrected: ino 0 off 788086235136 (dev /dev/sdb1 sector 1244531920) [ 1577.478249] BTRFS info (device sdb1): read error corrected: ino 0 off 788086239232 (dev /dev/sdb1 sector 1244531928) Thanks, Stephan > Thanks, Anand > > > Is there any way to find out, which files are affected by the errors > > above? > > No files are affected, but an essential tree, extent tree, is > corrupted. > > Normally this may prevent RW mount, and even it mounts it can still > cause problem when doing any write. > It could even
Re: bad tree block start, want 705757184 have 82362368
On 11/18/2018 03:56 PM, Stephan Olbrich wrote: Am Sonntag, 18. November 2018, 01:30:14 CET schrieb Qu Wenruo: Late on I got the same errors for my /home partition (on the same drive) as well. I have snapshots of all partitions on another drive made by btrbk. To get a working system, I made new (rw) snapshots of the most recent backup and setup grub and fstab, so my system would boot from the other drive. Unfortunately now I got the "bad tree block start" error again at least once in dmesg but I didn't save it and it's not in syslog :-( What I remember is, that it was followed by other btrfs error messages saying something about correcting something. And the filesystem was still read/write this time. At the moment I can't reproduce it. Today it happend again (sdb is my backup drive, which is my main drive at the moment): [ 286.325857] BTRFS error (device sdb1): bad tree block start, want 787719208960 have 11268016545161247416 [ 286.363245] BTRFS info (device sdb1): read error corrected: ino 0 off 787719208960 (dev /dev/sdb1 sector 1243815072) [ 286.364087] BTRFS info (device sdb1): read error corrected: ino 0 off 787719213056 (dev /dev/sdb1 sector 1243815080) [ 286.425946] BTRFS info (device sdb1): read error corrected: ino 0 off 787719217152 (dev /dev/sdb1 sector 1243815088) [ 286.427530] BTRFS info (device sdb1): read error corrected: ino 0 off 787719221248 (dev /dev/sdb1 sector 1243815096) Was there any hardware issues? How about the following data from the system.. btrfs fi df btrfs dev stat Thanks, Anand Is there any way to find out, which files are affected by the errors above? No files are affected, but an essential tree, extent tree, is corrupted. Normally this may prevent RW mount, and even it mounts it can still cause problem when doing any write. It could even prevent RO mount if the corrupted leaf contains block group item. But your data should be OK if there is no other corruption, and in that case btrfs-restore should work well. I don't really trust the data on the drive I'm using at the moment, as it has shown errors as well, but I have a less current backup on yet another drive but at it is a few weeks old, I don't want to use it to setup the system on the SSD again, but just copy the relevant files if possible. Or is it possible to repair the original file system? At least we need "btrfs check" output. I updated btrfs-progs and run btrfs check for / and /home No errors are found on / (sda2), but there are errors on /home ?? $ btrfs --version btrfs-progs v4.19 $ btrfs check /dev/sda2 Opening filesystem to check... Checking filesystem on /dev/sda2 UUID: 80368989-ffa8-463c-98fb-fe2e28ca7bf3 [1/7] checking root items [2/7] checking extents [3/7] checking free space cache [4/7] checking fs roots [5/7] checking only csums items (without verifying data) [6/7] checking root refs [7/7] checking quota groups skipped (not enabled on this FS) found 64816218112 bytes used, no error found total csum bytes: 59518732 total tree bytes: 2180268032 total fs tree bytes: 1965965312 total extent tree bytes: 123289600 btree space waste bytes: 478665777 file data blocks allocated: 151083261952 referenced 76879990784 This fs is completely fine, including your data. $ btrfs check /dev/sda4 Opening filesystem to check... Checking filesystem on /dev/sda4 UUID: 81c38df8-b7f9-412c-8c88-cfde8db68eb1 [1/7] checking root items [2/7] checking extents [3/7] checking free space cache [4/7] checking fs roots root 257 inode 7970563 errors 100, file extent discount Found file extent holes: start: 0, len: 20480 root 257 inode 7970564 errors 100, file extent discount Found file extent holes: start: 0, len: 77824 ERROR: errors found in fs roots These are just minor errors, won't even causing any data mismatch. So all your fses should be mostly OK. Would you please try to use v4.19-rc* to see if it changes anything? v4.19-rc1 is the only rc I found, but that is older than v4.19, right? Anyway, here is the output: $ btrfs --version btrfs-progs v4.19-rc1 $ btrfs check /dev/sda2 Opening filesystem to check... Checking filesystem on /dev/sda2 UUID: 80368989-ffa8-463c-98fb-fe2e28ca7bf3 [1/7] checking root items [2/7] checking extents [3/7] checking free space cache [4/7] checking fs roots [5/7] checking only csums items (without verifying data) [6/7] checking root refs [7/7] checking quota groups skipped (not enabled on this FS) found 64816218112 bytes used, no error found total csum bytes: 59518732 total tree bytes: 2180268032 total fs tree bytes: 1965965312 total extent tree bytes: 123289600 btree space waste bytes: 478665777 file data blocks allocated: 151083261952 referenced 76879990784 $ btrfs check /dev/sda4 Opening filesystem to check... Checking filesystem on /dev/sda4 UUID: 81c38df8-b7f9-412c-8c88-cfde8db68eb1 [1/7] checking root items [2/7] checking extents [3/7] checking free space cache [4/7] checking fs roots root 257 inode 7970563 errors
Re: bad tree block start, want 705757184 have 82362368
Am Sonntag, 18. November 2018, 01:30:14 CET schrieb Qu Wenruo: > >>> Late on I got the same errors for my /home partition (on the same drive) > >>> as well. I have snapshots of all partitions on another drive made by > >>> btrbk. To get a working system, I made new (rw) snapshots of the most > >>> recent backup and setup grub and fstab, so my system would boot from the > >>> other drive. Unfortunately now I got the "bad tree block start" error > >>> again at least once in dmesg but I didn't save it and it's not in syslog > >>> :-( What I remember is, that it was followed by other btrfs error > >>> messages saying something about correcting something. And the filesystem > >>> was still read/write this time. > >>> At the moment I can't reproduce it. Today it happend again (sdb is my backup drive, which is my main drive at the moment): [ 286.325857] BTRFS error (device sdb1): bad tree block start, want 787719208960 have 11268016545161247416 [ 286.363245] BTRFS info (device sdb1): read error corrected: ino 0 off 787719208960 (dev /dev/sdb1 sector 1243815072) [ 286.364087] BTRFS info (device sdb1): read error corrected: ino 0 off 787719213056 (dev /dev/sdb1 sector 1243815080) [ 286.425946] BTRFS info (device sdb1): read error corrected: ino 0 off 787719217152 (dev /dev/sdb1 sector 1243815088) [ 286.427530] BTRFS info (device sdb1): read error corrected: ino 0 off 787719221248 (dev /dev/sdb1 sector 1243815096) > >>> Is there any way to find out, which files are affected by the errors > >>> above? > >> > >> No files are affected, but an essential tree, extent tree, is corrupted. > >> > >> Normally this may prevent RW mount, and even it mounts it can still > >> cause problem when doing any write. > >> It could even prevent RO mount if the corrupted leaf contains block > >> group item. > >> > >> But your data should be OK if there is no other corruption, and in that > >> case btrfs-restore should work well. > >> > >>> I don't really trust the data on the drive I'm using at the > >>> moment, as it has shown errors as well, but I have a less current backup > >>> on yet another drive but at it is a few weeks old, I don't want to use > >>> it > >>> to setup the system on the SSD again, but just copy the relevant files > >>> if > >>> possible. Or is it possible to repair the original file system? > >> > >> At least we need "btrfs check" output. > > > > I updated btrfs-progs and run btrfs check for / and /home > > No errors are found on / (sda2), but there are errors on /home ?? > > > > $ btrfs --version > > btrfs-progs v4.19 > > > > $ btrfs check /dev/sda2 > > Opening filesystem to check... > > Checking filesystem on /dev/sda2 > > UUID: 80368989-ffa8-463c-98fb-fe2e28ca7bf3 > > [1/7] checking root items > > [2/7] checking extents > > [3/7] checking free space cache > > [4/7] checking fs roots > > [5/7] checking only csums items (without verifying data) > > [6/7] checking root refs > > [7/7] checking quota groups skipped (not enabled on this FS) > > found 64816218112 bytes used, no error found > > total csum bytes: 59518732 > > total tree bytes: 2180268032 > > total fs tree bytes: 1965965312 > > total extent tree bytes: 123289600 > > btree space waste bytes: 478665777 > > file data blocks allocated: 151083261952 > > > > referenced 76879990784 > > This fs is completely fine, including your data. > > > $ btrfs check /dev/sda4 > > Opening filesystem to check... > > Checking filesystem on /dev/sda4 > > UUID: 81c38df8-b7f9-412c-8c88-cfde8db68eb1 > > [1/7] checking root items > > [2/7] checking extents > > [3/7] checking free space cache > > [4/7] checking fs roots > > root 257 inode 7970563 errors 100, file extent discount > > > > Found file extent holes: > > start: 0, len: 20480 > > > > root 257 inode 7970564 errors 100, file extent discount > > > > Found file extent holes: > > start: 0, len: 77824 > > > > ERROR: errors found in fs roots > > These are just minor errors, won't even causing any data mismatch. > > So all your fses should be mostly OK. > > Would you please try to use v4.19-rc* to see if it changes anything? v4.19-rc1 is the only rc I found, but that is older than v4.19, right? Anyway, here is the output: $ btrfs --version btrfs-progs v4.19-rc1 $ btrfs check /dev/sda2 Opening filesystem to check... Checking filesystem on /dev/sda2 UUID: 80368989-ffa8-463c-98fb-fe2e28ca7bf3 [1/7] checking root items [2/7] checking extents [3/7] checking free space cache [4/7] checking fs roots [5/7] checking only csums items (without verifying data) [6/7] checking root refs [7/7] checking quota groups skipped (not enabled on this FS) found 64816218112 bytes used, no error found total csum bytes: 59518732 total tree bytes: 2180268032 total fs tree bytes: 1965965312 total extent tree bytes: 123289600 btree space waste bytes: 478665777 file data blocks allocated: 151083261952 referenced 76879990784 $ btrfs check /dev/sda4 Opening filesystem to check... Checking filesystem on
Re: bad tree block start, want 705757184 have 82362368
On 2018/11/18 上午4:28, Stephan Olbrich wrote: > Am Samstag, 17. November 2018, 09:03:04 CET schrieb Qu Wenruo: >> On 2018/11/17 上午12:17, Stephan Olbrich wrote: >>> Hi, >>> >>> a few days ago my root file system (simple btrfs on a SSD, no RAID or >>> anything) suddenly became read only. Looking at dmsg, I found this: >>> >>> [ 19.285020] BTRFS error (device sda2): bad tree block start, want >>> 705757184 have 82362368 >> Does this only occurs once? >> >> If only once, are you using SINGLE metadata profile (default for SSD)? >> If using DUP/RAID1 it may have a chance to recover. > It shows up once after reboot for each partition. > I'm using SINGLE metadata. > >> >>> [ 19.285042] BTRFS: error (device sda2) in __btrfs_free_extent:6804: >>> errno=-5 IO failure >> The problem is in extent tree. > > I checked my logs again, the __btrfs_free_extent error is missing for the > /home partition: > > [ 3154.066544] BTRFS error (device sda4): bad tree block start, want > 560119808 have 4330389667183373366 > [ 3154.066558] BTRFS: error (device sda4) in btrfs_run_delayed_refs:2934: > errno=-5 IO failure > [ 3154.066561] BTRFS info (device sda4): forced readonly > [ 3154.066904] BTRFS error (device sda4): pending csums is 360448 > >> >> If there is no other problem, your data should be OK. >> >> You could still try to mount the fs RO to salvage data. > > I'm not so much worried about salvaging data, as I have a current backup. I'm > more worried > how reliable the backup is, as I'm not sure, when the errors first started. > But if the data is > OK, then the backup (snapshots on another drive) should be OK as well I guess. > >>> [ 19.285048] BTRFS info (device sda2): forced readonly >>> [ 19.285051] BTRFS: error (device sda2) in btrfs_run_delayed_refs:2934: >>> errno=-5 IO failure [ 19.287213] BTRFS error (device sda2): pending >>> csums is 41889792 Still extent tree. >>> >>> Late on I got the same errors for my /home partition (on the same drive) >>> as well. I have snapshots of all partitions on another drive made by >>> btrbk. To get a working system, I made new (rw) snapshots of the most >>> recent backup and setup grub and fstab, so my system would boot from the >>> other drive. Unfortunately now I got the "bad tree block start" error >>> again at least once in dmesg but I didn't save it and it's not in syslog >>> :-( What I remember is, that it was followed by other btrfs error >>> messages saying something about correcting something. And the filesystem >>> was still read/write this time. >>> At the moment I can't reproduce it. >>> >>> Is there any way to find out, which files are affected by the errors >>> above? >> >> No files are affected, but an essential tree, extent tree, is corrupted. >> >> Normally this may prevent RW mount, and even it mounts it can still >> cause problem when doing any write. >> It could even prevent RO mount if the corrupted leaf contains block >> group item. >> >> But your data should be OK if there is no other corruption, and in that >> case btrfs-restore should work well. >> >>> I don't really trust the data on the drive I'm using at the >>> moment, as it has shown errors as well, but I have a less current backup >>> on yet another drive but at it is a few weeks old, I don't want to use it >>> to setup the system on the SSD again, but just copy the relevant files if >>> possible. Or is it possible to repair the original file system? >> >> At least we need "btrfs check" output. > I updated btrfs-progs and run btrfs check for / and /home > No errors are found on / (sda2), but there are errors on /home ?? > > $ btrfs --version > btrfs-progs v4.19 > > $ btrfs check /dev/sda2 > Opening filesystem to check... > Checking filesystem on /dev/sda2 > UUID: 80368989-ffa8-463c-98fb-fe2e28ca7bf3 > [1/7] checking root items > [2/7] checking extents > [3/7] checking free space cache > [4/7] checking fs roots > [5/7] checking only csums items (without verifying data) > [6/7] checking root refs > [7/7] checking quota groups skipped (not enabled on this FS) > found 64816218112 bytes used, no error found > total csum bytes: 59518732 > total tree bytes: 2180268032 > total fs tree bytes: 1965965312 > total extent tree bytes: 123289600 > btree space waste bytes: 478665777 > file data blocks allocated: 151083261952 > referenced 76879990784 This fs is completely fine, including your data. > > $ btrfs check /dev/sda4 > Opening filesystem to check... > Checking filesystem on /dev/sda4 > UUID: 81c38df8-b7f9-412c-8c88-cfde8db68eb1 > [1/7] checking root items > [2/7] checking extents > [3/7] checking free space cache > [4/7] checking fs roots > root 257 inode 7970563 errors 100, file extent discount > Found file extent holes: > start: 0, len: 20480 > root 257 inode 7970564 errors 100, file extent discount > Found file extent holes: > start: 0, len: 77824 > ERROR: errors found in fs roots These are just minor errors, won't even causing any data mismatch. So all your
Re: bad tree block start, want 705757184 have 82362368
Am Freitag, 16. November 2018, 17:44:35 CET schrieb Nikolay Borisov: > On 16.11.18 г. 18:17 ч., Stephan Olbrich wrote: > > Hi, > > > > a few days ago my root file system (simple btrfs on a SSD, no RAID or > > anything) suddenly became read only. Looking at dmsg, I found this: > > > > [ 19.285020] BTRFS error (device sda2): bad tree block start, want > > 705757184 have 82362368 [ 19.285042] BTRFS: error (device sda2) in > > __btrfs_free_extent:6804: errno=-5 IO failure [ 19.285048] BTRFS info > > (device sda2): forced readonly > > [ 19.285051] BTRFS: error (device sda2) in btrfs_run_delayed_refs:2934: > > errno=-5 IO failure [ 19.287213] BTRFS error (device sda2): pending > > csums is 41889792 > > > > Late on I got the same errors for my /home partition (on the same drive) > > as well. I have snapshots of all partitions on another drive made by > > btrbk. To get a working system, I made new (rw) snapshots of the most > > recent backup and setup grub and fstab, so my system would boot from the > > other drive. Unfortunately now I got the "bad tree block start" error > > again at least once in dmesg but I didn't save it and it's not in syslog > > :-( What I remember is, that it was followed by other btrfs error > > messages saying something about correcting something. And the filesystem > > was still read/write this time. > > At the moment I can't reproduce it. > > > > Is there any way to find out, which files are affected by the errors > > above? I don't really trust the data on the drive I'm using at the > > moment, as it has shown errors as well, but I have a less current backup > > on yet another drive but at it is a few weeks old, I don't want to use it > > to setup the system on the SSD again, but just copy the relevant files if > > possible. Or is it possible to repair the original file system? > > > > Some information about my system: > > Kubuntu 18.04 > > Kernel 4.19.1 when the problem occured, now 4.19.2 > > btrfs-tools 4.15.1 > > What is the SMART status of your SSD, how old is the ssd. This really > sounds like the drive going to lalal land. The SSD is 3.5 years old and was never really full, so I wouldn't expect it to die. SMART shows no errors. All values are way above the threshold and selftests show no errors. I checked RAM as well. No errors there, so I have no clue, where these errors come from. Regards, Stephan
Re: bad tree block start, want 705757184 have 82362368
Am Samstag, 17. November 2018, 09:03:04 CET schrieb Qu Wenruo: > On 2018/11/17 上午12:17, Stephan Olbrich wrote: > > Hi, > > > > a few days ago my root file system (simple btrfs on a SSD, no RAID or > > anything) suddenly became read only. Looking at dmsg, I found this: > > > > [ 19.285020] BTRFS error (device sda2): bad tree block start, want > > 705757184 have 82362368 > Does this only occurs once? > > If only once, are you using SINGLE metadata profile (default for SSD)? > If using DUP/RAID1 it may have a chance to recover. It shows up once after reboot for each partition. I'm using SINGLE metadata. > > > [ 19.285042] BTRFS: error (device sda2) in __btrfs_free_extent:6804: > > errno=-5 IO failure > The problem is in extent tree. I checked my logs again, the __btrfs_free_extent error is missing for the /home partition: [ 3154.066544] BTRFS error (device sda4): bad tree block start, want 560119808 have 4330389667183373366 [ 3154.066558] BTRFS: error (device sda4) in btrfs_run_delayed_refs:2934: errno=-5 IO failure [ 3154.066561] BTRFS info (device sda4): forced readonly [ 3154.066904] BTRFS error (device sda4): pending csums is 360448 > > If there is no other problem, your data should be OK. > > You could still try to mount the fs RO to salvage data. I'm not so much worried about salvaging data, as I have a current backup. I'm more worried how reliable the backup is, as I'm not sure, when the errors first started. But if the data is OK, then the backup (snapshots on another drive) should be OK as well I guess. > > [ 19.285048] BTRFS info (device sda2): forced readonly > > [ 19.285051] BTRFS: error (device sda2) in btrfs_run_delayed_refs:2934: > > errno=-5 IO failure [ 19.287213] BTRFS error (device sda2): pending > > csums is 41889792 > > > > Late on I got the same errors for my /home partition (on the same drive) > > as well. I have snapshots of all partitions on another drive made by > > btrbk. To get a working system, I made new (rw) snapshots of the most > > recent backup and setup grub and fstab, so my system would boot from the > > other drive. Unfortunately now I got the "bad tree block start" error > > again at least once in dmesg but I didn't save it and it's not in syslog > > :-( What I remember is, that it was followed by other btrfs error > > messages saying something about correcting something. And the filesystem > > was still read/write this time. > > At the moment I can't reproduce it. > > > > Is there any way to find out, which files are affected by the errors > > above? > > No files are affected, but an essential tree, extent tree, is corrupted. > > Normally this may prevent RW mount, and even it mounts it can still > cause problem when doing any write. > It could even prevent RO mount if the corrupted leaf contains block > group item. > > But your data should be OK if there is no other corruption, and in that > case btrfs-restore should work well. > > > I don't really trust the data on the drive I'm using at the > > moment, as it has shown errors as well, but I have a less current backup > > on yet another drive but at it is a few weeks old, I don't want to use it > > to setup the system on the SSD again, but just copy the relevant files if > > possible. Or is it possible to repair the original file system? > > At least we need "btrfs check" output. I updated btrfs-progs and run btrfs check for / and /home No errors are found on / (sda2), but there are errors on /home ?? $ btrfs --version btrfs-progs v4.19 $ btrfs check /dev/sda2 Opening filesystem to check... Checking filesystem on /dev/sda2 UUID: 80368989-ffa8-463c-98fb-fe2e28ca7bf3 [1/7] checking root items [2/7] checking extents [3/7] checking free space cache [4/7] checking fs roots [5/7] checking only csums items (without verifying data) [6/7] checking root refs [7/7] checking quota groups skipped (not enabled on this FS) found 64816218112 bytes used, no error found total csum bytes: 59518732 total tree bytes: 2180268032 total fs tree bytes: 1965965312 total extent tree bytes: 123289600 btree space waste bytes: 478665777 file data blocks allocated: 151083261952 referenced 76879990784 $ btrfs check /dev/sda4 Opening filesystem to check... Checking filesystem on /dev/sda4 UUID: 81c38df8-b7f9-412c-8c88-cfde8db68eb1 [1/7] checking root items [2/7] checking extents [3/7] checking free space cache [4/7] checking fs roots root 257 inode 7970563 errors 100, file extent discount Found file extent holes: start: 0, len: 20480 root 257 inode 7970564 errors 100, file extent discount Found file extent holes: start: 0, len: 77824 ERROR: errors found in fs roots found 303386652672 bytes used, error(s) found total csum bytes: 289501272 total tree bytes: 2336227328 total fs tree bytes: 1766473728 total extent tree bytes: 202014720 btree space waste bytes: 519245278 file data blocks allocated: 6851730792448 referenced 533348069376 Thanks, Stephan > > > Some information about my system: > >
Re: bad tree block start, want 705757184 have 82362368
On 2018/11/17 上午12:17, Stephan Olbrich wrote: > Hi, > > a few days ago my root file system (simple btrfs on a SSD, no RAID or > anything) suddenly became read only. > Looking at dmsg, I found this: > > [ 19.285020] BTRFS error (device sda2): bad tree block start, want > 705757184 have 82362368 Does this only occurs once? If only once, are you using SINGLE metadata profile (default for SSD)? If using DUP/RAID1 it may have a chance to recover. > [ 19.285042] BTRFS: error (device sda2) in __btrfs_free_extent:6804: > errno=-5 IO failure The problem is in extent tree. If there is no other problem, your data should be OK. You could still try to mount the fs RO to salvage data. > [ 19.285048] BTRFS info (device sda2): forced readonly > [ 19.285051] BTRFS: error (device sda2) in btrfs_run_delayed_refs:2934: > errno=-5 IO failure > [ 19.287213] BTRFS error (device sda2): pending csums is 41889792 > > Late on I got the same errors for my /home partition (on the same drive) as > well. > I have snapshots of all partitions on another drive made by btrbk. To get a > working system, I made new (rw) snapshots > of the most recent backup and setup grub and fstab, so my system would boot > from the other drive. > Unfortunately now I got the "bad tree block start" error again at least once > in dmesg but I didn't save it and it's not in syslog :-( > What I remember is, that it was followed by other btrfs error messages saying > something about correcting something. > And the filesystem was still read/write this time. > At the moment I can't reproduce it. > > Is there any way to find out, which files are affected by the errors above? No files are affected, but an essential tree, extent tree, is corrupted. Normally this may prevent RW mount, and even it mounts it can still cause problem when doing any write. It could even prevent RO mount if the corrupted leaf contains block group item. But your data should be OK if there is no other corruption, and in that case btrfs-restore should work well. > I don't really trust the data on the drive I'm using at the > moment, as it has shown errors as well, but I have a less current backup on > yet another drive but at it is a few weeks old, I don't > want to use it to setup the system on the SSD again, but just copy the > relevant files if possible. > Or is it possible to repair the original file system? At least we need "btrfs check" output. > > Some information about my system: > Kubuntu 18.04 > Kernel 4.19.1 when the problem occured, now 4.19.2 > btrfs-tools 4.15.1 And "btrfs check" should be executed using latest version. Thanks, Qu > > Regards, > Stephan > > > > > signature.asc Description: OpenPGP digital signature
Re: bad tree block start, want 705757184 have 82362368
On 16.11.18 г. 18:17 ч., Stephan Olbrich wrote: > Hi, > > a few days ago my root file system (simple btrfs on a SSD, no RAID or > anything) suddenly became read only. > Looking at dmsg, I found this: > > [ 19.285020] BTRFS error (device sda2): bad tree block start, want > 705757184 have 82362368 > [ 19.285042] BTRFS: error (device sda2) in __btrfs_free_extent:6804: > errno=-5 IO failure > [ 19.285048] BTRFS info (device sda2): forced readonly > [ 19.285051] BTRFS: error (device sda2) in btrfs_run_delayed_refs:2934: > errno=-5 IO failure > [ 19.287213] BTRFS error (device sda2): pending csums is 41889792 > > Late on I got the same errors for my /home partition (on the same drive) as > well. > I have snapshots of all partitions on another drive made by btrbk. To get a > working system, I made new (rw) snapshots > of the most recent backup and setup grub and fstab, so my system would boot > from the other drive. > Unfortunately now I got the "bad tree block start" error again at least once > in dmesg but I didn't save it and it's not in syslog :-( > What I remember is, that it was followed by other btrfs error messages saying > something about correcting something. > And the filesystem was still read/write this time. > At the moment I can't reproduce it. > > Is there any way to find out, which files are affected by the errors above? I > don't really trust the data on the drive I'm using at the > moment, as it has shown errors as well, but I have a less current backup on > yet another drive but at it is a few weeks old, I don't > want to use it to setup the system on the SSD again, but just copy the > relevant files if possible. > Or is it possible to repair the original file system? > > Some information about my system: > Kubuntu 18.04 > Kernel 4.19.1 when the problem occured, now 4.19.2 > btrfs-tools 4.15.1 What is the SMART status of your SSD, how old is the ssd. This really sounds like the drive going to lalal land. > > Regards, > Stephan > > > > > >