Hi
My btrfs-RAID6 seems to be broken again :(
When reading from it I get several of these:
[ 176.349943] BTRFS info (device dm-4): csum failed ino 1287707
extent 21274957705216 csum 2830458701 wanted 426660650 mirror 2
then followed by a "free_raid_bio"-crash:
[ 176.349961] ------------[ cut here ]------------
[ 176.349981] WARNING: CPU: 6 PID: 110 at
/home/kernel/COD/linux/fs/btrfs/raid56.c:831
__free_raid_bio+0xfc/0x130 [btrfs]()
[ 176.349982] Modules linked in: iosf_mbi kvm_intel kvm ppdev
crct10dif_pclmul crc32_pclmul dm_crypt ghash_clmulni_intel aesni_intel
aes_x86_64 lrw gf128mul glue_helper ablk_helper serio_raw 8250_fintek
i2c_piix4 pvpanic cryptd mac_hid virtio_rng parport_pc lp parport
btrfs xor raid6_pq cirrus syscopyarea sysfillrect sysimgblt ttm
drm_kms_helper mpt2sas drm raid_class psmouse floppy
scsi_transport_sas pata_acpi
[ 176.349998] CPU: 6 PID: 110 Comm: kworker/u16:2 Not tainted
4.1.2-040102-generic #201507101335
[ 176.349999] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996),
BIOS Bochs 01/01/2011
[ 176.350007] Workqueue: btrfs-endio btrfs_endio_helper [btrfs]
[ 176.350008] ffffffffc026fc18 ffff8800baa4f978 ffffffff817d076c
0000000000000000
[ 176.350010] 0000000000000000 ffff8800baa4f9b8 ffffffff81079b0a
0000000000000246
[ 176.350011] ffff88034e7baa68 ffff88008619b800 00000000fffffffb
0000000000000000
[ 176.350013] Call Trace:
[ 176.350023] [<ffffffff817d076c>] dump_stack+0x45/0x57
[ 176.350026] [<ffffffff81079b0a>] warn_slowpath_common+0x8a/0xc0
[ 176.350029] [<ffffffff81079bfa>] warn_slowpath_null+0x1a/0x20
[ 176.350036] [<ffffffffc025e91c>] __free_raid_bio+0xfc/0x130 [btrfs]
[ 176.350041] [<ffffffffc025f351>] rbio_orig_end_io+0x51/0xa0 [btrfs]
[ 176.350047] [<ffffffffc02610e3>] __raid56_parity_recover+0x1d3/0x210 [btrfs]
[ 176.350052] [<ffffffffc0261cb0>] raid56_parity_recover+0x110/0x180 [btrfs]
[ 176.350058] [<ffffffffc0216cdb>] btrfs_map_bio+0xdb/0x4e0 [btrfs]
[ 176.350065] [<ffffffffc0236024>]
btrfs_submit_compressed_read+0x354/0x4e0 [btrfs]
[ 176.350070] [<ffffffffc01ee681>] btrfs_submit_bio_hook+0x1d1/0x1e0 [btrfs]
[ 176.350076] [<ffffffff81376dbe>] ? bio_add_page+0x5e/0x70
[ 176.350083] [<ffffffffc020c176>] ?
btrfs_create_repair_bio+0xe6/0x110 [btrfs]
[ 176.350089] [<ffffffffc020c6ab>] end_bio_extent_readpage+0x50b/0x560 [btrfs]
[ 176.350094] [<ffffffffc020c1a0>] ?
btrfs_create_repair_bio+0x110/0x110 [btrfs]
[ 176.350096] [<ffffffff8137934b>] bio_endio+0x5b/0xa0
[ 176.350103] [<ffffffff811d9e19>] ? kmem_cache_free+0x1d9/0x1f0
[ 176.350104] [<ffffffff813793a2>] bio_endio_nodec+0x12/0x20
[ 176.350109] [<ffffffffc01e10df>] end_workqueue_fn+0x3f/0x50 [btrfs]
[ 176.350115] [<ffffffffc021b522>] normal_work_helper+0xc2/0x2b0 [btrfs]
[ 176.350121] [<ffffffffc021b7e2>] btrfs_endio_helper+0x12/0x20 [btrfs]
[ 176.350124] [<ffffffff8109324f>] process_one_work+0x14f/0x420
[ 176.350127] [<ffffffff81093a08>] worker_thread+0x118/0x530
[ 176.350128] [<ffffffff810938f0>] ? rescuer_thread+0x3d0/0x3d0
[ 176.350129] [<ffffffff81098f89>] kthread+0xc9/0xe0
[ 176.350130] [<ffffffff81098ec0>] ? kthread_create_on_node+0x180/0x180
[ 176.350134] [<ffffffff817d86a2>] ret_from_fork+0x42/0x70
[ 176.350135] [<ffffffff81098ec0>] ? kthread_create_on_node+0x180/0x180
[ 176.350136] ---[ end trace 81289955f20d48ee ]---
Did I found a kernel bug? What can/should I do?
Don't worry about my data, I have tape-backups of the important data,
I just want to help fixing RAID-related btrfs bugs.
Hardware: KVM with all drives attached to a passed through SAS-controller
System: Ubuntu 14.04.2
Kernel: 4.1.2
btrfs-tools: 4.0
It's a btrfs-RAID-6 on top of 6 LUKS-encrypted volumes, created with
"-O extref,raid56,skinny-metadata,no-holes". At normal it's mounted
with "defaults,compress=lzo,space_cache,autodefrag,subvol=raid".
One drive is broken, so at the moment it is mounted with "-O
defaults,ro,degraded,recovery,compress=lzo,space_cache,subvol=raid".
It's pretty much full, so "btrfs fi show" shows:
Label: 't-raid' uuid: 3938baeb-cb02-4909-8e75-6ec2f47d1d19
Total devices 6 FS bytes used 14.44TiB
devid 2 size 3.64TiB used 3.64TiB path /dev/mapper/sdb_crypt
devid 3 size 3.64TiB used 3.64TiB path /dev/mapper/sdc_crypt
devid 4 size 3.64TiB used 3.64TiB path /dev/mapper/sdd_crypt
devid 5 size 3.64TiB used 3.64TiB path /dev/mapper/sde_crypt
devid 6 size 3.64TiB used 3.64TiB path /dev/mapper/sdf_crypt
*** Some devices missing
and "btrfs fi df /raid" shows:
Data, RAID6: total=14.52TiB, used=14.42TiB
System, RAID6: total=64.00MiB, used=1.00MiB
Metadata, RAID6: total=24.00GiB, used=21.78GiB
GlobalReserve, single: total=512.00MiB, used=0.00B
Regards,
Tobias
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html