Hi

My btrfs-RAID6 seems to be broken again :(

When reading from it I get several of these:
[  176.349943] BTRFS info (device dm-4): csum failed ino 1287707
extent 21274957705216 csum 2830458701 wanted 426660650 mirror 2

then followed by a "free_raid_bio"-crash:

[  176.349961] ------------[ cut here ]------------
[  176.349981] WARNING: CPU: 6 PID: 110 at
/home/kernel/COD/linux/fs/btrfs/raid56.c:831
__free_raid_bio+0xfc/0x130 [btrfs]()
[  176.349982] Modules linked in: iosf_mbi kvm_intel kvm ppdev
crct10dif_pclmul crc32_pclmul dm_crypt ghash_clmulni_intel aesni_intel
aes_x86_64 lrw gf128mul glue_helper ablk_helper serio_raw 8250_fintek
i2c_piix4 pvpanic cryptd mac_hid virtio_rng parport_pc lp parport
btrfs xor raid6_pq cirrus syscopyarea sysfillrect sysimgblt ttm
drm_kms_helper mpt2sas drm raid_class psmouse floppy
scsi_transport_sas pata_acpi
[  176.349998] CPU: 6 PID: 110 Comm: kworker/u16:2 Not tainted
4.1.2-040102-generic #201507101335
[  176.349999] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996),
BIOS Bochs 01/01/2011
[  176.350007] Workqueue: btrfs-endio btrfs_endio_helper [btrfs]
[  176.350008]  ffffffffc026fc18 ffff8800baa4f978 ffffffff817d076c
0000000000000000
[  176.350010]  0000000000000000 ffff8800baa4f9b8 ffffffff81079b0a
0000000000000246
[  176.350011]  ffff88034e7baa68 ffff88008619b800 00000000fffffffb
0000000000000000
[  176.350013] Call Trace:
[  176.350023]  [<ffffffff817d076c>] dump_stack+0x45/0x57
[  176.350026]  [<ffffffff81079b0a>] warn_slowpath_common+0x8a/0xc0
[  176.350029]  [<ffffffff81079bfa>] warn_slowpath_null+0x1a/0x20
[  176.350036]  [<ffffffffc025e91c>] __free_raid_bio+0xfc/0x130 [btrfs]
[  176.350041]  [<ffffffffc025f351>] rbio_orig_end_io+0x51/0xa0 [btrfs]
[  176.350047]  [<ffffffffc02610e3>] __raid56_parity_recover+0x1d3/0x210 [btrfs]
[  176.350052]  [<ffffffffc0261cb0>] raid56_parity_recover+0x110/0x180 [btrfs]
[  176.350058]  [<ffffffffc0216cdb>] btrfs_map_bio+0xdb/0x4e0 [btrfs]
[  176.350065]  [<ffffffffc0236024>]
btrfs_submit_compressed_read+0x354/0x4e0 [btrfs]
[  176.350070]  [<ffffffffc01ee681>] btrfs_submit_bio_hook+0x1d1/0x1e0 [btrfs]
[  176.350076]  [<ffffffff81376dbe>] ? bio_add_page+0x5e/0x70
[  176.350083]  [<ffffffffc020c176>] ?
btrfs_create_repair_bio+0xe6/0x110 [btrfs]
[  176.350089]  [<ffffffffc020c6ab>] end_bio_extent_readpage+0x50b/0x560 [btrfs]
[  176.350094]  [<ffffffffc020c1a0>] ?
btrfs_create_repair_bio+0x110/0x110 [btrfs]
[  176.350096]  [<ffffffff8137934b>] bio_endio+0x5b/0xa0
[  176.350103]  [<ffffffff811d9e19>] ? kmem_cache_free+0x1d9/0x1f0
[  176.350104]  [<ffffffff813793a2>] bio_endio_nodec+0x12/0x20
[  176.350109]  [<ffffffffc01e10df>] end_workqueue_fn+0x3f/0x50 [btrfs]
[  176.350115]  [<ffffffffc021b522>] normal_work_helper+0xc2/0x2b0 [btrfs]
[  176.350121]  [<ffffffffc021b7e2>] btrfs_endio_helper+0x12/0x20 [btrfs]
[  176.350124]  [<ffffffff8109324f>] process_one_work+0x14f/0x420
[  176.350127]  [<ffffffff81093a08>] worker_thread+0x118/0x530
[  176.350128]  [<ffffffff810938f0>] ? rescuer_thread+0x3d0/0x3d0
[  176.350129]  [<ffffffff81098f89>] kthread+0xc9/0xe0
[  176.350130]  [<ffffffff81098ec0>] ? kthread_create_on_node+0x180/0x180
[  176.350134]  [<ffffffff817d86a2>] ret_from_fork+0x42/0x70
[  176.350135]  [<ffffffff81098ec0>] ? kthread_create_on_node+0x180/0x180
[  176.350136] ---[ end trace 81289955f20d48ee ]---

Did I found a kernel bug? What can/should I do?

Don't worry about my data, I have tape-backups of the important data,
I just want to help fixing RAID-related btrfs bugs.

Hardware: KVM with all drives attached to a passed through SAS-controller
System: Ubuntu 14.04.2
Kernel: 4.1.2
btrfs-tools: 4.0
It's a btrfs-RAID-6 on top of 6 LUKS-encrypted volumes, created with
"-O extref,raid56,skinny-metadata,no-holes". At normal it's mounted
with "defaults,compress=lzo,space_cache,autodefrag,subvol=raid".
One drive is broken, so at the moment it is mounted with "-O
defaults,ro,degraded,recovery,compress=lzo,space_cache,subvol=raid".

It's pretty much full, so "btrfs fi show" shows:
Label: 't-raid'  uuid: 3938baeb-cb02-4909-8e75-6ec2f47d1d19
        Total devices 6 FS bytes used 14.44TiB
        devid    2 size 3.64TiB used 3.64TiB path /dev/mapper/sdb_crypt
        devid    3 size 3.64TiB used 3.64TiB path /dev/mapper/sdc_crypt
        devid    4 size 3.64TiB used 3.64TiB path /dev/mapper/sdd_crypt
        devid    5 size 3.64TiB used 3.64TiB path /dev/mapper/sde_crypt
        devid    6 size 3.64TiB used 3.64TiB path /dev/mapper/sdf_crypt
        *** Some devices missing

and "btrfs fi df /raid" shows:
Data, RAID6: total=14.52TiB, used=14.42TiB
System, RAID6: total=64.00MiB, used=1.00MiB
Metadata, RAID6: total=24.00GiB, used=21.78GiB
GlobalReserve, single: total=512.00MiB, used=0.00B


Regards,
Tobias
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to