Hi Any ideas on this?
Regards, Tobias 2015-07-20 18:20 GMT+02:00 Tobias Holst <[email protected]>: > Hi > > My btrfs-RAID6 seems to be broken again :( > > When reading from it I get several of these: > [ 176.349943] BTRFS info (device dm-4): csum failed ino 1287707 > extent 21274957705216 csum 2830458701 wanted 426660650 mirror 2 > > then followed by a "free_raid_bio"-crash: > > [ 176.349961] ------------[ cut here ]------------ > [ 176.349981] WARNING: CPU: 6 PID: 110 at > /home/kernel/COD/linux/fs/btrfs/raid56.c:831 > __free_raid_bio+0xfc/0x130 [btrfs]() > [ 176.349982] Modules linked in: iosf_mbi kvm_intel kvm ppdev > crct10dif_pclmul crc32_pclmul dm_crypt ghash_clmulni_intel aesni_intel > aes_x86_64 lrw gf128mul glue_helper ablk_helper serio_raw 8250_fintek > i2c_piix4 pvpanic cryptd mac_hid virtio_rng parport_pc lp parport > btrfs xor raid6_pq cirrus syscopyarea sysfillrect sysimgblt ttm > drm_kms_helper mpt2sas drm raid_class psmouse floppy > scsi_transport_sas pata_acpi > [ 176.349998] CPU: 6 PID: 110 Comm: kworker/u16:2 Not tainted > 4.1.2-040102-generic #201507101335 > [ 176.349999] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), > BIOS Bochs 01/01/2011 > [ 176.350007] Workqueue: btrfs-endio btrfs_endio_helper [btrfs] > [ 176.350008] ffffffffc026fc18 ffff8800baa4f978 ffffffff817d076c > 0000000000000000 > [ 176.350010] 0000000000000000 ffff8800baa4f9b8 ffffffff81079b0a > 0000000000000246 > [ 176.350011] ffff88034e7baa68 ffff88008619b800 00000000fffffffb > 0000000000000000 > [ 176.350013] Call Trace: > [ 176.350023] [<ffffffff817d076c>] dump_stack+0x45/0x57 > [ 176.350026] [<ffffffff81079b0a>] warn_slowpath_common+0x8a/0xc0 > [ 176.350029] [<ffffffff81079bfa>] warn_slowpath_null+0x1a/0x20 > [ 176.350036] [<ffffffffc025e91c>] __free_raid_bio+0xfc/0x130 [btrfs] > [ 176.350041] [<ffffffffc025f351>] rbio_orig_end_io+0x51/0xa0 [btrfs] > [ 176.350047] [<ffffffffc02610e3>] __raid56_parity_recover+0x1d3/0x210 > [btrfs] > [ 176.350052] [<ffffffffc0261cb0>] raid56_parity_recover+0x110/0x180 [btrfs] > [ 176.350058] [<ffffffffc0216cdb>] btrfs_map_bio+0xdb/0x4e0 [btrfs] > [ 176.350065] [<ffffffffc0236024>] > btrfs_submit_compressed_read+0x354/0x4e0 [btrfs] > [ 176.350070] [<ffffffffc01ee681>] btrfs_submit_bio_hook+0x1d1/0x1e0 [btrfs] > [ 176.350076] [<ffffffff81376dbe>] ? bio_add_page+0x5e/0x70 > [ 176.350083] [<ffffffffc020c176>] ? > btrfs_create_repair_bio+0xe6/0x110 [btrfs] > [ 176.350089] [<ffffffffc020c6ab>] end_bio_extent_readpage+0x50b/0x560 > [btrfs] > [ 176.350094] [<ffffffffc020c1a0>] ? > btrfs_create_repair_bio+0x110/0x110 [btrfs] > [ 176.350096] [<ffffffff8137934b>] bio_endio+0x5b/0xa0 > [ 176.350103] [<ffffffff811d9e19>] ? kmem_cache_free+0x1d9/0x1f0 > [ 176.350104] [<ffffffff813793a2>] bio_endio_nodec+0x12/0x20 > [ 176.350109] [<ffffffffc01e10df>] end_workqueue_fn+0x3f/0x50 [btrfs] > [ 176.350115] [<ffffffffc021b522>] normal_work_helper+0xc2/0x2b0 [btrfs] > [ 176.350121] [<ffffffffc021b7e2>] btrfs_endio_helper+0x12/0x20 [btrfs] > [ 176.350124] [<ffffffff8109324f>] process_one_work+0x14f/0x420 > [ 176.350127] [<ffffffff81093a08>] worker_thread+0x118/0x530 > [ 176.350128] [<ffffffff810938f0>] ? rescuer_thread+0x3d0/0x3d0 > [ 176.350129] [<ffffffff81098f89>] kthread+0xc9/0xe0 > [ 176.350130] [<ffffffff81098ec0>] ? kthread_create_on_node+0x180/0x180 > [ 176.350134] [<ffffffff817d86a2>] ret_from_fork+0x42/0x70 > [ 176.350135] [<ffffffff81098ec0>] ? kthread_create_on_node+0x180/0x180 > [ 176.350136] ---[ end trace 81289955f20d48ee ]--- > > Did I found a kernel bug? What can/should I do? > > Don't worry about my data, I have tape-backups of the important data, > I just want to help fixing RAID-related btrfs bugs. > > Hardware: KVM with all drives attached to a passed through SAS-controller > System: Ubuntu 14.04.2 > Kernel: 4.1.2 > btrfs-tools: 4.0 > It's a btrfs-RAID-6 on top of 6 LUKS-encrypted volumes, created with > "-O extref,raid56,skinny-metadata,no-holes". At normal it's mounted > with "defaults,compress=lzo,space_cache,autodefrag,subvol=raid". > One drive is broken, so at the moment it is mounted with "-O > defaults,ro,degraded,recovery,compress=lzo,space_cache,subvol=raid". > > It's pretty much full, so "btrfs fi show" shows: > Label: 't-raid' uuid: 3938baeb-cb02-4909-8e75-6ec2f47d1d19 > Total devices 6 FS bytes used 14.44TiB > devid 2 size 3.64TiB used 3.64TiB path /dev/mapper/sdb_crypt > devid 3 size 3.64TiB used 3.64TiB path /dev/mapper/sdc_crypt > devid 4 size 3.64TiB used 3.64TiB path /dev/mapper/sdd_crypt > devid 5 size 3.64TiB used 3.64TiB path /dev/mapper/sde_crypt > devid 6 size 3.64TiB used 3.64TiB path /dev/mapper/sdf_crypt > *** Some devices missing > > and "btrfs fi df /raid" shows: > Data, RAID6: total=14.52TiB, used=14.42TiB > System, RAID6: total=64.00MiB, used=1.00MiB > Metadata, RAID6: total=24.00GiB, used=21.78GiB > GlobalReserve, single: total=512.00MiB, used=0.00B > > > Regards, > Tobias -- To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in the body of a message to [email protected] More majordomo info at http://vger.kernel.org/majordomo-info.html
