Hello, a trace of the kworker looks like this: kworker/u24:4-13405 [003] .... 344186.202535: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202535: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202535: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202535: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202535: btrfs_find_space_for_alloc <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202535: _raw_spin_lock <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202536: tree_search_offset.isra.25 <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202554: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202554: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202554: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202554: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202555: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202555: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202555: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202555: btrfs_find_space_for_alloc <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202555: _raw_spin_lock <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202556: tree_search_offset.isra.25 <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202560: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202560: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202561: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202561: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202561: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202561: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202561: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202562: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202562: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202562: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202562: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202562: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202563: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202563: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202563: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202563: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202563: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202563: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202564: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202564: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202564: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202564: btrfs_find_space_for_alloc <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202564: _raw_spin_lock <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202565: tree_search_offset.isra.25 <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202566: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202567: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202567: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202567: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202567: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202568: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202568: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202568: btrfs_find_space_for_alloc <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202568: _raw_spin_lock <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202569: tree_search_offset.isra.25 <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202576: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202576: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202577: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202577: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202577: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202577: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202578: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202578: btrfs_find_space_for_alloc <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202578: _raw_spin_lock <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202578: tree_search_offset.isra.25 <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202584: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202584: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202584: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202585: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202585: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202585: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202585: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202586: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202586: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202586: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202586: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202586: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202586: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202587: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202587: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202587: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202587: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202588: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202588: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202588: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202589: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202589: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202589: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202589: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202589: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202589: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202590: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202590: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202590: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202590: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202590: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202591: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202591: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202591: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202591: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202591: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202592: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202592: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202592: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202592: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202592: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202592: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202593: btrfs_find_space_for_alloc <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202593: _raw_spin_lock <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202593: tree_search_offset.isra.25 <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202597: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202597: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202597: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202598: _cond_resched <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202598: down_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202598: block_group_cache_done.isra.27 <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202598: _raw_spin_lock <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202598: btrfs_find_space_for_alloc <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202598: _raw_spin_lock <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202599: tree_search_offset.isra.25 <-btrfs_find_space_for_alloc kworker/u24:4-13405 [003] .... 344186.202623: __get_raid_index <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202623: up_read <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202623: btrfs_put_block_group <-find_free_extent kworker/u24:4-13405 [003] .... 344186.202623: _cond_resched <-find_free_extent
Greets, Stefan Am 20.08.2017 um 13:00 schrieb Stefan Priebe - Profihost AG: > Hello, > > this still happens with space_cache v2. I don't think it is space_cache > related? > > Stefan > > Am 17.08.2017 um 09:43 schrieb Stefan Priebe - Profihost AG: >> while mounting the device the dmesg is full of: >> [ 1320.325147] [<ffffffffa80a37b0>] ? kthread_park+0x60/0x60 >> [ 1440.330008] INFO: task btrfs-transacti:3701 blocked for more than 120 >> seconds. >> [ 1440.330014] Not tainted 4.4.82+525-ph #1 >> [ 1440.330015] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >> disables this message. >> [ 1440.330020] btrfs-transacti D ffff88080964fdd8 0 3701 2 >> 0x00080000 >> [ 1440.330024] ffff88080964fdd8 ffffffffa8e10500 ffff880859d4cb00 >> ffff880809650000 >> [ 1440.330026] ffff881056069800 ffff88080964fe08 ffff88080a100000 >> ffff88080a100068 >> [ 1440.330028] ffff88080964fdf0 ffffffffa86d2b75 ffff880036c92000 >> ffff88080964fe58 >> [ 1440.330028] Call Trace: >> [ 1440.330053] [<ffffffffa86d2b75>] schedule+0x35/0x80 >> [ 1440.330120] [<ffffffffc04cfa55>] >> btrfs_commit_transaction.part.24+0x245/0xa30 [btrfs] >> [ 1440.330159] [<ffffffffc04d027a>] btrfs_commit_transaction+0x3a/0x70 >> [btrfs] >> [ 1440.330186] [<ffffffffc04ca3c5>] transaction_kthread+0x1d5/0x240 [btrfs] >> [ 1440.330194] [<ffffffffa80a389b>] kthread+0xeb/0x110 >> [ 1440.330200] [<ffffffffa86d750f>] ret_from_fork+0x3f/0x70 >> [ 1440.333316] DWARF2 unwinder stuck at ret_from_fork+0x3f/0x70 >> >> [ 1440.333317] Leftover inexact backtrace: >> >> [ 1440.333322] [<ffffffffa80a37b0>] ? kthread_park+0x60/0x60 >> [ 1560.335839] INFO: task btrfs-transacti:3701 blocked for more than 120 >> seconds. >> [ 1560.335843] Not tainted 4.4.82+525-ph #1 >> [ 1560.335843] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >> disables this message. >> [ 1560.335848] btrfs-transacti D ffff88080964fdd8 0 3701 2 >> 0x00080000 >> [ 1560.335852] ffff88080964fdd8 ffffffffa8e10500 ffff880859d4cb00 >> ffff880809650000 >> [ 1560.335854] ffff881056069800 ffff88080964fe08 ffff88080a100000 >> ffff88080a100068 >> [ 1560.335856] ffff88080964fdf0 ffffffffa86d2b75 ffff880036c92000 >> ffff88080964fe58 >> [ 1560.335857] Call Trace: >> [ 1560.335875] [<ffffffffa86d2b75>] schedule+0x35/0x80 >> [ 1560.335953] [<ffffffffc04cfa55>] >> btrfs_commit_transaction.part.24+0x245/0xa30 [btrfs] >> [ 1560.335978] [<ffffffffc04d027a>] btrfs_commit_transaction+0x3a/0x70 >> [btrfs] >> [ 1560.335995] [<ffffffffc04ca3c5>] transaction_kthread+0x1d5/0x240 [btrfs] >> [ 1560.336001] [<ffffffffa80a389b>] kthread+0xeb/0x110 >> [ 1560.336006] [<ffffffffa86d750f>] ret_from_fork+0x3f/0x70 >> [ 1560.337829] DWARF2 unwinder stuck at ret_from_fork+0x3f/0x70 >> >> [ 1560.337830] Leftover inexact backtrace: >> >> [ 1560.337833] [<ffffffffa80a37b0>] ? kthread_park+0x60/0x60 >> [ 1680.341127] INFO: task btrfs-transacti:3701 blocked for more than 120 >> seconds. >> [ 1680.341130] Not tainted 4.4.82+525-ph #1 >> [ 1680.341131] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >> disables this message. >> [ 1680.341134] btrfs-transacti D ffff88080964fdd8 0 3701 2 >> 0x00080000 >> [ 1680.341137] ffff88080964fdd8 ffffffffa8e10500 ffff880859d4cb00 >> ffff880809650000 >> [ 1680.341138] ffff881056069800 ffff88080964fe08 ffff88080a100000 >> ffff88080a100068 >> [ 1680.341139] ffff88080964fdf0 ffffffffa86d2b75 ffff880036c92000 >> ffff88080964fe58 >> [ 1680.341140] Call Trace: >> [ 1680.341155] [<ffffffffa86d2b75>] schedule+0x35/0x80 >> [ 1680.341211] [<ffffffffc04cfa55>] >> btrfs_commit_transaction.part.24+0x245/0xa30 [btrfs] >> [ 1680.341237] [<ffffffffc04d027a>] btrfs_commit_transaction+0x3a/0x70 >> [btrfs] >> [ 1680.341252] [<ffffffffc04ca3c5>] transaction_kthread+0x1d5/0x240 [btrfs] >> [ 1680.341258] [<ffffffffa80a389b>] kthread+0xeb/0x110 >> [ 1680.341262] [<ffffffffa86d750f>] ret_from_fork+0x3f/0x70 >> [ 1680.343062] DWARF2 unwinder stuck at ret_from_fork+0x3f/0x70 >> >> Stefan >> >> Am 17.08.2017 um 07:47 schrieb Stefan Priebe - Profihost AG: >>> i've backported the free space cache tree to my kerne and hopefully any >>> fixes related to it. >>> >>> The first mount with clear_cache,space_cache=v2 took around 5 hours. >>> >>> Currently i do not see any kworker with 100CPU but i don't see much load >>> at all. >>> >>> btrfs-transaction tooks around 2-4% CPU together with a kworker process >>> and some 2-3% mdadm processes. I/O Wait is at 3%. >>> >>> That's it. It does not do much more. Writing a file does not work. >>> >>> Greets, >>> Stefan >>> >>> Am 16.08.2017 um 14:29 schrieb Konstantin V. Gavrilenko: >>>> Roman, initially I had a single process occupying 100% CPU, when sysrq it >>>> was indicating as "btrfs_find_space_for_alloc" >>>> but that's when I used the autodefrag, compress, forcecompress and >>>> commit=10 mount flags and space_cache was v1 by default. >>>> when I switched to "relatime,compress-force=zlib,space_cache=v2" the 100% >>>> cpu has dissapeared, but the shite performance remained. >>>> >>>> >>>> As to the chunk size, there is no information in the article about the >>>> type of data that was used. While in our case we are pretty certain about >>>> the compressed block size (32-128). I am currently inclining towards 32k >>>> as it might be ideal in a situation when we have a 5 disk raid5 array. >>>> >>>> In theory >>>> 1. The minimum compressed write (32k) would fill the chunk on a single >>>> disk, thus the IO cost of the operation would be 2 reads (original chunk + >>>> original parity) and 2 writes (new chunk + new parity) >>>> >>>> 2. The maximum compressed write (128k) would require the update of 1 chunk >>>> on each of the 4 data disks + 1 parity write >>>> >>>> >>>> >>>> Stefan what mount flags do you use? >>>> >>>> kos >>>> >>>> >>>> >>>> ----- Original Message ----- >>>> From: "Roman Mamedov" <r...@romanrm.net> >>>> To: "Konstantin V. Gavrilenko" <k.gavrile...@arhont.com> >>>> Cc: "Stefan Priebe - Profihost AG" <s.pri...@profihost.ag>, "Marat >>>> Khalili" <m...@rqc.ru>, linux-btrfs@vger.kernel.org, "Peter Grandi" >>>> <p...@btrfs.list.sabi.co.uk> >>>> Sent: Wednesday, 16 August, 2017 2:00:03 PM >>>> Subject: Re: slow btrfs with a single kworker process using 100% CPU >>>> >>>> On Wed, 16 Aug 2017 12:48:42 +0100 (BST) >>>> "Konstantin V. Gavrilenko" <k.gavrile...@arhont.com> wrote: >>>> >>>>> I believe the chunk size of 512kb is even worth for performance then the >>>>> default settings on my HW RAID of 256kb. >>>> >>>> It might be, but that does not explain the original problem reported at >>>> all. >>>> If mdraid performance would be the bottleneck, you would see high iowait, >>>> possibly some CPU load from the mdX_raidY threads. But not a single Btrfs >>>> thread pegging into 100% CPU. >>>> >>>>> So now I am moving the data from the array and will be rebuilding it with >>>>> 64 >>>>> or 32 chunk size and checking the performance. >>>> >>>> 64K is the sweet spot for RAID5/6: >>>> http://louwrentius.com/linux-raid-level-and-chunk-size-the-benchmarks.html >>>> -- To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html