On Fri, Apr 06, 2018 at 12:05:03AM +0800, Ming Lei wrote:
> On Wed, Apr 04, 2018 at 10:18:13AM +0200, Christian Borntraeger wrote:
> >
> >
> > On 03/30/2018 04:53 AM, Ming Lei wrote:
> > > On Thu, Mar 29, 2018 at 01:49:29PM +0200, Christian Borntraeger wrote:
> > >>
> > >>
> > >> On 03/29/2018 01:43 PM, Ming Lei wrote:
> > >>> On Thu, Mar 29, 2018 at 12:49:55PM +0200, Christian Borntraeger wrote:
> > >>>>
> > >>>>
> > >>>> On 03/29/2018 12:48 PM, Ming Lei wrote:
> > >>>>> On Thu, Mar 29, 2018 at 12:10:11PM +0200, Christian Borntraeger wrote:
> > >>>>>>
> > >>>>>>
> > >>>>>> On 03/29/2018 11:40 AM, Ming Lei wrote:
> > >>>>>>> On Thu, Mar 29, 2018 at 11:09:08AM +0200, Christian Borntraeger
> > >>>>>>> wrote:
> > >>>>>>>>
> > >>>>>>>>
> > >>>>>>>> On 03/29/2018 09:23 AM, Christian Borntraeger wrote:
> > >>>>>>>>>
> > >>>>>>>>>
> > >>>>>>>>> On 03/29/2018 04:00 AM, Ming Lei wrote:
> > >>>>>>>>>> On Wed, Mar 28, 2018 at 05:36:53PM +0200, Christian Borntraeger
> > >>>>>>>>>> wrote:
> > >>>>>>>>>>>
> > >>>>>>>>>>>
> > >>>>>>>>>>> On 03/28/2018 05:26 PM, Ming Lei wrote:
> > >>>>>>>>>>>> Hi Christian,
> > >>>>>>>>>>>>
> > >>>>>>>>>>>> On Wed, Mar 28, 2018 at 09:45:10AM +0200, Christian
> > >>>>>>>>>>>> Borntraeger wrote:
> > >>>>>>>>>>>>> FWIW, this patch does not fix the issue for me:
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> ostname=? addr=? terminal=? res=success'
> > >>>>>>>>>>>>> [ 21.454961] WARNING: CPU: 3 PID: 1882 at
> > >>>>>>>>>>>>> block/blk-mq.c:1410 __blk_mq_delay_run_hw_queue+0xbe/0xd8
> > >>>>>>>>>>>>> [ 21.454968] Modules linked in: scsi_dh_rdac scsi_dh_emc
> > >>>>>>>>>>>>> scsi_dh_alua dm_mirror dm_region_hash dm_log dm_multipath
> > >>>>>>>>>>>>> dm_mod autofs4
> > >>>>>>>>>>>>> [ 21.454984] CPU: 3 PID: 1882 Comm: dasdconf.sh Not tainted
> > >>>>>>>>>>>>> 4.16.0-rc7+ #26
> > >>>>>>>>>>>>> [ 21.454987] Hardware name: IBM 2964 NC9 704 (LPAR)
> > >>>>>>>>>>>>> [ 21.454990] Krnl PSW : 00000000c0131ea3 000000003ea2f7bf
> > >>>>>>>>>>>>> (__blk_mq_delay_run_hw_queue+0xbe/0xd8)
> > >>>>>>>>>>>>> [ 21.454996] R:0 T:1 IO:0 EX:0 Key:0 M:1 W:0 P:0
> > >>>>>>>>>>>>> AS:3 CC:0 PM:0 RI:0 EA:3
> > >>>>>>>>>>>>> [ 21.455005] Krnl GPRS: 0000013abb69a000 0000013a00000000
> > >>>>>>>>>>>>> 0000013ac6c0dc00 0000000000000001
> > >>>>>>>>>>>>> [ 21.455008] 0000000000000000 0000013abb69a710
> > >>>>>>>>>>>>> 0000013a00000000 00000001b691fd98
> > >>>>>>>>>>>>> [ 21.455011] 00000001b691fd98 0000013ace4775c8
> > >>>>>>>>>>>>> 0000000000000001 0000000000000000
> > >>>>>>>>>>>>> [ 21.455014] 0000013ac6c0dc00 0000000000b47238
> > >>>>>>>>>>>>> 00000001b691fc08 00000001b691fbd0
> > >>>>>>>>>>>>> [ 21.455032] Krnl Code: 000000000069c596: ebaff0a00004
> > >>>>>>>>>>>>> lmg %r10,%r15,160(%r15)
> > >>>>>>>>>>>>> 000000000069c59c: c0f4ffff7a5e
> > >>>>>>>>>>>>> brcl 15,68ba58
> > >>>>>>>>>>>>> #000000000069c5a2: a7f40001
> > >>>>>>>>>>>>> brc 15,69c5a4
> > >>>>>>>>>>>>> >000000000069c5a6: e340f0c00004
> > >>>>>>>>>>>>> lg %r4,192(%r15)
> > >>>>>>>>>>>>> 000000000069c5ac: ebaff0a00004
> > >>>>>>>>>>>>> lmg %r10,%r15,160(%r15)
> > >>>>>>>>>>>>> 000000000069c5b2: 07f4
> > >>>>>>>>>>>>> bcr 15,%r4
> > >>>>>>>>>>>>> 000000000069c5b4: c0e5fffffeea
> > >>>>>>>>>>>>> brasl %r14,69c388
> > >>>>>>>>>>>>> 000000000069c5ba: a7f4fff6
> > >>>>>>>>>>>>> brc 15,69c5a6
> > >>>>>>>>>>>>> [ 21.455067] Call Trace:
> > >>>>>>>>>>>>> [ 21.455072] ([<00000001b691fd98>] 0x1b691fd98)
> > >>>>>>>>>>>>> [ 21.455079] [<000000000069c692>]
> > >>>>>>>>>>>>> blk_mq_run_hw_queue+0xba/0x100
> > >>>>>>>>>>>>> [ 21.455083] [<000000000069c740>]
> > >>>>>>>>>>>>> blk_mq_run_hw_queues+0x68/0x88
> > >>>>>>>>>>>>> [ 21.455089] [<000000000069b956>]
> > >>>>>>>>>>>>> __blk_mq_complete_request+0x11e/0x1d8
> > >>>>>>>>>>>>> [ 21.455091] [<000000000069ba9c>]
> > >>>>>>>>>>>>> blk_mq_complete_request+0x8c/0xc8
> > >>>>>>>>>>>>> [ 21.455103] [<00000000008aa250>]
> > >>>>>>>>>>>>> dasd_block_tasklet+0x158/0x490
> > >>>>>>>>>>>>> [ 21.455110] [<000000000014c742>]
> > >>>>>>>>>>>>> tasklet_hi_action+0x92/0x120
> > >>>>>>>>>>>>> [ 21.455118] [<0000000000a7cfc0>] __do_softirq+0x120/0x348
> > >>>>>>>>>>>>> [ 21.455122] [<000000000014c212>] irq_exit+0xba/0xd0
> > >>>>>>>>>>>>> [ 21.455130] [<000000000010bf92>] do_IRQ+0x8a/0xb8
> > >>>>>>>>>>>>> [ 21.455133] [<0000000000a7c298>]
> > >>>>>>>>>>>>> io_int_handler+0x130/0x298
> > >>>>>>>>>>>>> [ 21.455136] Last Breaking-Event-Address:
> > >>>>>>>>>>>>> [ 21.455138] [<000000000069c5a2>]
> > >>>>>>>>>>>>> __blk_mq_delay_run_hw_queue+0xba/0xd8
> > >>>>>>>>>>>>> [ 21.455140] ---[ end trace be43f99a5d1e553e ]---
> > >>>>>>>>>>>>> [ 21.510046] dasdconf.sh Warning: 0.0.241e is already
> > >>>>>>>>>>>>> online, not configuring
> > >>>>>>>>>>>>
> > >>>>>>>>>>>> Thinking about this issue further, I can't understand the root
> > >>>>>>>>>>>> cause for
> > >>>>>>>>>>>> this issue.
> > >>>>>>>>
> > >>>>>>>> FWIW, Limiting CONFIG_NR_CPUS to 64 seems to make the problem go
> > >>>>>>>> away.
> > >>>>>>>
> > >>>>>>> I think the following patch is needed, and this way aligns to the
> > >>>>>>> mapping
> > >>>>>>> created via managed IRQ at least.
> > >>>>>>>
> > >>>>>>> diff --git a/block/blk-mq-cpumap.c b/block/blk-mq-cpumap.c
> > >>>>>>> index 9f8cffc8a701..638ab5c11b3c 100644
> > >>>>>>> --- a/block/blk-mq-cpumap.c
> > >>>>>>> +++ b/block/blk-mq-cpumap.c
> > >>>>>>> @@ -14,13 +14,12 @@
> > >>>>>>> #include "blk.h"
> > >>>>>>> #include "blk-mq.h"
> > >>>>>>>
> > >>>>>>> +/*
> > >>>>>>> + * Given there isn't CPU hotplug handler in blk-mq, map all
> > >>>>>>> possible CPUs to
> > >>>>>>> + * queues even it isn't present yet.
> > >>>>>>> + */
> > >>>>>>> static int cpu_to_queue_index(unsigned int nr_queues, const int
> > >>>>>>> cpu)
> > >>>>>>> {
> > >>>>>>> - /*
> > >>>>>>> - * Non present CPU will be mapped to queue index 0.
> > >>>>>>> - */
> > >>>>>>> - if (!cpu_present(cpu))
> > >>>>>>> - return 0;
> > >>>>>>> return cpu % nr_queues;
> > >>>>>>> }
> > >>>>>>>
> > >>>>>>> Thanks,
> > >>>>>>> Ming
> > >>>>>>>
> > >>>>>>
> > >>>>>> With that I no longer see the WARN_ON but the other warning instead:
> > >>>>>>
> > >>>>>> [ 31.903096] audit: type=1130 audit(1522318064.439:41): pid=1
> > >>>>>> uid=0 auid=4294967295 ses=4294967295
> > >>>>>> msg='unit=systemd-tmpfiles-setup-dev comm="systemd"
> > >>>>>> exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=?
> > >>>>>> res=success'
> > >>>>>> [ 31.903100] audit: type=1131 audit(1522318064.439:42): pid=1
> > >>>>>> uid=0 auid=4294967295 ses=4294967295
> > >>>>>> msg='unit=systemd-tmpfiles-setup-dev comm="systemd"
> > >>>>>> exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=?
> > >>>>>> res=success'
> > >>>>>> [ 31.985756] systemd-journald[379]: Received SIGTERM from PID 1
> > >>>>>> (systemd).
> > >>>>>> [ 32.000543] systemd: 18 output lines suppressed due to
> > >>>>>> ratelimiting
> > >>>>>> [ 32.209496] EXT4-fs (dasdc1): re-mounted. Opts: (null)
> > >>>>>> [ 32.234808] systemd-journald[2490]: Received request to flush
> > >>>>>> runtime journal from PID 1
> > >>>>>> [ 32.359832] tun: Universal TUN/TAP device driver, 1.6
> > >>>>>> [ 32.470841] run queue from wrong CPU 18, hctx active
> > >>>>>
> > >>>>> But your 'lscpu' log showed that you only have 16 CPUs online(0~15)
> > >>>>> and
> > >>>>> you also said CPU hotplug isn't involved in your test, so I am just
> > >>>>> wondering where the CPU 18 is from?
> > >>>>
> > >>>>
> > >>>> I have 2 test systems. One with 44CPU the other with 16.
> > >>>> The one with 16 now has the same flood of messages:
> > >>>>
> > >>>>
> > >>>> 4.454510] dasdc:VOL1/ 0X3F77: dasdc1
> > >>>> [ 4.454592] dasdd:VOL1/ 0X3F74: dasdd1
> > >>>> [ 4.593711] run queue from wrong CPU 0, hctx active
> > >>>
> > >>> Still can't reproduce your issue, could you please collect debugfs
> > >>> log again on your 16-core system after applying the patch on
> > >>> blk_mq_cpumap.c?
> > >>
> > >> done.
> > >
> > > OK, thanks, from the dumped mapping, looks everything is fine, each hctx
> > > is mapped to 4 CPUs, and only one of them is online.
> > >
> > > But still don't know why hctx is run from wrong CPU, looks we need more
> > > info to dump, could you apply the following debug patch and post the
> > > log?
> > >
> > > ---
> > > diff --git a/block/blk-mq-cpumap.c b/block/blk-mq-cpumap.c
> > > index 9f8cffc8a701..638ab5c11b3c 100644
> > > --- a/block/blk-mq-cpumap.c
> > > +++ b/block/blk-mq-cpumap.c
> > > @@ -14,13 +14,12 @@
> > > #include "blk.h"
> > > #include "blk-mq.h"
> > >
> > > +/*
> > > + * Given there isn't CPU hotplug handler in blk-mq, map all CPUs to
> > > + * queues even it isn't present yet.
> > > + */
> > > static int cpu_to_queue_index(unsigned int nr_queues, const int cpu)
> > > {
> > > - /*
> > > - * Non present CPU will be mapped to queue index 0.
> > > - */
> > > - if (!cpu_present(cpu))
> > > - return 0;
> > > return cpu % nr_queues;
> > > }
> > >
> > > diff --git a/block/blk-mq.c b/block/blk-mq.c
> > > index 16e83e6df404..65767be7927d 100644
> > > --- a/block/blk-mq.c
> > > +++ b/block/blk-mq.c
> > > @@ -1325,9 +1325,14 @@ static void __blk_mq_run_hw_queue(struct
> > > blk_mq_hw_ctx *hctx)
> > > */
> > > if (!cpumask_test_cpu(raw_smp_processor_id(), hctx->cpumask) &&
> > > cpu_online(hctx->next_cpu)) {
> > > - printk(KERN_WARNING "run queue from wrong CPU %d, hctx %s\n",
> > > - raw_smp_processor_id(),
> > > + int cpu;
> > > + printk(KERN_WARNING "run queue from wrong CPU %d/%d, hctx-%d
> > > %s\n",
> > > + raw_smp_processor_id(), hctx->next_cpu,
> > > + hctx->queue_num,
> > > cpumask_empty(hctx->cpumask) ? "inactive": "active");
> > > + for_each_cpu(cpu, hctx->cpumask)
> > > + printk("%d ", cpu);
> > > + printk("\n");
> > > dump_stack();
> > > }
> > >
> >
> >
> > attached. FWIW, it looks like these messages happen mostly during boot and
> > come less
> > and less often the longer the system runs. Could it be that the workqeue is
> > misplaced
> > before it runs for the first time, but then it is ok?
>
> Looks not workqueue's issue, and it shows that hctx->next_cpu is figured
> out as wrong by blk_mq_hctx_next_cpu(), maybe on your ARCH, either
> 'nr_cpu_ids'
> or 'cpu_online_mask' is too strange during booting and breaks
> blk_mq_hctx_next_cpu().
>
> Could you please apply the following patch and provide the dmesg boot log?
And please post out the 'lscpu' log together from the test machine too.
Thanks,
Ming