On 09/08/20 09:39, Igor Mammedov wrote: > On Mon, 7 Sep 2020 17:17:52 +0200 > Laszlo Ersek <ler...@redhat.com> wrote:
>> 1 Method (CSCN, 0, Serialized) >> 2 { >> 3 Acquire (\_SB.PCI0.PRES.CPLK, 0xFFFF) >> 4 Name (CNEW, Package (0xFF){}) >> 5 Local_Uid = Zero >> 6 Local_HasJob = One >> 7 While ((Local_HasJob == One)) >> 8 { >> 9 Local_HasJob = Zero >> 10 Local_HasEvent = One >> 11 Local_NumAddedCpus = Zero >> 12 While (((Local_HasEvent == One) && (Local_Uid < >> 0x08))) >> 13 { >> 14 Local_HasEvent = Zero >> 15 \_SB.PCI0.PRES.CSEL = Local_Uid >> 16 \_SB.PCI0.PRES.CCMD = Zero >> 17 If ((\_SB.PCI0.PRES.CDAT < Local_Uid)) >> 18 { >> 19 Break >> 20 } >> 21 >> 22 If ((Local_NumAddedCpus == 0xFF)) >> 23 { >> 24 Local_HasJob = One >> 25 Break >> 26 } >> 27 >> 28 Local_Uid = \_SB.PCI0.PRES.CDAT >> 29 If ((\_SB.PCI0.PRES.CINS == One)) >> 30 { >> 31 CNEW [Local_NumAddedCpus] = Local_Uid >> 32 Local_NumAddedCpus++ >> 33 Local_HasEvent = One >> 34 } >> 35 ElseIf ((\_SB.PCI0.PRES.CRMV == One)) >> 36 { >> 37 CTFY (Local_Uid, 0x03) >> 38 \_SB.PCI0.PRES.CRMV = One >> 39 Local_HasEvent = One >> 40 } >> 41 >> 42 Local_Uid++ >> 43 } >> 44 >> 45 If ((Local_NumAddedCpus > Zero)) >> 46 { >> 47 \_SB.PCI0.SMI0.SMIC = 0x04 >> 48 } >> 49 >> 50 Local_CpuIdx = Zero >> 51 While ((Local_CpuIdx < Local_NumAddedCpus)) >> 52 { >> 53 Local_Uid = DerefOf (CNEW [Local_CpuIdx]) >> 54 CTFY (Local_Uid, One) >> 55 Debug = Local_Uid >> 56 \_SB.PCI0.PRES.CSEL = Local_Uid >> 57 \_SB.PCI0.PRES.CINS = One >> 58 Local_CpuIdx++ >> 59 } >> 60 } >> 61 >> 62 Release (\_SB.PCI0.PRES.CPLK) >> 63 } >> >> When we take the Break on line 25, then: >> >> (a) on line 25, the following equality holds: >> >> Local_Uid == CNEW[Local_NumAddedCpus - 1] + 1 >> >> (b) on line 60, the following equality holds: >> >> Local_Uid == CNEW[Local_NumAddedCpus - 1] >> >> This means that, when we write Local_Uid to CSEL on line 15 again, then: >> >> - we effectively re-investigate the last-cached CPU (with selector value >> CNEW[Local_NumAddedCpus - 1]) >> >> - rather than resuming the scanning right after it -- that is, with >> selector value (CNEW[Local_NumAddedCpus - 1] + 1) --, in the spirit of >> line 42. >> >> My question is: is this "rewind" intentional? >> >> Now, I don't see a functionality problem with this, as on line 57, we >> clear the pending insert event for the last-cached CPU, so when we >> re-check it, the "get pending" command will simply seek past it. >> >> But I'd like to understand if this is *precisely* your intent, or if >> it's an oversight and it just ends up working OK. > it's the later (it should not have any adverse effects) so I didn't care > much about restarting from the last processed CPU. > > how about moving > > 22 If ((Local_NumAddedCpus == 0xFF)) > 23 { > 24 Local_HasJob = One > 25 Break > 26 } > > right after > 40 } > 41 > 42 Local_Uid++ > > instead of adding extra 'if' at the end of outer loop? That only seems to save a CSEL write on line 15, during the first iteration of the outer loop. And we would still re-select the last selector from CNEW, in the second iteration of the outer loop. But, again, there's no bug; I just wanted to understand your intent. Can you please squash the following patch: > diff --git a/hw/acpi/cpu.c b/hw/acpi/cpu.c > index 12839720018e..8dd4d8ebbf55 100644 > --- a/hw/acpi/cpu.c > +++ b/hw/acpi/cpu.c > @@ -601,6 +601,15 @@ void build_cpus_aml(Aml *table, MachineState *machine, > CPUHotplugFeatures opts, > aml_append(while_ctx, aml_increment(cpu_idx)); > } > aml_append(while_ctx2, while_ctx); > + /* > + * If another batch is needed, then it will resume scanning > + * exactly at -- and not after -- the last CPU that's > currently > + * in CPU_ADDED_LIST. In other words, the last CPU in > + * CPU_ADDED_LIST is going to be re-checked. That's OK: we've > + * just cleared the insert event for *all* CPUs in > + * CPU_ADDED_LIST, including the last one. So the scan will > + * simply seek past it. > + */ > } > aml_append(method, while_ctx2); > aml_append(method, aml_release(ctrl_lock)); With that: Reviewed-by: Laszlo Ersek <ler...@redhat.com> I'll also follow up with test results for this patch (meaning a lowered "max_cpus_per_pass"). Thanks! Laszlo