Was able to test the patch, here is the result: I have not tested with
bigger VMs though. Results make it difficult to talk about any side
effect of
patch if any.
System 16 core 32cpu (+ht) sandybridge
with 4 guests of 16vcpu each
+-----------+-----------+-----------+------------+-----------+
kernbench (time taken lower is better)
+-----------+-----------+-----------+------------+-----------+
base %stdev patched %stdev %improvement
+-----------+-----------+-----------+------------+-----------+
1x 53.1421 2.3086 54.6671 2.9673 -2.86966
2x 89.6858 6.4540 94.0626 6.8317 -4.88015
+-----------+-----------+-----------+------------+-----------+
+-----------+-----------+-----------+------------+-----------+
ebizzy (recors/sec higher is better)
+-----------+-----------+-----------+------------+-----------+
base %stdev patched %stdev %improvement
+-----------+-----------+-----------+------------+-----------+
1x 14523.2500 8.4388 14928.8750 3.0478 2.79294
2x 3338.8750 1.4592 3270.8750 2.3980 -2.03661
+-----------+-----------+-----------+------------+-----------+
+-----------+-----------+-----------+------------+-----------+
dbench (Throughput higher is better)
+-----------+-----------+-----------+------------+-----------+
base %stdev patched %stdev %improvement
+-----------+-----------+-----------+------------+-----------+
1x 6386.4737 1.0487 6703.9113 1.2298 4.97047
2x 2571.4712 1.3733 2571.8175 1.6919 0.01347
+-----------+-----------+-----------+------------+-----------+
Raghu
On Wed, Nov 26, 2014 at 3:01 PM, Christian Borntraeger
<[email protected]> wrote:
> Am 26.11.2014 um 10:23 schrieb David Hildenbrand:
>>> This change is a trade-off.
>>> PRO: This patch would improve the case of preemption on s390. This is
>>> probably a corner case as most distros have preemption off anyway.
>>> CON: The downside is that kvm_vcpu_yield_to is called also from
>>> kvm_vcpu_on_spin. Here we want to avoid the scheduler overhead for a wrong
>>> decision.
>>
>> Won't most of that part be covered by:
>> if (!ACCESS_ONCE(vcpu->preempted))
>
> Hmm, right. Checking vcpu->preempted and PF_VCPU might boil down to the same.
> Would be good if to have to performance regression test, though.
>
>>
>> vcpu->preempted is only set when scheduled out involuntarily. It is cleared
>> when scheduled in. s390 sets it manually, to speed up waking up a vcpu.
>>
>> So when our task is scheduled in (an PF_VCPU is set), this check will already
>> avoid scheduler overhead in kvm_vcpu_on_spin() or am I missing something?
>>
>
> CC Raghavendra KT. Could be rerun your kernbench/sysbench/ebizzy setup on x86
> to see if the patch in this thread causes any regression? If think your
> commit 7bc7ae25b143"kvm: Iterate over only vcpus that are preempted" might
> have really made the PF_VCPU check unnecessary
>
> CC Michael Mueller, do we still have our yield performance setup handy to
> check if this patch causes any regression?
>
>
> Christian
>
--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html