Instead of checking for each request linearly, use for_each_set_bit() to
iterate on just the requests that are set (should be 0 or 1 most of the
time).

To avoid a useless call to find_first_bit(), add an extra check for no
requests set.  To avoid an extra indent and an unreviewable patch, I
added a rather ugly goto.  This can be fixed in a later patch.

Signed-off-by: Avi Kivity <[email protected]>
---
 arch/x86/kvm/x86.c | 62 ++++++++++++++++++++++++++++++++++--------------------
 1 file changed, 39 insertions(+), 23 deletions(-)

diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
index 162231f..9296dce 100644
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -5217,6 +5217,7 @@ static void process_nmi(struct kvm_vcpu *vcpu)
 
 static int vcpu_enter_guest(struct kvm_vcpu *vcpu)
 {
+       unsigned req;
        int r;
        bool req_int_win = !irqchip_in_kernel(vcpu->kvm) &&
                vcpu->run->request_interrupt_window;
@@ -5225,57 +5226,67 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu)
        if (unlikely(req_int_win))
                kvm_make_request(KVM_REQ_EVENT, vcpu);
 
-       if (vcpu->requests) {
-               if (kvm_check_request(KVM_REQ_MMU_RELOAD, vcpu)) {
+       if (!vcpu->requests)
+               goto no_requests;
+
+       for_each_set_bit(req, &vcpu->requests, BITS_PER_LONG) {
+               clear_bit(req, &vcpu->requests);
+               switch (req) {
+               case KVM_REQ_MMU_RELOAD:
                        kvm_mmu_unload(vcpu);
                        r = kvm_mmu_reload(vcpu);
                        if (unlikely(r)) {
                                kvm_make_request(KVM_REQ_MMU_RELOAD, vcpu);
                                goto out;
                        }
-               }
-               if (kvm_check_request(KVM_REQ_MIGRATE_TIMER, vcpu))
+                       break;
+               case KVM_REQ_MIGRATE_TIMER:
                        __kvm_migrate_timers(vcpu);
-               if (kvm_check_request(KVM_REQ_CLOCK_UPDATE, vcpu)) {
+                       break;
+               case KVM_REQ_CLOCK_UPDATE:
                        r = kvm_guest_time_update(vcpu);
                        if (unlikely(r))
                                goto out;
-               }
-               if (kvm_check_request(KVM_REQ_MMU_SYNC, vcpu))
+                       break;
+               case KVM_REQ_MMU_SYNC:
                        kvm_mmu_sync_roots(vcpu);
-               if (kvm_check_request(KVM_REQ_TLB_FLUSH, vcpu))
+                       break;
+               case KVM_REQ_TLB_FLUSH:
                        kvm_x86_ops->tlb_flush(vcpu);
-               if (kvm_check_request(KVM_REQ_REPORT_TPR_ACCESS, vcpu)) {
+                       break;
+               case KVM_REQ_REPORT_TPR_ACCESS:
                        vcpu->run->exit_reason = KVM_EXIT_TPR_ACCESS;
                        r = 0;
                        goto out;
-               }
-               if (kvm_check_request(KVM_REQ_TRIPLE_FAULT, vcpu)) {
+               case KVM_REQ_TRIPLE_FAULT:
                        vcpu->run->exit_reason = KVM_EXIT_SHUTDOWN;
                        r = 0;
                        goto out;
-               }
-               if (kvm_check_request(KVM_REQ_DEACTIVATE_FPU, vcpu)) {
+               case KVM_REQ_DEACTIVATE_FPU:
                        vcpu->fpu_active = 0;
                        kvm_x86_ops->fpu_deactivate(vcpu);
-               }
-               if (kvm_check_request(KVM_REQ_APF_HALT, vcpu)) {
+                       break;
+               case KVM_REQ_APF_HALT:
                        /* Page is swapped out. Do synthetic halt */
                        vcpu->arch.apf.halted = true;
                        r = 1;
                        goto out;
-               }
-               if (kvm_check_request(KVM_REQ_STEAL_UPDATE, vcpu))
+               case KVM_REQ_STEAL_UPDATE:
                        record_steal_time(vcpu);
-               if (kvm_check_request(KVM_REQ_NMI, vcpu))
+                       break;
+               case KVM_REQ_NMI:
                        process_nmi(vcpu);
-               req_immediate_exit =
-                       kvm_check_request(KVM_REQ_IMMEDIATE_EXIT, vcpu);
-               if (kvm_check_request(KVM_REQ_PMU, vcpu))
+                       break;
+               case KVM_REQ_IMMEDIATE_EXIT:
+                       req_immediate_exit = true;
+                       break;
+               case KVM_REQ_PMU:
                        kvm_handle_pmu_event(vcpu);
-               if (kvm_check_request(KVM_REQ_PMI, vcpu))
+                       break;
+               case KVM_REQ_PMI:
                        kvm_deliver_pmi(vcpu);
-               if (kvm_check_request(KVM_REQ_EVENT, vcpu)) {
+                       break;
+               case KVM_REQ_EVENT:
                        inject_pending_event(vcpu);
 
                        /* enable NMI/IRQ window open exits if needed */
@@ -5288,9 +5299,14 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu)
                                update_cr8_intercept(vcpu);
                                kvm_lapic_sync_to_vapic(vcpu);
                        }
+                       break;
+               default:
+                       BUG();
                }
        }
 
+no_requests:
+
        preempt_disable();
 
        kvm_x86_ops->prepare_guest_switch(vcpu);
-- 
1.7.11

--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to