On Thu, May 17, 2012 at 01:24:42PM +0300, Avi Kivity wrote:
> Signed-off-by: Avi Kivity <[email protected]>
> ---
> virt/kvm/kvm_main.c | 16 ++++++++--------
> 1 file changed, 8 insertions(+), 8 deletions(-)
>
> diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c
> index 585ab45..9f6d15d 100644
> --- a/virt/kvm/kvm_main.c
> +++ b/virt/kvm/kvm_main.c
> @@ -302,11 +302,11 @@ static void kvm_mmu_notifier_invalidate_page(struct
> mmu_notifier *mn,
> kvm->mmu_notifier_seq++;
> if (kvm_unmap_hva(kvm, address))
> kvm_mark_tlb_dirty(kvm);
> - /* we've to flush the tlb before the pages can be freed */
> - kvm_cond_flush_remote_tlbs(kvm);
> -
> spin_unlock(&kvm->mmu_lock);
> srcu_read_unlock(&kvm->srcu, idx);
> +
> + /* we've to flush the tlb before the pages can be freed */
> + kvm_cond_flush_remote_tlbs(kvm);
> }
There are still sites that assumed implicitly that acquiring mmu_lock
guarantees that sptes and remote TLBs are in sync. Example:
void kvm_mmu_zap_all(struct kvm *kvm)
{
struct kvm_mmu_page *sp, *node;
LIST_HEAD(invalid_list);
spin_lock(&kvm->mmu_lock);
restart:
list_for_each_entry_safe(sp, node, &kvm->arch.active_mmu_pages,
link)
if (kvm_mmu_prepare_zap_page(kvm, sp, &invalid_list))
goto restart;
kvm_mmu_commit_zap_page(kvm, &invalid_list);
spin_unlock(&kvm->mmu_lock);
}
kvm_mmu_commit_zap_page only flushes if the TLB was dirtied by this
context, not before. kvm_mmu_unprotect_page is similar.
In general:
context 1 context 2
lock(mmu_lock)
modify spte
mark_tlb_dirty
unlock(mmu_lock)
lock(mmu_lock)
read spte
make a decision based on spte value
unlock(mmu_lock)
flush remote TLBs
Is scary.
Perhaps have a rule that says:
1) Conditionally flush remote TLB after acquiring mmu_lock,
before anything (even perhaps inside the lock macro).
2) Except special cases where it is clear that this is not
necessary.
--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html