On Wed, Jul 09, 2014 at 04:12:53PM -0300, [email protected] wrote:
> Reload remote vcpus MMU from GET_DIRTY_LOG codepath, before
> deleting a pinned spte.
>
> Signed-off-by: Marcelo Tosatti <[email protected]>
>
> ---
> arch/x86/kvm/mmu.c | 29 +++++++++++++++++++++++------
> 1 file changed, 23 insertions(+), 6 deletions(-)
>
> Index: kvm.pinned-sptes/arch/x86/kvm/mmu.c
> ===================================================================
> --- kvm.pinned-sptes.orig/arch/x86/kvm/mmu.c 2014-07-09 11:23:59.290744490
> -0300
> +++ kvm.pinned-sptes/arch/x86/kvm/mmu.c 2014-07-09 11:24:58.449632435
> -0300
> @@ -1208,7 +1208,8 @@
> *
> * Return true if tlb need be flushed.
> */
> -static bool spte_write_protect(struct kvm *kvm, u64 *sptep, bool pt_protect)
> +static bool spte_write_protect(struct kvm *kvm, u64 *sptep, bool pt_protect,
> + bool skip_pinned)
> {
> u64 spte = *sptep;
>
> @@ -1218,6 +1219,22 @@
>
> rmap_printk("rmap_write_protect: spte %p %llx\n", sptep, *sptep);
>
> + if (is_pinned_spte(spte)) {
> + /* keep pinned spte intact, mark page dirty again */
> + if (skip_pinned) {
> + struct kvm_mmu_page *sp;
> + gfn_t gfn;
> +
> + sp = page_header(__pa(sptep));
> + gfn = kvm_mmu_page_get_gfn(sp, sptep - sp->spt);
> +
> + mark_page_dirty(kvm, gfn);
> + return false;
Why not mark all pinned gfns as dirty in kvm_vm_ioctl_get_dirty_log() while
populating dirty_bitmap_buffer?
> + } else
> + mmu_reload_pinned_vcpus(kvm);
Can you explain why do you need this?
> + }
> +
> +
> if (pt_protect)
> spte &= ~SPTE_MMU_WRITEABLE;
> spte = spte & ~PT_WRITABLE_MASK;
> @@ -1226,7 +1243,7 @@
> }
>
> static bool __rmap_write_protect(struct kvm *kvm, unsigned long *rmapp,
> - bool pt_protect)
> + bool pt_protect, bool skip_pinned)
> {
> u64 *sptep;
> struct rmap_iterator iter;
> @@ -1235,7 +1252,7 @@
> for (sptep = rmap_get_first(*rmapp, &iter); sptep;) {
> BUG_ON(!(*sptep & PT_PRESENT_MASK));
>
> - flush |= spte_write_protect(kvm, sptep, pt_protect);
> + flush |= spte_write_protect(kvm, sptep, pt_protect,
> skip_pinned);
> sptep = rmap_get_next(&iter);
> }
>
> @@ -1261,7 +1278,7 @@
> while (mask) {
> rmapp = __gfn_to_rmap(slot->base_gfn + gfn_offset + __ffs(mask),
> PT_PAGE_TABLE_LEVEL, slot);
> - __rmap_write_protect(kvm, rmapp, false);
> + __rmap_write_protect(kvm, rmapp, false, true);
>
> /* clear the first set bit */
> mask &= mask - 1;
> @@ -1280,7 +1297,7 @@
> for (i = PT_PAGE_TABLE_LEVEL;
> i < PT_PAGE_TABLE_LEVEL + KVM_NR_PAGE_SIZES; ++i) {
> rmapp = __gfn_to_rmap(gfn, i, slot);
> - write_protected |= __rmap_write_protect(kvm, rmapp, true);
> + write_protected |= __rmap_write_protect(kvm, rmapp, true,
> false);
> }
>
> return write_protected;
> @@ -4565,7 +4582,7 @@
>
> for (index = 0; index <= last_index; ++index, ++rmapp) {
> if (*rmapp)
> - __rmap_write_protect(kvm, rmapp, false);
> + __rmap_write_protect(kvm, rmapp, false, false);
>
> if (need_resched() || spin_needbreak(&kvm->mmu_lock))
> cond_resched_lock(&kvm->mmu_lock);
>
>
--
Gleb.
--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html