On 11/07/18 17:05, Punit Agrawal wrote:
Suzuki K Poulose <suzuki.poul...@arm.com> writes:

On 09/07/18 15:41, Punit Agrawal wrote:
KVM only supports PMD hugepages at stage 2. Now that the various page
handling routines are updated, extend the stage 2 fault handling to
map in PUD hugepages.

Addition of PUD hugepage support enables additional page sizes (e.g.,
1G with 4K granule) which can be useful on cores that support mapping
larger block sizes in the TLB entries.

Signed-off-by: Punit Agrawal <punit.agra...@arm.com>
Cc: Christoffer Dall <christoffer.d...@arm.com>
Cc: Marc Zyngier <marc.zyng...@arm.com>
Cc: Russell King <li...@armlinux.org.uk>
Cc: Catalin Marinas <catalin.mari...@arm.com>
Cc: Will Deacon <will.dea...@arm.com>
---
   arch/arm/include/asm/kvm_mmu.h         | 19 +++++++
   arch/arm64/include/asm/kvm_mmu.h       | 15 +++++
   arch/arm64/include/asm/pgtable-hwdef.h |  2 +
   arch/arm64/include/asm/pgtable.h       |  2 +
   virt/kvm/arm/mmu.c                     | 78 ++++++++++++++++++++++++--
   5 files changed, 112 insertions(+), 4 deletions(-)


[...]

diff --git a/virt/kvm/arm/mmu.c b/virt/kvm/arm/mmu.c
index a6d3ac9d7c7a..d8e2497e5353 100644
--- a/virt/kvm/arm/mmu.c
+++ b/virt/kvm/arm/mmu.c

[...]

@@ -1100,6 +1139,7 @@ static int stage2_set_pte(struct kvm *kvm, struct 
kvm_mmu_memory_cache *cache,
                          phys_addr_t addr, const pte_t *new_pte,
                          unsigned long flags)
   {
+       pud_t *pud;
        pmd_t *pmd;
        pte_t *pte, old_pte;
        bool iomap = flags & KVM_S2PTE_FLAG_IS_IOMAP;
@@ -1108,6 +1148,22 @@ static int stage2_set_pte(struct kvm *kvm, struct 
kvm_mmu_memory_cache *cache,
        VM_BUG_ON(logging_active && !cache);
        /* Create stage-2 page table mapping - Levels 0 and 1 */
+       pud = stage2_get_pud(kvm, cache, addr);
+       if (!pud) {
+               /*
+                * Ignore calls from kvm_set_spte_hva for unallocated
+                * address ranges.
+                */
+               return 0;
+       }
+
+       /*
+        * While dirty page logging - dissolve huge PUD, then continue
+        * on to allocate page.

Punit,

We don't seem to allocate a page here for the PUD entry, in case if it is 
dissolved
or empty (i.e, stage2_pud_none(*pud) is true.).

I was trying to avoid duplicating the PUD allocation by reusing the
functionality in stage2_get_pmd().

Does the below updated comment help?

        /*
         * While dirty page logging - dissolve huge PUD, it'll be
         * allocated in stage2_get_pmd().
         */

The other option is to duplicate the stage2_pud_none() case from
stage2_get_pmd() here.

I think the explicit check for stage2_pud_none() suits better here.
That would make it explicit that we are tearing down the entries
from top to bottom. Also, we may be able to short cut for case
where we know we just allocated a PUD page and hence we need another
PMD level page.

Also, you are missing the comment about the assumption that stage2 PUD
level always exist with 4k fixed IPA.

Cheers
Suzuki
_______________________________________________
kvmarm mailing list
kvmarm@lists.cs.columbia.edu
https://lists.cs.columbia.edu/mailman/listinfo/kvmarm

Reply via email to