There was is no hugetlb specific routine for clearing soft dirty and
other referrences.  The 'default' routines would only clear the
VM_SOFTDIRTY flag in the vma.

Add new routine specifically for hugetlb vmas.

Signed-off-by: Mike Kravetz <[email protected]>
---
 fs/proc/task_mmu.c | 110 +++++++++++++++++++++++++++++++++++++++++++++
 1 file changed, 110 insertions(+)

diff --git a/fs/proc/task_mmu.c b/fs/proc/task_mmu.c
index 829b35016aaa..f06cf9b131a8 100644
--- a/fs/proc/task_mmu.c
+++ b/fs/proc/task_mmu.c
@@ -1116,6 +1116,115 @@ static inline void clear_soft_dirty_pmd(struct 
vm_area_struct *vma,
 }
 #endif
 
+#ifdef CONFIG_HUGETLB_PAGE
+static inline bool huge_pte_is_pinned(struct vm_area_struct *vma,
+                                       unsigned long addr, pte_t pte)
+{
+       struct page *page;
+
+       if (likely(!atomic_read(&vma->vm_mm->has_pinned)))
+               return false;
+       page = pte_page(pte);
+       if (!page)
+               return false;
+       return page_maybe_dma_pinned(page);
+}
+
+static int clear_refs_hugetlb_range(pte_t *ptep, unsigned long hmask,
+                               unsigned long addr, unsigned long end,
+                               struct mm_walk *walk)
+{
+       struct clear_refs_private *cp = walk->private;
+       struct vm_area_struct *vma = walk->vma;
+       struct hstate *h = hstate_vma(walk->vma);
+       unsigned long adj_start = addr, adj_end = end;
+       spinlock_t *ptl;
+       pte_t old_pte, pte;
+
+       /*
+        * clear_refs should only operate on complete vmas.  Therefore,
+        * values passed here should be huge page aligned and huge page
+        * size in length.  Quick validation before taking any action in
+        * case upstream code is changed.
+        */
+       if ((addr & hmask) != addr || end - addr != huge_page_size(h)) {
+               WARN_ONCE(1, "%s passed unaligned address\n", __func__);
+               return 1;
+       }
+
+       ptl = huge_pte_lock(hstate_vma(vma), walk->mm, ptep);
+
+       /* Soft dirty and pmd sharing do not mix */
+
+       pte = huge_ptep_get(ptep);
+       if (!pte_present(pte))
+               goto out;
+       if (unlikely(is_hugetlb_entry_hwpoisoned(pte)))
+               goto out;
+
+       if (cp->type == CLEAR_REFS_SOFT_DIRTY) {
+               if (huge_pte_is_pinned(vma, addr, pte))
+                       goto out;
+
+               /*
+                * soft dirty and pmd sharing do not work together as
+                * per-process is tracked in ptes, and pmd sharing allows
+                * processed to share ptes.  We unshare any pmds here.
+                */
+               adjust_range_if_pmd_sharing_possible(vma, &adj_start, &adj_end);
+               flush_cache_range(vma, adj_start, adj_end);
+               /*
+                * Only atttempt unshare if sharing possible.  If we unshare,
+                * then pte's for a PUD sized area are effectively removed for
+                * this process.  That clears soft dirty.
+                */
+               if (adj_start != addr || adj_end != end) {
+                       struct mmu_notifier_range range;
+                       int unshared;
+
+                       mmu_notifier_range_init(&range, MMU_NOTIFY_CLEAR,
+                                       0, vma, vma->vm_mm, adj_start, adj_end);
+                       mmu_notifier_invalidate_range_start(&range);
+                       i_mmap_lock_write(vma->vm_file->f_mapping);
+                       unshared = huge_pmd_unshare(vma->vm_mm, vma,
+                                                               &addr, ptep);
+                       i_mmap_unlock_write(vma->vm_file->f_mapping);
+                       mmu_notifier_invalidate_range_end(&range);
+                       if (unshared)
+                               goto unshare_done;
+               }
+
+               if (is_hugetlb_entry_migration(pte)) {
+                       pte = huge_pte_swp_clear_soft_dirty(pte);
+                       set_huge_pte_at(walk->mm, addr, ptep, pte);
+               } else {
+                       old_pte = huge_ptep_modify_prot_start(vma, addr, ptep);
+                       pte = pte_mkhuge(huge_pte_wrprotect(pte));
+                       pte = arch_make_huge_pte(pte, vma, NULL, 0);
+                       pte = huge_pte_clear_soft_dirty(pte);
+                       pte = huge_pte_mkyoung(pte);
+                       huge_ptep_modify_prot_commit(vma, addr, ptep,
+                                                       old_pte, pte);
+               }
+
+unshare_done:
+               flush_hugetlb_tlb_range(vma, addr, end);
+       }
+
+       /* reference bits in hugetlb pages are not reset/used */
+out:
+       spin_unlock(ptl);
+       return 0;
+}
+#else /* CONFIG_HUGETLB_PAGE */
+static int clear_refs_hugetlb_range(pte_t *ptep, unsigned long hmask,
+                               unsigned long addr, unsigned long end,
+                               struct mm_walk *walk)
+{
+       return 1;
+}
+#endif /* CONFIG_HUGETLB_PAGE */
+
 static int clear_refs_pte_range(pmd_t *pmd, unsigned long addr,
                                unsigned long end, struct mm_walk *walk)
 {
@@ -1198,6 +1307,7 @@ static int clear_refs_test_walk(unsigned long start, 
unsigned long end,
 }
 
 static const struct mm_walk_ops clear_refs_walk_ops = {
+       .hugetlb_entry          = clear_refs_hugetlb_range,
        .pmd_entry              = clear_refs_pte_range,
        .test_walk              = clear_refs_test_walk,
 };
-- 
2.29.2

Reply via email to