Gitweb:     
http://git.kernel.org/git/?p=linux/kernel/git/torvalds/linux-2.6.git;a=commit;h=721151d004dcf01a71b12bb6b893f9160284cf6e
Commit:     721151d004dcf01a71b12bb6b893f9160284cf6e
Parent:     1a38147ed0737a9c01dbf5f2ca47fd2a0aa5cb55
Author:     Paul Mackerras <[EMAIL PROTECTED]>
AuthorDate: Tue Apr 3 21:24:02 2007 +1000
Committer:  Paul Mackerras <[EMAIL PROTECTED]>
CommitDate: Fri Apr 13 03:55:18 2007 +1000

    [POWERPC] Allow drivers to map individual 4k pages to userspace
    
    Some drivers have resources that they want to be able to map into
    userspace that are 4k in size.  On a kernel configured with 64k pages
    we currently end up mapping the 4k we want plus another 60k of
    physical address space, which could contain anything.  This can
    introduce security problems, for example in the case of an infiniband
    adaptor where the other 60k could contain registers that some other
    program is using for its communications.
    
    This patch adds a new function, remap_4k_pfn, which drivers can use to
    map a single 4k page to userspace regardless of whether the kernel is
    using a 4k or a 64k page size.  Like remap_pfn_range, it would
    typically be called in a driver's mmap function.  It only maps a
    single 4k page, which on a 64k page kernel appears replicated 16 times
    throughout a 64k page.  On a 4k page kernel it reduces to a call to
    remap_pfn_range.
    
    The way this works on a 64k kernel is that a new bit, _PAGE_4K_PFN,
    gets set on the linux PTE.  This alters the way that __hash_page_4K
    computes the real address to put in the HPTE.  The RPN field of the
    linux PTE becomes the 4k RPN directly rather than being interpreted as
    a 64k RPN.  Since the RPN field is 32 bits, this means that physical
    addresses being mapped with remap_4k_pfn have to be below 2^44,
    i.e. 0x100000000000.
    
    The patch also factors out the code in arch/powerpc/mm/hash_utils_64.c
    that deals with demoting a process to use 4k pages into one function
    that gets called in the various different places where we need to do
    that.  There were some discrepancies between exactly what was done in
    the various places, such as a call to spu_flush_all_slbs in one case
    but not in others.
    
    Signed-off-by: Paul Mackerras <[EMAIL PROTECTED]>
---
 arch/powerpc/mm/hash_low_64.S     |    5 +++-
 arch/powerpc/mm/hash_utils_64.c   |   43 +++++++++++++++++++++++++-----------
 include/asm-powerpc/pgtable-4k.h  |    3 ++
 include/asm-powerpc/pgtable-64k.h |    5 ++++
 4 files changed, 42 insertions(+), 14 deletions(-)

diff --git a/arch/powerpc/mm/hash_low_64.S b/arch/powerpc/mm/hash_low_64.S
index 9bc0a9c..e64ce3e 100644
--- a/arch/powerpc/mm/hash_low_64.S
+++ b/arch/powerpc/mm/hash_low_64.S
@@ -445,9 +445,12 @@ END_FTR_SECTION(CPU_FTR_NOEXECUTE|CPU_FTR_COHERENT_ICACHE, 
CPU_FTR_NOEXECUTE)
 
 htab_insert_pte:
        /* real page number in r5, PTE RPN value + index */
-       rldicl  r5,r31,64-PTE_RPN_SHIFT,PTE_RPN_SHIFT
+       andis.  r0,r31,[EMAIL PROTECTED]
+       srdi    r5,r31,PTE_RPN_SHIFT
+       bne-    htab_special_pfn
        sldi    r5,r5,PAGE_SHIFT-HW_PAGE_SHIFT
        add     r5,r5,r25
+htab_special_pfn:
        sldi    r5,r5,HW_PAGE_SHIFT
 
        /* Calculate primary group hash */
diff --git a/arch/powerpc/mm/hash_utils_64.c b/arch/powerpc/mm/hash_utils_64.c
index 3c7fe2c..aae0853 100644
--- a/arch/powerpc/mm/hash_utils_64.c
+++ b/arch/powerpc/mm/hash_utils_64.c
@@ -573,6 +573,27 @@ unsigned int hash_page_do_lazy_icache(unsigned int pp, 
pte_t pte, int trap)
        return pp;
 }
 
+/*
+ * Demote a segment to using 4k pages.
+ * For now this makes the whole process use 4k pages.
+ */
+void demote_segment_4k(struct mm_struct *mm, unsigned long addr)
+{
+#ifdef CONFIG_PPC_64K_PAGES
+       if (mm->context.user_psize == MMU_PAGE_4K)
+               return;
+       mm->context.user_psize = MMU_PAGE_4K;
+       mm->context.sllp = SLB_VSID_USER | mmu_psize_defs[MMU_PAGE_4K].sllp;
+       get_paca()->context = mm->context;
+       slb_flush_and_rebolt();
+#ifdef CONFIG_SPE_BASE
+       spu_flush_all_slbs(mm);
+#endif
+#endif
+}
+
+EXPORT_SYMBOL_GPL(demote_segment_4k);
+
 /* Result code is:
  *  0 - handled
  *  1 - normal page fault
@@ -665,15 +686,19 @@ int hash_page(unsigned long ea, unsigned long access, 
unsigned long trap)
 #ifndef CONFIG_PPC_64K_PAGES
        rc = __hash_page_4K(ea, access, vsid, ptep, trap, local);
 #else
+       /* If _PAGE_4K_PFN is set, make sure this is a 4k segment */
+       if (pte_val(*ptep) & _PAGE_4K_PFN) {
+               demote_segment_4k(mm, ea);
+               psize = MMU_PAGE_4K;
+       }
+
        if (mmu_ci_restrictions) {
                /* If this PTE is non-cacheable, switch to 4k */
                if (psize == MMU_PAGE_64K &&
                    (pte_val(*ptep) & _PAGE_NO_CACHE)) {
                        if (user_region) {
+                               demote_segment_4k(mm, ea);
                                psize = MMU_PAGE_4K;
-                               mm->context.user_psize = MMU_PAGE_4K;
-                               mm->context.sllp = SLB_VSID_USER |
-                                       mmu_psize_defs[MMU_PAGE_4K].sllp;
                        } else if (ea < VMALLOC_END) {
                                /*
                                 * some driver did a non-cacheable mapping
@@ -756,16 +781,8 @@ void hash_preload(struct mm_struct *mm, unsigned long ea,
        if (mmu_ci_restrictions) {
                /* If this PTE is non-cacheable, switch to 4k */
                if (mm->context.user_psize == MMU_PAGE_64K &&
-                   (pte_val(*ptep) & _PAGE_NO_CACHE)) {
-                       mm->context.user_psize = MMU_PAGE_4K;
-                       mm->context.sllp = SLB_VSID_USER |
-                               mmu_psize_defs[MMU_PAGE_4K].sllp;
-                       get_paca()->context = mm->context;
-                       slb_flush_and_rebolt();
-#ifdef CONFIG_SPE_BASE
-                       spu_flush_all_slbs(mm);
-#endif
-               }
+                   (pte_val(*ptep) & _PAGE_NO_CACHE))
+                       demote_segment_4k(mm, ea);
        }
        if (mm->context.user_psize == MMU_PAGE_64K)
                __hash_page_64K(ea, access, vsid, ptep, trap, local);
diff --git a/include/asm-powerpc/pgtable-4k.h b/include/asm-powerpc/pgtable-4k.h
index 345d9b0..a28fa8b 100644
--- a/include/asm-powerpc/pgtable-4k.h
+++ b/include/asm-powerpc/pgtable-4k.h
@@ -97,3 +97,6 @@
 
 #define pud_ERROR(e) \
        printk("%s:%d: bad pud %08lx.\n", __FILE__, __LINE__, pud_val(e))
+
+#define remap_4k_pfn(vma, addr, pfn, prot)     \
+       remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE, (prot))
diff --git a/include/asm-powerpc/pgtable-64k.h 
b/include/asm-powerpc/pgtable-64k.h
index 4b7126c..5e84f07 100644
--- a/include/asm-powerpc/pgtable-64k.h
+++ b/include/asm-powerpc/pgtable-64k.h
@@ -35,6 +35,7 @@
 #define _PAGE_HPTE_SUB 0x0ffff000 /* combo only: sub pages HPTE bits */
 #define _PAGE_HPTE_SUB0        0x08000000 /* combo only: first sub page */
 #define _PAGE_COMBO    0x10000000 /* this is a combo 4k page */
+#define _PAGE_4K_PFN   0x20000000 /* PFN is for a single 4k page */
 #define _PAGE_F_SECOND  0x00008000 /* full page: hidx bits */
 #define _PAGE_F_GIX     0x00007000 /* full page: hidx bits */
 
@@ -93,6 +94,10 @@
 #define pte_pagesize_index(pte)        \
        (((pte) & _PAGE_COMBO)? MMU_PAGE_4K: MMU_PAGE_64K)
 
+#define remap_4k_pfn(vma, addr, pfn, prot)                             \
+       remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE,                \
+                       __pgprot(pgprot_val((prot)) | _PAGE_4K_PFN))
+
 #endif /*  __ASSEMBLY__ */
 #endif /* __KERNEL__ */
 #endif /* _ASM_POWERPC_PGTABLE_64K_H */
-
To unsubscribe from this list: send the line "unsubscribe git-commits-head" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to