By adding a helper that sets multiple PTEs at once, we can generalize
the break-before-make handling into one single place.

No functional change.

Signed-off-by: Ahmad Fatoum <a.fat...@pengutronix.de>
---
 arch/arm/cpu/mmu_32.c | 83 ++++++++++++++++++++++---------------------
 1 file changed, 42 insertions(+), 41 deletions(-)

diff --git a/arch/arm/cpu/mmu_32.c b/arch/arm/cpu/mmu_32.c
index a76d403e3477..7cf04ea9412a 100644
--- a/arch/arm/cpu/mmu_32.c
+++ b/arch/arm/cpu/mmu_32.c
@@ -106,6 +106,23 @@ static void set_pte(uint32_t *pt, uint32_t val)
        WRITE_ONCE(*pt, val);
 }
 
+static void set_pte_range(unsigned level, uint32_t *virt, phys_addr_t phys,
+                         size_t count, uint32_t attrs, bool bbm)
+{
+       unsigned granularity = granule_size(level);
+
+       if (!bbm)
+               goto write_attrs;
+
+        // TODO break-before-make missing
+
+write_attrs:
+       for (int i = 0; i < count; i++, phys += granularity)
+               set_pte(&virt[i], phys | attrs);
+
+       dma_flush_range(virt, count * sizeof(*virt));
+}
+
 #ifdef __PBL__
 static uint32_t *alloc_pte(void)
 {
@@ -203,11 +220,11 @@ void dma_inv_range(void *ptr, size_t size)
  * Not yet exported, but may be later if someone finds use for it.
  */
 static u32 *arm_create_pte(unsigned long virt, unsigned long phys,
-                          uint32_t flags)
+                          uint32_t flags, bool bbm)
 {
        uint32_t *ttb = get_ttb();
        u32 *table;
-       int i, ttb_idx;
+       int ttb_idx;
 
        virt = ALIGN_DOWN(virt, PGDIR_SIZE);
        phys = ALIGN_DOWN(phys, PGDIR_SIZE);
@@ -216,16 +233,9 @@ static u32 *arm_create_pte(unsigned long virt, unsigned 
long phys,
 
        ttb_idx = pgd_index(virt);
 
-       for (i = 0; i < PTRS_PER_PTE; i++) {
-               set_pte(&table[i], phys | PTE_TYPE_SMALL | flags);
-               virt += PAGE_SIZE;
-               phys += PAGE_SIZE;
-       }
-       dma_flush_range(table, PTRS_PER_PTE * sizeof(u32));
+       set_pte_range(2, table, phys, PTRS_PER_PTE, PTE_TYPE_SMALL | flags, 
bbm);
 
-       // TODO break-before-make missing
-       set_pte(&ttb[ttb_idx], (unsigned long)table | PMD_TYPE_TABLE);
-       dma_flush_range(&ttb[ttb_idx], sizeof(u32));
+       set_pte_range(1, &ttb[ttb_idx], (unsigned long)table, 1, 
PMD_TYPE_TABLE, bbm);
 
        return table;
 }
@@ -335,6 +345,7 @@ static void __arch_remap_range(void *_virt_addr, 
phys_addr_t phys_addr, size_t s
                               maptype_t map_type)
 {
        bool force_pages = map_type & ARCH_MAP_FLAG_PAGEWISE;
+       bool mmu_on;
        u32 virt_addr = (u32)_virt_addr;
        u32 pte_flags, pmd_flags;
        uint32_t *ttb = get_ttb();
@@ -351,30 +362,30 @@ static void __arch_remap_range(void *_virt_addr, 
phys_addr_t phys_addr, size_t s
        if (!size)
                return;
 
+       mmu_on = get_cr() & CR_M;
+
        while (size) {
                const bool pgdir_size_aligned = IS_ALIGNED(virt_addr, 
PGDIR_SIZE);
                u32 *pgd = (u32 *)&ttb[pgd_index(virt_addr)];
+               u32 flags;
                size_t chunk;
 
                if (size >= PGDIR_SIZE && pgdir_size_aligned &&
                    IS_ALIGNED(phys_addr, PGDIR_SIZE) &&
                    !pgd_type_table(*pgd) && !force_pages) {
-                       u32 val;
                        /*
                         * TODO: Add code to discard a page table and
                         * replace it with a section
                         */
                        chunk = PGDIR_SIZE;
-                       val = phys_addr | pmd_flags;
+                       flags = pmd_flags;
                        if (!maptype_is_compatible(map_type, MAP_FAULT))
-                               val |= PMD_TYPE_SECT;
-                       // TODO break-before-make missing
-                       set_pte(pgd, val);
-                       dma_flush_range(pgd, sizeof(*pgd));
+                               flags |= PMD_TYPE_SECT;
+                       set_pte_range(1, pgd, phys_addr, 1, flags, mmu_on);
                } else {
                        unsigned int num_ptes;
                        u32 *table = NULL;
-                       unsigned int i, level;
+                       unsigned int level;
                        u32 *pte;
                        /*
                         * We only want to cover pages up until next
@@ -401,23 +412,14 @@ static void __arch_remap_range(void *_virt_addr, 
phys_addr_t phys_addr, size_t s
                                 * and create a new page table for it
                                 */
                                table = arm_create_pte(virt_addr, phys_addr,
-                                                      pmd_flags_to_pte(*pgd));
+                                                      pmd_flags_to_pte(*pgd), 
mmu_on);
                                pte = find_pte(ttb, virt_addr, NULL);
                        }
 
-                       for (i = 0; i < num_ptes; i++) {
-                               u32 val;
-
-                               val = phys_addr + i * PAGE_SIZE;
-                               val |= pte_flags;
-                               if (!maptype_is_compatible(map_type, MAP_FAULT))
-                                       val |= PTE_TYPE_SMALL;
-
-                               // TODO break-before-make missing
-                               set_pte(&pte[i], val);
-                       }
-
-                       dma_flush_range(pte, num_ptes * sizeof(u32));
+                       flags = pte_flags;
+                       if (!maptype_is_compatible(map_type, MAP_FAULT))
+                               flags |= PTE_TYPE_SMALL;
+                       set_pte_range(2, pte, phys_addr, num_ptes, flags, 
mmu_on);
                }
 
                virt_addr += chunk;
@@ -461,6 +463,7 @@ static void early_create_sections(unsigned long first, 
unsigned long last,
        unsigned long ttb_end = pgd_index(last) + 1;
        unsigned int i, addr = first;
 
+       /* This always runs with MMU disabled, so just opencode the loop */
        for (i = ttb_start; i < ttb_end; i++) {
                set_pte(&ttb[i], addr | flags);
                addr += PGDIR_SIZE;
@@ -475,13 +478,11 @@ static inline void early_create_flat_mapping(void)
 
 void *map_io_sections(unsigned long phys, void *_start, size_t size)
 {
-       unsigned long start = (unsigned long)_start, sec;
+       unsigned long start = (unsigned long)_start;
        uint32_t *ttb = get_ttb();
 
-       for (sec = start; sec < start + size; sec += PGDIR_SIZE, phys += 
PGDIR_SIZE) {
-               // TODO break-before-make missing
-               set_pte(&ttb[pgd_index(sec)], phys | 
get_pmd_flags(MAP_UNCACHED));
-       }
+       set_pte_range(1, &ttb[pgd_index(start)], phys, size / PGDIR_SIZE,
+                     get_pmd_flags(MAP_UNCACHED), true);
 
        dma_flush_range(ttb, 0x4000);
        tlb_invalidate();
@@ -523,11 +524,11 @@ static void create_vector_table(unsigned long adr)
                vectors = xmemalign(PAGE_SIZE, PAGE_SIZE);
                pr_debug("Creating vector table, virt = 0x%p, phys = 0x%08lx\n",
                         vectors, adr);
-               arm_create_pte(adr, adr, get_pte_flags(MAP_UNCACHED));
+
+               arm_create_pte(adr, adr, get_pte_flags(MAP_UNCACHED), true);
                pte = find_pte(get_ttb(), adr, NULL);
-               // TODO break-before-make missing
-               set_pte(pte, (u32)vectors | PTE_TYPE_SMALL |
-                       get_pte_flags(MAP_CACHED));
+               set_pte_range(2, pte, (u32)vectors, 1, PTE_TYPE_SMALL |
+                             get_pte_flags(MAP_CACHED), true);
        }
 
        arm_fixup_vectors();
-- 
2.39.5


Reply via email to