Module Name: src Committed By: maxv Date: Fri Jul 1 11:39:46 UTC 2016
Modified Files: src/sys/arch/x86/x86: pmap.c Log Message: Create the direct map in a separate function. While here, add some comments to explain what we are doing. No functional change. To generate a diff of this commit: cvs rdiff -u -r1.200 -r1.201 src/sys/arch/x86/x86/pmap.c Please note that diffs are not public domain; they are subject to the copyright notices on the relevant files.
Modified files: Index: src/sys/arch/x86/x86/pmap.c diff -u src/sys/arch/x86/x86/pmap.c:1.200 src/sys/arch/x86/x86/pmap.c:1.201 --- src/sys/arch/x86/x86/pmap.c:1.200 Fri Jul 1 11:28:18 2016 +++ src/sys/arch/x86/x86/pmap.c Fri Jul 1 11:39:45 2016 @@ -1,4 +1,4 @@ -/* $NetBSD: pmap.c,v 1.200 2016/07/01 11:28:18 maxv Exp $ */ +/* $NetBSD: pmap.c,v 1.201 2016/07/01 11:39:45 maxv Exp $ */ /*- * Copyright (c) 2008, 2010 The NetBSD Foundation, Inc. @@ -171,7 +171,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.200 2016/07/01 11:28:18 maxv Exp $"); +__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.201 2016/07/01 11:39:45 maxv Exp $"); #include "opt_user_ldt.h" #include "opt_lockdebug.h" @@ -561,6 +561,10 @@ extern vaddr_t pentium_idt_vaddr; * local prototypes */ +#ifdef __HAVE_DIRECT_MAP +static void pmap_init_directmap(struct pmap *); +#endif + #ifndef XEN static void pmap_remap_largepages(void); #endif @@ -1219,19 +1223,12 @@ void pmap_bootstrap(vaddr_t kva_start) { struct pmap *kpm; - pt_entry_t *pte; int i; vaddr_t kva; #ifndef XEN unsigned long p1i; vaddr_t kva_end; #endif -#ifdef __HAVE_DIRECT_MAP - phys_ram_seg_t *mc; - long ndmpdp; - paddr_t lastpa, dmpd, dmpdp, pdp; - vaddr_t tmpva; -#endif pmap_pg_nx = (cpu_feature[2] & CPUID_NOX ? PG_NX : 0); @@ -1343,77 +1340,7 @@ pmap_bootstrap(vaddr_t kva_start) #endif /* !XEN */ #ifdef __HAVE_DIRECT_MAP - - pd_entry_t *pde; - - tmpva = (KERNBASE + NKL2_KIMG_ENTRIES * NBPD_L2); - pte = PTE_BASE + pl1_i(tmpva); - - /* - * Map the direct map RW. Use 1GB pages if they are available, - * otherwise use 2MB pages. Note that the unused parts of - * PTPs * must be zero outed, as they might be accessed due - * to speculative execution. Also, PG_G is not allowed on - * non-leaf PTPs. - */ - - lastpa = 0; - for (i = 0; i < mem_cluster_cnt; i++) { - mc = &mem_clusters[i]; - lastpa = MAX(lastpa, mc->start + mc->size); - } - - ndmpdp = (lastpa + NBPD_L3 - 1) >> L3_SHIFT; - dmpdp = avail_start; avail_start += PAGE_SIZE; - - *pte = dmpdp | PG_V | PG_RW | pmap_pg_nx; - pmap_update_pg(tmpva); - memset((void *)tmpva, 0, PAGE_SIZE); - - if (cpu_feature[2] & CPUID_P1GB) { - for (i = 0; i < ndmpdp; i++) { - pdp = (paddr_t)&(((pd_entry_t *)dmpdp)[i]); - *pte = (pdp & PG_FRAME) | PG_V | PG_RW | pmap_pg_nx; - pmap_update_pg(tmpva); - - pde = (pd_entry_t *)(tmpva + (pdp & ~PG_FRAME)); - *pde = ((paddr_t)i << L3_SHIFT) | PG_RW | pmap_pg_nx | - PG_V | PG_U | PG_PS | PG_G; - } - } else { - dmpd = avail_start; avail_start += ndmpdp * PAGE_SIZE; - - for (i = 0; i < ndmpdp; i++) { - pdp = dmpd + i * PAGE_SIZE; - *pte = (pdp & PG_FRAME) | PG_V | PG_RW | pmap_pg_nx; - pmap_update_pg(tmpva); - - memset((void *)tmpva, 0, PAGE_SIZE); - } - for (i = 0; i < NPDPG * ndmpdp; i++) { - pdp = (paddr_t)&(((pd_entry_t *)dmpd)[i]); - *pte = (pdp & PG_FRAME) | PG_V | PG_RW | pmap_pg_nx; - pmap_update_pg(tmpva); - - pde = (pd_entry_t *)(tmpva + (pdp & ~PG_FRAME)); - *pde = ((paddr_t)i << L2_SHIFT) | PG_RW | pmap_pg_nx | - PG_V | PG_U | PG_PS | PG_G; - } - for (i = 0; i < ndmpdp; i++) { - pdp = (paddr_t)&(((pd_entry_t *)dmpdp)[i]); - *pte = (pdp & PG_FRAME) | PG_V | PG_RW | pmap_pg_nx; - pmap_update_pg((vaddr_t)tmpva); - - pde = (pd_entry_t *)(tmpva + (pdp & ~PG_FRAME)); - *pde = (dmpd + (i << PAGE_SHIFT)) | PG_RW | pmap_pg_nx | - PG_V | PG_U; - } - } - - kpm->pm_pdir[PDIR_SLOT_DIRECT] = dmpdp | PG_KW | pmap_pg_nx | PG_V | PG_U; - - tlbflush(); - + pmap_init_directmap(kpm); #else if (VM_MIN_KERNEL_ADDRESS != KERNBASE) { /* @@ -1439,7 +1366,7 @@ pmap_bootstrap(vaddr_t kva_start) * mapping. */ - pte = PTE_BASE + pl1_i(virtual_avail); + pt_entry_t *pte = PTE_BASE + pl1_i(virtual_avail); #ifdef MULTIPROCESSOR /* @@ -1480,11 +1407,6 @@ pmap_bootstrap(vaddr_t kva_start) } #endif - /* - * Nothing after this point actually needs pte. - */ - pte = (void *)0xdeadbeef; - #ifdef XEN #ifdef __x86_64__ /* @@ -1561,6 +1483,101 @@ pmap_bootstrap(vaddr_t kva_start) pmap_maxkvaddr = kva; } + +#ifdef __HAVE_DIRECT_MAP +/* + * Create the amd64 direct map. Called only once at boot time. + */ +static void +pmap_init_directmap(struct pmap *kpm) +{ + extern phys_ram_seg_t mem_clusters[]; + extern int mem_cluster_cnt; + + paddr_t lastpa, dm_pd, dm_pdp, pdp; + vaddr_t tmpva; + pt_entry_t *pte; + pd_entry_t *pde; + phys_ram_seg_t *mc; + long n_dm_pdp; + int i; + + const pd_entry_t pteflags = PG_V | PG_KW | pmap_pg_nx; + + /* Get the last physical address available */ + lastpa = 0; + for (i = 0; i < mem_cluster_cnt; i++) { + mc = &mem_clusters[i]; + lastpa = MAX(lastpa, mc->start + mc->size); + } + + /* Allocate L3. */ + dm_pdp = pmap_bootstrap_palloc(1); + + /* Number of L3 entries. */ + n_dm_pdp = (lastpa + NBPD_L3 - 1) >> L3_SHIFT; + + tmpva = (KERNBASE + NKL2_KIMG_ENTRIES * NBPD_L2); + pte = PTE_BASE + pl1_i(tmpva); + *pte = dm_pdp | pteflags; + pmap_update_pg(tmpva); + memset((void *)tmpva, 0, PAGE_SIZE); + + /* + * Map the direct map RW. Use super pages (1GB) or large pages (2MB) if + * they are supported. Note: PG_G is not allowed on non-leaf PTPs. + */ + if (cpu_feature[2] & CPUID_P1GB) { + /* Super pages are supported. Just create L3. */ + for (i = 0; i < n_dm_pdp; i++) { + pdp = (paddr_t)&(((pd_entry_t *)dm_pdp)[i]); + *pte = (pdp & PG_FRAME) | pteflags; + pmap_update_pg(tmpva); + + pde = (pd_entry_t *)(tmpva + (pdp & ~PG_FRAME)); + *pde = ((paddr_t)i << L3_SHIFT) | pteflags | PG_U | + PG_PS | PG_G; + } + } else { + /* Allocate L2. */ + dm_pd = pmap_bootstrap_palloc(n_dm_pdp); + + /* Zero out the L2 pages. */ + for (i = 0; i < n_dm_pdp; i++) { + pdp = dm_pd + i * PAGE_SIZE; + *pte = (pdp & PG_FRAME) | pteflags; + pmap_update_pg(tmpva); + + memset((void *)tmpva, 0, PAGE_SIZE); + } + + for (i = 0; i < NPDPG * n_dm_pdp; i++) { + pdp = (paddr_t)&(((pd_entry_t *)dm_pd)[i]); + *pte = (pdp & PG_FRAME) | pteflags; + pmap_update_pg(tmpva); + + pde = (pd_entry_t *)(tmpva + (pdp & ~PG_FRAME)); + *pde = ((paddr_t)i << L2_SHIFT) | pteflags | + PG_U | PG_PS | PG_G; + } + + /* Fill in the L3 entries, linked to L2. */ + for (i = 0; i < n_dm_pdp; i++) { + pdp = (paddr_t)&(((pd_entry_t *)dm_pdp)[i]); + *pte = (pdp & PG_FRAME) | pteflags; + pmap_update_pg(tmpva); + + pde = (pd_entry_t *)(tmpva + (pdp & ~PG_FRAME)); + *pde = (dm_pd + (i << PAGE_SHIFT)) | pteflags | PG_U; + } + } + + kpm->pm_pdir[PDIR_SLOT_DIRECT] = dm_pdp | pteflags | PG_U; + + tlbflush(); +} +#endif /* __HAVE_DIRECT_MAP */ + #ifndef XEN /* * Remap several kernel segments with large pages. We cover as many pages as we