Module Name:    src
Committed By:   cherry
Date:           Sat Aug 20 19:22:47 UTC 2011

Modified Files:
        src/sys/arch/amd64/amd64 [cherry-xenmp]: machdep.c
        src/sys/arch/amd64/include [cherry-xenmp]: pmap.h
        src/sys/arch/i386/i386 [cherry-xenmp]: machdep.c
        src/sys/arch/i386/include [cherry-xenmp]: pmap.h
        src/sys/arch/x86/include [cherry-xenmp]: cpu.h
        src/sys/arch/x86/x86 [cherry-xenmp]: pmap.c
        src/sys/arch/xen/include [cherry-xenmp]: hypervisor.h
        src/sys/arch/xen/x86 [cherry-xenmp]: cpu.c x86_xpmap.c xen_pmap.c

Log Message:
PAE MP support (preliminary), amd64 per-cpu L4 model redesigned, i386 
pmap_pa_start/end fixup


To generate a diff of this commit:
cvs rdiff -u -r1.160.2.2 -r1.160.2.3 src/sys/arch/amd64/amd64/machdep.c
cvs rdiff -u -r1.24.2.1 -r1.24.2.2 src/sys/arch/amd64/include/pmap.h
cvs rdiff -u -r1.702.2.4 -r1.702.2.5 src/sys/arch/i386/i386/machdep.c
cvs rdiff -u -r1.109.2.1 -r1.109.2.2 src/sys/arch/i386/include/pmap.h
cvs rdiff -u -r1.34.2.4 -r1.34.2.5 src/sys/arch/x86/include/cpu.h
cvs rdiff -u -r1.121.2.5 -r1.121.2.6 src/sys/arch/x86/x86/pmap.c
cvs rdiff -u -r1.31.10.1 -r1.31.10.2 src/sys/arch/xen/include/hypervisor.h
cvs rdiff -u -r1.56.2.6 -r1.56.2.7 src/sys/arch/xen/x86/cpu.c
cvs rdiff -u -r1.26.2.6 -r1.26.2.7 src/sys/arch/xen/x86/x86_xpmap.c
cvs rdiff -u -r1.2.2.2 -r1.2.2.3 src/sys/arch/xen/x86/xen_pmap.c

Please note that diffs are not public domain; they are subject to the
copyright notices on the relevant files.

Modified files:

Index: src/sys/arch/amd64/amd64/machdep.c
diff -u src/sys/arch/amd64/amd64/machdep.c:1.160.2.2 src/sys/arch/amd64/amd64/machdep.c:1.160.2.3
--- src/sys/arch/amd64/amd64/machdep.c:1.160.2.2	Thu Jun 23 14:18:56 2011
+++ src/sys/arch/amd64/amd64/machdep.c	Sat Aug 20 19:22:46 2011
@@ -1,4 +1,4 @@
-/*	$NetBSD: machdep.c,v 1.160.2.2 2011/06/23 14:18:56 cherry Exp $	*/
+/*	$NetBSD: machdep.c,v 1.160.2.3 2011/08/20 19:22:46 cherry Exp $	*/
 
 /*-
  * Copyright (c) 1996, 1997, 1998, 2000, 2006, 2007, 2008
@@ -107,7 +107,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: machdep.c,v 1.160.2.2 2011/06/23 14:18:56 cherry Exp $");
+__KERNEL_RCSID(0, "$NetBSD: machdep.c,v 1.160.2.3 2011/08/20 19:22:46 cherry Exp $");
 
 /* #define XENDEBUG_LOW  */
 
@@ -1320,8 +1320,8 @@
 	/* Determine physical address space */
 	avail_start = first_avail;
 	avail_end = ctob(xen_start_info.nr_pages);
-	pmap_pa_start = (KERNTEXTOFF - KERNBASE);
-	pmap_pa_end = avail_end;
+	pmap_pa_start = XPMAP_OFFSET;
+	pmap_pa_end = pmap_pa_start + ctob(xen_start_info.nr_pages);
 	__PRINTK(("pmap_pa_start 0x%lx avail_start 0x%lx avail_end 0x%lx\n",
 	    pmap_pa_start, avail_start, avail_end));
 #endif	/* !XEN */

Index: src/sys/arch/amd64/include/pmap.h
diff -u src/sys/arch/amd64/include/pmap.h:1.24.2.1 src/sys/arch/amd64/include/pmap.h:1.24.2.2
--- src/sys/arch/amd64/include/pmap.h:1.24.2.1	Fri Jun  3 13:27:38 2011
+++ src/sys/arch/amd64/include/pmap.h	Sat Aug 20 19:22:46 2011
@@ -1,4 +1,4 @@
-/*	$NetBSD: pmap.h,v 1.24.2.1 2011/06/03 13:27:38 cherry Exp $	*/
+/*	$NetBSD: pmap.h,v 1.24.2.2 2011/08/20 19:22:46 cherry Exp $	*/
 
 /*
  * Copyright (c) 1997 Charles D. Cranor and Washington University.
@@ -191,7 +191,8 @@
 #define AL4_BASE ((pd_entry_t *)((char *)AL3_BASE + L4_SLOT_PTE * NBPD_L1))
 
 #define PDP_PDE		(L4_BASE + PDIR_SLOT_PTE)
-#define APDP_PDE	(L4_BASE + PDIR_SLOT_APTE)
+#define APDP_PDE	(&curcpu()->ci_kpm_pdir[PDIR_SLOT_APTE])
+#define APDP_PDE_SHADOW	(L4_BASE + PDIR_SLOT_APTE)
 
 #define PDP_BASE	L4_BASE
 #define APDP_BASE	AL4_BASE

Index: src/sys/arch/i386/i386/machdep.c
diff -u src/sys/arch/i386/i386/machdep.c:1.702.2.4 src/sys/arch/i386/i386/machdep.c:1.702.2.5
--- src/sys/arch/i386/i386/machdep.c:1.702.2.4	Sun Aug  7 19:48:07 2011
+++ src/sys/arch/i386/i386/machdep.c	Sat Aug 20 19:22:46 2011
@@ -1,4 +1,4 @@
-/*	$NetBSD: machdep.c,v 1.702.2.4 2011/08/07 19:48:07 cherry Exp $	*/
+/*	$NetBSD: machdep.c,v 1.702.2.5 2011/08/20 19:22:46 cherry Exp $	*/
 
 /*-
  * Copyright (c) 1996, 1997, 1998, 2000, 2004, 2006, 2008, 2009
@@ -67,7 +67,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: machdep.c,v 1.702.2.4 2011/08/07 19:48:07 cherry Exp $");
+__KERNEL_RCSID(0, "$NetBSD: machdep.c,v 1.702.2.5 2011/08/20 19:22:46 cherry Exp $");
 
 #include "opt_beep.h"
 #include "opt_compat_ibcs2.h"
@@ -1437,7 +1437,7 @@
 	avail_start = first_avail;
 	avail_end = ctob(xen_start_info.nr_pages) + XPMAP_OFFSET;
 	pmap_pa_start = (KERNTEXTOFF - KERNBASE);
-	pmap_pa_end = avail_end;
+	pmap_pa_end = pmap_pa_start + ctob(xen_start_info.nr_pages);
 	mem_clusters[0].start = avail_start;
 	mem_clusters[0].size = avail_end - avail_start;
 	mem_cluster_cnt++;

Index: src/sys/arch/i386/include/pmap.h
diff -u src/sys/arch/i386/include/pmap.h:1.109.2.1 src/sys/arch/i386/include/pmap.h:1.109.2.2
--- src/sys/arch/i386/include/pmap.h:1.109.2.1	Sun Jul 31 20:49:11 2011
+++ src/sys/arch/i386/include/pmap.h	Sat Aug 20 19:22:46 2011
@@ -1,4 +1,4 @@
-/*	$NetBSD: pmap.h,v 1.109.2.1 2011/07/31 20:49:11 cherry Exp $	*/
+/*	$NetBSD: pmap.h,v 1.109.2.2 2011/08/20 19:22:46 cherry Exp $	*/
 
 /*
  * Copyright (c) 1997 Charles D. Cranor and Washington University.
@@ -277,7 +277,7 @@
  * mapping, because it points to the shadow PD. Use the kernel PD instead,
  * which is static
  */
-#define APDP_PDE	(&pmap_kl2pd[l2tol2(PDIR_SLOT_APTE)])
+#define APDP_PDE	(&curcpu()->ci_kpm_pdir[l2tol2(PDIR_SLOT_APTE)])
 #define APDP_PDE_SHADOW	(L2_BASE + PDIR_SLOT_APTE)
 #else /* PAE && XEN */
 #define APDP_PDE	(L2_BASE + PDIR_SLOT_APTE)
@@ -440,13 +440,6 @@
 
 #endif
 
-#ifdef PAE
-/* Address of the static kernel's L2 page */
-pd_entry_t *pmap_kl2pd;
-paddr_t pmap_kl2paddr;
-#endif
-
-
 struct trapframe;
 
 int	pmap_exec_fixup(struct vm_map *, struct trapframe *, struct pcb *);

Index: src/sys/arch/x86/include/cpu.h
diff -u src/sys/arch/x86/include/cpu.h:1.34.2.4 src/sys/arch/x86/include/cpu.h:1.34.2.5
--- src/sys/arch/x86/include/cpu.h:1.34.2.4	Wed Aug 17 09:40:39 2011
+++ src/sys/arch/x86/include/cpu.h	Sat Aug 20 19:22:47 2011
@@ -1,4 +1,4 @@
-/*	$NetBSD: cpu.h,v 1.34.2.4 2011/08/17 09:40:39 cherry Exp $	*/
+/*	$NetBSD: cpu.h,v 1.34.2.5 2011/08/20 19:22:47 cherry Exp $	*/
 
 /*-
  * Copyright (c) 1990 The Regents of the University of California.
@@ -180,12 +180,14 @@
 	pd_entry_t *	ci_pae_l3_pdir; /* VA pointer to L3 PD */
 #endif
 
-#if defined(XEN) && defined(__x86_64__)
+#if defined(XEN) && (defined(PAE) || defined(__x86_64__))
 	/* Currently active user PGD (can't use rcr3() with Xen) */
-	pd_entry_t *	ci_kpm_pdir;	/* per-cpu L4 PD (va) */
-	paddr_t		ci_kpm_pdirpa; /* per-cpu L4 PD (pa) */
+	pd_entry_t *	ci_kpm_pdir;	/* per-cpu PMD (va) */
+	paddr_t		ci_kpm_pdirpa; /* per-cpu PMD (pa) */
+#if defined(__x86_64__)
 	paddr_t		ci_xen_current_user_pgd;
-#endif
+#endif /* __x86_64__ */
+#endif /* XEN et.al */
 
 	char *ci_doubleflt_stack;
 	char *ci_ddbipi_stack;
@@ -232,11 +234,6 @@
 	int		ci_padout __aligned(64);
 };
 
-#ifdef __x86_64__
-#define ci_pdirpa(ci, index) \
-	((ci)->ci_kpm_pdirpa + (index) * sizeof(pd_entry_t))
-#endif /* __x86_64__ */
-
 /*
  * Macros to handle (some) trapframe registers for common x86 code.
  */

Index: src/sys/arch/x86/x86/pmap.c
diff -u src/sys/arch/x86/x86/pmap.c:1.121.2.5 src/sys/arch/x86/x86/pmap.c:1.121.2.6
--- src/sys/arch/x86/x86/pmap.c:1.121.2.5	Wed Aug 17 09:40:39 2011
+++ src/sys/arch/x86/x86/pmap.c	Sat Aug 20 19:22:47 2011
@@ -1,4 +1,4 @@
-/*	$NetBSD: pmap.c,v 1.121.2.5 2011/08/17 09:40:39 cherry Exp $	*/
+/*	$NetBSD: pmap.c,v 1.121.2.6 2011/08/20 19:22:47 cherry Exp $	*/
 
 /*-
  * Copyright (c) 2008, 2010 The NetBSD Foundation, Inc.
@@ -171,7 +171,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.121.2.5 2011/08/17 09:40:39 cherry Exp $");
+__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.121.2.6 2011/08/20 19:22:47 cherry Exp $");
 
 #include "opt_user_ldt.h"
 #include "opt_lockdebug.h"
@@ -1503,10 +1503,18 @@
 			HYPERVISOR_update_va_mapping (newp + KERNBASE,
 			    xpmap_ptom_masked(newp) | PG_u | PG_V, UVMF_INVLPG);
 		xpq_queue_lock();
+		/* Update the pmap_kernel() L4 shadow */
 		xpq_queue_pte_update (
 		    xpmap_ptom_masked(pdes_pa)
 		    + (pl_i(0, level) * sizeof (pd_entry_t)),
 		    xpmap_ptom_masked(newp) | PG_RW | PG_u | PG_V);
+
+		/* sync to per-cpu PD */
+		xpq_queue_pte_update(
+			xpmap_ptom_masked(cpu_info_primary.ci_kpm_pdirpa +
+			    pl_i(0, PTP_LEVELS) *
+			    sizeof(pd_entry_t)),
+			pmap_kernel()->pm_pdir[pl_i(0, PTP_LEVELS)]);
 		xpq_queue_unlock();
 		pmap_pte_flush();
 
@@ -1515,15 +1523,6 @@
 			break;
 		pdes_pa = newp;
 	}
-
-	/* sync to per-cpu PD */
-	xpq_queue_lock();
-	xpq_queue_pte_update(
-		xpmap_ptom_masked(ci_pdirpa(&cpu_info_primary,
-					    pl_i(0, PTP_LEVELS))),
-				  pmap_kernel()->pm_pdir[pl_i(0, PTP_LEVELS)]);
-	xpq_queue_unlock();
-	pmap_pte_flush();
 #else /* XEN */
 	pd_entry_t *pdes;
 
@@ -1580,7 +1579,7 @@
 /*
  * pmap_cpu_init_late: perform late per-CPU initialization.
  */
-
+#ifndef XEN
 void
 pmap_cpu_init_late(struct cpu_info *ci)
 {
@@ -1620,31 +1619,8 @@
 
 	pmap_update(pmap_kernel());
 #endif
-#if defined(XEN) && defined (__x86_64)
-	KASSERT(ci != NULL);
-
-	ci->ci_kpm_pdir = (pd_entry_t *)uvm_km_alloc(kernel_map,
-						     PAGE_SIZE, 0, UVM_KMF_WIRED | UVM_KMF_ZERO | UVM_KMF_NOWAIT);
-	if (ci->ci_kpm_pdir == NULL) {
-		panic("%s: failed to allocate L4 per-cpu PD for CPU %d\n",
-		      __func__, cpu_index(ci));
-	}
-	ci->ci_kpm_pdirpa = vtophys((vaddr_t) ci->ci_kpm_pdir);
-	KASSERT(ci->ci_kpm_pdirpa != 0);
-
-	cpu_load_pmap(pmap_kernel());
-
-	pmap_kenter_pa((vaddr_t)ci->ci_kpm_pdir, ci->ci_kpm_pdirpa,
-		VM_PROT_READ, 0);
-
-	pmap_update(pmap_kernel());
-
-	xpq_queue_lock();
-	xpq_queue_pin_l4_table(xpmap_ptom_masked(ci->ci_kpm_pdirpa));
-	xpq_queue_unlock();
-
-#endif /* defined(XEN) && defined (__x86_64__) */
 }
+#endif
 
 /*
  * p v _ e n t r y   f u n c t i o n s
@@ -4163,7 +4139,6 @@
 #ifdef XEN
 	int s = splvm(); /* protect xpq_* */
 #endif
-
 	for (level = lvl; level > 1; level--) {
 		if (level == PTP_LEVELS){
 			pdep = pmap_kernel()->pm_pdir;
@@ -4179,6 +4154,7 @@
 
 		for (i = index; i <= endindex; i++) {
 			pt_entry_t pte;
+
 			KASSERT(!pmap_valid_entry(pdep[i]));
 			pmap_get_physpage(va, &pa);
 			pte = pmap_pa2pte(pa) | PG_k | PG_V | PG_RW;
@@ -4186,18 +4162,25 @@
 			xpq_queue_lock();
 			switch (level) {
 			case PTP_LEVELS: 
-#ifdef __x86_64__
-				/* update the per-cpu L4 */
-				xpq_queue_pte_update(
-					xpmap_ptom(ci_pdirpa(&cpu_info_primary, i)), pte);
-#endif /* __x86_64__ */
+#if defined(PAE) || defined(__x86_64__)
+				if (i >= PDIR_SLOT_KERN) {
+					/* update per-cpu PMDs on all cpus */
+					CPU_INFO_ITERATOR cii;
+					struct cpu_info *ci;
+					for (CPU_INFO_FOREACH(cii, ci)) {
+						if (ci == NULL) {
+							continue;
+						}
 #ifdef PAE
-				if (i > L2_SLOT_KERN) {
-					/* update real kernel PD too */
-					xpq_queue_pte_update(
-						xpmap_ptetomach(&pmap_kl2pd[l2tol2(i)]), pte);
+						xpq_queue_pte_update(
+							xpmap_ptetomach(&ci->ci_kpm_pdir[l2tol2(i)]), pte);
+#elif defined(__x86_64__)
+						xpq_queue_pte_update(
+							xpmap_ptetomach(&ci->ci_kpm_pdir[i]), pte);
+#endif /* PAE */
+					}
 				}
-#endif
+#endif /* PAE || __x86_64__ */
 				/* FALLTHROUGH */
 
 			default: /* All other levels */

Index: src/sys/arch/xen/include/hypervisor.h
diff -u src/sys/arch/xen/include/hypervisor.h:1.31.10.1 src/sys/arch/xen/include/hypervisor.h:1.31.10.2
--- src/sys/arch/xen/include/hypervisor.h:1.31.10.1	Thu Aug  4 09:07:46 2011
+++ src/sys/arch/xen/include/hypervisor.h	Sat Aug 20 19:22:47 2011
@@ -1,4 +1,4 @@
-/*	$NetBSD: hypervisor.h,v 1.31.10.1 2011/08/04 09:07:46 cherry Exp $	*/
+/*	$NetBSD: hypervisor.h,v 1.31.10.2 2011/08/20 19:22:47 cherry Exp $	*/
 
 /*
  * Copyright (c) 2006 Manuel Bouyer.
@@ -129,6 +129,7 @@
 
 /* hypervisor.c */
 struct intrframe;
+struct cpu_info;
 void do_hypervisor_callback(struct intrframe *regs);
 void hypervisor_enable_event(unsigned int);
 

Index: src/sys/arch/xen/x86/cpu.c
diff -u src/sys/arch/xen/x86/cpu.c:1.56.2.6 src/sys/arch/xen/x86/cpu.c:1.56.2.7
--- src/sys/arch/xen/x86/cpu.c:1.56.2.6	Wed Aug 17 09:40:39 2011
+++ src/sys/arch/xen/x86/cpu.c	Sat Aug 20 19:22:47 2011
@@ -1,4 +1,4 @@
-/*	$NetBSD: cpu.c,v 1.56.2.6 2011/08/17 09:40:39 cherry Exp $	*/
+/*	$NetBSD: cpu.c,v 1.56.2.7 2011/08/20 19:22:47 cherry Exp $	*/
 /* NetBSD: cpu.c,v 1.18 2004/02/20 17:35:01 yamt Exp  */
 
 /*-
@@ -66,7 +66,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: cpu.c,v 1.56.2.6 2011/08/17 09:40:39 cherry Exp $");
+__KERNEL_RCSID(0, "$NetBSD: cpu.c,v 1.56.2.7 2011/08/20 19:22:47 cherry Exp $");
 
 #include "opt_ddb.h"
 #include "opt_multiprocessor.h"
@@ -1022,7 +1022,11 @@
 	 * Use pmap_kernel() L4 PD directly, until we setup the
 	 * per-cpu L4 PD in pmap_cpu_init_late()
 	 */
+#ifdef PAE
+	initctx->ctrlreg[3] = xpmap_ptom(ci->ci_pae_l3_pdirpa);
+#else /* PAE */
 	initctx->ctrlreg[3] = xpmap_ptom(pcb->pcb_cr3);
+#endif /* PAE */
 	initctx->ctrlreg[4] = /* CR4_PAE |  */CR4_OSFXSR | CR4_OSXMMEXCPT;
 
 
@@ -1262,20 +1266,12 @@
 	new_pgd = pmap->pm_pdir;
 
 	xpq_queue_lock();
-	/* Copy source pmap L4 PDEs (in user addr. range) to shadow */
-	for (i = 0; i < PDIR_SLOT_PTE; i++) {
-		xpq_queue_pte_update(l3_shadow_pa + i * sizeof(pd_entry_t), new_pgd[i]);
-	}
 
-	/* Copy kernel mappings */
-	new_pgd = pmap_kernel()->pm_pdir;
-	for (i = PDIR_SLOT_KERN; i < nkptp[PTP_LEVELS - 1]; i++) {
+	/* Copy user pmap L4 PDEs (in user addr. range) to per-cpu L4 */
+	for (i = 0; i < PDIR_SLOT_PTE; i++) {
 		xpq_queue_pte_update(l3_shadow_pa + i * sizeof(pd_entry_t), new_pgd[i]);
 	}
 
-	xpq_queue_unlock();
-	tlbflush();
-	xpq_queue_lock();
 	if (__predict_true(pmap != pmap_kernel())) {
 		xen_set_user_pgd(pmap_pdirpa(pmap, 0));
 		ci->ci_xen_current_user_pgd = pmap_pdirpa(pmap, 0);
@@ -1285,12 +1281,120 @@
 		ci->ci_xen_current_user_pgd = 0;
 	}
 	xpq_queue_unlock();
+
+	tlbflush();
+
 	splx(s);
 
 #endif /* __x86_64__ */
 }
 
 /*
+ * pmap_cpu_init_late: perform late per-CPU initialization.
+ */
+
+void
+pmap_cpu_init_late(struct cpu_info *ci)
+{
+#if defined(PAE) || defined(__x86_64__)
+	/*
+	 * The BP has already its own PD page allocated during early
+	 * MD startup.
+	 */
+
+	if (ci == &cpu_info_primary)
+		return;
+
+	KASSERT(ci != NULL);
+
+#ifdef PAE
+	{
+		int ret;
+		struct pglist pg;
+		struct vm_page *vmap;
+
+		/*
+		 * Allocate a page for the per-CPU L3 PD. cr3 being 32 bits, PA musts
+		 * resides below the 4GB boundary.
+		 */
+		ret = uvm_pglistalloc(PAGE_SIZE, 0,
+		    0x100000000ULL, 32, 0, &pg, 1, 0);
+
+		vmap = TAILQ_FIRST(&pg);
+
+		if (ret != 0 || vmap == NULL)
+			panic("%s: failed to allocate L3 pglist for CPU %d (ret %d)\n",
+			    __func__, cpu_index(ci), ret);
+
+		ci->ci_pae_l3_pdirpa = vmap->phys_addr;
+
+		ci->ci_pae_l3_pdir = (paddr_t *)uvm_km_alloc(kernel_map, PAGE_SIZE, 0,
+		    UVM_KMF_VAONLY | UVM_KMF_NOWAIT);
+
+		if (ci->ci_pae_l3_pdir == NULL)
+			panic("%s: failed to allocate L3 PD for CPU %d\n",
+			    __func__, cpu_index(ci));
+		pmap_kenter_pa((vaddr_t)ci->ci_pae_l3_pdir, ci->ci_pae_l3_pdirpa,
+		    VM_PROT_READ | VM_PROT_WRITE, 0);
+	}
+	/* Initialise L2 entries 0 - 2: Point them to pmap_kernel() */
+	ci->ci_pae_l3_pdir[0] =
+	    xpmap_ptom_masked(pmap_kernel()->pm_pdirpa[0]) | PG_V;
+	ci->ci_pae_l3_pdir[1] =
+	    xpmap_ptom_masked(pmap_kernel()->pm_pdirpa[1]) | PG_V;
+	ci->ci_pae_l3_pdir[2] =
+	    xpmap_ptom_masked(pmap_kernel()->pm_pdirpa[2]) | PG_V;
+#endif /* PAE */
+
+	ci->ci_kpm_pdir = (pd_entry_t *)uvm_km_alloc(kernel_map, PAGE_SIZE, 0,
+	    UVM_KMF_WIRED | UVM_KMF_ZERO | UVM_KMF_NOWAIT);
+
+	if (ci->ci_kpm_pdir == NULL) {
+		panic("%s: failed to allocate L4 per-cpu PD for CPU %d\n",
+		      __func__, cpu_index(ci));
+	}
+	ci->ci_kpm_pdirpa = vtophys((vaddr_t) ci->ci_kpm_pdir);
+	KASSERT(ci->ci_kpm_pdirpa != 0);
+	/*
+	 * Copy over kernel pmd entries from boot
+	 * cpu. XXX:locking/races
+	 */
+
+	memcpy(ci->ci_kpm_pdir,
+	    &pmap_kernel()->pm_pdir[PDIR_SLOT_KERN],
+	    nkptp[PTP_LEVELS - 1] * sizeof(pd_entry_t));
+
+	/* Xen wants R/O */
+	pmap_kenter_pa((vaddr_t)ci->ci_kpm_pdir, ci->ci_kpm_pdirpa,
+	    VM_PROT_READ, 0);
+
+#ifdef PAE
+	/* Initialise L3 entry 3. This mapping is shared across all
+	 * pmaps and is static, ie; loading a new pmap will not update
+	 * this entry.
+	 */
+	
+	ci->ci_pae_l3_pdir[3] = xpmap_ptom_masked(ci->ci_kpm_pdirpa) | PG_V;
+
+	/* Mark L3 R/O (Xen wants this) */
+	pmap_kenter_pa((vaddr_t)ci->ci_pae_l3_pdir, ci->ci_pae_l3_pdirpa,
+		VM_PROT_READ, 0);
+
+	xpq_queue_lock();
+	xpq_queue_pin_l3_table(xpmap_ptom_masked(ci->ci_pae_l3_pdirpa));
+	xpq_queue_unlock();
+
+#elif defined(__x86_64__)	
+	xpq_queue_lock();
+	xpq_queue_pin_l4_table(xpmap_ptom_masked(ci->ci_kpm_pdirpa));
+	xpq_queue_unlock();
+#endif /* PAE */
+#endif /* defined(PAE) || defined(__x86_64__) */
+}
+
+
+
+/*
  * Notify all other cpus to halt.
  */
 
@@ -1307,5 +1411,5 @@
 void
 cpu_kick(struct cpu_info *ci)
 {
-	xen_send_ipi(ci, XEN_IPI_KICK);
+	(void)xen_send_ipi(ci, XEN_IPI_KICK);
 }

Index: src/sys/arch/xen/x86/x86_xpmap.c
diff -u src/sys/arch/xen/x86/x86_xpmap.c:1.26.2.6 src/sys/arch/xen/x86/x86_xpmap.c:1.26.2.7
--- src/sys/arch/xen/x86/x86_xpmap.c:1.26.2.6	Wed Aug 17 09:40:40 2011
+++ src/sys/arch/xen/x86/x86_xpmap.c	Sat Aug 20 19:22:47 2011
@@ -1,4 +1,4 @@
-/*	$NetBSD: x86_xpmap.c,v 1.26.2.6 2011/08/17 09:40:40 cherry Exp $	*/
+/*	$NetBSD: x86_xpmap.c,v 1.26.2.7 2011/08/20 19:22:47 cherry Exp $	*/
 
 /*
  * Copyright (c) 2006 Mathieu Ropert <[email protected]>
@@ -69,7 +69,7 @@
 
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: x86_xpmap.c,v 1.26.2.6 2011/08/17 09:40:40 cherry Exp $");
+__KERNEL_RCSID(0, "$NetBSD: x86_xpmap.c,v 1.26.2.7 2011/08/20 19:22:47 cherry Exp $");
 
 #include "opt_xen.h"
 #include "opt_ddb.h"
@@ -751,8 +751,9 @@
 	memset (bt_pgd, 0, PAGE_SIZE);
 	avail = new_pgd + PAGE_SIZE;
 #if PTP_LEVELS > 3
-	/* per-cpu "shadow" pmd */
+	/* per-cpu L4 PD */
 	pd_entry_t *bt_cpu_pgd = bt_pgd;
+	/* pmap_kernel() "shadow" L4 PD */
 	bt_pgd = (pd_entry_t *) avail;
 	memset(bt_pgd, 0, PAGE_SIZE);
 	avail += PAGE_SIZE;
@@ -909,8 +910,9 @@
 	 * pde[L2_SLOT_KERN] always point to the shadow.
 	 */
 	memcpy(&pde[L2_SLOT_KERN + NPDPG], &pde[L2_SLOT_KERN], PAGE_SIZE);
-	pmap_kl2pd = &pde[L2_SLOT_KERN + NPDPG];
-	pmap_kl2paddr = (u_long)pmap_kl2pd - KERNBASE;
+	cpu_info_primary.ci_kpm_pdir = &pde[L2_SLOT_KERN + NPDPG];
+	cpu_info_primary.ci_kpm_pdirpa =
+	    (vaddr_t) cpu_info_primary.ci_kpm_pdir - KERNBASE;
 
 	/*
 	 * We don't enter a recursive entry from the L3 PD. Instead,
@@ -1012,7 +1014,7 @@
 	}
 #elif defined(__x86_64__)
 	if (final) {
-		/* save the address of the shadow L4 pgd page */
+		/* save the address of the real per-cpu L4 pgd page */
 		cpu_info_primary.ci_kpm_pdir = bt_cpu_pgd;
 		cpu_info_primary.ci_kpm_pdirpa = ((paddr_t) bt_cpu_pgd - KERNBASE);
 	}

Index: src/sys/arch/xen/x86/xen_pmap.c
diff -u src/sys/arch/xen/x86/xen_pmap.c:1.2.2.2 src/sys/arch/xen/x86/xen_pmap.c:1.2.2.3
--- src/sys/arch/xen/x86/xen_pmap.c:1.2.2.2	Sat Jul 16 10:59:46 2011
+++ src/sys/arch/xen/x86/xen_pmap.c	Sat Aug 20 19:22:47 2011
@@ -1,4 +1,4 @@
-/*	$NetBSD: xen_pmap.c,v 1.2.2.2 2011/07/16 10:59:46 cherry Exp $	*/
+/*	$NetBSD: xen_pmap.c,v 1.2.2.3 2011/08/20 19:22:47 cherry Exp $	*/
 
 /*
  * Copyright (c) 2007 Manuel Bouyer.
@@ -102,7 +102,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: xen_pmap.c,v 1.2.2.2 2011/07/16 10:59:46 cherry Exp $");
+__KERNEL_RCSID(0, "$NetBSD: xen_pmap.c,v 1.2.2.3 2011/08/20 19:22:47 cherry Exp $");
 
 #include "opt_user_ldt.h"
 #include "opt_lockdebug.h"
@@ -183,10 +183,10 @@
 	int i;
 
 	for (i = 0; i < PDP_SIZE; i++) {
-		pmap_pte_set(APDP_PDE+i, 0);
-#if defined (PAE)
+		pmap_pte_set(APDP_PDE + i, 0);
+#if defined (PAE) || defined(__x86_64__)
 		/* clear shadow entries too */
-		pmap_pte_set(APDP_PDE_SHADOW+i, 0);
+		pmap_pte_set(APDP_PDE_SHADOW + i, 0);
 #endif
 	}
 }
@@ -212,6 +212,7 @@
 
 	/* the kernel's pmap is always accessible */
 	if (pmap == pmap_kernel()) {
+		mutex_enter(pmap->pm_lock);
 		*pmap2 = NULL;
 		*ptepp = PTE_BASE;
 		*pdeppp = normal_pdes;
@@ -276,27 +277,19 @@
 			npde = pmap_pa2pte(
 			    pmap_pdirpa(pmap, i * NPDPG)) | PG_k | PG_V;
 
-			xpq_queue_pte_update(xpmap_ptetomach(&APDP_PDE[i]),
-			    npde);
-
-			/* APDP_PDE is per-cpu */
-			xpq_queue_invlpg((vaddr_t) &APDP_PDE[i]);
-
-			/* 
-			 * Install temporary recursive mapping L4 in
-			 * the user pmap. XXX: What's this for ?
-			 */
 			xpq_queue_pte_update(
 			    xpmap_ptom(pmap_pdirpa(pmap, PDIR_SLOT_PTE + i)),
 			    npde);
+			xpq_queue_pte_update(xpmap_ptetomach(&APDP_PDE[i]),
+			    npde);
 
-			xen_bcast_invlpg((vaddr_t) &pmap->pm_pdir[PDIR_SLOT_PTE + i]);
-
-#ifdef PAE
+#if defined(PAE) || defined(__x86_64__)
 			/* update shadow entry too */
 			xpq_queue_pte_update(
 			    xpmap_ptetomach(&APDP_PDE_SHADOW[i]), npde);
-#endif /* PAE */
+#endif /* PAE || __x86_64__ */
+			xpq_queue_invlpg(
+			    (vaddr_t)&pmap->pm_pdir[PDIR_SLOT_PTE + i]);
 
 		}
 		if (pmap_valid_entry(opde))
@@ -334,6 +327,7 @@
 {
 
 	if (pmap == pmap_kernel()) {
+		mutex_exit(pmap->pm_lock);
 		return;
 	}
 	KASSERT(kpreempt_disabled());

Reply via email to