Module Name:    src
Committed By:   riastradh
Date:           Wed Aug 27 16:05:38 UTC 2014

Modified Files:
        src/sys/external/bsd/drm2/linux: linux_kmap.c

Log Message:
Convert linux_kmap_lock to an adaptive lock.

kmap/kunmap can't be used in interrupt context anyway, so there is no
need for this to be a spin lock.


To generate a diff of this commit:
cvs rdiff -u -r1.4 -r1.5 src/sys/external/bsd/drm2/linux/linux_kmap.c

Please note that diffs are not public domain; they are subject to the
copyright notices on the relevant files.

Modified files:

Index: src/sys/external/bsd/drm2/linux/linux_kmap.c
diff -u src/sys/external/bsd/drm2/linux/linux_kmap.c:1.4 src/sys/external/bsd/drm2/linux/linux_kmap.c:1.5
--- src/sys/external/bsd/drm2/linux/linux_kmap.c:1.4	Fri Mar 28 23:22:27 2014
+++ src/sys/external/bsd/drm2/linux/linux_kmap.c	Wed Aug 27 16:05:38 2014
@@ -1,4 +1,4 @@
-/*	$NetBSD: linux_kmap.c,v 1.4 2014/03/28 23:22:27 riastradh Exp $	*/
+/*	$NetBSD: linux_kmap.c,v 1.5 2014/08/27 16:05:38 riastradh Exp $	*/
 
 /*-
  * Copyright (c) 2013 The NetBSD Foundation, Inc.
@@ -30,7 +30,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: linux_kmap.c,v 1.4 2014/03/28 23:22:27 riastradh Exp $");
+__KERNEL_RCSID(0, "$NetBSD: linux_kmap.c,v 1.5 2014/08/27 16:05:38 riastradh Exp $");
 
 #include <sys/types.h>
 #include <sys/kmem.h>
@@ -106,7 +106,7 @@ int
 linux_kmap_init(void)
 {
 
-	/* IPL_VM is needed to block pmap_kenter_pa.  */
+	/* IPL_VM since interrupt handlers use kmap_atomic.  */
 	mutex_init(&linux_kmap_atomic_lock, MUTEX_DEFAULT, IPL_VM);
 
 	linux_kmap_atomic_vaddr = uvm_km_alloc(kernel_map, PAGE_SIZE, 0,
@@ -115,7 +115,7 @@ linux_kmap_init(void)
 	KASSERT(linux_kmap_atomic_vaddr != 0);
 	KASSERT(!pmap_extract(pmap_kernel(), linux_kmap_atomic_vaddr, NULL));
 
-	mutex_init(&linux_kmap_lock, MUTEX_DEFAULT, IPL_VM);
+	mutex_init(&linux_kmap_lock, MUTEX_DEFAULT, IPL_NONE);
 	rb_tree_init(&linux_kmap_entries, &linux_kmap_entry_ops);
 
 	return 0;
@@ -188,11 +188,11 @@ kmap(struct page *page)
 	lke->lke_paddr = paddr;
 	lke->lke_vaddr = vaddr;
 
-	mutex_spin_enter(&linux_kmap_lock);
+	mutex_enter(&linux_kmap_lock);
 	struct linux_kmap_entry *const collision __unused =
 	    rb_tree_insert_node(&linux_kmap_entries, lke);
 	KASSERT(collision == lke);
-	mutex_spin_exit(&linux_kmap_lock);
+	mutex_exit(&linux_kmap_lock);
 
 	KASSERT(!pmap_extract(pmap_kernel(), vaddr, NULL));
 	const int prot = (VM_PROT_READ | VM_PROT_WRITE);
@@ -208,12 +208,12 @@ kunmap(struct page *page)
 {
 	const paddr_t paddr = VM_PAGE_TO_PHYS(&page->p_vmp);
 
-	mutex_spin_enter(&linux_kmap_lock);
+	mutex_enter(&linux_kmap_lock);
 	struct linux_kmap_entry *const lke =
 	    rb_tree_find_node(&linux_kmap_entries, &paddr);
 	KASSERT(lke != NULL);
 	rb_tree_remove_node(&linux_kmap_entries, lke);
-	mutex_spin_exit(&linux_kmap_lock);
+	mutex_exit(&linux_kmap_lock);
 
 	const vaddr_t vaddr = lke->lke_vaddr;
 	kmem_free(lke, sizeof(*lke));

Reply via email to