Module Name:    src
Committed By:   maxv
Date:           Sun Aug  7 10:07:58 UTC 2016

Modified Files:
        src/sys/uvm: uvm_mmap.c

Log Message:
KNF a little.


To generate a diff of this commit:
cvs rdiff -u -r1.160 -r1.161 src/sys/uvm/uvm_mmap.c

Please note that diffs are not public domain; they are subject to the
copyright notices on the relevant files.

Modified files:

Index: src/sys/uvm/uvm_mmap.c
diff -u src/sys/uvm/uvm_mmap.c:1.160 src/sys/uvm/uvm_mmap.c:1.161
--- src/sys/uvm/uvm_mmap.c:1.160	Sun Aug  7 09:55:18 2016
+++ src/sys/uvm/uvm_mmap.c	Sun Aug  7 10:07:58 2016
@@ -1,4 +1,4 @@
-/*	$NetBSD: uvm_mmap.c,v 1.160 2016/08/07 09:55:18 maxv Exp $	*/
+/*	$NetBSD: uvm_mmap.c,v 1.161 2016/08/07 10:07:58 maxv Exp $	*/
 
 /*
  * Copyright (c) 1997 Charles D. Cranor and Washington University.
@@ -46,7 +46,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: uvm_mmap.c,v 1.160 2016/08/07 09:55:18 maxv Exp $");
+__KERNEL_RCSID(0, "$NetBSD: uvm_mmap.c,v 1.161 2016/08/07 10:07:58 maxv Exp $");
 
 #include "opt_compat_netbsd.h"
 #include "opt_pax.h"
@@ -64,7 +64,7 @@ __KERNEL_RCSID(0, "$NetBSD: uvm_mmap.c,v
 #include <uvm/uvm_device.h>
 
 static int uvm_mmap(struct vm_map *, vaddr_t *, vsize_t, vm_prot_t, vm_prot_t,
-		    int, int, struct uvm_object *, voff_t, vsize_t);
+    int, int, struct uvm_object *, voff_t, vsize_t);
 
 static int
 range_test(struct vm_map *map, vaddr_t addr, vsize_t size, bool ismmap)
@@ -104,7 +104,7 @@ sys_sbrk(struct lwp *l, const struct sys
 		syscallarg(intptr_t) incr;
 	} */
 
-	return (ENOSYS);
+	return ENOSYS;
 }
 
 /*
@@ -119,7 +119,7 @@ sys_sstk(struct lwp *l, const struct sys
 		syscallarg(int) incr;
 	} */
 
-	return (ENOSYS);
+	return ENOSYS;
 }
 
 /*
@@ -155,11 +155,11 @@ sys_mincore(struct lwp *l, const struct 
 	vec = SCARG(uap, vec);
 
 	if (start & PAGE_MASK)
-		return (EINVAL);
+		return EINVAL;
 	len = round_page(len);
 	end = start + len;
 	if (end <= start)
-		return (EINVAL);
+		return EINVAL;
 
 	/*
 	 * Lock down vec, so our returned status isn't outdated by
@@ -259,7 +259,7 @@ sys_mincore(struct lwp *l, const struct 
  out:
 	vm_map_unlock_read(map);
 	uvm_vsunlock(p->p_vmspace, SCARG(uap, vec), npgs);
-	return (error);
+	return error;
 }
 
 /*
@@ -328,7 +328,7 @@ sys_mmap(struct lwp *l, const struct sys
 #endif
 	}
 	if ((flags & (MAP_SHARED|MAP_PRIVATE)) == (MAP_SHARED|MAP_PRIVATE))
-		return (EINVAL);
+		return EINVAL;
 
 	/*
 	 * align file position and save offset.  adjust size.
@@ -340,26 +340,23 @@ sys_mmap(struct lwp *l, const struct sys
 	newsize = (vsize_t)round_page(newsize);	/* round up */
 
 	if (newsize < size)
-		return (ENOMEM);
+		return ENOMEM;
 	size = newsize;
 
 	/*
 	 * now check (MAP_FIXED) or get (!MAP_FIXED) the "addr"
 	 */
 	if (flags & MAP_FIXED) {
-
 		/* ensure address and file offset are aligned properly */
 		addr -= pageoff;
 		if (addr & PAGE_MASK)
-			return (EINVAL);
+			return EINVAL;
 
 		error = range_test(&p->p_vmspace->vm_map, addr, size, true);
 		if (error) {
 			return error;
 		}
-
 	} else if (addr == 0 || !(flags & MAP_TRYFIXED)) {
-
 		/*
 		 * not fixed: make sure we skip over the largest
 		 * possible heap for non-topdown mapping arrangements.
@@ -371,8 +368,7 @@ sys_mmap(struct lwp *l, const struct sys
 		    (vaddr_t)p->p_vmspace->vm_daddr, size,
 		    p->p_vmspace->vm_map.flags & VM_MAP_TOPDOWN);
 
-		if (addr == 0 ||
-		    !(p->p_vmspace->vm_map.flags & VM_MAP_TOPDOWN))
+		if (addr == 0 || !(p->p_vmspace->vm_map.flags & VM_MAP_TOPDOWN))
 			addr = MAX(addr, defaddr);
 		else
 			addr = MIN(addr, defaddr);
@@ -385,14 +381,14 @@ sys_mmap(struct lwp *l, const struct sys
 	advice = UVM_ADV_NORMAL;
 	if ((flags & MAP_ANON) == 0) {
 		if ((fp = fd_getfile(fd)) == NULL)
-			return (EBADF);
+			return EBADF;
 
 		if (fp->f_ops->fo_mmap == NULL) {
 			error = ENODEV;
 			goto out;
 		}
 		error = (*fp->f_ops->fo_mmap)(fp, &pos, size, prot, &flags,
-					      &advice, &uobj, &maxprot);
+		    &advice, &uobj, &maxprot);
 		if (error) {
 			goto out;
 		}
@@ -407,7 +403,7 @@ sys_mmap(struct lwp *l, const struct sys
 		 * XXX What do we do about (MAP_SHARED|MAP_PRIVATE) == 0?
 		 */
 		if (fd != -1)
-			return (EINVAL);
+			return EINVAL;
 
  is_anon:		/* label for SunOS style /dev/zero */
 		uobj = NULL;
@@ -430,10 +426,10 @@ sys_mmap(struct lwp *l, const struct sys
 	*retval = (register_t)(addr + pageoff);
 
  out:
-     	if (fp != NULL)
+	if (fp != NULL)
 		fd_putfile(fd);
 
-	return (error);
+	return error;
 }
 
 /*
@@ -468,7 +464,7 @@ sys___msync13(struct lwp *l, const struc
 	if ((flags & ~(MS_ASYNC | MS_SYNC | MS_INVALIDATE)) != 0 ||
 	    (flags & (MS_ASYNC | MS_SYNC | MS_INVALIDATE)) == 0 ||
 	    (flags & (MS_ASYNC | MS_SYNC)) == (MS_ASYNC | MS_SYNC))
-		return (EINVAL);
+		return EINVAL;
 	if ((flags & (MS_ASYNC | MS_SYNC)) == 0)
 		flags |= MS_SYNC;
 
@@ -513,7 +509,7 @@ sys___msync13(struct lwp *l, const struc
 		}
 		vm_map_unlock_read(map);
 		if (rv == false)
-			return (EINVAL);
+			return EINVAL;
 	}
 
 	/*
@@ -565,7 +561,7 @@ sys_munmap(struct lwp *l, const struct s
 	size = (vsize_t)round_page(size);
 
 	if (size == 0)
-		return (0);
+		return 0;
 
 	map = &p->p_vmspace->vm_map;
 
@@ -573,23 +569,22 @@ sys_munmap(struct lwp *l, const struct s
 	if (error)
 		return EINVAL;
 
+	vm_map_lock(map);
+#if 0
 	/*
 	 * interesting system call semantic: make sure entire range is
 	 * allocated before allowing an unmap.
 	 */
-
-	vm_map_lock(map);
-#if 0
 	if (!uvm_map_checkprot(map, addr, addr + size, VM_PROT_NONE)) {
 		vm_map_unlock(map);
-		return (EINVAL);
+		return EINVAL;
 	}
 #endif
 	uvm_unmap_remove(map, addr, addr + size, &dead_entries, 0);
 	vm_map_unlock(map);
 	if (dead_entries != NULL)
 		uvm_unmap_detach(dead_entries, 0);
-	return (0);
+	return 0;
 }
 
 /*
@@ -674,7 +669,7 @@ sys_minherit(struct lwp *l, const struct
 		return EINVAL;
 
 	error = uvm_map_inherit(&p->p_vmspace->vm_map, addr, addr + size,
-				inherit);
+	    inherit);
 	return error;
 }
 
@@ -769,10 +764,10 @@ sys_madvise(struct lwp *l, const struct 
 		 * There's also what to do for device/file/anonymous memory.
 		 */
 
-		return (EINVAL);
+		return EINVAL;
 
 	default:
-		return (EINVAL);
+		return EINVAL;
 	}
 
 	return error;
@@ -815,11 +810,11 @@ sys_mlock(struct lwp *l, const struct sy
 		return ENOMEM;
 
 	if (atop(size) + uvmexp.wired > uvmexp.wiredmax)
-		return (EAGAIN);
+		return EAGAIN;
 
 	if (size + ptoa(pmap_wired_count(vm_map_pmap(&p->p_vmspace->vm_map))) >
-			p->p_rlimit[RLIMIT_MEMLOCK].rlim_cur)
-		return (EAGAIN);
+	    p->p_rlimit[RLIMIT_MEMLOCK].rlim_cur)
+		return EAGAIN;
 
 	error = uvm_map_pageable(&p->p_vmspace->vm_map, addr, addr+size, false,
 	    0);
@@ -888,13 +883,12 @@ sys_mlockall(struct lwp *l, const struct
 
 	flags = SCARG(uap, flags);
 
-	if (flags == 0 ||
-	    (flags & ~(MCL_CURRENT|MCL_FUTURE)) != 0)
-		return (EINVAL);
+	if (flags == 0 || (flags & ~(MCL_CURRENT|MCL_FUTURE)) != 0)
+		return EINVAL;
 
 	error = uvm_map_pageable_all(&p->p_vmspace->vm_map, flags,
 	    p->p_rlimit[RLIMIT_MEMLOCK].rlim_cur);
-	return (error);
+	return error;
 }
 
 /*
@@ -907,7 +901,7 @@ sys_munlockall(struct lwp *l, const void
 	struct proc *p = l->l_proc;
 
 	(void) uvm_map_pageable_all(&p->p_vmspace->vm_map, 0, 0);
-	return (0);
+	return 0;
 }
 
 /*
@@ -932,11 +926,11 @@ uvm_mmap(struct vm_map *map, vaddr_t *ad
 	 */
 
 	if (size == 0)
-		return(0);
+		return 0;
 	if (foff & PAGE_MASK)
-		return(EINVAL);
+		return EINVAL;
 	if ((prot & maxprot) != prot)
-		return(EINVAL);
+		return EINVAL;
 
 	/*
 	 * for non-fixed mappings, round off the suggested address.
@@ -947,7 +941,7 @@ uvm_mmap(struct vm_map *map, vaddr_t *ad
 		*addr = round_page(*addr);
 	} else {
 		if (*addr & PAGE_MASK)
-			return(EINVAL);
+			return EINVAL;
 		uvmflag |= UVM_FLAG_FIXED;
 		(void) uvm_unmap(map, *addr, *addr + size);
 	}
@@ -963,15 +957,15 @@ uvm_mmap(struct vm_map *map, vaddr_t *ad
 	align = (flags & MAP_ALIGNMENT_MASK) >> MAP_ALIGNMENT_SHIFT;
 	if (align) {
 		if (align >= sizeof(vaddr_t) * NBBY)
-			return(EINVAL);
+			return EINVAL;
 		align = 1L << align;
 		if (align < PAGE_SIZE)
-			return(EINVAL);
+			return EINVAL;
 		if (align >= vm_map_max(map))
-			return(ENOMEM);
+			return ENOMEM;
 		if (flags & MAP_FIXED) {
 			if ((*addr & (align-1)) != 0)
-				return(EINVAL);
+				return EINVAL;
 			align = 0;
 		}
 	}
@@ -1008,8 +1002,8 @@ uvm_mmap(struct vm_map *map, vaddr_t *ad
 	}
 
 	uvmflag = UVM_MAPFLAG(prot, maxprot,
-			(flags & MAP_SHARED) ? UVM_INH_SHARE : UVM_INH_COPY,
-			advice, uvmflag);
+	    (flags & MAP_SHARED) ? UVM_INH_SHARE : UVM_INH_COPY, advice,
+	    uvmflag);
 	error = uvm_map(map, addr, size, uobj, foff, align, uvmflag);
 	if (error) {
 		if (uobj)
@@ -1030,7 +1024,7 @@ uvm_mmap(struct vm_map *map, vaddr_t *ad
 		 * No more work to do in this case.
 		 */
 
-		return (0);
+		return 0;
 	}
 	if ((flags & MAP_WIRED) != 0 || (map->flags & VM_MAP_WIREFUTURE) != 0) {
 		vm_map_lock(map);
@@ -1048,12 +1042,12 @@ uvm_mmap(struct vm_map *map, vaddr_t *ad
 		 */
 
 		error = uvm_map_pageable(map, *addr, *addr + size,
-					 false, UVM_LK_ENTER);
+		    false, UVM_LK_ENTER);
 		if (error) {
 			uvm_unmap(map, *addr, *addr + size);
 			return error;
 		}
-		return (0);
+		return 0;
 	}
 	return 0;
 }
@@ -1089,8 +1083,8 @@ uvm_mmap_dev(struct proc *p, void **addr
 		return EINVAL;
 
 	error = uvm_mmap(&p->p_vmspace->vm_map, (vaddr_t *)addrp,
-			 (vsize_t)len, prot, prot, flags, UVM_ADV_RANDOM,
-			 uobj, off, p->p_rlimit[RLIMIT_MEMLOCK].rlim_cur);
+	    (vsize_t)len, prot, prot, flags, UVM_ADV_RANDOM, uobj, off,
+	    p->p_rlimit[RLIMIT_MEMLOCK].rlim_cur);
 	return error;
 }
 
@@ -1109,7 +1103,7 @@ uvm_mmap_anon(struct proc *p, void **add
 		    p->p_vmspace->vm_map.flags & VM_MAP_TOPDOWN);
 
 	error = uvm_mmap(&p->p_vmspace->vm_map, (vaddr_t *)addrp,
-			 (vsize_t)len, prot, prot, flags, UVM_ADV_NORMAL,
-			 NULL, 0, p->p_rlimit[RLIMIT_MEMLOCK].rlim_cur);
+	    (vsize_t)len, prot, prot, flags, UVM_ADV_NORMAL, NULL, 0,
+	    p->p_rlimit[RLIMIT_MEMLOCK].rlim_cur);
 	return error;
 }

Reply via email to