On Tue, Mar 23, 2021 at 01:52:20PM +0100, Martin Pieuchot wrote:
> Diff below convert multiple "return(val)" and "return (val)" to
> "return val".  I only changed those that help decrease the size
> of the diff with NetBSD or didn't change anything.
>
> ok?
>

I read through these and agree this should not change any behaviour.

ok mlarkin if this helps you move forward by improving diffability.

> Index: uvm/uvm_amap.c
> ===================================================================
> RCS file: /cvs/src/sys/uvm/uvm_amap.c,v
> retrieving revision 1.88
> diff -u -p -r1.88 uvm_amap.c
> --- uvm/uvm_amap.c    20 Mar 2021 10:24:21 -0000      1.88
> +++ uvm/uvm_amap.c    23 Mar 2021 12:14:26 -0000
> @@ -342,7 +342,7 @@ amap_alloc1(int slots, int waitf, int la
>               amap = pool_get(&uvm_small_amap_pool[slots - 1],
>                   pwaitf | PR_ZERO);
>       if (amap == NULL)
> -             return(NULL);
> +             return NULL;
>
>       amap->am_lock = NULL;
>       amap->am_ref = 1;
> @@ -355,7 +355,7 @@ amap_alloc1(int slots, int waitf, int la
>
>       if (UVM_AMAP_SMALL(amap)) {
>               amap->am_small.ac_nslot = slots;
> -             return (amap);
> +             return amap;
>       }
>
>       amap->am_ncused = 0;
> @@ -392,14 +392,14 @@ amap_alloc1(int slots, int waitf, int la
>               }
>       }
>
> -     return(amap);
> +     return amap;
>
>  fail1:
>       free(amap->am_buckets, M_UVMAMAP, buckets * sizeof(*amap->am_buckets));
>       TAILQ_FOREACH_SAFE(chunk, &amap->am_chunks, ac_list, tmp)
>               pool_put(&uvm_amap_chunk_pool, chunk);
>       pool_put(&uvm_amap_pool, amap);
> -     return (NULL);
> +     return NULL;
>  }
>
>  static void
> @@ -423,7 +423,7 @@ amap_alloc(vaddr_t sz, int waitf, int la
>
>       AMAP_B2SLOT(slots, sz);         /* load slots */
>       if (slots > INT_MAX)
> -             return (NULL);
> +             return NULL;
>
>       amap = amap_alloc1(slots, waitf, lazyalloc);
>       if (amap != NULL) {
> @@ -431,7 +431,7 @@ amap_alloc(vaddr_t sz, int waitf, int la
>               amap_list_insert(amap);
>       }
>
> -     return(amap);
> +     return amap;
>  }
>
>
> Index: uvm/uvm_anon.c
> ===================================================================
> RCS file: /cvs/src/sys/uvm/uvm_anon.c,v
> retrieving revision 1.53
> diff -u -p -r1.53 uvm_anon.c
> --- uvm/uvm_anon.c    20 Mar 2021 10:24:21 -0000      1.53
> +++ uvm/uvm_anon.c    23 Mar 2021 12:01:03 -0000
> @@ -67,7 +67,7 @@ uvm_analloc(void)
>               anon->an_page = NULL;
>               anon->an_swslot = 0;
>       }
> -     return(anon);
> +     return anon;
>  }
>
>  /*
> Index: uvm/uvm_aobj.c
> ===================================================================
> RCS file: /cvs/src/sys/uvm/uvm_aobj.c,v
> retrieving revision 1.92
> diff -u -p -r1.92 uvm_aobj.c
> --- uvm/uvm_aobj.c    20 Mar 2021 10:24:21 -0000      1.92
> +++ uvm/uvm_aobj.c    23 Mar 2021 12:17:00 -0000
> @@ -211,7 +211,7 @@ uao_find_swhash_elt(struct uvm_aobj *aob
>        */
>       LIST_FOREACH(elt, swhash, list) {
>               if (elt->tag == page_tag)
> -                     return(elt);
> +                     return elt;
>       }
>
>       if (!create)
> @@ -234,7 +234,7 @@ uao_find_swhash_elt(struct uvm_aobj *aob
>       LIST_INSERT_HEAD(swhash, elt, list);
>       elt->tag = page_tag;
>
> -     return(elt);
> +     return elt;
>  }
>
>  /*
> @@ -248,7 +248,7 @@ uao_find_swslot(struct uvm_aobj *aobj, i
>        * if noswap flag is set, then we never return a slot
>        */
>       if (aobj->u_flags & UAO_FLAG_NOSWAP)
> -             return(0);
> +             return 0;
>
>       /*
>        * if hashing, look in hash table.
> @@ -258,15 +258,15 @@ uao_find_swslot(struct uvm_aobj *aobj, i
>                   uao_find_swhash_elt(aobj, pageidx, FALSE);
>
>               if (elt)
> -                     return(UAO_SWHASH_ELT_PAGESLOT(elt, pageidx));
> +                     return UAO_SWHASH_ELT_PAGESLOT(elt, pageidx);
>               else
> -                     return(0);
> +                     return 0;
>       }
>
>       /*
>        * otherwise, look in the array
>        */
> -     return(aobj->u_swslots[pageidx]);
> +     return aobj->u_swslots[pageidx];
>  }
>
>  /*
> @@ -289,7 +289,7 @@ uao_set_swslot(struct uvm_object *uobj,
>        */
>       if (aobj->u_flags & UAO_FLAG_NOSWAP) {
>               if (slot == 0)
> -                     return(0);              /* a clear is ok */
> +                     return 0;               /* a clear is ok */
>
>               /* but a set is not */
>               printf("uao_set_swslot: uobj = %p\n", uobj);
> @@ -309,7 +309,7 @@ uao_set_swslot(struct uvm_object *uobj,
>                   uao_find_swhash_elt(aobj, pageidx, slot ? TRUE : FALSE);
>               if (elt == NULL) {
>                       KASSERT(slot == 0);
> -                     return (0);
> +                     return 0;
>               }
>
>               oldslot = UAO_SWHASH_ELT_PAGESLOT(elt, pageidx);
> @@ -336,7 +336,7 @@ uao_set_swslot(struct uvm_object *uobj,
>               oldslot = aobj->u_swslots[pageidx];
>               aobj->u_swslots[pageidx] = slot;
>       }
> -     return (oldslot);
> +     return oldslot;
>  }
>  /*
>   * end of hash/array functions
> @@ -749,7 +749,7 @@ uao_create(vsize_t size, int flags)
>                       if (aobj->u_swhash == NULL) {
>                               if (flags & UAO_FLAG_CANFAIL) {
>                                       pool_put(&uvm_aobj_pool, aobj);
> -                                     return (NULL);
> +                                     return NULL;
>                               }
>                               panic("uao_create: hashinit swhash failed");
>                       }
> @@ -759,7 +759,7 @@ uao_create(vsize_t size, int flags)
>                       if (aobj->u_swslots == NULL) {
>                               if (flags & UAO_FLAG_CANFAIL) {
>                                       pool_put(&uvm_aobj_pool, aobj);
> -                                     return (NULL);
> +                                     return NULL;
>                               }
>                               panic("uao_create: malloc swslots failed");
>                       }
> @@ -767,7 +767,7 @@ uao_create(vsize_t size, int flags)
>
>               if (flags & UAO_FLAG_KERNSWAP) {
>                       aobj->u_flags &= ~UAO_FLAG_NOSWAP; /* clear noswap */
> -                     return(&aobj->u_obj);
> +                     return &aobj->u_obj;
>                       /* done! */
>               }
>       }
> @@ -784,7 +784,7 @@ uao_create(vsize_t size, int flags)
>       LIST_INSERT_HEAD(&uao_list, aobj, u_list);
>       mtx_leave(&uao_list_lock);
>
> -     return(&aobj->u_obj);
> +     return &aobj->u_obj;
>  }
>
>
> @@ -940,7 +940,7 @@ uao_flush(struct uvm_object *uobj, voff_
>        * or deactivating pages.
>        */
>       if ((flags & (PGO_DEACTIVATE|PGO_FREE)) == 0)
> -             return (TRUE);
> +             return TRUE;
>
>       curoff = start;
>       for (;;) {
> @@ -1016,7 +1016,7 @@ uao_flush(struct uvm_object *uobj, voff_
>               }
>       }
>
> -     return (TRUE);
> +     return TRUE;
>  }
>
>  /*
> @@ -1118,10 +1118,10 @@ uao_get(struct uvm_object *uobj, voff_t
>               *npagesp = gotpages;
>               if (done)
>                       /* bingo! */
> -                     return(VM_PAGER_OK);
> +                     return VM_PAGER_OK;
>               else
>                       /* EEK!   Need to unlock and I/O */
> -                     return(VM_PAGER_UNLOCK);
> +                     return VM_PAGER_UNLOCK;
>       }
>
>       /*
> @@ -1249,7 +1249,7 @@ uao_get(struct uvm_object *uobj, voff_t
>                               uvm_pagefree(ptmp);
>                               uvm_unlock_pageq();
>
> -                             return (rv);
> +                             return rv;
>                       }
>               }
>
> @@ -1269,7 +1269,7 @@ uao_get(struct uvm_object *uobj, voff_t
>
>       }       /* lcv loop */
>
> -     return(VM_PAGER_OK);
> +     return VM_PAGER_OK;
>  }
>
>  /*
> @@ -1284,7 +1284,7 @@ uao_dropswap(struct uvm_object *uobj, in
>       if (slot) {
>               uvm_swap_free(slot, 1);
>       }
> -     return (slot);
> +     return slot;
>  }
>
>  /*
> Index: uvm/uvm_fault.c
> ===================================================================
> RCS file: /cvs/src/sys/uvm/uvm_fault.c,v
> retrieving revision 1.119
> diff -u -p -r1.119 uvm_fault.c
> --- uvm/uvm_fault.c   20 Mar 2021 10:24:21 -0000      1.119
> +++ uvm/uvm_fault.c   23 Mar 2021 12:01:58 -0000
> @@ -1688,7 +1688,7 @@ uvmfault_lookup(struct uvm_faultinfo *uf
>       while (1) {
>               if (ufi->orig_rvaddr < ufi->map->min_offset ||
>                   ufi->orig_rvaddr >= ufi->map->max_offset)
> -                     return(FALSE);
> +                     return FALSE;
>
>               /* lock map */
>               if (write_lock) {
> @@ -1701,7 +1701,7 @@ uvmfault_lookup(struct uvm_faultinfo *uf
>               if (!uvm_map_lookup_entry(ufi->map, ufi->orig_rvaddr,
>                   &ufi->entry)) {
>                       uvmfault_unlockmaps(ufi, write_lock);
> -                     return(FALSE);
> +                     return FALSE;
>               }
>
>               /* reduce size if necessary */
> @@ -1723,7 +1723,7 @@ uvmfault_lookup(struct uvm_faultinfo *uf
>                * got it!
>                */
>               ufi->mapv = ufi->map->timestamp;
> -             return(TRUE);
> +             return TRUE;
>
>       }       /* while loop */
>
> @@ -1756,9 +1756,9 @@ uvmfault_relock(struct uvm_faultinfo *uf
>       vm_map_lock_read(ufi->map);
>       if (ufi->mapv != ufi->map->timestamp) {
>               vm_map_unlock_read(ufi->map);
> -             return(FALSE);
> +             return FALSE;
>       }
>
>       counters_inc(uvmexp_counters, flt_relckok);
> -     return(TRUE);           /* got it! */
> +     return TRUE;            /* got it! */
>  }
> Index: uvm/uvm_glue.c
> ===================================================================
> RCS file: /cvs/src/sys/uvm/uvm_glue.c,v
> retrieving revision 1.79
> diff -u -p -r1.79 uvm_glue.c
> --- uvm/uvm_glue.c    12 Mar 2021 14:15:49 -0000      1.79
> +++ uvm/uvm_glue.c    23 Mar 2021 12:01:46 -0000
> @@ -97,7 +97,7 @@ uvm_kernacc(caddr_t addr, size_t len, in
>       rv = uvm_map_checkprot(kernel_map, saddr, eaddr, prot);
>       vm_map_unlock_read(kernel_map);
>
> -     return(rv);
> +     return rv;
>  }
>
>  /*
> Index: uvm/uvm_km.c
> ===================================================================
> RCS file: /cvs/src/sys/uvm/uvm_km.c,v
> retrieving revision 1.142
> diff -u -p -r1.142 uvm_km.c
> --- uvm/uvm_km.c      20 Mar 2021 10:24:21 -0000      1.142
> +++ uvm/uvm_km.c      23 Mar 2021 12:26:41 -0000
> @@ -343,12 +343,12 @@ uvm_km_kmemalloc_pla(struct vm_map *map,
>       if (__predict_false(uvm_map(map, &kva, size, obj, UVM_UNKNOWN_OFFSET,
>           valign, UVM_MAPFLAG(PROT_READ | PROT_WRITE, PROT_READ | PROT_WRITE,
>           MAP_INHERIT_NONE, MADV_RANDOM, (flags & UVM_KMF_TRYLOCK))) != 0)) {
> -             return(0);
> +             return 0;
>       }
>
>       /* if all we wanted was VA, return now */
>       if (flags & UVM_KMF_VALLOC) {
> -             return(kva);
> +             return kva;
>       }
>
>       /* recover object offset from virtual address */
> @@ -405,7 +405,7 @@ uvm_km_kmemalloc_pla(struct vm_map *map,
>       KASSERT(TAILQ_EMPTY(&pgl));
>       pmap_update(pmap_kernel());
>
> -     return(kva);
> +     return kva;
>  }
>
>  /*
> @@ -461,7 +461,7 @@ uvm_km_alloc1(struct vm_map *map, vsize_
>           UVM_MAPFLAG(PROT_READ | PROT_WRITE,
>           PROT_READ | PROT_WRITE | PROT_EXEC,
>           MAP_INHERIT_NONE, MADV_RANDOM, 0)) != 0)) {
> -             return(0);
> +             return 0;
>       }
>
>       /* recover object offset from virtual address */
> @@ -512,7 +512,7 @@ uvm_km_alloc1(struct vm_map *map, vsize_
>       if (zeroit)
>               memset((caddr_t)kva, 0, loopva - kva);
>
> -     return(kva);
> +     return kva;
>  }
>
>  /*
> @@ -524,13 +524,13 @@ uvm_km_alloc1(struct vm_map *map, vsize_
>  vaddr_t
>  uvm_km_valloc(struct vm_map *map, vsize_t size)
>  {
> -     return(uvm_km_valloc_align(map, size, 0, 0));
> +     return uvm_km_valloc_align(map, size, 0, 0);
>  }
>
>  vaddr_t
>  uvm_km_valloc_try(struct vm_map *map, vsize_t size)
>  {
> -     return(uvm_km_valloc_align(map, size, 0, UVM_FLAG_TRYLOCK));
> +     return uvm_km_valloc_align(map, size, 0, UVM_FLAG_TRYLOCK);
>  }
>
>  vaddr_t
> @@ -550,10 +550,10 @@ uvm_km_valloc_align(struct vm_map *map,
>           UVM_UNKNOWN_OFFSET, align,
>           UVM_MAPFLAG(PROT_READ | PROT_WRITE, PROT_READ | PROT_WRITE,
>           MAP_INHERIT_NONE, MADV_RANDOM, flags)) != 0)) {
> -             return(0);
> +             return 0;
>       }
>
> -     return(kva);
> +     return kva;
>  }
>
>  /*
> @@ -572,7 +572,7 @@ uvm_km_valloc_prefer_wait(struct vm_map
>
>       size = round_page(size);
>       if (size > vm_map_max(map) - vm_map_min(map))
> -             return(0);
> +             return 0;
>
>       while (1) {
>               kva = vm_map_min(map);          /* hint */
> @@ -585,7 +585,7 @@ uvm_km_valloc_prefer_wait(struct vm_map
>                   prefer, 0,
>                   UVM_MAPFLAG(PROT_READ | PROT_WRITE, PROT_READ | PROT_WRITE,
>                   MAP_INHERIT_NONE, MADV_RANDOM, 0)) == 0)) {
> -                     return(kva);
> +                     return kva;
>               }
>
>               /* failed.  sleep for a while (on map) */
> Index: uvm/uvm_map.c
> ===================================================================
> RCS file: /cvs/src/sys/uvm/uvm_map.c,v
> retrieving revision 1.273
> diff -u -p -r1.273 uvm_map.c
> --- uvm/uvm_map.c     12 Mar 2021 14:15:49 -0000      1.273
> +++ uvm/uvm_map.c     23 Mar 2021 12:02:31 -0000
> @@ -1000,7 +1000,7 @@ uvm_mapanon(struct vm_map *map, vaddr_t
>        */
>       new = uvm_mapent_alloc(map, flags);
>       if (new == NULL)
> -             return(ENOMEM);
> +             return ENOMEM;
>
>       vm_map_lock(map);
>       first = last = NULL;
> @@ -1229,7 +1229,7 @@ uvm_map(struct vm_map *map, vaddr_t *add
>        */
>       new = uvm_mapent_alloc(map, flags);
>       if (new == NULL)
> -             return(ENOMEM);
> +             return ENOMEM;
>
>       if (flags & UVM_FLAG_TRYLOCK) {
>               if (vm_map_lock_try(map) == FALSE) {
> @@ -1759,7 +1759,7 @@ uvm_mapent_alloc(struct vm_map *map, int
>
>       RBT_POISON(uvm_map_addr, me, UVMMAP_DEADBEEF);
>  out:
> -     return(me);
> +     return me;
>  }
>
>  /*
> @@ -4229,7 +4229,7 @@ uvm_map_submap(struct vm_map *map, vaddr
>               result = EINVAL;
>
>       vm_map_unlock(map);
> -     return(result);
> +     return result;
>  }
>
>  /*
> Index: uvm/uvm_mmap.c
> ===================================================================
> RCS file: /cvs/src/sys/uvm/uvm_mmap.c,v
> retrieving revision 1.163
> diff -u -p -r1.163 uvm_mmap.c
> --- uvm/uvm_mmap.c    7 Oct 2020 12:26:20 -0000       1.163
> +++ uvm/uvm_mmap.c    23 Mar 2021 12:29:24 -0000
> @@ -92,14 +92,14 @@ int uvm_mmapfile(vm_map_t, vaddr_t *, vs
>       pageoff = (addr & PAGE_MASK);                                   \
>       if (pageoff != 0) {                                             \
>               if (size > SIZE_MAX - pageoff)                          \
> -                     return (EINVAL);        /* wraparound */        \
> +                     return EINVAL;  /* wraparound */        \
>               addr -= pageoff;                                        \
>               size += pageoff;                                        \
>       }                                                               \
>       if (size != 0) {                                                \
>               size = (vsize_t)round_page(size);                       \
>               if (size == 0)                                          \
> -                     return (EINVAL);        /* wraparound */        \
> +                     return EINVAL;  /* wraparound */        \
>       }                                                               \
>  } while (0)
>
> @@ -140,14 +140,14 @@ sys_mquery(struct proc *p, void *v, regi
>       fd = SCARG(uap, fd);
>
>       if ((prot & PROT_MASK) != prot)
> -             return (EINVAL);
> +             return EINVAL;
>
>       if (SCARG(uap, flags) & MAP_FIXED)
>               flags |= UVM_FLAG_FIXED;
>
>       if (fd >= 0) {
>               if ((error = getvnode(p, fd, &fp)) != 0)
> -                     return (error);
> +                     return error;
>               uoff = SCARG(uap, pos);
>       } else {
>               fp = NULL;
> @@ -165,7 +165,7 @@ sys_mquery(struct proc *p, void *v, regi
>
>       if (fp != NULL)
>               FRELE(fp, p);
> -     return (error);
> +     return error;
>  }
>
>  int  uvm_wxabort;
> @@ -181,7 +181,7 @@ uvm_wxcheck(struct proc *p, char *call)
>           (pr->ps_textvp->v_mount->mnt_flag & MNT_WXALLOWED));
>
>       if (wxallowed && (pr->ps_flags & PS_WXNEEDED))
> -             return (0);
> +             return 0;
>
>       if (uvm_wxabort) {
>               /* Report W^X failures */
> @@ -192,7 +192,7 @@ uvm_wxcheck(struct proc *p, char *call)
>               sigexit(p, SIGABRT);
>       }
>
> -     return (ENOTSUP);
> +     return ENOTSUP;
>  }
>
>  /*
> @@ -239,33 +239,33 @@ sys_mmap(struct proc *p, void *v, regist
>        * Validate the flags.
>        */
>       if ((prot & PROT_MASK) != prot)
> -             return (EINVAL);
> +             return EINVAL;
>       if ((prot & (PROT_WRITE | PROT_EXEC)) == (PROT_WRITE | PROT_EXEC) &&
>           (error = uvm_wxcheck(p, "mmap")))
> -             return (error);
> +             return error;
>
>       if ((flags & MAP_FLAGMASK) != flags)
> -             return (EINVAL);
> +             return EINVAL;
>       if ((flags & (MAP_SHARED|MAP_PRIVATE)) == (MAP_SHARED|MAP_PRIVATE))
> -             return (EINVAL);
> +             return EINVAL;
>       if ((flags & (MAP_FIXED|__MAP_NOREPLACE)) == __MAP_NOREPLACE)
> -             return (EINVAL);
> +             return EINVAL;
>       if (flags & MAP_STACK) {
>               if ((flags & (MAP_ANON|MAP_PRIVATE)) != (MAP_ANON|MAP_PRIVATE))
> -                     return (EINVAL);
> +                     return EINVAL;
>               if (flags & ~(MAP_STACK|MAP_FIXED|MAP_ANON|MAP_PRIVATE))
> -                     return (EINVAL);
> +                     return EINVAL;
>               if (pos != 0)
> -                     return (EINVAL);
> +                     return EINVAL;
>               if ((prot & (PROT_READ|PROT_WRITE)) != (PROT_READ|PROT_WRITE))
> -                     return (EINVAL);
> +                     return EINVAL;
>       }
>       if (size == 0)
> -             return (EINVAL);
> +             return EINVAL;
>
>       error = pledge_protexec(p, prot);
>       if (error)
> -             return (error);
> +             return error;
>
>       /* align file position and save offset.  adjust size. */
>       ALIGN_ADDR(pos, size, pageoff);
> @@ -275,15 +275,15 @@ sys_mmap(struct proc *p, void *v, regist
>               /* adjust address by the same amount as we did the offset */
>               addr -= pageoff;
>               if (addr & PAGE_MASK)
> -                     return (EINVAL);                /* not page aligned */
> +                     return EINVAL;          /* not page aligned */
>
>               if (addr > SIZE_MAX - size)
> -                     return (EINVAL);                /* no wrapping! */
> +                     return EINVAL;          /* no wrapping! */
>               if (VM_MAXUSER_ADDRESS > 0 &&
>                   (addr + size) > VM_MAXUSER_ADDRESS)
> -                     return (EINVAL);
> +                     return EINVAL;
>               if (vm_min_address > 0 && addr < vm_min_address)
> -                     return (EINVAL);
> +                     return EINVAL;
>       }
>
>       /* check for file mappings (i.e. not anonymous) and verify file. */
> @@ -430,13 +430,13 @@ is_anon:        /* label for SunOS style /dev/z
>               /* remember to add offset */
>               *retval = (register_t)(addr + pageoff);
>
> -     return (error);
> +     return error;
>
>  out:
>       KERNEL_UNLOCK();
>       if (fp)
>               FRELE(fp, p);
> -     return (error);
> +     return error;
>  }
>
>  /*
> @@ -465,14 +465,14 @@ sys_msync(struct proc *p, void *v, regis
>       if ((flags & ~(MS_ASYNC | MS_SYNC | MS_INVALIDATE)) != 0 ||
>                       (flags & (MS_ASYNC | MS_SYNC | MS_INVALIDATE)) == 0 ||
>                       (flags & (MS_ASYNC | MS_SYNC)) == (MS_ASYNC | MS_SYNC))
> -             return (EINVAL);
> +             return EINVAL;
>       if ((flags & (MS_ASYNC | MS_SYNC)) == 0)
>               flags |= MS_SYNC;
>
>       /* align the address to a page boundary, and adjust the size 
> accordingly */
>       ALIGN_ADDR(addr, size, pageoff);
>       if (addr > SIZE_MAX - size)
> -             return (EINVAL);                /* disallow wrap-around. */
> +             return EINVAL;          /* disallow wrap-around. */
>
>       /* get map */
>       map = &p->p_vmspace->vm_map;
> @@ -486,7 +486,7 @@ sys_msync(struct proc *p, void *v, regis
>       else
>               uvmflags |= PGO_SYNCIO;  /* XXXCDC: force sync for now! */
>
> -     return (uvm_map_clean(map, addr, addr+size, uvmflags));
> +     return uvm_map_clean(map, addr, addr+size, uvmflags);
>  }
>
>  /*
> @@ -517,11 +517,11 @@ sys_munmap(struct proc *p, void *v, regi
>        * Note that VM_*_ADDRESS are not constants due to casts (argh).
>        */
>       if (addr > SIZE_MAX - size)
> -             return (EINVAL);
> +             return EINVAL;
>       if (VM_MAXUSER_ADDRESS > 0 && addr + size > VM_MAXUSER_ADDRESS)
> -             return (EINVAL);
> +             return EINVAL;
>       if (vm_min_address > 0 && addr < vm_min_address)
> -             return (EINVAL);
> +             return EINVAL;
>       map = &p->p_vmspace->vm_map;
>
>
> @@ -533,7 +533,7 @@ sys_munmap(struct proc *p, void *v, regi
>        */
>       if (!uvm_map_checkprot(map, addr, addr + size, PROT_NONE)) {
>               vm_map_unlock(map);
> -             return (EINVAL);
> +             return EINVAL;
>       }
>
>       TAILQ_INIT(&dead_entries);
> @@ -542,7 +542,7 @@ sys_munmap(struct proc *p, void *v, regi
>
>       uvm_unmap_detach(&dead_entries, 0);
>
> -     return (0);
> +     return 0;
>  }
>
>  /*
> @@ -570,21 +570,21 @@ sys_mprotect(struct proc *p, void *v, re
>       prot = SCARG(uap, prot);
>
>       if ((prot & PROT_MASK) != prot)
> -             return (EINVAL);
> +             return EINVAL;
>       if ((prot & (PROT_WRITE | PROT_EXEC)) == (PROT_WRITE | PROT_EXEC) &&
>           (error = uvm_wxcheck(p, "mprotect")))
> -             return (error);
> +             return error;
>
>       error = pledge_protexec(p, prot);
>       if (error)
> -             return (error);
> +             return error;
>
>       /*
>        * align the address to a page boundary, and adjust the size accordingly
>        */
>       ALIGN_ADDR(addr, size, pageoff);
>       if (addr > SIZE_MAX - size)
> -             return (EINVAL);                /* disallow wrap-around. */
> +             return EINVAL;          /* disallow wrap-around. */
>
>       return (uvm_map_protect(&p->p_vmspace->vm_map, addr, addr+size,
>           prot, FALSE));
> @@ -611,9 +611,9 @@ sys_msyscall(struct proc *p, void *v, re
>        */
>       ALIGN_ADDR(addr, size, pageoff);
>       if (addr > SIZE_MAX - size)
> -             return (EINVAL);                /* disallow wrap-around. */
> +             return EINVAL;          /* disallow wrap-around. */
>
> -     return (uvm_map_syscall(&p->p_vmspace->vm_map, addr, addr+size));
> +     return uvm_map_syscall(&p->p_vmspace->vm_map, addr, addr+size);
>  }
>
>  /*
> @@ -640,7 +640,7 @@ sys_minherit(struct proc *p, void *v, re
>        */
>       ALIGN_ADDR(addr, size, pageoff);
>       if (addr > SIZE_MAX - size)
> -             return (EINVAL);                /* disallow wrap-around. */
> +             return EINVAL;          /* disallow wrap-around. */
>
>       return (uvm_map_inherit(&p->p_vmspace->vm_map, addr, addr+size,
>           inherit));
> @@ -671,7 +671,7 @@ sys_madvise(struct proc *p, void *v, reg
>        */
>       ALIGN_ADDR(addr, size, pageoff);
>       if (addr > SIZE_MAX - size)
> -             return (EINVAL);                /* disallow wrap-around. */
> +             return EINVAL;          /* disallow wrap-around. */
>
>       switch (advice) {
>       case MADV_NORMAL:
> @@ -691,7 +691,7 @@ sys_madvise(struct proc *p, void *v, reg
>                * Should invent a "weak" mode for uvm_fault()
>                * which would only do the PGO_LOCKED pgo_get().
>                */
> -             return (0);
> +             return 0;
>
>       case MADV_DONTNEED:
>               /*
> @@ -724,13 +724,13 @@ sys_madvise(struct proc *p, void *v, reg
>                * as it will free swap space allocated to pages in core.
>                * There's also what to do for device/file/anonymous memory.
>                */
> -             return (EINVAL);
> +             return EINVAL;
>
>       default:
> -             return (EINVAL);
> +             return EINVAL;
>       }
>
> -     return (error);
> +     return error;
>  }
>
>  /*
> @@ -755,23 +755,23 @@ sys_mlock(struct proc *p, void *v, regis
>       /* align address to a page boundary and adjust size accordingly */
>       ALIGN_ADDR(addr, size, pageoff);
>       if (addr > SIZE_MAX - size)
> -             return (EINVAL);                /* disallow wrap-around. */
> +             return EINVAL;          /* disallow wrap-around. */
>
>       if (atop(size) + uvmexp.wired > uvmexp.wiredmax)
> -             return (EAGAIN);
> +             return EAGAIN;
>
>  #ifdef pmap_wired_count
>       if (size + ptoa(pmap_wired_count(vm_map_pmap(&p->p_vmspace->vm_map))) >
>                       lim_cur(RLIMIT_MEMLOCK))
> -             return (EAGAIN);
> +             return EAGAIN;
>  #else
>       if ((error = suser(p)) != 0)
> -             return (error);
> +             return error;
>  #endif
>
>       error = uvm_map_pageable(&p->p_vmspace->vm_map, addr, addr+size, FALSE,
>           0);
> -     return (error == 0 ? 0 : ENOMEM);
> +     return error == 0 ? 0 : ENOMEM;
>  }
>
>  /*
> @@ -796,16 +796,16 @@ sys_munlock(struct proc *p, void *v, reg
>       /* align address to a page boundary, and adjust size accordingly */
>       ALIGN_ADDR(addr, size, pageoff);
>       if (addr > SIZE_MAX - size)
> -             return (EINVAL);                /* disallow wrap-around. */
> +             return EINVAL;          /* disallow wrap-around. */
>
>  #ifndef pmap_wired_count
>       if ((error = suser(p)) != 0)
> -             return (error);
> +             return error;
>  #endif
>
>       error = uvm_map_pageable(&p->p_vmspace->vm_map, addr, addr+size, TRUE,
>           0);
> -     return (error == 0 ? 0 : ENOMEM);
> +     return error == 0 ? 0 : ENOMEM;
>  }
>
>  /*
> @@ -823,18 +823,18 @@ sys_mlockall(struct proc *p, void *v, re
>
>       if (flags == 0 ||
>           (flags & ~(MCL_CURRENT|MCL_FUTURE)) != 0)
> -             return (EINVAL);
> +             return EINVAL;
>
>  #ifndef pmap_wired_count
>       if ((error = suser(p)) != 0)
> -             return (error);
> +             return error;
>  #endif
>
>       error = uvm_map_pageable_all(&p->p_vmspace->vm_map, flags,
>           lim_cur(RLIMIT_MEMLOCK));
>       if (error != 0 && error != ENOMEM)
> -             return (EAGAIN);
> -     return (error);
> +             return EAGAIN;
> +     return error;
>  }
>
>  /*
> @@ -845,7 +845,7 @@ sys_munlockall(struct proc *p, void *v,
>  {
>
>       (void) uvm_map_pageable_all(&p->p_vmspace->vm_map, 0, 0);
> -     return (0);
> +     return 0;
>  }
>
>  /*
> @@ -865,7 +865,7 @@ uvm_mmaplock(vm_map_t map, vaddr_t *addr
>               /*
>                * No more work to do in this case.
>                */
> -             return (0);
> +             return 0;
>       }
>
>       vm_map_lock(map);
> @@ -883,7 +883,7 @@ uvm_mmaplock(vm_map_t map, vaddr_t *addr
>                       /* unmap the region! */
>                       uvm_unmap(map, *addr, *addr + size);
>                       KERNEL_UNLOCK();
> -                     return (error);
> +                     return error;
>               }
>               /*
>                * uvm_map_pageable() always returns the map
> @@ -895,13 +895,13 @@ uvm_mmaplock(vm_map_t map, vaddr_t *addr
>                       /* unmap the region! */
>                       uvm_unmap(map, *addr, *addr + size);
>                       KERNEL_UNLOCK();
> -                     return (error);
> +                     return error;
>               }
>               KERNEL_UNLOCK();
> -             return (0);
> +             return 0;
>       }
>       vm_map_unlock(map);
> -     return (0);
> +     return 0;
>  }
>
>  /*
> @@ -926,7 +926,7 @@ uvm_mmapanon(vm_map_t map, vaddr_t *addr
>               *addr = round_page(*addr);      /* round */
>       } else {
>               if (*addr & PAGE_MASK)
> -                     return(EINVAL);
> +                     return EINVAL;
>
>               uvmflag |= UVM_FLAG_FIXED;
>               if ((flags & __MAP_NOREPLACE) == 0)
> @@ -983,7 +983,7 @@ uvm_mmapfile(vm_map_t map, vaddr_t *addr
>               *addr = round_page(*addr);      /* round */
>       } else {
>               if (*addr & PAGE_MASK)
> -                     return(EINVAL);
> +                     return EINVAL;
>
>               uvmflag |= UVM_FLAG_FIXED;
>               if ((flags & __MAP_NOREPLACE) == 0)
> @@ -1045,7 +1045,7 @@ uvm_mmapfile(vm_map_t map, vaddr_t *addr
>       }
>
>       if (uobj == NULL)
> -             return((vp->v_type == VREG) ? ENOMEM : EINVAL);
> +             return vp->v_type == VREG ? ENOMEM : EINVAL;
>
>       if ((flags & MAP_SHARED) == 0)
>               uvmflag |= UVM_FLAG_COPYONW;
> @@ -1070,7 +1070,7 @@ uvm_mmapfile(vm_map_t map, vaddr_t *addr
>       if (uobj)
>               uobj->pgops->pgo_detach(uobj);
>
> -     return (error);
> +     return error;
>  }
>
>  /* an address that can't be in userspace or kernelspace */
> @@ -1106,7 +1106,7 @@ sys_kbind(struct proc *p, void *v, regis
>       /* a NULL paramp disables the syscall for the process */
>       if (paramp == NULL) {
>               pr->ps_kbind_addr = BOGO_PC;
> -             return (0);
> +             return 0;
>       }
>
>       /* security checks */
> @@ -1119,9 +1119,9 @@ sys_kbind(struct proc *p, void *v, regis
>       else if (pr->ps_kbind_cookie != SCARG(uap, proc_cookie))
>               sigexit(p, SIGILL);
>       if (psize < sizeof(struct __kbind) || psize > sizeof(param))
> -             return (EINVAL);
> +             return EINVAL;
>       if ((error = copyin(paramp, &param, psize)))
> -             return (error);
> +             return error;
>
>       /*
>        * The param argument points to an array of __kbind structures
> @@ -1133,7 +1133,7 @@ sys_kbind(struct proc *p, void *v, regis
>       s = psize;
>       for (count = 0; s > 0 && count < KBIND_BLOCK_MAX; count++) {
>               if (s < sizeof(*paramp))
> -                     return (EINVAL);
> +                     return EINVAL;
>               s -= sizeof(*paramp);
>
>               baseva = (vaddr_t)paramp[count].kb_addr;
> @@ -1145,12 +1145,12 @@ sys_kbind(struct proc *p, void *v, regis
>                   endva >= VM_MAXUSER_ADDRESS ||
>                   trunc_page(baseva) != trunc_page(endva) ||
>                   s < paramp[count].kb_size)
> -                     return (EINVAL);
> +                     return EINVAL;
>
>               s -= paramp[count].kb_size;
>       }
>       if (s > 0)
> -             return (EINVAL);
> +             return EINVAL;
>       data = (const char *)&paramp[count];
>
>       /* all looks good, so do the bindings */
> @@ -1192,5 +1192,5 @@ sys_kbind(struct proc *p, void *v, regis
>       }
>       uvm_unmap_detach(&dead_entries, AMAP_REFALL);
>
> -     return (error);
> +     return error;
>  }
> Index: uvm/uvm_page.c
> ===================================================================
> RCS file: /cvs/src/sys/uvm/uvm_page.c,v
> retrieving revision 1.155
> diff -u -p -r1.155 uvm_page.c
> --- uvm/uvm_page.c    19 Jan 2021 13:21:36 -0000      1.155
> +++ uvm/uvm_page.c    23 Mar 2021 12:32:01 -0000
> @@ -84,7 +84,7 @@ RBT_GENERATE(uvm_objtree, vm_page, objt,
>  int
>  uvm_pagecmp(const struct vm_page *a, const struct vm_page *b)
>  {
> -     return (a->offset < b->offset ? -1 : a->offset > b->offset);
> +     return a->offset < b->offset ? -1 : a->offset > b->offset;
>  }
>
>  /*
> @@ -324,7 +324,7 @@ uvm_pageboot_alloc(vsize_t size)
>       addr = pmap_steal_memory(size, &virtual_space_start,
>           &virtual_space_end);
>
> -     return(addr);
> +     return addr;
>
>  #else /* !PMAP_STEAL_MEMORY */
>
> @@ -380,7 +380,7 @@ uvm_pageboot_alloc(vsize_t size)
>               pmap_kenter_pa(vaddr, paddr, PROT_READ | PROT_WRITE);
>       }
>       pmap_update(pmap_kernel());
> -     return(addr);
> +     return addr;
>  #endif       /* PMAP_STEAL_MEMORY */
>  }
>
> @@ -428,7 +428,7 @@ uvm_page_physget(paddr_t *paddrp)
>                                       /* structure copy */
>                                       seg[0] = seg[1];
>                       }
> -                     return (TRUE);
> +                     return TRUE;
>               }
>
>               /* try from rear */
> @@ -446,7 +446,7 @@ uvm_page_physget(paddr_t *paddrp)
>                                       /* structure copy */
>                                       seg[0] = seg[1];
>                       }
> -                     return (TRUE);
> +                     return TRUE;
>               }
>       }
>
> @@ -478,10 +478,10 @@ uvm_page_physget(paddr_t *paddrp)
>                               /* structure copy */
>                               seg[0] = seg[1];
>               }
> -             return (TRUE);
> +             return TRUE;
>       }
>
> -     return (FALSE);        /* whoops! */
> +     return FALSE;        /* whoops! */
>  }
>
>  #endif /* PMAP_STEAL_MEMORY */
> @@ -729,7 +729,7 @@ uvm_pglistalloc(psize_t size, paddr_t lo
>       KASSERT(!(flags & UVM_PLA_WAITOK) ^ !(flags & UVM_PLA_NOWAIT));
>
>       if (size == 0)
> -             return (EINVAL);
> +             return EINVAL;
>       size = atop(round_page(size));
>
>       /*
> @@ -896,10 +896,10 @@ uvm_pagealloc(struct uvm_object *obj, vo
>       else
>               atomic_setbits_int(&pg->pg_flags, PG_CLEAN);
>
> -     return(pg);
> +     return pg;
>
>  fail:
> -     return (NULL);
> +     return NULL;
>  }
>
>  /*
> @@ -1136,7 +1136,7 @@ vm_physseg_find(paddr_t pframe, int *off
>                       if (pframe < seg->end) {
>                               if (offp)
>                                       *offp = pframe - seg->start;
> -                             return(try);            /* got it */
> +                             return try;            /* got it */
>                       }
>                       start = try + 1;        /* next time, start here */
>                       len--;                  /* "adjust" */
> @@ -1147,7 +1147,7 @@ vm_physseg_find(paddr_t pframe, int *off
>                        */
>               }
>       }
> -     return(-1);
> +     return -1;
>
>  #else
>       /* linear search for it */
> @@ -1157,10 +1157,10 @@ vm_physseg_find(paddr_t pframe, int *off
>               if (pframe >= seg->start && pframe < seg->end) {
>                       if (offp)
>                               *offp = pframe - seg->start;
> -                     return(lcv);               /* got it */
> +                     return lcv;                /* got it */
>               }
>       }
> -     return(-1);
> +     return -1;
>
>  #endif
>  }
> @@ -1178,7 +1178,7 @@ PHYS_TO_VM_PAGE(paddr_t pa)
>
>       psi = vm_physseg_find(pf, &off);
>
> -     return ((psi == -1) ? NULL : &vm_physmem[psi].pgs[off]);
> +     return (psi == -1) ? NULL : &vm_physmem[psi].pgs[off];
>  }
>  #endif /* VM_PHYSSEG_MAX > 1 */
>
> @@ -1192,7 +1192,7 @@ uvm_pagelookup(struct uvm_object *obj, v
>       struct vm_page pg;
>
>       pg.offset = off;
> -     return (RBT_FIND(uvm_objtree, &obj->memt, &pg));
> +     return RBT_FIND(uvm_objtree, &obj->memt, &pg);
>  }
>
>  /*
> Index: uvm/uvm_page.h
> ===================================================================
> RCS file: /cvs/src/sys/uvm/uvm_page.h,v
> retrieving revision 1.65
> diff -u -p -r1.65 uvm_page.h
> --- uvm/uvm_page.h    22 Sep 2020 14:31:08 -0000      1.65
> +++ uvm/uvm_page.h    23 Mar 2021 12:02:07 -0000
> @@ -256,9 +256,9 @@ vm_physseg_find(paddr_t pframe, int *off
>       if (pframe >= vm_physmem[0].start && pframe < vm_physmem[0].end) {
>               if (offp)
>                       *offp = pframe - vm_physmem[0].start;
> -             return(0);
> +             return 0;
>       }
> -     return(-1);
> +     return -1;
>  }
>
>  /*
> Index: uvm/uvm_pager.c
> ===================================================================
> RCS file: /cvs/src/sys/uvm/uvm_pager.c,v
> retrieving revision 1.75
> diff -u -p -r1.75 uvm_pager.c
> --- uvm/uvm_pager.c   12 Mar 2021 14:15:49 -0000      1.75
> +++ uvm/uvm_pager.c   23 Mar 2021 12:02:26 -0000
> @@ -326,7 +326,7 @@ uvm_mk_pcluster(struct uvm_object *uobj,
>       if ((hi - lo) >> PAGE_SHIFT > *npages) { /* pps too small, bail out! */
>               pps[0] = center;
>               *npages = 1;
> -             return(pps);
> +             return pps;
>       }
>
>       /* now determine the center and attempt to cluster around the edges */
> @@ -410,7 +410,7 @@ uvm_mk_pcluster(struct uvm_object *uobj,
>       /*
>        * done!  return the cluster array to the caller!!!
>        */
> -     return(ppsp);
> +     return ppsp;
>  }
>
>  /*
> @@ -586,7 +586,7 @@ ReTry:
>        * to worry about.
>        */
>
> -     return(result);
> +     return result;
>  }
>
>  /*
> Index: uvm/uvm_swap.c
> ===================================================================
> RCS file: /cvs/src/sys/uvm/uvm_swap.c,v
> retrieving revision 1.149
> diff -u -p -r1.149 uvm_swap.c
> --- uvm/uvm_swap.c    4 Mar 2021 09:00:03 -0000       1.149
> +++ uvm/uvm_swap.c    23 Mar 2021 12:02:35 -0000
> @@ -1426,7 +1426,7 @@ ReTry:  /* XXXMRG */
>                       sdp->swd_npginuse += *nslots;
>                       uvmexp.swpginuse += *nslots;
>                       /* done!  return drum slot number */
> -                     return(result + sdp->swd_drumoffset);
> +                     return result + sdp->swd_drumoffset;
>               }
>       }
>
> Index: uvm/uvm_vnode.c
> ===================================================================
> RCS file: /cvs/src/sys/uvm/uvm_vnode.c,v
> retrieving revision 1.112
> diff -u -p -r1.112 uvm_vnode.c
> --- uvm/uvm_vnode.c   12 Mar 2021 14:15:49 -0000      1.112
> +++ uvm/uvm_vnode.c   23 Mar 2021 12:02:43 -0000
> @@ -155,7 +155,7 @@ uvn_attach(struct vnode *vp, vm_prot_t a
>
>       /* if we're mapping a BLK device, make sure it is a disk. */
>       if (vp->v_type == VBLK && bdevsw[major(vp->v_rdev)].d_type != D_DISK) {
> -             return(NULL);
> +             return NULL;
>       }
>
>       /*
> @@ -219,7 +219,7 @@ uvn_attach(struct vnode *vp, vm_prot_t a
>               if (uvn->u_flags & UVM_VNODE_WANTED)
>                       wakeup(uvn);
>               uvn->u_flags = 0;
> -             return(NULL);
> +             return NULL;
>       }
>
>       /*
> @@ -253,7 +253,7 @@ uvn_attach(struct vnode *vp, vm_prot_t a
>       if (oldflags & UVM_VNODE_WANTED)
>               wakeup(uvn);
>
> -     return(&uvn->u_obj);
> +     return &uvn->u_obj;
>  }
>
>
> @@ -835,7 +835,7 @@ ReTry:
>
>       uvm_pglistfree(&dead);
>
> -     return(retval);
> +     return retval;
>  }
>
>  /*
> @@ -885,11 +885,11 @@ uvn_put(struct uvm_object *uobj, struct
>
>       retval = uvm_vnode_lock(uvn);
>       if (retval)
> -             return(retval);
> +             return retval;
>       retval = uvn_io(uvn, pps, npages, flags, UIO_WRITE);
>       uvm_vnode_unlock(uvn);
>
> -     return(retval);
> +     return retval;
>  }
>
>  /*
> @@ -977,9 +977,9 @@ uvn_get(struct uvm_object *uobj, voff_t
>
>               *npagesp = gotpages;            /* let caller know */
>               if (done)
> -                     return(VM_PAGER_OK);            /* bingo! */
> +                     return VM_PAGER_OK;             /* bingo! */
>               else
> -                     return(VM_PAGER_UNLOCK);
> +                     return VM_PAGER_UNLOCK;
>       }
>
>       /*
> @@ -992,7 +992,7 @@ uvn_get(struct uvm_object *uobj, voff_t
>        */
>       retval = uvm_vnode_lock(uvn);
>       if (retval)
> -             return(retval);
> +             return retval;
>
>       /*
>        * step 2: get non-resident or busy pages.
> @@ -1098,7 +1098,7 @@ uvn_get(struct uvm_object *uobj, voff_t
>                       uvm_lock_pageq();
>                       uvm_pagefree(ptmp);
>                       uvm_unlock_pageq();
> -                     return(result);
> +                     return result;
>               }
>
>               /*
> @@ -1154,7 +1154,7 @@ uvn_io(struct uvm_vnode *uvn, vm_page_t
>       /* check for sync'ing I/O. */
>       while (uvn->u_flags & UVM_VNODE_IOSYNC) {
>               if (waitf == M_NOWAIT) {
> -                     return(VM_PAGER_AGAIN);
> +                     return VM_PAGER_AGAIN;
>               }
>               uvn->u_flags |= UVM_VNODE_IOSYNCWANTED;
>               tsleep_nsec(&uvn->u_flags, PVM, "uvn_iosync", INFSLP);
> @@ -1162,7 +1162,7 @@ uvn_io(struct uvm_vnode *uvn, vm_page_t
>
>       /* check size */
>       if (file_offset >= uvn->u_size) {
> -             return(VM_PAGER_BAD);
> +             return VM_PAGER_BAD;
>       }
>
>       /* first try and map the pages in (without waiting) */
> @@ -1171,7 +1171,7 @@ uvn_io(struct uvm_vnode *uvn, vm_page_t
>
>       kva = uvm_pagermapin(pps, npages, mapinflags);
>       if (kva == 0 && waitf == M_NOWAIT) {
> -             return(VM_PAGER_AGAIN);
> +             return VM_PAGER_AGAIN;
>       }
>
>       /*
> @@ -1245,13 +1245,13 @@ uvn_io(struct uvm_vnode *uvn, vm_page_t
>       }
>
>       if (result == 0)
> -             return(VM_PAGER_OK);
> +             return VM_PAGER_OK;
>
>       if (result == EIO) {
>               /* Signal back to uvm_vnode_unlock(). */
>               uvn->u_flags |= UVM_VNODE_IOERROR;
>       }
> -     return(VM_PAGER_ERROR);
> +     return VM_PAGER_ERROR;
>  }
>
>  /*
> @@ -1301,7 +1301,7 @@ uvm_vnp_uncache(struct vnode *vp)
>
>       if ((uvn->u_flags & UVM_VNODE_VALID) == 0 ||
>                       (uvn->u_flags & UVM_VNODE_BLOCKED) != 0) {
> -             return(TRUE);
> +             return TRUE;
>       }
>
>       /*
> @@ -1310,7 +1310,7 @@ uvm_vnp_uncache(struct vnode *vp)
>        */
>       uvn->u_flags &= ~UVM_VNODE_CANPERSIST;
>       if (uvn->u_obj.uo_refs) {
> -             return(FALSE);
> +             return FALSE;
>       }
>
>       /*
> @@ -1343,7 +1343,7 @@ uvm_vnp_uncache(struct vnode *vp)
>       uvn_detach(&uvn->u_obj);
>       vn_lock(vp, LK_EXCLUSIVE | LK_RETRY);
>
> -     return(TRUE);
> +     return TRUE;
>  }
>
>  /*
> @@ -1476,7 +1476,7 @@ uvm_vnode_lock(struct uvm_vnode *uvn)
>       int netunlocked = 0;
>
>       if (uvn->u_flags & UVM_VNODE_VNISLOCKED)
> -             return(VM_PAGER_OK);
> +             return VM_PAGER_OK;
>
>       /*
>        * This thread may already have the net lock, if we faulted in copyin()
> @@ -1499,7 +1499,7 @@ uvm_vnode_lock(struct uvm_vnode *uvn)
>       error = vn_lock(uvn->u_vnode, LK_EXCLUSIVE | LK_RECURSEFAIL);
>       if (netunlocked)
>               NET_LOCK();
> -     return(error ? VM_PAGER_ERROR : VM_PAGER_OK);
> +     return error ? VM_PAGER_ERROR : VM_PAGER_OK;
>  }
>
>  void
>

Reply via email to