Module Name: src Committed By: skrll Date: Thu Jul 9 05:57:15 UTC 2020
Modified Files: src/sys/uvm: uvm_amap.c uvm_anon.c uvm_aobj.c uvm_bio.c uvm_device.c uvm_fault.c uvm_km.c uvm_map.c uvm_page.c uvm_pager.c uvm_pdaemon.c uvm_swap.c uvm_vnode.c Log Message: Consistently use UVMHIST(__func__) Convert UVMHIST_{CALLED,LOG} into UVMHIST_CALLARGS To generate a diff of this commit: cvs rdiff -u -r1.121 -r1.122 src/sys/uvm/uvm_amap.c cvs rdiff -u -r1.78 -r1.79 src/sys/uvm/uvm_anon.c cvs rdiff -u -r1.148 -r1.149 src/sys/uvm/uvm_aobj.c cvs rdiff -u -r1.119 -r1.120 src/sys/uvm/uvm_bio.c cvs rdiff -u -r1.70 -r1.71 src/sys/uvm/uvm_device.c cvs rdiff -u -r1.227 -r1.228 src/sys/uvm/uvm_fault.c cvs rdiff -u -r1.158 -r1.159 src/sys/uvm/uvm_km.c cvs rdiff -u -r1.384 -r1.385 src/sys/uvm/uvm_map.c cvs rdiff -u -r1.243 -r1.244 src/sys/uvm/uvm_page.c cvs rdiff -u -r1.127 -r1.128 src/sys/uvm/uvm_pager.c cvs rdiff -u -r1.129 -r1.130 src/sys/uvm/uvm_pdaemon.c cvs rdiff -u -r1.196 -r1.197 src/sys/uvm/uvm_swap.c cvs rdiff -u -r1.114 -r1.115 src/sys/uvm/uvm_vnode.c Please note that diffs are not public domain; they are subject to the copyright notices on the relevant files.
Modified files: Index: src/sys/uvm/uvm_amap.c diff -u src/sys/uvm/uvm_amap.c:1.121 src/sys/uvm/uvm_amap.c:1.122 --- src/sys/uvm/uvm_amap.c:1.121 Wed Jul 8 13:26:22 2020 +++ src/sys/uvm/uvm_amap.c Thu Jul 9 05:57:15 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: uvm_amap.c,v 1.121 2020/07/08 13:26:22 skrll Exp $ */ +/* $NetBSD: uvm_amap.c,v 1.122 2020/07/09 05:57:15 skrll Exp $ */ /* * Copyright (c) 1997 Charles D. Cranor and Washington University. @@ -35,7 +35,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: uvm_amap.c,v 1.121 2020/07/08 13:26:22 skrll Exp $"); +__KERNEL_RCSID(0, "$NetBSD: uvm_amap.c,v 1.122 2020/07/09 05:57:15 skrll Exp $"); #include "opt_uvmhist.h" @@ -235,7 +235,7 @@ amap_alloc(vaddr_t sz, vaddr_t padsz, in { struct vm_amap *amap; int slots, padslots; - UVMHIST_FUNC("amap_alloc"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); AMAP_B2SLOT(slots, sz); AMAP_B2SLOT(padslots, padsz); @@ -321,7 +321,7 @@ amap_free(struct vm_amap *amap) { int slots; - UVMHIST_FUNC("amap_free"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); KASSERT(amap->am_ref == 0 && amap->am_nused == 0); KASSERT((amap->am_flags & AMAP_SWAPOFF) == 0); @@ -363,9 +363,8 @@ amap_extend(struct vm_map_entry *entry, const km_flag_t kmflags = (flags & AMAP_EXTEND_NOWAIT) ? KM_NOSLEEP : KM_SLEEP; - UVMHIST_FUNC("amap_extend"); UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist, " (entry=%#jx, addsize=%#jx, flags=%#jx)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist, " (entry=%#jx, addsize=%#jx, flags=%#jx)", (uintptr_t)entry, addsize, flags, 0); /* @@ -725,8 +724,8 @@ amap_wipeout(struct vm_amap *amap) { u_int lcv; - UVMHIST_FUNC("amap_wipeout"); UVMHIST_CALLED(maphist); - UVMHIST_LOG(maphist,"(amap=%#jx)", (uintptr_t)amap, 0,0,0); + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(amap=%#jx)", (uintptr_t)amap, 0,0,0); KASSERT(rw_write_held(amap->am_lock)); KASSERT(amap->am_ref == 0); @@ -797,9 +796,9 @@ amap_copy(struct vm_map *map, struct vm_ krwlock_t *oldlock; vsize_t len; - UVMHIST_FUNC("amap_copy"); UVMHIST_CALLED(maphist); - UVMHIST_LOG(maphist, " (map=%#j, entry=%#j, flags=%jd)", - (uintptr_t)map, (uintptr_t)entry, flags, 0); + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist, " (map=%#j, entry=%#j, flags=%jd)", + (uintptr_t)map, (uintptr_t)entry, flags, -2); KASSERT(map != kernel_map); /* we use nointr pool */ @@ -1416,7 +1415,7 @@ amap_lookup(struct vm_aref *aref, vaddr_ struct vm_anon *an; u_int slot; - UVMHIST_FUNC("amap_lookup"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); KASSERT(rw_lock_held(amap->am_lock)); AMAP_B2SLOT(slot, offset); @@ -1445,7 +1444,7 @@ amap_lookups(struct vm_aref *aref, vaddr struct vm_amap *amap = aref->ar_amap; u_int slot; - UVMHIST_FUNC("amap_lookups"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); KASSERT(rw_lock_held(amap->am_lock)); AMAP_B2SLOT(slot, offset); @@ -1483,7 +1482,7 @@ amap_add(struct vm_aref *aref, vaddr_t o struct vm_amap *amap = aref->ar_amap; u_int slot; - UVMHIST_FUNC("amap_add"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); KASSERT(rw_write_held(amap->am_lock)); KASSERT(anon->an_lock == amap->am_lock); @@ -1525,7 +1524,7 @@ amap_unadd(struct vm_aref *aref, vaddr_t struct vm_amap *amap = aref->ar_amap; u_int slot, ptr, last; - UVMHIST_FUNC("amap_unadd"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); KASSERT(rw_write_held(amap->am_lock)); AMAP_B2SLOT(slot, offset); @@ -1593,7 +1592,7 @@ amap_adjref_anons(struct vm_amap *amap, void amap_ref(struct vm_amap *amap, vaddr_t offset, vsize_t len, int flags) { - UVMHIST_FUNC("amap_ref"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); amap_lock(amap, RW_WRITER); if (flags & AMAP_SHARED) { @@ -1614,7 +1613,7 @@ amap_ref(struct vm_amap *amap, vaddr_t o void amap_unref(struct vm_amap *amap, vaddr_t offset, vsize_t len, bool all) { - UVMHIST_FUNC("amap_unref"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); amap_lock(amap, RW_WRITER); Index: src/sys/uvm/uvm_anon.c diff -u src/sys/uvm/uvm_anon.c:1.78 src/sys/uvm/uvm_anon.c:1.79 --- src/sys/uvm/uvm_anon.c:1.78 Wed Jul 8 13:26:22 2020 +++ src/sys/uvm/uvm_anon.c Thu Jul 9 05:57:15 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: uvm_anon.c,v 1.78 2020/07/08 13:26:22 skrll Exp $ */ +/* $NetBSD: uvm_anon.c,v 1.79 2020/07/09 05:57:15 skrll Exp $ */ /* * Copyright (c) 1997 Charles D. Cranor and Washington University. @@ -30,7 +30,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: uvm_anon.c,v 1.78 2020/07/08 13:26:22 skrll Exp $"); +__KERNEL_RCSID(0, "$NetBSD: uvm_anon.c,v 1.79 2020/07/09 05:57:15 skrll Exp $"); #include "opt_uvmhist.h" @@ -106,8 +106,8 @@ uvm_anfree(struct vm_anon *anon) { struct vm_page *pg = anon->an_page, *pg2 __diagused; - UVMHIST_FUNC("uvm_anon_dispose"); UVMHIST_CALLED(maphist); - UVMHIST_LOG(maphist,"(anon=%#jx)", (uintptr_t)anon, 0,0,0); + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(anon=%#jx)", (uintptr_t)anon, 0,0,0); KASSERT(anon->an_lock == NULL || rw_write_held(anon->an_lock)); KASSERT(anon->an_ref == 0); @@ -336,7 +336,7 @@ uvm_anon_pagein(struct vm_amap *amap, st void uvm_anon_dropswap(struct vm_anon *anon) { - UVMHIST_FUNC("uvm_anon_dropswap"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); if (anon->an_swslot == 0) return; Index: src/sys/uvm/uvm_aobj.c diff -u src/sys/uvm/uvm_aobj.c:1.148 src/sys/uvm/uvm_aobj.c:1.149 --- src/sys/uvm/uvm_aobj.c:1.148 Wed Jul 8 13:26:22 2020 +++ src/sys/uvm/uvm_aobj.c Thu Jul 9 05:57:15 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: uvm_aobj.c,v 1.148 2020/07/08 13:26:22 skrll Exp $ */ +/* $NetBSD: uvm_aobj.c,v 1.149 2020/07/09 05:57:15 skrll Exp $ */ /* * Copyright (c) 1998 Chuck Silvers, Charles D. Cranor and @@ -38,7 +38,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: uvm_aobj.c,v 1.148 2020/07/08 13:26:22 skrll Exp $"); +__KERNEL_RCSID(0, "$NetBSD: uvm_aobj.c,v 1.149 2020/07/09 05:57:15 skrll Exp $"); #ifdef _KERNEL_OPT #include "opt_uvmhist.h" @@ -290,8 +290,8 @@ uao_set_swslot(struct uvm_object *uobj, struct uvm_aobj *aobj = (struct uvm_aobj *)uobj; struct uao_swhash_elt *elt; int oldslot; - UVMHIST_FUNC("uao_set_swslot"); UVMHIST_CALLED(pdhist); - UVMHIST_LOG(pdhist, "aobj %#jx pageidx %jd slot %jd", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(pdhist, "aobj %#jx pageidx %jd slot %jd", (uintptr_t)aobj, pageidx, slot, 0); KASSERT(rw_write_held(uobj->vmobjlock) || uobj->uo_refs == 0); @@ -583,7 +583,7 @@ uao_detach(struct uvm_object *uobj) struct uvm_page_array a; struct vm_page *pg; - UVMHIST_FUNC("uao_detach"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); /* * Detaching from kernel object is a NOP. @@ -666,7 +666,7 @@ uao_put(struct uvm_object *uobj, voff_t struct uvm_page_array a; struct vm_page *pg; voff_t curoff; - UVMHIST_FUNC("uao_put"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); KASSERT(UVM_OBJ_IS_AOBJ(uobj)); KASSERT(rw_write_held(uobj->vmobjlock)); @@ -801,11 +801,11 @@ uao_get(struct uvm_object *uobj, voff_t voff_t current_offset; struct vm_page *ptmp; int lcv, gotpages, maxpages, swslot, pageidx; - UVMHIST_FUNC("uao_get"); UVMHIST_CALLED(pdhist); bool overwrite = ((flags & PGO_OVERWRITE) != 0); struct uvm_page_array a; - UVMHIST_LOG(pdhist, "aobj=%#jx offset=%jd, flags=%jd", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(pdhist, "aobj=%#jx offset=%jd, flags=%jd", (uintptr_t)uobj, offset, flags,0); /* Index: src/sys/uvm/uvm_bio.c diff -u src/sys/uvm/uvm_bio.c:1.119 src/sys/uvm/uvm_bio.c:1.120 --- src/sys/uvm/uvm_bio.c:1.119 Wed Jul 8 13:26:22 2020 +++ src/sys/uvm/uvm_bio.c Thu Jul 9 05:57:15 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: uvm_bio.c,v 1.119 2020/07/08 13:26:22 skrll Exp $ */ +/* $NetBSD: uvm_bio.c,v 1.120 2020/07/09 05:57:15 skrll Exp $ */ /* * Copyright (c) 1998 Chuck Silvers. @@ -34,7 +34,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: uvm_bio.c,v 1.119 2020/07/08 13:26:22 skrll Exp $"); +__KERNEL_RCSID(0, "$NetBSD: uvm_bio.c,v 1.120 2020/07/09 05:57:15 skrll Exp $"); #include "opt_uvmhist.h" #include "opt_ubc.h" @@ -311,7 +311,7 @@ ubc_fault(struct uvm_faultinfo *ufi, vad int i, error, npages; vm_prot_t prot; - UVMHIST_FUNC("ubc_fault"); UVMHIST_CALLED(ubchist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(ubchist); /* * no need to try with PGO_LOCKED... @@ -482,9 +482,8 @@ ubc_alloc(struct uvm_object *uobj, voff_ struct ubc_map *umap; voff_t umap_offset; int error; - UVMHIST_FUNC("ubc_alloc"); UVMHIST_CALLED(ubchist); - - UVMHIST_LOG(ubchist, "uobj %#jx offset 0x%jx len 0x%jx", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(ubchist, "uobj %#jx offset 0x%jx len 0x%jx", (uintptr_t)uobj, offset, *lenp, 0); KASSERT(*lenp > 0); @@ -640,9 +639,9 @@ ubc_release(void *va, int flags, struct struct uvm_object *uobj; vaddr_t umapva; bool unmapped; - UVMHIST_FUNC("ubc_release"); UVMHIST_CALLED(ubchist); + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(ubchist, "va %#jx", (uintptr_t)va, 0, 0, 0); - UVMHIST_LOG(ubchist, "va %#jx", (uintptr_t)va, 0, 0, 0); umap = &ubc_object.umap[((char *)va - ubc_object.kva) >> ubc_winshift]; umapva = UBC_UMAP_ADDR(umap); uobj = umap->uobj; @@ -844,7 +843,7 @@ ubc_alloc_direct(struct uvm_object *uobj int error; int gpflags = flags | PGO_NOTIMESTAMP | PGO_SYNCIO; int access_type = VM_PROT_READ; - UVMHIST_FUNC("ubc_alloc_direct"); UVMHIST_CALLED(ubchist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(ubchist); if (flags & UBC_WRITE) { if (flags & UBC_FAULTBUSY) Index: src/sys/uvm/uvm_device.c diff -u src/sys/uvm/uvm_device.c:1.70 src/sys/uvm/uvm_device.c:1.71 --- src/sys/uvm/uvm_device.c:1.70 Mon Feb 24 12:38:57 2020 +++ src/sys/uvm/uvm_device.c Thu Jul 9 05:57:15 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: uvm_device.c,v 1.70 2020/02/24 12:38:57 rin Exp $ */ +/* $NetBSD: uvm_device.c,v 1.71 2020/07/09 05:57:15 skrll Exp $ */ /* * Copyright (c) 1997 Charles D. Cranor and Washington University. @@ -32,7 +32,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: uvm_device.c,v 1.70 2020/02/24 12:38:57 rin Exp $"); +__KERNEL_RCSID(0, "$NetBSD: uvm_device.c,v 1.71 2020/07/09 05:57:15 skrll Exp $"); #include "opt_uvmhist.h" @@ -114,9 +114,8 @@ udv_attach(dev_t device, vm_prot_t acces const struct cdevsw *cdev; dev_type_mmap((*mapfn)); - UVMHIST_FUNC("udv_attach"); UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist, "(device=%#jx)", device,0,0,0); + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist, "(device=%#jx)", device,0,0,0); /* * before we do anything, ensure this device supports mmap @@ -269,7 +268,7 @@ udv_attach(dev_t device, vm_prot_t acces static void udv_reference(struct uvm_object *uobj) { - UVMHIST_FUNC("udv_reference"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); rw_enter(uobj->vmobjlock, RW_WRITER); uobj->uo_refs++; @@ -290,7 +289,7 @@ static void udv_detach(struct uvm_object *uobj) { struct uvm_device *udv = (struct uvm_device *)uobj; - UVMHIST_FUNC("udv_detach"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); /* * loop until done @@ -364,7 +363,7 @@ udv_fault(struct uvm_faultinfo *ufi, vad int lcv, retval; dev_t device; vm_prot_t mapprot; - UVMHIST_FUNC("udv_fault"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); UVMHIST_LOG(maphist," flags=%jd", flags,0,0,0); /* Index: src/sys/uvm/uvm_fault.c diff -u src/sys/uvm/uvm_fault.c:1.227 src/sys/uvm/uvm_fault.c:1.228 --- src/sys/uvm/uvm_fault.c:1.227 Sun May 17 19:38:17 2020 +++ src/sys/uvm/uvm_fault.c Thu Jul 9 05:57:15 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: uvm_fault.c,v 1.227 2020/05/17 19:38:17 ad Exp $ */ +/* $NetBSD: uvm_fault.c,v 1.228 2020/07/09 05:57:15 skrll Exp $ */ /* * Copyright (c) 1997 Charles D. Cranor and Washington University. @@ -32,7 +32,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: uvm_fault.c,v 1.227 2020/05/17 19:38:17 ad Exp $"); +__KERNEL_RCSID(0, "$NetBSD: uvm_fault.c,v 1.228 2020/07/09 05:57:15 skrll Exp $"); #include "opt_uvmhist.h" @@ -280,7 +280,7 @@ uvmfault_anonget(struct uvm_faultinfo *u krw_t lock_type; int error; - UVMHIST_FUNC("uvmfault_anonget"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); KASSERT(rw_lock_held(anon->an_lock)); KASSERT(anon->an_lock == amap->am_lock); @@ -859,9 +859,8 @@ uvm_fault_internal(struct vm_map *orig_m struct vm_page *pages_store[UVM_MAXRANGE], **pages; int error; - UVMHIST_FUNC("uvm_fault"); UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist, "(map=%#jx, vaddr=%#jx, at=%jd, ff=%jd)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist, "(map=%#jx, vaddr=%#jx, at=%jd, ff=%jd)", (uintptr_t)orig_map, vaddr, access_type, fault_flag); /* Don't count anything until user interaction is possible */ @@ -979,7 +978,7 @@ uvm_fault_check( struct uvm_object *uobj; vm_prot_t check_prot; int nback, nforw; - UVMHIST_FUNC("uvm_fault_check"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); /* * lookup and lock the maps @@ -1291,7 +1290,7 @@ uvm_fault_upper_lookup( vaddr_t currva; bool shadowed __unused; bool entered; - UVMHIST_FUNC("uvm_fault_upper_lookup"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); /* locked: maps(read), amap(if there) */ KASSERT(amap == NULL || @@ -1367,7 +1366,7 @@ uvm_fault_upper_neighbor( struct uvm_faultinfo *ufi, const struct uvm_faultctx *flt, vaddr_t currva, struct vm_page *pg, bool readonly) { - UVMHIST_FUNC("uvm_fault_upper_neighbor"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); /* locked: amap, anon */ @@ -1423,7 +1422,7 @@ uvm_fault_upper( struct vm_anon * const anon = anons[flt->centeridx]; struct uvm_object *uobj; int error; - UVMHIST_FUNC("uvm_fault_upper"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); /* locked: maps(read), amap, anon */ KASSERT(rw_lock_op(amap->am_lock) == flt->upper_lock_type); @@ -1533,7 +1532,7 @@ uvm_fault_upper_loan( { struct vm_amap * const amap = ufi->entry->aref.ar_amap; int error = 0; - UVMHIST_FUNC("uvm_fault_upper_loan"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); if (!flt->cow_now) { @@ -1599,7 +1598,7 @@ uvm_fault_upper_promote( struct vm_anon * const oanon = anon; struct vm_page *pg; int error; - UVMHIST_FUNC("uvm_fault_upper_promote"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); UVMHIST_LOG(maphist, " case 1B: COW fault",0,0,0,0); cpu_count(CPU_COUNT_FLT_ACOW, 1); @@ -1650,7 +1649,7 @@ uvm_fault_upper_direct( { struct vm_anon * const oanon = anon; struct vm_page *pg; - UVMHIST_FUNC("uvm_fault_upper_direct"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); cpu_count(CPU_COUNT_FLT_ANON, 1); pg = anon->an_page; @@ -1673,7 +1672,7 @@ uvm_fault_upper_enter( struct pmap *pmap = ufi->orig_map->pmap; vaddr_t va = ufi->orig_rvaddr; struct vm_amap * const amap = ufi->entry->aref.ar_amap; - UVMHIST_FUNC("uvm_fault_upper_enter"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); /* locked: maps(read), amap, oanon, anon(if different from oanon) */ KASSERT(rw_lock_op(amap->am_lock) == flt->upper_lock_type); @@ -1763,7 +1762,7 @@ uvm_fault_upper_done( { const bool wire_paging = flt->wire_paging; - UVMHIST_FUNC("uvm_fault_upper_done"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); /* * ... update the page queues. @@ -1857,7 +1856,7 @@ uvm_fault_lower( struct uvm_object *uobj = ufi->entry->object.uvm_obj; struct vm_page *uobjpage; int error; - UVMHIST_FUNC("uvm_fault_lower"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); /* * now, if the desired page is not shadowed by the amap and we have @@ -1979,7 +1978,7 @@ uvm_fault_lower_lookup( int lcv, gotpages; vaddr_t currva; bool entered; - UVMHIST_FUNC("uvm_fault_lower_lookup"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); rw_enter(uobj->vmobjlock, flt->lower_lock_type); @@ -2122,7 +2121,7 @@ uvm_fault_lower_io( voff_t uoff; vm_prot_t access_type; int advice; - UVMHIST_FUNC("uvm_fault_lower_io"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); /* update rusage counters */ curlwp->l_ru.ru_majflt++; @@ -2266,7 +2265,7 @@ uvm_fault_lower_direct( struct uvm_object *uobj, struct vm_page *uobjpage) { struct vm_page *pg; - UVMHIST_FUNC("uvm_fault_lower_direct"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); /* * we are not promoting. if the mapping is COW ensure that we @@ -2315,7 +2314,7 @@ uvm_fault_lower_direct_loan( struct vm_page *pg; struct vm_page *uobjpage = *ruobjpage; int error; - UVMHIST_FUNC("uvm_fault_lower_direct_loan"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); if (!flt->cow_now) { /* read fault: cap the protection at readonly */ @@ -2378,7 +2377,7 @@ uvm_fault_lower_promote( struct vm_anon *anon; struct vm_page *pg; int error; - UVMHIST_FUNC("uvm_fault_lower_promote"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); KASSERT(amap != NULL); @@ -2459,7 +2458,7 @@ uvm_fault_lower_enter( struct vm_amap * const amap = ufi->entry->aref.ar_amap; const bool readonly = uvm_pagereadonly_p(pg); int error; - UVMHIST_FUNC("uvm_fault_lower_enter"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); /* * Locked: @@ -2559,7 +2558,7 @@ uvm_fault_lower_done( struct uvm_object *uobj, struct vm_page *pg) { - UVMHIST_FUNC("uvm_fault_lower_done"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); if (flt->wire_paging) { uvm_pagelock(pg); Index: src/sys/uvm/uvm_km.c diff -u src/sys/uvm/uvm_km.c:1.158 src/sys/uvm/uvm_km.c:1.159 --- src/sys/uvm/uvm_km.c:1.158 Wed Jul 8 13:26:22 2020 +++ src/sys/uvm/uvm_km.c Thu Jul 9 05:57:15 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: uvm_km.c,v 1.158 2020/07/08 13:26:22 skrll Exp $ */ +/* $NetBSD: uvm_km.c,v 1.159 2020/07/09 05:57:15 skrll Exp $ */ /* * Copyright (c) 1997 Charles D. Cranor and Washington University. @@ -152,7 +152,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: uvm_km.c,v 1.158 2020/07/08 13:26:22 skrll Exp $"); +__KERNEL_RCSID(0, "$NetBSD: uvm_km.c,v 1.159 2020/07/09 05:57:15 skrll Exp $"); #include "opt_uvmhist.h" @@ -262,8 +262,8 @@ uvm_km_bootstrap(vaddr_t start, vaddr_t struct uvm_map_args args; int error; - UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); - UVMHIST_LOG(maphist, "start=%#jx end=%#jx", start, end, 0,0); + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist, "start=%#jx end=%#jx", start, end, 0,0); kmeminit_nkmempages(); kmemsize = (vsize_t)nkmempages * PAGE_SIZE; Index: src/sys/uvm/uvm_map.c diff -u src/sys/uvm/uvm_map.c:1.384 src/sys/uvm/uvm_map.c:1.385 --- src/sys/uvm/uvm_map.c:1.384 Sat May 30 08:50:31 2020 +++ src/sys/uvm/uvm_map.c Thu Jul 9 05:57:15 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: uvm_map.c,v 1.384 2020/05/30 08:50:31 maxv Exp $ */ +/* $NetBSD: uvm_map.c,v 1.385 2020/07/09 05:57:15 skrll Exp $ */ /* * Copyright (c) 1997 Charles D. Cranor and Washington University. @@ -66,7 +66,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: uvm_map.c,v 1.384 2020/05/30 08:50:31 maxv Exp $"); +__KERNEL_RCSID(0, "$NetBSD: uvm_map.c,v 1.385 2020/07/09 05:57:15 skrll Exp $"); #include "opt_ddb.h" #include "opt_pax.h" @@ -785,7 +785,7 @@ uvm_mapent_alloc(struct vm_map *map, int { struct vm_map_entry *me; int pflags = (flags & UVM_FLAG_NOWAIT) ? PR_NOWAIT : PR_WAITOK; - UVMHIST_FUNC("uvm_mapent_alloc"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); me = pool_cache_get(&uvm_map_entry_cache, pflags); if (__predict_false(me == NULL)) { @@ -805,9 +805,8 @@ uvm_mapent_alloc(struct vm_map *map, int static void uvm_mapent_free(struct vm_map_entry *me) { - UVMHIST_FUNC("uvm_mapent_free"); UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist,"<- freeing map entry=%#jx [flags=%jd]", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"<- freeing map entry=%#jx [flags=%jd]", (uintptr_t)me, me->flags, 0, 0); pool_cache_put(&uvm_map_entry_cache, me); } @@ -913,7 +912,7 @@ uvm_map_init(void) * first, init logging system. */ - UVMHIST_FUNC("uvm_map_init"); + UVMHIST_FUNC(__func__); UVMHIST_LINK_STATIC(maphist); UVMHIST_INIT_STATIC(pdhist, pdhistbuf); UVMHIST_CALLED(maphist); @@ -1128,10 +1127,8 @@ uvm_map_prepare(struct vm_map *map, vadd vm_prot_t prot = UVM_PROTECTION(flags); vm_prot_t maxprot = UVM_MAXPROTECTION(flags); - UVMHIST_FUNC("uvm_map_prepare"); - UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist, "(map=%#jx, start=%#jx, size=%ju, flags=%#jx)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist, "(map=%#jx, start=%#jx, size=%ju, flags=%#jx)", (uintptr_t)map, start, size, flags); UVMHIST_LOG(maphist, " uobj/offset %#jx/%jd", (uintptr_t)uobj, uoffset,0,0); @@ -1301,10 +1298,8 @@ uvm_map_enter(struct vm_map *map, const int error; int newetype; - UVMHIST_FUNC("uvm_map_enter"); - UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist, "(map=%#jx, start=%#jx, size=%ju, flags=%#jx)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist, "(map=%#jx, start=%#jx, size=%ju, flags=%#jx)", (uintptr_t)map, start, size, flags); UVMHIST_LOG(maphist, " uobj/offset %#jx/%jd", (uintptr_t)uobj, uoffset,0,0); @@ -1677,10 +1672,8 @@ uvm_map_lookup_entry(struct vm_map *map, struct vm_map_entry **entry /* OUT */) { struct vm_map_entry *cur; - UVMHIST_FUNC("uvm_map_lookup_entry"); - UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist,"(map=%#jx,addr=%#jx,ent=%#jx)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(map=%#jx,addr=%#jx,ent=%#jx)", (uintptr_t)map, address, (uintptr_t)entry, 0); /* @@ -1814,11 +1807,10 @@ uvm_map_findspace(struct vm_map *map, va struct vm_map_entry *child, *prev, *tmp; vaddr_t orig_hint __diagused; const int topdown = map->flags & VM_MAP_TOPDOWN; - UVMHIST_FUNC("uvm_map_findspace"); - UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist, "(map=%#jx, hint=%#jx, len=%ju, flags=%#jx)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist, "(map=%#jx, hint=%#jx, len=%ju, flags=%#jx)", (uintptr_t)map, hint, length, flags); + KASSERT((flags & UVM_FLAG_COLORMATCH) != 0 || powerof2(align)); KASSERT((flags & UVM_FLAG_COLORMATCH) == 0 || align < uvmexp.ncolors); KASSERT((flags & UVM_FLAG_FIXED) == 0 || align == 0); @@ -2161,9 +2153,8 @@ uvm_unmap_remove(struct vm_map *map, vad { struct vm_map_entry *entry, *first_entry, *next; vaddr_t len; - UVMHIST_FUNC("uvm_unmap_remove"); UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist,"(map=%#jx, start=%#jx, end=%#jx)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(map=%#jx, start=%#jx, end=%#jx)", (uintptr_t)map, start, end, 0); VM_MAP_RANGE_CHECK(map, start, end); @@ -2349,7 +2340,7 @@ void uvm_unmap_detach(struct vm_map_entry *first_entry, int flags) { struct vm_map_entry *next_entry; - UVMHIST_FUNC("uvm_unmap_detach"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); while (first_entry) { KASSERT(!VM_MAPENT_ISWIRED(first_entry)); @@ -2405,9 +2396,8 @@ uvm_map_reserve(struct vm_map *map, vsiz vaddr_t *raddr /* IN:hint, OUT: reserved VA */, uvm_flag_t flags /* UVM_FLAG_FIXED or UVM_FLAG_COLORMATCH or 0 */) { - UVMHIST_FUNC("uvm_map_reserve"); UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist, "(map=%#jx, size=%#jx, offset=%#jx, addr=%#jx)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist, "(map=%#jx, size=%#jx, offset=%#jx, addr=%#jx)", (uintptr_t)map, size, offset, (uintptr_t)raddr); size = round_page(size); @@ -2589,9 +2579,8 @@ uvm_map_extract(struct vm_map *srcmap, v vsize_t elen __unused; int nchain, error, copy_ok; vsize_t nsize; - UVMHIST_FUNC("uvm_map_extract"); UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist,"(srcmap=%#jx,start=%#jx, len=%#jx", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(srcmap=%#jx,start=%#jx, len=%#jx", (uintptr_t)srcmap, start, len, 0); UVMHIST_LOG(maphist," ...,dstmap=%#jx, flags=%#jx)", (uintptr_t)dstmap, flags, 0, 0); @@ -3007,8 +2996,8 @@ uvm_map_protect(struct vm_map *map, vadd { struct vm_map_entry *current, *entry; int error = 0; - UVMHIST_FUNC("uvm_map_protect"); UVMHIST_CALLED(maphist); - UVMHIST_LOG(maphist,"(map=%#jx,start=%#jx,end=%#jx,new_prot=%#jx)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(map=%#jx,start=%#jx,end=%#jx,new_prot=%#jx)", (uintptr_t)map, start, end, new_prot); vm_map_lock(map); @@ -3167,8 +3156,8 @@ uvm_map_inherit(struct vm_map *map, vadd vm_inherit_t new_inheritance) { struct vm_map_entry *entry, *temp_entry; - UVMHIST_FUNC("uvm_map_inherit"); UVMHIST_CALLED(maphist); - UVMHIST_LOG(maphist,"(map=%#jx,start=%#jx,end=%#jx,new_inh=%#jx)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(map=%#jx,start=%#jx,end=%#jx,new_inh=%#jx)", (uintptr_t)map, start, end, new_inheritance); switch (new_inheritance) { @@ -3210,8 +3199,8 @@ int uvm_map_advice(struct vm_map *map, vaddr_t start, vaddr_t end, int new_advice) { struct vm_map_entry *entry, *temp_entry; - UVMHIST_FUNC("uvm_map_advice"); UVMHIST_CALLED(maphist); - UVMHIST_LOG(maphist,"(map=%#jx,start=%#jx,end=%#jx,new_adv=%#jx)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(map=%#jx,start=%#jx,end=%#jx,new_adv=%#jx)", (uintptr_t)map, start, end, new_advice); vm_map_lock(map); @@ -3259,8 +3248,8 @@ int uvm_map_willneed(struct vm_map *map, vaddr_t start, vaddr_t end) { struct vm_map_entry *entry; - UVMHIST_FUNC("uvm_map_willneed"); UVMHIST_CALLED(maphist); - UVMHIST_LOG(maphist,"(map=%#jx,start=%#jx,end=%#jx)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(map=%#jx,start=%#jx,end=%#jx)", (uintptr_t)map, start, end, 0); vm_map_lock_read(map); @@ -3326,8 +3315,8 @@ uvm_map_pageable(struct vm_map *map, vad #ifdef DIAGNOSTIC u_int timestamp_save; #endif - UVMHIST_FUNC("uvm_map_pageable"); UVMHIST_CALLED(maphist); - UVMHIST_LOG(maphist,"(map=%#jx,start=%#jx,end=%#jx,new_pageable=%ju)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(map=%#jx,start=%#jx,end=%#jx,new_pageable=%ju)", (uintptr_t)map, start, end, new_pageable); KASSERT(map->flags & VM_MAP_PAGEABLE); @@ -3582,8 +3571,8 @@ uvm_map_pageable_all(struct vm_map *map, #ifdef DIAGNOSTIC u_int timestamp_save; #endif - UVMHIST_FUNC("uvm_map_pageable_all"); UVMHIST_CALLED(maphist); - UVMHIST_LOG(maphist,"(map=%#jx,flags=%#jx)", (uintptr_t)map, flags, + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(map=%#jx,flags=%#jx)", (uintptr_t)map, flags, 0, 0); KASSERT(map->flags & VM_MAP_PAGEABLE); @@ -3817,10 +3806,10 @@ uvm_map_clean(struct vm_map *map, vaddr_ vsize_t size; voff_t uoff; int error, refs; - UVMHIST_FUNC("uvm_map_clean"); UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist,"(map=%#jx,start=%#jx,end=%#jx,flags=%#jx)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(map=%#jx,start=%#jx,end=%#jx,flags=%#jx)", (uintptr_t)map, start, end, flags); + KASSERT((flags & (PGO_FREE|PGO_DEACTIVATE)) != (PGO_FREE|PGO_DEACTIVATE)); @@ -4018,7 +4007,7 @@ struct vmspace * uvmspace_alloc(vaddr_t vmin, vaddr_t vmax, bool topdown) { struct vmspace *vm; - UVMHIST_FUNC("uvmspace_alloc"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); vm = pool_cache_get(&uvm_vmspace_cache, PR_WAITOK); uvmspace_init(vm, NULL, vmin, vmax, topdown); @@ -4036,9 +4025,8 @@ void uvmspace_init(struct vmspace *vm, struct pmap *pmap, vaddr_t vmin, vaddr_t vmax, bool topdown) { - UVMHIST_FUNC("uvmspace_init"); UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist, "(vm=%#jx, pmap=%#jx, vmin=%#jx, vmax=%#jx", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist, "(vm=%#jx, pmap=%#jx, vmin=%#jx, vmax=%#jx", (uintptr_t)vm, (uintptr_t)pmap, vmin, vmax); UVMHIST_LOG(maphist, " topdown=%ju)", topdown, 0, 0, 0); @@ -4243,10 +4231,9 @@ uvmspace_free(struct vmspace *vm) struct vm_map *map = &vm->vm_map; int flags; - UVMHIST_FUNC("uvmspace_free"); UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist,"(vm=%#jx) ref=%jd", (uintptr_t)vm, vm->vm_refcnt, - 0, 0); + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(vm=%#jx) ref=%jd", (uintptr_t)vm, + vm->vm_refcnt, 0, 0); if (atomic_dec_uint_nv(&vm->vm_refcnt) > 0) return; @@ -4492,7 +4479,7 @@ uvmspace_fork(struct vmspace *vm1) struct vm_map *old_map = &vm1->vm_map; struct vm_map *new_map; struct vm_map_entry *old_entry; - UVMHIST_FUNC("uvmspace_fork"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); vm_map_lock(old_map); @@ -4717,13 +4704,13 @@ void uvm_unmap1(struct vm_map *map, vaddr_t start, vaddr_t end, int flags) { struct vm_map_entry *dead_entries; - UVMHIST_FUNC("uvm_unmap"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist, " (map=%#jx, start=%#jx, end=%#jx)", + (uintptr_t)map, start, end, 0); KASSERTMSG(start < end, "%s: map %p: start %#jx < end %#jx", __func__, map, (uintmax_t)start, (uintmax_t)end); - UVMHIST_LOG(maphist, " (map=%#jx, start=%#jx, end=%#jx)", - (uintptr_t)map, start, end, 0); if (map == kernel_map) { LOCKDEBUG_MEM_CHECK((void *)start, end - start); } @@ -4823,7 +4810,7 @@ uvm_voaddr_acquire(struct vm_map * const bool exclusive = false; void (*unlock_fn)(struct vm_map *); - UVMHIST_FUNC("uvm_voaddr_acquire"); UVMHIST_CALLED(maphist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist); UVMHIST_LOG(maphist,"(map=%#jx,va=%jx)", (uintptr_t)map, va, 0, 0); const vaddr_t start = trunc_page(va); Index: src/sys/uvm/uvm_page.c diff -u src/sys/uvm/uvm_page.c:1.243 src/sys/uvm/uvm_page.c:1.244 --- src/sys/uvm/uvm_page.c:1.243 Wed Jun 17 06:24:15 2020 +++ src/sys/uvm/uvm_page.c Thu Jul 9 05:57:15 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: uvm_page.c,v 1.243 2020/06/17 06:24:15 thorpej Exp $ */ +/* $NetBSD: uvm_page.c,v 1.244 2020/07/09 05:57:15 skrll Exp $ */ /*- * Copyright (c) 2019, 2020 The NetBSD Foundation, Inc. @@ -95,7 +95,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: uvm_page.c,v 1.243 2020/06/17 06:24:15 thorpej Exp $"); +__KERNEL_RCSID(0, "$NetBSD: uvm_page.c,v 1.244 2020/07/09 05:57:15 skrll Exp $"); #include "opt_ddb.h" #include "opt_uvm.h" @@ -1645,7 +1645,7 @@ uvm_page_unbusy(struct vm_page **pgs, in { struct vm_page *pg; int i; - UVMHIST_FUNC("uvm_page_unbusy"); UVMHIST_CALLED(ubchist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(ubchist); for (i = 0; i < npgs; i++) { pg = pgs[i]; @@ -1707,7 +1707,7 @@ uvm_pagewait(struct vm_page *pg, krwlock void uvm_pagewakeup(struct vm_page *pg) { - UVMHIST_FUNC("uvm_pagewakeup"); UVMHIST_CALLED(ubchist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(ubchist); KASSERT(mutex_owned(&pg->interlock)); Index: src/sys/uvm/uvm_pager.c diff -u src/sys/uvm/uvm_pager.c:1.127 src/sys/uvm/uvm_pager.c:1.128 --- src/sys/uvm/uvm_pager.c:1.127 Wed Jul 8 13:26:22 2020 +++ src/sys/uvm/uvm_pager.c Thu Jul 9 05:57:15 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: uvm_pager.c,v 1.127 2020/07/08 13:26:22 skrll Exp $ */ +/* $NetBSD: uvm_pager.c,v 1.128 2020/07/09 05:57:15 skrll Exp $ */ /* * Copyright (c) 1997 Charles D. Cranor and Washington University. @@ -32,7 +32,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: uvm_pager.c,v 1.127 2020/07/08 13:26:22 skrll Exp $"); +__KERNEL_RCSID(0, "$NetBSD: uvm_pager.c,v 1.128 2020/07/09 05:57:15 skrll Exp $"); #include "opt_uvmhist.h" #include "opt_readahead.h" @@ -189,9 +189,8 @@ uvm_pagermapin(struct vm_page **pps, int vm_prot_t prot; const bool pdaemon = (curlwp == uvm.pagedaemon_lwp); const u_int first_color = VM_PGCOLOR(*pps); - UVMHIST_FUNC("uvm_pagermapin"); UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist,"(pps=%#jx, npages=%jd, first_color=%ju)", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist,"(pps=%#jx, npages=%jd, first_color=%ju)", (uintptr_t)pps, npages, first_color, 0); #ifdef PMAP_DIRECT @@ -280,9 +279,8 @@ uvm_pagermapout(vaddr_t kva, int npages) { vsize_t size = ptoa(npages); struct vm_map_entry *entries; - UVMHIST_FUNC("uvm_pagermapout"); UVMHIST_CALLED(maphist); - - UVMHIST_LOG(maphist, " (kva=%#jx, npages=%jd)", kva, npages,0,0); + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(maphist, " (kva=%#jx, npages=%jd)", kva, npages,0,0); #ifdef PMAP_DIRECT /* @@ -335,7 +333,7 @@ uvm_aio_aiodone_pages(struct vm_page **p int swslot; int i; bool swap; - UVMHIST_FUNC("uvm_aio_aiodone_pages"); UVMHIST_CALLED(ubchist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(ubchist); swslot = 0; pageout_done = 0; @@ -523,8 +521,8 @@ uvm_aio_aiodone(struct buf *bp) struct vm_page *pgs[howmany(MAXPHYS, MIN_PAGE_SIZE)]; int i, error; bool write; - UVMHIST_FUNC("uvm_aio_aiodone"); UVMHIST_CALLED(ubchist); - UVMHIST_LOG(ubchist, "bp %#jx", (uintptr_t)bp, 0,0,0); + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(ubchist, "bp %#jx", (uintptr_t)bp, 0,0,0); KASSERT(bp->b_bufsize <= MAXPHYS); KASSERT(npages <= __arraycount(pgs)); Index: src/sys/uvm/uvm_pdaemon.c diff -u src/sys/uvm/uvm_pdaemon.c:1.129 src/sys/uvm/uvm_pdaemon.c:1.130 --- src/sys/uvm/uvm_pdaemon.c:1.129 Thu Jun 11 22:21:05 2020 +++ src/sys/uvm/uvm_pdaemon.c Thu Jul 9 05:57:15 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: uvm_pdaemon.c,v 1.129 2020/06/11 22:21:05 ad Exp $ */ +/* $NetBSD: uvm_pdaemon.c,v 1.130 2020/07/09 05:57:15 skrll Exp $ */ /* * Copyright (c) 1997 Charles D. Cranor and Washington University. @@ -66,7 +66,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: uvm_pdaemon.c,v 1.129 2020/06/11 22:21:05 ad Exp $"); +__KERNEL_RCSID(0, "$NetBSD: uvm_pdaemon.c,v 1.130 2020/07/09 05:57:15 skrll Exp $"); #include "opt_uvmhist.h" #include "opt_readahead.h" @@ -207,7 +207,7 @@ uvmpd_tune(void) { int val; - UVMHIST_FUNC("uvmpd_tune"); UVMHIST_CALLED(pdhist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(pdhist); /* * try to keep 0.5% of available RAM free, but limit to between @@ -244,8 +244,8 @@ uvm_pageout(void *arg) int npages = 0; int extrapages = 0; int fpages; - - UVMHIST_FUNC("uvm_pageout"); UVMHIST_CALLED(pdhist); + + UVMHIST_FUNC(__func__); UVMHIST_CALLED(pdhist); UVMHIST_LOG(pdhist,"<starting uvm pagedaemon>", 0, 0, 0, 0); @@ -634,7 +634,7 @@ uvmpd_scan_queue(void) #endif /* defined(VMSWAP) */ int dirtyreacts; krwlock_t *slock; - UVMHIST_FUNC("uvmpd_scan_queue"); UVMHIST_CALLED(pdhist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(pdhist); /* * swslot is non-zero if we are building a swap cluster. we want @@ -883,7 +883,7 @@ static void uvmpd_scan(void) { int swap_shortage, pages_freed, fpages; - UVMHIST_FUNC("uvmpd_scan"); UVMHIST_CALLED(pdhist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(pdhist); uvmexp.pdrevs++; @@ -1043,7 +1043,7 @@ uvmpd_pool_drain_thread(void *arg) mutex_exit(&bufcache_lock); /* - * drain a pool, and then re-enable the freelist cache. + * drain a pool, and then re-enable the freelist cache. */ (void)pool_drain(&curpool); KASSERT(curpool != NULL); Index: src/sys/uvm/uvm_swap.c diff -u src/sys/uvm/uvm_swap.c:1.196 src/sys/uvm/uvm_swap.c:1.197 --- src/sys/uvm/uvm_swap.c:1.196 Wed Jul 8 13:26:22 2020 +++ src/sys/uvm/uvm_swap.c Thu Jul 9 05:57:15 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: uvm_swap.c,v 1.196 2020/07/08 13:26:22 skrll Exp $ */ +/* $NetBSD: uvm_swap.c,v 1.197 2020/07/09 05:57:15 skrll Exp $ */ /* * Copyright (c) 1995, 1996, 1997, 2009 Matthew R. Green @@ -30,7 +30,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: uvm_swap.c,v 1.196 2020/07/08 13:26:22 skrll Exp $"); +__KERNEL_RCSID(0, "$NetBSD: uvm_swap.c,v 1.197 2020/07/09 05:57:15 skrll Exp $"); #include "opt_uvmhist.h" #include "opt_compat_netbsd.h" @@ -256,7 +256,7 @@ encmap_size(size_t npages) void uvm_swap_init(void) { - UVMHIST_FUNC("uvm_swap_init"); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(pdhist); /* @@ -315,7 +315,7 @@ static void swaplist_insert(struct swapdev *sdp, struct swappri *newspp, int priority) { struct swappri *spp, *pspp; - UVMHIST_FUNC("swaplist_insert"); UVMHIST_CALLED(pdhist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(pdhist); KASSERT(rw_write_held(&swap_syscall_lock)); KASSERT(mutex_owned(&uvm_swap_data_lock)); @@ -510,7 +510,7 @@ sys_swapctl(struct lwp *l, const struct size_t len = 0; int error; int priority; - UVMHIST_FUNC("sys_swapctl"); UVMHIST_CALLED(pdhist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(pdhist); /* * we handle the non-priv NSWAP and STATS request first. @@ -834,7 +834,7 @@ swap_on(struct lwp *l, struct swapdev *s vmem_addr_t result; struct vattr va; dev_t dev; - UVMHIST_FUNC("swap_on"); UVMHIST_CALLED(pdhist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(pdhist); /* * we want to enable swapping on sdp. the swd_vp contains @@ -1050,8 +1050,8 @@ swap_off(struct lwp *l, struct swapdev * int npages = sdp->swd_npages; int error = 0; - UVMHIST_FUNC("swap_off"); UVMHIST_CALLED(pdhist); - UVMHIST_LOG(pdhist, " dev=%jx, npages=%jd", sdp->swd_dev,npages, 0, 0); + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(pdhist, " dev=%jx, npages=%jd", sdp->swd_dev,npages, 0, 0); KASSERT(rw_write_held(&swap_syscall_lock)); KASSERT(mutex_owned(&uvm_swap_data_lock)); @@ -1196,7 +1196,7 @@ swstrategy(struct buf *bp) struct swapdev *sdp; struct vnode *vp; int pageno, bn; - UVMHIST_FUNC("swstrategy"); UVMHIST_CALLED(pdhist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(pdhist); /* * convert block number to swapdev. note that swapdev can't @@ -1284,9 +1284,9 @@ swstrategy(struct buf *bp) static int swread(dev_t dev, struct uio *uio, int ioflag) { - UVMHIST_FUNC("swread"); UVMHIST_CALLED(pdhist); + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(pdhist, " dev=%jx offset=%jx", dev, uio->uio_offset, 0, 0); - UVMHIST_LOG(pdhist, " dev=%jx offset=%jx", dev, uio->uio_offset, 0, 0); return (physio(swstrategy, NULL, dev, B_READ, minphys, uio)); } @@ -1297,9 +1297,9 @@ swread(dev_t dev, struct uio *uio, int i static int swwrite(dev_t dev, struct uio *uio, int ioflag) { - UVMHIST_FUNC("swwrite"); UVMHIST_CALLED(pdhist); + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(pdhist, " dev=%jx offset=%jx", dev, uio->uio_offset, 0, 0); - UVMHIST_LOG(pdhist, " dev=%jx offset=%jx", dev, uio->uio_offset, 0, 0); return (physio(swstrategy, NULL, dev, B_WRITE, minphys, uio)); } @@ -1341,7 +1341,7 @@ sw_reg_strategy(struct swapdev *sdp, str char *addr; off_t byteoff; int s, off, nra, error, sz, resid; - UVMHIST_FUNC("sw_reg_strategy"); UVMHIST_CALLED(pdhist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(pdhist); /* * allocate a vndxfer head for this transfer and point it to @@ -1490,7 +1490,7 @@ sw_reg_start(struct swapdev *sdp) { struct buf *bp; struct vnode *vp; - UVMHIST_FUNC("sw_reg_start"); UVMHIST_CALLED(pdhist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(pdhist); /* recursion control */ if ((sdp->swd_flags & SWF_BUSY) != 0) @@ -1543,9 +1543,8 @@ sw_reg_iodone(struct work *wk, void *dum struct swapdev *sdp = vnx->vx_sdp; int s, resid, error; KASSERT(&vbp->vb_buf.b_work == wk); - UVMHIST_FUNC("sw_reg_iodone"); UVMHIST_CALLED(pdhist); - - UVMHIST_LOG(pdhist, " vbp=%#jx vp=%#jx blkno=%jx addr=%#jx", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(pdhist, " vbp=%#jx vp=%#jx blkno=%jx addr=%#jx", (uintptr_t)vbp, (uintptr_t)vbp->vb_buf.b_vp, vbp->vb_buf.b_blkno, (uintptr_t)vbp->vb_buf.b_data); UVMHIST_LOG(pdhist, " cnt=%jx resid=%jx", @@ -1619,7 +1618,7 @@ uvm_swap_alloc(int *nslots /* IN/OUT */, { struct swapdev *sdp; struct swappri *spp; - UVMHIST_FUNC("uvm_swap_alloc"); UVMHIST_CALLED(pdhist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(pdhist); /* * no swap devices configured yet? definite failure. @@ -1720,7 +1719,7 @@ void uvm_swap_markbad(int startslot, int nslots) { struct swapdev *sdp; - UVMHIST_FUNC("uvm_swap_markbad"); UVMHIST_CALLED(pdhist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(pdhist); mutex_enter(&uvm_swap_data_lock); sdp = swapdrum_getsdp(startslot); @@ -1750,9 +1749,8 @@ void uvm_swap_free(int startslot, int nslots) { struct swapdev *sdp; - UVMHIST_FUNC("uvm_swap_free"); UVMHIST_CALLED(pdhist); - - UVMHIST_LOG(pdhist, "freeing %jd slots starting at %jd", nslots, + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(pdhist, "freeing %jd slots starting at %jd", nslots, startslot, 0, 0); /* @@ -1839,9 +1837,8 @@ uvm_swap_io(struct vm_page **pps, int st vaddr_t kva; int error, mapinflags; bool write, async, swap_encrypt; - UVMHIST_FUNC("uvm_swap_io"); UVMHIST_CALLED(pdhist); - - UVMHIST_LOG(pdhist, "<- called, startslot=%jd, npages=%jd, flags=%jd", + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(pdhist, "<- called, startslot=%jd, npages=%jd, flags=%jd", startslot, npages, flags, 0); write = (flags & B_READ) == 0; Index: src/sys/uvm/uvm_vnode.c diff -u src/sys/uvm/uvm_vnode.c:1.114 src/sys/uvm/uvm_vnode.c:1.115 --- src/sys/uvm/uvm_vnode.c:1.114 Mon May 25 21:15:10 2020 +++ src/sys/uvm/uvm_vnode.c Thu Jul 9 05:57:15 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: uvm_vnode.c,v 1.114 2020/05/25 21:15:10 ad Exp $ */ +/* $NetBSD: uvm_vnode.c,v 1.115 2020/07/09 05:57:15 skrll Exp $ */ /* * Copyright (c) 1997 Charles D. Cranor and Washington University. @@ -45,7 +45,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: uvm_vnode.c,v 1.114 2020/05/25 21:15:10 ad Exp $"); +__KERNEL_RCSID(0, "$NetBSD: uvm_vnode.c,v 1.115 2020/07/09 05:57:15 skrll Exp $"); #ifdef _KERNEL_OPT #include "opt_uvmhist.h" @@ -175,9 +175,8 @@ uvn_get(struct uvm_object *uobj, voff_t struct vnode *vp = (struct vnode *)uobj; int error; - UVMHIST_FUNC("uvn_get"); UVMHIST_CALLED(ubchist); - - UVMHIST_LOG(ubchist, "vp %#jx off 0x%jx", (uintptr_t)vp, (int)offset, + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(ubchist, "vp %#jx off 0x%jx", (uintptr_t)vp, offset, 0, 0); if (vp->v_type == VREG && (access_type & VM_PROT_WRITE) == 0 @@ -290,8 +289,8 @@ uvn_findpage(struct uvm_object *uobj, vo unsigned int flags, struct uvm_page_array *a, unsigned int nleft) { struct vm_page *pg; - UVMHIST_FUNC("uvn_findpage"); UVMHIST_CALLED(ubchist); - UVMHIST_LOG(ubchist, "vp %#jx off 0x%jx", (uintptr_t)uobj, offset, + UVMHIST_FUNC(__func__); + UVMHIST_CALLARGS(ubchist, "vp %#jx off 0x%jx", (uintptr_t)uobj, offset, 0, 0); /* @@ -439,7 +438,7 @@ uvm_vnp_setsize(struct vnode *vp, voff_t struct uvm_object *uobj = &vp->v_uobj; voff_t pgend = round_page(newsize); voff_t oldsize; - UVMHIST_FUNC("uvm_vnp_setsize"); UVMHIST_CALLED(ubchist); + UVMHIST_FUNC(__func__); UVMHIST_CALLED(ubchist); rw_enter(uobj->vmobjlock, RW_WRITER); UVMHIST_LOG(ubchist, "vp %#jx old 0x%jx new 0x%jx",