Module Name:    src
Committed By:   skrll
Date:           Sat Mar 13 15:29:56 UTC 2021

Modified Files:
        src/sys/uvm: uvm_amap.c uvm_aobj.c uvm_bio.c uvm_device.c uvm_km.c
            uvm_map.c uvm_swap.c uvm_vnode.c
        src/sys/uvm/pmap: pmap_segtab.c

Log Message:
Consistently use %#jx instead of 0x%jx or just %jx in UVMHIST_LOG formats


To generate a diff of this commit:
cvs rdiff -u -r1.125 -r1.126 src/sys/uvm/uvm_amap.c
cvs rdiff -u -r1.152 -r1.153 src/sys/uvm/uvm_aobj.c
cvs rdiff -u -r1.124 -r1.125 src/sys/uvm/uvm_bio.c
cvs rdiff -u -r1.71 -r1.72 src/sys/uvm/uvm_device.c
cvs rdiff -u -r1.159 -r1.160 src/sys/uvm/uvm_km.c
cvs rdiff -u -r1.385 -r1.386 src/sys/uvm/uvm_map.c
cvs rdiff -u -r1.202 -r1.203 src/sys/uvm/uvm_swap.c
cvs rdiff -u -r1.117 -r1.118 src/sys/uvm/uvm_vnode.c
cvs rdiff -u -r1.26 -r1.27 src/sys/uvm/pmap/pmap_segtab.c

Please note that diffs are not public domain; they are subject to the
copyright notices on the relevant files.

Modified files:

Index: src/sys/uvm/uvm_amap.c
diff -u src/sys/uvm/uvm_amap.c:1.125 src/sys/uvm/uvm_amap.c:1.126
--- src/sys/uvm/uvm_amap.c:1.125	Mon Sep 21 18:41:59 2020
+++ src/sys/uvm/uvm_amap.c	Sat Mar 13 15:29:55 2021
@@ -1,4 +1,4 @@
-/*	$NetBSD: uvm_amap.c,v 1.125 2020/09/21 18:41:59 chs Exp $	*/
+/*	$NetBSD: uvm_amap.c,v 1.126 2021/03/13 15:29:55 skrll Exp $	*/
 
 /*
  * Copyright (c) 1997 Charles D. Cranor and Washington University.
@@ -35,7 +35,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: uvm_amap.c,v 1.125 2020/09/21 18:41:59 chs Exp $");
+__KERNEL_RCSID(0, "$NetBSD: uvm_amap.c,v 1.126 2021/03/13 15:29:55 skrll Exp $");
 
 #include "opt_uvmhist.h"
 
@@ -827,7 +827,7 @@ amap_copy(struct vm_map *map, struct vm_
 	vsize_t len;
 
 	UVMHIST_FUNC(__func__);
-	UVMHIST_CALLARGS(maphist, "  (map=%#j, entry=%#j, flags=%jd)",
+	UVMHIST_CALLARGS(maphist, "  (map=%#jx, entry=%#jx, flags=%#jx)",
 	    (uintptr_t)map, (uintptr_t)entry, flags, -2);
 
 	KASSERT(map != kernel_map);	/* we use nointr pool */
@@ -903,7 +903,7 @@ amap_copy(struct vm_map *map, struct vm_
 		return;
 	}
 
-	UVMHIST_LOG(maphist,"  amap=%#j, ref=%jd, must copy it",
+	UVMHIST_LOG(maphist,"  amap=%#jx, ref=%jd, must copy it",
 	    (uintptr_t)srcamap, srcamap->am_ref, 0, 0);
 
 	/*

Index: src/sys/uvm/uvm_aobj.c
diff -u src/sys/uvm/uvm_aobj.c:1.152 src/sys/uvm/uvm_aobj.c:1.153
--- src/sys/uvm/uvm_aobj.c:1.152	Wed Nov  4 01:30:19 2020
+++ src/sys/uvm/uvm_aobj.c	Sat Mar 13 15:29:55 2021
@@ -1,4 +1,4 @@
-/*	$NetBSD: uvm_aobj.c,v 1.152 2020/11/04 01:30:19 chs Exp $	*/
+/*	$NetBSD: uvm_aobj.c,v 1.153 2021/03/13 15:29:55 skrll Exp $	*/
 
 /*
  * Copyright (c) 1998 Chuck Silvers, Charles D. Cranor and
@@ -38,7 +38,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: uvm_aobj.c,v 1.152 2020/11/04 01:30:19 chs Exp $");
+__KERNEL_RCSID(0, "$NetBSD: uvm_aobj.c,v 1.153 2021/03/13 15:29:55 skrll Exp $");
 
 #ifdef _KERNEL_OPT
 #include "opt_uvmhist.h"
@@ -810,7 +810,7 @@ uao_get(struct uvm_object *uobj, voff_t 
 	struct uvm_page_array a;
 
 	UVMHIST_FUNC(__func__);
-	UVMHIST_CALLARGS(pdhist, "aobj=%#jx offset=%jd, flags=%jd",
+	UVMHIST_CALLARGS(pdhist, "aobj=%#jx offset=%jd, flags=%#jx",
 		    (uintptr_t)uobj, offset, flags,0);
 
 	/*

Index: src/sys/uvm/uvm_bio.c
diff -u src/sys/uvm/uvm_bio.c:1.124 src/sys/uvm/uvm_bio.c:1.125
--- src/sys/uvm/uvm_bio.c:1.124	Tue Nov 10 04:27:22 2020
+++ src/sys/uvm/uvm_bio.c	Sat Mar 13 15:29:55 2021
@@ -1,4 +1,4 @@
-/*	$NetBSD: uvm_bio.c,v 1.124 2020/11/10 04:27:22 chs Exp $	*/
+/*	$NetBSD: uvm_bio.c,v 1.125 2021/03/13 15:29:55 skrll Exp $	*/
 
 /*
  * Copyright (c) 1998 Chuck Silvers.
@@ -34,7 +34,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: uvm_bio.c,v 1.124 2020/11/10 04:27:22 chs Exp $");
+__KERNEL_RCSID(0, "$NetBSD: uvm_bio.c,v 1.125 2021/03/13 15:29:55 skrll Exp $");
 
 #include "opt_uvmhist.h"
 #include "opt_ubc.h"
@@ -342,7 +342,7 @@ ubc_fault(struct uvm_faultinfo *ufi, vad
 	 */
 
 	access_type = umap->writelen ? VM_PROT_WRITE : VM_PROT_READ;
-	UVMHIST_LOG(ubchist, "va 0x%jx ubc_offset 0x%jx access_type %jd",
+	UVMHIST_LOG(ubchist, "va %#jx ubc_offset %#jx access_type %jd",
 	    va, ubc_offset, access_type, 0);
 
 	if ((access_type & VM_PROT_WRITE) != 0) {
@@ -374,9 +374,9 @@ again:
 	memset(pgs, 0, sizeof (pgs));
 	rw_enter(uobj->vmobjlock, RW_WRITER);
 
-	UVMHIST_LOG(ubchist, "slot_offset 0x%jx writeoff 0x%jx writelen 0x%jx ",
+	UVMHIST_LOG(ubchist, "slot_offset %#jx writeoff %#jx writelen %#jx ",
 	    slot_offset, umap->writeoff, umap->writelen, 0);
-	UVMHIST_LOG(ubchist, "getpages uobj %#jx offset 0x%jx npages %jd",
+	UVMHIST_LOG(ubchist, "getpages uobj %#jx offset %#jx npages %jd",
 	    (uintptr_t)uobj, umap->offset + slot_offset, npages, 0);
 
 	error = (*uobj->pgops->pgo_get)(uobj, umap->offset + slot_offset, pgs,
@@ -409,7 +409,7 @@ again:
 	va = ufi->orig_rvaddr;
 	eva = ufi->orig_rvaddr + (npages << PAGE_SHIFT);
 
-	UVMHIST_LOG(ubchist, "va 0x%jx eva 0x%jx", va, eva, 0, 0);
+	UVMHIST_LOG(ubchist, "va %#jx eva %#jx", va, eva, 0, 0);
 
 	/*
 	 * Note: normally all returned pages would have the same UVM object.
@@ -483,7 +483,7 @@ ubc_alloc(struct uvm_object *uobj, voff_
 	voff_t umap_offset;
 	int error;
 	UVMHIST_FUNC(__func__);
-	UVMHIST_CALLARGS(ubchist, "uobj %#jx offset 0x%jx len 0x%jx",
+	UVMHIST_CALLARGS(ubchist, "uobj %#jx offset %#jx len %#jx",
 	    (uintptr_t)uobj, offset, *lenp, 0);
 
 	KASSERT(*lenp > 0);
@@ -560,7 +560,7 @@ again:
 	umap->refcount++;
 	umap->advice = advice;
 	rw_exit(ubc_object.uobj.vmobjlock);
-	UVMHIST_LOG(ubchist, "umap %#jx refs %jd va %#jx flags 0x%jx",
+	UVMHIST_LOG(ubchist, "umap %#jx refs %jd va %#jx flags %#jx",
 	    (uintptr_t)umap, umap->refcount, (uintptr_t)va, flags);
 
 	if (flags & UBC_FAULTBUSY) {

Index: src/sys/uvm/uvm_device.c
diff -u src/sys/uvm/uvm_device.c:1.71 src/sys/uvm/uvm_device.c:1.72
--- src/sys/uvm/uvm_device.c:1.71	Thu Jul  9 05:57:15 2020
+++ src/sys/uvm/uvm_device.c	Sat Mar 13 15:29:55 2021
@@ -1,4 +1,4 @@
-/*	$NetBSD: uvm_device.c,v 1.71 2020/07/09 05:57:15 skrll Exp $	*/
+/*	$NetBSD: uvm_device.c,v 1.72 2021/03/13 15:29:55 skrll Exp $	*/
 
 /*
  * Copyright (c) 1997 Charles D. Cranor and Washington University.
@@ -32,7 +32,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: uvm_device.c,v 1.71 2020/07/09 05:57:15 skrll Exp $");
+__KERNEL_RCSID(0, "$NetBSD: uvm_device.c,v 1.72 2021/03/13 15:29:55 skrll Exp $");
 
 #include "opt_uvmhist.h"
 
@@ -364,7 +364,7 @@ udv_fault(struct uvm_faultinfo *ufi, vad
 	dev_t device;
 	vm_prot_t mapprot;
 	UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist);
-	UVMHIST_LOG(maphist,"  flags=%jd", flags,0,0,0);
+	UVMHIST_LOG(maphist,"  flags=%#jx", flags,0,0,0);
 
 	/*
 	 * we do not allow device mappings to be mapped copy-on-write

Index: src/sys/uvm/uvm_km.c
diff -u src/sys/uvm/uvm_km.c:1.159 src/sys/uvm/uvm_km.c:1.160
--- src/sys/uvm/uvm_km.c:1.159	Thu Jul  9 05:57:15 2020
+++ src/sys/uvm/uvm_km.c	Sat Mar 13 15:29:55 2021
@@ -1,4 +1,4 @@
-/*	$NetBSD: uvm_km.c,v 1.159 2020/07/09 05:57:15 skrll Exp $	*/
+/*	$NetBSD: uvm_km.c,v 1.160 2021/03/13 15:29:55 skrll Exp $	*/
 
 /*
  * Copyright (c) 1997 Charles D. Cranor and Washington University.
@@ -152,7 +152,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: uvm_km.c,v 1.159 2020/07/09 05:57:15 skrll Exp $");
+__KERNEL_RCSID(0, "$NetBSD: uvm_km.c,v 1.160 2021/03/13 15:29:55 skrll Exp $");
 
 #include "opt_uvmhist.h"
 
@@ -618,7 +618,7 @@ uvm_km_alloc(struct vm_map *map, vsize_t
 	kva = vm_map_min(map);	/* hint */
 	size = round_page(size);
 	obj = (flags & UVM_KMF_PAGEABLE) ? uvm_kernel_object : NULL;
-	UVMHIST_LOG(maphist,"  (map=%#jx, obj=%#jx, size=%#jx, flags=%jd)",
+	UVMHIST_LOG(maphist,"  (map=%#jx, obj=%#jx, size=%#jx, flags=%#jx)",
 	    (uintptr_t)map, (uintptr_t)obj, size, flags);
 
 	/*

Index: src/sys/uvm/uvm_map.c
diff -u src/sys/uvm/uvm_map.c:1.385 src/sys/uvm/uvm_map.c:1.386
--- src/sys/uvm/uvm_map.c:1.385	Thu Jul  9 05:57:15 2020
+++ src/sys/uvm/uvm_map.c	Sat Mar 13 15:29:55 2021
@@ -1,4 +1,4 @@
-/*	$NetBSD: uvm_map.c,v 1.385 2020/07/09 05:57:15 skrll Exp $	*/
+/*	$NetBSD: uvm_map.c,v 1.386 2021/03/13 15:29:55 skrll Exp $	*/
 
 /*
  * Copyright (c) 1997 Charles D. Cranor and Washington University.
@@ -66,7 +66,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: uvm_map.c,v 1.385 2020/07/09 05:57:15 skrll Exp $");
+__KERNEL_RCSID(0, "$NetBSD: uvm_map.c,v 1.386 2021/03/13 15:29:55 skrll Exp $");
 
 #include "opt_ddb.h"
 #include "opt_pax.h"
@@ -806,7 +806,7 @@ static void
 uvm_mapent_free(struct vm_map_entry *me)
 {
 	UVMHIST_FUNC(__func__);
-	UVMHIST_CALLARGS(maphist,"<- freeing map entry=%#jx [flags=%jd]",
+	UVMHIST_CALLARGS(maphist,"<- freeing map entry=%#jx [flags=%#jx]",
 		(uintptr_t)me, me->flags, 0, 0);
 	pool_cache_put(&uvm_map_entry_cache, me);
 }
@@ -2115,9 +2115,9 @@ nextgap:
 	SAVE_HINT(map, map->hint, entry);
 	*result = hint;
 	UVMHIST_LOG(maphist,"<- got it!  (result=%#jx)", hint, 0,0,0);
-	KASSERTMSG( topdown || hint >= orig_hint, "hint: %jx, orig_hint: %jx",
+	KASSERTMSG( topdown || hint >= orig_hint, "hint: %#jx, orig_hint: %#jx",
 	    (uintmax_t)hint, (uintmax_t)orig_hint);
-	KASSERTMSG(!topdown || hint <= orig_hint, "hint: %jx, orig_hint: %jx",
+	KASSERTMSG(!topdown || hint <= orig_hint, "hint: %#jx, orig_hint: %#jx",
 	    (uintmax_t)hint, (uintmax_t)orig_hint);
 	KASSERT(entry->end <= hint);
 	KASSERT(hint + length <= entry->next->start);
@@ -4811,7 +4811,7 @@ uvm_voaddr_acquire(struct vm_map * const
 	void (*unlock_fn)(struct vm_map *);
 
 	UVMHIST_FUNC(__func__); UVMHIST_CALLED(maphist);
-	UVMHIST_LOG(maphist,"(map=%#jx,va=%jx)", (uintptr_t)map, va, 0, 0);
+	UVMHIST_LOG(maphist,"(map=%#jx,va=%#jx)", (uintptr_t)map, va, 0, 0);
 
 	const vaddr_t start = trunc_page(va);
 	const vaddr_t end = round_page(va+1);
@@ -4969,7 +4969,7 @@ uvm_voaddr_acquire(struct vm_map * const
 
 	if (result) {
 		UVMHIST_LOG(maphist,
-		    "<- done OK (type=%jd,owner=#%jx,offset=%jx)",
+		    "<- done OK (type=%jd,owner=%#jx,offset=%#jx)",
 		    UVM_VOADDR_GET_TYPE(voaddr),
 		    UVM_VOADDR_GET_OBJECT(voaddr),
 		    voaddr->offset, 0);

Index: src/sys/uvm/uvm_swap.c
diff -u src/sys/uvm/uvm_swap.c:1.202 src/sys/uvm/uvm_swap.c:1.203
--- src/sys/uvm/uvm_swap.c:1.202	Fri Feb 19 13:20:43 2021
+++ src/sys/uvm/uvm_swap.c	Sat Mar 13 15:29:55 2021
@@ -1,4 +1,4 @@
-/*	$NetBSD: uvm_swap.c,v 1.202 2021/02/19 13:20:43 hannken Exp $	*/
+/*	$NetBSD: uvm_swap.c,v 1.203 2021/03/13 15:29:55 skrll Exp $	*/
 
 /*
  * Copyright (c) 1995, 1996, 1997, 2009 Matthew R. Green
@@ -30,7 +30,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: uvm_swap.c,v 1.202 2021/02/19 13:20:43 hannken Exp $");
+__KERNEL_RCSID(0, "$NetBSD: uvm_swap.c,v 1.203 2021/03/13 15:29:55 skrll Exp $");
 
 #include "opt_uvmhist.h"
 #include "opt_compat_netbsd.h"
@@ -933,7 +933,7 @@ swap_on(struct lwp *l, struct swapdev *s
 		goto bad;
 	}
 
-	UVMHIST_LOG(pdhist, "  dev=%jx: size=%jd addr=%jd", dev, size, addr, 0);
+	UVMHIST_LOG(pdhist, "  dev=%#jx: size=%jd addr=%jd", dev, size, addr, 0);
 
 	/*
 	 * now we need to allocate an extent to manage this swap device
@@ -1052,7 +1052,7 @@ swap_off(struct lwp *l, struct swapdev *
 	int error = 0;
 
 	UVMHIST_FUNC(__func__);
-	UVMHIST_CALLARGS(pdhist, "  dev=%jx, npages=%jd", sdp->swd_dev,npages, 0, 0);
+	UVMHIST_CALLARGS(pdhist, "  dev=%#jx, npages=%jd", sdp->swd_dev,npages, 0, 0);
 
 	KASSERT(rw_write_held(&swap_syscall_lock));
 	KASSERT(mutex_owned(&uvm_swap_data_lock));
@@ -1216,7 +1216,7 @@ swstrategy(struct buf *bp)
 	pageno -= sdp->swd_drumoffset;	/* page # on swapdev */
 	bn = btodb((uint64_t)pageno << PAGE_SHIFT); /* convert to diskblock */
 
-	UVMHIST_LOG(pdhist, "  Rd/Wr (0/1) %jd: mapoff=%jx bn=%jx bcount=%jd",
+	UVMHIST_LOG(pdhist, "  Rd/Wr (0/1) %jd: mapoff=%#jx bn=%#jx bcount=%jd",
 		((bp->b_flags & B_READ) == 0) ? 1 : 0,
 		sdp->swd_drumoffset, bn, bp->b_bcount);
 
@@ -1279,7 +1279,7 @@ static int
 swread(dev_t dev, struct uio *uio, int ioflag)
 {
 	UVMHIST_FUNC(__func__);
-	UVMHIST_CALLARGS(pdhist, "  dev=%jx offset=%jx", dev, uio->uio_offset, 0, 0);
+	UVMHIST_CALLARGS(pdhist, "  dev=%#jx offset=%#jx", dev, uio->uio_offset, 0, 0);
 
 	return (physio(swstrategy, NULL, dev, B_READ, minphys, uio));
 }
@@ -1292,7 +1292,7 @@ static int
 swwrite(dev_t dev, struct uio *uio, int ioflag)
 {
 	UVMHIST_FUNC(__func__);
-	UVMHIST_CALLARGS(pdhist, "  dev=%jx offset=%jx", dev, uio->uio_offset, 0, 0);
+	UVMHIST_CALLARGS(pdhist, "  dev=%#jx offset=%#jx", dev, uio->uio_offset, 0, 0);
 
 	return (physio(swstrategy, NULL, dev, B_WRITE, minphys, uio));
 }
@@ -1408,7 +1408,7 @@ sw_reg_strategy(struct swapdev *sdp, str
 			sz = resid;
 
 		UVMHIST_LOG(pdhist, "sw_reg_strategy: "
-		    "vp %#jx/%#jx offset 0x%jx/0x%jx",
+		    "vp %#jx/%#jx offset %#jx/%#jx",
 		    (uintptr_t)sdp->swd_vp, (uintptr_t)vp, byteoff, nbn);
 
 		/*
@@ -1499,7 +1499,7 @@ sw_reg_start(struct swapdev *sdp)
 		sdp->swd_active++;
 
 		UVMHIST_LOG(pdhist,
-		    "sw_reg_start:  bp %#jx vp %#jx blkno %#jx cnt %jx",
+		    "sw_reg_start:  bp %#jx vp %#jx blkno %#jx cnt %#jx",
 		    (uintptr_t)bp, (uintptr_t)bp->b_vp, (uintptr_t)bp->b_blkno,
 		    bp->b_bcount);
 		vp = bp->b_vp;
@@ -1538,10 +1538,10 @@ sw_reg_iodone(struct work *wk, void *dum
 	int s, resid, error;
 	KASSERT(&vbp->vb_buf.b_work == wk);
 	UVMHIST_FUNC(__func__);
-	UVMHIST_CALLARGS(pdhist, "  vbp=%#jx vp=%#jx blkno=%jx addr=%#jx",
+	UVMHIST_CALLARGS(pdhist, "  vbp=%#jx vp=%#jx blkno=%#jx addr=%#jx",
 	    (uintptr_t)vbp, (uintptr_t)vbp->vb_buf.b_vp, vbp->vb_buf.b_blkno,
 	    (uintptr_t)vbp->vb_buf.b_data);
-	UVMHIST_LOG(pdhist, "  cnt=%jx resid=%jx",
+	UVMHIST_LOG(pdhist, "  cnt=%#jx resid=%#jx",
 	    vbp->vb_buf.b_bcount, vbp->vb_buf.b_resid, 0, 0);
 
 	/*
@@ -1836,7 +1836,7 @@ uvm_swap_io(struct vm_page **pps, int st
 	int	error, mapinflags;
 	bool write, async, swap_encrypt;
 	UVMHIST_FUNC(__func__);
-	UVMHIST_CALLARGS(pdhist, "<- called, startslot=%jd, npages=%jd, flags=%jd",
+	UVMHIST_CALLARGS(pdhist, "<- called, startslot=%jd, npages=%jd, flags=%#jx",
 	    startslot, npages, flags, 0);
 
 	write = (flags & B_READ) == 0;
@@ -1967,7 +1967,7 @@ uvm_swap_io(struct vm_page **pps, int st
 		BIO_SETPRIO(bp, BPRIO_TIMECRITICAL);
 	}
 	UVMHIST_LOG(pdhist,
-	    "about to start io: data = %#jx blkno = 0x%jx, bcount = %jd",
+	    "about to start io: data = %#jx blkno = %#jx, bcount = %jd",
 	    (uintptr_t)bp->b_data, bp->b_blkno, bp->b_bcount, 0);
 
 	/*

Index: src/sys/uvm/uvm_vnode.c
diff -u src/sys/uvm/uvm_vnode.c:1.117 src/sys/uvm/uvm_vnode.c:1.118
--- src/sys/uvm/uvm_vnode.c:1.117	Sun Aug 16 00:24:41 2020
+++ src/sys/uvm/uvm_vnode.c	Sat Mar 13 15:29:55 2021
@@ -1,4 +1,4 @@
-/*	$NetBSD: uvm_vnode.c,v 1.117 2020/08/16 00:24:41 chs Exp $	*/
+/*	$NetBSD: uvm_vnode.c,v 1.118 2021/03/13 15:29:55 skrll Exp $	*/
 
 /*
  * Copyright (c) 1997 Charles D. Cranor and Washington University.
@@ -45,7 +45,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: uvm_vnode.c,v 1.117 2020/08/16 00:24:41 chs Exp $");
+__KERNEL_RCSID(0, "$NetBSD: uvm_vnode.c,v 1.118 2021/03/13 15:29:55 skrll Exp $");
 
 #ifdef _KERNEL_OPT
 #include "opt_uvmhist.h"
@@ -176,7 +176,7 @@ uvn_get(struct uvm_object *uobj, voff_t 
 	int error;
 
 	UVMHIST_FUNC(__func__);
-	UVMHIST_CALLARGS(ubchist, "vp %#jx off 0x%jx", (uintptr_t)vp, offset,
+	UVMHIST_CALLARGS(ubchist, "vp %#jx off %#jx", (uintptr_t)vp, offset,
 	    0, 0);
 
 	if (vp->v_type == VREG && (access_type & VM_PROT_WRITE) == 0
@@ -290,7 +290,7 @@ uvn_findpage(struct uvm_object *uobj, vo
 {
 	struct vm_page *pg;
 	UVMHIST_FUNC(__func__);
-	UVMHIST_CALLARGS(ubchist, "vp %#jx off 0x%jx", (uintptr_t)uobj, offset,
+	UVMHIST_CALLARGS(ubchist, "vp %#jx off %#jx", (uintptr_t)uobj, offset,
 	    0, 0);
 
 	/*
@@ -441,7 +441,7 @@ uvm_vnp_setsize(struct vnode *vp, voff_t
 	UVMHIST_FUNC(__func__); UVMHIST_CALLED(ubchist);
 
 	rw_enter(uobj->vmobjlock, RW_WRITER);
-	UVMHIST_LOG(ubchist, "vp %#jx old 0x%jx new 0x%jx",
+	UVMHIST_LOG(ubchist, "vp %#jx old %#jx new %#jx",
 	    (uintptr_t)vp, vp->v_size, newsize, 0);
 
 	/*

Index: src/sys/uvm/pmap/pmap_segtab.c
diff -u src/sys/uvm/pmap/pmap_segtab.c:1.26 src/sys/uvm/pmap/pmap_segtab.c:1.27
--- src/sys/uvm/pmap/pmap_segtab.c:1.26	Thu Oct  8 14:02:40 2020
+++ src/sys/uvm/pmap/pmap_segtab.c	Sat Mar 13 15:29:55 2021
@@ -1,4 +1,4 @@
-/*	$NetBSD: pmap_segtab.c,v 1.26 2020/10/08 14:02:40 skrll Exp $	*/
+/*	$NetBSD: pmap_segtab.c,v 1.27 2021/03/13 15:29:55 skrll Exp $	*/
 
 /*-
  * Copyright (c) 1998, 2001 The NetBSD Foundation, Inc.
@@ -67,7 +67,7 @@
 
 #include <sys/cdefs.h>
 
-__KERNEL_RCSID(0, "$NetBSD: pmap_segtab.c,v 1.26 2020/10/08 14:02:40 skrll Exp $");
+__KERNEL_RCSID(0, "$NetBSD: pmap_segtab.c,v 1.27 2021/03/13 15:29:55 skrll Exp $");
 
 /*
  *	Manages physical address maps.
@@ -265,7 +265,7 @@ pmap_segtab_release(pmap_t pmap, pmap_se
 	UVMHIST_FUNC(__func__);
 	UVMHIST_CALLARGS(pmapsegtabhist, "pm=%#jx stpp=%#jx free=%jd",
 	    (uintptr_t)pmap, (uintptr_t)stp_p, free_stp, 0);
-	UVMHIST_LOG(pmapsegtabhist, " callback=%#jx flags=%jx va=%#jx vinc=%#jx",
+	UVMHIST_LOG(pmapsegtabhist, " callback=%#jx flags=%#jx va=%#jx vinc=%#jx",
 	    (uintptr_t)callback, flags, (uintptr_t)va, (uintptr_t)vinc);
 	for (size_t i = (va / vinc) & (PMAP_SEGTABSIZE - 1);
 	     i < PMAP_SEGTABSIZE;

Reply via email to