Module Name:    src
Committed By:   mrg
Date:           Thu Mar  2 20:11:19 UTC 2017

Modified Files:
        src/sys/uvm/pmap: pmap.c

Log Message:
avoid using %s in UVMHIST.


To generate a diff of this commit:
cvs rdiff -u -r1.27 -r1.28 src/sys/uvm/pmap/pmap.c

Please note that diffs are not public domain; they are subject to the
copyright notices on the relevant files.

Modified files:

Index: src/sys/uvm/pmap/pmap.c
diff -u src/sys/uvm/pmap/pmap.c:1.27 src/sys/uvm/pmap/pmap.c:1.28
--- src/sys/uvm/pmap/pmap.c:1.27	Fri Dec 23 09:16:46 2016
+++ src/sys/uvm/pmap/pmap.c	Thu Mar  2 20:11:19 2017
@@ -1,4 +1,4 @@
-/*	$NetBSD: pmap.c,v 1.27 2016/12/23 09:16:46 skrll Exp $	*/
+/*	$NetBSD: pmap.c,v 1.28 2017/03/02 20:11:19 mrg Exp $	*/
 
 /*-
  * Copyright (c) 1998, 2001 The NetBSD Foundation, Inc.
@@ -67,7 +67,7 @@
 
 #include <sys/cdefs.h>
 
-__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.27 2016/12/23 09:16:46 skrll Exp $");
+__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.28 2017/03/02 20:11:19 mrg Exp $");
 
 /*
  *	Manages physical address maps.
@@ -668,7 +668,7 @@ pmap_activate(struct lwp *l)
 	((l)->l_name ? (l)->l_name : (l)->l_proc->p_comm)
 
 	UVMHIST_FUNC(__func__); UVMHIST_CALLED(pmaphist);
-	UVMHIST_LOG(pmaphist, "(l=%p pmap=%p (%s))", l, pmap, LNAME(l), 0);
+	UVMHIST_LOG(pmaphist, "(l=%p pmap=%p)", l, pmap, 0, 0);
 	PMAP_COUNT(activate);
 
 	kpreempt_disable();
@@ -680,7 +680,8 @@ pmap_activate(struct lwp *l)
 	pmap_md_tlb_miss_lock_exit();
 	kpreempt_enable();
 
-	UVMHIST_LOG(pmaphist, " <-- done (%u:%u)", l->l_proc->p_pid, l->l_lid, 0, 0);
+	UVMHIST_LOG(pmaphist, " <-- done (%u:%u)", l->l_proc->p_pid, l->l_lid,
+		    0, 0);
 }
 
 /*
@@ -698,8 +699,9 @@ pmap_page_remove(struct vm_page *pg)
 
 	UVMHIST_FUNC(__func__); UVMHIST_CALLED(pmaphist);
 
-	UVMHIST_LOG(pmapexechist, "pg %p (pa %#"PRIxPADDR")%s: %s",
-	    pg, VM_PAGE_TO_PHYS(pg), " [page removed]", "execpage cleared");
+	UVMHIST_LOG(pmapexechist, "pg %p (pa %#"PRIxPADDR") [page removed]: "
+				  "execpage cleared", pg, VM_PAGE_TO_PHYS(pg),
+				  0, 0);
 #ifdef PMAP_VIRTUAL_CACHE_ALIASES
 	pmap_page_clear_attributes(mdpg, VM_PAGEMD_EXECPAGE|VM_PAGEMD_UNCACHED);
 #else
@@ -811,7 +813,7 @@ pmap_deactivate(struct lwp *l)
 	pmap_t pmap = l->l_proc->p_vmspace->vm_map.pmap;
 
 	UVMHIST_FUNC(__func__); UVMHIST_CALLED(pmaphist);
-	UVMHIST_LOG(pmaphist, "(l=%p pmap=%p (%s))", l, pmap, LNAME(l), 0);
+	UVMHIST_LOG(pmaphist, "(l=%p pmap=%p)", l, pmap, 0, 0);
 	PMAP_COUNT(deactivate);
 
 	kpreempt_disable();
@@ -825,7 +827,8 @@ pmap_deactivate(struct lwp *l)
 	pmap_md_tlb_miss_lock_exit();
 	kpreempt_enable();
 
-	UVMHIST_LOG(pmaphist, " <-- done (%u:%u)", l->l_proc->p_pid, l->l_lid, 0, 0);
+	UVMHIST_LOG(pmaphist, " <-- done (%u:%u)", l->l_proc->p_pid, l->l_lid,
+		    0, 0);
 }
 
 void
@@ -858,8 +861,8 @@ pmap_update(struct pmap *pmap)
 	pmap_md_tlb_miss_lock_exit();
 	kpreempt_enable();
 
-	UVMHIST_LOG(pmaphist, " <-- done%s",
-	    (pmap == pmap_kernel()) ? " (kernel)" : "", 0, 0, 0);
+	UVMHIST_LOG(pmaphist, " <-- done (%c)",
+		    (pmap == pmap_kernel() ? 'k' : 'u'), 0, 0, 0);
 }
 
 /*
@@ -877,8 +880,8 @@ pmap_pte_remove(pmap_t pmap, vaddr_t sva
 	const bool is_kernel_pmap_p = (pmap == pmap_kernel());
 
 	UVMHIST_FUNC(__func__); UVMHIST_CALLED(pmaphist);
-	UVMHIST_LOG(pmaphist, "(pmap=%p %sva=%#"PRIxVADDR"..%#"PRIxVADDR,
-	    pmap, (is_kernel_pmap_p ? "(kernel) " : ""), sva, eva);
+	UVMHIST_LOG(pmaphist, "(pmap=%p %cva=%#"PRIxVADDR"..%#"PRIxVADDR,
+	    pmap, (is_kernel_pmap_p ? 'k' : 'u'), sva, eva);
 	UVMHIST_LOG(pmaphist, "ptep=%p, flags(npte)=%#"PRIxPTR")",
 	    ptep, flags, 0, 0);
 
@@ -1017,8 +1020,8 @@ pmap_pte_protect(pmap_t pmap, vaddr_t sv
 	const vm_prot_t prot = (flags & VM_PROT_ALL);
 
 	UVMHIST_FUNC(__func__); UVMHIST_CALLED(pmaphist);
-	UVMHIST_LOG(pmaphist, "(pmap=%p %sva=%#"PRIxVADDR"..%#"PRIxVADDR,
-	    pmap, (pmap == pmap_kernel() ? "(kernel) " : ""), sva, eva);
+	UVMHIST_LOG(pmaphist, "(pmap=%p %cva=%#"PRIxVADDR"..%#"PRIxVADDR")",
+	    pmap, (pmap == pmap_kernel() ? 'k' : 'u'), sva, eva);
 	UVMHIST_LOG(pmaphist, "ptep=%p, flags(npte)=%#"PRIxPTR")",
 	    ptep, flags, 0, 0);
 
@@ -1039,9 +1042,9 @@ pmap_pte_protect(pmap_t pmap, vaddr_t sv
 				if (VM_PAGEMD_CACHED_P(mdpg)) {
 #endif
 					UVMHIST_LOG(pmapexechist,
-					    "pg %p (pa %#"PRIxPADDR"): %s",
-					    pg, VM_PAGE_TO_PHYS(pg),
-					    "syncicached performed", 0);
+					    "pg %p (pa %#"PRIxPADDR"): "
+					    "syncicached performed",
+					    pg, VM_PAGE_TO_PHYS(pg), 0, 0);
 					pmap_page_syncicache(pg);
 					PMAP_COUNT(exec_synced_protect);
 #ifdef PMAP_VIRTUAL_CACHE_ALIASES
@@ -1108,8 +1111,8 @@ pmap_page_cache(struct vm_page *pg, bool
 	struct vm_page_md * const mdpg = VM_PAGE_TO_MD(pg);
 
 	UVMHIST_FUNC(__func__); UVMHIST_CALLED(pmaphist);
-	UVMHIST_LOG(pmaphist, "(pg=%p (pa %#"PRIxPADDR") cached=%s)",
-	    pg, VM_PAGE_TO_PHYS(pg), cached ? "true" : "false", 0);
+	UVMHIST_LOG(pmaphist, "(pg=%p (pa %#"PRIxPADDR") cached=%d)",
+	    pg, VM_PAGE_TO_PHYS(pg), cached, 0);
 
 	KASSERT(kpreempt_disabled());
 	KASSERT(VM_PAGEMD_PVLIST_LOCKED_P(mdpg));
@@ -1169,19 +1172,9 @@ pmap_enter(pmap_t pmap, vaddr_t va, padd
 #endif
 
 	UVMHIST_FUNC(__func__); UVMHIST_CALLED(*histp);
-#define VM_PROT_STRING(prot) \
-	&"\0     " \
-	 "(R)\0  " \
-	 "(W)\0  " \
-	 "(RW)\0 " \
-	 "(X)\0  " \
-	 "(RX)\0 " \
-	 "(WX)\0 " \
-	 "(RWX)\0"[UVM_PROTECTION(prot)*6]
 	UVMHIST_LOG(*histp, "(pmap=%p, va=%#"PRIxVADDR", pa=%#"PRIxPADDR,
 	    pmap, va, pa, 0);
-	UVMHIST_LOG(*histp, "prot=%#x%s flags=%#x%s)",
-	    prot, VM_PROT_STRING(prot), flags, VM_PROT_STRING(flags));
+	UVMHIST_LOG(*histp, "prot=%#x flags=%#x)", prot, flags, 0, 0);
 
 	const bool good_color = PMAP_PAGE_COLOROK_P(pa, va);
 	if (is_kernel_pmap_p) {
@@ -1289,25 +1282,22 @@ pmap_enter(pmap_t pmap, vaddr_t va, padd
 		PMAP_COUNT(exec_mappings);
 		if (!VM_PAGEMD_EXECPAGE_P(mdpg) && pte_cached_p(npte)) {
 			if (!pte_deferred_exec_p(npte)) {
-				UVMHIST_LOG(*histp,
-				    "va=%#"PRIxVADDR" pg %p: %s syncicache%s",
-				    va, pg, "immediate", "");
+				UVMHIST_LOG(*histp, "va=%#"PRIxVADDR
+				    " pg %p: immediate syncicache",
+				    va, pg, 0, 0);
 				pmap_page_syncicache(pg);
 				pmap_page_set_attributes(mdpg,
 				    VM_PAGEMD_EXECPAGE);
 				PMAP_COUNT(exec_synced_mappings);
 			} else {
 				UVMHIST_LOG(*histp, "va=%#"PRIxVADDR
-				    " pg %p: %s syncicache: pte %#x",
-				    va, pg, "defer", npte);
+				    " pg %p: defer syncicache: pte %#x",
+				    va, pg, npte, 0);
 			}
 		} else {
 			UVMHIST_LOG(*histp,
-			    "va=%#"PRIxVADDR" pg %p: %s syncicache%s",
-			    va, pg, "no",
-			    (pte_cached_p(npte)
-				? " (already exec)"
-				: " (uncached)"));
+			    "va=%#"PRIxVADDR" pg %p: no syncicache cached %d",
+			    va, pg, "no", pte_cached_p(npte));
 		}
 	} else if (pg != NULL && (prot & VM_PROT_EXECUTE)) {
 		KASSERT(mdpg != NULL);
@@ -1316,8 +1306,8 @@ pmap_enter(pmap_t pmap, vaddr_t va, padd
 		pmap_page_syncicache(pg);
 		pmap_page_clear_attributes(mdpg, VM_PAGEMD_EXECPAGE);
 		UVMHIST_LOG(*histp,
-		    "va=%#"PRIxVADDR" pg %p: %s syncicache%s",
-		    va, pg, "immediate", " (writeable)");
+		    "va=%#"PRIxVADDR" pg %p: immediate syncicache (writeable)",
+		    va, pg, 0, 0);
 	}
 
 	UVMHIST_LOG(*histp, " <-- 0 (OK)", 0, 0, 0, 0);
@@ -1595,7 +1585,7 @@ pmap_clear_reference(struct vm_page *pg)
 
 	bool rv = pmap_page_clear_attributes(mdpg, VM_PAGEMD_REFERENCED);
 
-	UVMHIST_LOG(pmaphist, " <-- %s", rv ? "true" : "false", 0, 0, 0);
+	UVMHIST_LOG(pmaphist, " <-- wasref %u", rv, 0, 0, 0);
 
 	return rv;
 }
@@ -1630,14 +1620,14 @@ pmap_clear_modify(struct vm_page *pg)
 	if (VM_PAGEMD_EXECPAGE_P(mdpg)) {
 		if (pv->pv_pmap == NULL) {
 			UVMHIST_LOG(pmapexechist,
-			    "pg %p (pa %#"PRIxPADDR"): %s",
-			    pg, VM_PAGE_TO_PHYS(pg), "execpage cleared", 0);
+			    "pg %p (pa %#"PRIxPADDR"): execpage cleared",
+			    pg, VM_PAGE_TO_PHYS(pg), 0, 0);
 			pmap_page_clear_attributes(mdpg, VM_PAGEMD_EXECPAGE);
 			PMAP_COUNT(exec_uncached_clear_modify);
 		} else {
 			UVMHIST_LOG(pmapexechist,
-			    "pg %p (pa %#"PRIxPADDR"): %s",
-			    pg, VM_PAGE_TO_PHYS(pg), "syncicache performed", 0);
+			    "pg %p (pa %#"PRIxPADDR"): syncicache performed",
+			    pg, VM_PAGE_TO_PHYS(pg), 0, 0);
 			pmap_page_syncicache(pg);
 			PMAP_COUNT(exec_synced_clear_modify);
 		}
@@ -1839,8 +1829,8 @@ again:
 				if (__predict_false(apv != NULL))
 					pmap_pv_free(apv);
 
-				UVMHIST_LOG(pmaphist, " <-- done pv=%p%s",
-				    pv, " (reused)", 0, 0);
+				UVMHIST_LOG(pmaphist, " <-- done pv=%p (reused)",
+				    pv, 0, 0, 0);
 				return;
 			}
 		}
@@ -1891,8 +1881,7 @@ again:
 	if (__predict_false(apv != NULL))
 		pmap_pv_free(apv);
 
-	UVMHIST_LOG(pmaphist, " <-- done pv=%p%s",
-	    pv, first ? " (first pv)" : "",0,0);
+	UVMHIST_LOG(pmaphist, " <-- done pv=%p (first %u)", pv, first, 0, 0);
 }
 
 /*
@@ -1913,7 +1902,7 @@ pmap_remove_pv(pmap_t pmap, vaddr_t va, 
 	UVMHIST_LOG(pmaphist,
 	    "(pmap=%p, va=%#"PRIxVADDR", pg=%p (pa %#"PRIxPADDR")",
 	    pmap, va, pg, VM_PAGE_TO_PHYS(pg));
-	UVMHIST_LOG(pmaphist, "dirty=%s)", dirty ? "true" : "false", 0, 0, 0);
+	UVMHIST_LOG(pmaphist, "dirty=%u)", dirty, 0, 0, 0);
 
 	KASSERT(kpreempt_disabled());
 	KASSERT((va & PAGE_MASK) == 0);
@@ -1973,10 +1962,8 @@ pmap_remove_pv(pmap_t pmap, vaddr_t va, 
 			 * care about its execness.
 			 */
 			UVMHIST_LOG(pmapexechist,
-			    "pg %p (pa %#"PRIxPADDR")%s: %s",
-			    pg, VM_PAGE_TO_PHYS(pg),
-			    last ? " [last mapping]" : "",
-			    "execpage cleared");
+			    "pg %p (pa %#"PRIxPADDR")last %u: execpage cleared",
+			    pg, VM_PAGE_TO_PHYS(pg), last, 0);
 			pmap_page_clear_attributes(mdpg, VM_PAGEMD_EXECPAGE);
 			PMAP_COUNT(exec_uncached_remove);
 		} else {
@@ -1985,10 +1972,8 @@ pmap_remove_pv(pmap_t pmap, vaddr_t va, 
 			 * so we must sync it.
 			 */
 			UVMHIST_LOG(pmapexechist,
-			    "pg %p (pa %#"PRIxPADDR")%s: %s",
-			    pg, VM_PAGE_TO_PHYS(pg),
-			    last ? " [last mapping]" : "",
-			    "performed syncicache");
+			    "pg %p (pa %#"PRIxPADDR")last %u: performed syncicache",
+			    pg, VM_PAGE_TO_PHYS(pg), last, 0);
 			pmap_page_syncicache(pg);
 			PMAP_COUNT(exec_synced_remove);
 		}

Reply via email to