Module Name:    src
Committed By:   skrll
Date:           Tue Aug  4 06:23:46 UTC 2020

Modified Files:
        src/sys/arch/x86/x86: pmap.c

Log Message:
Trailing whitespace


To generate a diff of this commit:
cvs rdiff -u -r1.402 -r1.403 src/sys/arch/x86/x86/pmap.c

Please note that diffs are not public domain; they are subject to the
copyright notices on the relevant files.

Modified files:

Index: src/sys/arch/x86/x86/pmap.c
diff -u src/sys/arch/x86/x86/pmap.c:1.402 src/sys/arch/x86/x86/pmap.c:1.403
--- src/sys/arch/x86/x86/pmap.c:1.402	Tue Aug  4 06:22:54 2020
+++ src/sys/arch/x86/x86/pmap.c	Tue Aug  4 06:23:46 2020
@@ -1,4 +1,4 @@
-/*	$NetBSD: pmap.c,v 1.402 2020/08/04 06:22:54 skrll Exp $	*/
+/*	$NetBSD: pmap.c,v 1.403 2020/08/04 06:23:46 skrll Exp $	*/
 
 /*
  * Copyright (c) 2008, 2010, 2016, 2017, 2019, 2020 The NetBSD Foundation, Inc.
@@ -130,7 +130,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.402 2020/08/04 06:22:54 skrll Exp $");
+__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.403 2020/08/04 06:23:46 skrll Exp $");
 
 #include "opt_user_ldt.h"
 #include "opt_lockdebug.h"
@@ -233,7 +233,7 @@ __KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.4
  * pg->uobject->vmobjlock, pg->uanon->an_lock
  *
  * 	For managed pages, these per-object locks are taken by the VM system
- *	before calling into the pmap module - either a read or write hold. 
+ *	before calling into the pmap module - either a read or write hold.
  *	The lock hold prevent pages from changing identity while the pmap is
  *	operating on them.  For example, the same lock is held across a call
  *	to pmap_remove() and the following call to pmap_update(), so that a
@@ -250,7 +250,7 @@ __KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.4
  *
  * pmaps_lock
  *
- *	This lock protects the list of active pmaps (headed by "pmaps"). 
+ *	This lock protects the list of active pmaps (headed by "pmaps").
  *	It's acquired when adding or removing pmaps or adjusting kernel PDEs.
  *
  * pp_lock
@@ -2066,7 +2066,7 @@ pmap_free_pv(struct pmap *pmap, struct p
 		/* part -> full */
 		LIST_REMOVE(pvp, pvp_list);
 		LIST_INSERT_HEAD(&pmap->pm_pvp_full, pvp, pvp_list);
-	} 
+	}
 }
 
 /*
@@ -2394,7 +2394,7 @@ pmap_freepage(struct pmap *pmap, struct 
 	/*
 	 * Enqueue the PTP to be freed by pmap_update().  We can't remove
 	 * the page from the uvm_object, as that can take further locks
-	 * (intolerable right now because the PTEs are likely mapped in). 
+	 * (intolerable right now because the PTEs are likely mapped in).
 	 * Instead mark the PTP as free and if we bump into it again, we'll
 	 * either ignore or reuse (depending on what's useful at the time).
 	 */
@@ -3091,7 +3091,7 @@ pmap_zap_ptp(struct pmap *pmap, struct v
 		 * tree by skipping to the next VA in the tree whenever
 		 * there is a match here.  The tree will be cleared out in
 		 * one pass before return to pmap_remove_all().
-		 */ 
+		 */
 		oattrs = pmap_pte_to_pp_attrs(opte);
 		if (pve != NULL && pve->pve_pte.pte_va == va) {
 			pp = pve->pve_pp;
@@ -3153,7 +3153,7 @@ pmap_zap_ptp(struct pmap *pmap, struct v
  *
  * Ordinarily when removing mappings it's important to hold the UVM object's
  * lock, so that pages do not gain a new identity while retaining stale TLB
- * entries (the same lock hold covers both pmap_remove() and pmap_update()). 
+ * entries (the same lock hold covers both pmap_remove() and pmap_update()).
  * Here it's known that the address space is no longer visible to any user
  * process, so we don't need to worry about that.
  */
@@ -3172,7 +3172,7 @@ pmap_remove_all(struct pmap *pmap)
 	if (pmap->pm_remove != NULL) {
 		return false;
 	}
- 
+
 	for (;;) {
 		/* Fetch a block of PTPs from tree. */
 		mutex_enter(&pmap->pm_lock);
@@ -4371,7 +4371,7 @@ pmap_pp_remove(struct pmap_page *pp, pad
 		    "va %lx pmap %p ptp %p is free", va, pmap, ptp);
 		KASSERTMSG(ptp == NULL || ptp->wire_count > 1,
 		    "va %lx pmap %p ptp %p is empty", va, pmap, ptp);
-		    
+
 #ifdef DEBUG
 		pmap_check_pv(pmap, ptp, pp, pvpte->pte_va, true);
 		rb_tree_t *tree = (ptp != NULL ?
@@ -4841,7 +4841,7 @@ pmap_enter_ma(struct pmap *pmap, vaddr_t
 	else
 #endif
 		new_pg = PHYS_TO_VM_PAGE(pa);
-		
+
 	if (new_pg != NULL) {
 		/* This is a managed page */
 		npte |= PTE_PVLIST;
@@ -5063,7 +5063,7 @@ same_pa:
 
 struct pmap_data_gnt {
 	SLIST_ENTRY(pmap_data_gnt) pd_gnt_list;
-	vaddr_t pd_gnt_sva; 
+	vaddr_t pd_gnt_sva;
 	vaddr_t pd_gnt_eva; /* range covered by this gnt */
 	int pd_gnt_refs; /* ref counter */
 	struct gnttab_map_grant_ref pd_gnt_ops[1]; /* variable length */

Reply via email to