Module Name: src Committed By: skrll Date: Thu Apr 30 20:54:27 UTC 2020
Modified Files: src/sys/kern: kern_softint.c Log Message: Trailing whitespace To generate a diff of this commit: cvs rdiff -u -r1.64 -r1.65 src/sys/kern/kern_softint.c Please note that diffs are not public domain; they are subject to the copyright notices on the relevant files.
Modified files: Index: src/sys/kern/kern_softint.c diff -u src/sys/kern/kern_softint.c:1.64 src/sys/kern/kern_softint.c:1.65 --- src/sys/kern/kern_softint.c:1.64 Fri Mar 27 00:13:52 2020 +++ src/sys/kern/kern_softint.c Thu Apr 30 20:54:27 2020 @@ -1,4 +1,4 @@ -/* $NetBSD: kern_softint.c,v 1.64 2020/03/27 00:13:52 ad Exp $ */ +/* $NetBSD: kern_softint.c,v 1.65 2020/04/30 20:54:27 skrll Exp $ */ /*- * Copyright (c) 2007, 2008, 2019, 2020 The NetBSD Foundation, Inc. @@ -112,7 +112,7 @@ * and so has a number of restrictions: * * 1) The software interrupts are not currently preemptive, so - * must wait for the currently executing LWP to yield the CPU. + * must wait for the currently executing LWP to yield the CPU. * This can introduce latency. * * 2) An expensive context switch is required for a software @@ -162,7 +162,7 @@ * * Once the soft interrupt has fired (and even if it has blocked), * no further soft interrupts at that level will be triggered by - * MI code until the soft interrupt handler has ceased execution. + * MI code until the soft interrupt handler has ceased execution. * If a soft interrupt handler blocks and is resumed, it resumes * execution as a normal LWP (kthread) and gains VM context. Only * when it has completed and is ready to fire again will it @@ -170,7 +170,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: kern_softint.c,v 1.64 2020/03/27 00:13:52 ad Exp $"); +__KERNEL_RCSID(0, "$NetBSD: kern_softint.c,v 1.65 2020/04/30 20:54:27 skrll Exp $"); #include <sys/param.h> #include <sys/proc.h> @@ -481,7 +481,7 @@ softint_schedule(void *arg) * If this assert fires, rather than disabling preemption explicitly * to make it stop, consider that you are probably using a softint * when you don't need to. - */ + */ KASSERT(kpreempt_disabled()); /* Find the handler record for this CPU. */ @@ -583,7 +583,7 @@ softint_execute(softint_t *si, lwp_t *l, (*sh->sh_func)(sh->sh_arg); KERNEL_UNLOCK_ONE(l); } - + /* Diagnostic: check that spin-locks have not leaked. */ KASSERTMSG(curcpu()->ci_mtx_count == 0, "%s: ci_mtx_count (%d) != 0, sh_func %p\n",