Module Name: src
Committed By: ad
Date: Fri Dec 6 18:33:19 UTC 2019
Modified Files:
src/sys/kern: sched_4bsd.c sched_m2.c
Log Message:
sched_tick(): don't try to optimise something that's called 10 times a
second, it's a fine way to introduce bugs (and I did). Use the MI
interface for rescheduling which always does the correct thing.
To generate a diff of this commit:
cvs rdiff -u -r1.40 -r1.41 src/sys/kern/sched_4bsd.c
cvs rdiff -u -r1.36 -r1.37 src/sys/kern/sched_m2.c
Please note that diffs are not public domain; they are subject to the
copyright notices on the relevant files.
Modified files:
Index: src/sys/kern/sched_4bsd.c
diff -u src/sys/kern/sched_4bsd.c:1.40 src/sys/kern/sched_4bsd.c:1.41
--- src/sys/kern/sched_4bsd.c:1.40 Sun Dec 1 15:34:46 2019
+++ src/sys/kern/sched_4bsd.c Fri Dec 6 18:33:19 2019
@@ -1,4 +1,4 @@
-/* $NetBSD: sched_4bsd.c,v 1.40 2019/12/01 15:34:46 ad Exp $ */
+/* $NetBSD: sched_4bsd.c,v 1.41 2019/12/06 18:33:19 ad Exp $ */
/*
* Copyright (c) 1999, 2000, 2004, 2006, 2007, 2008, 2019
@@ -69,7 +69,7 @@
*/
#include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: sched_4bsd.c,v 1.40 2019/12/01 15:34:46 ad Exp $");
+__KERNEL_RCSID(0, "$NetBSD: sched_4bsd.c,v 1.41 2019/12/06 18:33:19 ad Exp $");
#include "opt_ddb.h"
#include "opt_lockdebug.h"
@@ -104,13 +104,15 @@ void
sched_tick(struct cpu_info *ci)
{
struct schedstate_percpu *spc = &ci->ci_schedstate;
+ pri_t pri = PRI_NONE;
lwp_t *l;
spc->spc_ticks = rrticks;
if (CURCPU_IDLE_P()) {
- atomic_or_uint(&ci->ci_want_resched,
- RESCHED_IDLE | RESCHED_UPREEMPT);
+ spc_lock(ci);
+ sched_resched_cpu(ci, MAXPRI_KTHREAD, true);
+ /* spc now unlocked */
return;
}
l = ci->ci_onproc;
@@ -128,12 +130,7 @@ sched_tick(struct cpu_info *ci)
break;
case SCHED_RR:
/* Force it into mi_switch() to look for other jobs to run. */
-#ifdef __HAVE_PREEMPTION
- atomic_or_uint(&l->l_dopreempt, DOPREEMPT_ACTIVE);
- atomic_or_uint(&ci->ci_want_resched, RESCHED_KPREEMPT);
-#else
- atomic_or_uint(&ci->ci_want_resched, RESCHED_UPREEMPT);
-#endif
+ pri = MAXPRI_KERNEL_RT;
break;
default:
if (spc->spc_flags & SPCF_SHOULDYIELD) {
@@ -142,25 +139,25 @@ sched_tick(struct cpu_info *ci)
* due to buggy or inefficient code. Force a
* kernel preemption.
*/
-#ifdef __HAVE_PREEMPTION
- atomic_or_uint(&l->l_dopreempt, DOPREEMPT_ACTIVE);
- atomic_or_uint(&ci->ci_want_resched, RESCHED_KPREEMPT);
-#else
- atomic_or_uint(&ci->ci_want_resched, RESCHED_UPREEMPT);
-#endif
+ pri = MAXPRI_KERNEL_RT;
} else if (spc->spc_flags & SPCF_SEENRR) {
/*
* The process has already been through a roundrobin
* without switching and may be hogging the CPU.
* Indicate that the process should yield.
*/
- spc->spc_flags |= SPCF_SHOULDYIELD;
- atomic_or_uint(&ci->ci_want_resched, RESCHED_UPREEMPT);
+ pri = MAXPRI_KTHREAD;
} else {
spc->spc_flags |= SPCF_SEENRR;
}
break;
}
+
+ if (pri != PRI_NONE) {
+ spc_lock(ci);
+ sched_resched_cpu(ci, pri, true);
+ /* spc now unlocked */
+ }
}
/*
Index: src/sys/kern/sched_m2.c
diff -u src/sys/kern/sched_m2.c:1.36 src/sys/kern/sched_m2.c:1.37
--- src/sys/kern/sched_m2.c:1.36 Sun Dec 1 15:34:46 2019
+++ src/sys/kern/sched_m2.c Fri Dec 6 18:33:19 2019
@@ -1,4 +1,4 @@
-/* $NetBSD: sched_m2.c,v 1.36 2019/12/01 15:34:46 ad Exp $ */
+/* $NetBSD: sched_m2.c,v 1.37 2019/12/06 18:33:19 ad Exp $ */
/*
* Copyright (c) 2007, 2008 Mindaugas Rasiukevicius <rmind at NetBSD org>
@@ -33,7 +33,7 @@
*/
#include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: sched_m2.c,v 1.36 2019/12/01 15:34:46 ad Exp $");
+__KERNEL_RCSID(0, "$NetBSD: sched_m2.c,v 1.37 2019/12/06 18:33:19 ad Exp $");
#include <sys/param.h>
@@ -330,7 +330,9 @@ sched_tick(struct cpu_info *ci)
*/
if (lwp_eprio(l) <= spc->spc_maxpriority || l->l_target_cpu) {
spc->spc_flags |= SPCF_SHOULDYIELD;
- atomic_or_uint(&ci->ci_want_resched, RESCHED_UPREEMPT);
+ spc_lock(ci);
+ sched_resched_cpu(ci, MAXPRI_KTHREAD, true);
+ /* spc now unlocked */
} else
spc->spc_ticks = l->l_sched.timeslice;
lwp_unlock(l);