summaryrefslogtreecommitdiffstats
path: root/sys/kern/sched_ule.c
diff options
context:
space:
mode:
authormav <mav@FreeBSD.org>2014-09-17 14:06:21 +0000
committermav <mav@FreeBSD.org>2014-09-17 14:06:21 +0000
commit0e425be7bbb51e540023cafb50cb3a0de3b8abf9 (patch)
treeea47499283999aef711fb42cd91ef108ed09d29b /sys/kern/sched_ule.c
parent312a175bdaf969fe82f1472761a6ac1b811db765 (diff)
downloadFreeBSD-src-0e425be7bbb51e540023cafb50cb3a0de3b8abf9.zip
FreeBSD-src-0e425be7bbb51e540023cafb50cb3a0de3b8abf9.tar.gz
MFC r271604, r271616:
Add couple memory barriers to order tdq_cpu_idle and tdq_load accesses. This change fixes transient performance drops in some of my benchmarks, vanishing as soon as I am trying to collect any stats from the scheduler. It looks like reordered access to those variables sometimes caused loss of IPI_PREEMPT, that delayed thread execution until some later interrupt. Approved by: re (marius)
Diffstat (limited to 'sys/kern/sched_ule.c')
-rw-r--r--sys/kern/sched_ule.c14
1 files changed, 14 insertions, 0 deletions
diff --git a/sys/kern/sched_ule.c b/sys/kern/sched_ule.c
index a655440..961f80d 100644
--- a/sys/kern/sched_ule.c
+++ b/sys/kern/sched_ule.c
@@ -1037,6 +1037,14 @@ tdq_notify(struct tdq *tdq, struct thread *td)
ctd = pcpu_find(cpu)->pc_curthread;
if (!sched_shouldpreempt(pri, ctd->td_priority, 1))
return;
+
+ /*
+ * Make sure that tdq_load updated before calling this function
+ * is globally visible before we read tdq_cpu_idle. Idle thread
+ * accesses both of them without locks, and the order is important.
+ */
+ mb();
+
if (TD_IS_IDLETHREAD(ctd)) {
/*
* If the MD code has an idle wakeup routine try that before
@@ -2645,6 +2653,12 @@ sched_idletd(void *dummy)
/* Run main MD idle handler. */
tdq->tdq_cpu_idle = 1;
+ /*
+ * Make sure that tdq_cpu_idle update is globally visible
+ * before cpu_idle() read tdq_load. The order is important
+ * to avoid race with tdq_notify.
+ */
+ mb();
cpu_idle(switchcnt * 4 > sched_idlespinthresh);
tdq->tdq_cpu_idle = 0;
OpenPOWER on IntegriCloud