summaryrefslogtreecommitdiffstats
path: root/sys/kern/kern_idle.c
diff options
context:
space:
mode:
Diffstat (limited to 'sys/kern/kern_idle.c')
-rw-r--r--sys/kern/kern_idle.c126
1 files changed, 126 insertions, 0 deletions
diff --git a/sys/kern/kern_idle.c b/sys/kern/kern_idle.c
new file mode 100644
index 0000000..b4d43af
--- /dev/null
+++ b/sys/kern/kern_idle.c
@@ -0,0 +1,126 @@
+/*-
+ * Copyright (c) 2000, All rights reserved. See /usr/src/COPYRIGHT
+ *
+ */
+
+#include <sys/cdefs.h>
+__FBSDID("$FreeBSD$");
+
+#include "opt_ktrace.h"
+
+#include <sys/param.h>
+#include <sys/systm.h>
+#include <sys/kernel.h>
+#include <sys/ktr.h>
+#include <sys/kthread.h>
+#include <sys/lock.h>
+#include <sys/mutex.h>
+#include <sys/pcpu.h>
+#include <sys/proc.h>
+#include <sys/resourcevar.h>
+#include <sys/sched.h>
+#include <sys/smp.h>
+#include <sys/unistd.h>
+#ifdef KTRACE
+#include <sys/uio.h>
+#include <sys/ktrace.h>
+#endif
+
+static void idle_setup(void *dummy);
+SYSINIT(idle_setup, SI_SUB_SCHED_IDLE, SI_ORDER_FIRST, idle_setup, NULL)
+
+static void idle_proc(void *dummy);
+
+/*
+ * Setup per-cpu idle process contexts. The AP's shouldn't be running or
+ * accessing their idle processes at this point, so don't bother with
+ * locking.
+ */
+static void
+idle_setup(void *dummy)
+{
+#ifdef SMP
+ struct pcpu *pc;
+#endif
+ struct proc *p;
+ struct thread *td;
+ int error;
+
+#ifdef SMP
+ SLIST_FOREACH(pc, &cpuhead, pc_allcpu) {
+ error = kthread_create(idle_proc, NULL, &p,
+ RFSTOPPED | RFHIGHPID, 0, "idle: cpu%d", pc->pc_cpuid);
+ pc->pc_idlethread = FIRST_THREAD_IN_PROC(p);
+ if (pc->pc_curthread == NULL) {
+ pc->pc_curthread = pc->pc_idlethread;
+ pc->pc_idlethread->td_critnest = 0;
+ }
+#else
+ error = kthread_create(idle_proc, NULL, &p,
+ RFSTOPPED | RFHIGHPID, 0, "idle");
+ PCPU_SET(idlethread, FIRST_THREAD_IN_PROC(p));
+#endif
+ if (error)
+ panic("idle_setup: kthread_create error %d\n", error);
+
+ PROC_LOCK(p);
+ p->p_flag |= P_NOLOAD;
+ mtx_lock_spin(&sched_lock);
+ p->p_state = PRS_NORMAL;
+ td = FIRST_THREAD_IN_PROC(p);
+ td->td_state = TDS_CAN_RUN;
+ td->td_flags |= TDF_IDLETD;
+ mtx_unlock_spin(&sched_lock);
+ PROC_UNLOCK(p);
+#ifdef SMP
+ }
+#endif
+}
+
+/*
+ * idle process context
+ */
+static void
+idle_proc(void *dummy)
+{
+#ifdef DIAGNOSTIC
+ int count;
+#endif
+ struct thread *td;
+ struct proc *p;
+
+ td = curthread;
+ p = td->td_proc;
+ for (;;) {
+ mtx_assert(&Giant, MA_NOTOWNED);
+
+#ifdef DIAGNOSTIC
+ count = 0;
+
+ while (count >= 0 && sched_runnable() == 0) {
+#else
+ while (sched_runnable() == 0) {
+#endif
+ /*
+ * This is a good place to put things to be done in
+ * the background, including sanity checks.
+ */
+
+#ifdef DIAGNOSTIC
+ if (count++ < 0)
+ CTR0(KTR_PROC, "idle_proc: timed out waiting"
+ " for a process");
+#endif
+
+#ifdef __i386__
+ cpu_idle();
+#endif
+ }
+
+ mtx_lock_spin(&sched_lock);
+ p->p_stats->p_ru.ru_nvcsw++;
+ td->td_state = TDS_CAN_RUN;
+ mi_switch();
+ mtx_unlock_spin(&sched_lock);
+ }
+}
OpenPOWER on IntegriCloud