summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorbde <bde@FreeBSD.org>1999-11-27 12:32:27 +0000
committerbde <bde@FreeBSD.org>1999-11-27 12:32:27 +0000
commit4955977bad69ae0939c6eede4d50f3c8e9f843bf (patch)
tree5e379c99576655a86acc625c1e7b89043b3b1cfe
parent6661b59c9e36f23b9cbcf25f513565c3a64a7392 (diff)
downloadFreeBSD-src-4955977bad69ae0939c6eede4d50f3c8e9f843bf.zip
FreeBSD-src-4955977bad69ae0939c6eede4d50f3c8e9f843bf.tar.gz
Moved scheduling-related code to kern_synch.c so that it is easier to fix
and extend. The new function containing the code is named schedclock() as in NetBSD, but it has slightly different semantics (it already handles incrementation of p->p_cpticks, and it should handle any calling frequency). Agreed with in principle by: dufault
-rw-r--r--sys/amd64/amd64/mp_machdep.c9
-rw-r--r--sys/amd64/amd64/mptable.c9
-rw-r--r--sys/amd64/include/mptable.h9
-rw-r--r--sys/i386/i386/mp_machdep.c9
-rw-r--r--sys/i386/i386/mptable.c9
-rw-r--r--sys/i386/include/mptable.h9
-rw-r--r--sys/kern/kern_clock.c22
-rw-r--r--sys/kern/kern_synch.c26
-rw-r--r--sys/kern/kern_tc.c22
-rw-r--r--sys/kern/subr_smp.c9
-rw-r--r--sys/sys/proc.h1
11 files changed, 36 insertions, 98 deletions
diff --git a/sys/amd64/amd64/mp_machdep.c b/sys/amd64/amd64/mp_machdep.c
index d101973..b565cee 100644
--- a/sys/amd64/amd64/mp_machdep.c
+++ b/sys/amd64/amd64/mp_machdep.c
@@ -2323,14 +2323,7 @@ forwarded_statclock(int id, int pscnt, int *astmap)
cp_time[CP_INTR]++;
}
if (p != NULL) {
- p->p_cpticks++;
- if (++p->p_estcpu == 0)
- p->p_estcpu--;
- if ((p->p_estcpu & 3) == 0) {
- resetpriority(p);
- if (p->p_priority >= PUSER)
- p->p_priority = p->p_usrpri;
- }
+ schedclock(p);
/* Update resource usage integrals and maximums. */
if ((pstats = p->p_stats) != NULL &&
diff --git a/sys/amd64/amd64/mptable.c b/sys/amd64/amd64/mptable.c
index d101973..b565cee 100644
--- a/sys/amd64/amd64/mptable.c
+++ b/sys/amd64/amd64/mptable.c
@@ -2323,14 +2323,7 @@ forwarded_statclock(int id, int pscnt, int *astmap)
cp_time[CP_INTR]++;
}
if (p != NULL) {
- p->p_cpticks++;
- if (++p->p_estcpu == 0)
- p->p_estcpu--;
- if ((p->p_estcpu & 3) == 0) {
- resetpriority(p);
- if (p->p_priority >= PUSER)
- p->p_priority = p->p_usrpri;
- }
+ schedclock(p);
/* Update resource usage integrals and maximums. */
if ((pstats = p->p_stats) != NULL &&
diff --git a/sys/amd64/include/mptable.h b/sys/amd64/include/mptable.h
index d101973..b565cee 100644
--- a/sys/amd64/include/mptable.h
+++ b/sys/amd64/include/mptable.h
@@ -2323,14 +2323,7 @@ forwarded_statclock(int id, int pscnt, int *astmap)
cp_time[CP_INTR]++;
}
if (p != NULL) {
- p->p_cpticks++;
- if (++p->p_estcpu == 0)
- p->p_estcpu--;
- if ((p->p_estcpu & 3) == 0) {
- resetpriority(p);
- if (p->p_priority >= PUSER)
- p->p_priority = p->p_usrpri;
- }
+ schedclock(p);
/* Update resource usage integrals and maximums. */
if ((pstats = p->p_stats) != NULL &&
diff --git a/sys/i386/i386/mp_machdep.c b/sys/i386/i386/mp_machdep.c
index d101973..b565cee 100644
--- a/sys/i386/i386/mp_machdep.c
+++ b/sys/i386/i386/mp_machdep.c
@@ -2323,14 +2323,7 @@ forwarded_statclock(int id, int pscnt, int *astmap)
cp_time[CP_INTR]++;
}
if (p != NULL) {
- p->p_cpticks++;
- if (++p->p_estcpu == 0)
- p->p_estcpu--;
- if ((p->p_estcpu & 3) == 0) {
- resetpriority(p);
- if (p->p_priority >= PUSER)
- p->p_priority = p->p_usrpri;
- }
+ schedclock(p);
/* Update resource usage integrals and maximums. */
if ((pstats = p->p_stats) != NULL &&
diff --git a/sys/i386/i386/mptable.c b/sys/i386/i386/mptable.c
index d101973..b565cee 100644
--- a/sys/i386/i386/mptable.c
+++ b/sys/i386/i386/mptable.c
@@ -2323,14 +2323,7 @@ forwarded_statclock(int id, int pscnt, int *astmap)
cp_time[CP_INTR]++;
}
if (p != NULL) {
- p->p_cpticks++;
- if (++p->p_estcpu == 0)
- p->p_estcpu--;
- if ((p->p_estcpu & 3) == 0) {
- resetpriority(p);
- if (p->p_priority >= PUSER)
- p->p_priority = p->p_usrpri;
- }
+ schedclock(p);
/* Update resource usage integrals and maximums. */
if ((pstats = p->p_stats) != NULL &&
diff --git a/sys/i386/include/mptable.h b/sys/i386/include/mptable.h
index d101973..b565cee 100644
--- a/sys/i386/include/mptable.h
+++ b/sys/i386/include/mptable.h
@@ -2323,14 +2323,7 @@ forwarded_statclock(int id, int pscnt, int *astmap)
cp_time[CP_INTR]++;
}
if (p != NULL) {
- p->p_cpticks++;
- if (++p->p_estcpu == 0)
- p->p_estcpu--;
- if ((p->p_estcpu & 3) == 0) {
- resetpriority(p);
- if (p->p_priority >= PUSER)
- p->p_priority = p->p_usrpri;
- }
+ schedclock(p);
/* Update resource usage integrals and maximums. */
if ((pstats = p->p_stats) != NULL &&
diff --git a/sys/kern/kern_clock.c b/sys/kern/kern_clock.c
index 2d8a350..a8d492d 100644
--- a/sys/kern/kern_clock.c
+++ b/sys/kern/kern_clock.c
@@ -446,28 +446,8 @@ statclock(frame)
* programs: the amount of time in each cpu state.
*/
- /*
- * We adjust the priority of the current process. The priority of
- * a process gets worse as it accumulates CPU time. The cpu usage
- * estimator (p_estcpu) is increased here. The formula for computing
- * priorities (in kern_synch.c) will compute a different value each
- * time p_estcpu increases by 4. The cpu usage estimator ramps up
- * quite quickly when the process is running (linearly), and decays
- * away exponentially, at a rate which is proportionally slower when
- * the system is busy. The basic principal is that the system will
- * 90% forget that the process used a lot of CPU time in 5 * loadav
- * seconds. This causes the system to favor processes which haven't
- * run much recently, and to round-robin among other processes.
- */
if (p != NULL) {
- p->p_cpticks++;
- if (++p->p_estcpu == 0)
- p->p_estcpu--;
- if ((p->p_estcpu & 3) == 0) {
- resetpriority(p);
- if (p->p_priority >= PUSER)
- p->p_priority = p->p_usrpri;
- }
+ schedclock(p);
/* Update resource usage integrals and maximums. */
if ((pstats = p->p_stats) != NULL &&
diff --git a/sys/kern/kern_synch.c b/sys/kern/kern_synch.c
index 4811669..f14980b 100644
--- a/sys/kern/kern_synch.c
+++ b/sys/kern/kern_synch.c
@@ -899,3 +899,29 @@ sched_setup(dummy)
schedcpu(NULL);
}
+/*
+ * We adjust the priority of the current process. The priority of
+ * a process gets worse as it accumulates CPU time. The cpu usage
+ * estimator (p_estcpu) is increased here. The formula for computing
+ * priorities (in kern_synch.c) will compute a different value each
+ * time p_estcpu increases by 4. The cpu usage estimator ramps up
+ * quite quickly when the process is running (linearly), and decays
+ * away exponentially, at a rate which is proportionally slower when
+ * the system is busy. The basic principal is that the system will
+ * 90% forget that the process used a lot of CPU time in 5 * loadav
+ * seconds. This causes the system to favor processes which haven't
+ * run much recently, and to round-robin among other processes.
+ */
+void
+schedclock(p)
+ struct proc *p;
+{
+ p->p_cpticks++;
+ if (++p->p_estcpu == 0)
+ p->p_estcpu--;
+ if ((p->p_estcpu & 3) == 0) {
+ resetpriority(p);
+ if (p->p_priority >= PUSER)
+ p->p_priority = p->p_usrpri;
+ }
+}
diff --git a/sys/kern/kern_tc.c b/sys/kern/kern_tc.c
index 2d8a350..a8d492d 100644
--- a/sys/kern/kern_tc.c
+++ b/sys/kern/kern_tc.c
@@ -446,28 +446,8 @@ statclock(frame)
* programs: the amount of time in each cpu state.
*/
- /*
- * We adjust the priority of the current process. The priority of
- * a process gets worse as it accumulates CPU time. The cpu usage
- * estimator (p_estcpu) is increased here. The formula for computing
- * priorities (in kern_synch.c) will compute a different value each
- * time p_estcpu increases by 4. The cpu usage estimator ramps up
- * quite quickly when the process is running (linearly), and decays
- * away exponentially, at a rate which is proportionally slower when
- * the system is busy. The basic principal is that the system will
- * 90% forget that the process used a lot of CPU time in 5 * loadav
- * seconds. This causes the system to favor processes which haven't
- * run much recently, and to round-robin among other processes.
- */
if (p != NULL) {
- p->p_cpticks++;
- if (++p->p_estcpu == 0)
- p->p_estcpu--;
- if ((p->p_estcpu & 3) == 0) {
- resetpriority(p);
- if (p->p_priority >= PUSER)
- p->p_priority = p->p_usrpri;
- }
+ schedclock(p);
/* Update resource usage integrals and maximums. */
if ((pstats = p->p_stats) != NULL &&
diff --git a/sys/kern/subr_smp.c b/sys/kern/subr_smp.c
index d101973..b565cee 100644
--- a/sys/kern/subr_smp.c
+++ b/sys/kern/subr_smp.c
@@ -2323,14 +2323,7 @@ forwarded_statclock(int id, int pscnt, int *astmap)
cp_time[CP_INTR]++;
}
if (p != NULL) {
- p->p_cpticks++;
- if (++p->p_estcpu == 0)
- p->p_estcpu--;
- if ((p->p_estcpu & 3) == 0) {
- resetpriority(p);
- if (p->p_priority >= PUSER)
- p->p_priority = p->p_usrpri;
- }
+ schedclock(p);
/* Update resource usage integrals and maximums. */
if ((pstats = p->p_stats) != NULL &&
diff --git a/sys/sys/proc.h b/sys/sys/proc.h
index bd68a73..de6ccca 100644
--- a/sys/sys/proc.h
+++ b/sys/sys/proc.h
@@ -403,6 +403,7 @@ void procinit __P((void));
int p_trespass __P((struct proc *p1, struct proc *p2));
void resetpriority __P((struct proc *));
int roundrobin_interval __P((void));
+void schedclock __P((struct proc *));
void setrunnable __P((struct proc *));
void setrunqueue __P((struct proc *));
void sleepinit __P((void));
OpenPOWER on IntegriCloud