summaryrefslogtreecommitdiffstats
path: root/sys/kern
diff options
context:
space:
mode:
authorjulian <julian@FreeBSD.org>2003-02-27 02:05:19 +0000
committerjulian <julian@FreeBSD.org>2003-02-27 02:05:19 +0000
commit3fc9836d46e83fd0a5d0daa46f5639be3db193c1 (patch)
tree91973dfb2c47deb06ea09f1614aa486db1bd95dd /sys/kern
parentef32e2b20964fad2388f8146483b4dc4bc569c2f (diff)
downloadFreeBSD-src-3fc9836d46e83fd0a5d0daa46f5639be3db193c1.zip
FreeBSD-src-3fc9836d46e83fd0a5d0daa46f5639be3db193c1.tar.gz
Change the process flags P_KSES to be P_THREADED.
This is just a cosmetic change but I've been meaning to do it for about a year.
Diffstat (limited to 'sys/kern')
-rw-r--r--sys/kern/kern_clock.c6
-rw-r--r--sys/kern/kern_exec.c4
-rw-r--r--sys/kern/kern_exit.c4
-rw-r--r--sys/kern/kern_fork.c6
-rw-r--r--sys/kern/kern_kse.c10
-rw-r--r--sys/kern/kern_proc.c4
-rw-r--r--sys/kern/kern_sig.c2
-rw-r--r--sys/kern/kern_switch.c10
-rw-r--r--sys/kern/kern_synch.c2
-rw-r--r--sys/kern/kern_thread.c10
-rw-r--r--sys/kern/sched_4bsd.c2
-rw-r--r--sys/kern/sched_ule.c2
-rw-r--r--sys/kern/subr_trap.c4
-rw-r--r--sys/kern/tty.c2
14 files changed, 34 insertions, 34 deletions
diff --git a/sys/kern/kern_clock.c b/sys/kern/kern_clock.c
index 150b159..b772b94 100644
--- a/sys/kern/kern_clock.c
+++ b/sys/kern/kern_clock.c
@@ -166,7 +166,7 @@ hardclock_process(frame)
* Run current process's virtual and profile time, as needed.
*/
mtx_lock_spin_flags(&sched_lock, MTX_QUIET);
- if (p->p_flag & P_KSES) {
+ if (p->p_flag & P_THREADED) {
/* XXXKSE What to do? */
} else {
pstats = p->p_stats;
@@ -374,7 +374,7 @@ statclock(frame)
/*
* Charge the time as appropriate.
*/
- if (p->p_flag & P_KSES)
+ if (p->p_flag & P_THREADED)
thread_statclock(1);
p->p_uticks++;
if (ke->ke_ksegrp->kg_nice > NZERO)
@@ -398,7 +398,7 @@ statclock(frame)
p->p_iticks++;
cp_time[CP_INTR]++;
} else {
- if (p->p_flag & P_KSES)
+ if (p->p_flag & P_THREADED)
thread_statclock(0);
td->td_sticks++;
p->p_sticks++;
diff --git a/sys/kern/kern_exec.c b/sys/kern/kern_exec.c
index f239ab7..2c5df1c 100644
--- a/sys/kern/kern_exec.c
+++ b/sys/kern/kern_exec.c
@@ -200,7 +200,7 @@ kern_execve(td, fname, argv, envv, mac_p)
PROC_LOCK(p);
KASSERT((p->p_flag & P_INEXEC) == 0,
("%s(): process already has P_INEXEC flag", __func__));
- if (p->p_flag & P_KSES) {
+ if (p->p_flag & P_THREADED) {
if (thread_single(SINGLE_EXIT)) {
PROC_UNLOCK(p);
return (ERESTART); /* Try again later. */
@@ -209,7 +209,7 @@ kern_execve(td, fname, argv, envv, mac_p)
* If we get here all other threads are dead,
* so unset the associated flags and lose KSE mode.
*/
- p->p_flag &= ~P_KSES;
+ p->p_flag &= ~P_THREADED;
td->td_mailbox = NULL;
thread_single_end();
}
diff --git a/sys/kern/kern_exit.c b/sys/kern/kern_exit.c
index 2e0e709..59a1a4c 100644
--- a/sys/kern/kern_exit.c
+++ b/sys/kern/kern_exit.c
@@ -150,7 +150,7 @@ exit1(td, rv)
* XXXKSE: MUST abort all other threads before proceeding past here.
*/
PROC_LOCK(p);
- if (p->p_flag & P_KSES) {
+ if (p->p_flag & P_THREADED) {
/*
* First check if some other thread got here before us..
* if so, act apropriatly, (exit or suspend);
@@ -180,7 +180,7 @@ exit1(td, rv)
* ...
* Turn off threading support.
*/
- p->p_flag &= ~P_KSES;
+ p->p_flag &= ~P_THREADED;
thread_single_end(); /* Don't need this any more. */
}
/*
diff --git a/sys/kern/kern_fork.c b/sys/kern/kern_fork.c
index 19f506c..65a3360 100644
--- a/sys/kern/kern_fork.c
+++ b/sys/kern/kern_fork.c
@@ -289,7 +289,7 @@ fork1(td, flags, pages, procp)
return (0);
}
- if (p1->p_flag & P_KSES) {
+ if (p1->p_flag & P_THREADED) {
/*
* Idle the other threads for a second.
* Since the user space is copied, it must remain stable.
@@ -744,7 +744,7 @@ again:
/*
* If other threads are waiting, let them continue now
*/
- if (p1->p_flag & P_KSES) {
+ if (p1->p_flag & P_THREADED) {
PROC_LOCK(p1);
thread_single_end();
PROC_UNLOCK(p1);
@@ -758,7 +758,7 @@ again:
fail:
sx_xunlock(&allproc_lock);
uma_zfree(proc_zone, newproc);
- if (p1->p_flag & P_KSES) {
+ if (p1->p_flag & P_THREADED) {
PROC_LOCK(p1);
thread_single_end();
PROC_UNLOCK(p1);
diff --git a/sys/kern/kern_kse.c b/sys/kern/kern_kse.c
index aa96b4e..2029271 100644
--- a/sys/kern/kern_kse.c
+++ b/sys/kern/kern_kse.c
@@ -390,7 +390,7 @@ kse_thr_interrupt(struct thread *td, struct kse_thr_interrupt_args *uap)
struct thread *td2;
p = td->td_proc;
- if (!(p->p_flag & P_KSES) || (uap->tmbx == NULL))
+ if (!(p->p_flag & P_THREADED) || (uap->tmbx == NULL))
return (EINVAL);
mtx_lock_spin(&sched_lock);
FOREACH_THREAD_IN_PROC(p, td2) {
@@ -444,7 +444,7 @@ kse_exit(struct thread *td, struct kse_exit_args *uap)
upcall_remove(td);
if (p->p_numthreads == 1) {
kse_purge(p, td);
- p->p_flag &= ~P_KSES;
+ p->p_flag &= ~P_THREADED;
mtx_unlock_spin(&sched_lock);
PROC_UNLOCK(p);
} else {
@@ -535,7 +535,7 @@ kse_wakeup(struct thread *td, struct kse_wakeup_args *uap)
td2 = NULL;
ku = NULL;
/* KSE-enabled processes only, please. */
- if (!(p->p_flag & P_KSES))
+ if (!(p->p_flag & P_THREADED))
return (EINVAL);
PROC_LOCK(p);
mtx_lock_spin(&sched_lock);
@@ -609,7 +609,7 @@ kse_create(struct thread *td, struct kse_create_args *uap)
ncpus = virtual_cpu;
/* Easier to just set it than to test and set */
- p->p_flag |= P_KSES;
+ p->p_flag |= P_THREADED;
kg = td->td_ksegrp;
if (uap->newgroup) {
/* Have race condition but it is cheap */
@@ -1778,7 +1778,7 @@ thread_single(int force_exit)
PROC_LOCK_ASSERT(p, MA_OWNED);
KASSERT((td != NULL), ("curthread is NULL"));
- if ((p->p_flag & P_KSES) == 0)
+ if ((p->p_flag & P_THREADED) == 0)
return (0);
/* Is someone already single threading? */
diff --git a/sys/kern/kern_proc.c b/sys/kern/kern_proc.c
index 00f6186..c444cb0 100644
--- a/sys/kern/kern_proc.c
+++ b/sys/kern/kern_proc.c
@@ -681,7 +681,7 @@ fill_kinfo_proc(p, kp)
mtx_unlock_spin(&sched_lock);
return;
}
- if (!(p->p_flag & P_KSES)) {
+ if (!(p->p_flag & P_THREADED)) {
if (td->td_wmesg != NULL) {
strlcpy(kp->ki_wmesg, td->td_wmesg,
sizeof(kp->ki_wmesg));
@@ -715,7 +715,7 @@ fill_kinfo_proc(p, kp)
kp->ki_swtime = p->p_swtime;
kp->ki_pid = p->p_pid;
/* vvv XXXKSE */
- if (!(p->p_flag & P_KSES)) {
+ if (!(p->p_flag & P_THREADED)) {
kg = td->td_ksegrp;
ke = td->td_kse;
KASSERT((ke != NULL), ("fill_kinfo_proc: Null KSE"));
diff --git a/sys/kern/kern_sig.c b/sys/kern/kern_sig.c
index 7d45ea4..7aa9064 100644
--- a/sys/kern/kern_sig.c
+++ b/sys/kern/kern_sig.c
@@ -1902,7 +1902,7 @@ postsig(sig)
p->p_code = 0;
p->p_sig = 0;
}
- if (p->p_flag & P_KSES)
+ if (p->p_flag & P_THREADED)
thread_signal_add(curthread, sig);
else
(*p->p_sysent->sv_sendsig)(action, sig,
diff --git a/sys/kern/kern_switch.c b/sys/kern/kern_switch.c
index e76c34b..a50fa60 100644
--- a/sys/kern/kern_switch.c
+++ b/sys/kern/kern_switch.c
@@ -127,7 +127,7 @@ retry:
td = ke->ke_thread;
KASSERT((td->td_kse == ke), ("kse/thread mismatch"));
kg = ke->ke_ksegrp;
- if (td->td_proc->p_flag & P_KSES) {
+ if (td->td_proc->p_flag & P_THREADED) {
TAILQ_REMOVE(&kg->kg_runq, td, td_runq);
if (kg->kg_last_assigned == td) {
kg->kg_last_assigned = TAILQ_PREV(td,
@@ -255,7 +255,7 @@ remrunqueue(struct thread *td)
/*
* If it is not a threaded process, take the shortcut.
*/
- if ((td->td_proc->p_flag & P_KSES) == 0) {
+ if ((td->td_proc->p_flag & P_THREADED) == 0) {
/* Bring its kse with it, leave the thread attached */
sched_rem(ke);
ke->ke_state = KES_THREAD;
@@ -298,7 +298,7 @@ adjustrunqueue( struct thread *td, int newpri)
/*
* If it is not a threaded process, take the shortcut.
*/
- if ((td->td_proc->p_flag & P_KSES) == 0) {
+ if ((td->td_proc->p_flag & P_THREADED) == 0) {
/* We only care about the kse in the run queue. */
td->td_priority = newpri;
if (ke->ke_rqindex != (newpri / RQ_PPQ)) {
@@ -339,7 +339,7 @@ setrunqueue(struct thread *td)
TD_SET_RUNQ(td);
kg = td->td_ksegrp;
kg->kg_runnable++;
- if ((td->td_proc->p_flag & P_KSES) == 0) {
+ if ((td->td_proc->p_flag & P_THREADED) == 0) {
/*
* Common path optimisation: Only one of everything
* and the KSE is always already attached.
@@ -659,7 +659,7 @@ thread_sanity_check(struct thread *td, char *string)
}
}
- if ((p->p_flag & P_KSES) == 0) {
+ if ((p->p_flag & P_THREADED) == 0) {
if (ke == NULL) {
panc(string, "non KSE thread lost kse");
}
diff --git a/sys/kern/kern_synch.c b/sys/kern/kern_synch.c
index 9a1509f..2a89dde 100644
--- a/sys/kern/kern_synch.c
+++ b/sys/kern/kern_synch.c
@@ -160,7 +160,7 @@ msleep(ident, mtx, priority, wmesg, timo)
* the thread (recursion here might be bad).
* Hence the TDF_INMSLEEP flag.
*/
- if (p->p_flag & P_KSES) {
+ if (p->p_flag & P_THREADED) {
/*
* Just don't bother if we are exiting
* and not the exiting thread or thread was marked as
diff --git a/sys/kern/kern_thread.c b/sys/kern/kern_thread.c
index aa96b4e..2029271 100644
--- a/sys/kern/kern_thread.c
+++ b/sys/kern/kern_thread.c
@@ -390,7 +390,7 @@ kse_thr_interrupt(struct thread *td, struct kse_thr_interrupt_args *uap)
struct thread *td2;
p = td->td_proc;
- if (!(p->p_flag & P_KSES) || (uap->tmbx == NULL))
+ if (!(p->p_flag & P_THREADED) || (uap->tmbx == NULL))
return (EINVAL);
mtx_lock_spin(&sched_lock);
FOREACH_THREAD_IN_PROC(p, td2) {
@@ -444,7 +444,7 @@ kse_exit(struct thread *td, struct kse_exit_args *uap)
upcall_remove(td);
if (p->p_numthreads == 1) {
kse_purge(p, td);
- p->p_flag &= ~P_KSES;
+ p->p_flag &= ~P_THREADED;
mtx_unlock_spin(&sched_lock);
PROC_UNLOCK(p);
} else {
@@ -535,7 +535,7 @@ kse_wakeup(struct thread *td, struct kse_wakeup_args *uap)
td2 = NULL;
ku = NULL;
/* KSE-enabled processes only, please. */
- if (!(p->p_flag & P_KSES))
+ if (!(p->p_flag & P_THREADED))
return (EINVAL);
PROC_LOCK(p);
mtx_lock_spin(&sched_lock);
@@ -609,7 +609,7 @@ kse_create(struct thread *td, struct kse_create_args *uap)
ncpus = virtual_cpu;
/* Easier to just set it than to test and set */
- p->p_flag |= P_KSES;
+ p->p_flag |= P_THREADED;
kg = td->td_ksegrp;
if (uap->newgroup) {
/* Have race condition but it is cheap */
@@ -1778,7 +1778,7 @@ thread_single(int force_exit)
PROC_LOCK_ASSERT(p, MA_OWNED);
KASSERT((td != NULL), ("curthread is NULL"));
- if ((p->p_flag & P_KSES) == 0)
+ if ((p->p_flag & P_THREADED) == 0)
return (0);
/* Is someone already single threading? */
diff --git a/sys/kern/sched_4bsd.c b/sys/kern/sched_4bsd.c
index 3f8cc1a..0ea397b 100644
--- a/sys/kern/sched_4bsd.c
+++ b/sys/kern/sched_4bsd.c
@@ -544,7 +544,7 @@ sched_switchout(struct thread *td)
if (TD_IS_RUNNING(td)) {
/* Put us back on the run queue (kse and all). */
setrunqueue(td);
- } else if (p->p_flag & P_KSES) {
+ } else if (p->p_flag & P_THREADED) {
/*
* We will not be on the run queue. So we must be
* sleeping or similar. As it's available,
diff --git a/sys/kern/sched_ule.c b/sys/kern/sched_ule.c
index d6a2910..6861aad 100644
--- a/sys/kern/sched_ule.c
+++ b/sys/kern/sched_ule.c
@@ -504,7 +504,7 @@ sched_switchout(struct thread *td)
* We will not be on the run queue. So we must be
* sleeping or similar.
*/
- if (td->td_proc->p_flag & P_KSES)
+ if (td->td_proc->p_flag & P_THREADED)
kse_reassign(ke);
}
diff --git a/sys/kern/subr_trap.c b/sys/kern/subr_trap.c
index c0c0fd5..da4b7a8 100644
--- a/sys/kern/subr_trap.c
+++ b/sys/kern/subr_trap.c
@@ -109,7 +109,7 @@ userret(td, frame, oticks)
/*
* Do special thread processing, e.g. upcall tweaking and such.
*/
- if (p->p_flag & P_KSES) {
+ if (p->p_flag & P_THREADED) {
thread_userret(td, frame);
}
@@ -256,7 +256,7 @@ ast(struct trapframe *framep)
sigs++;
}
PROC_UNLOCK(p);
- if (p->p_flag & P_KSES && sigs)
+ if (p->p_flag & P_THREADED && sigs)
thread_signal_upcall(td);
}
diff --git a/sys/kern/tty.c b/sys/kern/tty.c
index 6f8114e..8cbcb65 100644
--- a/sys/kern/tty.c
+++ b/sys/kern/tty.c
@@ -2407,7 +2407,7 @@ ttyinfo(struct tty *tp)
PGRP_UNLOCK(tp->t_pgrp);
td = FIRST_THREAD_IN_PROC(pick);
- if (pick->p_flag & P_KSES) {
+ if (pick->p_flag & P_THREADED) {
stmp = "KSE" ; /* XXXKSE */
} else {
if (td) {
OpenPOWER on IntegriCloud