diff options
-rw-r--r-- | lib/libkvm/kvm_proc.c | 4 | ||||
-rw-r--r-- | sys/alpha/alpha/trap.c | 2 | ||||
-rw-r--r-- | sys/amd64/amd64/pmap.c | 2 | ||||
-rw-r--r-- | sys/amd64/amd64/sys_machdep.c | 2 | ||||
-rw-r--r-- | sys/amd64/amd64/trap.c | 2 | ||||
-rw-r--r-- | sys/ddb/db_ps.c | 6 | ||||
-rw-r--r-- | sys/fs/procfs/procfs_status.c | 2 | ||||
-rw-r--r-- | sys/i386/i386/pmap.c | 2 | ||||
-rw-r--r-- | sys/i386/i386/sys_machdep.c | 2 | ||||
-rw-r--r-- | sys/i386/i386/trap.c | 2 | ||||
-rw-r--r-- | sys/ia64/ia64/trap.c | 2 | ||||
-rw-r--r-- | sys/kern/kern_clock.c | 6 | ||||
-rw-r--r-- | sys/kern/kern_exec.c | 4 | ||||
-rw-r--r-- | sys/kern/kern_exit.c | 4 | ||||
-rw-r--r-- | sys/kern/kern_fork.c | 6 | ||||
-rw-r--r-- | sys/kern/kern_kse.c | 10 | ||||
-rw-r--r-- | sys/kern/kern_proc.c | 4 | ||||
-rw-r--r-- | sys/kern/kern_sig.c | 2 | ||||
-rw-r--r-- | sys/kern/kern_switch.c | 10 | ||||
-rw-r--r-- | sys/kern/kern_synch.c | 2 | ||||
-rw-r--r-- | sys/kern/kern_thread.c | 10 | ||||
-rw-r--r-- | sys/kern/sched_4bsd.c | 2 | ||||
-rw-r--r-- | sys/kern/sched_ule.c | 2 | ||||
-rw-r--r-- | sys/kern/subr_trap.c | 4 | ||||
-rw-r--r-- | sys/kern/tty.c | 2 | ||||
-rw-r--r-- | sys/sparc64/sparc64/trap.c | 2 | ||||
-rw-r--r-- | sys/sys/proc.h | 2 |
27 files changed, 50 insertions, 50 deletions
diff --git a/lib/libkvm/kvm_proc.c b/lib/libkvm/kvm_proc.c index 95b09d1..ed69394 100644 --- a/lib/libkvm/kvm_proc.c +++ b/lib/libkvm/kvm_proc.c @@ -126,7 +126,7 @@ kvm_proclist(kd, what, arg, p, bp, maxcnt) TAILQ_FIRST(&proc.p_threads)); return (-1); } - if (proc.p_flag & P_KSES == 0) { + if (proc.p_flag & P_THREADED == 0) { if (KREAD(kd, (u_long)TAILQ_FIRST(&proc.p_ksegrps), &mkg)) { @@ -364,7 +364,7 @@ nopgrp: kp->ki_lastcpu = mtd.td_lastcpu; kp->ki_wchan = mtd.td_wchan; - if (!(proc.p_flag & P_KSES)) { + if (!(proc.p_flag & P_THREADED)) { /* stuff from the ksegrp */ kp->ki_slptime = mkg.kg_slptime; kp->ki_pri.pri_class = mkg.kg_pri_class; diff --git a/sys/alpha/alpha/trap.c b/sys/alpha/alpha/trap.c index 0ee8227..718f677 100644 --- a/sys/alpha/alpha/trap.c +++ b/sys/alpha/alpha/trap.c @@ -663,7 +663,7 @@ syscall(code, framep) sticks = td->td_sticks; if (td->td_ucred != p->p_ucred) cred_update_thread(td); - if (p->p_flag & P_KSES) + if (p->p_flag & P_THREADED) thread_user_enter(p, td); #ifdef DIAGNOSTIC alpha_fpstate_check(td); diff --git a/sys/amd64/amd64/pmap.c b/sys/amd64/amd64/pmap.c index 7531770..d30be9c 100644 --- a/sys/amd64/amd64/pmap.c +++ b/sys/amd64/amd64/pmap.c @@ -3381,7 +3381,7 @@ pmap_activate(struct thread *td) /* XXXKSE this is wrong. * pmap_activate is for the current thread on the current cpu */ - if (p->p_flag & P_KSES) { + if (p->p_flag & P_THREADED) { /* Make sure all other cr3 entries are updated. */ /* what if they are running? XXXKSE (maybe abort them) */ FOREACH_THREAD_IN_PROC(p, td) { diff --git a/sys/amd64/amd64/sys_machdep.c b/sys/amd64/amd64/sys_machdep.c index 5fcad03..3e2ec74 100644 --- a/sys/amd64/amd64/sys_machdep.c +++ b/sys/amd64/amd64/sys_machdep.c @@ -129,7 +129,7 @@ i386_extend_pcb(struct thread *td) 0 /* granularity */ }; - if (td->td_proc->p_flag & P_KSES) + if (td->td_proc->p_flag & P_THREADED) return (EINVAL); /* XXXKSE */ /* XXXKSE All the code below only works in 1:1 needs changing */ ext = (struct pcb_ext *)kmem_alloc(kernel_map, ctob(IOPAGES+1)); diff --git a/sys/amd64/amd64/trap.c b/sys/amd64/amd64/trap.c index 85e4f30..ab1f1f0 100644 --- a/sys/amd64/amd64/trap.c +++ b/sys/amd64/amd64/trap.c @@ -958,7 +958,7 @@ syscall(frame) td->td_frame = &frame; if (td->td_ucred != p->p_ucred) cred_update_thread(td); - if (p->p_flag & P_KSES) + if (p->p_flag & P_THREADED) thread_user_enter(p, td); params = (caddr_t)frame.tf_esp + sizeof(int); code = frame.tf_eax; diff --git a/sys/ddb/db_ps.c b/sys/ddb/db_ps.c index a21764e..430dae8 100644 --- a/sys/ddb/db_ps.c +++ b/sys/ddb/db_ps.c @@ -125,7 +125,7 @@ db_ps(dummy1, dummy2, dummy3, dummy4) p->p_ucred != NULL ? p->p_ucred->cr_ruid : 0, pp->p_pid, p->p_pgrp != NULL ? p->p_pgrp->pg_id : 0, p->p_flag, state); - if (p->p_flag & P_KSES) + if (p->p_flag & P_THREADED) db_printf("(threaded) %s\n", p->p_comm); FOREACH_THREAD_IN_PROC(p, td) { dumpthread(p, td); @@ -141,7 +141,7 @@ db_ps(dummy1, dummy2, dummy3, dummy4) static void dumpthread(volatile struct proc *p, volatile struct thread *td) { - if (p->p_flag & P_KSES) + if (p->p_flag & P_THREADED) db_printf( " thread %p ksegrp %p ", td, td->td_ksegrp); if (TD_ON_SLEEPQ(td)) { if (td->td_flags & TDF_CVWAITQ) @@ -183,7 +183,7 @@ dumpthread(volatile struct proc *p, volatile struct thread *td) default: panic("unknown thread state"); } - if (p->p_flag & P_KSES) { + if (p->p_flag & P_THREADED) { if (td->td_kse) db_printf("[kse %p]", td->td_kse); db_printf("\n"); diff --git a/sys/fs/procfs/procfs_status.c b/sys/fs/procfs/procfs_status.c index 9e255ae..f411067 100644 --- a/sys/fs/procfs/procfs_status.c +++ b/sys/fs/procfs/procfs_status.c @@ -130,7 +130,7 @@ procfs_doprocstatus(PFS_FILL_ARGS) sbuf_printf(sb, " -1,-1 -1,-1 -1,-1"); } - if (p->p_flag & P_KSES) + if (p->p_flag & P_THREADED) sbuf_printf(sb, " %s", "-kse- "); else { tdfirst = FIRST_THREAD_IN_PROC(p); /* XXX diff from td? */ diff --git a/sys/i386/i386/pmap.c b/sys/i386/i386/pmap.c index 7531770..d30be9c 100644 --- a/sys/i386/i386/pmap.c +++ b/sys/i386/i386/pmap.c @@ -3381,7 +3381,7 @@ pmap_activate(struct thread *td) /* XXXKSE this is wrong. * pmap_activate is for the current thread on the current cpu */ - if (p->p_flag & P_KSES) { + if (p->p_flag & P_THREADED) { /* Make sure all other cr3 entries are updated. */ /* what if they are running? XXXKSE (maybe abort them) */ FOREACH_THREAD_IN_PROC(p, td) { diff --git a/sys/i386/i386/sys_machdep.c b/sys/i386/i386/sys_machdep.c index 5fcad03..3e2ec74 100644 --- a/sys/i386/i386/sys_machdep.c +++ b/sys/i386/i386/sys_machdep.c @@ -129,7 +129,7 @@ i386_extend_pcb(struct thread *td) 0 /* granularity */ }; - if (td->td_proc->p_flag & P_KSES) + if (td->td_proc->p_flag & P_THREADED) return (EINVAL); /* XXXKSE */ /* XXXKSE All the code below only works in 1:1 needs changing */ ext = (struct pcb_ext *)kmem_alloc(kernel_map, ctob(IOPAGES+1)); diff --git a/sys/i386/i386/trap.c b/sys/i386/i386/trap.c index 85e4f30..ab1f1f0 100644 --- a/sys/i386/i386/trap.c +++ b/sys/i386/i386/trap.c @@ -958,7 +958,7 @@ syscall(frame) td->td_frame = &frame; if (td->td_ucred != p->p_ucred) cred_update_thread(td); - if (p->p_flag & P_KSES) + if (p->p_flag & P_THREADED) thread_user_enter(p, td); params = (caddr_t)frame.tf_esp + sizeof(int); code = frame.tf_eax; diff --git a/sys/ia64/ia64/trap.c b/sys/ia64/ia64/trap.c index 1ce7d4e..67ba92d 100644 --- a/sys/ia64/ia64/trap.c +++ b/sys/ia64/ia64/trap.c @@ -808,7 +808,7 @@ syscall(int code, u_int64_t *args, struct trapframe *framep) framep->tf_cr_iip += 16; } - if (p->p_flag & P_KSES) + if (p->p_flag & P_THREADED) thread_user_enter(p, td); #ifdef DIAGNOSTIC ia64_fpstate_check(td); diff --git a/sys/kern/kern_clock.c b/sys/kern/kern_clock.c index 150b159..b772b94 100644 --- a/sys/kern/kern_clock.c +++ b/sys/kern/kern_clock.c @@ -166,7 +166,7 @@ hardclock_process(frame) * Run current process's virtual and profile time, as needed. */ mtx_lock_spin_flags(&sched_lock, MTX_QUIET); - if (p->p_flag & P_KSES) { + if (p->p_flag & P_THREADED) { /* XXXKSE What to do? */ } else { pstats = p->p_stats; @@ -374,7 +374,7 @@ statclock(frame) /* * Charge the time as appropriate. */ - if (p->p_flag & P_KSES) + if (p->p_flag & P_THREADED) thread_statclock(1); p->p_uticks++; if (ke->ke_ksegrp->kg_nice > NZERO) @@ -398,7 +398,7 @@ statclock(frame) p->p_iticks++; cp_time[CP_INTR]++; } else { - if (p->p_flag & P_KSES) + if (p->p_flag & P_THREADED) thread_statclock(0); td->td_sticks++; p->p_sticks++; diff --git a/sys/kern/kern_exec.c b/sys/kern/kern_exec.c index f239ab7..2c5df1c 100644 --- a/sys/kern/kern_exec.c +++ b/sys/kern/kern_exec.c @@ -200,7 +200,7 @@ kern_execve(td, fname, argv, envv, mac_p) PROC_LOCK(p); KASSERT((p->p_flag & P_INEXEC) == 0, ("%s(): process already has P_INEXEC flag", __func__)); - if (p->p_flag & P_KSES) { + if (p->p_flag & P_THREADED) { if (thread_single(SINGLE_EXIT)) { PROC_UNLOCK(p); return (ERESTART); /* Try again later. */ @@ -209,7 +209,7 @@ kern_execve(td, fname, argv, envv, mac_p) * If we get here all other threads are dead, * so unset the associated flags and lose KSE mode. */ - p->p_flag &= ~P_KSES; + p->p_flag &= ~P_THREADED; td->td_mailbox = NULL; thread_single_end(); } diff --git a/sys/kern/kern_exit.c b/sys/kern/kern_exit.c index 2e0e709..59a1a4c 100644 --- a/sys/kern/kern_exit.c +++ b/sys/kern/kern_exit.c @@ -150,7 +150,7 @@ exit1(td, rv) * XXXKSE: MUST abort all other threads before proceeding past here. */ PROC_LOCK(p); - if (p->p_flag & P_KSES) { + if (p->p_flag & P_THREADED) { /* * First check if some other thread got here before us.. * if so, act apropriatly, (exit or suspend); @@ -180,7 +180,7 @@ exit1(td, rv) * ... * Turn off threading support. */ - p->p_flag &= ~P_KSES; + p->p_flag &= ~P_THREADED; thread_single_end(); /* Don't need this any more. */ } /* diff --git a/sys/kern/kern_fork.c b/sys/kern/kern_fork.c index 19f506c..65a3360 100644 --- a/sys/kern/kern_fork.c +++ b/sys/kern/kern_fork.c @@ -289,7 +289,7 @@ fork1(td, flags, pages, procp) return (0); } - if (p1->p_flag & P_KSES) { + if (p1->p_flag & P_THREADED) { /* * Idle the other threads for a second. * Since the user space is copied, it must remain stable. @@ -744,7 +744,7 @@ again: /* * If other threads are waiting, let them continue now */ - if (p1->p_flag & P_KSES) { + if (p1->p_flag & P_THREADED) { PROC_LOCK(p1); thread_single_end(); PROC_UNLOCK(p1); @@ -758,7 +758,7 @@ again: fail: sx_xunlock(&allproc_lock); uma_zfree(proc_zone, newproc); - if (p1->p_flag & P_KSES) { + if (p1->p_flag & P_THREADED) { PROC_LOCK(p1); thread_single_end(); PROC_UNLOCK(p1); diff --git a/sys/kern/kern_kse.c b/sys/kern/kern_kse.c index aa96b4e..2029271 100644 --- a/sys/kern/kern_kse.c +++ b/sys/kern/kern_kse.c @@ -390,7 +390,7 @@ kse_thr_interrupt(struct thread *td, struct kse_thr_interrupt_args *uap) struct thread *td2; p = td->td_proc; - if (!(p->p_flag & P_KSES) || (uap->tmbx == NULL)) + if (!(p->p_flag & P_THREADED) || (uap->tmbx == NULL)) return (EINVAL); mtx_lock_spin(&sched_lock); FOREACH_THREAD_IN_PROC(p, td2) { @@ -444,7 +444,7 @@ kse_exit(struct thread *td, struct kse_exit_args *uap) upcall_remove(td); if (p->p_numthreads == 1) { kse_purge(p, td); - p->p_flag &= ~P_KSES; + p->p_flag &= ~P_THREADED; mtx_unlock_spin(&sched_lock); PROC_UNLOCK(p); } else { @@ -535,7 +535,7 @@ kse_wakeup(struct thread *td, struct kse_wakeup_args *uap) td2 = NULL; ku = NULL; /* KSE-enabled processes only, please. */ - if (!(p->p_flag & P_KSES)) + if (!(p->p_flag & P_THREADED)) return (EINVAL); PROC_LOCK(p); mtx_lock_spin(&sched_lock); @@ -609,7 +609,7 @@ kse_create(struct thread *td, struct kse_create_args *uap) ncpus = virtual_cpu; /* Easier to just set it than to test and set */ - p->p_flag |= P_KSES; + p->p_flag |= P_THREADED; kg = td->td_ksegrp; if (uap->newgroup) { /* Have race condition but it is cheap */ @@ -1778,7 +1778,7 @@ thread_single(int force_exit) PROC_LOCK_ASSERT(p, MA_OWNED); KASSERT((td != NULL), ("curthread is NULL")); - if ((p->p_flag & P_KSES) == 0) + if ((p->p_flag & P_THREADED) == 0) return (0); /* Is someone already single threading? */ diff --git a/sys/kern/kern_proc.c b/sys/kern/kern_proc.c index 00f6186..c444cb0 100644 --- a/sys/kern/kern_proc.c +++ b/sys/kern/kern_proc.c @@ -681,7 +681,7 @@ fill_kinfo_proc(p, kp) mtx_unlock_spin(&sched_lock); return; } - if (!(p->p_flag & P_KSES)) { + if (!(p->p_flag & P_THREADED)) { if (td->td_wmesg != NULL) { strlcpy(kp->ki_wmesg, td->td_wmesg, sizeof(kp->ki_wmesg)); @@ -715,7 +715,7 @@ fill_kinfo_proc(p, kp) kp->ki_swtime = p->p_swtime; kp->ki_pid = p->p_pid; /* vvv XXXKSE */ - if (!(p->p_flag & P_KSES)) { + if (!(p->p_flag & P_THREADED)) { kg = td->td_ksegrp; ke = td->td_kse; KASSERT((ke != NULL), ("fill_kinfo_proc: Null KSE")); diff --git a/sys/kern/kern_sig.c b/sys/kern/kern_sig.c index 7d45ea4..7aa9064 100644 --- a/sys/kern/kern_sig.c +++ b/sys/kern/kern_sig.c @@ -1902,7 +1902,7 @@ postsig(sig) p->p_code = 0; p->p_sig = 0; } - if (p->p_flag & P_KSES) + if (p->p_flag & P_THREADED) thread_signal_add(curthread, sig); else (*p->p_sysent->sv_sendsig)(action, sig, diff --git a/sys/kern/kern_switch.c b/sys/kern/kern_switch.c index e76c34b..a50fa60 100644 --- a/sys/kern/kern_switch.c +++ b/sys/kern/kern_switch.c @@ -127,7 +127,7 @@ retry: td = ke->ke_thread; KASSERT((td->td_kse == ke), ("kse/thread mismatch")); kg = ke->ke_ksegrp; - if (td->td_proc->p_flag & P_KSES) { + if (td->td_proc->p_flag & P_THREADED) { TAILQ_REMOVE(&kg->kg_runq, td, td_runq); if (kg->kg_last_assigned == td) { kg->kg_last_assigned = TAILQ_PREV(td, @@ -255,7 +255,7 @@ remrunqueue(struct thread *td) /* * If it is not a threaded process, take the shortcut. */ - if ((td->td_proc->p_flag & P_KSES) == 0) { + if ((td->td_proc->p_flag & P_THREADED) == 0) { /* Bring its kse with it, leave the thread attached */ sched_rem(ke); ke->ke_state = KES_THREAD; @@ -298,7 +298,7 @@ adjustrunqueue( struct thread *td, int newpri) /* * If it is not a threaded process, take the shortcut. */ - if ((td->td_proc->p_flag & P_KSES) == 0) { + if ((td->td_proc->p_flag & P_THREADED) == 0) { /* We only care about the kse in the run queue. */ td->td_priority = newpri; if (ke->ke_rqindex != (newpri / RQ_PPQ)) { @@ -339,7 +339,7 @@ setrunqueue(struct thread *td) TD_SET_RUNQ(td); kg = td->td_ksegrp; kg->kg_runnable++; - if ((td->td_proc->p_flag & P_KSES) == 0) { + if ((td->td_proc->p_flag & P_THREADED) == 0) { /* * Common path optimisation: Only one of everything * and the KSE is always already attached. @@ -659,7 +659,7 @@ thread_sanity_check(struct thread *td, char *string) } } - if ((p->p_flag & P_KSES) == 0) { + if ((p->p_flag & P_THREADED) == 0) { if (ke == NULL) { panc(string, "non KSE thread lost kse"); } diff --git a/sys/kern/kern_synch.c b/sys/kern/kern_synch.c index 9a1509f..2a89dde 100644 --- a/sys/kern/kern_synch.c +++ b/sys/kern/kern_synch.c @@ -160,7 +160,7 @@ msleep(ident, mtx, priority, wmesg, timo) * the thread (recursion here might be bad). * Hence the TDF_INMSLEEP flag. */ - if (p->p_flag & P_KSES) { + if (p->p_flag & P_THREADED) { /* * Just don't bother if we are exiting * and not the exiting thread or thread was marked as diff --git a/sys/kern/kern_thread.c b/sys/kern/kern_thread.c index aa96b4e..2029271 100644 --- a/sys/kern/kern_thread.c +++ b/sys/kern/kern_thread.c @@ -390,7 +390,7 @@ kse_thr_interrupt(struct thread *td, struct kse_thr_interrupt_args *uap) struct thread *td2; p = td->td_proc; - if (!(p->p_flag & P_KSES) || (uap->tmbx == NULL)) + if (!(p->p_flag & P_THREADED) || (uap->tmbx == NULL)) return (EINVAL); mtx_lock_spin(&sched_lock); FOREACH_THREAD_IN_PROC(p, td2) { @@ -444,7 +444,7 @@ kse_exit(struct thread *td, struct kse_exit_args *uap) upcall_remove(td); if (p->p_numthreads == 1) { kse_purge(p, td); - p->p_flag &= ~P_KSES; + p->p_flag &= ~P_THREADED; mtx_unlock_spin(&sched_lock); PROC_UNLOCK(p); } else { @@ -535,7 +535,7 @@ kse_wakeup(struct thread *td, struct kse_wakeup_args *uap) td2 = NULL; ku = NULL; /* KSE-enabled processes only, please. */ - if (!(p->p_flag & P_KSES)) + if (!(p->p_flag & P_THREADED)) return (EINVAL); PROC_LOCK(p); mtx_lock_spin(&sched_lock); @@ -609,7 +609,7 @@ kse_create(struct thread *td, struct kse_create_args *uap) ncpus = virtual_cpu; /* Easier to just set it than to test and set */ - p->p_flag |= P_KSES; + p->p_flag |= P_THREADED; kg = td->td_ksegrp; if (uap->newgroup) { /* Have race condition but it is cheap */ @@ -1778,7 +1778,7 @@ thread_single(int force_exit) PROC_LOCK_ASSERT(p, MA_OWNED); KASSERT((td != NULL), ("curthread is NULL")); - if ((p->p_flag & P_KSES) == 0) + if ((p->p_flag & P_THREADED) == 0) return (0); /* Is someone already single threading? */ diff --git a/sys/kern/sched_4bsd.c b/sys/kern/sched_4bsd.c index 3f8cc1a..0ea397b 100644 --- a/sys/kern/sched_4bsd.c +++ b/sys/kern/sched_4bsd.c @@ -544,7 +544,7 @@ sched_switchout(struct thread *td) if (TD_IS_RUNNING(td)) { /* Put us back on the run queue (kse and all). */ setrunqueue(td); - } else if (p->p_flag & P_KSES) { + } else if (p->p_flag & P_THREADED) { /* * We will not be on the run queue. So we must be * sleeping or similar. As it's available, diff --git a/sys/kern/sched_ule.c b/sys/kern/sched_ule.c index d6a2910..6861aad 100644 --- a/sys/kern/sched_ule.c +++ b/sys/kern/sched_ule.c @@ -504,7 +504,7 @@ sched_switchout(struct thread *td) * We will not be on the run queue. So we must be * sleeping or similar. */ - if (td->td_proc->p_flag & P_KSES) + if (td->td_proc->p_flag & P_THREADED) kse_reassign(ke); } diff --git a/sys/kern/subr_trap.c b/sys/kern/subr_trap.c index c0c0fd5..da4b7a8 100644 --- a/sys/kern/subr_trap.c +++ b/sys/kern/subr_trap.c @@ -109,7 +109,7 @@ userret(td, frame, oticks) /* * Do special thread processing, e.g. upcall tweaking and such. */ - if (p->p_flag & P_KSES) { + if (p->p_flag & P_THREADED) { thread_userret(td, frame); } @@ -256,7 +256,7 @@ ast(struct trapframe *framep) sigs++; } PROC_UNLOCK(p); - if (p->p_flag & P_KSES && sigs) + if (p->p_flag & P_THREADED && sigs) thread_signal_upcall(td); } diff --git a/sys/kern/tty.c b/sys/kern/tty.c index 6f8114e..8cbcb65 100644 --- a/sys/kern/tty.c +++ b/sys/kern/tty.c @@ -2407,7 +2407,7 @@ ttyinfo(struct tty *tp) PGRP_UNLOCK(tp->t_pgrp); td = FIRST_THREAD_IN_PROC(pick); - if (pick->p_flag & P_KSES) { + if (pick->p_flag & P_THREADED) { stmp = "KSE" ; /* XXXKSE */ } else { if (td) { diff --git a/sys/sparc64/sparc64/trap.c b/sys/sparc64/sparc64/trap.c index 190096a..fa31ad3 100644 --- a/sys/sparc64/sparc64/trap.c +++ b/sys/sparc64/sparc64/trap.c @@ -496,7 +496,7 @@ syscall(struct trapframe *tf) td->td_frame = tf; if (td->td_ucred != p->p_ucred) cred_update_thread(td); - if (p->p_flag & P_KSES) + if (p->p_flag & P_THREADED) thread_user_enter(p, td); code = tf->tf_global[1]; diff --git a/sys/sys/proc.h b/sys/sys/proc.h index cc77f64..368e2a1 100644 --- a/sys/sys/proc.h +++ b/sys/sys/proc.h @@ -624,7 +624,7 @@ struct proc { #define P_WAITED 0x01000 /* Someone is waiting for us */ #define P_WEXIT 0x02000 /* Working on exiting. */ #define P_EXEC 0x04000 /* Process called exec. */ -#define P_KSES 0x08000 /* Process is using KSEs. */ +#define P_THREADED 0x08000 /* Process is using threads. */ #define P_CONTINUED 0x10000 /* Proc has continued from a stopped state. */ /* flags that control how threads may be suspended for some reason */ |