diff options
author | davidxu <davidxu@FreeBSD.org> | 2002-11-17 11:47:03 +0000 |
---|---|---|
committer | davidxu <davidxu@FreeBSD.org> | 2002-11-17 11:47:03 +0000 |
commit | 7742d9b2867e95aff29d864698265dfce7b054d5 (patch) | |
tree | 1378f78156261fd06918dc6cddd6ff681f1ad56c /sys/kern | |
parent | 2e6e4a9c2a577bb3ef91a5e4beb65c499ad40dcf (diff) | |
download | FreeBSD-src-7742d9b2867e95aff29d864698265dfce7b054d5.zip FreeBSD-src-7742d9b2867e95aff29d864698265dfce7b054d5.tar.gz |
1.Add sysctls to control KSE resource allocation.
kern.threads.max_threads_per_proc
kern.threads.max_groups_per_proc
2.Temporary disable borrower thread stash itself as
owner thread's spare thread in thread_exit(). there
is a race between owner thread and borrow thread:
an owner thread may allocate a spare thread as this:
if (td->td_standin == NULL)
td->standin = thread_alloc();
but thread_alloc() can block the thread, then a borrower
thread would possible stash it self as owner's spare
thread in thread_exit(), after owner is resumed, result
is a thread leak in kernel, double check in owner can
avoid the race, but it may be ugly and not worth to do.
Diffstat (limited to 'sys/kern')
-rw-r--r-- | sys/kern/kern_kse.c | 38 | ||||
-rw-r--r-- | sys/kern/kern_thread.c | 38 |
2 files changed, 60 insertions, 16 deletions
diff --git a/sys/kern/kern_kse.c b/sys/kern/kern_kse.c index d4e74f0..ae329e5 100644 --- a/sys/kern/kern_kse.c +++ b/sys/kern/kern_kse.c @@ -64,14 +64,22 @@ static uma_zone_t thread_zone; /* DEBUG ONLY */ SYSCTL_NODE(_kern, OID_AUTO, threads, CTLFLAG_RW, 0, "thread allocation"); -static int oiks_debug = 1; /* 0 disable, 1 printf, 2 enter debugger */ +static int oiks_debug = 0; /* 0 disable, 1 printf, 2 enter debugger */ SYSCTL_INT(_kern_threads, OID_AUTO, oiks, CTLFLAG_RW, &oiks_debug, 0, "OIKS thread debug"); -static int max_threads_per_proc = 10; -SYSCTL_INT(_kern_threads, OID_AUTO, max_per_proc, CTLFLAG_RW, +static int oiks_max_threads_per_proc = 10; +SYSCTL_INT(_kern_threads, OID_AUTO, oiks_max_per_proc, CTLFLAG_RW, + &oiks_max_threads_per_proc, 0, "Debug limit on threads per proc"); + +static int max_threads_per_proc = 30; +SYSCTL_INT(_kern_threads, OID_AUTO, max_threads_per_proc, CTLFLAG_RW, &max_threads_per_proc, 0, "Limit on threads per proc"); +static int max_groups_per_proc = 5; +SYSCTL_INT(_kern_threads, OID_AUTO, max_groups_per_proc, CTLFLAG_RW, + &max_groups_per_proc, 0, "Limit on thread groups per proc"); + #define RANGEOF(type, start, end) (offsetof(type, end) - offsetof(type, start)) struct threadqueue zombie_threads = TAILQ_HEAD_INITIALIZER(zombie_threads); @@ -449,6 +457,8 @@ kse_create(struct thread *td, struct kse_create_args *uap) p->p_flag |= P_KSES; /* easier to just set it than to test and set */ kg = td->td_ksegrp; if (uap->newgroup) { + if (p->p_numksegrps >= max_groups_per_proc) + return (EPROCLIM); /* * If we want a new KSEGRP it doesn't matter whether * we have already fired up KSE mode before or not. @@ -483,12 +493,12 @@ kse_create(struct thread *td, struct kse_create_args *uap) * which is safe. */ if ((td->td_flags & TDF_UNBOUND) || td->td_kse->ke_mailbox) { -#if 0 /* while debugging */ + if (oiks_debug == 0) { #ifdef SMP if (kg->kg_kses > mp_ncpus) #endif return (EPROCLIM); -#endif + } newke = kse_alloc(); } else { newke = NULL; @@ -507,8 +517,15 @@ kse_create(struct thread *td, struct kse_create_args *uap) td->td_standin = thread_alloc(); } mtx_lock_spin(&sched_lock); - if (newkg) + if (newkg) { + if (p->p_numksegrps >= max_groups_per_proc) { + mtx_unlock_spin(&sched_lock); + ksegrp_free(newkg); + kse_free(newke); + return (EPROCLIM); + } ksegrp_link(newkg, p); + } else newkg = kg; kse_link(newke, newkg); @@ -980,6 +997,7 @@ thread_exit(void) } else { kse_reassign(ke); } +#if 0 if (ke->ke_bound) { /* * WE are a borrower.. @@ -990,12 +1008,15 @@ thread_exit(void) } ke->ke_bound->td_standin = td; } else { +#endif if (ke->ke_tdspare != NULL) { thread_stash(ke->ke_tdspare); ke->ke_tdspare = NULL; } ke->ke_tdspare = td; +#if 0 } +#endif PROC_UNLOCK(p); td->td_state = TDS_INACTIVE; td->td_proc = NULL; @@ -1034,7 +1055,7 @@ thread_link(struct thread *td, struct ksegrp *kg) TAILQ_INSERT_HEAD(&kg->kg_threads, td, td_kglist); p->p_numthreads++; kg->kg_numthreads++; - if (oiks_debug && p->p_numthreads > max_threads_per_proc) { + if (oiks_debug && (p->p_numthreads > oiks_max_threads_per_proc)) { printf("OIKS %d\n", p->p_numthreads); if (oiks_debug > 1) Debugger("OIKS"); @@ -1245,7 +1266,8 @@ thread_user_enter(struct proc *p, struct thread *td) (void *)fuword( (void *)&ke->ke_mailbox->km_curthread); #endif if ((td->td_mailbox == NULL) || - (td->td_mailbox == (void *)-1)) { + (td->td_mailbox == (void *)-1) || + (p->p_numthreads > max_threads_per_proc)) { td->td_mailbox = NULL; /* single thread it.. */ td->td_flags &= ~TDF_UNBOUND; } else { diff --git a/sys/kern/kern_thread.c b/sys/kern/kern_thread.c index d4e74f0..ae329e5 100644 --- a/sys/kern/kern_thread.c +++ b/sys/kern/kern_thread.c @@ -64,14 +64,22 @@ static uma_zone_t thread_zone; /* DEBUG ONLY */ SYSCTL_NODE(_kern, OID_AUTO, threads, CTLFLAG_RW, 0, "thread allocation"); -static int oiks_debug = 1; /* 0 disable, 1 printf, 2 enter debugger */ +static int oiks_debug = 0; /* 0 disable, 1 printf, 2 enter debugger */ SYSCTL_INT(_kern_threads, OID_AUTO, oiks, CTLFLAG_RW, &oiks_debug, 0, "OIKS thread debug"); -static int max_threads_per_proc = 10; -SYSCTL_INT(_kern_threads, OID_AUTO, max_per_proc, CTLFLAG_RW, +static int oiks_max_threads_per_proc = 10; +SYSCTL_INT(_kern_threads, OID_AUTO, oiks_max_per_proc, CTLFLAG_RW, + &oiks_max_threads_per_proc, 0, "Debug limit on threads per proc"); + +static int max_threads_per_proc = 30; +SYSCTL_INT(_kern_threads, OID_AUTO, max_threads_per_proc, CTLFLAG_RW, &max_threads_per_proc, 0, "Limit on threads per proc"); +static int max_groups_per_proc = 5; +SYSCTL_INT(_kern_threads, OID_AUTO, max_groups_per_proc, CTLFLAG_RW, + &max_groups_per_proc, 0, "Limit on thread groups per proc"); + #define RANGEOF(type, start, end) (offsetof(type, end) - offsetof(type, start)) struct threadqueue zombie_threads = TAILQ_HEAD_INITIALIZER(zombie_threads); @@ -449,6 +457,8 @@ kse_create(struct thread *td, struct kse_create_args *uap) p->p_flag |= P_KSES; /* easier to just set it than to test and set */ kg = td->td_ksegrp; if (uap->newgroup) { + if (p->p_numksegrps >= max_groups_per_proc) + return (EPROCLIM); /* * If we want a new KSEGRP it doesn't matter whether * we have already fired up KSE mode before or not. @@ -483,12 +493,12 @@ kse_create(struct thread *td, struct kse_create_args *uap) * which is safe. */ if ((td->td_flags & TDF_UNBOUND) || td->td_kse->ke_mailbox) { -#if 0 /* while debugging */ + if (oiks_debug == 0) { #ifdef SMP if (kg->kg_kses > mp_ncpus) #endif return (EPROCLIM); -#endif + } newke = kse_alloc(); } else { newke = NULL; @@ -507,8 +517,15 @@ kse_create(struct thread *td, struct kse_create_args *uap) td->td_standin = thread_alloc(); } mtx_lock_spin(&sched_lock); - if (newkg) + if (newkg) { + if (p->p_numksegrps >= max_groups_per_proc) { + mtx_unlock_spin(&sched_lock); + ksegrp_free(newkg); + kse_free(newke); + return (EPROCLIM); + } ksegrp_link(newkg, p); + } else newkg = kg; kse_link(newke, newkg); @@ -980,6 +997,7 @@ thread_exit(void) } else { kse_reassign(ke); } +#if 0 if (ke->ke_bound) { /* * WE are a borrower.. @@ -990,12 +1008,15 @@ thread_exit(void) } ke->ke_bound->td_standin = td; } else { +#endif if (ke->ke_tdspare != NULL) { thread_stash(ke->ke_tdspare); ke->ke_tdspare = NULL; } ke->ke_tdspare = td; +#if 0 } +#endif PROC_UNLOCK(p); td->td_state = TDS_INACTIVE; td->td_proc = NULL; @@ -1034,7 +1055,7 @@ thread_link(struct thread *td, struct ksegrp *kg) TAILQ_INSERT_HEAD(&kg->kg_threads, td, td_kglist); p->p_numthreads++; kg->kg_numthreads++; - if (oiks_debug && p->p_numthreads > max_threads_per_proc) { + if (oiks_debug && (p->p_numthreads > oiks_max_threads_per_proc)) { printf("OIKS %d\n", p->p_numthreads); if (oiks_debug > 1) Debugger("OIKS"); @@ -1245,7 +1266,8 @@ thread_user_enter(struct proc *p, struct thread *td) (void *)fuword( (void *)&ke->ke_mailbox->km_curthread); #endif if ((td->td_mailbox == NULL) || - (td->td_mailbox == (void *)-1)) { + (td->td_mailbox == (void *)-1) || + (p->p_numthreads > max_threads_per_proc)) { td->td_mailbox = NULL; /* single thread it.. */ td->td_flags &= ~TDF_UNBOUND; } else { |