summaryrefslogtreecommitdiffstats
path: root/sys/kern/kern_switch.c
Commit message (Collapse)AuthorAgeFilesLines
...
* Make !SMP kernels compile, and as far as I can tell, work again.phk2004-10-121-1/+2
|
* Prevent preemption in slot_fill.ups2004-10-121-1/+110
| | | | | | | | | | Implement preemption between threads in the same ksegp in out of slot situations to prevent priority inversion. Tested by: pho Reviewed by: jhb, julian Approved by: sam (mentor) MFC: ASAP
* Don't release the slot twice.. sched_rem() has already done it.julian2004-10-101-1/+0
| | | | | Submitted by: stephan uphoff (ups at tree dot com) MFC after: 3 days
* When preempting a thread, put it back on the HEAD of its run queue.julian2004-10-051-6/+8
| | | | | | (Only really implemented in 4bsd) MFC after: 4 days
* Use some macros to trach available scheduler slots to allowjulian2004-10-051-3/+8
| | | | | | easier debugging. MFC after: 4 days
* The zone from which proc structures are allocated is markeddas2004-09-191-16/+0
| | | | | | | UMA_ZONE_NOFREE to guarantee type stability, so proc_fini() should never be called. Move an assertion from proc_fini() to proc_dtor() and garbage-collect the rest of the unreachable code. I have retained vm_proc_dispose(), since I consider its disuse a bug.
* clean up thread runq accounting a bit.julian2004-09-161-20/+20
| | | | MFC after: 3 days
* e specific code to revert a partial add ot teh run queue, notjulian2004-09-161-1/+9
| | | | | | remrunqueue() which can't handle a partially added thread. MFC after: 1 week
* Oops accidentally removed #ifdef SCHED_4BSDjulian2004-09-151-0/+2
| | | | | as part of another commit This function is not yet used in ULE
* Commit a fix for some panics we've been seeing with preemption.julian2004-09-131-2/+10
| | | | MFC after: 2 days
* Add some kassertsjulian2004-09-131-0/+4
|
* Add some code to allow threads to nominat a sibling to run if theyu are ↵julian2004-09-101-5/+5
| | | | | | going to sleep. MFC after: 1 week
* Make debug printf less threatenning and make it only print out once.julian2004-09-071-2/+2
| | | | MFC after: 2 days
* Don't do IPIs on behalf of interrupt threads.julian2004-09-061-2/+2
| | | | | | | just punt straight on through to teh preemption code. Make a KASSSERT out of a condition that can no longer occur. MFC after: 1 week
* Refactor a bunch of scheduler code to give basically the same behaviourjulian2004-09-051-97/+184
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | but with slightly cleaned up interfaces. The KSE structure has become the same as the "per thread scheduler private data" structure. In order to not make the diffs too great one is #defined as the other at this time. The KSE (or td_sched) structure is now allocated per thread and has no allocation code of its own. Concurrency for a KSEGRP is now kept track of via a simple pair of counters rather than using KSE structures as tokens. Since the KSE structure is different in each scheduler, kern_switch.c is now included at the end of each scheduler. Nothing outside the scheduler knows the contents of the KSE (aka td_sched) structure. The fields in the ksegrp structure that are to do with the scheduler's queueing mechanisms are now moved to the kg_sched structure. (per ksegrp scheduler private data structure). In other words how the scheduler queues and keeps track of threads is no-one's business except the scheduler's. This should allow people to write experimental schedulers with completely different internal structuring. A scheduler call sched_set_concurrency(kg, N) has been added that notifies teh scheduler that no more than N threads from that ksegrp should be allowed to be on concurrently scheduled. This is also used to enforce 'fainess' at this time so that a ksegrp with 10000 threads can not swamp a the run queue and force out a process with 1 thread, since the current code will not set the concurrency above NCPU, and both schedulers will not allow more than that many onto the system run queue at a time. Each scheduler should eventualy develop their own methods to do this now that they are effectively separated. Rejig libthr's kernel interface to follow the same code paths as linkse for scope system threads. This has slightly hurt libthr's performance but I will work to recover as much of it as I can. Thread exit code has been cleaned up greatly. exit and exec code now transitions a process back to 'standard non-threaded mode' before taking the next step. Reviewed by: scottl, peter MFC after: 1 week
* remove unused codejulian2004-09-021-109/+0
| | | | MFC after: 2 days
* Turn PREEMPTION into a kernel option. Make sure that it's defined ifscottl2004-09-021-1/+5
| | | | | | FULL_PREEMPTION is defined. Add a runtime warning to ULE if PREEMPTION is enabled (code inspired by the PREEMPTION warning in kern_switch.c). This is a possible MT5 candidate.
* Give the 4bsd scheduler the ability to wake up idle processorsjulian2004-09-011-1/+33
| | | | | | when there is new work to be done. MFC after: 5 days
* Give setrunqueue() and sched_add() more of a clue as tojulian2004-09-011-6/+6
| | | | | | where they are coming from and what is expected from them. MFC after: 2 days
* Backout the previous backout (with scott's ok). sched_ule.c:1.122 ispeter2004-08-281-0/+2
| | | | believed to fix the problem with ULE that this change triggered.
* Revert the previous change. It works great for 4BSD but causes majorscottl2004-08-201-4/+0
| | | | problems for ULE. The reason is quite unknown and worrisome.
* In maybe_preempt(), ignore threads that are in an inconsistent state. Thisscottl2004-08-201-0/+4
| | | | | | | | is an effective band-aid for at least some of the scheduler corruption seen recently. The real fix will involve protecting threads while they are inconsistent, and will come later. Submitted by: julian
* Add a temporary debugging hack to detect a deadlock in setrunqueue(). Thisscottl2004-08-101-0/+7
| | | | | | is here so that we can gather stats on the nature of the recent rash of hard lockups, and in this particular case panic the machine instead of letting it deadlock forever.
* Make kg->kg_runnable actually count runnable threads in the ksegrp run queuejulian2004-08-091-4/+5
| | | | | instead of only doing it sometimes.. This is not used outdide of debugging code in the current code, but that will probably change.
* Increase the amount of data exported by KTR in the KTR_RUNQ setting.julian2004-08-091-6/+15
| | | | | This extra data is needed to really follow what is going on in the threaded case.
* Don't scare users with a warning about preemption being off when it isn'tjhb2004-08-061-0/+2
| | | | yet safe to have on by default.
* Pass a thread argument into cpu_critical_{enter,exit}() rather thanrwatson2004-07-271-2/+2
| | | | | | | | | dereference curthread. It is called only from critical_{enter,exit}(), which already dereferences curthread. This doesn't seem to affect SMP performance in my benchmarks, but improves MySQL transaction throughput by about 1% on UP on my Xeon. Head nodding: jhb, bmilekic
* Remove the previous hack since it doesn't make a difference and is gettingscottl2004-07-231-2/+0
| | | | in the way of debugging.
* Disable the PREEMPTION-enabled code in critical_exit() that encouragesscottl2004-07-221-0/+2
| | | | | switching to a different thread. This is just a hack to try to improve stability some more, but likely points closer to the real culprit.
* - Move TDF_OWEPREEMPT, TDF_OWEUPC, and TDF_USTATCLOCK over to td_pflagsjhb2004-07-161-3/+6
| | | | | | | | | since they are only accessed by curthread and thus do not need any locking. - Move pr_addr and pr_ticks out of struct uprof (which is per-process) and directly into struct thread as td_profil_addr and td_profil_ticks as these variables are really per-thread. (They are used to defer an addupc_intr() that was too "hard" until ast()).
* Update for the KDB framework:marcel2004-07-101-1/+2
| | | | | | | | | | | | | | | | | | | | | | o Make debugging code conditional upon KDB instead of DDB. o Call kdb_enter() instead of Debugger(). o Call kdb_backtrace() instead of db_print_backtrace() or backtrace(). kern_mutex.c: o Replace checks for db_active with checks for kdb_active and make them unconditional. kern_shutdown.c: o s/DDB_UNATTENDED/KDB_UNATTENDED/g o s/DDB_TRACE/KDB_TRACE/g o Save the TID of the thread doing the kernel dump so the debugger knows which thread to select as the current when debugging the kernel core file. o Clear kdb_active instead of db_active and do so unconditionally. o Remove backtrace() implementation. kern_synch.c: o Call kdb_reenter() instead of db_error().
* Unbreak build for the the !PREEMPTION case: don't define variablesmarcel2004-07-031-0/+2
| | | | that aren't used in that case.
* Implement preemption of kernel threads natively in the scheduler ratherjhb2004-07-021-4/+93
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | than as one-off hacks in various other parts of the kernel: - Add a function maybe_preempt() that is called from sched_add() to determine if a thread about to be added to a run queue should be preempted to directly. If it is not safe to preempt or if the new thread does not have a high enough priority, then the function returns false and sched_add() adds the thread to the run queue. If the thread should be preempted to but the current thread is in a nested critical section, then the flag TDF_OWEPREEMPT is set and the thread is added to the run queue. Otherwise, mi_switch() is called immediately and the thread is never added to the run queue since it is switch to directly. When exiting an outermost critical section, if TDF_OWEPREEMPT is set, then clear it and call mi_switch() to perform the deferred preemption. - Remove explicit preemption from ithread_schedule() as calling setrunqueue() now does all the correct work. This also removes the do_switch argument from ithread_schedule(). - Do not use the manual preemption code in mtx_unlock if the architecture supports native preemption. - Don't call mi_switch() in a loop during shutdown to give ithreads a chance to run if the architecture supports native preemption since the ithreads will just preempt DELAY(). - Don't call mi_switch() from the page zeroing idle thread for architectures that support native preemption as it is unnecessary. - Native preemption is enabled on the same archs that supported ithread preemption, namely alpha, i386, and amd64. This change should largely be a NOP for the default case as committed except that we will do fewer context switches in a few cases and will avoid the run queues completely when preempting. Approved by: scottl (with his re@ hat)
* - style fixes to the critical_exit() KASSERT().jeff2004-02-021-1/+2
| | | | Submitted by: bde
* Move KASSERT regarding td_critnest to after the value of td is set torwatson2004-02-011-2/+1
| | | | | | curthread, to avoid warning and incorrect behavior. Hoped not to mind: jeff
* - Assert that td_critnest > 0 in critical_exit() to catch cases ofjeff2004-02-011-0/+2
| | | | unbalanced uses of the critical_* api.
* Although sometimes to the uninitiated, it may seem like goup, KSEGOUPrwatson2003-12-121-1/+1
| | | | | | is actually spelt KSEGROUP. Go figure. Reported by: samy@kerneled.com
* Initial landing of SMP support for FreeBSD/amd64.peter2003-11-171-2/+2
| | | | | | | | | | | | | | | | - This is heavily derived from John Baldwin's apic/pci cleanup on i386. - I have completely rewritten or drastically cleaned up some other parts. (in particular, bootstrap) - This is still a WIP. It seems that there are some highly bogus bioses on nVidia nForce3-150 boards. I can't stress how broken these boards are. I have a workaround in mind, but right now the Asus SK8N is broken. The Gigabyte K8NPro (nVidia based) is also mind-numbingly hosed. - Most of my testing has been with SCHED_ULE. SCHED_4BSD works. - the apic and acpi components are 'standard'. - If you have an nVidia nForce3-150 board, you are stuck with 'device atpic' in addition, because they somehow managed to forget to connect the 8254 timer to the apic, even though its in the same silicon! ARGH! This directly violates the ACPI spec.
* - Remove the correct thread from the run queue in setrunqueue(). Thisjeff2003-10-171-1/+1
| | | | fixes ULE + KSE.
* - Update the sched api. sched_{add,rem,clock,pctcpu} now all accept a tdjeff2003-10-161-9/+9
| | | | argument rather than a kse.
* Rename P_THREADED to P_SA. P_SA means a process is using schedulerdavidxu2003-06-151-5/+5
| | | | activations.
* Use __FBSDID().obrien2003-06-111-5/+4
|
* When we are spilling threads out of the run queue during panic, make sure wejulian2003-05-211-3/+6
| | | | | | | | | | | keep the thread state variable consistent with its real state. i.e. Don't say it's on the run queue when it isn't. Also clarify the associated comment. Turns a double panic back to a single panic :-/ Approved by: re@ (jhb)
* Commit a partial lazy thread switch mechanism for i386. it isn't as lazypeter2003-04-021-1/+17
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | as it could be and can do with some more cleanup. Currently its under options LAZY_SWITCH. What this does is avoid %cr3 reloads for short context switches that do not involve another user process. ie: we can take an interrupt, switch to a kthread and return to the user without explicitly flushing the tlb. However, this isn't as exciting as it could be, the interrupt overhead is still high and too much blocks on Giant still. There are some debug sysctls, for stats and for an on/off switch. The main problem with doing this has been "what if the process that you're running on exits while we're borrowing its address space?" - in this case we use an IPI to give it a kick when we're about to reclaim the pmap. Its not compiled in unless you add the LAZY_SWITCH option. I want to fix a few more things and get some more feedback before turning it on by default. This is NOT a replacement for Bosko's lazy interrupt stuff. This was more meant for the kthread case, while his was for interrupts. Mine helps a little for interrupts, but his helps a lot more. The stats are enabled with options SWTCH_OPTIM_STATS - this has been a pseudo-option for years, I just added a bunch of stuff to it. One non-trivial change was to select a new thread before calling cpu_switch() in the first place. This allows us to catch the silly case of doing a cpu_switch() to the current process. This happens uncomfortably often. This simplifies a bit of the asm code in cpu_switch (no longer have to call choosethread() in the middle). This has been implemented on i386 and (thanks to jake) sparc64. The others will come soon. This is actually seperate to the lazy switch stuff. Glanced at by: jake, jhb
* Adjust code for userland preemptive. Userland can set a quantum indavidxu2003-03-191-27/+1
| | | | | | | | | | kse_mailbox to schedule an upcall, this is useful for userland timeout routine, for example pthread_cond_timedwait(). Also extract upcall scheduling code from kse_reassign and create a new function called thread_switchout to include these code. Reviewed by: julain
* Cosmetic change, make it QUEUE_MACRO_DEBUG friendlydavidxu2003-03-091-1/+1
|
* Change the process flags P_KSES to be P_THREADED.julian2003-02-271-5/+5
| | | | This is just a cosmetic change but I've been meaning to do it for about a year.
* Update comments to reflect new KSE code.davidxu2003-02-191-3/+1
|
* Move code for detecting PS_NEEDSIGCHK into thread_schedule_upcall,davidxu2003-02-171-2/+0
| | | | I think it is a better place to handle it.
* Move a bunch of flags from the KSE to the thread.julian2003-02-171-0/+2
| | | | | | | | I was in two minds as to where to put them in the first case.. I should have listenned to the other mind. Submitted by: parts by davidxu@ Reviewed by: jeff@ mini@
OpenPOWER on IntegriCloud