summaryrefslogtreecommitdiffstats
path: root/lib/libpthread
Commit message (Collapse)AuthorAgeFilesLines
...
* Correctly retrieve sigaction flags.davidxu2003-12-281-2/+2
|
* Replace a comment with more accurated one, memory heap is now protected bydavidxu2003-12-191-3/+2
| | | | new fork() wrapper.
* Code clean up, remove unused MACROS and function prototypes.davidxu2003-12-191-18/+0
|
* accept() returns a file descriptor when it succeeds which is verydeischen2003-12-092-2/+2
| | | | | | | | | likely to be non-zero. When leaving the cancellation point, check the return value against -1 to see if cancellation should be checked. While I'm here, make the same change to connect() just to be consisitent. Pointed out by: davidxu
* Remove an unused struct definition.deischen2003-12-091-12/+0
|
* Add cancellation points for accept() and connect().deischen2003-12-095-0/+106
|
* Use mutex instead of low level thread lock to implement spinlock, thisdavidxu2003-12-091-30/+18
| | | | avoids signal to be blocked when otherwise it can be handled.
* Rename _thr_enter_cancellation_point to _thr_cancel_enter, renamedavidxu2003-12-0930-136/+148
| | | | | | | | | | | | | | | | | | | | | _thr_leave_cancellation_point to _thr_cancel_leave, add a parameter to _thr_cancel_leave to indicate whether cancellation point should be checked, this gives us an option to not check cancallation point if a syscall successfully returns to avoid any leaks, current I have creat(), open() and fcntl(F_DUPFD) to not check cancellation point after they sucessfully returned. Replace some members in structure kse with bit flags to same some memory. Conditionally compile THR_ASSERT to nothing if _PTHREAD_INVARIANTS is not defined. Inline some small functions in thr_cancel.c. Use __predict_false in thr_kern.c for some executed only once code. Reviewd by: deischen
* More reliably check timeout for pthread_mutex_timedlock.davidxu2003-12-091-1/+1
|
* Go back to using rev 1.18 where thread locks are used instead of KSEdeischen2003-12-081-17/+16
| | | | | | | locks for [libc] spinlock implementation. This was previously backed out because it exposed a bug in ia64 implementation. OK'd by: marcel
* Simplify the contexts created by the kernel and remove the relatedmarcel2003-12-071-0/+6
| | | | | | | | | flags. We now create asynchronous contexts or syscall contexts only. Syscall contexts differ from the minimal ABI dictated contexts by having the scratch registers saved and restored because that's where we keep the syscall arguments and syscall return values. Since this change affects KSE, have it use kse_switchin(2) for the "new" syscall context.
* Apply a second fix for stack alignment with libkse. This time, enter thepeter2003-12-051-2/+2
| | | | | | | | | | | UTS with the stack correctly aligned. Also, while here, use an indirect jump rather than the pushq/ret hack. This fixes threaded apps that use floating point for me, although it hasn't solved all the problems. It is an improvement though. Preservation of the 128 byte red zone hasn't been resolved yet. Approved by: re (scottl)
* Eliminate two pushl by using call instruction directly, this reallydavidxu2003-11-291-2/+1
| | | | | | helps branch predict a lot for INTEL P4. Approved by: re (scottl)
* 1.Macro optimizing KSE_LOCK_ACQUIRE and THR_LOCK_ACQUIRE to use static falldavidxu2003-11-292-99/+31
| | | | | | | | | | | | | | through branch predict as suggested in INTEL IA32 optimization guide. 2.Allocate siginfo arrary separately to avoid pthread to be allocated at 2K boundary, which hits L1 address alias problem and causes context switch to be slow down. 3.Simplify context switch code by removing redundant code, code size is reduced, so it is expected to run faster. Reviewed by: deischen Approved by: re (scottl)
* Remove surplus mmap() call for stack guard page in init_private, it is donedavidxu2003-11-291-25/+0
| | | | | | | | in init_main_thread. Also don't initialize lock and lockuser again for initial thread, it is already done by _thr_alloc(). Reviewed by: deischen Approved by: re (scottl)
* Back out last change and go back to using KSE locks instead of threaddeischen2003-11-161-16/+17
| | | | | | locks until we know why this breaks ia64. Reported by: marcel
* If a thread in critical region got a synchronous signal, according currentdavidxu2003-11-091-0/+2
| | | | | | signal handling mode, there is no chance to handle the signal, something must be wrong in the library, just call kse_thr_interrupt to dump its core. I have the code for a long time, but forgot to commit it.
* Use THR lock instead of KSE lock to avoid scheduler be blocked in spinlock.davidxu2003-11-081-17/+16
| | | | Reviewed by: deischen
* style(9)deischen2003-11-051-40/+53
| | | | Reviewed by: bde
* Don't declare the malloc lock; use the declaration provided in libc.deischen2003-11-051-1/+6
| | | | Noticed by: bde
* Add pthread_atfork() source code. Dan forgot to commit this file.davidxu2003-11-051-0/+56
|
* Add an implementation for pthread_atfork().deischen2003-11-046-6/+73
| | | | | | | | Aside from the POSIX requirements for pthread_atfork(), when fork()ing, take the malloc lock to keep malloc state consistent in the child. Reviewed by: davidxu
* Add the ability to reinitialize libpthread's internal FIFO-queueingdeischen2003-11-042-15/+57
| | | | | | locks. Reviewed by: davidxu
* Add the ability to reinitialize a spinlock (libc/libpthreaddeischen2003-11-041-12/+17
| | | | | | internal lock, not a pthread spinlock). Reviewed by: davidxu
* s/foo()/foo(void)/deischen2003-11-041-2/+3
| | | | Add a blank line after a variable declaration.
* Libpthread uses the convention that all of its (non-weak) symbolsdeischen2003-11-041-5/+11
| | | | | begin with underscores and provide weak definitions without underscores. Make the pthread spinlock conform to this convention.
* Add the ability to reinitialize a mutex (internally, not a userlanddeischen2003-11-041-7/+20
| | | | | | API). Reviewed by: davidxu
* Use amd64_set_fsbase() instead of calling sysarch() directly.peter2003-10-231-6/+1
|
* This test relies on the concurrency level being 1; make it so.deischen2003-10-201-0/+3
|
* Update context code for my last ABI breakage of mcontext. I'm worriedpeter2003-10-171-7/+8
| | | | | | | about the fpu code here. It should be using fxsave/fxrstor instead of saving/restoring the control word. The SSE registers are used a lot in gcc generated code on amd64. I'm not sure how this all fits together though.
* Don't forget to initialize the fake tcb when the kcb is allocated.deischen2003-10-121-0/+3
|
* Reverse the order of the first two arguments to _sparc64_enter_uts().deischen2003-10-091-2/+2
| | | | | The first argument is the UTS function, the second argument is the first argument to the UTS function. Who's on first.
* Convert a couple of hardcoded values to constants. Make thr_getcontext()deischen2003-10-092-3/+7
| | | | | return 0 when called the first time, and return 1 when resumed by thr_setcontext().
* Add preliminary sparc64 support to libpthread. This does notdeischen2003-10-096-0/+509
| | | | | | | | yet work, but hopefully someone familiar with the sparc64 port can pick up the reins. Submitted by: jake With mods by: deischen
* Fix some comments for last commit.davidxu2003-10-081-5/+4
|
* Complete cancellation support for M:N threads, check cancelling flag whendavidxu2003-10-082-69/+157
| | | | | | | thread state is changed from RUNNING to WAIT state and do some cancellation operations for every cancellable state. Reviewed by: deischen
* Use thread lock instead of scheduler lock to eliminate lock contentiondavidxu2003-10-081-18/+21
| | | | | | for all wrapped syscalls under SMP. Reviewed by: deischen
* Fix a logic error; use beq to check for a register being NULL, not bne.deischen2003-10-021-1/+1
|
* Correct the library name.ru2003-10-021-1/+1
|
* - Remove error code that can't be returned (and original descriptionkensmith2003-10-021-2/+0
| | | | | | | | was not proper English) PR: docs/57434 Approved: blackend (mentor) Reviewed by: deischen
* Only generate code for _LCK_ASSERT if _LCK_DEBUG is defined.davidxu2003-10-021-1/+6
|
* When concurrency level is reduced and a kse is exiting, make sure no otherdavidxu2003-09-291-0/+13
| | | | | | threads are still referencing the kse by migrating them to initial kse. Reviewed by: deischen
* Remove unused variable.davidxu2003-09-281-2/+0
|
* Relink libc_r.a, libc_r.so and libc_r_p.so from libthr to libkse.marcel2003-09-271-0/+12
| | | | | | | | | | | On ia64, where there's no libc_r at all, libkse is now the default thread library by virtue of these links. The reasons for this change are: 1. libkse is slated to become the default thread library anyway, 2. active development and maintenance is only present for libkse, 3. GNOME and KDE, both in the process of being supported on ia64, work better with KSE; even on ia64.
* Add __volatile keyword.davidxu2003-09-261-2/+4
|
* s/ia64/alpha/gmarcel2003-09-261-5/+4
|
* pthread API should return error code in return value not in errno.davidxu2003-09-251-2/+2
| | | | Reviewed by: deischen
* If syscall failed, restore old sigaction and return error to thread.davidxu2003-09-251-11/+19
|
* As comments in _mutex_lock_backout state, only current threaddavidxu2003-09-241-6/+4
| | | | | | | | | | | | | | can clear the pointer to mutex, not the thread doing mutex handoff. Because _mutex_lock_backout does not hold scheduler lock while testing THR_FLAGS_IN_SYNCQ and then reading mutex pointer, it is possible mutex owner begin to unlock and handoff the mutex to the current thread, and mutex pointer will be cleared to NULL before current thread reading it, so current thread will end up with deferencing a NULL pointer, Fix the race by making mutex waiters to clear their mutex pointers. While I am here, also save inherited priority in mutex for PTHREAD_PRIO_INERIT mutex in mutex_trylock_common just like what we did in mutex_lock_common.
* Free thread name memory if there is.davidxu2003-09-231-0/+4
|
OpenPOWER on IntegriCloud