diff options
-rw-r--r-- | lib/libc_r/uthread/uthread_fd.c | 379 |
1 files changed, 333 insertions, 46 deletions
diff --git a/lib/libc_r/uthread/uthread_fd.c b/lib/libc_r/uthread/uthread_fd.c index 695dfd0..814fc7b 100644 --- a/lib/libc_r/uthread/uthread_fd.c +++ b/lib/libc_r/uthread/uthread_fd.c @@ -29,18 +29,19 @@ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * - * $Id: uthread_fd.c,v 1.6 1998/04/29 09:58:46 jb Exp $ + * $Id: uthread_fd.c,v 1.7 1998/05/27 00:41:22 jb Exp $ * */ #include <errno.h> #include <fcntl.h> #include <stdlib.h> +#include <string.h> #ifdef _THREAD_SAFE #include <pthread.h> #include "pthread_private.h" /* Static variables: */ -static long fd_table_lock = 0; +static spinlock_t fd_table_lock = _SPINLOCK_INITIALIZER; /* * This function *must* return -1 and set the thread specific errno @@ -53,9 +54,7 @@ int _thread_fd_table_init(int fd) { int ret = 0; - - /* Lock the file descriptor table: */ - _spinlock(&fd_table_lock); + struct fd_table_entry *entry; /* Check if the file descriptor is out of range: */ if (fd < 0 || fd >= _thread_dtablesize) { @@ -70,34 +69,31 @@ _thread_fd_table_init(int fd) */ else if (_thread_fd_table[fd] != NULL) { /* Memory has already been allocated. */ - } + /* Allocate memory for the file descriptor table entry: */ - else if ((_thread_fd_table[fd] = (struct fd_table_entry *) + } else if ((entry = (struct fd_table_entry *) malloc(sizeof(struct fd_table_entry))) == NULL) { - /* Return a bad file descriptor error: */ - errno = EBADF; + /* Return an insufficient memory error: */ + errno = ENOMEM; ret = -1; } else { - /* Assume that the operation will succeed: */ - ret = 0; - /* Initialise the file locks: */ - _thread_fd_table[fd]->access_lock = 0; - _thread_fd_table[fd]->r_owner = NULL; - _thread_fd_table[fd]->w_owner = NULL; - _thread_fd_table[fd]->r_fname = NULL; - _thread_fd_table[fd]->w_fname = NULL; - _thread_fd_table[fd]->r_lineno = 0;; - _thread_fd_table[fd]->w_lineno = 0;; - _thread_fd_table[fd]->r_lockcount = 0;; - _thread_fd_table[fd]->w_lockcount = 0;; + memset(&entry->lock, 0, sizeof(entry->lock)); + entry->r_owner = NULL; + entry->w_owner = NULL; + entry->r_fname = NULL; + entry->w_fname = NULL; + entry->r_lineno = 0;; + entry->w_lineno = 0;; + entry->r_lockcount = 0;; + entry->w_lockcount = 0;; /* Initialise the read/write queues: */ - _thread_queue_init(&_thread_fd_table[fd]->r_queue); - _thread_queue_init(&_thread_fd_table[fd]->w_queue); + _thread_queue_init(&entry->r_queue); + _thread_queue_init(&entry->w_queue); /* Get the flags for the file: */ - if (fd >= 3 && (_thread_fd_table[fd]->flags = + if (fd >= 3 && (entry->flags = _thread_sys_fcntl(fd, F_GETFL, 0)) == -1) { ret = -1; } @@ -112,8 +108,7 @@ _thread_fd_table_init(int fd) * stdio fd, is set on all stdio * fds. */ - _thread_fd_table[fd]->flags = - _pthread_stdio_flags[fd]; + entry->flags = _pthread_stdio_flags[fd]; /* * Make the file descriptor non-blocking. @@ -122,19 +117,39 @@ _thread_fd_table_init(int fd) * driver is naturally non-blocking. */ _thread_sys_fcntl(fd, F_SETFL, - _thread_fd_table[fd]->flags | O_NONBLOCK); - } + entry->flags | O_NONBLOCK); + + /* Lock the file descriptor table: */ + _SPINLOCK(&fd_table_lock); + + /* + * Check if another thread allocated the + * file descriptor entry while this thread + * was doing the same thing. The table wasn't + * kept locked during this operation because + * it has the potential to recurse. + */ + if (_thread_fd_table[fd] == NULL) { + /* This thread wins: */ + _thread_fd_table[fd] = entry; + entry = NULL; + } - /* Check if one of the fcntl calls failed: */ - if (ret != 0) { - /* Free the file descriptor table entry: */ - free(_thread_fd_table[fd]); - _thread_fd_table[fd] = NULL; + /* Unlock the file descriptor table: */ + _SPINUNLOCK(&fd_table_lock); } - } - /* Unlock the file descriptor table: */ - _atomic_unlock(&fd_table_lock); + /* + * Check if another thread initialised the table entry + * before this one could: + */ + if (entry != NULL) + /* + * Throw away the table entry that this thread + * prepared. The other thread wins. + */ + free(entry); + } /* Return the completion status: */ return (ret); @@ -155,7 +170,7 @@ _thread_fd_unlock(int fd, int lock_type) * other threads for clashing with the current * thread's accesses: */ - _spinlock(&_thread_fd_table[fd]->access_lock); + _SPINLOCK(&_thread_fd_table[fd]->lock); /* Check if the running thread owns the read lock: */ if (_thread_fd_table[fd]->r_owner == _thread_run) { @@ -227,7 +242,7 @@ _thread_fd_unlock(int fd, int lock_type) * Reset the number of write locks. * This will be incremented by the * new owner of the lock when it - * sees that it has the lock. + * sees that it has the lock. */ _thread_fd_table[fd]->w_lockcount = 0; } @@ -235,15 +250,286 @@ _thread_fd_unlock(int fd, int lock_type) } /* Unlock the file descriptor table entry: */ - _atomic_unlock(&_thread_fd_table[fd]->access_lock); + _SPINUNLOCK(&_thread_fd_table[fd]->lock); + } + + /* Nothing to return. */ + return; +} + +int +_thread_fd_lock(int fd, int lock_type, struct timespec * timeout) +{ + int ret; + + /* + * Check that the file descriptor table is initialised for this + * entry: + */ + if ((ret = _thread_fd_table_init(fd)) == 0) { + /* + * Lock the file descriptor table entry to prevent + * other threads for clashing with the current + * thread's accesses: + */ + _SPINLOCK(&_thread_fd_table[fd]->lock); + + /* Check the file descriptor and lock types: */ + if (lock_type == FD_READ || lock_type == FD_RDWR) { + /* + * Enter a loop to wait for the file descriptor to be + * locked for read for the current thread: + */ + while (_thread_fd_table[fd]->r_owner != _thread_run) { + /* + * Check if the file descriptor is locked by + * another thread: + */ + if (_thread_fd_table[fd]->r_owner != NULL) { + /* + * Another thread has locked the file + * descriptor for read, so join the + * queue of threads waiting for a + * read lock on this file descriptor: + */ + _thread_queue_enq(&_thread_fd_table[fd]->r_queue, _thread_run); + + /* + * Save the file descriptor details + * in the thread structure for the + * running thread: + */ + _thread_run->data.fd.fd = fd; + + /* Set the timeout: */ + _thread_kern_set_timeout(timeout); + + /* + * Unlock the file descriptor + * table entry: + */ + _SPINUNLOCK(&_thread_fd_table[fd]->lock); + + /* + * Schedule this thread to wait on + * the read lock. It will only be + * woken when it becomes the next in + * the queue and is granted access + * to the lock by the thread + * that is unlocking the file + * descriptor. + */ + _thread_kern_sched_state(PS_FDLR_WAIT, __FILE__, __LINE__); + + /* + * Lock the file descriptor + * table entry again: + */ + _SPINLOCK(&_thread_fd_table[fd]->lock); + + } else { + /* + * The running thread now owns the + * read lock on this file descriptor: + */ + _thread_fd_table[fd]->r_owner = _thread_run; + + /* + * Reset the number of read locks for + * this file descriptor: + */ + _thread_fd_table[fd]->r_lockcount = 0; + } + } + + /* Increment the read lock count: */ + _thread_fd_table[fd]->r_lockcount++; + } + /* Check the file descriptor and lock types: */ + if (lock_type == FD_WRITE || lock_type == FD_RDWR) { + /* + * Enter a loop to wait for the file descriptor to be + * locked for write for the current thread: + */ + while (_thread_fd_table[fd]->w_owner != _thread_run) { + /* + * Check if the file descriptor is locked by + * another thread: + */ + if (_thread_fd_table[fd]->w_owner != NULL) { + /* + * Another thread has locked the file + * descriptor for write, so join the + * queue of threads waiting for a + * write lock on this file + * descriptor: + */ + _thread_queue_enq(&_thread_fd_table[fd]->w_queue, _thread_run); + + /* + * Save the file descriptor details + * in the thread structure for the + * running thread: + */ + _thread_run->data.fd.fd = fd; + + /* Set the timeout: */ + _thread_kern_set_timeout(timeout); + + /* + * Unlock the file descriptor + * table entry: + */ + _SPINUNLOCK(&_thread_fd_table[fd]->lock); + + /* + * Schedule this thread to wait on + * the write lock. It will only be + * woken when it becomes the next in + * the queue and is granted access to + * the lock by the thread that is + * unlocking the file descriptor. + */ + _thread_kern_sched_state(PS_FDLW_WAIT, __FILE__, __LINE__); + + /* + * Lock the file descriptor + * table entry again: + */ + _SPINLOCK(&_thread_fd_table[fd]->lock); + } else { + /* + * The running thread now owns the + * write lock on this file + * descriptor: + */ + _thread_fd_table[fd]->w_owner = _thread_run; + + /* + * Reset the number of write locks + * for this file descriptor: + */ + _thread_fd_table[fd]->w_lockcount = 0; + } + } + + /* Increment the write lock count: */ + _thread_fd_table[fd]->w_lockcount++; + } + + /* Unlock the file descriptor table entry: */ + _SPINUNLOCK(&_thread_fd_table[fd]->lock); } + + /* Return the completion status: */ + return (ret); +} + +void +_thread_fd_unlock_debug(int fd, int lock_type, char *fname, int lineno) +{ + int ret; + + /* + * Check that the file descriptor table is initialised for this + * entry: + */ + if ((ret = _thread_fd_table_init(fd)) == 0) { + /* + * Lock the file descriptor table entry to prevent + * other threads for clashing with the current + * thread's accesses: + */ + _spinlock_debug(&_thread_fd_table[fd]->lock, fname, lineno); + + /* Check if the running thread owns the read lock: */ + if (_thread_fd_table[fd]->r_owner == _thread_run) { + /* Check the file descriptor and lock types: */ + if (lock_type == FD_READ || lock_type == FD_RDWR) { + /* + * Decrement the read lock count for the + * running thread: + */ + _thread_fd_table[fd]->r_lockcount--; + + /* + * Check if the running thread still has read + * locks on this file descriptor: + */ + if (_thread_fd_table[fd]->r_lockcount != 0) { + } + /* + * Get the next thread in the queue for a + * read lock on this file descriptor: + */ + else if ((_thread_fd_table[fd]->r_owner = _thread_queue_deq(&_thread_fd_table[fd]->r_queue)) == NULL) { + } else { + /* + * Set the state of the new owner of + * the thread to running: + */ + PTHREAD_NEW_STATE(_thread_fd_table[fd]->r_owner,PS_RUNNING); + + /* + * Reset the number of read locks. + * This will be incremented by the + * new owner of the lock when it sees + * that it has the lock. + */ + _thread_fd_table[fd]->r_lockcount = 0; + } + } + } + /* Check if the running thread owns the write lock: */ + if (_thread_fd_table[fd]->w_owner == _thread_run) { + /* Check the file descriptor and lock types: */ + if (lock_type == FD_WRITE || lock_type == FD_RDWR) { + /* + * Decrement the write lock count for the + * running thread: + */ + _thread_fd_table[fd]->w_lockcount--; + + /* + * Check if the running thread still has + * write locks on this file descriptor: + */ + if (_thread_fd_table[fd]->w_lockcount != 0) { + } + /* + * Get the next thread in the queue for a + * write lock on this file descriptor: + */ + else if ((_thread_fd_table[fd]->w_owner = _thread_queue_deq(&_thread_fd_table[fd]->w_queue)) == NULL) { + } else { + /* + * Set the state of the new owner of + * the thread to running: + */ + PTHREAD_NEW_STATE(_thread_fd_table[fd]->w_owner,PS_RUNNING); + + /* + * Reset the number of write locks. + * This will be incremented by the + * new owner of the lock when it + * sees that it has the lock. + */ + _thread_fd_table[fd]->w_lockcount = 0; + } + } + } + + /* Unlock the file descriptor table entry: */ + _SPINUNLOCK(&_thread_fd_table[fd]->lock); + } + /* Nothing to return. */ return; } int -_thread_fd_lock(int fd, int lock_type, struct timespec * timeout, +_thread_fd_lock_debug(int fd, int lock_type, struct timespec * timeout, char *fname, int lineno) { int ret; @@ -258,7 +544,7 @@ _thread_fd_lock(int fd, int lock_type, struct timespec * timeout, * other threads for clashing with the current * thread's accesses: */ - _spinlock(&_thread_fd_table[fd]->access_lock); + _spinlock_debug(&_thread_fd_table[fd]->lock, fname, lineno); /* Check the file descriptor and lock types: */ if (lock_type == FD_READ || lock_type == FD_RDWR) { @@ -296,7 +582,7 @@ _thread_fd_lock(int fd, int lock_type, struct timespec * timeout, * Unlock the file descriptor * table entry: */ - _atomic_unlock(&_thread_fd_table[fd]->access_lock); + _SPINUNLOCK(&_thread_fd_table[fd]->lock); /* * Schedule this thread to wait on @@ -313,7 +599,7 @@ _thread_fd_lock(int fd, int lock_type, struct timespec * timeout, * Lock the file descriptor * table entry again: */ - _spinlock(&_thread_fd_table[fd]->access_lock); + _SPINLOCK(&_thread_fd_table[fd]->lock); } else { /* @@ -340,6 +626,7 @@ _thread_fd_lock(int fd, int lock_type, struct timespec * timeout, /* Increment the read lock count: */ _thread_fd_table[fd]->r_lockcount++; } + /* Check the file descriptor and lock types: */ if (lock_type == FD_WRITE || lock_type == FD_RDWR) { /* @@ -377,7 +664,7 @@ _thread_fd_lock(int fd, int lock_type, struct timespec * timeout, * Unlock the file descriptor * table entry: */ - _atomic_unlock(&_thread_fd_table[fd]->access_lock); + _SPINUNLOCK(&_thread_fd_table[fd]->lock); /* * Schedule this thread to wait on @@ -393,7 +680,7 @@ _thread_fd_lock(int fd, int lock_type, struct timespec * timeout, * Lock the file descriptor * table entry again: */ - _spinlock(&_thread_fd_table[fd]->access_lock); + _SPINLOCK(&_thread_fd_table[fd]->lock); } else { /* * The running thread now owns the @@ -422,7 +709,7 @@ _thread_fd_lock(int fd, int lock_type, struct timespec * timeout, } /* Unlock the file descriptor table entry: */ - _atomic_unlock(&_thread_fd_table[fd]->access_lock); + _SPINUNLOCK(&_thread_fd_table[fd]->lock); } /* Return the completion status: */ |