summaryrefslogtreecommitdiffstats
path: root/lib/libpthread/sys
diff options
context:
space:
mode:
Diffstat (limited to 'lib/libpthread/sys')
-rw-r--r--lib/libpthread/sys/Makefile.inc5
-rw-r--r--lib/libpthread/sys/lock.c351
-rw-r--r--lib/libpthread/sys/lock.h95
-rw-r--r--lib/libpthread/sys/thr_error.c61
4 files changed, 0 insertions, 512 deletions
diff --git a/lib/libpthread/sys/Makefile.inc b/lib/libpthread/sys/Makefile.inc
deleted file mode 100644
index fb4a108..0000000
--- a/lib/libpthread/sys/Makefile.inc
+++ /dev/null
@@ -1,5 +0,0 @@
-# $FreeBSD$
-
-.PATH: ${.CURDIR}/sys
-
-SRCS+= lock.c thr_error.c
diff --git a/lib/libpthread/sys/lock.c b/lib/libpthread/sys/lock.c
deleted file mode 100644
index 2ac8c0c..0000000
--- a/lib/libpthread/sys/lock.c
+++ /dev/null
@@ -1,351 +0,0 @@
-/*-
- * Copyright (c) 2001, 2003 Daniel Eischen <deischen@freebsd.org>.
- * All rights reserved.
- *
- * Redistribution and use in source and binary forms, with or without
- * modification, are permitted provided that the following conditions
- * are met:
- * 1. Redistributions of source code must retain the above copyright
- * notice, this list of conditions and the following disclaimer.
- * 2. Redistributions in binary form must reproduce the above copyright
- * notice, this list of conditions and the following disclaimer in the
- * documentation and/or other materials provided with the distribution.
- *
- * THIS SOFTWARE IS PROVIDED BY THE AUTHORS AND CONTRIBUTORS ``AS IS'' AND
- * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
- * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
- * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
- * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
- * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
- * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
- * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
- * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
- * SUCH DAMAGE.
- *
- * $FreeBSD$
- */
-
-#include <sys/types.h>
-#include <machine/atomic.h>
-#include <assert.h>
-#include <stdlib.h>
-
-#include "atomic_ops.h"
-#include "lock.h"
-
-#ifdef _LOCK_DEBUG
-#define LCK_ASSERT(e) assert(e)
-#else
-#define LCK_ASSERT(e)
-#endif
-
-#define MAX_SPINS 500
-
-void
-_lock_destroy(struct lock *lck)
-{
- if ((lck != NULL) && (lck->l_head != NULL)) {
- free(lck->l_head);
- lck->l_head = NULL;
- lck->l_tail = NULL;
- }
-}
-
-int
-_lock_init(struct lock *lck, enum lock_type ltype,
- lock_handler_t *waitfunc, lock_handler_t *wakeupfunc)
-{
- if (lck == NULL)
- return (-1);
- else if ((lck->l_head = malloc(sizeof(struct lockreq))) == NULL)
- return (-1);
- else {
- lck->l_type = ltype;
- lck->l_wait = waitfunc;
- lck->l_wakeup = wakeupfunc;
- lck->l_head->lr_locked = 0;
- lck->l_head->lr_watcher = NULL;
- lck->l_head->lr_owner = NULL;
- lck->l_head->lr_active = 1;
- lck->l_tail = lck->l_head;
- }
- return (0);
-}
-
-int
-_lock_reinit(struct lock *lck, enum lock_type ltype,
- lock_handler_t *waitfunc, lock_handler_t *wakeupfunc)
-{
- if (lck == NULL)
- return (-1);
- else if (lck->l_head == NULL)
- return (_lock_init(lck, ltype, waitfunc, wakeupfunc));
- else {
- lck->l_head->lr_locked = 0;
- lck->l_head->lr_watcher = NULL;
- lck->l_head->lr_owner = NULL;
- lck->l_head->lr_active = 1;
- lck->l_tail = lck->l_head;
- }
- return (0);
-}
-
-int
-_lockuser_init(struct lockuser *lu, void *priv)
-{
- if (lu == NULL)
- return (-1);
- else if ((lu->lu_myreq == NULL) &&
- ((lu->lu_myreq = malloc(sizeof(struct lockreq))) == NULL))
- return (-1);
- else {
- lu->lu_myreq->lr_locked = 1;
- lu->lu_myreq->lr_watcher = NULL;
- lu->lu_myreq->lr_owner = lu;
- lu->lu_myreq->lr_active = 0;
- lu->lu_watchreq = NULL;
- lu->lu_priority = 0;
- lu->lu_private = priv;
- lu->lu_private2 = NULL;
- }
- return (0);
-}
-
-int
-_lockuser_reinit(struct lockuser *lu, void *priv)
-{
- if (lu == NULL)
- return (-1);
- /*
- * All lockusers keep their watch request and drop their
- * own (lu_myreq) request. Their own request is either
- * some other lockuser's watch request or is the head of
- * the lock.
- */
- lu->lu_myreq = lu->lu_watchreq;
- if (lu->lu_myreq == NULL)
- return (_lockuser_init(lu, priv));
- else {
- lu->lu_myreq->lr_locked = 1;
- lu->lu_myreq->lr_watcher = NULL;
- lu->lu_myreq->lr_owner = lu;
- lu->lu_myreq->lr_active = 0;
- lu->lu_watchreq = NULL;
- lu->lu_priority = 0;
- lu->lu_private = priv;
- lu->lu_private2 = NULL;
- }
- return (0);
-}
-
-void
-_lockuser_destroy(struct lockuser *lu)
-{
- if ((lu != NULL) && (lu->lu_myreq != NULL))
- free(lu->lu_myreq);
-}
-
-/*
- * Acquire a lock waiting (spin or sleep) for it to become available.
- */
-void
-_lock_acquire(struct lock *lck, struct lockuser *lu, int prio)
-{
- int i;
- int lval;
-
- /**
- * XXX - We probably want to remove these checks to optimize
- * performance. It is also a bug if any one of the
- * checks fail, so it's probably better to just let it
- * SEGV and fix it.
- */
-#if 0
- if (lck == NULL || lu == NULL || lck->l_head == NULL)
- return;
-#endif
- if ((lck->l_type & LCK_PRIORITY) != 0) {
- LCK_ASSERT(lu->lu_myreq->lr_locked == 1);
- LCK_ASSERT(lu->lu_myreq->lr_watcher == NULL);
- LCK_ASSERT(lu->lu_myreq->lr_owner == lu);
- LCK_ASSERT(lu->lu_watchreq == NULL);
-
- lu->lu_priority = prio;
- }
- /*
- * Atomically swap the head of the lock request with
- * this request.
- */
- atomic_swap_ptr(&lck->l_head, lu->lu_myreq, &lu->lu_watchreq);
-
- if (lu->lu_watchreq->lr_locked != 0) {
- atomic_store_rel_ptr
- ((volatile uintptr_t *)&lu->lu_watchreq->lr_watcher,
- (uintptr_t)lu);
- if ((lck->l_wait == NULL) ||
- ((lck->l_type & LCK_ADAPTIVE) == 0)) {
- while (lu->lu_watchreq->lr_locked != 0)
- ; /* spin, then yield? */
- } else {
- /*
- * Spin for a bit before invoking the wait function.
- *
- * We should be a little smarter here. If we're
- * running on a single processor, then the lock
- * owner got preempted and spinning will accomplish
- * nothing but waste time. If we're running on
- * multiple processors, the owner could be running
- * on another CPU and we might acquire the lock if
- * we spin for a bit.
- *
- * The other thing to keep in mind is that threads
- * acquiring these locks are considered to be in
- * critical regions; they will not be preempted by
- * the _UTS_ until they release the lock. It is
- * therefore safe to assume that if a lock can't
- * be acquired, it is currently held by a thread
- * running in another KSE.
- */
- for (i = 0; i < MAX_SPINS; i++) {
- if (lu->lu_watchreq->lr_locked == 0)
- return;
- if (lu->lu_watchreq->lr_active == 0)
- break;
- }
- atomic_swap_int((int *)&lu->lu_watchreq->lr_locked,
- 2, &lval);
- if (lval == 0)
- lu->lu_watchreq->lr_locked = 0;
- else
- lck->l_wait(lck, lu);
-
- }
- }
- lu->lu_myreq->lr_active = 1;
-}
-
-/*
- * Release a lock.
- */
-void
-_lock_release(struct lock *lck, struct lockuser *lu)
-{
- struct lockuser *lu_tmp, *lu_h;
- struct lockreq *myreq;
- int prio_h;
- int lval;
-
- /**
- * XXX - We probably want to remove these checks to optimize
- * performance. It is also a bug if any one of the
- * checks fail, so it's probably better to just let it
- * SEGV and fix it.
- */
-#if 0
- if ((lck == NULL) || (lu == NULL))
- return;
-#endif
- if ((lck->l_type & LCK_PRIORITY) != 0) {
- prio_h = 0;
- lu_h = NULL;
-
- /* Update tail if our request is last. */
- if (lu->lu_watchreq->lr_owner == NULL) {
- atomic_store_rel_ptr((volatile uintptr_t *)&lck->l_tail,
- (uintptr_t)lu->lu_myreq);
- atomic_store_rel_ptr
- ((volatile uintptr_t *)&lu->lu_myreq->lr_owner,
- (uintptr_t)NULL);
- } else {
- /* Remove ourselves from the list. */
- atomic_store_rel_ptr((volatile uintptr_t *)
- &lu->lu_myreq->lr_owner,
- (uintptr_t)lu->lu_watchreq->lr_owner);
- atomic_store_rel_ptr((volatile uintptr_t *)
- &lu->lu_watchreq->lr_owner->lu_myreq,
- (uintptr_t)lu->lu_myreq);
- }
- /*
- * The watch request now becomes our own because we've
- * traded away our previous request. Save our previous
- * request so that we can grant the lock.
- */
- myreq = lu->lu_myreq;
- lu->lu_myreq = lu->lu_watchreq;
- lu->lu_watchreq = NULL;
- lu->lu_myreq->lr_locked = 1;
- lu->lu_myreq->lr_owner = lu;
- lu->lu_myreq->lr_watcher = NULL;
- /*
- * Traverse the list of lock requests in reverse order
- * looking for the user with the highest priority.
- */
- for (lu_tmp = lck->l_tail->lr_watcher; lu_tmp != NULL;
- lu_tmp = lu_tmp->lu_myreq->lr_watcher) {
- if (lu_tmp->lu_priority > prio_h) {
- lu_h = lu_tmp;
- prio_h = lu_tmp->lu_priority;
- }
- }
- if (lu_h != NULL) {
- /* Give the lock to the highest priority user. */
- if (lck->l_wakeup != NULL) {
- atomic_swap_int(
- (int *)&lu_h->lu_watchreq->lr_locked,
- 0, &lval);
- if (lval == 2)
- /* Notify the sleeper */
- lck->l_wakeup(lck,
- lu_h->lu_myreq->lr_watcher);
- }
- else
- atomic_store_rel_int(
- &lu_h->lu_watchreq->lr_locked, 0);
- } else {
- if (lck->l_wakeup != NULL) {
- atomic_swap_int((int *)&myreq->lr_locked,
- 0, &lval);
- if (lval == 2)
- /* Notify the sleeper */
- lck->l_wakeup(lck, myreq->lr_watcher);
- }
- else
- /* Give the lock to the previous request. */
- atomic_store_rel_int(&myreq->lr_locked, 0);
- }
- } else {
- /*
- * The watch request now becomes our own because we've
- * traded away our previous request. Save our previous
- * request so that we can grant the lock.
- */
- myreq = lu->lu_myreq;
- lu->lu_myreq = lu->lu_watchreq;
- lu->lu_watchreq = NULL;
- lu->lu_myreq->lr_locked = 1;
- if (lck->l_wakeup) {
- atomic_swap_int((int *)&myreq->lr_locked, 0, &lval);
- if (lval == 2)
- /* Notify the sleeper */
- lck->l_wakeup(lck, myreq->lr_watcher);
- }
- else
- /* Give the lock to the previous request. */
- atomic_store_rel_int(&myreq->lr_locked, 0);
- }
- lu->lu_myreq->lr_active = 0;
-}
-
-void
-_lock_grant(struct lock *lck /* unused */, struct lockuser *lu)
-{
- atomic_store_rel_int(&lu->lu_watchreq->lr_locked, 3);
-}
-
-void
-_lockuser_setactive(struct lockuser *lu, int active)
-{
- lu->lu_myreq->lr_active = active;
-}
-
diff --git a/lib/libpthread/sys/lock.h b/lib/libpthread/sys/lock.h
deleted file mode 100644
index 6102a0b..0000000
--- a/lib/libpthread/sys/lock.h
+++ /dev/null
@@ -1,95 +0,0 @@
-/*
- * Copyright (c) 2001, 2003 Daniel Eischen <deischen@freebsd.org>.
- * All rights reserved.
- *
- * Redistribution and use in source and binary forms, with or without
- * modification, are permitted provided that the following conditions
- * are met:
- * 1. Redistributions of source code must retain the above copyright
- * notice, this list of conditions and the following disclaimer.
- * 2. Redistributions in binary form must reproduce the above copyright
- * notice, this list of conditions and the following disclaimer in the
- * documentation and/or other materials provided with the distribution.
- *
- * THIS SOFTWARE IS PROVIDED BY THE AUTHORS AND CONTRIBUTORS ``AS IS'' AND
- * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
- * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
- * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
- * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
- * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
- * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
- * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
- * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
- * SUCH DAMAGE.
- *
- * $FreeBSD$
- */
-
-#ifndef _LOCK_H_
-#define _LOCK_H_
-
-struct lockreq;
-struct lockuser;
-struct lock;
-
-enum lock_type {
- LCK_DEFAULT = 0x0000, /* default is FIFO spin locks */
- LCK_PRIORITY = 0x0001,
- LCK_ADAPTIVE = 0x0002 /* call user-supplied handlers */
-};
-
-typedef void lock_handler_t(struct lock *, struct lockuser *);
-
-struct lock {
- struct lockreq *l_head;
- struct lockreq *l_tail; /* only used for priority locks */
- enum lock_type l_type;
- lock_handler_t *l_wait; /* only used for adaptive locks */
- lock_handler_t *l_wakeup; /* only used for adaptive locks */
-};
-
-/* Try to make this >= CACHELINESIZE */
-struct lockreq {
- struct lockuser *lr_watcher; /* only used for priority locks */
- struct lockuser *lr_owner; /* only used for priority locks */
- volatile int lr_locked; /* lock granted = 0, busy otherwise */
- volatile int lr_active; /* non-zero if the lock is last lock for thread */
-};
-
-struct lockuser {
- struct lockreq *lu_myreq; /* request to give up/trade */
- struct lockreq *lu_watchreq; /* watch this request */
- int lu_priority; /* only used for priority locks */
- void *lu_private1; /* private{1,2} are initialized to */
- void *lu_private2; /* NULL and can be used by caller */
-#define lu_private lu_private1
-};
-
-#define _LCK_INITIALIZER(lck_req) { &lck_req, NULL, LCK_DEFAULT, \
- NULL, NULL }
-#define _LCK_REQUEST_INITIALIZER { 0, NULL, NULL, 0 }
-
-#define _LCK_BUSY(lu) ((lu)->lu_watchreq->lr_locked != 0)
-#define _LCK_ACTIVE(lu) ((lu)->lu_watchreq->lr_active != 0)
-#define _LCK_GRANTED(lu) ((lu)->lu_watchreq->lr_locked == 3)
-
-#define _LCK_SET_PRIVATE(lu, p) (lu)->lu_private = (void *)(p)
-#define _LCK_GET_PRIVATE(lu) (lu)->lu_private
-#define _LCK_SET_PRIVATE2(lu, p) (lu)->lu_private2 = (void *)(p)
-#define _LCK_GET_PRIVATE2(lu) (lu)->lu_private2
-
-void _lock_acquire(struct lock *, struct lockuser *, int);
-void _lock_destroy(struct lock *);
-void _lock_grant(struct lock *, struct lockuser *);
-int _lock_init(struct lock *, enum lock_type,
- lock_handler_t *, lock_handler_t *);
-int _lock_reinit(struct lock *, enum lock_type,
- lock_handler_t *, lock_handler_t *);
-void _lock_release(struct lock *, struct lockuser *);
-int _lockuser_init(struct lockuser *lu, void *priv);
-void _lockuser_destroy(struct lockuser *lu);
-int _lockuser_reinit(struct lockuser *lu, void *priv);
-void _lockuser_setactive(struct lockuser *lu, int active);
-
-#endif
diff --git a/lib/libpthread/sys/thr_error.c b/lib/libpthread/sys/thr_error.c
deleted file mode 100644
index 0ba34e0..0000000
--- a/lib/libpthread/sys/thr_error.c
+++ /dev/null
@@ -1,61 +0,0 @@
-/*
- * Copyright (c) 1995 John Birrell <jb@cimlogic.com.au>.
- * Copyright (c) 1994 by Chris Provenzano, proven@mit.edu
- * All rights reserved.
- *
- * Redistribution and use in source and binary forms, with or without
- * modification, are permitted provided that the following conditions
- * are met:
- * 1. Redistributions of source code must retain the above copyright
- * notice, this list of conditions and the following disclaimer.
- * 2. Redistributions in binary form must reproduce the above copyright
- * notice, this list of conditions and the following disclaimer in the
- * documentation and/or other materials provided with the distribution.
- * 3. All advertising materials mentioning features or use of this software
- * must display the following acknowledgement:
- * This product includes software developed by John Birrell
- * and Chris Provenzano.
- * 4. Neither the name of the author nor the names of any co-contributors
- * may be used to endorse or promote products derived from this software
- * without specific prior written permission.
- *
- * THIS SOFTWARE IS PROVIDED BY JOHN BIRRELL AND CONTRIBUTORS ``AS IS'' AND
- * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
- * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
- * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
- * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
- * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
- * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
- * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
- * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
- * SUCH DAMAGE.
- *
- * $FreeBSD$
- */
-#include <pthread.h>
-#include "libc_private.h"
-#include "thr_private.h"
-
-#undef errno
-extern int errno;
-
-LT10_COMPAT_DEFAULT(__error);
-
-int *
-__error(void)
-{
- struct pthread *curthread;
-
- if (__isthreaded == 0)
- return (&errno);
- else if (_kse_in_critical())
- return &(_get_curkse()->k_error);
- else {
- curthread = _get_curthread();
- if ((curthread == NULL) || (curthread == _thr_initial))
- return (&errno);
- else
- return (&curthread->error);
- }
-}
OpenPOWER on IntegriCloud