diff options
Diffstat (limited to 'sys/kern/sysv_sem.c')
-rw-r--r-- | sys/kern/sysv_sem.c | 1666 |
1 files changed, 1666 insertions, 0 deletions
diff --git a/sys/kern/sysv_sem.c b/sys/kern/sysv_sem.c new file mode 100644 index 0000000..f9ff217 --- /dev/null +++ b/sys/kern/sysv_sem.c @@ -0,0 +1,1666 @@ +/*- + * Implementation of SVID semaphores + * + * Author: Daniel Boulet + * + * This software is provided ``AS IS'' without any warranties of any kind. + */ +/*- + * Copyright (c) 2003-2005 McAfee, Inc. + * All rights reserved. + * + * This software was developed for the FreeBSD Project in part by McAfee + * Research, the Security Research Division of McAfee, Inc under DARPA/SPAWAR + * contract N66001-01-C-8035 ("CBOSS"), as part of the DARPA CHATS research + * program. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND + * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE + * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL + * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS + * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) + * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT + * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY + * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF + * SUCH DAMAGE. + */ + +#include <sys/cdefs.h> +__FBSDID("$FreeBSD$"); + +#include "opt_compat.h" +#include "opt_sysvipc.h" + +#include <sys/param.h> +#include <sys/systm.h> +#include <sys/sysproto.h> +#include <sys/eventhandler.h> +#include <sys/kernel.h> +#include <sys/proc.h> +#include <sys/lock.h> +#include <sys/module.h> +#include <sys/mutex.h> +#include <sys/racct.h> +#include <sys/sem.h> +#include <sys/syscall.h> +#include <sys/syscallsubr.h> +#include <sys/sysent.h> +#include <sys/sysctl.h> +#include <sys/uio.h> +#include <sys/malloc.h> +#include <sys/jail.h> + +#include <security/mac/mac_framework.h> + +FEATURE(sysv_sem, "System V semaphores support"); + +static MALLOC_DEFINE(M_SEM, "sem", "SVID compatible semaphores"); + +#ifdef SEM_DEBUG +#define DPRINTF(a) printf a +#else +#define DPRINTF(a) +#endif + +static int seminit(void); +static int sysvsem_modload(struct module *, int, void *); +static int semunload(void); +static void semexit_myhook(void *arg, struct proc *p); +static int sysctl_sema(SYSCTL_HANDLER_ARGS); +static int semvalid(int semid, struct semid_kernel *semakptr); + +#ifndef _SYS_SYSPROTO_H_ +struct __semctl_args; +int __semctl(struct thread *td, struct __semctl_args *uap); +struct semget_args; +int semget(struct thread *td, struct semget_args *uap); +struct semop_args; +int semop(struct thread *td, struct semop_args *uap); +#endif + +static struct sem_undo *semu_alloc(struct thread *td); +static int semundo_adjust(struct thread *td, struct sem_undo **supptr, + int semid, int semseq, int semnum, int adjval); +static void semundo_clear(int semid, int semnum); + +static struct mtx sem_mtx; /* semaphore global lock */ +static struct mtx sem_undo_mtx; +static int semtot = 0; +static struct semid_kernel *sema; /* semaphore id pool */ +static struct mtx *sema_mtx; /* semaphore id pool mutexes*/ +static struct sem *sem; /* semaphore pool */ +LIST_HEAD(, sem_undo) semu_list; /* list of active undo structures */ +LIST_HEAD(, sem_undo) semu_free_list; /* list of free undo structures */ +static int *semu; /* undo structure pool */ +static eventhandler_tag semexit_tag; + +#define SEMUNDO_MTX sem_undo_mtx +#define SEMUNDO_LOCK() mtx_lock(&SEMUNDO_MTX); +#define SEMUNDO_UNLOCK() mtx_unlock(&SEMUNDO_MTX); +#define SEMUNDO_LOCKASSERT(how) mtx_assert(&SEMUNDO_MTX, (how)); + +struct sem { + u_short semval; /* semaphore value */ + pid_t sempid; /* pid of last operation */ + u_short semncnt; /* # awaiting semval > cval */ + u_short semzcnt; /* # awaiting semval = 0 */ +}; + +/* + * Undo structure (one per process) + */ +struct sem_undo { + LIST_ENTRY(sem_undo) un_next; /* ptr to next active undo structure */ + struct proc *un_proc; /* owner of this structure */ + short un_cnt; /* # of active entries */ + struct undo { + short un_adjval; /* adjust on exit values */ + short un_num; /* semaphore # */ + int un_id; /* semid */ + unsigned short un_seq; + } un_ent[1]; /* undo entries */ +}; + +/* + * Configuration parameters + */ +#ifndef SEMMNI +#define SEMMNI 50 /* # of semaphore identifiers */ +#endif +#ifndef SEMMNS +#define SEMMNS 340 /* # of semaphores in system */ +#endif +#ifndef SEMUME +#define SEMUME 50 /* max # of undo entries per process */ +#endif +#ifndef SEMMNU +#define SEMMNU 150 /* # of undo structures in system */ +#endif + +/* shouldn't need tuning */ +#ifndef SEMMSL +#define SEMMSL SEMMNS /* max # of semaphores per id */ +#endif +#ifndef SEMOPM +#define SEMOPM 100 /* max # of operations per semop call */ +#endif + +#define SEMVMX 32767 /* semaphore maximum value */ +#define SEMAEM 16384 /* adjust on exit max value */ + +/* + * Due to the way semaphore memory is allocated, we have to ensure that + * SEMUSZ is properly aligned. + */ + +#define SEM_ALIGN(bytes) (((bytes) + (sizeof(long) - 1)) & ~(sizeof(long) - 1)) + +/* actual size of an undo structure */ +#define SEMUSZ SEM_ALIGN(offsetof(struct sem_undo, un_ent[SEMUME])) + +/* + * Macro to find a particular sem_undo vector + */ +#define SEMU(ix) \ + ((struct sem_undo *)(((intptr_t)semu)+ix * seminfo.semusz)) + +/* + * semaphore info struct + */ +struct seminfo seminfo = { + SEMMNI, /* # of semaphore identifiers */ + SEMMNS, /* # of semaphores in system */ + SEMMNU, /* # of undo structures in system */ + SEMMSL, /* max # of semaphores per id */ + SEMOPM, /* max # of operations per semop call */ + SEMUME, /* max # of undo entries per process */ + SEMUSZ, /* size in bytes of undo structure */ + SEMVMX, /* semaphore maximum value */ + SEMAEM /* adjust on exit max value */ +}; + +SYSCTL_INT(_kern_ipc, OID_AUTO, semmni, CTLFLAG_RDTUN, &seminfo.semmni, 0, + "Number of semaphore identifiers"); +SYSCTL_INT(_kern_ipc, OID_AUTO, semmns, CTLFLAG_RDTUN, &seminfo.semmns, 0, + "Maximum number of semaphores in the system"); +SYSCTL_INT(_kern_ipc, OID_AUTO, semmnu, CTLFLAG_RDTUN, &seminfo.semmnu, 0, + "Maximum number of undo structures in the system"); +SYSCTL_INT(_kern_ipc, OID_AUTO, semmsl, CTLFLAG_RW, &seminfo.semmsl, 0, + "Max semaphores per id"); +SYSCTL_INT(_kern_ipc, OID_AUTO, semopm, CTLFLAG_RDTUN, &seminfo.semopm, 0, + "Max operations per semop call"); +SYSCTL_INT(_kern_ipc, OID_AUTO, semume, CTLFLAG_RDTUN, &seminfo.semume, 0, + "Max undo entries per process"); +SYSCTL_INT(_kern_ipc, OID_AUTO, semusz, CTLFLAG_RDTUN, &seminfo.semusz, 0, + "Size in bytes of undo structure"); +SYSCTL_INT(_kern_ipc, OID_AUTO, semvmx, CTLFLAG_RW, &seminfo.semvmx, 0, + "Semaphore maximum value"); +SYSCTL_INT(_kern_ipc, OID_AUTO, semaem, CTLFLAG_RW, &seminfo.semaem, 0, + "Adjust on exit max value"); +SYSCTL_PROC(_kern_ipc, OID_AUTO, sema, CTLTYPE_OPAQUE | CTLFLAG_RD, + NULL, 0, sysctl_sema, "", "Semaphore id pool"); + +static struct syscall_helper_data sem_syscalls[] = { + SYSCALL_INIT_HELPER(__semctl), + SYSCALL_INIT_HELPER(semget), + SYSCALL_INIT_HELPER(semop), +#if defined(COMPAT_FREEBSD4) || defined(COMPAT_FREEBSD5) || \ + defined(COMPAT_FREEBSD6) || defined(COMPAT_FREEBSD7) + SYSCALL_INIT_HELPER(semsys), + SYSCALL_INIT_HELPER_COMPAT(freebsd7___semctl), +#endif + SYSCALL_INIT_LAST +}; + +#ifdef COMPAT_FREEBSD32 +#include <compat/freebsd32/freebsd32.h> +#include <compat/freebsd32/freebsd32_ipc.h> +#include <compat/freebsd32/freebsd32_proto.h> +#include <compat/freebsd32/freebsd32_signal.h> +#include <compat/freebsd32/freebsd32_syscall.h> +#include <compat/freebsd32/freebsd32_util.h> + +static struct syscall_helper_data sem32_syscalls[] = { + SYSCALL32_INIT_HELPER(freebsd32_semctl), + SYSCALL32_INIT_HELPER_COMPAT(semget), + SYSCALL32_INIT_HELPER_COMPAT(semop), + SYSCALL32_INIT_HELPER(freebsd32_semsys), +#if defined(COMPAT_FREEBSD4) || defined(COMPAT_FREEBSD5) || \ + defined(COMPAT_FREEBSD6) || defined(COMPAT_FREEBSD7) + SYSCALL32_INIT_HELPER(freebsd7_freebsd32_semctl), +#endif + SYSCALL_INIT_LAST +}; +#endif + +static int +seminit(void) +{ + int i, error; + + TUNABLE_INT_FETCH("kern.ipc.semmni", &seminfo.semmni); + TUNABLE_INT_FETCH("kern.ipc.semmns", &seminfo.semmns); + TUNABLE_INT_FETCH("kern.ipc.semmnu", &seminfo.semmnu); + TUNABLE_INT_FETCH("kern.ipc.semmsl", &seminfo.semmsl); + TUNABLE_INT_FETCH("kern.ipc.semopm", &seminfo.semopm); + TUNABLE_INT_FETCH("kern.ipc.semume", &seminfo.semume); + TUNABLE_INT_FETCH("kern.ipc.semusz", &seminfo.semusz); + TUNABLE_INT_FETCH("kern.ipc.semvmx", &seminfo.semvmx); + TUNABLE_INT_FETCH("kern.ipc.semaem", &seminfo.semaem); + + sem = malloc(sizeof(struct sem) * seminfo.semmns, M_SEM, M_WAITOK); + sema = malloc(sizeof(struct semid_kernel) * seminfo.semmni, M_SEM, + M_WAITOK); + sema_mtx = malloc(sizeof(struct mtx) * seminfo.semmni, M_SEM, + M_WAITOK | M_ZERO); + semu = malloc(seminfo.semmnu * seminfo.semusz, M_SEM, M_WAITOK); + + for (i = 0; i < seminfo.semmni; i++) { + sema[i].u.sem_base = 0; + sema[i].u.sem_perm.mode = 0; + sema[i].u.sem_perm.seq = 0; +#ifdef MAC + mac_sysvsem_init(&sema[i]); +#endif + } + for (i = 0; i < seminfo.semmni; i++) + mtx_init(&sema_mtx[i], "semid", NULL, MTX_DEF); + LIST_INIT(&semu_free_list); + for (i = 0; i < seminfo.semmnu; i++) { + struct sem_undo *suptr = SEMU(i); + suptr->un_proc = NULL; + LIST_INSERT_HEAD(&semu_free_list, suptr, un_next); + } + LIST_INIT(&semu_list); + mtx_init(&sem_mtx, "sem", NULL, MTX_DEF); + mtx_init(&sem_undo_mtx, "semu", NULL, MTX_DEF); + semexit_tag = EVENTHANDLER_REGISTER(process_exit, semexit_myhook, NULL, + EVENTHANDLER_PRI_ANY); + + error = syscall_helper_register(sem_syscalls); + if (error != 0) + return (error); +#ifdef COMPAT_FREEBSD32 + error = syscall32_helper_register(sem32_syscalls); + if (error != 0) + return (error); +#endif + return (0); +} + +static int +semunload(void) +{ + int i; + + /* XXXKIB */ + if (semtot != 0) + return (EBUSY); + +#ifdef COMPAT_FREEBSD32 + syscall32_helper_unregister(sem32_syscalls); +#endif + syscall_helper_unregister(sem_syscalls); + EVENTHANDLER_DEREGISTER(process_exit, semexit_tag); +#ifdef MAC + for (i = 0; i < seminfo.semmni; i++) + mac_sysvsem_destroy(&sema[i]); +#endif + free(sem, M_SEM); + free(sema, M_SEM); + free(semu, M_SEM); + for (i = 0; i < seminfo.semmni; i++) + mtx_destroy(&sema_mtx[i]); + free(sema_mtx, M_SEM); + mtx_destroy(&sem_mtx); + mtx_destroy(&sem_undo_mtx); + return (0); +} + +static int +sysvsem_modload(struct module *module, int cmd, void *arg) +{ + int error = 0; + + switch (cmd) { + case MOD_LOAD: + error = seminit(); + if (error != 0) + semunload(); + break; + case MOD_UNLOAD: + error = semunload(); + break; + case MOD_SHUTDOWN: + break; + default: + error = EINVAL; + break; + } + return (error); +} + +static moduledata_t sysvsem_mod = { + "sysvsem", + &sysvsem_modload, + NULL +}; + +DECLARE_MODULE(sysvsem, sysvsem_mod, SI_SUB_SYSV_SEM, SI_ORDER_FIRST); +MODULE_VERSION(sysvsem, 1); + +/* + * Allocate a new sem_undo structure for a process + * (returns ptr to structure or NULL if no more room) + */ + +static struct sem_undo * +semu_alloc(struct thread *td) +{ + struct sem_undo *suptr; + + SEMUNDO_LOCKASSERT(MA_OWNED); + if ((suptr = LIST_FIRST(&semu_free_list)) == NULL) + return (NULL); + LIST_REMOVE(suptr, un_next); + LIST_INSERT_HEAD(&semu_list, suptr, un_next); + suptr->un_cnt = 0; + suptr->un_proc = td->td_proc; + return (suptr); +} + +static int +semu_try_free(struct sem_undo *suptr) +{ + + SEMUNDO_LOCKASSERT(MA_OWNED); + + if (suptr->un_cnt != 0) + return (0); + LIST_REMOVE(suptr, un_next); + LIST_INSERT_HEAD(&semu_free_list, suptr, un_next); + return (1); +} + +/* + * Adjust a particular entry for a particular proc + */ + +static int +semundo_adjust(struct thread *td, struct sem_undo **supptr, int semid, + int semseq, int semnum, int adjval) +{ + struct proc *p = td->td_proc; + struct sem_undo *suptr; + struct undo *sunptr; + int i; + + SEMUNDO_LOCKASSERT(MA_OWNED); + /* Look for and remember the sem_undo if the caller doesn't provide + it */ + + suptr = *supptr; + if (suptr == NULL) { + LIST_FOREACH(suptr, &semu_list, un_next) { + if (suptr->un_proc == p) { + *supptr = suptr; + break; + } + } + if (suptr == NULL) { + if (adjval == 0) + return(0); + suptr = semu_alloc(td); + if (suptr == NULL) + return (ENOSPC); + *supptr = suptr; + } + } + + /* + * Look for the requested entry and adjust it (delete if adjval becomes + * 0). + */ + sunptr = &suptr->un_ent[0]; + for (i = 0; i < suptr->un_cnt; i++, sunptr++) { + if (sunptr->un_id != semid || sunptr->un_num != semnum) + continue; + if (adjval != 0) { + adjval += sunptr->un_adjval; + if (adjval > seminfo.semaem || adjval < -seminfo.semaem) + return (ERANGE); + } + sunptr->un_adjval = adjval; + if (sunptr->un_adjval == 0) { + suptr->un_cnt--; + if (i < suptr->un_cnt) + suptr->un_ent[i] = + suptr->un_ent[suptr->un_cnt]; + if (suptr->un_cnt == 0) + semu_try_free(suptr); + } + return (0); + } + + /* Didn't find the right entry - create it */ + if (adjval == 0) + return (0); + if (adjval > seminfo.semaem || adjval < -seminfo.semaem) + return (ERANGE); + if (suptr->un_cnt != seminfo.semume) { + sunptr = &suptr->un_ent[suptr->un_cnt]; + suptr->un_cnt++; + sunptr->un_adjval = adjval; + sunptr->un_id = semid; + sunptr->un_num = semnum; + sunptr->un_seq = semseq; + } else + return (EINVAL); + return (0); +} + +static void +semundo_clear(int semid, int semnum) +{ + struct sem_undo *suptr, *suptr1; + struct undo *sunptr; + int i; + + SEMUNDO_LOCKASSERT(MA_OWNED); + LIST_FOREACH_SAFE(suptr, &semu_list, un_next, suptr1) { + sunptr = &suptr->un_ent[0]; + for (i = 0; i < suptr->un_cnt; i++, sunptr++) { + if (sunptr->un_id != semid) + continue; + if (semnum == -1 || sunptr->un_num == semnum) { + suptr->un_cnt--; + if (i < suptr->un_cnt) { + suptr->un_ent[i] = + suptr->un_ent[suptr->un_cnt]; + continue; + } + semu_try_free(suptr); + } + if (semnum != -1) + break; + } + } +} + +static int +semvalid(int semid, struct semid_kernel *semakptr) +{ + + return ((semakptr->u.sem_perm.mode & SEM_ALLOC) == 0 || + semakptr->u.sem_perm.seq != IPCID_TO_SEQ(semid) ? EINVAL : 0); +} + +/* + * Note that the user-mode half of this passes a union, not a pointer. + */ +#ifndef _SYS_SYSPROTO_H_ +struct __semctl_args { + int semid; + int semnum; + int cmd; + union semun *arg; +}; +#endif +int +sys___semctl(struct thread *td, struct __semctl_args *uap) +{ + struct semid_ds dsbuf; + union semun arg, semun; + register_t rval; + int error; + + switch (uap->cmd) { + case SEM_STAT: + case IPC_SET: + case IPC_STAT: + case GETALL: + case SETVAL: + case SETALL: + error = copyin(uap->arg, &arg, sizeof(arg)); + if (error) + return (error); + break; + } + + switch (uap->cmd) { + case SEM_STAT: + case IPC_STAT: + semun.buf = &dsbuf; + break; + case IPC_SET: + error = copyin(arg.buf, &dsbuf, sizeof(dsbuf)); + if (error) + return (error); + semun.buf = &dsbuf; + break; + case GETALL: + case SETALL: + semun.array = arg.array; + break; + case SETVAL: + semun.val = arg.val; + break; + } + + error = kern_semctl(td, uap->semid, uap->semnum, uap->cmd, &semun, + &rval); + if (error) + return (error); + + switch (uap->cmd) { + case SEM_STAT: + case IPC_STAT: + error = copyout(&dsbuf, arg.buf, sizeof(dsbuf)); + break; + } + + if (error == 0) + td->td_retval[0] = rval; + return (error); +} + +int +kern_semctl(struct thread *td, int semid, int semnum, int cmd, + union semun *arg, register_t *rval) +{ + u_short *array; + struct ucred *cred = td->td_ucred; + int i, error; + struct semid_ds *sbuf; + struct semid_kernel *semakptr; + struct mtx *sema_mtxp; + u_short usval, count; + int semidx; + + DPRINTF(("call to semctl(%d, %d, %d, 0x%p)\n", + semid, semnum, cmd, arg)); + if (!prison_allow(td->td_ucred, PR_ALLOW_SYSVIPC)) + return (ENOSYS); + + array = NULL; + + switch(cmd) { + case SEM_STAT: + /* + * For this command we assume semid is an array index + * rather than an IPC id. + */ + if (semid < 0 || semid >= seminfo.semmni) + return (EINVAL); + semakptr = &sema[semid]; + sema_mtxp = &sema_mtx[semid]; + mtx_lock(sema_mtxp); + if ((semakptr->u.sem_perm.mode & SEM_ALLOC) == 0) { + error = EINVAL; + goto done2; + } + if ((error = ipcperm(td, &semakptr->u.sem_perm, IPC_R))) + goto done2; +#ifdef MAC + error = mac_sysvsem_check_semctl(cred, semakptr, cmd); + if (error != 0) + goto done2; +#endif + bcopy(&semakptr->u, arg->buf, sizeof(struct semid_ds)); + *rval = IXSEQ_TO_IPCID(semid, semakptr->u.sem_perm); + mtx_unlock(sema_mtxp); + return (0); + } + + semidx = IPCID_TO_IX(semid); + if (semidx < 0 || semidx >= seminfo.semmni) + return (EINVAL); + + semakptr = &sema[semidx]; + sema_mtxp = &sema_mtx[semidx]; + if (cmd == IPC_RMID) + mtx_lock(&sem_mtx); + mtx_lock(sema_mtxp); +#ifdef MAC + error = mac_sysvsem_check_semctl(cred, semakptr, cmd); + if (error != 0) + goto done2; +#endif + + error = 0; + *rval = 0; + + switch (cmd) { + case IPC_RMID: + if ((error = semvalid(semid, semakptr)) != 0) + goto done2; + if ((error = ipcperm(td, &semakptr->u.sem_perm, IPC_M))) + goto done2; + semakptr->u.sem_perm.cuid = cred->cr_uid; + semakptr->u.sem_perm.uid = cred->cr_uid; + semakptr->u.sem_perm.mode = 0; + racct_sub_cred(semakptr->cred, RACCT_NSEM, semakptr->u.sem_nsems); + crfree(semakptr->cred); + semakptr->cred = NULL; + SEMUNDO_LOCK(); + semundo_clear(semidx, -1); + SEMUNDO_UNLOCK(); +#ifdef MAC + mac_sysvsem_cleanup(semakptr); +#endif + wakeup(semakptr); + for (i = 0; i < seminfo.semmni; i++) { + if ((sema[i].u.sem_perm.mode & SEM_ALLOC) && + sema[i].u.sem_base > semakptr->u.sem_base) + mtx_lock_flags(&sema_mtx[i], LOP_DUPOK); + } + for (i = semakptr->u.sem_base - sem; i < semtot; i++) + sem[i] = sem[i + semakptr->u.sem_nsems]; + for (i = 0; i < seminfo.semmni; i++) { + if ((sema[i].u.sem_perm.mode & SEM_ALLOC) && + sema[i].u.sem_base > semakptr->u.sem_base) { + sema[i].u.sem_base -= semakptr->u.sem_nsems; + mtx_unlock(&sema_mtx[i]); + } + } + semtot -= semakptr->u.sem_nsems; + break; + + case IPC_SET: + if ((error = semvalid(semid, semakptr)) != 0) + goto done2; + if ((error = ipcperm(td, &semakptr->u.sem_perm, IPC_M))) + goto done2; + sbuf = arg->buf; + semakptr->u.sem_perm.uid = sbuf->sem_perm.uid; + semakptr->u.sem_perm.gid = sbuf->sem_perm.gid; + semakptr->u.sem_perm.mode = (semakptr->u.sem_perm.mode & + ~0777) | (sbuf->sem_perm.mode & 0777); + semakptr->u.sem_ctime = time_second; + break; + + case IPC_STAT: + if ((error = semvalid(semid, semakptr)) != 0) + goto done2; + if ((error = ipcperm(td, &semakptr->u.sem_perm, IPC_R))) + goto done2; + bcopy(&semakptr->u, arg->buf, sizeof(struct semid_ds)); + break; + + case GETNCNT: + if ((error = semvalid(semid, semakptr)) != 0) + goto done2; + if ((error = ipcperm(td, &semakptr->u.sem_perm, IPC_R))) + goto done2; + if (semnum < 0 || semnum >= semakptr->u.sem_nsems) { + error = EINVAL; + goto done2; + } + *rval = semakptr->u.sem_base[semnum].semncnt; + break; + + case GETPID: + if ((error = semvalid(semid, semakptr)) != 0) + goto done2; + if ((error = ipcperm(td, &semakptr->u.sem_perm, IPC_R))) + goto done2; + if (semnum < 0 || semnum >= semakptr->u.sem_nsems) { + error = EINVAL; + goto done2; + } + *rval = semakptr->u.sem_base[semnum].sempid; + break; + + case GETVAL: + if ((error = semvalid(semid, semakptr)) != 0) + goto done2; + if ((error = ipcperm(td, &semakptr->u.sem_perm, IPC_R))) + goto done2; + if (semnum < 0 || semnum >= semakptr->u.sem_nsems) { + error = EINVAL; + goto done2; + } + *rval = semakptr->u.sem_base[semnum].semval; + break; + + case GETALL: + /* + * Unfortunately, callers of this function don't know + * in advance how many semaphores are in this set. + * While we could just allocate the maximum size array + * and pass the actual size back to the caller, that + * won't work for SETALL since we can't copyin() more + * data than the user specified as we may return a + * spurious EFAULT. + * + * Note that the number of semaphores in a set is + * fixed for the life of that set. The only way that + * the 'count' could change while are blocked in + * malloc() is if this semaphore set were destroyed + * and a new one created with the same index. + * However, semvalid() will catch that due to the + * sequence number unless exactly 0x8000 (or a + * multiple thereof) semaphore sets for the same index + * are created and destroyed while we are in malloc! + * + */ + count = semakptr->u.sem_nsems; + mtx_unlock(sema_mtxp); + array = malloc(sizeof(*array) * count, M_TEMP, M_WAITOK); + mtx_lock(sema_mtxp); + if ((error = semvalid(semid, semakptr)) != 0) + goto done2; + KASSERT(count == semakptr->u.sem_nsems, ("nsems changed")); + if ((error = ipcperm(td, &semakptr->u.sem_perm, IPC_R))) + goto done2; + for (i = 0; i < semakptr->u.sem_nsems; i++) + array[i] = semakptr->u.sem_base[i].semval; + mtx_unlock(sema_mtxp); + error = copyout(array, arg->array, count * sizeof(*array)); + mtx_lock(sema_mtxp); + break; + + case GETZCNT: + if ((error = semvalid(semid, semakptr)) != 0) + goto done2; + if ((error = ipcperm(td, &semakptr->u.sem_perm, IPC_R))) + goto done2; + if (semnum < 0 || semnum >= semakptr->u.sem_nsems) { + error = EINVAL; + goto done2; + } + *rval = semakptr->u.sem_base[semnum].semzcnt; + break; + + case SETVAL: + if ((error = semvalid(semid, semakptr)) != 0) + goto done2; + if ((error = ipcperm(td, &semakptr->u.sem_perm, IPC_W))) + goto done2; + if (semnum < 0 || semnum >= semakptr->u.sem_nsems) { + error = EINVAL; + goto done2; + } + if (arg->val < 0 || arg->val > seminfo.semvmx) { + error = ERANGE; + goto done2; + } + semakptr->u.sem_base[semnum].semval = arg->val; + SEMUNDO_LOCK(); + semundo_clear(semidx, semnum); + SEMUNDO_UNLOCK(); + wakeup(semakptr); + break; + + case SETALL: + /* + * See comment on GETALL for why 'count' shouldn't change + * and why we require a userland buffer. + */ + count = semakptr->u.sem_nsems; + mtx_unlock(sema_mtxp); + array = malloc(sizeof(*array) * count, M_TEMP, M_WAITOK); + error = copyin(arg->array, array, count * sizeof(*array)); + mtx_lock(sema_mtxp); + if (error) + break; + if ((error = semvalid(semid, semakptr)) != 0) + goto done2; + KASSERT(count == semakptr->u.sem_nsems, ("nsems changed")); + if ((error = ipcperm(td, &semakptr->u.sem_perm, IPC_W))) + goto done2; + for (i = 0; i < semakptr->u.sem_nsems; i++) { + usval = array[i]; + if (usval > seminfo.semvmx) { + error = ERANGE; + break; + } + semakptr->u.sem_base[i].semval = usval; + } + SEMUNDO_LOCK(); + semundo_clear(semidx, -1); + SEMUNDO_UNLOCK(); + wakeup(semakptr); + break; + + default: + error = EINVAL; + break; + } + +done2: + mtx_unlock(sema_mtxp); + if (cmd == IPC_RMID) + mtx_unlock(&sem_mtx); + if (array != NULL) + free(array, M_TEMP); + return(error); +} + +#ifndef _SYS_SYSPROTO_H_ +struct semget_args { + key_t key; + int nsems; + int semflg; +}; +#endif +int +sys_semget(struct thread *td, struct semget_args *uap) +{ + int semid, error = 0; + int key = uap->key; + int nsems = uap->nsems; + int semflg = uap->semflg; + struct ucred *cred = td->td_ucred; + + DPRINTF(("semget(0x%x, %d, 0%o)\n", key, nsems, semflg)); + if (!prison_allow(td->td_ucred, PR_ALLOW_SYSVIPC)) + return (ENOSYS); + + mtx_lock(&sem_mtx); + if (key != IPC_PRIVATE) { + for (semid = 0; semid < seminfo.semmni; semid++) { + if ((sema[semid].u.sem_perm.mode & SEM_ALLOC) && + sema[semid].u.sem_perm.key == key) + break; + } + if (semid < seminfo.semmni) { + DPRINTF(("found public key\n")); + if ((error = ipcperm(td, &sema[semid].u.sem_perm, + semflg & 0700))) { + goto done2; + } + if (nsems > 0 && sema[semid].u.sem_nsems < nsems) { + DPRINTF(("too small\n")); + error = EINVAL; + goto done2; + } + if ((semflg & IPC_CREAT) && (semflg & IPC_EXCL)) { + DPRINTF(("not exclusive\n")); + error = EEXIST; + goto done2; + } +#ifdef MAC + error = mac_sysvsem_check_semget(cred, &sema[semid]); + if (error != 0) + goto done2; +#endif + goto found; + } + } + + DPRINTF(("need to allocate the semid_kernel\n")); + if (key == IPC_PRIVATE || (semflg & IPC_CREAT)) { + if (nsems <= 0 || nsems > seminfo.semmsl) { + DPRINTF(("nsems out of range (0<%d<=%d)\n", nsems, + seminfo.semmsl)); + error = EINVAL; + goto done2; + } + if (nsems > seminfo.semmns - semtot) { + DPRINTF(( + "not enough semaphores left (need %d, got %d)\n", + nsems, seminfo.semmns - semtot)); + error = ENOSPC; + goto done2; + } + for (semid = 0; semid < seminfo.semmni; semid++) { + if ((sema[semid].u.sem_perm.mode & SEM_ALLOC) == 0) + break; + } + if (semid == seminfo.semmni) { + DPRINTF(("no more semid_kernel's available\n")); + error = ENOSPC; + goto done2; + } +#ifdef RACCT + PROC_LOCK(td->td_proc); + error = racct_add(td->td_proc, RACCT_NSEM, nsems); + PROC_UNLOCK(td->td_proc); + if (error != 0) { + error = ENOSPC; + goto done2; + } +#endif + DPRINTF(("semid %d is available\n", semid)); + mtx_lock(&sema_mtx[semid]); + KASSERT((sema[semid].u.sem_perm.mode & SEM_ALLOC) == 0, + ("Lost semaphore %d", semid)); + sema[semid].u.sem_perm.key = key; + sema[semid].u.sem_perm.cuid = cred->cr_uid; + sema[semid].u.sem_perm.uid = cred->cr_uid; + sema[semid].u.sem_perm.cgid = cred->cr_gid; + sema[semid].u.sem_perm.gid = cred->cr_gid; + sema[semid].u.sem_perm.mode = (semflg & 0777) | SEM_ALLOC; + sema[semid].cred = crhold(cred); + sema[semid].u.sem_perm.seq = + (sema[semid].u.sem_perm.seq + 1) & 0x7fff; + sema[semid].u.sem_nsems = nsems; + sema[semid].u.sem_otime = 0; + sema[semid].u.sem_ctime = time_second; + sema[semid].u.sem_base = &sem[semtot]; + semtot += nsems; + bzero(sema[semid].u.sem_base, + sizeof(sema[semid].u.sem_base[0])*nsems); +#ifdef MAC + mac_sysvsem_create(cred, &sema[semid]); +#endif + mtx_unlock(&sema_mtx[semid]); + DPRINTF(("sembase = %p, next = %p\n", + sema[semid].u.sem_base, &sem[semtot])); + } else { + DPRINTF(("didn't find it and wasn't asked to create it\n")); + error = ENOENT; + goto done2; + } + +found: + td->td_retval[0] = IXSEQ_TO_IPCID(semid, sema[semid].u.sem_perm); +done2: + mtx_unlock(&sem_mtx); + return (error); +} + +#ifndef _SYS_SYSPROTO_H_ +struct semop_args { + int semid; + struct sembuf *sops; + size_t nsops; +}; +#endif +int +sys_semop(struct thread *td, struct semop_args *uap) +{ +#define SMALL_SOPS 8 + struct sembuf small_sops[SMALL_SOPS]; + int semid = uap->semid; + size_t nsops = uap->nsops; + struct sembuf *sops; + struct semid_kernel *semakptr; + struct sembuf *sopptr = 0; + struct sem *semptr = 0; + struct sem_undo *suptr; + struct mtx *sema_mtxp; + size_t i, j, k; + int error; + int do_wakeup, do_undos; + unsigned short seq; + +#ifdef SEM_DEBUG + sops = NULL; +#endif + DPRINTF(("call to semop(%d, %p, %u)\n", semid, sops, nsops)); + + if (!prison_allow(td->td_ucred, PR_ALLOW_SYSVIPC)) + return (ENOSYS); + + semid = IPCID_TO_IX(semid); /* Convert back to zero origin */ + + if (semid < 0 || semid >= seminfo.semmni) + return (EINVAL); + + /* Allocate memory for sem_ops */ + if (nsops <= SMALL_SOPS) + sops = small_sops; + else if (nsops > seminfo.semopm) { + DPRINTF(("too many sops (max=%d, nsops=%d)\n", seminfo.semopm, + nsops)); + return (E2BIG); + } else { +#ifdef RACCT + PROC_LOCK(td->td_proc); + if (nsops > racct_get_available(td->td_proc, RACCT_NSEMOP)) { + PROC_UNLOCK(td->td_proc); + return (E2BIG); + } + PROC_UNLOCK(td->td_proc); +#endif + + sops = malloc(nsops * sizeof(*sops), M_TEMP, M_WAITOK); + } + if ((error = copyin(uap->sops, sops, nsops * sizeof(sops[0]))) != 0) { + DPRINTF(("error = %d from copyin(%p, %p, %d)\n", error, + uap->sops, sops, nsops * sizeof(sops[0]))); + if (sops != small_sops) + free(sops, M_SEM); + return (error); + } + + semakptr = &sema[semid]; + sema_mtxp = &sema_mtx[semid]; + mtx_lock(sema_mtxp); + if ((semakptr->u.sem_perm.mode & SEM_ALLOC) == 0) { + error = EINVAL; + goto done2; + } + seq = semakptr->u.sem_perm.seq; + if (seq != IPCID_TO_SEQ(uap->semid)) { + error = EINVAL; + goto done2; + } + /* + * Initial pass thru sops to see what permissions are needed. + * Also perform any checks that don't need repeating on each + * attempt to satisfy the request vector. + */ + j = 0; /* permission needed */ + do_undos = 0; + for (i = 0; i < nsops; i++) { + sopptr = &sops[i]; + if (sopptr->sem_num >= semakptr->u.sem_nsems) { + error = EFBIG; + goto done2; + } + if (sopptr->sem_flg & SEM_UNDO && sopptr->sem_op != 0) + do_undos = 1; + j |= (sopptr->sem_op == 0) ? SEM_R : SEM_A; + } + + if ((error = ipcperm(td, &semakptr->u.sem_perm, j))) { + DPRINTF(("error = %d from ipaccess\n", error)); + goto done2; + } +#ifdef MAC + error = mac_sysvsem_check_semop(td->td_ucred, semakptr, j); + if (error != 0) + goto done2; +#endif + + /* + * Loop trying to satisfy the vector of requests. + * If we reach a point where we must wait, any requests already + * performed are rolled back and we go to sleep until some other + * process wakes us up. At this point, we start all over again. + * + * This ensures that from the perspective of other tasks, a set + * of requests is atomic (never partially satisfied). + */ + for (;;) { + do_wakeup = 0; + error = 0; /* error return if necessary */ + + for (i = 0; i < nsops; i++) { + sopptr = &sops[i]; + semptr = &semakptr->u.sem_base[sopptr->sem_num]; + + DPRINTF(( + "semop: semakptr=%p, sem_base=%p, " + "semptr=%p, sem[%d]=%d : op=%d, flag=%s\n", + semakptr, semakptr->u.sem_base, semptr, + sopptr->sem_num, semptr->semval, sopptr->sem_op, + (sopptr->sem_flg & IPC_NOWAIT) ? + "nowait" : "wait")); + + if (sopptr->sem_op < 0) { + if (semptr->semval + sopptr->sem_op < 0) { + DPRINTF(("semop: can't do it now\n")); + break; + } else { + semptr->semval += sopptr->sem_op; + if (semptr->semval == 0 && + semptr->semzcnt > 0) + do_wakeup = 1; + } + } else if (sopptr->sem_op == 0) { + if (semptr->semval != 0) { + DPRINTF(("semop: not zero now\n")); + break; + } + } else if (semptr->semval + sopptr->sem_op > + seminfo.semvmx) { + error = ERANGE; + break; + } else { + if (semptr->semncnt > 0) + do_wakeup = 1; + semptr->semval += sopptr->sem_op; + } + } + + /* + * Did we get through the entire vector? + */ + if (i >= nsops) + goto done; + + /* + * No ... rollback anything that we've already done + */ + DPRINTF(("semop: rollback 0 through %d\n", i-1)); + for (j = 0; j < i; j++) + semakptr->u.sem_base[sops[j].sem_num].semval -= + sops[j].sem_op; + + /* If we detected an error, return it */ + if (error != 0) + goto done2; + + /* + * If the request that we couldn't satisfy has the + * NOWAIT flag set then return with EAGAIN. + */ + if (sopptr->sem_flg & IPC_NOWAIT) { + error = EAGAIN; + goto done2; + } + + if (sopptr->sem_op == 0) + semptr->semzcnt++; + else + semptr->semncnt++; + + DPRINTF(("semop: good night!\n")); + error = msleep(semakptr, sema_mtxp, (PZERO - 4) | PCATCH, + "semwait", 0); + DPRINTF(("semop: good morning (error=%d)!\n", error)); + /* return code is checked below, after sem[nz]cnt-- */ + + /* + * Make sure that the semaphore still exists + */ + seq = semakptr->u.sem_perm.seq; + if ((semakptr->u.sem_perm.mode & SEM_ALLOC) == 0 || + seq != IPCID_TO_SEQ(uap->semid)) { + error = EIDRM; + goto done2; + } + + /* + * Renew the semaphore's pointer after wakeup since + * during msleep sem_base may have been modified and semptr + * is not valid any more + */ + semptr = &semakptr->u.sem_base[sopptr->sem_num]; + + /* + * The semaphore is still alive. Readjust the count of + * waiting processes. + */ + if (sopptr->sem_op == 0) + semptr->semzcnt--; + else + semptr->semncnt--; + + /* + * Is it really morning, or was our sleep interrupted? + * (Delayed check of msleep() return code because we + * need to decrement sem[nz]cnt either way.) + */ + if (error != 0) { + error = EINTR; + goto done2; + } + DPRINTF(("semop: good morning!\n")); + } + +done: + /* + * Process any SEM_UNDO requests. + */ + if (do_undos) { + SEMUNDO_LOCK(); + suptr = NULL; + for (i = 0; i < nsops; i++) { + /* + * We only need to deal with SEM_UNDO's for non-zero + * op's. + */ + int adjval; + + if ((sops[i].sem_flg & SEM_UNDO) == 0) + continue; + adjval = sops[i].sem_op; + if (adjval == 0) + continue; + error = semundo_adjust(td, &suptr, semid, seq, + sops[i].sem_num, -adjval); + if (error == 0) + continue; + + /* + * Oh-Oh! We ran out of either sem_undo's or undo's. + * Rollback the adjustments to this point and then + * rollback the semaphore ups and down so we can return + * with an error with all structures restored. We + * rollback the undo's in the exact reverse order that + * we applied them. This guarantees that we won't run + * out of space as we roll things back out. + */ + for (j = 0; j < i; j++) { + k = i - j - 1; + if ((sops[k].sem_flg & SEM_UNDO) == 0) + continue; + adjval = sops[k].sem_op; + if (adjval == 0) + continue; + if (semundo_adjust(td, &suptr, semid, seq, + sops[k].sem_num, adjval) != 0) + panic("semop - can't undo undos"); + } + + for (j = 0; j < nsops; j++) + semakptr->u.sem_base[sops[j].sem_num].semval -= + sops[j].sem_op; + + DPRINTF(("error = %d from semundo_adjust\n", error)); + SEMUNDO_UNLOCK(); + goto done2; + } /* loop through the sops */ + SEMUNDO_UNLOCK(); + } /* if (do_undos) */ + + /* We're definitely done - set the sempid's and time */ + for (i = 0; i < nsops; i++) { + sopptr = &sops[i]; + semptr = &semakptr->u.sem_base[sopptr->sem_num]; + semptr->sempid = td->td_proc->p_pid; + } + semakptr->u.sem_otime = time_second; + + /* + * Do a wakeup if any semaphore was up'd whilst something was + * sleeping on it. + */ + if (do_wakeup) { + DPRINTF(("semop: doing wakeup\n")); + wakeup(semakptr); + DPRINTF(("semop: back from wakeup\n")); + } + DPRINTF(("semop: done\n")); + td->td_retval[0] = 0; +done2: + mtx_unlock(sema_mtxp); + if (sops != small_sops) + free(sops, M_SEM); + return (error); +} + +/* + * Go through the undo structures for this process and apply the adjustments to + * semaphores. + */ +static void +semexit_myhook(void *arg, struct proc *p) +{ + struct sem_undo *suptr; + struct semid_kernel *semakptr; + struct mtx *sema_mtxp; + int semid, semnum, adjval, ix; + unsigned short seq; + + /* + * Go through the chain of undo vectors looking for one + * associated with this process. + */ + SEMUNDO_LOCK(); + LIST_FOREACH(suptr, &semu_list, un_next) { + if (suptr->un_proc == p) + break; + } + if (suptr == NULL) { + SEMUNDO_UNLOCK(); + return; + } + LIST_REMOVE(suptr, un_next); + + DPRINTF(("proc @%p has undo structure with %d entries\n", p, + suptr->un_cnt)); + + /* + * If there are any active undo elements then process them. + */ + if (suptr->un_cnt > 0) { + SEMUNDO_UNLOCK(); + for (ix = 0; ix < suptr->un_cnt; ix++) { + semid = suptr->un_ent[ix].un_id; + semnum = suptr->un_ent[ix].un_num; + adjval = suptr->un_ent[ix].un_adjval; + seq = suptr->un_ent[ix].un_seq; + semakptr = &sema[semid]; + sema_mtxp = &sema_mtx[semid]; + + mtx_lock(sema_mtxp); + if ((semakptr->u.sem_perm.mode & SEM_ALLOC) == 0 || + (semakptr->u.sem_perm.seq != seq)) { + mtx_unlock(sema_mtxp); + continue; + } + if (semnum >= semakptr->u.sem_nsems) + panic("semexit - semnum out of range"); + + DPRINTF(( + "semexit: %p id=%d num=%d(adj=%d) ; sem=%d\n", + suptr->un_proc, suptr->un_ent[ix].un_id, + suptr->un_ent[ix].un_num, + suptr->un_ent[ix].un_adjval, + semakptr->u.sem_base[semnum].semval)); + + if (adjval < 0 && semakptr->u.sem_base[semnum].semval < + -adjval) + semakptr->u.sem_base[semnum].semval = 0; + else + semakptr->u.sem_base[semnum].semval += adjval; + + wakeup(semakptr); + DPRINTF(("semexit: back from wakeup\n")); + mtx_unlock(sema_mtxp); + } + SEMUNDO_LOCK(); + } + + /* + * Deallocate the undo vector. + */ + DPRINTF(("removing vector\n")); + suptr->un_proc = NULL; + suptr->un_cnt = 0; + LIST_INSERT_HEAD(&semu_free_list, suptr, un_next); + SEMUNDO_UNLOCK(); +} + +static int +sysctl_sema(SYSCTL_HANDLER_ARGS) +{ + + return (SYSCTL_OUT(req, sema, + sizeof(struct semid_kernel) * seminfo.semmni)); +} + +#if defined(COMPAT_FREEBSD4) || defined(COMPAT_FREEBSD5) || \ + defined(COMPAT_FREEBSD6) || defined(COMPAT_FREEBSD7) + +/* XXX casting to (sy_call_t *) is bogus, as usual. */ +static sy_call_t *semcalls[] = { + (sy_call_t *)freebsd7___semctl, (sy_call_t *)sys_semget, + (sy_call_t *)sys_semop +}; + +/* + * Entry point for all SEM calls. + */ +int +sys_semsys(td, uap) + struct thread *td; + /* XXX actually varargs. */ + struct semsys_args /* { + int which; + int a2; + int a3; + int a4; + int a5; + } */ *uap; +{ + int error; + + if (!prison_allow(td->td_ucred, PR_ALLOW_SYSVIPC)) + return (ENOSYS); + if (uap->which < 0 || + uap->which >= sizeof(semcalls)/sizeof(semcalls[0])) + return (EINVAL); + error = (*semcalls[uap->which])(td, &uap->a2); + return (error); +} + +#ifndef CP +#define CP(src, dst, fld) do { (dst).fld = (src).fld; } while (0) +#endif + +#ifndef _SYS_SYSPROTO_H_ +struct freebsd7___semctl_args { + int semid; + int semnum; + int cmd; + union semun_old *arg; +}; +#endif +int +freebsd7___semctl(struct thread *td, struct freebsd7___semctl_args *uap) +{ + struct semid_ds_old dsold; + struct semid_ds dsbuf; + union semun_old arg; + union semun semun; + register_t rval; + int error; + + switch (uap->cmd) { + case SEM_STAT: + case IPC_SET: + case IPC_STAT: + case GETALL: + case SETVAL: + case SETALL: + error = copyin(uap->arg, &arg, sizeof(arg)); + if (error) + return (error); + break; + } + + switch (uap->cmd) { + case SEM_STAT: + case IPC_STAT: + semun.buf = &dsbuf; + break; + case IPC_SET: + error = copyin(arg.buf, &dsold, sizeof(dsold)); + if (error) + return (error); + ipcperm_old2new(&dsold.sem_perm, &dsbuf.sem_perm); + CP(dsold, dsbuf, sem_base); + CP(dsold, dsbuf, sem_nsems); + CP(dsold, dsbuf, sem_otime); + CP(dsold, dsbuf, sem_ctime); + semun.buf = &dsbuf; + break; + case GETALL: + case SETALL: + semun.array = arg.array; + break; + case SETVAL: + semun.val = arg.val; + break; + } + + error = kern_semctl(td, uap->semid, uap->semnum, uap->cmd, &semun, + &rval); + if (error) + return (error); + + switch (uap->cmd) { + case SEM_STAT: + case IPC_STAT: + bzero(&dsold, sizeof(dsold)); + ipcperm_new2old(&dsbuf.sem_perm, &dsold.sem_perm); + CP(dsbuf, dsold, sem_base); + CP(dsbuf, dsold, sem_nsems); + CP(dsbuf, dsold, sem_otime); + CP(dsbuf, dsold, sem_ctime); + error = copyout(&dsold, arg.buf, sizeof(dsold)); + break; + } + + if (error == 0) + td->td_retval[0] = rval; + return (error); +} + +#endif /* COMPAT_FREEBSD{4,5,6,7} */ + +#ifdef COMPAT_FREEBSD32 + +int +freebsd32_semsys(struct thread *td, struct freebsd32_semsys_args *uap) +{ + +#if defined(COMPAT_FREEBSD4) || defined(COMPAT_FREEBSD5) || \ + defined(COMPAT_FREEBSD6) || defined(COMPAT_FREEBSD7) + switch (uap->which) { + case 0: + return (freebsd7_freebsd32_semctl(td, + (struct freebsd7_freebsd32_semctl_args *)&uap->a2)); + default: + return (sys_semsys(td, (struct semsys_args *)uap)); + } +#else + return (nosys(td, NULL)); +#endif +} + +#if defined(COMPAT_FREEBSD4) || defined(COMPAT_FREEBSD5) || \ + defined(COMPAT_FREEBSD6) || defined(COMPAT_FREEBSD7) +int +freebsd7_freebsd32_semctl(struct thread *td, + struct freebsd7_freebsd32_semctl_args *uap) +{ + struct semid_ds32_old dsbuf32; + struct semid_ds dsbuf; + union semun semun; + union semun32 arg; + register_t rval; + int error; + + switch (uap->cmd) { + case SEM_STAT: + case IPC_SET: + case IPC_STAT: + case GETALL: + case SETVAL: + case SETALL: + error = copyin(uap->arg, &arg, sizeof(arg)); + if (error) + return (error); + break; + } + + switch (uap->cmd) { + case SEM_STAT: + case IPC_STAT: + semun.buf = &dsbuf; + break; + case IPC_SET: + error = copyin(PTRIN(arg.buf), &dsbuf32, sizeof(dsbuf32)); + if (error) + return (error); + freebsd32_ipcperm_old_in(&dsbuf32.sem_perm, &dsbuf.sem_perm); + PTRIN_CP(dsbuf32, dsbuf, sem_base); + CP(dsbuf32, dsbuf, sem_nsems); + CP(dsbuf32, dsbuf, sem_otime); + CP(dsbuf32, dsbuf, sem_ctime); + semun.buf = &dsbuf; + break; + case GETALL: + case SETALL: + semun.array = PTRIN(arg.array); + break; + case SETVAL: + semun.val = arg.val; + break; + } + + error = kern_semctl(td, uap->semid, uap->semnum, uap->cmd, &semun, + &rval); + if (error) + return (error); + + switch (uap->cmd) { + case SEM_STAT: + case IPC_STAT: + bzero(&dsbuf32, sizeof(dsbuf32)); + freebsd32_ipcperm_old_out(&dsbuf.sem_perm, &dsbuf32.sem_perm); + PTROUT_CP(dsbuf, dsbuf32, sem_base); + CP(dsbuf, dsbuf32, sem_nsems); + CP(dsbuf, dsbuf32, sem_otime); + CP(dsbuf, dsbuf32, sem_ctime); + error = copyout(&dsbuf32, PTRIN(arg.buf), sizeof(dsbuf32)); + break; + } + + if (error == 0) + td->td_retval[0] = rval; + return (error); +} +#endif + +int +freebsd32_semctl(struct thread *td, struct freebsd32_semctl_args *uap) +{ + struct semid_ds32 dsbuf32; + struct semid_ds dsbuf; + union semun semun; + union semun32 arg; + register_t rval; + int error; + + switch (uap->cmd) { + case SEM_STAT: + case IPC_SET: + case IPC_STAT: + case GETALL: + case SETVAL: + case SETALL: + error = copyin(uap->arg, &arg, sizeof(arg)); + if (error) + return (error); + break; + } + + switch (uap->cmd) { + case SEM_STAT: + case IPC_STAT: + semun.buf = &dsbuf; + break; + case IPC_SET: + error = copyin(PTRIN(arg.buf), &dsbuf32, sizeof(dsbuf32)); + if (error) + return (error); + freebsd32_ipcperm_in(&dsbuf32.sem_perm, &dsbuf.sem_perm); + PTRIN_CP(dsbuf32, dsbuf, sem_base); + CP(dsbuf32, dsbuf, sem_nsems); + CP(dsbuf32, dsbuf, sem_otime); + CP(dsbuf32, dsbuf, sem_ctime); + semun.buf = &dsbuf; + break; + case GETALL: + case SETALL: + semun.array = PTRIN(arg.array); + break; + case SETVAL: + semun.val = arg.val; + break; + } + + error = kern_semctl(td, uap->semid, uap->semnum, uap->cmd, &semun, + &rval); + if (error) + return (error); + + switch (uap->cmd) { + case SEM_STAT: + case IPC_STAT: + bzero(&dsbuf32, sizeof(dsbuf32)); + freebsd32_ipcperm_out(&dsbuf.sem_perm, &dsbuf32.sem_perm); + PTROUT_CP(dsbuf, dsbuf32, sem_base); + CP(dsbuf, dsbuf32, sem_nsems); + CP(dsbuf, dsbuf32, sem_otime); + CP(dsbuf, dsbuf32, sem_ctime); + error = copyout(&dsbuf32, PTRIN(arg.buf), sizeof(dsbuf32)); + break; + } + + if (error == 0) + td->td_retval[0] = rval; + return (error); +} + +#endif /* COMPAT_FREEBSD32 */ |