summaryrefslogtreecommitdiffstats
path: root/sys
diff options
context:
space:
mode:
Diffstat (limited to 'sys')
-rw-r--r--sys/compat/ndis/kern_windrv.c12
-rw-r--r--sys/i386/i386/machdep.c4
-rw-r--r--sys/i386/i386/mp_machdep.c2
-rw-r--r--sys/i386/i386/sys_machdep.c117
-rw-r--r--sys/i386/i386/vm_machdep.c15
-rw-r--r--sys/i386/include/proc.h6
6 files changed, 80 insertions, 76 deletions
diff --git a/sys/compat/ndis/kern_windrv.c b/sys/compat/ndis/kern_windrv.c
index 79ec7395..00d8e8d 100644
--- a/sys/compat/ndis/kern_windrv.c
+++ b/sys/compat/ndis/kern_windrv.c
@@ -883,10 +883,10 @@ x86_oldldt(dummy)
struct gdt gtable;
uint16_t ltable;
- mtx_lock_spin(&sched_lock);
-
t = curthread;
+ mtx_lock_spin(&dt_lock);
+
/* Grab location of existing GDT. */
x86_getldt(&gtable, &ltable);
@@ -904,7 +904,7 @@ x86_oldldt(dummy)
x86_setldt(&gtable, ltable);
- mtx_unlock_spin(&sched_lock);
+ mtx_unlock_spin(&dt_lock);
return;
}
@@ -918,10 +918,10 @@ x86_newldt(dummy)
struct x86desc *l;
struct thread *t;
- mtx_lock_spin(&sched_lock);
-
t = curthread;
+ mtx_lock_spin(&dt_lock);
+
/* Grab location of existing GDT. */
x86_getldt(&gtable, &ltable);
@@ -952,7 +952,7 @@ x86_newldt(dummy)
x86_setldt(&gtable, ltable);
- mtx_unlock_spin(&sched_lock);
+ mtx_unlock_spin(&dt_lock);
/* Whew. */
diff --git a/sys/i386/i386/machdep.c b/sys/i386/i386/machdep.c
index d2c15c8..4b04ab9 100644
--- a/sys/i386/i386/machdep.c
+++ b/sys/i386/i386/machdep.c
@@ -1171,8 +1171,10 @@ exec_setregs(td, entry, stack, ps_strings)
pcb->pcb_gs = _udatasel;
load_gs(_udatasel);
+ mtx_lock_spin(&dt_lock);
if (td->td_proc->p_md.md_ldt)
user_ldt_free(td);
+ mtx_unlock_spin(&dt_lock);
bzero((char *)regs, sizeof(struct trapframe));
regs->tf_eip = entry;
@@ -1278,6 +1280,7 @@ static struct gate_descriptor idt0[NIDT];
struct gate_descriptor *idt = &idt0[0]; /* interrupt descriptor table */
union descriptor ldt[NLDT]; /* local descriptor table */
struct region_descriptor r_gdt, r_idt; /* table descriptors */
+struct mtx dt_lock; /* lock for GDT and LDT */
#if defined(I586_CPU) && !defined(NO_F00F_HACK)
extern int has_f00f_bug;
@@ -2101,6 +2104,7 @@ init386(first)
r_gdt.rd_limit = NGDT * sizeof(gdt[0]) - 1;
r_gdt.rd_base = (int) gdt;
+ mtx_init(&dt_lock, "descriptor tables", NULL, MTX_SPIN);
lgdt(&r_gdt);
pcpu_init(pc, 0, sizeof(struct pcpu));
diff --git a/sys/i386/i386/mp_machdep.c b/sys/i386/i386/mp_machdep.c
index 352bfcb..e67a99d 100644
--- a/sys/i386/i386/mp_machdep.c
+++ b/sys/i386/i386/mp_machdep.c
@@ -1388,11 +1388,9 @@ release_aps(void *dummy __unused)
if (mp_ncpus == 1)
return;
- mtx_lock_spin(&sched_lock);
atomic_store_rel_int(&aps_ready, 1);
while (smp_started == 0)
ia32_pause();
- mtx_unlock_spin(&sched_lock);
}
SYSINIT(start_aps, SI_SUB_SMP, SI_ORDER_FIRST, release_aps, NULL);
diff --git a/sys/i386/i386/sys_machdep.c b/sys/i386/i386/sys_machdep.c
index 738d77e..74a7e6c 100644
--- a/sys/i386/i386/sys_machdep.c
+++ b/sys/i386/i386/sys_machdep.c
@@ -42,6 +42,7 @@ __FBSDID("$FreeBSD$");
#include <sys/mutex.h>
#include <sys/priv.h>
#include <sys/proc.h>
+#include <sys/refcount.h>
#include <sys/smp.h>
#include <sys/sysproto.h>
@@ -115,7 +116,6 @@ sysarch(td, uap)
break;
}
- mtx_lock(&Giant);
switch(uap->op) {
case I386_GET_LDT:
error = i386_get_ldt(td, &kargs.largs);
@@ -215,7 +215,6 @@ sysarch(td, uap)
error = EINVAL;
break;
}
- mtx_unlock(&Giant);
return (error);
}
@@ -351,16 +350,19 @@ done:
/*
* Update the GDT entry pointing to the LDT to point to the LDT of the
- * current process.
- *
- * This must be called with sched_lock held. Unfortunately, we can't use a
- * mtx_assert() here because cpu_switch() calls this function after changing
- * curproc but before sched_lock's owner is updated in mi_switch().
+ * current process. Manage dt_lock holding/unholding autonomously.
*/
void
set_user_ldt(struct mdproc *mdp)
{
struct proc_ldt *pldt;
+ int dtlocked;
+
+ dtlocked = 0;
+ if (!mtx_owned(&dt_lock)) {
+ mtx_lock_spin(&dt_lock);
+ dtlocked = 1;
+ }
pldt = mdp->md_ldt;
#ifdef SMP
@@ -370,6 +372,8 @@ set_user_ldt(struct mdproc *mdp)
#endif
lldt(GSEL(GUSERLDT_SEL, SEL_KPL));
PCPU_SET(currentldt, GSEL(GUSERLDT_SEL, SEL_KPL));
+ if (dtlocked)
+ mtx_unlock_spin(&dt_lock);
}
#ifdef SMP
@@ -385,17 +389,15 @@ set_user_ldt_rv(struct thread *td)
#endif
/*
- * Must be called with either sched_lock free or held but not recursed.
- * If it does not return NULL, it will return with it owned.
+ * dt_lock must be held. Returns with dt_lock held.
*/
struct proc_ldt *
user_ldt_alloc(struct mdproc *mdp, int len)
{
struct proc_ldt *pldt, *new_ldt;
- if (mtx_owned(&sched_lock))
- mtx_unlock_spin(&sched_lock);
- mtx_assert(&sched_lock, MA_NOTOWNED);
+ mtx_assert(&dt_lock, MA_OWNED);
+ mtx_unlock_spin(&dt_lock);
MALLOC(new_ldt, struct proc_ldt *, sizeof(struct proc_ldt),
M_SUBPROC, M_WAITOK);
@@ -406,54 +408,49 @@ user_ldt_alloc(struct mdproc *mdp, int len)
FREE(new_ldt, M_SUBPROC);
return NULL;
}
- new_ldt->ldt_refcnt = 1;
+ refcount_init(&new_ldt->ldt_refcnt, 1);
new_ldt->ldt_active = 0;
- mtx_lock_spin(&sched_lock);
+ mtx_lock_spin(&dt_lock);
gdt_segs[GUSERLDT_SEL].ssd_base = (unsigned)new_ldt->ldt_base;
gdt_segs[GUSERLDT_SEL].ssd_limit = len * sizeof(union descriptor) - 1;
ssdtosd(&gdt_segs[GUSERLDT_SEL], &new_ldt->ldt_sd);
- if ((pldt = mdp->md_ldt)) {
+ if ((pldt = mdp->md_ldt) != NULL) {
if (len > pldt->ldt_len)
len = pldt->ldt_len;
bcopy(pldt->ldt_base, new_ldt->ldt_base,
len * sizeof(union descriptor));
- } else {
+ } else
bcopy(ldt, new_ldt->ldt_base, sizeof(ldt));
- }
- return new_ldt;
+
+ return (new_ldt);
}
/*
- * Must be called either with sched_lock free or held but not recursed.
- * If md_ldt is not NULL, it will return with sched_lock released.
+ * Must be called with dt_lock held.
*/
void
user_ldt_free(struct thread *td)
{
struct mdproc *mdp = &td->td_proc->p_md;
- struct proc_ldt *pldt = mdp->md_ldt;
+ struct proc_ldt *pldt;
- if (pldt == NULL)
+ mtx_assert(&dt_lock, MA_OWNED);
+ if ((pldt = mdp->md_ldt) == NULL)
return;
- if (!mtx_owned(&sched_lock))
- mtx_lock_spin(&sched_lock);
- mtx_assert(&sched_lock, MA_OWNED | MA_NOTRECURSED);
if (td == PCPU_GET(curthread)) {
lldt(_default_ldt);
PCPU_SET(currentldt, _default_ldt);
}
mdp->md_ldt = NULL;
- if (--pldt->ldt_refcnt == 0) {
- mtx_unlock_spin(&sched_lock);
+ if (refcount_release(&pldt->ldt_refcnt)) {
kmem_free(kernel_map, (vm_offset_t)pldt->ldt_base,
pldt->ldt_len * sizeof(union descriptor));
FREE(pldt, M_SUBPROC);
- } else
- mtx_unlock_spin(&sched_lock);
+ }
}
/*
@@ -468,7 +465,7 @@ i386_get_ldt(td, uap)
struct i386_ldt_args *uap;
{
int error = 0;
- struct proc_ldt *pldt = td->td_proc->p_md.md_ldt;
+ struct proc_ldt *pldt;
int nldt, num;
union descriptor *lp;
@@ -477,11 +474,14 @@ i386_get_ldt(td, uap)
uap->start, uap->num, (void *)uap->descs);
#endif
- if (pldt) {
+ mtx_lock_spin(&dt_lock);
+ if ((pldt = td->td_proc->p_md.md_ldt) != NULL) {
nldt = pldt->ldt_len;
- num = min(uap->num, nldt);
lp = &((union descriptor *)(pldt->ldt_base))[uap->start];
+ mtx_unlock_spin(&dt_lock);
+ num = min(uap->num, nldt);
} else {
+ mtx_unlock_spin(&dt_lock);
nldt = sizeof(ldt)/sizeof(ldt[0]);
num = min(uap->num, nldt);
lp = &ldt[uap->start];
@@ -531,10 +531,10 @@ i386_set_ldt(td, uap, descs)
}
if (uap->num <= 0)
return (EINVAL);
- mtx_lock_spin(&sched_lock);
- pldt = mdp->md_ldt;
- if (pldt == NULL || uap->start >= pldt->ldt_len) {
- mtx_unlock_spin(&sched_lock);
+ mtx_lock_spin(&dt_lock);
+ if ((pldt = mdp->md_ldt) == NULL ||
+ uap->start >= pldt->ldt_len) {
+ mtx_unlock_spin(&dt_lock);
return (0);
}
largest_ld = uap->start + uap->num;
@@ -543,7 +543,7 @@ i386_set_ldt(td, uap, descs)
i = largest_ld - uap->start;
bzero(&((union descriptor *)(pldt->ldt_base))[uap->start],
sizeof(union descriptor) * i);
- mtx_unlock_spin(&sched_lock);
+ mtx_unlock_spin(&dt_lock);
return (0);
}
@@ -626,15 +626,15 @@ i386_set_ldt(td, uap, descs)
if (uap->start == LDT_AUTO_ALLOC && uap->num == 1) {
/* Allocate a free slot */
- pldt = mdp->md_ldt;
- if (pldt == NULL) {
- error = i386_ldt_grow(td, NLDT + 1);
- if (error)
+ mtx_lock_spin(&dt_lock);
+ if ((pldt = mdp->md_ldt) == NULL) {
+ if ((error = i386_ldt_grow(td, NLDT + 1))) {
+ mtx_unlock_spin(&dt_lock);
return (error);
+ }
pldt = mdp->md_ldt;
}
again:
- mtx_lock_spin(&sched_lock);
/*
* start scanning a bit up to leave room for NVidia and
* Wine, which still user the "Blat" method of allocation.
@@ -646,24 +646,23 @@ again:
dp++;
}
if (i >= pldt->ldt_len) {
- mtx_unlock_spin(&sched_lock);
- error = i386_ldt_grow(td, pldt->ldt_len+1);
- if (error)
+ if ((error = i386_ldt_grow(td, pldt->ldt_len+1))) {
+ mtx_unlock_spin(&dt_lock);
return (error);
+ }
goto again;
}
uap->start = i;
error = i386_set_ldt_data(td, i, 1, descs);
- mtx_unlock_spin(&sched_lock);
+ mtx_unlock_spin(&dt_lock);
} else {
largest_ld = uap->start + uap->num;
- error = i386_ldt_grow(td, largest_ld);
- if (error == 0) {
- mtx_lock_spin(&sched_lock);
+ mtx_lock_spin(&dt_lock);
+ if (!(error = i386_ldt_grow(td, largest_ld))) {
error = i386_set_ldt_data(td, uap->start, uap->num,
descs);
- mtx_unlock_spin(&sched_lock);
}
+ mtx_unlock_spin(&dt_lock);
}
if (error == 0)
td->td_retval[0] = uap->start;
@@ -677,7 +676,7 @@ i386_set_ldt_data(struct thread *td, int start, int num,
struct mdproc *mdp = &td->td_proc->p_md;
struct proc_ldt *pldt = mdp->md_ldt;
- mtx_assert(&sched_lock, MA_OWNED);
+ mtx_assert(&dt_lock, MA_OWNED);
/* Fill in range */
bcopy(descs,
@@ -694,14 +693,15 @@ i386_ldt_grow(struct thread *td, int len)
caddr_t old_ldt_base;
int old_ldt_len;
+ mtx_assert(&dt_lock, MA_OWNED);
+
if (len > MAX_LD)
return (ENOMEM);
if (len < NLDT + 1)
len = NLDT + 1;
/* Allocate a user ldt. */
- pldt = mdp->md_ldt;
- if (!pldt || len > pldt->ldt_len) {
+ if ((pldt = mdp->md_ldt) != NULL || len > pldt->ldt_len) {
struct proc_ldt *new_ldt;
new_ldt = user_ldt_alloc(mdp, len);
@@ -709,42 +709,35 @@ i386_ldt_grow(struct thread *td, int len)
return (ENOMEM);
pldt = mdp->md_ldt;
- /* sched_lock was acquired by user_ldt_alloc. */
- if (pldt) {
+ if (pldt != NULL) {
if (new_ldt->ldt_len > pldt->ldt_len) {
old_ldt_base = pldt->ldt_base;
old_ldt_len = pldt->ldt_len;
pldt->ldt_sd = new_ldt->ldt_sd;
pldt->ldt_base = new_ldt->ldt_base;
pldt->ldt_len = new_ldt->ldt_len;
- mtx_unlock_spin(&sched_lock);
kmem_free(kernel_map, (vm_offset_t)old_ldt_base,
old_ldt_len * sizeof(union descriptor));
FREE(new_ldt, M_SUBPROC);
- mtx_lock_spin(&sched_lock);
} else {
/*
* If other threads already did the work,
* do nothing.
*/
- mtx_unlock_spin(&sched_lock);
kmem_free(kernel_map,
(vm_offset_t)new_ldt->ldt_base,
new_ldt->ldt_len * sizeof(union descriptor));
FREE(new_ldt, M_SUBPROC);
return (0);
}
- } else {
+ } else
mdp->md_ldt = pldt = new_ldt;
- }
#ifdef SMP
- mtx_unlock_spin(&sched_lock);
/* signal other cpus to reload ldt */
smp_rendezvous(NULL, (void (*)(void *))set_user_ldt_rv,
NULL, td);
#else
set_user_ldt(mdp);
- mtx_unlock_spin(&sched_lock);
#endif
}
return (0);
diff --git a/sys/i386/i386/vm_machdep.c b/sys/i386/i386/vm_machdep.c
index b0e4d21..116f1df 100644
--- a/sys/i386/i386/vm_machdep.c
+++ b/sys/i386/i386/vm_machdep.c
@@ -62,6 +62,7 @@ __FBSDID("$FreeBSD$");
#include <sys/mutex.h>
#include <sys/pioctl.h>
#include <sys/proc.h>
+#include <sys/refcount.h>
#include <sys/sf_buf.h>
#include <sys/smp.h>
#include <sys/sched.h>
@@ -158,8 +159,9 @@ cpu_fork(td1, p2, td2, flags)
struct mdproc *mdp1 = &p1->p_md;
struct proc_ldt *pldt;
- pldt = mdp1->md_ldt;
- if (pldt && pldt->ldt_refcnt > 1) {
+ mtx_lock_spin(&dt_lock);
+ if ((pldt = mdp1->md_ldt) != NULL &&
+ pldt->ldt_refcnt > 1) {
pldt = user_ldt_alloc(mdp1, pldt->ldt_len);
if (pldt == NULL)
panic("could not copy LDT");
@@ -167,6 +169,7 @@ cpu_fork(td1, p2, td2, flags)
set_user_ldt(mdp1);
user_ldt_free(td1);
}
+ mtx_unlock_spin(&dt_lock);
}
return;
}
@@ -248,10 +251,10 @@ cpu_fork(td1, p2, td2, flags)
pcb2->pcb_ext = 0;
/* Copy the LDT, if necessary. */
- mtx_lock_spin(&sched_lock);
+ mtx_lock_spin(&dt_lock);
if (mdp2->md_ldt != NULL) {
if (flags & RFMEM) {
- mdp2->md_ldt->ldt_refcnt++;
+ refcount_acquire(&mdp2->md_ldt->ldt_refcnt);
} else {
mdp2->md_ldt = user_ldt_alloc(mdp2,
mdp2->md_ldt->ldt_len);
@@ -259,7 +262,7 @@ cpu_fork(td1, p2, td2, flags)
panic("could not copy LDT");
}
}
- mtx_unlock_spin(&sched_lock);
+ mtx_unlock_spin(&dt_lock);
/* Setup to release sched_lock in fork_exit(). */
td2->td_md.md_spinlock_count = 1;
@@ -304,11 +307,13 @@ cpu_exit(struct thread *td)
* If this process has a custom LDT, release it. Reset pc->pcb_gs
* and %gs before we free it in case they refer to an LDT entry.
*/
+ mtx_lock_spin(&dt_lock);
if (td->td_proc->p_md.md_ldt) {
td->td_pcb->pcb_gs = _udatasel;
load_gs(_udatasel);
user_ldt_free(td);
}
+ mtx_unlock_spin(&dt_lock);
}
void
diff --git a/sys/i386/include/proc.h b/sys/i386/include/proc.h
index ec2e6b4..368c0a4 100644
--- a/sys/i386/include/proc.h
+++ b/sys/i386/include/proc.h
@@ -45,6 +45,8 @@ struct proc_ldt {
/*
* Machine-dependent part of the proc structure for i386.
+ * Table of MD locks:
+ * t - Descriptor tables lock
*/
struct mdthread {
int md_spinlock_count; /* (k) */
@@ -52,7 +54,7 @@ struct mdthread {
};
struct mdproc {
- struct proc_ldt *md_ldt; /* (j) per-process ldt */
+ struct proc_ldt *md_ldt; /* (t) per-process ldt */
};
#ifdef _KERNEL
@@ -61,6 +63,8 @@ void set_user_ldt(struct mdproc *);
struct proc_ldt *user_ldt_alloc(struct mdproc *, int);
void user_ldt_free(struct thread *);
+extern struct mtx dt_lock;
+
#endif /* _KERNEL */
#endif /* !_MACHINE_PROC_H_ */
OpenPOWER on IntegriCloud