summaryrefslogtreecommitdiffstats
path: root/sys
diff options
context:
space:
mode:
authorattilio <attilio@FreeBSD.org>2013-02-27 18:12:13 +0000
committerattilio <attilio@FreeBSD.org>2013-02-27 18:12:13 +0000
commit8d28f9479090da520ef392412da9129ec48f85ea (patch)
tree960a120b0bfa6e8ab680fe148604e67ac3880e7d /sys
parent5661e1fe461be2da03359fdaaf886b1a1155062b (diff)
downloadFreeBSD-src-8d28f9479090da520ef392412da9129ec48f85ea.zip
FreeBSD-src-8d28f9479090da520ef392412da9129ec48f85ea.tar.gz
Merge from vmobj-rwlock:
VM_OBJECT_LOCKED() macro is only used to implement a custom version of lock assertions right now (which likely spread out thanks to copy and paste). Remove it and implement actual assertions. Sponsored by: EMC / Isilon storage division Reviewed by: alc Tested by: pho
Diffstat (limited to 'sys')
-rw-r--r--sys/amd64/amd64/pmap.c5
-rw-r--r--sys/i386/i386/pmap.c5
-rw-r--r--sys/i386/xen/pmap.c5
-rw-r--r--sys/powerpc/aim/mmu_oea.c5
-rw-r--r--sys/powerpc/aim/mmu_oea64.c5
-rw-r--r--sys/powerpc/booke/pmap.c5
-rw-r--r--sys/sparc64/sparc64/pmap.c5
-rw-r--r--sys/vm/vm_object.h1
8 files changed, 14 insertions, 22 deletions
diff --git a/sys/amd64/amd64/pmap.c b/sys/amd64/amd64/pmap.c
index c6c62ae..f2d8967 100644
--- a/sys/amd64/amd64/pmap.c
+++ b/sys/amd64/amd64/pmap.c
@@ -3492,9 +3492,8 @@ pmap_enter(pmap_t pmap, vm_offset_t va, vm_prot_t access, vm_page_t m,
KASSERT((m->oflags & VPO_UNMANAGED) != 0 || va < kmi.clean_sva ||
va >= kmi.clean_eva,
("pmap_enter: managed mapping within the clean submap"));
- KASSERT((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) != 0 ||
- VM_OBJECT_LOCKED(m->object),
- ("pmap_enter: page %p is not busy", m));
+ if ((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) == 0)
+ VM_OBJECT_LOCK_ASSERT(m->object, MA_OWNED);
pa = VM_PAGE_TO_PHYS(m);
newpte = (pt_entry_t)(pa | PG_A | PG_V);
if ((access & VM_PROT_WRITE) != 0)
diff --git a/sys/i386/i386/pmap.c b/sys/i386/i386/pmap.c
index 5fee565..3c47a45 100644
--- a/sys/i386/i386/pmap.c
+++ b/sys/i386/i386/pmap.c
@@ -3456,9 +3456,8 @@ pmap_enter(pmap_t pmap, vm_offset_t va, vm_prot_t access, vm_page_t m,
KASSERT(va < UPT_MIN_ADDRESS || va >= UPT_MAX_ADDRESS,
("pmap_enter: invalid to pmap_enter page table pages (va: 0x%x)",
va));
- KASSERT((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) != 0 ||
- VM_OBJECT_LOCKED(m->object),
- ("pmap_enter: page %p is not busy", m));
+ if ((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) == 0)
+ VM_OBJECT_LOCK_ASSERT(m->object, MA_OWNED);
mpte = NULL;
diff --git a/sys/i386/xen/pmap.c b/sys/i386/xen/pmap.c
index a8f11a4..4989e07 100644
--- a/sys/i386/xen/pmap.c
+++ b/sys/i386/xen/pmap.c
@@ -2666,9 +2666,8 @@ pmap_enter(pmap_t pmap, vm_offset_t va, vm_prot_t access, vm_page_t m,
KASSERT(va < UPT_MIN_ADDRESS || va >= UPT_MAX_ADDRESS,
("pmap_enter: invalid to pmap_enter page table pages (va: 0x%x)",
va));
- KASSERT((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) != 0 ||
- VM_OBJECT_LOCKED(m->object),
- ("pmap_enter: page %p is not busy", m));
+ if ((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) == 0)
+ VM_OBJECT_LOCK_ASSERT(m->object, MA_OWNED);
mpte = NULL;
diff --git a/sys/powerpc/aim/mmu_oea.c b/sys/powerpc/aim/mmu_oea.c
index 24350c4..e0af4e5 100644
--- a/sys/powerpc/aim/mmu_oea.c
+++ b/sys/powerpc/aim/mmu_oea.c
@@ -1121,9 +1121,8 @@ moea_enter_locked(pmap_t pmap, vm_offset_t va, vm_page_t m, vm_prot_t prot,
if (pmap_bootstrapped)
rw_assert(&pvh_global_lock, RA_WLOCKED);
PMAP_LOCK_ASSERT(pmap, MA_OWNED);
- KASSERT((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) != 0 ||
- VM_OBJECT_LOCKED(m->object),
- ("moea_enter_locked: page %p is not busy", m));
+ if ((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) == 0)
+ VM_OBJECT_LOCK_ASSERT(m->object, MA_OWNED);
/* XXX change the pvo head for fake pages */
if ((m->oflags & VPO_UNMANAGED) != 0) {
diff --git a/sys/powerpc/aim/mmu_oea64.c b/sys/powerpc/aim/mmu_oea64.c
index 0d77b57..296b546 100644
--- a/sys/powerpc/aim/mmu_oea64.c
+++ b/sys/powerpc/aim/mmu_oea64.c
@@ -1183,9 +1183,8 @@ moea64_enter(mmu_t mmu, pmap_t pmap, vm_offset_t va, vm_page_t m,
pvo_flags = PVO_MANAGED;
}
- KASSERT((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) != 0 ||
- VM_OBJECT_LOCKED(m->object),
- ("moea64_enter: page %p is not busy", m));
+ if ((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) == 0)
+ VM_OBJECT_LOCK_ASSERT(m->object, MA_OWNED);
/* XXX change the pvo head for fake pages */
if ((m->oflags & VPO_UNMANAGED) != 0) {
diff --git a/sys/powerpc/booke/pmap.c b/sys/powerpc/booke/pmap.c
index 5172009..8e0b76f 100644
--- a/sys/powerpc/booke/pmap.c
+++ b/sys/powerpc/booke/pmap.c
@@ -1560,9 +1560,8 @@ mmu_booke_enter_locked(mmu_t mmu, pmap_t pmap, vm_offset_t va, vm_page_t m,
KASSERT((va <= VM_MAXUSER_ADDRESS),
("mmu_booke_enter_locked: user pmap, non user va"));
}
- KASSERT((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) != 0 ||
- VM_OBJECT_LOCKED(m->object),
- ("mmu_booke_enter_locked: page %p is not busy", m));
+ if ((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) == 0)
+ VM_OBJECT_LOCK_ASSERT(m->object, MA_OWNED);
PMAP_LOCK_ASSERT(pmap, MA_OWNED);
diff --git a/sys/sparc64/sparc64/pmap.c b/sys/sparc64/sparc64/pmap.c
index 08f008c..f689b17 100644
--- a/sys/sparc64/sparc64/pmap.c
+++ b/sys/sparc64/sparc64/pmap.c
@@ -1494,9 +1494,8 @@ pmap_enter_locked(pmap_t pm, vm_offset_t va, vm_page_t m, vm_prot_t prot,
rw_assert(&tte_list_global_lock, RA_WLOCKED);
PMAP_LOCK_ASSERT(pm, MA_OWNED);
- KASSERT((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) != 0 ||
- VM_OBJECT_LOCKED(m->object),
- ("pmap_enter_locked: page %p is not busy", m));
+ if ((m->oflags & (VPO_UNMANAGED | VPO_BUSY)) == 0)
+ VM_OBJECT_LOCK_ASSERT(m->object, MA_OWNED);
PMAP_STATS_INC(pmap_nenter);
pa = VM_PAGE_TO_PHYS(m);
diff --git a/sys/vm/vm_object.h b/sys/vm/vm_object.h
index e085929..76caa75 100644
--- a/sys/vm/vm_object.h
+++ b/sys/vm/vm_object.h
@@ -206,7 +206,6 @@ extern struct vm_object kmem_object_store;
#define VM_OBJECT_LOCK(object) mtx_lock(&(object)->mtx)
#define VM_OBJECT_LOCK_ASSERT(object, type) \
mtx_assert(&(object)->mtx, (type))
-#define VM_OBJECT_LOCKED(object) mtx_owned(&(object)->mtx)
#define VM_OBJECT_SLEEP(object, wchan, pri, wmesg, timo) \
msleep((wchan), &(object)->mtx, (pri), \
(wmesg), (timo))
OpenPOWER on IntegriCloud