summaryrefslogtreecommitdiffstats
path: root/sys/vm/vm_fault.c
diff options
context:
space:
mode:
authorattilio <attilio@FreeBSD.org>2013-02-20 12:03:20 +0000
committerattilio <attilio@FreeBSD.org>2013-02-20 12:03:20 +0000
commit15bf891afe5ecb096114725fc8e6dc1cc3ef70d6 (patch)
tree430a1525becbd2674f05fbcf81b84fccc4aa7865 /sys/vm/vm_fault.c
parent1f1e13ca0304c5d3cab0d4c118678ec546f935bc (diff)
downloadFreeBSD-src-15bf891afe5ecb096114725fc8e6dc1cc3ef70d6.zip
FreeBSD-src-15bf891afe5ecb096114725fc8e6dc1cc3ef70d6.tar.gz
Rename VM_OBJECT_LOCK(), VM_OBJECT_UNLOCK() and VM_OBJECT_TRYLOCK() to
their "write" versions. Sponsored by: EMC / Isilon storage division
Diffstat (limited to 'sys/vm/vm_fault.c')
-rw-r--r--sys/vm/vm_fault.c78
1 files changed, 39 insertions, 39 deletions
diff --git a/sys/vm/vm_fault.c b/sys/vm/vm_fault.c
index 9517cd3..23a0d6d 100644
--- a/sys/vm/vm_fault.c
+++ b/sys/vm/vm_fault.c
@@ -163,14 +163,14 @@ unlock_and_deallocate(struct faultstate *fs)
{
vm_object_pip_wakeup(fs->object);
- VM_OBJECT_UNLOCK(fs->object);
+ VM_OBJECT_WUNLOCK(fs->object);
if (fs->object != fs->first_object) {
- VM_OBJECT_LOCK(fs->first_object);
+ VM_OBJECT_WLOCK(fs->first_object);
vm_page_lock(fs->first_m);
vm_page_free(fs->first_m);
vm_page_unlock(fs->first_m);
vm_object_pip_wakeup(fs->first_object);
- VM_OBJECT_UNLOCK(fs->first_object);
+ VM_OBJECT_WUNLOCK(fs->first_object);
fs->first_m = NULL;
}
vm_object_deallocate(fs->first_object);
@@ -290,7 +290,7 @@ RetryFault:;
* truncation operations) during I/O. This must be done after
* obtaining the vnode lock in order to avoid possible deadlocks.
*/
- VM_OBJECT_LOCK(fs.first_object);
+ VM_OBJECT_WLOCK(fs.first_object);
vm_object_reference_locked(fs.first_object);
vm_object_pip_add(fs.first_object, 1);
@@ -363,17 +363,17 @@ RetryFault:;
vm_page_aflag_set(fs.m, PGA_REFERENCED);
vm_page_unlock(fs.m);
if (fs.object != fs.first_object) {
- if (!VM_OBJECT_TRYLOCK(
+ if (!VM_OBJECT_TRYWLOCK(
fs.first_object)) {
- VM_OBJECT_UNLOCK(fs.object);
- VM_OBJECT_LOCK(fs.first_object);
- VM_OBJECT_LOCK(fs.object);
+ VM_OBJECT_WUNLOCK(fs.object);
+ VM_OBJECT_WLOCK(fs.first_object);
+ VM_OBJECT_WLOCK(fs.object);
}
vm_page_lock(fs.first_m);
vm_page_free(fs.first_m);
vm_page_unlock(fs.first_m);
vm_object_pip_wakeup(fs.first_object);
- VM_OBJECT_UNLOCK(fs.first_object);
+ VM_OBJECT_WUNLOCK(fs.first_object);
fs.first_m = NULL;
}
unlock_map(&fs);
@@ -383,7 +383,7 @@ RetryFault:;
"vmpfw");
}
vm_object_pip_wakeup(fs.object);
- VM_OBJECT_UNLOCK(fs.object);
+ VM_OBJECT_WUNLOCK(fs.object);
PCPU_INC(cnt.v_intrans);
vm_object_deallocate(fs.first_object);
goto RetryFault;
@@ -646,12 +646,12 @@ vnode_locked:
*/
if (fs.object != fs.first_object) {
vm_object_pip_wakeup(fs.object);
- VM_OBJECT_UNLOCK(fs.object);
+ VM_OBJECT_WUNLOCK(fs.object);
fs.object = fs.first_object;
fs.pindex = fs.first_pindex;
fs.m = fs.first_m;
- VM_OBJECT_LOCK(fs.object);
+ VM_OBJECT_WLOCK(fs.object);
}
fs.first_m = NULL;
@@ -669,11 +669,11 @@ vnode_locked:
} else {
KASSERT(fs.object != next_object,
("object loop %p", next_object));
- VM_OBJECT_LOCK(next_object);
+ VM_OBJECT_WLOCK(next_object);
vm_object_pip_add(next_object, 1);
if (fs.object != fs.first_object)
vm_object_pip_wakeup(fs.object);
- VM_OBJECT_UNLOCK(fs.object);
+ VM_OBJECT_WUNLOCK(fs.object);
fs.object = next_object;
}
}
@@ -725,7 +725,7 @@ vnode_locked:
*/
((fs.object->type == OBJT_DEFAULT) ||
(fs.object->type == OBJT_SWAP)) &&
- (is_first_object_locked = VM_OBJECT_TRYLOCK(fs.first_object)) &&
+ (is_first_object_locked = VM_OBJECT_TRYWLOCK(fs.first_object)) &&
/*
* We don't chase down the shadow chain
*/
@@ -774,7 +774,7 @@ vnode_locked:
* conditional
*/
vm_object_pip_wakeup(fs.object);
- VM_OBJECT_UNLOCK(fs.object);
+ VM_OBJECT_WUNLOCK(fs.object);
/*
* Only use the new page below...
*/
@@ -782,7 +782,7 @@ vnode_locked:
fs.pindex = fs.first_pindex;
fs.m = fs.first_m;
if (!is_first_object_locked)
- VM_OBJECT_LOCK(fs.object);
+ VM_OBJECT_WLOCK(fs.object);
PCPU_INC(cnt.v_cow_faults);
curthread->td_cow++;
} else {
@@ -903,7 +903,7 @@ vnode_locked:
*/
KASSERT(fs.m->valid == VM_PAGE_BITS_ALL,
("vm_fault: page %p partially invalid", fs.m));
- VM_OBJECT_UNLOCK(fs.object);
+ VM_OBJECT_WUNLOCK(fs.object);
/*
* Put this page into the physical map. We had to do the unlock above
@@ -914,7 +914,7 @@ vnode_locked:
pmap_enter(fs.map->pmap, vaddr, fault_type, fs.m, prot, wired);
if ((fault_flags & VM_FAULT_CHANGE_WIRING) == 0 && wired == 0)
vm_fault_prefault(fs.map->pmap, vaddr, fs.entry);
- VM_OBJECT_LOCK(fs.object);
+ VM_OBJECT_WLOCK(fs.object);
vm_page_lock(fs.m);
/*
@@ -963,10 +963,10 @@ vm_fault_cache_behind(const struct faultstate *fs, int distance)
VM_OBJECT_LOCK_ASSERT(object, RA_WLOCKED);
first_object = fs->first_object;
if (first_object != object) {
- if (!VM_OBJECT_TRYLOCK(first_object)) {
- VM_OBJECT_UNLOCK(object);
- VM_OBJECT_LOCK(first_object);
- VM_OBJECT_LOCK(object);
+ if (!VM_OBJECT_TRYWLOCK(first_object)) {
+ VM_OBJECT_WUNLOCK(object);
+ VM_OBJECT_WLOCK(first_object);
+ VM_OBJECT_WLOCK(object);
}
}
/* Neither fictitious nor unmanaged pages can be cached. */
@@ -999,7 +999,7 @@ vm_fault_cache_behind(const struct faultstate *fs, int distance)
}
}
if (first_object != object)
- VM_OBJECT_UNLOCK(first_object);
+ VM_OBJECT_WUNLOCK(first_object);
}
/*
@@ -1044,28 +1044,28 @@ vm_fault_prefault(pmap_t pmap, vm_offset_t addra, vm_map_entry_t entry)
pindex = ((addr - entry->start) + entry->offset) >> PAGE_SHIFT;
lobject = object;
- VM_OBJECT_LOCK(lobject);
+ VM_OBJECT_WLOCK(lobject);
while ((m = vm_page_lookup(lobject, pindex)) == NULL &&
lobject->type == OBJT_DEFAULT &&
(backing_object = lobject->backing_object) != NULL) {
KASSERT((lobject->backing_object_offset & PAGE_MASK) ==
0, ("vm_fault_prefault: unaligned object offset"));
pindex += lobject->backing_object_offset >> PAGE_SHIFT;
- VM_OBJECT_LOCK(backing_object);
- VM_OBJECT_UNLOCK(lobject);
+ VM_OBJECT_WLOCK(backing_object);
+ VM_OBJECT_WUNLOCK(lobject);
lobject = backing_object;
}
/*
* give-up when a page is not in memory
*/
if (m == NULL) {
- VM_OBJECT_UNLOCK(lobject);
+ VM_OBJECT_WUNLOCK(lobject);
break;
}
if (m->valid == VM_PAGE_BITS_ALL &&
(m->flags & PG_FICTITIOUS) == 0)
pmap_enter_quick(pmap, addr, m, entry->protection);
- VM_OBJECT_UNLOCK(lobject);
+ VM_OBJECT_WUNLOCK(lobject);
}
}
@@ -1257,7 +1257,7 @@ vm_fault_copy_entry(vm_map_t dst_map, vm_map_t src_map,
dst_object->pg_color = atop(dst_entry->start);
#endif
- VM_OBJECT_LOCK(dst_object);
+ VM_OBJECT_WLOCK(dst_object);
KASSERT(upgrade || dst_entry->object.vm_object == NULL,
("vm_fault_copy_entry: vm_object not NULL"));
dst_entry->object.vm_object = dst_object;
@@ -1307,9 +1307,9 @@ vm_fault_copy_entry(vm_map_t dst_map, vm_map_t src_map,
dst_m = vm_page_alloc(dst_object, dst_pindex,
VM_ALLOC_NORMAL);
if (dst_m == NULL) {
- VM_OBJECT_UNLOCK(dst_object);
+ VM_OBJECT_WUNLOCK(dst_object);
VM_WAIT;
- VM_OBJECT_LOCK(dst_object);
+ VM_OBJECT_WLOCK(dst_object);
}
} while (dst_m == NULL);
@@ -1318,7 +1318,7 @@ vm_fault_copy_entry(vm_map_t dst_map, vm_map_t src_map,
* (Because the source is wired down, the page will be in
* memory.)
*/
- VM_OBJECT_LOCK(src_object);
+ VM_OBJECT_WLOCK(src_object);
object = src_object;
pindex = src_pindex + dst_pindex;
while ((src_m = vm_page_lookup(object, pindex)) == NULL &&
@@ -1327,18 +1327,18 @@ vm_fault_copy_entry(vm_map_t dst_map, vm_map_t src_map,
/*
* Allow fallback to backing objects if we are reading.
*/
- VM_OBJECT_LOCK(backing_object);
+ VM_OBJECT_WLOCK(backing_object);
pindex += OFF_TO_IDX(object->backing_object_offset);
- VM_OBJECT_UNLOCK(object);
+ VM_OBJECT_WUNLOCK(object);
object = backing_object;
}
if (src_m == NULL)
panic("vm_fault_copy_wired: page missing");
pmap_copy_page(src_m, dst_m);
- VM_OBJECT_UNLOCK(object);
+ VM_OBJECT_WUNLOCK(object);
dst_m->valid = VM_PAGE_BITS_ALL;
dst_m->dirty = VM_PAGE_BITS_ALL;
- VM_OBJECT_UNLOCK(dst_object);
+ VM_OBJECT_WUNLOCK(dst_object);
/*
* Enter it in the pmap. If a wired, copy-on-write
@@ -1350,7 +1350,7 @@ vm_fault_copy_entry(vm_map_t dst_map, vm_map_t src_map,
/*
* Mark it no longer busy, and put it on the active list.
*/
- VM_OBJECT_LOCK(dst_object);
+ VM_OBJECT_WLOCK(dst_object);
if (upgrade) {
vm_page_lock(src_m);
@@ -1367,7 +1367,7 @@ vm_fault_copy_entry(vm_map_t dst_map, vm_map_t src_map,
}
vm_page_wakeup(dst_m);
}
- VM_OBJECT_UNLOCK(dst_object);
+ VM_OBJECT_WUNLOCK(dst_object);
if (upgrade) {
dst_entry->eflags &= ~(MAP_ENTRY_COW | MAP_ENTRY_NEEDS_COPY);
vm_object_deallocate(src_object);
OpenPOWER on IntegriCloud