summaryrefslogtreecommitdiffstats
path: root/sys/vm/vm_contig.c
diff options
context:
space:
mode:
authoralc <alc@FreeBSD.org>2004-01-08 20:48:26 +0000
committeralc <alc@FreeBSD.org>2004-01-08 20:48:26 +0000
commit9f7878e05ab8473614dbd02b348aced4f69c8f39 (patch)
tree93969e6896f2afbaefcb25fb23dc6cfbc7fa03cb /sys/vm/vm_contig.c
parentc87939ff3b0d7321c03632991a1435fd009a8763 (diff)
downloadFreeBSD-src-9f7878e05ab8473614dbd02b348aced4f69c8f39.zip
FreeBSD-src-9f7878e05ab8473614dbd02b348aced4f69c8f39.tar.gz
- Enable recursive acquisition of the mutex synchronizing access to the
free pages queue. This is presently needed by contigmalloc1(). - Move a sanity check against attempted double allocation of two pages to the same vm object offset from vm_page_alloc() to vm_page_insert(). This provides better protection because double allocation could occur through a direct call to vm_page_insert(), such as that by vm_page_rename(). - Modify contigmalloc1() to hold the mutex synchronizing access to the free pages queue while it scans vm_page_array in search of free pages. - Correct a potential leak of pages by contigmalloc1() that I introduced in revision 1.20: We must convert all cache queue pages to free pages before we begin removing free pages from the free queue. Otherwise, if we have to restart the scan because we are unable to acquire the vm object lock that is necessary to convert a cache queue page to a free page, we leak those free pages already removed from the free queue.
Diffstat (limited to 'sys/vm/vm_contig.c')
-rw-r--r--sys/vm/vm_contig.c8
1 files changed, 6 insertions, 2 deletions
diff --git a/sys/vm/vm_contig.c b/sys/vm/vm_contig.c
index 3628b37..68adc46 100644
--- a/sys/vm/vm_contig.c
+++ b/sys/vm/vm_contig.c
@@ -168,6 +168,7 @@ contigmalloc1(
for (pass = 0; pass <= 1; pass++) {
s = splvm();
vm_page_lock_queues();
+ mtx_lock_spin(&vm_page_queue_free_mtx);
again:
/*
* Find first page in array that is free, within range,
@@ -188,6 +189,7 @@ again:
*/
if ((i == cnt.v_page_count) ||
((VM_PAGE_TO_PHYS(&pga[i]) + size) > high)) {
+ mtx_unlock_spin(&vm_page_queue_free_mtx);
again1:
if (vm_contig_launder(PQ_INACTIVE))
goto again1;
@@ -224,7 +226,9 @@ again1:
vm_page_free(m);
VM_OBJECT_UNLOCK(object);
}
- mtx_lock_spin(&vm_page_queue_free_mtx);
+ }
+ for (i = start; i < (start + size / PAGE_SIZE); i++) {
+ vm_page_t m = &pga[i];
vm_pageq_remove_nowakeup(m);
m->valid = VM_PAGE_BITS_ALL;
if (m->flags & PG_ZERO)
@@ -236,8 +240,8 @@ again1:
m->wire_count = 0;
m->busy = 0;
m->object = NULL;
- mtx_unlock_spin(&vm_page_queue_free_mtx);
}
+ mtx_unlock_spin(&vm_page_queue_free_mtx);
vm_page_unlock_queues();
/*
* We've found a contiguous chunk that meets are requirements.
OpenPOWER on IntegriCloud