summaryrefslogtreecommitdiffstats
path: root/sys/vm/vm_contig.c
diff options
context:
space:
mode:
authoralc <alc@FreeBSD.org>2007-06-16 04:57:06 +0000
committeralc <alc@FreeBSD.org>2007-06-16 04:57:06 +0000
commita8415c5a0d7d39fe501a93e2f31b2532cf6dfd91 (patch)
treed4adea7a0cd4062bbcbfbb6bc1a5a30eecf59d97 /sys/vm/vm_contig.c
parent98cd3abe8b02f0ae3d389eb9e6ecbf4bf39032f6 (diff)
downloadFreeBSD-src-a8415c5a0d7d39fe501a93e2f31b2532cf6dfd91.zip
FreeBSD-src-a8415c5a0d7d39fe501a93e2f31b2532cf6dfd91.tar.gz
Enable the new physical memory allocator.
This allocator uses a binary buddy system with a twist. First and foremost, this allocator is required to support the implementation of superpages. As a side effect, it enables a more robust implementation of contigmalloc(9). Moreover, this reimplementation of contigmalloc(9) eliminates the acquisition of Giant by contigmalloc(..., M_NOWAIT, ...). The twist is that this allocator tries to reduce the number of TLB misses incurred by accesses through a direct map to small, UMA-managed objects and page table pages. Roughly speaking, the physical pages that are allocated for such purposes are clustered together in the physical address space. The performance benefits vary. In the most extreme case, a uniprocessor kernel running on an Opteron, I measured an 18% reduction in system time during a buildworld. This allocator does not implement page coloring. The reason is that superpages have much the same effect. The contiguous physical memory allocation necessary for a superpage is inherently colored. Finally, the one caveat is that this allocator does not effectively support prezeroed pages. I hope this is temporary. On i386, this is a slight pessimization. However, on amd64, the beneficial effects of the direct-map optimization outweigh the ill effects. I speculate that this is true in general of machines with a direct map. Approved by: re
Diffstat (limited to 'sys/vm/vm_contig.c')
-rw-r--r--sys/vm/vm_contig.c416
1 files changed, 42 insertions, 374 deletions
diff --git a/sys/vm/vm_contig.c b/sys/vm/vm_contig.c
index bd9233b..955df30 100644
--- a/sys/vm/vm_contig.c
+++ b/sys/vm/vm_contig.c
@@ -84,6 +84,7 @@ __FBSDID("$FreeBSD$");
#include <vm/vm_page.h>
#include <vm/vm_pageout.h>
#include <vm/vm_pager.h>
+#include <vm/vm_phys.h>
#include <vm/vm_extern.h>
static int
@@ -165,191 +166,6 @@ vm_contig_launder(int queue)
return (FALSE);
}
-/*
- * This interface is for merging with malloc() someday.
- * Even if we never implement compaction so that contiguous allocation
- * works after initialization time, malloc()'s data structures are good
- * for statistics and for allocations of less than a page.
- */
-static void *
-contigmalloc1(
- unsigned long size, /* should be size_t here and for malloc() */
- struct malloc_type *type,
- int flags,
- vm_paddr_t low,
- vm_paddr_t high,
- unsigned long alignment,
- unsigned long boundary,
- vm_map_t map)
-{
- int i, start;
- vm_paddr_t phys;
- vm_object_t object;
- vm_offset_t addr, tmp_addr;
- int pass, pqtype;
- int inactl, actl, inactmax, actmax;
- vm_page_t pga = vm_page_array;
-
- size = round_page(size);
- if (size == 0)
- panic("contigmalloc1: size must not be 0");
- if ((alignment & (alignment - 1)) != 0)
- panic("contigmalloc1: alignment must be a power of 2");
- if ((boundary & (boundary - 1)) != 0)
- panic("contigmalloc1: boundary must be a power of 2");
-
- start = 0;
- for (pass = 2; pass >= 0; pass--) {
- vm_page_lock_queues();
-again0:
- mtx_lock(&vm_page_queue_free_mtx);
-again:
- /*
- * Find first page in array that is free, within range,
- * aligned, and such that the boundary won't be crossed.
- */
- for (i = start; i < cnt.v_page_count; i++) {
- phys = VM_PAGE_TO_PHYS(&pga[i]);
- pqtype = pga[i].queue - pga[i].pc;
- if (((pqtype == PQ_FREE) || (pqtype == PQ_CACHE)) &&
- (phys >= low) && (phys < high) &&
- ((phys & (alignment - 1)) == 0) &&
- (((phys ^ (phys + size - 1)) & ~(boundary - 1)) == 0))
- break;
- }
-
- /*
- * If the above failed or we will exceed the upper bound, fail.
- */
- if ((i == cnt.v_page_count) ||
- ((VM_PAGE_TO_PHYS(&pga[i]) + size) > high)) {
- mtx_unlock(&vm_page_queue_free_mtx);
- /*
- * Instead of racing to empty the inactive/active
- * queues, give up, even with more left to free,
- * if we try more than the initial amount of pages.
- *
- * There's no point attempting this on the last pass.
- */
- if (pass > 0) {
- inactl = actl = 0;
- inactmax = vm_page_queues[PQ_INACTIVE].lcnt;
- actmax = vm_page_queues[PQ_ACTIVE].lcnt;
-again1:
- if (inactl < inactmax &&
- vm_contig_launder(PQ_INACTIVE)) {
- inactl++;
- goto again1;
- }
- if (actl < actmax &&
- vm_contig_launder(PQ_ACTIVE)) {
- actl++;
- goto again1;
- }
- }
- vm_page_unlock_queues();
- continue;
- }
- start = i;
-
- /*
- * Check successive pages for contiguous and free.
- */
- for (i = start + 1; i < (start + size / PAGE_SIZE); i++) {
- pqtype = pga[i].queue - pga[i].pc;
- if ((VM_PAGE_TO_PHYS(&pga[i]) !=
- (VM_PAGE_TO_PHYS(&pga[i - 1]) + PAGE_SIZE)) ||
- ((pqtype != PQ_FREE) && (pqtype != PQ_CACHE))) {
- start++;
- goto again;
- }
- }
- mtx_unlock(&vm_page_queue_free_mtx);
- for (i = start; i < (start + size / PAGE_SIZE); i++) {
- vm_page_t m = &pga[i];
-
- if (VM_PAGE_INQUEUE1(m, PQ_CACHE)) {
- if (m->hold_count != 0) {
- start++;
- goto again0;
- }
- object = m->object;
- if (!VM_OBJECT_TRYLOCK(object)) {
- start++;
- goto again0;
- }
- if ((m->oflags & VPO_BUSY) || m->busy != 0) {
- VM_OBJECT_UNLOCK(object);
- start++;
- goto again0;
- }
- vm_page_free(m);
- VM_OBJECT_UNLOCK(object);
- }
- }
- mtx_lock(&vm_page_queue_free_mtx);
- for (i = start; i < (start + size / PAGE_SIZE); i++) {
- pqtype = pga[i].queue - pga[i].pc;
- if (pqtype != PQ_FREE) {
- start++;
- goto again;
- }
- }
- for (i = start; i < (start + size / PAGE_SIZE); i++) {
- vm_page_t m = &pga[i];
- vm_pageq_remove_nowakeup(m);
- m->valid = VM_PAGE_BITS_ALL;
- if (m->flags & PG_ZERO)
- vm_page_zero_count--;
- /* Don't clear the PG_ZERO flag, we'll need it later. */
- m->flags = PG_UNMANAGED | (m->flags & PG_ZERO);
- KASSERT(m->dirty == 0,
- ("contigmalloc1: page %p was dirty", m));
- m->wire_count = 0;
- m->busy = 0;
- }
- mtx_unlock(&vm_page_queue_free_mtx);
- vm_page_unlock_queues();
- /*
- * We've found a contiguous chunk that meets are requirements.
- * Allocate kernel VM, unfree and assign the physical pages to
- * it and return kernel VM pointer.
- */
- vm_map_lock(map);
- if (vm_map_findspace(map, vm_map_min(map), size, &addr) !=
- KERN_SUCCESS) {
- /*
- * XXX We almost never run out of kernel virtual
- * space, so we don't make the allocated memory
- * above available.
- */
- vm_map_unlock(map);
- return (NULL);
- }
- vm_object_reference(kernel_object);
- vm_map_insert(map, kernel_object, addr - VM_MIN_KERNEL_ADDRESS,
- addr, addr + size, VM_PROT_ALL, VM_PROT_ALL, 0);
- vm_map_unlock(map);
-
- tmp_addr = addr;
- VM_OBJECT_LOCK(kernel_object);
- for (i = start; i < (start + size / PAGE_SIZE); i++) {
- vm_page_t m = &pga[i];
- vm_page_insert(m, kernel_object,
- OFF_TO_IDX(tmp_addr - VM_MIN_KERNEL_ADDRESS));
- if ((flags & M_ZERO) && !(m->flags & PG_ZERO))
- pmap_zero_page(m);
- tmp_addr += PAGE_SIZE;
- }
- VM_OBJECT_UNLOCK(kernel_object);
- vm_map_wire(map, addr, addr + size,
- VM_MAP_WIRE_SYSTEM|VM_MAP_WIRE_NOHOLES);
-
- return ((void *)addr);
- }
- return (NULL);
-}
-
static void
vm_page_release_contigl(vm_page_t m, vm_pindex_t count)
{
@@ -367,173 +183,6 @@ vm_page_release_contig(vm_page_t m, vm_pindex_t count)
vm_page_unlock_queues();
}
-static int
-vm_contig_unqueue_free(vm_page_t m)
-{
- int error = 0;
-
- mtx_lock(&vm_page_queue_free_mtx);
- if ((m->queue - m->pc) == PQ_FREE)
- vm_pageq_remove_nowakeup(m);
- else
- error = EAGAIN;
- mtx_unlock(&vm_page_queue_free_mtx);
- if (error)
- return (error);
- m->valid = VM_PAGE_BITS_ALL;
- if (m->flags & PG_ZERO)
- vm_page_zero_count--;
- /* Don't clear the PG_ZERO flag; we'll need it later. */
- m->flags = PG_UNMANAGED | (m->flags & PG_ZERO);
- m->oflags = 0;
- KASSERT(m->dirty == 0,
- ("contigmalloc2: page %p was dirty", m));
- m->wire_count = 0;
- m->busy = 0;
- return (error);
-}
-
-vm_page_t
-vm_page_alloc_contig(vm_pindex_t npages, vm_paddr_t low, vm_paddr_t high,
- vm_offset_t alignment, vm_offset_t boundary)
-{
- vm_object_t object;
- vm_offset_t size;
- vm_paddr_t phys;
- vm_page_t pga = vm_page_array;
- static vm_pindex_t np = 0;
- static vm_pindex_t start = 0;
- vm_pindex_t startl = 0;
- int i, pass, pqtype;
-
- size = npages << PAGE_SHIFT;
- if (size == 0)
- panic("vm_page_alloc_contig: size must not be 0");
- if ((alignment & (alignment - 1)) != 0)
- panic("vm_page_alloc_contig: alignment must be a power of 2");
- if ((boundary & (boundary - 1)) != 0)
- panic("vm_page_alloc_contig: boundary must be a power of 2");
-
- /*
- * Two simple optimizations. First, don't scan high ordered pages
- * if they are outside of the requested address range. Second, cache
- * the starting page index across calls and reuse it instead of
- * restarting the scan from the top. This is conditional on the
- * requested number of pages being the same or greater than the
- * cached amount.
- */
- for (pass = 0; pass < 2; pass++) {
- vm_page_lock_queues();
- if ((np == 0) || (np > npages)) {
- if (atop(high) < vm_page_array_size)
- start = atop(high) - npages + 1;
- else
- start = vm_page_array_size - npages + 1;
- }
- np = 0;
-retry:
- start--;
- /*
- * Find last page in array that is free, within range,
- * aligned, and such that the boundary won't be crossed.
- */
- for (i = start; i >= 0; i--) {
- phys = VM_PAGE_TO_PHYS(&pga[i]);
- pqtype = pga[i].queue - pga[i].pc;
- if (pass == 0) {
- if (pqtype != PQ_FREE && pqtype != PQ_CACHE)
- continue;
- } else if (pqtype != PQ_FREE && pqtype != PQ_CACHE &&
- pga[i].queue != PQ_ACTIVE &&
- pga[i].queue != PQ_INACTIVE)
- continue;
- if (phys >= low && phys + size <= high &&
- ((phys & (alignment - 1)) == 0) &&
- ((phys ^ (phys + size - 1)) & ~(boundary - 1)) == 0)
- break;
- }
- /* There are no candidates at all. */
- if (i < 0) {
- vm_page_unlock_queues();
- continue;
- }
- start = i;
- /*
- * Check successive pages for contiguous and free.
- */
- for (i = start + npages - 1; i > start; i--) {
- pqtype = pga[i].queue - pga[i].pc;
- if (VM_PAGE_TO_PHYS(&pga[i]) !=
- VM_PAGE_TO_PHYS(&pga[i - 1]) + PAGE_SIZE) {
- start = i - npages + 1;
- goto retry;
- }
- if (pass == 0) {
- if (pqtype != PQ_FREE && pqtype != PQ_CACHE) {
- start = i - npages + 1;
- goto retry;
- }
- } else if (pqtype != PQ_FREE && pqtype != PQ_CACHE &&
- pga[i].queue != PQ_ACTIVE &&
- pga[i].queue != PQ_INACTIVE) {
- start = i - npages + 1;
- goto retry;
- }
- }
- for (i = start + npages - 1; i >= start; i--) {
- vm_page_t m = &pga[i];
-
-retry_page:
- pqtype = m->queue - m->pc;
- if (pass != 0 && pqtype != PQ_FREE &&
- pqtype != PQ_CACHE) {
- if (m->queue == PQ_ACTIVE ||
- m->queue == PQ_INACTIVE) {
- if (vm_contig_launder_page(m) != 0)
- goto cleanup_freed;
- pqtype = m->queue - m->pc;
- if (pqtype != PQ_FREE &&
- pqtype != PQ_CACHE)
- goto cleanup_freed;
- } else {
-cleanup_freed:
- vm_page_release_contigl(&pga[i + 1],
- start + npages - 1 - i);
- start = i - npages + 1;
- goto retry;
- }
- }
- if (pqtype == PQ_CACHE) {
- if (m->hold_count != 0)
- goto cleanup_freed;
- object = m->object;
- if (!VM_OBJECT_TRYLOCK(object))
- goto cleanup_freed;
- if ((m->oflags & VPO_BUSY) || m->busy != 0) {
- VM_OBJECT_UNLOCK(object);
- goto cleanup_freed;
- }
- vm_page_free(m);
- VM_OBJECT_UNLOCK(object);
- }
- /*
- * There is no good API for freeing a page
- * directly to PQ_NONE on our behalf, so spin.
- */
- if (vm_contig_unqueue_free(m) != 0)
- goto retry_page;
- }
- /*
- * We've found a contiguous chunk that meets are requirements.
- */
- np = npages;
- startl = start;
- vm_page_unlock_queues();
- return (&pga[startl]);
- }
- return (NULL);
-}
-
static void *
contigmalloc2(vm_page_t m, vm_pindex_t npages, int flags)
{
@@ -571,11 +220,6 @@ contigmalloc2(vm_page_t m, vm_pindex_t npages, int flags)
return ((void *)addr);
}
-static int vm_old_contigmalloc = 0;
-SYSCTL_INT(_vm, OID_AUTO, old_contigmalloc,
- CTLFLAG_RW, &vm_old_contigmalloc, 0, "Use the old contigmalloc algorithm");
-TUNABLE_INT("vm.old_contigmalloc", &vm_old_contigmalloc);
-
void *
contigmalloc(
unsigned long size, /* should be size_t here and for malloc() */
@@ -587,27 +231,51 @@ contigmalloc(
unsigned long boundary)
{
void * ret;
- vm_page_t pages;
- vm_pindex_t npgs;
+ vm_object_t object;
+ vm_page_t m, m_next, pages;
+ unsigned long npgs;
+ int actl, actmax, inactl, inactmax, tries;
npgs = round_page(size) >> PAGE_SHIFT;
- mtx_lock(&Giant);
- if (vm_old_contigmalloc) {
- ret = contigmalloc1(size, type, flags, low, high, alignment,
- boundary, kernel_map);
- } else {
- pages = vm_page_alloc_contig(npgs, low, high,
- alignment, boundary);
- if (pages == NULL) {
- ret = NULL;
- } else {
- ret = contigmalloc2(pages, npgs, flags);
- if (ret == NULL)
- vm_page_release_contig(pages, npgs);
+ tries = 0;
+retry:
+ pages = vm_phys_alloc_contig(npgs, low, high, alignment, boundary);
+ if (pages == NULL) {
+ if (tries < ((flags & M_NOWAIT) != 0 ? 1 : 3)) {
+ vm_page_lock_queues();
+ inactl = 0;
+ inactmax = tries < 1 ? 0 : cnt.v_inactive_count;
+ actl = 0;
+ actmax = tries < 2 ? 0 : cnt.v_active_count;
+again:
+ if (inactl < inactmax &&
+ vm_contig_launder(PQ_INACTIVE)) {
+ inactl++;
+ goto again;
+ }
+ if (actl < actmax &&
+ vm_contig_launder(PQ_ACTIVE)) {
+ actl++;
+ goto again;
+ }
+ TAILQ_FOREACH_SAFE(m, &vm_page_queues[PQ_CACHE].pl,
+ pageq, m_next) {
+ if (m->hold_count == 0 &&
+ VM_OBJECT_TRYLOCK(object = m->object)) {
+ vm_page_free(m);
+ VM_OBJECT_UNLOCK(object);
+ }
+ }
+ vm_page_unlock_queues();
+ tries++;
+ goto retry;
}
-
+ ret = NULL;
+ } else {
+ ret = contigmalloc2(pages, npgs, flags);
+ if (ret == NULL)
+ vm_page_release_contig(pages, npgs);
}
- mtx_unlock(&Giant);
malloc_type_allocated(type, ret == NULL ? 0 : npgs << PAGE_SHIFT);
return (ret);
}
OpenPOWER on IntegriCloud