diff options
Diffstat (limited to 'sys/vm/vm_pageq.c')
-rw-r--r-- | sys/vm/vm_pageq.c | 233 |
1 files changed, 233 insertions, 0 deletions
diff --git a/sys/vm/vm_pageq.c b/sys/vm/vm_pageq.c new file mode 100644 index 0000000..b0b050b --- /dev/null +++ b/sys/vm/vm_pageq.c @@ -0,0 +1,233 @@ +/* + * (c)Copyright 1998, Matthew Dillon. Terms for use and redistribution + * are covered by the BSD Copyright as found in /usr/src/COPYRIGHT. + * + * $FreeBSD$ + */ + +#include <sys/param.h> +#include <sys/systm.h> +#include <sys/lock.h> +#include <sys/malloc.h> +#include <sys/mutex.h> +#include <sys/proc.h> +#include <sys/vmmeter.h> +#include <sys/vnode.h> + +#include <vm/vm.h> +#include <vm/vm_param.h> +#include <vm/vm_kern.h> +#include <vm/vm_object.h> +#include <vm/vm_page.h> +#include <vm/vm_pageout.h> +#include <vm/vm_pager.h> +#include <vm/vm_extern.h> + +struct vpgqueues vm_page_queues[PQ_COUNT]; +static struct mtx vm_pageq_mtx[PQ_COUNT]; + +void +vm_pageq_init(void) +{ + int i; + + for (i = 0; i < PQ_L2_SIZE; i++) { + vm_page_queues[PQ_FREE+i].cnt = &cnt.v_free_count; + } + for (i = 0; i < PQ_L2_SIZE; i++) { + vm_page_queues[PQ_CACHE+i].cnt = &cnt.v_cache_count; + } + vm_page_queues[PQ_INACTIVE].cnt = &cnt.v_inactive_count; + vm_page_queues[PQ_ACTIVE].cnt = &cnt.v_active_count; + + for (i = 0; i < PQ_COUNT; i++) { + TAILQ_INIT(&vm_page_queues[i].pl); + mtx_init(&vm_pageq_mtx[i], "vm pageq mutex", MTX_DEF); + } +} + +struct vpgqueues * +vm_pageq_aquire(int queue) +{ + struct vpgqueues *vpq = NULL; + + if (queue != PQ_NONE) { + vpq = &vm_page_queues[queue]; +#if 0 + mtx_lock(&vm_pageq_mtx[queue]); +#endif + } + return(vpq); +} + +void +vm_pageq_release(struct vpgqueues *vpq) +{ +#if 0 + mtx_unlock(&vm_pageq_mtx[vpq - &vm_page_queues[0]]); +#endif +} + +void +vm_pageq_requeue(vm_page_t m) +{ + int queue = m->queue; + struct vpgqueues *vpq; + + vpq = vm_pageq_aquire(queue); + TAILQ_REMOVE(&vpq->pl, m, pageq); + TAILQ_INSERT_TAIL(&vpq->pl, m, pageq); + vm_pageq_release(vpq); +} + +/* + * vm_add_new_page: + * + * Add a new page to the freelist for use by the system. + * Must be called at splhigh(). + */ +vm_page_t +vm_pageq_add_new_page(vm_offset_t pa) +{ + vm_page_t m; + + GIANT_REQUIRED; + + ++cnt.v_page_count; + ++cnt.v_free_count; + m = PHYS_TO_VM_PAGE(pa); + m->phys_addr = pa; + m->flags = 0; + m->pc = (pa >> PAGE_SHIFT) & PQ_L2_MASK; + m->queue = m->pc + PQ_FREE; + TAILQ_INSERT_TAIL(&vm_page_queues[m->queue].pl, m, pageq); + vm_page_queues[m->queue].lcnt++; + return (m); +} + +/* + * vm_pageq_remove_nowakeup: + * + * vm_page_unqueue() without any wakeup + * + * This routine must be called at splhigh(). + * This routine may not block. + */ + +void +vm_pageq_remove_nowakeup(vm_page_t m) +{ + int queue = m->queue; + struct vpgqueues *pq; + if (queue != PQ_NONE) { + pq = &vm_page_queues[queue]; + m->queue = PQ_NONE; + TAILQ_REMOVE(&pq->pl, m, pageq); + (*pq->cnt)--; + pq->lcnt--; + } +} + +/* + * vm_pageq_remove: + * + * Remove a page from its queue. + * + * This routine must be called at splhigh(). + * This routine may not block. + */ + +void +vm_pageq_remove(vm_page_t m) +{ + int queue = m->queue; + struct vpgqueues *pq; + + GIANT_REQUIRED; + if (queue != PQ_NONE) { + m->queue = PQ_NONE; + pq = &vm_page_queues[queue]; + TAILQ_REMOVE(&pq->pl, m, pageq); + (*pq->cnt)--; + pq->lcnt--; + if ((queue - m->pc) == PQ_CACHE) { + if (vm_paging_needed()) + pagedaemon_wakeup(); + } + } +} + +#if PQ_L2_SIZE > 1 + +/* + * vm_pageq_find: + * + * Find a page on the specified queue with color optimization. + * + * The page coloring optimization attempts to locate a page + * that does not overload other nearby pages in the object in + * the cpu's L1 or L2 caches. We need this optimization because + * cpu caches tend to be physical caches, while object spaces tend + * to be virtual. + * + * This routine must be called at splvm(). + * This routine may not block. + * + * This routine may only be called from the vm_page_list_find() macro + * in vm_page.h + */ + +static __inline vm_page_t +_vm_pageq_find(int basequeue, int index) +{ + int i; + vm_page_t m = NULL; + struct vpgqueues *pq; + + GIANT_REQUIRED; + pq = &vm_page_queues[basequeue]; + + /* + * Note that for the first loop, index+i and index-i wind up at the + * same place. Even though this is not totally optimal, we've already + * blown it by missing the cache case so we do not care. + */ + + for(i = PQ_L2_SIZE / 2; i > 0; --i) { + if ((m = TAILQ_FIRST(&pq[(index + i) & PQ_L2_MASK].pl)) != NULL) + break; + + if ((m = TAILQ_FIRST(&pq[(index - i) & PQ_L2_MASK].pl)) != NULL) + break; + } + return(m); +} + +#endif + +vm_page_t +vm_pageq_find(int basequeue, int index, boolean_t prefer_zero) +{ + vm_page_t m; + + GIANT_REQUIRED; + +#if PQ_L2_SIZE > 1 + if (prefer_zero) { + m = TAILQ_LAST(&vm_page_queues[basequeue+index].pl, pglist); + } else { + m = TAILQ_FIRST(&vm_page_queues[basequeue+index].pl); + } + if (m == NULL) { + m = _vm_pageq_find(basequeue, index); + } +#else + if (prefer_zero) { + m = TAILQ_LAST(&vm_page_queues[basequeue].pl, pglist); + } else { + m = TAILQ_FIRST(&vm_page_queues[basequeue].pl); + } +#endif + return(m); +} + |