diff options
author | gonzo <gonzo@FreeBSD.org> | 2009-01-15 18:31:36 +0000 |
---|---|---|
committer | gonzo <gonzo@FreeBSD.org> | 2009-01-15 18:31:36 +0000 |
commit | 2d3842287a220a3cdfb3ebdc2a413e5ed57798dd (patch) | |
tree | 50bdc2f1e37a9333171379f9245aa4080edca6fb /sys/mips | |
parent | 4b7227fe8be166f47d0d8586b8b34645a188d349 (diff) | |
download | FreeBSD-src-2d3842287a220a3cdfb3ebdc2a413e5ed57798dd.zip FreeBSD-src-2d3842287a220a3cdfb3ebdc2a413e5ed57798dd.tar.gz |
MFp4:
- Add debug output
- Fix pmap_zero_page and related places: use uncached segments and invalidate
cache after zeroing memory.
- Do not test for modified bit if it's not neccessary
(merged from mips-juniper p4 branch)
- Some #includes reorganization
Diffstat (limited to 'sys/mips')
-rw-r--r-- | sys/mips/include/pmap.h | 18 | ||||
-rw-r--r-- | sys/mips/mips/pmap.c | 38 |
2 files changed, 37 insertions, 19 deletions
diff --git a/sys/mips/include/pmap.h b/sys/mips/include/pmap.h index 2151348..625e92b3 100644 --- a/sys/mips/include/pmap.h +++ b/sys/mips/include/pmap.h @@ -1,4 +1,4 @@ -/* +/*- * Copyright (c) 1991 Regents of the University of California. * All rights reserved. * @@ -47,10 +47,8 @@ #define _MACHINE_PMAP_H_ #include <machine/vmparam.h> +#include <machine/pte.h> -/* - * Pte related macros - */ #define VADDR(pdi, pti) ((vm_offset_t)(((pdi)<<PDRSHIFT)|((pti)<<PAGE_SHIFT))) #define NKPT 120 /* actual number of kernel page tables */ @@ -65,11 +63,9 @@ #ifndef LOCORE #include <sys/queue.h> -#include <machine/pte.h> #include <sys/_lock.h> #include <sys/_mutex.h> - /* * Pmap stuff */ @@ -104,10 +100,7 @@ struct pmap { typedef struct pmap *pmap_t; -#ifdef _KERNEL -#include <sys/lock.h> -#include <sys/proc.h> -#include <vm/vm_map.h> +#ifdef _KERNEL pt_entry_t *pmap_pte(pmap_t, vm_offset_t); pd_entry_t pmap_segmap(pmap_t pmap, vm_offset_t va); @@ -132,8 +125,6 @@ extern pmap_t kernel_pmap; #define PMAP_LGMEM_UNLOCK(sysmap) mtx_unlock(&(sysmap)->lock) #define PMAP_LGMEM_DESTROY(sysmap) mtx_destroy(&(sysmap)->lock) -#endif /* _KERNEL */ - /* * For each vm_page_t, there is a list of all currently valid virtual * mappings of that page. An entry is a pv_entry_t, the list is pv_table. @@ -148,8 +139,6 @@ typedef struct pv_entry { } *pv_entry_t; -#ifdef _KERNEL - #if defined(DIAGNOSTIC) #define PMAP_DIAGNOSTIC #endif @@ -182,7 +171,6 @@ vm_offset_t pmap_steal_memory(vm_size_t size); void pmap_set_modified(vm_offset_t pa); int page_is_managed(vm_offset_t pa); void pmap_page_is_free(vm_page_t m); -void pmap_kushmem_reattach(struct proc *); /* PMAP_INLINE */ void pmap_kenter(vm_offset_t va, vm_paddr_t pa); /* PMAP_INLINE */ void pmap_kremove(vm_offset_t va); void *pmap_kenter_temporary(vm_paddr_t pa, int i); diff --git a/sys/mips/mips/pmap.c b/sys/mips/mips/pmap.c index 7256fc2..8212aa3 100644 --- a/sys/mips/mips/pmap.c +++ b/sys/mips/mips/pmap.c @@ -103,6 +103,8 @@ __FBSDID("$FreeBSD$"); #define PMAP_DIAGNOSTIC #endif +#undef PMAP_DEBUG + #ifndef PMAP_SHPGPERPROC #define PMAP_SHPGPERPROC 200 #endif @@ -489,6 +491,24 @@ pmap_nw_modified(pt_entry_t pte) #endif +/* + * this routine defines the region(s) of memory that should + * not be tested for the modified bit. + */ +static PMAP_INLINE int +pmap_track_modified(vm_offset_t va) +{ + /* + * Kernel submap initialization has been moved for MD to MI code. ie + * from cpu_startup() to vm_ksubmap_init(). clean_sva and clean_eva + * are part of the kmi structure. + */ + if ((va < kmi.clean_sva) || (va >= kmi.clean_eva)) + return (1); + else + return (0); +} + static void pmap_invalidate_all(pmap_t pmap) { @@ -672,6 +692,9 @@ pmap_kenter(vm_offset_t va, vm_paddr_t pa) register pt_entry_t *pte; pt_entry_t npte, opte; +#ifdef PMAP_DEBUG + printf("pmap_kenter: va: 0x%08x -> pa: 0x%08x\n", va, pa); +#endif npte = mips_paddr_to_tlbpfn(pa) | PTE_RW | PTE_V | PTE_G | PTE_W; if (is_cacheable_mem(pa)) @@ -1421,7 +1444,8 @@ pmap_remove_pte(struct pmap *pmap, pt_entry_t *ptq, vm_offset_t va) va, oldpte); } #endif - vm_page_dirty(m); + if (pmap_track_modified(va)) + vm_page_dirty(m); } if (m->md.pv_flags & PV_TABLE_REF) vm_page_flag_set(m, PG_REFERENCED); @@ -1778,6 +1802,9 @@ pmap_enter(pmap_t pmap, vm_offset_t va, vm_prot_t fault_type, vm_page_t m, vm_pr validate: rw = init_pte_prot(va, m, prot); +#ifdef PMAP_DEBUG + printf("pmap_enter: va: 0x%08x -> pa: 0x%08x\n", va, pa); +#endif /* * Now validate mapping with desired protection/wiring. */ @@ -2147,9 +2174,10 @@ pmap_zero_page(vm_page_t m) #endif if (phys < MIPS_KSEG0_LARGEST_PHYS) { - va = MIPS_PHYS_TO_CACHED(phys); + va = MIPS_PHYS_TO_UNCACHED(phys); bzero((caddr_t)va, PAGE_SIZE); + mips_dcache_wbinv_range(va, PAGE_SIZE); } else { int cpu; struct local_sysmaps *sysm; @@ -2202,8 +2230,9 @@ pmap_zero_page_area(vm_page_t m, int off, int size) } else #endif if (phys < MIPS_KSEG0_LARGEST_PHYS) { - va = MIPS_PHYS_TO_CACHED(phys); + va = MIPS_PHYS_TO_UNCACHED(phys); bzero((char *)(caddr_t)va + off, size); + mips_dcache_wbinv_range(va + off, size); } else { int cpu; struct local_sysmaps *sysm; @@ -2240,8 +2269,9 @@ pmap_zero_page_idle(vm_page_t m) } else #endif if (phys < MIPS_KSEG0_LARGEST_PHYS) { - va = MIPS_PHYS_TO_CACHED(phys); + va = MIPS_PHYS_TO_UNCACHED(phys); bzero((caddr_t)va, PAGE_SIZE); + mips_dcache_wbinv_range(va, PAGE_SIZE); } else { int cpu; struct local_sysmaps *sysm; |