summaryrefslogtreecommitdiffstats
path: root/sys/arm/include
diff options
context:
space:
mode:
Diffstat (limited to 'sys/arm/include')
-rw-r--r--sys/arm/include/cpufunc.h107
-rw-r--r--sys/arm/include/param.h1
-rw-r--r--sys/arm/include/pmap-v6.h57
-rw-r--r--sys/arm/include/pmap.h172
-rw-r--r--sys/arm/include/smp.h8
-rw-r--r--sys/arm/include/vm.h4
6 files changed, 21 insertions, 328 deletions
diff --git a/sys/arm/include/cpufunc.h b/sys/arm/include/cpufunc.h
index 8bbb1dd..aad0feb 100644
--- a/sys/arm/include/cpufunc.h
+++ b/sys/arm/include/cpufunc.h
@@ -60,23 +60,17 @@ struct cpu_functions {
/* CPU functions */
- u_int (*cf_id) (void);
void (*cf_cpwait) (void);
/* MMU functions */
u_int (*cf_control) (u_int bic, u_int eor);
- void (*cf_domains) (u_int domains);
void (*cf_setttb) (u_int ttb);
- u_int (*cf_faultstatus) (void);
- u_int (*cf_faultaddress) (void);
/* TLB functions */
void (*cf_tlb_flushID) (void);
void (*cf_tlb_flushID_SE) (u_int va);
- void (*cf_tlb_flushI) (void);
- void (*cf_tlb_flushI_SE) (u_int va);
void (*cf_tlb_flushD) (void);
void (*cf_tlb_flushD_SE) (u_int va);
@@ -155,18 +149,12 @@ struct cpu_functions {
/* Other functions */
- void (*cf_flush_prefetchbuf) (void);
void (*cf_drain_writebuf) (void);
- void (*cf_flush_brnchtgt_C) (void);
- void (*cf_flush_brnchtgt_E) (u_int va);
void (*cf_sleep) (int mode);
/* Soft functions */
- int (*cf_dataabt_fixup) (void *arg);
- int (*cf_prefetchabt_fixup) (void *arg);
-
void (*cf_context_switch) (void);
void (*cf_setup) (void);
@@ -175,69 +163,16 @@ struct cpu_functions {
extern struct cpu_functions cpufuncs;
extern u_int cputype;
-#define cpu_ident() cpufuncs.cf_id()
#define cpu_cpwait() cpufuncs.cf_cpwait()
#define cpu_control(c, e) cpufuncs.cf_control(c, e)
-#define cpu_domains(d) cpufuncs.cf_domains(d)
#define cpu_setttb(t) cpufuncs.cf_setttb(t)
-#define cpu_faultstatus() cpufuncs.cf_faultstatus()
-#define cpu_faultaddress() cpufuncs.cf_faultaddress()
-
-#ifndef SMP
#define cpu_tlb_flushID() cpufuncs.cf_tlb_flushID()
#define cpu_tlb_flushID_SE(e) cpufuncs.cf_tlb_flushID_SE(e)
-#define cpu_tlb_flushI() cpufuncs.cf_tlb_flushI()
-#define cpu_tlb_flushI_SE(e) cpufuncs.cf_tlb_flushI_SE(e)
#define cpu_tlb_flushD() cpufuncs.cf_tlb_flushD()
#define cpu_tlb_flushD_SE(e) cpufuncs.cf_tlb_flushD_SE(e)
-#else
-void tlb_broadcast(int);
-
-#if defined(CPU_CORTEXA) || defined(CPU_MV_PJ4B) || defined(CPU_KRAIT)
-#define TLB_BROADCAST /* No need to explicitely send an IPI */
-#else
-#define TLB_BROADCAST tlb_broadcast(7)
-#endif
-
-#define cpu_tlb_flushID() do { \
- cpufuncs.cf_tlb_flushID(); \
- TLB_BROADCAST; \
-} while(0)
-
-#define cpu_tlb_flushID_SE(e) do { \
- cpufuncs.cf_tlb_flushID_SE(e); \
- TLB_BROADCAST; \
-} while(0)
-
-
-#define cpu_tlb_flushI() do { \
- cpufuncs.cf_tlb_flushI(); \
- TLB_BROADCAST; \
-} while(0)
-
-
-#define cpu_tlb_flushI_SE(e) do { \
- cpufuncs.cf_tlb_flushI_SE(e); \
- TLB_BROADCAST; \
-} while(0)
-
-
-#define cpu_tlb_flushD() do { \
- cpufuncs.cf_tlb_flushD(); \
- TLB_BROADCAST; \
-} while(0)
-
-
-#define cpu_tlb_flushD_SE(e) do { \
- cpufuncs.cf_tlb_flushD_SE(e); \
- TLB_BROADCAST; \
-} while(0)
-
-#endif
-
#define cpu_icache_sync_all() cpufuncs.cf_icache_sync_all()
#define cpu_icache_sync_range(a, s) cpufuncs.cf_icache_sync_range((a), (s))
@@ -255,19 +190,9 @@ void tlb_broadcast(int);
#define cpu_l2cache_wbinv_range(a, s) cpufuncs.cf_l2cache_wbinv_range((a), (s))
#define cpu_l2cache_drain_writebuf() cpufuncs.cf_l2cache_drain_writebuf()
-#define cpu_flush_prefetchbuf() cpufuncs.cf_flush_prefetchbuf()
#define cpu_drain_writebuf() cpufuncs.cf_drain_writebuf()
-#define cpu_flush_brnchtgt_C() cpufuncs.cf_flush_brnchtgt_C()
-#define cpu_flush_brnchtgt_E(e) cpufuncs.cf_flush_brnchtgt_E(e)
-
#define cpu_sleep(m) cpufuncs.cf_sleep(m)
-#define cpu_dataabt_fixup(a) cpufuncs.cf_dataabt_fixup(a)
-#define cpu_prefetchabt_fixup(a) cpufuncs.cf_prefetchabt_fixup(a)
-#define ABORT_FIXUP_OK 0 /* fixup succeeded */
-#define ABORT_FIXUP_FAILED 1 /* fixup failed */
-#define ABORT_FIXUP_RETURN 2 /* abort handler should return */
-
#define cpu_setup() cpufuncs.cf_setup()
int set_cpufuncs (void);
@@ -275,15 +200,11 @@ int set_cpufuncs (void);
#define ARCHITECTURE_NOT_SUPPORTED 2 /* not known */
void cpufunc_nullop (void);
-int cpufunc_null_fixup (void *);
-int early_abort_fixup (void *);
-int late_abort_fixup (void *);
-u_int cpufunc_id (void);
-u_int cpufunc_cpuid (void);
+u_int cpu_ident (void);
u_int cpufunc_control (u_int clear, u_int bic);
-void cpufunc_domains (u_int domains);
-u_int cpufunc_faultstatus (void);
-u_int cpufunc_faultaddress (void);
+void cpu_domains (u_int domains);
+u_int cpu_faultstatus (void);
+u_int cpu_faultaddress (void);
u_int cpu_pfr (int);
#if defined(CPU_FA526)
@@ -291,10 +212,7 @@ void fa526_setup (void);
void fa526_setttb (u_int ttb);
void fa526_context_switch (void);
void fa526_cpu_sleep (int);
-void fa526_tlb_flushI_SE (u_int);
void fa526_tlb_flushID_SE (u_int);
-void fa526_flush_prefetchbuf (void);
-void fa526_flush_brnchtgt_E (u_int);
void fa526_icache_sync_all (void);
void fa526_icache_sync_range(vm_offset_t start, vm_size_t end);
@@ -307,11 +225,13 @@ void fa526_idcache_wbinv_range(vm_offset_t start, vm_size_t end);
#endif
-#ifdef CPU_ARM9
+#if defined(CPU_ARM9) || defined(CPU_ARM9E)
void arm9_setttb (u_int);
-
void arm9_tlb_flushID_SE (u_int va);
+void arm9_context_switch (void);
+#endif
+#if defined(CPU_ARM9)
void arm9_icache_sync_all (void);
void arm9_icache_sync_range (vm_offset_t, vm_size_t);
@@ -323,8 +243,6 @@ void arm9_dcache_wb_range (vm_offset_t, vm_size_t);
void arm9_idcache_wbinv_all (void);
void arm9_idcache_wbinv_range (vm_offset_t, vm_size_t);
-void arm9_context_switch (void);
-
void arm9_setup (void);
extern unsigned arm9_dcache_sets_max;
@@ -334,11 +252,6 @@ extern unsigned arm9_dcache_index_inc;
#endif
#if defined(CPU_ARM9E)
-void arm10_tlb_flushID_SE (u_int);
-void arm10_tlb_flushI_SE (u_int);
-
-void arm10_context_switch (void);
-
void arm10_setup (void);
u_int sheeva_control_ext (u_int, u_int);
@@ -390,8 +303,6 @@ void pj4bv7_setup (void);
#if defined(CPU_ARM1176)
void arm11_tlb_flushID (void);
void arm11_tlb_flushID_SE (u_int);
-void arm11_tlb_flushI (void);
-void arm11_tlb_flushI_SE (u_int);
void arm11_tlb_flushD (void);
void arm11_tlb_flushD_SE (u_int va);
@@ -409,7 +320,6 @@ void arm11x6_setttb (u_int);
void arm11x6_idcache_wbinv_all (void);
void arm11x6_dcache_wbinv_all (void);
void arm11x6_icache_sync_all (void);
-void arm11x6_flush_prefetchbuf (void);
void arm11x6_icache_sync_range (vm_offset_t, vm_size_t);
void arm11x6_idcache_wbinv_range (vm_offset_t, vm_size_t);
void arm11x6_setup (void);
@@ -438,7 +348,6 @@ void armv5_ec_idcache_wbinv_range(vm_offset_t, vm_size_t);
defined(CPU_XSCALE_80219) || defined(CPU_XSCALE_81342)
void armv4_tlb_flushID (void);
-void armv4_tlb_flushI (void);
void armv4_tlb_flushD (void);
void armv4_tlb_flushD_SE (u_int va);
diff --git a/sys/arm/include/param.h b/sys/arm/include/param.h
index bbe9bcb..384891d 100644
--- a/sys/arm/include/param.h
+++ b/sys/arm/include/param.h
@@ -110,7 +110,6 @@
#define PAGE_SHIFT 12
#define PAGE_SIZE (1 << PAGE_SHIFT) /* Page size */
#define PAGE_MASK (PAGE_SIZE - 1)
-#define NPTEPG (PAGE_SIZE/(sizeof (pt_entry_t)))
#define PDR_SHIFT 20 /* log2(NBPDR) */
#define NBPDR (1 << PDR_SHIFT)
diff --git a/sys/arm/include/pmap-v6.h b/sys/arm/include/pmap-v6.h
index beaf638..d522384 100644
--- a/sys/arm/include/pmap-v6.h
+++ b/sys/arm/include/pmap-v6.h
@@ -216,28 +216,8 @@ vm_paddr_t pmap_preboot_get_pages(u_int );
void pmap_preboot_map_pages(vm_paddr_t , vm_offset_t , u_int );
vm_offset_t pmap_preboot_reserve_pages(u_int );
vm_offset_t pmap_preboot_get_vpages(u_int );
-void pmap_preboot_map_attr(vm_paddr_t , vm_offset_t , vm_size_t ,
- int , int );
-static __inline void
-pmap_map_chunk(vm_offset_t l1pt, vm_offset_t va, vm_offset_t pa,
- vm_size_t size, int prot, int cache)
-{
- pmap_preboot_map_attr(pa, va, size, prot, cache);
-}
-
-/*
- * This structure is used by machine-dependent code to describe
- * static mappings of devices, created at bootstrap time.
- */
-struct pmap_devmap {
- vm_offset_t pd_va; /* virtual address */
- vm_paddr_t pd_pa; /* physical address */
- vm_size_t pd_size; /* size of region */
- vm_prot_t pd_prot; /* protection code */
- int pd_cache; /* cache attributes */
-};
-
-void pmap_devmap_bootstrap(const struct pmap_devmap *);
+void pmap_preboot_map_attr(vm_paddr_t, vm_offset_t, vm_size_t, vm_prot_t,
+ vm_memattr_t);
#endif /* _KERNEL */
@@ -268,41 +248,8 @@ void pmap_devmap_bootstrap(const struct pmap_devmap *);
/*
* sys/arm/arm/cpufunc.c
*/
-void pmap_pte_init_mmu_v6(void);
void vector_page_setprot(int);
-
-/*
- * sys/arm/arm/db_interface.c
- * sys/arm/arm/machdep.c
- * sys/arm/arm/minidump_machdep.c
- * sys/arm/arm/pmap.c
- */
-#define pmap_kernel() kernel_pmap
-
-/*
- * sys/arm/arm/bus_space_generic.c (just comment)
- * sys/arm/arm/devmap.c
- * sys/arm/arm/pmap.c (just comment)
- * sys/arm/at91/at91_machdep.c
- * sys/arm/cavium/cns11xx/econa_machdep.c
- * sys/arm/freescale/imx/imx6_machdep.c (just comment)
- * sys/arm/mv/orion/db88f5xxx.c
- * sys/arm/mv/mv_localbus.c
- * sys/arm/mv/mv_machdep.c
- * sys/arm/mv/mv_pci.c
- * sys/arm/s3c2xx0/s3c24x0_machdep.c
- * sys/arm/versatile/versatile_machdep.c
- * sys/arm/xscale/ixp425/avila_machdep.c
- * sys/arm/xscale/i8134x/crb_machdep.c
- * sys/arm/xscale/i80321/ep80219_machdep.c
- * sys/arm/xscale/i80321/iq31244_machdep.c
- * sys/arm/xscale/pxa/pxa_machdep.c
- */
-#define PTE_DEVICE PTE2_ATTR_DEVICE
-
-
-
#endif /* _KERNEL */
// -----------------------------------------------------------------------------
diff --git a/sys/arm/include/pmap.h b/sys/arm/include/pmap.h
index dc7fd38..8222652 100644
--- a/sys/arm/include/pmap.h
+++ b/sys/arm/include/pmap.h
@@ -60,21 +60,10 @@
/*
* Pte related macros
*/
-#if ARM_ARCH_6 || ARM_ARCH_7A
-#ifdef SMP
-#define PTE_NOCACHE 2
-#else
-#define PTE_NOCACHE 1
-#endif
-#define PTE_CACHE 6
-#define PTE_DEVICE 2
-#define PTE_PAGETABLE 6
-#else
#define PTE_NOCACHE 1
#define PTE_CACHE 2
#define PTE_DEVICE PTE_NOCACHE
#define PTE_PAGETABLE 3
-#endif
enum mem_type {
STRONG_ORD = 0,
@@ -104,11 +93,7 @@ enum mem_type {
#define pmap_page_get_memattr(m) ((m)->md.pv_memattr)
#define pmap_page_is_write_mapped(m) (((m)->aflags & PGA_WRITEABLE) != 0)
-#if (ARM_MMU_V6 + ARM_MMU_V7) > 0
-boolean_t pmap_page_is_mapped(vm_page_t);
-#else
#define pmap_page_is_mapped(m) (!TAILQ_EMPTY(&(m)->md.pv_list))
-#endif
void pmap_page_set_memattr(vm_page_t m, vm_memattr_t ma);
/*
@@ -131,9 +116,7 @@ struct pv_chunk;
struct md_page {
int pvh_attrs;
vm_memattr_t pv_memattr;
-#if (ARM_MMU_V6 + ARM_MMU_V7) == 0
vm_offset_t pv_kva; /* first kernel VA mapping */
-#endif
TAILQ_HEAD(,pv_entry) pv_list;
};
@@ -164,11 +147,7 @@ struct pmap {
struct l2_dtable *pm_l2[L2_SIZE];
cpuset_t pm_active; /* active on cpus */
struct pmap_statistics pm_stats; /* pmap statictics */
-#if (ARM_MMU_V6 + ARM_MMU_V7) != 0
- TAILQ_HEAD(,pv_chunk) pm_pvchunk; /* list of mappings in pmap */
-#else
TAILQ_HEAD(,pv_entry) pm_pvlist; /* list of mappings in pmap */
-#endif
};
typedef struct pmap *pmap_t;
@@ -176,7 +155,6 @@ typedef struct pmap *pmap_t;
#ifdef _KERNEL
extern struct pmap kernel_pmap_store;
#define kernel_pmap (&kernel_pmap_store)
-#define pmap_kernel() kernel_pmap
#define PMAP_ASSERT_LOCKED(pmap) \
mtx_assert(&(pmap)->pm_mtx, MA_OWNED)
@@ -199,10 +177,8 @@ typedef struct pv_entry {
vm_offset_t pv_va; /* virtual address for mapping */
TAILQ_ENTRY(pv_entry) pv_list;
int pv_flags; /* flags (wired, etc...) */
-#if (ARM_MMU_V6 + ARM_MMU_V7) == 0
pmap_t pv_pmap; /* pmap where mapping lies */
TAILQ_ENTRY(pv_entry) pv_plist;
-#endif
} *pv_entry_t;
/*
@@ -247,7 +223,7 @@ vtopte(vm_offset_t va)
pd_entry_t *pdep;
pt_entry_t *ptep;
- if (pmap_get_pde_pte(pmap_kernel(), va, &pdep, &ptep) == FALSE)
+ if (pmap_get_pde_pte(kernel_pmap, va, &pdep, &ptep) == FALSE)
return (NULL);
return (ptep);
}
@@ -271,9 +247,7 @@ void *pmap_mapdev(vm_offset_t, vm_size_t);
void pmap_unmapdev(vm_offset_t, vm_size_t);
vm_page_t pmap_use_pt(pmap_t, vm_offset_t);
void pmap_debug(int);
-#if (ARM_MMU_V6 + ARM_MMU_V7) == 0
void pmap_map_section(vm_offset_t, vm_offset_t, vm_offset_t, int, int);
-#endif
void pmap_link_l2pt(vm_offset_t, vm_offset_t, struct pv_addr *);
vm_size_t pmap_map_chunk(vm_offset_t, vm_offset_t, vm_offset_t, vm_size_t, int, int);
void
@@ -341,119 +315,9 @@ extern int pmap_needs_pte_sync;
/*
* User-visible names for the ones that vary with MMU class.
*/
-#if (ARM_MMU_V6 + ARM_MMU_V7) != 0
-#define L2_AP(x) (L2_AP0(x))
-#else
#define L2_AP(x) (L2_AP0(x) | L2_AP1(x) | L2_AP2(x) | L2_AP3(x))
-#endif
-#if (ARM_MMU_V6 + ARM_MMU_V7) != 0
-/*
- * AP[2:1] access permissions model:
- *
- * AP[2](APX) - Write Disable
- * AP[1] - User Enable
- * AP[0] - Reference Flag
- *
- * AP[2] AP[1] Kernel User
- * 0 0 R/W N
- * 0 1 R/W R/W
- * 1 0 R N
- * 1 1 R R
- *
- */
-#define L2_S_PROT_R (0) /* kernel read */
-#define L2_S_PROT_U (L2_AP0(2)) /* user read */
-#define L2_S_REF (L2_AP0(1)) /* reference flag */
-
-#define L2_S_PROT_MASK (L2_S_PROT_U|L2_S_PROT_R|L2_APX)
-#define L2_S_EXECUTABLE(pte) (!(pte & L2_XN))
-#define L2_S_WRITABLE(pte) (!(pte & L2_APX))
-#define L2_S_REFERENCED(pte) (!!(pte & L2_S_REF))
-
-#ifndef SMP
-#define L1_S_CACHE_MASK (L1_S_TEX_MASK|L1_S_B|L1_S_C)
-#define L2_L_CACHE_MASK (L2_L_TEX_MASK|L2_B|L2_C)
-#define L2_S_CACHE_MASK (L2_S_TEX_MASK|L2_B|L2_C)
-#else
-#define L1_S_CACHE_MASK (L1_S_TEX_MASK|L1_S_B|L1_S_C|L1_SHARED)
-#define L2_L_CACHE_MASK (L2_L_TEX_MASK|L2_B|L2_C|L2_SHARED)
-#define L2_S_CACHE_MASK (L2_S_TEX_MASK|L2_B|L2_C|L2_SHARED)
-#endif /* SMP */
-
-#define L1_S_PROTO (L1_TYPE_S)
-#define L1_C_PROTO (L1_TYPE_C)
-#define L2_S_PROTO (L2_TYPE_S)
-
-/*
- * Promotion to a 1MB (SECTION) mapping requires that the corresponding
- * 4KB (SMALL) page mappings have identical settings for the following fields:
- */
-#define L2_S_PROMOTE (L2_S_REF | L2_SHARED | L2_S_PROT_MASK | \
- L2_XN | L2_S_PROTO)
-
-/*
- * In order to compare 1MB (SECTION) entry settings with the 4KB (SMALL)
- * page mapping it is necessary to read and shift appropriate bits from
- * L1 entry to positions of the corresponding bits in the L2 entry.
- */
-#define L1_S_DEMOTE(l1pd) ((((l1pd) & L1_S_PROTO) >> 0) | \
- (((l1pd) & L1_SHARED) >> 6) | \
- (((l1pd) & L1_S_REF) >> 6) | \
- (((l1pd) & L1_S_PROT_MASK) >> 6) | \
- (((l1pd) & L1_S_XN) >> 4))
-
-#ifndef SMP
-#define ARM_L1S_STRONG_ORD (0)
-#define ARM_L1S_DEVICE_NOSHARE (L1_S_TEX(2))
-#define ARM_L1S_DEVICE_SHARE (L1_S_B)
-#define ARM_L1S_NRML_NOCACHE (L1_S_TEX(1))
-#define ARM_L1S_NRML_IWT_OWT (L1_S_C)
-#define ARM_L1S_NRML_IWB_OWB (L1_S_C|L1_S_B)
-#define ARM_L1S_NRML_IWBA_OWBA (L1_S_TEX(1)|L1_S_C|L1_S_B)
-
-#define ARM_L2L_STRONG_ORD (0)
-#define ARM_L2L_DEVICE_NOSHARE (L2_L_TEX(2))
-#define ARM_L2L_DEVICE_SHARE (L2_B)
-#define ARM_L2L_NRML_NOCACHE (L2_L_TEX(1))
-#define ARM_L2L_NRML_IWT_OWT (L2_C)
-#define ARM_L2L_NRML_IWB_OWB (L2_C|L2_B)
-#define ARM_L2L_NRML_IWBA_OWBA (L2_L_TEX(1)|L2_C|L2_B)
-
-#define ARM_L2S_STRONG_ORD (0)
-#define ARM_L2S_DEVICE_NOSHARE (L2_S_TEX(2))
-#define ARM_L2S_DEVICE_SHARE (L2_B)
-#define ARM_L2S_NRML_NOCACHE (L2_S_TEX(1))
-#define ARM_L2S_NRML_IWT_OWT (L2_C)
-#define ARM_L2S_NRML_IWB_OWB (L2_C|L2_B)
-#define ARM_L2S_NRML_IWBA_OWBA (L2_S_TEX(1)|L2_C|L2_B)
-#else
-#define ARM_L1S_STRONG_ORD (0)
-#define ARM_L1S_DEVICE_NOSHARE (L1_S_TEX(2))
-#define ARM_L1S_DEVICE_SHARE (L1_S_B)
-#define ARM_L1S_NRML_NOCACHE (L1_S_TEX(1)|L1_SHARED)
-#define ARM_L1S_NRML_IWT_OWT (L1_S_C|L1_SHARED)
-#define ARM_L1S_NRML_IWB_OWB (L1_S_C|L1_S_B|L1_SHARED)
-#define ARM_L1S_NRML_IWBA_OWBA (L1_S_TEX(1)|L1_S_C|L1_S_B|L1_SHARED)
-
-#define ARM_L2L_STRONG_ORD (0)
-#define ARM_L2L_DEVICE_NOSHARE (L2_L_TEX(2))
-#define ARM_L2L_DEVICE_SHARE (L2_B)
-#define ARM_L2L_NRML_NOCACHE (L2_L_TEX(1)|L2_SHARED)
-#define ARM_L2L_NRML_IWT_OWT (L2_C|L2_SHARED)
-#define ARM_L2L_NRML_IWB_OWB (L2_C|L2_B|L2_SHARED)
-#define ARM_L2L_NRML_IWBA_OWBA (L2_L_TEX(1)|L2_C|L2_B|L2_SHARED)
-
-#define ARM_L2S_STRONG_ORD (0)
-#define ARM_L2S_DEVICE_NOSHARE (L2_S_TEX(2))
-#define ARM_L2S_DEVICE_SHARE (L2_B)
-#define ARM_L2S_NRML_NOCACHE (L2_S_TEX(1)|L2_SHARED)
-#define ARM_L2S_NRML_IWT_OWT (L2_C|L2_SHARED)
-#define ARM_L2S_NRML_IWB_OWB (L2_C|L2_B|L2_SHARED)
-#define ARM_L2S_NRML_IWBA_OWBA (L2_S_TEX(1)|L2_C|L2_B|L2_SHARED)
-#endif /* SMP */
-
-#elif ARM_NMMUS > 1
+#if ARM_NMMUS > 1
/* More than one MMU class configured; use variables. */
#define L2_S_PROT_U pte_l2_s_prot_u
#define L2_S_PROT_W pte_l2_s_prot_w
@@ -495,7 +359,7 @@ extern int pmap_needs_pte_sync;
#endif /* ARM_NMMUS > 1 */
-#if defined(CPU_XSCALE_81342) || ARM_ARCH_6 || ARM_ARCH_7A
+#if defined(CPU_XSCALE_81342)
#define PMAP_NEEDS_PTE_SYNC 1
#define PMAP_INCLUDE_PTE_SYNC
#else
@@ -506,8 +370,6 @@ extern int pmap_needs_pte_sync;
* These macros return various bits based on kernel/user and protection.
* Note that the compiler will usually fold these at compile time.
*/
-#if (ARM_MMU_V6 + ARM_MMU_V7) == 0
-
#define L1_S_PROT_U (L1_S_AP(AP_U))
#define L1_S_PROT_W (L1_S_AP(AP_W))
#define L1_S_PROT_MASK (L1_S_PROT_U|L1_S_PROT_W)
@@ -525,27 +387,6 @@ extern int pmap_needs_pte_sync;
#define L2_S_PROT(ku, pr) ((((ku) == PTE_USER) ? L2_S_PROT_U : 0) | \
(((pr) & VM_PROT_WRITE) ? L2_S_PROT_W : 0))
-#else
-#define L1_S_PROT_U (L1_S_AP(AP_U))
-#define L1_S_PROT_W (L1_S_APX) /* Write disable */
-#define L1_S_PROT_MASK (L1_S_PROT_W|L1_S_PROT_U)
-#define L1_S_REF (L1_S_AP(AP_REF)) /* Reference flag */
-#define L1_S_WRITABLE(pd) (!((pd) & L1_S_PROT_W))
-#define L1_S_EXECUTABLE(pd) (!((pd) & L1_S_XN))
-#define L1_S_REFERENCED(pd) ((pd) & L1_S_REF)
-
-#define L1_S_PROT(ku, pr) (((((ku) == PTE_KERNEL) ? 0 : L1_S_PROT_U) | \
- (((pr) & VM_PROT_WRITE) ? 0 : L1_S_PROT_W) | \
- (((pr) & VM_PROT_EXECUTE) ? 0 : L1_S_XN)))
-
-#define L2_L_PROT_MASK (L2_APX|L2_AP0(0x3))
-#define L2_L_PROT(ku, pr) (L2_L_PROT_MASK & ~((((ku) == PTE_KERNEL) ? L2_S_PROT_U : 0) | \
- (((pr) & VM_PROT_WRITE) ? L2_APX : 0)))
-
-#define L2_S_PROT(ku, pr) (L2_S_PROT_MASK & ~((((ku) == PTE_KERNEL) ? L2_S_PROT_U : 0) | \
- (((pr) & VM_PROT_WRITE) ? L2_APX : 0)))
-
-#endif
/*
* Macros to test if a mapping is mappable with an L1 Section mapping
@@ -620,15 +461,12 @@ extern void (*pmap_copy_page_offs_func)(vm_paddr_t a_phys,
vm_offset_t a_offs, vm_paddr_t b_phys, vm_offset_t b_offs, int cnt);
extern void (*pmap_zero_page_func)(vm_paddr_t, int, int);
-#if (ARM_MMU_GENERIC + ARM_MMU_V6 + ARM_MMU_V7) != 0 || defined(CPU_XSCALE_81342)
+#if ARM_MMU_GENERIC != 0 || defined(CPU_XSCALE_81342)
void pmap_copy_page_generic(vm_paddr_t, vm_paddr_t);
void pmap_zero_page_generic(vm_paddr_t, int, int);
void pmap_pte_init_generic(void);
-#if (ARM_MMU_V6 + ARM_MMU_V7) != 0
-void pmap_pte_init_mmu_v6(void);
-#endif /* (ARM_MMU_V6 + ARM_MMU_V7) != 0 */
-#endif /* (ARM_MMU_GENERIC + ARM_MMU_V6 + ARM_MMU_V7) != 0 */
+#endif /* ARM_MMU_GENERIC != 0 */
#if ARM_MMU_XSCALE == 1
void pmap_copy_page_xscale(vm_paddr_t, vm_paddr_t);
diff --git a/sys/arm/include/smp.h b/sys/arm/include/smp.h
index c50b99d..a993cc0 100644
--- a/sys/arm/include/smp.h
+++ b/sys/arm/include/smp.h
@@ -14,8 +14,8 @@ enum {
IPI_STOP,
IPI_STOP_HARD = IPI_STOP, /* These are synonyms on arm. */
IPI_HARDCLOCK,
- IPI_TLB,
- IPI_CACHE,
+ IPI_TLB, /* Not used now, but keep it reserved. */
+ IPI_CACHE, /* Not used now, but keep it reserved. */
INTR_IPI_COUNT
};
#else
@@ -25,8 +25,8 @@ enum {
#define IPI_STOP 4
#define IPI_STOP_HARD 4
#define IPI_HARDCLOCK 6
-#define IPI_TLB 7
-#define IPI_CACHE 8
+#define IPI_TLB 7 /* Not used now, but keep it reserved. */
+#define IPI_CACHE 8 /* Not used now, but keep it reserved. */
#endif /* INTRNG */
void init_secondary(int cpu);
diff --git a/sys/arm/include/vm.h b/sys/arm/include/vm.h
index 70a4ab9..552460e 100644
--- a/sys/arm/include/vm.h
+++ b/sys/arm/include/vm.h
@@ -38,13 +38,13 @@
#define VM_MEMATTR_NOCACHE ((vm_memattr_t)PTE2_ATTR_NOCACHE)
#define VM_MEMATTR_DEVICE ((vm_memattr_t)PTE2_ATTR_DEVICE)
#define VM_MEMATTR_SO ((vm_memattr_t)PTE2_ATTR_SO)
-#define VM_MEMATTR_WT ((vm_memattr_t)PTE2_ATTR_WT)
+#define VM_MEMATTR_WRITE_THROUGH ((vm_memattr_t)PTE2_ATTR_WT)
#define VM_MEMATTR_DEFAULT VM_MEMATTR_WB_WA
#define VM_MEMATTR_UNCACHEABLE VM_MEMATTR_SO /* misused by DMA */
#ifdef _KERNEL
/* Don't export aliased VM_MEMATTR to userland */
-#define VM_MEMATTR_WRITE_COMBINING VM_MEMATTR_WT /* for DRM */
+#define VM_MEMATTR_WRITE_COMBINING VM_MEMATTR_WRITE_THROUGH /* for DRM */
#define VM_MEMATTR_WRITE_BACK VM_MEMATTR_WB_WA /* for DRM */
#endif
#else
OpenPOWER on IntegriCloud