diff options
author | fsmp <fsmp@FreeBSD.org> | 1997-07-23 20:47:19 +0000 |
---|---|---|
committer | fsmp <fsmp@FreeBSD.org> | 1997-07-23 20:47:19 +0000 |
commit | 31c730f3a6d3460c60a94601d51508874b1db14a (patch) | |
tree | 2adba51b4c63717f817e035de2715ad8337782c5 | |
parent | 8f17d1fae9a1616ab948ef9c678cc693ca2844b9 (diff) | |
download | FreeBSD-src-31c730f3a6d3460c60a94601d51508874b1db14a.zip FreeBSD-src-31c730f3a6d3460c60a94601d51508874b1db14a.tar.gz |
New simple_lock code in asm:
- s_lock_init()
- s_lock()
- s_lock_try()
- s_unlock()
Created lock for IO APIC and apic_imen (SMP version of imen)
- imen_lock
Code to use imen_lock for access from apic_ipl.s and apic_vector.s.
Moved this code *outside* of mp_lock.
It seems to work!!!
-rw-r--r-- | sys/amd64/amd64/apic_vector.S | 74 | ||||
-rw-r--r-- | sys/amd64/amd64/exception.S | 19 | ||||
-rw-r--r-- | sys/amd64/amd64/exception.s | 19 | ||||
-rw-r--r-- | sys/amd64/amd64/mp_machdep.c | 32 | ||||
-rw-r--r-- | sys/amd64/amd64/mptable.c | 32 | ||||
-rw-r--r-- | sys/amd64/include/mptable.h | 32 | ||||
-rw-r--r-- | sys/i386/i386/apic_vector.s | 74 | ||||
-rw-r--r-- | sys/i386/i386/exception.s | 19 | ||||
-rw-r--r-- | sys/i386/i386/mp_machdep.c | 32 | ||||
-rw-r--r-- | sys/i386/i386/mpapic.c | 10 | ||||
-rw-r--r-- | sys/i386/i386/mptable.c | 32 | ||||
-rw-r--r-- | sys/i386/include/mptable.h | 32 | ||||
-rw-r--r-- | sys/i386/isa/apic_ipl.s | 268 | ||||
-rw-r--r-- | sys/i386/isa/apic_vector.s | 74 | ||||
-rw-r--r-- | sys/i386/isa/ipl.s | 4 | ||||
-rw-r--r-- | sys/kern/subr_smp.c | 32 |
16 files changed, 550 insertions, 235 deletions
diff --git a/sys/amd64/amd64/apic_vector.S b/sys/amd64/amd64/apic_vector.S index 9ea96ca..cb4c4ff 100644 --- a/sys/amd64/amd64/apic_vector.S +++ b/sys/amd64/amd64/apic_vector.S @@ -1,6 +1,6 @@ /* * from: vector.s, 386BSD 0.1 unknown origin - * $Id: apic_vector.s,v 1.12 1997/07/18 19:47:13 smp Exp smp $ + * $Id: apic_vector.s,v 1.14 1997/07/23 20:18:14 smp Exp smp $ */ @@ -16,18 +16,21 @@ /* * 'lazy masking' code suggested by Bruce Evans <bde@zeta.org.au> */ +#if 1 #define MAYBE_MASK_IRQ(irq_num) \ - testl $IRQ_BIT(irq_num),iactive ; /* lazy masking */ \ - je 1f ; /* NOT currently active */ \ - orl $IRQ_BIT(irq_num),_imen ; /* set the mask bit */ \ + lock ; /* MP-safe */ \ + btsl $(irq_num),iactive ; /* lazy masking */ \ + jnc 1f ; /* NOT active */ \ + IMASK_LOCK ; /* enter critical reg */\ + orl $IRQ_BIT(irq_num),_apic_imen ; /* set the mask bit */ \ movl _ioapic,%ecx ; /* ioapic[0]addr */ \ movl $REDTBL_IDX(irq_num),(%ecx) ; /* write the index */ \ movl IOAPIC_WINDOW(%ecx),%eax ; /* current value */ \ orl $IOART_INTMASK,%eax ; /* set the mask */ \ movl %eax,IOAPIC_WINDOW(%ecx) ; /* new value */ \ - movl $0, lapic_eoi ; \ - orl $IRQ_BIT(irq_num), _ipending ; \ - REL_MPLOCK ; /* SMP release global lock */ \ + orl $IRQ_BIT(irq_num), _ipending ; /* set _ipending bit */ \ + movl $0, lapic_eoi ; /* do the EOI */ \ + IMASK_UNLOCK ; /* exit critical reg */ \ popl %es ; \ popl %ds ; \ popal ; \ @@ -35,16 +38,59 @@ iret ; \ ; \ ALIGN_TEXT ; \ -1: ; \ - orl $IRQ_BIT(irq_num),iactive - +1: GET_MPLOCK /* SMP Spin lock */ +#else +#define MAYBE_MASK_IRQ(irq_num) \ + GET_MPLOCK ; /* SMP Spin lock */ \ + lock ; /* MP-safe */ \ + btsl $(irq_num),iactive ; /* lazy masking */ \ + jnc 1f ; /* NOT active */ \ + /* XXX atomic access */ \ + orl $IRQ_BIT(irq_num),_apic_imen ; /* set the mask bit */ \ + movl _ioapic,%ecx ; /* ioapic[0]addr */ \ + movl $REDTBL_IDX(irq_num),(%ecx) ; /* write the index */ \ + movl IOAPIC_WINDOW(%ecx),%eax ; /* current value */ \ + orl $IOART_INTMASK,%eax ; /* set the mask */ \ + movl %eax,IOAPIC_WINDOW(%ecx) ; /* new value */ \ + movl $0, lapic_eoi ; /* do the EOI */ \ + orl $IRQ_BIT(irq_num), _ipending ; /* set _ipending bit */ \ + REL_MPLOCK ; /* release SMP GL */ \ + popl %es ; \ + popl %ds ; \ + popal ; \ + addl $4+4,%esp ; \ + iret ; \ +; \ + ALIGN_TEXT ; \ +1: +#endif +#if 1 #define MAYBE_UNMASK_IRQ(irq_num) \ - cli ; /* must unmask _imen and icu atomically */ \ + cli ; /* must unmask _apic_imen and IO APIC atomically */ \ + lock ; /* MP-safe */ \ andl $~IRQ_BIT(irq_num),iactive ; \ - testl $IRQ_BIT(irq_num),_imen ; \ + IMASK_LOCK ; /* enter critical reg */\ + testl $IRQ_BIT(irq_num),_apic_imen ; \ je 2f ; \ - andl $~IRQ_BIT(irq_num),_imen ; /* clear mask bit */ \ + andl $~IRQ_BIT(irq_num),_apic_imen ; /* clear mask bit */ \ + movl _ioapic,%ecx ; /* ioapic[0]addr */ \ + movl $REDTBL_IDX(irq_num),(%ecx) ; /* write the index */ \ + movl IOAPIC_WINDOW(%ecx),%eax ; /* current value */ \ + andl $~IOART_INTMASK,%eax ; /* clear the mask */ \ + movl %eax,IOAPIC_WINDOW(%ecx) ; /* new value */ \ +2: ; \ + IMASK_UNLOCK ; /* exit critical reg */ \ + sti /* XXX _doreti repeats the cli/sti */ +#else +#define MAYBE_UNMASK_IRQ(irq_num) \ + cli ; /* must unmask _apic_imen and IO APIC atomically */ \ + lock ; /* MP-safe */ \ + andl $~IRQ_BIT(irq_num),iactive ; \ + /* XXX atomic access */ \ + testl $IRQ_BIT(irq_num),_apic_imen ; \ + je 2f ; \ + andl $~IRQ_BIT(irq_num),_apic_imen ; /* clear mask bit */ \ movl _ioapic,%ecx ; /* ioapic[0]addr */ \ movl $REDTBL_IDX(irq_num),(%ecx) ; /* write the index */ \ movl IOAPIC_WINDOW(%ecx),%eax ; /* current value */ \ @@ -52,6 +98,7 @@ movl %eax,IOAPIC_WINDOW(%ecx) ; /* new value */ \ 2: ; \ sti ; /* XXX _doreti repeats the cli/sti */ +#endif /* @@ -131,7 +178,6 @@ IDTVEC(vec_name) ; \ movl $KDSEL,%eax ; /* ... and reload with kernel's ... */ \ movl %ax,%ds ; /* ... early for obsolete reasons */ \ movl %ax,%es ; \ - GET_MPLOCK ; /* SMP Spin lock */ \ MAYBE_MASK_IRQ(irq_num) ; \ movl $0, lapic_eoi ; \ movl _cpl,%eax ; \ diff --git a/sys/amd64/amd64/exception.S b/sys/amd64/amd64/exception.S index 991ff7b..ab55025 100644 --- a/sys/amd64/amd64/exception.S +++ b/sys/amd64/amd64/exception.S @@ -30,7 +30,7 @@ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * - * $Id: exception.s,v 1.31 1997/07/01 01:02:03 bde Exp $ + * $Id: exception.s,v 1.2 1997/07/23 20:24:21 smp Exp smp $ */ #include "npx.h" /* NNPX */ @@ -41,14 +41,31 @@ #include <machine/asmacros.h> /* miscellaneous macros */ #ifdef SMP + #include <machine/apic.h> /* for apic_vector.s */ #define GET_MPLOCK call _get_mplock #define REL_MPLOCK call _rel_mplock #define MP_INSTR_LOCK lock + +/* protects the IO APIC and apic_imen as a critical region */ +#define IMASK_LOCK \ + pushl $_imen_lock ; /* address of lock */ \ + call _s_lock ; /* MP-safe */ \ + addl $4,%esp + +#define IMASK_UNLOCK \ + pushl $_imen_lock ; /* address of lock */ \ + call _s_unlock ; /* MP-safe */ \ + addl $4,%esp + #else + #define GET_MPLOCK /* NOP get Kernel Mutex */ #define REL_MPLOCK /* NOP release mutex */ #define MP_INSTR_LOCK /* NOP instruction lock */ +#define IMASK_LOCK /* NOP IO APIC & apic_imen lock */ +#define IMASK_UNLOCK /* NOP IO APIC & apic_imen lock */ + #endif /* SMP */ #define KCSEL 0x08 /* kernel code selector */ diff --git a/sys/amd64/amd64/exception.s b/sys/amd64/amd64/exception.s index 991ff7b..ab55025 100644 --- a/sys/amd64/amd64/exception.s +++ b/sys/amd64/amd64/exception.s @@ -30,7 +30,7 @@ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * - * $Id: exception.s,v 1.31 1997/07/01 01:02:03 bde Exp $ + * $Id: exception.s,v 1.2 1997/07/23 20:24:21 smp Exp smp $ */ #include "npx.h" /* NNPX */ @@ -41,14 +41,31 @@ #include <machine/asmacros.h> /* miscellaneous macros */ #ifdef SMP + #include <machine/apic.h> /* for apic_vector.s */ #define GET_MPLOCK call _get_mplock #define REL_MPLOCK call _rel_mplock #define MP_INSTR_LOCK lock + +/* protects the IO APIC and apic_imen as a critical region */ +#define IMASK_LOCK \ + pushl $_imen_lock ; /* address of lock */ \ + call _s_lock ; /* MP-safe */ \ + addl $4,%esp + +#define IMASK_UNLOCK \ + pushl $_imen_lock ; /* address of lock */ \ + call _s_unlock ; /* MP-safe */ \ + addl $4,%esp + #else + #define GET_MPLOCK /* NOP get Kernel Mutex */ #define REL_MPLOCK /* NOP release mutex */ #define MP_INSTR_LOCK /* NOP instruction lock */ +#define IMASK_LOCK /* NOP IO APIC & apic_imen lock */ +#define IMASK_UNLOCK /* NOP IO APIC & apic_imen lock */ + #endif /* SMP */ #define KCSEL 0x08 /* kernel code selector */ diff --git a/sys/amd64/amd64/mp_machdep.c b/sys/amd64/amd64/mp_machdep.c index c9724a1..8a3f0c2 100644 --- a/sys/amd64/amd64/mp_machdep.c +++ b/sys/amd64/amd64/mp_machdep.c @@ -22,7 +22,7 @@ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * - * $Id: mp_machdep.c,v 1.18 1997/07/20 18:13:01 smp Exp smp $ + * $Id: mp_machdep.c,v 1.20 1997/07/23 20:20:21 smp Exp smp $ */ #include "opt_smp.h" @@ -248,6 +248,7 @@ static void mp_enable(u_int boot_addr); static int mptable_pass1(void); static int mptable_pass2(void); static void default_mp_table(int type); +static void init_locks(void); static int start_all_aps(u_int boot_addr); static void install_ap_tramp(u_int boot_addr); static int start_ap(int logicalCpu, u_int boot_addr); @@ -498,6 +499,9 @@ mp_enable(u_int boot_addr) #endif /* APIC_IO */ + /* initialize all SMP locks */ + init_locks(); + /* start each Application Processor */ start_all_aps(boot_addr); @@ -1415,6 +1419,23 @@ default_mp_table(int type) /* + * initialize all the SMP locks + */ +static void +init_locks(void) +{ + /* + * Get the initial mp_lock with a count of 1 for the BSP. + * This uses a LOGICAL cpu ID, ie BSP == 0. + */ + mp_lock = 0x00000001; + + /* locks the IO APIC and apic_imen accesses */ + s_lock_init(&imen_lock); +} + + +/* * start each AP in our list */ static int @@ -1429,15 +1450,6 @@ start_all_aps(u_int boot_addr) POSTCODE(START_ALL_APS_POST); - /** - * NOTE: this needs further thought: - * where does it get released? - * should it be set to empy? - * - * get the initial mp_lock with a count of 1 for the BSP - */ - mp_lock = 1; /* this uses a LOGICAL cpu ID, ie BSP == 0 */ - /* initialize BSP's local APIC */ apic_initialize(); diff --git a/sys/amd64/amd64/mptable.c b/sys/amd64/amd64/mptable.c index c9724a1..8a3f0c2 100644 --- a/sys/amd64/amd64/mptable.c +++ b/sys/amd64/amd64/mptable.c @@ -22,7 +22,7 @@ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * - * $Id: mp_machdep.c,v 1.18 1997/07/20 18:13:01 smp Exp smp $ + * $Id: mp_machdep.c,v 1.20 1997/07/23 20:20:21 smp Exp smp $ */ #include "opt_smp.h" @@ -248,6 +248,7 @@ static void mp_enable(u_int boot_addr); static int mptable_pass1(void); static int mptable_pass2(void); static void default_mp_table(int type); +static void init_locks(void); static int start_all_aps(u_int boot_addr); static void install_ap_tramp(u_int boot_addr); static int start_ap(int logicalCpu, u_int boot_addr); @@ -498,6 +499,9 @@ mp_enable(u_int boot_addr) #endif /* APIC_IO */ + /* initialize all SMP locks */ + init_locks(); + /* start each Application Processor */ start_all_aps(boot_addr); @@ -1415,6 +1419,23 @@ default_mp_table(int type) /* + * initialize all the SMP locks + */ +static void +init_locks(void) +{ + /* + * Get the initial mp_lock with a count of 1 for the BSP. + * This uses a LOGICAL cpu ID, ie BSP == 0. + */ + mp_lock = 0x00000001; + + /* locks the IO APIC and apic_imen accesses */ + s_lock_init(&imen_lock); +} + + +/* * start each AP in our list */ static int @@ -1429,15 +1450,6 @@ start_all_aps(u_int boot_addr) POSTCODE(START_ALL_APS_POST); - /** - * NOTE: this needs further thought: - * where does it get released? - * should it be set to empy? - * - * get the initial mp_lock with a count of 1 for the BSP - */ - mp_lock = 1; /* this uses a LOGICAL cpu ID, ie BSP == 0 */ - /* initialize BSP's local APIC */ apic_initialize(); diff --git a/sys/amd64/include/mptable.h b/sys/amd64/include/mptable.h index c9724a1..8a3f0c2 100644 --- a/sys/amd64/include/mptable.h +++ b/sys/amd64/include/mptable.h @@ -22,7 +22,7 @@ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * - * $Id: mp_machdep.c,v 1.18 1997/07/20 18:13:01 smp Exp smp $ + * $Id: mp_machdep.c,v 1.20 1997/07/23 20:20:21 smp Exp smp $ */ #include "opt_smp.h" @@ -248,6 +248,7 @@ static void mp_enable(u_int boot_addr); static int mptable_pass1(void); static int mptable_pass2(void); static void default_mp_table(int type); +static void init_locks(void); static int start_all_aps(u_int boot_addr); static void install_ap_tramp(u_int boot_addr); static int start_ap(int logicalCpu, u_int boot_addr); @@ -498,6 +499,9 @@ mp_enable(u_int boot_addr) #endif /* APIC_IO */ + /* initialize all SMP locks */ + init_locks(); + /* start each Application Processor */ start_all_aps(boot_addr); @@ -1415,6 +1419,23 @@ default_mp_table(int type) /* + * initialize all the SMP locks + */ +static void +init_locks(void) +{ + /* + * Get the initial mp_lock with a count of 1 for the BSP. + * This uses a LOGICAL cpu ID, ie BSP == 0. + */ + mp_lock = 0x00000001; + + /* locks the IO APIC and apic_imen accesses */ + s_lock_init(&imen_lock); +} + + +/* * start each AP in our list */ static int @@ -1429,15 +1450,6 @@ start_all_aps(u_int boot_addr) POSTCODE(START_ALL_APS_POST); - /** - * NOTE: this needs further thought: - * where does it get released? - * should it be set to empy? - * - * get the initial mp_lock with a count of 1 for the BSP - */ - mp_lock = 1; /* this uses a LOGICAL cpu ID, ie BSP == 0 */ - /* initialize BSP's local APIC */ apic_initialize(); diff --git a/sys/i386/i386/apic_vector.s b/sys/i386/i386/apic_vector.s index 9ea96ca..cb4c4ff 100644 --- a/sys/i386/i386/apic_vector.s +++ b/sys/i386/i386/apic_vector.s @@ -1,6 +1,6 @@ /* * from: vector.s, 386BSD 0.1 unknown origin - * $Id: apic_vector.s,v 1.12 1997/07/18 19:47:13 smp Exp smp $ + * $Id: apic_vector.s,v 1.14 1997/07/23 20:18:14 smp Exp smp $ */ @@ -16,18 +16,21 @@ /* * 'lazy masking' code suggested by Bruce Evans <bde@zeta.org.au> */ +#if 1 #define MAYBE_MASK_IRQ(irq_num) \ - testl $IRQ_BIT(irq_num),iactive ; /* lazy masking */ \ - je 1f ; /* NOT currently active */ \ - orl $IRQ_BIT(irq_num),_imen ; /* set the mask bit */ \ + lock ; /* MP-safe */ \ + btsl $(irq_num),iactive ; /* lazy masking */ \ + jnc 1f ; /* NOT active */ \ + IMASK_LOCK ; /* enter critical reg */\ + orl $IRQ_BIT(irq_num),_apic_imen ; /* set the mask bit */ \ movl _ioapic,%ecx ; /* ioapic[0]addr */ \ movl $REDTBL_IDX(irq_num),(%ecx) ; /* write the index */ \ movl IOAPIC_WINDOW(%ecx),%eax ; /* current value */ \ orl $IOART_INTMASK,%eax ; /* set the mask */ \ movl %eax,IOAPIC_WINDOW(%ecx) ; /* new value */ \ - movl $0, lapic_eoi ; \ - orl $IRQ_BIT(irq_num), _ipending ; \ - REL_MPLOCK ; /* SMP release global lock */ \ + orl $IRQ_BIT(irq_num), _ipending ; /* set _ipending bit */ \ + movl $0, lapic_eoi ; /* do the EOI */ \ + IMASK_UNLOCK ; /* exit critical reg */ \ popl %es ; \ popl %ds ; \ popal ; \ @@ -35,16 +38,59 @@ iret ; \ ; \ ALIGN_TEXT ; \ -1: ; \ - orl $IRQ_BIT(irq_num),iactive - +1: GET_MPLOCK /* SMP Spin lock */ +#else +#define MAYBE_MASK_IRQ(irq_num) \ + GET_MPLOCK ; /* SMP Spin lock */ \ + lock ; /* MP-safe */ \ + btsl $(irq_num),iactive ; /* lazy masking */ \ + jnc 1f ; /* NOT active */ \ + /* XXX atomic access */ \ + orl $IRQ_BIT(irq_num),_apic_imen ; /* set the mask bit */ \ + movl _ioapic,%ecx ; /* ioapic[0]addr */ \ + movl $REDTBL_IDX(irq_num),(%ecx) ; /* write the index */ \ + movl IOAPIC_WINDOW(%ecx),%eax ; /* current value */ \ + orl $IOART_INTMASK,%eax ; /* set the mask */ \ + movl %eax,IOAPIC_WINDOW(%ecx) ; /* new value */ \ + movl $0, lapic_eoi ; /* do the EOI */ \ + orl $IRQ_BIT(irq_num), _ipending ; /* set _ipending bit */ \ + REL_MPLOCK ; /* release SMP GL */ \ + popl %es ; \ + popl %ds ; \ + popal ; \ + addl $4+4,%esp ; \ + iret ; \ +; \ + ALIGN_TEXT ; \ +1: +#endif +#if 1 #define MAYBE_UNMASK_IRQ(irq_num) \ - cli ; /* must unmask _imen and icu atomically */ \ + cli ; /* must unmask _apic_imen and IO APIC atomically */ \ + lock ; /* MP-safe */ \ andl $~IRQ_BIT(irq_num),iactive ; \ - testl $IRQ_BIT(irq_num),_imen ; \ + IMASK_LOCK ; /* enter critical reg */\ + testl $IRQ_BIT(irq_num),_apic_imen ; \ je 2f ; \ - andl $~IRQ_BIT(irq_num),_imen ; /* clear mask bit */ \ + andl $~IRQ_BIT(irq_num),_apic_imen ; /* clear mask bit */ \ + movl _ioapic,%ecx ; /* ioapic[0]addr */ \ + movl $REDTBL_IDX(irq_num),(%ecx) ; /* write the index */ \ + movl IOAPIC_WINDOW(%ecx),%eax ; /* current value */ \ + andl $~IOART_INTMASK,%eax ; /* clear the mask */ \ + movl %eax,IOAPIC_WINDOW(%ecx) ; /* new value */ \ +2: ; \ + IMASK_UNLOCK ; /* exit critical reg */ \ + sti /* XXX _doreti repeats the cli/sti */ +#else +#define MAYBE_UNMASK_IRQ(irq_num) \ + cli ; /* must unmask _apic_imen and IO APIC atomically */ \ + lock ; /* MP-safe */ \ + andl $~IRQ_BIT(irq_num),iactive ; \ + /* XXX atomic access */ \ + testl $IRQ_BIT(irq_num),_apic_imen ; \ + je 2f ; \ + andl $~IRQ_BIT(irq_num),_apic_imen ; /* clear mask bit */ \ movl _ioapic,%ecx ; /* ioapic[0]addr */ \ movl $REDTBL_IDX(irq_num),(%ecx) ; /* write the index */ \ movl IOAPIC_WINDOW(%ecx),%eax ; /* current value */ \ @@ -52,6 +98,7 @@ movl %eax,IOAPIC_WINDOW(%ecx) ; /* new value */ \ 2: ; \ sti ; /* XXX _doreti repeats the cli/sti */ +#endif /* @@ -131,7 +178,6 @@ IDTVEC(vec_name) ; \ movl $KDSEL,%eax ; /* ... and reload with kernel's ... */ \ movl %ax,%ds ; /* ... early for obsolete reasons */ \ movl %ax,%es ; \ - GET_MPLOCK ; /* SMP Spin lock */ \ MAYBE_MASK_IRQ(irq_num) ; \ movl $0, lapic_eoi ; \ movl _cpl,%eax ; \ diff --git a/sys/i386/i386/exception.s b/sys/i386/i386/exception.s index 991ff7b..ab55025 100644 --- a/sys/i386/i386/exception.s +++ b/sys/i386/i386/exception.s @@ -30,7 +30,7 @@ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * - * $Id: exception.s,v 1.31 1997/07/01 01:02:03 bde Exp $ + * $Id: exception.s,v 1.2 1997/07/23 20:24:21 smp Exp smp $ */ #include "npx.h" /* NNPX */ @@ -41,14 +41,31 @@ #include <machine/asmacros.h> /* miscellaneous macros */ #ifdef SMP + #include <machine/apic.h> /* for apic_vector.s */ #define GET_MPLOCK call _get_mplock #define REL_MPLOCK call _rel_mplock #define MP_INSTR_LOCK lock + +/* protects the IO APIC and apic_imen as a critical region */ +#define IMASK_LOCK \ + pushl $_imen_lock ; /* address of lock */ \ + call _s_lock ; /* MP-safe */ \ + addl $4,%esp + +#define IMASK_UNLOCK \ + pushl $_imen_lock ; /* address of lock */ \ + call _s_unlock ; /* MP-safe */ \ + addl $4,%esp + #else + #define GET_MPLOCK /* NOP get Kernel Mutex */ #define REL_MPLOCK /* NOP release mutex */ #define MP_INSTR_LOCK /* NOP instruction lock */ +#define IMASK_LOCK /* NOP IO APIC & apic_imen lock */ +#define IMASK_UNLOCK /* NOP IO APIC & apic_imen lock */ + #endif /* SMP */ #define KCSEL 0x08 /* kernel code selector */ diff --git a/sys/i386/i386/mp_machdep.c b/sys/i386/i386/mp_machdep.c index c9724a1..8a3f0c2 100644 --- a/sys/i386/i386/mp_machdep.c +++ b/sys/i386/i386/mp_machdep.c @@ -22,7 +22,7 @@ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * - * $Id: mp_machdep.c,v 1.18 1997/07/20 18:13:01 smp Exp smp $ + * $Id: mp_machdep.c,v 1.20 1997/07/23 20:20:21 smp Exp smp $ */ #include "opt_smp.h" @@ -248,6 +248,7 @@ static void mp_enable(u_int boot_addr); static int mptable_pass1(void); static int mptable_pass2(void); static void default_mp_table(int type); +static void init_locks(void); static int start_all_aps(u_int boot_addr); static void install_ap_tramp(u_int boot_addr); static int start_ap(int logicalCpu, u_int boot_addr); @@ -498,6 +499,9 @@ mp_enable(u_int boot_addr) #endif /* APIC_IO */ + /* initialize all SMP locks */ + init_locks(); + /* start each Application Processor */ start_all_aps(boot_addr); @@ -1415,6 +1419,23 @@ default_mp_table(int type) /* + * initialize all the SMP locks + */ +static void +init_locks(void) +{ + /* + * Get the initial mp_lock with a count of 1 for the BSP. + * This uses a LOGICAL cpu ID, ie BSP == 0. + */ + mp_lock = 0x00000001; + + /* locks the IO APIC and apic_imen accesses */ + s_lock_init(&imen_lock); +} + + +/* * start each AP in our list */ static int @@ -1429,15 +1450,6 @@ start_all_aps(u_int boot_addr) POSTCODE(START_ALL_APS_POST); - /** - * NOTE: this needs further thought: - * where does it get released? - * should it be set to empy? - * - * get the initial mp_lock with a count of 1 for the BSP - */ - mp_lock = 1; /* this uses a LOGICAL cpu ID, ie BSP == 0 */ - /* initialize BSP's local APIC */ apic_initialize(); diff --git a/sys/i386/i386/mpapic.c b/sys/i386/i386/mpapic.c index 3df8000..bb369c0 100644 --- a/sys/i386/i386/mpapic.c +++ b/sys/i386/i386/mpapic.c @@ -22,7 +22,7 @@ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * - * $Id: mpapic.c,v 1.21 1997/07/22 21:21:35 smp Exp smp $ + * $Id: mpapic.c,v 1.23 1997/07/23 20:20:21 smp Exp smp $ */ #include "opt_smp.h" @@ -432,19 +432,19 @@ bad: /* - * Print contents of imen, keeps imen 'opaque'. + * Print contents of apic_imen. */ +extern u_int apic_imen; /* keep apic_imen 'opaque' */ void imen_dump(void) { - extern unsigned imen; /* interrupt mask enable */ int x; printf("SMP: enabled INTs: "); for (x = 0; x < 24; ++x) - if ((imen & (1 << x)) == 0) + if ((apic_imen & (1 << x)) == 0) printf("%d, ", x); - printf("imen: 0x%08x\n", imen); + printf("apic_imen: 0x%08x\n", apic_imen); } diff --git a/sys/i386/i386/mptable.c b/sys/i386/i386/mptable.c index c9724a1..8a3f0c2 100644 --- a/sys/i386/i386/mptable.c +++ b/sys/i386/i386/mptable.c @@ -22,7 +22,7 @@ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * - * $Id: mp_machdep.c,v 1.18 1997/07/20 18:13:01 smp Exp smp $ + * $Id: mp_machdep.c,v 1.20 1997/07/23 20:20:21 smp Exp smp $ */ #include "opt_smp.h" @@ -248,6 +248,7 @@ static void mp_enable(u_int boot_addr); static int mptable_pass1(void); static int mptable_pass2(void); static void default_mp_table(int type); +static void init_locks(void); static int start_all_aps(u_int boot_addr); static void install_ap_tramp(u_int boot_addr); static int start_ap(int logicalCpu, u_int boot_addr); @@ -498,6 +499,9 @@ mp_enable(u_int boot_addr) #endif /* APIC_IO */ + /* initialize all SMP locks */ + init_locks(); + /* start each Application Processor */ start_all_aps(boot_addr); @@ -1415,6 +1419,23 @@ default_mp_table(int type) /* + * initialize all the SMP locks + */ +static void +init_locks(void) +{ + /* + * Get the initial mp_lock with a count of 1 for the BSP. + * This uses a LOGICAL cpu ID, ie BSP == 0. + */ + mp_lock = 0x00000001; + + /* locks the IO APIC and apic_imen accesses */ + s_lock_init(&imen_lock); +} + + +/* * start each AP in our list */ static int @@ -1429,15 +1450,6 @@ start_all_aps(u_int boot_addr) POSTCODE(START_ALL_APS_POST); - /** - * NOTE: this needs further thought: - * where does it get released? - * should it be set to empy? - * - * get the initial mp_lock with a count of 1 for the BSP - */ - mp_lock = 1; /* this uses a LOGICAL cpu ID, ie BSP == 0 */ - /* initialize BSP's local APIC */ apic_initialize(); diff --git a/sys/i386/include/mptable.h b/sys/i386/include/mptable.h index c9724a1..8a3f0c2 100644 --- a/sys/i386/include/mptable.h +++ b/sys/i386/include/mptable.h @@ -22,7 +22,7 @@ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * - * $Id: mp_machdep.c,v 1.18 1997/07/20 18:13:01 smp Exp smp $ + * $Id: mp_machdep.c,v 1.20 1997/07/23 20:20:21 smp Exp smp $ */ #include "opt_smp.h" @@ -248,6 +248,7 @@ static void mp_enable(u_int boot_addr); static int mptable_pass1(void); static int mptable_pass2(void); static void default_mp_table(int type); +static void init_locks(void); static int start_all_aps(u_int boot_addr); static void install_ap_tramp(u_int boot_addr); static int start_ap(int logicalCpu, u_int boot_addr); @@ -498,6 +499,9 @@ mp_enable(u_int boot_addr) #endif /* APIC_IO */ + /* initialize all SMP locks */ + init_locks(); + /* start each Application Processor */ start_all_aps(boot_addr); @@ -1415,6 +1419,23 @@ default_mp_table(int type) /* + * initialize all the SMP locks + */ +static void +init_locks(void) +{ + /* + * Get the initial mp_lock with a count of 1 for the BSP. + * This uses a LOGICAL cpu ID, ie BSP == 0. + */ + mp_lock = 0x00000001; + + /* locks the IO APIC and apic_imen accesses */ + s_lock_init(&imen_lock); +} + + +/* * start each AP in our list */ static int @@ -1429,15 +1450,6 @@ start_all_aps(u_int boot_addr) POSTCODE(START_ALL_APS_POST); - /** - * NOTE: this needs further thought: - * where does it get released? - * should it be set to empy? - * - * get the initial mp_lock with a count of 1 for the BSP - */ - mp_lock = 1; /* this uses a LOGICAL cpu ID, ie BSP == 0 */ - /* initialize BSP's local APIC */ apic_initialize(); diff --git a/sys/i386/isa/apic_ipl.s b/sys/i386/isa/apic_ipl.s index 153c77a..d11a6ea 100644 --- a/sys/i386/isa/apic_ipl.s +++ b/sys/i386/isa/apic_ipl.s @@ -22,7 +22,7 @@ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * - * $Id: apic_ipl.s,v 1.16 1997/07/23 05:40:55 smp Exp smp $ + * $Id: apic_ipl.s,v 1.18 1997/07/23 20:18:14 smp Exp smp $ */ @@ -66,6 +66,26 @@ _vec: .long vec8, vec9, vec10, vec11, vec12, vec13, vec14, vec15 .long vec16, vec17, vec18, vec19, vec20, vec21, vec22, vec23 + .align 4 /* MUST be 32bit aligned */ + .globl _imen_lock +_imen_lock: + .long 0 + +/* + * Note: + * This is the UP equivilant of _imen. + * It is OPAQUE, and must NOT be accessed directly. + * It MUST be accessed along with the IO APIC as a 'critical region'. + * Accessed by: + * INTREN() + * INTRDIS() + * MAYBE_MASK_IRQ + * MAYBE_UNMASK_IRQ + * imen_dump() + */ + .globl _apic_imen +_apic_imen: + .long HWI_MASK /* * @@ -92,9 +112,10 @@ _vec8254: pushfl pushl $KCSEL pushl %eax - cli movl _mask8254, %eax /* lazy masking */ notl %eax + cli + lock /* MP-safe */ andl %eax, iactive MEXITCOUNT movl _Xintr8254, %eax @@ -109,7 +130,8 @@ vec0: pushl $KCSEL pushl %eax cli - andl $~IRQ_BIT(0), iactive ; /* lazy masking */ + lock /* MP-safe */ + andl $~IRQ_BIT(0), iactive /* lazy masking */ MEXITCOUNT jmp _Xintr0 /* XXX might need _Xfastintr0 */ #else @@ -119,9 +141,10 @@ _vec8254: pushfl pushl $KCSEL pushl %eax - cli movl _mask8254, %eax /* lazy masking */ notl %eax + cli + lock /* MP-safe */ andl %eax, iactive MEXITCOUNT movl _Xintr8254, %eax @@ -141,7 +164,8 @@ vec8: pushl $KCSEL pushl %eax cli - andl $~IRQ_BIT(8), iactive ; /* lazy masking */ + lock /* MP-safe */ + andl $~IRQ_BIT(8), iactive /* lazy masking */ MEXITCOUNT jmp _Xintr8 /* XXX might need _Xfastintr8 */ @@ -157,6 +181,7 @@ __CONCAT(vec,irq_num): ; \ pushl $KCSEL ; \ pushl %eax ; \ cli ; \ + lock ; /* MP-safe */ \ andl $~IRQ_BIT(irq_num), iactive ; /* lazy masking */ \ MEXITCOUNT ; \ jmp __CONCAT(_Xintr,irq_num) @@ -203,15 +228,15 @@ __CONCAT(vec,irq_num): ; \ */ /* this nonsense is to verify that masks ALWAYS have 1 and only 1 bit set */ -#define QUALIFY_MASKS +#define QUALIFY_MASKS_NOT #ifdef QUALIFY_MASKS #define QUALIFY_MASK \ - btrl %ecx, %eax ; \ - andl %eax, %eax ; \ - jz 1f ; \ - pushl $bad_mask ; \ - call _panic ; \ + btrl %ecx, %eax ; \ + andl %eax, %eax ; \ + jz 1f ; \ + pushl $bad_mask ; \ + call _panic ; \ 1: bad_mask: .asciz "bad mask" @@ -223,56 +248,62 @@ bad_mask: .asciz "bad mask" * MULTIPLE_IOAPICSXXX: cannot assume apic #0 in the following function. * (soon to be) MP-safe function to clear ONE INT mask bit. * The passed arg is a 32bit u_int MASK. - * It sets the associated bit in imen. + * It sets the associated bit in _apic_imen. * It sets the mask bit of the associated IO APIC register. */ ALIGN_TEXT .globl _INTREN _INTREN: - movl 4(%esp), %eax /* mask into %eax */ - bsfl %eax, %ecx /* get pin index */ - btrl %ecx, _imen /* update imen */ + IMASK_LOCK /* enter critical reg */ + + movl 4(%esp), %eax /* mask into %eax */ + bsfl %eax, %ecx /* get pin index */ + btrl %ecx, _apic_imen /* update _apic_imen */ QUALIFY_MASK - leal 16(,%ecx,2), %ecx /* calculate register index */ + leal 16(,%ecx,2), %ecx /* calculate register index */ - movl $0, %edx /* XXX FIXME: APIC # */ - movl _ioapic(,%edx,4), %edx /* %edx holds APIC base address */ + movl $0, %edx /* XXX FIXME: APIC # */ + movl _ioapic(,%edx,4), %edx /* %edx holds APIC base address */ - movl %ecx, (%edx) /* write the target register index */ - movl 16(%edx), %eax /* read the target register data */ - andl $~IOART_INTMASK, %eax /* clear mask bit */ - movl %eax, 16(%edx) /* write the APIC register data */ + movl %ecx, (%edx) /* write the target register index */ + movl 16(%edx), %eax /* read the target register data */ + andl $~IOART_INTMASK, %eax /* clear mask bit */ + movl %eax, 16(%edx) /* write the APIC register data */ + IMASK_UNLOCK /* exit critical reg */ ret /* * MULTIPLE_IOAPICSXXX: cannot assume apic #0 in the following function. * (soon to be) MP-safe function to set ONE INT mask bit. * The passed arg is a 32bit u_int MASK. - * It clears the associated bit in imen. + * It clears the associated bit in _apic_imen. * It clears the mask bit of the associated IO APIC register. */ ALIGN_TEXT .globl _INTRDIS _INTRDIS: - movl 4(%esp), %eax /* mask into %eax */ - bsfl %eax, %ecx /* get pin index */ - btsl %ecx, _imen /* update imen */ + IMASK_LOCK /* enter critical reg */ + + movl 4(%esp), %eax /* mask into %eax */ + bsfl %eax, %ecx /* get pin index */ + btsl %ecx, _apic_imen /* update _apic_imen */ QUALIFY_MASK - leal 16(,%ecx,2), %ecx /* calculate register index */ + leal 16(,%ecx,2), %ecx /* calculate register index */ - movl $0, %edx /* XXX FIXME: APIC # */ - movl _ioapic(,%edx,4), %edx /* %edx holds APIC base address */ + movl $0, %edx /* XXX FIXME: APIC # */ + movl _ioapic(,%edx,4), %edx /* %edx holds APIC base address */ - movl %ecx, (%edx) /* write the target register index */ - movl 16(%edx), %eax /* read the target register data */ - orl $IOART_INTMASK, %eax /* set mask bit */ - movl %eax, 16(%edx) /* write the APIC register data */ + movl %ecx, (%edx) /* write the target register index */ + movl 16(%edx), %eax /* read the target register data */ + orl $IOART_INTMASK, %eax /* set mask bit */ + movl %eax, 16(%edx) /* write the APIC register data */ + IMASK_UNLOCK /* exit critical reg */ ret @@ -282,21 +313,6 @@ _INTRDIS: /* - * void clr_ioapic_maskbit(int apic, int bit); - */ - .align 2 -clr_ioapic_maskbit: - ret - -/* - * void set_ioapic_maskbit(int apic, int bit); - */ - .align 2 -set_ioapic_maskbit: - ret - - -/* * void write_ioapic_mask(int apic, u_int mask); */ @@ -314,37 +330,37 @@ write_ioapic_mask: pushl %ebx /* scratch */ pushl %esi /* scratch */ - movl _imen, %ebx - xorl _MASK, %ebx /* %ebx = imen ^ mask */ - andl $_PIN_MASK, %ebx /* %ebx = imen & 0x00ffffff */ - jz all_done /* no change, return */ + movl _apic_imen, %ebx + xorl _MASK, %ebx /* %ebx = _apic_imen ^ mask */ + andl $_PIN_MASK, %ebx /* %ebx = _apic_imen & 0x00ffffff */ + jz all_done /* no change, return */ - movl _APIC, %esi /* APIC # */ - movl _ioapic(,%esi,4), %esi /* %esi holds APIC base address */ + movl _APIC, %esi /* APIC # */ + movl _ioapic(,%esi,4), %esi /* %esi holds APIC base address */ next_loop: /* %ebx = diffs, %esi = APIC base */ - bsfl %ebx, %ecx /* %ecx = index if 1st/next set bit */ - jz all_done + bsfl %ebx, %ecx /* %ecx = index if 1st/next set bit */ + jz all_done - btrl %ecx, %ebx /* clear this bit in diffs */ - leal 16(,%ecx,2), %edx /* calculate register index */ + btrl %ecx, %ebx /* clear this bit in diffs */ + leal 16(,%ecx,2), %edx /* calculate register index */ - movl %edx, (%esi) /* write the target register index */ - movl 16(%esi), %eax /* read the target register data */ + movl %edx, (%esi) /* write the target register index */ + movl 16(%esi), %eax /* read the target register data */ - btl %ecx, _MASK /* test for mask or unmask */ - jnc clear /* bit is clear */ - orl $_INT_MASK, %eax /* set mask bit */ - jmp write -clear: andl $~_INT_MASK, %eax /* clear mask bit */ + btl %ecx, _MASK /* test for mask or unmask */ + jnc clear /* bit is clear */ + orl $_INT_MASK, %eax /* set mask bit */ + jmp write +clear: andl $~_INT_MASK, %eax /* clear mask bit */ -write: movl %eax, 16(%esi) /* write the APIC register data */ +write: movl %eax, 16(%esi) /* write the APIC register data */ - jmp next_loop /* try another pass */ + jmp next_loop /* try another pass */ all_done: - popl %esi - popl %ebx + popl %esi + popl %ebx ret #undef _OLD_ESI @@ -359,28 +375,28 @@ all_done: #ifdef oldcode _INTREN: - movl _imen, %eax + movl _apic_imen, %eax notl %eax /* mask = ~mask */ - andl _imen, %eax /* %eax = imen & ~mask */ + andl _apic_imen, %eax /* %eax = _apic_imen & ~mask */ - pushl %eax /* new (future) imen value */ + pushl %eax /* new (future) _apic_imen value */ pushl $0 /* APIC# arg */ call write_ioapic_mask /* modify the APIC registers */ addl $4, %esp /* remove APIC# arg from stack */ - popl _imen /* imen |= mask */ + popl _apic_imen /* _apic_imen |= mask */ ret _INTRDIS: - movl _imen, %eax - orl 4(%esp), %eax /* %eax = imen | mask */ + movl _apic_imen, %eax + orl 4(%esp), %eax /* %eax = _apic_imen | mask */ - pushl %eax /* new (future) imen value */ + pushl %eax /* new (future) _apic_imen value */ pushl $0 /* APIC# arg */ call write_ioapic_mask /* modify the APIC registers */ addl $4, %esp /* remove APIC# arg from stack */ - popl _imen /* imen |= mask */ + popl _apic_imen /* _apic_imen |= mask */ ret #endif /* oldcode */ @@ -398,30 +414,35 @@ read_io_apic_mask: /* * Set INT mask bit for each bit set in 'mask'. * Ignore INT mask bit for all others. - * Only consider lower 24 bits in mask. * - * void set_io_apic_mask24(apic, u_int32_t bits); + * void set_io_apic_mask(apic, u_int32_t bits); */ .align 2 set_io_apic_mask: ret /* + * void set_ioapic_maskbit(int apic, int bit); + */ + .align 2 +set_ioapic_maskbit: + ret + +/* * Clear INT mask bit for each bit set in 'mask'. * Ignore INT mask bit for all others. - * Only consider lower 24 bits in mask. * - * void clr_io_apic_mask24(int apic, u_int32_t bits); + * void clr_io_apic_mask(int apic, u_int32_t bits); */ .align 2 -clr_io_apic_mask24: +clr_io_apic_mask: ret /* - * + * void clr_ioapic_maskbit(int apic, int bit); */ .align 2 - +clr_ioapic_maskbit: ret #endif /** ready */ @@ -436,11 +457,11 @@ clr_io_apic_mask24: .align 2 .globl _io_apic_read _io_apic_read: - movl 4(%esp), %ecx /* APIC # */ - movl _ioapic(,%ecx,4), %edx /* APIC base register address */ - movl 8(%esp), %eax /* target register index */ - movl %eax, (%edx) /* write the target register index */ - movl 16(%edx), %eax /* read the APIC register data */ + movl 4(%esp), %ecx /* APIC # */ + movl _ioapic(,%ecx,4), %edx /* APIC base register address */ + movl 8(%esp), %eax /* target register index */ + movl %eax, (%edx) /* write the target register index */ + movl 16(%edx), %eax /* read the APIC register data */ ret /* %eax = register value */ /* @@ -449,12 +470,12 @@ _io_apic_read: .align 2 .globl _io_apic_write _io_apic_write: - movl 4(%esp), %ecx /* APIC # */ - movl _ioapic(,%ecx,4), %edx /* APIC base register address */ - movl 8(%esp), %eax /* target register index */ - movl %eax, (%edx) /* write the target register index */ - movl 12(%esp), %eax /* target register value */ - movl %eax, 16(%edx) /* write the APIC register data */ + movl 4(%esp), %ecx /* APIC # */ + movl _ioapic(,%ecx,4), %edx /* APIC base register address */ + movl 8(%esp), %eax /* target register index */ + movl %eax, (%edx) /* write the target register index */ + movl 12(%esp), %eax /* target register value */ + movl %eax, 16(%edx) /* write the APIC register data */ ret /* %eax = void */ /* @@ -463,7 +484,7 @@ _io_apic_write: .align 2 .globl _apic_eoi _apic_eoi: - movl $0, _lapic+0xb0 + movl $0, _lapic+0xb0 ret @@ -477,11 +498,11 @@ _apic_eoi: .align 2 .globl _test_and_set _test_and_set: - movl 4(%esp), %eax /* get the address of the lock */ + movl 4(%esp), %eax /* get the address of the lock */ lock /* do it atomically */ - btsl $0, (%eax) /* set 0th bit, old value in CF */ - setc %al /* previous value into %al */ - movzbl %al, %eax /* return as an int */ + btsl $0, (%eax) /* set 0th bit, old value in CF */ + setc %al /* previous value into %al */ + movzbl %al, %eax /* return as an int */ ret /****************************************************************************** @@ -511,9 +532,13 @@ _test_and_set: .align 2 .globl _s_lock_init _s_lock_init: - movl 4(%esp), %eax /* get the address of the lock */ - xorl %ecx, %ecx /* cheap clear */ - xchgl %ecx, (%eax) /* in case its not 32bit aligned */ + movl 4(%esp), %eax /* get the address of the lock */ +#ifdef ANAL_SIMPLE_LOCK + xorl %ecx, %ecx /* cheap clear */ + xchgl %ecx, (%eax) /* in case its not 32bit aligned */ +#else + movl $0, (%eax) +#endif /* ANAL_SIMPLE_LOCK */ ret @@ -524,15 +549,26 @@ _s_lock_init: * while (test_and_set(&lkp->lock_data)) * continue; * } + * + * Note: + * If the acquire fails we do a loop of reads waiting for the lock to + * become free instead of continually beating on the lock with btsl. + * The theory here is that the CPU will stay within its cache until + * a write by the other CPU updates it, instead of continually updating + * the local cache (and thus causing external bus writes) with repeated + * writes to the lock. */ .align 2 .globl _s_lock _s_lock: - movl 4(%esp), %eax /* get the address of the lock */ + movl 4(%esp), %eax /* get the address of the lock */ 1: lock /* do it atomically */ - btsl $0, (%eax) /* set 0th bit, old value in CF */ - jc 1b /* already set, try again */ - ret + btsl $0, (%eax) /* set 0th bit, old value in CF */ + jnc 3f /* it was clear, return */ +2: btl $0, (%eax) /* wait to empty */ + jc 2b /* still set... */ + jmp 1b /* empty again, try once more */ +3: ret /* @@ -545,11 +581,11 @@ _s_lock: .align 2 .globl _s_lock_try _s_lock_try: - movl 4(%esp), %eax /* get the address of the lock */ + movl 4(%esp), %eax /* get the address of the lock */ lock /* do it atomically */ - btsl $0, (%eax) /* set 0th bit, old value in CF */ - setnc %al /* 1 if previous value was 0 */ - movzbl %al, %eax /* convert to an int */ + btsl $0, (%eax) /* set 0th bit, old value in CF */ + setnc %al /* 1 if previous value was 0 */ + movzbl %al, %eax /* convert to an int */ ret @@ -563,7 +599,11 @@ _s_lock_try: .align 2 .globl _s_unlock _s_unlock: - movl 4(%esp), %eax /* get the address of the lock */ - xorl %ecx, %ecx /* cheap clear */ - xchgl %ecx, (%eax) /* in case its not 32bit aligned */ + movl 4(%esp), %eax /* get the address of the lock */ +#ifdef ANAL_SIMPLE_LOCK + xorl %ecx, %ecx /* cheap clear */ + xchgl %ecx, (%eax) /* in case its not 32bit aligned */ +#else + movl $0, (%eax) +#endif /* ANAL_SIMPLE_LOCK */ ret diff --git a/sys/i386/isa/apic_vector.s b/sys/i386/isa/apic_vector.s index 9ea96ca..cb4c4ff 100644 --- a/sys/i386/isa/apic_vector.s +++ b/sys/i386/isa/apic_vector.s @@ -1,6 +1,6 @@ /* * from: vector.s, 386BSD 0.1 unknown origin - * $Id: apic_vector.s,v 1.12 1997/07/18 19:47:13 smp Exp smp $ + * $Id: apic_vector.s,v 1.14 1997/07/23 20:18:14 smp Exp smp $ */ @@ -16,18 +16,21 @@ /* * 'lazy masking' code suggested by Bruce Evans <bde@zeta.org.au> */ +#if 1 #define MAYBE_MASK_IRQ(irq_num) \ - testl $IRQ_BIT(irq_num),iactive ; /* lazy masking */ \ - je 1f ; /* NOT currently active */ \ - orl $IRQ_BIT(irq_num),_imen ; /* set the mask bit */ \ + lock ; /* MP-safe */ \ + btsl $(irq_num),iactive ; /* lazy masking */ \ + jnc 1f ; /* NOT active */ \ + IMASK_LOCK ; /* enter critical reg */\ + orl $IRQ_BIT(irq_num),_apic_imen ; /* set the mask bit */ \ movl _ioapic,%ecx ; /* ioapic[0]addr */ \ movl $REDTBL_IDX(irq_num),(%ecx) ; /* write the index */ \ movl IOAPIC_WINDOW(%ecx),%eax ; /* current value */ \ orl $IOART_INTMASK,%eax ; /* set the mask */ \ movl %eax,IOAPIC_WINDOW(%ecx) ; /* new value */ \ - movl $0, lapic_eoi ; \ - orl $IRQ_BIT(irq_num), _ipending ; \ - REL_MPLOCK ; /* SMP release global lock */ \ + orl $IRQ_BIT(irq_num), _ipending ; /* set _ipending bit */ \ + movl $0, lapic_eoi ; /* do the EOI */ \ + IMASK_UNLOCK ; /* exit critical reg */ \ popl %es ; \ popl %ds ; \ popal ; \ @@ -35,16 +38,59 @@ iret ; \ ; \ ALIGN_TEXT ; \ -1: ; \ - orl $IRQ_BIT(irq_num),iactive - +1: GET_MPLOCK /* SMP Spin lock */ +#else +#define MAYBE_MASK_IRQ(irq_num) \ + GET_MPLOCK ; /* SMP Spin lock */ \ + lock ; /* MP-safe */ \ + btsl $(irq_num),iactive ; /* lazy masking */ \ + jnc 1f ; /* NOT active */ \ + /* XXX atomic access */ \ + orl $IRQ_BIT(irq_num),_apic_imen ; /* set the mask bit */ \ + movl _ioapic,%ecx ; /* ioapic[0]addr */ \ + movl $REDTBL_IDX(irq_num),(%ecx) ; /* write the index */ \ + movl IOAPIC_WINDOW(%ecx),%eax ; /* current value */ \ + orl $IOART_INTMASK,%eax ; /* set the mask */ \ + movl %eax,IOAPIC_WINDOW(%ecx) ; /* new value */ \ + movl $0, lapic_eoi ; /* do the EOI */ \ + orl $IRQ_BIT(irq_num), _ipending ; /* set _ipending bit */ \ + REL_MPLOCK ; /* release SMP GL */ \ + popl %es ; \ + popl %ds ; \ + popal ; \ + addl $4+4,%esp ; \ + iret ; \ +; \ + ALIGN_TEXT ; \ +1: +#endif +#if 1 #define MAYBE_UNMASK_IRQ(irq_num) \ - cli ; /* must unmask _imen and icu atomically */ \ + cli ; /* must unmask _apic_imen and IO APIC atomically */ \ + lock ; /* MP-safe */ \ andl $~IRQ_BIT(irq_num),iactive ; \ - testl $IRQ_BIT(irq_num),_imen ; \ + IMASK_LOCK ; /* enter critical reg */\ + testl $IRQ_BIT(irq_num),_apic_imen ; \ je 2f ; \ - andl $~IRQ_BIT(irq_num),_imen ; /* clear mask bit */ \ + andl $~IRQ_BIT(irq_num),_apic_imen ; /* clear mask bit */ \ + movl _ioapic,%ecx ; /* ioapic[0]addr */ \ + movl $REDTBL_IDX(irq_num),(%ecx) ; /* write the index */ \ + movl IOAPIC_WINDOW(%ecx),%eax ; /* current value */ \ + andl $~IOART_INTMASK,%eax ; /* clear the mask */ \ + movl %eax,IOAPIC_WINDOW(%ecx) ; /* new value */ \ +2: ; \ + IMASK_UNLOCK ; /* exit critical reg */ \ + sti /* XXX _doreti repeats the cli/sti */ +#else +#define MAYBE_UNMASK_IRQ(irq_num) \ + cli ; /* must unmask _apic_imen and IO APIC atomically */ \ + lock ; /* MP-safe */ \ + andl $~IRQ_BIT(irq_num),iactive ; \ + /* XXX atomic access */ \ + testl $IRQ_BIT(irq_num),_apic_imen ; \ + je 2f ; \ + andl $~IRQ_BIT(irq_num),_apic_imen ; /* clear mask bit */ \ movl _ioapic,%ecx ; /* ioapic[0]addr */ \ movl $REDTBL_IDX(irq_num),(%ecx) ; /* write the index */ \ movl IOAPIC_WINDOW(%ecx),%eax ; /* current value */ \ @@ -52,6 +98,7 @@ movl %eax,IOAPIC_WINDOW(%ecx) ; /* new value */ \ 2: ; \ sti ; /* XXX _doreti repeats the cli/sti */ +#endif /* @@ -131,7 +178,6 @@ IDTVEC(vec_name) ; \ movl $KDSEL,%eax ; /* ... and reload with kernel's ... */ \ movl %ax,%ds ; /* ... early for obsolete reasons */ \ movl %ax,%es ; \ - GET_MPLOCK ; /* SMP Spin lock */ \ MAYBE_MASK_IRQ(irq_num) ; \ movl $0, lapic_eoi ; \ movl _cpl,%eax ; \ diff --git a/sys/i386/isa/ipl.s b/sys/i386/isa/ipl.s index daf1760..5382ff3 100644 --- a/sys/i386/isa/ipl.s +++ b/sys/i386/isa/ipl.s @@ -36,7 +36,7 @@ * * @(#)ipl.s * - * $Id: ipl.s,v 1.2 1997/05/31 09:03:52 peter Exp $ + * $Id: ipl.s,v 1.2 1997/07/23 18:41:25 smp Exp smp $ */ @@ -52,9 +52,11 @@ .globl _cpl _cpl: .long HWI_MASK | SWI_MASK +#ifndef APIC_IO /* interrupt mask enable (all h/w off) */ .globl _imen _imen: .long HWI_MASK +#endif /* APIC_IO */ .globl _tty_imask _tty_imask: .long 0 diff --git a/sys/kern/subr_smp.c b/sys/kern/subr_smp.c index c9724a1..8a3f0c2 100644 --- a/sys/kern/subr_smp.c +++ b/sys/kern/subr_smp.c @@ -22,7 +22,7 @@ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * - * $Id: mp_machdep.c,v 1.18 1997/07/20 18:13:01 smp Exp smp $ + * $Id: mp_machdep.c,v 1.20 1997/07/23 20:20:21 smp Exp smp $ */ #include "opt_smp.h" @@ -248,6 +248,7 @@ static void mp_enable(u_int boot_addr); static int mptable_pass1(void); static int mptable_pass2(void); static void default_mp_table(int type); +static void init_locks(void); static int start_all_aps(u_int boot_addr); static void install_ap_tramp(u_int boot_addr); static int start_ap(int logicalCpu, u_int boot_addr); @@ -498,6 +499,9 @@ mp_enable(u_int boot_addr) #endif /* APIC_IO */ + /* initialize all SMP locks */ + init_locks(); + /* start each Application Processor */ start_all_aps(boot_addr); @@ -1415,6 +1419,23 @@ default_mp_table(int type) /* + * initialize all the SMP locks + */ +static void +init_locks(void) +{ + /* + * Get the initial mp_lock with a count of 1 for the BSP. + * This uses a LOGICAL cpu ID, ie BSP == 0. + */ + mp_lock = 0x00000001; + + /* locks the IO APIC and apic_imen accesses */ + s_lock_init(&imen_lock); +} + + +/* * start each AP in our list */ static int @@ -1429,15 +1450,6 @@ start_all_aps(u_int boot_addr) POSTCODE(START_ALL_APS_POST); - /** - * NOTE: this needs further thought: - * where does it get released? - * should it be set to empy? - * - * get the initial mp_lock with a count of 1 for the BSP - */ - mp_lock = 1; /* this uses a LOGICAL cpu ID, ie BSP == 0 */ - /* initialize BSP's local APIC */ apic_initialize(); |