summaryrefslogtreecommitdiffstats
path: root/sys/i386/isa
diff options
context:
space:
mode:
Diffstat (limited to 'sys/i386/isa')
-rw-r--r--sys/i386/isa/apic_ipl.s10
-rw-r--r--sys/i386/isa/apic_vector.s118
-rw-r--r--sys/i386/isa/atpic_vector.s20
-rw-r--r--sys/i386/isa/icu_ipl.s4
-rw-r--r--sys/i386/isa/icu_vector.s20
-rw-r--r--sys/i386/isa/ipl.s10
-rw-r--r--sys/i386/isa/vector.s12
7 files changed, 97 insertions, 97 deletions
diff --git a/sys/i386/isa/apic_ipl.s b/sys/i386/isa/apic_ipl.s
index 5abd500..8b1f703 100644
--- a/sys/i386/isa/apic_ipl.s
+++ b/sys/i386/isa/apic_ipl.s
@@ -39,8 +39,8 @@
* imen_dump()
*/
.p2align 2 /* MUST be 32bit aligned */
- .globl _apic_imen
-_apic_imen:
+ .globl apic_imen
+apic_imen:
.long HWI_MASK
.text
@@ -59,7 +59,7 @@ _apic_imen:
andl %eax, %eax ; \
jz 1f ; \
pushl $bad_mask ; \
- call _panic ; \
+ call panic ; \
1:
bad_mask: .asciz "bad mask"
@@ -80,7 +80,7 @@ ENTRY(INTREN)
movl 8(%esp), %eax /* mask into %eax */
bsfl %eax, %ecx /* get pin index */
- btrl %ecx, _apic_imen /* update _apic_imen */
+ btrl %ecx, apic_imen /* update apic_imen */
QUALIFY_MASK
@@ -112,7 +112,7 @@ ENTRY(INTRDIS)
movl 8(%esp), %eax /* mask into %eax */
bsfl %eax, %ecx /* get pin index */
- btsl %ecx, _apic_imen /* update _apic_imen */
+ btsl %ecx, apic_imen /* update apic_imen */
QUALIFY_MASK
diff --git a/sys/i386/isa/apic_vector.s b/sys/i386/isa/apic_vector.s
index 68b6c77..007d52f 100644
--- a/sys/i386/isa/apic_vector.s
+++ b/sys/i386/isa/apic_vector.s
@@ -50,27 +50,27 @@ IDTVEC(vec_name) ; \
FAKE_MCOUNT(13*4(%esp)) ; \
movl PCPU(CURPROC),%ebx ; \
incl P_INTR_NESTING_LEVEL(%ebx) ; \
- pushl _intr_unit + (irq_num) * 4 ; \
- call *_intr_handler + (irq_num) * 4 ; /* do the work ASAP */ \
+ pushl intr_unit + (irq_num) * 4 ; \
+ call *intr_handler + (irq_num) * 4 ; /* do the work ASAP */ \
addl $4, %esp ; \
- movl $0, _lapic+LA_EOI ; \
+ movl $0, lapic+LA_EOI ; \
lock ; \
- incl _cnt+V_INTR ; /* book-keeping can wait */ \
- movl _intr_countp + (irq_num) * 4, %eax ; \
+ incl cnt+V_INTR ; /* book-keeping can wait */ \
+ movl intr_countp + (irq_num) * 4, %eax ; \
lock ; \
incl (%eax) ; \
decl P_INTR_NESTING_LEVEL(%ebx) ; \
MEXITCOUNT ; \
- jmp _doreti
+ jmp doreti
#define IOAPICADDR(irq_num) CNAME(int_to_apicintpin) + 16 * (irq_num) + 8
#define REDIRIDX(irq_num) CNAME(int_to_apicintpin) + 16 * (irq_num) + 12
#define MASK_IRQ(irq_num) \
IMASK_LOCK ; /* into critical reg */ \
- testl $IRQ_BIT(irq_num), _apic_imen ; \
+ testl $IRQ_BIT(irq_num), apic_imen ; \
jne 7f ; /* masked, don't mask */ \
- orl $IRQ_BIT(irq_num), _apic_imen ; /* set the mask bit */ \
+ orl $IRQ_BIT(irq_num), apic_imen ; /* set the mask bit */ \
movl IOAPICADDR(irq_num), %ecx ; /* ioapic addr */ \
movl REDIRIDX(irq_num), %eax ; /* get the index */ \
movl %eax, (%ecx) ; /* write the index */ \
@@ -85,7 +85,7 @@ IDTVEC(vec_name) ; \
* and the EOI cycle would cause redundant INTs to occur.
*/
#define MASK_LEVEL_IRQ(irq_num) \
- testl $IRQ_BIT(irq_num), _apic_pin_trigger ; \
+ testl $IRQ_BIT(irq_num), apic_pin_trigger ; \
jz 9f ; /* edge, don't mask */ \
MASK_IRQ(irq_num) ; \
9:
@@ -93,18 +93,18 @@ IDTVEC(vec_name) ; \
#ifdef APIC_INTR_REORDER
#define EOI_IRQ(irq_num) \
- movl _apic_isrbit_location + 8 * (irq_num), %eax ; \
+ movl apic_isrbit_location + 8 * (irq_num), %eax ; \
movl (%eax), %eax ; \
- testl _apic_isrbit_location + 4 + 8 * (irq_num), %eax ; \
+ testl apic_isrbit_location + 4 + 8 * (irq_num), %eax ; \
jz 9f ; /* not active */ \
- movl $0, _lapic+LA_EOI ; \
+ movl $0, lapic+LA_EOI ; \
9:
#else
#define EOI_IRQ(irq_num) \
- testl $IRQ_BIT(irq_num), _lapic+LA_ISR1; \
+ testl $IRQ_BIT(irq_num), lapic+LA_ISR1; \
jz 9f ; /* not active */ \
- movl $0, _lapic+LA_EOI; \
+ movl $0, lapic+LA_EOI; \
9:
#endif
@@ -160,12 +160,12 @@ __CONCAT(Xresume,irq_num): ; \
FAKE_MCOUNT(13*4(%esp)) ; /* XXX avoid dbl cnt */ \
pushl $irq_num; /* pass the IRQ */ \
sti ; \
- call _sched_ithd ; \
+ call sched_ithd ; \
addl $4, %esp ; /* discard the parameter */ \
; \
decl P_INTR_NESTING_LEVEL(%ebx) ; \
MEXITCOUNT ; \
- jmp _doreti
+ jmp doreti
/*
* Handle "spurious INTerrupts".
@@ -176,8 +176,8 @@ __CONCAT(Xresume,irq_num): ; \
*/
.text
SUPERALIGN_TEXT
- .globl _Xspuriousint
-_Xspuriousint:
+ .globl Xspuriousint
+Xspuriousint:
/* No EOI cycle used here */
@@ -189,8 +189,8 @@ _Xspuriousint:
*/
.text
SUPERALIGN_TEXT
- .globl _Xinvltlb
-_Xinvltlb:
+ .globl Xinvltlb
+Xinvltlb:
pushl %eax
#ifdef COUNT_XINVLTLB_HITS
@@ -207,7 +207,7 @@ _Xinvltlb:
movl %eax, %cr3
ss /* stack segment, avoid %ds load */
- movl $0, _lapic+LA_EOI /* End Of Interrupt to APIC */
+ movl $0, lapic+LA_EOI /* End Of Interrupt to APIC */
popl %eax
iret
@@ -229,11 +229,11 @@ _Xinvltlb:
.text
SUPERALIGN_TEXT
- .globl _Xcpucheckstate
- .globl _checkstate_cpustate
- .globl _checkstate_curproc
- .globl _checkstate_pc
-_Xcpucheckstate:
+ .globl Xcpucheckstate
+ .globl checkstate_cpustate
+ .globl checkstate_curproc
+ .globl checkstate_pc
+Xcpucheckstate:
pushl %eax
pushl %ebx
pushl %ds /* save current data segment */
@@ -244,7 +244,7 @@ _Xcpucheckstate:
movl $KPSEL, %eax
mov %ax, %fs
- movl $0, _lapic+LA_EOI /* End Of Interrupt to APIC */
+ movl $0, lapic+LA_EOI /* End Of Interrupt to APIC */
movl $0, %ebx
movl 20(%esp), %eax
@@ -256,15 +256,15 @@ _Xcpucheckstate:
incl %ebx /* system or interrupt */
1:
movl PCPU(CPUID), %eax
- movl %ebx, _checkstate_cpustate(,%eax,4)
+ movl %ebx, checkstate_cpustate(,%eax,4)
movl PCPU(CURPROC), %ebx
- movl %ebx, _checkstate_curproc(,%eax,4)
+ movl %ebx, checkstate_curproc(,%eax,4)
movl 16(%esp), %ebx
- movl %ebx, _checkstate_pc(,%eax,4)
+ movl %ebx, checkstate_pc(,%eax,4)
lock /* checkstate_probed_cpus |= (1<<id) */
- btsl %eax, _checkstate_probed_cpus
+ btsl %eax, checkstate_probed_cpus
popl %fs
popl %ds /* restore previous data segment */
@@ -284,8 +284,8 @@ _Xcpucheckstate:
.text
SUPERALIGN_TEXT
- .globl _Xcpuast
-_Xcpuast:
+ .globl Xcpuast
+Xcpuast:
PUSH_FRAME
movl $KDSEL, %eax
mov %ax, %ds /* use KERNEL data segment */
@@ -295,11 +295,11 @@ _Xcpuast:
movl PCPU(CPUID), %eax
lock /* checkstate_need_ast &= ~(1<<id) */
- btrl %eax, _checkstate_need_ast
- movl $0, _lapic+LA_EOI /* End Of Interrupt to APIC */
+ btrl %eax, checkstate_need_ast
+ movl $0, lapic+LA_EOI /* End Of Interrupt to APIC */
lock
- btsl %eax, _checkstate_pending_ast
+ btsl %eax, checkstate_pending_ast
jc 1f
FAKE_MCOUNT(13*4(%esp))
@@ -310,7 +310,7 @@ _Xcpuast:
movl PCPU(CPUID), %eax
lock
- btrl %eax, _checkstate_pending_ast
+ btrl %eax, checkstate_pending_ast
lock
btrl %eax, CNAME(resched_cpus)
jnc 2f
@@ -322,7 +322,7 @@ _Xcpuast:
lock
incl CNAME(cpuast_cnt)
MEXITCOUNT
- jmp _doreti
+ jmp doreti
1:
/* We are already in the process of delivering an ast for this CPU */
POP_FRAME
@@ -338,8 +338,8 @@ _Xcpuast:
.text
SUPERALIGN_TEXT
- .globl _Xcpustop
-_Xcpustop:
+ .globl Xcpustop
+Xcpustop:
pushl %ebp
movl %esp, %ebp
pushl %eax
@@ -353,7 +353,7 @@ _Xcpustop:
movl $KPSEL, %eax
mov %ax, %fs
- movl $0, _lapic+LA_EOI /* End Of Interrupt to APIC */
+ movl $0, lapic+LA_EOI /* End Of Interrupt to APIC */
movl PCPU(CPUID), %eax
imull $PCB_SIZE, %eax
@@ -366,15 +366,15 @@ _Xcpustop:
movl PCPU(CPUID), %eax
lock
- btsl %eax, _stopped_cpus /* stopped_cpus |= (1<<id) */
+ btsl %eax, stopped_cpus /* stopped_cpus |= (1<<id) */
1:
- btl %eax, _started_cpus /* while (!(started_cpus & (1<<id))) */
+ btl %eax, started_cpus /* while (!(started_cpus & (1<<id))) */
jnc 1b
lock
- btrl %eax, _started_cpus /* started_cpus &= ~(1<<id) */
+ btrl %eax, started_cpus /* started_cpus &= ~(1<<id) */
lock
- btrl %eax, _stopped_cpus /* stopped_cpus &= ~(1<<id) */
+ btrl %eax, stopped_cpus /* stopped_cpus &= ~(1<<id) */
test %eax, %eax
jnz 2f
@@ -472,8 +472,8 @@ MCOUNT_LABEL(eintr)
*/
.text
SUPERALIGN_TEXT
- .globl _Xrendezvous
-_Xrendezvous:
+ .globl Xrendezvous
+Xrendezvous:
PUSH_FRAME
movl $KDSEL, %eax
mov %ax, %ds /* use KERNEL data segment */
@@ -481,9 +481,9 @@ _Xrendezvous:
movl $KPSEL, %eax
mov %ax, %fs
- call _smp_rendezvous_action
+ call smp_rendezvous_action
- movl $0, _lapic+LA_EOI /* End Of Interrupt to APIC */
+ movl $0, lapic+LA_EOI /* End Of Interrupt to APIC */
POP_FRAME
iret
@@ -497,21 +497,21 @@ _xhits:
#endif /* COUNT_XINVLTLB_HITS */
/* variables used by stop_cpus()/restart_cpus()/Xcpustop */
- .globl _stopped_cpus, _started_cpus
-_stopped_cpus:
+ .globl stopped_cpus, started_cpus
+stopped_cpus:
.long 0
-_started_cpus:
+started_cpus:
.long 0
#ifdef BETTER_CLOCK
- .globl _checkstate_probed_cpus
-_checkstate_probed_cpus:
+ .globl checkstate_probed_cpus
+checkstate_probed_cpus:
.long 0
#endif /* BETTER_CLOCK */
- .globl _checkstate_need_ast
-_checkstate_need_ast:
+ .globl checkstate_need_ast
+checkstate_need_ast:
.long 0
-_checkstate_pending_ast:
+checkstate_pending_ast:
.long 0
.globl CNAME(resched_cpus)
.globl CNAME(want_resched_cnt)
@@ -526,8 +526,8 @@ CNAME(cpuast_cnt):
CNAME(cpustop_restartfunc):
.long 0
- .globl _apic_pin_trigger
-_apic_pin_trigger:
+ .globl apic_pin_trigger
+apic_pin_trigger:
.long 0
.text
diff --git a/sys/i386/isa/atpic_vector.s b/sys/i386/isa/atpic_vector.s
index 0a3456f..ca0ebc0 100644
--- a/sys/i386/isa/atpic_vector.s
+++ b/sys/i386/isa/atpic_vector.s
@@ -63,16 +63,16 @@ IDTVEC(vec_name) ; \
FAKE_MCOUNT((12+ACTUALLY_PUSHED)*4(%esp)) ; \
movl PCPU(CURPROC),%ebx ; \
incl P_INTR_NESTING_LEVEL(%ebx) ; \
- pushl _intr_unit + (irq_num) * 4 ; \
- call *_intr_handler + (irq_num) * 4 ; /* do the work ASAP */ \
+ pushl intr_unit + (irq_num) * 4 ; \
+ call *intr_handler + (irq_num) * 4 ; /* do the work ASAP */ \
enable_icus ; /* (re)enable ASAP (helps edge trigger?) */ \
addl $4,%esp ; \
- incl _cnt+V_INTR ; /* book-keeping can wait */ \
- movl _intr_countp + (irq_num) * 4,%eax ; \
+ incl cnt+V_INTR ; /* book-keeping can wait */ \
+ movl intr_countp + (irq_num) * 4,%eax ; \
incl (%eax) ; \
decl P_INTR_NESTING_LEVEL(%ebx) ; \
MEXITCOUNT ; \
- jmp _doreti
+ jmp doreti
/*
* Slow, threaded interrupts.
@@ -99,9 +99,9 @@ IDTVEC(vec_name) ; \
mov $KPSEL,%ax ; \
mov %ax,%fs ; \
maybe_extra_ipending ; \
- movb _imen + IRQ_BYTE(irq_num),%al ; \
+ movb imen + IRQ_BYTE(irq_num),%al ; \
orb $IRQ_BIT(irq_num),%al ; \
- movb %al,_imen + IRQ_BYTE(irq_num) ; \
+ movb %al,imen + IRQ_BYTE(irq_num) ; \
outb %al,$icu+ICU_IMR_OFFSET ; \
enable_icus ; \
movl PCPU(CURPROC),%ebx ; \
@@ -110,13 +110,13 @@ __CONCAT(Xresume,irq_num): ; \
FAKE_MCOUNT(13*4(%esp)) ; /* XXX late to avoid double count */ \
pushl $irq_num; /* pass the IRQ */ \
sti ; \
- call _sched_ithd ; \
+ call sched_ithd ; \
addl $4, %esp ; /* discard the parameter */ \
decl P_INTR_NESTING_LEVEL(%ebx) ; \
MEXITCOUNT ; \
/* We could usually avoid the following jmp by inlining some of */ \
- /* _doreti, but it's probably better to use less cache. */ \
- jmp _doreti /* and catch up inside doreti */
+ /* doreti, but it's probably better to use less cache. */ \
+ jmp doreti /* and catch up inside doreti */
MCOUNT_LABEL(bintr)
FAST_INTR(0,fastintr0, ENABLE_ICU1)
diff --git a/sys/i386/isa/icu_ipl.s b/sys/i386/isa/icu_ipl.s
index 6ea9052..36804aa 100644
--- a/sys/i386/isa/icu_ipl.s
+++ b/sys/i386/isa/icu_ipl.s
@@ -41,8 +41,8 @@
ALIGN_DATA
/* interrupt mask enable (all h/w off) */
- .globl _imen
-_imen: .long HWI_MASK
+ .globl imen
+imen: .long HWI_MASK
.text
SUPERALIGN_TEXT
diff --git a/sys/i386/isa/icu_vector.s b/sys/i386/isa/icu_vector.s
index 0a3456f..ca0ebc0 100644
--- a/sys/i386/isa/icu_vector.s
+++ b/sys/i386/isa/icu_vector.s
@@ -63,16 +63,16 @@ IDTVEC(vec_name) ; \
FAKE_MCOUNT((12+ACTUALLY_PUSHED)*4(%esp)) ; \
movl PCPU(CURPROC),%ebx ; \
incl P_INTR_NESTING_LEVEL(%ebx) ; \
- pushl _intr_unit + (irq_num) * 4 ; \
- call *_intr_handler + (irq_num) * 4 ; /* do the work ASAP */ \
+ pushl intr_unit + (irq_num) * 4 ; \
+ call *intr_handler + (irq_num) * 4 ; /* do the work ASAP */ \
enable_icus ; /* (re)enable ASAP (helps edge trigger?) */ \
addl $4,%esp ; \
- incl _cnt+V_INTR ; /* book-keeping can wait */ \
- movl _intr_countp + (irq_num) * 4,%eax ; \
+ incl cnt+V_INTR ; /* book-keeping can wait */ \
+ movl intr_countp + (irq_num) * 4,%eax ; \
incl (%eax) ; \
decl P_INTR_NESTING_LEVEL(%ebx) ; \
MEXITCOUNT ; \
- jmp _doreti
+ jmp doreti
/*
* Slow, threaded interrupts.
@@ -99,9 +99,9 @@ IDTVEC(vec_name) ; \
mov $KPSEL,%ax ; \
mov %ax,%fs ; \
maybe_extra_ipending ; \
- movb _imen + IRQ_BYTE(irq_num),%al ; \
+ movb imen + IRQ_BYTE(irq_num),%al ; \
orb $IRQ_BIT(irq_num),%al ; \
- movb %al,_imen + IRQ_BYTE(irq_num) ; \
+ movb %al,imen + IRQ_BYTE(irq_num) ; \
outb %al,$icu+ICU_IMR_OFFSET ; \
enable_icus ; \
movl PCPU(CURPROC),%ebx ; \
@@ -110,13 +110,13 @@ __CONCAT(Xresume,irq_num): ; \
FAKE_MCOUNT(13*4(%esp)) ; /* XXX late to avoid double count */ \
pushl $irq_num; /* pass the IRQ */ \
sti ; \
- call _sched_ithd ; \
+ call sched_ithd ; \
addl $4, %esp ; /* discard the parameter */ \
decl P_INTR_NESTING_LEVEL(%ebx) ; \
MEXITCOUNT ; \
/* We could usually avoid the following jmp by inlining some of */ \
- /* _doreti, but it's probably better to use less cache. */ \
- jmp _doreti /* and catch up inside doreti */
+ /* doreti, but it's probably better to use less cache. */ \
+ jmp doreti /* and catch up inside doreti */
MCOUNT_LABEL(bintr)
FAST_INTR(0,fastintr0, ENABLE_ICU1)
diff --git a/sys/i386/isa/ipl.s b/sys/i386/isa/ipl.s
index 0c7cfe5..e7088ce 100644
--- a/sys/i386/isa/ipl.s
+++ b/sys/i386/isa/ipl.s
@@ -53,22 +53,22 @@
* Handle return from interrupts, traps and syscalls.
*/
SUPERALIGN_TEXT
- .type _doreti,@function
-_doreti:
+ .type doreti,@function
+doreti:
- FAKE_MCOUNT(_bintr) /* init "from" _bintr -> _doreti */
+ FAKE_MCOUNT(bintr) /* init "from" bintr -> doreti */
doreti_next:
/* Check for ASTs that can be handled now. */
testb $SEL_RPL_MASK,TF_CS(%esp) /* are we in user mode? */
jne doreti_ast /* yes, do it now. */
testl $PSL_VM,TF_EFLAGS(%esp) /* kernel mode */
je doreti_exit /* and not VM86 mode, defer */
- cmpl $1,_in_vm86call /* are we in a VM86 call? */
+ cmpl $1,in_vm86call /* are we in a VM86 call? */
je doreti_exit /* no, defer */
doreti_ast:
pushl %esp /* pass a pointer to the trapframe */
- call _ast
+ call ast
add $4, %esp
/*
diff --git a/sys/i386/isa/vector.s b/sys/i386/isa/vector.s
index 79f2320..68b09b0 100644
--- a/sys/i386/isa/vector.s
+++ b/sys/i386/isa/vector.s
@@ -45,15 +45,15 @@
*/
#define NR_INTRNAMES (1 + ICU_LEN + 2 * ICU_LEN)
- .globl _intrcnt, _eintrcnt
-_intrcnt:
+ .globl intrcnt, eintrcnt
+intrcnt:
.space NR_INTRNAMES * 4
-_eintrcnt:
+eintrcnt:
- .globl _intrnames, _eintrnames
-_intrnames:
+ .globl intrnames, eintrnames
+intrnames:
.space NR_INTRNAMES * 16
-_eintrnames:
+eintrnames:
.text
/*
OpenPOWER on IntegriCloud