version 1.1.2.5, 2002/12/19 00:33:48 |
version 1.17, 2003/09/10 19:48:49 |
|
|
* 2. Redistributions in binary form must reproduce the above copyright |
* 2. Redistributions in binary form must reproduce the above copyright |
* notice, this list of conditions and the following disclaimer in the |
* notice, this list of conditions and the following disclaimer in the |
* documentation and/or other materials provided with the distribution. |
* documentation and/or other materials provided with the distribution. |
* 3. All advertising materials mentioning features or use of this software |
* 3. Neither the name of the University nor the names of its contributors |
* must display the following acknowledgement: |
|
* This product includes software developed by the University of |
|
* California, Berkeley and its contributors. |
|
* 4. Neither the name of the University nor the names of its contributors |
|
* may be used to endorse or promote products derived from this software |
* may be used to endorse or promote products derived from this software |
* without specific prior written permission. |
* without specific prior written permission. |
* |
* |
|
|
#include "apm.h" |
#include "apm.h" |
#include "lapic.h" |
#include "lapic.h" |
#include "ioapic.h" |
#include "ioapic.h" |
|
#include "ksyms.h" |
|
|
#include <sys/errno.h> |
#include <sys/errno.h> |
#include <sys/syscall.h> |
#include <sys/syscall.h> |
|
|
* XXX 4 == sizeof pde |
* XXX 4 == sizeof pde |
*/ |
*/ |
.set _C_LABEL(PTmap),(PDSLOT_PTE << PDSHIFT) |
.set _C_LABEL(PTmap),(PDSLOT_PTE << PDSHIFT) |
.set _C_LABEL(PTD),(_C_LABEL(PTmap) + PDSLOT_PTE * NBPG) |
.set _C_LABEL(PTD),(_C_LABEL(PTmap) + PDSLOT_PTE * PAGE_SIZE) |
.set _C_LABEL(PTDpde),(_C_LABEL(PTD) + PDSLOT_PTE * 4) |
.set _C_LABEL(PTDpde),(_C_LABEL(PTD) + PDSLOT_PTE * 4) |
|
|
/* |
/* |
|
|
* XXX 4 == sizeof pde |
* XXX 4 == sizeof pde |
*/ |
*/ |
.set _C_LABEL(APTmap),(PDSLOT_APTE << PDSHIFT) |
.set _C_LABEL(APTmap),(PDSLOT_APTE << PDSHIFT) |
.set _C_LABEL(APTD),(_C_LABEL(APTmap) + PDSLOT_APTE * NBPG) |
.set _C_LABEL(APTD),(_C_LABEL(APTmap) + PDSLOT_APTE * PAGE_SIZE) |
.set _C_LABEL(APTDpde),(_C_LABEL(PTD) + PDSLOT_APTE * 4) |
.set _C_LABEL(APTDpde),(_C_LABEL(PTD) + PDSLOT_APTE * 4) |
|
|
|
|
|
|
.data |
.data |
|
|
.globl _C_LABEL(cpu) |
.globl _C_LABEL(cpu) |
.globl _C_LABEL(cpu_feature) |
|
.globl _C_LABEL(esym),_C_LABEL(boothowto) |
.globl _C_LABEL(esym),_C_LABEL(boothowto) |
.globl _C_LABEL(bootinfo),_C_LABEL(atdevbase) |
.globl _C_LABEL(bootinfo),_C_LABEL(atdevbase) |
#ifdef COMPAT_OLDBOOT |
#ifdef COMPAT_OLDBOOT |
|
|
|
|
#if NLAPIC > 0 |
#if NLAPIC > 0 |
#ifdef __ELF__ |
#ifdef __ELF__ |
.align NBPG |
.align PAGE_SIZE |
#else |
#else |
.align 12 |
.align 12 |
#endif |
#endif |
Line 214 _C_LABEL(lapic_tpr): |
|
Line 210 _C_LABEL(lapic_tpr): |
|
_C_LABEL(lapic_ppr): |
_C_LABEL(lapic_ppr): |
.space LAPIC_ISR-LAPIC_PPRI |
.space LAPIC_ISR-LAPIC_PPRI |
_C_LABEL(lapic_isr): |
_C_LABEL(lapic_isr): |
.space NBPG-LAPIC_ISR |
.space PAGE_SIZE-LAPIC_ISR |
#else |
#else |
_C_LABEL(lapic_tpr): |
_C_LABEL(lapic_tpr): |
.long 0 |
.long 0 |
Line 223 _C_LABEL(lapic_tpr): |
|
Line 219 _C_LABEL(lapic_tpr): |
|
|
|
_C_LABEL(cpu): .long 0 # are we 386, 386sx, or 486, |
_C_LABEL(cpu): .long 0 # are we 386, 386sx, or 486, |
# or Pentium, or.. |
# or Pentium, or.. |
_C_LABEL(cpu_feature): .long 0 # feature flags from 'cpuid' |
|
# instruction |
|
_C_LABEL(esym): .long 0 # ptr to end of syms |
_C_LABEL(esym): .long 0 # ptr to end of syms |
_C_LABEL(atdevbase): .long 0 # location of start of iomem in virtual |
_C_LABEL(atdevbase): .long 0 # location of start of iomem in virtual |
_C_LABEL(proc0paddr): .long 0 |
_C_LABEL(proc0paddr): .long 0 |
Line 244 _C_LABEL(biosextmem): .long REALEXTMEM |
|
Line 238 _C_LABEL(biosextmem): .long REALEXTMEM |
|
tmpstk: |
tmpstk: |
|
|
|
|
#define _RELOC(x) ((x) - KERNBASE) |
#define _RELOC(x) ((x) - KERNBASE_LOCORE) |
#define RELOC(x) _RELOC(_C_LABEL(x)) |
#define RELOC(x) _RELOC(_C_LABEL(x)) |
|
|
.text |
.text |
Line 298 start: movw $0x1234,0x472 # warm boot |
|
Line 292 start: movw $0x1234,0x472 # warm boot |
|
movl 16(%esp),%eax |
movl 16(%esp),%eax |
testl %eax,%eax |
testl %eax,%eax |
jz 1f |
jz 1f |
addl $KERNBASE,%eax |
addl $KERNBASE_LOCORE,%eax |
1: movl %eax,RELOC(esym) |
1: movl %eax,RELOC(esym) |
|
|
movl RELOC(biosextmem),%eax |
movl RELOC(biosextmem),%eax |
Line 506 try586: /* Use the `cpuid' instruction. |
|
Line 500 try586: /* Use the `cpuid' instruction. |
|
* text | data | bss | [syms] | page dir | proc0 kstack |
* text | data | bss | [syms] | page dir | proc0 kstack |
* 0 1 2 3 |
* 0 1 2 3 |
*/ |
*/ |
#define PROC0PDIR ((0) * NBPG) |
#define PROC0PDIR ((0) * PAGE_SIZE) |
#define PROC0STACK ((1) * NBPG) |
#define PROC0STACK ((1) * PAGE_SIZE) |
#define SYSMAP ((1+UPAGES) * NBPG) |
#define SYSMAP ((1+UPAGES) * PAGE_SIZE) |
#define TABLESIZE ((1+UPAGES) * NBPG) /* + nkpde * NBPG */ |
#define TABLESIZE ((1+UPAGES) * PAGE_SIZE) /* + nkpde * PAGE_SIZE */ |
|
|
/* Find end of kernel image. */ |
/* Find end of kernel image. */ |
movl $RELOC(end),%edi |
movl $RELOC(end),%edi |
#if defined(DDB) && !defined(SYMTAB_SPACE) |
#if (NKSYMS || defined(DDB) || defined(LKM)) && !defined(SYMTAB_SPACE) |
/* Save the symbols (if loaded). */ |
/* Save the symbols (if loaded). */ |
movl RELOC(esym),%eax |
movl RELOC(esym),%eax |
testl %eax,%eax |
testl %eax,%eax |
jz 1f |
jz 1f |
subl $KERNBASE,%eax |
subl $KERNBASE_LOCORE,%eax |
movl %eax,%edi |
movl %eax,%edi |
1: |
1: |
#endif |
#endif |
Line 561 try586: /* Use the `cpuid' instruction. |
|
Line 555 try586: /* Use the `cpuid' instruction. |
|
*/ |
*/ |
#define fillkpt \ |
#define fillkpt \ |
1: movl %eax,(%ebx) ; \ |
1: movl %eax,(%ebx) ; \ |
addl $NBPG,%eax ; /* increment physical address */ \ |
addl $PAGE_SIZE,%eax ; /* increment physical address */ \ |
addl $4,%ebx ; /* next pte */ \ |
addl $4,%ebx ; /* next pte */ \ |
loop 1b ; |
loop 1b ; |
|
|
Line 634 try586: /* Use the `cpuid' instruction. |
|
Line 628 try586: /* Use the `cpuid' instruction. |
|
ret |
ret |
|
|
begin: |
begin: |
/* Now running relocated at KERNBASE. Remove double mapping. */ |
/* Now running relocated at KERNBASE_LOCORE. Remove double mapping. */ |
movl _C_LABEL(nkpde),%ecx # for this many pde s, |
movl _C_LABEL(nkpde),%ecx # for this many pde s, |
leal (PROC0PDIR+0*4)(%esi),%ebx # which is where temp maps! |
leal (PROC0PDIR+0*4)(%esi),%ebx # which is where temp maps! |
addl $(KERNBASE), %ebx # now use relocated address |
addl $(KERNBASE_LOCORE), %ebx # now use relocated address |
1: movl $0,(%ebx) |
1: movl $0,(%ebx) |
addl $4,%ebx # next pde |
addl $4,%ebx # next pde |
loop 1b |
loop 1b |
|
|
/* Relocate atdevbase. */ |
/* Relocate atdevbase. */ |
movl _C_LABEL(nkpde),%edx |
movl _C_LABEL(nkpde),%edx |
shll $PGSHIFT,%edx |
shll $PGSHIFT,%edx |
addl $(TABLESIZE+KERNBASE),%edx |
addl $(TABLESIZE+KERNBASE_LOCORE),%edx |
addl %esi,%edx |
addl %esi,%edx |
movl %edx,_C_LABEL(atdevbase) |
movl %edx,_C_LABEL(atdevbase) |
|
|
/* Set up bootstrap stack. */ |
/* Set up bootstrap stack. */ |
leal (PROC0STACK+KERNBASE)(%esi),%eax |
leal (PROC0STACK+KERNBASE_LOCORE)(%esi),%eax |
movl %eax,_C_LABEL(proc0paddr) |
movl %eax,_C_LABEL(proc0paddr) |
leal (USPACE-FRAMESIZE)(%eax),%esp |
leal (USPACE-FRAMESIZE)(%eax),%esp |
movl %esi,PCB_CR3(%eax) # pcb->pcb_cr3 |
movl %esi,PCB_CR3(%eax) # pcb->pcb_cr3 |
Line 711 NENTRY(proc_trampoline) |
|
Line 705 NENTRY(proc_trampoline) |
|
/* NOTREACHED */ |
/* NOTREACHED */ |
|
|
/*****************************************************************************/ |
/*****************************************************************************/ |
|
#ifdef COMPAT_16 |
/* |
/* |
* Signal trampoline; copied to top of user stack. |
* Signal trampoline; copied to top of user stack. |
*/ |
*/ |
|
|
leal 12(%esp),%eax # get pointer to sigcontext |
leal 12(%esp),%eax # get pointer to sigcontext |
movl %eax,4(%esp) # put it in the argument slot |
movl %eax,4(%esp) # put it in the argument slot |
# fake return address already there |
# fake return address already there |
movl $SYS___sigreturn14,%eax |
movl $SYS_compat_16___sigreturn14,%eax |
int $0x80 # enter kernel with args on stack |
int $0x80 # enter kernel with args on stack |
movl $SYS_exit,%eax |
movl $SYS_exit,%eax |
int $0x80 # exit if sigreturn fails |
int $0x80 # exit if sigreturn fails |
.globl _C_LABEL(esigcode) |
.globl _C_LABEL(esigcode) |
_C_LABEL(esigcode): |
_C_LABEL(esigcode): |
|
#endif |
|
|
/*****************************************************************************/ |
/*****************************************************************************/ |
|
|
Line 1095 ENTRY(copyoutstr) |
|
Line 1090 ENTRY(copyoutstr) |
|
/* Compute number of bytes in first page. */ |
/* Compute number of bytes in first page. */ |
movl %edi,%eax |
movl %edi,%eax |
andl $PGOFSET,%eax |
andl $PGOFSET,%eax |
movl $NBPG,%ecx |
movl $PAGE_SIZE,%ecx |
subl %eax,%ecx # ecx = NBPG - (src % NBPG) |
subl %eax,%ecx # ecx = PAGE_SIZE - (src % PAGE_SIZE) |
|
|
GET_CURPCB(%eax) |
GET_CURPCB(%eax) |
movl $6f,PCB_ONFAULT(%eax) |
movl $6f,PCB_ONFAULT(%eax) |
Line 1143 ENTRY(copyoutstr) |
|
Line 1138 ENTRY(copyoutstr) |
|
jmp copystr_return |
jmp copystr_return |
|
|
4: /* Go to next page, if any. */ |
4: /* Go to next page, if any. */ |
movl $NBPG,%ecx |
movl $PAGE_SIZE,%ecx |
testl %edx,%edx |
testl %edx,%edx |
jnz 1b |
jnz 1b |
|
|
Line 1683 ENTRY(cpu_switch) |
|
Line 1678 ENTRY(cpu_switch) |
|
1: |
1: |
#endif /* DEBUG */ |
#endif /* DEBUG */ |
|
|
movl CPUVAR(CURLWP),%esi |
movl 16(%esp),%esi # current |
|
|
/* |
/* |
* Clear curlwp so that we don't accumulate system time while idle. |
* Clear curlwp so that we don't accumulate system time while idle. |
|
|
jz idle_loop |
jz idle_loop |
idle_exit: |
idle_exit: |
movl $IPL_HIGH,CPUVAR(ILEVEL) # splhigh |
movl $IPL_HIGH,CPUVAR(ILEVEL) # splhigh |
|
sti |
#if defined(MULTIPROCESSOR) || defined(LOCKDEBUG) |
#if defined(MULTIPROCESSOR) || defined(LOCKDEBUG) |
call _C_LABEL(sched_lock_idle) |
call _C_LABEL(sched_lock_idle) |
#endif |
#endif |
|
|
SET_CURLWP(%edi,%ecx) |
SET_CURLWP(%edi,%ecx) |
|
|
/* Skip context switch if same lwp. */ |
/* Skip context switch if same lwp. */ |
movl $1, %eax |
xorl %ebx,%ebx |
cmpl %edi,%esi |
cmpl %edi,%esi |
je switch_return |
je switch_return |
|
|
Line 1960 switch_restored: |
|
Line 1956 switch_restored: |
|
je 1f |
je 1f |
movl %eax,TF_EIP(%ebx) |
movl %eax,TF_EIP(%ebx) |
1: |
1: |
xor %eax,%eax |
movl $1,%ebx |
|
|
switch_return: |
switch_return: |
#if defined(MULTIPROCESSOR) || defined(LOCKDEBUG) |
#if defined(MULTIPROCESSOR) || defined(LOCKDEBUG) |
|
|
call _C_LABEL(Xspllower) # process pending interrupts |
call _C_LABEL(Xspllower) # process pending interrupts |
addl $4,%esp |
addl $4,%esp |
movl $IPL_HIGH,CPUVAR(ILEVEL) # splhigh() |
movl $IPL_HIGH,CPUVAR(ILEVEL) # splhigh() |
|
|
|
movl %ebx,%eax |
|
|
popl %edi |
popl %edi |
popl %esi |
popl %esi |
popl %ebx |
popl %ebx |
ret |
ret |
|
|
#ifdef DIAGNOSTIC |
|
NENTRY(preempt_error) |
|
pushl $1f |
|
call _C_LABEL(panic) |
|
/* NOTREACHED */ |
|
1: .asciz "cpu_preempt" |
|
#endif /* DIAGNOSTIC */ |
|
|
|
/* |
/* |
* void cpu_preempt(struct lwp *current, struct lwp *next) |
* void cpu_switchto(struct lwp *current, struct lwp *next) |
* Switch to the specified nexxt LWP. |
* Switch to the specified next LWP. |
*/ |
*/ |
ENTRY(cpu_preempt) |
ENTRY(cpu_switchto) |
pushl %ebx |
pushl %ebx |
pushl %esi |
pushl %esi |
pushl %edi |
pushl %edi |
Line 1999 ENTRY(cpu_preempt) |
|
Line 1989 ENTRY(cpu_preempt) |
|
pushl $2f |
pushl $2f |
call _C_LABEL(panic) |
call _C_LABEL(panic) |
/* NOTREACHED */ |
/* NOTREACHED */ |
2: .asciz "not splsched() in cpu_preempt!" |
2: .asciz "not splsched() in cpu_switchto!" |
1: |
1: |
#endif /* DEBUG */ |
#endif /* DEBUG */ |
|
|
movl CPUVAR(CURLWP),%esi |
movl 16(%esp),%esi # current |
movl 20(%esp),%edi # next |
movl 20(%esp),%edi # next |
|
|
/* |
/* |
Line 2012 ENTRY(cpu_preempt) |
|
Line 2002 ENTRY(cpu_preempt) |
|
* the correct queue if it happens to get called from the spllower() |
* the correct queue if it happens to get called from the spllower() |
* below and changes the priority. (See corresponding comment in |
* below and changes the priority. (See corresponding comment in |
* usrret()). |
* usrret()). |
|
* |
|
* XXX Is this necessary? We know we won't go idle. |
*/ |
*/ |
movl $0,CPUVAR(CURLWP) |
movl $0,CPUVAR(CURLWP) |
|
|
preempt_dequeue: |
|
/* |
/* |
* We're running at splhigh(), but it's otherwise okay to take |
* We're running at splhigh(), but it's otherwise okay to take |
* interrupts here. |
* interrupts here. |
*/ |
*/ |
sti |
sti |
|
|
movzbl L_PRIORITY(%edi),%eax |
|
#ifdef DIAGNOSTIC |
|
shrl $2,%eax |
|
btl %eax,_C_LABEL(sched_whichqs) |
|
jnc _C_LABEL(preempt_error) |
|
#endif /* DIAGNOSTIC */ |
|
movl L_BACK(%edi),%edx # unlink process |
|
movl $0,L_BACK(%edi) # zap reverse link to indicate off list |
|
movl L_FORW(%edi),%ecx |
|
movl %ecx,L_FORW(%edx) |
|
movl %edx,L_BACK(%ecx) |
|
cmpl %ecx,%edx # q still has something? |
|
jne 3f |
|
#ifndef DIAGNOSTIC |
|
shrl $2,%eax |
|
#endif |
|
btrl %eax,_C_LABEL(sched_whichqs) # no; clear bit |
|
|
|
3: |
|
/* Jump into the middle of cpu_switch */ |
/* Jump into the middle of cpu_switch */ |
xorl %eax,%eax |
xorl %eax,%eax |
jmp switch_resume |
jmp switch_resume |
|
|
/* |
/* |
* void switch_exit(struct lwp *l); |
* void switch_exit(struct lwp *l, void (*exit)(struct lwp *)); |
* Switch to the appropriate idle context (lwp0's if uniprocessor; the cpu's |
* Switch to the appropriate idle context (lwp0's if uniprocessor; the cpu's |
* if multiprocessor) and deallocate the address space and kernel stack for p. |
* if multiprocessor) and deallocate the address space and kernel stack for p. |
* Then jump into cpu_switch(), as if we were in the idle proc all along. |
* Then jump into cpu_switch(), as if we were in the idle proc all along. |
Line 2056 preempt_dequeue: |
|
Line 2028 preempt_dequeue: |
|
#endif |
#endif |
.globl _C_LABEL(uvmspace_free),_C_LABEL(kernel_map) |
.globl _C_LABEL(uvmspace_free),_C_LABEL(kernel_map) |
.globl _C_LABEL(uvm_km_free),_C_LABEL(tss_free) |
.globl _C_LABEL(uvm_km_free),_C_LABEL(tss_free) |
/* LINTSTUB: Func: void switch_exit(struct lwp *l) */ |
/* LINTSTUB: Func: void switch_exit(struct lwp *l, void (*exit)(struct lwp *)) */ |
ENTRY(switch_exit) |
ENTRY(switch_exit) |
movl 4(%esp),%edi # old process |
movl 4(%esp),%edi # old process |
|
movl 8(%esp),%eax # exit func |
#ifndef MULTIPROCESSOR |
#ifndef MULTIPROCESSOR |
movl $_C_LABEL(lwp0),%ebx |
movl $_C_LABEL(lwp0),%ebx |
movl L_ADDR(%ebx),%esi |
movl L_ADDR(%ebx),%esi |
Line 2077 ENTRY(switch_exit) |
|
Line 2050 ENTRY(switch_exit) |
|
movl PCB_ESP(%esi),%esp |
movl PCB_ESP(%esi),%esp |
movl PCB_EBP(%esi),%ebp |
movl PCB_EBP(%esi),%ebp |
|
|
/* Load TSS info. */ |
/* Save exit func. */ |
#ifdef MULTIPROCESSOR |
pushl %eax |
movl CPUVAR(GDT),%eax |
|
#else |
|
/* Load TSS info. */ |
|
movl _C_LABEL(gdt),%eax |
|
#endif |
|
|
|
/* Switch address space. */ |
|
movl PCB_CR3(%esi),%ecx |
|
movl %ecx,%cr3 |
|
|
|
/* Switch TSS. */ |
|
andl $~0x0200,4-SEL_KPL(%eax,%edx,1) |
|
ltr %dx |
|
|
|
/* We're always in the kernel, so we don't need the LDT. */ |
|
|
|
/* Restore cr0 (including FPU state). */ |
|
movl PCB_CR0(%esi),%ecx |
|
movl %ecx,%cr0 |
|
|
|
/* Record new pcb. */ |
|
SET_CURPCB(%esi) |
|
|
|
/* Interrupts are okay again. */ |
|
sti |
|
|
|
/* |
|
* Schedule the dead process's vmspace and stack to be freed. |
|
*/ |
|
pushl %edi /* exit2(l) */ |
|
call _C_LABEL(exit2) |
|
addl $4,%esp |
|
|
|
/* Jump into cpu_switch() with the right state. */ |
|
xorl %esi,%esi |
|
movl %esi,CPUVAR(CURLWP) |
|
jmp idle_start |
|
|
|
/* |
|
* switch_lwp_exit(struct lwp *l); |
|
* Switch to the appropriate idle context (lwp0's if uniprocessor; the cpu's |
|
* if multiprocessor) and deallocate the kernel stack for l. |
|
* Then jump into cpu_switch(), as if we were in the idle proc all along. |
|
*/ |
|
#ifndef MULTIPROCESSOR |
|
.globl _C_LABEL(lwp0) |
|
#endif |
|
.globl _C_LABEL(uvmspace_free),_C_LABEL(kernel_map) |
|
.globl _C_LABEL(uvm_km_free),_C_LABEL(tss_free) |
|
/* LINTSTUB: Func: void switch_lwp_exit(struct lwp *l) */ |
|
ENTRY(switch_lwp_exit) |
|
movl 4(%esp),%edi # old process |
|
#ifndef MULTIPROCESSOR |
|
movl $_C_LABEL(lwp0),%ebx |
|
movl L_ADDR(%ebx),%esi |
|
movl L_MD_TSS_SEL(%ebx),%edx |
|
#else |
|
movl CPUVAR(IDLE_PCB),%esi |
|
movl CPUVAR(IDLE_TSS_SEL),%edx |
|
#endif |
|
/* In case we fault... */ |
|
movl $0,CPUVAR(CURLWP) |
|
|
|
/* Restore the idle context. */ |
|
cli |
|
|
|
/* Restore stack pointers. */ |
|
movl PCB_ESP(%esi),%esp |
|
movl PCB_EBP(%esi),%ebp |
|
|
|
/* Load TSS info. */ |
/* Load TSS info. */ |
#ifdef MULTIPROCESSOR |
#ifdef MULTIPROCESSOR |
Line 2180 ENTRY(switch_lwp_exit) |
|
Line 2084 ENTRY(switch_lwp_exit) |
|
/* |
/* |
* Schedule the dead process's vmspace and stack to be freed. |
* Schedule the dead process's vmspace and stack to be freed. |
*/ |
*/ |
pushl %edi /* lwp_exit2(l) */ |
movl 0(%esp),%eax /* %eax = exit func */ |
call _C_LABEL(lwp_exit2) |
movl %edi,0(%esp) /* {lwp_}exit2(l) */ |
|
call *%eax |
addl $4,%esp |
addl $4,%esp |
|
|
/* Jump into cpu_switch() with the right state. */ |
/* Jump into cpu_switch() with the right state. */ |
|
|
movl CPUVAR(CURLWP),%edx |
movl CPUVAR(CURLWP),%edx |
movl %esp,L_MD_REGS(%edx) # save pointer to frame |
movl %esp,L_MD_REGS(%edx) # save pointer to frame |
movl L_PROC(%edx),%edx |
movl L_PROC(%edx),%edx |
|
pushl %esp |
call *P_MD_SYSCALL(%edx) # get pointer to syscall() function |
call *P_MD_SYSCALL(%edx) # get pointer to syscall() function |
|
addl $4,%esp |
2: /* Check for ASTs on exit to user mode. */ |
2: /* Check for ASTs on exit to user mode. */ |
cli |
cli |
CHECK_ASTPENDING(%eax) |
CHECK_ASTPENDING(%eax) |
|
|
CLEAR_ASTPENDING(%eax) |
CLEAR_ASTPENDING(%eax) |
sti |
sti |
/* Pushed T_ASTFLT into tf_trapno on entry. */ |
/* Pushed T_ASTFLT into tf_trapno on entry. */ |
|
pushl %esp |
call _C_LABEL(trap) |
call _C_LABEL(trap) |
|
addl $4,%esp |
jmp 2b |
jmp 2b |
#ifndef DIAGNOSTIC |
#ifndef DIAGNOSTIC |
1: INTRFASTEXIT |
1: INTRFASTEXIT |