version 1.34.2.5, 2007/10/27 11:26:37 |
version 1.72, 2008/05/03 05:56:44 |
|
|
*/ |
*/ |
|
|
/* |
/* |
|
* Copyright (c) 2006 Manuel Bouyer. |
|
* |
|
* Redistribution and use in source and binary forms, with or without |
|
* modification, are permitted provided that the following conditions |
|
* are met: |
|
* 1. Redistributions of source code must retain the above copyright |
|
* notice, this list of conditions and the following disclaimer. |
|
* 2. Redistributions in binary form must reproduce the above copyright |
|
* notice, this list of conditions and the following disclaimer in the |
|
* documentation and/or other materials provided with the distribution. |
|
* 3. All advertising materials mentioning features or use of this software |
|
* must display the following acknowledgement: |
|
* This product includes software developed by Manuel Bouyer. |
|
* 4. The name of the author may not be used to endorse or promote products |
|
* derived from this software without specific prior written permission. |
|
* |
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
|
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
|
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. |
|
* IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, |
|
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT |
|
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
|
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
|
* |
|
*/ |
|
|
|
/* |
* Copyright (c) 2001 Wasabi Systems, Inc. |
* Copyright (c) 2001 Wasabi Systems, Inc. |
* All rights reserved. |
* All rights reserved. |
* |
* |
|
|
* 2. Redistributions in binary form must reproduce the above copyright |
* 2. Redistributions in binary form must reproduce the above copyright |
* notice, this list of conditions and the following disclaimer in the |
* notice, this list of conditions and the following disclaimer in the |
* documentation and/or other materials provided with the distribution. |
* documentation and/or other materials provided with the distribution. |
* 3. All advertising materials mentioning features or use of this software |
|
* must display the following acknowledgement: |
|
* This product includes software developed by the NetBSD |
|
* Foundation, Inc. and its contributors. |
|
* 4. Neither the name of The NetBSD Foundation nor the names of its |
|
* contributors may be used to endorse or promote products derived |
|
* from this software without specific prior written permission. |
|
* |
* |
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS |
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS |
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED |
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED |
|
|
* @(#)locore.s 7.3 (Berkeley) 5/13/91 |
* @(#)locore.s 7.3 (Berkeley) 5/13/91 |
*/ |
*/ |
|
|
|
#include <machine/asm.h> |
|
__KERNEL_RCSID(0, "$NetBSD$"); |
|
|
#include "opt_compat_oldboot.h" |
#include "opt_compat_oldboot.h" |
#include "opt_cputype.h" |
|
#include "opt_ddb.h" |
#include "opt_ddb.h" |
#include "opt_realmem.h" |
#include "opt_realmem.h" |
#include "opt_vm86.h" |
#include "opt_vm86.h" |
|
#include "opt_xen.h" |
|
|
#include "npx.h" |
#include "npx.h" |
#include "assym.h" |
#include "assym.h" |
|
|
#include <machine/specialreg.h> |
#include <machine/specialreg.h> |
#include <machine/trap.h> |
#include <machine/trap.h> |
#include <machine/i82489reg.h> |
#include <machine/i82489reg.h> |
#include <machine/multiboot.h> |
|
#include <machine/asm.h> |
|
#include <machine/frameasm.h> |
#include <machine/frameasm.h> |
#include <machine/i82489reg.h> |
#include <machine/i82489reg.h> |
|
#ifndef XEN |
|
#include <machine/multiboot.h> |
|
#endif |
|
|
/* XXX temporary kluge; these should not be here */ |
/* XXX temporary kluge; these should not be here */ |
/* Get definitions for IOM_BEGIN, IOM_END, and IOM_SIZE */ |
/* Get definitions for IOM_BEGIN, IOM_END, and IOM_SIZE */ |
#include <dev/isa/isareg.h> |
#include <dev/isa/isareg.h> |
|
|
|
#ifdef XEN |
|
/* |
|
* Xen guest identifier and loader selection |
|
*/ |
|
.section __xen_guest |
|
#ifdef XEN3 |
|
.ascii "GUEST_OS=netbsd,GUEST_VER=3.0,XEN_VER=xen-3.0" |
|
#if defined(DOM0OPS) || !defined(XEN_COMPAT_030001) |
|
.ascii ",VIRT_BASE=0xc0000000" /* KERNBASE */ |
|
.ascii ",ELF_PADDR_OFFSET=0xc0000000" /* KERNBASE */ |
|
#else |
|
.ascii ",VIRT_BASE=0xc0100000" /* KERNTEXTOFF */ |
|
.ascii ",ELF_PADDR_OFFSET=0xc0100000" /* KERNTEXTOFF */ |
|
#endif |
|
.ascii ",VIRT_ENTRY=0xc0100000" /* KERNTEXTOFF */ |
|
#if !defined(XEN_COMPAT_030001) |
|
.ascii ",HYPERCALL_PAGE=0x00000101" |
|
/* (???+HYPERCALL_PAGE_OFFSET)/PAGE_SIZE) */ |
|
#endif |
|
#ifdef PAE |
|
.ascii ",PAE=yes[extended-cr3]" |
|
#endif |
|
#else /* !XEN3 */ |
|
.ascii "GUEST_OS=netbsd,GUEST_VER=2.0,XEN_VER=2.0" |
|
#endif /* XEN3 */ |
|
.ascii ",LOADER=generic" |
|
#if (NKSYMS || defined(DDB) || defined(LKM)) && !defined(SYMTAB_SPACE) |
|
.ascii ",BSD_SYMTAB=yes" |
|
#endif |
|
.byte 0 |
|
#endif |
|
|
/* |
/* |
* Initialization |
* Initialization |
*/ |
*/ |
|
|
|
|
.globl _C_LABEL(cpu) |
.globl _C_LABEL(cpu) |
.globl _C_LABEL(esym) |
.globl _C_LABEL(esym) |
|
.globl _C_LABEL(eblob) |
.globl _C_LABEL(atdevbase) |
.globl _C_LABEL(atdevbase) |
.globl _C_LABEL(proc0uarea),_C_LABEL(PDPpaddr) |
.globl _C_LABEL(proc0uarea),_C_LABEL(PDPpaddr) |
.globl _C_LABEL(gdt) |
.globl _C_LABEL(gdt) |
Line 184 _C_LABEL(tablesize): .long 0 |
|
Line 244 _C_LABEL(tablesize): .long 0 |
|
.space 512 |
.space 512 |
tmpstk: |
tmpstk: |
|
|
|
#ifndef XEN |
#define _RELOC(x) ((x) - KERNBASE) |
#define _RELOC(x) ((x) - KERNBASE) |
|
#else |
|
#define _RELOC(x) ((x)) |
|
#endif /* XEN */ |
#define RELOC(x) _RELOC(_C_LABEL(x)) |
#define RELOC(x) _RELOC(_C_LABEL(x)) |
|
|
.text |
.text |
|
|
.set _C_LABEL(kernel_text),KERNTEXTOFF |
.set _C_LABEL(kernel_text),KERNTEXTOFF |
|
|
.globl start |
.globl start |
|
#ifndef XEN |
start: movw $0x1234,0x472 # warm boot |
start: movw $0x1234,0x472 # warm boot |
|
|
#if defined(MULTIBOOT) |
#if defined(MULTIBOOT) |
Line 425 try586: /* Use the `cpuid' instruction. |
|
Line 489 try586: /* Use the `cpuid' instruction. |
|
/* |
/* |
* Virtual address space of kernel: |
* Virtual address space of kernel: |
* |
* |
* text | data | bss | [syms] | page dir | proc0 kstack | L1 ptp |
* text | data | bss | [syms] | [blobs] | page dir | proc0 kstack | L1 ptp |
* 0 1 2 3 |
* 0 1 2 3 |
*/ |
*/ |
|
|
#define PROC0_PDIR_OFF 0 |
#define PROC0_PDIR_OFF 0 |
Line 449 try586: /* Use the `cpuid' instruction. |
|
Line 513 try586: /* Use the `cpuid' instruction. |
|
|
|
/* Find end of kernel image. */ |
/* Find end of kernel image. */ |
movl $RELOC(end),%edi |
movl $RELOC(end),%edi |
|
|
#if (NKSYMS || defined(DDB) || defined(LKM)) && !defined(SYMTAB_SPACE) |
#if (NKSYMS || defined(DDB) || defined(LKM)) && !defined(SYMTAB_SPACE) |
/* Save the symbols (if loaded). */ |
/* Save the symbols (if loaded). */ |
movl RELOC(esym),%eax |
movl RELOC(esym),%eax |
Line 459 try586: /* Use the `cpuid' instruction. |
|
Line 524 try586: /* Use the `cpuid' instruction. |
|
1: |
1: |
#endif |
#endif |
|
|
|
/* Skip over any modules/blobs. */ |
|
movl RELOC(eblob),%eax |
|
testl %eax,%eax |
|
jz 1f |
|
subl $KERNBASE,%eax |
|
movl %eax,%edi |
|
1: |
/* Compute sizes */ |
/* Compute sizes */ |
movl %edi,%esi # edi = esym ? esym : end |
movl %edi,%esi |
addl $PGOFSET,%esi # page align up |
addl $PGOFSET,%esi # page align up |
andl $~PGOFSET,%esi |
andl $~PGOFSET,%esi |
|
|
|
|
#endif /* SAFARI_FIFO_HACK */ |
#endif /* SAFARI_FIFO_HACK */ |
|
|
call _C_LABEL(main) |
call _C_LABEL(main) |
|
#else /* XEN */ |
|
start: |
|
/* First, reset the PSL. */ |
|
pushl $PSL_MBO |
|
popfl |
|
|
|
cld |
|
#ifdef XEN3 |
|
movl %esp, %ebx # save start of available space |
|
#else |
|
movl %esi,%ebx # save start_info pointer |
|
#endif |
|
movl $_RELOC(tmpstk),%esp # bootstrap stack end location |
|
|
|
/* Clear BSS first so that there are no surprises... */ |
|
xorl %eax,%eax |
|
movl $RELOC(__bss_start),%edi |
|
movl $RELOC(_end),%ecx |
|
subl %edi,%ecx |
|
rep stosb |
|
|
|
/* Copy the necessary stuff from start_info structure. */ |
|
/* We need to copy shared_info early, so that sti/cli work */ |
|
movl $RELOC(start_info_union),%edi |
|
movl $128,%ecx |
|
rep movsl |
|
|
|
/* Clear segment registers; always null in proc0. */ |
|
xorl %eax,%eax |
|
movw %ax,%fs |
|
movw %ax,%gs |
|
decl %eax |
|
movl %eax,RELOC(cpu_info_primary)+CPU_INFO_LEVEL |
|
|
|
xorl %eax,%eax |
|
cpuid |
|
movl %eax,RELOC(cpu_info_primary)+CPU_INFO_LEVEL |
|
|
|
call xen_pmap_bootstrap |
|
/* |
|
* First avail returned by xen_pmap_bootstrap in %eax |
|
*/ |
|
movl %eax, %esi; |
|
movl %esi, _C_LABEL(proc0uarea) |
|
|
|
#define PROC0PDIR ((0) * PAGE_SIZE) |
|
#define PROC0STACK ((1) * PAGE_SIZE) |
|
|
|
/* Set up bootstrap stack. */ |
|
leal (KSTACK_SIZE-FRAMESIZE)(%eax),%esp |
|
xorl %ebp,%ebp # mark end of frames |
|
|
|
addl $USPACE, %esi |
|
subl $KERNBASE, %esi #init386 want a physical address |
|
pushl %esi |
|
call _C_LABEL(init386) # wire 386 chip for unix operation |
|
addl $4,%esp |
|
call _C_LABEL(main) |
|
|
|
#if defined(XEN3) && !defined(XEN_COMPAT_030001) |
|
/* space for the hypercall call page */ |
|
#define HYPERCALL_PAGE_OFFSET 0x1000 |
|
.org HYPERCALL_PAGE_OFFSET |
|
ENTRY(hypercall_page) |
|
.skip 0x1000 |
|
#endif /* defined(XEN3) && !defined(XEN_COMPAT_030001) */ |
|
|
|
/* |
|
* void lgdt_finish(void); |
|
* Finish load a new GDT pointer (do any necessary cleanup). |
|
* XXX It's somewhat questionable whether reloading all the segment registers |
|
* is necessary, since the actual descriptor data is not changed except by |
|
* process creation and exit, both of which clean up via task switches. OTOH, |
|
* this only happens at run time when the GDT is resized. |
|
*/ |
|
/* LINTSTUB: Func: void lgdt_finish(void) */ |
|
NENTRY(lgdt_finish) |
|
movl $GSEL(GDATA_SEL, SEL_KPL),%eax |
|
movw %ax,%ds |
|
movw %ax,%es |
|
movw %ax,%gs |
|
movw %ax,%ss |
|
movl $GSEL(GCPU_SEL, SEL_KPL),%eax |
|
movw %ax,%fs |
|
/* Reload code selector by doing intersegment return. */ |
|
popl %eax |
|
pushl $GSEL(GCODE_SEL, SEL_KPL) |
|
pushl %eax |
|
lret |
|
END(lgdt_finish) |
|
|
|
#endif /* XEN */ |
|
|
/* |
/* |
* void lwp_trampoline(void); |
* void lwp_trampoline(void); |
|
|
* NOTE: This function does not have a normal calling sequence! |
* NOTE: This function does not have a normal calling sequence! |
*/ |
*/ |
NENTRY(lwp_trampoline) |
NENTRY(lwp_trampoline) |
pushl %ebp |
movl %ebp,%edi /* for .Lsyscall_checkast */ |
xorl %ebp,%ebp |
xorl %ebp,%ebp |
|
pushl %edi |
pushl %eax |
pushl %eax |
call _C_LABEL(lwp_startup) |
call _C_LABEL(lwp_startup) |
addl $8,%esp |
addl $8,%esp |
pushl %ebx |
pushl %ebx |
call *%esi |
call *%esi |
addl $4,%esp |
addl $4,%esp |
DO_DEFERRED_SWITCH |
jmp .Lsyscall_checkast |
INTRFASTEXIT |
|
/* NOTREACHED */ |
/* NOTREACHED */ |
|
END(lwp_trampoline) |
|
|
/* |
/* |
* sigcode() |
* sigcode() |
|
|
int $0x80 # exit if sigreturn fails |
int $0x80 # exit if sigreturn fails |
.globl _C_LABEL(esigcode) |
.globl _C_LABEL(esigcode) |
_C_LABEL(esigcode): |
_C_LABEL(esigcode): |
|
END(sigcode) |
|
|
/* |
/* |
* int setjmp(label_t *) |
* int setjmp(label_t *) |
|
|
movl %edx,20(%eax) # save eip |
movl %edx,20(%eax) # save eip |
xorl %eax,%eax # return 0 |
xorl %eax,%eax # return 0 |
ret |
ret |
|
END(setjmp) |
|
|
/* |
/* |
* int longjmp(label_t *) |
* int longjmp(label_t *) |
|
|
movl %edx,(%esp) # put in return frame |
movl %edx,(%esp) # put in return frame |
movl $1,%eax # return 1 |
movl $1,%eax # return 1 |
ret |
ret |
|
END(longjmp) |
|
|
|
/* |
|
* void dumpsys(void) |
|
* |
|
* Mimic cpu_switchto() for postmortem debugging. |
|
*/ |
|
ENTRY(dumpsys) |
|
pushl %ebx # set up fake switchframe |
|
pushl %esi # and save context |
|
pushl %edi |
|
movl %esp,_C_LABEL(dumppcb)+PCB_ESP |
|
movl %ebp,_C_LABEL(dumppcb)+PCB_EBP |
|
call _C_LABEL(dodumpsys) # dump! |
|
addl $(3*4), %esp # unwind switchframe |
|
ret |
|
END(dumpsys) |
|
|
/* |
/* |
* struct lwp *cpu_switchto(struct lwp *oldlwp, struct newlwp) |
* struct lwp *cpu_switchto(struct lwp *oldlwp, struct newlwp, |
|
* bool returning) |
* |
* |
* 1. if (oldlwp != NULL), save its context. |
* 1. if (oldlwp != NULL), save its context. |
* 2. then, restore context of newlwp. |
* 2. then, restore context of newlwp. |
Line 737 ENTRY(cpu_switchto) |
|
Line 922 ENTRY(cpu_switchto) |
|
pushl %esi |
pushl %esi |
pushl %edi |
pushl %edi |
|
|
|
#if defined(DIAGNOSTIC) && !defined(XEN) |
|
cmpl $IPL_SCHED,CPUVAR(ILEVEL) |
|
jbe 0f |
|
pushl $.Lstr |
|
call _C_LABEL(panic) |
|
addl $4,%esp |
|
.Lstr: .string "cpu_switcho: switching above IPL_SCHED\0" |
|
0: |
|
#endif |
|
|
movl 16(%esp),%esi # oldlwp |
movl 16(%esp),%esi # oldlwp |
movl 20(%esp),%edi # newlwp |
movl 20(%esp),%edi # newlwp |
|
movl 24(%esp),%edx # returning |
testl %esi,%esi |
testl %esi,%esi |
jz 1f |
jz 1f |
|
|
Line 752 ENTRY(cpu_switchto) |
|
Line 948 ENTRY(cpu_switchto) |
|
movl PCB_EBP(%ebx),%ebp |
movl PCB_EBP(%ebx),%ebp |
movl PCB_ESP(%ebx),%esp |
movl PCB_ESP(%ebx),%esp |
|
|
/* Switch TSS. Reset "task busy" flag before loading. */ |
/* |
movl %cr3,%eax |
* Set curlwp. This must be globally visible in order to permit |
movl %eax,PCB_CR3(%ebx) # for TSS gates |
* non-interlocked mutex release. |
movl CPUVAR(GDT),%eax |
*/ |
movl L_MD_TSS_SEL(%edi),%edx |
movl %edi,%ecx |
andl $~0x0200,4(%eax,%edx, 1) |
xchgl %ecx,CPUVAR(CURLWP) |
ltr %dx |
|
|
/* Skip the rest if returning to a pinned LWP. */ |
|
testl %edx,%edx |
|
jnz 4f |
|
|
/* Set curlwp. */ |
#ifdef XEN |
movl %edi,CPUVAR(CURLWP) |
pushl %edi |
|
call _C_LABEL(i386_switch_context) |
|
addl $4,%esp |
|
#else /* XEN */ |
|
/* Switch ring0 esp */ |
|
movl PCB_ESP0(%ebx),%eax |
|
movl %eax,CPUVAR(ESP0) |
|
|
/* Don't bother with the rest if switching to a system process. */ |
/* Don't bother with the rest if switching to a system process. */ |
testl $LW_SYSTEM,L_FLAG(%edi) |
testl $LW_SYSTEM,L_FLAG(%edi) |
jnz 4f |
jnz 4f |
|
|
|
/* Restore thread-private %fs/%gs descriptors. */ |
|
movl CPUVAR(GDT),%ecx |
|
movl PCB_FSD(%ebx), %eax |
|
movl PCB_FSD+4(%ebx), %edx |
|
movl %eax, (GUFS_SEL*8)(%ecx) |
|
movl %edx, (GUFS_SEL*8+4)(%ecx) |
|
movl PCB_GSD(%ebx), %eax |
|
movl PCB_GSD+4(%ebx), %edx |
|
movl %eax, (GUGS_SEL*8)(%ecx) |
|
movl %edx, (GUGS_SEL*8+4)(%ecx) |
|
#endif /* XEN */ |
|
|
|
/* Switch I/O bitmap */ |
|
movl PCB_IOMAP(%ebx),%eax |
|
orl %eax,%eax |
|
jnz,pn .Lcopy_iobitmap |
|
movl $(IOMAP_INVALOFF << 16),CPUVAR(IOBASE) |
|
.Liobitmap_done: |
|
|
/* Is this process using RAS (restartable atomic sequences)? */ |
/* Is this process using RAS (restartable atomic sequences)? */ |
movl L_PROC(%edi),%eax |
movl L_PROC(%edi),%eax |
cmpl $0,P_RASLIST(%eax) |
cmpl $0,P_RASLIST(%eax) |
Line 777 ENTRY(cpu_switchto) |
|
Line 1001 ENTRY(cpu_switchto) |
|
* FPU IPIs can alter the LWP's saved cr0. Dropping the priority |
* FPU IPIs can alter the LWP's saved cr0. Dropping the priority |
* is deferred until mi_switch(), when cpu_switchto() returns. |
* is deferred until mi_switch(), when cpu_switchto() returns. |
*/ |
*/ |
2: movl $IPL_IPI,CPUVAR(ILEVEL) |
2: |
|
#ifndef XEN |
|
movl $IPL_IPI,CPUVAR(ILEVEL) |
movl PCB_CR0(%ebx),%ecx |
movl PCB_CR0(%ebx),%ecx |
movl %cr0,%edx |
movl %cr0,%edx |
|
|
Line 794 ENTRY(cpu_switchto) |
|
Line 1020 ENTRY(cpu_switchto) |
|
3: cmpl %edx,%ecx |
3: cmpl %edx,%ecx |
je 4f |
je 4f |
movl %ecx,%cr0 |
movl %ecx,%cr0 |
|
#endif /* XEN */ |
|
|
/* Return to the new LWP, returning 'oldlwp' in %eax. */ |
/* Return to the new LWP, returning 'oldlwp' in %eax. */ |
4: movl %esi,%eax |
4: movl %esi,%eax |
Line 814 ENTRY(cpu_switchto) |
|
Line 1041 ENTRY(cpu_switchto) |
|
movl %eax,TF_EIP(%ecx) |
movl %eax,TF_EIP(%ecx) |
jmp 2b |
jmp 2b |
|
|
|
.Lcopy_iobitmap: |
|
/* Copy I/O bitmap. */ |
|
movl $(IOMAPSIZE/4),%ecx |
|
pushl %esi |
|
pushl %edi |
|
movl %eax,%esi /* pcb_iomap */ |
|
movl CPUVAR(SELF),%edi |
|
leal CPU_INFO_IOMAP(%edi),%edi |
|
rep |
|
movsl |
|
popl %edi |
|
popl %esi |
|
movl $((CPU_INFO_IOMAP - CPU_INFO_TSS) << 16),CPUVAR(IOBASE) |
|
jmp .Liobitmap_done |
|
END(cpu_switchto) |
|
|
/* |
/* |
* void savectx(struct pcb *pcb); |
* void savectx(struct pcb *pcb); |
* |
* |
|
|
movl %esp,PCB_ESP(%edx) |
movl %esp,PCB_ESP(%edx) |
movl %ebp,PCB_EBP(%edx) |
movl %ebp,PCB_EBP(%edx) |
ret |
ret |
|
END(savectx) |
|
|
/* |
/* |
* osyscall() |
* osyscall() |
Line 835 IDTVEC(osyscall) |
|
Line 1079 IDTVEC(osyscall) |
|
popl 8(%esp) |
popl 8(%esp) |
pushl $7 # size of instruction for restart |
pushl $7 # size of instruction for restart |
jmp syscall1 |
jmp syscall1 |
|
IDTVEC_END(osyscall) |
|
|
/* |
/* |
* syscall() |
* syscall() |
|
Line 1091 IDTVEC(syscall) |
|
syscall1: |
syscall1: |
pushl $T_ASTFLT # trap # for doing ASTs |
pushl $T_ASTFLT # trap # for doing ASTs |
INTRENTRY |
INTRENTRY |
|
|
#ifdef DIAGNOSTIC |
#ifdef DIAGNOSTIC |
cmpl $0, CPUVAR(WANT_PMAPLOAD) |
|
jz 1f |
|
pushl $6f |
|
call _C_LABEL(printf) |
|
addl $4, %esp |
|
1: |
|
movl CPUVAR(ILEVEL),%ebx |
movl CPUVAR(ILEVEL),%ebx |
testl %ebx,%ebx |
testl %ebx,%ebx |
jz 1f |
jz 1f |
|
|
#endif |
#endif |
1: |
1: |
#endif /* DIAGNOSTIC */ |
#endif /* DIAGNOSTIC */ |
movl CPUVAR(CURLWP),%edx |
incl CPUVAR(NSYSCALL) # count it atomically |
movl %esp,L_MD_REGS(%edx) # save pointer to frame |
movl CPUVAR(CURLWP),%edi |
movl L_PROC(%edx),%edx |
movl L_PROC(%edi),%edx |
|
movl %esp,L_MD_REGS(%edi) # save pointer to frame |
pushl %esp |
pushl %esp |
call *P_MD_SYSCALL(%edx) # get pointer to syscall() function |
call *P_MD_SYSCALL(%edx) # get pointer to syscall() function |
addl $4,%esp |
addl $4,%esp |
.Lsyscall_checkast: |
.Lsyscall_checkast: |
/* Check for ASTs on exit to user mode. */ |
/* Check for ASTs on exit to user mode. */ |
cli |
CLI(%eax) |
CHECK_ASTPENDING(%eax) |
movl L_MD_ASTPENDING(%edi), %eax |
je 1f |
orl CPUVAR(WANT_PMAPLOAD), %eax |
/* Always returning to user mode here. */ |
|
CLEAR_ASTPENDING(%eax) |
|
sti |
|
/* Pushed T_ASTFLT into tf_trapno on entry. */ |
|
pushl %esp |
|
call _C_LABEL(trap) |
|
addl $4,%esp |
|
jmp .Lsyscall_checkast /* re-check ASTs */ |
|
1: CHECK_DEFERRED_SWITCH |
|
jnz 9f |
jnz 9f |
|
#ifdef XEN |
|
STIC(%eax) |
|
jz 14f |
|
call _C_LABEL(stipending) |
|
testl %eax,%eax |
|
jz 14f |
|
/* process pending interrupts */ |
|
CLI(%eax) |
|
movl CPUVAR(ILEVEL), %ebx |
|
movl $.Lsyscall_resume, %esi # address to resume loop at |
|
.Lsyscall_resume: |
|
movl %ebx,%eax # get cpl |
|
movl CPUVAR(IUNMASK)(,%eax,4),%eax |
|
andl CPUVAR(IPENDING),%eax # any non-masked bits left? |
|
jz 17f |
|
bsrl %eax,%eax |
|
btrl %eax,CPUVAR(IPENDING) |
|
movl CPUVAR(ISOURCES)(,%eax,4),%eax |
|
jmp *IS_RESUME(%eax) |
|
17: movl %ebx, CPUVAR(ILEVEL) #restore cpl |
|
jmp .Lsyscall_checkast |
|
14: |
|
#endif /* XEN */ |
#ifndef DIAGNOSTIC |
#ifndef DIAGNOSTIC |
INTRFASTEXIT |
INTRFASTEXIT |
#else /* DIAGNOSTIC */ |
#else /* DIAGNOSTIC */ |
cmpl $IPL_NONE,CPUVAR(ILEVEL) |
cmpl $IPL_NONE,CPUVAR(ILEVEL) |
jne 3f |
jne 3f |
INTRFASTEXIT |
INTRFASTEXIT |
3: sti |
3: STI(%eax) |
pushl $4f |
pushl $4f |
call _C_LABEL(printf) |
call _C_LABEL(printf) |
addl $4,%esp |
addl $4,%esp |
#ifdef DDB |
pushl $IPL_NONE |
int $3 |
call _C_LABEL(spllower) |
#endif /* DDB */ |
addl $4,%esp |
movl $IPL_NONE,CPUVAR(ILEVEL) |
|
jmp .Lsyscall_checkast |
jmp .Lsyscall_checkast |
4: .asciz "WARNING: SPL NOT LOWERED ON SYSCALL EXIT\n" |
4: .asciz "WARNING: SPL NOT LOWERED ON SYSCALL EXIT\n" |
5: .asciz "WARNING: SPL NOT ZERO ON SYSCALL ENTRY\n" |
5: .asciz "WARNING: SPL NOT ZERO ON SYSCALL ENTRY\n" |
6: .asciz "WARNING: WANT PMAPLOAD ON SYSCALL ENTRY\n" |
|
#endif /* DIAGNOSTIC */ |
#endif /* DIAGNOSTIC */ |
9: sti |
9: |
|
cmpl $0, CPUVAR(WANT_PMAPLOAD) |
|
jz 10f |
|
STI(%eax) |
call _C_LABEL(pmap_load) |
call _C_LABEL(pmap_load) |
jmp .Lsyscall_checkast /* re-check ASTs */ |
jmp .Lsyscall_checkast /* re-check ASTs */ |
|
10: |
|
/* Always returning to user mode here. */ |
|
movl $0, L_MD_ASTPENDING(%edi) |
|
STI(%eax) |
|
/* Pushed T_ASTFLT into tf_trapno on entry. */ |
|
pushl %esp |
|
call _C_LABEL(trap) |
|
addl $4,%esp |
|
jmp .Lsyscall_checkast /* re-check ASTs */ |
|
IDTVEC_END(syscall) |
|
|
#if NNPX > 0 |
#if NNPX > 0 |
/* |
/* |
Line 930 NENTRY(probeintr) |
|
Line 1194 NENTRY(probeintr) |
|
outb %al,$0xf0 # clear BUSY# latch |
outb %al,$0xf0 # clear BUSY# latch |
popl %eax |
popl %eax |
iret |
iret |
|
END(probeintr) |
|
|
/* |
/* |
* void probetrap(void) |
* void probetrap(void) |
Line 939 NENTRY(probetrap) |
|
Line 1204 NENTRY(probetrap) |
|
incl _C_LABEL(npx_traps_while_probing) |
incl _C_LABEL(npx_traps_while_probing) |
fnclex |
fnclex |
iret |
iret |
|
END(probetrap) |
|
|
/* |
/* |
* int npx586bug1(int a, int b) |
* int npx586bug1(int a, int b) |
Line 954 NENTRY(npx586bug1) |
|
Line 1220 NENTRY(npx586bug1) |
|
fistpl (%esp) |
fistpl (%esp) |
popl %eax |
popl %eax |
ret |
ret |
|
END(npx586bug1) |
#endif /* NNPX > 0 */ |
#endif /* NNPX > 0 */ |
|
|
/* |
/* |
Line 983 ENTRY(sse2_zero_page) |
|
Line 1250 ENTRY(sse2_zero_page) |
|
sfence |
sfence |
pop %ebp |
pop %ebp |
ret |
ret |
|
END(sse2_zero_page) |
|
|
/* |
/* |
* void sse2_copy_page(void *src, void *dst) |
* void sse2_copy_page(void *src, void *dst) |
Line 1017 ENTRY(sse2_copy_page) |
|
Line 1285 ENTRY(sse2_copy_page) |
|
popl %ebx |
popl %ebx |
popl %ebp |
popl %ebp |
ret |
ret |
|
END(sse2_copy_page) |