version 1.76, 2008/06/02 14:41:41 |
version 1.100, 2012/06/16 17:30:19 |
|
|
* 2. Redistributions in binary form must reproduce the above copyright |
* 2. Redistributions in binary form must reproduce the above copyright |
* notice, this list of conditions and the following disclaimer in the |
* notice, this list of conditions and the following disclaimer in the |
* documentation and/or other materials provided with the distribution. |
* documentation and/or other materials provided with the distribution. |
* 3. All advertising materials mentioning features or use of this software |
|
* must display the following acknowledgement: |
|
* This product includes software developed by Manuel Bouyer. |
|
* 4. The name of the author may not be used to endorse or promote products |
|
* derived from this software without specific prior written permission. |
|
* |
* |
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
|
|
|
|
|
|
/*- |
/*- |
* Copyright (c) 1998, 2000, 2004, 2006, 2007 The NetBSD Foundation, Inc. |
* Copyright (c) 1998, 2000, 2004, 2006, 2007, 2009 The NetBSD Foundation, Inc. |
* All rights reserved. |
* All rights reserved. |
* |
* |
* This code is derived from software contributed to The NetBSD Foundation |
* This code is derived from software contributed to The NetBSD Foundation |
* by Charles M. Hannum. |
* by Charles M. Hannum, and by Andrew Doran. |
* |
* |
* Redistribution and use in source and binary forms, with or without |
* Redistribution and use in source and binary forms, with or without |
* modification, are permitted provided that the following conditions |
* modification, are permitted provided that the following conditions |
Line 138 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 133 __KERNEL_RCSID(0, "$NetBSD$"); |
|
|
|
#include "opt_compat_oldboot.h" |
#include "opt_compat_oldboot.h" |
#include "opt_ddb.h" |
#include "opt_ddb.h" |
|
#include "opt_modular.h" |
|
#include "opt_multiboot.h" |
#include "opt_realmem.h" |
#include "opt_realmem.h" |
#include "opt_vm86.h" |
#include "opt_vm86.h" |
#include "opt_xen.h" |
#include "opt_xen.h" |
Line 171 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 168 __KERNEL_RCSID(0, "$NetBSD$"); |
|
* Xen guest identifier and loader selection |
* Xen guest identifier and loader selection |
*/ |
*/ |
.section __xen_guest |
.section __xen_guest |
#ifdef XEN3 |
|
.ascii "GUEST_OS=netbsd,GUEST_VER=3.0,XEN_VER=xen-3.0" |
.ascii "GUEST_OS=netbsd,GUEST_VER=3.0,XEN_VER=xen-3.0" |
#if defined(DOM0OPS) || !defined(XEN_COMPAT_030001) |
#if defined(DOM0OPS) || !defined(XEN_COMPAT_030001) |
.ascii ",VIRT_BASE=0xc0000000" /* KERNBASE */ |
.ascii ",VIRT_BASE=0xc0000000" /* KERNBASE */ |
Line 188 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 184 __KERNEL_RCSID(0, "$NetBSD$"); |
|
#ifdef PAE |
#ifdef PAE |
.ascii ",PAE=yes[extended-cr3]" |
.ascii ",PAE=yes[extended-cr3]" |
#endif |
#endif |
#else /* !XEN3 */ |
|
.ascii "GUEST_OS=netbsd,GUEST_VER=2.0,XEN_VER=2.0" |
|
#endif /* XEN3 */ |
|
.ascii ",LOADER=generic" |
.ascii ",LOADER=generic" |
#if (NKSYMS || defined(DDB) || defined(LKM)) && !defined(SYMTAB_SPACE) |
#if (NKSYMS || defined(DDB) || defined(MODULAR)) && !defined(SYMTAB_SPACE) |
.ascii ",BSD_SYMTAB=yes" |
.ascii ",BSD_SYMTAB=yes" |
#endif |
#endif |
.byte 0 |
.byte 0 |
Line 203 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 196 __KERNEL_RCSID(0, "$NetBSD$"); |
|
*/ |
*/ |
.data |
.data |
|
|
.globl _C_LABEL(cpu) |
.globl _C_LABEL(cputype) |
.globl _C_LABEL(cpuid_level) |
.globl _C_LABEL(cpuid_level) |
.globl _C_LABEL(esym) |
.globl _C_LABEL(esym) |
.globl _C_LABEL(eblob) |
.globl _C_LABEL(eblob) |
.globl _C_LABEL(atdevbase) |
.globl _C_LABEL(atdevbase) |
.globl _C_LABEL(proc0uarea),_C_LABEL(PDPpaddr) |
.globl _C_LABEL(lwp0uarea),_C_LABEL(PDPpaddr) |
.globl _C_LABEL(gdt) |
.globl _C_LABEL(gdt) |
.globl _C_LABEL(idt) |
.globl _C_LABEL(idt) |
.globl _C_LABEL(lapic_tpr) |
.globl _C_LABEL(lapic_tpr) |
Line 236 _C_LABEL(lapic_tpr): |
|
Line 229 _C_LABEL(lapic_tpr): |
|
.long 0 |
.long 0 |
#endif |
#endif |
|
|
_C_LABEL(cpu): .long 0 # are we 80486, Pentium, or.. |
_C_LABEL(cputype): .long 0 # are we 80486, Pentium, or.. |
_C_LABEL(cpuid_level): .long 0 |
_C_LABEL(cpuid_level): .long 0 |
_C_LABEL(atdevbase): .long 0 # location of start of iomem in virtual |
_C_LABEL(atdevbase): .long 0 # location of start of iomem in virtual |
_C_LABEL(proc0uarea): .long 0 |
_C_LABEL(lwp0uarea): .long 0 |
_C_LABEL(PDPpaddr): .long 0 # paddr of PDP, for libkvm |
_C_LABEL(PDPpaddr): .long 0 # paddr of PDP, for libkvm |
_C_LABEL(tablesize): .long 0 |
_C_LABEL(tablesize): .long 0 |
|
|
.space 512 |
.space 512 |
tmpstk: |
tmpstk: |
|
#ifdef XEN |
|
.align PAGE_SIZE, 0x0 # Align on page boundary |
|
_C_LABEL(tmpgdt): |
|
.space PAGE_SIZE # Xen expects a page |
|
#endif /* XEN */ |
|
.globl tmpgdt |
#ifndef XEN |
#ifndef XEN |
#define _RELOC(x) ((x) - KERNBASE) |
#define _RELOC(x) ((x) - KERNBASE) |
#else |
#else |
|
|
* Don't try cpuid, as Nx586s reportedly don't support the |
* Don't try cpuid, as Nx586s reportedly don't support the |
* PSL_ID bit. |
* PSL_ID bit. |
*/ |
*/ |
movl $CPU_NX586,RELOC(cpu) |
movl $CPU_NX586,RELOC(cputype) |
jmp 2f |
jmp 2f |
|
|
is386: |
is386: |
movl $CPU_386,RELOC(cpu) |
movl $CPU_386,RELOC(cputype) |
jmp 2f |
jmp 2f |
|
|
try486: /* Try to toggle identification flag; does not exist on early 486s. */ |
try486: /* Try to toggle identification flag; does not exist on early 486s. */ |
Line 371 try486: /* Try to toggle identification |
|
Line 369 try486: /* Try to toggle identification |
|
|
|
testl %eax,%eax |
testl %eax,%eax |
jnz try586 |
jnz try586 |
is486: movl $CPU_486,RELOC(cpu) |
is486: movl $CPU_486,RELOC(cputype) |
/* |
/* |
* Check Cyrix CPU |
* Check Cyrix CPU |
* Cyrix CPUs do not change the undefined flags following |
* Cyrix CPUs do not change the undefined flags following |
Line 389 is486: movl $CPU_486,RELOC(cpu) |
|
Line 387 is486: movl $CPU_486,RELOC(cpu) |
|
popfl |
popfl |
jmp 2f |
jmp 2f |
trycyrix486: |
trycyrix486: |
movl $CPU_6x86,RELOC(cpu) # set CPU type |
movl $CPU_6x86,RELOC(cputype) # set CPU type |
/* |
/* |
* Check for Cyrix 486 CPU by seeing if the flags change during a |
* Check for Cyrix 486 CPU by seeing if the flags change during a |
* divide. This is documented in the Cx486SLC/e SMM Programmer's |
* divide. This is documented in the Cx486SLC/e SMM Programmer's |
|
|
xorl %ecx,%eax # are the flags different? |
xorl %ecx,%eax # are the flags different? |
testl $0x8d5,%eax # only check C|PF|AF|Z|N|V |
testl $0x8d5,%eax # only check C|PF|AF|Z|N|V |
jne 2f # yes; must be Cyrix 6x86 CPU |
jne 2f # yes; must be Cyrix 6x86 CPU |
movl $CPU_486DLC,RELOC(cpu) # set CPU type |
movl $CPU_486DLC,RELOC(cputype) # set CPU type |
|
|
#ifndef CYRIX_CACHE_WORKS |
#ifndef CYRIX_CACHE_WORKS |
/* Disable caching of the ISA hole only. */ |
/* Disable caching of the ISA hole only. */ |
Line 489 try586: /* Use the `cpuid' instruction. |
|
Line 487 try586: /* Use the `cpuid' instruction. |
|
movl $_RELOC(tmpstk),%esp # bootstrap stack end location |
movl $_RELOC(tmpstk),%esp # bootstrap stack end location |
|
|
/* |
/* |
* Virtual address space of kernel: |
* Virtual address space of kernel, without PAE. The page dir is 1 page long. |
* |
* |
* text | data | bss | [syms] | [blobs] | page dir | proc0 kstack | L1 ptp |
* text | data | bss | [syms] | [blobs] | page dir | proc0 kstack | L1 ptp |
* 0 1 2 3 |
* 0 1 2 3 |
|
* |
|
* Virtual address space of kernel, with PAE. We need 4 pages for the page dir |
|
* and 1 page for the L3. |
|
* text | data | bss | [syms] | [blobs] | L3 | page dir | proc0 kstack | L1 ptp |
|
* 0 1 5 6 7 |
*/ |
*/ |
|
#ifndef PAE |
#define PROC0_PDIR_OFF 0 |
#define PROC0_PDIR_OFF 0 |
#define PROC0_STK_OFF (PROC0_PDIR_OFF + PAGE_SIZE) |
#else |
|
#define PROC0_L3_OFF 0 |
|
#define PROC0_PDIR_OFF 1 * PAGE_SIZE |
|
#endif |
|
|
|
#define PROC0_STK_OFF (PROC0_PDIR_OFF + PDP_SIZE * PAGE_SIZE) |
#define PROC0_PTP1_OFF (PROC0_STK_OFF + UPAGES * PAGE_SIZE) |
#define PROC0_PTP1_OFF (PROC0_STK_OFF + UPAGES * PAGE_SIZE) |
|
|
/* |
/* |
* fillkpt |
* fillkpt - Fill in a kernel page table |
* eax = pte (page frame | control | status) |
* eax = pte (page frame | control | status) |
* ebx = page table address |
* ebx = page table address |
* ecx = number of pages to map |
* ecx = number of pages to map |
|
* |
|
* For PAE, each entry is 8 bytes long: we must set the 4 upper bytes to 0. |
|
* This is done by the first instruction of fillkpt. In the non-PAE case, this |
|
* instruction just clears the page table entry. |
*/ |
*/ |
|
|
#define fillkpt \ |
#define fillkpt \ |
1: movl %eax,(%ebx) ; /* store phys addr */ \ |
1: movl $0,(PDE_SIZE-4)(%ebx) ; /* clear bits */ \ |
addl $4,%ebx ; /* next pte/pde */ \ |
movl %eax,(%ebx) ; /* store phys addr */ \ |
addl $PAGE_SIZE,%eax ; /* next phys page */ \ |
addl $PDE_SIZE,%ebx ; /* next pte/pde */ \ |
loop 1b ; \ |
addl $PAGE_SIZE,%eax ; /* next phys page */ \ |
|
loop 1b ; |
|
|
/* Find end of kernel image. */ |
/* Find end of kernel image. */ |
movl $RELOC(end),%edi |
movl $RELOC(end),%edi |
|
|
#if (NKSYMS || defined(DDB) || defined(LKM)) && !defined(SYMTAB_SPACE) |
#if (NKSYMS || defined(DDB) || defined(MODULAR)) && !defined(SYMTAB_SPACE) |
/* Save the symbols (if loaded). */ |
/* Save the symbols (if loaded). */ |
movl RELOC(esym),%eax |
movl RELOC(esym),%eax |
testl %eax,%eax |
testl %eax,%eax |
Line 545 try586: /* Use the `cpuid' instruction. |
|
Line 557 try586: /* Use the `cpuid' instruction. |
|
incl %eax /* one more ptp for VAs stolen by bootstrap */ |
incl %eax /* one more ptp for VAs stolen by bootstrap */ |
1: movl %eax,RELOC(nkptp)+1*4 |
1: movl %eax,RELOC(nkptp)+1*4 |
|
|
/* tablesize = (1 + UPAGES + nkptp) << PGSHIFT; */ |
/* tablesize = (PDP_SIZE + UPAGES + nkptp) << PGSHIFT; */ |
addl $(1+UPAGES),%eax |
addl $(PDP_SIZE+UPAGES),%eax |
|
#ifdef PAE |
|
incl %eax /* one more page for the L3 PD */ |
|
shll $PGSHIFT+1,%eax /* PTP tables are twice larger with PAE */ |
|
#else |
shll $PGSHIFT,%eax |
shll $PGSHIFT,%eax |
|
#endif |
movl %eax,RELOC(tablesize) |
movl %eax,RELOC(tablesize) |
|
|
/* ensure that nkptp covers bootstrap tables */ |
/* ensure that nkptp covers bootstrap tables */ |
Line 585 try586: /* Use the `cpuid' instruction. |
|
Line 602 try586: /* Use the `cpuid' instruction. |
|
*/ |
*/ |
movl $_RELOC(KERNTEXTOFF),%eax |
movl $_RELOC(KERNTEXTOFF),%eax |
movl %eax,%ecx |
movl %eax,%ecx |
shrl $(PGSHIFT-2),%ecx /* ((n >> PGSHIFT) << 2) for # pdes */ |
shrl $(PGSHIFT-2),%ecx /* ((n >> PGSHIFT) << 2) for # pdes */ |
|
#ifdef PAE |
|
shll $1,%ecx /* pdes are twice larger with PAE */ |
|
#endif |
addl %ecx,%ebx |
addl %ecx,%ebx |
|
|
/* Map the kernel text read-only. */ |
/* Map the kernel text read-only. */ |
Line 612 try586: /* Use the `cpuid' instruction. |
|
Line 632 try586: /* Use the `cpuid' instruction. |
|
* Construct a page table directory. |
* Construct a page table directory. |
*/ |
*/ |
/* Set up top level entries for identity mapping */ |
/* Set up top level entries for identity mapping */ |
leal (PROC0_PDIR_OFF)(%esi),%ebx |
leal (PROC0_PDIR_OFF)(%esi),%ebx |
leal (PROC0_PTP1_OFF)(%esi),%eax |
leal (PROC0_PTP1_OFF)(%esi),%eax |
orl $(PG_V|PG_KW), %eax |
orl $(PG_V|PG_KW), %eax |
movl RELOC(nkptp)+1*4,%ecx |
movl RELOC(nkptp)+1*4,%ecx |
fillkpt |
fillkpt |
|
|
/* Set up top level entries for actual kernel mapping */ |
/* Set up top level entries for actual kernel mapping */ |
leal (PROC0_PDIR_OFF + L2_SLOT_KERNBASE*4)(%esi),%ebx |
leal (PROC0_PDIR_OFF + L2_SLOT_KERNBASE*PDE_SIZE)(%esi),%ebx |
leal (PROC0_PTP1_OFF)(%esi),%eax |
leal (PROC0_PTP1_OFF)(%esi),%eax |
orl $(PG_V|PG_KW), %eax |
orl $(PG_V|PG_KW), %eax |
movl RELOC(nkptp)+1*4,%ecx |
movl RELOC(nkptp)+1*4,%ecx |
fillkpt |
fillkpt |
|
|
/* Install a PDE recursively mapping page directory as a page table! */ |
/* Install a PDE recursively mapping page directory as a page table! */ |
leal (PROC0_PDIR_OFF + PDIR_SLOT_PTE*4)(%esi),%ebx |
leal (PROC0_PDIR_OFF + PDIR_SLOT_PTE*PDE_SIZE)(%esi),%ebx |
leal (PROC0_PDIR_OFF)(%esi),%eax |
leal (PROC0_PDIR_OFF)(%esi),%eax |
orl $(PG_V|PG_KW),%eax |
orl $(PG_V|PG_KW),%eax |
movl %eax,(%ebx) |
movl $PDP_SIZE,%ecx |
|
fillkpt |
|
|
|
#ifdef PAE |
|
/* Fill in proc0 L3 page with entries pointing to the page dirs */ |
|
leal (PROC0_L3_OFF)(%esi),%ebx |
|
leal (PROC0_PDIR_OFF)(%esi),%eax |
|
orl $(PG_V),%eax |
|
movl $PDP_SIZE,%ecx |
|
fillkpt |
|
|
|
/* Enable PAE mode */ |
|
movl %cr4,%eax |
|
orl $CR4_PAE,%eax |
|
movl %eax,%cr4 |
|
#endif |
|
|
/* Save phys. addr of PDP, for libkvm. */ |
/* Save phys. addr of PDP, for libkvm. */ |
movl %esi,RELOC(PDPpaddr) |
leal (PROC0_PDIR_OFF)(%esi),%eax |
|
movl %eax,RELOC(PDPpaddr) |
|
|
/* |
/* |
* Startup checklist: |
* Startup checklist: |
* 1. Load %cr3 with pointer to PDIR. |
* 1. Load %cr3 with pointer to PDIR (or L3 PD page for PAE). |
*/ |
*/ |
movl %esi,%eax # phys address of ptd in proc 0 |
movl %esi,%eax # phys address of ptd in proc 0 |
movl %eax,%cr3 # load ptd addr into mmu |
movl %eax,%cr3 # load ptd addr into mmu |
|
|
/* |
/* |
* 2. Enable paging and the rest of it. |
* 2. Enable paging and the rest of it. |
*/ |
*/ |
movl %cr0,%eax # get control word |
movl %cr0,%eax # get control word |
# enable paging & NPX emulation |
# enable paging & NPX emulation |
orl $(CR0_PE|CR0_PG|CR0_NE|CR0_TS|CR0_EM|CR0_MP),%eax |
orl $(CR0_PE|CR0_PG|CR0_NE|CR0_TS|CR0_EM|CR0_MP|CR0_WP|CR0_AM),%eax |
movl %eax,%cr0 # and let's page NOW! |
movl %eax,%cr0 # and page NOW! |
|
|
pushl $begin # jump to high mem |
pushl $begin # jump to high mem |
ret |
ret |
|
|
* memory, remove it. |
* memory, remove it. |
*/ |
*/ |
movl _C_LABEL(nkptp)+1*4,%ecx |
movl _C_LABEL(nkptp)+1*4,%ecx |
leal (PROC0_PDIR_OFF)(%esi),%ebx # old, phys address of PDIR |
leal (PROC0_PDIR_OFF)(%esi),%ebx # old, phys address of PDIR |
addl $(KERNBASE), %ebx # new, virtual address of PDIR |
addl $(KERNBASE), %ebx # new, virtual address of PDIR |
1: movl $0,(%ebx) |
1: movl $0,(PDE_SIZE-4)(%ebx) # Upper bits (for PAE) |
addl $4,%ebx |
movl $0,(%ebx) |
|
addl $PDE_SIZE,%ebx |
loop 1b |
loop 1b |
|
|
/* Relocate atdevbase. */ |
/* Relocate atdevbase. */ |
|
|
|
|
/* Set up bootstrap stack. */ |
/* Set up bootstrap stack. */ |
leal (PROC0_STK_OFF+KERNBASE)(%esi),%eax |
leal (PROC0_STK_OFF+KERNBASE)(%esi),%eax |
movl %eax,_C_LABEL(proc0uarea) |
movl %eax,_C_LABEL(lwp0uarea) |
leal (KSTACK_SIZE-FRAMESIZE)(%eax),%esp |
leal (KSTACK_SIZE-FRAMESIZE)(%eax),%esp |
movl %esi,(KSTACK_SIZE+PCB_CR3)(%eax) # pcb->pcb_cr3 |
movl %esi,(KSTACK_SIZE+PCB_CR3)(%eax) # pcb->pcb_cr3 |
xorl %ebp,%ebp # mark end of frames |
xorl %ebp,%ebp # mark end of frames |
|
|
movl _C_LABEL(tablesize),%eax |
movl _C_LABEL(tablesize),%eax |
addl %esi,%eax # skip past stack and page tables |
addl %esi,%eax # skip past stack and page tables |
|
|
|
#ifdef PAE |
|
pushl $0 # init386() expects a 64 bits paddr_t with PAE |
|
#endif |
pushl %eax |
pushl %eax |
call _C_LABEL(init386) # wire 386 chip for unix operation |
call _C_LABEL(init386) # wire 386 chip for unix operation |
addl $4+NGDT*8,%esp # pop temporary gdt |
addl $PDE_SIZE,%esp # pop paddr_t |
|
addl $NGDT*8,%esp # pop temporary gdt |
|
|
#ifdef SAFARI_FIFO_HACK |
#ifdef SAFARI_FIFO_HACK |
movb $5,%al |
movb $5,%al |
|
|
popfl |
popfl |
|
|
cld |
cld |
#ifdef XEN3 |
|
movl %esp, %ebx # save start of available space |
movl %esp, %ebx # save start of available space |
#else |
|
movl %esi,%ebx # save start_info pointer |
|
#endif |
|
movl $_RELOC(tmpstk),%esp # bootstrap stack end location |
movl $_RELOC(tmpstk),%esp # bootstrap stack end location |
|
|
/* Clear BSS first so that there are no surprises... */ |
/* Clear BSS first so that there are no surprises... */ |
|
|
cpuid |
cpuid |
movl %eax,RELOC(cpuid_level) |
movl %eax,RELOC(cpuid_level) |
|
|
|
/* |
|
* Use a temp page. We'll re- add it to uvm(9) once we're |
|
* done using it. |
|
*/ |
|
movl $RELOC(tmpgdt), %eax |
|
pushl %eax # start of temporary gdt |
|
call _C_LABEL(initgdt) |
|
addl $4,%esp |
|
|
call xen_pmap_bootstrap |
call xen_pmap_bootstrap |
|
|
/* |
/* |
* First avail returned by xen_pmap_bootstrap in %eax |
* First avail returned by xen_pmap_bootstrap in %eax |
*/ |
*/ |
movl %eax, %esi; |
movl %eax, %esi; |
movl %esi, _C_LABEL(proc0uarea) |
movl %esi, _C_LABEL(lwp0uarea) |
|
|
#define PROC0PDIR ((0) * PAGE_SIZE) |
|
#define PROC0STACK ((1) * PAGE_SIZE) |
|
|
|
/* Set up bootstrap stack. */ |
/* Set up bootstrap stack. */ |
leal (KSTACK_SIZE-FRAMESIZE)(%eax),%esp |
leal (KSTACK_SIZE-FRAMESIZE)(%eax),%esp |
|
|
|
|
addl $USPACE, %esi |
addl $USPACE, %esi |
subl $KERNBASE, %esi #init386 want a physical address |
subl $KERNBASE, %esi #init386 want a physical address |
|
|
|
#ifdef PAE |
|
pushl $0 # init386() expects a 64 bits paddr_t with PAE |
|
#endif |
pushl %esi |
pushl %esi |
call _C_LABEL(init386) # wire 386 chip for unix operation |
call _C_LABEL(init386) # wire 386 chip for unix operation |
addl $4,%esp |
addl $PDE_SIZE,%esp # pop paddr_t |
call _C_LABEL(main) |
call _C_LABEL(main) |
|
|
#if defined(XEN3) && !defined(XEN_COMPAT_030001) |
#if defined(XEN) && !defined(XEN_COMPAT_030001) |
/* space for the hypercall call page */ |
/* space for the hypercall call page */ |
#define HYPERCALL_PAGE_OFFSET 0x1000 |
#define HYPERCALL_PAGE_OFFSET 0x1000 |
.org HYPERCALL_PAGE_OFFSET |
.org HYPERCALL_PAGE_OFFSET |
ENTRY(hypercall_page) |
ENTRY(hypercall_page) |
.skip 0x1000 |
.skip 0x1000 |
#endif /* defined(XEN3) && !defined(XEN_COMPAT_030001) */ |
#endif /* defined(XEN) && !defined(XEN_COMPAT_030001) */ |
|
|
/* |
/* |
* void lgdt_finish(void); |
* void lgdt_finish(void); |
|
|
END(dumpsys) |
END(dumpsys) |
|
|
/* |
/* |
* struct lwp *cpu_switchto(struct lwp *oldlwp, struct newlwp, |
* struct lwp *cpu_switchto(struct lwp *oldlwp, struct *newlwp, |
* bool returning) |
* bool returning) |
* |
* |
* 1. if (oldlwp != NULL), save its context. |
* 1. if (oldlwp != NULL), save its context. |
Line 942 ENTRY(cpu_switchto) |
|
Line 989 ENTRY(cpu_switchto) |
|
jz 1f |
jz 1f |
|
|
/* Save old context. */ |
/* Save old context. */ |
movl L_ADDR(%esi),%eax |
movl L_PCB(%esi),%eax |
movl %esp,PCB_ESP(%eax) |
movl %esp,PCB_ESP(%eax) |
movl %ebp,PCB_EBP(%eax) |
movl %ebp,PCB_EBP(%eax) |
|
|
/* Switch to newlwp's stack. */ |
/* Switch to newlwp's stack. */ |
1: movl L_ADDR(%edi),%ebx |
1: movl L_PCB(%edi),%ebx |
movl PCB_EBP(%ebx),%ebp |
movl PCB_EBP(%ebx),%ebp |
movl PCB_ESP(%ebx),%esp |
movl PCB_ESP(%ebx),%esp |
|
|
Line 966 ENTRY(cpu_switchto) |
|
Line 1013 ENTRY(cpu_switchto) |
|
pushl %edi |
pushl %edi |
call _C_LABEL(i386_switch_context) |
call _C_LABEL(i386_switch_context) |
addl $4,%esp |
addl $4,%esp |
#else /* XEN */ |
#else /* !XEN */ |
/* Switch ring0 esp */ |
/* Switch ring0 esp */ |
movl PCB_ESP0(%ebx),%eax |
movl PCB_ESP0(%ebx),%eax |
movl %eax,CPUVAR(ESP0) |
movl %eax,CPUVAR(ESP0) |
|
#endif /* !XEN */ |
|
|
/* Don't bother with the rest if switching to a system process. */ |
/* Don't bother with the rest if switching to a system process. */ |
testl $LW_SYSTEM,L_FLAG(%edi) |
testl $LW_SYSTEM,L_FLAG(%edi) |
jnz 4f |
jnz 4f |
|
|
|
#ifndef XEN |
/* Restore thread-private %fs/%gs descriptors. */ |
/* Restore thread-private %fs/%gs descriptors. */ |
movl CPUVAR(GDT),%ecx |
movl CPUVAR(GDT),%ecx |
movl PCB_FSD(%ebx), %eax |
movl PCB_FSD(%ebx), %eax |
Line 985 ENTRY(cpu_switchto) |
|
Line 1034 ENTRY(cpu_switchto) |
|
movl PCB_GSD+4(%ebx), %edx |
movl PCB_GSD+4(%ebx), %edx |
movl %eax, (GUGS_SEL*8)(%ecx) |
movl %eax, (GUGS_SEL*8)(%ecx) |
movl %edx, (GUGS_SEL*8+4)(%ecx) |
movl %edx, (GUGS_SEL*8+4)(%ecx) |
#endif /* XEN */ |
#endif /* !XEN */ |
|
|
/* Switch I/O bitmap */ |
/* Switch I/O bitmap */ |
movl PCB_IOMAP(%ebx),%eax |
movl PCB_IOMAP(%ebx),%eax |
orl %eax,%eax |
orl %eax,%eax |
jnz,pn .Lcopy_iobitmap |
jnz .Lcopy_iobitmap |
movl $(IOMAP_INVALOFF << 16),CPUVAR(IOBASE) |
movl $(IOMAP_INVALOFF << 16),CPUVAR(IOBASE) |
.Liobitmap_done: |
.Liobitmap_done: |
|
|
Line 1000 ENTRY(cpu_switchto) |
|
Line 1049 ENTRY(cpu_switchto) |
|
jne 5f |
jne 5f |
|
|
/* |
/* |
* Restore cr0 (including FPU state). Raise the IPL to IPL_IPI. |
* Restore cr0 (including FPU state). Raise the IPL to IPL_HIGH. |
* FPU IPIs can alter the LWP's saved cr0. Dropping the priority |
* FPU IPIs can alter the LWP's saved cr0. Dropping the priority |
* is deferred until mi_switch(), when cpu_switchto() returns. |
* is deferred until mi_switch(), when cpu_switchto() returns. |
*/ |
*/ |
2: |
2: |
#ifndef XEN |
#ifdef XEN |
movl $IPL_IPI,CPUVAR(ILEVEL) |
pushl %edi |
movl PCB_CR0(%ebx),%ecx |
call _C_LABEL(i386_tls_switch) |
|
addl $4,%esp |
|
#else /* !XEN */ |
|
movl $IPL_HIGH,CPUVAR(ILEVEL) |
|
movl PCB_CR0(%ebx),%ecx /* has CR0_TS clear */ |
movl %cr0,%edx |
movl %cr0,%edx |
|
|
/* |
/* |
* If our floating point registers are on a different CPU, |
* If our floating point registers are on a different CPU, |
* set CR0_TS so we'll trap rather than reuse bogus state. |
* set CR0_TS so we'll trap rather than reuse bogus state. |
*/ |
*/ |
movl PCB_FPCPU(%ebx),%eax |
cmpl CPUVAR(FPCURLWP),%edi |
cmpl CPUVAR(SELF),%eax |
|
je 3f |
je 3f |
orl $CR0_TS,%ecx |
orl $CR0_TS,%ecx |
|
|
Line 1023 ENTRY(cpu_switchto) |
|
Line 1075 ENTRY(cpu_switchto) |
|
3: cmpl %edx,%ecx |
3: cmpl %edx,%ecx |
je 4f |
je 4f |
movl %ecx,%cr0 |
movl %ecx,%cr0 |
#endif /* XEN */ |
#endif /* !XEN */ |
|
|
/* Return to the new LWP, returning 'oldlwp' in %eax. */ |
/* Return to the new LWP, returning 'oldlwp' in %eax. */ |
4: movl %esi,%eax |
4: movl %esi,%eax |
Line 1046 ENTRY(cpu_switchto) |
|
Line 1098 ENTRY(cpu_switchto) |
|
|
|
.Lcopy_iobitmap: |
.Lcopy_iobitmap: |
/* Copy I/O bitmap. */ |
/* Copy I/O bitmap. */ |
|
incl _C_LABEL(pmap_iobmp_evcnt)+EV_COUNT |
movl $(IOMAPSIZE/4),%ecx |
movl $(IOMAPSIZE/4),%ecx |
pushl %esi |
pushl %esi |
pushl %edi |
pushl %edi |
|
|
* Old call gate entry for syscall |
* Old call gate entry for syscall |
*/ |
*/ |
IDTVEC(osyscall) |
IDTVEC(osyscall) |
|
#ifndef XEN |
|
/* XXX we are in trouble! interrupts be off here. */ |
|
cli # must be first instruction |
|
#endif |
pushfl # set eflags in trap frame |
pushfl # set eflags in trap frame |
popl 8(%esp) |
popl 8(%esp) |
|
orl $PSL_I,(%esp) # re-enable ints on return to user |
pushl $7 # size of instruction for restart |
pushl $7 # size of instruction for restart |
jmp syscall1 |
jmp syscall1 |
IDTVEC_END(osyscall) |
IDTVEC_END(osyscall) |
Line 1094 IDTVEC(syscall) |
|
Line 1152 IDTVEC(syscall) |
|
syscall1: |
syscall1: |
pushl $T_ASTFLT # trap # for doing ASTs |
pushl $T_ASTFLT # trap # for doing ASTs |
INTRENTRY |
INTRENTRY |
|
STI(%eax) |
#ifdef DIAGNOSTIC |
#ifdef DIAGNOSTIC |
movl CPUVAR(ILEVEL),%ebx |
movl CPUVAR(ILEVEL),%ebx |
testl %ebx,%ebx |
testl %ebx,%ebx |
jz 1f |
jz 1f |
pushl $5f |
pushl $5f |
call _C_LABEL(printf) |
call _C_LABEL(panic) |
addl $4,%esp |
addl $4,%esp |
#ifdef DDB |
#ifdef DDB |
int $3 |
int $3 |
#endif |
#endif |
1: |
1: |
#endif /* DIAGNOSTIC */ |
#endif /* DIAGNOSTIC */ |
incl CPUVAR(NSYSCALL) # count it atomically |
addl $1,CPUVAR(NSYSCALL) # count it atomically |
|
adcl $0,CPUVAR(NSYSCALL)+4 # count it atomically |
movl CPUVAR(CURLWP),%edi |
movl CPUVAR(CURLWP),%edi |
movl L_PROC(%edi),%edx |
movl L_PROC(%edi),%edx |
movl %esp,L_MD_REGS(%edi) # save pointer to frame |
movl %esp,L_MD_REGS(%edi) # save pointer to frame |
|
|
INTRFASTEXIT |
INTRFASTEXIT |
3: STI(%eax) |
3: STI(%eax) |
pushl $4f |
pushl $4f |
call _C_LABEL(printf) |
call _C_LABEL(panic) |
addl $4,%esp |
addl $4,%esp |
pushl $IPL_NONE |
pushl $IPL_NONE |
call _C_LABEL(spllower) |
call _C_LABEL(spllower) |
addl $4,%esp |
addl $4,%esp |
jmp .Lsyscall_checkast |
jmp .Lsyscall_checkast |
4: .asciz "WARNING: SPL NOT LOWERED ON SYSCALL EXIT\n" |
4: .asciz "SPL NOT LOWERED ON SYSCALL EXIT\n" |
5: .asciz "WARNING: SPL NOT ZERO ON SYSCALL ENTRY\n" |
5: .asciz "SPL NOT ZERO ON SYSCALL ENTRY\n" |
#endif /* DIAGNOSTIC */ |
#endif /* DIAGNOSTIC */ |
9: |
9: |
cmpl $0, CPUVAR(WANT_PMAPLOAD) |
cmpl $0, CPUVAR(WANT_PMAPLOAD) |
|
|
jmp .Lsyscall_checkast /* re-check ASTs */ |
jmp .Lsyscall_checkast /* re-check ASTs */ |
IDTVEC_END(syscall) |
IDTVEC_END(syscall) |
|
|
|
IDTVEC(svr4_fasttrap) |
|
pushl $2 # size of instruction for restart |
|
pushl $T_ASTFLT # trap # for doing ASTs |
|
INTRENTRY |
|
STI(%eax) |
|
pushl $RW_READER |
|
pushl $_C_LABEL(svr4_fasttrap_lock) |
|
call _C_LABEL(rw_enter) |
|
addl $8,%esp |
|
call *_C_LABEL(svr4_fasttrap_vec) |
|
pushl $_C_LABEL(svr4_fasttrap_lock) |
|
call _C_LABEL(rw_exit) |
|
addl $4,%esp |
|
2: /* Check for ASTs on exit to user mode. */ |
|
cli |
|
CHECK_ASTPENDING(%eax) |
|
je 1f |
|
/* Always returning to user mode here. */ |
|
CLEAR_ASTPENDING(%eax) |
|
sti |
|
/* Pushed T_ASTFLT into tf_trapno on entry. */ |
|
pushl %esp |
|
call _C_LABEL(trap) |
|
addl $4,%esp |
|
jmp 2b |
|
1: CHECK_DEFERRED_SWITCH |
|
jnz 9f |
|
INTRFASTEXIT |
|
9: sti |
|
call _C_LABEL(pmap_load) |
|
cli |
|
jmp 2b |
|
|
#if NNPX > 0 |
#if NNPX > 0 |
/* |
/* |
* Special interrupt handlers. Someday intr0-intr15 will be used to count |
* Special interrupt handlers. Someday intr0-intr15 will be used to count |
Line 1240 ENTRY(sse2_idlezero_page) |
|
Line 1333 ENTRY(sse2_idlezero_page) |
|
xorl %eax, %eax |
xorl %eax, %eax |
.align 16 |
.align 16 |
1: |
1: |
cmpl $0, CPUVAR(RESCHED) |
testl $RESCHED_KPREEMPT, CPUVAR(RESCHED) |
jnz 2f |
jnz 2f |
movnti %eax, 0(%edx) |
movnti %eax, 0(%edx) |
movnti %eax, 4(%edx) |
movnti %eax, 4(%edx) |