version 1.138, 2016/10/16 10:51:31 |
version 1.158, 2018/06/29 19:21:43 |
|
|
#include <machine/asm.h> |
#include <machine/asm.h> |
__KERNEL_RCSID(0, "$NetBSD$"); |
__KERNEL_RCSID(0, "$NetBSD$"); |
|
|
#include "opt_compat_oldboot.h" |
|
#include "opt_copy_symtab.h" |
#include "opt_copy_symtab.h" |
#include "opt_ddb.h" |
#include "opt_ddb.h" |
#include "opt_modular.h" |
#include "opt_modular.h" |
#include "opt_multiboot.h" |
#include "opt_multiboot.h" |
#include "opt_realmem.h" |
#include "opt_realmem.h" |
#include "opt_vm86.h" |
|
#include "opt_xen.h" |
#include "opt_xen.h" |
|
|
#include "assym.h" |
#include "assym.h" |
Line 270 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 268 __KERNEL_RCSID(0, "$NetBSD$"); |
|
*/ |
*/ |
.data |
.data |
|
|
#if NLAPIC > 0 |
|
.align PAGE_SIZE |
|
.globl _C_LABEL(local_apic) |
|
|
|
.type _C_LABEL(local_apic), @object |
|
LABEL(local_apic) |
|
.space PAGE_SIZE |
|
END(local_apic) |
|
#endif |
|
|
|
.globl _C_LABEL(tablesize) |
.globl _C_LABEL(tablesize) |
.globl _C_LABEL(nox_flag) |
.globl _C_LABEL(nox_flag) |
.globl _C_LABEL(cputype) |
.globl _C_LABEL(cputype) |
Line 353 _C_LABEL(Multiboot_Header): |
|
Line 341 _C_LABEL(Multiboot_Header): |
|
jne 1f |
jne 1f |
|
|
/* |
/* |
* Indeed, a multiboot-compliant boot loader executed us. We copy |
* Indeed, a multiboot-compliant boot loader executed us. We switch |
* the received Multiboot information structure into kernel's data |
* to the temporary stack, and copy the received Multiboot information |
* space to process it later -- after we are relocated. It will |
* structure into kernel's data space to process it later -- after we |
* be safer to run complex C code than doing it at this point. |
* are relocated. It will be safer to run complex C code than doing it |
|
* at this point. |
*/ |
*/ |
|
movl $_RELOC(tmpstk),%esp |
pushl %ebx /* Address of Multiboot information */ |
pushl %ebx /* Address of Multiboot information */ |
call _C_LABEL(multiboot_pre_reloc) |
call _C_LABEL(multiboot_pre_reloc) |
addl $4,%esp |
addl $4,%esp |
|
|
*/ |
*/ |
|
|
/* Find end of kernel image; brings us on (1). */ |
/* Find end of kernel image; brings us on (1). */ |
movl $RELOC(end),%edi |
movl $RELOC(__kernel_end),%edi |
|
|
#if (NKSYMS || defined(DDB) || defined(MODULAR)) && !defined(makeoptions_COPY_SYMTAB) |
#if (NKSYMS || defined(DDB) || defined(MODULAR)) && !defined(makeoptions_COPY_SYMTAB) |
/* Save the symbols (if loaded); brinds us on (2). */ |
/* Save the symbols (if loaded); brinds us on (2). */ |
|
|
incl %eax /* one more PTP for VAs stolen by bootstrap */ |
incl %eax /* one more PTP for VAs stolen by bootstrap */ |
1: movl %eax,RELOC(nkptp)+1*4 |
1: movl %eax,RELOC(nkptp)+1*4 |
|
|
/* tablesize = (PDP_SIZE + UPAGES + nkptp) << PGSHIFT; */ |
/* tablesize = (PDP_SIZE + UPAGES + nkptp[1]) << PGSHIFT; */ |
addl $(PDP_SIZE+UPAGES),%eax |
addl $(PDP_SIZE+UPAGES),%eax |
#ifdef PAE |
#ifdef PAE |
incl %eax /* one more page for L3 */ |
incl %eax /* one more page for L3 */ |
|
|
#endif |
#endif |
movl %eax,RELOC(tablesize) |
movl %eax,RELOC(tablesize) |
|
|
/* Ensure that nkptp covers BOOTSTRAP TABLES. */ |
/* Ensure that nkptp[1] covers BOOTSTRAP TABLES, ie: |
|
* (esi + tablesize) >> L2_SHIFT + 1 < nkptp[1] */ |
addl %esi,%eax |
addl %esi,%eax |
addl $~L2_FRAME,%eax |
addl $~L2_FRAME,%eax |
shrl $L2_SHIFT,%eax |
shrl $L2_SHIFT,%eax |
|
|
*/ |
*/ |
leal (PROC0_PTP1_OFF)(%esi),%ebx |
leal (PROC0_PTP1_OFF)(%esi),%ebx |
|
|
/* Skip the first MB. */ |
/* Skip the area below the kernel text. */ |
movl $(KERNTEXTOFF - KERNBASE),%ecx |
movl $(KERNTEXTOFF - KERNBASE),%ecx |
shrl $PGSHIFT,%ecx |
shrl $PGSHIFT,%ecx |
fillkpt_blank |
fillkpt_blank |
|
|
pushl $0 /* init386() expects a 64 bits paddr_t with PAE */ |
pushl $0 /* init386() expects a 64 bits paddr_t with PAE */ |
#endif |
#endif |
pushl %eax |
pushl %eax |
call _C_LABEL(init386) /* wire 386 chip for unix operation */ |
call _C_LABEL(init_bootspace) |
|
call _C_LABEL(init386) |
addl $PDE_SIZE,%esp /* pop paddr_t */ |
addl $PDE_SIZE,%esp /* pop paddr_t */ |
addl $NGDT*8,%esp /* pop temporary gdt */ |
addl $NGDT*8,%esp /* pop temporary gdt */ |
|
|
#ifdef SAFARI_FIFO_HACK |
|
movb $5,%al |
|
movw $0x37b,%dx |
|
outb %al,%dx |
|
movw $0x37f,%dx |
|
inb %dx,%al |
|
movb %al,%cl |
|
|
|
orb $1,%cl |
|
|
|
movb $5,%al |
|
movw $0x37b,%dx |
|
outb %al,%dx |
|
movw $0x37f,%dx |
|
movb %cl,%al |
|
outb %al,%dx |
|
#endif /* SAFARI_FIFO_HACK */ |
|
|
|
call _C_LABEL(main) |
call _C_LABEL(main) |
#else /* XEN */ |
#else /* XEN */ |
/* First, reset the PSL. */ |
/* First, reset the PSL. */ |
|
|
popfl |
popfl |
|
|
cld |
cld |
movl %esp, %ebx /* save start of available space */ |
|
movl $_RELOC(tmpstk),%esp /* bootstrap stack end location */ |
/* |
|
* Xen info: |
|
* - %esp -> stack, *theoretically* the last used page by Xen bootstrap |
|
*/ |
|
movl %esp,%ebx |
|
movl $_RELOC(tmpstk),%esp |
|
|
/* Clear BSS. */ |
/* Clear BSS. */ |
xorl %eax,%eax |
xorl %eax,%eax |
movl $RELOC(__bss_start),%edi |
movl $RELOC(__bss_start),%edi |
movl $RELOC(_end),%ecx |
movl $RELOC(_end),%ecx |
subl %edi,%ecx |
subl %edi,%ecx |
rep stosb |
rep |
|
stosb |
|
|
/* Copy the necessary stuff from start_info structure. */ |
/* Copy the necessary stuff from start_info structure. */ |
/* We need to copy shared_info early, so that sti/cli work */ |
/* We need to copy shared_info early, so that sti/cli work */ |
movl $RELOC(start_info_union),%edi |
movl $RELOC(start_info_union),%edi |
movl $128,%ecx |
movl $128,%ecx |
rep movsl |
rep |
|
movsl |
|
|
/* Clear segment registers; always null in proc0. */ |
/* Clear segment registers. */ |
xorl %eax,%eax |
xorl %eax,%eax |
movw %ax,%fs |
movw %ax,%fs |
movw %ax,%gs |
movw %ax,%gs |
decl %eax |
|
movl %eax,RELOC(cpuid_level) |
|
|
|
xorl %eax,%eax |
xorl %eax,%eax |
cpuid |
cpuid |
movl %eax,RELOC(cpuid_level) |
movl %eax,RELOC(cpuid_level) |
|
|
/* |
/* |
* Use a temp page. We'll re- add it to uvm(9) once we're |
* Use a temporary GDT page. We'll re-add it to uvm(9) once we're done |
* done using it. |
* using it. |
*/ |
*/ |
movl $RELOC(tmpgdt), %eax |
movl $RELOC(tmpgdt),%eax |
pushl %eax /* start of temporary gdt */ |
pushl %eax /* start of temporary gdt */ |
call _C_LABEL(initgdt) |
call _C_LABEL(initgdt) |
addl $4,%esp |
addl $4,%esp |
|
|
call xen_pmap_bootstrap |
call xen_locore |
|
|
/* |
/* |
* First avail returned by xen_pmap_bootstrap in %eax |
* The first VA available is returned by xen_locore in %eax. We |
|
* use it as the UAREA, and set up the stack here. |
*/ |
*/ |
movl %eax, %esi; |
movl %eax,%esi |
movl %esi, _C_LABEL(lwp0uarea) |
movl %esi,_C_LABEL(lwp0uarea) |
|
|
/* Set up bootstrap stack. */ |
|
leal (USPACE-FRAMESIZE)(%eax),%esp |
leal (USPACE-FRAMESIZE)(%eax),%esp |
xorl %ebp,%ebp /* mark end of frames */ |
xorl %ebp,%ebp /* mark end of frames */ |
|
|
addl $USPACE, %esi |
/* Set first_avail after the DUMMY PAGE (see xen_locore). */ |
subl $KERNBASE, %esi /* init386 wants a physical address */ |
addl $(USPACE+PAGE_SIZE),%esi |
|
subl $KERNBASE,%esi /* init386 wants a physical address */ |
|
|
#ifdef PAE |
#ifdef PAE |
pushl $0 /* init386() expects a 64 bits paddr_t with PAE */ |
pushl $0 /* init386() expects a 64 bits paddr_t with PAE */ |
#endif |
#endif |
pushl %esi |
pushl %esi |
call _C_LABEL(init386) /* wire 386 chip for unix operation */ |
call _C_LABEL(init_bootspace) |
|
call _C_LABEL(init386) |
addl $PDE_SIZE,%esp /* pop paddr_t */ |
addl $PDE_SIZE,%esp /* pop paddr_t */ |
call _C_LABEL(main) |
call _C_LABEL(main) |
#endif /* XEN */ |
#endif /* XEN */ |
|
|
addl $4,%esp |
addl $4,%esp |
#else |
#else |
movl PCB_ESP0(%ebx),%eax |
movl PCB_ESP0(%ebx),%eax |
movl %eax,CPUVAR(ESP0) |
movl CPUVAR(TSS),%ecx |
|
movl %eax,TSS_ESP0(%ecx) |
|
#endif |
|
|
|
#ifndef XEN |
|
pushl %edx |
|
movb _C_LABEL(x86_fpu_eager),%dl |
|
testb %dl,%dl |
|
jz .Lno_eagerfpu |
|
pushl %edi |
|
pushl %esi |
|
call _C_LABEL(fpu_eagerswitch) |
|
addl $8,%esp |
|
.Lno_eagerfpu: |
|
popl %edx |
#endif |
#endif |
|
|
/* Don't bother with the rest if switching to a system process. */ |
/* Don't bother with the rest if switching to a system process. */ |
|
|
movl PCB_IOMAP(%ebx),%eax |
movl PCB_IOMAP(%ebx),%eax |
orl %eax,%eax |
orl %eax,%eax |
jnz .Lcopy_iobitmap |
jnz .Lcopy_iobitmap |
movl $(IOMAP_INVALOFF << 16),CPUVAR(IOBASE) |
movl CPUVAR(TSS),%eax |
|
movl $(IOMAP_INVALOFF << 16),TSS_IOBASE(%eax) |
.Liobitmap_done: |
.Liobitmap_done: |
|
|
/* Is this process using RAS (restartable atomic sequences)? */ |
/* Is this process using RAS (restartable atomic sequences)? */ |
|
|
pushl %esi |
pushl %esi |
pushl %edi |
pushl %edi |
movl %eax,%esi /* pcb_iomap */ |
movl %eax,%esi /* pcb_iomap */ |
movl CPUVAR(SELF),%edi |
movl CPUVAR(TSS),%edi |
leal CPU_INFO_IOMAP(%edi),%edi |
leal TSS_IOMAP(%edi),%edi |
rep |
rep |
movsl |
movsl |
popl %edi |
popl %edi |
popl %esi |
popl %esi |
movl $((CPU_INFO_IOMAP - CPU_INFO_TSS) << 16),CPUVAR(IOBASE) |
movl CPUVAR(TSS),%eax |
|
movl $(IOMAP_VALIDOFF << 16),TSS_IOBASE(%eax) |
jmp .Liobitmap_done |
jmp .Liobitmap_done |
END(cpu_switchto) |
END(cpu_switchto) |
|
|
|
|
END(savectx) |
END(savectx) |
|
|
/* |
/* |
* osyscall() |
|
* |
|
* Old call gate entry for syscall |
|
*/ |
|
IDTVEC(osyscall) |
|
#ifndef XEN |
|
/* XXX we are in trouble! interrupts be off here. */ |
|
cli /* must be first instruction */ |
|
#endif |
|
pushfl /* set eflags in trap frame */ |
|
popl 8(%esp) |
|
orl $PSL_I,8(%esp) /* re-enable ints on return to user */ |
|
pushl $7 /* size of instruction for restart */ |
|
jmp syscall1 |
|
IDTVEC_END(osyscall) |
|
|
|
/* |
|
* syscall() |
* syscall() |
* |
* |
* Trap gate entry for syscall |
* Trap gate entry for syscall |
*/ |
*/ |
IDTVEC(syscall) |
IDTVEC(syscall) |
pushl $2 /* size of instruction for restart */ |
pushl $2 /* size of instruction for restart */ |
syscall1: |
|
pushl $T_ASTFLT /* trap # for doing ASTs */ |
pushl $T_ASTFLT /* trap # for doing ASTs */ |
INTRENTRY |
INTRENTRY |
STI(%eax) |
STI(%eax) |
|
|
jmp .Lsyscall_checkast /* re-check ASTs */ |
jmp .Lsyscall_checkast /* re-check ASTs */ |
IDTVEC_END(syscall) |
IDTVEC_END(syscall) |
|
|
IDTVEC(svr4_fasttrap) |
|
pushl $2 /* size of instruction for restart */ |
|
pushl $T_ASTFLT /* trap # for doing ASTs */ |
|
INTRENTRY |
|
STI(%eax) |
|
pushl $RW_READER |
|
pushl $_C_LABEL(svr4_fasttrap_lock) |
|
call _C_LABEL(rw_enter) |
|
addl $8,%esp |
|
call *_C_LABEL(svr4_fasttrap_vec) |
|
pushl $_C_LABEL(svr4_fasttrap_lock) |
|
call _C_LABEL(rw_exit) |
|
addl $4,%esp |
|
2: /* Check for ASTs on exit to user mode. */ |
|
cli |
|
CHECK_ASTPENDING(%eax) |
|
je 1f |
|
/* Always returning to user mode here. */ |
|
CLEAR_ASTPENDING(%eax) |
|
sti |
|
/* Pushed T_ASTFLT into tf_trapno on entry. */ |
|
pushl %esp |
|
call _C_LABEL(trap) |
|
addl $4,%esp |
|
jmp 2b |
|
1: CHECK_DEFERRED_SWITCH |
|
jnz 9f |
|
INTRFASTEXIT |
|
9: sti |
|
call _C_LABEL(pmap_load) |
|
cli |
|
jmp 2b |
|
|
|
/* |
/* |
* int npx586bug1(int a, int b) |
* int npx586bug1(int a, int b) |
* Used when checking for the FDIV bug on first generations pentiums. |
* Used when checking for the FDIV bug on first generations pentiums. |
Line 1440 ENTRY(sse2_idlezero_page) |
|
Line 1385 ENTRY(sse2_idlezero_page) |
|
popl %ebp |
popl %ebp |
ret |
ret |
END(sse2_idlezero_page) |
END(sse2_idlezero_page) |
|
|
|
ENTRY(intrfastexit) |
|
movw TF_GS(%esp),%gs |
|
movw TF_FS(%esp),%fs |
|
movw TF_ES(%esp),%es |
|
movw TF_DS(%esp),%ds |
|
movl TF_EDI(%esp),%edi |
|
movl TF_ESI(%esp),%esi |
|
movl TF_EBP(%esp),%ebp |
|
movl TF_EBX(%esp),%ebx |
|
movl TF_EDX(%esp),%edx |
|
movl TF_ECX(%esp),%ecx |
|
movl TF_EAX(%esp),%eax |
|
addl $(TF_PUSHSIZE+8),%esp |
|
iret |
|
END(intrfastexit) |
|
|