version 1.65, 2008/01/21 08:22:42 |
version 1.82.4.1, 2009/05/13 17:17:49 |
|
|
|
|
|
|
/*- |
/*- |
* Copyright (c) 1998, 2000, 2004, 2006, 2007 The NetBSD Foundation, Inc. |
* Copyright (c) 1998, 2000, 2004, 2006, 2007, 2009 The NetBSD Foundation, Inc. |
* All rights reserved. |
* All rights reserved. |
* |
* |
* This code is derived from software contributed to The NetBSD Foundation |
* This code is derived from software contributed to The NetBSD Foundation |
* by Charles M. Hannum. |
* by Charles M. Hannum, and by Andrew Doran. |
* |
* |
* Redistribution and use in source and binary forms, with or without |
* Redistribution and use in source and binary forms, with or without |
* modification, are permitted provided that the following conditions |
* modification, are permitted provided that the following conditions |
|
|
* 2. Redistributions in binary form must reproduce the above copyright |
* 2. Redistributions in binary form must reproduce the above copyright |
* notice, this list of conditions and the following disclaimer in the |
* notice, this list of conditions and the following disclaimer in the |
* documentation and/or other materials provided with the distribution. |
* documentation and/or other materials provided with the distribution. |
* 3. All advertising materials mentioning features or use of this software |
|
* must display the following acknowledgement: |
|
* This product includes software developed by the NetBSD |
|
* Foundation, Inc. and its contributors. |
|
* 4. Neither the name of The NetBSD Foundation nor the names of its |
|
* contributors may be used to endorse or promote products derived |
|
* from this software without specific prior written permission. |
|
* |
* |
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS |
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS |
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED |
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED |
Line 145 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 138 __KERNEL_RCSID(0, "$NetBSD$"); |
|
|
|
#include "opt_compat_oldboot.h" |
#include "opt_compat_oldboot.h" |
#include "opt_ddb.h" |
#include "opt_ddb.h" |
|
#include "opt_modular.h" |
|
#include "opt_multiboot.h" |
#include "opt_realmem.h" |
#include "opt_realmem.h" |
#include "opt_vm86.h" |
#include "opt_vm86.h" |
#include "opt_xen.h" |
#include "opt_xen.h" |
Line 192 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 187 __KERNEL_RCSID(0, "$NetBSD$"); |
|
.ascii ",HYPERCALL_PAGE=0x00000101" |
.ascii ",HYPERCALL_PAGE=0x00000101" |
/* (???+HYPERCALL_PAGE_OFFSET)/PAGE_SIZE) */ |
/* (???+HYPERCALL_PAGE_OFFSET)/PAGE_SIZE) */ |
#endif |
#endif |
|
#ifdef PAE |
|
.ascii ",PAE=yes[extended-cr3]" |
|
#endif |
#else /* !XEN3 */ |
#else /* !XEN3 */ |
.ascii "GUEST_OS=netbsd,GUEST_VER=2.0,XEN_VER=2.0" |
.ascii "GUEST_OS=netbsd,GUEST_VER=2.0,XEN_VER=2.0" |
#endif /* XEN3 */ |
#endif /* XEN3 */ |
.ascii ",LOADER=generic" |
.ascii ",LOADER=generic" |
#if (NKSYMS || defined(DDB) || defined(LKM)) && !defined(SYMTAB_SPACE) |
#if (NKSYMS || defined(DDB) || defined(MODULAR)) && !defined(SYMTAB_SPACE) |
.ascii ",BSD_SYMTAB=yes" |
.ascii ",BSD_SYMTAB=yes" |
#endif |
#endif |
.byte 0 |
.byte 0 |
Line 208 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 206 __KERNEL_RCSID(0, "$NetBSD$"); |
|
.data |
.data |
|
|
.globl _C_LABEL(cpu) |
.globl _C_LABEL(cpu) |
|
.globl _C_LABEL(cpuid_level) |
.globl _C_LABEL(esym) |
.globl _C_LABEL(esym) |
|
.globl _C_LABEL(eblob) |
.globl _C_LABEL(atdevbase) |
.globl _C_LABEL(atdevbase) |
.globl _C_LABEL(proc0uarea),_C_LABEL(PDPpaddr) |
.globl _C_LABEL(proc0uarea),_C_LABEL(PDPpaddr) |
.globl _C_LABEL(gdt) |
.globl _C_LABEL(gdt) |
Line 239 _C_LABEL(lapic_tpr): |
|
Line 239 _C_LABEL(lapic_tpr): |
|
#endif |
#endif |
|
|
_C_LABEL(cpu): .long 0 # are we 80486, Pentium, or.. |
_C_LABEL(cpu): .long 0 # are we 80486, Pentium, or.. |
|
_C_LABEL(cpuid_level): .long 0 |
_C_LABEL(atdevbase): .long 0 # location of start of iomem in virtual |
_C_LABEL(atdevbase): .long 0 # location of start of iomem in virtual |
_C_LABEL(proc0uarea): .long 0 |
_C_LABEL(proc0uarea): .long 0 |
_C_LABEL(PDPpaddr): .long 0 # paddr of PDP, for libkvm |
_C_LABEL(PDPpaddr): .long 0 # paddr of PDP, for libkvm |
Line 312 _C_LABEL(Multiboot_Header): |
|
Line 313 _C_LABEL(Multiboot_Header): |
|
movw %ax,%fs |
movw %ax,%fs |
movw %ax,%gs |
movw %ax,%gs |
decl %eax |
decl %eax |
movl %eax,RELOC(cpu_info_primary)+CPU_INFO_LEVEL |
movl %eax,RELOC(cpuid_level) |
|
|
/* Find out our CPU type. */ |
/* Find out our CPU type. */ |
|
|
|
|
try586: /* Use the `cpuid' instruction. */ |
try586: /* Use the `cpuid' instruction. */ |
xorl %eax,%eax |
xorl %eax,%eax |
cpuid |
cpuid |
movl %eax,RELOC(cpu_info_primary)+CPU_INFO_LEVEL |
movl %eax,RELOC(cpuid_level) |
|
|
2: |
2: |
/* |
/* |
Line 492 try586: /* Use the `cpuid' instruction. |
|
Line 493 try586: /* Use the `cpuid' instruction. |
|
/* |
/* |
* Virtual address space of kernel: |
* Virtual address space of kernel: |
* |
* |
* text | data | bss | [syms] | page dir | proc0 kstack | L1 ptp |
* text | data | bss | [syms] | [blobs] | page dir | proc0 kstack | L1 ptp |
* 0 1 2 3 |
* 0 1 2 3 |
*/ |
*/ |
|
|
#define PROC0_PDIR_OFF 0 |
#define PROC0_PDIR_OFF 0 |
Line 516 try586: /* Use the `cpuid' instruction. |
|
Line 517 try586: /* Use the `cpuid' instruction. |
|
|
|
/* Find end of kernel image. */ |
/* Find end of kernel image. */ |
movl $RELOC(end),%edi |
movl $RELOC(end),%edi |
#if (NKSYMS || defined(DDB) || defined(LKM)) && !defined(SYMTAB_SPACE) |
|
|
#if (NKSYMS || defined(DDB) || defined(MODULAR)) && !defined(SYMTAB_SPACE) |
/* Save the symbols (if loaded). */ |
/* Save the symbols (if loaded). */ |
movl RELOC(esym),%eax |
movl RELOC(esym),%eax |
testl %eax,%eax |
testl %eax,%eax |
Line 526 try586: /* Use the `cpuid' instruction. |
|
Line 528 try586: /* Use the `cpuid' instruction. |
|
1: |
1: |
#endif |
#endif |
|
|
|
/* Skip over any modules/blobs. */ |
|
movl RELOC(eblob),%eax |
|
testl %eax,%eax |
|
jz 1f |
|
subl $KERNBASE,%eax |
|
movl %eax,%edi |
|
1: |
/* Compute sizes */ |
/* Compute sizes */ |
movl %edi,%esi # edi = esym ? esym : end |
movl %edi,%esi |
addl $PGOFSET,%esi # page align up |
addl $PGOFSET,%esi # page align up |
andl $~PGOFSET,%esi |
andl $~PGOFSET,%esi |
|
|
Line 641 try586: /* Use the `cpuid' instruction. |
|
Line 650 try586: /* Use the `cpuid' instruction. |
|
movl %cr0,%eax # get control word |
movl %cr0,%eax # get control word |
# enable paging & NPX emulation |
# enable paging & NPX emulation |
orl $(CR0_PE|CR0_PG|CR0_NE|CR0_TS|CR0_EM|CR0_MP),%eax |
orl $(CR0_PE|CR0_PG|CR0_NE|CR0_TS|CR0_EM|CR0_MP),%eax |
movl %eax,%cr0 # and let's page NOW! |
movl %eax,%cr0 # and page NOW! |
|
|
pushl $begin # jump to high mem |
pushl $begin # jump to high mem |
ret |
ret |
|
|
movw %ax,%fs |
movw %ax,%fs |
movw %ax,%gs |
movw %ax,%gs |
decl %eax |
decl %eax |
movl %eax,RELOC(cpu_info_primary)+CPU_INFO_LEVEL |
movl %eax,RELOC(cpuid_level) |
|
|
xorl %eax,%eax |
xorl %eax,%eax |
cpuid |
cpuid |
movl %eax,RELOC(cpu_info_primary)+CPU_INFO_LEVEL |
movl %eax,RELOC(cpuid_level) |
|
|
call xen_pmap_bootstrap |
call xen_pmap_bootstrap |
/* |
/* |
Line 800 NENTRY(lgdt_finish) |
|
Line 809 NENTRY(lgdt_finish) |
|
pushl $GSEL(GCODE_SEL, SEL_KPL) |
pushl $GSEL(GCODE_SEL, SEL_KPL) |
pushl %eax |
pushl %eax |
lret |
lret |
|
END(lgdt_finish) |
|
|
#endif /* XEN */ |
#endif /* XEN */ |
|
|
Line 813 NENTRY(lgdt_finish) |
|
Line 823 NENTRY(lgdt_finish) |
|
* NOTE: This function does not have a normal calling sequence! |
* NOTE: This function does not have a normal calling sequence! |
*/ |
*/ |
NENTRY(lwp_trampoline) |
NENTRY(lwp_trampoline) |
pushl %ebp |
movl %ebp,%edi /* for .Lsyscall_checkast */ |
xorl %ebp,%ebp |
xorl %ebp,%ebp |
|
pushl %edi |
pushl %eax |
pushl %eax |
call _C_LABEL(lwp_startup) |
call _C_LABEL(lwp_startup) |
addl $8,%esp |
addl $8,%esp |
pushl %ebx |
pushl %ebx |
call *%esi |
call *%esi |
addl $4,%esp |
addl $4,%esp |
DO_DEFERRED_SWITCH |
jmp .Lsyscall_checkast |
INTRFASTEXIT |
|
/* NOTREACHED */ |
/* NOTREACHED */ |
|
END(lwp_trampoline) |
|
|
/* |
/* |
* sigcode() |
* sigcode() |
|
|
int $0x80 # exit if sigreturn fails |
int $0x80 # exit if sigreturn fails |
.globl _C_LABEL(esigcode) |
.globl _C_LABEL(esigcode) |
_C_LABEL(esigcode): |
_C_LABEL(esigcode): |
|
END(sigcode) |
|
|
/* |
/* |
* int setjmp(label_t *) |
* int setjmp(label_t *) |
|
|
movl %edx,20(%eax) # save eip |
movl %edx,20(%eax) # save eip |
xorl %eax,%eax # return 0 |
xorl %eax,%eax # return 0 |
ret |
ret |
|
END(setjmp) |
|
|
/* |
/* |
* int longjmp(label_t *) |
* int longjmp(label_t *) |
|
|
movl %edx,(%esp) # put in return frame |
movl %edx,(%esp) # put in return frame |
movl $1,%eax # return 1 |
movl $1,%eax # return 1 |
ret |
ret |
|
END(longjmp) |
|
|
/* |
/* |
* void dumpsys(void) |
* void dumpsys(void) |
|
|
call _C_LABEL(dodumpsys) # dump! |
call _C_LABEL(dodumpsys) # dump! |
addl $(3*4), %esp # unwind switchframe |
addl $(3*4), %esp # unwind switchframe |
ret |
ret |
|
END(dumpsys) |
|
|
/* |
/* |
* struct lwp *cpu_switchto(struct lwp *oldlwp, struct newlwp, |
* struct lwp *cpu_switchto(struct lwp *oldlwp, struct *newlwp, |
* bool returning) |
* bool returning) |
* |
* |
* 1. if (oldlwp != NULL), save its context. |
* 1. if (oldlwp != NULL), save its context. |
Line 911 ENTRY(cpu_switchto) |
|
Line 926 ENTRY(cpu_switchto) |
|
pushl %esi |
pushl %esi |
pushl %edi |
pushl %edi |
|
|
|
#if defined(DIAGNOSTIC) && !defined(XEN) |
|
cmpl $IPL_SCHED,CPUVAR(ILEVEL) |
|
jbe 0f |
|
pushl CPUVAR(ILEVEL) |
|
pushl $.Lstr |
|
call _C_LABEL(panic) |
|
addl $8,%esp |
|
.Lstr: .string "cpu_switchto: switching above IPL_SCHED (%d)\0" |
|
0: |
|
#endif |
|
|
movl 16(%esp),%esi # oldlwp |
movl 16(%esp),%esi # oldlwp |
movl 20(%esp),%edi # newlwp |
movl 20(%esp),%edi # newlwp |
movl 24(%esp),%edx # returning |
movl 24(%esp),%edx # returning |
Line 983 ENTRY(cpu_switchto) |
|
Line 1009 ENTRY(cpu_switchto) |
|
2: |
2: |
#ifndef XEN |
#ifndef XEN |
movl $IPL_IPI,CPUVAR(ILEVEL) |
movl $IPL_IPI,CPUVAR(ILEVEL) |
movl PCB_CR0(%ebx),%ecx |
movl PCB_CR0(%ebx),%ecx /* has CR0_TS clear */ |
movl %cr0,%edx |
movl %cr0,%edx |
|
|
/* |
/* |
* If our floating point registers are on a different CPU, |
* If our floating point registers are on a different CPU, |
* set CR0_TS so we'll trap rather than reuse bogus state. |
* set CR0_TS so we'll trap rather than reuse bogus state. |
*/ |
*/ |
movl PCB_FPCPU(%ebx),%eax |
cmpl CPUVAR(FPCURLWP),%edi |
cmpl CPUVAR(SELF),%eax |
|
je 3f |
je 3f |
orl $CR0_TS,%ecx |
orl $CR0_TS,%ecx |
|
|
Line 1022 ENTRY(cpu_switchto) |
|
Line 1047 ENTRY(cpu_switchto) |
|
|
|
.Lcopy_iobitmap: |
.Lcopy_iobitmap: |
/* Copy I/O bitmap. */ |
/* Copy I/O bitmap. */ |
|
incl _C_LABEL(pmap_iobmp_evcnt)+EV_COUNT |
movl $(IOMAPSIZE/4),%ecx |
movl $(IOMAPSIZE/4),%ecx |
pushl %esi |
pushl %esi |
pushl %edi |
pushl %edi |
Line 1034 ENTRY(cpu_switchto) |
|
Line 1060 ENTRY(cpu_switchto) |
|
popl %esi |
popl %esi |
movl $((CPU_INFO_IOMAP - CPU_INFO_TSS) << 16),CPUVAR(IOBASE) |
movl $((CPU_INFO_IOMAP - CPU_INFO_TSS) << 16),CPUVAR(IOBASE) |
jmp .Liobitmap_done |
jmp .Liobitmap_done |
|
END(cpu_switchto) |
|
|
/* |
/* |
* void savectx(struct pcb *pcb); |
* void savectx(struct pcb *pcb); |
|
|
movl %esp,PCB_ESP(%edx) |
movl %esp,PCB_ESP(%edx) |
movl %ebp,PCB_EBP(%edx) |
movl %ebp,PCB_EBP(%edx) |
ret |
ret |
|
END(savectx) |
|
|
/* |
/* |
* osyscall() |
* osyscall() |
|
|
* Old call gate entry for syscall |
* Old call gate entry for syscall |
*/ |
*/ |
IDTVEC(osyscall) |
IDTVEC(osyscall) |
|
#ifndef XEN |
|
/* XXX we are in trouble! interrupts be off here. */ |
|
cli # must be first instruction |
|
#endif |
pushfl # set eflags in trap frame |
pushfl # set eflags in trap frame |
popl 8(%esp) |
popl 8(%esp) |
|
orl $PSL_I,(%esp) # re-enable ints on return to user |
pushl $7 # size of instruction for restart |
pushl $7 # size of instruction for restart |
jmp syscall1 |
jmp syscall1 |
|
IDTVEC_END(osyscall) |
|
|
/* |
/* |
* syscall() |
* syscall() |
Line 1067 IDTVEC(syscall) |
|
Line 1101 IDTVEC(syscall) |
|
syscall1: |
syscall1: |
pushl $T_ASTFLT # trap # for doing ASTs |
pushl $T_ASTFLT # trap # for doing ASTs |
INTRENTRY |
INTRENTRY |
|
STI(%eax) |
#ifdef DIAGNOSTIC |
#ifdef DIAGNOSTIC |
cmpl $0, CPUVAR(WANT_PMAPLOAD) |
|
jz 1f |
|
pushl $6f |
|
call _C_LABEL(printf) |
|
addl $4, %esp |
|
1: |
|
movl CPUVAR(ILEVEL),%ebx |
movl CPUVAR(ILEVEL),%ebx |
testl %ebx,%ebx |
testl %ebx,%ebx |
jz 1f |
jz 1f |
pushl $5f |
pushl $5f |
call _C_LABEL(printf) |
call _C_LABEL(panic) |
addl $4,%esp |
addl $4,%esp |
#ifdef DDB |
#ifdef DDB |
int $3 |
int $3 |
#endif |
#endif |
1: |
1: |
#endif /* DIAGNOSTIC */ |
#endif /* DIAGNOSTIC */ |
movl CPUVAR(CURLWP),%edx |
incl CPUVAR(NSYSCALL) # count it atomically |
movl %esp,L_MD_REGS(%edx) # save pointer to frame |
movl CPUVAR(CURLWP),%edi |
movl L_PROC(%edx),%edx |
movl L_PROC(%edi),%edx |
|
movl %esp,L_MD_REGS(%edi) # save pointer to frame |
pushl %esp |
pushl %esp |
call *P_MD_SYSCALL(%edx) # get pointer to syscall() function |
call *P_MD_SYSCALL(%edx) # get pointer to syscall() function |
addl $4,%esp |
addl $4,%esp |
.Lsyscall_checkast: |
.Lsyscall_checkast: |
/* Check for ASTs on exit to user mode. */ |
/* Check for ASTs on exit to user mode. */ |
CLI(%eax) |
CLI(%eax) |
CHECK_ASTPENDING(%eax) |
movl L_MD_ASTPENDING(%edi), %eax |
je 1f |
orl CPUVAR(WANT_PMAPLOAD), %eax |
/* Always returning to user mode here. */ |
|
CLEAR_ASTPENDING(%eax) |
|
STI(%eax) |
|
/* Pushed T_ASTFLT into tf_trapno on entry. */ |
|
pushl %esp |
|
call _C_LABEL(trap) |
|
addl $4,%esp |
|
jmp .Lsyscall_checkast /* re-check ASTs */ |
|
1: CHECK_DEFERRED_SWITCH |
|
jnz 9f |
jnz 9f |
#ifdef XEN |
#ifdef XEN |
STIC(%eax) |
STIC(%eax) |
|
|
INTRFASTEXIT |
INTRFASTEXIT |
3: STI(%eax) |
3: STI(%eax) |
pushl $4f |
pushl $4f |
call _C_LABEL(printf) |
call _C_LABEL(panic) |
addl $4,%esp |
addl $4,%esp |
pushl $IPL_NONE |
pushl $IPL_NONE |
call _C_LABEL(spllower) |
call _C_LABEL(spllower) |
addl $4,%esp |
addl $4,%esp |
jmp .Lsyscall_checkast |
jmp .Lsyscall_checkast |
4: .asciz "WARNING: SPL NOT LOWERED ON SYSCALL EXIT\n" |
4: .asciz "SPL NOT LOWERED ON SYSCALL EXIT\n" |
5: .asciz "WARNING: SPL NOT ZERO ON SYSCALL ENTRY\n" |
5: .asciz "SPL NOT ZERO ON SYSCALL ENTRY\n" |
6: .asciz "WARNING: WANT PMAPLOAD ON SYSCALL ENTRY\n" |
|
#endif /* DIAGNOSTIC */ |
#endif /* DIAGNOSTIC */ |
9: STI(%eax) |
9: |
|
cmpl $0, CPUVAR(WANT_PMAPLOAD) |
|
jz 10f |
|
STI(%eax) |
call _C_LABEL(pmap_load) |
call _C_LABEL(pmap_load) |
jmp .Lsyscall_checkast /* re-check ASTs */ |
jmp .Lsyscall_checkast /* re-check ASTs */ |
|
10: |
|
/* Always returning to user mode here. */ |
|
movl $0, L_MD_ASTPENDING(%edi) |
|
STI(%eax) |
|
/* Pushed T_ASTFLT into tf_trapno on entry. */ |
|
pushl %esp |
|
call _C_LABEL(trap) |
|
addl $4,%esp |
|
jmp .Lsyscall_checkast /* re-check ASTs */ |
|
IDTVEC_END(syscall) |
|
|
|
IDTVEC(svr4_fasttrap) |
|
pushl $2 # size of instruction for restart |
|
pushl $T_ASTFLT # trap # for doing ASTs |
|
INTRENTRY |
|
STI(%eax) |
|
pushl $RW_READER |
|
pushl $_C_LABEL(svr4_fasttrap_lock) |
|
call _C_LABEL(rw_enter) |
|
addl $8,%esp |
|
call *_C_LABEL(svr4_fasttrap_vec) |
|
pushl $_C_LABEL(svr4_fasttrap_lock) |
|
call _C_LABEL(rw_exit) |
|
addl $4,%esp |
|
2: /* Check for ASTs on exit to user mode. */ |
|
cli |
|
CHECK_ASTPENDING(%eax) |
|
je 1f |
|
/* Always returning to user mode here. */ |
|
CLEAR_ASTPENDING(%eax) |
|
sti |
|
/* Pushed T_ASTFLT into tf_trapno on entry. */ |
|
pushl %esp |
|
call _C_LABEL(trap) |
|
addl $4,%esp |
|
jmp 2b |
|
1: CHECK_DEFERRED_SWITCH |
|
jnz 9f |
|
INTRFASTEXIT |
|
9: sti |
|
call _C_LABEL(pmap_load) |
|
cli |
|
jmp 2b |
|
|
#if NNPX > 0 |
#if NNPX > 0 |
/* |
/* |
Line 1173 NENTRY(probeintr) |
|
Line 1238 NENTRY(probeintr) |
|
outb %al,$0xf0 # clear BUSY# latch |
outb %al,$0xf0 # clear BUSY# latch |
popl %eax |
popl %eax |
iret |
iret |
|
END(probeintr) |
|
|
/* |
/* |
* void probetrap(void) |
* void probetrap(void) |
Line 1182 NENTRY(probetrap) |
|
Line 1248 NENTRY(probetrap) |
|
incl _C_LABEL(npx_traps_while_probing) |
incl _C_LABEL(npx_traps_while_probing) |
fnclex |
fnclex |
iret |
iret |
|
END(probetrap) |
|
|
/* |
/* |
* int npx586bug1(int a, int b) |
* int npx586bug1(int a, int b) |
Line 1197 NENTRY(npx586bug1) |
|
Line 1264 NENTRY(npx586bug1) |
|
fistpl (%esp) |
fistpl (%esp) |
popl %eax |
popl %eax |
ret |
ret |
|
END(npx586bug1) |
#endif /* NNPX > 0 */ |
#endif /* NNPX > 0 */ |
|
|
/* |
/* |
* void sse2_zero_page(void *pg) |
* void sse2_idlezero_page(void *pg) |
* |
* |
* Zero a page without polluting the cache. |
* Zero a page without polluting the cache. Preemption must be |
|
* disabled by the caller. Abort if a preemption is pending. |
*/ |
*/ |
ENTRY(sse2_zero_page) |
ENTRY(sse2_idlezero_page) |
pushl %ebp |
pushl %ebp |
movl %esp,%ebp |
movl %esp,%ebp |
movl 8(%esp), %edx |
movl 8(%esp), %edx |
movl $PAGE_SIZE, %ecx |
movl $(PAGE_SIZE/32), %ecx |
xorl %eax, %eax |
xorl %eax, %eax |
.align 16 |
.align 16 |
1: |
1: |
|
testl $RESCHED_KPREEMPT, CPUVAR(RESCHED) |
|
jnz 2f |
movnti %eax, 0(%edx) |
movnti %eax, 0(%edx) |
movnti %eax, 4(%edx) |
movnti %eax, 4(%edx) |
movnti %eax, 8(%edx) |
movnti %eax, 8(%edx) |
Line 1220 ENTRY(sse2_zero_page) |
|
Line 1291 ENTRY(sse2_zero_page) |
|
movnti %eax, 20(%edx) |
movnti %eax, 20(%edx) |
movnti %eax, 24(%edx) |
movnti %eax, 24(%edx) |
movnti %eax, 28(%edx) |
movnti %eax, 28(%edx) |
subl $32, %ecx |
addl $32, %edx |
leal 32(%edx), %edx |
decl %ecx |
jnz 1b |
jnz 1b |
sfence |
sfence |
|
incl %eax |
pop %ebp |
pop %ebp |
ret |
ret |
|
2: |
/* |
|
* void sse2_copy_page(void *src, void *dst) |
|
* |
|
* Copy a page without polluting the cache. |
|
*/ |
|
ENTRY(sse2_copy_page) |
|
pushl %ebp |
|
pushl %ebx |
|
pushl %esi |
|
pushl %edi |
|
movl 20(%esp), %esi |
|
movl 24(%esp), %edi |
|
movl $PAGE_SIZE, %ebp |
|
.align 16 |
|
1: |
|
movl 0(%esi), %eax |
|
movl 4(%esi), %ebx |
|
movl 8(%esi), %ecx |
|
movl 12(%esi), %edx |
|
movnti %eax, 0(%edi) |
|
movnti %ebx, 4(%edi) |
|
movnti %ecx, 8(%edi) |
|
movnti %edx, 12(%edi) |
|
subl $16, %ebp |
|
leal 16(%esi), %esi |
|
leal 16(%edi), %edi |
|
jnz 1b |
|
sfence |
sfence |
popl %edi |
|
popl %esi |
|
popl %ebx |
|
popl %ebp |
popl %ebp |
ret |
ret |
|
END(sse2_idlezero_page) |