version 1.14, 2003/08/11 21:15:50 |
version 1.21, 2004/01/04 11:44:52 |
|
|
* @(#)locore.s 7.3 (Berkeley) 5/13/91 |
* @(#)locore.s 7.3 (Berkeley) 5/13/91 |
*/ |
*/ |
|
|
|
#include "opt_compat_netbsd.h" |
|
#include "opt_compat_oldboot.h" |
#include "opt_cputype.h" |
#include "opt_cputype.h" |
#include "opt_ddb.h" |
#include "opt_ddb.h" |
#include "opt_ipkdb.h" |
|
#include "opt_vm86.h" |
|
#include "opt_user_ldt.h" |
|
#include "opt_dummy_nops.h" |
#include "opt_dummy_nops.h" |
#include "opt_compat_oldboot.h" |
#include "opt_ipkdb.h" |
#include "opt_multiprocessor.h" |
|
#include "opt_lockdebug.h" |
#include "opt_lockdebug.h" |
|
#include "opt_multiprocessor.h" |
#include "opt_realmem.h" |
#include "opt_realmem.h" |
|
#include "opt_user_ldt.h" |
|
#include "opt_vm86.h" |
|
|
#include "npx.h" |
#include "npx.h" |
#include "assym.h" |
#include "assym.h" |
Line 705 NENTRY(proc_trampoline) |
|
Line 706 NENTRY(proc_trampoline) |
|
/* NOTREACHED */ |
/* NOTREACHED */ |
|
|
/*****************************************************************************/ |
/*****************************************************************************/ |
|
#ifdef COMPAT_16 |
/* |
/* |
* Signal trampoline; copied to top of user stack. |
* Signal trampoline; copied to top of user stack. |
*/ |
*/ |
|
|
leal 12(%esp),%eax # get pointer to sigcontext |
leal 12(%esp),%eax # get pointer to sigcontext |
movl %eax,4(%esp) # put it in the argument slot |
movl %eax,4(%esp) # put it in the argument slot |
# fake return address already there |
# fake return address already there |
|
#if defined(SYS_compat_16___sigreturn14) |
|
movl $SYS_compat_16___sigreturn14,%eax |
|
#elif defined(SYS___sigreturn14) |
movl $SYS___sigreturn14,%eax |
movl $SYS___sigreturn14,%eax |
|
#else |
|
#error "no sigreturn14 syscall" |
|
#endif |
int $0x80 # enter kernel with args on stack |
int $0x80 # enter kernel with args on stack |
movl $SYS_exit,%eax |
movl $SYS_exit,%eax |
int $0x80 # exit if sigreturn fails |
int $0x80 # exit if sigreturn fails |
.globl _C_LABEL(esigcode) |
.globl _C_LABEL(esigcode) |
_C_LABEL(esigcode): |
_C_LABEL(esigcode): |
|
#endif |
|
|
/*****************************************************************************/ |
/*****************************************************************************/ |
|
|
Line 1943 switch_restored: |
|
Line 1951 switch_restored: |
|
*/ |
*/ |
movl CPUVAR(CURLWP),%edi |
movl CPUVAR(CURLWP),%edi |
movl L_PROC(%edi),%esi |
movl L_PROC(%edi),%esi |
cmpl $0,P_NRAS(%esi) |
cmpl $0,P_RASLIST(%esi) |
je 1f |
jne 2f |
movl L_MD_REGS(%edi),%ebx |
|
movl TF_EIP(%ebx),%eax |
|
pushl %eax |
|
pushl %esi |
|
call _C_LABEL(ras_lookup) |
|
addl $8,%esp |
|
cmpl $-1,%eax |
|
je 1f |
|
movl %eax,TF_EIP(%ebx) |
|
1: |
1: |
movl $1,%ebx |
movl $1,%ebx |
|
|
|
|
popl %ebx |
popl %ebx |
ret |
ret |
|
|
|
2: # check RAS list |
|
movl L_MD_REGS(%edi),%ebx |
|
movl TF_EIP(%ebx),%eax |
|
pushl %eax |
|
pushl %esi |
|
call _C_LABEL(ras_lookup) |
|
addl $8,%esp |
|
cmpl $-1,%eax |
|
je 1b |
|
movl %eax,TF_EIP(%ebx) |
|
jmp 1b |
|
|
/* |
/* |
* void cpu_switchto(struct lwp *current, struct lwp *next) |
* void cpu_switchto(struct lwp *current, struct lwp *next) |
* Switch to the specified next LWP. |
* Switch to the specified next LWP. |
Line 2017 ENTRY(cpu_switchto) |
|
Line 2028 ENTRY(cpu_switchto) |
|
jmp switch_resume |
jmp switch_resume |
|
|
/* |
/* |
* void switch_exit(struct lwp *l, void (*exit)(struct lwp *)); |
* void cpu_exit(struct lwp *l) |
* Switch to the appropriate idle context (lwp0's if uniprocessor; the cpu's |
* Switch to the appropriate idle context (lwp0's if uniprocessor; the cpu's |
* if multiprocessor) and deallocate the address space and kernel stack for p. |
* if multiprocessor) and deallocate the address space and kernel stack for p. |
* Then jump into cpu_switch(), as if we were in the idle proc all along. |
* Then jump into cpu_switch(), as if we were in the idle proc all along. |
Line 2027 ENTRY(cpu_switchto) |
|
Line 2038 ENTRY(cpu_switchto) |
|
#endif |
#endif |
.globl _C_LABEL(uvmspace_free),_C_LABEL(kernel_map) |
.globl _C_LABEL(uvmspace_free),_C_LABEL(kernel_map) |
.globl _C_LABEL(uvm_km_free),_C_LABEL(tss_free) |
.globl _C_LABEL(uvm_km_free),_C_LABEL(tss_free) |
/* LINTSTUB: Func: void switch_exit(struct lwp *l, void (*exit)(struct lwp *)) */ |
/* LINTSTUB: Func: void cpu_exit(struct lwp *l) */ |
ENTRY(switch_exit) |
ENTRY(cpu_exit) |
movl 4(%esp),%edi # old process |
movl 4(%esp),%edi # old process |
movl 8(%esp),%eax # exit func |
|
#ifndef MULTIPROCESSOR |
#ifndef MULTIPROCESSOR |
movl $_C_LABEL(lwp0),%ebx |
movl $_C_LABEL(lwp0),%ebx |
movl L_ADDR(%ebx),%esi |
movl L_ADDR(%ebx),%esi |
Line 2049 ENTRY(switch_exit) |
|
Line 2059 ENTRY(switch_exit) |
|
movl PCB_ESP(%esi),%esp |
movl PCB_ESP(%esi),%esp |
movl PCB_EBP(%esi),%ebp |
movl PCB_EBP(%esi),%ebp |
|
|
/* Save exit func. */ |
|
pushl %eax |
|
|
|
/* Load TSS info. */ |
/* Load TSS info. */ |
#ifdef MULTIPROCESSOR |
#ifdef MULTIPROCESSOR |
movl CPUVAR(GDT),%eax |
movl CPUVAR(GDT),%eax |
Line 2081 ENTRY(switch_exit) |
|
Line 2088 ENTRY(switch_exit) |
|
sti |
sti |
|
|
/* |
/* |
* Schedule the dead process's vmspace and stack to be freed. |
* Schedule the dead LWP's stack to be freed. |
*/ |
*/ |
movl 0(%esp),%eax /* %eax = exit func */ |
pushl %edi |
movl %edi,0(%esp) /* {lwp_}exit2(l) */ |
call _C_LABEL(lwp_exit2) |
call *%eax |
|
addl $4,%esp |
addl $4,%esp |
|
|
/* Jump into cpu_switch() with the right state. */ |
/* Jump into cpu_switch() with the right state. */ |
|
|
movl CPUVAR(CURLWP),%edx |
movl CPUVAR(CURLWP),%edx |
movl %esp,L_MD_REGS(%edx) # save pointer to frame |
movl %esp,L_MD_REGS(%edx) # save pointer to frame |
movl L_PROC(%edx),%edx |
movl L_PROC(%edx),%edx |
|
pushl %esp |
call *P_MD_SYSCALL(%edx) # get pointer to syscall() function |
call *P_MD_SYSCALL(%edx) # get pointer to syscall() function |
|
addl $4,%esp |
2: /* Check for ASTs on exit to user mode. */ |
2: /* Check for ASTs on exit to user mode. */ |
cli |
cli |
CHECK_ASTPENDING(%eax) |
CHECK_ASTPENDING(%eax) |
|
|
CLEAR_ASTPENDING(%eax) |
CLEAR_ASTPENDING(%eax) |
sti |
sti |
/* Pushed T_ASTFLT into tf_trapno on entry. */ |
/* Pushed T_ASTFLT into tf_trapno on entry. */ |
|
pushl %esp |
call _C_LABEL(trap) |
call _C_LABEL(trap) |
|
addl $4,%esp |
jmp 2b |
jmp 2b |
#ifndef DIAGNOSTIC |
#ifndef DIAGNOSTIC |
1: INTRFASTEXIT |
1: INTRFASTEXIT |