version 1.269, 2010/11/12 07:59:27 |
version 1.269.2.1, 2011/06/06 09:06:53 |
|
|
#define PDB_REMOVE 0x000004 |
#define PDB_REMOVE 0x000004 |
#define PDB_CHANGEPROT 0x000008 |
#define PDB_CHANGEPROT 0x000008 |
#define PDB_ENTER 0x000010 |
#define PDB_ENTER 0x000010 |
#define PDB_DEMAP 0x000020 |
#define PDB_DEMAP 0x000020 /* used in locore */ |
#define PDB_REF 0x000040 |
#define PDB_REF 0x000040 |
#define PDB_COPY 0x000080 |
#define PDB_COPY 0x000080 |
#define PDB_MMU_ALLOC 0x000100 |
#define PDB_MMU_ALLOC 0x000100 |
#define PDB_MMU_STEAL 0x000200 |
#define PDB_MMU_STEAL 0x000200 |
#define PDB_CTX_ALLOC 0x000400 |
#define PDB_CTX_ALLOC 0x000400 |
Line 684 pmap_bootstrap(u_long kernelstart, u_lon |
|
Line 684 pmap_bootstrap(u_long kernelstart, u_lon |
|
|
|
BDPRINTF(PDB_BOOT, ("Entered pmap_bootstrap.\n")); |
BDPRINTF(PDB_BOOT, ("Entered pmap_bootstrap.\n")); |
|
|
|
cache_setup_funcs(); |
|
|
/* |
/* |
* Calculate kernel size. |
* Calculate kernel size. |
*/ |
*/ |
Line 1975 pmap_remove_all(struct pmap *pm) |
|
Line 1977 pmap_remove_all(struct pmap *pm) |
|
* XXXMRG: couldn't we do something less severe here, and |
* XXXMRG: couldn't we do something less severe here, and |
* only flush the right context on each CPU? |
* only flush the right context on each CPU? |
*/ |
*/ |
#ifdef MULTIPROCESSOR |
blast_dcache(); |
smp_blast_dcache(pmap_cpus_active); |
|
#else |
|
sp_blast_dcache(dcache_size, dcache_line_size); |
|
#endif |
|
} |
} |
|
|
/* |
/* |
Line 2338 pmap_dumpmmu(int (*dump)(dev_t, daddr_t, |
|
Line 2336 pmap_dumpmmu(int (*dump)(dev_t, daddr_t, |
|
|
|
/* Fill in MD segment header (interpreted by MD part of libkvm) */ |
/* Fill in MD segment header (interpreted by MD part of libkvm) */ |
kcpu = (cpu_kcore_hdr_t *)((long)bp + ALIGN(sizeof(kcore_seg_t))); |
kcpu = (cpu_kcore_hdr_t *)((long)bp + ALIGN(sizeof(kcore_seg_t))); |
kcpu->cputype = CPU_SUN4U; |
kcpu->cputype = cputyp; |
kcpu->kernbase = (uint64_t)KERNBASE; |
kcpu->kernbase = (uint64_t)KERNBASE; |
kcpu->cpubase = (uint64_t)CPUINFO_VA; |
kcpu->cpubase = (uint64_t)CPUINFO_VA; |
|
|
Line 3350 pmap_page_cache(struct pmap *pm, paddr_t |
|
Line 3348 pmap_page_cache(struct pmap *pm, paddr_t |
|
vaddr_t va; |
vaddr_t va; |
int rv; |
int rv; |
|
|
|
#if 0 |
|
/* |
|
* Why is this? |
|
*/ |
|
if (CPU_ISSUN4US || CPU_ISSUN4V) |
|
return; |
|
#endif |
|
|
KASSERT(mutex_owned(&pmap_lock)); |
KASSERT(mutex_owned(&pmap_lock)); |
|
|
DPRINTF(PDB_ENTER, ("pmap_page_uncache(%llx)\n", |
DPRINTF(PDB_ENTER, ("pmap_page_uncache(%llx)\n", |