clang-format
This commit is contained in:
@@ -30,7 +30,6 @@
|
||||
#ifndef _MIPS_CURRENT_H_
|
||||
#define _MIPS_CURRENT_H_
|
||||
|
||||
|
||||
/*
|
||||
* Macro for current thread, or current cpu.
|
||||
*
|
||||
@@ -60,7 +59,7 @@
|
||||
*/
|
||||
|
||||
#ifdef __GNUC__
|
||||
register struct thread *curthread __asm("$23"); /* s7 register */
|
||||
register struct thread *curthread __asm("$23"); /* s7 register */
|
||||
#else
|
||||
#error "Don't know how to declare curthread in this compiler"
|
||||
#endif
|
||||
|
||||
@@ -34,25 +34,23 @@
|
||||
* MIPS machine-dependent definitions for the ELF binary format.
|
||||
*/
|
||||
|
||||
|
||||
/* The ELF executable type. */
|
||||
#define EM_MACHINE EM_MIPS
|
||||
#define EM_MACHINE EM_MIPS
|
||||
|
||||
/* Linker relocation codes. SIZE DESCRIPTION */
|
||||
#define R_MIPS_NONE 0 /* --- nop */
|
||||
#define R_MIPS_16 1 /* u16 value */
|
||||
#define R_MIPS_32 2 /* u32 value */
|
||||
#define R_MIPS_REL32 3 /* u32 value relative to patched address */
|
||||
#define R_MIPS_26 4 /* u26 j/jal instruction address field */
|
||||
#define R_MIPS_HI16 5 /* u16 %hi(sym) as below */
|
||||
#define R_MIPS_LO16 6 /* s16 %lo(sym) as below */
|
||||
#define R_MIPS_GPREL16 7 /* s16 offset from GP register */
|
||||
#define R_MIPS_LITERAL 8 /* s16 GPREL16 for file-local symbols (?) */
|
||||
#define R_MIPS_GOT16 9 /* u16 offset into global offset table */
|
||||
#define R_MIPS_PC16 10 /* s16 PC-relative reference */
|
||||
#define R_MIPS_CALL16 11 /* u16 call through global offset table */
|
||||
#define R_MIPS_GPREL32 12 /* s32 offset from GP register */
|
||||
#define R_MIPS_NONE 0 /* --- nop */
|
||||
#define R_MIPS_16 1 /* u16 value */
|
||||
#define R_MIPS_32 2 /* u32 value */
|
||||
#define R_MIPS_REL32 3 /* u32 value relative to patched address */
|
||||
#define R_MIPS_26 4 /* u26 j/jal instruction address field */
|
||||
#define R_MIPS_HI16 5 /* u16 %hi(sym) as below */
|
||||
#define R_MIPS_LO16 6 /* s16 %lo(sym) as below */
|
||||
#define R_MIPS_GPREL16 7 /* s16 offset from GP register */
|
||||
#define R_MIPS_LITERAL 8 /* s16 GPREL16 for file-local symbols (?) */
|
||||
#define R_MIPS_GOT16 9 /* u16 offset into global offset table */
|
||||
#define R_MIPS_PC16 10 /* s16 PC-relative reference */
|
||||
#define R_MIPS_CALL16 11 /* u16 call through global offset table */
|
||||
#define R_MIPS_GPREL32 12 /* s32 offset from GP register */
|
||||
/* %hi/%lo are defined so %hi(sym) << 16 + %lo(sym) = sym */
|
||||
|
||||
|
||||
#endif /* _MIPS_ELF_H_ */
|
||||
|
||||
@@ -36,39 +36,37 @@
|
||||
#ifndef _KERN_MIPS_REGDEFS_H_
|
||||
#define _KERN_MIPS_REGDEFS_H_
|
||||
|
||||
|
||||
#define z0 $0 /* always zero register */
|
||||
#define AT $1 /* assembler temp register */
|
||||
#define v0 $2 /* value 0 */
|
||||
#define v1 $3 /* value 1 */
|
||||
#define a0 $4 /* argument 0 */
|
||||
#define a1 $5 /* argument 1 */
|
||||
#define a2 $6 /* argument 2 */
|
||||
#define a3 $7 /* argument 3 */
|
||||
#define t0 $8 /* temporary (caller-save) 0 */
|
||||
#define t1 $9 /* temporary (caller-save) 1 */
|
||||
#define t2 $10 /* temporary (caller-save) 2 */
|
||||
#define t3 $11 /* temporary (caller-save) 3 */
|
||||
#define t4 $12 /* temporary (caller-save) 4 */
|
||||
#define t5 $13 /* temporary (caller-save) 5 */
|
||||
#define t6 $14 /* temporary (caller-save) 6 */
|
||||
#define t7 $15 /* temporary (caller-save) 7 */
|
||||
#define s0 $16 /* saved (callee-save) 0 */
|
||||
#define s1 $17 /* saved (callee-save) 1 */
|
||||
#define s2 $18 /* saved (callee-save) 2 */
|
||||
#define s3 $19 /* saved (callee-save) 3 */
|
||||
#define s4 $20 /* saved (callee-save) 4 */
|
||||
#define s5 $21 /* saved (callee-save) 5 */
|
||||
#define s6 $22 /* saved (callee-save) 6 */
|
||||
#define s7 $23 /* saved (callee-save) 7 */
|
||||
#define t8 $24 /* temporary (caller-save) 8 */
|
||||
#define t9 $25 /* temporary (caller-save) 9 */
|
||||
#define k0 $26 /* kernel temporary 0 */
|
||||
#define k1 $27 /* kernel temporary 1 */
|
||||
#define gp $28 /* global pointer */
|
||||
#define sp $29 /* stack pointer */
|
||||
#define s8 $30 /* saved (callee-save) 8 = frame pointer */
|
||||
#define ra $31 /* return address */
|
||||
|
||||
#define z0 $0 /* always zero register */
|
||||
#define AT $1 /* assembler temp register */
|
||||
#define v0 $2 /* value 0 */
|
||||
#define v1 $3 /* value 1 */
|
||||
#define a0 $4 /* argument 0 */
|
||||
#define a1 $5 /* argument 1 */
|
||||
#define a2 $6 /* argument 2 */
|
||||
#define a3 $7 /* argument 3 */
|
||||
#define t0 $8 /* temporary (caller-save) 0 */
|
||||
#define t1 $9 /* temporary (caller-save) 1 */
|
||||
#define t2 $10 /* temporary (caller-save) 2 */
|
||||
#define t3 $11 /* temporary (caller-save) 3 */
|
||||
#define t4 $12 /* temporary (caller-save) 4 */
|
||||
#define t5 $13 /* temporary (caller-save) 5 */
|
||||
#define t6 $14 /* temporary (caller-save) 6 */
|
||||
#define t7 $15 /* temporary (caller-save) 7 */
|
||||
#define s0 $16 /* saved (callee-save) 0 */
|
||||
#define s1 $17 /* saved (callee-save) 1 */
|
||||
#define s2 $18 /* saved (callee-save) 2 */
|
||||
#define s3 $19 /* saved (callee-save) 3 */
|
||||
#define s4 $20 /* saved (callee-save) 4 */
|
||||
#define s5 $21 /* saved (callee-save) 5 */
|
||||
#define s6 $22 /* saved (callee-save) 6 */
|
||||
#define s7 $23 /* saved (callee-save) 7 */
|
||||
#define t8 $24 /* temporary (caller-save) 8 */
|
||||
#define t9 $25 /* temporary (caller-save) 9 */
|
||||
#define k0 $26 /* kernel temporary 0 */
|
||||
#define k1 $27 /* kernel temporary 1 */
|
||||
#define gp $28 /* global pointer */
|
||||
#define sp $29 /* stack pointer */
|
||||
#define s8 $30 /* saved (callee-save) 8 = frame pointer */
|
||||
#define ra $31 /* return address */
|
||||
|
||||
#endif /* _KERN_MIPS_REGDEFS_H_ */
|
||||
|
||||
@@ -38,10 +38,9 @@
|
||||
* Must save: s0-s8, sp, ra (11 registers)
|
||||
* Don't change __JB_REGS without adjusting mips_setjmp.S accordingly.
|
||||
*/
|
||||
#define __JB_REGS 11
|
||||
#define __JB_REGS 11
|
||||
|
||||
/* A jmp_buf is an array of __JB_REGS registers */
|
||||
typedef uint32_t jmp_buf[__JB_REGS];
|
||||
|
||||
|
||||
#endif /* _MIPS_SETJMP_H_ */
|
||||
|
||||
@@ -27,7 +27,6 @@
|
||||
* SUCH DAMAGE.
|
||||
*/
|
||||
|
||||
|
||||
#ifndef _KERN_MIPS_SIGNAL_H_
|
||||
#define _KERN_MIPS_SIGNAL_H_
|
||||
|
||||
@@ -39,7 +38,7 @@
|
||||
* probably won't.)
|
||||
*/
|
||||
struct sigcontext {
|
||||
/* Dummy. */
|
||||
/* Dummy. */
|
||||
};
|
||||
|
||||
#endif /* _KERN_MIPS_SIGNAL_H_ */
|
||||
|
||||
@@ -38,21 +38,20 @@
|
||||
* See kern/types.h for an explanation of the underscores.
|
||||
*/
|
||||
|
||||
|
||||
/* Sized integer types, with convenient short names */
|
||||
typedef char __i8; /* 8-bit signed integer */
|
||||
typedef short __i16; /* 16-bit signed integer */
|
||||
typedef int __i32; /* 32-bit signed integer */
|
||||
typedef long long __i64; /* 64-bit signed integer */
|
||||
typedef char __i8; /* 8-bit signed integer */
|
||||
typedef short __i16; /* 16-bit signed integer */
|
||||
typedef int __i32; /* 32-bit signed integer */
|
||||
typedef long long __i64; /* 64-bit signed integer */
|
||||
|
||||
typedef unsigned char __u8; /* 8-bit unsigned integer */
|
||||
typedef unsigned short __u16; /* 16-bit unsigned integer */
|
||||
typedef unsigned int __u32; /* 32-bit unsigned integer */
|
||||
typedef unsigned long long __u64; /* 64-bit unsigned integer */
|
||||
typedef unsigned char __u8; /* 8-bit unsigned integer */
|
||||
typedef unsigned short __u16; /* 16-bit unsigned integer */
|
||||
typedef unsigned int __u32; /* 32-bit unsigned integer */
|
||||
typedef unsigned long long __u64; /* 64-bit unsigned integer */
|
||||
|
||||
/* Further standard C types */
|
||||
typedef long __intptr_t; /* Signed pointer-sized integer */
|
||||
typedef unsigned long __uintptr_t; /* Unsigned pointer-sized integer */
|
||||
typedef long __intptr_t; /* Signed pointer-sized integer */
|
||||
typedef unsigned long __uintptr_t; /* Unsigned pointer-sized integer */
|
||||
|
||||
/*
|
||||
* Since we're a 32-bit platform, size_t, ssize_t, and ptrdiff_t can
|
||||
@@ -62,17 +61,16 @@ typedef unsigned long __uintptr_t; /* Unsigned pointer-sized integer */
|
||||
* errors involving size_t, try changing this.
|
||||
*/
|
||||
#if 1
|
||||
typedef unsigned __size_t; /* Size of a memory region */
|
||||
typedef int __ssize_t; /* Signed type of same size */
|
||||
typedef int __ptrdiff_t; /* Difference of two pointers */
|
||||
typedef unsigned __size_t; /* Size of a memory region */
|
||||
typedef int __ssize_t; /* Signed type of same size */
|
||||
typedef int __ptrdiff_t; /* Difference of two pointers */
|
||||
#else
|
||||
typedef unsigned long __size_t; /* Size of a memory region */
|
||||
typedef long __ssize_t; /* Signed type of same size */
|
||||
typedef long __ptrdiff_t; /* Difference of two pointers */
|
||||
typedef unsigned long __size_t; /* Size of a memory region */
|
||||
typedef long __ssize_t; /* Signed type of same size */
|
||||
typedef long __ptrdiff_t; /* Difference of two pointers */
|
||||
#endif
|
||||
|
||||
/* Number of bits per byte. */
|
||||
#define __CHAR_BIT 8
|
||||
|
||||
#define __CHAR_BIT 8
|
||||
|
||||
#endif /* _KERN_MIPS_TYPES_H_ */
|
||||
|
||||
@@ -41,17 +41,14 @@
|
||||
*/
|
||||
|
||||
MEMBAR_INLINE
|
||||
void
|
||||
membar_any_any(void)
|
||||
{
|
||||
__asm volatile(
|
||||
".set push;" /* save assembler mode */
|
||||
".set mips32;" /* allow MIPS32 instructions */
|
||||
"sync;" /* do it */
|
||||
".set pop" /* restore assembler mode */
|
||||
: /* no outputs */
|
||||
: /* no inputs */
|
||||
: "memory"); /* "changes" memory */
|
||||
void membar_any_any(void) {
|
||||
__asm volatile(".set push;" /* save assembler mode */
|
||||
".set mips32;" /* allow MIPS32 instructions */
|
||||
"sync;" /* do it */
|
||||
".set pop" /* restore assembler mode */
|
||||
: /* no outputs */
|
||||
: /* no inputs */
|
||||
: "memory"); /* "changes" memory */
|
||||
}
|
||||
|
||||
MEMBAR_INLINE void membar_load_load(void) { membar_any_any(); }
|
||||
@@ -59,5 +56,4 @@ MEMBAR_INLINE void membar_store_store(void) { membar_any_any(); }
|
||||
MEMBAR_INLINE void membar_store_any(void) { membar_any_any(); }
|
||||
MEMBAR_INLINE void membar_any_store(void) { membar_any_any(); }
|
||||
|
||||
|
||||
#endif /* _MIPS_MEMBAR_H_ */
|
||||
|
||||
@@ -30,62 +30,61 @@
|
||||
#ifndef _MIPS_SPECIALREG_H_
|
||||
#define _MIPS_SPECIALREG_H_
|
||||
|
||||
|
||||
/*
|
||||
* Coprocessor 0 (system processor) register numbers
|
||||
*/
|
||||
#define c0_index $0 /* TLB entry index register */
|
||||
#define c0_random $1 /* TLB random slot register */
|
||||
#define c0_entrylo $2 /* TLB entry contents (low-order half) */
|
||||
/* c0_entrylo0 $2 */ /* MIPS-II and up only */
|
||||
/* c0_entrylo1 $3 */ /* MIPS-II and up only */
|
||||
#define c0_context $4 /* some precomputed pagetable stuff */
|
||||
/* c0_pagemask $5 */ /* MIPS-II and up only */
|
||||
/* c0_wired $6 */ /* MIPS-II and up only */
|
||||
#define c0_vaddr $8 /* virtual addr of failing memory access */
|
||||
#define c0_count $9 /* cycle counter (MIPS-II and up) */
|
||||
#define c0_entryhi $10 /* TLB entry contents (high-order half) */
|
||||
#define c0_compare $11 /* on-chip timer control (MIPS-II and up) */
|
||||
#define c0_status $12 /* processor status register */
|
||||
#define c0_cause $13 /* exception cause register */
|
||||
#define c0_epc $14 /* exception PC register */
|
||||
#define c0_prid $15 /* processor ID register */
|
||||
/* c0_config $16 */ /* MIPS-II and up only */
|
||||
/* c0_lladdr $17 */ /* MIPS-II and up only */
|
||||
/* c0_watchlo $18 */ /* MIPS-II and up only */
|
||||
/* c0_watchhi $19 */ /* MIPS-II and up only */
|
||||
#define c0_index $0 /* TLB entry index register */
|
||||
#define c0_random $1 /* TLB random slot register */
|
||||
#define c0_entrylo $2 /* TLB entry contents (low-order half) */
|
||||
/* c0_entrylo0 $2 */ /* MIPS-II and up only */
|
||||
/* c0_entrylo1 $3 */ /* MIPS-II and up only */
|
||||
#define c0_context $4 /* some precomputed pagetable stuff */
|
||||
/* c0_pagemask $5 */ /* MIPS-II and up only */
|
||||
/* c0_wired $6 */ /* MIPS-II and up only */
|
||||
#define c0_vaddr $8 /* virtual addr of failing memory access */
|
||||
#define c0_count $9 /* cycle counter (MIPS-II and up) */
|
||||
#define c0_entryhi $10 /* TLB entry contents (high-order half) */
|
||||
#define c0_compare $11 /* on-chip timer control (MIPS-II and up) */
|
||||
#define c0_status $12 /* processor status register */
|
||||
#define c0_cause $13 /* exception cause register */
|
||||
#define c0_epc $14 /* exception PC register */
|
||||
#define c0_prid $15 /* processor ID register */
|
||||
/* c0_config $16 */ /* MIPS-II and up only */
|
||||
/* c0_lladdr $17 */ /* MIPS-II and up only */
|
||||
/* c0_watchlo $18 */ /* MIPS-II and up only */
|
||||
/* c0_watchhi $19 */ /* MIPS-II and up only */
|
||||
|
||||
/*
|
||||
* Mode bits in c0_status
|
||||
*/
|
||||
#define CST_IEc 0x00000001 /* current: interrupt enable */
|
||||
#define CST_KUc 0x00000002 /* current: user mode */
|
||||
#define CST_IEp 0x00000004 /* previous: interrupt enable */
|
||||
#define CST_KUp 0x00000008 /* previous: user mode */
|
||||
#define CST_IEo 0x00000010 /* old: interrupt enable */
|
||||
#define CST_KUo 0x00000020 /* old: user mode */
|
||||
#define CST_IEc 0x00000001 /* current: interrupt enable */
|
||||
#define CST_KUc 0x00000002 /* current: user mode */
|
||||
#define CST_IEp 0x00000004 /* previous: interrupt enable */
|
||||
#define CST_KUp 0x00000008 /* previous: user mode */
|
||||
#define CST_IEo 0x00000010 /* old: interrupt enable */
|
||||
#define CST_KUo 0x00000020 /* old: user mode */
|
||||
#define CST_MODEMASK 0x0000003f /* mask for the above */
|
||||
#define CST_IRQMASK 0x0000ff00 /* mask for the individual irq enable bits */
|
||||
#define CST_BEV 0x00400000 /* bootstrap exception vectors flag */
|
||||
#define CST_IRQMASK 0x0000ff00 /* mask for the individual irq enable bits */
|
||||
#define CST_BEV 0x00400000 /* bootstrap exception vectors flag */
|
||||
|
||||
/*
|
||||
* Fields of the c0_cause register
|
||||
*/
|
||||
#define CCA_UTLB 0x00000001 /* true if UTLB exception (set by our asm) */
|
||||
#define CCA_CODE 0x0000003c /* EX_foo in trapframe.h */
|
||||
#define CCA_IRQS 0x0000ff00 /* Currently pending interrupts */
|
||||
#define CCA_COPN 0x30000000 /* Coprocessor number for EX_CPU */
|
||||
#define CCA_JD 0x80000000 /* True if exception happened in jump delay */
|
||||
#define CCA_UTLB 0x00000001 /* true if UTLB exception (set by our asm) */
|
||||
#define CCA_CODE 0x0000003c /* EX_foo in trapframe.h */
|
||||
#define CCA_IRQS 0x0000ff00 /* Currently pending interrupts */
|
||||
#define CCA_COPN 0x30000000 /* Coprocessor number for EX_CPU */
|
||||
#define CCA_JD 0x80000000 /* True if exception happened in jump delay */
|
||||
|
||||
#define CCA_CODESHIFT 2 /* shift for CCA_CODE field */
|
||||
#define CCA_CODESHIFT 2 /* shift for CCA_CODE field */
|
||||
|
||||
/*
|
||||
* Fields of the c0_index register
|
||||
*/
|
||||
#define CIN_P 0x80000000 /* nonzero -> TLB probe found nothing */
|
||||
#define CIN_INDEX 0x00003f00 /* 6-bit index into TLB */
|
||||
#define CIN_P 0x80000000 /* nonzero -> TLB probe found nothing */
|
||||
#define CIN_INDEX 0x00003f00 /* 6-bit index into TLB */
|
||||
|
||||
#define CIN_INDEXSHIFT 8 /* shift for CIN_INDEX field */
|
||||
#define CIN_INDEXSHIFT 8 /* shift for CIN_INDEX field */
|
||||
|
||||
/*
|
||||
* Fields of the c0_context register
|
||||
@@ -102,16 +101,15 @@
|
||||
* there's no other good place in the chip to put it. See discussions
|
||||
* elsewhere.
|
||||
*/
|
||||
#define CTX_VSHIFT 0x001ffffc /* shifted/masked copy of c0_vaddr */
|
||||
#define CTX_PTBASE 0xffe00000 /* page table base address */
|
||||
#define CTX_VSHIFT 0x001ffffc /* shifted/masked copy of c0_vaddr */
|
||||
#define CTX_PTBASE 0xffe00000 /* page table base address */
|
||||
|
||||
#define CTX_PTBASESHIFT 21 /* shift for CTX_PBASE field */
|
||||
#define CTX_PTBASESHIFT 21 /* shift for CTX_PBASE field */
|
||||
|
||||
/*
|
||||
* Hardwired exception handler addresses.
|
||||
*/
|
||||
#define EXADDR_UTLB 0x80000000
|
||||
#define EXADDR_GENERAL 0x80000080
|
||||
|
||||
#define EXADDR_UTLB 0x80000000
|
||||
#define EXADDR_GENERAL 0x80000080
|
||||
|
||||
#endif /* _MIPS_SPECIALREG_H_ */
|
||||
|
||||
@@ -32,12 +32,11 @@
|
||||
|
||||
#include <cdefs.h>
|
||||
|
||||
|
||||
/* Type of value needed to actually spin on */
|
||||
typedef unsigned spinlock_data_t;
|
||||
|
||||
/* Initializer for use by SPINLOCK_INITIALIZER */
|
||||
#define SPINLOCK_DATA_INITIALIZER 0
|
||||
#define SPINLOCK_DATA_INITIALIZER 0
|
||||
|
||||
/* Atomic operations on spinlock_data_t */
|
||||
SPINLOCK_INLINE
|
||||
@@ -55,10 +54,8 @@ spinlock_data_t spinlock_data_testandset(volatile spinlock_data_t *sd);
|
||||
* memory.
|
||||
*/
|
||||
SPINLOCK_INLINE
|
||||
void
|
||||
spinlock_data_set(volatile spinlock_data_t *sd, unsigned val)
|
||||
{
|
||||
*sd = val;
|
||||
void spinlock_data_set(volatile spinlock_data_t *sd, unsigned val) {
|
||||
*sd = val;
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -66,11 +63,7 @@ spinlock_data_set(volatile spinlock_data_t *sd, unsigned val)
|
||||
* instruction, and instructions are atomic with respect to memory.
|
||||
*/
|
||||
SPINLOCK_INLINE
|
||||
spinlock_data_t
|
||||
spinlock_data_get(volatile spinlock_data_t *sd)
|
||||
{
|
||||
return *sd;
|
||||
}
|
||||
spinlock_data_t spinlock_data_get(volatile spinlock_data_t *sd) { return *sd; }
|
||||
|
||||
/*
|
||||
* Test-and-set a spinlock_data_t. Use the LL/SC instructions to
|
||||
@@ -86,37 +79,34 @@ spinlock_data_get(volatile spinlock_data_t *sd)
|
||||
* to atomically update one machine word.
|
||||
*/
|
||||
SPINLOCK_INLINE
|
||||
spinlock_data_t
|
||||
spinlock_data_testandset(volatile spinlock_data_t *sd)
|
||||
{
|
||||
spinlock_data_t x;
|
||||
spinlock_data_t y;
|
||||
spinlock_data_t spinlock_data_testandset(volatile spinlock_data_t *sd) {
|
||||
spinlock_data_t x;
|
||||
spinlock_data_t y;
|
||||
|
||||
/*
|
||||
* Test-and-set using LL/SC.
|
||||
*
|
||||
* Load the existing value into X, and use Y to store 1.
|
||||
* After the SC, Y contains 1 if the store succeeded,
|
||||
* 0 if it failed.
|
||||
*
|
||||
* On failure, return 1 to pretend that the spinlock
|
||||
* was already held.
|
||||
*/
|
||||
/*
|
||||
* Test-and-set using LL/SC.
|
||||
*
|
||||
* Load the existing value into X, and use Y to store 1.
|
||||
* After the SC, Y contains 1 if the store succeeded,
|
||||
* 0 if it failed.
|
||||
*
|
||||
* On failure, return 1 to pretend that the spinlock
|
||||
* was already held.
|
||||
*/
|
||||
|
||||
y = 1;
|
||||
__asm volatile(
|
||||
".set push;" /* save assembler mode */
|
||||
".set mips32;" /* allow MIPS32 instructions */
|
||||
".set volatile;" /* avoid unwanted optimization */
|
||||
"ll %0, 0(%2);" /* x = *sd */
|
||||
"sc %1, 0(%2);" /* *sd = y; y = success? */
|
||||
".set pop" /* restore assembler mode */
|
||||
: "=&r" (x), "+r" (y) : "r" (sd));
|
||||
if (y == 0) {
|
||||
return 1;
|
||||
}
|
||||
return x;
|
||||
y = 1;
|
||||
__asm volatile(".set push;" /* save assembler mode */
|
||||
".set mips32;" /* allow MIPS32 instructions */
|
||||
".set volatile;" /* avoid unwanted optimization */
|
||||
"ll %0, 0(%2);" /* x = *sd */
|
||||
"sc %1, 0(%2);" /* *sd = y; y = success? */
|
||||
".set pop" /* restore assembler mode */
|
||||
: "=&r"(x), "+r"(y)
|
||||
: "r"(sd));
|
||||
if (y == 0) {
|
||||
return 1;
|
||||
}
|
||||
return x;
|
||||
}
|
||||
|
||||
|
||||
#endif /* _MIPS_SPINLOCK_H_ */
|
||||
|
||||
@@ -30,7 +30,6 @@
|
||||
#ifndef _MIPS_THREAD_H_
|
||||
#define _MIPS_THREAD_H_
|
||||
|
||||
|
||||
/*
|
||||
* Machine-dependent thread bits.
|
||||
*/
|
||||
@@ -40,9 +39,8 @@
|
||||
typedef void (*badfaultfunc_t)(void);
|
||||
|
||||
struct thread_machdep {
|
||||
badfaultfunc_t tm_badfaultfunc; /* fault hook used by copyin/out */
|
||||
jmp_buf tm_copyjmp; /* longjmp area used by copyin/out */
|
||||
badfaultfunc_t tm_badfaultfunc; /* fault hook used by copyin/out */
|
||||
jmp_buf tm_copyjmp; /* longjmp area used by copyin/out */
|
||||
};
|
||||
|
||||
|
||||
#endif /* _MIPS_THREAD_H_ */
|
||||
|
||||
@@ -77,14 +77,14 @@ int tlb_probe(uint32_t entryhi, uint32_t entrylo);
|
||||
*/
|
||||
|
||||
/* Fields in the high-order word */
|
||||
#define TLBHI_VPAGE 0xfffff000
|
||||
#define TLBHI_VPAGE 0xfffff000
|
||||
/* TLBHI_PID 0x00000fc0 */
|
||||
|
||||
/* Fields in the low-order word */
|
||||
#define TLBLO_PPAGE 0xfffff000
|
||||
#define TLBLO_PPAGE 0xfffff000
|
||||
#define TLBLO_NOCACHE 0x00000800
|
||||
#define TLBLO_DIRTY 0x00000400
|
||||
#define TLBLO_VALID 0x00000200
|
||||
#define TLBLO_DIRTY 0x00000400
|
||||
#define TLBLO_VALID 0x00000200
|
||||
/* TLBLO_GLOBAL 0x00000100 */
|
||||
|
||||
/*
|
||||
@@ -92,14 +92,13 @@ int tlb_probe(uint32_t entryhi, uint32_t entrylo);
|
||||
* be passed to TLBHI_INVALID; this prevents loading the same invalid
|
||||
* entry into multiple TLB slots.
|
||||
*/
|
||||
#define TLBHI_INVALID(entryno) ((0x80000+(entryno))<<12)
|
||||
#define TLBLO_INVALID() (0)
|
||||
#define TLBHI_INVALID(entryno) ((0x80000 + (entryno)) << 12)
|
||||
#define TLBLO_INVALID() (0)
|
||||
|
||||
/*
|
||||
* Number of TLB entries in the processor.
|
||||
*/
|
||||
|
||||
#define NUM_TLB 64
|
||||
|
||||
#define NUM_TLB 64
|
||||
|
||||
#endif /* _MIPS_TLB_H_ */
|
||||
|
||||
@@ -38,59 +38,59 @@
|
||||
*/
|
||||
|
||||
struct trapframe {
|
||||
uint32_t tf_vaddr; /* coprocessor 0 vaddr register */
|
||||
uint32_t tf_status; /* coprocessor 0 status register */
|
||||
uint32_t tf_cause; /* coprocessor 0 cause register */
|
||||
uint32_t tf_lo;
|
||||
uint32_t tf_hi;
|
||||
uint32_t tf_ra; /* Saved register 31 */
|
||||
uint32_t tf_at; /* Saved register 1 (AT) */
|
||||
uint32_t tf_v0; /* Saved register 2 (v0) */
|
||||
uint32_t tf_v1; /* etc. */
|
||||
uint32_t tf_a0;
|
||||
uint32_t tf_a1;
|
||||
uint32_t tf_a2;
|
||||
uint32_t tf_a3;
|
||||
uint32_t tf_t0;
|
||||
uint32_t tf_t1;
|
||||
uint32_t tf_t2;
|
||||
uint32_t tf_t3;
|
||||
uint32_t tf_t4;
|
||||
uint32_t tf_t5;
|
||||
uint32_t tf_t6;
|
||||
uint32_t tf_t7;
|
||||
uint32_t tf_s0;
|
||||
uint32_t tf_s1;
|
||||
uint32_t tf_s2;
|
||||
uint32_t tf_s3;
|
||||
uint32_t tf_s4;
|
||||
uint32_t tf_s5;
|
||||
uint32_t tf_s6;
|
||||
uint32_t tf_s7;
|
||||
uint32_t tf_t8;
|
||||
uint32_t tf_t9;
|
||||
uint32_t tf_gp;
|
||||
uint32_t tf_sp;
|
||||
uint32_t tf_s8;
|
||||
uint32_t tf_epc; /* coprocessor 0 epc register */
|
||||
uint32_t tf_vaddr; /* coprocessor 0 vaddr register */
|
||||
uint32_t tf_status; /* coprocessor 0 status register */
|
||||
uint32_t tf_cause; /* coprocessor 0 cause register */
|
||||
uint32_t tf_lo;
|
||||
uint32_t tf_hi;
|
||||
uint32_t tf_ra; /* Saved register 31 */
|
||||
uint32_t tf_at; /* Saved register 1 (AT) */
|
||||
uint32_t tf_v0; /* Saved register 2 (v0) */
|
||||
uint32_t tf_v1; /* etc. */
|
||||
uint32_t tf_a0;
|
||||
uint32_t tf_a1;
|
||||
uint32_t tf_a2;
|
||||
uint32_t tf_a3;
|
||||
uint32_t tf_t0;
|
||||
uint32_t tf_t1;
|
||||
uint32_t tf_t2;
|
||||
uint32_t tf_t3;
|
||||
uint32_t tf_t4;
|
||||
uint32_t tf_t5;
|
||||
uint32_t tf_t6;
|
||||
uint32_t tf_t7;
|
||||
uint32_t tf_s0;
|
||||
uint32_t tf_s1;
|
||||
uint32_t tf_s2;
|
||||
uint32_t tf_s3;
|
||||
uint32_t tf_s4;
|
||||
uint32_t tf_s5;
|
||||
uint32_t tf_s6;
|
||||
uint32_t tf_s7;
|
||||
uint32_t tf_t8;
|
||||
uint32_t tf_t9;
|
||||
uint32_t tf_gp;
|
||||
uint32_t tf_sp;
|
||||
uint32_t tf_s8;
|
||||
uint32_t tf_epc; /* coprocessor 0 epc register */
|
||||
};
|
||||
|
||||
/*
|
||||
* MIPS exception codes.
|
||||
*/
|
||||
#define EX_IRQ 0 /* Interrupt */
|
||||
#define EX_MOD 1 /* TLB Modify (write to read-only page) */
|
||||
#define EX_TLBL 2 /* TLB miss on load */
|
||||
#define EX_TLBS 3 /* TLB miss on store */
|
||||
#define EX_ADEL 4 /* Address error on load */
|
||||
#define EX_ADES 5 /* Address error on store */
|
||||
#define EX_IBE 6 /* Bus error on instruction fetch */
|
||||
#define EX_DBE 7 /* Bus error on data load *or* store */
|
||||
#define EX_SYS 8 /* Syscall */
|
||||
#define EX_BP 9 /* Breakpoint */
|
||||
#define EX_RI 10 /* Reserved (illegal) instruction */
|
||||
#define EX_CPU 11 /* Coprocessor unusable */
|
||||
#define EX_OVF 12 /* Arithmetic overflow */
|
||||
#define EX_IRQ 0 /* Interrupt */
|
||||
#define EX_MOD 1 /* TLB Modify (write to read-only page) */
|
||||
#define EX_TLBL 2 /* TLB miss on load */
|
||||
#define EX_TLBS 3 /* TLB miss on store */
|
||||
#define EX_ADEL 4 /* Address error on load */
|
||||
#define EX_ADES 5 /* Address error on store */
|
||||
#define EX_IBE 6 /* Bus error on instruction fetch */
|
||||
#define EX_DBE 7 /* Bus error on data load *or* store */
|
||||
#define EX_SYS 8 /* Syscall */
|
||||
#define EX_BP 9 /* Breakpoint */
|
||||
#define EX_RI 10 /* Reserved (illegal) instruction */
|
||||
#define EX_CPU 11 /* Coprocessor unusable */
|
||||
#define EX_OVF 12 /* Arithmetic overflow */
|
||||
|
||||
/*
|
||||
* Function to enter user mode. Does not return. The trapframe must
|
||||
@@ -104,5 +104,4 @@ __DEAD void mips_usermode(struct trapframe *tf);
|
||||
extern vaddr_t cpustacks[];
|
||||
extern vaddr_t cputhreads[];
|
||||
|
||||
|
||||
#endif /* _MIPS_TRAPFRAME_H_ */
|
||||
|
||||
@@ -30,13 +30,12 @@
|
||||
#ifndef _MIPS_VM_H_
|
||||
#define _MIPS_VM_H_
|
||||
|
||||
|
||||
/*
|
||||
* Machine-dependent VM system definitions.
|
||||
*/
|
||||
|
||||
#define PAGE_SIZE 4096 /* size of VM page */
|
||||
#define PAGE_FRAME 0xfffff000 /* mask for getting page number from addr */
|
||||
#define PAGE_SIZE 4096 /* size of VM page */
|
||||
#define PAGE_FRAME 0xfffff000 /* mask for getting page number from addr */
|
||||
|
||||
/*
|
||||
* MIPS-I hardwired memory layout:
|
||||
@@ -48,10 +47,10 @@
|
||||
* (mips32 is a little different)
|
||||
*/
|
||||
|
||||
#define MIPS_KUSEG 0x00000000
|
||||
#define MIPS_KSEG0 0x80000000
|
||||
#define MIPS_KSEG1 0xa0000000
|
||||
#define MIPS_KSEG2 0xc0000000
|
||||
#define MIPS_KUSEG 0x00000000
|
||||
#define MIPS_KSEG0 0x80000000
|
||||
#define MIPS_KSEG1 0xa0000000
|
||||
#define MIPS_KSEG2 0xc0000000
|
||||
|
||||
/*
|
||||
* The first 512 megs of physical space can be addressed in both kseg0 and
|
||||
@@ -65,13 +64,13 @@
|
||||
* exception handler code) when converted to a vaddr it's *not* NULL, *is*
|
||||
* a valid address, and will make a *huge* mess if you scribble on it.
|
||||
*/
|
||||
#define PADDR_TO_KVADDR(paddr) ((paddr)+MIPS_KSEG0)
|
||||
#define PADDR_TO_KVADDR(paddr) ((paddr) + MIPS_KSEG0)
|
||||
|
||||
/*
|
||||
* The top of user space. (Actually, the address immediately above the
|
||||
* last valid user address.)
|
||||
*/
|
||||
#define USERSPACETOP MIPS_KSEG0
|
||||
#define USERSPACETOP MIPS_KSEG0
|
||||
|
||||
/*
|
||||
* The starting value for the stack pointer at user level. Because
|
||||
@@ -81,7 +80,7 @@
|
||||
* We put the stack at the very top of user virtual memory because it
|
||||
* grows downwards.
|
||||
*/
|
||||
#define USERSTACK USERSPACETOP
|
||||
#define USERSTACK USERSPACETOP
|
||||
|
||||
/*
|
||||
* Interface to the low-level module that looks after the amount of
|
||||
@@ -117,13 +116,12 @@ paddr_t ram_getfirstfree(void);
|
||||
*/
|
||||
|
||||
struct tlbshootdown {
|
||||
/*
|
||||
* Change this to what you need for your VM design.
|
||||
*/
|
||||
int ts_placeholder;
|
||||
/*
|
||||
* Change this to what you need for your VM design.
|
||||
*/
|
||||
int ts_placeholder;
|
||||
};
|
||||
|
||||
#define TLBSHOOTDOWN_MAX 16
|
||||
|
||||
|
||||
#endif /* _MIPS_VM_H_ */
|
||||
|
||||
@@ -40,81 +40,68 @@
|
||||
#include <mainbus.h>
|
||||
#include <syscall.h>
|
||||
|
||||
|
||||
/* in exception-*.S */
|
||||
extern __DEAD void asm_usermode(struct trapframe *tf);
|
||||
|
||||
/* called only from assembler, so not declared in a header */
|
||||
void mips_trap(struct trapframe *tf);
|
||||
|
||||
|
||||
/* Names for trap codes */
|
||||
#define NTRAPCODES 13
|
||||
static const char *const trapcodenames[NTRAPCODES] = {
|
||||
"Interrupt",
|
||||
"TLB modify trap",
|
||||
"TLB miss on load",
|
||||
"TLB miss on store",
|
||||
"Address error on load",
|
||||
"Address error on store",
|
||||
"Bus error on code",
|
||||
"Bus error on data",
|
||||
"System call",
|
||||
"Break instruction",
|
||||
"Illegal instruction",
|
||||
"Coprocessor unusable",
|
||||
"Arithmetic overflow",
|
||||
"Interrupt", "TLB modify trap", "TLB miss on load",
|
||||
"TLB miss on store", "Address error on load", "Address error on store",
|
||||
"Bus error on code", "Bus error on data", "System call",
|
||||
"Break instruction", "Illegal instruction", "Coprocessor unusable",
|
||||
"Arithmetic overflow",
|
||||
};
|
||||
|
||||
/*
|
||||
* Function called when user-level code hits a fatal fault.
|
||||
*/
|
||||
static
|
||||
void
|
||||
kill_curthread(vaddr_t epc, unsigned code, vaddr_t vaddr)
|
||||
{
|
||||
int sig = 0;
|
||||
static void kill_curthread(vaddr_t epc, unsigned code, vaddr_t vaddr) {
|
||||
int sig = 0;
|
||||
|
||||
KASSERT(code < NTRAPCODES);
|
||||
switch (code) {
|
||||
case EX_IRQ:
|
||||
case EX_IBE:
|
||||
case EX_DBE:
|
||||
case EX_SYS:
|
||||
/* should not be seen */
|
||||
KASSERT(0);
|
||||
sig = SIGABRT;
|
||||
break;
|
||||
case EX_MOD:
|
||||
case EX_TLBL:
|
||||
case EX_TLBS:
|
||||
sig = SIGSEGV;
|
||||
break;
|
||||
case EX_ADEL:
|
||||
case EX_ADES:
|
||||
sig = SIGBUS;
|
||||
break;
|
||||
case EX_BP:
|
||||
sig = SIGTRAP;
|
||||
break;
|
||||
case EX_RI:
|
||||
sig = SIGILL;
|
||||
break;
|
||||
case EX_CPU:
|
||||
sig = SIGSEGV;
|
||||
break;
|
||||
case EX_OVF:
|
||||
sig = SIGFPE;
|
||||
break;
|
||||
}
|
||||
KASSERT(code < NTRAPCODES);
|
||||
switch (code) {
|
||||
case EX_IRQ:
|
||||
case EX_IBE:
|
||||
case EX_DBE:
|
||||
case EX_SYS:
|
||||
/* should not be seen */
|
||||
KASSERT(0);
|
||||
sig = SIGABRT;
|
||||
break;
|
||||
case EX_MOD:
|
||||
case EX_TLBL:
|
||||
case EX_TLBS:
|
||||
sig = SIGSEGV;
|
||||
break;
|
||||
case EX_ADEL:
|
||||
case EX_ADES:
|
||||
sig = SIGBUS;
|
||||
break;
|
||||
case EX_BP:
|
||||
sig = SIGTRAP;
|
||||
break;
|
||||
case EX_RI:
|
||||
sig = SIGILL;
|
||||
break;
|
||||
case EX_CPU:
|
||||
sig = SIGSEGV;
|
||||
break;
|
||||
case EX_OVF:
|
||||
sig = SIGFPE;
|
||||
break;
|
||||
}
|
||||
|
||||
/*
|
||||
* You will probably want to change this.
|
||||
*/
|
||||
/*
|
||||
* You will probably want to change this.
|
||||
*/
|
||||
|
||||
kprintf("Fatal user mode trap %u sig %d (%s, epc 0x%x, vaddr 0x%x)\n",
|
||||
code, sig, trapcodenames[code], epc, vaddr);
|
||||
panic("I don't know how to handle this\n");
|
||||
kprintf("Fatal user mode trap %u sig %d (%s, epc 0x%x, vaddr 0x%x)\n", code,
|
||||
sig, trapcodenames[code], epc, vaddr);
|
||||
panic("I don't know how to handle this\n");
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -122,229 +109,223 @@ kill_curthread(vaddr_t epc, unsigned code, vaddr_t vaddr)
|
||||
* This is called by the assembly-language exception handler once
|
||||
* the trapframe has been set up.
|
||||
*/
|
||||
void
|
||||
mips_trap(struct trapframe *tf)
|
||||
{
|
||||
uint32_t code;
|
||||
/*bool isutlb; -- not used */
|
||||
bool iskern;
|
||||
int spl;
|
||||
void mips_trap(struct trapframe *tf) {
|
||||
uint32_t code;
|
||||
/*bool isutlb; -- not used */
|
||||
bool iskern;
|
||||
int spl;
|
||||
|
||||
/* The trap frame is supposed to be 35 registers long. */
|
||||
KASSERT(sizeof(struct trapframe)==(35*4));
|
||||
/* The trap frame is supposed to be 35 registers long. */
|
||||
KASSERT(sizeof(struct trapframe) == (35 * 4));
|
||||
|
||||
/*
|
||||
* Extract the exception code info from the register fields.
|
||||
*/
|
||||
code = (tf->tf_cause & CCA_CODE) >> CCA_CODESHIFT;
|
||||
/*isutlb = (tf->tf_cause & CCA_UTLB) != 0;*/
|
||||
iskern = (tf->tf_status & CST_KUp) == 0;
|
||||
/*
|
||||
* Extract the exception code info from the register fields.
|
||||
*/
|
||||
code = (tf->tf_cause & CCA_CODE) >> CCA_CODESHIFT;
|
||||
/*isutlb = (tf->tf_cause & CCA_UTLB) != 0;*/
|
||||
iskern = (tf->tf_status & CST_KUp) == 0;
|
||||
|
||||
KASSERT(code < NTRAPCODES);
|
||||
KASSERT(code < NTRAPCODES);
|
||||
|
||||
/* Make sure we haven't run off our stack */
|
||||
if (curthread != NULL && curthread->t_stack != NULL) {
|
||||
KASSERT((vaddr_t)tf > (vaddr_t)curthread->t_stack);
|
||||
KASSERT((vaddr_t)tf < (vaddr_t)(curthread->t_stack
|
||||
+ STACK_SIZE));
|
||||
}
|
||||
/* Make sure we haven't run off our stack */
|
||||
if (curthread != NULL && curthread->t_stack != NULL) {
|
||||
KASSERT((vaddr_t)tf > (vaddr_t)curthread->t_stack);
|
||||
KASSERT((vaddr_t)tf < (vaddr_t)(curthread->t_stack + STACK_SIZE));
|
||||
}
|
||||
|
||||
/* Interrupt? Call the interrupt handler and return. */
|
||||
if (code == EX_IRQ) {
|
||||
int old_in;
|
||||
bool doadjust;
|
||||
/* Interrupt? Call the interrupt handler and return. */
|
||||
if (code == EX_IRQ) {
|
||||
int old_in;
|
||||
bool doadjust;
|
||||
|
||||
old_in = curthread->t_in_interrupt;
|
||||
curthread->t_in_interrupt = 1;
|
||||
old_in = curthread->t_in_interrupt;
|
||||
curthread->t_in_interrupt = 1;
|
||||
|
||||
/*
|
||||
* The processor has turned interrupts off; if the
|
||||
* currently recorded interrupt state is interrupts on
|
||||
* (spl of 0), adjust the recorded state to match, and
|
||||
* restore after processing the interrupt.
|
||||
*
|
||||
* How can we get an interrupt if the recorded state
|
||||
* is interrupts off? Well, as things currently stand
|
||||
* when the CPU finishes idling it flips interrupts on
|
||||
* and off to allow things to happen, but leaves
|
||||
* curspl high while doing so.
|
||||
*
|
||||
* While we're here, assert that the interrupt
|
||||
* handling code hasn't leaked a spinlock or an
|
||||
* splhigh().
|
||||
*/
|
||||
/*
|
||||
* The processor has turned interrupts off; if the
|
||||
* currently recorded interrupt state is interrupts on
|
||||
* (spl of 0), adjust the recorded state to match, and
|
||||
* restore after processing the interrupt.
|
||||
*
|
||||
* How can we get an interrupt if the recorded state
|
||||
* is interrupts off? Well, as things currently stand
|
||||
* when the CPU finishes idling it flips interrupts on
|
||||
* and off to allow things to happen, but leaves
|
||||
* curspl high while doing so.
|
||||
*
|
||||
* While we're here, assert that the interrupt
|
||||
* handling code hasn't leaked a spinlock or an
|
||||
* splhigh().
|
||||
*/
|
||||
|
||||
if (curthread->t_curspl == 0) {
|
||||
KASSERT(curthread->t_curspl == 0);
|
||||
KASSERT(curthread->t_iplhigh_count == 0);
|
||||
curthread->t_curspl = IPL_HIGH;
|
||||
curthread->t_iplhigh_count++;
|
||||
doadjust = true;
|
||||
}
|
||||
else {
|
||||
doadjust = false;
|
||||
}
|
||||
if (curthread->t_curspl == 0) {
|
||||
KASSERT(curthread->t_curspl == 0);
|
||||
KASSERT(curthread->t_iplhigh_count == 0);
|
||||
curthread->t_curspl = IPL_HIGH;
|
||||
curthread->t_iplhigh_count++;
|
||||
doadjust = true;
|
||||
} else {
|
||||
doadjust = false;
|
||||
}
|
||||
|
||||
mainbus_interrupt(tf);
|
||||
mainbus_interrupt(tf);
|
||||
|
||||
if (doadjust) {
|
||||
KASSERT(curthread->t_curspl == IPL_HIGH);
|
||||
KASSERT(curthread->t_iplhigh_count == 1);
|
||||
curthread->t_iplhigh_count--;
|
||||
curthread->t_curspl = 0;
|
||||
}
|
||||
if (doadjust) {
|
||||
KASSERT(curthread->t_curspl == IPL_HIGH);
|
||||
KASSERT(curthread->t_iplhigh_count == 1);
|
||||
curthread->t_iplhigh_count--;
|
||||
curthread->t_curspl = 0;
|
||||
}
|
||||
|
||||
curthread->t_in_interrupt = old_in;
|
||||
goto done2;
|
||||
}
|
||||
curthread->t_in_interrupt = old_in;
|
||||
goto done2;
|
||||
}
|
||||
|
||||
/*
|
||||
* The processor turned interrupts off when it took the trap.
|
||||
*
|
||||
* While we're in the kernel, and not actually handling an
|
||||
* interrupt, restore the interrupt state to where it was in
|
||||
* the previous context, which may be low (interrupts on).
|
||||
*
|
||||
* Do this by forcing splhigh(), which may do a redundant
|
||||
* cpu_irqoff() but forces the stored MI interrupt state into
|
||||
* sync, then restoring the previous state.
|
||||
*/
|
||||
spl = splhigh();
|
||||
splx(spl);
|
||||
/*
|
||||
* The processor turned interrupts off when it took the trap.
|
||||
*
|
||||
* While we're in the kernel, and not actually handling an
|
||||
* interrupt, restore the interrupt state to where it was in
|
||||
* the previous context, which may be low (interrupts on).
|
||||
*
|
||||
* Do this by forcing splhigh(), which may do a redundant
|
||||
* cpu_irqoff() but forces the stored MI interrupt state into
|
||||
* sync, then restoring the previous state.
|
||||
*/
|
||||
spl = splhigh();
|
||||
splx(spl);
|
||||
|
||||
/* Syscall? Call the syscall handler and return. */
|
||||
if (code == EX_SYS) {
|
||||
/* Interrupts should have been on while in user mode. */
|
||||
KASSERT(curthread->t_curspl == 0);
|
||||
KASSERT(curthread->t_iplhigh_count == 0);
|
||||
/* Syscall? Call the syscall handler and return. */
|
||||
if (code == EX_SYS) {
|
||||
/* Interrupts should have been on while in user mode. */
|
||||
KASSERT(curthread->t_curspl == 0);
|
||||
KASSERT(curthread->t_iplhigh_count == 0);
|
||||
|
||||
DEBUG(DB_SYSCALL, "syscall: #%d, args %x %x %x %x\n",
|
||||
tf->tf_v0, tf->tf_a0, tf->tf_a1, tf->tf_a2, tf->tf_a3);
|
||||
DEBUG(DB_SYSCALL, "syscall: #%d, args %x %x %x %x\n", tf->tf_v0, tf->tf_a0,
|
||||
tf->tf_a1, tf->tf_a2, tf->tf_a3);
|
||||
|
||||
syscall(tf);
|
||||
goto done;
|
||||
}
|
||||
syscall(tf);
|
||||
goto done;
|
||||
}
|
||||
|
||||
/*
|
||||
* Ok, it wasn't any of the really easy cases.
|
||||
* Call vm_fault on the TLB exceptions.
|
||||
* Panic on the bus error exceptions.
|
||||
*/
|
||||
switch (code) {
|
||||
case EX_MOD:
|
||||
if (vm_fault(VM_FAULT_READONLY, tf->tf_vaddr)==0) {
|
||||
goto done;
|
||||
}
|
||||
break;
|
||||
case EX_TLBL:
|
||||
if (vm_fault(VM_FAULT_READ, tf->tf_vaddr)==0) {
|
||||
goto done;
|
||||
}
|
||||
break;
|
||||
case EX_TLBS:
|
||||
if (vm_fault(VM_FAULT_WRITE, tf->tf_vaddr)==0) {
|
||||
goto done;
|
||||
}
|
||||
break;
|
||||
case EX_IBE:
|
||||
case EX_DBE:
|
||||
/*
|
||||
* This means you loaded invalid TLB entries, or
|
||||
* touched invalid parts of the direct-mapped
|
||||
* segments. These are serious kernel errors, so
|
||||
* panic.
|
||||
*
|
||||
* The MIPS won't even tell you what invalid address
|
||||
* caused the bus error.
|
||||
*/
|
||||
panic("Bus error exception, PC=0x%x\n", tf->tf_epc);
|
||||
break;
|
||||
}
|
||||
/*
|
||||
* Ok, it wasn't any of the really easy cases.
|
||||
* Call vm_fault on the TLB exceptions.
|
||||
* Panic on the bus error exceptions.
|
||||
*/
|
||||
switch (code) {
|
||||
case EX_MOD:
|
||||
if (vm_fault(VM_FAULT_READONLY, tf->tf_vaddr) == 0) {
|
||||
goto done;
|
||||
}
|
||||
break;
|
||||
case EX_TLBL:
|
||||
if (vm_fault(VM_FAULT_READ, tf->tf_vaddr) == 0) {
|
||||
goto done;
|
||||
}
|
||||
break;
|
||||
case EX_TLBS:
|
||||
if (vm_fault(VM_FAULT_WRITE, tf->tf_vaddr) == 0) {
|
||||
goto done;
|
||||
}
|
||||
break;
|
||||
case EX_IBE:
|
||||
case EX_DBE:
|
||||
/*
|
||||
* This means you loaded invalid TLB entries, or
|
||||
* touched invalid parts of the direct-mapped
|
||||
* segments. These are serious kernel errors, so
|
||||
* panic.
|
||||
*
|
||||
* The MIPS won't even tell you what invalid address
|
||||
* caused the bus error.
|
||||
*/
|
||||
panic("Bus error exception, PC=0x%x\n", tf->tf_epc);
|
||||
break;
|
||||
}
|
||||
|
||||
/*
|
||||
* If we get to this point, it's a fatal fault - either it's
|
||||
* one of the other exceptions, like illegal instruction, or
|
||||
* it was a page fault we couldn't handle.
|
||||
*/
|
||||
/*
|
||||
* If we get to this point, it's a fatal fault - either it's
|
||||
* one of the other exceptions, like illegal instruction, or
|
||||
* it was a page fault we couldn't handle.
|
||||
*/
|
||||
|
||||
if (!iskern) {
|
||||
/*
|
||||
* Fatal fault in user mode.
|
||||
* Kill the current user process.
|
||||
*/
|
||||
kill_curthread(tf->tf_epc, code, tf->tf_vaddr);
|
||||
goto done;
|
||||
}
|
||||
if (!iskern) {
|
||||
/*
|
||||
* Fatal fault in user mode.
|
||||
* Kill the current user process.
|
||||
*/
|
||||
kill_curthread(tf->tf_epc, code, tf->tf_vaddr);
|
||||
goto done;
|
||||
}
|
||||
|
||||
/*
|
||||
* Fatal fault in kernel mode.
|
||||
*
|
||||
* If pcb_badfaultfunc is set, we do not panic; badfaultfunc is
|
||||
* set by copyin/copyout and related functions to signify that
|
||||
* the addresses they're accessing are userlevel-supplied and
|
||||
* not trustable. What we actually want to do is resume
|
||||
* execution at the function pointed to by badfaultfunc. That's
|
||||
* going to be "copyfail" (see copyinout.c), which longjmps
|
||||
* back to copyin/copyout or wherever and returns EFAULT.
|
||||
*
|
||||
* Note that we do not just *call* this function, because that
|
||||
* won't necessarily do anything. We want the control flow
|
||||
* that is currently executing in copyin (or whichever), and
|
||||
* is stopped while we process the exception, to *teleport* to
|
||||
* copyfail.
|
||||
*
|
||||
* This is accomplished by changing tf->tf_epc and returning
|
||||
* from the exception handler.
|
||||
*/
|
||||
/*
|
||||
* Fatal fault in kernel mode.
|
||||
*
|
||||
* If pcb_badfaultfunc is set, we do not panic; badfaultfunc is
|
||||
* set by copyin/copyout and related functions to signify that
|
||||
* the addresses they're accessing are userlevel-supplied and
|
||||
* not trustable. What we actually want to do is resume
|
||||
* execution at the function pointed to by badfaultfunc. That's
|
||||
* going to be "copyfail" (see copyinout.c), which longjmps
|
||||
* back to copyin/copyout or wherever and returns EFAULT.
|
||||
*
|
||||
* Note that we do not just *call* this function, because that
|
||||
* won't necessarily do anything. We want the control flow
|
||||
* that is currently executing in copyin (or whichever), and
|
||||
* is stopped while we process the exception, to *teleport* to
|
||||
* copyfail.
|
||||
*
|
||||
* This is accomplished by changing tf->tf_epc and returning
|
||||
* from the exception handler.
|
||||
*/
|
||||
|
||||
if (curthread != NULL &&
|
||||
curthread->t_machdep.tm_badfaultfunc != NULL) {
|
||||
tf->tf_epc = (vaddr_t) curthread->t_machdep.tm_badfaultfunc;
|
||||
goto done;
|
||||
}
|
||||
if (curthread != NULL && curthread->t_machdep.tm_badfaultfunc != NULL) {
|
||||
tf->tf_epc = (vaddr_t)curthread->t_machdep.tm_badfaultfunc;
|
||||
goto done;
|
||||
}
|
||||
|
||||
/*
|
||||
* Really fatal kernel-mode fault.
|
||||
*/
|
||||
/*
|
||||
* Really fatal kernel-mode fault.
|
||||
*/
|
||||
|
||||
kprintf("panic: Fatal exception %u (%s) in kernel mode\n", code,
|
||||
trapcodenames[code]);
|
||||
kprintf("panic: EPC 0x%x, exception vaddr 0x%x\n",
|
||||
tf->tf_epc, tf->tf_vaddr);
|
||||
kprintf("panic: Fatal exception %u (%s) in kernel mode\n", code,
|
||||
trapcodenames[code]);
|
||||
kprintf("panic: EPC 0x%x, exception vaddr 0x%x\n", tf->tf_epc, tf->tf_vaddr);
|
||||
|
||||
panic("I can't handle this... I think I'll just die now...\n");
|
||||
panic("I can't handle this... I think I'll just die now...\n");
|
||||
|
||||
done:
|
||||
/*
|
||||
* Turn interrupts off on the processor, without affecting the
|
||||
* stored interrupt state.
|
||||
*/
|
||||
cpu_irqoff();
|
||||
done2:
|
||||
done:
|
||||
/*
|
||||
* Turn interrupts off on the processor, without affecting the
|
||||
* stored interrupt state.
|
||||
*/
|
||||
cpu_irqoff();
|
||||
done2:
|
||||
|
||||
/*
|
||||
* The boot thread can get here (e.g. on interrupt return) but
|
||||
* since it doesn't go to userlevel, it can't be returning to
|
||||
* userlevel, so there's no need to set cputhreads[] and
|
||||
* cpustacks[]. Just return.
|
||||
*/
|
||||
if (curthread->t_stack == NULL) {
|
||||
return;
|
||||
}
|
||||
/*
|
||||
* The boot thread can get here (e.g. on interrupt return) but
|
||||
* since it doesn't go to userlevel, it can't be returning to
|
||||
* userlevel, so there's no need to set cputhreads[] and
|
||||
* cpustacks[]. Just return.
|
||||
*/
|
||||
if (curthread->t_stack == NULL) {
|
||||
return;
|
||||
}
|
||||
|
||||
cputhreads[curcpu->c_number] = (vaddr_t)curthread;
|
||||
cpustacks[curcpu->c_number] = (vaddr_t)curthread->t_stack + STACK_SIZE;
|
||||
cputhreads[curcpu->c_number] = (vaddr_t)curthread;
|
||||
cpustacks[curcpu->c_number] = (vaddr_t)curthread->t_stack + STACK_SIZE;
|
||||
|
||||
/*
|
||||
* This assertion will fail if either
|
||||
* (1) curthread->t_stack is corrupted, or
|
||||
* (2) the trap frame is somehow on the wrong kernel stack.
|
||||
*
|
||||
* If cpustacks[] is corrupted, the next trap back to the
|
||||
* kernel will (most likely) hang the system, so it's better
|
||||
* to find out now.
|
||||
*/
|
||||
KASSERT(SAME_STACK(cpustacks[curcpu->c_number]-1, (vaddr_t)tf));
|
||||
/*
|
||||
* This assertion will fail if either
|
||||
* (1) curthread->t_stack is corrupted, or
|
||||
* (2) the trap frame is somehow on the wrong kernel stack.
|
||||
*
|
||||
* If cpustacks[] is corrupted, the next trap back to the
|
||||
* kernel will (most likely) hang the system, so it's better
|
||||
* to find out now.
|
||||
*/
|
||||
KASSERT(SAME_STACK(cpustacks[curcpu->c_number] - 1, (vaddr_t)tf));
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -364,43 +345,41 @@ mips_trap(struct trapframe *tf)
|
||||
* - enter_new_process, for use by exec and equivalent.
|
||||
* - enter_forked_process, in syscall.c, for use by fork.
|
||||
*/
|
||||
void
|
||||
mips_usermode(struct trapframe *tf)
|
||||
{
|
||||
void mips_usermode(struct trapframe *tf) {
|
||||
|
||||
/*
|
||||
* Interrupts should be off within the kernel while entering
|
||||
* user mode. However, while in user mode, interrupts should
|
||||
* be on. To interact properly with the spl-handling logic
|
||||
* above, we explicitly call spl0() and then call cpu_irqoff().
|
||||
*/
|
||||
spl0();
|
||||
cpu_irqoff();
|
||||
/*
|
||||
* Interrupts should be off within the kernel while entering
|
||||
* user mode. However, while in user mode, interrupts should
|
||||
* be on. To interact properly with the spl-handling logic
|
||||
* above, we explicitly call spl0() and then call cpu_irqoff().
|
||||
*/
|
||||
spl0();
|
||||
cpu_irqoff();
|
||||
|
||||
cputhreads[curcpu->c_number] = (vaddr_t)curthread;
|
||||
cpustacks[curcpu->c_number] = (vaddr_t)curthread->t_stack + STACK_SIZE;
|
||||
cputhreads[curcpu->c_number] = (vaddr_t)curthread;
|
||||
cpustacks[curcpu->c_number] = (vaddr_t)curthread->t_stack + STACK_SIZE;
|
||||
|
||||
/*
|
||||
* This assertion will fail if either
|
||||
* (1) cpustacks[] is corrupted, or
|
||||
* (2) the trap frame is not on our own kernel stack, or
|
||||
* (3) the boot thread tries to enter user mode.
|
||||
*
|
||||
* If cpustacks[] is corrupted, the next trap back to the
|
||||
* kernel will (most likely) hang the system, so it's better
|
||||
* to find out now.
|
||||
*
|
||||
* It's necessary for the trap frame used here to be on the
|
||||
* current thread's own stack. It cannot correctly be on
|
||||
* either another thread's stack or in the kernel heap.
|
||||
* (Exercise: why?)
|
||||
*/
|
||||
KASSERT(SAME_STACK(cpustacks[curcpu->c_number]-1, (vaddr_t)tf));
|
||||
/*
|
||||
* This assertion will fail if either
|
||||
* (1) cpustacks[] is corrupted, or
|
||||
* (2) the trap frame is not on our own kernel stack, or
|
||||
* (3) the boot thread tries to enter user mode.
|
||||
*
|
||||
* If cpustacks[] is corrupted, the next trap back to the
|
||||
* kernel will (most likely) hang the system, so it's better
|
||||
* to find out now.
|
||||
*
|
||||
* It's necessary for the trap frame used here to be on the
|
||||
* current thread's own stack. It cannot correctly be on
|
||||
* either another thread's stack or in the kernel heap.
|
||||
* (Exercise: why?)
|
||||
*/
|
||||
KASSERT(SAME_STACK(cpustacks[curcpu->c_number] - 1, (vaddr_t)tf));
|
||||
|
||||
/*
|
||||
* This actually does it. See exception-*.S.
|
||||
*/
|
||||
asm_usermode(tf);
|
||||
/*
|
||||
* This actually does it. See exception-*.S.
|
||||
*/
|
||||
asm_usermode(tf);
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -419,20 +398,18 @@ mips_usermode(struct trapframe *tf)
|
||||
*
|
||||
* Works by creating an ersatz trapframe.
|
||||
*/
|
||||
void
|
||||
enter_new_process(int argc, userptr_t argv, userptr_t env,
|
||||
vaddr_t stack, vaddr_t entry)
|
||||
{
|
||||
struct trapframe tf;
|
||||
void enter_new_process(int argc, userptr_t argv, userptr_t env, vaddr_t stack,
|
||||
vaddr_t entry) {
|
||||
struct trapframe tf;
|
||||
|
||||
bzero(&tf, sizeof(tf));
|
||||
bzero(&tf, sizeof(tf));
|
||||
|
||||
tf.tf_status = CST_IRQMASK | CST_IEp | CST_KUp;
|
||||
tf.tf_epc = entry;
|
||||
tf.tf_a0 = argc;
|
||||
tf.tf_a1 = (vaddr_t)argv;
|
||||
tf.tf_a2 = (vaddr_t)env;
|
||||
tf.tf_sp = stack;
|
||||
tf.tf_status = CST_IRQMASK | CST_IEp | CST_KUp;
|
||||
tf.tf_epc = entry;
|
||||
tf.tf_a0 = argc;
|
||||
tf.tf_a1 = (vaddr_t)argv;
|
||||
tf.tf_a2 = (vaddr_t)env;
|
||||
tf.tf_sp = stack;
|
||||
|
||||
mips_usermode(&tf);
|
||||
mips_usermode(&tf);
|
||||
}
|
||||
|
||||
@@ -36,7 +36,6 @@
|
||||
#include <current.h>
|
||||
#include <syscall.h>
|
||||
|
||||
|
||||
/*
|
||||
* System call dispatcher.
|
||||
*
|
||||
@@ -75,75 +74,70 @@
|
||||
* stack, starting at sp+16 to skip over the slots for the
|
||||
* registerized values, with copyin().
|
||||
*/
|
||||
void
|
||||
syscall(struct trapframe *tf)
|
||||
{
|
||||
int callno;
|
||||
int32_t retval;
|
||||
int err;
|
||||
void syscall(struct trapframe *tf) {
|
||||
int callno;
|
||||
int32_t retval;
|
||||
int err;
|
||||
|
||||
KASSERT(curthread != NULL);
|
||||
KASSERT(curthread->t_curspl == 0);
|
||||
KASSERT(curthread->t_iplhigh_count == 0);
|
||||
KASSERT(curthread != NULL);
|
||||
KASSERT(curthread->t_curspl == 0);
|
||||
KASSERT(curthread->t_iplhigh_count == 0);
|
||||
|
||||
callno = tf->tf_v0;
|
||||
callno = tf->tf_v0;
|
||||
|
||||
/*
|
||||
* Initialize retval to 0. Many of the system calls don't
|
||||
* really return a value, just 0 for success and -1 on
|
||||
* error. Since retval is the value returned on success,
|
||||
* initialize it to 0 by default; thus it's not necessary to
|
||||
* deal with it except for calls that return other values,
|
||||
* like write.
|
||||
*/
|
||||
/*
|
||||
* Initialize retval to 0. Many of the system calls don't
|
||||
* really return a value, just 0 for success and -1 on
|
||||
* error. Since retval is the value returned on success,
|
||||
* initialize it to 0 by default; thus it's not necessary to
|
||||
* deal with it except for calls that return other values,
|
||||
* like write.
|
||||
*/
|
||||
|
||||
retval = 0;
|
||||
retval = 0;
|
||||
|
||||
switch (callno) {
|
||||
case SYS_reboot:
|
||||
err = sys_reboot(tf->tf_a0);
|
||||
break;
|
||||
switch (callno) {
|
||||
case SYS_reboot:
|
||||
err = sys_reboot(tf->tf_a0);
|
||||
break;
|
||||
|
||||
case SYS___time:
|
||||
err = sys___time((userptr_t)tf->tf_a0,
|
||||
(userptr_t)tf->tf_a1);
|
||||
break;
|
||||
case SYS___time:
|
||||
err = sys___time((userptr_t)tf->tf_a0, (userptr_t)tf->tf_a1);
|
||||
break;
|
||||
|
||||
/* Add stuff here */
|
||||
/* Add stuff here */
|
||||
|
||||
default:
|
||||
kprintf("Unknown syscall %d\n", callno);
|
||||
err = ENOSYS;
|
||||
break;
|
||||
}
|
||||
default:
|
||||
kprintf("Unknown syscall %d\n", callno);
|
||||
err = ENOSYS;
|
||||
break;
|
||||
}
|
||||
|
||||
if (err) {
|
||||
/*
|
||||
* Return the error code. This gets converted at
|
||||
* userlevel to a return value of -1 and the error
|
||||
* code in errno.
|
||||
*/
|
||||
tf->tf_v0 = err;
|
||||
tf->tf_a3 = 1; /* signal an error */
|
||||
} else {
|
||||
/* Success. */
|
||||
tf->tf_v0 = retval;
|
||||
tf->tf_a3 = 0; /* signal no error */
|
||||
}
|
||||
|
||||
if (err) {
|
||||
/*
|
||||
* Return the error code. This gets converted at
|
||||
* userlevel to a return value of -1 and the error
|
||||
* code in errno.
|
||||
*/
|
||||
tf->tf_v0 = err;
|
||||
tf->tf_a3 = 1; /* signal an error */
|
||||
}
|
||||
else {
|
||||
/* Success. */
|
||||
tf->tf_v0 = retval;
|
||||
tf->tf_a3 = 0; /* signal no error */
|
||||
}
|
||||
/*
|
||||
* Now, advance the program counter, to avoid restarting
|
||||
* the syscall over and over again.
|
||||
*/
|
||||
|
||||
/*
|
||||
* Now, advance the program counter, to avoid restarting
|
||||
* the syscall over and over again.
|
||||
*/
|
||||
tf->tf_epc += 4;
|
||||
|
||||
tf->tf_epc += 4;
|
||||
|
||||
/* Make sure the syscall code didn't forget to lower spl */
|
||||
KASSERT(curthread->t_curspl == 0);
|
||||
/* ...or leak any spinlocks */
|
||||
KASSERT(curthread->t_iplhigh_count == 0);
|
||||
/* Make sure the syscall code didn't forget to lower spl */
|
||||
KASSERT(curthread->t_curspl == 0);
|
||||
/* ...or leak any spinlocks */
|
||||
KASSERT(curthread->t_iplhigh_count == 0);
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -154,8 +148,4 @@ syscall(struct trapframe *tf)
|
||||
*
|
||||
* Thus, you can trash it and do things another way if you prefer.
|
||||
*/
|
||||
void
|
||||
enter_forked_process(struct trapframe *tf)
|
||||
{
|
||||
(void)tf;
|
||||
}
|
||||
void enter_forked_process(struct trapframe *tf) { (void)tf; }
|
||||
|
||||
@@ -72,30 +72,27 @@ vaddr_t cputhreads[MAXCPUS];
|
||||
* associated with a new cpu. Note that we're not running on the new
|
||||
* cpu when this is called.
|
||||
*/
|
||||
void
|
||||
cpu_machdep_init(struct cpu *c)
|
||||
{
|
||||
vaddr_t stackpointer;
|
||||
void cpu_machdep_init(struct cpu *c) {
|
||||
vaddr_t stackpointer;
|
||||
|
||||
KASSERT(c->c_number < MAXCPUS);
|
||||
KASSERT(c->c_number < MAXCPUS);
|
||||
|
||||
if (c->c_curthread->t_stack == NULL) {
|
||||
/* boot cpu; don't need to do anything here */
|
||||
}
|
||||
else {
|
||||
/*
|
||||
* Stick the stack in cpustacks[], and thread pointer
|
||||
* in cputhreads[].
|
||||
*/
|
||||
if (c->c_curthread->t_stack == NULL) {
|
||||
/* boot cpu; don't need to do anything here */
|
||||
} else {
|
||||
/*
|
||||
* Stick the stack in cpustacks[], and thread pointer
|
||||
* in cputhreads[].
|
||||
*/
|
||||
|
||||
/* stack base address */
|
||||
stackpointer = (vaddr_t) c->c_curthread->t_stack;
|
||||
/* since stacks grow down, get the top */
|
||||
stackpointer += STACK_SIZE;
|
||||
/* stack base address */
|
||||
stackpointer = (vaddr_t)c->c_curthread->t_stack;
|
||||
/* since stacks grow down, get the top */
|
||||
stackpointer += STACK_SIZE;
|
||||
|
||||
cpustacks[c->c_number] = stackpointer;
|
||||
cputhreads[c->c_number] = (vaddr_t)c->c_curthread;
|
||||
}
|
||||
cpustacks[c->c_number] = stackpointer;
|
||||
cputhreads[c->c_number] = (vaddr_t)c->c_curthread;
|
||||
}
|
||||
}
|
||||
|
||||
////////////////////////////////////////////////////////////
|
||||
@@ -107,73 +104,61 @@ cpu_machdep_init(struct cpu *c)
|
||||
* System/161 processor-ID values.
|
||||
*/
|
||||
|
||||
#define SYS161_PRID_ORIG 0x000003ff
|
||||
#define SYS161_PRID_2X 0x000000a1
|
||||
#define SYS161_PRID_ORIG 0x000003ff
|
||||
#define SYS161_PRID_2X 0x000000a1
|
||||
|
||||
static inline
|
||||
uint32_t
|
||||
cpu_getprid(void)
|
||||
{
|
||||
uint32_t prid;
|
||||
static inline uint32_t cpu_getprid(void) {
|
||||
uint32_t prid;
|
||||
|
||||
__asm volatile("mfc0 %0,$15" : "=r" (prid));
|
||||
return prid;
|
||||
__asm volatile("mfc0 %0,$15" : "=r"(prid));
|
||||
return prid;
|
||||
}
|
||||
|
||||
static inline
|
||||
uint32_t
|
||||
cpu_getfeatures(void)
|
||||
{
|
||||
uint32_t features;
|
||||
static inline uint32_t cpu_getfeatures(void) {
|
||||
uint32_t features;
|
||||
|
||||
__asm volatile(".set push;" /* save assembler mode */
|
||||
".set mips32;" /* allow mips32 instructions */
|
||||
"mfc0 %0,$15,1;" /* get cop0 reg 15 sel 1 */
|
||||
".set pop" /* restore assembler mode */
|
||||
: "=r" (features));
|
||||
return features;
|
||||
__asm volatile(".set push;" /* save assembler mode */
|
||||
".set mips32;" /* allow mips32 instructions */
|
||||
"mfc0 %0,$15,1;" /* get cop0 reg 15 sel 1 */
|
||||
".set pop" /* restore assembler mode */
|
||||
: "=r"(features));
|
||||
return features;
|
||||
}
|
||||
|
||||
static inline
|
||||
uint32_t
|
||||
cpu_getifeatures(void)
|
||||
{
|
||||
uint32_t features;
|
||||
static inline uint32_t cpu_getifeatures(void) {
|
||||
uint32_t features;
|
||||
|
||||
__asm volatile(".set push;" /* save assembler mode */
|
||||
".set mips32;" /* allow mips32 instructions */
|
||||
"mfc0 %0,$15,2;" /* get cop0 reg 15 sel 2 */
|
||||
".set pop" /* restore assembler mode */
|
||||
: "=r" (features));
|
||||
return features;
|
||||
__asm volatile(".set push;" /* save assembler mode */
|
||||
".set mips32;" /* allow mips32 instructions */
|
||||
"mfc0 %0,$15,2;" /* get cop0 reg 15 sel 2 */
|
||||
".set pop" /* restore assembler mode */
|
||||
: "=r"(features));
|
||||
return features;
|
||||
}
|
||||
|
||||
void
|
||||
cpu_identify(char *buf, size_t max)
|
||||
{
|
||||
uint32_t prid;
|
||||
uint32_t features;
|
||||
void cpu_identify(char *buf, size_t max) {
|
||||
uint32_t prid;
|
||||
uint32_t features;
|
||||
|
||||
prid = cpu_getprid();
|
||||
switch (prid) {
|
||||
case SYS161_PRID_ORIG:
|
||||
snprintf(buf, max, "MIPS/161 (System/161 1.x and pre-2.x)");
|
||||
break;
|
||||
case SYS161_PRID_2X:
|
||||
features = cpu_getfeatures();
|
||||
snprintf(buf, max, "MIPS/161 (System/161 2.x) features 0x%x",
|
||||
features);
|
||||
features = cpu_getifeatures();
|
||||
if (features != 0) {
|
||||
kprintf("WARNING: unknown CPU incompatible features "
|
||||
"0x%x\n", features);
|
||||
}
|
||||
break;
|
||||
default:
|
||||
snprintf(buf, max, "32-bit MIPS (unknown type, CPU ID 0x%x)",
|
||||
prid);
|
||||
break;
|
||||
}
|
||||
prid = cpu_getprid();
|
||||
switch (prid) {
|
||||
case SYS161_PRID_ORIG:
|
||||
snprintf(buf, max, "MIPS/161 (System/161 1.x and pre-2.x)");
|
||||
break;
|
||||
case SYS161_PRID_2X:
|
||||
features = cpu_getfeatures();
|
||||
snprintf(buf, max, "MIPS/161 (System/161 2.x) features 0x%x", features);
|
||||
features = cpu_getifeatures();
|
||||
if (features != 0) {
|
||||
kprintf("WARNING: unknown CPU incompatible features "
|
||||
"0x%x\n",
|
||||
features);
|
||||
}
|
||||
break;
|
||||
default:
|
||||
snprintf(buf, max, "32-bit MIPS (unknown type, CPU ID 0x%x)", prid);
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
////////////////////////////////////////////////////////////
|
||||
@@ -200,50 +185,43 @@ cpu_identify(char *buf, size_t max)
|
||||
* These considerations do not (currently) apply to System/161,
|
||||
* however.
|
||||
*/
|
||||
#define GET_STATUS(x) __asm volatile("mfc0 %0,$12" : "=r" (x))
|
||||
#define SET_STATUS(x) __asm volatile("mtc0 %0,$12" :: "r" (x))
|
||||
#define GET_STATUS(x) __asm volatile("mfc0 %0,$12" : "=r"(x))
|
||||
#define SET_STATUS(x) __asm volatile("mtc0 %0,$12" ::"r"(x))
|
||||
|
||||
/*
|
||||
* Interrupts on.
|
||||
*/
|
||||
void
|
||||
cpu_irqon(void)
|
||||
{
|
||||
uint32_t x;
|
||||
void cpu_irqon(void) {
|
||||
uint32_t x;
|
||||
|
||||
GET_STATUS(x);
|
||||
x |= CST_IEc;
|
||||
SET_STATUS(x);
|
||||
GET_STATUS(x);
|
||||
x |= CST_IEc;
|
||||
SET_STATUS(x);
|
||||
}
|
||||
|
||||
/*
|
||||
* Interrupts off.
|
||||
*/
|
||||
void
|
||||
cpu_irqoff(void)
|
||||
{
|
||||
uint32_t x;
|
||||
void cpu_irqoff(void) {
|
||||
uint32_t x;
|
||||
|
||||
GET_STATUS(x);
|
||||
x &= ~(uint32_t)CST_IEc;
|
||||
SET_STATUS(x);
|
||||
GET_STATUS(x);
|
||||
x &= ~(uint32_t)CST_IEc;
|
||||
SET_STATUS(x);
|
||||
}
|
||||
|
||||
/*
|
||||
* Used below.
|
||||
*/
|
||||
static
|
||||
void
|
||||
cpu_irqonoff(void)
|
||||
{
|
||||
uint32_t x, xon, xoff;
|
||||
static void cpu_irqonoff(void) {
|
||||
uint32_t x, xon, xoff;
|
||||
|
||||
GET_STATUS(x);
|
||||
xon = x | CST_IEc;
|
||||
xoff = x & ~(uint32_t)CST_IEc;
|
||||
SET_STATUS(xon);
|
||||
__asm volatile("nop; nop; nop; nop");
|
||||
SET_STATUS(xoff);
|
||||
GET_STATUS(x);
|
||||
xon = x | CST_IEc;
|
||||
xoff = x & ~(uint32_t)CST_IEc;
|
||||
SET_STATUS(xon);
|
||||
__asm volatile("nop; nop; nop; nop");
|
||||
SET_STATUS(xoff);
|
||||
}
|
||||
|
||||
////////////////////////////////////////////////////////////
|
||||
@@ -261,49 +239,40 @@ cpu_irqonoff(void)
|
||||
* appropriate the mips32 WAIT instruction.
|
||||
*/
|
||||
|
||||
static
|
||||
inline
|
||||
void
|
||||
wait(void)
|
||||
{
|
||||
/*
|
||||
* The WAIT instruction goes into powersave mode until an
|
||||
* interrupt is trying to occur.
|
||||
*
|
||||
* Then switch interrupts on and off again, so we actually
|
||||
* take the interrupt.
|
||||
*
|
||||
* Note that the precise behavior of this instruction in the
|
||||
* System/161 simulator is partly guesswork. This code may not
|
||||
* work on a real mips.
|
||||
*/
|
||||
__asm volatile(
|
||||
".set push;" /* save assembler mode */
|
||||
".set mips32;" /* allow MIPS32 instructions */
|
||||
".set volatile;" /* avoid unwanted optimization */
|
||||
"wait;" /* suspend until interrupted */
|
||||
".set pop" /* restore assembler mode */
|
||||
);
|
||||
static inline void wait(void) {
|
||||
/*
|
||||
* The WAIT instruction goes into powersave mode until an
|
||||
* interrupt is trying to occur.
|
||||
*
|
||||
* Then switch interrupts on and off again, so we actually
|
||||
* take the interrupt.
|
||||
*
|
||||
* Note that the precise behavior of this instruction in the
|
||||
* System/161 simulator is partly guesswork. This code may not
|
||||
* work on a real mips.
|
||||
*/
|
||||
__asm volatile(".set push;" /* save assembler mode */
|
||||
".set mips32;" /* allow MIPS32 instructions */
|
||||
".set volatile;" /* avoid unwanted optimization */
|
||||
"wait;" /* suspend until interrupted */
|
||||
".set pop" /* restore assembler mode */
|
||||
);
|
||||
}
|
||||
|
||||
/*
|
||||
* Idle the processor until something happens.
|
||||
*/
|
||||
void
|
||||
cpu_idle(void)
|
||||
{
|
||||
wait();
|
||||
cpu_irqonoff();
|
||||
void cpu_idle(void) {
|
||||
wait();
|
||||
cpu_irqonoff();
|
||||
}
|
||||
|
||||
/*
|
||||
* Halt the CPU permanently.
|
||||
*/
|
||||
void
|
||||
cpu_halt(void)
|
||||
{
|
||||
cpu_irqoff();
|
||||
while (1) {
|
||||
wait();
|
||||
}
|
||||
void cpu_halt(void) {
|
||||
cpu_irqoff();
|
||||
while (1) {
|
||||
wait();
|
||||
}
|
||||
}
|
||||
|
||||
@@ -37,7 +37,6 @@
|
||||
/* in threadstart.S */
|
||||
extern void mips_threadstart(/* arguments are in unusual registers */);
|
||||
|
||||
|
||||
/*
|
||||
* Function to initialize the switchframe of a new thread, which is
|
||||
* *not* the one that is currently running.
|
||||
@@ -51,48 +50,46 @@ extern void mips_threadstart(/* arguments are in unusual registers */);
|
||||
* store the arguments in the s* registers, and use a bit of asm
|
||||
* (mips_threadstart) to move them and then jump to thread_startup.
|
||||
*/
|
||||
void
|
||||
switchframe_init(struct thread *thread,
|
||||
void (*entrypoint)(void *data1, unsigned long data2),
|
||||
void *data1, unsigned long data2)
|
||||
{
|
||||
vaddr_t stacktop;
|
||||
struct switchframe *sf;
|
||||
void switchframe_init(struct thread *thread,
|
||||
void (*entrypoint)(void *data1, unsigned long data2),
|
||||
void *data1, unsigned long data2) {
|
||||
vaddr_t stacktop;
|
||||
struct switchframe *sf;
|
||||
|
||||
/*
|
||||
* MIPS stacks grow down. t_stack is just a hunk of memory, so
|
||||
* get the other end of it. Then set up a switchframe on the
|
||||
* top of the stack.
|
||||
*/
|
||||
stacktop = ((vaddr_t)thread->t_stack) + STACK_SIZE;
|
||||
sf = ((struct switchframe *) stacktop) - 1;
|
||||
/*
|
||||
* MIPS stacks grow down. t_stack is just a hunk of memory, so
|
||||
* get the other end of it. Then set up a switchframe on the
|
||||
* top of the stack.
|
||||
*/
|
||||
stacktop = ((vaddr_t)thread->t_stack) + STACK_SIZE;
|
||||
sf = ((struct switchframe *)stacktop) - 1;
|
||||
|
||||
/* Zero out the switchframe. */
|
||||
bzero(sf, sizeof(*sf));
|
||||
/* Zero out the switchframe. */
|
||||
bzero(sf, sizeof(*sf));
|
||||
|
||||
/*
|
||||
* Now set the important parts: pass through the three arguments,
|
||||
* and set the return address register to the place we want
|
||||
* execution to begin.
|
||||
*
|
||||
* Thus, when switchframe_switch does its "j ra", it will
|
||||
* actually jump to mips_threadstart, which will move the
|
||||
* arguments into the right register and jump to
|
||||
* thread_startup().
|
||||
*
|
||||
* Note that this means that when we call switchframe_switch()
|
||||
* in thread_switch(), we may not come back out the same way
|
||||
* in the next thread. (Though we will come back out the same
|
||||
* way when we later come back to the same thread again.)
|
||||
*
|
||||
* This has implications for code at the bottom of
|
||||
* thread_switch, described in thread.c.
|
||||
*/
|
||||
sf->sf_s0 = (uint32_t)entrypoint;
|
||||
sf->sf_s1 = (uint32_t)data1;
|
||||
sf->sf_s2 = (uint32_t)data2;
|
||||
sf->sf_ra = (uint32_t)mips_threadstart;
|
||||
/*
|
||||
* Now set the important parts: pass through the three arguments,
|
||||
* and set the return address register to the place we want
|
||||
* execution to begin.
|
||||
*
|
||||
* Thus, when switchframe_switch does its "j ra", it will
|
||||
* actually jump to mips_threadstart, which will move the
|
||||
* arguments into the right register and jump to
|
||||
* thread_startup().
|
||||
*
|
||||
* Note that this means that when we call switchframe_switch()
|
||||
* in thread_switch(), we may not come back out the same way
|
||||
* in the next thread. (Though we will come back out the same
|
||||
* way when we later come back to the same thread again.)
|
||||
*
|
||||
* This has implications for code at the bottom of
|
||||
* thread_switch, described in thread.c.
|
||||
*/
|
||||
sf->sf_s0 = (uint32_t)entrypoint;
|
||||
sf->sf_s1 = (uint32_t)data1;
|
||||
sf->sf_s2 = (uint32_t)data2;
|
||||
sf->sf_ra = (uint32_t)mips_threadstart;
|
||||
|
||||
/* Set ->t_context, and we're done. */
|
||||
thread->t_context = sf;
|
||||
/* Set ->t_context, and we're done. */
|
||||
thread->t_context = sf;
|
||||
}
|
||||
|
||||
@@ -37,16 +37,16 @@
|
||||
*/
|
||||
|
||||
struct switchframe {
|
||||
uint32_t sf_s0;
|
||||
uint32_t sf_s1;
|
||||
uint32_t sf_s2;
|
||||
uint32_t sf_s3;
|
||||
uint32_t sf_s4;
|
||||
uint32_t sf_s5;
|
||||
uint32_t sf_s6;
|
||||
uint32_t sf_s8;
|
||||
uint32_t sf_gp;
|
||||
uint32_t sf_ra;
|
||||
uint32_t sf_s0;
|
||||
uint32_t sf_s1;
|
||||
uint32_t sf_s2;
|
||||
uint32_t sf_s3;
|
||||
uint32_t sf_s4;
|
||||
uint32_t sf_s5;
|
||||
uint32_t sf_s6;
|
||||
uint32_t sf_s8;
|
||||
uint32_t sf_gp;
|
||||
uint32_t sf_ra;
|
||||
};
|
||||
|
||||
#endif /* _MIPS_SWITCHFRAME_H_ */
|
||||
|
||||
@@ -36,14 +36,10 @@
|
||||
#include <thread.h>
|
||||
#include <threadprivate.h>
|
||||
|
||||
void
|
||||
thread_machdep_init(struct thread_machdep *tm)
|
||||
{
|
||||
tm->tm_badfaultfunc = NULL;
|
||||
void thread_machdep_init(struct thread_machdep *tm) {
|
||||
tm->tm_badfaultfunc = NULL;
|
||||
}
|
||||
|
||||
void
|
||||
thread_machdep_cleanup(struct thread_machdep *tm)
|
||||
{
|
||||
KASSERT(tm->tm_badfaultfunc == NULL);
|
||||
void thread_machdep_cleanup(struct thread_machdep *tm) {
|
||||
KASSERT(tm->tm_badfaultfunc == NULL);
|
||||
}
|
||||
|
||||
@@ -57,18 +57,14 @@
|
||||
|
||||
/* under dumbvm, always have 72k of user stack */
|
||||
/* (this must be > 64K so argument blocks of size ARG_MAX will fit) */
|
||||
#define DUMBVM_STACKPAGES 18
|
||||
#define DUMBVM_STACKPAGES 18
|
||||
|
||||
/*
|
||||
* Wrap ram_stealmem in a spinlock.
|
||||
*/
|
||||
static struct spinlock stealmem_lock = SPINLOCK_INITIALIZER;
|
||||
|
||||
void
|
||||
vm_bootstrap(void)
|
||||
{
|
||||
/* Do nothing. */
|
||||
}
|
||||
void vm_bootstrap(void) { /* Do nothing. */ }
|
||||
|
||||
/*
|
||||
* Check if we're in a context that can sleep. While most of the
|
||||
@@ -77,351 +73,311 @@ vm_bootstrap(void)
|
||||
* avoid the situation where syscall-layer code that works ok with
|
||||
* dumbvm starts blowing up during the VM assignment.
|
||||
*/
|
||||
static
|
||||
void
|
||||
dumbvm_can_sleep(void)
|
||||
{
|
||||
if (CURCPU_EXISTS()) {
|
||||
/* must not hold spinlocks */
|
||||
KASSERT(curcpu->c_spinlocks == 0);
|
||||
static void dumbvm_can_sleep(void) {
|
||||
if (CURCPU_EXISTS()) {
|
||||
/* must not hold spinlocks */
|
||||
KASSERT(curcpu->c_spinlocks == 0);
|
||||
|
||||
/* must not be in an interrupt handler */
|
||||
KASSERT(curthread->t_in_interrupt == 0);
|
||||
}
|
||||
/* must not be in an interrupt handler */
|
||||
KASSERT(curthread->t_in_interrupt == 0);
|
||||
}
|
||||
}
|
||||
|
||||
static
|
||||
paddr_t
|
||||
getppages(unsigned long npages)
|
||||
{
|
||||
paddr_t addr;
|
||||
static paddr_t getppages(unsigned long npages) {
|
||||
paddr_t addr;
|
||||
|
||||
spinlock_acquire(&stealmem_lock);
|
||||
spinlock_acquire(&stealmem_lock);
|
||||
|
||||
addr = ram_stealmem(npages);
|
||||
addr = ram_stealmem(npages);
|
||||
|
||||
spinlock_release(&stealmem_lock);
|
||||
return addr;
|
||||
spinlock_release(&stealmem_lock);
|
||||
return addr;
|
||||
}
|
||||
|
||||
/* Allocate/free some kernel-space virtual pages */
|
||||
vaddr_t
|
||||
alloc_kpages(unsigned npages)
|
||||
{
|
||||
paddr_t pa;
|
||||
vaddr_t alloc_kpages(unsigned npages) {
|
||||
paddr_t pa;
|
||||
|
||||
dumbvm_can_sleep();
|
||||
pa = getppages(npages);
|
||||
if (pa==0) {
|
||||
return 0;
|
||||
}
|
||||
return PADDR_TO_KVADDR(pa);
|
||||
dumbvm_can_sleep();
|
||||
pa = getppages(npages);
|
||||
if (pa == 0) {
|
||||
return 0;
|
||||
}
|
||||
return PADDR_TO_KVADDR(pa);
|
||||
}
|
||||
|
||||
void
|
||||
free_kpages(vaddr_t addr)
|
||||
{
|
||||
/* nothing - leak the memory. */
|
||||
void free_kpages(vaddr_t addr) {
|
||||
/* nothing - leak the memory. */
|
||||
|
||||
(void)addr;
|
||||
(void)addr;
|
||||
}
|
||||
|
||||
void
|
||||
vm_tlbshootdown(const struct tlbshootdown *ts)
|
||||
{
|
||||
(void)ts;
|
||||
panic("dumbvm tried to do tlb shootdown?!\n");
|
||||
void vm_tlbshootdown(const struct tlbshootdown *ts) {
|
||||
(void)ts;
|
||||
panic("dumbvm tried to do tlb shootdown?!\n");
|
||||
}
|
||||
|
||||
int
|
||||
vm_fault(int faulttype, vaddr_t faultaddress)
|
||||
{
|
||||
vaddr_t vbase1, vtop1, vbase2, vtop2, stackbase, stacktop;
|
||||
paddr_t paddr;
|
||||
int i;
|
||||
uint32_t ehi, elo;
|
||||
struct addrspace *as;
|
||||
int spl;
|
||||
int vm_fault(int faulttype, vaddr_t faultaddress) {
|
||||
vaddr_t vbase1, vtop1, vbase2, vtop2, stackbase, stacktop;
|
||||
paddr_t paddr;
|
||||
int i;
|
||||
uint32_t ehi, elo;
|
||||
struct addrspace *as;
|
||||
int spl;
|
||||
|
||||
faultaddress &= PAGE_FRAME;
|
||||
faultaddress &= PAGE_FRAME;
|
||||
|
||||
DEBUG(DB_VM, "dumbvm: fault: 0x%x\n", faultaddress);
|
||||
DEBUG(DB_VM, "dumbvm: fault: 0x%x\n", faultaddress);
|
||||
|
||||
switch (faulttype) {
|
||||
case VM_FAULT_READONLY:
|
||||
/* We always create pages read-write, so we can't get this */
|
||||
panic("dumbvm: got VM_FAULT_READONLY\n");
|
||||
case VM_FAULT_READ:
|
||||
case VM_FAULT_WRITE:
|
||||
break;
|
||||
default:
|
||||
return EINVAL;
|
||||
}
|
||||
switch (faulttype) {
|
||||
case VM_FAULT_READONLY:
|
||||
/* We always create pages read-write, so we can't get this */
|
||||
panic("dumbvm: got VM_FAULT_READONLY\n");
|
||||
case VM_FAULT_READ:
|
||||
case VM_FAULT_WRITE:
|
||||
break;
|
||||
default:
|
||||
return EINVAL;
|
||||
}
|
||||
|
||||
if (curproc == NULL) {
|
||||
/*
|
||||
* No process. This is probably a kernel fault early
|
||||
* in boot. Return EFAULT so as to panic instead of
|
||||
* getting into an infinite faulting loop.
|
||||
*/
|
||||
return EFAULT;
|
||||
}
|
||||
if (curproc == NULL) {
|
||||
/*
|
||||
* No process. This is probably a kernel fault early
|
||||
* in boot. Return EFAULT so as to panic instead of
|
||||
* getting into an infinite faulting loop.
|
||||
*/
|
||||
return EFAULT;
|
||||
}
|
||||
|
||||
as = proc_getas();
|
||||
if (as == NULL) {
|
||||
/*
|
||||
* No address space set up. This is probably also a
|
||||
* kernel fault early in boot.
|
||||
*/
|
||||
return EFAULT;
|
||||
}
|
||||
as = proc_getas();
|
||||
if (as == NULL) {
|
||||
/*
|
||||
* No address space set up. This is probably also a
|
||||
* kernel fault early in boot.
|
||||
*/
|
||||
return EFAULT;
|
||||
}
|
||||
|
||||
/* Assert that the address space has been set up properly. */
|
||||
KASSERT(as->as_vbase1 != 0);
|
||||
KASSERT(as->as_pbase1 != 0);
|
||||
KASSERT(as->as_npages1 != 0);
|
||||
KASSERT(as->as_vbase2 != 0);
|
||||
KASSERT(as->as_pbase2 != 0);
|
||||
KASSERT(as->as_npages2 != 0);
|
||||
KASSERT(as->as_stackpbase != 0);
|
||||
KASSERT((as->as_vbase1 & PAGE_FRAME) == as->as_vbase1);
|
||||
KASSERT((as->as_pbase1 & PAGE_FRAME) == as->as_pbase1);
|
||||
KASSERT((as->as_vbase2 & PAGE_FRAME) == as->as_vbase2);
|
||||
KASSERT((as->as_pbase2 & PAGE_FRAME) == as->as_pbase2);
|
||||
KASSERT((as->as_stackpbase & PAGE_FRAME) == as->as_stackpbase);
|
||||
/* Assert that the address space has been set up properly. */
|
||||
KASSERT(as->as_vbase1 != 0);
|
||||
KASSERT(as->as_pbase1 != 0);
|
||||
KASSERT(as->as_npages1 != 0);
|
||||
KASSERT(as->as_vbase2 != 0);
|
||||
KASSERT(as->as_pbase2 != 0);
|
||||
KASSERT(as->as_npages2 != 0);
|
||||
KASSERT(as->as_stackpbase != 0);
|
||||
KASSERT((as->as_vbase1 & PAGE_FRAME) == as->as_vbase1);
|
||||
KASSERT((as->as_pbase1 & PAGE_FRAME) == as->as_pbase1);
|
||||
KASSERT((as->as_vbase2 & PAGE_FRAME) == as->as_vbase2);
|
||||
KASSERT((as->as_pbase2 & PAGE_FRAME) == as->as_pbase2);
|
||||
KASSERT((as->as_stackpbase & PAGE_FRAME) == as->as_stackpbase);
|
||||
|
||||
vbase1 = as->as_vbase1;
|
||||
vtop1 = vbase1 + as->as_npages1 * PAGE_SIZE;
|
||||
vbase2 = as->as_vbase2;
|
||||
vtop2 = vbase2 + as->as_npages2 * PAGE_SIZE;
|
||||
stackbase = USERSTACK - DUMBVM_STACKPAGES * PAGE_SIZE;
|
||||
stacktop = USERSTACK;
|
||||
vbase1 = as->as_vbase1;
|
||||
vtop1 = vbase1 + as->as_npages1 * PAGE_SIZE;
|
||||
vbase2 = as->as_vbase2;
|
||||
vtop2 = vbase2 + as->as_npages2 * PAGE_SIZE;
|
||||
stackbase = USERSTACK - DUMBVM_STACKPAGES * PAGE_SIZE;
|
||||
stacktop = USERSTACK;
|
||||
|
||||
if (faultaddress >= vbase1 && faultaddress < vtop1) {
|
||||
paddr = (faultaddress - vbase1) + as->as_pbase1;
|
||||
}
|
||||
else if (faultaddress >= vbase2 && faultaddress < vtop2) {
|
||||
paddr = (faultaddress - vbase2) + as->as_pbase2;
|
||||
}
|
||||
else if (faultaddress >= stackbase && faultaddress < stacktop) {
|
||||
paddr = (faultaddress - stackbase) + as->as_stackpbase;
|
||||
}
|
||||
else {
|
||||
return EFAULT;
|
||||
}
|
||||
if (faultaddress >= vbase1 && faultaddress < vtop1) {
|
||||
paddr = (faultaddress - vbase1) + as->as_pbase1;
|
||||
} else if (faultaddress >= vbase2 && faultaddress < vtop2) {
|
||||
paddr = (faultaddress - vbase2) + as->as_pbase2;
|
||||
} else if (faultaddress >= stackbase && faultaddress < stacktop) {
|
||||
paddr = (faultaddress - stackbase) + as->as_stackpbase;
|
||||
} else {
|
||||
return EFAULT;
|
||||
}
|
||||
|
||||
/* make sure it's page-aligned */
|
||||
KASSERT((paddr & PAGE_FRAME) == paddr);
|
||||
/* make sure it's page-aligned */
|
||||
KASSERT((paddr & PAGE_FRAME) == paddr);
|
||||
|
||||
/* Disable interrupts on this CPU while frobbing the TLB. */
|
||||
spl = splhigh();
|
||||
/* Disable interrupts on this CPU while frobbing the TLB. */
|
||||
spl = splhigh();
|
||||
|
||||
for (i=0; i<NUM_TLB; i++) {
|
||||
tlb_read(&ehi, &elo, i);
|
||||
if (elo & TLBLO_VALID) {
|
||||
continue;
|
||||
}
|
||||
ehi = faultaddress;
|
||||
elo = paddr | TLBLO_DIRTY | TLBLO_VALID;
|
||||
DEBUG(DB_VM, "dumbvm: 0x%x -> 0x%x\n", faultaddress, paddr);
|
||||
tlb_write(ehi, elo, i);
|
||||
splx(spl);
|
||||
return 0;
|
||||
}
|
||||
for (i = 0; i < NUM_TLB; i++) {
|
||||
tlb_read(&ehi, &elo, i);
|
||||
if (elo & TLBLO_VALID) {
|
||||
continue;
|
||||
}
|
||||
ehi = faultaddress;
|
||||
elo = paddr | TLBLO_DIRTY | TLBLO_VALID;
|
||||
DEBUG(DB_VM, "dumbvm: 0x%x -> 0x%x\n", faultaddress, paddr);
|
||||
tlb_write(ehi, elo, i);
|
||||
splx(spl);
|
||||
return 0;
|
||||
}
|
||||
|
||||
kprintf("dumbvm: Ran out of TLB entries - cannot handle page fault\n");
|
||||
splx(spl);
|
||||
return EFAULT;
|
||||
kprintf("dumbvm: Ran out of TLB entries - cannot handle page fault\n");
|
||||
splx(spl);
|
||||
return EFAULT;
|
||||
}
|
||||
|
||||
struct addrspace *
|
||||
as_create(void)
|
||||
{
|
||||
struct addrspace *as = kmalloc(sizeof(struct addrspace));
|
||||
if (as==NULL) {
|
||||
return NULL;
|
||||
}
|
||||
struct addrspace *as_create(void) {
|
||||
struct addrspace *as = kmalloc(sizeof(struct addrspace));
|
||||
if (as == NULL) {
|
||||
return NULL;
|
||||
}
|
||||
|
||||
as->as_vbase1 = 0;
|
||||
as->as_pbase1 = 0;
|
||||
as->as_npages1 = 0;
|
||||
as->as_vbase2 = 0;
|
||||
as->as_pbase2 = 0;
|
||||
as->as_npages2 = 0;
|
||||
as->as_stackpbase = 0;
|
||||
as->as_vbase1 = 0;
|
||||
as->as_pbase1 = 0;
|
||||
as->as_npages1 = 0;
|
||||
as->as_vbase2 = 0;
|
||||
as->as_pbase2 = 0;
|
||||
as->as_npages2 = 0;
|
||||
as->as_stackpbase = 0;
|
||||
|
||||
return as;
|
||||
return as;
|
||||
}
|
||||
|
||||
void
|
||||
as_destroy(struct addrspace *as)
|
||||
{
|
||||
dumbvm_can_sleep();
|
||||
kfree(as);
|
||||
void as_destroy(struct addrspace *as) {
|
||||
dumbvm_can_sleep();
|
||||
kfree(as);
|
||||
}
|
||||
|
||||
void
|
||||
as_activate(void)
|
||||
{
|
||||
int i, spl;
|
||||
struct addrspace *as;
|
||||
void as_activate(void) {
|
||||
int i, spl;
|
||||
struct addrspace *as;
|
||||
|
||||
as = proc_getas();
|
||||
if (as == NULL) {
|
||||
return;
|
||||
}
|
||||
as = proc_getas();
|
||||
if (as == NULL) {
|
||||
return;
|
||||
}
|
||||
|
||||
/* Disable interrupts on this CPU while frobbing the TLB. */
|
||||
spl = splhigh();
|
||||
/* Disable interrupts on this CPU while frobbing the TLB. */
|
||||
spl = splhigh();
|
||||
|
||||
for (i=0; i<NUM_TLB; i++) {
|
||||
tlb_write(TLBHI_INVALID(i), TLBLO_INVALID(), i);
|
||||
}
|
||||
for (i = 0; i < NUM_TLB; i++) {
|
||||
tlb_write(TLBHI_INVALID(i), TLBLO_INVALID(), i);
|
||||
}
|
||||
|
||||
splx(spl);
|
||||
splx(spl);
|
||||
}
|
||||
|
||||
void
|
||||
as_deactivate(void)
|
||||
{
|
||||
/* nothing */
|
||||
void as_deactivate(void) { /* nothing */ }
|
||||
|
||||
int as_define_region(struct addrspace *as, vaddr_t vaddr, size_t sz,
|
||||
int readable, int writeable, int executable) {
|
||||
size_t npages;
|
||||
|
||||
dumbvm_can_sleep();
|
||||
|
||||
/* Align the region. First, the base... */
|
||||
sz += vaddr & ~(vaddr_t)PAGE_FRAME;
|
||||
vaddr &= PAGE_FRAME;
|
||||
|
||||
/* ...and now the length. */
|
||||
sz = (sz + PAGE_SIZE - 1) & PAGE_FRAME;
|
||||
|
||||
npages = sz / PAGE_SIZE;
|
||||
|
||||
/* We don't use these - all pages are read-write */
|
||||
(void)readable;
|
||||
(void)writeable;
|
||||
(void)executable;
|
||||
|
||||
if (as->as_vbase1 == 0) {
|
||||
as->as_vbase1 = vaddr;
|
||||
as->as_npages1 = npages;
|
||||
return 0;
|
||||
}
|
||||
|
||||
if (as->as_vbase2 == 0) {
|
||||
as->as_vbase2 = vaddr;
|
||||
as->as_npages2 = npages;
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* Support for more than two regions is not available.
|
||||
*/
|
||||
kprintf("dumbvm: Warning: too many regions\n");
|
||||
return ENOSYS;
|
||||
}
|
||||
|
||||
int
|
||||
as_define_region(struct addrspace *as, vaddr_t vaddr, size_t sz,
|
||||
int readable, int writeable, int executable)
|
||||
{
|
||||
size_t npages;
|
||||
|
||||
dumbvm_can_sleep();
|
||||
|
||||
/* Align the region. First, the base... */
|
||||
sz += vaddr & ~(vaddr_t)PAGE_FRAME;
|
||||
vaddr &= PAGE_FRAME;
|
||||
|
||||
/* ...and now the length. */
|
||||
sz = (sz + PAGE_SIZE - 1) & PAGE_FRAME;
|
||||
|
||||
npages = sz / PAGE_SIZE;
|
||||
|
||||
/* We don't use these - all pages are read-write */
|
||||
(void)readable;
|
||||
(void)writeable;
|
||||
(void)executable;
|
||||
|
||||
if (as->as_vbase1 == 0) {
|
||||
as->as_vbase1 = vaddr;
|
||||
as->as_npages1 = npages;
|
||||
return 0;
|
||||
}
|
||||
|
||||
if (as->as_vbase2 == 0) {
|
||||
as->as_vbase2 = vaddr;
|
||||
as->as_npages2 = npages;
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* Support for more than two regions is not available.
|
||||
*/
|
||||
kprintf("dumbvm: Warning: too many regions\n");
|
||||
return ENOSYS;
|
||||
static void as_zero_region(paddr_t paddr, unsigned npages) {
|
||||
bzero((void *)PADDR_TO_KVADDR(paddr), npages * PAGE_SIZE);
|
||||
}
|
||||
|
||||
static
|
||||
void
|
||||
as_zero_region(paddr_t paddr, unsigned npages)
|
||||
{
|
||||
bzero((void *)PADDR_TO_KVADDR(paddr), npages * PAGE_SIZE);
|
||||
int as_prepare_load(struct addrspace *as) {
|
||||
KASSERT(as->as_pbase1 == 0);
|
||||
KASSERT(as->as_pbase2 == 0);
|
||||
KASSERT(as->as_stackpbase == 0);
|
||||
|
||||
dumbvm_can_sleep();
|
||||
|
||||
as->as_pbase1 = getppages(as->as_npages1);
|
||||
if (as->as_pbase1 == 0) {
|
||||
return ENOMEM;
|
||||
}
|
||||
|
||||
as->as_pbase2 = getppages(as->as_npages2);
|
||||
if (as->as_pbase2 == 0) {
|
||||
return ENOMEM;
|
||||
}
|
||||
|
||||
as->as_stackpbase = getppages(DUMBVM_STACKPAGES);
|
||||
if (as->as_stackpbase == 0) {
|
||||
return ENOMEM;
|
||||
}
|
||||
|
||||
as_zero_region(as->as_pbase1, as->as_npages1);
|
||||
as_zero_region(as->as_pbase2, as->as_npages2);
|
||||
as_zero_region(as->as_stackpbase, DUMBVM_STACKPAGES);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
int
|
||||
as_prepare_load(struct addrspace *as)
|
||||
{
|
||||
KASSERT(as->as_pbase1 == 0);
|
||||
KASSERT(as->as_pbase2 == 0);
|
||||
KASSERT(as->as_stackpbase == 0);
|
||||
|
||||
dumbvm_can_sleep();
|
||||
|
||||
as->as_pbase1 = getppages(as->as_npages1);
|
||||
if (as->as_pbase1 == 0) {
|
||||
return ENOMEM;
|
||||
}
|
||||
|
||||
as->as_pbase2 = getppages(as->as_npages2);
|
||||
if (as->as_pbase2 == 0) {
|
||||
return ENOMEM;
|
||||
}
|
||||
|
||||
as->as_stackpbase = getppages(DUMBVM_STACKPAGES);
|
||||
if (as->as_stackpbase == 0) {
|
||||
return ENOMEM;
|
||||
}
|
||||
|
||||
as_zero_region(as->as_pbase1, as->as_npages1);
|
||||
as_zero_region(as->as_pbase2, as->as_npages2);
|
||||
as_zero_region(as->as_stackpbase, DUMBVM_STACKPAGES);
|
||||
|
||||
return 0;
|
||||
int as_complete_load(struct addrspace *as) {
|
||||
dumbvm_can_sleep();
|
||||
(void)as;
|
||||
return 0;
|
||||
}
|
||||
|
||||
int
|
||||
as_complete_load(struct addrspace *as)
|
||||
{
|
||||
dumbvm_can_sleep();
|
||||
(void)as;
|
||||
return 0;
|
||||
int as_define_stack(struct addrspace *as, vaddr_t *stackptr) {
|
||||
KASSERT(as->as_stackpbase != 0);
|
||||
|
||||
*stackptr = USERSTACK;
|
||||
return 0;
|
||||
}
|
||||
|
||||
int
|
||||
as_define_stack(struct addrspace *as, vaddr_t *stackptr)
|
||||
{
|
||||
KASSERT(as->as_stackpbase != 0);
|
||||
int as_copy(struct addrspace *old, struct addrspace **ret) {
|
||||
struct addrspace *new;
|
||||
|
||||
*stackptr = USERSTACK;
|
||||
return 0;
|
||||
}
|
||||
|
||||
int
|
||||
as_copy(struct addrspace *old, struct addrspace **ret)
|
||||
{
|
||||
struct addrspace *new;
|
||||
|
||||
dumbvm_can_sleep();
|
||||
|
||||
new = as_create();
|
||||
if (new==NULL) {
|
||||
return ENOMEM;
|
||||
}
|
||||
|
||||
new->as_vbase1 = old->as_vbase1;
|
||||
new->as_npages1 = old->as_npages1;
|
||||
new->as_vbase2 = old->as_vbase2;
|
||||
new->as_npages2 = old->as_npages2;
|
||||
|
||||
/* (Mis)use as_prepare_load to allocate some physical memory. */
|
||||
if (as_prepare_load(new)) {
|
||||
as_destroy(new);
|
||||
return ENOMEM;
|
||||
}
|
||||
|
||||
KASSERT(new->as_pbase1 != 0);
|
||||
KASSERT(new->as_pbase2 != 0);
|
||||
KASSERT(new->as_stackpbase != 0);
|
||||
|
||||
memmove((void *)PADDR_TO_KVADDR(new->as_pbase1),
|
||||
(const void *)PADDR_TO_KVADDR(old->as_pbase1),
|
||||
old->as_npages1*PAGE_SIZE);
|
||||
|
||||
memmove((void *)PADDR_TO_KVADDR(new->as_pbase2),
|
||||
(const void *)PADDR_TO_KVADDR(old->as_pbase2),
|
||||
old->as_npages2*PAGE_SIZE);
|
||||
|
||||
memmove((void *)PADDR_TO_KVADDR(new->as_stackpbase),
|
||||
(const void *)PADDR_TO_KVADDR(old->as_stackpbase),
|
||||
DUMBVM_STACKPAGES*PAGE_SIZE);
|
||||
|
||||
*ret = new;
|
||||
return 0;
|
||||
dumbvm_can_sleep();
|
||||
|
||||
new = as_create();
|
||||
if (new == NULL) {
|
||||
return ENOMEM;
|
||||
}
|
||||
|
||||
new->as_vbase1 = old->as_vbase1;
|
||||
new->as_npages1 = old->as_npages1;
|
||||
new->as_vbase2 = old->as_vbase2;
|
||||
new->as_npages2 = old->as_npages2;
|
||||
|
||||
/* (Mis)use as_prepare_load to allocate some physical memory. */
|
||||
if (as_prepare_load(new)) {
|
||||
as_destroy(new);
|
||||
return ENOMEM;
|
||||
}
|
||||
|
||||
KASSERT(new->as_pbase1 != 0);
|
||||
KASSERT(new->as_pbase2 != 0);
|
||||
KASSERT(new->as_stackpbase != 0);
|
||||
|
||||
memmove((void *)PADDR_TO_KVADDR(new->as_pbase1),
|
||||
(const void *)PADDR_TO_KVADDR(old->as_pbase1),
|
||||
old->as_npages1 * PAGE_SIZE);
|
||||
|
||||
memmove((void *)PADDR_TO_KVADDR(new->as_pbase2),
|
||||
(const void *)PADDR_TO_KVADDR(old->as_pbase2),
|
||||
old->as_npages2 * PAGE_SIZE);
|
||||
|
||||
memmove((void *)PADDR_TO_KVADDR(new->as_stackpbase),
|
||||
(const void *)PADDR_TO_KVADDR(old->as_stackpbase),
|
||||
DUMBVM_STACKPAGES * PAGE_SIZE);
|
||||
|
||||
*ret = new;
|
||||
return 0;
|
||||
}
|
||||
|
||||
@@ -32,45 +32,41 @@
|
||||
#include <vm.h>
|
||||
#include <mainbus.h>
|
||||
|
||||
vaddr_t firstfree; /* first free virtual address; set by start.S */
|
||||
|
||||
vaddr_t firstfree; /* first free virtual address; set by start.S */
|
||||
|
||||
static paddr_t firstpaddr; /* address of first free physical page */
|
||||
static paddr_t lastpaddr; /* one past end of last free physical page */
|
||||
static paddr_t firstpaddr; /* address of first free physical page */
|
||||
static paddr_t lastpaddr; /* one past end of last free physical page */
|
||||
|
||||
/*
|
||||
* Called very early in system boot to figure out how much physical
|
||||
* RAM is available.
|
||||
*/
|
||||
void
|
||||
ram_bootstrap(void)
|
||||
{
|
||||
size_t ramsize;
|
||||
void ram_bootstrap(void) {
|
||||
size_t ramsize;
|
||||
|
||||
/* Get size of RAM. */
|
||||
ramsize = mainbus_ramsize();
|
||||
/* Get size of RAM. */
|
||||
ramsize = mainbus_ramsize();
|
||||
|
||||
/*
|
||||
* This is the same as the last physical address, as long as
|
||||
* we have less than 512 megabytes of memory. If we had more,
|
||||
* we wouldn't be able to access it all through kseg0 and
|
||||
* everything would get a lot more complicated. This is not a
|
||||
* case we are going to worry about.
|
||||
*/
|
||||
if (ramsize > 512*1024*1024) {
|
||||
ramsize = 512*1024*1024;
|
||||
}
|
||||
/*
|
||||
* This is the same as the last physical address, as long as
|
||||
* we have less than 512 megabytes of memory. If we had more,
|
||||
* we wouldn't be able to access it all through kseg0 and
|
||||
* everything would get a lot more complicated. This is not a
|
||||
* case we are going to worry about.
|
||||
*/
|
||||
if (ramsize > 512 * 1024 * 1024) {
|
||||
ramsize = 512 * 1024 * 1024;
|
||||
}
|
||||
|
||||
lastpaddr = ramsize;
|
||||
lastpaddr = ramsize;
|
||||
|
||||
/*
|
||||
* Get first free virtual address from where start.S saved it.
|
||||
* Convert to physical address.
|
||||
*/
|
||||
firstpaddr = firstfree - MIPS_KSEG0;
|
||||
/*
|
||||
* Get first free virtual address from where start.S saved it.
|
||||
* Convert to physical address.
|
||||
*/
|
||||
firstpaddr = firstfree - MIPS_KSEG0;
|
||||
|
||||
kprintf("%uk physical memory available\n",
|
||||
(lastpaddr-firstpaddr)/1024);
|
||||
kprintf("%uk physical memory available\n", (lastpaddr - firstpaddr) / 1024);
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -91,22 +87,20 @@ ram_bootstrap(void)
|
||||
* This function should not be called once the VM system is initialized,
|
||||
* so it is not synchronized.
|
||||
*/
|
||||
paddr_t
|
||||
ram_stealmem(unsigned long npages)
|
||||
{
|
||||
size_t size;
|
||||
paddr_t paddr;
|
||||
paddr_t ram_stealmem(unsigned long npages) {
|
||||
size_t size;
|
||||
paddr_t paddr;
|
||||
|
||||
size = npages * PAGE_SIZE;
|
||||
size = npages * PAGE_SIZE;
|
||||
|
||||
if (firstpaddr + size > lastpaddr) {
|
||||
return 0;
|
||||
}
|
||||
if (firstpaddr + size > lastpaddr) {
|
||||
return 0;
|
||||
}
|
||||
|
||||
paddr = firstpaddr;
|
||||
firstpaddr += size;
|
||||
paddr = firstpaddr;
|
||||
firstpaddr += size;
|
||||
|
||||
return paddr;
|
||||
return paddr;
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -124,11 +118,7 @@ ram_stealmem(unsigned long npages)
|
||||
* initialize the VM system, after which the VM system should take
|
||||
* charge of knowing what memory exists.
|
||||
*/
|
||||
paddr_t
|
||||
ram_getsize(void)
|
||||
{
|
||||
return lastpaddr;
|
||||
}
|
||||
paddr_t ram_getsize(void) { return lastpaddr; }
|
||||
|
||||
/*
|
||||
* This function is intended to be called by the VM system when it
|
||||
@@ -142,12 +132,10 @@ ram_getsize(void)
|
||||
* This function should not be called once the VM system is initialized,
|
||||
* so it is not synchronized.
|
||||
*/
|
||||
paddr_t
|
||||
ram_getfirstfree(void)
|
||||
{
|
||||
paddr_t ret;
|
||||
paddr_t ram_getfirstfree(void) {
|
||||
paddr_t ret;
|
||||
|
||||
ret = firstpaddr;
|
||||
firstpaddr = lastpaddr = 0;
|
||||
return ret;
|
||||
ret = firstpaddr;
|
||||
firstpaddr = lastpaddr = 0;
|
||||
return ret;
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user