* Copyright (c) 1992 Terrence R. Lambert.
* Copyright (c) 1982, 1987, 1990 The Regents of the University of California.
* This code is derived from software contributed to Berkeley by
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. All advertising materials mentioning features or use of this software
* must display the following acknowledgement:
* This product includes software developed by the University of
* California, Berkeley and its contributors.
* 4. Neither the name of the University nor the names of its contributors
* may be used to endorse or promote products derived from this software
* without specific prior written permission.
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* from: @(#)machdep.c 7.4 (Berkeley) 6/3/91
* $Id: machdep.c,v 1.36.2.4 1994/04/18 03:03:34 rgrimes Exp $
#include "exec.h" /* for PS_STRINGS */
extern vm_offset_t avail_start
, avail_end
;
#include "machine/specialreg.h"
#include "machine/sysarch.h"
#include "machine/cons.h"
#include "i386/isa/isa.h"
#include "i386/isa/rtc.h"
static void identifycpu(void);
static void initcpu(void);
static int test_page(int *, int);
extern int grow(struct proc
*,int);
#ifndef PANIC_REBOOT_WAIT_TIME
#define PANIC_REBOOT_WAIT_TIME 15 /* default to 15 seconds */
* Declare these as initialized data so we can patch them.
int _udatasel
, _ucodesel
;
* Machine-dependent startup code
int boothowto
= 0, Maxmem
= 0, maxmem
= 0, badpages
= 0, physmem
= 0;
#if defined(GENERICAH) || defined(GENERICBT)
vm_offset_t phys_avail
[6];
void dumpsys
__P((void));
register struct pte
*pte
;
int maxbufs
, base
, residual
;
vm_offset_t minaddr
, maxaddr
;
* Initialize error message buffer (at end of core).
/* avail_end was pre-decremented in init_386() to compensate */
for (i
= 0; i
< btoc(sizeof (struct msgbuf
)); i
++)
pmap_enter(pmap_kernel(), (vm_offset_t
)msgbufp
,
* Good {morning,afternoon,evening,night}.
printf("real memory = %d (%d pages)\n", ptoa(physmem
), physmem
);
printf("bad memory = %d (%d pages)\n", ptoa(badpages
), badpages
);
* Allocate space for system data structures.
* The first available kernel virtual address is in "v".
* As pages of kernel virtual memory are allocated, "v" is incremented.
* As pages of memory are allocated and cleared,
* "firstaddr" is incremented.
* An index into the kernel page table corresponding to the
* virtual memory address maintained in "v" is kept in "mapaddr".
* Make two passes. The first pass calculates how much memory is
* needed and allocates it. The second pass assigns virtual
* addresses to the various data structures.
#define valloc(name, type, num) \
(name) = (type *)v; v = (caddr_t)((name)+(num))
#define valloclim(name, type, num, lim) \
(name) = (type *)v; v = (caddr_t)((lim) = ((name)+(num)))
valloc(callout
, struct callout
, ncallout
);
valloc(shmsegs
, struct shmid_ds
, shminfo
.shmmni
);
valloc(sema
, struct semid_ds
, seminfo
.semmni
);
valloc(sem
, struct sem
, seminfo
.semmns
);
/* This is pretty disgusting! */
valloc(semu
, int, (seminfo
.semmnu
* seminfo
.semusz
) / sizeof(int));
valloc(msgpool
, char, msginfo
.msgmax
);
valloc(msgmaps
, struct msgmap
, msginfo
.msgseg
);
valloc(msghdrs
, struct msg
, msginfo
.msgtql
);
valloc(msqids
, struct msqid_ds
, msginfo
.msgmni
);
* Determine how many buffers to allocate.
* Use 20% of memory of memory beyond the first 2MB
* Insure a minimum of 16 fs buffers.
* We allocate 1/2 as many swap buffer headers as file i/o buffers.
bufpages
= ((physmem
<< PGSHIFT
) - 2048*1024) / NBPG
/ 5;
* We must still limit the maximum number of buffers to be no
* more than 2/5's of the size of the kernal malloc region, this
* will only take effect for machines with lots of memory
bufpages
= min(bufpages
, (VM_KMEM_SIZE
/ NBPG
) * 2 / 5);
freebufspace
= bufpages
* NBPG
;
nswbuf
= (nbuf
/ 2) &~ 1; /* force even */
nswbuf
= 256; /* sanity */
valloc(swbuf
, struct buf
, nswbuf
);
valloc(buf
, struct buf
, nbuf
);
* End of first pass, size has been calculated so allocate memory
size
= (vm_size_t
)(v
- firstaddr
);
firstaddr
= (int)kmem_alloc(kernel_map
, round_page(size
));
panic("startup: no room for tables");
* End of second pass, addresses have been assigned
if ((vm_size_t
)(v
- firstaddr
) != size
)
panic("startup: table size inconsistency");
* Allocate a submap for buffer space allocations.
* XXX we are NOT using buffer_map, but due to
* the references to it we will just allocate 1 page of
* vm (not real memory) to make things happy...
buffer_map
= kmem_suballoc(kernel_map
, &minaddr
, &maxaddr
,
/* bufpages * */NBPG
, TRUE
);
* Allocate a submap for physio
phys_map
= kmem_suballoc(kernel_map
, &minaddr
, &maxaddr
,
* Finally, allocate mbuf pool. Since mclrefcnt is an off-size
* we use the more space efficient malloc in place of kmem_alloc.
mclrefcnt
= (char *)malloc(NMBCLUSTERS
+CLBYTES
/MCLBYTES
,
bzero(mclrefcnt
, NMBCLUSTERS
+CLBYTES
/MCLBYTES
);
mb_map
= kmem_suballoc(kmem_map
, (vm_offset_t
)&mbutl
, &maxaddr
,
for (i
= 1; i
< ncallout
; i
++)
callout
[i
-1].c_next
= &callout
[i
];
printf("avail memory = %d (%d pages)\n", ptoa(vm_page_free_count
), vm_page_free_count
);
printf("using %d buffers containing %d bytes of memory\n",
nbuf
, bufpages
* CLBYTES
);
* Set up CPU-specific registers, cache, etc.
* Set up buffers, so they can be used to read disk labels.
struct cpu_nameclass i386_cpus
[] = {
{ "Intel 80286", CPUCLASS_286
}, /* CPU_286 */
{ "i386SX", CPUCLASS_386
}, /* CPU_386SX */
{ "i386DX", CPUCLASS_386
}, /* CPU_386 */
{ "i486SX", CPUCLASS_486
}, /* CPU_486SX */
{ "i486DX", CPUCLASS_486
}, /* CPU_486 */
{ "i586", CPUCLASS_586
}, /* CPU_586 */
if (cpu
>= 0 && cpu
< (sizeof i386_cpus
/sizeof(struct cpu_nameclass
))) {
printf("%s", i386_cpus
[cpu
].cpu_name
);
cpu_class
= i386_cpus
[cpu
].cpu_class
;
printf("unknown cpu type %d\n", cpu
);
panic("startup: bad cpu id");
printf("unknown"); /* will panic below... */
printf("\n"); /* cpu speed would be nice, but how? */
* Now that we have told the user what they have,
* let them know if that machine type isn't configured.
case CPUCLASS_286
: /* a 286 should not make it this far, anyway */
#if !defined(I386_CPU) && !defined(I486_CPU) && !defined(I586_CPU)
#error This kernel is not configured for one of the supported CPUs
panic("CPU class not configured");
* Return the difference (in microseconds)
* between the current time and a previous
* time as represented by the arguments.
* If there is a pending clock interrupt
* which has not been serviced due to high
* ipl, return error code.
vmtime(otime
, olbolt
, oicr
)
register int otime
, olbolt
, oicr
;
return (((time
.tv_sec
-otime
)*60 + lbolt
-olbolt
)*16667);
* Send an interrupt to process.
* Stack is set up to allow sigcode stored
* in u. to call routine, followed by kcall
* to sigreturn routine below. After sigreturn
* resets the signal mask, the stack, and the
* frame pointer, it returns to the user
sendsig(catcher
, sig
, mask
, code
)
register struct proc
*p
= curproc
;
register struct sigframe
*fp
;
struct sigacts
*ps
= p
->p_sigacts
;
oonstack
= ps
->ps_onstack
;
* Allocate and validate space for the signal handler
* context. Note that if the stack is in P0 space, the
* call to grow() is a nop, and the useracc() check
* will fail if the process has not already allocated
* the space with a `brk'.
if (!ps
->ps_onstack
&& (ps
->ps_sigonstack
& sigmask(sig
))) {
fp
= (struct sigframe
*)(ps
->ps_sigsp
- sizeof(struct sigframe
));
fp
= (struct sigframe
*)(regs
[tESP
]
- sizeof(struct sigframe
));
* grow() will return FALSE if the fp will not fit inside the stack
* and the stack can not be grown. useracc will return FALSE
if ((grow(p
, (int)fp
) == FALSE
) ||
(useracc((caddr_t
)fp
, sizeof (struct sigframe
), B_WRITE
) == FALSE
)) {
* Process has trashed its stack; give it an illegal
* instruction to halt it in its tracks.
SIGACTION(p
, SIGILL
) = SIG_DFL
;
* Build the argument list for the signal handler.
fp
->sf_addr
= (char *) regs
[tERR
];
fp
->sf_handler
= catcher
;
/* save scratch registers */
fp
->sf_sc
.sc_eax
= regs
[tEAX
];
fp
->sf_sc
.sc_ebx
= regs
[tEBX
];
fp
->sf_sc
.sc_ecx
= regs
[tECX
];
fp
->sf_sc
.sc_edx
= regs
[tEDX
];
fp
->sf_sc
.sc_esi
= regs
[tESI
];
fp
->sf_sc
.sc_edi
= regs
[tEDI
];
fp
->sf_sc
.sc_cs
= regs
[tCS
];
fp
->sf_sc
.sc_ds
= regs
[tDS
];
fp
->sf_sc
.sc_ss
= regs
[tSS
];
fp
->sf_sc
.sc_es
= regs
[tES
];
fp
->sf_sc
.sc_isp
= regs
[tISP
];
* Build the signal context to be used by sigreturn.
fp
->sf_sc
.sc_onstack
= oonstack
;
fp
->sf_sc
.sc_mask
= mask
;
fp
->sf_sc
.sc_sp
= regs
[tESP
];
fp
->sf_sc
.sc_fp
= regs
[tEBP
];
fp
->sf_sc
.sc_pc
= regs
[tEIP
];
fp
->sf_sc
.sc_ps
= regs
[tEFLAGS
];
regs
[tEIP
] = (int)((struct pcb
*)kstack
)->pcb_sigc
;
regs
[tEFLAGS
] &= ~PSL_VM
;
* System call to cleanup state after a signal
* has been taken. Reset signal mask and
* stack state from context left by sendsig (above).
* Return to previous pc and psl as specified by
* context left by sendsig. Check carefully to
* make sure that the user has not modified the
* psl to gain improper priviledges or to cause
struct sigcontext
*sigcntxp
;
sigreturn(p
, uap
, retval
)
struct sigreturn_args
*uap
;
register struct sigcontext
*scp
;
register struct sigframe
*fp
;
register int *regs
= p
->p_regs
;
* (XXX old comment) regs[tESP] points to the return address.
* The user scp pointer is above that.
* The return address is faked in the signal trampoline code
((caddr_t
)scp
- offsetof(struct sigframe
, sf_sc
));
if (useracc((caddr_t
)fp
, sizeof (*fp
), 0) == 0)
if ((eflags
& PSL_USERCLR
) != 0 ||
(eflags
& PSL_USERSET
) != PSL_USERSET
||
(eflags
& PSL_IOPL
) < (regs
[tEFLAGS
] & PSL_IOPL
)) {
printf("sigreturn: eflags=0x%x\n", eflags
);
* Sanity check the user's selectors and error if they
#define max_ldt_sel(pcb) \
((pcb)->pcb_ldt ? (pcb)->pcb_ldt_len : (sizeof(ldt) / sizeof(ldt[0])))
#define valid_ldt_sel(sel) \
(ISLDT(sel) && ISPL(sel) == SEL_UPL && \
IDXSEL(sel) < max_ldt_sel(&p->p_addr->u_pcb))
(!ISLDT(sel) && IDXSEL(sel) == 0)
if ((scp
->sc_cs
&0xffff != _ucodesel
&& !valid_ldt_sel(scp
->sc_cs
)) ||
(scp
->sc_ss
&0xffff != _udatasel
&& !valid_ldt_sel(scp
->sc_ss
)) ||
(scp
->sc_ds
&0xffff != _udatasel
&& !valid_ldt_sel(scp
->sc_ds
) &&
!null_sel(scp
->sc_ds
)) ||
(scp
->sc_es
&0xffff != _udatasel
&& !valid_ldt_sel(scp
->sc_es
) &&
!null_sel(scp
->sc_es
))) {
printf("sigreturn: cs=0x%x ss=0x%x ds=0x%x es=0x%x\n",
scp
->sc_cs
, scp
->sc_ss
, scp
->sc_ds
, scp
->sc_es
);
trapsignal(p
, SIGBUS
, T_PROTFLT
);
/* restore scratch registers */
regs
[tEAX
] = scp
->sc_eax
;
regs
[tEBX
] = scp
->sc_ebx
;
regs
[tECX
] = scp
->sc_ecx
;
regs
[tEDX
] = scp
->sc_edx
;
regs
[tESI
] = scp
->sc_esi
;
regs
[tEDI
] = scp
->sc_edi
;
regs
[tISP
] = scp
->sc_isp
;
if (useracc((caddr_t
)scp
, sizeof (*scp
), 0) == 0)
p
->p_sigacts
->ps_onstack
= scp
->sc_onstack
& 01;
p
->p_sigmask
= scp
->sc_mask
&~
(sigmask(SIGKILL
)|sigmask(SIGCONT
)|sigmask(SIGSTOP
));
* a simple function to make the system panic (and dump a vmcore)
* in a predictable fashion
panic("because you said to!");
register long dummy
; /* r12 is reserved */
register int howto
; /* r11 == how to boot */
register int devtype
; /* r10 == major of root dev */
printf("hit reset please");
if ((howto
&RB_NOSYNC
) == 0 && waittime
< 0 && bfreelist
[0].b_forw
) {
printf("syncing disks... ");
* Release inodes held by texts before update.
vnode_pager_umount(NULL
);
sync(curproc
, NULL
, NULL
);
for (iter
= 0; iter
< 20; iter
++) {
for (bp
= &buf
[nbuf
]; --bp
>= buf
; )
if ((bp
->b_flags
& (B_BUSY
|B_INVAL
)) == B_BUSY
)
printf("updating disks before rebooting... ");
DELAY(10000); /* wait for printf to finish */
devtype
= major(rootdev
);
printf("The operating system has halted.\n");
printf("Please press any key to reboot.\n\n");
dumppcb
.pcb_ptd
= rcr3();
if (PANIC_REBOOT_WAIT_TIME
!= 0) {
if (PANIC_REBOOT_WAIT_TIME
!= -1) {
printf("Automatic reboot in %d seconds - press a key on the console to abort\n",
for (loop
= PANIC_REBOOT_WAIT_TIME
; loop
> 0; --loop
) {
DELAY(1000 * 1000); /* one second */
if (sgetc(1)) /* Did user type a key? */
} else { /* zero time specified - reboot NOW */
printf("--> Press a key on the console to reboot <--\n");
dummy
= 0; dummy
= dummy
;
printf("howto %d, devtype %d\n", arghowto
, devtype
);
printf("Rebooting...\n");
DELAY (100000); /* wait 100ms for printf's to complete */
unsigned long dumpmag
= 0x8fca0101UL
; /* magic number for savecore */
int dumpsize
= 0; /* also for savecore */
* Doadump comes here after turning off memory management and
* getting on the dump stack, either when called above, or by
if ((minor(dumpdev
)&07) != 1)
printf("\ndumping to dev %x, offset %d\n", dumpdev
, dumplo
);
switch ((*bdevsw
[major(dumpdev
)].d_dump
)(dumpdev
)) {
printf("device not ready\n");
printf("area improper\n");
printf("aborted from console\n");
* If HZ is defined we use this code, otherwise the code in
* /sys/i386/i386/microtime.s is used. The othercode only works
register struct timeval
*tvp
;
while (tvp
->tv_usec
> 1000000) {
physstrat(bp
, strat
, prio
)
bp
->b_iodone
= physstratdone
;
/* pageout daemon doesn't wait for pushed pages */
if (bp
->b_flags
& B_DIRTY
)
while ((bp
->b_flags
& B_DONE
) == 0)
tsleep((caddr_t
)bp
, prio
, "physstr", 0);
* Clear registers on exec
p
->p_regs
[tEBP
] = 0; /* bottom of the fp chain */
p
->p_regs
[tSS
] = _udatasel
;
p
->p_regs
[tDS
] = _udatasel
;
p
->p_regs
[tES
] = _udatasel
;
p
->p_regs
[tCS
] = _ucodesel
;
p
->p_addr
->u_pcb
.pcb_flags
= 0; /* no fp at all */
load_cr0(rcr0() | CR0_TS
); /* start emulating */
npxinit(__INITIAL_NPXCW__
);
* Initialize 386 and configure to run kernel
* Initialize segments & interrupt table
union descriptor gdt
[NGDT
];
union descriptor ldt
[NLDT
]; /* local descriptor table */
struct gate_descriptor idt
[NIDT
]; /* interrupt descriptor table */
int _default_ldt
, currentldt
;
struct i386tss tss
, panic_tss
;
extern struct user
*proc0paddr
;
/* software prototypes -- in more palatable form */
struct soft_segment_descriptor gdt_segs
[] = {
{ 0x0, /* segment base address */
0, /* segment descriptor priority level */
0, /* segment descriptor present */
0, /* default 32 vs 16 bit size */
0 /* limit granularity (byte/page units)*/ },
/* Code Descriptor for kernel */
{ 0x0, /* segment base address */
0xfffff, /* length - all address space */
SDT_MEMERA
, /* segment type */
0, /* segment descriptor priority level */
1, /* segment descriptor present */
1, /* default 32 vs 16 bit size */
1 /* limit granularity (byte/page units)*/ },
/* Data Descriptor for kernel */
{ 0x0, /* segment base address */
0xfffff, /* length - all address space */
SDT_MEMRWA
, /* segment type */
0, /* segment descriptor priority level */
1, /* segment descriptor present */
1, /* default 32 vs 16 bit size */
1 /* limit granularity (byte/page units)*/ },
{ (int) ldt
, /* segment base address */
sizeof(ldt
)-1, /* length - all address space */
SDT_SYSLDT
, /* segment type */
0, /* segment descriptor priority level */
1, /* segment descriptor present */
0, /* unused - default 32 vs 16 bit size */
0 /* limit granularity (byte/page units)*/ },
/* Null Descriptor - Placeholder */
{ 0x0, /* segment base address */
0x0, /* length - all address space */
0, /* segment descriptor priority level */
0, /* segment descriptor present */
0, /* default 32 vs 16 bit size */
0 /* limit granularity (byte/page units)*/ },
/* Panic Tss Descriptor */
{ (int) &panic_tss
, /* segment base address */
sizeof(tss
)-1, /* length - all address space */
SDT_SYS386TSS
, /* segment type */
0, /* segment descriptor priority level */
1, /* segment descriptor present */
0, /* unused - default 32 vs 16 bit size */
0 /* limit granularity (byte/page units)*/ },
/* Proc 0 Tss Descriptor */
{ (int) kstack
, /* segment base address */
sizeof(tss
)-1, /* length - all address space */
SDT_SYS386TSS
, /* segment type */
0, /* segment descriptor priority level */
1, /* segment descriptor present */
0, /* unused - default 32 vs 16 bit size */
0 /* limit granularity (byte/page units)*/ },
/* User LDT Descriptor per process */
{ (int) ldt
, /* segment base address */
(512 * sizeof(union descriptor
)-1), /* length */
SDT_SYSLDT
, /* segment type */
0, /* segment descriptor priority level */
1, /* segment descriptor present */
0, /* unused - default 32 vs 16 bit size */
0 /* limit granularity (byte/page units)*/ },
struct soft_segment_descriptor ldt_segs
[] = {
/* Null Descriptor - overwritten by call gate */
{ 0x0, /* segment base address */
0x0, /* length - all address space */
0, /* segment descriptor priority level */
0, /* segment descriptor present */
0, /* default 32 vs 16 bit size */
0 /* limit granularity (byte/page units)*/ },
/* Null Descriptor - overwritten by call gate */
{ 0x0, /* segment base address */
0x0, /* length - all address space */
0, /* segment descriptor priority level */
0, /* segment descriptor present */
0, /* default 32 vs 16 bit size */
0 /* limit granularity (byte/page units)*/ },
/* Null Descriptor - overwritten by call gate */
{ 0x0, /* segment base address */
0x0, /* length - all address space */
0, /* segment descriptor priority level */
0, /* segment descriptor present */
0, /* default 32 vs 16 bit size */
0 /* limit granularity (byte/page units)*/ },
/* Code Descriptor for user */
{ 0x0, /* segment base address */
0xfffff, /* length - all address space */
SDT_MEMERA
, /* segment type */
SEL_UPL
, /* segment descriptor priority level */
1, /* segment descriptor present */
1, /* default 32 vs 16 bit size */
1 /* limit granularity (byte/page units)*/ },
/* Data Descriptor for user */
{ 0x0, /* segment base address */
0xfffff, /* length - all address space */
SDT_MEMRWA
, /* segment type */
SEL_UPL
, /* segment descriptor priority level */
1, /* segment descriptor present */
1, /* default 32 vs 16 bit size */
1 /* limit granularity (byte/page units)*/ } };
setidt(idx
, func
, typ
, dpl
)
struct gate_descriptor
*ip
= idt
+ idx
;
ip
->gd_looffset
= (int)func
;
ip
->gd_hioffset
= ((int)func
)>>16 ;
#define IDTVEC(name) __CONCAT(X, name)
IDTVEC(div
), IDTVEC(dbg
), IDTVEC(nmi
), IDTVEC(bpt
), IDTVEC(ofl
),
IDTVEC(bnd
), IDTVEC(ill
), IDTVEC(dna
), IDTVEC(dble
), IDTVEC(fpusegm
),
IDTVEC(tss
), IDTVEC(missing
), IDTVEC(stk
), IDTVEC(prot
),
IDTVEC(page
), IDTVEC(rsvd
), IDTVEC(fpu
), IDTVEC(rsvd0
),
IDTVEC(rsvd1
), IDTVEC(rsvd2
), IDTVEC(rsvd3
), IDTVEC(rsvd4
),
IDTVEC(rsvd5
), IDTVEC(rsvd6
), IDTVEC(rsvd7
), IDTVEC(rsvd8
),
IDTVEC(rsvd9
), IDTVEC(rsvd10
), IDTVEC(rsvd11
), IDTVEC(rsvd12
),
IDTVEC(rsvd13
), IDTVEC(rsvd14
), IDTVEC(rsvd14
), IDTVEC(syscall
);
extern ssdtosd(), lgdt(), lidt(), lldt(), etext
;
unsigned biosbasemem
, biosextmem
;
struct gate_descriptor
*gdp
;
extern int sigcode
,szsigcode
;
/* table descriptors - used to load tables by microp */
struct region_descriptor r_gdt
, r_idt
;
int pagesinbase
, pagesinext
;
proc0
.p_addr
= proc0paddr
;
* Initialize the console before we print anything out.
* make gdt memory segments, the code segment goes up to end of the
* page with etext in it, the data segment goes to the end of
gdt_segs
[GCODE_SEL
].ssd_limit
= i386_btop(i386_round_page(&etext
)) - 1;
gdt_segs
[GDATA_SEL
].ssd_limit
= 0xffffffffUL
; /* XXX constant? */
for (x
=0; x
< NGDT
; x
++) ssdtosd(gdt_segs
+x
, gdt
+x
);
/* make ldt memory segments */
* The data segment limit must not cover the user area because we
* don't want the user area to be writable in copyout() etc. (page
* level protection is lost in kernel mode on 386's). Also, we
* don't want the user area to be writable directly (page level
* protection of the user area is not available on 486's with
* CR0_WP set, because there is no user-read/kernel-write mode).
* XXX - VM_MAXUSER_ADDRESS is an end address, not a max. And it
* should be spelled ...MAX_USER...
#define VM_END_USER_RW_ADDRESS VM_MAXUSER_ADDRESS
* The code segment limit has to cover the user area until we move
* the signal trampoline out of the user area. This is safe because
* the code segment cannot be written to directly.
#define VM_END_USER_R_ADDRESS (VM_END_USER_RW_ADDRESS + UPAGES * NBPG)
ldt_segs
[LUCODE_SEL
].ssd_limit
= i386_btop(VM_END_USER_R_ADDRESS
) - 1;
ldt_segs
[LUDATA_SEL
].ssd_limit
= i386_btop(VM_END_USER_RW_ADDRESS
) - 1;
/* Note. eventually want private ldts per process */
for (x
=0; x
< 5; x
++) ssdtosd(ldt_segs
+x
, ldt
+x
);
setidt(0, &IDTVEC(div
), SDT_SYS386TGT
, SEL_KPL
);
setidt(1, &IDTVEC(dbg
), SDT_SYS386TGT
, SEL_KPL
);
setidt(2, &IDTVEC(nmi
), SDT_SYS386TGT
, SEL_KPL
);
setidt(3, &IDTVEC(bpt
), SDT_SYS386TGT
, SEL_UPL
);
setidt(4, &IDTVEC(ofl
), SDT_SYS386TGT
, SEL_UPL
);
setidt(5, &IDTVEC(bnd
), SDT_SYS386TGT
, SEL_KPL
);
setidt(6, &IDTVEC(ill
), SDT_SYS386TGT
, SEL_KPL
);
setidt(7, &IDTVEC(dna
), SDT_SYS386TGT
, SEL_KPL
);
setidt(8, &IDTVEC(dble
), SDT_SYS386TGT
, SEL_KPL
);
setidt(9, &IDTVEC(fpusegm
), SDT_SYS386TGT
, SEL_KPL
);
setidt(10, &IDTVEC(tss
), SDT_SYS386TGT
, SEL_KPL
);
setidt(11, &IDTVEC(missing
), SDT_SYS386TGT
, SEL_KPL
);
setidt(12, &IDTVEC(stk
), SDT_SYS386TGT
, SEL_KPL
);
setidt(13, &IDTVEC(prot
), SDT_SYS386TGT
, SEL_KPL
);
setidt(14, &IDTVEC(page
), SDT_SYS386TGT
, SEL_KPL
);
setidt(15, &IDTVEC(rsvd
), SDT_SYS386TGT
, SEL_KPL
);
setidt(16, &IDTVEC(fpu
), SDT_SYS386TGT
, SEL_KPL
);
setidt(17, &IDTVEC(rsvd0
), SDT_SYS386TGT
, SEL_KPL
);
setidt(18, &IDTVEC(rsvd1
), SDT_SYS386TGT
, SEL_KPL
);
setidt(19, &IDTVEC(rsvd2
), SDT_SYS386TGT
, SEL_KPL
);
setidt(20, &IDTVEC(rsvd3
), SDT_SYS386TGT
, SEL_KPL
);
setidt(21, &IDTVEC(rsvd4
), SDT_SYS386TGT
, SEL_KPL
);
setidt(22, &IDTVEC(rsvd5
), SDT_SYS386TGT
, SEL_KPL
);
setidt(23, &IDTVEC(rsvd6
), SDT_SYS386TGT
, SEL_KPL
);
setidt(24, &IDTVEC(rsvd7
), SDT_SYS386TGT
, SEL_KPL
);
setidt(25, &IDTVEC(rsvd8
), SDT_SYS386TGT
, SEL_KPL
);
setidt(26, &IDTVEC(rsvd9
), SDT_SYS386TGT
, SEL_KPL
);
setidt(27, &IDTVEC(rsvd10
), SDT_SYS386TGT
, SEL_KPL
);
setidt(28, &IDTVEC(rsvd11
), SDT_SYS386TGT
, SEL_KPL
);
setidt(29, &IDTVEC(rsvd12
), SDT_SYS386TGT
, SEL_KPL
);
setidt(30, &IDTVEC(rsvd13
), SDT_SYS386TGT
, SEL_KPL
);
setidt(31, &IDTVEC(rsvd14
), SDT_SYS386TGT
, SEL_KPL
);
r_gdt
.rd_limit
= sizeof(gdt
) - 1;
r_gdt
.rd_base
= (int) gdt
;
r_idt
.rd_limit
= sizeof(idt
) - 1;
r_idt
.rd_base
= (int) idt
;
_default_ldt
= GSEL(GLDT_SEL
, SEL_KPL
);
currentldt
= _default_ldt
;
Debugger("Boot flags requested debugger");
/* Use BIOS values stored in RTC CMOS RAM, since probing
* breaks certain 386 AT relics.
biosbasemem
= rtcin(RTC_BASELO
)+ (rtcin(RTC_BASEHI
)<<8);
biosextmem
= rtcin(RTC_EXTLO
)+ (rtcin(RTC_EXTHI
)<<8);
* If BIOS tells us that it has more than 640k in the basemem,
* don't believe it - set it to 640k.
* Some 386 machines might give us a bogus number for extended
* mem. If this happens, stop now.
if (biosextmem
> 65536) {
panic("extended memory beyond limit of 64MB");
pagesinbase
= biosbasemem
* 1024 / NBPG
;
pagesinext
= biosextmem
* 1024 / NBPG
;
* Special hack for chipsets that still remap the 384k hole when
* there's 16MB of memory - this really confuses people that
* are trying to use bus mastering ISA controllers with the
* "16MB limit"; they only have 16MB, but the remapping puts
* XXX - this should be removed when bounce buffers are
* If extended memory is between 15-16MB (16-17MB phys address range),
if ((pagesinext
> 3840) && (pagesinext
< 4096))
#if defined(GENERICAH) || defined(GENERICBT)
/* XXX This is an ugle hack so that machines with >16MB of memory
* can be booted using the GENERIC* kernels and not have to worry
* about bus mastered DMA on the ISA bus. It is ONLY compiled into
* the GENERIC* kernels and can be disabled by tweaking the global
* generic_hack to be zero using gdb.
printf("WARNING WARNING WARNING WARNING WARNING WARNING\n");
printf("GENERIC* kernels only USE the first 16MB of your ");
printf("%dMB.\n", (pagesinext
+ 256) / 256);
printf("Read the RELNOTES.FreeBSD file for the reason.\n");
printf("WARNING WARNING WARNING WARNING WARNING WARNING\n");
#endif /* defined (GENERICAH) || defiend(GENERICBT) */
* Maxmem isn't the "maximum memory", it's the highest page of
* of the physical address space. It should be "Maxphyspage".
Maxmem
= pagesinext
+ 0x100000/PAGE_SIZE
;
* Calculate number of physical pages, but account for Maxmem
(min(pagesinext
+ 0x100000/PAGE_SIZE
, Maxmem
) - 0x100000/PAGE_SIZE
);
/* call pmap initialization to make new kernel address space */
pmap_bootstrap (first
, 0);
* Do simple memory test over range of extended memory that BIOS
* indicates exists. Adjust Maxmem to the highest page of
printf("Testing memory (%dMB)...", ptoa(Maxmem
)/1024/1024);
for (target_page
= Maxmem
- 1; target_page
>= atop(first
); target_page
--) {
extern struct pte
*CMAP1
;
* map page into kernel: valid, read/write, non-cacheable
*(int *)CMAP1
= PG_V
| PG_KW
| PG_N
| ptoa(target_page
);
* Test for alternating 1's and 0's
filli(0xaaaaaaaa, CADDR1
, PAGE_SIZE
/sizeof(int));
if (test_page((int *)CADDR1
, 0xaaaaaaaa)) {
* Test for alternating 0's and 1's
filli(0x55555555, CADDR1
, PAGE_SIZE
/sizeof(int));
if (test_page((int *)CADDR1
, 0x55555555)) {
filli(0xffffffff, CADDR1
, PAGE_SIZE
/sizeof(int));
if (test_page((int *)CADDR1
, 0xffffffff)) {
bzero(CADDR1
, PAGE_SIZE
);
if (test_page((int *)CADDR1
, 0)) {
maxmem
= Maxmem
- 1; /* highest page of usable memory */
avail_end
= (maxmem
<< PAGE_SHIFT
) - i386_round_page(sizeof(struct msgbuf
));
* Initialize pointers to the two chunks of memory; for use
* later in vm_page_startup.
/* avail_start is initialized in pmap_bootstrap */
phys_avail
[x
++] = NBPG
; /* skip first page of memory */
phys_avail
[x
++] = pagesinbase
* NBPG
; /* memory up to the ISA hole */
phys_avail
[x
++] = avail_start
; /* memory up to the end */
phys_avail
[x
++] = avail_end
;
phys_avail
[x
++] = 0; /* no more chunks */
/* now running on new page tables, configured,and u/iom is accessible */
/* make a initial tss so microp can get interrupt stack on syscall! */
proc0
.p_addr
->u_pcb
.pcb_tss
.tss_esp0
= (int) kstack
+ UPAGES
*NBPG
;
proc0
.p_addr
->u_pcb
.pcb_tss
.tss_ss0
= GSEL(GDATA_SEL
, SEL_KPL
) ;
_gsel_tss
= GSEL(GPROC0_SEL
, SEL_KPL
);
((struct i386tss
*)gdt_segs
[GPROC0_SEL
].ssd_base
)->tss_ioopt
=
/* make a call gate to reenter kernel with */
gdp
= &ldt
[LSYS5CALLS_SEL
].gd
;
x
= (int) &IDTVEC(syscall
);
gdp
->gd_selector
= GSEL(GCODE_SEL
,SEL_KPL
);
gdp
->gd_type
= SDT_SYS386CGT
;
gdp
->gd_hioffset
= ((int) &IDTVEC(syscall
)) >>16;
/* transfer to user mode */
_ucodesel
= LSEL(LUCODE_SEL
, SEL_UPL
);
_udatasel
= LSEL(LUDATA_SEL
, SEL_UPL
);
bcopy(&sigcode
, proc0
.p_addr
->u_pcb
.pcb_sigc
, szsigcode
);
proc0
.p_addr
->u_pcb
.pcb_flags
= 0;
proc0
.p_addr
->u_pcb
.pcb_ptd
= IdlePTD
;
test_page(address
, pattern
)
for (x
= address
; x
< (int *)((char *)address
+ PAGE_SIZE
); x
++) {
schednetisr(NETISR_SCLK
);
* insert an element into a queue
void /* XXX replace with inline FIXME! */
register struct prochd
*element
, *head
;
element
->ph_link
= head
->ph_link
;
head
->ph_link
= (struct proc
*)element
;
element
->ph_rlink
= (struct proc
*)head
;
((struct prochd
*)(element
->ph_link
))->ph_rlink
=(struct proc
*)element
;
* remove an element from a queue
void /* XXX replace with inline FIXME! */
register struct prochd
*element
;
((struct prochd
*)(element
->ph_link
))->ph_rlink
= element
->ph_rlink
;
((struct prochd
*)(element
->ph_rlink
))->ph_link
= element
->ph_link
;
element
->ph_rlink
= (struct proc
*)0;
* The registers are in the frame; the frame is in the user area of
* the process in question; when the process is active, the registers
* are in "the kernel stack"; when it's not, they're still there, but
* things get flipped around. So, since p->p_regs is the whole address
* of the register set, take its offset from the kernel stack, and
* index into the user block. Don't you just *love* virtual memory?
* (I'm starting to think seymour is right...)
ptrace_set_pc (struct proc
*p
, unsigned int addr
) {
void *regs
= (char*)p
->p_addr
+
((char*) p
->p_regs
- (char*) kstack
);
((struct trapframe
*)regs
)->tf_eip
= addr
;
ptrace_single_step (struct proc
*p
) {
void *regs
= (char*)p
->p_addr
+
((char*) p
->p_regs
- (char*) kstack
);
((struct trapframe
*)regs
)->tf_eflags
|= PSL_T
;
* Copy the registers to user-space.
ptrace_getregs (struct proc
*p
, unsigned int *addr
) {
if (error
= fill_regs (p
, ®s
))
return copyout (®s
, addr
, sizeof (regs
));
ptrace_setregs (struct proc
*p
, unsigned int *addr
) {
if (error
= copyin (addr
, ®s
, sizeof(regs
)))
return set_regs (p
, ®s
);
fill_regs(struct proc
*p
, struct regs
*regs
) {
void *ptr
= (char*)p
->p_addr
+
((char*) p
->p_regs
- (char*) kstack
);
regs
->r_edi
= tp
->tf_edi
;
regs
->r_esi
= tp
->tf_esi
;
regs
->r_ebp
= tp
->tf_ebp
;
regs
->r_ebx
= tp
->tf_ebx
;
regs
->r_edx
= tp
->tf_edx
;
regs
->r_ecx
= tp
->tf_ecx
;
regs
->r_eax
= tp
->tf_eax
;
regs
->r_eip
= tp
->tf_eip
;
regs
->r_eflags
= tp
->tf_eflags
;
regs
->r_esp
= tp
->tf_esp
;
set_regs (struct proc
*p
, struct regs
*regs
) {
void *ptr
= (char*)p
->p_addr
+
((char*) p
->p_regs
- (char*) kstack
);
tp
->tf_edi
= regs
->r_edi
;
tp
->tf_esi
= regs
->r_esi
;
tp
->tf_ebp
= regs
->r_ebp
;
tp
->tf_ebx
= regs
->r_ebx
;
tp
->tf_edx
= regs
->r_edx
;
tp
->tf_ecx
= regs
->r_ecx
;
tp
->tf_eax
= regs
->r_eax
;
tp
->tf_eip
= regs
->r_eip
;
tp
->tf_eflags
= regs
->r_eflags
;
tp
->tf_esp
= regs
->r_esp
;
Debugger(const char *msg
)
printf("Debugger(\"%s\") called.", msg
);