1998-06-10 19:59:41 +00:00
|
|
|
/*-
|
|
|
|
* Copyright (c) 1982, 1986 The Regents of the University of California.
|
|
|
|
* Copyright (c) 1989, 1990 William Jolitz
|
|
|
|
* Copyright (c) 1994 John Dyson
|
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* This code is derived from software contributed to Berkeley by
|
|
|
|
* the Systems Programming Group of the University of Utah Computer
|
|
|
|
* Science Department, and William Jolitz.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
|
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
|
|
* must display the following acknowledgement:
|
|
|
|
* This product includes software developed by the University of
|
|
|
|
* California, Berkeley and its contributors.
|
|
|
|
* 4. Neither the name of the University nor the names of its contributors
|
|
|
|
* may be used to endorse or promote products derived from this software
|
|
|
|
* without specific prior written permission.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
|
|
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
|
|
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
|
|
* SUCH DAMAGE.
|
|
|
|
*
|
|
|
|
* from: @(#)vm_machdep.c 7.3 (Berkeley) 5/13/91
|
|
|
|
* Utah $Hdr: vm_machdep.c 1.16.1.1 89/06/23$
|
1999-08-28 01:08:13 +00:00
|
|
|
* $FreeBSD$
|
1998-06-10 19:59:41 +00:00
|
|
|
*/
|
|
|
|
/*
|
|
|
|
* Copyright (c) 1994, 1995, 1996 Carnegie-Mellon University.
|
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* Author: Chris G. Demetriou
|
|
|
|
*
|
|
|
|
* Permission to use, copy, modify and distribute this software and
|
|
|
|
* its documentation is hereby granted, provided that both the copyright
|
|
|
|
* notice and this permission notice appear in all copies of the
|
|
|
|
* software, derivative works or modified versions, and any portions
|
|
|
|
* thereof, and that both notices appear in supporting documentation.
|
|
|
|
*
|
|
|
|
* CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
|
|
|
|
* CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND
|
|
|
|
* FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
|
|
|
|
*
|
|
|
|
* Carnegie Mellon requests users of this software to return to
|
|
|
|
*
|
|
|
|
* Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU
|
|
|
|
* School of Computer Science
|
|
|
|
* Carnegie Mellon University
|
|
|
|
* Pittsburgh PA 15213-3890
|
|
|
|
*
|
|
|
|
* any improvements or extensions that they make and grant Carnegie the
|
|
|
|
* rights to redistribute these changes.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <sys/param.h>
|
|
|
|
#include <sys/systm.h>
|
|
|
|
#include <sys/proc.h>
|
|
|
|
#include <sys/malloc.h>
|
2000-05-05 09:59:14 +00:00
|
|
|
#include <sys/bio.h>
|
1998-06-10 19:59:41 +00:00
|
|
|
#include <sys/buf.h>
|
2000-10-20 07:58:15 +00:00
|
|
|
#include <sys/mutex.h>
|
1998-06-10 19:59:41 +00:00
|
|
|
#include <sys/vnode.h>
|
|
|
|
#include <sys/vmmeter.h>
|
|
|
|
#include <sys/kernel.h>
|
|
|
|
#include <sys/sysctl.h>
|
1999-12-06 18:12:29 +00:00
|
|
|
#include <sys/unistd.h>
|
1998-06-10 19:59:41 +00:00
|
|
|
|
|
|
|
#include <machine/clock.h>
|
|
|
|
#include <machine/cpu.h>
|
1998-12-04 10:52:48 +00:00
|
|
|
#include <machine/fpu.h>
|
1998-06-10 19:59:41 +00:00
|
|
|
#include <machine/md_var.h>
|
|
|
|
#include <machine/prom.h>
|
|
|
|
|
|
|
|
#include <vm/vm.h>
|
|
|
|
#include <vm/vm_param.h>
|
|
|
|
#include <sys/lock.h>
|
|
|
|
#include <vm/vm_kern.h>
|
|
|
|
#include <vm/vm_page.h>
|
|
|
|
#include <vm/vm_map.h>
|
|
|
|
#include <vm/vm_extern.h>
|
|
|
|
|
|
|
|
#include <sys/user.h>
|
|
|
|
|
|
|
|
/*
|
|
|
|
* quick version of vm_fault
|
|
|
|
*/
|
1999-09-20 19:08:48 +00:00
|
|
|
int
|
1998-06-10 19:59:41 +00:00
|
|
|
vm_fault_quick(v, prot)
|
|
|
|
caddr_t v;
|
|
|
|
int prot;
|
|
|
|
{
|
1999-09-20 19:08:48 +00:00
|
|
|
int r;
|
1998-06-10 19:59:41 +00:00
|
|
|
if (prot & VM_PROT_WRITE)
|
1999-09-20 19:08:48 +00:00
|
|
|
r = subyte(v, fubyte(v));
|
1998-06-10 19:59:41 +00:00
|
|
|
else
|
1999-09-20 19:08:48 +00:00
|
|
|
r = fubyte(v);
|
|
|
|
return(r);
|
1998-06-10 19:59:41 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Finish a fork operation, with process p2 nearly set up.
|
|
|
|
* Copy and update the pcb, set up the stack so that the child
|
|
|
|
* ready to run and return to user mode.
|
|
|
|
*/
|
|
|
|
void
|
1999-12-06 04:53:08 +00:00
|
|
|
cpu_fork(p1, p2, flags)
|
1998-06-10 19:59:41 +00:00
|
|
|
register struct proc *p1, *p2;
|
1999-12-06 04:53:08 +00:00
|
|
|
int flags;
|
1998-06-10 19:59:41 +00:00
|
|
|
{
|
1999-12-06 04:53:08 +00:00
|
|
|
if ((flags & RFPROC) == 0)
|
|
|
|
return;
|
|
|
|
|
1998-06-10 19:59:41 +00:00
|
|
|
p2->p_md.md_tf = p1->p_md.md_tf;
|
2000-01-16 07:07:33 +00:00
|
|
|
p2->p_md.md_flags = p1->p_md.md_flags & (MDP_FPUSED | MDP_UAC_MASK);
|
1998-06-10 19:59:41 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Cache the physical address of the pcb, so we can
|
|
|
|
* swap to it easily.
|
|
|
|
*/
|
2000-06-11 06:22:01 +00:00
|
|
|
p2->p_md.md_pcbpaddr = (void*)vtophys((vm_offset_t)&p2->p_addr->u_pcb);
|
1998-06-10 19:59:41 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Copy floating point state from the FP chip to the PCB
|
|
|
|
* if this process has state stored there.
|
|
|
|
*/
|
1999-11-10 21:14:25 +00:00
|
|
|
alpha_fpstate_save(p1, 0);
|
1998-06-10 19:59:41 +00:00
|
|
|
|
|
|
|
/*
|
1999-11-10 21:14:25 +00:00
|
|
|
* Copy pcb and stack from proc p1 to p2. We do this as
|
|
|
|
* cheaply as possible, copying only the active part of the
|
|
|
|
* stack. The stack and pcb need to agree. Make sure that the
|
|
|
|
* new process has FEN disabled.
|
1998-06-10 19:59:41 +00:00
|
|
|
*/
|
|
|
|
p2->p_addr->u_pcb = p1->p_addr->u_pcb;
|
|
|
|
p2->p_addr->u_pcb.pcb_hw.apcb_usp = alpha_pal_rdusp();
|
1999-11-10 21:14:25 +00:00
|
|
|
p2->p_addr->u_pcb.pcb_hw.apcb_flags &= ~ALPHA_PCB_FLAGS_FEN;
|
1998-06-10 19:59:41 +00:00
|
|
|
|
1998-12-04 10:52:48 +00:00
|
|
|
/*
|
|
|
|
* Set the floating point state.
|
|
|
|
*/
|
|
|
|
if ((p2->p_addr->u_pcb.pcb_fp_control & IEEE_INHERIT) == 0) {
|
1998-12-30 10:38:59 +00:00
|
|
|
p2->p_addr->u_pcb.pcb_fp_control = 0;
|
1998-12-04 10:52:48 +00:00
|
|
|
p2->p_addr->u_pcb.pcb_fp.fpr_cr = (FPCR_DYN_NORMAL
|
1998-12-30 10:38:59 +00:00
|
|
|
| FPCR_INVD | FPCR_DZED
|
|
|
|
| FPCR_OVFD | FPCR_INED
|
|
|
|
| FPCR_UNFD);
|
1998-12-04 10:52:48 +00:00
|
|
|
}
|
|
|
|
|
1998-06-10 19:59:41 +00:00
|
|
|
/*
|
|
|
|
* Arrange for a non-local goto when the new process
|
|
|
|
* is started, to resume here, returning nonzero from setjmp.
|
|
|
|
*/
|
|
|
|
#ifdef DIAGNOSTIC
|
|
|
|
if (p1 != curproc)
|
|
|
|
panic("cpu_fork: curproc");
|
1999-11-10 21:14:25 +00:00
|
|
|
alpha_fpstate_check(p1);
|
1998-06-10 19:59:41 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
/*
|
|
|
|
* create the child's kernel stack, from scratch.
|
|
|
|
*/
|
|
|
|
{
|
2000-06-11 06:22:01 +00:00
|
|
|
struct user *up = p2->p_addr;
|
1998-06-10 19:59:41 +00:00
|
|
|
struct trapframe *p2tf;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Pick a stack pointer, leaving room for a trapframe;
|
|
|
|
* copy trapframe from parent so return to user mode
|
|
|
|
* will be to right address, with correct registers.
|
|
|
|
*/
|
|
|
|
p2tf = p2->p_md.md_tf = (struct trapframe *)
|
|
|
|
((char *)p2->p_addr + USPACE - sizeof(struct trapframe));
|
|
|
|
bcopy(p1->p_md.md_tf, p2->p_md.md_tf,
|
|
|
|
sizeof(struct trapframe));
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Set up return-value registers as fork() libc stub expects.
|
|
|
|
*/
|
1999-08-27 14:47:23 +00:00
|
|
|
p2tf->tf_regs[FRAME_V0] = 0; /* child's pid (linux) */
|
|
|
|
p2tf->tf_regs[FRAME_A3] = 0; /* no error */
|
|
|
|
p2tf->tf_regs[FRAME_A4] = 1; /* is child (FreeBSD) */
|
1998-06-10 19:59:41 +00:00
|
|
|
|
|
|
|
/*
|
2001-01-24 21:59:25 +00:00
|
|
|
* Arrange for continuation at fork_return(), which
|
1998-06-10 19:59:41 +00:00
|
|
|
* will return to exception_return(). Note that the child
|
|
|
|
* process doesn't stay in the kernel for long!
|
|
|
|
*
|
|
|
|
* This is an inlined version of cpu_set_kpc.
|
|
|
|
*/
|
|
|
|
up->u_pcb.pcb_hw.apcb_ksp = (u_int64_t)p2tf;
|
|
|
|
up->u_pcb.pcb_context[0] =
|
2001-01-26 23:32:38 +00:00
|
|
|
(u_int64_t)fork_return; /* s0: a0 */
|
1998-06-10 19:59:41 +00:00
|
|
|
up->u_pcb.pcb_context[1] =
|
|
|
|
(u_int64_t)exception_return; /* s1: ra */
|
2001-01-26 23:32:38 +00:00
|
|
|
up->u_pcb.pcb_context[2] = (u_long) p2; /* s2: a1 */
|
1998-06-10 19:59:41 +00:00
|
|
|
up->u_pcb.pcb_context[7] =
|
2001-02-22 16:56:53 +00:00
|
|
|
(u_int64_t)fork_trampoline; /* ra: assembly magic */
|
1998-06-10 19:59:41 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Intercept the return address from a freshly forked process that has NOT
|
|
|
|
* been scheduled yet.
|
|
|
|
*
|
|
|
|
* This is needed to make kernel threads stay in kernel mode.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
cpu_set_fork_handler(p, func, arg)
|
|
|
|
struct proc *p;
|
1999-07-01 13:21:46 +00:00
|
|
|
void (*func) __P((void *));
|
|
|
|
void *arg;
|
1998-06-10 19:59:41 +00:00
|
|
|
{
|
|
|
|
/*
|
|
|
|
* Note that the trap frame follows the args, so the function
|
|
|
|
* is really called like this: func(arg, frame);
|
|
|
|
*/
|
|
|
|
p->p_addr->u_pcb.pcb_context[0] = (u_long) func;
|
|
|
|
p->p_addr->u_pcb.pcb_context[2] = (u_long) arg;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* cpu_exit is called as the last action during exit.
|
|
|
|
* We release the address space of the process, block interrupts,
|
|
|
|
* and call switch_exit. switch_exit switches to proc0's PCB and stack,
|
|
|
|
* then jumps into the middle of cpu_switch, as if it were switching
|
|
|
|
* from proc0.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
cpu_exit(p)
|
|
|
|
register struct proc *p;
|
|
|
|
{
|
1999-11-10 21:14:25 +00:00
|
|
|
alpha_fpstate_drop(p);
|
1998-06-10 19:59:41 +00:00
|
|
|
|
2001-03-07 03:20:15 +00:00
|
|
|
PROC_LOCK(p);
|
Change and clean the mutex lock interface.
mtx_enter(lock, type) becomes:
mtx_lock(lock) for sleep locks (MTX_DEF-initialized locks)
mtx_lock_spin(lock) for spin locks (MTX_SPIN-initialized)
similarily, for releasing a lock, we now have:
mtx_unlock(lock) for MTX_DEF and mtx_unlock_spin(lock) for MTX_SPIN.
We change the caller interface for the two different types of locks
because the semantics are entirely different for each case, and this
makes it explicitly clear and, at the same time, it rids us of the
extra `type' argument.
The enter->lock and exit->unlock change has been made with the idea
that we're "locking data" and not "entering locked code" in mind.
Further, remove all additional "flags" previously passed to the
lock acquire/release routines with the exception of two:
MTX_QUIET and MTX_NOSWITCH
The functionality of these flags is preserved and they can be passed
to the lock/unlock routines by calling the corresponding wrappers:
mtx_{lock, unlock}_flags(lock, flag(s)) and
mtx_{lock, unlock}_spin_flags(lock, flag(s)) for MTX_DEF and MTX_SPIN
locks, respectively.
Re-inline some lock acq/rel code; in the sleep lock case, we only
inline the _obtain_lock()s in order to ensure that the inlined code
fits into a cache line. In the spin lock case, we inline recursion and
actually only perform a function call if we need to spin. This change
has been made with the idea that we generally tend to avoid spin locks
and that also the spin locks that we do have and are heavily used
(i.e. sched_lock) do recurse, and therefore in an effort to reduce
function call overhead for some architectures (such as alpha), we
inline recursion for this case.
Create a new malloc type for the witness code and retire from using
the M_DEV type. The new type is called M_WITNESS and is only declared
if WITNESS is enabled.
Begin cleaning up some machdep/mutex.h code - specifically updated the
"optimized" inlined code in alpha/mutex.h and wrote MTX_LOCK_SPIN
and MTX_UNLOCK_SPIN asm macros for the i386/mutex.h as we presently
need those.
Finally, caught up to the interface changes in all sys code.
Contributors: jake, jhb, jasone (in no particular order)
2001-02-09 06:11:45 +00:00
|
|
|
mtx_lock_spin(&sched_lock);
|
|
|
|
mtx_unlock_flags(&Giant, MTX_NOSWITCH);
|
2000-11-15 19:44:38 +00:00
|
|
|
mtx_assert(&Giant, MA_NOTOWNED);
|
2000-12-23 19:43:10 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* We have to wait until after releasing all locks before
|
|
|
|
* changing p_stat. If we block on a mutex then we will be
|
|
|
|
* back at SRUN when we resume and our parent will never
|
|
|
|
* harvest us.
|
|
|
|
*/
|
|
|
|
p->p_stat = SZOMB;
|
|
|
|
|
|
|
|
wakeup(p->p_pptr);
|
2001-03-07 03:20:15 +00:00
|
|
|
PROC_UNLOCK_NOSWITCH(p);
|
2000-12-23 19:43:10 +00:00
|
|
|
|
1998-06-10 19:59:41 +00:00
|
|
|
cnt.v_swtch++;
|
2000-09-07 01:33:02 +00:00
|
|
|
cpu_switch();
|
1998-06-10 19:59:41 +00:00
|
|
|
panic("cpu_exit");
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
cpu_wait(p)
|
|
|
|
struct proc *p;
|
|
|
|
{
|
|
|
|
/* drop per-process resources */
|
|
|
|
pmap_dispose_proc(p);
|
|
|
|
|
|
|
|
/* and clean-out the vmspace */
|
|
|
|
vmspace_free(p->p_vmspace);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Dump the machine specific header information at the start of a core dump.
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
cpu_coredump(p, vp, cred)
|
|
|
|
struct proc *p;
|
|
|
|
struct vnode *vp;
|
|
|
|
struct ucred *cred;
|
|
|
|
{
|
|
|
|
|
|
|
|
return (vn_rdwr(UIO_WRITE, vp, (caddr_t) p->p_addr, ctob(UPAGES),
|
|
|
|
(off_t)0, UIO_SYSSPACE, IO_NODELOCKED|IO_UNIT, cred, (int *)NULL,
|
|
|
|
p));
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef notyet
|
|
|
|
static void
|
|
|
|
setredzone(pte, vaddr)
|
|
|
|
u_short *pte;
|
|
|
|
caddr_t vaddr;
|
|
|
|
{
|
|
|
|
/* eventually do this by setting up an expand-down stack segment
|
|
|
|
for ss0: selector, allowing stack access down to top of u.
|
|
|
|
this means though that protection violations need to be handled
|
|
|
|
thru a double fault exception that must do an integral task
|
|
|
|
switch to a known good context, within which a dump can be
|
|
|
|
taken. a sensible scheme might be to save the initial context
|
|
|
|
used by sched (that has physical memory mapped 1:1 at bottom)
|
|
|
|
and take the dump while still in mapped mode */
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Map an IO request into kernel virtual address space.
|
|
|
|
*
|
|
|
|
* All requests are (re)mapped into kernel VA space.
|
|
|
|
* Notice that we use b_bufsize for the size of the buffer
|
|
|
|
* to be mapped. b_bcount might be modified by the driver.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
vmapbuf(bp)
|
|
|
|
register struct buf *bp;
|
|
|
|
{
|
|
|
|
register caddr_t addr, v, kva;
|
|
|
|
vm_offset_t pa;
|
|
|
|
|
|
|
|
if ((bp->b_flags & B_PHYS) == 0)
|
|
|
|
panic("vmapbuf");
|
|
|
|
|
|
|
|
for (v = bp->b_saveaddr, addr = (caddr_t)trunc_page(bp->b_data);
|
|
|
|
addr < bp->b_data + bp->b_bufsize;
|
|
|
|
addr += PAGE_SIZE, v += PAGE_SIZE) {
|
|
|
|
/*
|
|
|
|
* Do the vm_fault if needed; do the copy-on-write thing
|
|
|
|
* when reading stuff off device into memory.
|
|
|
|
*/
|
|
|
|
vm_fault_quick(addr,
|
2000-03-20 10:44:49 +00:00
|
|
|
(bp->b_iocmd == BIO_READ)?(VM_PROT_READ|VM_PROT_WRITE):VM_PROT_READ);
|
1998-06-10 19:59:41 +00:00
|
|
|
pa = trunc_page(pmap_kextract((vm_offset_t) addr));
|
|
|
|
if (pa == 0)
|
|
|
|
panic("vmapbuf: page not present");
|
|
|
|
vm_page_hold(PHYS_TO_VM_PAGE(pa));
|
|
|
|
pmap_kenter((vm_offset_t) v, pa);
|
|
|
|
}
|
|
|
|
|
|
|
|
kva = bp->b_saveaddr;
|
|
|
|
bp->b_saveaddr = bp->b_data;
|
|
|
|
bp->b_data = kva + (((vm_offset_t) bp->b_data) & PAGE_MASK);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Free the io map PTEs associated with this IO operation.
|
|
|
|
* We also invalidate the TLB entries and restore the original b_addr.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
vunmapbuf(bp)
|
|
|
|
register struct buf *bp;
|
|
|
|
{
|
|
|
|
register caddr_t addr;
|
|
|
|
vm_offset_t pa;
|
|
|
|
|
|
|
|
if ((bp->b_flags & B_PHYS) == 0)
|
|
|
|
panic("vunmapbuf");
|
|
|
|
|
|
|
|
for (addr = (caddr_t)trunc_page(bp->b_data);
|
|
|
|
addr < bp->b_data + bp->b_bufsize;
|
|
|
|
addr += PAGE_SIZE) {
|
|
|
|
pa = trunc_page(pmap_kextract((vm_offset_t) addr));
|
|
|
|
pmap_kremove((vm_offset_t) addr);
|
|
|
|
vm_page_unhold(PHYS_TO_VM_PAGE(pa));
|
|
|
|
}
|
|
|
|
|
|
|
|
bp->b_data = bp->b_saveaddr;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2000-09-07 01:33:02 +00:00
|
|
|
* Reset back to firmware.
|
1998-06-10 19:59:41 +00:00
|
|
|
*/
|
|
|
|
void
|
|
|
|
cpu_reset()
|
|
|
|
{
|
|
|
|
prom_halt(0);
|
|
|
|
}
|
|
|
|
|
1999-01-26 02:49:52 +00:00
|
|
|
int
|
|
|
|
grow_stack(p, sp)
|
|
|
|
struct proc *p;
|
|
|
|
size_t sp;
|
|
|
|
{
|
|
|
|
int rv;
|
|
|
|
|
|
|
|
rv = vm_map_growstack (p, sp);
|
|
|
|
if (rv != KERN_SUCCESS)
|
|
|
|
return (0);
|
|
|
|
|
|
|
|
return (1);
|
|
|
|
}
|
|
|
|
|
1998-06-10 19:59:41 +00:00
|
|
|
|
|
|
|
static int cnt_prezero;
|
|
|
|
|
|
|
|
SYSCTL_INT(_machdep, OID_AUTO, cnt_prezero, CTLFLAG_RD, &cnt_prezero, 0, "");
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Implement the pre-zeroed page mechanism.
|
|
|
|
* This routine is called from the idle loop.
|
|
|
|
*/
|
1999-02-08 02:42:13 +00:00
|
|
|
|
|
|
|
#define ZIDLE_LO(v) ((v) * 2 / 3)
|
|
|
|
#define ZIDLE_HI(v) ((v) * 4 / 5)
|
|
|
|
|
1998-06-10 19:59:41 +00:00
|
|
|
int
|
|
|
|
vm_page_zero_idle()
|
|
|
|
{
|
|
|
|
static int free_rover;
|
1999-02-08 00:47:32 +00:00
|
|
|
static int zero_state;
|
1998-06-10 19:59:41 +00:00
|
|
|
vm_page_t m;
|
|
|
|
int s;
|
|
|
|
|
|
|
|
/*
|
1999-02-08 00:47:32 +00:00
|
|
|
* Attempt to maintain approximately 1/2 of our free pages in a
|
|
|
|
* PG_ZERO'd state. Add some hysteresis to (attempt to) avoid
|
|
|
|
* generally zeroing a page when the system is near steady-state.
|
|
|
|
* Otherwise we might get 'flutter' during disk I/O / IPC or
|
|
|
|
* fast sleeps. We also do not want to be continuously zeroing
|
|
|
|
* pages because doing so may flush our L1 and L2 caches too much.
|
1998-06-10 19:59:41 +00:00
|
|
|
*/
|
1999-02-08 00:47:32 +00:00
|
|
|
|
1999-02-08 02:42:13 +00:00
|
|
|
if (zero_state && vm_page_zero_count >= ZIDLE_LO(cnt.v_free_count))
|
1999-02-08 00:47:32 +00:00
|
|
|
return(0);
|
1999-02-08 02:42:13 +00:00
|
|
|
if (vm_page_zero_count >= ZIDLE_HI(cnt.v_free_count))
|
1999-02-08 00:47:32 +00:00
|
|
|
return(0);
|
|
|
|
|
Change and clean the mutex lock interface.
mtx_enter(lock, type) becomes:
mtx_lock(lock) for sleep locks (MTX_DEF-initialized locks)
mtx_lock_spin(lock) for spin locks (MTX_SPIN-initialized)
similarily, for releasing a lock, we now have:
mtx_unlock(lock) for MTX_DEF and mtx_unlock_spin(lock) for MTX_SPIN.
We change the caller interface for the two different types of locks
because the semantics are entirely different for each case, and this
makes it explicitly clear and, at the same time, it rids us of the
extra `type' argument.
The enter->lock and exit->unlock change has been made with the idea
that we're "locking data" and not "entering locked code" in mind.
Further, remove all additional "flags" previously passed to the
lock acquire/release routines with the exception of two:
MTX_QUIET and MTX_NOSWITCH
The functionality of these flags is preserved and they can be passed
to the lock/unlock routines by calling the corresponding wrappers:
mtx_{lock, unlock}_flags(lock, flag(s)) and
mtx_{lock, unlock}_spin_flags(lock, flag(s)) for MTX_DEF and MTX_SPIN
locks, respectively.
Re-inline some lock acq/rel code; in the sleep lock case, we only
inline the _obtain_lock()s in order to ensure that the inlined code
fits into a cache line. In the spin lock case, we inline recursion and
actually only perform a function call if we need to spin. This change
has been made with the idea that we generally tend to avoid spin locks
and that also the spin locks that we do have and are heavily used
(i.e. sched_lock) do recurse, and therefore in an effort to reduce
function call overhead for some architectures (such as alpha), we
inline recursion for this case.
Create a new malloc type for the witness code and retire from using
the M_DEV type. The new type is called M_WITNESS and is only declared
if WITNESS is enabled.
Begin cleaning up some machdep/mutex.h code - specifically updated the
"optimized" inlined code in alpha/mutex.h and wrote MTX_LOCK_SPIN
and MTX_UNLOCK_SPIN asm macros for the i386/mutex.h as we presently
need those.
Finally, caught up to the interface changes in all sys code.
Contributors: jake, jhb, jasone (in no particular order)
2001-02-09 06:11:45 +00:00
|
|
|
if (mtx_trylock(&Giant)) {
|
1998-06-10 19:59:41 +00:00
|
|
|
s = splvm();
|
1999-02-08 00:37:36 +00:00
|
|
|
m = vm_page_list_find(PQ_FREE, free_rover, FALSE);
|
1999-02-08 00:47:32 +00:00
|
|
|
zero_state = 0;
|
1999-02-08 00:37:36 +00:00
|
|
|
if (m != NULL && (m->flags & PG_ZERO) == 0) {
|
1999-08-05 23:38:13 +00:00
|
|
|
vm_page_queues[m->queue].lcnt--;
|
1999-10-30 07:37:14 +00:00
|
|
|
TAILQ_REMOVE(&vm_page_queues[m->queue].pl, m, pageq);
|
1998-06-10 19:59:41 +00:00
|
|
|
m->queue = PQ_NONE;
|
|
|
|
splx(s);
|
|
|
|
#if 0
|
|
|
|
rel_mplock();
|
|
|
|
#endif
|
|
|
|
pmap_zero_page(VM_PAGE_TO_PHYS(m));
|
|
|
|
#if 0
|
|
|
|
get_mplock();
|
|
|
|
#endif
|
|
|
|
(void)splvm();
|
1999-02-08 00:37:36 +00:00
|
|
|
vm_page_flag_set(m, PG_ZERO);
|
|
|
|
m->queue = PQ_FREE + m->pc;
|
1999-08-05 23:38:13 +00:00
|
|
|
vm_page_queues[m->queue].lcnt++;
|
1999-10-30 07:37:14 +00:00
|
|
|
TAILQ_INSERT_TAIL(&vm_page_queues[m->queue].pl, m,
|
1998-06-10 19:59:41 +00:00
|
|
|
pageq);
|
|
|
|
++vm_page_zero_count;
|
|
|
|
++cnt_prezero;
|
1999-02-08 02:42:13 +00:00
|
|
|
if (vm_page_zero_count >= ZIDLE_HI(cnt.v_free_count))
|
1999-02-08 00:47:32 +00:00
|
|
|
zero_state = 1;
|
1998-06-10 19:59:41 +00:00
|
|
|
}
|
1999-07-22 06:04:17 +00:00
|
|
|
free_rover = (free_rover + PQ_PRIME2) & PQ_L2_MASK;
|
1998-06-10 19:59:41 +00:00
|
|
|
splx(s);
|
Change and clean the mutex lock interface.
mtx_enter(lock, type) becomes:
mtx_lock(lock) for sleep locks (MTX_DEF-initialized locks)
mtx_lock_spin(lock) for spin locks (MTX_SPIN-initialized)
similarily, for releasing a lock, we now have:
mtx_unlock(lock) for MTX_DEF and mtx_unlock_spin(lock) for MTX_SPIN.
We change the caller interface for the two different types of locks
because the semantics are entirely different for each case, and this
makes it explicitly clear and, at the same time, it rids us of the
extra `type' argument.
The enter->lock and exit->unlock change has been made with the idea
that we're "locking data" and not "entering locked code" in mind.
Further, remove all additional "flags" previously passed to the
lock acquire/release routines with the exception of two:
MTX_QUIET and MTX_NOSWITCH
The functionality of these flags is preserved and they can be passed
to the lock/unlock routines by calling the corresponding wrappers:
mtx_{lock, unlock}_flags(lock, flag(s)) and
mtx_{lock, unlock}_spin_flags(lock, flag(s)) for MTX_DEF and MTX_SPIN
locks, respectively.
Re-inline some lock acq/rel code; in the sleep lock case, we only
inline the _obtain_lock()s in order to ensure that the inlined code
fits into a cache line. In the spin lock case, we inline recursion and
actually only perform a function call if we need to spin. This change
has been made with the idea that we generally tend to avoid spin locks
and that also the spin locks that we do have and are heavily used
(i.e. sched_lock) do recurse, and therefore in an effort to reduce
function call overhead for some architectures (such as alpha), we
inline recursion for this case.
Create a new malloc type for the witness code and retire from using
the M_DEV type. The new type is called M_WITNESS and is only declared
if WITNESS is enabled.
Begin cleaning up some machdep/mutex.h code - specifically updated the
"optimized" inlined code in alpha/mutex.h and wrote MTX_LOCK_SPIN
and MTX_UNLOCK_SPIN asm macros for the i386/mutex.h as we presently
need those.
Finally, caught up to the interface changes in all sys code.
Contributors: jake, jhb, jasone (in no particular order)
2001-02-09 06:11:45 +00:00
|
|
|
mtx_unlock(&Giant);
|
1998-06-10 19:59:41 +00:00
|
|
|
return (1);
|
|
|
|
}
|
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Software interrupt handler for queued VM system processing.
|
|
|
|
*/
|
|
|
|
void
|
2000-10-25 05:19:40 +00:00
|
|
|
swi_vm(void *dummy)
|
1998-06-10 19:59:41 +00:00
|
|
|
{
|
|
|
|
if (busdma_swi_pending != 0)
|
|
|
|
busdma_swi();
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Tell whether this address is in some physical memory region.
|
|
|
|
* Currently used by the kernel coredump code in order to avoid
|
|
|
|
* dumping the ``ISA memory hole'' which could cause indefinite hangs,
|
|
|
|
* or other unpredictable behaviour.
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
|
|
int
|
|
|
|
is_physical_memory(addr)
|
|
|
|
vm_offset_t addr;
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* stuff other tests for known memory-mapped devices (PCI?)
|
|
|
|
* here
|
|
|
|
*/
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|