c4a7cdb3b6
and is module aware. Yes, this means that kvm_nlist(3) will find symbols in loaded modules. The emulation of the nlist struct is pretty crude but seems to work well enough for all the users in the tree that I found.
241 lines
6.8 KiB
C
241 lines
6.8 KiB
C
/*-
|
|
* Copyright (c) 1992, 1993
|
|
* The Regents of the University of California. All rights reserved.
|
|
*
|
|
* This code is derived from software developed by the Computer Systems
|
|
* Engineering group at Lawrence Berkeley Laboratory under DARPA contract
|
|
* BG 91-66 and contributed to Berkeley.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
* must display the following acknowledgement:
|
|
* This product includes software developed by the University of
|
|
* California, Berkeley and its contributors.
|
|
* 4. Neither the name of the University nor the names of its contributors
|
|
* may be used to endorse or promote products derived from this software
|
|
* without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
|
|
#if defined(LIBC_SCCS) && !defined(lint)
|
|
#if 0
|
|
static char sccsid[] = "@(#)kvm_sparc.c 8.1 (Berkeley) 6/4/93";
|
|
#else
|
|
static const char rcsid[] =
|
|
"$FreeBSD$";
|
|
#endif
|
|
#endif /* LIBC_SCCS and not lint */
|
|
|
|
/*
|
|
* Sparc machine dependent routines for kvm. Hopefully, the forthcoming
|
|
* vm code will one day obsolete this module.
|
|
*/
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/user.h>
|
|
#include <sys/proc.h>
|
|
#include <sys/stat.h>
|
|
#include <unistd.h>
|
|
#include <nlist.h>
|
|
#include <kvm.h>
|
|
|
|
#include <vm/vm.h>
|
|
#include <vm/vm_param.h>
|
|
|
|
#include <limits.h>
|
|
|
|
#include "kvm_private.h"
|
|
|
|
#define NPMEG 128
|
|
|
|
/* XXX from sparc/pmap.c */
|
|
#define MAXMEM (128 * 1024 * 1024) /* no more than 128 MB phys mem */
|
|
#define NPGBANK 16 /* 2^4 pages per bank (64K / bank) */
|
|
#define BSHIFT 4 /* log2(NPGBANK) */
|
|
#define BOFFSET (NPGBANK - 1)
|
|
#define BTSIZE (MAXMEM / NBPG / NPGBANK)
|
|
#define HWTOSW(pmap_stod, pg) (pmap_stod[(pg) >> BSHIFT] | ((pg) & BOFFSET))
|
|
|
|
struct vmstate {
|
|
pmeg_t segmap[NKSEG];
|
|
int pmeg[NPMEG][NPTESG];
|
|
int pmap_stod[BTSIZE]; /* dense to sparse */
|
|
};
|
|
|
|
void
|
|
_kvm_freevtop(kd)
|
|
kvm_t *kd;
|
|
{
|
|
if (kd->vmst != 0)
|
|
free(kd->vmst);
|
|
}
|
|
|
|
int
|
|
_kvm_initvtop(kd)
|
|
kvm_t *kd;
|
|
{
|
|
register int i;
|
|
register int off;
|
|
register struct vmstate *vm;
|
|
struct stat st;
|
|
struct nlist nlist[2];
|
|
|
|
vm = (struct vmstate *)_kvm_malloc(kd, sizeof(*vm));
|
|
if (vm == 0)
|
|
return (-1);
|
|
|
|
kd->vmst = vm;
|
|
|
|
if (fstat(kd->pmfd, &st) < 0)
|
|
return (-1);
|
|
/*
|
|
* Read segment table.
|
|
*/
|
|
off = st.st_size - ctob(btoc(sizeof(vm->segmap)));
|
|
errno = 0;
|
|
if (lseek(kd->pmfd, (off_t)off, 0) == -1 && errno != 0 ||
|
|
read(kd->pmfd, (char *)vm->segmap, sizeof(vm->segmap)) < 0) {
|
|
_kvm_err(kd, kd->program, "cannot read segment map");
|
|
return (-1);
|
|
}
|
|
/*
|
|
* Read PMEGs.
|
|
*/
|
|
off = st.st_size - ctob(btoc(sizeof(vm->pmeg)) +
|
|
btoc(sizeof(vm->segmap)));
|
|
errno = 0;
|
|
if (lseek(kd->pmfd, (off_t)off, 0) == -1 && errno != 0 ||
|
|
read(kd->pmfd, (char *)vm->pmeg, sizeof(vm->pmeg)) < 0) {
|
|
_kvm_err(kd, kd->program, "cannot read PMEG table");
|
|
return (-1);
|
|
}
|
|
/*
|
|
* Make pmap_stod be an identity map so we can bootstrap it in.
|
|
* We assume it's in the first contiguous chunk of physical memory.
|
|
*/
|
|
for (i = 0; i < BTSIZE; ++i)
|
|
vm->pmap_stod[i] = i << 4;
|
|
|
|
/*
|
|
* It's okay to do this nlist separately from the one kvm_getprocs()
|
|
* does, since the only time we could gain anything by combining
|
|
* them is if we do a kvm_getprocs() on a dead kernel, which is
|
|
* not too common.
|
|
*/
|
|
nlist[0].n_name = "_pmap_stod";
|
|
nlist[1].n_name = 0;
|
|
if (kvm_nlist(kd, nlist) != 0) {
|
|
_kvm_err(kd, kd->program, "pmap_stod: no such symbol");
|
|
return (-1);
|
|
}
|
|
if (kvm_read(kd, (u_long)nlist[0].n_value,
|
|
(char *)vm->pmap_stod, sizeof(vm->pmap_stod))
|
|
!= sizeof(vm->pmap_stod)) {
|
|
_kvm_err(kd, kd->program, "cannot read pmap_stod");
|
|
return (-1);
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
#define VA_OFF(va) (va & (NBPG - 1))
|
|
|
|
/*
|
|
* Translate a user virtual address to a physical address.
|
|
*/
|
|
int
|
|
_kvm_uvatop(kd, p, va, pa)
|
|
kvm_t *kd;
|
|
const struct proc *p;
|
|
u_long va;
|
|
u_long *pa;
|
|
{
|
|
int kva, pte;
|
|
register int off, frame;
|
|
register struct vmspace *vms = p->p_vmspace;
|
|
|
|
if ((u_long)vms < KERNBASE) {
|
|
_kvm_err(kd, kd->program, "_kvm_uvatop: corrupt proc");
|
|
return (0);
|
|
}
|
|
if (va >= KERNBASE)
|
|
return (0);
|
|
/*
|
|
* Get the PTE. This takes two steps. We read the
|
|
* base address of the table, then we index it.
|
|
* Note that the index pte table is indexed by
|
|
* virtual segment rather than physical segment.
|
|
*/
|
|
kva = (u_long)&vms->vm_pmap.pm_rpte[VA_VSEG(va)];
|
|
if (kvm_read(kd, kva, (char *)&kva, 4) != 4 || kva == 0)
|
|
goto invalid;
|
|
kva += sizeof(vms->vm_pmap.pm_rpte[0]) * VA_VPG(va);
|
|
if (kvm_read(kd, kva, (char *)&pte, 4) == 4 && (pte & PG_V)) {
|
|
off = VA_OFF(va);
|
|
/*
|
|
* /dev/mem adheres to the hardware model of physical memory
|
|
* (with holes in the address space), while crashdumps
|
|
* adhere to the contiguous software model.
|
|
*/
|
|
if (ISALIVE(kd))
|
|
frame = pte & PG_PFNUM;
|
|
else
|
|
frame = HWTOSW(kd->vmst->pmap_stod, pte & PG_PFNUM);
|
|
*pa = (frame << PGSHIFT) | off;
|
|
return (NBPG - off);
|
|
}
|
|
invalid:
|
|
_kvm_err(kd, 0, "invalid address (%x)", va);
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* Translate a kernel virtual address to a physical address using the
|
|
* mapping information in kd->vm. Returns the result in pa, and returns
|
|
* the number of bytes that are contiguously available from this
|
|
* physical address. This routine is used only for crashdumps.
|
|
*/
|
|
int
|
|
_kvm_kvatop(kd, va, pa)
|
|
kvm_t *kd;
|
|
u_long va;
|
|
u_long *pa;
|
|
{
|
|
register struct vmstate *vm;
|
|
register int s;
|
|
register int pte;
|
|
register int off;
|
|
|
|
if (va >= KERNBASE) {
|
|
vm = kd->vmst;
|
|
s = vm->segmap[VA_VSEG(va) - NUSEG];
|
|
pte = vm->pmeg[s][VA_VPG(va)];
|
|
if ((pte & PG_V) != 0) {
|
|
off = VA_OFF(va);
|
|
*pa = (HWTOSW(vm->pmap_stod, pte & PG_PFNUM)
|
|
<< PGSHIFT) | off;
|
|
|
|
return (NBPG - off);
|
|
}
|
|
}
|
|
_kvm_err(kd, 0, "invalid address (%x)", va);
|
|
return (0);
|
|
}
|