1995-05-30 08:16:23 +00:00
|
|
|
/*
|
1993-06-12 14:58:17 +00:00
|
|
|
* Copyright (c) 1991 Regents of the University of California.
|
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* This code is derived from software contributed to Berkeley by
|
|
|
|
* the Systems Programming Group of the University of Utah Computer
|
|
|
|
* Science Department and William Jolitz of UUNET Technologies Inc.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
|
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
|
|
* must display the following acknowledgement:
|
|
|
|
* This product includes software developed by the University of
|
|
|
|
* California, Berkeley and its contributors.
|
|
|
|
* 4. Neither the name of the University nor the names of its contributors
|
|
|
|
* may be used to endorse or promote products derived from this software
|
|
|
|
* without specific prior written permission.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
|
|
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
|
|
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
|
|
* SUCH DAMAGE.
|
|
|
|
*
|
|
|
|
* Derived from hp300 version by Mike Hibler, this version by William
|
|
|
|
* Jolitz uses a recursive map [a pde points to the page directory] to
|
|
|
|
* map the page tables using the pagetables themselves. This is done to
|
|
|
|
* reduce the impact on kernel virtual memory for lots of sparse address
|
|
|
|
* space, and to reduce the cost of memory to each process.
|
|
|
|
*
|
1993-10-15 10:07:45 +00:00
|
|
|
* from: hp300: @(#)pmap.h 7.2 (Berkeley) 12/16/90
|
|
|
|
* from: @(#)pmap.h 7.4 (Berkeley) 5/12/91
|
1999-08-28 01:08:13 +00:00
|
|
|
* $FreeBSD$
|
1993-06-12 14:58:17 +00:00
|
|
|
*/
|
|
|
|
|
1994-11-14 14:12:24 +00:00
|
|
|
#ifndef _MACHINE_PMAP_H_
|
|
|
|
#define _MACHINE_PMAP_H_
|
1993-06-12 14:58:17 +00:00
|
|
|
|
1996-05-02 22:25:18 +00:00
|
|
|
/*
|
|
|
|
* Page-directory and page-table entires follow this format, with a few
|
|
|
|
* of the fields not present here and there, depending on a lot of things.
|
|
|
|
*/
|
|
|
|
/* ---- Intel Nomenclature ---- */
|
|
|
|
#define PG_V 0x001 /* P Valid */
|
|
|
|
#define PG_RW 0x002 /* R/W Read/Write */
|
|
|
|
#define PG_U 0x004 /* U/S User/Supervisor */
|
|
|
|
#define PG_NC_PWT 0x008 /* PWT Write through */
|
|
|
|
#define PG_NC_PCD 0x010 /* PCD Cache disable */
|
|
|
|
#define PG_A 0x020 /* A Accessed */
|
|
|
|
#define PG_M 0x040 /* D Dirty */
|
|
|
|
#define PG_PS 0x080 /* PS Page size (0=4k,1=4M) */
|
|
|
|
#define PG_G 0x100 /* G Global */
|
|
|
|
#define PG_AVAIL1 0x200 /* / Available for system */
|
|
|
|
#define PG_AVAIL2 0x400 /* < programmers use */
|
|
|
|
#define PG_AVAIL3 0x800 /* \ */
|
|
|
|
|
|
|
|
|
|
|
|
/* Our various interpretations of the above */
|
|
|
|
#define PG_W PG_AVAIL1 /* "Wired" pseudoflag */
|
1996-05-18 03:38:05 +00:00
|
|
|
#define PG_MANAGED PG_AVAIL2
|
2003-03-25 00:07:06 +00:00
|
|
|
#define PG_FRAME (~((vm_paddr_t)PAGE_MASK))
|
1996-05-02 22:25:18 +00:00
|
|
|
#define PG_PROT (PG_RW|PG_U) /* all protection bits . */
|
|
|
|
#define PG_N (PG_NC_PWT|PG_NC_PCD) /* Non-cacheable */
|
1993-06-12 14:58:17 +00:00
|
|
|
|
1996-05-02 14:21:14 +00:00
|
|
|
/*
|
|
|
|
* Page Protection Exception bits
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define PGEX_P 0x01 /* Protection violation vs. not present */
|
|
|
|
#define PGEX_W 0x02 /* during a Write cycle */
|
|
|
|
#define PGEX_U 0x04 /* access from User mode (UPL) */
|
|
|
|
|
|
|
|
/*
|
2003-05-01 01:05:25 +00:00
|
|
|
* Pte related macros. This is complicated by having to deal with
|
|
|
|
* the sign extension of the 48th bit.
|
1996-05-02 14:21:14 +00:00
|
|
|
*/
|
2003-05-01 01:05:25 +00:00
|
|
|
#define VADDR_SIGN(l4) \
|
|
|
|
((l4) >= NPML4EPG/2 ? ((unsigned long)-1 << 47) : 0ul)
|
|
|
|
#define VADDR(l4, l3, l2, l1) ( \
|
|
|
|
((unsigned long)(l4) << PML4SHIFT) | VADDR_SIGN(l4) | \
|
|
|
|
((unsigned long)(l3) << PDPSHIFT) | \
|
|
|
|
((unsigned long)(l2) << PDRSHIFT) | \
|
|
|
|
((unsigned long)(l1) << PAGE_SHIFT))
|
|
|
|
|
1993-06-12 14:58:17 +00:00
|
|
|
|
1994-01-14 16:25:31 +00:00
|
|
|
#ifndef NKPT
|
2003-05-01 01:05:25 +00:00
|
|
|
#define NKPT 120 /* initial number of kernel page tables */
|
2001-09-21 06:23:03 +00:00
|
|
|
#endif
|
2003-05-23 05:04:54 +00:00
|
|
|
|
|
|
|
#define NKPML4E 1 /* number of kernel PML4 slots */
|
|
|
|
#define NKPDPE 1 /* number of kernel PDP slots */
|
|
|
|
#define NKPDE (NKPDPE*NPDEPG) /* number of kernel PD slots */
|
|
|
|
|
|
|
|
#define NUPML4E 1 /* number of userland PML4 pages */
|
|
|
|
#define NUPDPE (NUPML4E*NPDPEPG)/* number of userland PDP pages */
|
|
|
|
#define NUPDE (NUPDPE*NPDEPG) /* number of userland PD entries */
|
|
|
|
|
|
|
|
#define NDMPML4E 1 /* number of dmap PML4 slots */
|
1994-01-14 16:25:31 +00:00
|
|
|
|
1993-10-12 13:58:01 +00:00
|
|
|
/*
|
2003-05-23 05:04:54 +00:00
|
|
|
* The *PDI values control the layout of virtual memory
|
1993-10-12 13:58:01 +00:00
|
|
|
*/
|
2003-05-23 05:04:54 +00:00
|
|
|
#define PML4PML4I (NPML4EPG/2) /* Index of recursive pml4 mapping */
|
|
|
|
|
2003-05-23 06:35:45 +00:00
|
|
|
#define KPML4I (NPML4EPG-1) /* Top 512GB for KVM */
|
|
|
|
#define DMPML4I (KPML4I-1) /* Next 512GB down for direct map */
|
2003-05-23 05:04:54 +00:00
|
|
|
|
2003-05-23 06:35:45 +00:00
|
|
|
#define KPDPI (NPDPEPG-1) /* kernbase at -1GB */
|
First steps in rewriting locore.s, and making info useful
when the machine panics.
i386/i386/locore.s:
1) got rid of most .set directives that were being used like
#define's, and replaced them with appropriate #define's in
the appropriate header files (accessed via genassym).
2) added comments to header inclusions and global definitions,
and global variables
3) replaced some hardcoded constants with cpp defines (such as
PDESIZE and others)
4) aligned all comments to the same column to make them easier to
read
5) moved macro definitions for ENTRY, ALIGN, NOP, etc. to
/sys/i386/include/asmacros.h
6) added #ifdef BDE_DEBUGGER around all of Bruce's debugger code
7) added new global '_KERNend' to store last location+1 of kernel
8) cleaned up zeroing of bss so that only bss is zeroed
9) fix zeroing of page tables so that it really does zero them all
- not just if they follow the bss.
10) rewrote page table initialization code so that 1) works correctly
and 2) write protects the kernel text by default
11) properly initialize the kernel page directory, upages, p0stack PT,
and page tables. The previous scheme was more than a bit
screwy.
12) change allocation of virtual area of IO hole so that it is
fixed at KERNBASE + 0xa0000. The previous scheme put it
right after the kernel page tables and then later expected
it to be at KERNBASE +0xa0000
13) change multiple bogus settings of user read/write of various
areas of kernel VM - including the IO hole; we should never
be accessing the IO hole in user mode through the kernel
page tables
14) split kernel support routines such as bcopy, bzero, copyin,
copyout, etc. into a seperate file 'support.s'
15) split swtch and related routines into a seperate 'swtch.s'
16) split routines related to traps, syscalls, and interrupts
into a seperate file 'exception.s'
17) remove some unused global variables from locore that got
inserted by Garrett when he pulled them out of some .h
files.
i386/isa/icu.s:
1) clean up global variable declarations
2) move in declaration of astpending and netisr
i386/i386/pmap.c:
1) fix calculation of virtual_avail. It previously was calculated
to be right in the middle of the kernel page tables - not
a good place to start allocating kernel VM.
2) properly allocate kernel page dir/tables etc out of kernel map
- previously only took out 2 pages.
i386/i386/machdep.c:
1) modify boot() to print a warning that the system will reboot in
PANIC_REBOOT_WAIT_TIME amount of seconds, and let the user
abort with a key on the console. The machine will wait for
ever if a key is typed before the reboot. The default is
15 seconds, but can be set to 0 to mean don't wait at all,
-1 to mean wait forever, or any positive value to wait for
that many seconds.
2) print "Rebooting..." just before doing it.
kern/subr_prf.c:
1) remove PANICWAIT as it is deprecated by the change to machdep.c
i386/i386/trap.c:
1) add table of trap type strings and use it to print a real trap/
panic message rather than just a number. Lot's of work to
be done here, but this is the first step. Symbolic traceback
is in the TODO.
i386/i386/Makefile.i386:
1) add support in to build support.s, exception.s and swtch.s
...and various changes to various header files to make all of the
above happen.
1993-11-13 02:25:21 +00:00
|
|
|
|
1996-05-02 22:25:18 +00:00
|
|
|
/*
|
|
|
|
* XXX doesn't really belong here I guess...
|
|
|
|
*/
|
|
|
|
#define ISA_HOLE_START 0xa0000
|
|
|
|
#define ISA_HOLE_LENGTH (0x100000-ISA_HOLE_START)
|
|
|
|
|
1996-05-02 14:21:14 +00:00
|
|
|
#ifndef LOCORE
|
1996-09-08 16:57:53 +00:00
|
|
|
|
|
|
|
#include <sys/queue.h>
|
|
|
|
|
2003-05-01 01:05:25 +00:00
|
|
|
typedef u_int64_t pd_entry_t;
|
|
|
|
typedef u_int64_t pt_entry_t;
|
|
|
|
typedef u_int64_t pdp_entry_t;
|
|
|
|
typedef u_int64_t pml4_entry_t;
|
2003-03-30 05:24:52 +00:00
|
|
|
|
2003-05-01 01:05:25 +00:00
|
|
|
#define PML4ESHIFT (3)
|
|
|
|
#define PDPESHIFT (3)
|
2003-03-30 05:24:52 +00:00
|
|
|
#define PTESHIFT (3)
|
|
|
|
#define PDESHIFT (3)
|
|
|
|
|
1993-06-12 14:58:17 +00:00
|
|
|
/*
|
|
|
|
* Address of current and alternate address space page table maps
|
|
|
|
* and directories.
|
2003-05-01 01:05:25 +00:00
|
|
|
* XXX it might be saner to just direct map all of physical memory
|
|
|
|
* into the kernel using 2MB pages. We have enough space to do
|
|
|
|
* it (2^47 bits of KVM, while current max physical addressability
|
|
|
|
* is 2^40 physical bits). Then we can get rid of the evil hole
|
|
|
|
* in the page tables and the evil overlapping.
|
1993-06-12 14:58:17 +00:00
|
|
|
*/
|
1999-12-29 04:46:21 +00:00
|
|
|
#ifdef _KERNEL
|
2003-05-23 05:04:54 +00:00
|
|
|
#define addr_PTmap (VADDR(PML4PML4I, 0, 0, 0))
|
|
|
|
#define addr_PDmap (VADDR(PML4PML4I, PML4PML4I, 0, 0))
|
|
|
|
#define addr_PDPmap (VADDR(PML4PML4I, PML4PML4I, PML4PML4I, 0))
|
|
|
|
#define addr_PML4map (VADDR(PML4PML4I, PML4PML4I, PML4PML4I, PML4PML4I))
|
|
|
|
#define addr_PML4pml4e (addr_PML4map + (PML4PML4I * sizeof(pml4_entry_t)))
|
|
|
|
#define PTmap ((pt_entry_t *)(addr_PTmap))
|
|
|
|
#define PDmap ((pd_entry_t *)(addr_PDmap))
|
|
|
|
#define PDPmap ((pd_entry_t *)(addr_PDPmap))
|
|
|
|
#define PML4map ((pd_entry_t *)(addr_PML4map))
|
|
|
|
#define PML4pml4e ((pd_entry_t *)(addr_PML4pml4e))
|
|
|
|
|
|
|
|
extern u_int64_t KPML4phys; /* physical address of kernel level 4 */
|
1993-06-12 14:58:17 +00:00
|
|
|
#endif
|
|
|
|
|
1999-12-29 04:46:21 +00:00
|
|
|
#ifdef _KERNEL
|
1993-06-12 14:58:17 +00:00
|
|
|
/*
|
|
|
|
* virtual address to page table entry and
|
|
|
|
* to physical address. Likewise for alternate address space.
|
|
|
|
* Note: these work recursively, thus vtopte of a pte will give
|
|
|
|
* the corresponding pde that in turn maps it.
|
|
|
|
*/
|
2003-05-23 05:04:54 +00:00
|
|
|
pt_entry_t *vtopte(vm_offset_t);
|
|
|
|
vm_paddr_t pmap_kextract(vm_offset_t);
|
1997-07-17 04:34:03 +00:00
|
|
|
|
|
|
|
#define vtophys(va) pmap_kextract(((vm_offset_t) (va)))
|
2003-03-30 05:24:52 +00:00
|
|
|
|
2003-04-28 20:35:36 +00:00
|
|
|
static __inline pt_entry_t
|
|
|
|
pte_load(pt_entry_t *ptep)
|
|
|
|
{
|
|
|
|
pt_entry_t r;
|
|
|
|
|
|
|
|
r = *ptep;
|
|
|
|
return (r);
|
|
|
|
}
|
|
|
|
|
|
|
|
static __inline pt_entry_t
|
|
|
|
pte_load_store(pt_entry_t *ptep, pt_entry_t pte)
|
|
|
|
{
|
|
|
|
pt_entry_t r;
|
|
|
|
|
|
|
|
r = *ptep;
|
|
|
|
*ptep = pte;
|
|
|
|
return (r);
|
|
|
|
}
|
2003-03-30 05:24:52 +00:00
|
|
|
|
2003-05-01 01:05:25 +00:00
|
|
|
#define pte_load_clear(pte) atomic_readandclear_long(pte)
|
2003-03-30 05:24:52 +00:00
|
|
|
|
2003-04-28 20:35:36 +00:00
|
|
|
#define pte_clear(ptep) pte_load_store((ptep), (pt_entry_t)0ULL)
|
|
|
|
#define pte_store(ptep, pte) pte_load_store((ptep), (pt_entry_t)pte)
|
|
|
|
|
|
|
|
#define pde_store(pdep, pde) pte_store((pdep), (pde))
|
|
|
|
|
|
|
|
#endif /* _KERNEL */
|
1997-07-17 04:34:03 +00:00
|
|
|
|
1993-06-12 14:58:17 +00:00
|
|
|
/*
|
|
|
|
* Pmap stuff
|
|
|
|
*/
|
1996-09-08 16:57:53 +00:00
|
|
|
struct pv_entry;
|
2000-05-21 12:50:18 +00:00
|
|
|
|
|
|
|
struct md_page {
|
1996-09-08 16:57:53 +00:00
|
|
|
int pv_list_count;
|
2000-05-26 02:09:24 +00:00
|
|
|
TAILQ_HEAD(,pv_entry) pv_list;
|
2000-05-21 12:50:18 +00:00
|
|
|
};
|
1993-06-12 14:58:17 +00:00
|
|
|
|
|
|
|
struct pmap {
|
2003-05-23 05:04:54 +00:00
|
|
|
pml4_entry_t *pm_pml4; /* KVA of level 4 page table */
|
1996-05-18 03:38:05 +00:00
|
|
|
vm_object_t pm_pteobj; /* Container for pte's */
|
2000-05-26 02:09:24 +00:00
|
|
|
TAILQ_HEAD(,pv_entry) pm_pvlist; /* list of mappings in pmap */
|
2003-05-01 01:05:25 +00:00
|
|
|
u_long pm_active; /* active on cpus */
|
1993-06-12 14:58:17 +00:00
|
|
|
struct pmap_statistics pm_stats; /* pmap statistics */
|
2000-08-16 21:24:44 +00:00
|
|
|
LIST_ENTRY(pmap) pm_list; /* List of all pmaps */
|
1993-06-12 14:58:17 +00:00
|
|
|
};
|
|
|
|
|
2002-08-05 03:40:28 +00:00
|
|
|
#define pmap_page_is_mapped(m) (!TAILQ_EMPTY(&(m)->md.pv_list))
|
1996-10-13 03:14:57 +00:00
|
|
|
#define pmap_resident_count(pmap) (pmap)->pm_stats.resident_count
|
|
|
|
|
1993-06-12 14:58:17 +00:00
|
|
|
typedef struct pmap *pmap_t;
|
|
|
|
|
1999-12-29 04:46:21 +00:00
|
|
|
#ifdef _KERNEL
|
2002-04-29 07:43:16 +00:00
|
|
|
extern struct pmap kernel_pmap_store;
|
|
|
|
#define kernel_pmap (&kernel_pmap_store)
|
1993-06-12 14:58:17 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
/*
|
|
|
|
* For each vm_page_t, there is a list of all currently valid virtual
|
|
|
|
* mappings of that page. An entry is a pv_entry_t, the list is pv_table.
|
|
|
|
*/
|
|
|
|
typedef struct pv_entry {
|
|
|
|
pmap_t pv_pmap; /* pmap where mapping lies */
|
|
|
|
vm_offset_t pv_va; /* virtual address for mapping */
|
2000-05-26 02:09:24 +00:00
|
|
|
TAILQ_ENTRY(pv_entry) pv_list;
|
|
|
|
TAILQ_ENTRY(pv_entry) pv_plist;
|
1996-02-25 03:02:53 +00:00
|
|
|
vm_page_t pv_ptem; /* VM page for pte */
|
1993-06-12 14:58:17 +00:00
|
|
|
} *pv_entry_t;
|
|
|
|
|
1999-12-29 04:46:21 +00:00
|
|
|
#ifdef _KERNEL
|
1993-06-12 14:58:17 +00:00
|
|
|
|
1998-05-11 01:06:08 +00:00
|
|
|
#define NPPROVMTRR 8
|
|
|
|
#define PPRO_VMTRRphysBase0 0x200
|
|
|
|
#define PPRO_VMTRRphysMask0 0x201
|
1998-11-24 20:25:52 +00:00
|
|
|
struct ppro_vmtrr {
|
1998-05-11 01:06:08 +00:00
|
|
|
u_int64_t base, mask;
|
1998-11-24 20:25:52 +00:00
|
|
|
};
|
|
|
|
extern struct ppro_vmtrr PPro_vmtrr[NPPROVMTRR];
|
1998-05-11 01:06:08 +00:00
|
|
|
|
1995-03-16 18:17:34 +00:00
|
|
|
extern caddr_t CADDR1;
|
|
|
|
extern pt_entry_t *CMAP1;
|
2003-03-25 00:07:06 +00:00
|
|
|
extern vm_paddr_t avail_end;
|
|
|
|
extern vm_paddr_t avail_start;
|
1997-11-20 19:30:35 +00:00
|
|
|
extern vm_offset_t clean_eva;
|
|
|
|
extern vm_offset_t clean_sva;
|
2003-03-25 00:07:06 +00:00
|
|
|
extern vm_paddr_t phys_avail[];
|
1997-11-20 19:30:35 +00:00
|
|
|
extern char *ptvmmap; /* poor name! */
|
1995-03-16 18:17:34 +00:00
|
|
|
extern vm_offset_t virtual_avail;
|
|
|
|
extern vm_offset_t virtual_end;
|
1993-06-12 14:58:17 +00:00
|
|
|
|
2003-05-23 05:04:54 +00:00
|
|
|
void pmap_bootstrap(vm_paddr_t *);
|
2003-03-25 00:07:06 +00:00
|
|
|
void pmap_kenter(vm_offset_t va, vm_paddr_t pa);
|
2003-03-16 04:16:03 +00:00
|
|
|
void pmap_kremove(vm_offset_t);
|
2003-03-25 00:07:06 +00:00
|
|
|
void *pmap_mapdev(vm_paddr_t, vm_size_t);
|
2002-03-20 05:48:58 +00:00
|
|
|
void pmap_unmapdev(vm_offset_t, vm_size_t);
|
2003-03-30 01:16:19 +00:00
|
|
|
pt_entry_t *pmap_pte_quick(pmap_t, vm_offset_t) __pure2;
|
2002-07-12 07:56:11 +00:00
|
|
|
void pmap_invalidate_page(pmap_t, vm_offset_t);
|
|
|
|
void pmap_invalidate_range(pmap_t, vm_offset_t, vm_offset_t);
|
|
|
|
void pmap_invalidate_all(pmap_t);
|
1993-12-19 00:55:01 +00:00
|
|
|
|
1999-12-29 04:46:21 +00:00
|
|
|
#endif /* _KERNEL */
|
1996-10-12 20:36:15 +00:00
|
|
|
|
1996-05-02 14:21:14 +00:00
|
|
|
#endif /* !LOCORE */
|
1993-06-12 14:58:17 +00:00
|
|
|
|
1994-11-14 14:12:24 +00:00
|
|
|
#endif /* !_MACHINE_PMAP_H_ */
|