As part of a general cleanup and reconvergence of related pmap code,

start tidying up some loose ends.  The DEBUG_VA stuff has long since
passed its use-by date.  It wasn't used on ia64 but got cut/pasted there.
This commit is contained in:
Peter Wemm 2001-11-16 01:56:34 +00:00
parent 0721252ad2
commit 024c3fd9d5
2 changed files with 1 additions and 46 deletions

View File

@ -197,21 +197,6 @@
#define PMAP_INLINE
#endif
#if 0
static void
pmap_break(void)
{
}
/* #define PMAP_DEBUG_VA(va) if ((va) == 0x120058000) pmap_break(); else */
#endif
#ifndef PMAP_DEBUG_VA
#define PMAP_DEBUG_VA(va) do {} while(0)
#endif
/*
* Some macros for manipulating virtual addresses
*/
@ -824,7 +809,6 @@ pmap_qenter(vm_offset_t va, vm_page_t *m, int count)
pt_entry_t opte;
pte = vtopte(tva);
opte = *pte;
PMAP_DEBUG_VA(va);
*pte = npte;
if (opte)
pmap_invalidate_page(kernel_pmap, tva);
@ -845,7 +829,6 @@ pmap_qremove(va, count)
for (i = 0; i < count; i++) {
pte = vtopte(va);
PMAP_DEBUG_VA(va);
*pte = 0;
pmap_invalidate_page(kernel_pmap, va);
va += PAGE_SIZE;
@ -866,7 +849,6 @@ pmap_kenter(vm_offset_t va, vm_offset_t pa)
npte = pmap_phys_to_pte(pa) | PG_ASM | PG_KRE | PG_KWE | PG_V;
pte = vtopte(va);
opte = *pte;
PMAP_DEBUG_VA(va);
*pte = npte;
if (opte)
pmap_invalidate_page(kernel_pmap, va);
@ -881,7 +863,6 @@ pmap_kremove(vm_offset_t va)
register pt_entry_t *pte;
pte = vtopte(va);
PMAP_DEBUG_VA(va);
*pte = 0;
pmap_invalidate_page(kernel_pmap, va);
}
@ -1942,7 +1923,6 @@ pmap_remove_pte(pmap_t pmap, pt_entry_t* ptq, vm_offset_t va)
vm_page_t m;
oldpte = *ptq;
PMAP_DEBUG_VA(va);
*ptq = 0;
if (oldpte & PG_W)
pmap->pm_stats.wired_count -= 1;
@ -2075,7 +2055,6 @@ pmap_remove_all(vm_page_t m)
tpte = *pte;
PMAP_DEBUG_VA(pv->pv_va);
*pte = 0;
if (tpte & PG_W)
pv->pv_pmap->pm_stats.wired_count--;
@ -2328,7 +2307,6 @@ validate:
* to update the pte.
*/
if (origpte != newpte) {
PMAP_DEBUG_VA(va);
*pte = newpte;
if (origpte)
pmap_invalidate_page(pmap, va);
@ -2417,7 +2395,6 @@ retry:
* raise IPL while manipulating pv_table since pmap_enter can be
* called at interrupt time.
*/
PMAP_DEBUG_VA(va);
pmap_insert_entry(pmap, va, mpte, m);
/*
@ -2847,7 +2824,6 @@ pmap_remove_pages(pmap, sva, eva)
npv = TAILQ_NEXT(pv, pv_plist);
continue;
}
PMAP_DEBUG_VA(pv->pv_va);
*pte = 0;
m = PHYS_TO_VM_PAGE(pmap_pte_pa(&tpte));

View File

@ -145,21 +145,6 @@ MALLOC_DEFINE(M_PMAP, "PMAP", "PMAP Structures");
#define PMAP_INLINE
#endif
#if 0
static void
pmap_break(void)
{
}
/* #define PMAP_DEBUG_VA(va) if ((va) == 0x120058000) pmap_break(); else */
#endif
#ifndef PMAP_DEBUG_VA
#define PMAP_DEBUG_VA(va) do {} while(0)
#endif
/*
* Get PDEs and PTEs for user/kernel address space
*/
@ -1376,7 +1361,6 @@ pmap_qenter(vm_offset_t va, vm_page_t *m, int count)
for (i = 0; i < count; i++) {
vm_offset_t tva = va + i * PAGE_SIZE;
int wasvalid;
PMAP_DEBUG_VA(tva);
pte = pmap_find_kpte(tva);
wasvalid = pte->pte_p;
pmap_set_pte(pte, tva, VM_PAGE_TO_PHYS(m[i]),
@ -1397,7 +1381,6 @@ pmap_qremove(vm_offset_t va, int count)
struct ia64_lpte *pte;
for (i = 0; i < count; i++) {
PMAP_DEBUG_VA(va);
pte = pmap_find_kpte(va);
pmap_clear_pte(pte, va);
va += PAGE_SIZE;
@ -1781,10 +1764,8 @@ validate:
* if the mapping or permission bits are different, we need
* to invalidate the page.
*/
if (!pmap_equal_pte(&origpte, pte)) {
PMAP_DEBUG_VA(va);
if (!pmap_equal_pte(&origpte, pte))
pmap_invalidate_page(pmap, va);
}
pmap_install(oldpmap);
}
@ -1814,8 +1795,6 @@ pmap_enter_quick(pmap_t pmap, vm_offset_t va, vm_page_t m)
if (pte->pte_p)
return;
PMAP_DEBUG_VA(va);
/*
* Enter on the PV list since its part of our managed memory.
*/