b10ed4a911
to this pmap.c. This new r/w lock is used primarily to synchronize access to the TTE lists. However, it will be used in a somewhat unconventional way. As finer-grained TTE list locking is added to each of the pmap functions that acquire this r/w lock, its acquisition will be changed from write to read, enabling concurrent execution of the pmap functions with finer-grained locking. Reviewed by: attilio Tested by: flo MFC after: 10 days
230 lines
6.0 KiB
C
230 lines
6.0 KiB
C
/*-
|
|
* Copyright (c) 1997 Berkeley Software Design, Inc. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. Berkeley Software Design Inc's name may not be used to endorse or
|
|
* promote products derived from this software without specific prior
|
|
* written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY BERKELEY SOFTWARE DESIGN INC ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL BERKELEY SOFTWARE DESIGN INC BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* from BSDI: pmap.c,v 1.28.2.15 2000/04/27 03:10:31 cp Exp
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
#include "opt_ddb.h"
|
|
#include "opt_pmap.h"
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/queue.h>
|
|
#include <sys/ktr.h>
|
|
#include <sys/lock.h>
|
|
#include <sys/mutex.h>
|
|
#include <sys/proc.h>
|
|
#include <sys/rwlock.h>
|
|
#include <sys/smp.h>
|
|
#include <sys/sysctl.h>
|
|
#include <sys/systm.h>
|
|
|
|
#include <vm/vm.h>
|
|
#include <vm/vm_param.h>
|
|
#include <vm/vm_kern.h>
|
|
#include <vm/vm_page.h>
|
|
#include <vm/vm_map.h>
|
|
#include <vm/vm_object.h>
|
|
#include <vm/vm_extern.h>
|
|
#include <vm/vm_pageout.h>
|
|
#include <vm/vm_pager.h>
|
|
|
|
#include <machine/cpufunc.h>
|
|
#include <machine/frame.h>
|
|
#include <machine/trap.h>
|
|
#include <machine/pmap.h>
|
|
#include <machine/smp.h>
|
|
#include <machine/tlb.h>
|
|
#include <machine/tsb.h>
|
|
#include <machine/tte.h>
|
|
|
|
CTASSERT((1 << TTE_SHIFT) == sizeof(struct tte));
|
|
CTASSERT(TSB_BUCKET_MASK < (1 << 12));
|
|
|
|
PMAP_STATS_VAR(tsb_nrepl);
|
|
PMAP_STATS_VAR(tsb_nlookup_k);
|
|
PMAP_STATS_VAR(tsb_nlookup_u);
|
|
PMAP_STATS_VAR(tsb_nenter_k);
|
|
PMAP_STATS_VAR(tsb_nenter_k_oc);
|
|
PMAP_STATS_VAR(tsb_nenter_u);
|
|
PMAP_STATS_VAR(tsb_nenter_u_oc);
|
|
PMAP_STATS_VAR(tsb_nforeach);
|
|
|
|
struct tte *tsb_kernel;
|
|
vm_size_t tsb_kernel_mask;
|
|
vm_size_t tsb_kernel_size;
|
|
vm_paddr_t tsb_kernel_phys;
|
|
u_int tsb_kernel_ldd_phys;
|
|
|
|
struct tte *
|
|
tsb_tte_lookup(pmap_t pm, vm_offset_t va)
|
|
{
|
|
struct tte *bucket;
|
|
struct tte *tp;
|
|
u_long sz;
|
|
u_int i;
|
|
|
|
if (pm == kernel_pmap) {
|
|
PMAP_STATS_INC(tsb_nlookup_k);
|
|
tp = tsb_kvtotte(va);
|
|
if (tte_match(tp, va))
|
|
return (tp);
|
|
} else {
|
|
PMAP_LOCK_ASSERT(pm, MA_OWNED);
|
|
PMAP_STATS_INC(tsb_nlookup_u);
|
|
for (sz = TS_MIN; sz <= TS_MAX; sz++) {
|
|
bucket = tsb_vtobucket(pm, sz, va);
|
|
for (i = 0; i < TSB_BUCKET_SIZE; i++) {
|
|
tp = &bucket[i];
|
|
if (tte_match(tp, va))
|
|
return (tp);
|
|
}
|
|
}
|
|
}
|
|
return (NULL);
|
|
}
|
|
|
|
struct tte *
|
|
tsb_tte_enter(pmap_t pm, vm_page_t m, vm_offset_t va, u_long sz, u_long data)
|
|
{
|
|
struct tte *bucket;
|
|
struct tte *rtp;
|
|
struct tte *tp;
|
|
vm_offset_t ova;
|
|
int b0;
|
|
int i;
|
|
|
|
if (DCACHE_COLOR(VM_PAGE_TO_PHYS(m)) != DCACHE_COLOR(va)) {
|
|
CTR5(KTR_SPARE2,
|
|
"tsb_tte_enter: off colour va=%#lx pa=%#lx o=%p ot=%d pi=%#lx",
|
|
va, VM_PAGE_TO_PHYS(m), m->object,
|
|
m->object ? m->object->type : -1,
|
|
m->pindex);
|
|
if (pm == kernel_pmap)
|
|
PMAP_STATS_INC(tsb_nenter_k_oc);
|
|
else
|
|
PMAP_STATS_INC(tsb_nenter_u_oc);
|
|
}
|
|
|
|
rw_assert(&tte_list_global_lock, RA_WLOCKED);
|
|
PMAP_LOCK_ASSERT(pm, MA_OWNED);
|
|
if (pm == kernel_pmap) {
|
|
PMAP_STATS_INC(tsb_nenter_k);
|
|
tp = tsb_kvtotte(va);
|
|
KASSERT((tp->tte_data & TD_V) == 0,
|
|
("tsb_tte_enter: replacing valid kernel mapping"));
|
|
goto enter;
|
|
}
|
|
PMAP_STATS_INC(tsb_nenter_u);
|
|
|
|
bucket = tsb_vtobucket(pm, sz, va);
|
|
|
|
tp = NULL;
|
|
rtp = NULL;
|
|
b0 = rd(tick) & (TSB_BUCKET_SIZE - 1);
|
|
i = b0;
|
|
do {
|
|
if ((bucket[i].tte_data & TD_V) == 0) {
|
|
tp = &bucket[i];
|
|
break;
|
|
}
|
|
if (tp == NULL) {
|
|
if ((bucket[i].tte_data & TD_REF) == 0)
|
|
tp = &bucket[i];
|
|
else if (rtp == NULL)
|
|
rtp = &bucket[i];
|
|
}
|
|
} while ((i = (i + 1) & (TSB_BUCKET_SIZE - 1)) != b0);
|
|
|
|
if (tp == NULL)
|
|
tp = rtp;
|
|
if ((tp->tte_data & TD_V) != 0) {
|
|
PMAP_STATS_INC(tsb_nrepl);
|
|
ova = TTE_GET_VA(tp);
|
|
pmap_remove_tte(pm, NULL, tp, ova);
|
|
tlb_page_demap(pm, ova);
|
|
}
|
|
|
|
enter:
|
|
if ((m->flags & PG_FICTITIOUS) == 0) {
|
|
data |= TD_CP;
|
|
if ((m->oflags & VPO_UNMANAGED) == 0) {
|
|
pm->pm_stats.resident_count++;
|
|
data |= TD_PV;
|
|
}
|
|
if (pmap_cache_enter(m, va) != 0)
|
|
data |= TD_CV;
|
|
TAILQ_INSERT_TAIL(&m->md.tte_list, tp, tte_link);
|
|
} else
|
|
data |= TD_FAKE | TD_E;
|
|
|
|
tp->tte_vpn = TV_VPN(va, sz);
|
|
tp->tte_data = data;
|
|
|
|
return (tp);
|
|
}
|
|
|
|
/*
|
|
* Traverse the tsb of a pmap, calling the callback function for any tte entry
|
|
* that has a virtual address between start and end. If this function returns 0,
|
|
* tsb_foreach() terminates.
|
|
* This is used by pmap_remove(), pmap_protect(), and pmap_copy() in the case
|
|
* that the number of pages in the range given to them reaches the
|
|
* dimensions of the tsb size as an optimization.
|
|
*/
|
|
void
|
|
tsb_foreach(pmap_t pm1, pmap_t pm2, vm_offset_t start, vm_offset_t end,
|
|
tsb_callback_t *callback)
|
|
{
|
|
vm_offset_t va;
|
|
struct tte *tp;
|
|
struct tte *tsbp;
|
|
uintptr_t i;
|
|
uintptr_t n;
|
|
|
|
PMAP_STATS_INC(tsb_nforeach);
|
|
if (pm1 == kernel_pmap) {
|
|
tsbp = tsb_kernel;
|
|
n = tsb_kernel_size / sizeof(struct tte);
|
|
} else {
|
|
tsbp = pm1->pm_tsb;
|
|
n = TSB_SIZE;
|
|
}
|
|
for (i = 0; i < n; i++) {
|
|
tp = &tsbp[i];
|
|
if ((tp->tte_data & TD_V) != 0) {
|
|
va = TTE_GET_VA(tp);
|
|
if (va >= start && va < end) {
|
|
if (!callback(pm1, pm2, tp, va))
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|