powerpc: Re-merge isa3 HPT with moea64 native HPT

r345402 fixed the bug that led to the split of the ISA 3.0 HPT handling from
the existing manager.  The cause of the bug was gcc moving the register
holding VPN to a different register (not r0), which triggered bizarre
behaviors.  With the fix, things work, so they can be re-merged.  No
performance lost with the merge.
This commit is contained in:
Justin Hibbits 2019-03-22 22:14:14 +00:00
parent ac1a10efad
commit bc94b70098
Notes: svn2git 2020-12-20 02:59:44 +00:00
svn path=/head/; revision=345426
4 changed files with 51 additions and 646 deletions

View File

@ -103,7 +103,6 @@ libkern/qdivrem.c optional powerpc | powerpcspe
libkern/ucmpdi2.c optional powerpc | powerpcspe
libkern/udivdi3.c optional powerpc | powerpcspe
libkern/umoddi3.c optional powerpc | powerpcspe
powerpc/aim/isa3_hashtb.c optional aim powerpc64
powerpc/aim/locore.S optional aim no-obj
powerpc/aim/aim_machdep.c optional aim
powerpc/aim/mmu_oea.c optional aim powerpc

View File

@ -421,9 +421,7 @@ aim_cpu_init(vm_offset_t toc)
* in case the platform module had a better idea of what we
* should do.
*/
if (cpu_features2 & PPC_FEATURE2_ARCH_3_00)
pmap_mmu_install(MMU_TYPE_P9H, BUS_PROBE_GENERIC);
else if (cpu_features & PPC_FEATURE_64)
if (cpu_features & PPC_FEATURE_64)
pmap_mmu_install(MMU_TYPE_G5, BUS_PROBE_GENERIC);
else
pmap_mmu_install(MMU_TYPE_OEA, BUS_PROBE_GENERIC);

View File

@ -1,632 +0,0 @@
/*-
* SPDX-License-Identifier: BSD-2-Clause-FreeBSD AND 4-Clause-BSD
*
* Copyright (c) 2001 The NetBSD Foundation, Inc.
* All rights reserved.
*
* This code is derived from software contributed to The NetBSD Foundation
* by Matt Thomas <matt@3am-software.com> of Allegro Networks, Inc.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/
/*-
* Copyright (C) 1995, 1996 Wolfgang Solfrank.
* Copyright (C) 1995, 1996 TooLs GmbH.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. All advertising materials mentioning features or use of this software
* must display the following acknowledgement:
* This product includes software developed by TooLs GmbH.
* 4. The name of TooLs GmbH may not be used to endorse or promote products
* derived from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY TOOLS GMBH ``AS IS'' AND ANY EXPRESS OR
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
* IN NO EVENT SHALL TOOLS GMBH BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
* OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
* WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
* OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
* ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*
* $NetBSD: pmap.c,v 1.28 2000/03/26 20:42:36 kleink Exp $
*/
/*-
* Copyright (C) 2001 Benno Rice.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY Benno Rice ``AS IS'' AND ANY EXPRESS OR
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
* IN NO EVENT SHALL TOOLS GMBH BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
* OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
* WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
* OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
* ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include <sys/cdefs.h>
__FBSDID("$FreeBSD$");
/*
* Native 64-bit page table operations for running without a hypervisor.
*/
#include <sys/param.h>
#include <sys/kernel.h>
#include <sys/ktr.h>
#include <sys/lock.h>
#include <sys/mutex.h>
#include <sys/proc.h>
#include <sys/sched.h>
#include <sys/sysctl.h>
#include <sys/systm.h>
#include <sys/rwlock.h>
#include <sys/endian.h>
#include <sys/kdb.h>
#include <vm/vm.h>
#include <vm/vm_param.h>
#include <vm/vm_kern.h>
#include <vm/vm_page.h>
#include <vm/vm_map.h>
#include <vm/vm_object.h>
#include <vm/vm_extern.h>
#include <vm/vm_pageout.h>
#include <machine/cpu.h>
#include <machine/hid.h>
#include <machine/md_var.h>
#include <machine/mmuvar.h>
#include "mmu_oea64.h"
#include "mmu_if.h"
#include "moea64_if.h"
#define PTESYNC() __asm __volatile("ptesync");
#define TLBSYNC() __asm __volatile("tlbsync; ptesync");
#define SYNC() __asm __volatile("sync");
#define EIEIO() __asm __volatile("eieio");
#define VSID_HASH_MASK 0x0000007fffffffffULL
/* POWER9 only permits a 64k partition table size. */
#define PART_SIZE 0x10000
static __inline void
TLBIE(uint64_t vpn)
{
vpn <<= ADDR_PIDX_SHFT;
__asm __volatile(".long 0x7c000264 | (%0 << 11) | (%1 << 21)"
:: "r"(vpn), "r"(0) : "memory");
__asm __volatile("eieio; tlbsync; ptesync" ::: "memory");
}
#define DISABLE_TRANS(msr) msr = mfmsr(); mtmsr(msr & ~PSL_DR)
#define ENABLE_TRANS(msr) mtmsr(msr)
/*
* PTEG data.
*/
static volatile struct pate *isa3_part_table;
static volatile struct lpte *isa3_hashtb_pteg_table;
static struct rwlock isa3_hashtb_eviction_lock;
/*
* PTE calls.
*/
static int isa3_hashtb_pte_insert(mmu_t, struct pvo_entry *);
static int64_t isa3_hashtb_pte_synch(mmu_t, struct pvo_entry *);
static int64_t isa3_hashtb_pte_clear(mmu_t, struct pvo_entry *, uint64_t);
static int64_t isa3_hashtb_pte_replace(mmu_t, struct pvo_entry *, int);
static int64_t isa3_hashtb_pte_unset(mmu_t mmu, struct pvo_entry *);
/*
* Utility routines.
*/
static void isa3_hashtb_bootstrap(mmu_t mmup,
vm_offset_t kernelstart, vm_offset_t kernelend);
static void isa3_hashtb_cpu_bootstrap(mmu_t, int ap);
static void tlbia(void);
static mmu_method_t isa3_hashtb_methods[] = {
/* Internal interfaces */
MMUMETHOD(mmu_bootstrap, isa3_hashtb_bootstrap),
MMUMETHOD(mmu_cpu_bootstrap, isa3_hashtb_cpu_bootstrap),
MMUMETHOD(moea64_pte_synch, isa3_hashtb_pte_synch),
MMUMETHOD(moea64_pte_clear, isa3_hashtb_pte_clear),
MMUMETHOD(moea64_pte_unset, isa3_hashtb_pte_unset),
MMUMETHOD(moea64_pte_replace, isa3_hashtb_pte_replace),
MMUMETHOD(moea64_pte_insert, isa3_hashtb_pte_insert),
{ 0, 0 }
};
MMU_DEF_INHERIT(isa3_mmu_native, MMU_TYPE_P9H, isa3_hashtb_methods,
0, oea64_mmu);
static int64_t
isa3_hashtb_pte_synch(mmu_t mmu, struct pvo_entry *pvo)
{
volatile struct lpte *pt = isa3_hashtb_pteg_table + pvo->pvo_pte.slot;
struct lpte properpt;
uint64_t ptelo;
PMAP_LOCK_ASSERT(pvo->pvo_pmap, MA_OWNED);
moea64_pte_from_pvo(pvo, &properpt);
rw_rlock(&isa3_hashtb_eviction_lock);
if ((be64toh(pt->pte_hi) & LPTE_AVPN_MASK) !=
(properpt.pte_hi & LPTE_AVPN_MASK)) {
/* Evicted */
rw_runlock(&isa3_hashtb_eviction_lock);
return (-1);
}
PTESYNC();
ptelo = be64toh(pt->pte_lo);
rw_runlock(&isa3_hashtb_eviction_lock);
return (ptelo & (LPTE_REF | LPTE_CHG));
}
static int64_t
isa3_hashtb_pte_clear(mmu_t mmu, struct pvo_entry *pvo, uint64_t ptebit)
{
volatile struct lpte *pt = isa3_hashtb_pteg_table + pvo->pvo_pte.slot;
struct lpte properpt;
uint64_t ptelo;
PMAP_LOCK_ASSERT(pvo->pvo_pmap, MA_OWNED);
moea64_pte_from_pvo(pvo, &properpt);
rw_rlock(&isa3_hashtb_eviction_lock);
if ((be64toh(pt->pte_hi) & LPTE_AVPN_MASK) !=
(properpt.pte_hi & LPTE_AVPN_MASK)) {
/* Evicted */
rw_runlock(&isa3_hashtb_eviction_lock);
return (-1);
}
if (ptebit == LPTE_REF) {
/* See "Resetting the Reference Bit" in arch manual */
PTESYNC();
/* 2-step here safe: precision is not guaranteed */
ptelo = be64toh(pt->pte_lo);
/* One-byte store to avoid touching the C bit */
((volatile uint8_t *)(&pt->pte_lo))[6] =
#if BYTE_ORDER == BIG_ENDIAN
((uint8_t *)(&properpt.pte_lo))[6];
#else
((uint8_t *)(&properpt.pte_lo))[1];
#endif
rw_runlock(&isa3_hashtb_eviction_lock);
critical_enter();
TLBIE(pvo->pvo_vpn);
critical_exit();
} else {
rw_runlock(&isa3_hashtb_eviction_lock);
ptelo = isa3_hashtb_pte_unset(mmu, pvo);
isa3_hashtb_pte_insert(mmu, pvo);
}
return (ptelo & (LPTE_REF | LPTE_CHG));
}
static int64_t
isa3_hashtb_pte_unset(mmu_t mmu, struct pvo_entry *pvo)
{
volatile struct lpte *pt = isa3_hashtb_pteg_table + pvo->pvo_pte.slot;
struct lpte properpt;
uint64_t ptelo;
moea64_pte_from_pvo(pvo, &properpt);
rw_rlock(&isa3_hashtb_eviction_lock);
if ((be64toh(pt->pte_hi & LPTE_AVPN_MASK)) !=
(properpt.pte_hi & LPTE_AVPN_MASK)) {
/* Evicted */
moea64_pte_overflow--;
rw_runlock(&isa3_hashtb_eviction_lock);
return (-1);
}
/*
* Invalidate the pte, briefly locking it to collect RC bits. No
* atomics needed since this is protected against eviction by the lock.
*/
isync();
critical_enter();
pt->pte_hi = be64toh((pt->pte_hi & ~LPTE_VALID) | LPTE_LOCKED);
PTESYNC();
TLBIE(pvo->pvo_vpn);
ptelo = be64toh(pt->pte_lo);
*((volatile int32_t *)(&pt->pte_hi) + 1) = 0; /* Release lock */
critical_exit();
rw_runlock(&isa3_hashtb_eviction_lock);
/* Keep statistics */
moea64_pte_valid--;
return (ptelo & (LPTE_CHG | LPTE_REF));
}
static int64_t
isa3_hashtb_pte_replace(mmu_t mmu, struct pvo_entry *pvo, int flags)
{
volatile struct lpte *pt = isa3_hashtb_pteg_table + pvo->pvo_pte.slot;
struct lpte properpt;
int64_t ptelo;
if (flags == 0) {
/* Just some software bits changing. */
moea64_pte_from_pvo(pvo, &properpt);
rw_rlock(&isa3_hashtb_eviction_lock);
if ((be64toh(pt->pte_hi) & LPTE_AVPN_MASK) !=
(properpt.pte_hi & LPTE_AVPN_MASK)) {
rw_runlock(&isa3_hashtb_eviction_lock);
return (-1);
}
pt->pte_hi = htobe64(properpt.pte_hi);
ptelo = be64toh(pt->pte_lo);
rw_runlock(&isa3_hashtb_eviction_lock);
} else {
/* Otherwise, need reinsertion and deletion */
ptelo = isa3_hashtb_pte_unset(mmu, pvo);
isa3_hashtb_pte_insert(mmu, pvo);
}
return (ptelo);
}
static void
isa3_hashtb_cpu_bootstrap(mmu_t mmup, int ap)
{
int i = 0;
struct slb *slb = PCPU_GET(aim.slb);
register_t seg0;
/*
* Initialize segment registers and MMU
*/
mtmsr(mfmsr() & ~PSL_DR & ~PSL_IR);
switch (mfpvr() >> 16) {
case IBMPOWER9:
mtspr(SPR_HID0, mfspr(SPR_HID0) & ~HID0_RADIX);
break;
}
/*
* Install kernel SLB entries
*/
__asm __volatile ("slbia");
__asm __volatile ("slbmfee %0,%1; slbie %0;" : "=r"(seg0) :
"r"(0));
for (i = 0; i < n_slbs; i++) {
if (!(slb[i].slbe & SLBE_VALID))
continue;
__asm __volatile ("slbmte %0, %1" ::
"r"(slb[i].slbv), "r"(slb[i].slbe));
}
/*
* Install page table
*/
mtspr(SPR_PTCR,
((uintptr_t)isa3_part_table & ~DMAP_BASE_ADDRESS) |
flsl((PART_SIZE >> 12) - 1));
tlbia();
}
static void
isa3_hashtb_bootstrap(mmu_t mmup, vm_offset_t kernelstart,
vm_offset_t kernelend)
{
vm_size_t size;
vm_offset_t off;
vm_paddr_t pa;
register_t msr;
moea64_early_bootstrap(mmup, kernelstart, kernelend);
/*
* Allocate PTEG table.
*/
size = moea64_pteg_count * sizeof(struct lpteg);
CTR2(KTR_PMAP, "moea64_bootstrap: %d PTEGs, %lu bytes",
moea64_pteg_count, size);
rw_init(&isa3_hashtb_eviction_lock, "pte eviction");
/*
* We now need to allocate memory. This memory, to be allocated,
* has to reside in a page table. The page table we are about to
* allocate. We don't have BAT. So drop to data real mode for a minute
* as a measure of last resort. We do this a couple times.
*/
isa3_part_table =
(struct pate *)moea64_bootstrap_alloc(PART_SIZE, PART_SIZE);
if (hw_direct_map)
isa3_part_table = (struct pate *)PHYS_TO_DMAP(
(vm_offset_t)isa3_part_table);
/*
* PTEG table must be aligned on a 256k boundary, but can be placed
* anywhere with that alignment on POWER ISA 3+ systems.
*/
isa3_hashtb_pteg_table = (struct lpte *)moea64_bootstrap_alloc(size,
MAX(256*1024, size));
if (hw_direct_map)
isa3_hashtb_pteg_table =
(struct lpte *)PHYS_TO_DMAP((vm_offset_t)isa3_hashtb_pteg_table);
DISABLE_TRANS(msr);
bzero(__DEVOLATILE(void *, isa3_part_table), PART_SIZE);
isa3_part_table[0].pagetab =
((uintptr_t)isa3_hashtb_pteg_table & ~DMAP_BASE_ADDRESS) |
(uintptr_t)(flsl((moea64_pteg_count - 1) >> 11));
bzero(__DEVOLATILE(void *, isa3_hashtb_pteg_table), moea64_pteg_count *
sizeof(struct lpteg));
ENABLE_TRANS(msr);
CTR1(KTR_PMAP, "moea64_bootstrap: PTEG table at %p", isa3_hashtb_pteg_table);
moea64_mid_bootstrap(mmup, kernelstart, kernelend);
/*
* Add a mapping for the page table itself if there is no direct map.
*/
if (!hw_direct_map) {
size = moea64_pteg_count * sizeof(struct lpteg);
off = (vm_offset_t)(isa3_hashtb_pteg_table);
DISABLE_TRANS(msr);
for (pa = off; pa < off + size; pa += PAGE_SIZE)
pmap_kenter(pa, pa);
ENABLE_TRANS(msr);
}
/* Bring up virtual memory */
moea64_late_bootstrap(mmup, kernelstart, kernelend);
}
static void
tlbia(void)
{
vm_offset_t i;
i = 0xc00; /* IS = 11 */
TLBSYNC();
for (; i < 0x200000; i += 0x00001000) {
__asm __volatile("tlbiel %0" :: "r"(i));
}
EIEIO();
TLBSYNC();
}
static int
atomic_pte_lock(volatile struct lpte *pte, uint64_t bitmask, uint64_t *oldhi)
{
int ret;
uint32_t oldhihalf;
/*
* Note: in principle, if just the locked bit were set here, we
* could avoid needing the eviction lock. However, eviction occurs
* so rarely that it isn't worth bothering about in practice.
*/
__asm __volatile (
"1:\tlwarx %1, 0, %3\n\t" /* load old value */
"and. %0,%1,%4\n\t" /* check if any bits set */
"bne 2f\n\t" /* exit if any set */
"stwcx. %5, 0, %3\n\t" /* attempt to store */
"bne- 1b\n\t" /* spin if failed */
"li %0, 1\n\t" /* success - retval = 1 */
"b 3f\n\t" /* we've succeeded */
"2:\n\t"
"stwcx. %1, 0, %3\n\t" /* clear reservation (74xx) */
"li %0, 0\n\t" /* failure - retval = 0 */
"3:\n\t"
: "=&r" (ret), "=&r"(oldhihalf), "=m" (pte->pte_hi)
: "r" ((volatile char *)&pte->pte_hi + 4),
"r" ((uint32_t)bitmask), "r" ((uint32_t)LPTE_LOCKED),
"m" (pte->pte_hi)
: "cr0", "cr1", "cr2", "memory");
*oldhi = (pte->pte_hi & 0xffffffff00000000ULL) | oldhihalf;
return (ret);
}
static uintptr_t
isa3_hashtb_insert_to_pteg(struct lpte *pvo_pt, uintptr_t slotbase,
uint64_t mask)
{
volatile struct lpte *pt;
uint64_t oldptehi, va;
uintptr_t k;
int i, j;
/* Start at a random slot */
i = mftb() % 8;
for (j = 0; j < 8; j++) {
k = slotbase + (i + j) % 8;
pt = &isa3_hashtb_pteg_table[k];
/* Invalidate and seize lock only if no bits in mask set */
if (atomic_pte_lock(pt, mask, &oldptehi)) /* Lock obtained */
break;
}
if (j == 8)
return (-1);
if (oldptehi & LPTE_VALID) {
KASSERT(!(oldptehi & LPTE_WIRED), ("Unmapped wired entry"));
/*
* Need to invalidate old entry completely: see
* "Modifying a Page Table Entry". Need to reconstruct
* the virtual address for the outgoing entry to do that.
*/
if (oldptehi & LPTE_BIG)
va = oldptehi >> moea64_large_page_shift;
else
va = oldptehi >> ADDR_PIDX_SHFT;
if (oldptehi & LPTE_HID)
va = (((k >> 3) ^ moea64_pteg_mask) ^ va) &
VSID_HASH_MASK;
else
va = ((k >> 3) ^ va) & VSID_HASH_MASK;
va |= (oldptehi & LPTE_AVPN_MASK) <<
(ADDR_API_SHFT64 - ADDR_PIDX_SHFT);
PTESYNC();
TLBIE(va);
moea64_pte_valid--;
moea64_pte_overflow++;
}
/*
* Update the PTE as per "Adding a Page Table Entry". Lock is released
* by setting the high doubleworld.
*/
pt->pte_lo = htobe64(pvo_pt->pte_lo);
EIEIO();
pt->pte_hi = htobe64(pvo_pt->pte_hi);
PTESYNC();
/* Keep statistics */
moea64_pte_valid++;
return (k);
}
static int
isa3_hashtb_pte_insert(mmu_t mmu, struct pvo_entry *pvo)
{
struct lpte insertpt;
uintptr_t slot;
/* Initialize PTE */
moea64_pte_from_pvo(pvo, &insertpt);
/* Make sure further insertion is locked out during evictions */
rw_rlock(&isa3_hashtb_eviction_lock);
/*
* First try primary hash.
*/
pvo->pvo_pte.slot &= ~7ULL; /* Base slot address */
slot = isa3_hashtb_insert_to_pteg(&insertpt, pvo->pvo_pte.slot,
LPTE_VALID | LPTE_WIRED | LPTE_LOCKED);
if (slot != -1) {
rw_runlock(&isa3_hashtb_eviction_lock);
pvo->pvo_pte.slot = slot;
return (0);
}
/*
* Now try secondary hash.
*/
pvo->pvo_vaddr ^= PVO_HID;
insertpt.pte_hi ^= LPTE_HID;
pvo->pvo_pte.slot ^= (moea64_pteg_mask << 3);
slot = isa3_hashtb_insert_to_pteg(&insertpt, pvo->pvo_pte.slot,
LPTE_VALID | LPTE_WIRED | LPTE_LOCKED);
if (slot != -1) {
rw_runlock(&isa3_hashtb_eviction_lock);
pvo->pvo_pte.slot = slot;
return (0);
}
/*
* Out of luck. Find a PTE to sacrifice.
*/
/* Lock out all insertions for a bit */
if (!rw_try_upgrade(&isa3_hashtb_eviction_lock)) {
rw_runlock(&isa3_hashtb_eviction_lock);
rw_wlock(&isa3_hashtb_eviction_lock);
}
slot = isa3_hashtb_insert_to_pteg(&insertpt, pvo->pvo_pte.slot,
LPTE_WIRED | LPTE_LOCKED);
if (slot != -1) {
rw_wunlock(&isa3_hashtb_eviction_lock);
pvo->pvo_pte.slot = slot;
return (0);
}
/* Try other hash table. Now we're getting desperate... */
pvo->pvo_vaddr ^= PVO_HID;
insertpt.pte_hi ^= LPTE_HID;
pvo->pvo_pte.slot ^= (moea64_pteg_mask << 3);
slot = isa3_hashtb_insert_to_pteg(&insertpt, pvo->pvo_pte.slot,
LPTE_WIRED | LPTE_LOCKED);
if (slot != -1) {
rw_wunlock(&isa3_hashtb_eviction_lock);
pvo->pvo_pte.slot = slot;
return (0);
}
/* No freeable slots in either PTEG? We're hosed. */
rw_wunlock(&isa3_hashtb_eviction_lock);
panic("moea64_pte_insert: overflow");
return (-1);
}

View File

@ -134,7 +134,8 @@ __FBSDID("$FreeBSD$");
/* POWER9 only permits a 64k partition table size. */
#define PART_SIZE 0x10000
static int moea64_crop_tlbie;
static bool moea64_crop_tlbie;
static bool moea64_need_lock;
static __inline void
TLBIE(uint64_t vpn) {
@ -149,14 +150,26 @@ TLBIE(uint64_t vpn) {
vpn <<= ADDR_PIDX_SHFT;
/* Hobo spinlock: we need stronger guarantees than mutexes provide */
while (!atomic_cmpset_int(&tlbie_lock, 0, 1));
isync(); /* Flush instruction queue once lock acquired */
if (moea64_need_lock) {
while (!atomic_cmpset_int(&tlbie_lock, 0, 1));
isync(); /* Flush instruction queue once lock acquired */
}
if (moea64_crop_tlbie)
vpn &= ~(0xffffULL << 48);
#ifdef __powerpc64__
__asm __volatile("li 0, 0; tlbie %0" :: "r"(vpn) : "0","memory");
/*
* Explicitly clobber r0. The tlbie instruction has two forms: an old
* one used by PowerISA 2.03 and prior, and a newer one used by PowerISA
* 2.06 (maybe 2.05?) and later. We need to support both, and it just
* so happens that since we use 4k pages we can simply zero out r0, and
* clobber it, and the assembler will interpret the single-operand form
* of tlbie as having RB set, and everything else as 0. The RS operand
* in the newer form is in the same position as the L(page size) bit of
* the old form, so a slong as RS is 0, we're good on both sides.
*/
__asm __volatile("li 0, 0 \n tlbie %0" :: "r"(vpn) : "r0", "memory");
__asm __volatile("eieio; tlbsync; ptesync" ::: "memory");
#else
vpn_hi = (uint32_t)(vpn >> 32);
@ -183,7 +196,8 @@ TLBIE(uint64_t vpn) {
#endif
/* No barriers or special ops -- taken care of by ptesync above */
tlbie_lock = 0;
if (moea64_need_lock)
tlbie_lock = 0;
}
#define DISABLE_TRANS(msr) msr = mfmsr(); mtmsr(msr & ~PSL_DR)
@ -195,6 +209,8 @@ TLBIE(uint64_t vpn) {
static volatile struct lpte *moea64_pteg_table;
static struct rwlock moea64_eviction_lock;
static volatile struct pate *moea64_part_table;
/*
* PTE calls.
*/
@ -409,9 +425,14 @@ moea64_cpu_bootstrap_native(mmu_t mmup, int ap)
* Install page table
*/
__asm __volatile ("ptesync; mtsdr1 %0; isync"
:: "r"(((uintptr_t)moea64_pteg_table & ~DMAP_BASE_ADDRESS)
| (uintptr_t)(flsl(moea64_pteg_mask >> 11))));
if (cpu_features2 & PPC_FEATURE2_ARCH_3_00)
mtspr(SPR_PTCR,
((uintptr_t)moea64_part_table & ~DMAP_BASE_ADDRESS) |
flsl((PART_SIZE >> 12) - 1));
else
__asm __volatile ("ptesync; mtsdr1 %0; isync"
:: "r"(((uintptr_t)moea64_pteg_table & ~DMAP_BASE_ADDRESS)
| (uintptr_t)(flsl(moea64_pteg_mask >> 11))));
tlbia();
}
@ -427,13 +448,18 @@ moea64_bootstrap_native(mmu_t mmup, vm_offset_t kernelstart,
moea64_early_bootstrap(mmup, kernelstart, kernelend);
switch (mfpvr() >> 16) {
case IBMPOWER9:
moea64_need_lock = false;
break;
case IBMPOWER4:
case IBMPOWER4PLUS:
case IBM970:
case IBM970FX:
case IBM970GX:
case IBM970MP:
moea64_crop_tlbie = true;
moea64_crop_tlbie = true;
default:
moea64_need_lock = true;
}
/*
* Allocate PTEG table.
@ -463,9 +489,23 @@ moea64_bootstrap_native(mmu_t mmup, vm_offset_t kernelstart,
if (hw_direct_map)
moea64_pteg_table =
(struct lpte *)PHYS_TO_DMAP((vm_offset_t)moea64_pteg_table);
/* Allocate partition table (ISA 3.0). */
if (cpu_features2 & PPC_FEATURE2_ARCH_3_00) {
moea64_part_table =
(struct pate *)moea64_bootstrap_alloc(PART_SIZE, PART_SIZE);
if (hw_direct_map)
moea64_part_table =
(struct pate *)PHYS_TO_DMAP((vm_offset_t)moea64_part_table);
}
DISABLE_TRANS(msr);
bzero(__DEVOLATILE(void *, moea64_pteg_table), moea64_pteg_count *
sizeof(struct lpteg));
if (cpu_features2 & PPC_FEATURE2_ARCH_3_00) {
bzero(__DEVOLATILE(void *, moea64_part_table), PART_SIZE);
moea64_part_table[0].pagetab =
(DMAP_TO_PHYS((vm_offset_t)moea64_pteg_table)) |
(uintptr_t)(flsl((moea64_pteg_count - 1) >> 11));
}
ENABLE_TRANS(msr);
CTR1(KTR_PMAP, "moea64_bootstrap: PTEG table at %p", moea64_pteg_table);
@ -512,7 +552,7 @@ tlbia(void)
TLBSYNC();
for (; i < 0x200000; i += 0x00001000) {
for (; i < 0x400000; i += 0x00001000) {
#ifdef __powerpc64__
__asm __volatile("tlbiel %0" :: "r"(i));
#else