6859a304c6
Note that the APs still blow up in sched_throw(). MFC after: 1 month
229 lines
7.7 KiB
C
229 lines
7.7 KiB
C
/*
|
|
*
|
|
* Copyright (c) 2004 Christian Limpach.
|
|
* Copyright (c) 2004,2005 Kip Macy
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
* must display the following acknowledgement:
|
|
* This product includes software developed by Christian Limpach.
|
|
* 4. The name of the author may not be used to endorse or promote products
|
|
* derived from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
|
|
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
|
|
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
|
|
* IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
|
|
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
|
|
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
|
|
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
*
|
|
* $FreeBSD$
|
|
*/
|
|
|
|
#ifndef _XEN_XENPMAP_H_
|
|
#define _XEN_XENPMAP_H_
|
|
void _xen_queue_pt_update(vm_paddr_t, vm_paddr_t, char *, int);
|
|
void xen_pt_switch(vm_paddr_t);
|
|
void xen_set_ldt(vm_paddr_t, unsigned long);
|
|
void xen_pgdpt_pin(vm_paddr_t);
|
|
void xen_pgd_pin(vm_paddr_t);
|
|
void xen_pgd_unpin(vm_paddr_t);
|
|
void xen_pt_pin(vm_paddr_t);
|
|
void xen_pt_unpin(vm_paddr_t);
|
|
void xen_flush_queue(void);
|
|
void pmap_ref(pt_entry_t *pte, vm_paddr_t ma);
|
|
void xen_check_queue(void);
|
|
|
|
#ifdef INVARIANTS
|
|
#define xen_queue_pt_update(a, b) _xen_queue_pt_update((a), (b), __FILE__, __LINE__)
|
|
#else
|
|
#define xen_queue_pt_update(a, b) _xen_queue_pt_update((a), (b), NULL, 0)
|
|
#endif
|
|
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/pcpu.h>
|
|
|
|
#ifdef PMAP_DEBUG
|
|
#define PMAP_REF pmap_ref
|
|
#define PMAP_DEC_REF_PAGE pmap_dec_ref_page
|
|
#define PMAP_MARK_PRIV pmap_mark_privileged
|
|
#define PMAP_MARK_UNPRIV pmap_mark_unprivileged
|
|
#else
|
|
#define PMAP_MARK_PRIV(a)
|
|
#define PMAP_MARK_UNPRIV(a)
|
|
#define PMAP_REF(a, b)
|
|
#define PMAP_DEC_REF_PAGE(a)
|
|
#endif
|
|
|
|
#define ALWAYS_SYNC 0
|
|
|
|
#ifdef PT_DEBUG
|
|
#define PT_LOG() printk("WP PT_SET %s:%d\n", __FILE__, __LINE__)
|
|
#else
|
|
#define PT_LOG()
|
|
#endif
|
|
|
|
#define INVALID_P2M_ENTRY (~0UL)
|
|
|
|
#define pmap_valid_entry(E) ((E) & PG_V) /* is PDE or PTE valid? */
|
|
|
|
#define SH_PD_SET_VA 1
|
|
#define SH_PD_SET_VA_MA 2
|
|
#define SH_PD_SET_VA_CLEAR 3
|
|
|
|
struct pmap;
|
|
void pd_set(struct pmap *pmap, int ptepindex, vm_paddr_t val, int type);
|
|
#ifdef notyet
|
|
static vm_paddr_t
|
|
vptetomachpte(vm_paddr_t *pte)
|
|
{
|
|
vm_offset_t offset, ppte;
|
|
vm_paddr_t pgoffset, retval, *pdir_shadow_ptr;
|
|
int pgindex;
|
|
|
|
ppte = (vm_offset_t)pte;
|
|
pgoffset = (ppte & PAGE_MASK);
|
|
offset = ppte - (vm_offset_t)PTmap;
|
|
pgindex = ppte >> PDRSHIFT;
|
|
|
|
pdir_shadow_ptr = (vm_paddr_t *)PCPU_GET(pdir_shadow);
|
|
retval = (pdir_shadow_ptr[pgindex] & ~PAGE_MASK) + pgoffset;
|
|
return (retval);
|
|
}
|
|
#endif
|
|
#define PT_GET(_ptp) \
|
|
(pmap_valid_entry(*(_ptp)) ? xpmap_mtop(*(_ptp)) : (0))
|
|
|
|
#ifdef WRITABLE_PAGETABLES
|
|
|
|
#define PT_SET_VA(_ptp,_npte,sync) do { \
|
|
PMAP_REF((_ptp), xpmap_ptom(_npte)); \
|
|
PT_LOG(); \
|
|
*(_ptp) = xpmap_ptom((_npte)); \
|
|
} while (/*CONSTCOND*/0)
|
|
#define PT_SET_VA_MA(_ptp,_npte,sync) do { \
|
|
PMAP_REF((_ptp), (_npte)); \
|
|
PT_LOG(); \
|
|
*(_ptp) = (_npte); \
|
|
} while (/*CONSTCOND*/0)
|
|
#define PT_CLEAR_VA(_ptp, sync) do { \
|
|
PMAP_REF((pt_entry_t *)(_ptp), 0); \
|
|
PT_LOG(); \
|
|
*(_ptp) = 0; \
|
|
} while (/*CONSTCOND*/0)
|
|
|
|
#define PD_SET_VA(_pmap, _ptp, _npte, sync) do { \
|
|
PMAP_REF((_ptp), xpmap_ptom(_npte)); \
|
|
pd_set((_pmap),(_ptp),(_npte), SH_PD_SET_VA); \
|
|
if (sync || ALWAYS_SYNC) xen_flush_queue(); \
|
|
} while (/*CONSTCOND*/0)
|
|
#define PD_SET_VA_MA(_pmap, _ptp, _npte, sync) do { \
|
|
PMAP_REF((_ptp), (_npte)); \
|
|
pd_set((_pmap),(_ptp),(_npte), SH_PD_SET_VA_MA); \
|
|
if (sync || ALWAYS_SYNC) xen_flush_queue(); \
|
|
} while (/*CONSTCOND*/0)
|
|
#define PD_CLEAR_VA(_pmap, _ptp, sync) do { \
|
|
PMAP_REF((pt_entry_t *)(_ptp), 0); \
|
|
pd_set((_pmap),(_ptp), 0, SH_PD_SET_VA_CLEAR); \
|
|
if (sync || ALWAYS_SYNC) xen_flush_queue(); \
|
|
} while (/*CONSTCOND*/0)
|
|
|
|
#else /* !WRITABLE_PAGETABLES */
|
|
|
|
#define PT_SET_VA(_ptp,_npte,sync) do { \
|
|
PMAP_REF((_ptp), xpmap_ptom(_npte)); \
|
|
xen_queue_pt_update(vtomach(_ptp), \
|
|
xpmap_ptom(_npte)); \
|
|
if (sync || ALWAYS_SYNC) xen_flush_queue(); \
|
|
} while (/*CONSTCOND*/0)
|
|
#define PT_SET_VA_MA(_ptp,_npte,sync) do { \
|
|
PMAP_REF((_ptp), (_npte)); \
|
|
xen_queue_pt_update(vtomach(_ptp), _npte); \
|
|
if (sync || ALWAYS_SYNC) xen_flush_queue(); \
|
|
} while (/*CONSTCOND*/0)
|
|
#define PT_CLEAR_VA(_ptp, sync) do { \
|
|
PMAP_REF((pt_entry_t *)(_ptp), 0); \
|
|
xen_queue_pt_update(vtomach(_ptp), 0); \
|
|
if (sync || ALWAYS_SYNC) \
|
|
xen_flush_queue(); \
|
|
} while (/*CONSTCOND*/0)
|
|
|
|
#define PD_SET_VA(_pmap, _ptepindex,_npte,sync) do { \
|
|
PMAP_REF((_ptp), xpmap_ptom(_npte)); \
|
|
pd_set((_pmap),(_ptepindex),(_npte), SH_PD_SET_VA); \
|
|
if (sync || ALWAYS_SYNC) xen_flush_queue(); \
|
|
} while (/*CONSTCOND*/0)
|
|
#define PD_SET_VA_MA(_pmap, _ptepindex,_npte,sync) do { \
|
|
PMAP_REF((_ptp), (_npte)); \
|
|
pd_set((_pmap),(_ptepindex),(_npte), SH_PD_SET_VA_MA); \
|
|
if (sync || ALWAYS_SYNC) xen_flush_queue(); \
|
|
} while (/*CONSTCOND*/0)
|
|
#define PD_CLEAR_VA(_pmap, _ptepindex, sync) do { \
|
|
PMAP_REF((pt_entry_t *)(_ptp), 0); \
|
|
pd_set((_pmap),(_ptepindex), 0, SH_PD_SET_VA_CLEAR); \
|
|
if (sync || ALWAYS_SYNC) xen_flush_queue(); \
|
|
} while (/*CONSTCOND*/0)
|
|
|
|
#endif
|
|
|
|
#define PT_SET_MA(_va, _ma) \
|
|
do { \
|
|
PANIC_IF(HYPERVISOR_update_va_mapping(((unsigned long)(_va)),\
|
|
(_ma), \
|
|
UVMF_INVLPG| UVMF_ALL) < 0); \
|
|
} while (/*CONSTCOND*/0)
|
|
|
|
#define PT_UPDATES_FLUSH() do { \
|
|
xen_flush_queue(); \
|
|
} while (/*CONSTCOND*/0)
|
|
|
|
static __inline vm_paddr_t
|
|
xpmap_mtop(vm_paddr_t mpa)
|
|
{
|
|
vm_paddr_t tmp = (mpa & PG_FRAME);
|
|
|
|
return machtophys(tmp) | (mpa & ~PG_FRAME);
|
|
}
|
|
|
|
static __inline vm_paddr_t
|
|
xpmap_ptom(vm_paddr_t ppa)
|
|
{
|
|
vm_paddr_t tmp = (ppa & PG_FRAME);
|
|
|
|
return phystomach(tmp) | (ppa & ~PG_FRAME);
|
|
}
|
|
|
|
static __inline void
|
|
set_phys_to_machine(unsigned long pfn, unsigned long mfn)
|
|
{
|
|
#ifdef notyet
|
|
PANIC_IF(max_mapnr && pfn >= max_mapnr);
|
|
#endif
|
|
if (xen_feature(XENFEAT_auto_translated_physmap)) {
|
|
#ifdef notyet
|
|
PANIC_IF((pfn != mfn && mfn != INVALID_P2M_ENTRY));
|
|
#endif
|
|
return;
|
|
}
|
|
xen_phys_machine[pfn] = mfn;
|
|
}
|
|
|
|
|
|
|
|
|
|
#endif /* _XEN_XENPMAP_H_ */
|