3e1f2ae835
Add virtualized XSAVE support to bhyve which permits guests to use XSAVE and XSAVE-enabled features like AVX. - Store a per-cpu guest xcr0 register and handle xsetbv VM exits by emulating the instruction. - Only expose XSAVE to guests if XSAVE is enabled in the host. Only expose a subset of XSAVE features currently supported by the guest and for which the proper emulation of xsetbv is known. Currently this includes X87, SSE, AVX, AVX-512, and Intel MPX. - Add support for injecting hardware exceptions into the guest and use this to trigger exceptions in the guest for invalid xsetbv operations instead of potentially faulting in the host. - Queue pending exceptions in the 'struct vcpu' instead of directly updating the processor-specific VMCS or VMCB. The pending exception will be delivered right before entering the guest. - Rename the unused ioctl VM_INJECT_EVENT to VM_INJECT_EXCEPTION and restrict it to only deliver x86 hardware exceptions. This new ioctl is now used to inject a protection fault when the guest accesses an unimplemented MSR. - Expose a subset of known-safe features from leaf 0 of the structured extended features to guests if they are supported on the host including RDFSBASE/RDGSBASE, BMI1/2, AVX2, AVX-512, HLE, ERMS, and RTM. Aside from AVX-512, these features are all new instructions available for use in ring 3 with no additional hypervisor changes needed.
162 lines
3.5 KiB
C
162 lines
3.5 KiB
C
/*-
|
|
* Copyright (c) 2012 NetApp, Inc.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY NETAPP, INC ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL NETAPP, INC OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* $FreeBSD$
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/pcpu.h>
|
|
|
|
#include <machine/cpufunc.h>
|
|
#include <machine/segments.h>
|
|
#include <machine/specialreg.h>
|
|
|
|
#include "vmm_host.h"
|
|
|
|
static uint64_t vmm_host_efer, vmm_host_pat, vmm_host_cr0, vmm_host_cr4,
|
|
vmm_host_xcr0;
|
|
static struct xsave_limits vmm_xsave_limits;
|
|
|
|
void
|
|
vmm_host_state_init(void)
|
|
{
|
|
int regs[4];
|
|
|
|
vmm_host_efer = rdmsr(MSR_EFER);
|
|
vmm_host_pat = rdmsr(MSR_PAT);
|
|
|
|
/*
|
|
* We always want CR0.TS to be set when the processor does a VM exit.
|
|
*
|
|
* With emulation turned on unconditionally after a VM exit, we are
|
|
* able to trap inadvertent use of the FPU until the guest FPU state
|
|
* has been safely squirreled away.
|
|
*/
|
|
vmm_host_cr0 = rcr0() | CR0_TS;
|
|
|
|
vmm_host_cr4 = rcr4();
|
|
|
|
/*
|
|
* Only permit a guest to use XSAVE if the host is using
|
|
* XSAVE. Only permit a guest to use XSAVE features supported
|
|
* by the host. This ensures that the FPU state used by the
|
|
* guest is always a subset of the saved guest FPU state.
|
|
*
|
|
* In addition, only permit known XSAVE features where the
|
|
* rules for which features depend on other features is known
|
|
* to properly emulate xsetbv.
|
|
*/
|
|
if (vmm_host_cr4 & CR4_XSAVE) {
|
|
vmm_xsave_limits.xsave_enabled = 1;
|
|
vmm_host_xcr0 = rxcr(0);
|
|
vmm_xsave_limits.xcr0_allowed = vmm_host_xcr0 &
|
|
(XFEATURE_AVX | XFEATURE_MPX | XFEATURE_AVX512);
|
|
|
|
cpuid_count(0xd, 0x0, regs);
|
|
vmm_xsave_limits.xsave_max_size = regs[1];
|
|
}
|
|
}
|
|
|
|
uint64_t
|
|
vmm_get_host_pat(void)
|
|
{
|
|
|
|
return (vmm_host_pat);
|
|
}
|
|
|
|
uint64_t
|
|
vmm_get_host_efer(void)
|
|
{
|
|
|
|
return (vmm_host_efer);
|
|
}
|
|
|
|
uint64_t
|
|
vmm_get_host_cr0(void)
|
|
{
|
|
|
|
return (vmm_host_cr0);
|
|
}
|
|
|
|
uint64_t
|
|
vmm_get_host_cr4(void)
|
|
{
|
|
|
|
return (vmm_host_cr4);
|
|
}
|
|
|
|
uint64_t
|
|
vmm_get_host_xcr0(void)
|
|
{
|
|
|
|
return (vmm_host_xcr0);
|
|
}
|
|
|
|
uint64_t
|
|
vmm_get_host_datasel(void)
|
|
{
|
|
|
|
return (GSEL(GDATA_SEL, SEL_KPL));
|
|
|
|
}
|
|
|
|
uint64_t
|
|
vmm_get_host_codesel(void)
|
|
{
|
|
|
|
return (GSEL(GCODE_SEL, SEL_KPL));
|
|
}
|
|
|
|
uint64_t
|
|
vmm_get_host_tsssel(void)
|
|
{
|
|
|
|
return (GSEL(GPROC0_SEL, SEL_KPL));
|
|
}
|
|
|
|
uint64_t
|
|
vmm_get_host_fsbase(void)
|
|
{
|
|
|
|
return (0);
|
|
}
|
|
|
|
uint64_t
|
|
vmm_get_host_idtrbase(void)
|
|
{
|
|
|
|
return (r_idt.rd_base);
|
|
}
|
|
|
|
const struct xsave_limits *
|
|
vmm_get_xsave_limits(void)
|
|
{
|
|
|
|
return (&vmm_xsave_limits);
|
|
}
|