aboutsummaryrefslogtreecommitdiff
path: root/usr.sbin/bhyve/aarch64/vmexit.c
diff options
context:
space:
mode:
Diffstat (limited to 'usr.sbin/bhyve/aarch64/vmexit.c')
-rw-r--r--usr.sbin/bhyve/aarch64/vmexit.c305
1 files changed, 305 insertions, 0 deletions
diff --git a/usr.sbin/bhyve/aarch64/vmexit.c b/usr.sbin/bhyve/aarch64/vmexit.c
new file mode 100644
index 000000000000..2457cbe76b5e
--- /dev/null
+++ b/usr.sbin/bhyve/aarch64/vmexit.c
@@ -0,0 +1,305 @@
+/*-
+ * SPDX-License-Identifier: BSD-2-Clause
+ *
+ * Copyright (c) 2011 NetApp, Inc.
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ *
+ * THIS SOFTWARE IS PROVIDED BY NETAPP, INC ``AS IS'' AND
+ * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL NETAPP, INC OR CONTRIBUTORS BE LIABLE
+ * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
+ * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
+ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
+ * SUCH DAMAGE.
+ */
+
+#include <sys/types.h>
+#include <sys/cpuset.h>
+
+#include <dev/psci/psci.h>
+#include <dev/psci/smccc.h>
+
+#include <machine/armreg.h>
+#include <machine/cpu.h>
+#include <machine/vmm.h>
+#include <machine/vmm_dev.h>
+#include <machine/vmm_instruction_emul.h>
+
+#include <assert.h>
+#include <errno.h>
+#include <stdbool.h>
+#include <stdio.h>
+#include <stdlib.h>
+#include <unistd.h>
+
+#include <vmmapi.h>
+
+#include "bhyve_machdep.h"
+#include "bhyverun.h"
+#include "config.h"
+#include "debug.h"
+#include "gdb.h"
+#include "mem.h"
+#include "vmexit.h"
+
+cpuset_t running_cpumask;
+
+static int
+vmexit_inst_emul(struct vmctx *ctx __unused, struct vcpu *vcpu,
+ struct vm_run *vmrun)
+{
+ struct vm_exit *vme;
+ struct vie *vie;
+ int err;
+
+ vme = vmrun->vm_exit;
+ vie = &vme->u.inst_emul.vie;
+
+ err = emulate_mem(vcpu, vme->u.inst_emul.gpa, vie,
+ &vme->u.inst_emul.paging);
+ if (err) {
+ if (err == ESRCH) {
+ EPRINTLN("Unhandled memory access to 0x%lx\n",
+ vme->u.inst_emul.gpa);
+ }
+ goto fail;
+ }
+
+ return (VMEXIT_CONTINUE);
+
+fail:
+ fprintf(stderr, "Failed to emulate instruction ");
+ FPRINTLN(stderr, "at 0x%lx", vme->pc);
+ return (VMEXIT_ABORT);
+}
+
+static int
+vmexit_reg_emul(struct vmctx *ctx __unused, struct vcpu *vcpu __unused,
+ struct vm_run *vmrun)
+{
+ struct vm_exit *vme;
+ struct vre *vre;
+
+ vme = vmrun->vm_exit;
+ vre = &vme->u.reg_emul.vre;
+
+ EPRINTLN("Unhandled register access: pc %#lx syndrome %#x reg %d\n",
+ vme->pc, vre->inst_syndrome, vre->reg);
+ return (VMEXIT_ABORT);
+}
+
+static int
+vmexit_suspend(struct vmctx *ctx, struct vcpu *vcpu, struct vm_run *vmrun)
+{
+ struct vm_exit *vme;
+ enum vm_suspend_how how;
+ int vcpuid = vcpu_id(vcpu);
+
+ vme = vmrun->vm_exit;
+ how = vme->u.suspended.how;
+
+ fbsdrun_deletecpu(vcpuid);
+
+ switch (how) {
+ case VM_SUSPEND_RESET:
+ exit(0);
+ case VM_SUSPEND_POWEROFF:
+ if (get_config_bool_default("destroy_on_poweroff", false))
+ vm_destroy(ctx);
+ exit(1);
+ case VM_SUSPEND_HALT:
+ exit(2);
+ case VM_SUSPEND_DESTROY:
+ exit(4);
+ default:
+ fprintf(stderr, "vmexit_suspend: invalid reason %d\n", how);
+ exit(100);
+ }
+ return (0); /* NOTREACHED */
+}
+
+static int
+vmexit_debug(struct vmctx *ctx __unused, struct vcpu *vcpu,
+ struct vm_run *vmrun __unused)
+{
+ gdb_cpu_suspend(vcpu);
+ /*
+ * XXX-MJ sleep for a short period to avoid chewing up the CPU in the
+ * window between activation of the vCPU thread and the STARTUP IPI.
+ */
+ usleep(1000);
+ return (VMEXIT_CONTINUE);
+}
+
+static int
+vmexit_bogus(struct vmctx *ctx __unused, struct vcpu *vcpu __unused,
+ struct vm_run *vmrun __unused)
+{
+ return (VMEXIT_CONTINUE);
+}
+
+static uint64_t
+smccc_affinity_info(uint64_t target_affinity, uint32_t lowest_affinity_level)
+{
+ uint64_t mask = 0;
+
+ switch (lowest_affinity_level) {
+ case 0:
+ mask |= CPU_AFF0_MASK;
+ /* FALLTHROUGH */
+ case 1:
+ mask |= CPU_AFF1_MASK;
+ /* FALLTHROUGH */
+ case 2:
+ mask |= CPU_AFF2_MASK;
+ /* FALLTHROUGH */
+ case 3:
+ mask |= CPU_AFF3_MASK;
+ break;
+ default:
+ return (PSCI_RETVAL_INVALID_PARAMS);
+ }
+
+ for (int vcpu = 0; vcpu < guest_ncpus; vcpu++) {
+ if ((cpu_to_mpidr[vcpu] & mask) == (target_affinity & mask) &&
+ CPU_ISSET(vcpu, &running_cpumask)) {
+ /* Return ON if any CPUs are on */
+ return (PSCI_AFFINITY_INFO_ON);
+ }
+ }
+
+ /* No CPUs in the affinity mask are on, return OFF */
+ return (PSCI_AFFINITY_INFO_OFF);
+}
+
+static int
+vmexit_smccc(struct vmctx *ctx, struct vcpu *vcpu, struct vm_run *vmrun)
+{
+ struct vcpu *newvcpu;
+ struct vm_exit *vme;
+ uint64_t mpidr, smccc_rv;
+ enum vm_suspend_how how;
+ int error, newcpu;
+
+ /* Return the Unknown Function Identifier by default */
+ smccc_rv = SMCCC_RET_NOT_SUPPORTED;
+
+ vme = vmrun->vm_exit;
+ switch (vme->u.smccc_call.func_id) {
+ case PSCI_FNID_VERSION:
+ /* We implement PSCI 1.0 */
+ smccc_rv = PSCI_VER(1, 0);
+ break;
+ case PSCI_FNID_CPU_SUSPEND:
+ break;
+ case PSCI_FNID_CPU_OFF:
+ CPU_CLR_ATOMIC(vcpu_id(vcpu), &running_cpumask);
+ vm_suspend_cpu(vcpu);
+ break;
+ case PSCI_FNID_CPU_ON:
+ mpidr = vme->u.smccc_call.args[0];
+ for (newcpu = 0; newcpu < guest_ncpus; newcpu++) {
+ if (cpu_to_mpidr[newcpu] == mpidr)
+ break;
+ }
+
+ if (newcpu == guest_ncpus) {
+ smccc_rv = PSCI_RETVAL_INVALID_PARAMS;
+ break;
+ }
+
+ if (CPU_TEST_SET_ATOMIC(newcpu, &running_cpumask)) {
+ smccc_rv = PSCI_RETVAL_ALREADY_ON;
+ break;
+ }
+
+ newvcpu = fbsdrun_vcpu(newcpu);
+ assert(newvcpu != NULL);
+
+ /* Set the context ID */
+ error = vm_set_register(newvcpu, VM_REG_GUEST_X0,
+ vme->u.smccc_call.args[2]);
+ assert(error == 0);
+
+ /* Set the start program counter */
+ error = vm_set_register(newvcpu, VM_REG_GUEST_PC,
+ vme->u.smccc_call.args[1]);
+ assert(error == 0);
+
+ vm_resume_cpu(newvcpu);
+
+ smccc_rv = PSCI_RETVAL_SUCCESS;
+ break;
+ case PSCI_FNID_AFFINITY_INFO:
+ smccc_rv = smccc_affinity_info(vme->u.smccc_call.args[0],
+ vme->u.smccc_call.args[1]);
+ break;
+ case PSCI_FNID_SYSTEM_OFF:
+ case PSCI_FNID_SYSTEM_RESET:
+ if (vme->u.smccc_call.func_id == PSCI_FNID_SYSTEM_OFF)
+ how = VM_SUSPEND_POWEROFF;
+ else
+ how = VM_SUSPEND_RESET;
+ error = vm_suspend(ctx, how);
+ assert(error == 0 || errno == EALREADY);
+ break;
+ default:
+ break;
+ }
+
+ error = vm_set_register(vcpu, VM_REG_GUEST_X0, smccc_rv);
+ assert(error == 0);
+
+ return (VMEXIT_CONTINUE);
+}
+
+static int
+vmexit_hyp(struct vmctx *ctx __unused, struct vcpu *vcpu, struct vm_run *vmrun)
+{
+ /* Raise an unknown reason exception */
+ if (vm_inject_exception(vcpu,
+ (EXCP_UNKNOWN << ESR_ELx_EC_SHIFT) | ESR_ELx_IL,
+ vmrun->vm_exit->u.hyp.far_el2) != 0)
+ return (VMEXIT_ABORT);
+
+ return (VMEXIT_CONTINUE);
+}
+
+static int
+vmexit_brk(struct vmctx *ctx __unused, struct vcpu *vcpu, struct vm_run *vmrun)
+{
+ gdb_cpu_breakpoint(vcpu, vmrun->vm_exit);
+ return (VMEXIT_CONTINUE);
+}
+
+static int
+vmexit_ss(struct vmctx *ctx __unused, struct vcpu *vcpu, struct vm_run *vmrun)
+{
+ gdb_cpu_debug(vcpu, vmrun->vm_exit);
+ return (VMEXIT_CONTINUE);
+}
+
+const vmexit_handler_t vmexit_handlers[VM_EXITCODE_MAX] = {
+ [VM_EXITCODE_BOGUS] = vmexit_bogus,
+ [VM_EXITCODE_INST_EMUL] = vmexit_inst_emul,
+ [VM_EXITCODE_REG_EMUL] = vmexit_reg_emul,
+ [VM_EXITCODE_SUSPENDED] = vmexit_suspend,
+ [VM_EXITCODE_DEBUG] = vmexit_debug,
+ [VM_EXITCODE_SMCCC] = vmexit_smccc,
+ [VM_EXITCODE_HYP] = vmexit_hyp,
+ [VM_EXITCODE_BRK] = vmexit_brk,
+ [VM_EXITCODE_SS] = vmexit_ss,
+};