summaryrefslogtreecommitdiff
path: root/src/VBox/VMM/VMMR3/EMHM.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'src/VBox/VMM/VMMR3/EMHM.cpp')
-rw-r--r--src/VBox/VMM/VMMR3/EMHM.cpp604
1 files changed, 604 insertions, 0 deletions
diff --git a/src/VBox/VMM/VMMR3/EMHM.cpp b/src/VBox/VMM/VMMR3/EMHM.cpp
new file mode 100644
index 00000000..39dd8df2
--- /dev/null
+++ b/src/VBox/VMM/VMMR3/EMHM.cpp
@@ -0,0 +1,604 @@
+/* $Id: EMHM.cpp $ */
+/** @file
+ * EM - Execution Monitor / Manager - hardware virtualization
+ */
+
+/*
+ * Copyright (C) 2006-2013 Oracle Corporation
+ *
+ * This file is part of VirtualBox Open Source Edition (OSE), as
+ * available from http://www.virtualbox.org. This file is free software;
+ * you can redistribute it and/or modify it under the terms of the GNU
+ * General Public License (GPL) as published by the Free Software
+ * Foundation, in version 2 as it comes in the "COPYING" file of the
+ * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
+ * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
+ */
+
+/*******************************************************************************
+* Header Files *
+*******************************************************************************/
+#define LOG_GROUP LOG_GROUP_EM
+#include <VBox/vmm/em.h>
+#include <VBox/vmm/vmm.h>
+#include <VBox/vmm/csam.h>
+#include <VBox/vmm/selm.h>
+#include <VBox/vmm/trpm.h>
+#include <VBox/vmm/iem.h>
+#include <VBox/vmm/iom.h>
+#include <VBox/vmm/dbgf.h>
+#include <VBox/vmm/pgm.h>
+#ifdef VBOX_WITH_REM
+# include <VBox/vmm/rem.h>
+#endif
+#include <VBox/vmm/tm.h>
+#include <VBox/vmm/mm.h>
+#include <VBox/vmm/ssm.h>
+#include <VBox/vmm/pdmapi.h>
+#include <VBox/vmm/pdmcritsect.h>
+#include <VBox/vmm/pdmqueue.h>
+#include <VBox/vmm/hm.h>
+#include "EMInternal.h"
+#include <VBox/vmm/vm.h>
+#include <VBox/vmm/cpumdis.h>
+#include <VBox/dis.h>
+#include <VBox/disopcode.h>
+#include <VBox/vmm/dbgf.h>
+#include "VMMTracing.h"
+
+#include <iprt/asm.h>
+
+
+/*******************************************************************************
+* Defined Constants And Macros *
+*******************************************************************************/
+#if 0 /* Disabled till after 2.1.0 when we've time to test it. */
+#define EM_NOTIFY_HM
+#endif
+
+
+/*******************************************************************************
+* Internal Functions *
+*******************************************************************************/
+DECLINLINE(int) emR3HmExecuteInstruction(PVM pVM, PVMCPU pVCpu, const char *pszPrefix, int rcGC = VINF_SUCCESS);
+static int emR3HmExecuteIOInstruction(PVM pVM, PVMCPU pVCpu);
+static int emR3HmForcedActions(PVM pVM, PVMCPU pVCpu, PCPUMCTX pCtx);
+
+#define EMHANDLERC_WITH_HM
+#define emR3ExecuteInstruction emR3HmExecuteInstruction
+#define emR3ExecuteIOInstruction emR3HmExecuteIOInstruction
+#include "EMHandleRCTmpl.h"
+
+
+/**
+ * Executes instruction in HM mode if we can.
+ *
+ * This is somewhat comparable to REMR3EmulateInstruction.
+ *
+ * @returns VBox strict status code.
+ * @retval VINF_EM_DBG_STEPPED on success.
+ * @retval VERR_EM_CANNOT_EXEC_GUEST if we cannot execute guest instructions in
+ * HM right now.
+ *
+ * @param pVM Pointer to the cross context VM structure.
+ * @param pVCpu Pointer to the cross context CPU structure for
+ * the calling EMT.
+ * @param fFlags Combinations of EM_ONE_INS_FLAGS_XXX.
+ * @thread EMT.
+ */
+VMMR3_INT_DECL(VBOXSTRICTRC) EMR3HmSingleInstruction(PVM pVM, PVMCPU pVCpu, uint32_t fFlags)
+{
+ PCPUMCTX pCtx = pVCpu->em.s.pCtx;
+ Assert(!(fFlags & ~EM_ONE_INS_FLAGS_MASK));
+
+ if (!HMR3CanExecuteGuest(pVM, pCtx))
+ return VINF_EM_RESCHEDULE;
+
+ uint64_t const uOldRip = pCtx->rip;
+ for (;;)
+ {
+ /*
+ * Service necessary FFs before going into HM.
+ */
+ if ( VM_FF_IS_PENDING(pVM, VM_FF_HIGH_PRIORITY_PRE_RAW_MASK)
+ || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HIGH_PRIORITY_PRE_RAW_MASK))
+ {
+ VBOXSTRICTRC rcStrict = emR3HmForcedActions(pVM, pVCpu, pCtx);
+ if (rcStrict != VINF_SUCCESS)
+ {
+ Log(("EMR3HmSingleInstruction: FFs before -> %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
+ return rcStrict;
+ }
+ }
+
+ /*
+ * Go execute it.
+ */
+ bool fOld = HMSetSingleInstruction(pVCpu, true);
+ VBOXSTRICTRC rcStrict = VMMR3HmRunGC(pVM, pVCpu);
+ HMSetSingleInstruction(pVCpu, fOld);
+ LogFlow(("EMR3HmSingleInstruction: %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
+
+ /*
+ * Handle high priority FFs and informational status codes. We don't do
+ * normal FF processing the caller or the next call can deal with them.
+ */
+ VMCPU_FF_CLEAR(pVCpu, VMCPU_FF_RESUME_GUEST_MASK);
+ if ( VM_FF_IS_PENDING(pVM, VM_FF_HIGH_PRIORITY_POST_MASK)
+ || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HIGH_PRIORITY_POST_MASK))
+ {
+ rcStrict = emR3HighPriorityPostForcedActions(pVM, pVCpu, VBOXSTRICTRC_TODO(rcStrict));
+ LogFlow(("EMR3HmSingleInstruction: FFs after -> %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
+ }
+
+ if (rcStrict != VINF_SUCCESS && (rcStrict < VINF_EM_FIRST || rcStrict > VINF_EM_LAST))
+ {
+ rcStrict = emR3HmHandleRC(pVM, pVCpu, pCtx, VBOXSTRICTRC_TODO(rcStrict));
+ Log(("EMR3HmSingleInstruction: emR3HmHandleRC -> %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
+ }
+
+ /*
+ * Done?
+ */
+ if ( (rcStrict != VINF_SUCCESS && rcStrict != VINF_EM_DBG_STEPPED)
+ || !(fFlags & EM_ONE_INS_FLAGS_RIP_CHANGE)
+ || pCtx->rip != uOldRip)
+ {
+ if (rcStrict == VINF_SUCCESS && pCtx->rip != uOldRip)
+ rcStrict = VINF_EM_DBG_STEPPED;
+ Log(("EMR3HmSingleInstruction: returns %Rrc (rip %llx -> %llx)\n", VBOXSTRICTRC_VAL(rcStrict), uOldRip, pCtx->rip));
+ return rcStrict;
+ }
+ }
+}
+
+
+/**
+ * Executes one (or perhaps a few more) instruction(s).
+ *
+ * @returns VBox status code suitable for EM.
+ *
+ * @param pVM Pointer to the VM.
+ * @param pVCpu Pointer to the VMCPU.
+ * @param rcRC Return code from RC.
+ * @param pszPrefix Disassembly prefix. If not NULL we'll disassemble the
+ * instruction and prefix the log output with this text.
+ */
+#ifdef LOG_ENABLED
+static int emR3HmExecuteInstructionWorker(PVM pVM, PVMCPU pVCpu, int rcRC, const char *pszPrefix)
+#else
+static int emR3HmExecuteInstructionWorker(PVM pVM, PVMCPU pVCpu, int rcRC)
+#endif
+{
+#ifdef LOG_ENABLED
+ PCPUMCTX pCtx = pVCpu->em.s.pCtx;
+#endif
+ int rc;
+ NOREF(rcRC);
+
+#ifdef LOG_ENABLED
+ /*
+ * Log it.
+ */
+ Log(("EMINS: %04x:%RGv RSP=%RGv\n", pCtx->cs.Sel, (RTGCPTR)pCtx->rip, (RTGCPTR)pCtx->rsp));
+ if (pszPrefix)
+ {
+ DBGFR3_INFO_LOG(pVM, "cpumguest", pszPrefix);
+ DBGFR3_DISAS_INSTR_CUR_LOG(pVCpu, pszPrefix);
+ }
+#endif
+
+ /*
+ * Use IEM and fallback on REM if the functionality is missing.
+ * Once IEM gets mature enough, nothing should ever fall back.
+ */
+#if defined(VBOX_WITH_FIRST_IEM_STEP) || !defined(VBOX_WITH_REM)
+ STAM_PROFILE_START(&pVCpu->em.s.StatIEMEmu, a);
+ rc = VBOXSTRICTRC_TODO(IEMExecOne(pVCpu));
+ STAM_PROFILE_STOP(&pVCpu->em.s.StatIEMEmu, a);
+
+ if ( rc == VERR_IEM_ASPECT_NOT_IMPLEMENTED
+ || rc == VERR_IEM_INSTR_NOT_IMPLEMENTED)
+#endif
+ {
+#ifdef VBOX_WITH_REM
+ STAM_PROFILE_START(&pVCpu->em.s.StatREMEmu, b);
+# ifndef VBOX_WITH_FIRST_IEM_STEP
+ Log(("EMINS[rem]: %04x:%RGv RSP=%RGv\n", pCtx->cs.Sel, (RTGCPTR)pCtx->rip, (RTGCPTR)pCtx->rsp));
+//# elif defined(DEBUG_bird)
+// AssertFailed();
+# endif
+ EMRemLock(pVM);
+ /* Flush the recompiler TLB if the VCPU has changed. */
+ if (pVM->em.s.idLastRemCpu != pVCpu->idCpu)
+ CPUMSetChangedFlags(pVCpu, CPUM_CHANGED_ALL);
+ pVM->em.s.idLastRemCpu = pVCpu->idCpu;
+
+ rc = REMR3EmulateInstruction(pVM, pVCpu);
+ EMRemUnlock(pVM);
+ STAM_PROFILE_STOP(&pVCpu->em.s.StatREMEmu, b);
+#else /* !VBOX_WITH_REM */
+ NOREF(pVM);
+#endif /* !VBOX_WITH_REM */
+ }
+
+#ifdef EM_NOTIFY_HM
+ if (pVCpu->em.s.enmState == EMSTATE_DEBUG_GUEST_HM)
+ HMR3NotifyEmulated(pVCpu);
+#endif
+ return rc;
+}
+
+
+/**
+ * Executes one (or perhaps a few more) instruction(s).
+ * This is just a wrapper for discarding pszPrefix in non-logging builds.
+ *
+ * @returns VBox status code suitable for EM.
+ * @param pVM Pointer to the VM.
+ * @param pVCpu Pointer to the VMCPU.
+ * @param pszPrefix Disassembly prefix. If not NULL we'll disassemble the
+ * instruction and prefix the log output with this text.
+ * @param rcGC GC return code
+ */
+DECLINLINE(int) emR3HmExecuteInstruction(PVM pVM, PVMCPU pVCpu, const char *pszPrefix, int rcGC)
+{
+#ifdef LOG_ENABLED
+ return emR3HmExecuteInstructionWorker(pVM, pVCpu, rcGC, pszPrefix);
+#else
+ return emR3HmExecuteInstructionWorker(pVM, pVCpu, rcGC);
+#endif
+}
+
+/**
+ * Executes one (or perhaps a few more) IO instruction(s).
+ *
+ * @returns VBox status code suitable for EM.
+ * @param pVM Pointer to the VM.
+ * @param pVCpu Pointer to the VMCPU.
+ */
+static int emR3HmExecuteIOInstruction(PVM pVM, PVMCPU pVCpu)
+{
+ PCPUMCTX pCtx = pVCpu->em.s.pCtx;
+
+ STAM_PROFILE_START(&pVCpu->em.s.StatIOEmu, a);
+
+ /*
+ * Try to restart the io instruction that was refused in ring-0.
+ */
+ VBOXSTRICTRC rcStrict = HMR3RestartPendingIOInstr(pVM, pVCpu, pCtx);
+ if (IOM_SUCCESS(rcStrict))
+ {
+ STAM_COUNTER_INC(&pVCpu->em.s.CTX_SUFF(pStats)->StatIoRestarted);
+ STAM_PROFILE_STOP(&pVCpu->em.s.StatIOEmu, a);
+ return VBOXSTRICTRC_TODO(rcStrict); /* rip already updated. */
+ }
+ AssertMsgReturn(rcStrict == VERR_NOT_FOUND, ("%Rrc\n", VBOXSTRICTRC_VAL(rcStrict)),
+ RT_SUCCESS_NP(rcStrict) ? VERR_IPE_UNEXPECTED_INFO_STATUS : VBOXSTRICTRC_TODO(rcStrict));
+
+#ifdef VBOX_WITH_FIRST_IEM_STEP
+ /*
+ * Hand it over to the interpreter.
+ */
+ rcStrict = IEMExecOne(pVCpu);
+ LogFlow(("emR3HmExecuteIOInstruction: %Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
+ STAM_COUNTER_INC(&pVCpu->em.s.CTX_SUFF(pStats)->StatIoIem);
+ STAM_PROFILE_STOP(&pVCpu->em.s.StatIOEmu, a);
+ return VBOXSTRICTRC_TODO(rcStrict);
+
+#else
+ /** @todo probably we should fall back to the recompiler; otherwise we'll go back and forth between HC & GC
+ * as io instructions tend to come in packages of more than one
+ */
+ DISCPUSTATE Cpu;
+ int rc2 = CPUMR3DisasmInstrCPU(pVM, pVCpu, pCtx, pCtx->rip, &Cpu, "IO EMU");
+ if (RT_SUCCESS(rc2))
+ {
+ rcStrict = VINF_EM_RAW_EMULATE_INSTR;
+
+ if (!(Cpu.fPrefix & (DISPREFIX_REP | DISPREFIX_REPNE)))
+ {
+ switch (Cpu.pCurInstr->uOpcode)
+ {
+ case OP_IN:
+ {
+ STAM_COUNTER_INC(&pVCpu->em.s.CTX_SUFF(pStats)->StatIn);
+ rcStrict = IOMInterpretIN(pVM, pVCpu, CPUMCTX2CORE(pCtx), &Cpu);
+ break;
+ }
+
+ case OP_OUT:
+ {
+ STAM_COUNTER_INC(&pVCpu->em.s.CTX_SUFF(pStats)->StatOut);
+ rcStrict = IOMInterpretOUT(pVM, pVCpu, CPUMCTX2CORE(pCtx), &Cpu);
+ break;
+ }
+ }
+ }
+ else if (Cpu.fPrefix & DISPREFIX_REP)
+ {
+ switch (Cpu.pCurInstr->uOpcode)
+ {
+ case OP_INSB:
+ case OP_INSWD:
+ {
+ STAM_COUNTER_INC(&pVCpu->em.s.CTX_SUFF(pStats)->StatIn);
+ rcStrict = IOMInterpretINS(pVM, pVCpu, CPUMCTX2CORE(pCtx), &Cpu);
+ break;
+ }
+
+ case OP_OUTSB:
+ case OP_OUTSWD:
+ {
+ STAM_COUNTER_INC(&pVCpu->em.s.CTX_SUFF(pStats)->StatOut);
+ rcStrict = IOMInterpretOUTS(pVM, pVCpu, CPUMCTX2CORE(pCtx), &Cpu);
+ break;
+ }
+ }
+ }
+
+ /*
+ * Handled the I/O return codes.
+ * (The unhandled cases end up with rcStrict == VINF_EM_RAW_EMULATE_INSTR.)
+ */
+ if (IOM_SUCCESS(rcStrict))
+ {
+ pCtx->rip += Cpu.cbInstr;
+ STAM_PROFILE_STOP(&pVCpu->em.s.StatIOEmu, a);
+ LogFlow(("emR3HmExecuteIOInstruction: %Rrc 1\n", VBOXSTRICTRC_VAL(rcStrict)));
+ return VBOXSTRICTRC_TODO(rcStrict);
+ }
+
+ if (rcStrict == VINF_EM_RAW_GUEST_TRAP)
+ {
+ /* The active trap will be dispatched. */
+ Assert(TRPMHasTrap(pVCpu));
+ STAM_PROFILE_STOP(&pVCpu->em.s.StatIOEmu, a);
+ LogFlow(("emR3HmExecuteIOInstruction: VINF_SUCCESS 2\n"));
+ return VINF_SUCCESS;
+ }
+ AssertMsg(rcStrict != VINF_TRPM_XCPT_DISPATCHED, ("Handle VINF_TRPM_XCPT_DISPATCHED\n"));
+
+ if (RT_FAILURE(rcStrict))
+ {
+ STAM_PROFILE_STOP(&pVCpu->em.s.StatIOEmu, a);
+ LogFlow(("emR3HmExecuteIOInstruction: %Rrc 3\n", VBOXSTRICTRC_VAL(rcStrict)));
+ return VBOXSTRICTRC_TODO(rcStrict);
+ }
+ AssertMsg(rcStrict == VINF_EM_RAW_EMULATE_INSTR || rcStrict == VINF_EM_RESCHEDULE_REM, ("rcStrict=%Rrc\n", VBOXSTRICTRC_VAL(rcStrict)));
+ }
+
+ STAM_PROFILE_STOP(&pVCpu->em.s.StatIOEmu, a);
+ int rc3 = emR3HmExecuteInstruction(pVM, pVCpu, "IO: ");
+ LogFlow(("emR3HmExecuteIOInstruction: %Rrc 4 (rc2=%Rrc, rc3=%Rrc)\n", VBOXSTRICTRC_VAL(rcStrict), rc2, rc3));
+ return rc3;
+#endif
+}
+
+
+/**
+ * Process raw-mode specific forced actions.
+ *
+ * This function is called when any FFs in the VM_FF_HIGH_PRIORITY_PRE_RAW_MASK is pending.
+ *
+ * @returns VBox status code. May return VINF_EM_NO_MEMORY but none of the other
+ * EM statuses.
+ * @param pVM Pointer to the VM.
+ * @param pVCpu Pointer to the VMCPU.
+ * @param pCtx Pointer to the guest CPU context.
+ */
+static int emR3HmForcedActions(PVM pVM, PVMCPU pVCpu, PCPUMCTX pCtx)
+{
+ /*
+ * Sync page directory.
+ */
+ if (VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_PGM_SYNC_CR3 | VMCPU_FF_PGM_SYNC_CR3_NON_GLOBAL))
+ {
+ Assert(pVCpu->em.s.enmState != EMSTATE_WAIT_SIPI);
+ int rc = PGMSyncCR3(pVCpu, pCtx->cr0, pCtx->cr3, pCtx->cr4, VMCPU_FF_IS_SET(pVCpu, VMCPU_FF_PGM_SYNC_CR3));
+ if (RT_FAILURE(rc))
+ return rc;
+
+#ifdef VBOX_WITH_RAW_MODE
+ Assert(!VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_SELM_SYNC_GDT | VMCPU_FF_SELM_SYNC_LDT));
+#endif
+
+ /* Prefetch pages for EIP and ESP. */
+ /** @todo This is rather expensive. Should investigate if it really helps at all. */
+ rc = PGMPrefetchPage(pVCpu, SELMToFlat(pVM, DISSELREG_CS, CPUMCTX2CORE(pCtx), pCtx->rip));
+ if (rc == VINF_SUCCESS)
+ rc = PGMPrefetchPage(pVCpu, SELMToFlat(pVM, DISSELREG_SS, CPUMCTX2CORE(pCtx), pCtx->rsp));
+ if (rc != VINF_SUCCESS)
+ {
+ if (rc != VINF_PGM_SYNC_CR3)
+ {
+ AssertLogRelMsgReturn(RT_FAILURE(rc), ("%Rrc\n", rc), VERR_IPE_UNEXPECTED_INFO_STATUS);
+ return rc;
+ }
+ rc = PGMSyncCR3(pVCpu, pCtx->cr0, pCtx->cr3, pCtx->cr4, VMCPU_FF_IS_SET(pVCpu, VMCPU_FF_PGM_SYNC_CR3));
+ if (RT_FAILURE(rc))
+ return rc;
+ }
+ /** @todo maybe prefetch the supervisor stack page as well */
+#ifdef VBOX_WITH_RAW_MODE
+ Assert(!VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_SELM_SYNC_GDT | VMCPU_FF_SELM_SYNC_LDT));
+#endif
+ }
+
+ /*
+ * Allocate handy pages (just in case the above actions have consumed some pages).
+ */
+ if (VM_FF_IS_PENDING_EXCEPT(pVM, VM_FF_PGM_NEED_HANDY_PAGES, VM_FF_PGM_NO_MEMORY))
+ {
+ int rc = PGMR3PhysAllocateHandyPages(pVM);
+ if (RT_FAILURE(rc))
+ return rc;
+ }
+
+ /*
+ * Check whether we're out of memory now.
+ *
+ * This may stem from some of the above actions or operations that has been executed
+ * since we ran FFs. The allocate handy pages must for instance always be followed by
+ * this check.
+ */
+ if (VM_FF_IS_PENDING(pVM, VM_FF_PGM_NO_MEMORY))
+ return VINF_EM_NO_MEMORY;
+
+ return VINF_SUCCESS;
+}
+
+
+/**
+ * Executes hardware accelerated raw code. (Intel VT-x & AMD-V)
+ *
+ * This function contains the raw-mode version of the inner
+ * execution loop (the outer loop being in EMR3ExecuteVM()).
+ *
+ * @returns VBox status code. The most important ones are: VINF_EM_RESCHEDULE, VINF_EM_RESCHEDULE_RAW,
+ * VINF_EM_RESCHEDULE_REM, VINF_EM_SUSPEND, VINF_EM_RESET and VINF_EM_TERMINATE.
+ *
+ * @param pVM Pointer to the VM.
+ * @param pVCpu Pointer to the VMCPU.
+ * @param pfFFDone Where to store an indicator telling whether or not
+ * FFs were done before returning.
+ */
+int emR3HmExecute(PVM pVM, PVMCPU pVCpu, bool *pfFFDone)
+{
+ int rc = VERR_IPE_UNINITIALIZED_STATUS;
+ PCPUMCTX pCtx = pVCpu->em.s.pCtx;
+
+ LogFlow(("emR3HmExecute%d: (cs:eip=%04x:%RGv)\n", pVCpu->idCpu, pCtx->cs.Sel, (RTGCPTR)pCtx->rip));
+ *pfFFDone = false;
+
+ STAM_COUNTER_INC(&pVCpu->em.s.StatHmExecuteEntry);
+
+#ifdef EM_NOTIFY_HM
+ HMR3NotifyScheduled(pVCpu);
+#endif
+
+ /*
+ * Spin till we get a forced action which returns anything but VINF_SUCCESS.
+ */
+ for (;;)
+ {
+ STAM_PROFILE_ADV_START(&pVCpu->em.s.StatHmEntry, a);
+
+ /* Check if a forced reschedule is pending. */
+ if (HMR3IsRescheduleRequired(pVM, pCtx))
+ {
+ rc = VINF_EM_RESCHEDULE;
+ break;
+ }
+
+ /*
+ * Process high priority pre-execution raw-mode FFs.
+ */
+#ifdef VBOX_WITH_RAW_MODE
+ Assert(!VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_SELM_SYNC_TSS | VMCPU_FF_SELM_SYNC_GDT | VMCPU_FF_SELM_SYNC_LDT));
+#endif
+ if ( VM_FF_IS_PENDING(pVM, VM_FF_HIGH_PRIORITY_PRE_RAW_MASK)
+ || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HIGH_PRIORITY_PRE_RAW_MASK))
+ {
+ rc = emR3HmForcedActions(pVM, pVCpu, pCtx);
+ if (rc != VINF_SUCCESS)
+ break;
+ }
+
+#ifdef LOG_ENABLED
+ /*
+ * Log important stuff before entering GC.
+ */
+ if (TRPMHasTrap(pVCpu))
+ Log(("CPU%d: Pending hardware interrupt=0x%x cs:rip=%04X:%RGv\n", pVCpu->idCpu, TRPMGetTrapNo(pVCpu), pCtx->cs.Sel, (RTGCPTR)pCtx->rip));
+
+ uint32_t cpl = CPUMGetGuestCPL(pVCpu);
+
+ if (pVM->cCpus == 1)
+ {
+ if (pCtx->eflags.Bits.u1VM)
+ Log(("HWV86: %08X IF=%d\n", pCtx->eip, pCtx->eflags.Bits.u1IF));
+ else if (CPUMIsGuestIn64BitCodeEx(pCtx))
+ Log(("HWR%d: %04X:%RGv ESP=%RGv IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pCtx->cs.Sel, (RTGCPTR)pCtx->rip, pCtx->rsp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER));
+ else
+ Log(("HWR%d: %04X:%08X ESP=%08X IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pCtx->cs.Sel, pCtx->eip, pCtx->esp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER));
+ }
+ else
+ {
+ if (pCtx->eflags.Bits.u1VM)
+ Log(("HWV86-CPU%d: %08X IF=%d\n", pVCpu->idCpu, pCtx->eip, pCtx->eflags.Bits.u1IF));
+ else if (CPUMIsGuestIn64BitCodeEx(pCtx))
+ Log(("HWR%d-CPU%d: %04X:%RGv ESP=%RGv IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pVCpu->idCpu, pCtx->cs.Sel, (RTGCPTR)pCtx->rip, pCtx->rsp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER));
+ else
+ Log(("HWR%d-CPU%d: %04X:%08X ESP=%08X IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pVCpu->idCpu, pCtx->cs.Sel, pCtx->eip, pCtx->esp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER));
+ }
+#endif /* LOG_ENABLED */
+
+ /*
+ * Execute the code.
+ */
+ STAM_PROFILE_ADV_STOP(&pVCpu->em.s.StatHmEntry, a);
+
+ if (RT_LIKELY(emR3IsExecutionAllowed(pVM, pVCpu)))
+ {
+ STAM_PROFILE_START(&pVCpu->em.s.StatHmExec, x);
+ rc = VMMR3HmRunGC(pVM, pVCpu);
+ STAM_PROFILE_STOP(&pVCpu->em.s.StatHmExec, x);
+ }
+ else
+ {
+ /* Give up this time slice; virtual time continues */
+ STAM_REL_PROFILE_ADV_START(&pVCpu->em.s.StatCapped, u);
+ RTThreadSleep(5);
+ STAM_REL_PROFILE_ADV_STOP(&pVCpu->em.s.StatCapped, u);
+ rc = VINF_SUCCESS;
+ }
+
+
+ /*
+ * Deal with high priority post execution FFs before doing anything else.
+ */
+ VMCPU_FF_CLEAR(pVCpu, VMCPU_FF_RESUME_GUEST_MASK);
+ if ( VM_FF_IS_PENDING(pVM, VM_FF_HIGH_PRIORITY_POST_MASK)
+ || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_HIGH_PRIORITY_POST_MASK))
+ rc = emR3HighPriorityPostForcedActions(pVM, pVCpu, rc);
+
+ /*
+ * Process the returned status code.
+ */
+ if (rc >= VINF_EM_FIRST && rc <= VINF_EM_LAST)
+ break;
+
+ rc = emR3HmHandleRC(pVM, pVCpu, pCtx, rc);
+ if (rc != VINF_SUCCESS)
+ break;
+
+ /*
+ * Check and execute forced actions.
+ */
+#ifdef VBOX_HIGH_RES_TIMERS_HACK
+ TMTimerPollVoid(pVM, pVCpu);
+#endif
+ if ( VM_FF_IS_PENDING(pVM, VM_FF_ALL_MASK)
+ || VMCPU_FF_IS_PENDING(pVCpu, VMCPU_FF_ALL_MASK))
+ {
+ rc = emR3ForcedActions(pVM, pVCpu, rc);
+ VBOXVMM_EM_FF_ALL_RET(pVCpu, rc);
+ if ( rc != VINF_SUCCESS
+ && rc != VINF_EM_RESCHEDULE_HM)
+ {
+ *pfFFDone = true;
+ break;
+ }
+ }
+ }
+
+ /*
+ * Return to outer loop.
+ */
+#if defined(LOG_ENABLED) && defined(DEBUG)
+ RTLogFlush(NULL);
+#endif
+ return rc;
+}
+