VMM.cpp revision 8bb2a695da71f0eea8a6f88873fa78c1f350341b
/* $Id$ */
/** @file
* VMM - The Virtual Machine Monitor Core.
*/
/*
* Copyright (C) 2006-2007 Oracle Corporation
*
* This file is part of VirtualBox Open Source Edition (OSE), as
* available from http://www.virtualbox.org. This file is free software;
* General Public License (GPL) as published by the Free Software
* Foundation, in version 2 as it comes in the "COPYING" file of the
* VirtualBox OSE distribution. VirtualBox OSE is distributed in the
* hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
*/
//#define NO_SUPCALLR0VMM
/** @page pg_vmm VMM - The Virtual Machine Monitor
*
* The VMM component is two things at the moment, it's a component doing a few
* management and routing tasks, and it's the whole virtual machine monitor
* thing. For hysterical reasons, it is not doing all the management that one
* would expect, this is instead done by @ref pg_vm. We'll address this
* misdesign eventually.
*
* @see grp_vmm, grp_vm
*
*
* @section sec_vmmstate VMM State
*
* @image html VM_Statechart_Diagram.gif
*
* To be written.
*
*
* @subsection subsec_vmm_init VMM Initialization
*
* To be written.
*
*
* @subsection subsec_vmm_term VMM Termination
*
* To be written.
*
*/
/*******************************************************************************
* Header Files *
*******************************************************************************/
#define LOG_GROUP LOG_GROUP_VMM
#include <VBox/pdmqueue.h>
#include <VBox/pdmcritsect.h>
#include "VMMInternal.h"
#include "VMMSwitcher/VMMSwitcher.h"
#include <iprt/semaphore.h>
/*******************************************************************************
* Defined Constants And Macros *
*******************************************************************************/
/** The saved state version. */
#define VMM_SAVED_STATE_VERSION 4
/** The saved state version used by v3.0 and earlier. (Teleportation) */
#define VMM_SAVED_STATE_VERSION_3_0 3
/*******************************************************************************
* Internal Functions *
*******************************************************************************/
/**
* Initializes the VMM.
*
* @returns VBox status code.
* @param pVM The VM to operate on.
*/
{
LogFlow(("VMMR3Init\n"));
/*
* Assert alignment, sizes and order.
*/
/*
* Init basic VM VMM members.
*/
/** @cfgm{YieldEMTInterval, uint32_t, 1, UINT32_MAX, 23, ms}
* The EMT yield interval. The EMT yielding is a hack we employ to play a
* bit nicer with the rest of the system (like for instance the GUI).
*/
23 /* Value arrived at after experimenting with the grub boot prompt. */);
AssertMsgRCReturn(rc, ("Configuration error. Failed to query \"YieldEMTInterval\", rc=%Rrc\n", rc), rc);
/** @cfgm{VMM/UsePeriodicPreemptionTimers, boolean, true}
* Controls whether we employ per-cpu preemption timers to limit the time
* spent executing guest code. This option is not available on all
* platforms and we will silently ignore this setting then. If we are
* running in VT-x mode, we will use the VMX-preemption timer instead of
* this one when possible.
*/
rc = CFGMR3QueryBoolDef(pCfgVMM, "UsePeriodicPreemptionTimers", &pVM->vmm.s.fUsePeriodicPreemptionTimers, true);
AssertMsgRCReturn(rc, ("Configuration error. Failed to query \"VMM/UsePeriodicPreemptionTimers\", rc=%Rrc\n", rc), rc);
/*
* Initialize the VMM sync critical section and semaphores.
*/
pVM->vmm.s.pahEvtRendezvousEnterOrdered = (PRTSEMEVENT)MMR3HeapAlloc(pVM, MM_TAG_VMM, sizeof(RTSEMEVENT) * pVM->cCpus);
return VERR_NO_MEMORY;
{
}
/* GC switchers are enabled by default. Turned off by HWACCM. */
/*
* Register the saved state data unit.
*/
rc = SSMR3RegisterInternal(pVM, "vmm", 1, VMM_SAVED_STATE_VERSION, VMM_STACK_SIZE + sizeof(RTGCPTR),
if (RT_FAILURE(rc))
return rc;
/*
* Register the Ring-0 VM handle with the session for fast ioctl calls.
*/
if (RT_FAILURE(rc))
return rc;
/*
* Init various sub-components.
*/
if (RT_SUCCESS(rc))
{
if (RT_SUCCESS(rc))
{
#ifdef VBOX_WITH_NMI
/*
* Allocate mapping for the host APIC.
*/
if (RT_SUCCESS(rc))
{
}
#endif
if (RT_SUCCESS(rc))
{
/*
* Debug info and statistics.
*/
return VINF_SUCCESS;
}
}
/** @todo: Need failure cleanup. */
//more todo in here?
//if (RT_SUCCESS(rc))
//{
//}
//int rc2 = vmmR3TermCoreCode(pVM);
//AssertRC(rc2));
}
return rc;
}
/**
* Allocate & setup the VMM RC stack(s) (for EMTs).
*
* The stacks are also used for long jumps in Ring-0.
*
* @returns VBox status code.
* @param pVM Pointer to the shared VM structure.
*
* @remarks The optional guard page gets it protection setup up during R3 init
* completion because of init order issues.
*/
{
int rc = VINF_SUCCESS;
#ifdef VMM_R0_SWITCH_STACK
#else
#endif
{
#ifdef VBOX_STRICT_VMM_STACK
#else
#endif
if (RT_SUCCESS(rc))
{
#ifdef VBOX_STRICT_VMM_STACK
#endif
#ifdef VBOX_WITH_2X_4GB_ADDR_SPACE
/* MMHyperR3ToR0 returns R3 when not doing hardware assisted virtualization. */
if (!VMMIsHwVirtExtForced(pVM))
else
#endif
}
}
return rc;
}
/**
* Initialize the loggers.
*
* @returns VBox status code.
* @param pVM Pointer to the shared VM structure.
*/
{
int rc;
/*
* Allocate RC & R0 Logger instances (they are finalized in the relocator).
*/
#ifdef LOG_ENABLED
if (pLogger)
{
rc = MMR3HyperAllocOnceNoRel(pVM, pVM->vmm.s.cbRCLogger, 0, MM_TAG_VMM, (void **)&pVM->vmm.s.pRCLoggerR3);
if (RT_FAILURE(rc))
return rc;
# ifdef VBOX_WITH_R0_LOGGING
{
if (RT_FAILURE(rc))
return rc;
//pVCpu->vmm.s.pR0LoggerR3->fCreated = false;
}
# endif
}
#endif /* LOG_ENABLED */
#ifdef VBOX_WITH_RC_RELEASE_LOGGING
/*
* Allocate RC release logger instances (finalized in the relocator).
*/
if (pRelLogger)
{
rc = MMR3HyperAllocOnceNoRel(pVM, pVM->vmm.s.cbRCRelLogger, 0, MM_TAG_VMM, (void **)&pVM->vmm.s.pRCRelLoggerR3);
if (RT_FAILURE(rc))
return rc;
}
#endif /* VBOX_WITH_RC_RELEASE_LOGGING */
return VINF_SUCCESS;
}
/**
* VMMR3Init worker that register the statistics with STAM.
*
* @param pVM The shared VM structure.
*/
{
/*
* Statistics.
*/
STAM_REG(pVM, &pVM->vmm.s.StatRunRC, STAMTYPE_COUNTER, "/VMM/RunRC", STAMUNIT_OCCURENCES, "Number of context switches.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetNormal, STAMTYPE_COUNTER, "/VMM/RZRet/Normal", STAMUNIT_OCCURENCES, "Number of VINF_SUCCESS returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetInterrupt, STAMTYPE_COUNTER, "/VMM/RZRet/Interrupt", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_INTERRUPT returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetInterruptHyper, STAMTYPE_COUNTER, "/VMM/RZRet/InterruptHyper", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_INTERRUPT_HYPER returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetGuestTrap, STAMTYPE_COUNTER, "/VMM/RZRet/GuestTrap", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_GUEST_TRAP returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetRingSwitch, STAMTYPE_COUNTER, "/VMM/RZRet/RingSwitch", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_RING_SWITCH returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetRingSwitchInt, STAMTYPE_COUNTER, "/VMM/RZRet/RingSwitchInt", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_RING_SWITCH_INT returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetStaleSelector, STAMTYPE_COUNTER, "/VMM/RZRet/StaleSelector", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_STALE_SELECTOR returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetIRETTrap, STAMTYPE_COUNTER, "/VMM/RZRet/IRETTrap", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_IRET_TRAP returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetEmulate, STAMTYPE_COUNTER, "/VMM/RZRet/Emulate", STAMUNIT_OCCURENCES, "Number of VINF_EM_EXECUTE_INSTRUCTION returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetIOBlockEmulate, STAMTYPE_COUNTER, "/VMM/RZRet/EmulateIOBlock", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_EMULATE_IO_BLOCK returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetPatchEmulate, STAMTYPE_COUNTER, "/VMM/RZRet/PatchEmulate", STAMUNIT_OCCURENCES, "Number of VINF_PATCH_EMULATE_INSTR returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetIORead, STAMTYPE_COUNTER, "/VMM/RZRet/IORead", STAMUNIT_OCCURENCES, "Number of VINF_IOM_HC_IOPORT_READ returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetIOWrite, STAMTYPE_COUNTER, "/VMM/RZRet/IOWrite", STAMUNIT_OCCURENCES, "Number of VINF_IOM_HC_IOPORT_WRITE returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetMMIORead, STAMTYPE_COUNTER, "/VMM/RZRet/MMIORead", STAMUNIT_OCCURENCES, "Number of VINF_IOM_HC_MMIO_READ returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetMMIOWrite, STAMTYPE_COUNTER, "/VMM/RZRet/MMIOWrite", STAMUNIT_OCCURENCES, "Number of VINF_IOM_HC_MMIO_WRITE returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetMMIOReadWrite, STAMTYPE_COUNTER, "/VMM/RZRet/MMIOReadWrite", STAMUNIT_OCCURENCES, "Number of VINF_IOM_HC_MMIO_READ_WRITE returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetMMIOPatchRead, STAMTYPE_COUNTER, "/VMM/RZRet/MMIOPatchRead", STAMUNIT_OCCURENCES, "Number of VINF_IOM_HC_MMIO_PATCH_READ returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetMMIOPatchWrite, STAMTYPE_COUNTER, "/VMM/RZRet/MMIOPatchWrite", STAMUNIT_OCCURENCES, "Number of VINF_IOM_HC_MMIO_PATCH_WRITE returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetLDTFault, STAMTYPE_COUNTER, "/VMM/RZRet/LDTFault", STAMUNIT_OCCURENCES, "Number of VINF_EM_EXECUTE_INSTRUCTION_GDT_FAULT returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetGDTFault, STAMTYPE_COUNTER, "/VMM/RZRet/GDTFault", STAMUNIT_OCCURENCES, "Number of VINF_EM_EXECUTE_INSTRUCTION_LDT_FAULT returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetIDTFault, STAMTYPE_COUNTER, "/VMM/RZRet/IDTFault", STAMUNIT_OCCURENCES, "Number of VINF_EM_EXECUTE_INSTRUCTION_IDT_FAULT returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetTSSFault, STAMTYPE_COUNTER, "/VMM/RZRet/TSSFault", STAMUNIT_OCCURENCES, "Number of VINF_EM_EXECUTE_INSTRUCTION_TSS_FAULT returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetPDFault, STAMTYPE_COUNTER, "/VMM/RZRet/PDFault", STAMUNIT_OCCURENCES, "Number of VINF_EM_EXECUTE_INSTRUCTION_PD_FAULT returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetCSAMTask, STAMTYPE_COUNTER, "/VMM/RZRet/CSAMTask", STAMUNIT_OCCURENCES, "Number of VINF_CSAM_PENDING_ACTION returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetSyncCR3, STAMTYPE_COUNTER, "/VMM/RZRet/SyncCR", STAMUNIT_OCCURENCES, "Number of VINF_PGM_SYNC_CR3 returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetMisc, STAMTYPE_COUNTER, "/VMM/RZRet/Misc", STAMUNIT_OCCURENCES, "Number of misc returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetPatchInt3, STAMTYPE_COUNTER, "/VMM/RZRet/PatchInt3", STAMUNIT_OCCURENCES, "Number of VINF_PATM_PATCH_INT3 returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetPatchPF, STAMTYPE_COUNTER, "/VMM/RZRet/PatchPF", STAMUNIT_OCCURENCES, "Number of VINF_PATM_PATCH_TRAP_PF returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetPatchGP, STAMTYPE_COUNTER, "/VMM/RZRet/PatchGP", STAMUNIT_OCCURENCES, "Number of VINF_PATM_PATCH_TRAP_GP returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetPatchIretIRQ, STAMTYPE_COUNTER, "/VMM/RZRet/PatchIret", STAMUNIT_OCCURENCES, "Number of VINF_PATM_PENDING_IRQ_AFTER_IRET returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetRescheduleREM, STAMTYPE_COUNTER, "/VMM/RZRet/ScheduleREM", STAMUNIT_OCCURENCES, "Number of VINF_EM_RESCHEDULE_REM returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetToR3, STAMTYPE_COUNTER, "/VMM/RZRet/ToR3", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_TO_R3 returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetToR3Unknown, STAMTYPE_COUNTER, "/VMM/RZRet/ToR3/Unknown", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_TO_R3 returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetToR3TMVirt, STAMTYPE_COUNTER, "/VMM/RZRet/ToR3/TMVirt", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_TO_R3 returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetToR3HandyPages, STAMTYPE_COUNTER, "/VMM/RZRet/ToR3/Handy", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_TO_R3 returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetToR3PDMQueues, STAMTYPE_COUNTER, "/VMM/RZRet/ToR3/PDMQueue", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_TO_R3 returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetToR3Rendezvous, STAMTYPE_COUNTER, "/VMM/RZRet/ToR3/Rendezvous", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_TO_R3 returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetToR3Timer, STAMTYPE_COUNTER, "/VMM/RZRet/ToR3/Timer", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_TO_R3 returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetTimerPending, STAMTYPE_COUNTER, "/VMM/RZRet/TimerPending", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_TIMER_PENDING returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetInterruptPending, STAMTYPE_COUNTER, "/VMM/RZRet/InterruptPending", STAMUNIT_OCCURENCES, "Number of VINF_EM_RAW_INTERRUPT_PENDING returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetPATMDuplicateFn, STAMTYPE_COUNTER, "/VMM/RZRet/PATMDuplicateFn", STAMUNIT_OCCURENCES, "Number of VINF_PATM_DUPLICATE_FUNCTION returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetPGMChangeMode, STAMTYPE_COUNTER, "/VMM/RZRet/PGMChangeMode", STAMUNIT_OCCURENCES, "Number of VINF_PGM_CHANGE_MODE returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetPGMFlushPending, STAMTYPE_COUNTER, "/VMM/RZRet/PGMFlushPending", STAMUNIT_OCCURENCES, "Number of VINF_PGM_POOL_FLUSH_PENDING returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetPendingRequest, STAMTYPE_COUNTER, "/VMM/RZRet/PendingRequest", STAMUNIT_OCCURENCES, "Number of VINF_EM_PENDING_REQUEST returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetPatchTPR, STAMTYPE_COUNTER, "/VMM/RZRet/PatchTPR", STAMUNIT_OCCURENCES, "Number of VINF_EM_HWACCM_PATCH_TPR_INSTR returns.");
STAM_REG(pVM, &pVM->vmm.s.StatRZRetCallRing3, STAMTYPE_COUNTER, "/VMM/RZCallR3/Misc", STAMUNIT_OCCURENCES, "Number of Other ring-3 calls.");
STAM_REG(pVM, &pVM->vmm.s.StatRZCallPDMLock, STAMTYPE_COUNTER, "/VMM/RZCallR3/PDMLock", STAMUNIT_OCCURENCES, "Number of VMMCALLRING3_PDM_LOCK calls.");
STAM_REG(pVM, &pVM->vmm.s.StatRZCallPGMLock, STAMTYPE_COUNTER, "/VMM/RZCallR3/PGMLock", STAMUNIT_OCCURENCES, "Number of VMMCALLRING3_PGM_LOCK calls.");
STAM_REG(pVM, &pVM->vmm.s.StatRZCallPGMPoolGrow, STAMTYPE_COUNTER, "/VMM/RZCallR3/PGMPoolGrow", STAMUNIT_OCCURENCES, "Number of VMMCALLRING3_PGM_POOL_GROW calls.");
STAM_REG(pVM, &pVM->vmm.s.StatRZCallPGMMapChunk, STAMTYPE_COUNTER, "/VMM/RZCallR3/PGMMapChunk", STAMUNIT_OCCURENCES, "Number of VMMCALLRING3_PGM_MAP_CHUNK calls.");
STAM_REG(pVM, &pVM->vmm.s.StatRZCallPGMAllocHandy, STAMTYPE_COUNTER, "/VMM/RZCallR3/PGMAllocHandy", STAMUNIT_OCCURENCES, "Number of VMMCALLRING3_PGM_ALLOCATE_HANDY_PAGES calls.");
STAM_REG(pVM, &pVM->vmm.s.StatRZCallRemReplay, STAMTYPE_COUNTER, "/VMM/RZCallR3/REMReplay", STAMUNIT_OCCURENCES, "Number of VMMCALLRING3_REM_REPLAY_HANDLER_NOTIFICATIONS calls.");
STAM_REG(pVM, &pVM->vmm.s.StatRZCallLogFlush, STAMTYPE_COUNTER, "/VMM/RZCallR3/VMMLogFlush", STAMUNIT_OCCURENCES, "Number of VMMCALLRING3_VMM_LOGGER_FLUSH calls.");
STAM_REG(pVM, &pVM->vmm.s.StatRZCallVMSetError, STAMTYPE_COUNTER, "/VMM/RZCallR3/VMSetError", STAMUNIT_OCCURENCES, "Number of VMMCALLRING3_VM_SET_ERROR calls.");
STAM_REG(pVM, &pVM->vmm.s.StatRZCallVMSetRuntimeError, STAMTYPE_COUNTER, "/VMM/RZCallR3/VMRuntimeError", STAMUNIT_OCCURENCES, "Number of VMMCALLRING3_VM_SET_RUNTIME_ERROR calls.");
#ifdef VBOX_WITH_STATISTICS
{
STAMR3RegisterF(pVM, &pVM->aCpus[i].vmm.s.CallRing3JmpBufR0.cbUsedMax, STAMTYPE_U32_RESET, STAMVISIBILITY_ALWAYS, STAMUNIT_BYTES, "Max amount of stack used.", "/VMM/Stack/CPU%u/Max", i);
STAMR3RegisterF(pVM, &pVM->aCpus[i].vmm.s.CallRing3JmpBufR0.cbUsedAvg, STAMTYPE_U32, STAMVISIBILITY_ALWAYS, STAMUNIT_BYTES, "Average stack usage.", "/VMM/Stack/CPU%u/Avg", i);
STAMR3RegisterF(pVM, &pVM->aCpus[i].vmm.s.CallRing3JmpBufR0.cUsedTotal, STAMTYPE_U64, STAMVISIBILITY_ALWAYS, STAMUNIT_OCCURENCES, "Number of stack usages.", "/VMM/Stack/CPU%u/Uses", i);
}
#endif
}
/**
* Initializes the per-VCPU VMM.
*
* @returns VBox status code.
* @param pVM The VM to operate on.
*/
{
LogFlow(("VMMR3InitCPU\n"));
return VINF_SUCCESS;
}
/**
* Ring-3 init finalizing.
*
* @returns VBox status code.
* @param pVM The VM handle.
*/
{
int rc;
/*
* Set page attributes to r/w for stack pages.
*/
{
}
/*
* Create the EMT yield timer.
*/
rc = TMR3TimerCreateInternal(pVM, TMCLOCK_REAL, vmmR3YieldEMT, NULL, "EMT Yielder", &pVM->vmm.s.pYieldTimer);
#ifdef VBOX_WITH_NMI
/*
*/
#endif
#ifdef VBOX_STRICT_VMM_STACK
/*
* Setup the stack guard pages: Two inaccessible pages at each sides of the
* stack to catch over/under-flows.
*/
{
}
#endif
/*
* Disable the periodic preemption timers if we can use the VMX-preemption
* timer instead.
*/
return VINF_SUCCESS;
}
/**
* Initializes the R0 VMM.
*
* @returns VBox status code.
* @param pVM The VM to operate on.
*/
{
int rc;
#ifdef LOG_ENABLED
/*
* Initialize the ring-0 logger if we haven't done so yet.
*/
{
if (RT_FAILURE(rc))
return rc;
}
#endif
/*
* Call Ring-0 entry with init code.
*/
for (;;)
{
#ifdef NO_SUPCALLR0VMM
//rc = VERR_GENERAL_FAILURE;
rc = VINF_SUCCESS;
#else
#endif
/*
* Flush the logs.
*/
#ifdef LOG_ENABLED
#endif
if (rc != VINF_VMM_CALL_HOST)
break;
break;
/* Resume R0 */
}
{
if (RT_SUCCESS(rc))
}
return rc;
}
/**
* Initializes the RC VMM.
*
* @returns VBox status code.
* @param pVM The VM to operate on.
*/
{
/* In VMX mode, there's no need to init RC. */
return VINF_SUCCESS;
/*
* Call VMMGCInit():
* -# resolve the address.
* -# setup stackframe and EIP to use the trampoline.
* -# do a generic hypervisor call.
*/
if (RT_SUCCESS(rc))
{
for (;;)
{
#ifdef NO_SUPCALLR0VMM
//rc = VERR_GENERAL_FAILURE;
rc = VINF_SUCCESS;
#else
#endif
#ifdef LOG_ENABLED
if ( pLogger
&& pLogger->offScratch > 0)
#endif
#ifdef VBOX_WITH_RC_RELEASE_LOGGING
#endif
if (rc != VINF_VMM_CALL_HOST)
break;
break;
}
{
}
}
return rc;
}
/**
* Terminate the VMM bits.
*
* @returns VINF_SUCCESS.
* @param pVM The VM handle.
*/
{
/*
* Call Ring-0 entry with termination code.
*/
int rc;
for (;;)
{
#ifdef NO_SUPCALLR0VMM
//rc = VERR_GENERAL_FAILURE;
rc = VINF_SUCCESS;
#else
#endif
/*
* Flush the logs.
*/
#ifdef LOG_ENABLED
#endif
if (rc != VINF_VMM_CALL_HOST)
break;
break;
/* Resume R0 */
}
{
if (RT_SUCCESS(rc))
}
{
}
#ifdef VBOX_STRICT_VMM_STACK
/*
* Make the two stack guard pages present again.
*/
{
{
}
}
#endif
return rc;
}
/**
* Terminates the per-VCPU VMM.
*
* Termination means cleaning up and freeing all resources,
* the VM it self is at this point powered off or suspended.
*
* @returns VBox status code.
* @param pVM The VM to operate on.
*/
{
return VINF_SUCCESS;
}
/**
* Applies relocations to data and code managed by this
* component. This function will be called at init and
* whenever the VMM need to relocate it self inside the GC.
*
* The VMM will need to apply relocations to the core code.
*
* @param pVM The VM handle.
* @param offDelta The relocation delta.
*/
{
/*
* Recalc the RC address.
*/
#ifdef VBOX_WITH_RAW_MODE
#endif
/*
* The stack.
*/
{
}
/*
* All the switchers.
*/
/*
* Get other RC entry points.
*/
int rc = PDMR3LdrGetSymbolRC(pVM, VMMGC_MAIN_MODULE_NAME, "CPUMGCResumeGuest", &pVM->vmm.s.pfnCPUMRCResumeGuest);
rc = PDMR3LdrGetSymbolRC(pVM, VMMGC_MAIN_MODULE_NAME, "CPUMGCResumeGuestV86", &pVM->vmm.s.pfnCPUMRCResumeGuestV86);
/*
* Update the logger.
*/
}
/**
* Updates the settings for the RC and R0 loggers.
*
* @returns VBox status code.
* @param pVM The VM handle.
*/
{
/*
* Simply clone the logger instance (for RC).
*/
int rc = VINF_SUCCESS;
RTRCPTR RCPtrLoggerFlush = 0;
#ifdef VBOX_WITH_RC_RELEASE_LOGGING
#endif
)
{
}
{
}
#ifdef VBOX_WITH_RC_RELEASE_LOGGING
{
rc = PDMR3LdrGetSymbolRC(pVM, VMMGC_MAIN_MODULE_NAME, "vmmGCRelLoggerWrapper", &RCPtrLoggerWrapper);
}
#endif /* VBOX_WITH_RC_RELEASE_LOGGING */
#ifdef LOG_ENABLED
/*
* For the ring-0 EMT logger, we use a per-thread logger instance
* in ring-0. Only initialize it once.
*/
{
if (pR0LoggerR3)
{
if (!pR0LoggerR3->fCreated)
{
pR0LoggerR3->idCpu = i;
pR0LoggerR3->fCreated = true;
pR0LoggerR3->fFlushingDisabled = false;
}
rc = RTLogCopyGroupsAndFlags(&pR0LoggerR3->Logger, NULL /* default */, pVM->vmm.s.pRCLoggerR3->fFlags, RTLOGFLAGS_BUFFERED);
}
}
#endif
return rc;
}
/**
*
* @returns Pointer to the buffer.
* @param pVM The VM handle.
*/
{
if (HWACCMIsEnabled(pVM))
if (RT_SUCCESS(rc))
return NULL;
}
/**
*
* @returns Pointer to the buffer.
* @param pVM The VM handle.
*/
{
if (HWACCMIsEnabled(pVM))
if (RT_SUCCESS(rc))
return NULL;
}
/**
* Execute state save operation.
*
* @returns VBox status code.
* @param pVM VM Handle.
* @param pSSM SSM operation handle.
*/
{
LogFlow(("vmmR3Save:\n"));
/*
* be running. This avoids breaking the saved state version. :-)
*/
}
/**
* Execute state load operation.
*
* @returns VBox status code.
* @param pVM VM Handle.
* @param pSSM SSM operation handle.
* @param uVersion Data layout version.
* @param uPass The data pass.
*/
{
LogFlow(("vmmR3Load:\n"));
/*
* Validate version.
*/
if ( uVersion != VMM_SAVED_STATE_VERSION
{
}
if (uVersion <= VMM_SAVED_STATE_VERSION_3_0)
{
/* Ignore the stack bottom, stack pointer and stack bits. */
#ifdef RT_OS_DARWIN
)
else
#else
#endif
}
/*
* Restore the VMCPU states. VCPU 0 is always started.
*/
{
bool fStarted;
if (RT_FAILURE(rc))
return rc;
}
/* terminator */
if (RT_FAILURE(rc))
return rc;
if (u32 != UINT32_MAX)
{
}
return VINF_SUCCESS;
}
/**
* Resolve a builtin RC symbol.
*
* Called by PDM when loading or relocating RC modules.
*
* @returns VBox status
* @param pVM VM Handle.
* @param pszSymbol Symbol to resolv
* @param pRCPtrValue Where to store the symbol value.
*
* @remark This has to work before VMMR3Relocate() is called.
*/
{
{
}
{
#ifdef VBOX_WITH_RC_RELEASE_LOGGING
#else
#endif
}
else
return VERR_SYMBOL_NOT_FOUND;
return VINF_SUCCESS;
}
/**
* Suspends the CPU yielder.
*
* @param pVM The VM handle.
*/
{
{
else
}
}
/**
* Stops the CPU yielder.
*
* @param pVM The VM handle.
*/
{
}
/**
* Resumes the CPU yielder when it has been a suspended or stopped.
*
* @param pVM The VM handle.
*/
{
{
}
}
/**
* Internal timer callback function.
*
* @param pVM The VM.
* @param pTimer The timer handle.
* @param pvUser User argument specified upon timer creation.
*/
{
/*
* This really needs some careful tuning. While we shouldn't be too greedy since
* that'll cause the rest of the system to stop up, we shouldn't be too nice either
* because that'll cause us to stop up.
*
* The current logic is to use the default interval when there is no lag worth
* mentioning, but when we start accumulating lag we don't bother yielding at all.
*
* (This depends on the TMCLOCK_VIRTUAL_SYNC to be scheduled before TMCLOCK_REAL
* so the lag is up to date.)
*/
)
{
#ifdef LOG_ENABLED
#endif
}
}
/**
* Executes guest code in the raw-mode context.
*
* @param pVM VM handle.
* @param pVCpu The VMCPU to operate on.
*/
{
/*
* Set the EIP and ESP.
*/
/*
* We hide log flushes (outer) and hypervisor interrupts (inner).
*/
for (;;)
{
#ifdef VBOX_STRICT
#endif
int rc;
do
{
#ifdef NO_SUPCALLR0VMM
#else
#endif
} while (rc == VINF_EM_RAW_INTERRUPT_HYPER);
/*
* Flush the logs.
*/
#ifdef LOG_ENABLED
if ( pLogger
&& pLogger->offScratch > 0)
#endif
#ifdef VBOX_WITH_RC_RELEASE_LOGGING
#endif
if (rc != VINF_VMM_CALL_HOST)
{
Log2(("VMMR3RawRunGC: returns %Rrc (cs:eip=%04x:%08x)\n", rc, CPUMGetGuestCS(pVCpu), CPUMGetGuestEIP(pVCpu)));
return rc;
}
if (RT_FAILURE(rc))
return rc;
/* Resume GC */
}
}
/**
* Executes guest code (Intel VT-x and AMD-V).
*
* @param pVM VM handle.
* @param pVCpu The VMCPU to operate on.
*/
{
for (;;)
{
int rc;
do
{
#ifdef NO_SUPCALLR0VMM
#else
#endif
} while (rc == VINF_EM_RAW_INTERRUPT_HYPER);
#ifdef LOG_ENABLED
/*
* Flush the log
*/
if ( pR0LoggerR3
#endif /* !LOG_ENABLED */
if (rc != VINF_VMM_CALL_HOST)
{
Log2(("VMMR3HwAccRunGC: returns %Rrc (cs:eip=%04x:%08x)\n", rc, CPUMGetGuestCS(pVCpu), CPUMGetGuestEIP(pVCpu)));
return rc;
}
if (RT_FAILURE(rc))
return rc;
/* Resume R0 */
}
}
/**
* VCPU worker for VMMSendSipi.
*
* @param pVM The VM to operate on.
* @param idCpu Virtual CPU to perform SIPI on
* @param uVector SIPI vector
*/
{
/** @todo what are we supposed to do if the processor is already running? */
return VERR_ACCESS_DENIED;
# if 1 /* If we keep the EMSTATE_WAIT_SIPI method, then move this to EM.cpp. */
return VINF_EM_RESCHEDULE;
# else /* And if we go the VMCPU::enmState way it can stay here. */
return VINF_SUCCESS;
# endif
}
{
return VINF_EM_WAIT_SIPI;
}
/**
* Sends SIPI to the virtual CPU by setting CS:EIP into vector-dependent state
* and unhalting processor
*
* @param pVM The VM to operate on.
* @param idCpu Virtual CPU to perform SIPI on
* @param uVector SIPI vector
*/
{
}
/**
* Sends init IPI to the virtual CPU.
*
* @param pVM The VM to operate on.
* @param idCpu Virtual CPU to perform int IPI on
*/
{
}
/**
* Registers the guest memory range that can be used for patching
*
* @returns VBox status code.
* @param pVM The VM to operate on.
* @param pPatchMem Patch memory range
* @param cbPatchMem Size of the memory range
*/
{
if (HWACCMIsEnabled(pVM))
return VERR_NOT_SUPPORTED;
}
/**
* Deregisters the guest memory range that can be used for patching
*
* @returns VBox status code.
* @param pVM The VM to operate on.
* @param pPatchMem Patch memory range
* @param cbPatchMem Size of the memory range
*/
{
if (HWACCMIsEnabled(pVM))
return VINF_SUCCESS;
}
/**
* VCPU worker for VMMR3SynchronizeAllVCpus.
*
* @param pVM The VM to operate on.
* @param idCpu Virtual CPU to perform SIPI on
* @param uVector SIPI vector
*/
{
/* Block until the job in the caller has finished. */
return VINF_SUCCESS;
}
/**
* Atomically execute a callback handler
* Note: This is very expensive; avoid using it frequently!
*
* @param pVM The VM to operate on.
* @param pfnHandler Callback handler
* @param pvUser User specified parameter
*
* @thread EMT
*/
{
int rc;
/* Shortcut for the uniprocessor case. */
{
{
}
}
/* Wait until all other VCPUs are waiting for us. */
RTThreadSleep(1);
return rc;
}
/**
* Count returns and have the last non-caller EMT wake up the caller.
*
* @returns VBox strict informational status code for EM scheduling. No failures
* will be returned here, those are for the caller only.
*
* @param pVM The VM handle.
*/
{
{
}
("%Rrc\n", rcRet),
}
/**
* Common worker for VMMR3EmtRendezvous and VMMR3EmtRendezvousFF.
*
* @returns VBox strict informational status code for EM scheduling. No failures
* will be returned here, those are for the caller only. When
* fIsCaller is set, VINF_SUCESS is always returned.
*
* @param pVM The VM handle.
* @param pVCpu The VMCPU structure for the calling EMT.
* @param fIsCaller Whether we're the VMMR3EmtRendezvous caller or
* not.
* @param fFlags The flags.
* @param pfnRendezvous The callback.
* @param pvUser The user argument for the callback.
*/
{
int rc;
/*
* Enter, the last EMT triggers the next callback phase.
*/
{
{
/* Wait for our turn. */
}
{
/* Wait for the last EMT to arrive and wake everyone up. */
}
{
/* Wait for our turn. */
}
else
{
/*
* The execute once is handled specially to optimize the code flow.
*
* The last EMT to arrive will perform the callback and the other
* EMTs will wait on the Done/DoneCaller semaphores (instead of
* the EnterOneByOne/AllAtOnce) in the meanwhile. When the callback
* returns, that EMT will initiate the normal return sequence.
*/
if (!fIsCaller)
{
return vmmR3EmtRendezvousNonCallerReturn(pVM);
}
return VINF_SUCCESS;
}
}
else
{
/*
* All EMTs are waiting, clear the FF and take action according to the
* execution method.
*/
{
/* Wake up everyone. */
}
{
/* Figure out who to wake up and wake it up. If it's ourself, then
it's easy otherwise wait for our turn. */
VMCPUID iFirst = (fFlags & VMMEMTRENDEZVOUS_FLAGS_TYPE_MASK) == VMMEMTRENDEZVOUS_FLAGS_TYPE_ASCENDING
? 0
{
}
}
/* else: execute the handler on the current EMT and wake up one or more threads afterwards. */
}
/*
* Do the callback and update the status if necessary.
*/
if ( !(fFlags & VMMEMTRENDEZVOUS_FLAGS_STOP_ON_ERROR)
{
if (rcStrict != VINF_SUCCESS)
{
do
{
if ( rcStrict == i32RendezvousStatus
|| ( i32RendezvousStatus != VINF_SUCCESS
&& rcStrict > i32RendezvousStatus))
break;
} while (!ASMAtomicCmpXchgS32(&pVM->vmm.s.i32RendezvousStatus, VBOXSTRICTRC_VAL(rcStrict), i32RendezvousStatus));
}
}
/*
* Increment the done counter and take action depending on whether we're
* the last to finish callback execution.
*/
{
/* Signal the next EMT? */
{
}
{
}
{
}
/* Wait for the rest to finish (the caller waits on hEvtRendezvousDoneCaller). */
if (!fIsCaller)
{
}
}
else
{
/* Callback execution is all done, tell the rest to return. */
}
if (!fIsCaller)
return vmmR3EmtRendezvousNonCallerReturn(pVM);
return VINF_SUCCESS;
}
/**
* Called in response to VM_FF_EMT_RENDEZVOUS.
*
* @returns VBox strict status code - EM scheduling. No errors will be returned
* here, nor will any non-EM scheduling status codes be returned.
*
* @param pVM The VM handle
* @param pVCpu The handle of the calling EMT.
*
* @thread EMT
*/
{
}
/**
* EMT rendezvous.
*
* Gathers all the EMTs and execute some code on each of them, either in a one
* by one fashion or all at once.
*
* @returns VBox strict status code. This will be the the first error,
* VINF_SUCCESS, or an EM scheduling status code.
*
* @param pVM The VM handle.
* @param fFlags Flags indicating execution methods. See
* grp_VMMR3EmtRendezvous_fFlags.
* @param pfnRendezvous The callback.
* @param pvUser User argument for the callback.
*
* @thread Any.
*/
VMMR3DECL(int) VMMR3EmtRendezvous(PVM pVM, uint32_t fFlags, PFNVMMEMTRENDEZVOUS pfnRendezvous, void *pvUser)
{
/*
* Validate input.
*/
if (!pVCpu)
/*
* Forward the request to an EMT thread.
*/
/*
* Shortcut for the single EMT case.
*/
else
{
/*
* Spin lock. If busy, wait for the other EMT to finish while keeping a
* lookout of the RENDEZVOUS FF.
*/
int rc;
{
{
{
if ( rc != VINF_SUCCESS
&& ( rcStrict == VINF_SUCCESS
/** @todo Perhaps deal with termination here? */
}
ASMNopPause();
}
}
/*
* Clear the slate. This is a semaphore ping-pong orgy. :-)
*/
{
}
rc = RTSemEventWait(pVM->vmm.s.hEvtRendezvousEnterOneByOne, 0); AssertLogRelMsg(rc == VERR_TIMEOUT || rc == VINF_SUCCESS, ("%Rrc\n", rc));
rc = RTSemEventWait(pVM->vmm.s.hEvtRendezvousDoneCaller, 0); AssertLogRelMsg(rc == VERR_TIMEOUT || rc == VINF_SUCCESS, ("%Rrc\n", rc));
/*
* Set the FF and poke the other EMTs.
*/
/*
* Do the same ourselves.
*/
/*
* The caller waits for the other EMTs to be done and return before doing
* the cleanup. This makes away with wakeup / reset races we would otherwise
* risk in the multiple release event semaphore code (hEvtRendezvousDoneCaller).
*/
/*
* Get the return code and clean up a little bit.
*/
/*
* Merge rcStrict and rcMy.
*/
if ( rcMy != VINF_SUCCESS
&& ( rcStrict == VINF_SUCCESS
}
return VBOXSTRICTRC_VAL(rcStrict);
}
/**
* Read from the ring 0 jump buffer stack
*
* @returns VBox status code.
*
* @param pVM Pointer to the shared VM structure.
* @param idCpu The ID of the source CPU context (for the address).
* @param R0Addr Where to start reading.
* @param pvBuf Where to store the data we've read.
* @param cbRead The number of bytes to read.
*/
VMMR3DECL(int) VMMR3ReadR0Stack(PVM pVM, VMCPUID idCpu, RTHCUINTPTR R0Addr, void *pvBuf, size_t cbRead)
{
#ifdef VMM_R0_SWITCH_STACK
#else
RTHCUINTPTR off = pVCpu->vmm.s.CallRing3JmpBufR0.cbSavedStack - (pVCpu->vmm.s.CallRing3JmpBufR0.SpCheck - R0Addr);
#endif
if ( off > VMM_STACK_SIZE
return VERR_INVALID_POINTER;
return VINF_SUCCESS;
}
/**
* Calls a RC function.
*
* @param pVM The VM handle.
* @param RCPtrEntry The address of the RC function.
* @param cArgs The number of arguments in the ....
* @param ... Arguments to the function.
*/
{
return rc;
}
/**
* Calls a RC function.
*
* @param pVM The VM handle.
* @param RCPtrEntry The address of the RC function.
* @param cArgs The number of arguments in the ....
* @param args Arguments to the function.
*/
{
/* Raw mode implies 1 VCPU. */
/*
* Setup the call frame using the trampoline.
*/
int i = cArgs;
while (i-- > 0)
/*
* We hide log flushes (outer) and hypervisor interrupts (inner).
*/
for (;;)
{
int rc;
do
{
#ifdef NO_SUPCALLR0VMM
#else
#endif
} while (rc == VINF_EM_RAW_INTERRUPT_HYPER);
/*
* Flush the logs.
*/
#ifdef LOG_ENABLED
if ( pLogger
&& pLogger->offScratch > 0)
#endif
#ifdef VBOX_WITH_RC_RELEASE_LOGGING
#endif
if (rc != VINF_VMM_CALL_HOST)
{
Log2(("VMMR3CallGCV: returns %Rrc (cs:eip=%04x:%08x)\n", rc, CPUMGetGuestCS(pVCpu), CPUMGetGuestEIP(pVCpu)));
return rc;
}
if (RT_FAILURE(rc))
return rc;
}
}
/**
* Wrapper for SUPR3CallVMMR0Ex which will deal with VINF_VMM_CALL_HOST returns.
*
* @returns VBox status code.
* @param pVM The VM to operate on.
* @param uOperation Operation to execute.
* @param u64Arg Constant argument.
* @param pReqHdr Pointer to a request header. See SUPR3CallVMMR0Ex for
* details.
*/
{
/*
* Call Ring-0 entry with init code.
*/
int rc;
for (;;)
{
#ifdef NO_SUPCALLR0VMM
#else
#endif
/*
* Flush the logs.
*/
#ifdef LOG_ENABLED
#endif
if (rc != VINF_VMM_CALL_HOST)
break;
break;
/* Resume R0 */
}
return rc;
}
/**
* Resumes executing hypervisor code when interrupted by a queue flush or a
* debug event.
*
* @returns VBox status code.
* @param pVM VM handle.
* @param pVCpu VMCPU handle.
*/
{
/*
* We hide log flushes (outer) and hypervisor interrupts (inner).
*/
for (;;)
{
int rc;
do
{
#ifdef NO_SUPCALLR0VMM
#else
#endif
} while (rc == VINF_EM_RAW_INTERRUPT_HYPER);
/*
* Flush the loggers,
*/
#ifdef LOG_ENABLED
if ( pLogger
&& pLogger->offScratch > 0)
#endif
#ifdef VBOX_WITH_RC_RELEASE_LOGGING
#endif
if (rc != VINF_VMM_CALL_HOST)
{
return rc;
}
if (RT_FAILURE(rc))
return rc;
}
}
/**
* Service a call to the ring-3 host code.
*
* @returns VBox status code.
* @param pVM VM handle.
* @param pVCpu VMCPU handle
* @remark Careful with critsects.
*/
{
/*
* We must also check for pending critsect exits or else we can deadlock
* when entering other critsects here.
*/
{
/*
* Acquire the PDM lock.
*/
case VMMCALLRING3_PDM_LOCK:
{
break;
}
/*
* Grow the PGM pool.
*/
{
break;
}
/*
* Maps an page allocation chunk into ring-3 so ring-0 can use it.
*/
{
break;
}
/*
* Allocates more handy pages.
*/
{
break;
}
/*
* Allocates a large page.
*/
{
break;
}
/*
* Acquire the PGM lock.
*/
case VMMCALLRING3_PGM_LOCK:
{
break;
}
/*
* Acquire the MM hypervisor heap lock.
*/
{
break;
}
/*
* Flush REM handler notifications.
*/
{
break;
}
/*
* This is a noop. We just take this route to avoid unnecessary
* tests in the loops.
*/
LogAlways(("*FLUSH*\n"));
break;
/*
* Set the VM error message.
*/
break;
/*
* Set the VM runtime error message.
*/
break;
/*
* Signal a ring 0 hypervisor assertion.
* Cancel the longjmp operation that's in progress.
*/
#ifdef RT_ARCH_X86
#else
#endif
#ifdef VMM_R0_SWITCH_STACK
#endif
return VERR_VMM_RING0_ASSERTION;
/*
* A forced switch to ring 0 for preemption purposes.
*/
break;
pVCpu->vmm.s.rcCallRing3 = FTMR3SetCheckpoint(pVM, (FTMCHECKPOINTTYPE)pVCpu->vmm.s.u64CallRing3Arg);
break;
default:
return VERR_INTERNAL_ERROR;
}
return VINF_SUCCESS;
}
/**
* Displays the Force action Flags.
*
* @param pVM The VM handle.
* @param pHlp The output helpers.
* @param pszArgs The additional arguments (ignored).
*/
{
int c;
uint32_t f;
{ \
if (!(c % 6)) \
else \
c++; \
} \
} while (0)
{ \
if (!(c % 5)) \
else \
c++; \
} \
} while (0)
/*
* The global flags.
*/
/* show the flag mnemonics */
c = 0;
f = fGlobalForcedActions;
if (f)
else
/* the groups */
c = 0;
f = fGlobalForcedActions;
if (c)
/*
* Per CPU flags.
*/
{
/* show the flag mnemonics */
c = 0;
f = fLocalForcedActions;
if (f)
else
/* the groups */
c = 0;
f = fLocalForcedActions;
if (c)
}
}