vmm.h revision ae20b83f0c94402a3e3ac021c3d4e5f827e4905c
/** @file
* VMM - The Virtual Machine Monitor.
*/
/*
* Copyright (C) 2006-2013 Oracle Corporation
*
* This file is part of VirtualBox Open Source Edition (OSE), as
* available from http://www.virtualbox.org. This file is free software;
* you can redistribute it and/or modify it under the terms of the GNU
* General Public License (GPL) as published by the Free Software
* Foundation, in version 2 as it comes in the "COPYING" file of the
* VirtualBox OSE distribution. VirtualBox OSE is distributed in the
* hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
*
* The contents of this file may alternatively be used under the terms
* of the Common Development and Distribution License Version 1.0
* (CDDL) only, as it comes in the "COPYING.CDDL" file of the
* VirtualBox OSE distribution, in which case the provisions of the
* CDDL are applicable instead of those of the GPL.
*
* You may elect to license modified versions of this file under the
* terms and conditions of either the GPL or the CDDL or both.
*/
#ifndef ___VBox_vmm_vmm_h
#define ___VBox_vmm_vmm_h
#include <VBox/types.h>
#include <VBox/vmm/vmapi.h>
#include <VBox/sup.h>
#include <VBox/log.h>
#include <iprt/stdarg.h>
RT_C_DECLS_BEGIN
/** @defgroup grp_vmm The Virtual Machine Monitor API
* @{
*/
/**
* World switcher identifiers.
*/
typedef enum VMMSWITCHER
{
/** The usual invalid 0. */
VMMSWITCHER_INVALID = 0,
/** Switcher for 32-bit host to 32-bit shadow paging. */
VMMSWITCHER_32_TO_32,
/** Switcher for 32-bit host paging to PAE shadow paging. */
VMMSWITCHER_32_TO_PAE,
/** Switcher for 32-bit host paging to AMD64 shadow paging. */
VMMSWITCHER_32_TO_AMD64,
/** Switcher for PAE host to 32-bit shadow paging. */
VMMSWITCHER_PAE_TO_32,
/** Switcher for PAE host to PAE shadow paging. */
VMMSWITCHER_PAE_TO_PAE,
/** Switcher for PAE host paging to AMD64 shadow paging. */
VMMSWITCHER_PAE_TO_AMD64,
/** Switcher for AMD64 host paging to 32-bit shadow paging. */
VMMSWITCHER_AMD64_TO_32,
/** Switcher for AMD64 host paging to PAE shadow paging. */
VMMSWITCHER_AMD64_TO_PAE,
/** Switcher for AMD64 host paging to AMD64 shadow paging. */
VMMSWITCHER_AMD64_TO_AMD64,
/** Stub switcher for 32-bit and PAE. */
VMMSWITCHER_X86_STUB,
/** Stub switcher for AMD64. */
VMMSWITCHER_AMD64_STUB,
/** Used to make a count for array declarations and suchlike. */
VMMSWITCHER_MAX,
/** The usual 32-bit paranoia. */
VMMSWITCHER_32BIT_HACK = 0x7fffffff
} VMMSWITCHER;
/**
* VMMRZCallRing3 operations.
*/
typedef enum VMMCALLRING3
{
/** Invalid operation. */
VMMCALLRING3_INVALID = 0,
/** Acquire the PDM lock. */
VMMCALLRING3_PDM_LOCK,
/** Acquire the critical section specified as argument. */
VMMCALLRING3_PDM_CRIT_SECT_ENTER,
/** Enter the R/W critical section (in argument) exclusively. */
VMMCALLRING3_PDM_CRIT_SECT_RW_ENTER_EXCL,
/** Enter the R/W critical section (in argument) shared. */
VMMCALLRING3_PDM_CRIT_SECT_RW_ENTER_SHARED,
/** Acquire the PGM lock. */
VMMCALLRING3_PGM_LOCK,
/** Grow the PGM shadow page pool. */
VMMCALLRING3_PGM_POOL_GROW,
/** Maps a chunk into ring-3. */
VMMCALLRING3_PGM_MAP_CHUNK,
/** Allocates more handy pages. */
VMMCALLRING3_PGM_ALLOCATE_HANDY_PAGES,
/** Allocates a large (2MB) page. */
VMMCALLRING3_PGM_ALLOCATE_LARGE_HANDY_PAGE,
/** Acquire the MM hypervisor heap lock. */
VMMCALLRING3_MMHYPER_LOCK,
/** Replay the REM handler notifications. */
VMMCALLRING3_REM_REPLAY_HANDLER_NOTIFICATIONS,
/** Flush the GC/R0 logger. */
VMMCALLRING3_VMM_LOGGER_FLUSH,
/** Set the VM error message. */
VMMCALLRING3_VM_SET_ERROR,
/** Set the VM runtime error message. */
VMMCALLRING3_VM_SET_RUNTIME_ERROR,
/** Signal a ring 0 assertion. */
VMMCALLRING3_VM_R0_ASSERTION,
/** Ring switch to force preemption. */
VMMCALLRING3_VM_R0_PREEMPT,
/** Sync the FTM state with the standby node. */
VMMCALLRING3_FTM_SET_CHECKPOINT,
/** The usual 32-bit hack. */
VMMCALLRING3_32BIT_HACK = 0x7fffffff
} VMMCALLRING3;
/**
* VMMRZCallRing3 notification callback.
*
* @param pVCpu Pointer to the VMCPU.
* @param enmOperation The operation causing the ring-3 jump.
* @param pvUser The user argument.
*/
typedef DECLCALLBACK(void) FNVMMR0CALLRING3NOTIFICATION(PVMCPU pVCpu, VMMCALLRING3 enmOperation, void *pvUser);
/** Pointer to a FNRTMPNOTIFICATION(). */
typedef FNVMMR0CALLRING3NOTIFICATION *PFNVMMR0CALLRING3NOTIFICATION;
/**
* Rendezvous callback.
*
* @returns VBox strict status code - EM scheduling. Do not return
* informational status code other than the ones used by EM for
* scheduling.
*
* @param pVM The VM handle.
* @param pVCpu The handle of the calling virtual CPU.
* @param pvUser The user argument.
*/
typedef DECLCALLBACK(VBOXSTRICTRC) FNVMMEMTRENDEZVOUS(PVM pVM, PVMCPU pVCpu, void *pvUser);
/** Pointer to a rendezvous callback function. */
typedef FNVMMEMTRENDEZVOUS *PFNVMMEMTRENDEZVOUS;
/**
* Method table that the VMM uses to call back the user of the VMM.
*/
typedef struct VMM2USERMETHODS
{
/** Magic value (VMM2USERMETHODS_MAGIC). */
uint32_t u32Magic;
/** Structure version (VMM2USERMETHODS_VERSION). */
uint32_t u32Version;
/**
* Save the VM state.
*
* @returns VBox status code.
* @param pThis Pointer to the callback method table.
* @param pUVM The user mode VM handle.
*
* @remarks This member shall be set to NULL if the operation is not
* supported.
*/
DECLR3CALLBACKMEMBER(int, pfnSaveState,(PCVMM2USERMETHODS pThis, PUVM pUVM));
/** @todo Move pfnVMAtError and pfnCFGMConstructor here? */
/**
* EMT initialization notification callback.
*
* This is intended for doing per-thread initialization for EMTs (like COM
* init).
*
* @param pThis Pointer to the callback method table.
* @param pUVM The user mode VM handle.
* @param pUVCpu The user mode virtual CPU handle.
*
* @remarks This is optional and shall be set to NULL if not wanted.
*/
DECLR3CALLBACKMEMBER(void, pfnNotifyEmtInit,(PCVMM2USERMETHODS pThis, PUVM pUVM, PUVMCPU pUVCpu));
/**
* EMT termination notification callback.
*
* This is intended for doing per-thread cleanups for EMTs (like COM).
*
* @param pThis Pointer to the callback method table.
* @param pUVM The user mode VM handle.
* @param pUVCpu The user mode virtual CPU handle.
*
* @remarks This is optional and shall be set to NULL if not wanted.
*/
DECLR3CALLBACKMEMBER(void, pfnNotifyEmtTerm,(PCVMM2USERMETHODS pThis, PUVM pUVM, PUVMCPU pUVCpu));
/**
* PDM thread initialization notification callback.
*
* This is intended for doing per-thread initialization (like COM init).
*
* @param pThis Pointer to the callback method table.
* @param pUVM The user mode VM handle.
*
* @remarks This is optional and shall be set to NULL if not wanted.
*/
DECLR3CALLBACKMEMBER(void, pfnNotifyPdmtInit,(PCVMM2USERMETHODS pThis, PUVM pUVM));
/**
* EMT termination notification callback.
*
* This is intended for doing per-thread cleanups for EMTs (like COM).
*
* @param pThis Pointer to the callback method table.
* @param pUVM The user mode VM handle.
*
* @remarks This is optional and shall be set to NULL if not wanted.
*/
DECLR3CALLBACKMEMBER(void, pfnNotifyPdmtTerm,(PCVMM2USERMETHODS pThis, PUVM pUVM));
/** Magic value (VMM2USERMETHODS_MAGIC) marking the end of the structure. */
uint32_t u32EndMagic;
} VMM2USERMETHODS;
/** Magic value of the VMM2USERMETHODS (Franz Kafka). */
#define VMM2USERMETHODS_MAGIC UINT32_C(0x18830703)
/** The VMM2USERMETHODS structure version. */
#define VMM2USERMETHODS_VERSION UINT32_C(0x00020000)
VMM_INT_DECL(RTRCPTR) VMMGetStackRC(PVMCPU pVCpu);
VMMDECL(VMCPUID) VMMGetCpuId(PVM pVM);
VMMDECL(PVMCPU) VMMGetCpu(PVM pVM);
VMMDECL(PVMCPU) VMMGetCpu0(PVM pVM);
VMMDECL(PVMCPU) VMMGetCpuById(PVM pVM, VMCPUID idCpu);
VMMR3DECL(PVMCPU) VMMR3GetCpuByIdU(PUVM pVM, VMCPUID idCpu);
VMM_INT_DECL(uint32_t) VMMGetSvnRev(void);
VMM_INT_DECL(VMMSWITCHER) VMMGetSwitcher(PVM pVM);
VMM_INT_DECL(void) VMMTrashVolatileXMMRegs(void);
#ifdef IN_RING3
/** @defgroup grp_vmm_r3 The VMM Host Context Ring 3 API
* @ingroup grp_vmm
* @{
*/
VMMR3_INT_DECL(int) VMMR3Init(PVM pVM);
VMMR3_INT_DECL(int) VMMR3InitR0(PVM pVM);
# ifdef VBOX_WITH_RAW_MODE
VMMR3_INT_DECL(int) VMMR3InitRC(PVM pVM);
# endif
VMMR3_INT_DECL(int) VMMR3InitCompleted(PVM pVM, VMINITCOMPLETED enmWhat);
VMMR3_INT_DECL(int) VMMR3Term(PVM pVM);
VMMR3_INT_DECL(void) VMMR3Relocate(PVM pVM, RTGCINTPTR offDelta);
VMMR3_INT_DECL(int) VMMR3UpdateLoggers(PVM pVM);
VMMR3DECL(const char *) VMMR3GetRZAssertMsg1(PVM pVM);
VMMR3DECL(const char *) VMMR3GetRZAssertMsg2(PVM pVM);
VMMR3_INT_DECL(int) VMMR3SelectSwitcher(PVM pVM, VMMSWITCHER enmSwitcher);
VMMR3_INT_DECL(RTR0PTR) VMMR3GetHostToGuestSwitcher(PVM pVM, VMMSWITCHER enmSwitcher);
VMMR3_INT_DECL(int) VMMR3HmRunGC(PVM pVM, PVMCPU pVCpu);
# ifdef VBOX_WITH_RAW_MODE
VMMR3_INT_DECL(int) VMMR3RawRunGC(PVM pVM, PVMCPU pVCpu);
VMMR3DECL(int) VMMR3ResumeHyper(PVM pVM, PVMCPU pVCpu);
VMMR3_INT_DECL(int) VMMR3GetImportRC(PVM pVM, const char *pszSymbol, PRTRCPTR pRCPtrValue);
VMMR3DECL(int) VMMR3CallRC(PVM pVM, RTRCPTR RCPtrEntry, unsigned cArgs, ...);
VMMR3DECL(int) VMMR3CallRCV(PVM pVM, RTRCPTR RCPtrEntry, unsigned cArgs, va_list args);
# endif
VMMR3DECL(int) VMMR3CallR0(PVM pVM, uint32_t uOperation, uint64_t u64Arg, PSUPVMMR0REQHDR pReqHdr);
VMMR3DECL(void) VMMR3FatalDump(PVM pVM, PVMCPU pVCpu, int rcErr);
VMMR3_INT_DECL(void) VMMR3YieldSuspend(PVM pVM);
VMMR3_INT_DECL(void) VMMR3YieldStop(PVM pVM);
VMMR3_INT_DECL(void) VMMR3YieldResume(PVM pVM);
VMMR3_INT_DECL(void) VMMR3SendSipi(PVM pVM, VMCPUID idCpu, uint32_t uVector);
VMMR3_INT_DECL(void) VMMR3SendInitIpi(PVM pVM, VMCPUID idCpu);
VMMR3DECL(int) VMMR3RegisterPatchMemory(PVM pVM, RTGCPTR pPatchMem, unsigned cbPatchMem);
VMMR3DECL(int) VMMR3DeregisterPatchMemory(PVM pVM, RTGCPTR pPatchMem, unsigned cbPatchMem);
VMMR3DECL(int) VMMR3EmtRendezvous(PVM pVM, uint32_t fFlags, PFNVMMEMTRENDEZVOUS pfnRendezvous, void *pvUser);
VMMR3_INT_DECL(bool) VMMR3EmtRendezvousSetDisabled(PVMCPU pVCpu, bool fDisabled);
/** @defgroup grp_VMMR3EmtRendezvous_fFlags VMMR3EmtRendezvous flags
* @{ */
/** Execution type mask. */
#define VMMEMTRENDEZVOUS_FLAGS_TYPE_MASK UINT32_C(0x00000007)
/** Invalid execution type. */
#define VMMEMTRENDEZVOUS_FLAGS_TYPE_INVALID UINT32_C(0)
/** Let the EMTs execute the callback one by one (in no particular order). */
#define VMMEMTRENDEZVOUS_FLAGS_TYPE_ONE_BY_ONE UINT32_C(1)
/** Let all the EMTs execute the callback at the same time. */
#define VMMEMTRENDEZVOUS_FLAGS_TYPE_ALL_AT_ONCE UINT32_C(2)
/** Only execute the callback on one EMT (no particular one). */
#define VMMEMTRENDEZVOUS_FLAGS_TYPE_ONCE UINT32_C(3)
/** Let the EMTs execute the callback one by one in ascending order. */
#define VMMEMTRENDEZVOUS_FLAGS_TYPE_ASCENDING UINT32_C(4)
/** Let the EMTs execute the callback one by one in descending order. */
#define VMMEMTRENDEZVOUS_FLAGS_TYPE_DESCENDING UINT32_C(5)
/** Stop after the first error.
* This is not valid for any execution type where more than one EMT is active
* at a time. */
#define VMMEMTRENDEZVOUS_FLAGS_STOP_ON_ERROR UINT32_C(0x00000008)
/** The valid flags. */
#define VMMEMTRENDEZVOUS_FLAGS_VALID_MASK UINT32_C(0x0000000f)
/** @} */
VMMR3_INT_DECL(int) VMMR3EmtRendezvousFF(PVM pVM, PVMCPU pVCpu);
VMMR3_INT_DECL(int) VMMR3ReadR0Stack(PVM pVM, VMCPUID idCpu, RTHCUINTPTR R0Addr, void *pvBuf, size_t cbRead);
/** @} */
#endif /* IN_RING3 */
/** @defgroup grp_vmm_r0 The VMM Host Context Ring 0 API
* @ingroup grp_vmm
* @{
*/
/**
* The VMMR0Entry() codes.
*/
typedef enum VMMR0OPERATION
{
/** Run guest context. */
VMMR0_DO_RAW_RUN = SUP_VMMR0_DO_RAW_RUN,
/** Run guest code using the available hardware acceleration technology. */
VMMR0_DO_HM_RUN = SUP_VMMR0_DO_HM_RUN,
/** Official NOP that we use for profiling. */
VMMR0_DO_NOP = SUP_VMMR0_DO_NOP,
/** Official slow iocl NOP that we use for profiling. */
VMMR0_DO_SLOW_NOP,
/** Ask the GVMM to create a new VM. */
VMMR0_DO_GVMM_CREATE_VM,
/** Ask the GVMM to destroy the VM. */
VMMR0_DO_GVMM_DESTROY_VM,
/** Call GVMMR0SchedHalt(). */
VMMR0_DO_GVMM_SCHED_HALT,
/** Call GVMMR0SchedWakeUp(). */
VMMR0_DO_GVMM_SCHED_WAKE_UP,
/** Call GVMMR0SchedPoke(). */
VMMR0_DO_GVMM_SCHED_POKE,
/** Call GVMMR0SchedWakeUpAndPokeCpus(). */
VMMR0_DO_GVMM_SCHED_WAKE_UP_AND_POKE_CPUS,
/** Call GVMMR0SchedPoll(). */
VMMR0_DO_GVMM_SCHED_POLL,
/** Call GVMMR0QueryStatistics(). */
VMMR0_DO_GVMM_QUERY_STATISTICS,
/** Call GVMMR0ResetStatistics(). */
VMMR0_DO_GVMM_RESET_STATISTICS,
/** Call GVMMR0RegisterVCpu(). */
VMMR0_DO_GVMM_REGISTER_VMCPU,
/** Call VMMR0 Per VM Init. */
VMMR0_DO_VMMR0_INIT,
/** Call VMMR0 Per VM Termination. */
VMMR0_DO_VMMR0_TERM,
/** Setup the hardware accelerated raw-mode session. */
VMMR0_DO_HM_SETUP_VM,
/** Attempt to enable or disable hardware accelerated raw-mode. */
VMMR0_DO_HM_ENABLE,
/** Calls function in the hypervisor.
* The caller must setup the hypervisor context so the call will be performed.
* The difference between VMMR0_DO_RUN_GC and this one is the handling of
* the return GC code. The return code will not be interpreted by this operation.
*/
VMMR0_DO_CALL_HYPERVISOR,
/** Call PGMR0PhysAllocateHandyPages(). */
VMMR0_DO_PGM_ALLOCATE_HANDY_PAGES,
/** Call PGMR0PhysFlushHandyPages(). */
VMMR0_DO_PGM_FLUSH_HANDY_PAGES,
/** Call PGMR0AllocateLargePage(). */
VMMR0_DO_PGM_ALLOCATE_LARGE_HANDY_PAGE,
/** Call PGMR0PhysSetupIommu(). */
VMMR0_DO_PGM_PHYS_SETUP_IOMMU,
/** Call GMMR0InitialReservation(). */
VMMR0_DO_GMM_INITIAL_RESERVATION,
/** Call GMMR0UpdateReservation(). */
VMMR0_DO_GMM_UPDATE_RESERVATION,
/** Call GMMR0AllocatePages(). */
VMMR0_DO_GMM_ALLOCATE_PAGES,
/** Call GMMR0FreePages(). */
VMMR0_DO_GMM_FREE_PAGES,
/** Call GMMR0FreeLargePage(). */
VMMR0_DO_GMM_FREE_LARGE_PAGE,
/** Call GMMR0QueryHypervisorMemoryStatsReq(). */
VMMR0_DO_GMM_QUERY_HYPERVISOR_MEM_STATS,
/** Call GMMR0QueryMemoryStatsReq(). */
VMMR0_DO_GMM_QUERY_MEM_STATS,
/** Call GMMR0BalloonedPages(). */
VMMR0_DO_GMM_BALLOONED_PAGES,
/** Call GMMR0MapUnmapChunk(). */
VMMR0_DO_GMM_MAP_UNMAP_CHUNK,
/** Call GMMR0SeedChunk(). */
VMMR0_DO_GMM_SEED_CHUNK,
/** Call GMMR0RegisterSharedModule. */
VMMR0_DO_GMM_REGISTER_SHARED_MODULE,
/** Call GMMR0UnregisterSharedModule. */
VMMR0_DO_GMM_UNREGISTER_SHARED_MODULE,
/** Call GMMR0ResetSharedModules. */
VMMR0_DO_GMM_RESET_SHARED_MODULES,
/** Call GMMR0CheckSharedModules. */
VMMR0_DO_GMM_CHECK_SHARED_MODULES,
/** Call GMMR0FindDuplicatePage. */
VMMR0_DO_GMM_FIND_DUPLICATE_PAGE,
/** Call GMMR0QueryStatistics(). */
VMMR0_DO_GMM_QUERY_STATISTICS,
/** Call GMMR0ResetStatistics(). */
VMMR0_DO_GMM_RESET_STATISTICS,
/** Set a GVMM or GMM configuration value. */
VMMR0_DO_GCFGM_SET_VALUE,
/** Query a GVMM or GMM configuration value. */
VMMR0_DO_GCFGM_QUERY_VALUE,
/** Call PDMR0DriverCallReqHandler. */
VMMR0_DO_PDM_DRIVER_CALL_REQ_HANDLER,
/** Call PDMR0DeviceCallReqHandler. */
VMMR0_DO_PDM_DEVICE_CALL_REQ_HANDLER,
/** The start of the R0 service operations. */
VMMR0_DO_SRV_START,
/** Call IntNetR0Open(). */
VMMR0_DO_INTNET_OPEN,
/** Call IntNetR0IfClose(). */
VMMR0_DO_INTNET_IF_CLOSE,
/** Call IntNetR0IfGetBufferPtrs(). */
VMMR0_DO_INTNET_IF_GET_BUFFER_PTRS,
/** Call IntNetR0IfSetPromiscuousMode(). */
VMMR0_DO_INTNET_IF_SET_PROMISCUOUS_MODE,
/** Call IntNetR0IfSetMacAddress(). */
VMMR0_DO_INTNET_IF_SET_MAC_ADDRESS,
/** Call IntNetR0IfSetActive(). */
VMMR0_DO_INTNET_IF_SET_ACTIVE,
/** Call IntNetR0IfSend(). */
VMMR0_DO_INTNET_IF_SEND,
/** Call IntNetR0IfWait(). */
VMMR0_DO_INTNET_IF_WAIT,
/** Call IntNetR0IfAbortWait(). */
VMMR0_DO_INTNET_IF_ABORT_WAIT,
/** Forward call to the PCI driver */
VMMR0_DO_PCIRAW_REQ,
/** The end of the R0 service operations. */
VMMR0_DO_SRV_END,
/** Official call we use for testing Ring-0 APIs. */
VMMR0_DO_TESTS,
/** Test the 32->64 bits switcher. */
VMMR0_DO_TEST_SWITCHER3264,
/** The usual 32-bit type blow up. */
VMMR0_DO_32BIT_HACK = 0x7fffffff
} VMMR0OPERATION;
/**
* Request buffer for VMMR0_DO_GCFGM_SET_VALUE and VMMR0_DO_GCFGM_QUERY_VALUE.
* @todo Move got GCFGM.h when it's implemented.
*/
typedef struct GCFGMVALUEREQ
{
/** The request header.*/
SUPVMMR0REQHDR Hdr;
/** The support driver session handle. */
PSUPDRVSESSION pSession;
/** The value.
* This is input for the set request and output for the query. */
uint64_t u64Value;
/** The variable name.
* This is fixed sized just to make things simple for the mock-up. */
char szName[48];
} GCFGMVALUEREQ;
/** Pointer to a VMMR0_DO_GCFGM_SET_VALUE and VMMR0_DO_GCFGM_QUERY_VALUE request buffer.
* @todo Move got GCFGM.h when it's implemented.
*/
typedef GCFGMVALUEREQ *PGCFGMVALUEREQ;
VMMR0DECL(int) VMMR0EntryInt(PVM pVM, VMMR0OPERATION enmOperation, void *pvArg);
VMMR0DECL(void) VMMR0EntryFast(PVM pVM, VMCPUID idCpu, VMMR0OPERATION enmOperation);
VMMR0DECL(int) VMMR0EntryEx(PVM pVM, VMCPUID idCpu, VMMR0OPERATION enmOperation, PSUPVMMR0REQHDR pReq, uint64_t u64Arg, PSUPDRVSESSION);
VMMR0DECL(int) VMMR0TermVM(PVM pVM, PGVM pGVM);
#ifdef LOG_ENABLED
VMMR0DECL(void) VMMR0LogFlushDisable(PVMCPU pVCpu);
VMMR0DECL(void) VMMR0LogFlushEnable(PVMCPU pVCpu);
VMMR0DECL(bool) VMMR0IsLogFlushDisabled(PVMCPU pVCpu);
#else
#define VMMR0LogFlushDisable(pVCpu) do { } while(0)
#define VMMR0LogFlushEnable(pVCpu) do { } while(0)
#define VMMR0IsLogFlushDisabled(pVCpu) (true)
#endif
/** @} */
#ifdef IN_RC
/** @defgroup grp_vmm_rc The VMM Raw-Mode Context API
* @ingroup grp_vmm
* @{
*/
VMMRCDECL(int) VMMGCEntry(PVM pVM, unsigned uOperation, unsigned uArg, ...);
VMMRCDECL(void) VMMGCGuestToHost(PVM pVM, int rc);
VMMRCDECL(void) VMMGCLogFlushIfFull(PVM pVM);
/** @} */
#endif /* IN_RC */
#if defined(IN_RC) || defined(IN_RING0)
/** @defgroup grp_vmm_rz The VMM Raw-Mode and Ring-0 Context API
* @ingroup grp_vmm
* @{
*/
VMMRZDECL(int) VMMRZCallRing3(PVM pVM, PVMCPU pVCpu, VMMCALLRING3 enmOperation, uint64_t uArg);
VMMRZDECL(int) VMMRZCallRing3NoCpu(PVM pVM, VMMCALLRING3 enmOperation, uint64_t uArg);
VMMRZDECL(void) VMMRZCallRing3Disable(PVMCPU pVCpu);
VMMRZDECL(void) VMMRZCallRing3Enable(PVMCPU pVCpu);
VMMRZDECL(bool) VMMRZCallRing3IsEnabled(PVMCPU pVCpu);
VMMRZDECL(int) VMMRZCallRing3SetNotification(PVMCPU pVCpu, R0PTRTYPE(PFNVMMR0CALLRING3NOTIFICATION) pfnCallback, RTR0PTR pvUser);
VMMRZDECL(void) VMMRZCallRing3RemoveNotification(PVMCPU pVCpu);
/** @} */
#endif
/** @} */
RT_C_DECLS_END
#endif