PGMInternal.h revision 6db04572233249ed0e75ec100beb6418f475737c
af062818b47340eef15700d2f0211576ba3506eevboxsync * PGM - Internal header file.
af062818b47340eef15700d2f0211576ba3506eevboxsync * Copyright (C) 2006-2010 Oracle Corporation
af062818b47340eef15700d2f0211576ba3506eevboxsync * This file is part of VirtualBox Open Source Edition (OSE), as
af062818b47340eef15700d2f0211576ba3506eevboxsync * available from http://www.virtualbox.org. This file is free software;
af062818b47340eef15700d2f0211576ba3506eevboxsync * you can redistribute it and/or modify it under the terms of the GNU
af062818b47340eef15700d2f0211576ba3506eevboxsync * General Public License (GPL) as published by the Free Software
af062818b47340eef15700d2f0211576ba3506eevboxsync * Foundation, in version 2 as it comes in the "COPYING" file of the
af062818b47340eef15700d2f0211576ba3506eevboxsync * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
af062818b47340eef15700d2f0211576ba3506eevboxsync * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @defgroup grp_pgm_int Internals
af062818b47340eef15700d2f0211576ba3506eevboxsync * @ingroup grp_pgm
af062818b47340eef15700d2f0211576ba3506eevboxsync * @internal
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @name PGM Compile Time Config
af062818b47340eef15700d2f0211576ba3506eevboxsync * Indicates that there are no guest mappings to care about.
af062818b47340eef15700d2f0211576ba3506eevboxsync * Currently on raw-mode related code uses mappings, i.e. RC and R3 code.
af062818b47340eef15700d2f0211576ba3506eevboxsync#if defined(IN_RING0) || !defined(VBOX_WITH_RAW_MODE)
af062818b47340eef15700d2f0211576ba3506eevboxsync * Check and skip global PDEs for non-global flushes
af062818b47340eef15700d2f0211576ba3506eevboxsync * Optimization for PAE page tables that are modified often
af062818b47340eef15700d2f0211576ba3506eevboxsync//#if 0 /* disabled again while debugging */
af062818b47340eef15700d2f0211576ba3506eevboxsync * Large page support enabled only on 64 bits hosts; applies to nested paging only.
af062818b47340eef15700d2f0211576ba3506eevboxsync * Enables optimizations for MMIO handlers that exploits X86_TRAP_PF_RSVD and
af062818b47340eef15700d2f0211576ba3506eevboxsync * VMX_EXIT_EPT_MISCONFIG.
af062818b47340eef15700d2f0211576ba3506eevboxsync * Chunk unmapping code activated on 32-bit hosts for > 1.5/2 GB guest memory support
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_WITH_LARGE_ADDRESS_SPACE_ON_32_BIT_HOST
af062818b47340eef15700d2f0211576ba3506eevboxsync * Sync N pages instead of a whole page table
af062818b47340eef15700d2f0211576ba3506eevboxsync * Number of pages to sync during a page fault
af062818b47340eef15700d2f0211576ba3506eevboxsync * When PGMPOOL_WITH_GCPHYS_TRACKING is enabled using high values here
af062818b47340eef15700d2f0211576ba3506eevboxsync * causes a lot of unnecessary extents and also is slower than taking more \#PFs.
af062818b47340eef15700d2f0211576ba3506eevboxsync * Note that \#PFs are much more expensive in the VT-x/AMD-V case due to
af062818b47340eef15700d2f0211576ba3506eevboxsync * world switch overhead, so let's sync more.
af062818b47340eef15700d2f0211576ba3506eevboxsync/* Chose 32 based on the compile test in #4219; 64 shows worse stats.
af062818b47340eef15700d2f0211576ba3506eevboxsync * 32 again shows better results than 16; slightly more overhead in the \#PF handler,
af062818b47340eef15700d2f0211576ba3506eevboxsync * but ~5% fewer faults.
af062818b47340eef15700d2f0211576ba3506eevboxsync * Number of PGMPhysRead/Write cache entries (must be <= sizeof(uint64_t))
af062818b47340eef15700d2f0211576ba3506eevboxsync#define PGM_MAX_PHYSCACHE_ENTRIES_MASK (PGM_MAX_PHYSCACHE_ENTRIES-1)
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def PGMPOOL_CFG_MAX_GROW
af062818b47340eef15700d2f0211576ba3506eevboxsync * The maximum number of pages to add to the pool in one go.
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def VBOX_STRICT_PGM_HANDLER_VIRTUAL
af062818b47340eef15700d2f0211576ba3506eevboxsync * Enables some extra assertions for virtual handlers (mainly phys2virt related).
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def VBOX_WITH_NEW_LAZY_PAGE_ALLOC
af062818b47340eef15700d2f0211576ba3506eevboxsync * Enables the experimental lazy page allocation code. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/*#define VBOX_WITH_NEW_LAZY_PAGE_ALLOC */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def VBOX_WITH_REAL_WRITE_MONITORED_PAGES
af062818b47340eef15700d2f0211576ba3506eevboxsync * Enables real write monitoring of pages, i.e. mapping them read-only and
af062818b47340eef15700d2f0211576ba3506eevboxsync * only making them writable when getting a write access #PF. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @name PDPT and PML4 flags.
af062818b47340eef15700d2f0211576ba3506eevboxsync * These are placed in the three bits available for system programs in
af062818b47340eef15700d2f0211576ba3506eevboxsync * the PDPT and PML4 entries.
af062818b47340eef15700d2f0211576ba3506eevboxsync/** The entry is a permanent one and it's must always be present.
af062818b47340eef15700d2f0211576ba3506eevboxsync * Never free such an entry. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Mapping (hypervisor allocated pagetable). */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @name Page directory flags.
af062818b47340eef15700d2f0211576ba3506eevboxsync * These are placed in the three bits available for system programs in
af062818b47340eef15700d2f0211576ba3506eevboxsync * the page directory entries.
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Mapping (hypervisor allocated pagetable). */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Made read-only to facilitate dirty bit tracking. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @name Page flags.
af062818b47340eef15700d2f0211576ba3506eevboxsync * These are placed in the three bits available for system programs in
af062818b47340eef15700d2f0211576ba3506eevboxsync * the page entries.
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Made read-only to facilitate dirty bit tracking. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Scanned and approved by CSAM (tm).
af062818b47340eef15700d2f0211576ba3506eevboxsync * NOTE: Must be identical to the one defined in CSAMInternal.h!!
af062818b47340eef15700d2f0211576ba3506eevboxsync * @todo Move PGM_PTFLAGS_* and PGM_PDFLAGS_* to VBox/vmm/pgm.h. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @name Defines used to indicate the shadow and guest paging in the templates.
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Macro for checking if the guest is using paging.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param uGstType PGM_TYPE_*
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param uShwType PGM_TYPE_*
af062818b47340eef15700d2f0211576ba3506eevboxsync * @remark ASSUMES certain order of the PGM_TYPE_* values.
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Macro for checking if the guest supports the NX bit.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param uGstType PGM_TYPE_*
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param uShwType PGM_TYPE_*
af062818b47340eef15700d2f0211576ba3506eevboxsync * @remark ASSUMES certain order of the PGM_TYPE_* values.
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def PGM_HCPHYS_2_PTR
af062818b47340eef15700d2f0211576ba3506eevboxsync * Maps a HC physical page pool address to a virtual address.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @returns VBox status code.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVM The VM handle.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVCpu The current CPU.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param HCPhys The HC physical address to map to a virtual one.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param ppv Where to store the virtual address. No need to cast
af062818b47340eef15700d2f0211576ba3506eevboxsync * @remark Use with care as we don't have so much dynamic mapping space in
af062818b47340eef15700d2f0211576ba3506eevboxsync * ring-0 on 32-bit darwin and in RC.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @remark There is no need to assert on the result.
af062818b47340eef15700d2f0211576ba3506eevboxsync#if defined(VBOX_WITH_2X_4GB_ADDR_SPACE_IN_R0) || defined(IN_RC)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_HCPHYS_2_PTR(pVM, pVCpu, HCPhys, ppv) \
af062818b47340eef15700d2f0211576ba3506eevboxsync pgmRZDynMapHCPageInlined(pVCpu, HCPhys, (void **)(ppv) RTLOG_COMMA_SRC_POS)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_HCPHYS_2_PTR(pVM, pVCpu, HCPhys, ppv) \
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def PGM_GCPHYS_2_PTR_V2
af062818b47340eef15700d2f0211576ba3506eevboxsync * Maps a GC physical page address to a virtual address.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @returns VBox status code.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVM The VM handle.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVCpu The current CPU.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param GCPhys The GC physical address to map to a virtual one.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param ppv Where to store the virtual address. No need to cast this.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @remark Use with care as we don't have so much dynamic mapping space in
af062818b47340eef15700d2f0211576ba3506eevboxsync * ring-0 on 32-bit darwin and in RC.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @remark There is no need to assert on the result.
af062818b47340eef15700d2f0211576ba3506eevboxsync#if defined(VBOX_WITH_2X_4GB_ADDR_SPACE_IN_R0) || defined(IN_RC)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_GCPHYS_2_PTR_V2(pVM, pVCpu, GCPhys, ppv) \
af062818b47340eef15700d2f0211576ba3506eevboxsync pgmRZDynMapGCPageV2Inlined(pVM, pVCpu, GCPhys, (void **)(ppv) RTLOG_COMMA_SRC_POS)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_GCPHYS_2_PTR_V2(pVM, pVCpu, GCPhys, ppv) \
af062818b47340eef15700d2f0211576ba3506eevboxsync PGMPhysGCPhys2R3Ptr(pVM, GCPhys, 1 /* one page only */, (PRTR3PTR)(ppv)) /** @todo this isn't asserting, use PGMRamGCPhys2HCPtr! */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def PGM_GCPHYS_2_PTR
af062818b47340eef15700d2f0211576ba3506eevboxsync * Maps a GC physical page address to a virtual address.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @returns VBox status code.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVM The VM handle.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param GCPhys The GC physical address to map to a virtual one.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param ppv Where to store the virtual address. No need to cast this.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @remark Use with care as we don't have so much dynamic mapping space in
af062818b47340eef15700d2f0211576ba3506eevboxsync * ring-0 on 32-bit darwin and in RC.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @remark There is no need to assert on the result.
af062818b47340eef15700d2f0211576ba3506eevboxsync#define PGM_GCPHYS_2_PTR(pVM, GCPhys, ppv) PGM_GCPHYS_2_PTR_V2(pVM, VMMGetCpu(pVM), GCPhys, ppv)
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def PGM_GCPHYS_2_PTR_BY_VMCPU
af062818b47340eef15700d2f0211576ba3506eevboxsync * Maps a GC physical page address to a virtual address.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @returns VBox status code.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVCpu The current CPU.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param GCPhys The GC physical address to map to a virtual one.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param ppv Where to store the virtual address. No need to cast this.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @remark Use with care as we don't have so much dynamic mapping space in
af062818b47340eef15700d2f0211576ba3506eevboxsync * ring-0 on 32-bit darwin and in RC.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @remark There is no need to assert on the result.
af062818b47340eef15700d2f0211576ba3506eevboxsync#define PGM_GCPHYS_2_PTR_BY_VMCPU(pVCpu, GCPhys, ppv) PGM_GCPHYS_2_PTR_V2((pVCpu)->CTX_SUFF(pVM), pVCpu, GCPhys, ppv)
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def PGM_GCPHYS_2_PTR_EX
af062818b47340eef15700d2f0211576ba3506eevboxsync * Maps a unaligned GC physical page address to a virtual address.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @returns VBox status code.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVM The VM handle.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param GCPhys The GC physical address to map to a virtual one.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param ppv Where to store the virtual address. No need to cast this.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @remark Use with care as we don't have so much dynamic mapping space in
af062818b47340eef15700d2f0211576ba3506eevboxsync * ring-0 on 32-bit darwin and in RC.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @remark There is no need to assert on the result.
af062818b47340eef15700d2f0211576ba3506eevboxsync#if defined(IN_RC) || defined(VBOX_WITH_2X_4GB_ADDR_SPACE_IN_R0)
af062818b47340eef15700d2f0211576ba3506eevboxsync pgmRZDynMapGCPageOffInlined(VMMGetCpu(pVM), GCPhys, (void **)(ppv) RTLOG_COMMA_SRC_POS)
af062818b47340eef15700d2f0211576ba3506eevboxsync PGMPhysGCPhys2R3Ptr(pVM, GCPhys, 1 /* one page only */, (PRTR3PTR)(ppv)) /** @todo this isn't asserting, use PGMRamGCPhys2HCPtr! */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def PGM_DYNMAP_UNUSED_HINT
af062818b47340eef15700d2f0211576ba3506eevboxsync * Hints to the dynamic mapping code in RC and R0/darwin that the specified page
af062818b47340eef15700d2f0211576ba3506eevboxsync * is no longer used.
af062818b47340eef15700d2f0211576ba3506eevboxsync * For best effect only apply this to the page that was mapped most recently.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVCpu The current CPU.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pvPage The pool page.
af062818b47340eef15700d2f0211576ba3506eevboxsync#if defined(IN_RC) || defined(VBOX_WITH_2X_4GB_ADDR_SPACE_IN_R0)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_DYNMAP_UNUSED_HINT(pVCpu, pvPage) pgmRZDynMapUnusedHint(pVCpu, pvPage, RT_SRC_POS)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_DYNMAP_UNUSED_HINT(pVCpu, pvPage) pgmRZDynMapUnusedHint(pVCpu, pvPage)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_DYNMAP_UNUSED_HINT(pVCpu, pvPage) do {} while (0)
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def PGM_DYNMAP_UNUSED_HINT_VM
af062818b47340eef15700d2f0211576ba3506eevboxsync * Hints to the dynamic mapping code in RC and R0/darwin that the specified page
af062818b47340eef15700d2f0211576ba3506eevboxsync * is no longer used.
af062818b47340eef15700d2f0211576ba3506eevboxsync * For best effect only apply this to the page that was mapped most recently.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVM The VM handle.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pvPage The pool page.
af062818b47340eef15700d2f0211576ba3506eevboxsync#define PGM_DYNMAP_UNUSED_HINT_VM(pVM, pvPage) PGM_DYNMAP_UNUSED_HINT(VMMGetCpu(pVM), pvPage)
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def PGM_INVL_PG
af062818b47340eef15700d2f0211576ba3506eevboxsync * Invalidates a page.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVCpu The VMCPU handle.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param GCVirt The virtual address of the page to invalidate.
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_INVL_PG(pVCpu, GCVirt) ASMInvalidatePage((void *)(uintptr_t)(GCVirt))
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_INVL_PG(pVCpu, GCVirt) HWACCMInvalidatePage(pVCpu, (RTGCPTR)(GCVirt))
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_INVL_PG(pVCpu, GCVirt) HWACCMInvalidatePage(pVCpu, (RTGCPTR)(GCVirt))
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def PGM_INVL_PG_ALL_VCPU
af062818b47340eef15700d2f0211576ba3506eevboxsync * Invalidates a page on all VCPUs
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVM The VM handle.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param GCVirt The virtual address of the page to invalidate.
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_INVL_PG_ALL_VCPU(pVM, GCVirt) ASMInvalidatePage((void *)(uintptr_t)(GCVirt))
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_INVL_PG_ALL_VCPU(pVM, GCVirt) HWACCMInvalidatePageOnAllVCpus(pVM, (RTGCPTR)(GCVirt))
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_INVL_PG_ALL_VCPU(pVM, GCVirt) HWACCMInvalidatePageOnAllVCpus(pVM, (RTGCPTR)(GCVirt))
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def PGM_INVL_BIG_PG
af062818b47340eef15700d2f0211576ba3506eevboxsync * Invalidates a 4MB page directory entry.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVCpu The VMCPU handle.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param GCVirt The virtual address within the page directory to invalidate.
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_INVL_BIG_PG(pVCpu, GCVirt) ASMReloadCR3()
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_INVL_BIG_PG(pVCpu, GCVirt) HWACCMFlushTLB(pVCpu)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_INVL_BIG_PG(pVCpu, GCVirt) HWACCMFlushTLB(pVCpu)
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def PGM_INVL_VCPU_TLBS()
af062818b47340eef15700d2f0211576ba3506eevboxsync * Invalidates the TLBs of the specified VCPU
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVCpu The VMCPU handle.
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_INVL_VCPU_TLBS(pVCpu) HWACCMFlushTLB(pVCpu)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_INVL_VCPU_TLBS(pVCpu) HWACCMFlushTLB(pVCpu)
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @def PGM_INVL_ALL_VCPU_TLBS()
af062818b47340eef15700d2f0211576ba3506eevboxsync * Invalidates the TLBs of all VCPUs
af062818b47340eef15700d2f0211576ba3506eevboxsync * @param pVM The VM handle.
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_INVL_ALL_VCPU_TLBS(pVM) HWACCMFlushTLBOnAllVCpus(pVM)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGM_INVL_ALL_VCPU_TLBS(pVM) HWACCMFlushTLBOnAllVCpus(pVM)
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @name Safer Shadow PAE PT/PTE
af062818b47340eef15700d2f0211576ba3506eevboxsync * For helping avoid misinterpreting invalid PAE/AMD64 page table entries as
af062818b47340eef15700d2f0211576ba3506eevboxsync * For making sure that u1Present and X86_PTE_P checks doesn't mistake
af062818b47340eef15700d2f0211576ba3506eevboxsync * invalid entries for present.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @sa X86PTEPAE.
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Unsigned integer view */
af062818b47340eef15700d2f0211576ba3506eevboxsync /* Not other views. */
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_IS_P(Pte) ( ((Pte).uCareful & (X86_PTE_P | X86_PTE_PAE_MBZ_MASK_NX)) == X86_PTE_P )
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_IS_RW(Pte) ( !!((Pte).uCareful & X86_PTE_RW))
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_IS_US(Pte) ( !!((Pte).uCareful & X86_PTE_US))
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_IS_A(Pte) ( !!((Pte).uCareful & X86_PTE_A))
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_IS_D(Pte) ( !!((Pte).uCareful & X86_PTE_D))
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_IS_TRACK_DIRTY(Pte) ( !!((Pte).uCareful & PGM_PTFLAGS_TRACK_DIRTY) )
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_IS_P_RW(Pte) ( ((Pte).uCareful & (X86_PTE_P | X86_PTE_RW | X86_PTE_PAE_MBZ_MASK_NX)) == (X86_PTE_P | X86_PTE_RW) )
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_GET_LOG(Pte) ( (Pte).uCareful )
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_GET_HCPHYS(Pte) ( (Pte).uCareful & X86_PTE_PAE_PG_MASK )
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_GET_U(Pte) ( (Pte).uCareful ) /**< Use with care. */
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_SET(Pte, uVal) do { (Pte).uCareful = (uVal); } while (0)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_SET2(Pte, Pte2) do { (Pte).uCareful = (Pte2).uCareful; } while (0)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_ATOMIC_SET(Pte, uVal) do { ASMAtomicWriteU64(&(Pte).uCareful, (uVal)); } while (0)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_ATOMIC_SET2(Pte, Pte2) do { ASMAtomicWriteU64(&(Pte).uCareful, (Pte2).uCareful); } while (0)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_SET_RO(Pte) do { (Pte).uCareful &= ~(X86PGPAEUINT)X86_PTE_RW; } while (0)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_SET_RW(Pte) do { (Pte).uCareful |= X86_PTE_RW; } while (0)
af062818b47340eef15700d2f0211576ba3506eevboxsync * For making sure that u1Present and X86_PTE_P checks doesn't mistake
af062818b47340eef15700d2f0211576ba3506eevboxsync * invalid entries for present.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @sa X86PTPAE.
af062818b47340eef15700d2f0211576ba3506eevboxsynctypedef struct PGMSHWPTPAE
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_IS_P(Pte) ( (Pte).n.u1Present )
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_IS_RW(Pte) ( (Pte).n.u1Write )
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_IS_A(Pte) ( (Pte).n.u1Accessed )
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_IS_TRACK_DIRTY(Pte) ( !!((Pte).u & PGM_PTFLAGS_TRACK_DIRTY) )
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_IS_P_RW(Pte) ( ((Pte).u & (X86_PTE_P | X86_PTE_RW)) == (X86_PTE_P | X86_PTE_RW) )
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_GET_HCPHYS(Pte) ( (Pte).u & X86_PTE_PAE_PG_MASK )
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_GET_U(Pte) ( (Pte).u ) /**< Use with care. */
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_SET(Pte, uVal) do { (Pte).u = (uVal); } while (0)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_SET2(Pte, Pte2) do { (Pte).u = (Pte2).u; } while (0)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_ATOMIC_SET(Pte, uVal) do { ASMAtomicWriteU64(&(Pte).u, (uVal)); } while (0)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_ATOMIC_SET2(Pte, Pte2) do { ASMAtomicWriteU64(&(Pte).u, (Pte2).u); } while (0)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_SET_RO(Pte) do { (Pte).u &= ~(X86PGPAEUINT)X86_PTE_RW; } while (0)
af062818b47340eef15700d2f0211576ba3506eevboxsync# define PGMSHWPTEPAE_SET_RW(Pte) do { (Pte).u |= X86_PTE_RW; } while (0)
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Pointer to a shadow PAE PTE. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Pointer to a const shadow PAE PTE. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Pointer to a shadow PAE page table. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Pointer to a const shadow PAE page table. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Size of the GCPtrConflict array in PGMMAPPING.
af062818b47340eef15700d2f0211576ba3506eevboxsync * @remarks Must be a power of two. */
af062818b47340eef15700d2f0211576ba3506eevboxsync * Structure for tracking GC Mappings.
af062818b47340eef15700d2f0211576ba3506eevboxsync * This structure is used by linked list in both GC and HC.
af062818b47340eef15700d2f0211576ba3506eevboxsynctypedef struct PGMMAPPING
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Pointer to next entry. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Pointer to next entry. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Pointer to next entry. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Indicate whether this entry is finalized. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Start Virtual address. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Last Virtual address (inclusive). */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Range size (bytes). */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Pointer to relocation callback function. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** User argument to the callback. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Mapping description / name. For easing debugging. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Last 8 addresses that caused conflicts. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Number of conflicts for this hypervisor mapping. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Number of page tables. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Array of page table mapping data. Each entry
af062818b47340eef15700d2f0211576ba3506eevboxsync * describes one page table. The array can be longer
af062818b47340eef15700d2f0211576ba3506eevboxsync * than the declared length.
af062818b47340eef15700d2f0211576ba3506eevboxsync /** The HC physical address of the page table. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** The HC physical address of the first PAE page table. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** The HC physical address of the second PAE page table. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** The HC virtual address of the 32-bit page table. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** The HC virtual address of the two PAE page table. (i.e 1024 entries instead of 512) */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** The RC virtual address of the 32-bit page table. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** The RC virtual address of the two PAE page table. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** The R0 virtual address of the 32-bit page table. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** The R0 virtual address of the two PAE page table. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Pointer to structure for tracking GC Mappings. */
af062818b47340eef15700d2f0211576ba3506eevboxsync * Physical page access handler structure.
af062818b47340eef15700d2f0211576ba3506eevboxsync * This is used to keep track of physical address ranges
af062818b47340eef15700d2f0211576ba3506eevboxsync * which are being monitored in some kind of way.
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Access type. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Number of pages to update. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Set if we have pages that have been aliased. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Set if we have pages that have temporarily been disabled. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Pointer to R3 callback function. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** User argument for R3 handlers. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Pointer to R0 callback function. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** User argument for R0 handlers. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Pointer to RC callback function. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** User argument for RC handlers. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Description / Name. For easing debugging. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Profiling of this handler. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Pointer to a physical page access handler structure. */
af062818b47340eef15700d2f0211576ba3506eevboxsync * Cache node for the physical addresses covered by a virtual handler.
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Core node for the tree based on physical ranges. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Offset from this struct to the PGMVIRTHANDLER structure. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Offset of the next alias relative to this one.
af062818b47340eef15700d2f0211576ba3506eevboxsync * Bit 0 is used for indicating whether we're in the tree.
af062818b47340eef15700d2f0211576ba3506eevboxsync * Bit 1 is used for indicating that we're the head node.
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Pointer to a phys to virtual handler structure. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** The bit in PGMPHYS2VIRTHANDLER::offNextAlias used to indicate that the
af062818b47340eef15700d2f0211576ba3506eevboxsync * node is in the tree. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** The bit in PGMPHYS2VIRTHANDLER::offNextAlias used to indicate that the
af062818b47340eef15700d2f0211576ba3506eevboxsync * node is in the head of an alias chain.
af062818b47340eef15700d2f0211576ba3506eevboxsync * The PGMPHYS2VIRTHANDLER_IN_TREE is always set if this bit is set. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** The mask to apply to PGMPHYS2VIRTHANDLER::offNextAlias to get the offset. */
af062818b47340eef15700d2f0211576ba3506eevboxsync * Virtual page access handler structure.
af062818b47340eef15700d2f0211576ba3506eevboxsync * This is used to keep track of virtual address ranges
af062818b47340eef15700d2f0211576ba3506eevboxsync * which are being monitored in some kind of way.
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Core node for the tree based on virtual ranges. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Size of the range (in bytes). */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Number of cache pages. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Access type. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Pointer to the RC callback function. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Pointer to the R3 callback function for invalidation. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Pointer to the R3 callback function. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Description / Name. For easing debugging. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Profiling of this handler. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Array of cached physical addresses for the monitored ranged. */
af062818b47340eef15700d2f0211576ba3506eevboxsync PGMPHYS2VIRTHANDLER aPhysToVirt[HC_ARCH_BITS == 32 ? 1 : 2];
af062818b47340eef15700d2f0211576ba3506eevboxsync/** Pointer to a virtual page access handler structure. */
af062818b47340eef15700d2f0211576ba3506eevboxsync/** @name Page type predicates.
af062818b47340eef15700d2f0211576ba3506eevboxsync#define PGMPAGETYPE_IS_READABLE(type) ( (type) <= PGMPAGETYPE_ROM )
af062818b47340eef15700d2f0211576ba3506eevboxsync#define PGMPAGETYPE_IS_WRITEABLE(type) ( (type) <= PGMPAGETYPE_ROM_SHADOW )
af062818b47340eef15700d2f0211576ba3506eevboxsync#define PGMPAGETYPE_IS_RWX(type) ( (type) <= PGMPAGETYPE_ROM_SHADOW )
af062818b47340eef15700d2f0211576ba3506eevboxsync#define PGMPAGETYPE_IS_ROX(type) ( (type) == PGMPAGETYPE_ROM )
af062818b47340eef15700d2f0211576ba3506eevboxsync#define PGMPAGETYPE_IS_NP(type) ( (type) == PGMPAGETYPE_MMIO )
af062818b47340eef15700d2f0211576ba3506eevboxsync * A Physical Guest Page tracking structure.
af062818b47340eef15700d2f0211576ba3506eevboxsync * The format of this structure is complicated because we have to fit a lot
af062818b47340eef15700d2f0211576ba3506eevboxsync * of information into as few bits as possible. The format is also subject
af062818b47340eef15700d2f0211576ba3506eevboxsync * to change (there is one coming up soon). Which means that for we'll be
af062818b47340eef15700d2f0211576ba3506eevboxsync * using PGM_PAGE_GET_*, PGM_PAGE_IS_ and PGM_PAGE_SET_* macros for *all*
af062818b47340eef15700d2f0211576ba3506eevboxsync * accesses to the structure.
af062818b47340eef15700d2f0211576ba3506eevboxsynctypedef union PGMPAGE
af062818b47340eef15700d2f0211576ba3506eevboxsync /** Structured view. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** 1:0 - The physical handler state (PGM_PAGE_HNDL_PHYS_STATE_*). */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** 3:2 - Paging structure needed to map the page
af062818b47340eef15700d2f0211576ba3506eevboxsync * (PGM_PAGE_PDE_TYPE_*). */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** 4 - Indicator of dirty page for fault tolerance tracking. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** 5 - Flag indicating that a write monitored page was written to
af062818b47340eef15700d2f0211576ba3506eevboxsync * when set. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** 7:6 - Unused. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** 9:8 - The physical handler state (PGM_PAGE_HNDL_VIRT_STATE_*). */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** 11:10 - Unused. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** 12:48 - The host physical frame number (shift left to get the
af062818b47340eef15700d2f0211576ba3506eevboxsync * address). */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** 50:48 - The page state. */
af062818b47340eef15700d2f0211576ba3506eevboxsync /** 51:53 - The page type (PGMPAGETYPE). */
} PGMPAGE;
PGM_PAGE_INIT((a_pPage), (a_pVM)->pgm.s.HCPhysZeroPg, NIL_GMM_PAGEID, (a_uType), PGM_PAGE_STATE_ZERO)
#define PGM_PAGE_STATE_ZERO 0
#define PGM_PAGE_SET_WRITTEN_TO(a_pPage) do { (a_pPage)->au8[1] |= UINT8_C(0x80); } while (0) /// FIXME FIXME
#define PGM_PAGE_PDE_TYPE_DONTCARE 0
/** Can use a page directory entry to map the continuous range - temporarily disabled (by page monitoring). */
#define PGM_PAGE_HNDL_PHYS_STATE_NONE 0
#define PGM_PAGE_HNDL_VIRT_STATE_NONE 0
# define PGMLIVESAVERAMPAGE_WITH_CRC32
typedef struct PGMLIVESAVERAMPAGE
#define PGM_USE_RAMRANGE_SEARCH_TREES
typedef struct PGMRAMRANGE
} PGMRAMRANGE;
(!!( (pRam)->fFlags & (PGM_RAM_RANGE_FLAGS_AD_HOC_ROM | PGM_RAM_RANGE_FLAGS_AD_HOC_MMIO | PGM_RAM_RANGE_FLAGS_AD_HOC_MMIO2) ) )
typedef struct PGMROMPAGE
bool fWrittenTo;
bool fDirty;
bool fDirtiedRecently;
} LiveSave;
} PGMROMPAGE;
typedef struct PGMROMRANGE
} PGMROMRANGE;
typedef struct PGMLIVESAVEMMIO2PAGE
bool fDirty;
bool fZero;
bool fReserved;
typedef struct PGMMMIO2RANGE
bool fMapped;
bool fOverlapping;
* PGMPhysRead/Write cache entry
typedef struct PGMPHYSCACHEENTRY
* PGMPhysRead/Write cache to reduce REM memory access overhead
typedef struct PGMPHYSCACHE
} PGMPHYSCACHE;
typedef struct PGMCHUNKR3MAP
void *pv;
typedef struct PGMCHUNKR3MAPTLBE
#ifdef VBOX_WITH_2X_4GB_ADDR_SPACE
typedef struct PGMCHUNKR3MAPTLB
typedef struct PGMPAGER3MAPTLBE
#ifdef VBOX_WITH_2X_4GB_ADDR_SPACE
#ifdef VBOX_WITH_2X_4GB_ADDR_SPACE
#ifdef VBOX_WITH_2X_4GB_ADDR_SPACE
typedef struct PGMPAGER3MAPTLB
typedef struct PGMRCDYNMAPENTRY
struct PGMRCDYNMAPENTRY_PPTE
} uPte;
* paDynPageMap* PGM members. However, it has to be defined in PGMInternal.h
typedef struct PGMRCDYNMAP
} PGMRCDYNMAP;
typedef struct PGMMAPSETENTRY
#ifndef IN_RC
typedef struct PGMMAPSET
} PGMMAPSET;
#ifdef IN_RC
typedef void * PPGMPAGEMAP;
typedef void ** PPPGMPAGEMAP;
#define NIL_PGMPOOL_IDX 0
typedef struct PGMPOOLUSER
#pragma pack()
typedef struct PGMPOOLPHYSEXT
#pragma pack()
typedef enum PGMPOOLKIND
PGMPOOLKIND_INVALID = 0,
} PGMPOOLKIND;
typedef struct PGMPOOLPAGE
#ifdef VBOX_WITH_2X_4GB_ADDR_SPACE
bool fZeroed;
bool fSeenNonGlobal;
bool fMonitored;
bool fCached;
bool volatile fReusedFlushPending;
bool fDirty;
/** Used to indicate that this page can't be flushed. Important for cr3 root pages or shadow pae pd pages). */
typedef struct PGMPOOL
bool fCacheEnabled;
#ifdef VBOX_WITH_STATISTICS
#ifdef VBOX_WITH_STATISTICS
# define PGMPOOL_PAGE_2_PTR(a_pVM, a_pPage) pgmPoolMapPageInlined((a_pVM), (a_pPage) RTLOG_COMMA_SRC_POS)
# define PGMPOOL_PAGE_2_PTR_V2(a_pVM, a_pVCpu, a_pPage) pgmPoolMapPageV2Inlined((a_pVM), (a_pVCpu), (a_pPage) RTLOG_COMMA_SRC_POS)
#define PGMPOOL_TD_IDX_SHIFT 0
typedef struct PGMTREES
} PGMTREES;
typedef struct PGMPTWALKCORE
bool fSucceeded;
bool fNotPresent;
bool fBadPhysAddr;
bool fRsvdError;
bool fBigPage;
bool fGigantPage;
bool fEffectiveUS;
bool fEffectiveRW;
bool fEffectiveNX;
typedef struct PGMPTWALKGSTAMD64
typedef struct PGMPTWALKGSTPAE
typedef struct PGMPTWALKGST32BIT
#ifdef IN_RC
# ifdef IN_RING3
typedef struct PGMMODEDATA
DECLR3CALLBACKMEMBER(int, pfnR3ShwGetPage,(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTHCPHYS pHCPhys));
DECLR3CALLBACKMEMBER(int, pfnR3ShwModifyPage,(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cbPages, uint64_t fFlags, uint64_t fMask, uint32_t fOpFlags));
DECLRCCALLBACKMEMBER(int, pfnRCShwGetPage,(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTHCPHYS pHCPhys));
DECLRCCALLBACKMEMBER(int, pfnRCShwModifyPage,(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cbPages, uint64_t fFlags, uint64_t fMask, uint32_t fOpFlags));
DECLR0CALLBACKMEMBER(int, pfnR0ShwGetPage,(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTHCPHYS pHCPhys));
DECLR0CALLBACKMEMBER(int, pfnR0ShwModifyPage,(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cbPages, uint64_t fFlags, uint64_t fMask, uint32_t fOpFlags));
DECLR3CALLBACKMEMBER(int, pfnR3GstGetPage,(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTGCPHYS pGCPhys));
DECLR3CALLBACKMEMBER(int, pfnR3GstModifyPage,(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cbPages, uint64_t fFlags, uint64_t fMask));
DECLRCCALLBACKMEMBER(int, pfnRCGstGetPage,(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTGCPHYS pGCPhys));
DECLRCCALLBACKMEMBER(int, pfnRCGstModifyPage,(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cbPages, uint64_t fFlags, uint64_t fMask));
DECLR0CALLBACKMEMBER(int, pfnR0GstGetPage,(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTGCPHYS pGCPhys));
DECLR0CALLBACKMEMBER(int, pfnR0GstModifyPage,(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cbPages, uint64_t fFlags, uint64_t fMask));
DECLR3CALLBACKMEMBER(int, pfnR3BthSyncCR3,(PVMCPU pVCpu, uint64_t cr0, uint64_t cr3, uint64_t cr4, bool fGlobal));
DECLR3CALLBACKMEMBER(int, pfnR3BthVerifyAccessSyncPage,(PVMCPU pVCpu, RTGCPTR GCPtrPage, unsigned fFlags, unsigned uError));
#ifdef VBOX_STRICT
DECLR3CALLBACKMEMBER(unsigned, pfnR3BthAssertCR3,(PVMCPU pVCpu, uint64_t cr3, uint64_t cr4, RTGCPTR GCPtr, RTGCPTR cb));
DECLRCCALLBACKMEMBER(int, pfnRCBthTrap0eHandler,(PVMCPU pVCpu, RTGCUINT uErr, PCPUMCTXCORE pRegFrame, RTGCPTR pvFault, bool *pfLockTaken));
DECLRCCALLBACKMEMBER(int, pfnRCBthSyncCR3,(PVMCPU pVCpu, uint64_t cr0, uint64_t cr3, uint64_t cr4, bool fGlobal));
DECLRCCALLBACKMEMBER(int, pfnRCBthVerifyAccessSyncPage,(PVMCPU pVCpu, RTGCPTR GCPtrPage, unsigned fFlags, unsigned uError));
#ifdef VBOX_STRICT
DECLRCCALLBACKMEMBER(unsigned, pfnRCBthAssertCR3,(PVMCPU pVCpu, uint64_t cr3, uint64_t cr4, RTGCPTR GCPtr, RTGCPTR cb));
DECLR0CALLBACKMEMBER(int, pfnR0BthTrap0eHandler,(PVMCPU pVCpu, RTGCUINT uErr, PCPUMCTXCORE pRegFrame, RTGCPTR pvFault, bool *pfLockTaken));
DECLR0CALLBACKMEMBER(int, pfnR0BthSyncCR3,(PVMCPU pVCpu, uint64_t cr0, uint64_t cr3, uint64_t cr4, bool fGlobal));
DECLR0CALLBACKMEMBER(int, pfnR0BthVerifyAccessSyncPage,(PVMCPU pVCpu, RTGCPTR GCPtrPage, unsigned fFlags, unsigned uError));
#ifdef VBOX_STRICT
DECLR0CALLBACKMEMBER(unsigned, pfnR0BthAssertCR3,(PVMCPU pVCpu, uint64_t cr3, uint64_t cr4, RTGCPTR GCPtr, RTGCPTR cb));
#ifdef VBOX_WITH_STATISTICS
typedef struct PGMSTATS
STAMCOUNTER StatR3DetectedConflicts; /**< R3: Number of times PGMR3MapHasConflicts() detected a conflict. */
STAMPROFILE StatR3ResolveConflict; /**< R3: pgmR3SyncPTResolveConflict() profiling (includes the entire relocation). */
STAMPROFILE StatRZSyncCR3HandlerVirtualReset; /**< RC/R0: Profiling of the virtual handler resets. */
STAMPROFILE StatRZSyncCR3HandlerVirtualUpdate; /**< RC/R0: Profiling of the virtual handler updates. */
STAMPROFILE StatR3SyncCR3HandlerVirtualUpdate; /**< R3: Profiling of the virtual handler updates. */
STAMCOUNTER StatR3PhysHandlerReset; /**< R3: The number of times PGMHandlerPhysicalReset is called. */
STAMCOUNTER StatRZPhysHandlerReset; /**< RC/R0: The number of times PGMHandlerPhysicalReset is called. */
STAMCOUNTER StatR3PhysHandlerLookupHits; /**< R3: Number of cache hits when looking up physical handlers. */
STAMCOUNTER StatR3PhysHandlerLookupMisses; /**< R3: Number of cache misses when looking up physical handlers. */
STAMCOUNTER StatRZPhysHandlerLookupHits; /**< RC/R0: Number of cache hits when lookup up physical handlers. */
STAMCOUNTER StatRZPhysHandlerLookupMisses; /**< RC/R0: Number of cache misses when looking up physical handlers */
STAMPROFILE StatRZVirtHandlerSearchByPhys; /**< RC/R0: Profiling of pgmHandlerVirtualFindByPhysAddr. */
STAMPROFILE StatR3VirtHandlerSearchByPhys; /**< R3: Profiling of pgmHandlerVirtualFindByPhysAddr. */
STAMCOUNTER StatRZPageReplaceShared; /**< RC/R0: Times a shared page has been replaced by a private one. */
STAMCOUNTER StatRZPageReplaceZero; /**< RC/R0: Times the zero page has been replaced by a private one. */
/// @todo STAMCOUNTER StatRZPageHandyAllocs; /**< RC/R0: The number of times we've executed GMMR3AllocateHandyPages. */
STAMCOUNTER StatR3PageReplaceShared; /**< R3: Times a shared page has been replaced by a private one. */
STAMCOUNTER StatR3PageReplaceZero; /**< R3: Times the zero page has been replaced by a private one. */
/// @todo STAMCOUNTER StatR3PageHandyAllocs; /**< R3: The number of times we've executed GMMR3AllocateHandyPages. */
STAMCOUNTER StatRCInvlPgConflict; /**< RC: Number of times PGMInvalidatePage() detected a mapping conflict. */
STAMCOUNTER StatRCInvlPgSyncMonCR3; /**< RC: Number of times PGMInvalidatePage() ran into PGM_SYNC_MONITOR_CR3. */
STAMCOUNTER StatTrackAliased; /**< The number of times switching to cRef2, i.e. the page is being shadowed by two PTs. */
STAMCOUNTER StatTrackAliasedLots; /**< The number of times we're hitting pages which has overflowed cRef2. */
} PGMSTATS;
typedef struct PGM
bool fRamPreAlloc;
bool fNestedPaging;
bool fNoMorePhysWrites;
bool fPciPassthrough;
bool fFinalizedMappings;
bool fMappingsFixed;
bool fMappingsFixedRestored;
bool fMappingsDisabled;
* @todo The plan of keeping PGMRCDYNMAP private to PGMRZDynMap.cpp didn't
#ifdef VBOX_WITH_2X_4GB_ADDR_SPACE
uint32_t c;
* @cfgm PGM/MaxRing3Chunks */
} ChunkR3Map;
} Rom,
Ram;
bool fActive;
} LiveSave;
bool volatile fErrInjHandyPages;
uint32_t cAllPages; /**< The total number of pages. (Should be Private + Shared + Zero + Pure MMIO.) */
#ifdef VBOX_WITH_STATISTICS
} PGM;
typedef struct PGMCPUSTATS
STAMCOUNTER StatR0NpMiscfgSyncPage; /**< R0: SyncPage calls from PGMR0Trap0eHandlerNPMisconfig(). */
STAMPROFILE StatRZTrap0eTime2Ballooned; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is read access to a ballooned page. */
STAMPROFILE StatRZTrap0eTime2CSAM; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is CSAM. */
STAMPROFILE StatRZTrap0eTime2DirtyAndAccessed; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is dirty and/or accessed bit emulation. */
STAMPROFILE StatRZTrap0eTime2GuestTrap; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is a guest trap. */
STAMPROFILE StatRZTrap0eTime2HndPhys; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is a physical handler. */
STAMPROFILE StatRZTrap0eTime2HndVirt; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is a virtual handler. */
STAMPROFILE StatRZTrap0eTime2HndUnhandled; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is access outside the monitored areas of a monitored page. */
STAMPROFILE StatRZTrap0eTime2InvalidPhys; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is access to an invalid physical guest address. */
STAMPROFILE StatRZTrap0eTime2MakeWritable; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is a page that needed to be made writable. */
STAMPROFILE StatRZTrap0eTime2Mapping; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is the guest mappings. */
STAMPROFILE StatRZTrap0eTime2Misc; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is not known. */
STAMPROFILE StatRZTrap0eTime2OutOfSync; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is an out-of-sync page. */
STAMPROFILE StatRZTrap0eTime2OutOfSyncHndPhys; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is an out-of-sync physical handler page. */
STAMPROFILE StatRZTrap0eTime2OutOfSyncHndVirt; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is an out-of-sync virtual handler page. */
STAMPROFILE StatRZTrap0eTime2OutOfSyncHndObs; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is an obsolete handler page. */
STAMPROFILE StatRZTrap0eTime2SyncPT; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is lazy syncing of a PT. */
STAMPROFILE StatRZTrap0eTime2WPEmulation; /**< RC/R0: Profiling of the Trap0eHandler body when the cause is CR0.WP emulation. */
STAMCOUNTER StatRZTrap0eConflicts; /**< RC/R0: The number of times \#PF was caused by an undetected conflict. */
STAMCOUNTER StatRZTrap0eHandlersMapping; /**< RC/R0: Number of traps due to access handlers in mappings. */
STAMCOUNTER StatRZTrap0eHandlersPhysAll; /**< RC/R0: Number of traps due to physical all-access handlers. */
STAMCOUNTER StatRZTrap0eHandlersPhysAllOpt; /**< RC/R0: Number of the physical all-access handler traps using the optimization. */
STAMCOUNTER StatRZTrap0eHandlersPhysWrite; /**< RC/R0: Number of traps due to write-physical access handlers. */
STAMCOUNTER StatRZTrap0eHandlersVirtual; /**< RC/R0: Number of traps due to virtual access handlers. */
STAMCOUNTER StatRZTrap0eHandlersVirtualByPhys; /**< RC/R0: Number of traps due to virtual access handlers found by physical address. */
STAMCOUNTER StatRZTrap0eHandlersVirtualUnmarked;/**< RC/R0: Number of traps due to virtual access handlers found by virtual address (without proper physical flags). */
STAMCOUNTER StatRZTrap0eHandlersUnhandled; /**< RC/R0: Number of traps due to access outside range of monitored page(s). */
STAMCOUNTER StatRZTrap0eHandlersInvalid; /**< RC/R0: Number of traps due to access to invalid physical memory. */
STAMCOUNTER StatRZGuestCR3WriteHandled; /**< RC/R0: The number of times WriteHandlerCR3() was successfully called. */
STAMCOUNTER StatRZGuestCR3WriteUnhandled; /**< RC/R0: The number of times WriteHandlerCR3() was called and we had to fall back to the recompiler. */
STAMCOUNTER StatRZGuestCR3WriteConflict; /**< RC/R0: The number of times WriteHandlerCR3() was called and a conflict was detected. */
STAMCOUNTER StatRZGuestROMWriteHandled; /**< RC/R0: The number of times pgmPhysRomWriteHandler() was successfully called. */
STAMCOUNTER StatRZGuestROMWriteUnhandled; /**< RC/R0: The number of times pgmPhysRomWriteHandler() was called and we had to fall back to the recompiler */
STAMCOUNTER StatRZDynMapPageSlowLoopMisses; /**< RZ: Misses in the pgmR0DynMapPageSlow search loop. */
STAMPROFILE StatRZSyncCR3Handlers; /**< RC/R0: Profiling of the PGMSyncCR3() update handler section. */
STAMCOUNTER StatRZSyncCR3DstCacheHit; /**< RC/R0: The number of times we got some kind of cache hit on a page table. */
STAMCOUNTER StatRZSyncCR3DstFreed; /**< RC/R0: The number of times we've had to free a shadow entry. */
STAMCOUNTER StatRZSyncCR3DstFreedSrcNP; /**< RC/R0: The number of times we've had to free a shadow entry for which the source entry was not present. */
STAMCOUNTER StatRZSyncCR3DstNotPresent; /**< RC/R0: The number of times we've encountered a not present shadow entry for a present guest entry. */
STAMCOUNTER StatRZSyncCR3DstSkippedGlobalPD; /**< RC/R0: The number of times a global page directory wasn't flushed. */
STAMCOUNTER StatRZSyncCR3DstSkippedGlobalPT; /**< RC/R0: The number of times a page table with only global entries wasn't flushed. */
STAMCOUNTER StatRZSyncPagePDNAs; /**< RC/R0: The number of time we've marked a PD not present from SyncPage to virtualize the accessed bit. */
STAMCOUNTER StatRZSyncPagePDOutOfSync; /**< RC/R0: The number of time we've encountered an out-of-sync PD in SyncPage. */
STAMCOUNTER StatRZAccessedPage; /**< RC/R0: The number of pages marked not present for accessed bit emulation. */
STAMPROFILE StatRZDirtyBitTracking; /**< RC/R0: Profiling the dirty bit tracking in CheckPageFault().. */
STAMCOUNTER StatRZDirtyPage; /**< RC/R0: The number of pages marked read-only for dirty bit tracking. */
STAMCOUNTER StatRZDirtyPageBig; /**< RC/R0: The number of pages marked read-only for dirty bit tracking. */
STAMCOUNTER StatRZDirtyPageTrap; /**< RC/R0: The number of traps generated for dirty bit tracking. */
STAMCOUNTER StatRZDirtyPageStale; /**< RC/R0: The number of traps generated for dirty bit tracking. (stale tlb entries) */
STAMCOUNTER StatRZDirtyTrackRealPF; /**< RC/R0: The number of real pages faults during dirty bit tracking. */
STAMCOUNTER StatRZDirtiedPage; /**< RC/R0: The number of pages marked dirty because of write accesses. */
STAMCOUNTER StatRZPageAlreadyDirty; /**< RC/R0: The number of pages already marked dirty because of write accesses. */
STAMCOUNTER StatRZInvalidatePage4KBPages; /**< RC/R0: The number of times PGMInvalidatePage() was called for a 4KB page. */
STAMCOUNTER StatRZInvalidatePage4MBPages; /**< RC/R0: The number of times PGMInvalidatePage() was called for a 4MB page. */
STAMCOUNTER StatRZInvalidatePage4MBPagesSkip; /**< RC/R0: The number of times PGMInvalidatePage() skipped a 4MB page. */
STAMCOUNTER StatRZInvalidatePagePDMappings; /**< RC/R0: The number of times PGMInvalidatePage() was called for a page directory containing mappings (no conflict). */
STAMCOUNTER StatRZInvalidatePagePDNAs; /**< RC/R0: The number of times PGMInvalidatePage() was called for a not accessed page directory. */
STAMCOUNTER StatRZInvalidatePagePDNPs; /**< RC/R0: The number of times PGMInvalidatePage() was called for a not present page directory. */
STAMCOUNTER StatRZInvalidatePagePDOutOfSync; /**< RC/R0: The number of times PGMInvalidatePage() was called for an out of sync page directory. */
STAMCOUNTER StatRZInvalidatePageSkipped; /**< RC/R0: The number of times PGMInvalidatePage() was skipped due to not present shw or pending pending SyncCR3. */
STAMCOUNTER StatRZPageOutOfSyncUser; /**< RC/R0: The number of times user page is out of sync was detected in \#PF or VerifyAccessSyncPage. */
STAMCOUNTER StatRZPageOutOfSyncSupervisor; /**< RC/R0: The number of times supervisor page is out of sync was detected in in \#PF or VerifyAccessSyncPage. */
STAMCOUNTER StatRZPageOutOfSyncUserWrite; /**< RC/R0: The number of times user page is out of sync was detected in \#PF. */
STAMCOUNTER StatRZPageOutOfSyncSupervisorWrite; /**< RC/R0: The number of times supervisor page is out of sync was detected in in \#PF. */
STAMCOUNTER StatRZPageOutOfSyncBallloon; /**< RC/R0: The number of times a ballooned page was accessed (read). */
STAMCOUNTER StatRZFlushTLBNewCR3; /**< RC/R0: The number of times PGMFlushTLB was called with a new CR3, non-global. (switch) */
STAMCOUNTER StatRZFlushTLBNewCR3Global; /**< RC/R0: The number of times PGMFlushTLB was called with a new CR3, global. (switch) */
STAMCOUNTER StatRZFlushTLBSameCR3; /**< RC/R0: The number of times PGMFlushTLB was called with the same CR3, non-global. (flush) */
STAMCOUNTER StatRZFlushTLBSameCR3Global; /**< RC/R0: The number of times PGMFlushTLB was called with the same CR3, global. (flush) */
STAMPROFILE StatR3SyncCR3Handlers; /**< R3: Profiling of the PGMSyncCR3() update handler section. */
STAMCOUNTER StatR3SyncCR3DstFreed; /**< R3: The number of times we've had to free a shadow entry. */
STAMCOUNTER StatR3SyncCR3DstFreedSrcNP; /**< R3: The number of times we've had to free a shadow entry for which the source entry was not present. */
STAMCOUNTER StatR3SyncCR3DstNotPresent; /**< R3: The number of times we've encountered a not present shadow entry for a present guest entry. */
STAMCOUNTER StatR3SyncCR3DstSkippedGlobalPD; /**< R3: The number of times a global page directory wasn't flushed. */
STAMCOUNTER StatR3SyncCR3DstSkippedGlobalPT; /**< R3: The number of times a page table with only global entries wasn't flushed. */
STAMCOUNTER StatR3SyncCR3DstCacheHit; /**< R3: The number of times we got some kind of cache hit on a page table. */
STAMCOUNTER StatR3SyncPagePDNAs; /**< R3: The number of time we've marked a PD not present from SyncPage to virtualize the accessed bit. */
STAMCOUNTER StatR3SyncPagePDOutOfSync; /**< R3: The number of time we've encountered an out-of-sync PD in SyncPage. */
STAMCOUNTER StatR3AccessedPage; /**< R3: The number of pages marked not present for accessed bit emulation. */
STAMPROFILE StatR3DirtyBitTracking; /**< R3: Profiling the dirty bit tracking in CheckPageFault(). */
STAMCOUNTER StatR3DirtyPage; /**< R3: The number of pages marked read-only for dirty bit tracking. */
STAMCOUNTER StatR3DirtyPageBig; /**< R3: The number of pages marked read-only for dirty bit tracking. */
STAMCOUNTER StatR3DirtyTrackRealPF; /**< R3: The number of real pages faults during dirty bit tracking. */
STAMCOUNTER StatR3DirtiedPage; /**< R3: The number of pages marked dirty because of write accesses. */
STAMCOUNTER StatR3PageAlreadyDirty; /**< R3: The number of pages already marked dirty because of write accesses. */
STAMCOUNTER StatR3InvalidatePage4KBPages; /**< R3: The number of times PGMInvalidatePage() was called for a 4KB page. */
STAMCOUNTER StatR3InvalidatePage4MBPages; /**< R3: The number of times PGMInvalidatePage() was called for a 4MB page. */
STAMCOUNTER StatR3InvalidatePage4MBPagesSkip; /**< R3: The number of times PGMInvalidatePage() skipped a 4MB page. */
STAMCOUNTER StatR3InvalidatePagePDNAs; /**< R3: The number of times PGMInvalidatePage() was called for a not accessed page directory. */
STAMCOUNTER StatR3InvalidatePagePDNPs; /**< R3: The number of times PGMInvalidatePage() was called for a not present page directory. */
STAMCOUNTER StatR3InvalidatePagePDMappings; /**< R3: The number of times PGMInvalidatePage() was called for a page directory containing mappings (no conflict). */
STAMCOUNTER StatR3InvalidatePagePDOutOfSync; /**< R3: The number of times PGMInvalidatePage() was called for an out of sync page directory. */
STAMCOUNTER StatR3InvalidatePageSkipped; /**< R3: The number of times PGMInvalidatePage() was skipped due to not present shw or pending pending SyncCR3. */
STAMCOUNTER StatR3PageOutOfSyncUser; /**< R3: The number of times user page is out of sync was detected in \#PF or VerifyAccessSyncPage. */
STAMCOUNTER StatR3PageOutOfSyncSupervisor; /**< R3: The number of times supervisor page is out of sync was detected in in \#PF or VerifyAccessSyncPage. */
STAMCOUNTER StatR3PageOutOfSyncUserWrite; /**< R3: The number of times user page is out of sync was detected in \#PF. */
STAMCOUNTER StatR3PageOutOfSyncSupervisorWrite; /**< R3: The number of times supervisor page is out of sync was detected in in \#PF. */
STAMCOUNTER StatR3PageOutOfSyncBallloon; /**< R3: The number of times a ballooned page was accessed (read). */
STAMCOUNTER StatR3FlushTLBNewCR3; /**< R3: The number of times PGMFlushTLB was called with a new CR3, non-global. (switch) */
STAMCOUNTER StatR3FlushTLBNewCR3Global; /**< R3: The number of times PGMFlushTLB was called with a new CR3, global. (switch) */
STAMCOUNTER StatR3FlushTLBSameCR3; /**< R3: The number of times PGMFlushTLB was called with the same CR3, non-global. (flush) */
STAMCOUNTER StatR3FlushTLBSameCR3Global; /**< R3: The number of times PGMFlushTLB was called with the same CR3, global. (flush) */
} PGMCPUSTATS;
typedef struct PGMCPU
bool fA20Enabled;
bool fNoExecuteEnabled;
#ifndef VBOX_WITH_2X_4GB_ADDR_SPACE
#ifndef VBOX_WITH_2X_4GB_ADDR_SPACE
#ifndef VBOX_WITH_2X_4GB_ADDR_SPACE
/** The physical addresses of the guest page directories (PAE) pointed to by apGstPagePDsHC/GC. */
#ifndef VBOX_WITH_2X_4GB_ADDR_SPACE
/* The shadow page pool index of the user table as specified during allocation; useful for freeing root pages */
/* The index into the user table (shadowed) as specified during allocation; useful for freeing root pages. */
DECLR3CALLBACKMEMBER(int, pfnR3ShwGetPage,(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTHCPHYS pHCPhys));
DECLR3CALLBACKMEMBER(int, pfnR3ShwModifyPage,(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cbPages, uint64_t fFlags, uint64_t fMask, uint32_t fOpFlags));
DECLRCCALLBACKMEMBER(int, pfnRCShwGetPage,(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTHCPHYS pHCPhys));
DECLRCCALLBACKMEMBER(int, pfnRCShwModifyPage,(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cbPages, uint64_t fFlags, uint64_t fMask, uint32_t fOpFlags));
DECLR0CALLBACKMEMBER(int, pfnR0ShwGetPage,(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTHCPHYS pHCPhys));
DECLR0CALLBACKMEMBER(int, pfnR0ShwModifyPage,(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cbPages, uint64_t fFlags, uint64_t fMask, uint32_t fOpFlags));
DECLR3CALLBACKMEMBER(int, pfnR3GstGetPage,(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTGCPHYS pGCPhys));
DECLR3CALLBACKMEMBER(int, pfnR3GstModifyPage,(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cbPages, uint64_t fFlags, uint64_t fMask));
DECLRCCALLBACKMEMBER(int, pfnRCGstGetPage,(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTGCPHYS pGCPhys));
DECLRCCALLBACKMEMBER(int, pfnRCGstModifyPage,(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cbPages, uint64_t fFlags, uint64_t fMask));
DECLR0CALLBACKMEMBER(int, pfnR0GstGetPage,(PVMCPU pVCpu, RTGCPTR GCPtr, uint64_t *pfFlags, PRTGCPHYS pGCPhys));
DECLR0CALLBACKMEMBER(int, pfnR0GstModifyPage,(PVMCPU pVCpu, RTGCPTR GCPtr, size_t cbPages, uint64_t fFlags, uint64_t fMask));
DECLR3CALLBACKMEMBER(int, pfnR3BthSyncCR3,(PVMCPU pVCpu, uint64_t cr0, uint64_t cr3, uint64_t cr4, bool fGlobal));
DECLR3CALLBACKMEMBER(int, pfnR3BthVerifyAccessSyncPage,(PVMCPU pVCpu, RTGCPTR GCPtrPage, unsigned fFlags, unsigned uError));
DECLR3CALLBACKMEMBER(unsigned, pfnR3BthAssertCR3,(PVMCPU pVCpu, uint64_t cr3, uint64_t cr4, RTGCPTR GCPtr, RTGCPTR cb));
DECLR0CALLBACKMEMBER(int, pfnR0BthTrap0eHandler,(PVMCPU pVCpu, RTGCUINT uErr, PCPUMCTXCORE pRegFrame, RTGCPTR pvFault, bool *pfLockTaken));
DECLR0CALLBACKMEMBER(int, pfnR0BthSyncCR3,(PVMCPU pVCpu, uint64_t cr0, uint64_t cr3, uint64_t cr4, bool fGlobal));
DECLR0CALLBACKMEMBER(int, pfnR0BthVerifyAccessSyncPage,(PVMCPU pVCpu, RTGCPTR GCPtrPage, unsigned fFlags, unsigned uError));
DECLR0CALLBACKMEMBER(unsigned, pfnR0BthAssertCR3,(PVMCPU pVCpu, uint64_t cr3, uint64_t cr4, RTGCPTR GCPtr, RTGCPTR cb));
DECLRCCALLBACKMEMBER(int, pfnRCBthTrap0eHandler,(PVMCPU pVCpu, RTGCUINT uErr, PCPUMCTXCORE pRegFrame, RTGCPTR pvFault, bool *pfLockTaken));
DECLRCCALLBACKMEMBER(int, pfnRCBthSyncCR3,(PVMCPU pVCpu, uint64_t cr0, uint64_t cr3, uint64_t cr4, bool fGlobal));
DECLRCCALLBACKMEMBER(int, pfnRCBthVerifyAccessSyncPage,(PVMCPU pVCpu, RTGCPTR GCPtrPage, unsigned fFlags, unsigned uError));
DECLRCCALLBACKMEMBER(unsigned, pfnRCBthAssertCR3,(PVMCPU pVCpu, uint64_t cr3, uint64_t cr4, RTGCPTR GCPtr, RTGCPTR cb));
} PGMCPU;
int pgmR3SyncPTResolveConflict(PVM pVM, PPGMMAPPING pMapping, PX86PD pPDSrc, RTGCPTR GCPtrOldMapping);
void pgmHandlerPhysicalResetAliasedPage(PVM pVM, PPGMPAGE pPage, RTGCPHYS GCPhysPage, bool fDoAccounting);
int pgmHandlerVirtualFindByPhysAddr(PVM pVM, RTGCPHYS GCPhys, PPGMVIRTHANDLER *ppVirt, unsigned *piPage);
# define pgmHandlerVirtualDumpPhysPages(a) do { } while (0)
int pgmPhysGCPhys2CCPtrInternalReadOnly(PVM pVM, PPGMPAGE pPage, RTGCPHYS GCPhys, const void **ppv);
VMMDECL(int) pgmPhysHandlerRedirectToHC(PVM pVM, RTGCUINT uErrorCode, PCPUMCTXCORE pRegFrame, RTGCPTR pvFault, RTGCPHYS GCPhysFault, void *pvUser);
VMMDECL(int) pgmPhysRomWriteHandler(PVM pVM, RTGCUINT uErrorCode, PCPUMCTXCORE pRegFrame, RTGCPTR pvFault, RTGCPHYS GCPhysFault, void *pvUser);
int pgmPhysFreePage(PVM pVM, PGMMFREEPAGESREQ pReq, uint32_t *pcPendingPages, PPGMPAGE pPage, RTGCPHYS GCPhys);
#ifdef IN_RING3
DECLCALLBACK(VBOXSTRICTRC) pgmR3PoolClearAllRendezvous(PVM pVM, PVMCPU pVCpu, void *fpvFlushRemTbl);
int pgmRZDynMapGCPageCommon(PVM pVM, PVMCPU pVCpu, RTGCPHYS GCPhys, void **ppv RTLOG_COMMA_SRC_POS_DECL);
# ifdef LOG_ENABLED
int pgmPoolAllocEx(PVM pVM, RTGCPHYS GCPhys, PGMPOOLKIND enmKind, PGMPOOLACCESS enmAccess, uint16_t iUser,
DECLINLINE(int) pgmPoolAlloc(PVM pVM, RTGCPHYS GCPhys, PGMPOOLKIND enmKind, uint16_t iUser, uint32_t iUserTable,
return pgmPoolAllocEx(pVM, GCPhys, enmKind, PGMPOOLACCESS_DONTCARE, iUser, iUserTable, false, ppPage);
int pgmPoolFlushPage(PPGMPOOL pPool, PPGMPOOLPAGE pPage, bool fFlush = true /* DO NOT USE false UNLESS YOU KNOWN WHAT YOU'RE DOING!! */);
int pgmPoolTrackUpdateGCPhys(PVM pVM, RTGCPHYS GCPhysPage, PPGMPAGE pPhysPage, bool fFlushPTEs, bool *pfFlushTLBs);
void pgmPoolTracDerefGCPhysHint(PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTHCPHYS HCPhys, RTGCPHYS GCPhysHint, uint16_t iPte);
uint16_t pgmPoolTrackPhysExtAddref(PVM pVM, PPGMPAGE pPhysPage, uint16_t u16, uint16_t iShwPT, uint16_t iPte);
void pgmPoolTrackPhysExtDerefGCPhys(PPGMPOOL pPool, PPGMPOOLPAGE pPoolPage, PPGMPAGE pPhysPage, uint16_t iPte);
void pgmPoolMonitorChainChanging(PVMCPU pVCpu, PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTGCPHYS GCPhysFault, CTXTYPE(RTGCPTR, RTHCPTR, RTGCPTR) pvAddress, unsigned cbWrite);
void pgmMapClearShadowPDEs(PVM pVM, PPGMPOOLPAGE pShwPageCR3, PPGMMAPPING pMap, unsigned iOldPDE, bool fDeactivateCR3);
int pgmShwSyncNestedPageLocked(PVMCPU pVCpu, RTGCPHYS GCPhysFault, uint32_t cPages, PGMMODE enmShwPagingMode);
DECLCALLBACK(int) pgmR3CmdCheckDuplicatePages(PCDBGCCMD pCmd, PDBGCCMDHLP pCmdHlp, PVM pVM, PCDBGCVAR paArgs, unsigned cArgs);
DECLCALLBACK(int) pgmR3CmdShowSharedModules(PCDBGCCMD pCmd, PDBGCCMDHLP pCmdHlp, PVM pVM, PCDBGCVAR paArgs, unsigned cArgs);