PGMAllPool.cpp revision f37fdece8b97c04e00923c1ec924706682f89b2b
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * PGM Shadow Page Pool.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Copyright (C) 2006-2007 Sun Microsystems, Inc.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * This file is part of VirtualBox Open Source Edition (OSE), as
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * available from http://www.virtualbox.org. This file is free software;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * you can redistribute it and/or modify it under the terms of the GNU
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * General Public License (GPL) as published by the Free Software
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Foundation, in version 2 as it comes in the "COPYING" file of the
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Clara, CA 95054 USA or visit http://www.sun.com if you need
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * additional information or have any questions.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync/*******************************************************************************
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync* Header Files *
7b9f0c34e9ea328981c99e97054bdf8684d9d620vboxsync*******************************************************************************/
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync/*******************************************************************************
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync* Internal Functions *
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync*******************************************************************************/
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncDECLINLINE(unsigned) pgmPoolTrackGetShadowEntrySize(PGMPOOLKIND enmKind);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncDECLINLINE(unsigned) pgmPoolTrackGetGuestEntrySize(PGMPOOLKIND enmKind);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic void pgmPoolTrackDeref(PPGMPOOL pPool, PPGMPOOLPAGE pPage);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic void pgmPoolTracDerefGCPhysHint(PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTHCPHYS HCPhys, RTGCPHYS GCPhysHint);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic int pgmPoolTrackAddUser(PPGMPOOL pPool, PPGMPOOLPAGE pPage, uint16_t iUser, uint32_t iUserTable);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic void pgmPoolMonitorModifiedRemove(PPGMPOOL pPool, PPGMPOOLPAGE pPage);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncDECLEXPORT(int) pgmPoolAccessHandler(PVM pVM, RTGCUINT uErrorCode, PCPUMCTXCORE pRegFrame, RTGCPTR pvFault, RTGCPHYS GCPhysFault, void *pvUser);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Checks if the specified page pool kind is for a 4MB or 2MB guest page.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @returns true if it's the shadow of a 4MB or 2MB guest page, otherwise false.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param enmKind The page kind.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncDECLINLINE(bool) pgmPoolIsBigPage(PGMPOOLKIND enmKind)
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync return true;
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync return false;
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync#if defined(IN_RC) || defined(VBOX_WITH_2X_4GB_ADDR_SPACE_IN_R0)
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * Maps a pool page into the current context.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * @returns Pointer to the mapping.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * @param pVM The VM handle.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * @param pPage The page to map.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync /* general pages. */
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync Assert(pPage->idx < pVM->pgm.s.CTX_SUFF(pPool)->cCurPages);
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync int rc = PGMDynMapHCPage(pVM, pPage->Core.Key, &pv);
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync /* special pages. */
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync AssertReleaseMsgFailed(("Invalid index %d\n", pPage->idx));
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync# else /* VBOX_WITH_2X_4GB_ADDR_SPACE_IN_R0 */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync AssertReleaseMsgFailed(("Invalid index %d\n", pPage->idx));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync int rc = PGMDynMapHCPage(pVM, pPage->Core.Key, &pv);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync# endif /* VBOX_WITH_2X_4GB_ADDR_SPACE_IN_R0 */
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync#endif /* IN_RC || VBOX_WITH_2X_4GB_ADDR_SPACE_IN_R0 */
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Determin the size of a write instruction.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @returns number of bytes written.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pDis The disassembler state.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic unsigned pgmPoolDisasWriteSize(PDISCPUSTATE pDis)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * This is very crude and possibly wrong for some opcodes,
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * but since it's not really supposed to be called we can
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * probably live with that.
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * Flushes a chain of pages sharing the same access monitor.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @returns VBox status code suitable for scheduling.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pPool The pool.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pPage A page in the chain.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncint pgmPoolMonitorChainFlush(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync LogFlow(("pgmPoolMonitorChainFlush: Flush page %RGp type=%d\n", pPage->GCPhys, pPage->enmKind));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Find the list head.
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * Iterate the list flushing each shadow page.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync if (rc2 == VERR_PGM_POOL_CLEARED && rc == VINF_SUCCESS)
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * Wrapper for getting the current context pointer to the entry being modified.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * @returns Pointer to the current context mapping of the entry.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * @param pPool The pool.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * @param pvFault The fault virtual address.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * @param GCPhysFault The fault physical address.
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync * @param cbEntry The entry size.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsyncDECLINLINE(const void *) pgmPoolMonitorGCPtr2CCPtr(PPGMPOOL pPool, RTHCPTR pvFault, RTGCPHYS GCPhysFault, const unsigned cbEntry)
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsyncDECLINLINE(const void *) pgmPoolMonitorGCPtr2CCPtr(PPGMPOOL pPool, RTGCPTR pvFault, RTGCPHYS GCPhysFault, const unsigned cbEntry)
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync return (const void *)((RTGCUINTPTR)pvFault & ~(RTGCUINTPTR)(cbEntry - 1));
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync int rc = PGMDynMapGCPageOff(pPool->pVMR0, GCPhysFault & ~(RTGCPHYS)(cbEntry - 1), &pvRet);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync int rc = pgmRamGCPhys2HCPtr(&pPool->pVMR0->pgm.s, GCPhysFault & ~(RTGCPHYS)(cbEntry - 1), &pvRet);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync return (RTHCPTR)((uintptr_t)pvFault & ~(RTHCUINTPTR)(cbEntry - 1));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Process shadow entries before they are changed by the guest.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * For PT entries we will clear them. For PD entries, we'll simply check
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * for mapping conflicts and set the SyncCR3 FF if found.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * @param pPool The pool.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPage The head page.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param GCPhysFault The guest physical fault address.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param uAddress In R0 and GC this is the guest context fault address (flat).
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * In R3 this is the host context 'fault' address.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pCpu The disassembler state for figuring out the write size.
9523921c89c66f4bececdbd5ac95aed0039eda1bvboxsync * This need not be specified if the caller knows we won't do cross entry accesses.
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsyncvoid pgmPoolMonitorChainChanging(PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTGCPHYS GCPhysFault, RTHCPTR pvAddress, PDISCPUSTATE pCpu)
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsyncvoid pgmPoolMonitorChainChanging(PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTGCPHYS GCPhysFault, RTGCPTR pvAddress, PDISCPUSTATE pCpu)
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync const unsigned off = GCPhysFault & PAGE_OFFSET_MASK;
9523921c89c66f4bececdbd5ac95aed0039eda1bvboxsync const unsigned cbWrite = (pCpu) ? pgmPoolDisasWriteSize(pCpu) : 0;
9523921c89c66f4bececdbd5ac95aed0039eda1bvboxsync LogFlow(("pgmPoolMonitorChainChanging: %RGv phys=%RGp kind=%d cbWrite=%d\n", pvAddress, GCPhysFault, pPage->enmKind, cbWrite));
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync uShw.pv = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage);
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync PCX86PTE pGstPte = (PCX86PTE)pgmPoolMonitorGCPtr2CCPtr(pPool, pvAddress, GCPhysFault, sizeof(*pGstPte));
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync Log4(("pgmPoolMonitorChainChanging 32_32: deref %016RX64 GCPhys %08RX32\n", uShw.pPT->a[iShw].u & X86_PTE_PAE_PG_MASK, pGstPte->u & X86_PTE_PG_MASK));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* page/2 sized */
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync const unsigned iShw = (off / sizeof(X86PTE)) & (X86_PG_PAE_ENTRIES - 1);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PCX86PTE pGstPte = (PCX86PTE)pgmPoolMonitorGCPtr2CCPtr(pPool, pvAddress, GCPhysFault, sizeof(*pGstPte));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Log4(("pgmPoolMonitorChainChanging pae_32: deref %016RX64 GCPhys %08RX32\n", uShw.pPT->a[iShw].u & X86_PTE_PAE_PG_MASK, pGstPte->u & X86_PTE_PG_MASK));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync PCX86PTEPAE pGstPte = (PCX86PTEPAE)pgmPoolMonitorGCPtr2CCPtr(pPool, pvAddress, GCPhysFault, sizeof(*pGstPte));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Log4(("pgmPoolMonitorChainChanging pae: deref %016RX64 GCPhys %016RX64\n", uShw.pPTPae->a[iShw].u & X86_PTE_PAE_PG_MASK, pGstPte->u & X86_PTE_PAE_PG_MASK));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* paranoia / a bit assumptive. */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync const unsigned iShw2 = (off + cbWrite - 1) / sizeof(X86PTEPAE);
06ea6bcf23874b662d499b3f130024c98b2dd7a6vboxsync AssertReturnVoid(iShw2 < RT_ELEMENTS(uShw.pPTPae->a));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync PCX86PTEPAE pGstPte = (PCX86PTEPAE)pgmPoolMonitorGCPtr2CCPtr(pPool, pvAddress, GCPhysFault, sizeof(*pGstPte));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync Log4(("pgmPoolMonitorChainChanging pae: deref %016RX64 GCPhys %016RX64\n", uShw.pPTPae->a[iShw2].u & X86_PTE_PAE_PG_MASK, pGstPte->u & X86_PTE_PAE_PG_MASK));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync const unsigned iShw = off / sizeof(X86PTE); // ASSUMING 32-bit guest paging!
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw=%#x!\n", iShw));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync /* paranoia / a bit assumptive. */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync const unsigned iShw2 = (off + cbWrite - 1) / sizeof(X86PTE);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw2=%#x!\n", iShw2));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync#if 0 /* useful when running PGMAssertCR3(), a bit too troublesome for general use (TLBs). */
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync && !VM_FF_ISSET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3))
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync LogFlow(("pgmPoolMonitorChainChanging: iShw=%#x: %RX32 -> freeing it!\n", iShw, uShw.pPD->a[iShw].u));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync# ifdef IN_RC /* TLB load - we're pushing things a bit... */
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync pgmPoolFree(pPool->CTX_SUFF(pVM), uShw.pPD->a[iShw].u & X86_PDE_PG_MASK, pPage->idx, iShw);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync unsigned iShw = (off / sizeof(X86PTE)) * 2; // ASSUMING 32-bit guest paging!
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync if ((uShw.pPDPae->a[iShw].u & (PGM_PDFLAGS_MAPPING | X86_PDE_P)) == (PGM_PDFLAGS_MAPPING | X86_PDE_P))
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw=%#x!\n", iShw));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync /* paranoia / a bit assumptive. */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync && (uShw.pPDPae->a[iShw2].u & (PGM_PDFLAGS_MAPPING | X86_PDE_P)) == (PGM_PDFLAGS_MAPPING | X86_PDE_P))
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw2=%#x!\n", iShw2));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync#if 0 /* useful when running PGMAssertCR3(), a bit too troublesome for general use (TLBs). */
7e8ef90d3160234df0f254131b87af4243d79476vboxsync && !VM_FF_ISSET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3))
7e8ef90d3160234df0f254131b87af4243d79476vboxsync LogFlow(("pgmPoolMonitorChainChanging: iShw=%#x: %RX64 -> freeing it!\n", iShw, uShw.pPDPae->a[iShw].u));
7e8ef90d3160234df0f254131b87af4243d79476vboxsync# ifdef IN_RC /* TLB load - we're pushing things a bit... */
7e8ef90d3160234df0f254131b87af4243d79476vboxsync pgmPoolFree(pPool->CTX_SUFF(pVM), uShw.pPDPae->a[iShw].u & X86_PDE_PAE_PG_MASK, pPage->idx, iShw);
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw=%#x!\n", iShw));
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync#ifdef PGMPOOL_INVALIDATE_UPPER_SHADOW_TABLE_ENTRIES
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync * Causes trouble when the guest uses a PDE to refer to the whole page table level
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync * structure. (Invalidate here; faults later on when it tries to change the page
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync * table entries -> recheck; probably only applies to the RC case.)
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync LogFlow(("pgmPoolMonitorChainChanging: pae pd iShw=%#x: %RX64 -> freeing it!\n", iShw, uShw.pPDPae->a[iShw].u));
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync /* Note: hardcoded PAE implementation dependency */
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync (pPage->enmKind == PGMPOOLKIND_PAE_PD_FOR_PAE_PD) ? PGMPOOL_IDX_PAE_PD : pPage->idx,
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync (pPage->enmKind == PGMPOOLKIND_PAE_PD_FOR_PAE_PD) ? iShw + (pPage->idx - PGMPOOL_IDX_PAE_PD_0) * X86_PG_PAE_ENTRIES : iShw);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* paranoia / a bit assumptive. */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync const unsigned iShw2 = (off + cbWrite - 1) / sizeof(X86PDEPAE);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync AssertReturnVoid(iShw2 < RT_ELEMENTS(uShw.pPDPae->a));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw2=%#x!\n", iShw2));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync#ifdef PGMPOOL_INVALIDATE_UPPER_SHADOW_TABLE_ENTRIES
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync LogFlow(("pgmPoolMonitorChainChanging: pae pd iShw2=%#x: %RX64 -> freeing it!\n", iShw2, uShw.pPDPae->a[iShw2].u));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync /* Note: hardcoded PAE implementation dependency */
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync (pPage->enmKind == PGMPOOLKIND_PAE_PD_FOR_PAE_PD) ? PGMPOOL_IDX_PAE_PD : pPage->idx,
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync (pPage->enmKind == PGMPOOLKIND_PAE_PD_FOR_PAE_PD) ? iShw2 + (pPage->idx - PGMPOOL_IDX_PAE_PD_0) * X86_PG_PAE_ENTRIES : iShw2);
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync * Hopefully this doesn't happen very often:
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync * - touching unused parts of the page
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync * - messing with the bits of pd pointers without changing the physical address
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync if (iShw < X86_PG_PAE_PDPE_ENTRIES) /* don't use RT_ELEMENTS(uShw.pPDPT->a), because that's for long mode only */
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw=%#x!\n", iShw));
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync /* paranoia / a bit assumptive. */
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync const unsigned iShw2 = (off + cbWrite - 1) / sizeof(X86PDPE);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw2=%#x!\n", iShw2));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Assert(pPage->enmKind == PGMPOOLKIND_64BIT_PD_FOR_64BIT_PD);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw=%#x!\n", iShw));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolMonitorChainChanging: pae pd iShw=%#x: %RX64 -> freeing it!\n", iShw, uShw.pPDPae->a[iShw].u));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* paranoia / a bit assumptive. */
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync const unsigned iShw2 = (off + cbWrite - 1) / sizeof(X86PDEPAE);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync AssertReturnVoid(iShw2 < RT_ELEMENTS(uShw.pPDPae->a));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw2=%#x!\n", iShw2));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolMonitorChainChanging: pae pd iShw2=%#x: %RX64 -> freeing it!\n", iShw2, uShw.pPDPae->a[iShw2].u));
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * Hopefully this doesn't happen very often:
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * - messing with the bits of pd pointers without changing the physical address
fb41ad77bcfbdb3aaa1fc9503a37ee6a70dc6461vboxsync if (!VM_FF_ISSET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3))
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync LogFlow(("pgmPoolMonitorChainChanging: pdpt iShw=%#x: %RX64 -> freeing it!\n", iShw, uShw.pPDPT->a[iShw].u));
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync pgmPoolFree(pPool->CTX_SUFF(pVM), uShw.pPDPT->a[iShw].u & X86_PDPE_PG_MASK, pPage->idx, iShw);
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync /* paranoia / a bit assumptive. */
fb41ad77bcfbdb3aaa1fc9503a37ee6a70dc6461vboxsync const unsigned iShw2 = (off + cbWrite - 1) / sizeof(X86PDPE);
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync LogFlow(("pgmPoolMonitorChainChanging: pdpt iShw2=%#x: %RX64 -> freeing it!\n", iShw2, uShw.pPDPT->a[iShw2].u));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync pgmPoolFree(pPool->CTX_SUFF(pVM), uShw.pPDPT->a[iShw2].u & X86_PDPE_PG_MASK, pPage->idx, iShw2);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Hopefully this doesn't happen very often:
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * - messing with the bits of pd pointers without changing the physical address
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync if (!VM_FF_ISSET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3))
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync LogFlow(("pgmPoolMonitorChainChanging: pml4 iShw=%#x: %RX64 -> freeing it!\n", iShw, uShw.pPML4->a[iShw].u));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync pgmPoolFree(pPool->CTX_SUFF(pVM), uShw.pPML4->a[iShw].u & X86_PML4E_PG_MASK, pPage->idx, iShw);
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync /* paranoia / a bit assumptive. */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync const unsigned iShw2 = (off + cbWrite - 1) / sizeof(X86PML4E);
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync LogFlow(("pgmPoolMonitorChainChanging: pml4 iShw2=%#x: %RX64 -> freeing it!\n", iShw2, uShw.pPML4->a[iShw2].u));
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync pgmPoolFree(pPool->CTX_SUFF(pVM), uShw.pPML4->a[iShw2].u & X86_PML4E_PG_MASK, pPage->idx, iShw2);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync#endif /* IN_RING0 */
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync AssertFatalMsgFailed(("enmKind=%d\n", pPage->enmKind));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Checks if a access could be a fork operation in progress.
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync * Meaning, that the guest is setuping up the parent process for Copy-On-Write.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @returns true if it's likly that we're forking, otherwise false.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pPool The pool.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pCpu The disassembled instruction.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * @param offFault The access offset.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsyncDECLINLINE(bool) pgmPoolMonitorIsForking(PPGMPOOL pPool, PDISCPUSTATE pCpu, unsigned offFault)
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * i386 linux is using btr to clear X86_PTE_RW.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * The functions involved are (2.6.16 source inspection):
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * clear_bit
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * ptep_set_wrprotect
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * copy_one_pte
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * copy_pte_range
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * copy_pmd_range
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * copy_pud_range
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * copy_page_range
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * copy_process
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync /** @todo Validate that the bit index is X86_PTE_RW. */
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync STAM_COUNTER_INC(&pPool->CTX_MID_Z(StatMonitor,Fork));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync return true;
b8bb9c9f6b8ebfd0a7d6df0c0289f9fe80241750vboxsync return false;
5366e994777f9d9391cf809dc77610f57270d75dvboxsync * Determine whether the page is likely to have been reused.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @returns true if we consider the page as being reused for a different purpose.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * @returns false if we consider it to still be a paging page.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pVM VM Handle.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * @param pPage The page in question.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * @param pRegFrame Trap register frame.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * @param pCpu The disassembly info for the faulting instruction.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * @param pvFault The fault address.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * @remark The REP prefix check is left to the caller because of STOSD/W.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsyncDECLINLINE(bool) pgmPoolMonitorIsReused(PVM pVM, PPGMPOOLPAGE pPage, PCPUMCTXCORE pRegFrame, PDISCPUSTATE pCpu, RTGCPTR pvFault)
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync /** @todo could make this general, faulting close to rsp should be safe reuse heuristic. */
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync /* Fault caused by stack writes while trying to inject an interrupt event. */
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync Log(("pgmPoolMonitorIsReused: reused %RGv for interrupt stack (rsp=%RGv).\n", pvFault, pRegFrame->rsp));
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync return true;
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync /* call implies the actual push of the return address faulted */
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync return true;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync return true;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync return true;
b8bb9c9f6b8ebfd0a7d6df0c0289f9fe80241750vboxsync return true;
b8bb9c9f6b8ebfd0a7d6df0c0289f9fe80241750vboxsync return true;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync return true;
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync case OP_MOVNTDQ: /* solaris - hwblkclr & hwblkpagecopy */
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync return true;
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync return true;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync return false;
ca3db470494a8b6eaec69ea37468a5cda65e2da8vboxsync return true;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync //if (pPage->fCR3Mix)
ca3db470494a8b6eaec69ea37468a5cda65e2da8vboxsync // return false;
ca3db470494a8b6eaec69ea37468a5cda65e2da8vboxsync return false;
ca3db470494a8b6eaec69ea37468a5cda65e2da8vboxsync * Flushes the page being accessed.
ca3db470494a8b6eaec69ea37468a5cda65e2da8vboxsync * @returns VBox status code suitable for scheduling.
ca3db470494a8b6eaec69ea37468a5cda65e2da8vboxsync * @param pVM The VM handle.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pPool The pool.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pPage The pool page (head).
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pCpu The disassembly of the write instruction.
ca3db470494a8b6eaec69ea37468a5cda65e2da8vboxsync * @param pRegFrame The trap register frame.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param GCPhysFault The fault address as guest physical address.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pvFault The fault address.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsyncstatic int pgmPoolAccessHandlerFlush(PVM pVM, PPGMPOOL pPool, PPGMPOOLPAGE pPage, PDISCPUSTATE pCpu,
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync PCPUMCTXCORE pRegFrame, RTGCPHYS GCPhysFault, RTGCPTR pvFault)
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * First, do the flushing.
b8bb9c9f6b8ebfd0a7d6df0c0289f9fe80241750vboxsync * Emulate the instruction (xp/w2k problem, requires pc/cr2/sp detection).
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync int rc2 = EMInterpretInstructionCPU(pVM, pCpu, pRegFrame, pvFault, &cbWritten);
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync if (PATMIsPatchGCAddr(pVM, (RTRCPTR)pRegFrame->eip))
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync LogFlow(("pgmPoolAccessHandlerPTWorker: Interpretation failed for patch code %04x:%RGv, ignoring.\n",
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsync STAM_COUNTER_INC(&pPool->StatMonitorRZIntrFailPatch2);
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync STAM_COUNTER_INC(&pPool->CTX_MID_Z(StatMonitor,EmulateInstr));
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsync /* See use in pgmPoolAccessHandlerSimple(). */
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsync LogFlow(("pgmPoolAccessHandlerPT: returns %Rrc (flushed)\n", rc));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Handles the STOSD write accesses.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @returns VBox status code suitable for scheduling.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pVM The VM handle.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pPool The pool.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pPage The pool page (head).
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync * @param pCpu The disassembly of the write instruction.
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync * @param pRegFrame The trap register frame.
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync * @param GCPhysFault The fault address as guest physical address.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pvFault The fault address.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsyncDECLINLINE(int) pgmPoolAccessHandlerSTOSD(PVM pVM, PPGMPOOL pPool, PPGMPOOLPAGE pPage, PDISCPUSTATE pCpu,
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync PCPUMCTXCORE pRegFrame, RTGCPHYS GCPhysFault, RTGCPTR pvFault)
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Increment the modification counter and insert it into the list
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * of modified pages the first time.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Execute REP STOSD.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * This ASSUMES that we're not invoked by Trap0e on in a out-of-sync
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * write situation, meaning that it's safe to write here.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync pgmPoolMonitorChainChanging(pPool, pPage, GCPhysFault, (RTGCPTR)pu32, NULL);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync PGMPhysSimpleWriteGCPhys(pVM, GCPhysFault, &pRegFrame->eax, 4);
cc1ef2ef9bbc6a0ff964928d61b7298e5bfcce5fvboxsync /* See use in pgmPoolAccessHandlerSimple(). */
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync * Handles the simple write accesses.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @returns VBox status code suitable for scheduling.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pVM The VM handle.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pPool The pool.
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync * @param pPage The pool page (head).
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pCpu The disassembly of the write instruction.
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync * @param pRegFrame The trap register frame.
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync * @param GCPhysFault The fault address as guest physical address.
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync * @param pvFault The fault address.
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsyncDECLINLINE(int) pgmPoolAccessHandlerSimple(PVM pVM, PPGMPOOL pPool, PPGMPOOLPAGE pPage, PDISCPUSTATE pCpu,
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync PCPUMCTXCORE pRegFrame, RTGCPHYS GCPhysFault, RTGCPTR pvFault)
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Increment the modification counter and insert it into the list
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * of modified pages the first time.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Clear all the pages. ASSUMES that pvFault is readable.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync pgmPoolMonitorChainChanging(pPool, pPage, GCPhysFault, pvFault, pCpu);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Interpret the instruction.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync int rc = EMInterpretInstructionCPU(pVM, pCpu, pRegFrame, pvFault, &cb);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync LogFlow(("pgmPoolAccessHandlerPTWorker: Interpretation failed for %04x:%RGv - opcode=%d\n",
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync pRegFrame->cs, (RTGCPTR)pRegFrame->rip, pCpu->pCurInstr->opcode));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync STAM_COUNTER_INC(&pPool->CTX_MID_Z(StatMonitor,EmulateInstr));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Quick hack, with logging enabled we're getting stale
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * code TLBs but no data TLB for EIP and crash in EMInterpretDisasOne.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Flushing here is BAD and expensive, I think EMInterpretDisasOne will
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * have to be fixed to support this. But that'll have to wait till next week.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * An alternative is to keep track of the changed PTEs together with the
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * GCPhys from the guest PT. This may proove expensive though.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * At the moment, it's VITAL that it's done AFTER the instruction interpreting
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * because we need the stale TLBs in some cases (XP boot). This MUST be fixed properly!
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync LogFlow(("pgmPoolAccessHandlerSimple: returns %Rrc cb=%d\n", rc, cb));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * \#PF Handler callback for PT write accesses.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * @returns VBox status code (appropriate for GC return).
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * @param pVM VM Handle.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * @param uErrorCode CPU Error code.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * @param pRegFrame Trap register frame.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * NULL on DMA and other non CPU access.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * @param pvFault The fault address (cr2).
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param GCPhysFault The GC physical address corresponding to pvFault.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * @param pvUser User argument.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsyncDECLEXPORT(int) pgmPoolAccessHandler(PVM pVM, RTGCUINT uErrorCode, PCPUMCTXCORE pRegFrame, RTGCPTR pvFault, RTGCPHYS GCPhysFault, void *pvUser)
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync STAM_PROFILE_START(&pVM->pgm.s.CTX_SUFF(pPool)->CTX_SUFF_Z(StatMonitor), a);
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync LogFlow(("pgmPoolAccessHandler: pvFault=%RGv pPage=%p:{.idx=%d} GCPhysFault=%RGp\n", pvFault, pPage, pPage->idx, GCPhysFault));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * We should ALWAYS have the list head as user parameter. This
15c6c4ce0082362b8b81e15c3605f2d3aca69a21vboxsync * is because we use that page to record the changes.
15c6c4ce0082362b8b81e15c3605f2d3aca69a21vboxsync * Disassemble the faulting instruction.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync int rc = EMInterpretDisasOne(pVM, pRegFrame, &Cpu, NULL);
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * Check if it's worth dealing with.
004d74842597dacc4009803171296dfcf9398c69vboxsync bool fReused = false;
004d74842597dacc4009803171296dfcf9398c69vboxsync if ( ( pPage->cModifications < 48 /** @todo #define */ /** @todo need to check that it's not mapping EIP. */ /** @todo adjust this! */
004d74842597dacc4009803171296dfcf9398c69vboxsync && !(fReused = pgmPoolMonitorIsReused(pVM, pPage, pRegFrame, &Cpu, pvFault))
004d74842597dacc4009803171296dfcf9398c69vboxsync && !pgmPoolMonitorIsForking(pPool, &Cpu, GCPhysFault & PAGE_OFFSET_MASK))
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Simple instructions, no REP prefix.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync rc = pgmPoolAccessHandlerSimple(pVM, pPool, pPage, &Cpu, pRegFrame, GCPhysFault, pvFault);
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync STAM_PROFILE_STOP_EX(&pVM->pgm.s.CTX_SUFF(pPool)->CTX_SUFF_Z(StatMonitor), &pPool->CTX_MID_Z(StatMonitor,Handled), a);
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Windows is frequently doing small memset() operations (netio test 4k+).
004d74842597dacc4009803171296dfcf9398c69vboxsync * We have to deal with these or we'll kill the cache and performance.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync && pRegFrame->ecx * 4 <= PAGE_SIZE - ((uintptr_t)pvFault & PAGE_OFFSET_MASK)
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync && (pRegFrame->eax == 0 || pRegFrame->eax == 0x80) /* the two values observed. */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync rc = pgmPoolAccessHandlerSTOSD(pVM, pPool, pPage, &Cpu, pRegFrame, GCPhysFault, pvFault);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync STAM_PROFILE_STOP_EX(&pVM->pgm.s.CTX_SUFF(pPool)->CTX_SUFF_Z(StatMonitor), &pPool->CTX_MID_Z(StatMonitor,RepStosd), a);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync /* REP prefix, don't bother. */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync STAM_COUNTER_INC(&pPool->CTX_MID_Z(StatMonitor,RepPrefix));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync Log4(("pgmPoolAccessHandler: eax=%#x ecx=%#x edi=%#x esi=%#x rip=%RGv opcode=%d prefix=%#x\n",
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync pRegFrame->eax, pRegFrame->ecx, pRegFrame->edi, pRegFrame->esi, (RTGCPTR)pRegFrame->rip, Cpu.pCurInstr->opcode, Cpu.prefix));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Not worth it, so flush it.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * If we considered it to be reused, don't to back to ring-3
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * to emulate failed instructions since we usually cannot
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * interpret then. This may be a bit risky, in which case
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * the reuse detection must be fixed.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync rc = pgmPoolAccessHandlerFlush(pVM, pPool, pPage, &Cpu, pRegFrame, GCPhysFault, pvFault);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync STAM_PROFILE_STOP_EX(&pVM->pgm.s.CTX_SUFF(pPool)->CTX_SUFF_Z(StatMonitor), &pPool->CTX_MID_Z(StatMonitor,FlushPage), a);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync# endif /* !IN_RING3 */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync#endif /* PGMPOOL_WITH_MONITORING */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Inserts a page into the GCPhys hash table.
5366e994777f9d9391cf809dc77610f57270d75dvboxsync * @param pPool The pool.
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsync * @param pPage The page.
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsyncDECLINLINE(void) pgmPoolHashInsert(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
5366e994777f9d9391cf809dc77610f57270d75dvboxsync Log3(("pgmPoolHashInsert: %RGp\n", pPage->GCPhys));
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync Assert(pPage->GCPhys != NIL_RTGCPHYS); Assert(pPage->iNext == NIL_PGMPOOL_IDX);
d8e12fa5dd1c35282b98cb165e42b6b395cf971bvboxsync * Removes a page from the GCPhys hash table.
d8e12fa5dd1c35282b98cb165e42b6b395cf971bvboxsync * @param pPool The pool.
d8e12fa5dd1c35282b98cb165e42b6b395cf971bvboxsync * @param pPage The page.
d8e12fa5dd1c35282b98cb165e42b6b395cf971bvboxsyncDECLINLINE(void) pgmPoolHashRemove(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
d8e12fa5dd1c35282b98cb165e42b6b395cf971bvboxsync Log3(("pgmPoolHashRemove: %RGp\n", pPage->GCPhys));
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync AssertReleaseMsgFailed(("GCPhys=%RGp idx=%#x\n", pPage->GCPhys, pPage->idx));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Frees up one cache page.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * @returns VBox status code.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @retval VINF_SUCCESS on success.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * @retval VERR_PGM_POOL_CLEARED if the deregistration of a physical handler will cause a light weight pool flush.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * @param pPool The pool.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * @param iUser The user index.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsyncstatic int pgmPoolCacheFreeOne(PPGMPOOL pPool, uint16_t iUser)
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync Assert(pPool->iAgeHead != pPool->iAgeTail); /* We shouldn't be here if there < 2 cached entries! */
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * Select one page from the tail of the age list.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync/* This is the alternative to the SyncCR3 pgmPoolCacheUsed calls.
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync if (pPool->aPages[iToFree].iUserHead != NIL_PGMPOOL_USER_INDEX)
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync uint16_t i = pPool->aPages[iToFree].iAgePrev;
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync for (unsigned j = 0; j < 10 && i != NIL_PGMPOOL_USER_INDEX; j++, i = pPool->aPages[i].iAgePrev)
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync if (pPool->aPages[iToFree].iUserHead == NIL_PGMPOOL_USER_INDEX)
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync iToFree = i;
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync * Reject any attempts at flushing the currently active shadow CR3 mapping
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync if (PGMGetHyperCR3(pPool->CTX_SUFF(pVM)) == pPage->Core.Key)
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync /* Refresh the cr3 mapping by putting it at the head of the age list. */
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync * Checks if a kind mismatch is really a page being reused
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync * or if it's just normal remappings.
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync * @returns true if reused and the cached page (enmKind1) should be flushed
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync * @returns false if not reused.
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync * @param enmKind1 The kind of the cached page.
40c1a23e86c79b24a917a43c186b2e54504d12c1vboxsync * @param enmKind2 The kind of the requested page.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic bool pgmPoolCacheReusedByKind(PGMPOOLKIND enmKind1, PGMPOOLKIND enmKind2)
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Never reuse them. There is no remapping in non-paging mode.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync return true;
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * It's perfectly fine to reuse these, except for PAE and non-paging stuff.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync return true;
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync return false;
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * It's perfectly fine to reuse these, except for PAE and non-paging stuff.
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync return true;
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync return false;
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * These cannot be flushed, and it's common to reuse the PDs as PTs.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync return false;
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * Attempts to satisfy a pgmPoolAlloc request from the cache.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @returns VBox status code.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @retval VINF_PGM_CACHED_PAGE on success.
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * @retval VERR_FILE_NOT_FOUND if not found.
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * @param pPool The pool.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param GCPhys The GC physical address of the page we're gonna shadow.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param enmKind The kind of mapping.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param iUser The shadow page pool index of the user table.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param iUserTable The index into the user table (shadowed).
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * @param ppPage Where to store the pointer to the page.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic int pgmPoolCacheAlloc(PPGMPOOL pPool, RTGCPHYS GCPhys, PGMPOOLKIND enmKind, uint16_t iUser, uint32_t iUserTable, PPPGMPOOLPAGE ppPage)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Look up the GCPhys in the hash.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync Log3(("pgmPoolCacheAlloc: %RGp kind %d iUser=%d iUserTable=%x SLOT=%d\n", GCPhys, enmKind, iUser, iUserTable, i));
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync Log3(("pgmPoolCacheAlloc: slot %d found page %RGp\n", i, pPage->GCPhys));
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync int rc = pgmPoolTrackAddUser(pPool, pPage, iUser, iUserTable);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * The kind is different. In some cases we should now flush the page
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * as it has been reused, but in most cases this is normal remapping
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * of PDs as PT or big pages using the GCPhys field in a slightly
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * different way than the other kinds.
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync if (pgmPoolCacheReusedByKind((PGMPOOLKIND)pPage->enmKind, enmKind))
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync pgmPoolFlushPage(pPool, pPage); /* ASSUMES that VERR_PGM_POOL_CLEARED will be returned by pgmPoolTracInsert. */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync } while (i != NIL_PGMPOOL_IDX);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync Log3(("pgmPoolCacheAlloc: Missed GCPhys=%RGp enmKind=%d\n", GCPhys, enmKind));
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync * Inserts a page into the cache.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pPool The pool.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pPage The cached page.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param fCanBeCached Set if the page is fit for caching from the caller's point of view.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic void pgmPoolCacheInsert(PPGMPOOL pPool, PPGMPOOLPAGE pPage, bool fCanBeCached)
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Insert into the GCPhys hash if the page is fit for that.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync Log3(("pgmPoolCacheInsert: Caching %p:{.Core=%RHp, .idx=%d, .enmKind=%d, GCPhys=%RGp}\n",
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync pPage, pPage->Core.Key, pPage->idx, pPage->enmKind, pPage->GCPhys));
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync Log3(("pgmPoolCacheInsert: Not caching %p:{.Core=%RHp, .idx=%d, .enmKind=%d, GCPhys=%RGp}\n",
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync pPage, pPage->Core.Key, pPage->idx, pPage->enmKind, pPage->GCPhys));
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * Insert at the head of the age list.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync pPool->aPages[pPool->iAgeHead].iAgePrev = pPage->idx;
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * Flushes a cached page.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * @param pPool The pool.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * @param pPage The cached page.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsyncstatic void pgmPoolCacheFlushPage(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync Log3(("pgmPoolCacheFlushPage: %RGp\n", pPage->GCPhys));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Remove the page from the hash.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Remove it from the age list.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync pPool->aPages[pPage->iAgeNext].iAgePrev = pPage->iAgePrev;
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync pPool->aPages[pPage->iAgePrev].iAgeNext = pPage->iAgeNext;
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync#endif /* PGMPOOL_WITH_CACHE */
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Looks for pages sharing the monitor.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @returns Pointer to the head page.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @returns NULL if not found.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pPool The Pool
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pNewPage The page which is going to be monitored.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic PPGMPOOLPAGE pgmPoolMonitorGetPageByGCPhys(PPGMPOOL pPool, PPGMPOOLPAGE pNewPage)
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Look up the GCPhys in the hash.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync RTGCPHYS GCPhys = pNewPage->GCPhys & ~(RTGCPHYS)(PAGE_SIZE - 1);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync /* find the head */
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync /* ignore, no monitoring. */
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync AssertFatalMsgFailed(("enmKind=%d idx=%d\n", pPage->enmKind, pPage->idx));
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync } while (i != NIL_PGMPOOL_IDX);
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync * Enabled write monitoring of a guest page.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * @returns VBox status code.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * @retval VINF_SUCCESS on success.
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * @retval VERR_PGM_POOL_CLEARED if the registration of the physical handler will cause a light weight pool flush.
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * @param pPool The pool.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * @param pPage The cached page.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsyncstatic int pgmPoolMonitorInsert(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync LogFlow(("pgmPoolMonitorInsert %RGp\n", pPage->GCPhys & ~(RTGCPHYS)(PAGE_SIZE - 1)));
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * Filter out the relevant kinds.
5366e994777f9d9391cf809dc77610f57270d75dvboxsync /* Nothing to monitor here. */
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync AssertFatalMsgFailed(("This can't happen! enmKind=%d\n", pPage->enmKind));
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * Install handler.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync PPGMPOOLPAGE pPageHead = pgmPoolMonitorGetPageByGCPhys(pPool, pPage);
e2489bd9ef063ae006feaebc3318ffa4143f6e16vboxsync Assert(pPageHead != pPage); Assert(pPageHead->iMonitoredNext != pPage->idx);
e2489bd9ef063ae006feaebc3318ffa4143f6e16vboxsync pPool->aPages[pPageHead->iMonitoredNext].iMonitoredPrev = pPage->idx;
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync Assert(pPage->iMonitoredNext == NIL_PGMPOOL_IDX); Assert(pPage->iMonitoredPrev == NIL_PGMPOOL_IDX);
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync const RTGCPHYS GCPhysPage = pPage->GCPhys & ~(RTGCPHYS)(PAGE_SIZE - 1);
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync rc = PGMHandlerPhysicalRegisterEx(pVM, PGMPHYSHANDLERTYPE_PHYSICAL_WRITE,
cc1ef2ef9bbc6a0ff964928d61b7298e5bfcce5fvboxsync pPool->pfnAccessHandlerR3, MMHyperCCToR3(pVM, pPage),
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync pPool->pfnAccessHandlerR0, MMHyperCCToR0(pVM, pPage),
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync pPool->pfnAccessHandlerRC, MMHyperCCToRC(pVM, pPage),
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync /** @todo we should probably deal with out-of-memory conditions here, but for now increasing
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync * the heap size should suffice. */
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync if (pVM->pgm.s.fSyncFlags & PGM_SYNC_CLEAR_PGM_POOL)
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync * Disables write monitoring of a guest page.
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync * @returns VBox status code.
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync * @retval VINF_SUCCESS on success.
6b9d50a0f466bd5a61458ed53925480ab28a3c17vboxsync * @retval VERR_PGM_POOL_CLEARED if the deregistration of the physical handler will cause a light weight pool flush.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * @param pPool The pool.
6b9d50a0f466bd5a61458ed53925480ab28a3c17vboxsync * @param pPage The cached page.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic int pgmPoolMonitorFlush(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync * Filter out the relevant kinds.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync /* Nothing to monitor here. */
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync AssertFatalMsgFailed(("This can't happen! enmKind=%d\n", pPage->enmKind));
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync * Remove the page from the monitored list or uninstall it if last.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync PPGMPOOLPAGE pNewHead = &pPool->aPages[pPage->iMonitoredNext];
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync rc = PGMHandlerPhysicalChangeCallbacks(pVM, pPage->GCPhys & ~(RTGCPHYS)(PAGE_SIZE - 1),
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync pPool->pfnAccessHandlerR3, MMHyperCCToR3(pVM, pNewHead),
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync pPool->pfnAccessHandlerR0, MMHyperCCToR0(pVM, pNewHead),
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync pPool->pfnAccessHandlerRC, MMHyperCCToRC(pVM, pNewHead),
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync pPool->aPages[pPage->iMonitoredPrev].iMonitoredNext = pPage->iMonitoredNext;
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync pPool->aPages[pPage->iMonitoredNext].iMonitoredPrev = pPage->iMonitoredPrev;
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync rc = PGMHandlerPhysicalDeregister(pVM, pPage->GCPhys & ~(RTGCPHYS)(PAGE_SIZE - 1));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync if (pVM->pgm.s.fSyncFlags & PGM_SYNC_CLEAR_PGM_POOL)
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Remove it from the list of modified pages (if in it).
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Set or clear the fCR3Mix attribute in a chain of monitored pages.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pPool The Pool.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pPage A page in the chain.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param fCR3Mix The new fCR3Mix value.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic void pgmPoolMonitorChainChangeCR3Mix(PPGMPOOL pPool, PPGMPOOLPAGE pPage, bool fCR3Mix)
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync /* current */
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync /* before */
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync /* after */
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * Installs or modifies monitoring of a CR3 page (special).
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * We're pretending the CR3 page is shadowed by the pool so we can use the
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * generic mechanisms in detecting chained monitoring. (This also gives us a
9e7e0b61d29309a0ed7af9472c8d6d865f9e8a2dvboxsync * tast of what code changes are required to really pool CR3 shadow pages.)
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * @returns VBox status code.
9e7e0b61d29309a0ed7af9472c8d6d865f9e8a2dvboxsync * @param pPool The pool.
9e7e0b61d29309a0ed7af9472c8d6d865f9e8a2dvboxsync * @param idxRoot The CR3 (root) page index.
9e7e0b61d29309a0ed7af9472c8d6d865f9e8a2dvboxsync * @param GCPhysCR3 The (new) CR3 value.
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsyncint pgmPoolMonitorMonitorCR3(PPGMPOOL pPool, uint16_t idxRoot, RTGCPHYS GCPhysCR3)
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync Assert(idxRoot != NIL_PGMPOOL_IDX && idxRoot < PGMPOOL_IDX_FIRST);
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync LogFlow(("pgmPoolMonitorMonitorCR3: idxRoot=%d pPage=%p:{.GCPhys=%RGp, .fMonitored=%d} GCPhysCR3=%RGp\n",
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync idxRoot, pPage, pPage->GCPhys, pPage->fMonitored, GCPhysCR3));
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * The unlikely case where it already matches.
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * Flush the current monitoring and remove it from the hash.
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync pgmPoolMonitorChainChangeCR3Mix(pPool, pPage, false);
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsync * Monitor the page at the new location and insert it into the hash.
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsync pgmPoolMonitorChainChangeCR3Mix(pPool, pPage, true);
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * Removes the monitoring of a CR3 page (special).
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * @returns VBox status code.
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * @param pPool The pool.
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * @param idxRoot The CR3 (root) page index.
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsyncint pgmPoolMonitorUnmonitorCR3(PPGMPOOL pPool, uint16_t idxRoot)
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsync Assert(idxRoot != NIL_PGMPOOL_IDX && idxRoot < PGMPOOL_IDX_FIRST);
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsync LogFlow(("pgmPoolMonitorUnmonitorCR3: idxRoot=%d pPage=%p:{.GCPhys=%RGp, .fMonitored=%d}\n",
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsync idxRoot, pPage, pPage->GCPhys, pPage->fMonitored));
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsync pgmPoolMonitorChainChangeCR3Mix(pPool, pPage, false);
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync# endif /* PGMPOOL_WITH_MIXED_PT_CR3 */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Inserts the page into the list of modified pages.
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsync * @param pPool The pool.
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync * @param pPage The page.
cd5df721f068659172f3bf95de8fedeb465f057dvboxsyncvoid pgmPoolMonitorModifiedInsert(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsync Log3(("pgmPoolMonitorModifiedInsert: idx=%d\n", pPage->idx));
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync ("Next=%d Prev=%d idx=%d cModifications=%d Head=%d cModifiedPages=%d\n",
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsync pPage->iModifiedNext, pPage->iModifiedPrev, pPage->idx, pPage->cModifications,
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pPool->aPages[pPool->iModifiedHead].iModifiedPrev = pPage->idx;
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync if (pPool->cModifiedPages > pPool->cModifiedPagesHigh)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Removes the page from the list of modified pages and resets the
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * moficiation counter.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPool The pool.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPage The page which is believed to be in the list of modified pages.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsyncstatic void pgmPoolMonitorModifiedRemove(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync Log3(("pgmPoolMonitorModifiedRemove: idx=%d cModifications=%d\n", pPage->idx, pPage->cModifications));
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync pPool->aPages[pPage->iModifiedNext].iModifiedPrev = NIL_PGMPOOL_IDX;
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pPool->aPages[pPage->iModifiedPrev].iModifiedNext = pPage->iModifiedNext;
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pPool->aPages[pPage->iModifiedNext].iModifiedPrev = pPage->iModifiedPrev;
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Zaps the list of modified pages, resetting their modification counters in the process.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pVM The VM handle.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync LogFlow(("pgmPoolMonitorModifiedClearAll: cModifiedPages=%d\n", pPool->cModifiedPages));
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync AssertMsg(cPages == pPool->cModifiedPages, ("%d != %d\n", cPages, pPool->cModifiedPages));
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Clear all shadow pages and clear all modification counters.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pVM The VM handle.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @remark Should only be used when monitoring is available, thus placed in
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * the PGMPOOL_WITH_MONITORING #ifdef.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync LogFlow(("pgmPoolClearAll: cUsedPages=%d\n", pPool->cUsedPages));
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Iterate all the pages until we've encountered all that in use.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * This is simple but not quite optimal solution.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync unsigned cModifiedPages = 0; NOREF(cModifiedPages);
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * We only care about shadow page tables.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync void *pvShw = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync /* fall thru */
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync Assert(!pPage->cModifications || ++cModifiedPages);
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync Assert(pPage->iModifiedNext == NIL_PGMPOOL_IDX || pPage->cModifications);
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync Assert(pPage->iModifiedPrev == NIL_PGMPOOL_IDX || pPage->cModifications);
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync /* swipe the special pages too. */
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync for (iPage = PGMPOOL_IDX_FIRST_SPECIAL; iPage < PGMPOOL_IDX_FIRST; iPage++)
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync Assert(!pPage->cModifications || ++cModifiedPages);
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync Assert(pPage->iModifiedNext == NIL_PGMPOOL_IDX || pPage->cModifications);
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync Assert(pPage->iModifiedPrev == NIL_PGMPOOL_IDX || pPage->cModifications);
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync AssertMsg(cModifiedPages == pPool->cModifiedPages, ("%d != %d\n", cModifiedPages, pPool->cModifiedPages));
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * Clear all the GCPhys links and rebuild the phys ext free list.
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync for (PPGMRAMRANGE pRam = pPool->CTX_SUFF(pVM)->pgm.s.CTX_SUFF(pRamRanges);
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync while (iPage-- > 0)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pRam->aPages[iPage].HCPhys &= MM_RAM_FLAGS_NO_REFS_MASK; /** @todo PAGE FLAGS */
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync PPGMPOOLPHYSEXT paPhysExts = pPool->CTX_SUFF(paPhysExts);
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync for (unsigned i = 0; i < cMaxPhysExts; i++)
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync paPhysExts[cMaxPhysExts - 1].iNext = NIL_PGMPOOL_PHYSEXT_INDEX;
ecf100db90e8e3af96312908282d3c20e754fbe8vboxsync * Handle SyncCR3 pool tasks
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @returns VBox status code.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @retval VINF_SUCCESS if successfully added.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @retval VINF_PGM_SYNC_CR3 is it needs to be deferred to ring 3 (GC only)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pVM The VM handle.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @remark Should only be used when monitoring is available, thus placed in
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * the PGMPOOL_WITH_MONITORING #ifdef.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * When monitoring shadowed pages, we reset the modification counters on CR3 sync.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Occasionally we will have to clear all the shadow page tables because we wanted
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * to monitor a page which was mapped by too many shadowed page tables. This operation
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * sometimes refered to as a 'lightweight flush'.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync if (!(pVM->pgm.s.fSyncFlags & PGM_SYNC_CLEAR_PGM_POOL))
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync LogFlow(("SyncCR3: PGM_SYNC_CLEAR_PGM_POOL is set -> VINF_PGM_SYNC_CR3\n"));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync VM_FF_SET(pVM, VM_FF_PGM_SYNC_CR3); /** @todo no need to do global sync, right? */
cd5df721f068659172f3bf95de8fedeb465f057dvboxsync#endif /* PGMPOOL_WITH_MONITORING */
d8e12fa5dd1c35282b98cb165e42b6b395cf971bvboxsync * Frees up at least one user entry.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @returns VBox status code.
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * @retval VINF_SUCCESS if successfully added.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @retval VERR_PGM_POOL_FLUSHED if the pool was flushed.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pPool The pool.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * @param iUser The user index.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsyncstatic int pgmPoolTrackFreeOneUser(PPGMPOOL pPool, uint16_t iUser)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Just free cached pages in a braindead fashion.
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync /** @todo walk the age list backwards and free the first with usage. */
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync } while (pPool->iUserFreeHead == NIL_PGMPOOL_USER_INDEX);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Lazy approach.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* @todo incompatible with long mode paging (cr3 root will be flushed) */
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * Inserts a page into the cache.
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * This will create user node for the page, insert it into the GCPhys
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * hash, and insert it into the age list.
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * @returns VBox status code.
c99b597540585068d22dde4c9f74730305f24097vboxsync * @retval VINF_SUCCESS if successfully added.
c99b597540585068d22dde4c9f74730305f24097vboxsync * @retval VERR_PGM_POOL_FLUSHED if the pool was flushed.
c99b597540585068d22dde4c9f74730305f24097vboxsync * @retval VERR_PGM_POOL_CLEARED if the deregistration of the physical handler will cause a light weight pool flush.
c99b597540585068d22dde4c9f74730305f24097vboxsync * @param pPool The pool.
c99b597540585068d22dde4c9f74730305f24097vboxsync * @param pPage The cached page.
c99b597540585068d22dde4c9f74730305f24097vboxsync * @param GCPhys The GC physical address of the page we're gonna shadow.
c99b597540585068d22dde4c9f74730305f24097vboxsync * @param iUser The user index.
c99b597540585068d22dde4c9f74730305f24097vboxsync * @param iUserTable The user table index.
c99b597540585068d22dde4c9f74730305f24097vboxsyncDECLINLINE(int) pgmPoolTrackInsert(PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTGCPHYS GCPhys, uint16_t iUser, uint32_t iUserTable)
c99b597540585068d22dde4c9f74730305f24097vboxsync LogFlow(("pgmPoolTrackInsert iUser %d iUserTable %d\n", iUser, iUserTable));
c99b597540585068d22dde4c9f74730305f24097vboxsync * Find free a user node.
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * Unlink the user node from the free list,
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * initialize and insert it into the user list.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Insert into cache and enable monitoring of the guest page if enabled.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Until we implement caching of all levels, including the CR3 one, we'll
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * have to make sure we don't try monitor & cache any recursive reuse of
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * a monitored CR3 page. Because all windows versions are doing this we'll
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * have to be able to do combined access monitoring, CR3 + PT and
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * PD + PT (guest PAE).
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * We're now cooperating with the CR3 monitor if an uncachable page is found.
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync#if defined(PGMPOOL_WITH_MONITORING) || defined(PGMPOOL_WITH_CACHE)
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync const bool fCanBeMonitored = true;
c99b597540585068d22dde4c9f74730305f24097vboxsync bool fCanBeMonitored = pPool->CTX_SUFF(pVM)->pgm.s.GCPhysGstCR3Monitored == NIL_RTGCPHYS
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync || (GCPhys & X86_PTE_PAE_PG_MASK) != (pPool->CTX_SUFF(pVM)->pgm.s.GCPhysGstCR3Monitored & X86_PTE_PAE_PG_MASK)
02651f98b4320e70a300ba1ebe95270096ebfd4dvboxsync pgmPoolCacheInsert(pPool, pPage, fCanBeMonitored); /* This can be expanded. */
c99b597540585068d22dde4c9f74730305f24097vboxsync /* 'Failed' - free the usage, and keep it in the cache (if enabled). */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync#endif /* PGMPOOL_WITH_MONITORING */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync# ifdef PGMPOOL_WITH_CACHE /* (only used when the cache is enabled.) */
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * Adds a user reference to a page.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * This will
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * This will move the page to the head of the
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @returns VBox status code.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @retval VINF_SUCCESS if successfully added.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @retval VERR_PGM_POOL_FLUSHED if the pool was flushed.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * @param pPool The pool.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * @param pPage The cached page.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * @param iUser The user index.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * @param iUserTable The user table.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsyncstatic int pgmPoolTrackAddUser(PPGMPOOL pPool, PPGMPOOLPAGE pPage, uint16_t iUser, uint32_t iUserTable)
5366e994777f9d9391cf809dc77610f57270d75dvboxsync LogFlow(("pgmPoolTrackAddUser iUser %d iUserTable %d\n", iUser, iUserTable));
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * Check that the entry doesn't already exists.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync AssertMsg(paUsers[i].iUser != iUser || paUsers[i].iUserTable != iUserTable, ("%x %x vs new %x %x\n", paUsers[i].iUser, paUsers[i].iUserTable, iUser, iUserTable));
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * Allocate a user node.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * Initialize the user node and insert it.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Tell the cache to update its replacement stats for this page.
36f3c24e4ad9c6b813767db1faeabbe7e2ecc057vboxsync# endif /* PGMPOOL_WITH_CACHE */
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Frees a user record associated with a page.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * This does not clear the entry in the user table, it simply replaces the
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * user record to the chain of free records.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pPool The pool.
36f3c24e4ad9c6b813767db1faeabbe7e2ecc057vboxsync * @param HCPhys The HC physical address of the shadow page.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param iUser The shadow page pool index of the user table.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param iUserTable The index into the user table (shadowed).
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsyncstatic void pgmPoolTrackFreeUser(PPGMPOOL pPool, PPGMPOOLPAGE pPage, uint16_t iUser, uint32_t iUserTable)
ca3db470494a8b6eaec69ea37468a5cda65e2da8vboxsync * Unlink and free the specified user entry.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync /* Special: For PAE and 32-bit paging, there is usually no more than one user. */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* General: Linear search. */
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync /* Fatal: didn't find it */
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync AssertFatalMsgFailed(("Didn't find the user entry! iUser=%#x iUserTable=%#x GCPhys=%RGp\n",
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync * Gets the entry size of a shadow table.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param enmKind The kind of page.
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync * @returns The size of the entry in bytes. That is, 4 or 8.
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync * @returns If the kind is not for a table, an assertion is raised and 0 is
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * returned.
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsyncDECLINLINE(unsigned) pgmPoolTrackGetShadowEntrySize(PGMPOOLKIND enmKind)
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync * Gets the entry size of a guest table.
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync * @param enmKind The kind of page.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @returns The size of the entry in bytes. That is, 0, 4 or 8.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @returns If the kind is not for a table, an assertion is raised and 0 is
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync * returned.
8bc8d66f188d5357155b8340e2d489573be2b607vboxsyncDECLINLINE(unsigned) pgmPoolTrackGetGuestEntrySize(PGMPOOLKIND enmKind)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /** @todo can we return 0? (nobody is calling this...) */
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * Scans one shadow page table for mappings of a physical page.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * @param pVM The VM handle.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pPhysPage The guest page in question.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param iShw The shadow page table.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param cRefs The number of references made in that PT.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsyncstatic void pgmPoolTrackFlushGCPhysPTInt(PVM pVM, PCPGMPAGE pPhysPage, uint16_t iShw, uint16_t cRefs)
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync LogFlow(("pgmPoolTrackFlushGCPhysPT: HCPhys=%RHp iShw=%d cRefs=%d\n", pPhysPage->HCPhys, iShw, cRefs));
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * Assert sanity.
b8bb9c9f6b8ebfd0a7d6df0c0289f9fe80241750vboxsync AssertFatalMsg(iShw < pPool->cCurPages && iShw != NIL_PGMPOOL_IDX, ("iShw=%d\n", iShw));
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * Then, clear the actual mappings to the page in the shadow PT.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync const uint32_t u32 = PGM_PAGE_GET_HCPHYS(pPhysPage) | X86_PTE_P;
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync PX86PT pPT = (PX86PT)PGMPOOL_PAGE_2_PTR(pVM, pPage);
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync for (unsigned i = pPage->iFirstPresent; i < RT_ELEMENTS(pPT->a); i++)
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync if ((pPT->a[i].u & (X86_PTE_PG_MASK | X86_PTE_P)) == u32)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Log4(("pgmPoolTrackFlushGCPhysPTs: i=%d pte=%RX32 cRefs=%#x\n", i, pPT->a[i], cRefs));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pPT->a[i].u = 0;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync#if defined(DEBUG) && !defined(IN_RING0) ///@todo RTLogPrintf is missing in R0.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync RTLogPrintf("cRefs=%d iFirstPresent=%d cPresent=%d\n", cRefs, pPage->iFirstPresent, pPage->cPresent);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync if ((pPT->a[i].u & (X86_PTE_PG_MASK | X86_PTE_P)) == u32)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pPT->a[i].u = 0;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync AssertFatalMsgFailed(("cRefs=%d iFirstPresent=%d cPresent=%d\n", cRefs, pPage->iFirstPresent, pPage->cPresent));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync const uint64_t u64 = PGM_PAGE_GET_HCPHYS(pPhysPage) | X86_PTE_P;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PX86PTPAE pPT = (PX86PTPAE)PGMPOOL_PAGE_2_PTR(pVM, pPage);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync for (unsigned i = pPage->iFirstPresent; i < RT_ELEMENTS(pPT->a); i++)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync if ((pPT->a[i].u & (X86_PTE_PAE_PG_MASK | X86_PTE_P)) == u64)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Log4(("pgmPoolTrackFlushGCPhysPTs: i=%d pte=%RX64 cRefs=%#x\n", i, pPT->a[i], cRefs));
06ea6bcf23874b662d499b3f130024c98b2dd7a6vboxsync pPT->a[i].u = 0;
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync#if defined(DEBUG) && !defined(IN_RING0) ///@todo RTLogPrintf is missing in R0.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync RTLogPrintf("cRefs=%d iFirstPresent=%d cPresent=%d\n", cRefs, pPage->iFirstPresent, pPage->cPresent);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync if ((pPT->a[i].u & (X86_PTE_PAE_PG_MASK | X86_PTE_P)) == u64)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pPT->a[i].u = 0;
06ea6bcf23874b662d499b3f130024c98b2dd7a6vboxsync AssertFatalMsgFailed(("cRefs=%d iFirstPresent=%d cPresent=%d\n", cRefs, pPage->iFirstPresent, pPage->cPresent));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync const uint64_t u64 = PGM_PAGE_GET_HCPHYS(pPhysPage) | X86_PTE_P;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PEPTPT pPT = (PEPTPT)PGMPOOL_PAGE_2_PTR(pVM, pPage);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync for (unsigned i = pPage->iFirstPresent; i < RT_ELEMENTS(pPT->a); i++)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync if ((pPT->a[i].u & (EPT_PTE_PG_MASK | X86_PTE_P)) == u64)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Log4(("pgmPoolTrackFlushGCPhysPTs: i=%d pte=%RX64 cRefs=%#x\n", i, pPT->a[i], cRefs));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pPT->a[i].u = 0;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync#if defined(DEBUG) && !defined(IN_RING0) ///@todo RTLogPrintf is missing in R0.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync RTLogPrintf("cRefs=%d iFirstPresent=%d cPresent=%d\n", cRefs, pPage->iFirstPresent, pPage->cPresent);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync if ((pPT->a[i].u & (EPT_PTE_PG_MASK | X86_PTE_P)) == u64)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pPT->a[i].u = 0;
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync AssertFatalMsgFailed(("cRefs=%d iFirstPresent=%d cPresent=%d\n", cRefs, pPage->iFirstPresent, pPage->cPresent));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync AssertFatalMsgFailed(("enmKind=%d iShw=%d\n", pPage->enmKind, iShw));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * Scans one shadow page table for mappings of a physical page.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * @param pVM The VM handle.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * @param pPhysPage The guest page in question.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * @param iShw The shadow page table.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * @param cRefs The number of references made in that PT.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsyncvoid pgmPoolTrackFlushGCPhysPT(PVM pVM, PPGMPAGE pPhysPage, uint16_t iShw, uint16_t cRefs)
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync PPGMPOOL pPool = pVM->pgm.s.CTX_SUFF(pPool); NOREF(pPool);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync LogFlow(("pgmPoolTrackFlushGCPhysPT: HCPhys=%RHp iShw=%d cRefs=%d\n", pPhysPage->HCPhys, iShw, cRefs));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync STAM_PROFILE_START(&pPool->StatTrackFlushGCPhysPT, f);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync pgmPoolTrackFlushGCPhysPTInt(pVM, pPhysPage, iShw, cRefs);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync pPhysPage->HCPhys &= MM_RAM_FLAGS_NO_REFS_MASK; /** @todo PAGE FLAGS */
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync STAM_PROFILE_STOP(&pPool->StatTrackFlushGCPhysPT, f);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Flushes a list of shadow page tables mapping the same physical page.
06ea6bcf23874b662d499b3f130024c98b2dd7a6vboxsync * @param pVM The VM handle.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPhysPage The guest page in question.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param iPhysExt The physical cross reference extent list to flush.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsyncvoid pgmPoolTrackFlushGCPhysPTs(PVM pVM, PPGMPAGE pPhysPage, uint16_t iPhysExt)
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync STAM_PROFILE_START(&pPool->StatTrackFlushGCPhysPTs, f);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync LogFlow(("pgmPoolTrackFlushGCPhysPTs: HCPhys=%RHp iPhysExt\n", pPhysPage->HCPhys, iPhysExt));
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync for (unsigned i = 0; i < RT_ELEMENTS(pPhysExt->aidx); i++)
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync pgmPoolTrackFlushGCPhysPTInt(pVM, pPhysPage, pPhysExt->aidx[i], 1);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* insert the list into the free list and clear the ram range entry. */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pPhysPage->HCPhys &= MM_RAM_FLAGS_NO_REFS_MASK; /** @todo PAGE FLAGS */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync STAM_PROFILE_STOP(&pPool->StatTrackFlushGCPhysPTs, f);
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync#endif /* PGMPOOL_WITH_GCPHYS_TRACKING */
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync * Scans all shadow page tables for mappings of a physical page.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * This may be slow, but it's most likely more efficient than cleaning
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * out the entire page pool / cache.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @returns VBox status code.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @retval VINF_SUCCESS if all references has been successfully cleared.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @retval VINF_PGM_GCPHYS_ALIASED if we're better off with a CR3 sync and
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * a page pool cleaning.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * @param pVM The VM handle.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPhysPage The guest page in question.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsyncint pgmPoolTrackFlushGCPhysPTsSlow(PVM pVM, PPGMPAGE pPhysPage)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync STAM_PROFILE_START(&pPool->StatTrackFlushGCPhysPTsSlow, s);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolTrackFlushGCPhysPTsSlow: cUsedPages=%d cPresent=%d HCPhys=%RHp\n",
9523921c89c66f4bececdbd5ac95aed0039eda1bvboxsync pPool->cUsedPages, pPool->cPresent, pPhysPage->HCPhys));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * There is a limit to what makes sense.
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync LogFlow(("pgmPoolTrackFlushGCPhysPTsSlow: giving up... (cPresent=%d)\n", pPool->cPresent));
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync STAM_PROFILE_STOP(&pPool->StatTrackFlushGCPhysPTsSlow, s);
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Iterate all the pages until we've encountered all that in use.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * This is simple but not quite optimal solution.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync const uint64_t u64 = PGM_PAGE_GET_HCPHYS(pPhysPage) | X86_PTE_P;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * We only care about shadow page tables.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync PX86PT pPT = (PX86PT)PGMPOOL_PAGE_2_PTR(pVM, pPage);
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync for (unsigned i = pPage->iFirstPresent; i < RT_ELEMENTS(pPT->a); i++)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync if ((pPT->a[i].u & (X86_PTE_PG_MASK | X86_PTE_P)) == u32)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync //Log4(("pgmPoolTrackFlushGCPhysPTsSlow: idx=%d i=%d pte=%RX32\n", iPage, i, pPT->a[i]));
36f3c24e4ad9c6b813767db1faeabbe7e2ecc057vboxsync pPT->a[i].u = 0;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync PX86PTPAE pPT = (PX86PTPAE)PGMPOOL_PAGE_2_PTR(pVM, pPage);
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync for (unsigned i = pPage->iFirstPresent; i < RT_ELEMENTS(pPT->a); i++)
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync if ((pPT->a[i].u & (X86_PTE_PAE_PG_MASK | X86_PTE_P)) == u64)
b8bb9c9f6b8ebfd0a7d6df0c0289f9fe80241750vboxsync //Log4(("pgmPoolTrackFlushGCPhysPTsSlow: idx=%d i=%d pte=%RX64\n", iPage, i, pPT->a[i]));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync pPT->a[i].u = 0;
9e17ca2e9d797e845e3284141dd4086a4b817ae5vboxsync pPhysPage->HCPhys &= MM_RAM_FLAGS_NO_REFS_MASK; /** @todo PAGE FLAGS */
06ea6bcf23874b662d499b3f130024c98b2dd7a6vboxsync STAM_PROFILE_STOP(&pPool->StatTrackFlushGCPhysPTsSlow, s);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * Clears the user entry in a user table.
5366e994777f9d9391cf809dc77610f57270d75dvboxsync * This is used to remove all references to a page when flushing it.
5366e994777f9d9391cf809dc77610f57270d75dvboxsyncstatic void pgmPoolTrackClearPageUser(PPGMPOOL pPool, PPGMPOOLPAGE pPage, PCPGMPOOLUSER pUser)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Map the user page.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PPGMPOOLPAGE pUserPage = &pPool->aPages[pUser->iUser];
5366e994777f9d9391cf809dc77610f57270d75dvboxsync u.pau64 = (uint64_t *)PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pUserPage);
5366e994777f9d9391cf809dc77610f57270d75dvboxsync /* Safety precaution in case we change the paging for other modes too in the future. */
5366e994777f9d9391cf809dc77610f57270d75dvboxsync Assert(PGMGetHyperCR3(pPool->CTX_SUFF(pVM)) != pPage->Core.Key);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Some sanity checks.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync Assert(!(u.pau32[pUser->iUserTable] & PGM_PDFLAGS_MAPPING));
5366e994777f9d9391cf809dc77610f57270d75dvboxsync Assert(pUser->iUserTable < 2048 && pUser->iUser == PGMPOOL_IDX_PAE_PD);
5366e994777f9d9391cf809dc77610f57270d75dvboxsync Assert(!(u.pau64[pUser->iUserTable] & PGM_PDFLAGS_MAPPING));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync Assert(!(u.pau64[pUser->iUserTable] & PGM_PLXFLAGS_PERMANENT));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync Assert(!(u.pau64[pUser->iUserTable] & PGM_PDFLAGS_MAPPING));
5366e994777f9d9391cf809dc77610f57270d75dvboxsync Assert(!(u.pau64[pUser->iUserTable] & PGM_PLXFLAGS_PERMANENT));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Assert(!(u.pau64[pUser->iUserTable] & PGM_PLXFLAGS_PERMANENT));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* GCPhys >> PAGE_SHIFT is the index here */
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync AssertMsgFailed(("enmKind=%d\n", pUserPage->enmKind));
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync#endif /* VBOX_STRICT */
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Clear the entry in the user page.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync /* 32-bit entries */
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync /* 64-bit entries */
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync AssertFatalMsgFailed(("enmKind=%d iUser=%#x iUserTable=%#x\n", pUserPage->enmKind, pUser->iUser, pUser->iUserTable));
9523921c89c66f4bececdbd5ac95aed0039eda1bvboxsync * Clears all users of a page.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsyncstatic void pgmPoolTrackClearPageUsers(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * Free all the user records.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* Clear enter in user table. */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pgmPoolTrackClearPageUser(pPool, pPage, &paUsers[i]);
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync /* Free it. */
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync /* Next. */
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync * Allocates a new physical cross reference extent.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @returns Pointer to the allocated extent on success. NULL if we're out of them.
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync * @param pVM The VM handle.
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync * @param piPhysExt Where to store the phys ext index.
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsyncPPGMPOOLPHYSEXT pgmPoolTrackPhysExtAlloc(PVM pVM, uint16_t *piPhysExt)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync STAM_COUNTER_INC(&pPool->StamTrackPhysExtAllocFailures);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PPGMPOOLPHYSEXT pPhysExt = &pPool->CTX_SUFF(paPhysExts)[iPhysExt];
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Frees a physical cross reference extent.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pVM The VM handle.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param iPhysExt The extent to free.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsyncvoid pgmPoolTrackPhysExtFree(PVM pVM, uint16_t iPhysExt)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PPGMPOOLPHYSEXT pPhysExt = &pPool->CTX_SUFF(paPhysExts)[iPhysExt];
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync for (unsigned i = 0; i < RT_ELEMENTS(pPhysExt->aidx); i++)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Frees a physical cross reference extent.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pVM The VM handle.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * @param iPhysExt The extent to free.
cd5df721f068659172f3bf95de8fedeb465f057dvboxsyncvoid pgmPoolTrackPhysExtFreeList(PVM pVM, uint16_t iPhysExt)
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync for (unsigned i = 0; i < RT_ELEMENTS(pPhysExt->aidx); i++)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Insert a reference into a list of physical cross reference extents.
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * @returns The new ram range flags (top 16-bits).
c55bf74b54ecdfb5ebc4e5d90b620d0fee31737evboxsync * @param pVM The VM handle.
5366e994777f9d9391cf809dc77610f57270d75dvboxsync * @param iPhysExt The physical extent index of the list head.
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync * @param iShwPT The shadow page table index.
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsyncstatic uint16_t pgmPoolTrackPhysExtInsert(PVM pVM, uint16_t iPhysExt, uint16_t iShwPT)
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync PPGMPOOLPHYSEXT paPhysExts = pPool->CTX_SUFF(paPhysExts);
45655563f818c5d5bbf4b3d14aa48cbd92a871f1vboxsync /* special common case. */
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync if (paPhysExts[iPhysExt].aidx[2] == NIL_PGMPOOL_IDX)
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync STAM_COUNTER_INC(&pVM->pgm.s.StatTrackAliasedMany);
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync LogFlow(("pgmPoolTrackPhysExtAddref: %d:{,,%d}\n", iPhysExt, iShwPT));
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync return iPhysExt | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync /* general treatment. */
cc1ef2ef9bbc6a0ff964928d61b7298e5bfcce5fvboxsync for (unsigned i = 0; i < RT_ELEMENTS(paPhysExts[iPhysExt].aidx); i++)
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync if (paPhysExts[iPhysExt].aidx[i] == NIL_PGMPOOL_IDX)
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync STAM_COUNTER_INC(&pVM->pgm.s.StatTrackAliasedMany);
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync LogFlow(("pgmPoolTrackPhysExtAddref: %d:{%d} i=%d cMax=%d\n", iPhysExt, iShwPT, i, cMax));
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync return iPhysExtStart | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync LogFlow(("pgmPoolTrackPhysExtAddref: overflow (1) iShwPT=%d\n", iShwPT));
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync return MM_RAM_FLAGS_IDX_OVERFLOWED | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync /* add another extent to the list. */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync PPGMPOOLPHYSEXT pNew = pgmPoolTrackPhysExtAlloc(pVM, &iPhysExt);
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync return MM_RAM_FLAGS_IDX_OVERFLOWED | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync LogFlow(("pgmPoolTrackPhysExtAddref: added new extent %d:{%d}->%d\n", iPhysExt, iShwPT, iPhysExtStart));
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync return iPhysExt | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync * Add a reference to guest physical page where extents are in use.
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync * @returns The new ram range flags (top 16-bits).
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync * @param pVM The VM handle.
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync * @param u16 The ram range flags (top 16-bits).
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * @param iShwPT The shadow page table index.
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsyncuint16_t pgmPoolTrackPhysExtAddref(PVM pVM, uint16_t u16, uint16_t iShwPT)
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync if ((u16 >> (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT)) != MM_RAM_FLAGS_CREFS_PHYSEXT)
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * Convert to extent list.
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync Assert((u16 >> (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT)) == 1);
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync PPGMPOOLPHYSEXT pPhysExt = pgmPoolTrackPhysExtAlloc(pVM, &iPhysExt);
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync LogFlow(("pgmPoolTrackPhysExtAddref: new extent: %d:{%d, %d}\n", iPhysExt, u16 & MM_RAM_FLAGS_IDX_MASK, iShwPT));
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync u16 = iPhysExt | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync u16 = MM_RAM_FLAGS_IDX_OVERFLOWED | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync else if (u16 != (MM_RAM_FLAGS_IDX_OVERFLOWED | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT))))
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Insert into the extent list.
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync u16 = pgmPoolTrackPhysExtInsert(pVM, u16 & MM_RAM_FLAGS_IDX_MASK, iShwPT);
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync STAM_COUNTER_INC(&pVM->pgm.s.StatTrackAliasedLots);
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync * Clear references to guest physical memory.
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync * @param pPool The pool.
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync * @param pPage The page.
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync * @param pPhysPage Pointer to the aPages entry in the ram range.
687794577e2e35c3cae67e692a7f2130d1262a82vboxsyncvoid pgmPoolTrackPhysExtDerefGCPhys(PPGMPOOL pPool, PPGMPOOLPAGE pPage, PPGMPAGE pPhysPage)
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync const unsigned cRefs = pPhysPage->HCPhys >> MM_RAM_FLAGS_CREFS_SHIFT; /** @todo PAGE FLAGS */
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync AssertFatalMsg(cRefs == MM_RAM_FLAGS_CREFS_PHYSEXT, ("cRefs=%d HCPhys=%RHp pPage=%p:{.idx=%d}\n", cRefs, pPhysPage->HCPhys, pPage, pPage->idx));
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync uint16_t iPhysExt = (pPhysPage->HCPhys >> MM_RAM_FLAGS_IDX_SHIFT) & MM_RAM_FLAGS_IDX_MASK;
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync PPGMPOOLPHYSEXT paPhysExts = pPool->CTX_SUFF(paPhysExts);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Look for the shadow page and check if it's all freed.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync for (unsigned i = 0; i < RT_ELEMENTS(paPhysExts[iPhysExt].aidx); i++)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync for (i = 0; i < RT_ELEMENTS(paPhysExts[iPhysExt].aidx); i++)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync if (paPhysExts[iPhysExt].aidx[i] != NIL_PGMPOOL_IDX)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolTrackPhysExtDerefGCPhys: HCPhys=%RX64 idx=%d\n", pPhysPage->HCPhys, pPage->idx));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* we can free the node. */
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync const uint16_t iPhysExtNext = paPhysExts[iPhysExt].iNext;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* lonely node */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolTrackPhysExtDerefGCPhys: HCPhys=%RX64 idx=%d lonely\n", pPhysPage->HCPhys, pPage->idx));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pPhysPage->HCPhys &= MM_RAM_FLAGS_NO_REFS_MASK; /** @todo PAGE FLAGS */
06ea6bcf23874b662d499b3f130024c98b2dd7a6vboxsync else if (iPhysExtPrev == NIL_PGMPOOL_PHYSEXT_INDEX)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolTrackPhysExtDerefGCPhys: HCPhys=%RX64 idx=%d head\n", pPhysPage->HCPhys, pPage->idx));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pPhysPage->HCPhys = (pPhysPage->HCPhys & MM_RAM_FLAGS_NO_REFS_MASK) /** @todo PAGE FLAGS */
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync | ((uint64_t)MM_RAM_FLAGS_CREFS_PHYSEXT << MM_RAM_FLAGS_CREFS_SHIFT)
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync | ((uint64_t)iPhysExtNext << MM_RAM_FLAGS_IDX_SHIFT);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* in list */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolTrackPhysExtDerefGCPhys: HCPhys=%RX64 idx=%d\n", pPhysPage->HCPhys, pPage->idx));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync AssertFatalMsgFailed(("not-found! cRefs=%d HCPhys=%RHp pPage=%p:{.idx=%d}\n", cRefs, pPhysPage->HCPhys, pPage, pPage->idx));
9523921c89c66f4bececdbd5ac95aed0039eda1bvboxsync else /* nothing to do */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolTrackPhysExtDerefGCPhys: HCPhys=%RX64\n", pPhysPage->HCPhys));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * Clear references to guest physical memory.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * This is the same as pgmPoolTracDerefGCPhys except that the guest physical address
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * is assumed to be correct, so the linear search can be skipped and we can assert
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * at an earlier point.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPool The pool.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPage The page.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param HCPhys The host physical address corresponding to the guest page.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param GCPhys The guest physical address corresponding to HCPhys.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsyncstatic void pgmPoolTracDerefGCPhys(PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTHCPHYS HCPhys, RTGCPHYS GCPhys)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Walk range list.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync PPGMRAMRANGE pRam = pPool->CTX_SUFF(pVM)->pgm.s.CTX_SUFF(pRamRanges);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync /* does it match? */
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsyncRTHCPHYS HCPhysPage = PGM_PAGE_GET_HCPHYS(&pRam->aPages[iPage]);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsyncLog(("pgmPoolTracDerefGCPhys %RHp vs %RHp\n", HCPhysPage, HCPhys));
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync if (PGM_PAGE_GET_HCPHYS(&pRam->aPages[iPage]) == HCPhys)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pgmTrackDerefGCPhys(pPool, pPage, &pRam->aPages[iPage]);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync AssertFatalMsgFailed(("HCPhys=%RHp GCPhys=%RGp\n", HCPhys, GCPhys));
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Clear references to guest physical memory.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPool The pool.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPage The page.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param HCPhys The host physical address corresponding to the guest page.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param GCPhysHint The guest physical address which may corresponding to HCPhys.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsyncstatic void pgmPoolTracDerefGCPhysHint(PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTHCPHYS HCPhys, RTGCPHYS GCPhysHint)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Walk range list.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync PPGMRAMRANGE pRam = pPool->CTX_SUFF(pVM)->pgm.s.CTX_SUFF(pRamRanges);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync /* does it match? */
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync if (PGM_PAGE_GET_HCPHYS(&pRam->aPages[iPage]) == HCPhys)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pgmTrackDerefGCPhys(pPool, pPage, &pRam->aPages[iPage]);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Damn, the hint didn't work. We'll have to do an expensive linear search.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync STAM_COUNTER_INC(&pPool->StatTrackLinearRamSearches);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pRam = pPool->CTX_SUFF(pVM)->pgm.s.CTX_SUFF(pRamRanges);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync while (iPage-- > 0)
DECLINLINE(void) pgmPoolTrackDerefPT32Bit32Bit(PPGMPOOL pPool, PPGMPOOLPAGE pPage, PX86PT pShwPT, PCX86PT pGstPT)
pgmPoolTracDerefGCPhysHint(pPool, pPage, pShwPT->a[i].u & X86_PTE_PG_MASK, pGstPT->a[i].u & X86_PTE_PG_MASK);
DECLINLINE(void) pgmPoolTrackDerefPTPae32Bit(PPGMPOOL pPool, PPGMPOOLPAGE pPage, PX86PTPAE pShwPT, PCX86PT pGstPT)
pgmPoolTracDerefGCPhysHint(pPool, pPage, pShwPT->a[i].u & X86_PTE_PAE_PG_MASK, pGstPT->a[i].u & X86_PTE_PG_MASK);
DECLINLINE(void) pgmPoolTrackDerefPTPaePae(PPGMPOOL pPool, PPGMPOOLPAGE pPage, PX86PTPAE pShwPT, PCX86PTPAE pGstPT)
pgmPoolTracDerefGCPhysHint(pPool, pPage, pShwPT->a[i].u & X86_PTE_PAE_PG_MASK, pGstPT->a[i].u & X86_PTE_PAE_PG_MASK);
PPGMPOOLPAGE pSubPage = (PPGMPOOLPAGE)RTAvloHCPhysGet(&pPool->HCPhysTree, pShwPD->a[i].u & X86_PDE_PAE_PG_MASK);
if (pSubPage)
PPGMPOOLPAGE pSubPage = (PPGMPOOLPAGE)RTAvloHCPhysGet(&pPool->HCPhysTree, pShwPDPT->a[i].u & X86_PDPE_PG_MASK);
if (pSubPage)
PPGMPOOLPAGE pSubPage = (PPGMPOOLPAGE)RTAvloHCPhysGet(&pPool->HCPhysTree, pShwPML4->a[i].u & X86_PDPE_PG_MASK);
if (pSubPage)
PPGMPOOLPAGE pSubPage = (PPGMPOOLPAGE)RTAvloHCPhysGet(&pPool->HCPhysTree, pShwPD->a[i].u & EPT_PDE_PG_MASK);
if (pSubPage)
PPGMPOOLPAGE pSubPage = (PPGMPOOLPAGE)RTAvloHCPhysGet(&pPool->HCPhysTree, pShwPDPT->a[i].u & EPT_PDPTE_PG_MASK);
if (pSubPage)
#ifdef PGMPOOL_WITH_GCPHYS_TRACKING
void *pvGst;
void *pvGst;
void *pvGst;
case PGMPOOLKIND_ROOT_PAE_PD:
case PGMPOOLKIND_ROOT_PDPT:
case PGMPOOLKIND_ROOT_NESTED:
#ifdef IN_RING3
#ifdef PGMPOOL_WITH_MONITORING
#ifdef PGMPOOL_WITH_USER_TRACKING
#ifdef PGMPOOL_WITH_CACHE
#ifdef PGMPOOL_WITH_USER_TRACKING
for (unsigned i = 0; i < cMaxUsers; i++)
#ifdef PGMPOOL_WITH_GCPHYS_TRACKING
pRam;
while (iPage-- > 0)
for (unsigned i = 0; i < cMaxPhysExts; i++)
#ifdef PGMPOOL_WITH_MONITORING
#ifdef PGMPOOL_WITH_CACHE
#ifdef PGMPOOL_WITH_MONITORING
# ifdef PGMPOOL_WITH_CACHE
#ifdef PGMPOOL_WITH_USER_TRACKING
#ifdef PGMPOOL_WITH_CACHE
* @retval VERR_PGM_POOL_CLEARED if the deregistration of the physical handler will cause a light weight pool flush.
Log(("pgmPoolFlushPage: special root page, rejected. enmKind=%d idx=%d\n", pPage->enmKind, pPage->idx));
return VINF_SUCCESS;
("Can't free the shadow CR3! (%RHp vs %RHp kind=%d\n", PGMGetHyperCR3(pPool->CTX_SUFF(pVM)), pPage->Core.Key, pPage->enmKind));
Log(("pgmPoolFlushPage: current active shadow CR3, rejected. enmKind=%d idx=%d\n", pPage->enmKind, pPage->idx));
return VINF_SUCCESS;
#ifdef PGMPOOL_WITH_USER_TRACKING
#ifdef PGMPOOL_WITH_CACHE
#ifdef PGMPOOL_WITH_MONITORING
return rc;
#ifdef PGMPOOL_WITH_USER_TRACKING
#ifdef PGMPOOL_WITH_CACHE
#ifdef IN_RING3
return rc;
return VINF_SUCCESS;
#ifdef PGMPOOL_WITH_CACHE
return VERR_PGM_POOL_FLUSHED;
int pgmPoolAlloc(PVM pVM, RTGCPHYS GCPhys, PGMPOOLKIND enmKind, uint16_t iUser, uint32_t iUserTable, PPPGMPOOLPAGE ppPage)
LogFlow(("pgmPoolAlloc: GCPhys=%RGp enmKind=%d iUser=%#x iUserTable=%#x\n", GCPhys, enmKind, iUser, iUserTable));
#ifdef PGMPOOL_WITH_CACHE
LogFlow(("pgmPoolAlloc: cached returns %Rrc *ppPage=%p:{.Key=%RHp, .idx=%d}\n", rc2, *ppPage, (*ppPage)->Core.Key, (*ppPage)->idx));
return rc2;
return rc;
pPool->cUsedPages++; /* physical handler registration / pgmPoolTrackFlushGCPhysPTsSlow requirement. */
#ifdef PGMPOOL_WITH_MONITORING
#ifdef PGMPOOL_WITH_USER_TRACKING
return rc3;
#ifdef VBOX_WITH_STATISTICS
LogFlow(("pgmPoolAlloc: returns %Rrc *ppPage=%p:{.Key=%RHp, .idx=%d, .fCached=%RTbool, .fMonitored=%RTbool}\n",
return rc;
return pPage;