PGMAllPool.cpp revision 0aeb5f500280603a51364071670efeb2617f771c
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * PGM Shadow Page Pool.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Copyright (C) 2006-2007 Sun Microsystems, Inc.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * This file is part of VirtualBox Open Source Edition (OSE), as
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * available from http://www.virtualbox.org. This file is free software;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * you can redistribute it and/or modify it under the terms of the GNU
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * General Public License (GPL) as published by the Free Software
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Foundation, in version 2 as it comes in the "COPYING" file of the
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Clara, CA 95054 USA or visit http://www.sun.com if you need
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * additional information or have any questions.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync/*******************************************************************************
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync* Header Files *
7b9f0c34e9ea328981c99e97054bdf8684d9d620vboxsync*******************************************************************************/
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync/*******************************************************************************
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync* Internal Functions *
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync*******************************************************************************/
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncDECLINLINE(unsigned) pgmPoolTrackGetShadowEntrySize(PGMPOOLKIND enmKind);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncDECLINLINE(unsigned) pgmPoolTrackGetGuestEntrySize(PGMPOOLKIND enmKind);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic void pgmPoolTrackDeref(PPGMPOOL pPool, PPGMPOOLPAGE pPage);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic void pgmPoolTracDerefGCPhysHint(PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTHCPHYS HCPhys, RTGCPHYS GCPhysHint);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic int pgmPoolTrackAddUser(PPGMPOOL pPool, PPGMPOOLPAGE pPage, uint16_t iUser, uint32_t iUserTable);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic void pgmPoolMonitorModifiedRemove(PPGMPOOL pPool, PPGMPOOLPAGE pPage);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncDECLEXPORT(int) pgmPoolAccessHandler(PVM pVM, RTGCUINT uErrorCode, PCPUMCTXCORE pRegFrame, RTGCPTR pvFault, RTGCPHYS GCPhysFault, void *pvUser);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Checks if the specified page pool kind is for a 4MB or 2MB guest page.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @returns true if it's the shadow of a 4MB or 2MB guest page, otherwise false.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param enmKind The page kind.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncDECLINLINE(bool) pgmPoolIsBigPage(PGMPOOLKIND enmKind)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync return true;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync return false;
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync#if defined(IN_RC) || defined(VBOX_WITH_2X_4GB_ADDR_SPACE_IN_R0)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Maps a pool page into the current context.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @returns Pointer to the mapping.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPGM Pointer to the PGM instance data.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPage The page to map.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsyncvoid *pgmPoolMapPageFallback(PPGM pPGM, PPGMPOOLPAGE pPage)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* general pages are take care of by the inlined part, it
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync only ends up here in case of failure. */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync AssertReleaseReturn(pPage->idx < PGMPOOL_IDX_FIRST, NULL);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync/** @todo make sure HCPhys is valid for *all* indexes. */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* special pages. */
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync AssertReleaseMsgFailed(("Invalid index %d\n", pPage->idx));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync# else /* VBOX_WITH_2X_4GB_ADDR_SPACE_IN_R0 */
9523921c89c66f4bececdbd5ac95aed0039eda1bvboxsync AssertReleaseMsgFailed(("PGMPOOL_IDX_PAE_PD is not usable in VBOX_WITH_2X_4GB_ADDR_SPACE_IN_R0 context\n"));
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync AssertReleaseMsgFailed(("Invalid index %d\n", pPage->idx));
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync AssertMsg(HCPhys && HCPhys != NIL_RTHCPHYS && !(PAGE_OFFSET_MASK & HCPhys), ("%RHp\n", HCPhys));
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync# endif /* VBOX_WITH_2X_4GB_ADDR_SPACE_IN_R0 */
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync#endif /* IN_RC || VBOX_WITH_2X_4GB_ADDR_SPACE_IN_R0 */
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * Determin the size of a write instruction.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * @returns number of bytes written.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * @param pDis The disassembler state.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsyncstatic unsigned pgmPoolDisasWriteSize(PDISCPUSTATE pDis)
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * This is very crude and possibly wrong for some opcodes,
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * but since it's not really supposed to be called we can
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * probably live with that.
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync * Flushes a chain of pages sharing the same access monitor.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * @returns VBox status code suitable for scheduling.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * @param pPool The pool.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * @param pPage A page in the chain.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsyncint pgmPoolMonitorChainFlush(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
9523921c89c66f4bececdbd5ac95aed0039eda1bvboxsync LogFlow(("pgmPoolMonitorChainFlush: Flush page %RGp type=%d\n", pPage->GCPhys, pPage->enmKind));
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * Find the list head.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * Iterate the list flushing each shadow page.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync if (rc2 == VERR_PGM_POOL_CLEARED && rc == VINF_SUCCESS)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Wrapper for getting the current context pointer to the entry being modified.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @returns Pointer to the current context mapping of the entry.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPool The pool.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pvFault The fault virtual address.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param GCPhysFault The fault physical address.
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync * @param cbEntry The entry size.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsyncDECLINLINE(const void *) pgmPoolMonitorGCPtr2CCPtr(PPGMPOOL pPool, RTHCPTR pvFault, RTGCPHYS GCPhysFault, const unsigned cbEntry)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsyncDECLINLINE(const void *) pgmPoolMonitorGCPtr2CCPtr(PPGMPOOL pPool, RTGCPTR pvFault, RTGCPHYS GCPhysFault, const unsigned cbEntry)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync return (const void *)((RTGCUINTPTR)pvFault & ~(RTGCUINTPTR)(cbEntry - 1));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync int rc = PGMDynMapGCPageOff(pPool->pVMR0, GCPhysFault & ~(RTGCPHYS)(cbEntry - 1), &pvRet);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync int rc = pgmRamGCPhys2HCPtr(&pPool->pVMR0->pgm.s, GCPhysFault & ~(RTGCPHYS)(cbEntry - 1), &pvRet);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync return (RTHCPTR)((uintptr_t)pvFault & ~(RTHCUINTPTR)(cbEntry - 1));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Process shadow entries before they are changed by the guest.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * For PT entries we will clear them. For PD entries, we'll simply check
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * for mapping conflicts and set the SyncCR3 FF if found.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPool The pool.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPage The head page.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param GCPhysFault The guest physical fault address.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param uAddress In R0 and GC this is the guest context fault address (flat).
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * In R3 this is the host context 'fault' address.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pCpu The disassembler state for figuring out the write size.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * This need not be specified if the caller knows we won't do cross entry accesses.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncvoid pgmPoolMonitorChainChanging(PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTGCPHYS GCPhysFault, RTHCPTR pvAddress, PDISCPUSTATE pCpu)
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncvoid pgmPoolMonitorChainChanging(PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTGCPHYS GCPhysFault, RTGCPTR pvAddress, PDISCPUSTATE pCpu)
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync const unsigned off = GCPhysFault & PAGE_OFFSET_MASK;
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync const unsigned cbWrite = (pCpu) ? pgmPoolDisasWriteSize(pCpu) : 0;
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync LogFlow(("pgmPoolMonitorChainChanging: %RGv phys=%RGp kind=%d cbWrite=%d\n", pvAddress, GCPhysFault, pPage->enmKind, cbWrite));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync uShw.pv = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PCX86PTE pGstPte = (PCX86PTE)pgmPoolMonitorGCPtr2CCPtr(pPool, pvAddress, GCPhysFault, sizeof(*pGstPte));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Log4(("pgmPoolMonitorChainChanging 32_32: deref %016RX64 GCPhys %08RX32\n", uShw.pPT->a[iShw].u & X86_PTE_PAE_PG_MASK, pGstPte->u & X86_PTE_PG_MASK));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* page/2 sized */
7e8ef90d3160234df0f254131b87af4243d79476vboxsync uShw.pv = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage);
7e8ef90d3160234df0f254131b87af4243d79476vboxsync const unsigned iShw = (off / sizeof(X86PTE)) & (X86_PG_PAE_ENTRIES - 1);
7e8ef90d3160234df0f254131b87af4243d79476vboxsync PCX86PTE pGstPte = (PCX86PTE)pgmPoolMonitorGCPtr2CCPtr(pPool, pvAddress, GCPhysFault, sizeof(*pGstPte));
7e8ef90d3160234df0f254131b87af4243d79476vboxsync Log4(("pgmPoolMonitorChainChanging pae_32: deref %016RX64 GCPhys %08RX32\n", uShw.pPT->a[iShw].u & X86_PTE_PAE_PG_MASK, pGstPte->u & X86_PTE_PG_MASK));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync uShw.pv = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolMonitorChainChanging PAE for 32 bits: iGst=%x idx = %d page idx=%d\n", iGst, iShwPdpt, pPage->enmKind - PGMPOOLKIND_PAE_PD0_FOR_32BIT_PD));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync if (iShwPdpt == pPage->enmKind - PGMPOOLKIND_PAE_PD0_FOR_32BIT_PD)
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync for (unsigned i=0;i<2;i++)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync if ((uShw.pPDPae->a[iShw + i].u & (PGM_PDFLAGS_MAPPING | X86_PDE_P)) == (PGM_PDFLAGS_MAPPING | X86_PDE_P))
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShwPdpt=%#x iShw=%#x!\n", iShwPdpt, iShw+i));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolMonitorChainChanging: pae pd iShw=%#x: %RX64 -> freeing it!\n", iShw+i, uShw.pPDPae->a[iShw+i].u));
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync /* paranoia / a bit assumptive. */
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync if ((uShw.pPDPae->a[iShw2].u & (PGM_PDFLAGS_MAPPING | X86_PDE_P)) == (PGM_PDFLAGS_MAPPING | X86_PDE_P))
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShwPdpt=%#x iShw2=%#x!\n", iShwPdpt, iShw2));
1cd59fdf671ca60c64d77e3f7046aaecf7003824vboxsync LogFlow(("pgmPoolMonitorChainChanging: pae pd iShw=%#x: %RX64 -> freeing it!\n", iShw2, uShw.pPDPae->a[iShw2].u));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync uShw.pv = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PCX86PTEPAE pGstPte = (PCX86PTEPAE)pgmPoolMonitorGCPtr2CCPtr(pPool, pvAddress, GCPhysFault, sizeof(*pGstPte));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Log4(("pgmPoolMonitorChainChanging pae: deref %016RX64 GCPhys %016RX64\n", uShw.pPTPae->a[iShw].u & X86_PTE_PAE_PG_MASK, pGstPte->u & X86_PTE_PAE_PG_MASK));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync /* paranoia / a bit assumptive. */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync const unsigned iShw2 = (off + cbWrite - 1) / sizeof(X86PTEPAE);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync AssertReturnVoid(iShw2 < RT_ELEMENTS(uShw.pPTPae->a));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PCX86PTEPAE pGstPte = (PCX86PTEPAE)pgmPoolMonitorGCPtr2CCPtr(pPool, pvAddress, GCPhysFault, sizeof(*pGstPte));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync Log4(("pgmPoolMonitorChainChanging pae: deref %016RX64 GCPhys %016RX64\n", uShw.pPTPae->a[iShw2].u & X86_PTE_PAE_PG_MASK, pGstPte->u & X86_PTE_PAE_PG_MASK));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync uShw.pv = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync const unsigned iShw = off / sizeof(X86PTE); // ASSUMING 32-bit guest paging!
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync STAM_COUNTER_INC(&(pPool->CTX_SUFF(pVM)->pgm.s.StatRZGuestCR3WriteConflict));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw=%#x!\n", iShw));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync LogFlow(("pgmPoolMonitorChainChanging: 32 bit pd iShw=%#x: %RX64 -> freeing it!\n", iShw, uShw.pPD->a[iShw].u));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync /* paranoia / a bit assumptive. */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync const unsigned iShw2 = (off + cbWrite - 1) / sizeof(X86PTE);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync STAM_COUNTER_INC(&(pPool->CTX_SUFF(pVM)->pgm.s.StatRZGuestCR3WriteConflict));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw2=%#x!\n", iShw2));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync LogFlow(("pgmPoolMonitorChainChanging: 32 bit pd iShw=%#x: %RX64 -> freeing it!\n", iShw2, uShw.pPD->a[iShw2].u));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync#if 0 /* useful when running PGMAssertCR3(), a bit too troublesome for general use (TLBs). */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync && !VM_FF_ISSET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3))
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync LogFlow(("pgmPoolMonitorChainChanging: iShw=%#x: %RX32 -> freeing it!\n", iShw, uShw.pPD->a[iShw].u));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync# ifdef IN_RC /* TLB load - we're pushing things a bit... */
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync pgmPoolFree(pPool->CTX_SUFF(pVM), uShw.pPD->a[iShw].u & X86_PDE_PG_MASK, pPage->idx, iShw);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync unsigned iGst = off / sizeof(X86PDE); // ASSUMING 32-bit guest paging!
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync Assert(pPage2->idx == PGMPOOL_IDX_PAE_PD_0 + iShwPdpt);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync uShw.pv = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage2);
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync if ((uShw.pPDPae->a[iShw].u & (PGM_PDFLAGS_MAPPING | X86_PDE_P)) == (PGM_PDFLAGS_MAPPING | X86_PDE_P))
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShwPdpt=%#x iShw=%#x!\n", iShwPdpt, iShw));
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync /* paranoia / a bit assumptive. */
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync if ( iShw2 < RT_ELEMENTS(uShw.pPDPae->a) /** @todo was completely wrong, it's better now after #1865 but still wrong from cross PD. */
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync && (uShw.pPDPae->a[iShw2].u & (PGM_PDFLAGS_MAPPING | X86_PDE_P)) == (PGM_PDFLAGS_MAPPING | X86_PDE_P))
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShwPdpt=%#x iShw2=%#x!\n", iShwPdpt, iShw2));
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync#if 0 /* useful when running PGMAssertCR3(), a bit too troublesome for general use (TLBs). */
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync && !VM_FF_ISSET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3))
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync LogFlow(("pgmPoolMonitorChainChanging: iShwPdpt=%#x iShw=%#x: %RX64 -> freeing it!\n", iShwPdpt, iShw, uShw.pPDPae->a[iShw].u));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync# ifdef IN_RC /* TLB load - we're pushing things a bit... */
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync pgmPoolFree(pPool->CTX_SUFF(pVM), uShw.pPDPae->a[iShw].u & X86_PDE_PAE_PG_MASK, pPage->idx, iShw + iShwPdpt * X86_PG_PAE_ENTRIES);
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync# endif /* !VBOX_WITH_PGMPOOL_PAGING_ONLY */
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync uShw.pv = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage);
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync STAM_COUNTER_INC(&(pPool->CTX_SUFF(pVM)->pgm.s.StatRZGuestCR3WriteConflict));
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw=%#x!\n", iShw));
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * Causes trouble when the guest uses a PDE to refer to the whole page table level
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * structure. (Invalidate here; faults later on when it tries to change the page
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * table entries -> recheck; probably only applies to the RC case.)
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync LogFlow(("pgmPoolMonitorChainChanging: pae pd iShw=%#x: %RX64 -> freeing it!\n", iShw, uShw.pPDPae->a[iShw].u));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync /* Note: hardcoded PAE implementation dependency */
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync (pPage->enmKind == PGMPOOLKIND_PAE_PD_FOR_PAE_PD) ? PGMPOOL_IDX_PAE_PD : pPage->idx,
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync (pPage->enmKind == PGMPOOLKIND_PAE_PD_FOR_PAE_PD) ? iShw + (pPage->idx - PGMPOOL_IDX_PAE_PD_0) * X86_PG_PAE_ENTRIES : iShw);
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync /* paranoia / a bit assumptive. */
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync const unsigned iShw2 = (off + cbWrite - 1) / sizeof(X86PDEPAE);
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync AssertReturnVoid(iShw2 < RT_ELEMENTS(uShw.pPDPae->a));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync STAM_COUNTER_INC(&(pPool->CTX_SUFF(pVM)->pgm.s.StatRZGuestCR3WriteConflict));
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw2=%#x!\n", iShw2));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync LogFlow(("pgmPoolMonitorChainChanging: pae pd iShw2=%#x: %RX64 -> freeing it!\n", iShw2, uShw.pPDPae->a[iShw2].u));
ca3db470494a8b6eaec69ea37468a5cda65e2da8vboxsync /* Note: hardcoded PAE implementation dependency */
ca3db470494a8b6eaec69ea37468a5cda65e2da8vboxsync (pPage->enmKind == PGMPOOLKIND_PAE_PD_FOR_PAE_PD) ? PGMPOOL_IDX_PAE_PD : pPage->idx,
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync (pPage->enmKind == PGMPOOLKIND_PAE_PD_FOR_PAE_PD) ? iShw2 + (pPage->idx - PGMPOOL_IDX_PAE_PD_0) * X86_PG_PAE_ENTRIES : iShw2);
ca3db470494a8b6eaec69ea37468a5cda65e2da8vboxsync * Hopefully this doesn't happen very often:
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * - touching unused parts of the page
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * - messing with the bits of pd pointers without changing the physical address
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync /* PDPT roots are not page aligned; 32 byte only! */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync const unsigned offPdpt = GCPhysFault - pPage->GCPhys;
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync uShw.pv = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage);
ca3db470494a8b6eaec69ea37468a5cda65e2da8vboxsync if (iShw < X86_PG_PAE_PDPE_ENTRIES) /* don't use RT_ELEMENTS(uShw.pPDPT->a), because that's for long mode only */
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync STAM_COUNTER_INC(&(pPool->CTX_SUFF(pVM)->pgm.s.StatRZGuestCR3WriteConflict));
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw=%#x!\n", iShw));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync LogFlow(("pgmPoolMonitorChainChanging: pae pdpt iShw=%#x: %RX64 -> freeing it!\n", iShw, uShw.pPDPT->a[iShw].u));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync /* paranoia / a bit assumptive. */
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync const unsigned iShw2 = (offPdpt + cbWrite - 1) / sizeof(X86PDPE);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync STAM_COUNTER_INC(&(pPool->CTX_SUFF(pVM)->pgm.s.StatRZGuestCR3WriteConflict));
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw2=%#x!\n", iShw2));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync LogFlow(("pgmPoolMonitorChainChanging: pae pdpt iShw=%#x: %RX64 -> freeing it!\n", iShw2, uShw.pPDPT->a[iShw2].u));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync uShw.pv = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw=%#x!\n", iShw));
cc1ef2ef9bbc6a0ff964928d61b7298e5bfcce5fvboxsync LogFlow(("pgmPoolMonitorChainChanging: pae pd iShw=%#x: %RX64 -> freeing it!\n", iShw, uShw.pPDPae->a[iShw].u));
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync /* paranoia / a bit assumptive. */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync const unsigned iShw2 = (off + cbWrite - 1) / sizeof(X86PDEPAE);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync AssertReturnVoid(iShw2 < RT_ELEMENTS(uShw.pPDPae->a));
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync Assert(pgmMapAreMappingsEnabled(&pPool->CTX_SUFF(pVM)->pgm.s));
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync VM_FF_SET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3);
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync LogFlow(("pgmPoolMonitorChainChanging: Detected conflict at iShw2=%#x!\n", iShw2));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync LogFlow(("pgmPoolMonitorChainChanging: pae pd iShw2=%#x: %RX64 -> freeing it!\n", iShw2, uShw.pPDPae->a[iShw2].u));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Hopefully this doesn't happen very often:
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * - messing with the bits of pd pointers without changing the physical address
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync if (!VM_FF_ISSET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3))
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync uShw.pv = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync LogFlow(("pgmPoolMonitorChainChanging: pdpt iShw=%#x: %RX64 -> freeing it!\n", iShw, uShw.pPDPT->a[iShw].u));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync pgmPoolFree(pPool->CTX_SUFF(pVM), uShw.pPDPT->a[iShw].u & X86_PDPE_PG_MASK, pPage->idx, iShw);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync /* paranoia / a bit assumptive. */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync const unsigned iShw2 = (off + cbWrite - 1) / sizeof(X86PDPE);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync LogFlow(("pgmPoolMonitorChainChanging: pdpt iShw2=%#x: %RX64 -> freeing it!\n", iShw2, uShw.pPDPT->a[iShw2].u));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync pgmPoolFree(pPool->CTX_SUFF(pVM), uShw.pPDPT->a[iShw2].u & X86_PDPE_PG_MASK, pPage->idx, iShw2);
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * Hopefully this doesn't happen very often:
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * - messing with the bits of pd pointers without changing the physical address
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync if (!VM_FF_ISSET(pPool->CTX_SUFF(pVM), VM_FF_PGM_SYNC_CR3))
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync uShw.pv = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage);
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync LogFlow(("pgmPoolMonitorChainChanging: pml4 iShw=%#x: %RX64 -> freeing it!\n", iShw, uShw.pPML4->a[iShw].u));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync pgmPoolFree(pPool->CTX_SUFF(pVM), uShw.pPML4->a[iShw].u & X86_PML4E_PG_MASK, pPage->idx, iShw);
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync /* paranoia / a bit assumptive. */
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync const unsigned iShw2 = (off + cbWrite - 1) / sizeof(X86PML4E);
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync LogFlow(("pgmPoolMonitorChainChanging: pml4 iShw2=%#x: %RX64 -> freeing it!\n", iShw2, uShw.pPML4->a[iShw2].u));
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync pgmPoolFree(pPool->CTX_SUFF(pVM), uShw.pPML4->a[iShw2].u & X86_PML4E_PG_MASK, pPage->idx, iShw2);
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync#endif /* IN_RING0 */
004d74842597dacc4009803171296dfcf9398c69vboxsync AssertFatalMsgFailed(("enmKind=%d\n", pPage->enmKind));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Checks if a access could be a fork operation in progress.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Meaning, that the guest is setuping up the parent process for Copy-On-Write.
004d74842597dacc4009803171296dfcf9398c69vboxsync * @returns true if it's likly that we're forking, otherwise false.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pPool The pool.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pCpu The disassembled instruction.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param offFault The access offset.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsyncDECLINLINE(bool) pgmPoolMonitorIsForking(PPGMPOOL pPool, PDISCPUSTATE pCpu, unsigned offFault)
004d74842597dacc4009803171296dfcf9398c69vboxsync * i386 linux is using btr to clear X86_PTE_RW.
004d74842597dacc4009803171296dfcf9398c69vboxsync * The functions involved are (2.6.16 source inspection):
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * clear_bit
5366e994777f9d9391cf809dc77610f57270d75dvboxsync * ptep_set_wrprotect
5366e994777f9d9391cf809dc77610f57270d75dvboxsync * copy_one_pte
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * copy_pte_range
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * copy_pmd_range
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * copy_pud_range
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * copy_page_range
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * copy_process
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync /** @todo Validate that the bit index is X86_PTE_RW. */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync STAM_COUNTER_INC(&pPool->CTX_MID_Z(StatMonitor,Fork));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync return true;
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync return false;
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Determine whether the page is likely to have been reused.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @returns true if we consider the page as being reused for a different purpose.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @returns false if we consider it to still be a paging page.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pVM VM Handle.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pPage The page in question.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pRegFrame Trap register frame.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pCpu The disassembly info for the faulting instruction.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pvFault The fault address.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @remark The REP prefix check is left to the caller because of STOSD/W.
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsyncDECLINLINE(bool) pgmPoolMonitorIsReused(PVM pVM, PPGMPOOLPAGE pPage, PCPUMCTXCORE pRegFrame, PDISCPUSTATE pCpu, RTGCPTR pvFault)
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsync /** @todo could make this general, faulting close to rsp should be safe reuse heuristic. */
5366e994777f9d9391cf809dc77610f57270d75dvboxsync /* Fault caused by stack writes while trying to inject an interrupt event. */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync Log(("pgmPoolMonitorIsReused: reused %RGv for interrupt stack (rsp=%RGv).\n", pvFault, pRegFrame->rsp));
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync return true;
5366e994777f9d9391cf809dc77610f57270d75dvboxsync /* call implies the actual push of the return address faulted */
5366e994777f9d9391cf809dc77610f57270d75dvboxsync return true;
5366e994777f9d9391cf809dc77610f57270d75dvboxsync return true;
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync return true;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync return true;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync return true;
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync return true;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync case OP_MOVNTDQ: /* solaris - hwblkclr & hwblkpagecopy */
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync return true;
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync return true;
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync return false;
3ff8aa7d3c74cfbe8da5f77b8ea6c748cc79213avboxsync return true;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync //if (pPage->fCR3Mix)
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync // return false;
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync return false;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Flushes the page being accessed.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @returns VBox status code suitable for scheduling.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pVM The VM handle.
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync * @param pPool The pool.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pPage The pool page (head).
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pCpu The disassembly of the write instruction.
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync * @param pRegFrame The trap register frame.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param GCPhysFault The fault address as guest physical address.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pvFault The fault address.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsyncstatic int pgmPoolAccessHandlerFlush(PVM pVM, PPGMPOOL pPool, PPGMPOOLPAGE pPage, PDISCPUSTATE pCpu,
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync PCPUMCTXCORE pRegFrame, RTGCPHYS GCPhysFault, RTGCPTR pvFault)
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync * First, do the flushing.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Emulate the instruction (xp/w2k problem, requires pc/cr2/sp detection).
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync int rc2 = EMInterpretInstructionCPU(pVM, pCpu, pRegFrame, pvFault, &cbWritten);
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync if (PATMIsPatchGCAddr(pVM, (RTRCPTR)pRegFrame->eip))
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync LogFlow(("pgmPoolAccessHandlerPTWorker: Interpretation failed for patch code %04x:%RGv, ignoring.\n",
40c1a23e86c79b24a917a43c186b2e54504d12c1vboxsync STAM_COUNTER_INC(&pPool->StatMonitorRZIntrFailPatch2);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync STAM_COUNTER_INC(&pPool->CTX_MID_Z(StatMonitor,EmulateInstr));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync /* See use in pgmPoolAccessHandlerSimple(). */
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync LogFlow(("pgmPoolAccessHandlerPT: returns %Rrc (flushed)\n", rc));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Handles the STOSD write accesses.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @returns VBox status code suitable for scheduling.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pVM The VM handle.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pPool The pool.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pPage The pool page (head).
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pCpu The disassembly of the write instruction.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pRegFrame The trap register frame.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param GCPhysFault The fault address as guest physical address.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pvFault The fault address.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncDECLINLINE(int) pgmPoolAccessHandlerSTOSD(PVM pVM, PPGMPOOL pPool, PPGMPOOLPAGE pPage, PDISCPUSTATE pCpu,
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync PCPUMCTXCORE pRegFrame, RTGCPHYS GCPhysFault, RTGCPTR pvFault)
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Increment the modification counter and insert it into the list
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * of modified pages the first time.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Execute REP STOSD.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * This ASSUMES that we're not invoked by Trap0e on in a out-of-sync
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * write situation, meaning that it's safe to write here.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync uint32_t iPrevSubset = PGMDynMapPushAutoSubset(pVCpu);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync pgmPoolMonitorChainChanging(pPool, pPage, GCPhysFault, (RTGCPTR)pu32, NULL);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync pgmPoolMonitorChainChanging(pPool, pPage, GCPhysFault, (RTGCPTR)pu32, NULL);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync PGMPhysSimpleWriteGCPhys(pVM, GCPhysFault, &pRegFrame->eax, 4);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync /* See use in pgmPoolAccessHandlerSimple(). */
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Handles the simple write accesses.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @returns VBox status code suitable for scheduling.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pVM The VM handle.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPool The pool.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPage The pool page (head).
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pCpu The disassembly of the write instruction.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pRegFrame The trap register frame.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param GCPhysFault The fault address as guest physical address.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pvFault The fault address.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsyncDECLINLINE(int) pgmPoolAccessHandlerSimple(PVM pVM, PPGMPOOL pPool, PPGMPOOLPAGE pPage, PDISCPUSTATE pCpu,
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync PCPUMCTXCORE pRegFrame, RTGCPHYS GCPhysFault, RTGCPTR pvFault)
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Increment the modification counter and insert it into the list
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * of modified pages the first time.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Clear all the pages. ASSUMES that pvFault is readable.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync uint32_t iPrevSubset = PGMDynMapPushAutoSubset(pVCpu);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync pgmPoolMonitorChainChanging(pPool, pPage, GCPhysFault, pvFault, pCpu);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync pgmPoolMonitorChainChanging(pPool, pPage, GCPhysFault, pvFault, pCpu);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Interpret the instruction.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync int rc = EMInterpretInstructionCPU(pVM, pCpu, pRegFrame, pvFault, &cb);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync LogFlow(("pgmPoolAccessHandlerPTWorker: Interpretation failed for %04x:%RGv - opcode=%d\n",
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync pRegFrame->cs, (RTGCPTR)pRegFrame->rip, pCpu->pCurInstr->opcode));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync STAM_COUNTER_INC(&pPool->CTX_MID_Z(StatMonitor,EmulateInstr));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Quick hack, with logging enabled we're getting stale
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * code TLBs but no data TLB for EIP and crash in EMInterpretDisasOne.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Flushing here is BAD and expensive, I think EMInterpretDisasOne will
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * have to be fixed to support this. But that'll have to wait till next week.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * An alternative is to keep track of the changed PTEs together with the
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * GCPhys from the guest PT. This may proove expensive though.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * At the moment, it's VITAL that it's done AFTER the instruction interpreting
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * because we need the stale TLBs in some cases (XP boot). This MUST be fixed properly!
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync LogFlow(("pgmPoolAccessHandlerSimple: returns %Rrc cb=%d\n", rc, cb));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * \#PF Handler callback for PT write accesses.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @returns VBox status code (appropriate for GC return).
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pVM VM Handle.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param uErrorCode CPU Error code.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pRegFrame Trap register frame.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * NULL on DMA and other non CPU access.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pvFault The fault address (cr2).
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param GCPhysFault The GC physical address corresponding to pvFault.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pvUser User argument.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsyncDECLEXPORT(int) pgmPoolAccessHandler(PVM pVM, RTGCUINT uErrorCode, PCPUMCTXCORE pRegFrame, RTGCPTR pvFault, RTGCPHYS GCPhysFault, void *pvUser)
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync STAM_PROFILE_START(&pVM->pgm.s.CTX_SUFF(pPool)->CTX_SUFF_Z(StatMonitor), a);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync LogFlow(("pgmPoolAccessHandler: pvFault=%RGv pPage=%p:{.idx=%d} GCPhysFault=%RGp\n", pvFault, pPage, pPage->idx, GCPhysFault));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * We should ALWAYS have the list head as user parameter. This
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * is because we use that page to record the changes.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Disassemble the faulting instruction.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync int rc = EMInterpretDisasOne(pVM, pRegFrame, &Cpu, NULL);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Check if it's worth dealing with.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync bool fReused = false;
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync if ( ( pPage->cModifications < 48 /** @todo #define */ /** @todo need to check that it's not mapping EIP. */ /** @todo adjust this! */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync && !(fReused = pgmPoolMonitorIsReused(pVM, pPage, pRegFrame, &Cpu, pvFault))
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync && !pgmPoolMonitorIsForking(pPool, &Cpu, GCPhysFault & PAGE_OFFSET_MASK))
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Simple instructions, no REP prefix.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync rc = pgmPoolAccessHandlerSimple(pVM, pPool, pPage, &Cpu, pRegFrame, GCPhysFault, pvFault);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync STAM_PROFILE_STOP_EX(&pVM->pgm.s.CTX_SUFF(pPool)->CTX_SUFF_Z(StatMonitor), &pPool->CTX_MID_Z(StatMonitor,Handled), a);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Windows is frequently doing small memset() operations (netio test 4k+).
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * We have to deal with these or we'll kill the cache and performance.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync && pRegFrame->ecx * 4 <= PAGE_SIZE - ((uintptr_t)pvFault & PAGE_OFFSET_MASK)
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync && (pRegFrame->eax == 0 || pRegFrame->eax == 0x80) /* the two values observed. */
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync rc = pgmPoolAccessHandlerSTOSD(pVM, pPool, pPage, &Cpu, pRegFrame, GCPhysFault, pvFault);
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync STAM_PROFILE_STOP_EX(&pVM->pgm.s.CTX_SUFF(pPool)->CTX_SUFF_Z(StatMonitor), &pPool->CTX_MID_Z(StatMonitor,RepStosd), a);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync /* REP prefix, don't bother. */
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync STAM_COUNTER_INC(&pPool->CTX_MID_Z(StatMonitor,RepPrefix));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync Log4(("pgmPoolAccessHandler: eax=%#x ecx=%#x edi=%#x esi=%#x rip=%RGv opcode=%d prefix=%#x\n",
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync pRegFrame->eax, pRegFrame->ecx, pRegFrame->edi, pRegFrame->esi, (RTGCPTR)pRegFrame->rip, Cpu.pCurInstr->opcode, Cpu.prefix));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Not worth it, so flush it.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * If we considered it to be reused, don't to back to ring-3
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * to emulate failed instructions since we usually cannot
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * interpret then. This may be a bit risky, in which case
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * the reuse detection must be fixed.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync rc = pgmPoolAccessHandlerFlush(pVM, pPool, pPage, &Cpu, pRegFrame, GCPhysFault, pvFault);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync STAM_PROFILE_STOP_EX(&pVM->pgm.s.CTX_SUFF(pPool)->CTX_SUFF_Z(StatMonitor), &pPool->CTX_MID_Z(StatMonitor,FlushPage), a);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync# endif /* !IN_RING3 */
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync#endif /* PGMPOOL_WITH_MONITORING */
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Inserts a page into the GCPhys hash table.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pPool The pool.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pPage The page.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsyncDECLINLINE(void) pgmPoolHashInsert(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync Log3(("pgmPoolHashInsert: %RGp\n", pPage->GCPhys));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync Assert(pPage->GCPhys != NIL_RTGCPHYS); Assert(pPage->iNext == NIL_PGMPOOL_IDX);
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Removes a page from the GCPhys hash table.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPool The pool.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pPage The page.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncDECLINLINE(void) pgmPoolHashRemove(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync Log3(("pgmPoolHashRemove: %RGp\n", pPage->GCPhys));
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync AssertReleaseMsgFailed(("GCPhys=%RGp idx=%#x\n", pPage->GCPhys, pPage->idx));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Frees up one cache page.
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync * @returns VBox status code.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * @retval VINF_SUCCESS on success.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * @retval VERR_PGM_POOL_CLEARED if the deregistration of a physical handler will cause a light weight pool flush.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * @param pPool The pool.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * @param iUser The user index.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsyncstatic int pgmPoolCacheFreeOne(PPGMPOOL pPool, uint16_t iUser)
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync Assert(pPool->iAgeHead != pPool->iAgeTail); /* We shouldn't be here if there < 2 cached entries! */
5366e994777f9d9391cf809dc77610f57270d75dvboxsync * Select one page from the tail of the age list.
5366e994777f9d9391cf809dc77610f57270d75dvboxsync/* This is the alternative to the SyncCR3 pgmPoolCacheUsed calls.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync if (pPool->aPages[iToFree].iUserHead != NIL_PGMPOOL_USER_INDEX)
cc1ef2ef9bbc6a0ff964928d61b7298e5bfcce5fvboxsync uint16_t i = pPool->aPages[iToFree].iAgePrev;
5366e994777f9d9391cf809dc77610f57270d75dvboxsync for (unsigned j = 0; j < 10 && i != NIL_PGMPOOL_USER_INDEX; j++, i = pPool->aPages[i].iAgePrev)
5366e994777f9d9391cf809dc77610f57270d75dvboxsync if (pPool->aPages[iToFree].iUserHead == NIL_PGMPOOL_USER_INDEX)
5366e994777f9d9391cf809dc77610f57270d75dvboxsync iToFree = i;
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * Reject any attempts at flushing the currently active shadow CR3 mapping
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync if (PGMGetHyperCR3(pPool->CTX_SUFF(pVM)) == pPage->Core.Key)
7e8ef90d3160234df0f254131b87af4243d79476vboxsync /* Refresh the cr3 mapping by putting it at the head of the age list. */
e2489bd9ef063ae006feaebc3318ffa4143f6e16vboxsync * Checks if a kind mismatch is really a page being reused
e2489bd9ef063ae006feaebc3318ffa4143f6e16vboxsync * or if it's just normal remappings.
e2489bd9ef063ae006feaebc3318ffa4143f6e16vboxsync * @returns true if reused and the cached page (enmKind1) should be flushed
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * @returns false if not reused.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync * @param enmKind1 The kind of the cached page.
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync * @param enmKind2 The kind of the requested page.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsyncstatic bool pgmPoolCacheReusedByKind(PGMPOOLKIND enmKind1, PGMPOOLKIND enmKind2)
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync * Never reuse them. There is no remapping in non-paging mode.
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync case PGMPOOLKIND_PAE_PDPT_FOR_32BIT: /* never reuse them for other types */
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync return false;
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync return true;
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync * It's perfectly fine to reuse these, except for PAE and non-paging stuff.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync return true;
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync return false;
ccbdc11833996cb9f3be7868f1ebaefcacafb94dvboxsync * It's perfectly fine to reuse these, except for PAE and non-paging stuff.
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync return true;
c10a6f0c7041e4d1ee50ad38425aab9d43c55522vboxsync return false;
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync * These cannot be flushed, and it's common to reuse the PDs as PTs.
d4a9d525e6f2111d462d2d96462dced6b9ec00efvboxsync return false;
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Attempts to satisfy a pgmPoolAlloc request from the cache.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @returns VBox status code.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @retval VINF_PGM_CACHED_PAGE on success.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @retval VERR_FILE_NOT_FOUND if not found.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param pPool The pool.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param GCPhys The GC physical address of the page we're gonna shadow.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param enmKind The kind of mapping.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param iUser The shadow page pool index of the user table.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param iUserTable The index into the user table (shadowed).
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * @param ppPage Where to store the pointer to the page.
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsyncstatic int pgmPoolCacheAlloc(PPGMPOOL pPool, RTGCPHYS GCPhys, PGMPOOLKIND enmKind, uint16_t iUser, uint32_t iUserTable, PPPGMPOOLPAGE ppPage)
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync * Look up the GCPhys in the hash.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync Log3(("pgmPoolCacheAlloc: %RGp kind %d iUser=%d iUserTable=%x SLOT=%d\n", GCPhys, enmKind, iUser, iUserTable, i));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync Log3(("pgmPoolCacheAlloc: slot %d found page %RGp\n", i, pPage->GCPhys));
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync int rc = pgmPoolTrackAddUser(pPool, pPage, iUser, iUserTable);
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsync * The kind is different. In some cases we should now flush the page
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsync * as it has been reused, but in most cases this is normal remapping
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * of PDs as PT or big pages using the GCPhys field in a slightly
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsync * different way than the other kinds.
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync if (pgmPoolCacheReusedByKind((PGMPOOLKIND)pPage->enmKind, enmKind))
88cc9bf61296bc5526344415167bb2625ae1dd99vboxsync pgmPoolFlushPage(pPool, pPage); /* ASSUMES that VERR_PGM_POOL_CLEARED will be returned by pgmPoolTracInsert. */
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync } while (i != NIL_PGMPOOL_IDX);
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsync Log3(("pgmPoolCacheAlloc: Missed GCPhys=%RGp enmKind=%s\n", GCPhys, pgmPoolPoolKindToStr(enmKind)));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Inserts a page into the cache.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pPool The pool.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * @param pPage The cached page.
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * @param fCanBeCached Set if the page is fit for caching from the caller's point of view.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsyncstatic void pgmPoolCacheInsert(PPGMPOOL pPool, PPGMPOOLPAGE pPage, bool fCanBeCached)
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsync * Insert into the GCPhys hash if the page is fit for that.
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsync Log3(("pgmPoolCacheInsert: Caching %p:{.Core=%RHp, .idx=%d, .enmKind=%s, GCPhys=%RGp}\n",
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsync pPage, pPage->Core.Key, pPage->idx, pgmPoolPoolKindToStr(pPage->enmKind), pPage->GCPhys));
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsync Log3(("pgmPoolCacheInsert: Not caching %p:{.Core=%RHp, .idx=%d, .enmKind=%s, GCPhys=%RGp}\n",
f9ce005e61f0fbb51a2cabc53d58c3485151faa9vboxsync pPage, pPage->Core.Key, pPage->idx, pgmPoolPoolKindToStr(pPage->enmKind), pPage->GCPhys));
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync * Insert at the head of the age list.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pPool->aPages[pPool->iAgeHead].iAgePrev = pPage->idx;
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Flushes a cached page.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPool The pool.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPage The cached page.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsyncstatic void pgmPoolCacheFlushPage(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync Log3(("pgmPoolCacheFlushPage: %RGp\n", pPage->GCPhys));
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Remove the page from the hash.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Remove it from the age list.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pPool->aPages[pPage->iAgeNext].iAgePrev = pPage->iAgePrev;
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pPool->aPages[pPage->iAgePrev].iAgeNext = pPage->iAgeNext;
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync#endif /* PGMPOOL_WITH_CACHE */
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Looks for pages sharing the monitor.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @returns Pointer to the head page.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @returns NULL if not found.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPool The Pool
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pNewPage The page which is going to be monitored.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsyncstatic PPGMPOOLPAGE pgmPoolMonitorGetPageByGCPhys(PPGMPOOL pPool, PPGMPOOLPAGE pNewPage)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Look up the GCPhys in the hash.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync RTGCPHYS GCPhys = pNewPage->GCPhys & ~(RTGCPHYS)(PAGE_SIZE - 1);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync /* find the head */
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync /* ignore, no monitoring. */
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync AssertFatalMsgFailed(("enmKind=%d idx=%d\n", pPage->enmKind, pPage->idx));
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync } while (i != NIL_PGMPOOL_IDX);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Enabled write monitoring of a guest page.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @returns VBox status code.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @retval VINF_SUCCESS on success.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @retval VERR_PGM_POOL_CLEARED if the registration of the physical handler will cause a light weight pool flush.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPool The pool.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPage The cached page.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsyncstatic int pgmPoolMonitorInsert(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync LogFlow(("pgmPoolMonitorInsert %RGp\n", pPage->GCPhys & ~(RTGCPHYS)(PAGE_SIZE - 1)));
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Filter out the relevant kinds.
3fb3de312d1ff675e0f7cc62a7d46cbb1d5d9353vboxsync /* Nothing to monitor here. */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* Nothing to monitor here. */
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync AssertFatalMsgFailed(("This can't happen! enmKind=%d\n", pPage->enmKind));
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * Install handler.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PPGMPOOLPAGE pPageHead = pgmPoolMonitorGetPageByGCPhys(pPool, pPage);
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync Assert(pPageHead != pPage); Assert(pPageHead->iMonitoredNext != pPage->idx);
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync pPool->aPages[pPageHead->iMonitoredNext].iMonitoredPrev = pPage->idx;
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync Assert(pPage->iMonitoredNext == NIL_PGMPOOL_IDX); Assert(pPage->iMonitoredPrev == NIL_PGMPOOL_IDX);
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync const RTGCPHYS GCPhysPage = pPage->GCPhys & ~(RTGCPHYS)(PAGE_SIZE - 1);
c99b597540585068d22dde4c9f74730305f24097vboxsync rc = PGMHandlerPhysicalRegisterEx(pVM, PGMPHYSHANDLERTYPE_PHYSICAL_WRITE,
c99b597540585068d22dde4c9f74730305f24097vboxsync pPool->pfnAccessHandlerR3, MMHyperCCToR3(pVM, pPage),
c99b597540585068d22dde4c9f74730305f24097vboxsync pPool->pfnAccessHandlerR0, MMHyperCCToR0(pVM, pPage),
c99b597540585068d22dde4c9f74730305f24097vboxsync pPool->pfnAccessHandlerRC, MMHyperCCToRC(pVM, pPage),
c99b597540585068d22dde4c9f74730305f24097vboxsync /** @todo we should probably deal with out-of-memory conditions here, but for now increasing
c99b597540585068d22dde4c9f74730305f24097vboxsync * the heap size should suffice. */
c99b597540585068d22dde4c9f74730305f24097vboxsync if (pVM->pgm.s.fSyncFlags & PGM_SYNC_CLEAR_PGM_POOL)
c99b597540585068d22dde4c9f74730305f24097vboxsync * Disables write monitoring of a guest page.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @returns VBox status code.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @retval VINF_SUCCESS on success.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @retval VERR_PGM_POOL_CLEARED if the deregistration of the physical handler will cause a light weight pool flush.
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * @param pPool The pool.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPage The cached page.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsyncstatic int pgmPoolMonitorFlush(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
0fd108a555ae02f2fb557d5f2c40281999b60d15vboxsync * Filter out the relevant kinds.
c99b597540585068d22dde4c9f74730305f24097vboxsync /* Nothing to monitor here. */
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync AssertFatalMsgFailed(("This can't happen! enmKind=%d\n", pPage->enmKind));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Remove the page from the monitored list or uninstall it if last.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PPGMPOOLPAGE pNewHead = &pPool->aPages[pPage->iMonitoredNext];
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync rc = PGMHandlerPhysicalChangeCallbacks(pVM, pPage->GCPhys & ~(RTGCPHYS)(PAGE_SIZE - 1),
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync pPool->pfnAccessHandlerR3, MMHyperCCToR3(pVM, pNewHead),
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync pPool->pfnAccessHandlerR0, MMHyperCCToR0(pVM, pNewHead),
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync pPool->pfnAccessHandlerRC, MMHyperCCToRC(pVM, pNewHead),
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync pPool->aPages[pPage->iMonitoredPrev].iMonitoredNext = pPage->iMonitoredNext;
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync pPool->aPages[pPage->iMonitoredNext].iMonitoredPrev = pPage->iMonitoredPrev;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync rc = PGMHandlerPhysicalDeregister(pVM, pPage->GCPhys & ~(RTGCPHYS)(PAGE_SIZE - 1));
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync if (pVM->pgm.s.fSyncFlags & PGM_SYNC_CLEAR_PGM_POOL)
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * Remove it from the list of modified pages (if in it).
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * Set or clear the fCR3Mix attribute in a chain of monitored pages.
35e6d303696e46d969aaf9a59cc381333a483b0bvboxsync * @param pPool The Pool.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPage A page in the chain.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param fCR3Mix The new fCR3Mix value.
3641904508e6a0671c21a083a2ed03ebad8d976cvboxsyncstatic void pgmPoolMonitorChainChangeCR3Mix(PPGMPOOL pPool, PPGMPOOLPAGE pPage, bool fCR3Mix)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* current */
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync /* before */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* after */
ca3db470494a8b6eaec69ea37468a5cda65e2da8vboxsync * Installs or modifies monitoring of a CR3 page (special).
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * We're pretending the CR3 page is shadowed by the pool so we can use the
36f3c24e4ad9c6b813767db1faeabbe7e2ecc057vboxsync * generic mechanisms in detecting chained monitoring. (This also gives us a
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * tast of what code changes are required to really pool CR3 shadow pages.)
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @returns VBox status code.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pPool The pool.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param idxRoot The CR3 (root) page index.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param GCPhysCR3 The (new) CR3 value.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsyncint pgmPoolMonitorMonitorCR3(PPGMPOOL pPool, uint16_t idxRoot, RTGCPHYS GCPhysCR3)
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync Assert(idxRoot != NIL_PGMPOOL_IDX && idxRoot < PGMPOOL_IDX_FIRST);
36f3c24e4ad9c6b813767db1faeabbe7e2ecc057vboxsync LogFlow(("pgmPoolMonitorMonitorCR3: idxRoot=%d pPage=%p:{.GCPhys=%RGp, .fMonitored=%d} GCPhysCR3=%RGp\n",
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync idxRoot, pPage, pPage->GCPhys, pPage->fMonitored, GCPhysCR3));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * The unlikely case where it already matches.
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync * Flush the current monitoring and remove it from the hash.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pgmPoolMonitorChainChangeCR3Mix(pPool, pPage, false);
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync * Monitor the page at the new location and insert it into the hash.
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync pgmPoolMonitorChainChangeCR3Mix(pPool, pPage, true);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Removes the monitoring of a CR3 page (special).
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync * @returns VBox status code.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPool The pool.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param idxRoot The CR3 (root) page index.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsyncint pgmPoolMonitorUnmonitorCR3(PPGMPOOL pPool, uint16_t idxRoot)
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync Assert(idxRoot != NIL_PGMPOOL_IDX && idxRoot < PGMPOOL_IDX_FIRST);
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync LogFlow(("pgmPoolMonitorUnmonitorCR3: idxRoot=%d pPage=%p:{.GCPhys=%RGp, .fMonitored=%d}\n",
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync idxRoot, pPage, pPage->GCPhys, pPage->fMonitored));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pgmPoolMonitorChainChangeCR3Mix(pPool, pPage, false);
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync# endif /* PGMPOOL_WITH_MIXED_PT_CR3 */
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync * Inserts the page into the list of modified pages.
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync * @param pPool The pool.
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync * @param pPage The page.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsyncvoid pgmPoolMonitorModifiedInsert(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync Log3(("pgmPoolMonitorModifiedInsert: idx=%d\n", pPage->idx));
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync ("Next=%d Prev=%d idx=%d cModifications=%d Head=%d cModifiedPages=%d\n",
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync pPage->iModifiedNext, pPage->iModifiedPrev, pPage->idx, pPage->cModifications,
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync pPool->aPages[pPool->iModifiedHead].iModifiedPrev = pPage->idx;
8bc8d66f188d5357155b8340e2d489573be2b607vboxsync if (pPool->cModifiedPages > pPool->cModifiedPagesHigh)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Removes the page from the list of modified pages and resets the
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * moficiation counter.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPool The pool.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * @param pPage The page which is believed to be in the list of modified pages.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsyncstatic void pgmPoolMonitorModifiedRemove(PPGMPOOL pPool, PPGMPOOLPAGE pPage)
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync Log3(("pgmPoolMonitorModifiedRemove: idx=%d cModifications=%d\n", pPage->idx, pPage->cModifications));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync pPool->aPages[pPage->iModifiedNext].iModifiedPrev = NIL_PGMPOOL_IDX;
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync pPool->aPages[pPage->iModifiedPrev].iModifiedNext = pPage->iModifiedNext;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync pPool->aPages[pPage->iModifiedNext].iModifiedPrev = pPage->iModifiedPrev;
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Zaps the list of modified pages, resetting their modification counters in the process.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pVM The VM handle.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync LogFlow(("pgmPoolMonitorModifiedClearAll: cModifiedPages=%d\n", pPool->cModifiedPages));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync AssertMsg(cPages == pPool->cModifiedPages, ("%d != %d\n", cPages, pPool->cModifiedPages));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Clear all shadow pages and clear all modification counters.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pVM The VM handle.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @remark Should only be used when monitoring is available, thus placed in
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * the PGMPOOL_WITH_MONITORING #ifdef.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("pgmPoolClearAll: cUsedPages=%d\n", pPool->cUsedPages));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Iterate all the pages until we've encountered all that in use.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * This is simple but not quite optimal solution.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync unsigned cModifiedPages = 0; NOREF(cModifiedPages);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * We only care about shadow page tables.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync void *pvShw = PGMPOOL_PAGE_2_PTR(pPool->CTX_SUFF(pVM), pPage);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* fall thru */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Assert(!pPage->cModifications || ++cModifiedPages);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Assert(pPage->iModifiedNext == NIL_PGMPOOL_IDX || pPage->cModifications);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Assert(pPage->iModifiedPrev == NIL_PGMPOOL_IDX || pPage->cModifications);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync /* swipe the special pages too. */
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync for (iPage = PGMPOOL_IDX_FIRST_SPECIAL; iPage < PGMPOOL_IDX_FIRST; iPage++)
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync Assert(!pPage->cModifications || ++cModifiedPages);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync Assert(pPage->iModifiedNext == NIL_PGMPOOL_IDX || pPage->cModifications);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync Assert(pPage->iModifiedPrev == NIL_PGMPOOL_IDX || pPage->cModifications);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync AssertMsg(cModifiedPages == pPool->cModifiedPages, ("%d != %d\n", cModifiedPages, pPool->cModifiedPages));
06ea6bcf23874b662d499b3f130024c98b2dd7a6vboxsync * Clear all the GCPhys links and rebuild the phys ext free list.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync for (PPGMRAMRANGE pRam = pPool->CTX_SUFF(pVM)->pgm.s.CTX_SUFF(pRamRanges);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync while (iPage-- > 0)
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync pRam->aPages[iPage].HCPhys &= MM_RAM_FLAGS_NO_REFS_MASK; /** @todo PAGE FLAGS */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PPGMPOOLPHYSEXT paPhysExts = pPool->CTX_SUFF(paPhysExts);
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync for (unsigned i = 0; i < cMaxPhysExts; i++)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync paPhysExts[cMaxPhysExts - 1].iNext = NIL_PGMPOOL_PHYSEXT_INDEX;
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync#endif /* IN_RING3 */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Handle SyncCR3 pool tasks
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync * @returns VBox status code.
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync * @retval VINF_SUCCESS if successfully added.
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync * @retval VINF_PGM_SYNC_CR3 is it needs to be deferred to ring 3 (GC only)
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * @param pVM The VM handle.
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync * @remark Should only be used when monitoring is available, thus placed in
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * the PGMPOOL_WITH_MONITORING #ifdef.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * When monitoring shadowed pages, we reset the modification counters on CR3 sync.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Occasionally we will have to clear all the shadow page tables because we wanted
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * to monitor a page which was mapped by too many shadowed page tables. This operation
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * sometimes refered to as a 'lightweight flush'.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync if (!(pVM->pgm.s.fSyncFlags & PGM_SYNC_CLEAR_PGM_POOL))
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync# ifdef IN_RING3 /* Don't flush in ring-0 or raw mode, it's taking too long. */
9523921c89c66f4bececdbd5ac95aed0039eda1bvboxsync# else /* !IN_RING3 */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync LogFlow(("SyncCR3: PGM_SYNC_CLEAR_PGM_POOL is set -> VINF_PGM_SYNC_CR3\n"));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync VM_FF_SET(pVM, VM_FF_PGM_SYNC_CR3); /** @todo no need to do global sync, right? */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync# endif /* !IN_RING3 */
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync#endif /* PGMPOOL_WITH_MONITORING */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Frees up at least one user entry.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @returns VBox status code.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @retval VINF_SUCCESS if successfully added.
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync * @retval VERR_PGM_POOL_FLUSHED if the pool was flushed.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pPool The pool.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param iUser The user index.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsyncstatic int pgmPoolTrackFreeOneUser(PPGMPOOL pPool, uint16_t iUser)
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Just free cached pages in a braindead fashion.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync /** @todo walk the age list backwards and free the first with usage. */
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync } while (pPool->iUserFreeHead == NIL_PGMPOOL_USER_INDEX);
58c0567dee3cc3ebe62dec1e27f8e35bac4ddeb0vboxsync * Lazy approach.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* @todo incompatible with long mode paging (cr3 root will be flushed) */
9e17ca2e9d797e845e3284141dd4086a4b817ae5vboxsync * Inserts a page into the cache.
9e17ca2e9d797e845e3284141dd4086a4b817ae5vboxsync * This will create user node for the page, insert it into the GCPhys
9e17ca2e9d797e845e3284141dd4086a4b817ae5vboxsync * hash, and insert it into the age list.
9e17ca2e9d797e845e3284141dd4086a4b817ae5vboxsync * @returns VBox status code.
9e17ca2e9d797e845e3284141dd4086a4b817ae5vboxsync * @retval VINF_SUCCESS if successfully added.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @retval VERR_PGM_POOL_FLUSHED if the pool was flushed.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @retval VERR_PGM_POOL_CLEARED if the deregistration of the physical handler will cause a light weight pool flush.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pPool The pool.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param pPage The cached page.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param GCPhys The GC physical address of the page we're gonna shadow.
9e17ca2e9d797e845e3284141dd4086a4b817ae5vboxsync * @param iUser The user index.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @param iUserTable The user table index.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsyncDECLINLINE(int) pgmPoolTrackInsert(PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTGCPHYS GCPhys, uint16_t iUser, uint32_t iUserTable)
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync LogFlow(("pgmPoolTrackInsert iUser %d iUserTable %d\n", iUser, iUserTable));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Find free a user node.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * Unlink the user node from the free list,
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * initialize and insert it into the user list.
5366e994777f9d9391cf809dc77610f57270d75dvboxsync * Insert into cache and enable monitoring of the guest page if enabled.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Until we implement caching of all levels, including the CR3 one, we'll
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * have to make sure we don't try monitor & cache any recursive reuse of
5366e994777f9d9391cf809dc77610f57270d75dvboxsync * a monitored CR3 page. Because all windows versions are doing this we'll
5366e994777f9d9391cf809dc77610f57270d75dvboxsync * have to be able to do combined access monitoring, CR3 + PT and
5366e994777f9d9391cf809dc77610f57270d75dvboxsync * PD + PT (guest PAE).
5366e994777f9d9391cf809dc77610f57270d75dvboxsync * We're now cooperating with the CR3 monitor if an uncachable page is found.
5366e994777f9d9391cf809dc77610f57270d75dvboxsync#if defined(PGMPOOL_WITH_MONITORING) || defined(PGMPOOL_WITH_CACHE)
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync const bool fCanBeMonitored = true;
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync bool fCanBeMonitored = pPool->CTX_SUFF(pVM)->pgm.s.GCPhysGstCR3Monitored == NIL_RTGCPHYS
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync || (GCPhys & X86_PTE_PAE_PG_MASK) != (pPool->CTX_SUFF(pVM)->pgm.s.GCPhysGstCR3Monitored & X86_PTE_PAE_PG_MASK)
462e60a19d02a99b2b1a5c08dff74bb0808d707cvboxsync pgmPoolCacheInsert(pPool, pPage, fCanBeMonitored); /* This can be expanded. */
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync /* 'Failed' - free the usage, and keep it in the cache (if enabled). */
5366e994777f9d9391cf809dc77610f57270d75dvboxsync#endif /* PGMPOOL_WITH_MONITORING */
9523921c89c66f4bececdbd5ac95aed0039eda1bvboxsync# ifdef PGMPOOL_WITH_CACHE /* (only used when the cache is enabled.) */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * Adds a user reference to a page.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * This will
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * This will move the page to the head of the
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * @returns VBox status code.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * @retval VINF_SUCCESS if successfully added.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * @retval VERR_PGM_POOL_FLUSHED if the pool was flushed.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * @param pPool The pool.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPage The cached page.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param iUser The user index.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param iUserTable The user table.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsyncstatic int pgmPoolTrackAddUser(PPGMPOOL pPool, PPGMPOOLPAGE pPage, uint16_t iUser, uint32_t iUserTable)
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync LogFlow(("pgmPoolTrackAddUser iUser %d iUserTable %d\n", iUser, iUserTable));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Check that the entry doesn't already exists.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync AssertMsg(paUsers[i].iUser != iUser || paUsers[i].iUserTable != iUserTable, ("%x %x vs new %x %x\n", paUsers[i].iUser, paUsers[i].iUserTable, iUser, iUserTable));
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Allocate a user node.
4121d226ac899f17e13aff3aff42b603c8b5c1fevboxsync * Initialize the user node and insert it.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * Tell the cache to update its replacement stats for this page.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync# endif /* PGMPOOL_WITH_CACHE */
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync * Frees a user record associated with a page.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * This does not clear the entry in the user table, it simply replaces the
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * user record to the chain of free records.
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync * @param pPool The pool.
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync * @param HCPhys The HC physical address of the shadow page.
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync * @param iUser The shadow page pool index of the user table.
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync * @param iUserTable The index into the user table (shadowed).
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsyncstatic void pgmPoolTrackFreeUser(PPGMPOOL pPool, PPGMPOOLPAGE pPage, uint16_t iUser, uint32_t iUserTable)
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * Unlink and free the specified user entry.
e378dfdadd62aadc0a012c9953322d979d7606e6vboxsync /* Special: For PAE and 32-bit paging, there is usually no more than one user. */
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync /* General: Linear search. */
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync /* Fatal: didn't find it */
9523921c89c66f4bececdbd5ac95aed0039eda1bvboxsync AssertFatalMsgFailed(("Didn't find the user entry! iUser=%#x iUserTable=%#x GCPhys=%RGp\n",
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * Gets the entry size of a shadow table.
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync * @param enmKind The kind of page.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @returns The size of the entry in bytes. That is, 4 or 8.
5050fc8de0b121eab1b738d7c1007cde4903284dvboxsync * @returns If the kind is not for a table, an assertion is raised and 0 is
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * returned.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsyncDECLINLINE(unsigned) pgmPoolTrackGetShadowEntrySize(PGMPOOLKIND enmKind)
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync * Gets the entry size of a guest table.
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync * @param enmKind The kind of page.
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsync * @returns The size of the entry in bytes. That is, 0, 4 or 8.
5f2b03bf7695dabd71222dba123532a3f76828c1vboxsync * @returns If the kind is not for a table, an assertion is raised and 0 is
7862f4bd000f1eb6c86289f5ac2849e9cf943ca9vboxsync * returned.
907b6adfa052386a0666d5557bee9bdbc100c2e5vboxsyncDECLINLINE(unsigned) pgmPoolTrackGetGuestEntrySize(PGMPOOLKIND enmKind)
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync /** @todo can we return 0? (nobody is calling this...) */
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync * Scans one shadow page table for mappings of a physical page.
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync * @param pVM The VM handle.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param pPhysPage The guest page in question.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param iShw The shadow page table.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync * @param cRefs The number of references made in that PT.
687794577e2e35c3cae67e692a7f2130d1262a82vboxsyncstatic void pgmPoolTrackFlushGCPhysPTInt(PVM pVM, PCPGMPAGE pPhysPage, uint16_t iShw, uint16_t cRefs)
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync LogFlow(("pgmPoolTrackFlushGCPhysPT: HCPhys=%RHp iShw=%d cRefs=%d\n", pPhysPage->HCPhys, iShw, cRefs));
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync * Assert sanity.
687794577e2e35c3cae67e692a7f2130d1262a82vboxsync AssertFatalMsg(iShw < pPool->cCurPages && iShw != NIL_PGMPOOL_IDX, ("iShw=%d\n", iShw));
ed9d3db07648c7e3a979105c15ad752ee9ea18devboxsync * Then, clear the actual mappings to the page in the shadow PT.
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync const uint32_t u32 = PGM_PAGE_GET_HCPHYS(pPhysPage) | X86_PTE_P;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PX86PT pPT = (PX86PT)PGMPOOL_PAGE_2_PTR(pVM, pPage);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync for (unsigned i = pPage->iFirstPresent; i < RT_ELEMENTS(pPT->a); i++)
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync if ((pPT->a[i].u & (X86_PTE_PG_MASK | X86_PTE_P)) == u32)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Log4(("pgmPoolTrackFlushGCPhysPTs: i=%d pte=%RX32 cRefs=%#x\n", i, pPT->a[i], cRefs));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pPT->a[i].u = 0;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync RTLogPrintf("cRefs=%d iFirstPresent=%d cPresent=%d\n", cRefs, pPage->iFirstPresent, pPage->cPresent);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync if ((pPT->a[i].u & (X86_PTE_PG_MASK | X86_PTE_P)) == u32)
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync pPT->a[i].u = 0;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync AssertFatalMsgFailed(("cRefs=%d iFirstPresent=%d cPresent=%d\n", cRefs, pPage->iFirstPresent, pPage->cPresent));
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync const uint64_t u64 = PGM_PAGE_GET_HCPHYS(pPhysPage) | X86_PTE_P;
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync PX86PTPAE pPT = (PX86PTPAE)PGMPOOL_PAGE_2_PTR(pVM, pPage);
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync for (unsigned i = pPage->iFirstPresent; i < RT_ELEMENTS(pPT->a); i++)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync if ((pPT->a[i].u & (X86_PTE_PAE_PG_MASK | X86_PTE_P)) == u64)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Log4(("pgmPoolTrackFlushGCPhysPTs: i=%d pte=%RX64 cRefs=%#x\n", i, pPT->a[i], cRefs));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pPT->a[i].u = 0;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync RTLogPrintf("cRefs=%d iFirstPresent=%d cPresent=%d\n", cRefs, pPage->iFirstPresent, pPage->cPresent);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync if ((pPT->a[i].u & (X86_PTE_PAE_PG_MASK | X86_PTE_P)) == u64)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pPT->a[i].u = 0;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync AssertFatalMsgFailed(("cRefs=%d iFirstPresent=%d cPresent=%d u64=%RX64\n", cRefs, pPage->iFirstPresent, pPage->cPresent, u64));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync const uint64_t u64 = PGM_PAGE_GET_HCPHYS(pPhysPage) | X86_PTE_P;
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync PEPTPT pPT = (PEPTPT)PGMPOOL_PAGE_2_PTR(pVM, pPage);
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync for (unsigned i = pPage->iFirstPresent; i < RT_ELEMENTS(pPT->a); i++)
702a8ee2dc1de96f2f77e97135015d3e243186fdvboxsync if ((pPT->a[i].u & (EPT_PTE_PG_MASK | X86_PTE_P)) == u64)
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync Log4(("pgmPoolTrackFlushGCPhysPTs: i=%d pte=%RX64 cRefs=%#x\n", i, pPT->a[i], cRefs));
2f3883b126a405f92b19e829472f614c7352b4f9vboxsync pPT->a[i].u = 0;
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync RTLogPrintf("cRefs=%d iFirstPresent=%d cPresent=%d\n", cRefs, pPage->iFirstPresent, pPage->cPresent);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync if ((pPT->a[i].u & (EPT_PTE_PG_MASK | X86_PTE_P)) == u64)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pPT->a[i].u = 0;
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync AssertFatalMsgFailed(("cRefs=%d iFirstPresent=%d cPresent=%d\n", cRefs, pPage->iFirstPresent, pPage->cPresent));
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync AssertFatalMsgFailed(("enmKind=%d iShw=%d\n", pPage->enmKind, iShw));
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Scans one shadow page table for mappings of a physical page.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pVM The VM handle.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPhysPage The guest page in question.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param iShw The shadow page table.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param cRefs The number of references made in that PT.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsyncvoid pgmPoolTrackFlushGCPhysPT(PVM pVM, PPGMPAGE pPhysPage, uint16_t iShw, uint16_t cRefs)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync PPGMPOOL pPool = pVM->pgm.s.CTX_SUFF(pPool); NOREF(pPool);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync LogFlow(("pgmPoolTrackFlushGCPhysPT: HCPhys=%RHp iShw=%d cRefs=%d\n", pPhysPage->HCPhys, iShw, cRefs));
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync STAM_PROFILE_START(&pPool->StatTrackFlushGCPhysPT, f);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pgmPoolTrackFlushGCPhysPTInt(pVM, pPhysPage, iShw, cRefs);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pPhysPage->HCPhys &= MM_RAM_FLAGS_NO_REFS_MASK; /** @todo PAGE FLAGS */
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync STAM_PROFILE_STOP(&pPool->StatTrackFlushGCPhysPT, f);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * Flushes a list of shadow page tables mapping the same physical page.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pVM The VM handle.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param pPhysPage The guest page in question.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync * @param iPhysExt The physical cross reference extent list to flush.
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsyncvoid pgmPoolTrackFlushGCPhysPTs(PVM pVM, PPGMPAGE pPhysPage, uint16_t iPhysExt)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync STAM_PROFILE_START(&pPool->StatTrackFlushGCPhysPTs, f);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync LogFlow(("pgmPoolTrackFlushGCPhysPTs: HCPhys=%RHp iPhysExt\n", pPhysPage->HCPhys, iPhysExt));
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync for (unsigned i = 0; i < RT_ELEMENTS(pPhysExt->aidx); i++)
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pgmPoolTrackFlushGCPhysPTInt(pVM, pPhysPage, pPhysExt->aidx[i], 1);
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync /* insert the list into the free list and clear the ram range entry. */
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync pPhysPage->HCPhys &= MM_RAM_FLAGS_NO_REFS_MASK; /** @todo PAGE FLAGS */
9bff17fe6983cfda2ddd98f1979841bcb48e78e7vboxsync STAM_PROFILE_STOP(&pPool->StatTrackFlushGCPhysPTs, f);
return VINF_PGM_GCPHYS_ALIASED;
pPT->a[i].u = 0;
if (!--cPresent)
pPT->a[i].u = 0;
if (!--cPresent)
if (!--cLeft)
return VINF_SUCCESS;
#ifdef VBOX_STRICT
# ifdef VBOX_WITH_PGMPOOL_PAGING_ONLY
case PGMPOOLKIND_32BIT_PD:
case PGMPOOLKIND_ROOT_PAE_PD:
AssertMsg(!(u.pau64[iUserTable] & PGM_PDFLAGS_MAPPING), ("%llx %d\n", u.pau64[iUserTable], iUserTable));
# ifdef VBOX_WITH_PGMPOOL_PAGING_ONLY
case PGMPOOLKIND_PAE_PDPT:
case PGMPOOLKIND_ROOT_PDPT:
case PGMPOOLKIND_PAE_PD_PHYS:
case PGMPOOLKIND_64BIT_PML4:
case PGMPOOLKIND_ROOT_NESTED:
case PGMPOOLKIND_32BIT_PD:
case PGMPOOLKIND_PAE_PD_PHYS:
case PGMPOOLKIND_64BIT_PML4:
case PGMPOOLKIND_ROOT_PAE_PD:
case PGMPOOLKIND_PAE_PDPT:
case PGMPOOLKIND_ROOT_PDPT:
case PGMPOOLKIND_ROOT_NESTED:
AssertFatalMsgFailed(("enmKind=%d iUser=%#x iUserTable=%#x\n", pUserPage->enmKind, pUser->iUser, pUser->iUserTable));
while (i != NIL_PGMPOOL_USER_INDEX)
i = iNext;
#ifdef PGMPOOL_WITH_GCPHYS_TRACKING
return NULL;
return pPhysExt;
return iPhysExt | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
return iPhysExtStart | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
if (!--cMax)
return MM_RAM_FLAGS_IDX_OVERFLOWED | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
if (!pNew)
return MM_RAM_FLAGS_IDX_OVERFLOWED | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
LogFlow(("pgmPoolTrackPhysExtAddref: added new extent %d:{%d}->%d\n", iPhysExt, iShwPT, iPhysExtStart));
return iPhysExt | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
if (pPhysExt)
LogFlow(("pgmPoolTrackPhysExtAddref: new extent: %d:{%d, %d}\n", iPhysExt, u16 & MM_RAM_FLAGS_IDX_MASK, iShwPT));
u16 = iPhysExt | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
u16 = MM_RAM_FLAGS_IDX_OVERFLOWED | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT));
else if (u16 != (MM_RAM_FLAGS_IDX_OVERFLOWED | (MM_RAM_FLAGS_CREFS_PHYSEXT << (MM_RAM_FLAGS_CREFS_SHIFT - MM_RAM_FLAGS_IDX_SHIFT))))
return u16;
AssertFatalMsg(cRefs == MM_RAM_FLAGS_CREFS_PHYSEXT, ("cRefs=%d HCPhys=%RHp pPage=%p:{.idx=%d}\n", cRefs, pPhysPage->HCPhys, pPage, pPage->idx));
LogFlow(("pgmPoolTrackPhysExtDerefGCPhys: HCPhys=%RX64 idx=%d lonely\n", pPhysPage->HCPhys, pPage->idx));
LogFlow(("pgmPoolTrackPhysExtDerefGCPhys: HCPhys=%RX64 idx=%d head\n", pPhysPage->HCPhys, pPage->idx));
AssertFatalMsgFailed(("not-found! cRefs=%d HCPhys=%RHp pPage=%p:{.idx=%d}\n", cRefs, pPhysPage->HCPhys, pPage, pPage->idx));
static void pgmPoolTracDerefGCPhys(PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTHCPHYS HCPhys, RTGCPHYS GCPhys)
while (pRam)
#ifdef LOG_ENABLED
static void pgmPoolTracDerefGCPhysHint(PPGMPOOL pPool, PPGMPOOLPAGE pPage, RTHCPHYS HCPhys, RTGCPHYS GCPhysHint)
while (pRam)
while (pRam)
while (iPage-- > 0)
DECLINLINE(void) pgmPoolTrackDerefPT32Bit32Bit(PPGMPOOL pPool, PPGMPOOLPAGE pPage, PX86PT pShwPT, PCX86PT pGstPT)
pgmPoolTracDerefGCPhysHint(pPool, pPage, pShwPT->a[i].u & X86_PTE_PG_MASK, pGstPT->a[i].u & X86_PTE_PG_MASK);
DECLINLINE(void) pgmPoolTrackDerefPTPae32Bit(PPGMPOOL pPool, PPGMPOOLPAGE pPage, PX86PTPAE pShwPT, PCX86PT pGstPT)
pgmPoolTracDerefGCPhysHint(pPool, pPage, pShwPT->a[i].u & X86_PTE_PAE_PG_MASK, pGstPT->a[i].u & X86_PTE_PG_MASK);
DECLINLINE(void) pgmPoolTrackDerefPTPaePae(PPGMPOOL pPool, PPGMPOOLPAGE pPage, PX86PTPAE pShwPT, PCX86PTPAE pGstPT)
pgmPoolTracDerefGCPhysHint(pPool, pPage, pShwPT->a[i].u & X86_PTE_PAE_PG_MASK, pGstPT->a[i].u & X86_PTE_PAE_PG_MASK);
PPGMPOOLPAGE pSubPage = (PPGMPOOLPAGE)RTAvloHCPhysGet(&pPool->HCPhysTree, pShwPD->a[i].u & X86_PDE_PG_MASK);
if (pSubPage)
PPGMPOOLPAGE pSubPage = (PPGMPOOLPAGE)RTAvloHCPhysGet(&pPool->HCPhysTree, pShwPD->a[i].u & X86_PDE_PAE_PG_MASK);
if (pSubPage)
PPGMPOOLPAGE pSubPage = (PPGMPOOLPAGE)RTAvloHCPhysGet(&pPool->HCPhysTree, pShwPDPT->a[i].u & X86_PDPE_PG_MASK);
if (pSubPage)
PPGMPOOLPAGE pSubPage = (PPGMPOOLPAGE)RTAvloHCPhysGet(&pPool->HCPhysTree, pShwPML4->a[i].u & X86_PDPE_PG_MASK);
if (pSubPage)
PPGMPOOLPAGE pSubPage = (PPGMPOOLPAGE)RTAvloHCPhysGet(&pPool->HCPhysTree, pShwPD->a[i].u & EPT_PDE_PG_MASK);
if (pSubPage)
PPGMPOOLPAGE pSubPage = (PPGMPOOLPAGE)RTAvloHCPhysGet(&pPool->HCPhysTree, pShwPDPT->a[i].u & EPT_PDPTE_PG_MASK);
if (pSubPage)
#ifdef PGMPOOL_WITH_GCPHYS_TRACKING
void *pvGst;
void *pvGst;
void *pvGst;
case PGMPOOLKIND_PAE_PD_PHYS:
case PGMPOOLKIND_32BIT_PD:
case PGMPOOLKIND_PAE_PDPT:
case PGMPOOLKIND_64BIT_PML4:
#ifndef VBOX_WITH_PGMPOOL_PAGING_ONLY
case PGMPOOLKIND_ROOT_PDPT:
case PGMPOOLKIND_ROOT_NESTED:
#ifdef IN_RING3
#ifdef PGMPOOL_WITH_MONITORING
#ifdef PGMPOOL_WITH_USER_TRACKING
#ifdef PGMPOOL_WITH_CACHE
#ifdef PGMPOOL_WITH_USER_TRACKING
for (unsigned i = 0; i < cMaxUsers; i++)
#ifdef PGMPOOL_WITH_GCPHYS_TRACKING
pRam;
while (iPage-- > 0)
for (unsigned i = 0; i < cMaxPhysExts; i++)
#ifdef PGMPOOL_WITH_MONITORING
#ifdef PGMPOOL_WITH_CACHE
#ifdef PGMPOOL_WITH_MONITORING
# ifdef PGMPOOL_WITH_CACHE
#ifdef PGMPOOL_WITH_USER_TRACKING
#ifdef PGMPOOL_WITH_CACHE
* @retval VERR_PGM_POOL_CLEARED if the deregistration of the physical handler will cause a light weight pool flush.
Log(("pgmPoolFlushPage: special root page, rejected. enmKind=%s idx=%d\n", pgmPoolPoolKindToStr(pPage->enmKind), pPage->idx));
return VINF_SUCCESS;
("Can't free the shadow CR3! (%RHp vs %RHp kind=%d\n", PGMGetHyperCR3(pPool->CTX_SUFF(pVM)), pPage->Core.Key, pPage->enmKind));
("Can't free the shadow CR3! (%RHp vs %RHp kind=%d\n", PGMGetHyperCR3(pPool->CTX_SUFF(pVM)), pPage->Core.Key, pPage->enmKind));
Log(("pgmPoolFlushPage: current active shadow CR3, rejected. enmKind=%s idx=%d\n", pgmPoolPoolKindToStr(pPage->enmKind), pPage->idx));
return VINF_SUCCESS;
#ifdef PGMPOOL_WITH_USER_TRACKING
#ifdef PGMPOOL_WITH_CACHE
#ifdef PGMPOOL_WITH_MONITORING
return rc;
#ifdef PGMPOOL_WITH_USER_TRACKING
#ifdef PGMPOOL_WITH_CACHE
#ifdef IN_RING3
return rc;
return VINF_SUCCESS;
#ifdef PGMPOOL_WITH_CACHE
return VERR_PGM_POOL_FLUSHED;
int pgmPoolAlloc(PVM pVM, RTGCPHYS GCPhys, PGMPOOLKIND enmKind, uint16_t iUser, uint32_t iUserTable, PPPGMPOOLPAGE ppPage)
LogFlow(("pgmPoolAlloc: GCPhys=%RGp enmKind=%s iUser=%#x iUserTable=%#x\n", GCPhys, pgmPoolPoolKindToStr(enmKind), iUser, iUserTable));
/** @todo CSAM/PGMPrefetchPage messes up here during CSAMR3CheckGates
* Assert(!(pVM->pgm.s.fSyncFlags & PGM_SYNC_CLEAR_PGM_POOL)); */
#ifdef PGMPOOL_WITH_CACHE
LogFlow(("pgmPoolAlloc: cached returns %Rrc *ppPage=%p:{.Key=%RHp, .idx=%d}\n", rc2, *ppPage, (*ppPage)->Core.Key, (*ppPage)->idx));
return rc2;
return rc;
pPool->cUsedPages++; /* physical handler registration / pgmPoolTrackFlushGCPhysPTsSlow requirement. */
#ifdef PGMPOOL_WITH_MONITORING
#ifdef PGMPOOL_WITH_USER_TRACKING
return rc3;
#ifdef VBOX_WITH_STATISTICS
LogFlow(("pgmPoolAlloc: returns %Rrc *ppPage=%p:{.Key=%RHp, .idx=%d, .fCached=%RTbool, .fMonitored=%RTbool}\n",
return rc;
return pPage;
#ifdef LOG_ENABLED
switch(enmKind)
case PGMPOOLKIND_INVALID:
case PGMPOOLKIND_FREE:
case PGMPOOLKIND_32BIT_PD:
case PGMPOOLKIND_PAE_PD_PHYS:
case PGMPOOLKIND_PAE_PDPT:
case PGMPOOLKIND_64BIT_PML4:
#ifndef VBOX_WITH_PGMPOOL_PAGING_ONLY
case PGMPOOLKIND_ROOT_PAE_PD:
case PGMPOOLKIND_ROOT_PDPT:
case PGMPOOLKIND_ROOT_NESTED: