/vbox/src/VBox/VMM/VMMR3/ |
H A D | MMHyper.cpp | 409 * @param pvR3 Ring-3 address of the memory. Must be page aligned! 417 VMMR3DECL(int) MMR3HyperMapHCPhys(PVM pVM, void *pvR3, RTR0PTR pvR0, RTHCPHYS HCPhys, size_t cb, argument 420 LogFlow(("MMR3HyperMapHCPhys: pvR3=%p pvR0=%p HCPhys=%RHp cb=%d pszDesc=%p:{%s} pGCPtr=%p\n", 421 pvR3, pvR0, HCPhys, (int)cb, pszDesc, pszDesc, pGCPtr)); 426 AssertReturn(RT_ALIGN_P(pvR3, PAGE_SIZE) == pvR3, VERR_INVALID_PARAMETER); 442 pLookup->u.HCPhys.pvR3 = pvR3; 614 * @param pvR3 The ring-3 address of the memory, must be page aligned. 619 * @param pGCPtr Where to store the GC address corresponding to pvR3 621 MMR3HyperMapPages(PVM pVM, void *pvR3, RTR0PTR pvR0, size_t cPages, PCSUPPAGE paPages, const char *pszDesc, PRTGCPTR pGCPtr) argument 1086 mmR3HyperLookupR3(PVM pVM, void *pvR3) argument 1182 MMR3HyperHCVirt2HCPhys(PVM pVM, void *pvR3) argument [all...] |
H A D | GMM.cpp | 358 Req.pvR3 = NULL; 361 *ppvR3 = Req.pvR3; 382 GMMR3DECL(int) GMMR3SeedChunk(PVM pVM, RTR3PTR pvR3) argument 384 return VMMR3CallR0(pVM, VMMR0_DO_GMM_SEED_CHUNK, (uintptr_t)pvR3, NULL);
|
H A D | VMMSwitcher.cpp | 256 void *pvR3; 264 paBadTries[i].pvR3 = pVM->vmm.s.pvCoreCodeR3; 280 paBadTries[i].pvR3 = pVM->vmm.s.pvCoreCodeR3; 289 LogRel(("VMM: Core code alloc attempt #%d: pvR3=%p pvR0=%p HCPhys=%RHp\n", 290 i, paBadTries[i].pvR3, paBadTries[i].pvR0, paBadTries[i].HCPhys)); 291 SUPR3ContFree(paBadTries[i].pvR3, paBadTries[i].cb >> PAGE_SHIFT);
|
H A D | MM.cpp | 744 *ppv = (char *)pCur->u.Locked.pvR3 + (iPage << PAGE_SHIFT) + off; 753 *ppv = (uint8_t *)pCur->u.HCPhys.pvR3 + pCur->u.HCPhys.HCPhys - HCPhys + off;
|
H A D | PGMPhys.cpp | 1483 pNew->pvR3 = NULL; 2289 pNew->pvR3 = NULL; 2573 pNew->pvR3 = pvPages; 2587 pNew->RamRange.pvR3 = pvPages; 2696 int rc2 = SUPR3PageFreeEx(pCur->pvR3, pCur->RamRange.cb >> PAGE_SHIFT); 3113 int rc = SUPR3PageMapKernel(pCur->pvR3, off, cb, 0, pR0Ptr); 3282 pRamNew->pvR3 = NULL; 4023 Req.pvR3 = NULL; 4148 Req.pvR3 = NULL; 4158 pChunk->pv = Req.pvR3; [all...] |
H A D | PGMSavedState.cpp | 883 uint8_t const *pbPage = (uint8_t const *)pMmio2->pvR3 + iPage * PAGE_SIZE; 920 uint8_t const *pbPage = (uint8_t const *)pMmio2->RamRange.pvR3; 977 uint8_t const *pbPage = (uint8_t const *)pMmio2->RamRange.pvR3; 2466 pRam->GCPhys, pRam->GCPhysLast, pRam->cb, pRam->pvR3 ? "bits" : "nobits", pRam->pszDesc, 2477 pRam->GCPhys, pRam->GCPhysLast, pRam->cb, pRam->pvR3 ? "bits" : "nobits", pRam->pszDesc); 2559 else if (pRam->pvR3) 2567 AssertLogRelMsgReturn(pRam->pvR3, 2571 rc = SSMR3GetMem(pSSM, pRam->pvR3, pRam->cb); 2831 void *pvDstPage = (uint8_t *)pMmio2->RamRange.pvR3 + ((size_t)iPage << PAGE_SHIFT);
|
H A D | PGM.cpp | 2735 pCur->pvR3,
|
/vbox/src/VBox/Runtime/r0drv/os2/ |
H A D | memobj-r0drv-os2.cpp | 408 void *pvR3 = NULL; local 430 rc = KernVMAlloc(pMemToMapOs2->Core.cb, VMDHA_PHYS | VMDHA_PROCESS, &pvR3, (PPVOID)&ulPhys, NULL); 459 if (!pvR3) 467 rc = RTR0Os2DHVMGlobalToProcess(flFlags, pvR0, pMemToMapOs2->Core.cb, &pvR3); 471 Assert(pvR3); 476 PRTR0MEMOBJOS2 pMemOs2 = (PRTR0MEMOBJOS2)rtR0MemObjNew(RT_OFFSETOF(RTR0MEMOBJOS2, Lock), RTR0MEMOBJTYPE_MAPPING, pvR3, pMemToMapOs2->Core.cb); 479 Assert(pMemOs2->Core.pv == pvR3); 484 KernVMFree(pvR3);
|
/vbox/src/VBox/HostDrivers/Support/ |
H A D | SUPDrvIOC.h | 590 RTR3PTR pvR3; member in struct:SUPLOWALLOC::__anon16302::__anon16304 618 RTR3PTR pvR3; member in struct:SUPLOWFREE::__anon16305::__anon16306 660 RTR3PTR pvR3; member in struct:SUPPAGEALLOCEX::__anon16307::__anon16309 695 RTR3PTR pvR3; member in struct:SUPPAGEMAPKERNEL::__anon16310::__anon16311 705 /** The ring-0 address corresponding to pvR3 + offSub. */ 735 RTR3PTR pvR3; member in struct:SUPPAGEPROTECT::__anon16313::__anon16314 768 RTR3PTR pvR3; member in struct:SUPPAGEFREE::__anon16315::__anon16316 794 RTR3PTR pvR3; member in struct:SUPPAGELOCK::__anon16317::__anon16318 825 RTR3PTR pvR3; member in struct:SUPPAGEUNLOCK::__anon16320::__anon16321 857 RTR3PTR pvR3; member in struct:SUPCONTALLOC::__anon16322::__anon16324 883 RTR3PTR pvR3; member in struct:SUPCONTFREE::__anon16325::__anon16326 [all...] |
H A D | SUPLib.cpp | 980 pReq->u.In.pvR3 = pvStart; 1032 Req.u.In.pvR3 = pvStart; 1148 *ppvPages = pReq->u.Out.pvR3; 1159 supR3TouchPages(pReq->u.Out.pvR3, cPages); 1179 SUPR3DECL(int) SUPR3PageMapKernel(void *pvR3, uint32_t off, uint32_t cb, uint32_t fFlags, PRTR0PTR pR0Ptr) argument 1184 AssertPtrReturn(pvR3, VERR_INVALID_POINTER); 1205 Req.u.In.pvR3 = pvR3; 1218 SUPR3DECL(int) SUPR3PageProtect(void *pvR3, RTR0PTR R0Ptr, uint32_t off, uint32_t cb, uint32_t fProt) argument 1223 AssertPtrReturn(pvR3, VERR_INVALID_POINTE [all...] |
H A D | SUPDrv.cpp | 898 Log2(("eType=%d pvR0=%p pvR3=%p cb=%ld\n", pBundle->aMem[i].eType, RTR0MemObjAddress(pBundle->aMem[i].MemObj), 1579 REQ_CHECK_EXPR(SUP_IOCTL_PAGE_LOCK, pReq->u.In.pvR3 >= PAGE_SIZE); 1582 pReq->Hdr.rc = SUPR0LockMem(pSession, pReq->u.In.pvR3, pReq->u.In.cPages, &pReq->u.Out.aPages[0]); 1595 pReq->Hdr.rc = SUPR0UnlockMem(pSession, pReq->u.In.pvR3); 1606 pReq->Hdr.rc = SUPR0ContAlloc(pSession, pReq->u.In.cPages, &pReq->u.Out.pvR0, &pReq->u.Out.pvR3, &pReq->u.Out.HCPhys); 1619 pReq->Hdr.rc = SUPR0ContFree(pSession, (RTHCUINTPTR)pReq->u.In.pvR3); 1810 pReq->Hdr.rc = SUPR0LowAlloc(pSession, pReq->u.In.cPages, &pReq->u.Out.pvR0, &pReq->u.Out.pvR3, &pReq->u.Out.aPages[0]); 1823 pReq->Hdr.rc = SUPR0LowFree(pSession, (RTHCUINTPTR)pReq->u.In.pvR3); 1881 pReq->u.In.fUserMapping ? &pReq->u.Out.pvR3 : NULL, 1900 pReq->Hdr.rc = SUPR0PageMapKernel(pSession, pReq->u.In.pvR3, pRe 2944 SUPR0LockMem(PSUPDRVSESSION pSession, RTR3PTR pvR3, uint32_t cPages, PRTHCPHYS paPages) argument 2971 AssertMsg(RTR0MemObjAddressR3(Mem.MemObj) == pvR3, ("%p == %p\\n", RTR0MemObjAddressR3(Mem.MemObj), pvR3)); local 3004 SUPR0UnlockMem(PSUPDRVSESSION pSession, RTR3PTR pvR3) argument 3412 SUPR0PageMapKernel(PSUPDRVSESSION pSession, RTR3PTR pvR3, uint32_t offSub, uint32_t cbSub, uint32_t fFlags, PRTR0PTR ppvR0) argument 3498 SUPR0PageProtect(PSUPDRVSESSION pSession, RTR3PTR pvR3, RTR0PTR pvR0, uint32_t offSub, uint32_t cbSub, uint32_t fProt) argument 3581 SUPR0PageFree(PSUPDRVSESSION pSession, RTR3PTR pvR3) argument [all...] |
/vbox/src/VBox/VMM/VMMAll/ |
H A D | MMAll.cpp | 53 const RTR3UINTPTR off = (RTR3UINTPTR)R3Ptr - (RTR3UINTPTR)pLookup->u.Locked.pvR3; 64 const RTR3UINTPTR off = (RTR3UINTPTR)R3Ptr - (RTR3UINTPTR)pLookup->u.HCPhys.pvR3; 234 return (RTR3PTR)((RTR3UINTPTR)pLookup->u.Locked.pvR3 + off); 236 return (RTR3PTR)((RTR3UINTPTR)pLookup->u.HCPhys.pvR3 + off);
|
H A D | PGMAllPhys.cpp | 1047 *ppv = (uint8_t *)pMmio2Range->RamRange.pvR3 + ((uintptr_t)iPage << PAGE_SHIFT);
|
/vbox/include/VBox/ |
H A D | sup.h | 1260 * @param pvR3 The address SUPR3PageAllocEx return. 1267 SUPR3DECL(int) SUPR3PageMapKernel(void *pvR3, uint32_t off, uint32_t cb, uint32_t fFlags, PRTR0PTR pR0Ptr); 1276 * @param pvR3 The ring-3 address SUPR3PageAllocEx returned. 1289 SUPR3DECL(int) SUPR3PageProtect(void *pvR3, RTR0PTR R0Ptr, uint32_t off, uint32_t cb, uint32_t fProt); 1810 SUPR0DECL(int) SUPR0LockMem(PSUPDRVSESSION pSession, RTR3PTR pvR3, uint32_t cPages, PRTHCPHYS paPages); 1811 SUPR0DECL(int) SUPR0UnlockMem(PSUPDRVSESSION pSession, RTR3PTR pvR3); 1820 SUPR0DECL(int) SUPR0PageMapKernel(PSUPDRVSESSION pSession, RTR3PTR pvR3, uint32_t offSub, uint32_t cbSub, uint32_t fFlags, PRTR0PTR ppvR0); 1821 SUPR0DECL(int) SUPR0PageProtect(PSUPDRVSESSION pSession, RTR3PTR pvR3, RTR0PTR pvR0, uint32_t offSub, uint32_t cbSub, uint32_t fProt); 1822 SUPR0DECL(int) SUPR0PageFree(PSUPDRVSESSION pSession, RTR3PTR pvR3);
|
/vbox/include/VBox/vmm/ |
H A D | gmm.h | 411 GMMR0DECL(int) GMMR0SeedChunk(PVM pVM, VMCPUID idCpu, RTR3PTR pvR3); 560 RTR3PTR pvR3; member in struct:GMMMAPUNMAPCHUNKREQ 787 GMMR3DECL(int) GMMR3SeedChunk(PVM pVM, RTR3PTR pvR3);
|
H A D | mm.h | 272 VMMR3DECL(int) MMR3HyperMapHCPhys(PVM pVM, void *pvR3, RTR0PTR pvR0, RTHCPHYS HCPhys, size_t cb, const char *pszDesc, PRTGCPTR pGCPtr); 275 VMMR3DECL(int) MMR3HyperMapPages(PVM pVM, void *pvR3, RTR0PTR pvR0, size_t cPages, PCSUPPAGE paPages, const char *pszDesc, PRTGCPTR pGCPtr);
|
/vbox/src/VBox/VMM/include/ |
H A D | MMInternal.h | 646 R3PTRTYPE(void *) pvR3; member in struct:MMLOOKUPHYPER::__anon16923::__anon16924 657 R3PTRTYPE(void *) pvR3; member in struct:MMLOOKUPHYPER::__anon16923::__anon16925 660 /** HC physical address corresponding to pvR3/pvR0. */
|
H A D | PGMInternal.h | 1393 R3PTRTYPE(void *) pvR3; member in struct:PGMRAMRANGE 1603 RTR3PTR pvR3; member in struct:PGMMMIO2RANGE
|
/vbox/src/VBox/VMM/VMMR0/ |
H A D | GMMR0.cpp | 4218 return GMMR0MapUnmapChunk(pVM, pReq->idChunkMap, pReq->idChunkUnmap, &pReq->pvR3); 4231 * @param pvR3 Pointer to the chunk size memory block to lock down. 4233 GMMR0DECL(int) GMMR0SeedChunk(PVM pVM, VMCPUID idCpu, RTR3PTR pvR3) 4245 AssertPtrReturn(pvR3, VERR_INVALID_POINTER); 4246 AssertReturn(!(PAGE_OFFSET_MASK & pvR3), VERR_INVALID_POINTER); 4259 rc = RTR0MemObjLockUser(&MemObj, pvR3, GMM_CHUNK_SIZE, RTMEM_PROT_READ | RTMEM_PROT_WRITE, NIL_RTR0PROCESS); 4269 LogFlow(("GMMR0SeedChunk: rc=%d (pvR3=%p)\n", rc, pvR3));
|
/vbox/include/iprt/ |
H A D | asm.h | 582 * @param pvR3 The pointer value to assign to *ppv. 584 DECLINLINE(RTR3PTR) ASMAtomicXchgR3Ptr(RTR3PTR volatile *ppvR3, RTR3PTR pvR3) argument 587 return (RTR3PTR)ASMAtomicXchgU32((volatile uint32_t *)(void *)ppvR3, (uint32_t)pvR3); 589 return (RTR3PTR)ASMAtomicXchgU64((volatile uint64_t *)(void *)ppvR3, (uint64_t)pvR3);
|
/vbox/src/VBox/VMM/testcase/ |
H A D | tstVMStruct.h | 380 GEN_CHECK_OFF_DOT(MMLOOKUPHYPER, u.Locked.pvR3); 383 GEN_CHECK_OFF_DOT(MMLOOKUPHYPER, u.HCPhys.pvR3); 813 GEN_CHECK_OFF(PGMRAMRANGE, pvR3);
|