/vbox/src/VBox/Runtime/r0drv/os2/ |
H A D | the-os2-kernel.h | 51 DECLASM(int) RTR0Os2DHVMGlobalToProcess(ULONG fFlags, PVOID pvR0, ULONG cb, PPVOID ppvR3);
|
H A D | memobj-r0drv-os2.cpp | 334 void *pvR0 = NULL; local 344 pvR0 = pMemToMapOs2->Core.pv; 348 pvR0 = pMemToMapOs2->Core.pv; 349 if (!pvR0) 355 rc = KernVMAlloc(pMemToMapOs2->Core.cb, VMDHA_PHYS, &pvR0, (PPVOID)&ulPhys, NULL); 358 pMemToMapOs2->Core.pv = pvR0; 370 pvR0 = pMemToMapOs2->Core.pv; 388 PRTR0MEMOBJOS2 pMemOs2 = (PRTR0MEMOBJOS2)rtR0MemObjNew(RT_OFFSETOF(RTR0MEMOBJOS2, Lock), RTR0MEMOBJTYPE_MAPPING, pvR0, pMemToMapOs2->Core.cb); 407 void *pvR0; local 418 pvR0 [all...] |
H A D | RTR0Os2DHVMGlobalToProcess.asm | 60 ; @param pvR0 [ebp + 0ch] Ring-0 memory. 76 mov ebx, [ebp + 0ch] ; pvR0
|
/vbox/src/VBox/VMM/VMMR3/ |
H A D | MMHyper.cpp | 410 * @param pvR0 Optional ring-0 address of the memory. 417 VMMR3DECL(int) MMR3HyperMapHCPhys(PVM pVM, void *pvR3, RTR0PTR pvR0, RTHCPHYS HCPhys, size_t cb, argument 420 LogFlow(("MMR3HyperMapHCPhys: pvR3=%p pvR0=%p HCPhys=%RHp cb=%d pszDesc=%p:{%s} pGCPtr=%p\n", 421 pvR3, pvR0, HCPhys, (int)cb, pszDesc, pszDesc, pGCPtr)); 427 AssertReturn(RT_ALIGN_T(pvR0, PAGE_SIZE, RTR0PTR) == pvR0, VERR_INVALID_PARAMETER); 443 pLookup->u.HCPhys.pvR0 = pvR0; 615 * @param pvR0 The ring-0 address of the memory, must be page aligned. (optional) 621 VMMR3DECL(int) MMR3HyperMapPages(PVM pVM, void *pvR3, RTR0PTR pvR0, size_ argument 814 RTR0PTR pvR0 = NIL_RTR0PTR; local 1010 RTR0PTR pvR0 = NIL_RTR0PTR; local [all...] |
H A D | MMUkHeap.cpp | 218 int rc = SUPR3PageAllocEx(pSubHeap->cb >> PAGE_SHIFT, 0, &pSubHeap->pv, &pSubHeap->pvR0, NULL); 369 *pR0Ptr = (uintptr_t)pv - (uintptr_t)pSubHeap->pv + pSubHeap->pvR0;
|
H A D | VMMSwitcher.cpp | 255 RTR0PTR pvR0; 265 paBadTries[i].pvR0 = pVM->vmm.s.pvCoreCodeR0; 281 paBadTries[i].pvR0 = pVM->vmm.s.pvCoreCodeR0; 289 LogRel(("VMM: Core code alloc attempt #%d: pvR3=%p pvR0=%p HCPhys=%RHp\n", 290 i, paBadTries[i].pvR3, paBadTries[i].pvR0, paBadTries[i].HCPhys));
|
/vbox/src/VBox/VMM/VMMAll/ |
H A D | MMAll.cpp | 115 const RTR0UINTPTR off = (RTR0UINTPTR)R0Ptr - (RTR0UINTPTR)pLookup->u.Locked.pvR0; 116 if (off < pLookup->cb && pLookup->u.Locked.pvR0) 126 const RTR0UINTPTR off = (RTR0UINTPTR)R0Ptr - (RTR0UINTPTR)pLookup->u.HCPhys.pvR0; 127 if (off < pLookup->cb && pLookup->u.HCPhys.pvR0) 257 if (pLookup->u.Locked.pvR0) 258 return (RTR0PTR)((RTR0UINTPTR)pLookup->u.Locked.pvR0 + off); 267 if (pLookup->u.HCPhys.pvR0) 268 return (RTR0PTR)((RTR0UINTPTR)pLookup->u.HCPhys.pvR0 + off);
|
/vbox/src/VBox/VMM/testcase/ |
H A D | tstMMHyperHeap.cpp | 55 RTR0PTR pvR0; local 59 rc = SUPR3LowAlloc(RT_ELEMENTS(aPages), (void **)&pVM, &pvR0, &aPages[0]); 67 pVM->pVMR0 = pvR0;
|
H A D | tstVMStruct.h | 381 GEN_CHECK_OFF_DOT(MMLOOKUPHYPER, u.Locked.pvR0);
|
/vbox/src/VBox/VMM/include/ |
H A D | MMInternal.h | 193 RTR0PTR pvR0; member in struct:MMUKHEAPSUB 648 RTR0PTR pvR0; member in struct:MMLOOKUPHYPER::__anon16923::__anon16924 659 RTR0PTR pvR0; member in struct:MMLOOKUPHYPER::__anon16923::__anon16925 660 /** HC physical address corresponding to pvR3/pvR0. */
|
/vbox/src/VBox/HostDrivers/Support/ |
H A D | SUPDrvIOC.h | 592 RTR0PTR pvR0; member in struct:SUPLOWALLOC::__anon16302::__anon16304 662 RTR0PTR pvR0; member in struct:SUPPAGEALLOCEX::__anon16307::__anon16309 706 RTR0PTR pvR0; member in struct:SUPPAGEMAPKERNEL::__anon16310::__anon16312 738 RTR0PTR pvR0; member in struct:SUPPAGEPROTECT::__anon16313::__anon16314 855 RTR0PTR pvR0; member in struct:SUPCONTALLOC::__anon16322::__anon16324
|
H A D | SUPLib.cpp | 1150 *pR0Ptr = pReq->u.Out.pvR0; 1213 *pR0Ptr = Req.u.Out.pvR0; 1249 Req.u.In.pvR0 = R0Ptr; 1345 *pR0Ptr = Req.u.Out.pvR0; 1437 *ppvPagesR0 = pReq->u.Out.pvR0;
|
H A D | SUPDrv.cpp | 898 Log2(("eType=%d pvR0=%p pvR3=%p cb=%ld\n", pBundle->aMem[i].eType, RTR0MemObjAddress(pBundle->aMem[i].MemObj), 1606 pReq->Hdr.rc = SUPR0ContAlloc(pSession, pReq->u.In.cPages, &pReq->u.Out.pvR0, &pReq->u.Out.pvR3, &pReq->u.Out.HCPhys); 1810 pReq->Hdr.rc = SUPR0LowAlloc(pSession, pReq->u.In.cPages, &pReq->u.Out.pvR0, &pReq->u.Out.pvR3, &pReq->u.Out.aPages[0]); 1882 pReq->u.In.fKernelMapping ? &pReq->u.Out.pvR0 : NULL, 1901 pReq->u.In.fFlags, &pReq->u.Out.pvR0); 1919 pReq->Hdr.rc = SUPR0PageProtect(pSession, pReq->u.In.pvR3, pReq->u.In.pvR0, pReq->u.In.offSub, pReq->u.In.cbSub, pReq->u.In.fProt); 3492 * @param pvR0 The ring-0 address returned by SUPR0PageAllocEx. 3498 SUPR0DECL(int) SUPR0PageProtect(PSUPDRVSESSION pSession, RTR3PTR pvR3, RTR0PTR pvR0, uint32_t offSub, uint32_t cbSub, uint32_t fProt) argument 3504 LogFlow(("SUPR0PageProtect: pSession=%p pvR3=%p pvR0=%p offSub=%#x cbSub=%#x fProt-%#x\n", pSession, pvR3, pvR0, offSu [all...] |
/vbox/include/VBox/vmm/ |
H A D | mm.h | 272 VMMR3DECL(int) MMR3HyperMapHCPhys(PVM pVM, void *pvR3, RTR0PTR pvR0, RTHCPHYS HCPhys, size_t cb, const char *pszDesc, PRTGCPTR pGCPtr); 275 VMMR3DECL(int) MMR3HyperMapPages(PVM pVM, void *pvR3, RTR0PTR pvR0, size_t cPages, PCSUPPAGE paPages, const char *pszDesc, PRTGCPTR pGCPtr);
|
/vbox/src/VBox/VMM/VMMR0/ |
H A D | HMR0.cpp | 953 void *pvR0 = RTR0MemObjAddress(g_HvmR0.aCpuInfo[i].hMemObj); Assert(pvR0); local 954 ASMMemZeroPage(pvR0);
|
/vbox/include/iprt/ |
H A D | asm.h | 563 * @param pvR0 The pointer value to assign to *ppv. 565 DECLINLINE(RTR0PTR) ASMAtomicXchgR0Ptr(RTR0PTR volatile *ppvR0, RTR0PTR pvR0) argument 568 return (RTR0PTR)ASMAtomicXchgU32((volatile uint32_t *)(void *)ppvR0, (uint32_t)pvR0); 570 return (RTR0PTR)ASMAtomicXchgU64((volatile uint64_t *)(void *)ppvR0, (uint64_t)pvR0);
|
/vbox/include/VBox/ |
H A D | sup.h | 1821 SUPR0DECL(int) SUPR0PageProtect(PSUPDRVSESSION pSession, RTR3PTR pvR3, RTR0PTR pvR0, uint32_t offSub, uint32_t cbSub, uint32_t fProt);
|