/vbox/src/VBox/VMM/VMMR3/ |
H A D | IEMR3.cpp | 46 PVMCPU pVCpu = &pVM->aCpus[idCpu]; 47 pVCpu->iem.s.offVM = -RT_OFFSETOF(VM, aCpus[idCpu].iem.s); 78 pVCpu->iem.s.enmCpuVendor = pVM->aCpus[0].iem.s.enmCpuVendor; 79 pVCpu->iem.s.enmHostCpuVendor = pVM->aCpus[0].iem.s.enmHostCpuVendor; 103 pVM->aCpus[idCpu].iem.s.pCtxRC = VM_RC_ADDR(pVM, pVM->aCpus[idCpu].iem.s.pCtxR3);
|
H A D | DBGFR3Trace.cpp | 306 pVM->aCpus[iCpu].fTraceGroups = UINT32_MAX; 309 pVM->aCpus[iCpu].fTraceGroups = 0; 328 pVM->aCpus[iCpu].fTraceGroups |= g_aVmmTpGroups[i].fMask; 331 pVM->aCpus[iCpu].fTraceGroups &= ~g_aVmmTpGroups[i].fMask; 376 uint32_t const fTraceGroups = pVM->aCpus[0].fTraceGroups;
|
H A D | GIMKvm.cpp | 148 VMMHypercallsEnable(&pVM->aCpus[i]); 248 PGIMKVMCPU pKvmCpu = &pVM->aCpus[iCpu].gim.s.u.KvmCpu; 275 PCGIMKVMCPU pcKvmCpu = &pVM->aCpus[i].gim.s.u.KvmCpu; 328 PVMCPU pVCpu = &pVM->aCpus[i];
|
H A D | TRPM.cpp | 470 pVM->trpm.s.offTRPMCPU = RT_OFFSETOF(VM, aCpus[0].trpm) - RT_OFFSETOF(VM, trpm); 474 PVMCPU pVCpu = &pVM->aCpus[i]; 476 pVCpu->trpm.s.offVM = RT_OFFSETOF(VM, aCpus[i].trpm); 594 PVMCPU pVCpu = &pVM->aCpus[0]; /* raw mode implies on VCPU */ 618 PVMCPU pVCpu = &pVM->aCpus[0]; 789 TRPMR3ResetCpu(&pVM->aCpus[i]); 800 PVMCPU pVCpu = &pVM->aCpus[0]; /* raw mode implies on VCPU */ 825 *pRCPtrValue = VM_RC_ADDR(pVM, &pVM->aCpus[0].trpm); 870 PTRPMCPU pTrpmCpu = &pVM->aCpus[i].trpm.s; 882 PVMCPU pVCpu = &pVM->aCpus[ [all...] |
H A D | VM.cpp | 473 PUVM pUVM = (PUVM)RTMemPageAllocZ(RT_OFFSETOF(UVM, aCpus[cCpus])); 492 pUVM->aCpus[i].pUVM = pUVM; 493 pUVM->aCpus[i].idCpu = i; 503 pUVM->aCpus[i].vm.s.EventSemWait = NIL_RTSEMEVENT; 506 rc = RTSemEventCreate(&pUVM->aCpus[i].vm.s.EventSemWait); 535 rc = RTThreadCreateF(&pUVM->aCpus[i].vm.s.ThreadEMT, vmR3EmulationThread, &pUVM->aCpus[i], 541 pUVM->aCpus[i].vm.s.NativeThreadEMT = RTThreadGetNative(pUVM->aCpus[i].vm.s.ThreadEMT); 568 RTSemEventDestroy(pUVM->aCpus[ [all...] |
H A D | VMM.cpp | 161 AssertCompile(sizeof(pVM->aCpus[0].vmm.s) <= sizeof(pVM->aCpus[0].vmm.padding)); 301 PVMCPU pVCpu = &pVM->aCpus[idCpu]; 364 PVMCPU pVCpu = &pVM->aCpus[i]; 474 STAMR3RegisterF(pVM, &pVM->aCpus[i].vmm.s.CallRing3JmpBufR0.cbUsedMax, STAMTYPE_U32_RESET, STAMVISIBILITY_ALWAYS, STAMUNIT_BYTES, "Max amount of stack used.", "/VMM/Stack/CPU%u/Max", i); 475 STAMR3RegisterF(pVM, &pVM->aCpus[i].vmm.s.CallRing3JmpBufR0.cbUsedAvg, STAMTYPE_U32, STAMVISIBILITY_ALWAYS, STAMUNIT_BYTES, "Average stack usage.", "/VMM/Stack/CPU%u/Avg", i); 476 STAMR3RegisterF(pVM, &pVM->aCpus[i].vmm.s.CallRing3JmpBufR0.cUsedTotal, STAMTYPE_U64, STAMVISIBILITY_ALWAYS, STAMUNIT_OCCURENCES, "Number of stack usages.", "/VMM/Stack/CPU%u/Uses", i); 543 if (pVM->aCpus[0].vmm.s.hR0ThreadCtx != NIL_RTTHREADCTX) 658 rc = PGMMapSetPage(pVM, pVM->aCpus[idCpu].vmm.s.pbEMTStackRC, VMM_STACK_SIZE, 688 uint8_t *pbEMTStackR3 = pVM->aCpus[idCp [all...] |
H A D | TM.cpp | 755 STAMR3RegisterF(pVM, &pVM->aCpus[i].tm.s.offTSCRawSrc, STAMTYPE_U64, STAMVISIBILITY_ALWAYS, STAMUNIT_TICKS, "TSC offset relative the raw source", "/TM/TSC/offCPU%u", i); 758 STAMR3RegisterF(pVM, &pVM->aCpus[i].tm.s.StatNsTotal, STAMTYPE_COUNTER, STAMVISIBILITY_ALWAYS, STAMUNIT_NS, "Resettable: Total CPU run time.", "/TM/CPU/%02u", i); 759 STAMR3RegisterF(pVM, &pVM->aCpus[i].tm.s.StatNsExecuting, STAMTYPE_PROFILE, STAMVISIBILITY_ALWAYS, STAMUNIT_NS_PER_OCCURENCE, "Resettable: Time spent executing guest code.", "/TM/CPU/%02u/PrfExecuting", i); 760 STAMR3RegisterF(pVM, &pVM->aCpus[i].tm.s.StatNsExecLong, STAMTYPE_PROFILE, STAMVISIBILITY_ALWAYS, STAMUNIT_NS_PER_OCCURENCE, "Resettable: Time spent executing guest code - long hauls.", "/TM/CPU/%02u/PrfExecLong", i); 761 STAMR3RegisterF(pVM, &pVM->aCpus[i].tm.s.StatNsExecShort, STAMTYPE_PROFILE, STAMVISIBILITY_ALWAYS, STAMUNIT_NS_PER_OCCURENCE, "Resettable: Time spent executing guest code - short stretches.", "/TM/CPU/%02u/PrfExecShort", i); 762 STAMR3RegisterF(pVM, &pVM->aCpus[i].tm.s.StatNsExecTiny, STAMTYPE_PROFILE, STAMVISIBILITY_ALWAYS, STAMUNIT_NS_PER_OCCURENCE, "Resettable: Time spent executing guest code - tiny bits.", "/TM/CPU/%02u/PrfExecTiny", i); 763 STAMR3RegisterF(pVM, &pVM->aCpus[i].tm.s.StatNsHalted, STAMTYPE_PROFILE, STAMVISIBILITY_ALWAYS, STAMUNIT_NS_PER_OCCURENCE, "Resettable: Time spent halted.", "/TM/CPU/%02u/PrfHalted", i); 764 STAMR3RegisterF(pVM, &pVM->aCpus[i].tm.s.StatNsOther, STAMTYPE_PROFILE, STAMVISIBILITY_ALWAYS, STAMUNIT_NS_PER_OCCURENCE, "Resettable: Time spent in the VMM or preempted.", "/TM/CPU/%02u/PrfOther", i); 766 STAMR3RegisterF(pVM, &pVM->aCpus[i].tm.s.cNsTotal, STAMTYPE_U64, STAMVISIBILITY_ALWAYS, STAMUNIT_NS, "Total CPU run time.", "/TM/CPU/%02u/cNsTotal", i); 767 STAMR3RegisterF(pVM, &pVM->aCpus[ [all...] |
H A D | CPUM.cpp | 560 pVM->aCpus[i].cpum.s.fUseFlags |= CPUM_USE_FFXSR_LEAKY; 587 AssertCompileMemberAlignment(VM, aCpus, 64); 589 AssertCompileMemberSizeAlignment(VM, aCpus[0].cpum.s, 64); 600 pVM->cpum.s.offCPUMCPU0 = RT_OFFSETOF(VM, aCpus[0].cpum) - RT_OFFSETOF(VM, cpum); 601 Assert((uintptr_t)&pVM->cpum + pVM->cpum.s.offCPUMCPU0 == (uintptr_t)&pVM->aCpus[0].cpum); 607 PVMCPU pVCpu = &pVM->aCpus[i]; 609 pVCpu->cpum.s.offCPUM = RT_OFFSETOF(VM, aCpus[i].cpum) - RT_OFFSETOF(VM, cpum); 679 PVMCPU pVCpu = &pVM->aCpus[i]; 761 PVMCPU pVCpu = &pVM->aCpus[iCpu]; 810 PVMCPU pVCpu = &pVM->aCpus[ [all...] |
H A D | VMReq.cpp | 658 RTSemEventAddSignaller(pReq->EventSem, pUVM->aCpus[idCpu].vm.s.ThreadEMT); 705 RTSemEventAddSignaller(pReq->EventSem, pUVM->aCpus[idCpu].vm.s.ThreadEMT); 867 PVMCPU pVCpu = &pUVM->pVM->aCpus[idTarget]; 871 pUVCpu = &pUVM->aCpus[idTarget]; 1026 VMCPU_FF_SET(&pUVM->pVM->aCpus[idDstCpu], VMCPU_FF_REQUEST); 1121 Assert(pUVM->aCpus[idDstCpu].vm.s.NativeThreadEMT == RTThreadNativeSelf()); 1122 ppPriorityReqs = &pUVM->aCpus[idDstCpu].vm.s.pPriorityReqs; 1123 ppNormalReqs = !fPriorityOnly ? &pUVM->aCpus[idDstCpu].vm.s.pNormalReqs : ppPriorityReqs; 1146 VMCPU_FF_CLEAR(&pUVM->pVM->aCpus[idDstCpu], VMCPU_FF_REQUEST);
|
H A D | PDMQueue.cpp | 216 VMCPU_ASSERT_EMT(&pVM->aCpus[0]); 266 VMCPU_ASSERT_EMT(&pVM->aCpus[0]); 312 VMCPU_ASSERT_EMT(&pVM->aCpus[0]); 356 VMCPU_ASSERT_EMT(&pVM->aCpus[0]);
|
H A D | GIM.cpp | 249 rc = SSMR3PutXYZ(pSSM, pVM->aCpus[i].gim.s.XYZ); 303 rc = SSMR3PutXYZ(pSSM, pVM->aCpus[i].gim.s.XYZ);
|
H A D | PGM.cpp | 1228 AssertCompile(sizeof(pVM->aCpus[0].pgm.s) <= sizeof(pVM->aCpus[0].pgm.padding)); 1254 PVMCPU pVCpu = &pVM->aCpus[idCpu]; 1360 pVM->aCpus[iCpu].pgm.s.pStatsR3 = (PGMCPUSTATS *)pv; 1361 pVM->aCpus[iCpu].pgm.s.pStatsR0 = MMHyperCCToR0(pVM, pv); 1362 pVM->aCpus[iCpu].pgm.s.pStatsRC = MMHyperCCToRC(pVM, pv); 1446 PVMCPU pVCpu = &pVM->aCpus[i]; 1518 PVMCPU pVCpu = &pVM->aCpus[i]; 1848 PPGMCPU pPgmCpu = &pVM->aCpus[idCpu].pgm.s; 1861 PGMCPUSTATS *pCpuStats = pVM->aCpus[idCp [all...] |
H A D | PGMMap.cpp | 211 PVMCPU pVCpu = &pVM->aCpus[i]; 270 PVMCPU pVCpu = &pVM->aCpus[i]; 526 PVMCPU pVCpu = &pVM->aCpus[0]; 595 PVMCPU pVCpu = &pVM->aCpus[0]; 678 pVM->aCpus[idCpu].pgm.s.fSyncFlags &= ~PGM_SYNC_MONITOR_CR3; 679 VMCPU_FF_SET(&pVM->aCpus[idCpu], VMCPU_FF_PGM_SYNC_CR3); 713 VMCPU_FF_SET(&pVM->aCpus[i], VMCPU_FF_PGM_SYNC_CR3);
|
/vbox/include/VBox/vmm/ |
H A D | gvm.h | 82 /** Number of Virtual CPUs, i.e. how many entries there are in aCpus. 116 GVMCPU aCpus[1]; member in struct:GVM
|
H A D | uvm.h | 149 UVMCPU aCpus[1]; member in struct:UVM 155 AssertCompileMemberAlignment(UVM, aCpus, 32);
|
/vbox/src/VBox/VMM/VMMAll/ |
H A D | VMMAll.cpp | 215 PVMCPU pVCpu = &pVM->aCpus[idCpu]; 228 PVMCPU pVCpu = &pVM->aCpus[idCpu]; 257 return &pVM->aCpus[idCpu]; 261 return &pVM->aCpus[0]; 277 PVMCPU pVCpu = &pVM->aCpus[idCpu]; 291 PVMCPU pVCpu = &pVM->aCpus[idCpu]; 299 return &pVM->aCpus[0]; 314 return &pVM->aCpus[0]; 330 return &pVM->aCpus[idCpu];
|
H A D | HMAll.cpp | 218 STAM_COUNTER_INC(&pVM->aCpus[idCurCpu].hm.s.StatFlushPage); 222 PVMCPU pVCpu = &pVM->aCpus[idCpu]; 251 return HMFlushTLB(&pVM->aCpus[0]); 255 STAM_COUNTER_INC(&pVM->aCpus[idThisCpu].hm.s.StatFlushTlb); 259 PVMCPU pVCpu = &pVM->aCpus[idCpu]; 381 PVMCPU pVCpu = &pVM->aCpus[idCpu];
|
H A D | GIMAllKvm.cpp | 90 PVMCPU pVCpuTarget = &pVM->aCpus[uHyperArg1]; /** ASSUMES pVCpu index == ApicId of the VCPU. */ 146 PVMCPU pVCpu = &pVM->aCpus[i];
|
/vbox/src/VBox/VMM/testcase/ |
H A D | tstVMStructSize.cpp | 271 CHECK_MEMBER_ALIGNMENT(VM, aCpus[0], PAGE_SIZE); 272 CHECK_MEMBER_ALIGNMENT(VM, aCpus[1], PAGE_SIZE); 273 CHECK_MEMBER_ALIGNMENT(VM, aCpus[0].cpum.s.Host, 64); 274 CHECK_MEMBER_ALIGNMENT(VM, aCpus[0].cpum.s.Guest, 64); 275 CHECK_MEMBER_ALIGNMENT(VM, aCpus[1].cpum.s.Host, 64); 276 CHECK_MEMBER_ALIGNMENT(VM, aCpus[1].cpum.s.Guest, 64); 277 CHECK_MEMBER_ALIGNMENT(VM, aCpus[0].cpum.s.Hyper, 64); 278 CHECK_MEMBER_ALIGNMENT(VM, aCpus[1].cpum.s.Hyper, 64); 280 CHECK_MEMBER_ALIGNMENT(VM, aCpus[0].cpum.s.pvApicBase, 8); 296 CHECK_MEMBER_ALIGNMENT(VM, aCpus, 409 [all...] |
H A D | tstPDMAsyncCompletion.cpp | 99 RTTlsSet(pVM->pUVM->vm.s.idxTLS, &pVM->pUVM->aCpus[0]); 100 pVM->pUVM->aCpus[0].pUVM = pVM->pUVM; 101 pVM->pUVM->aCpus[0].vm.s.NativeThreadEMT = RTThreadNativeSelf();
|
H A D | tstSSM.cpp | 639 int rc = RTTlsSet(pUVM->vm.s.idxTLS, &pUVM->aCpus[0]); 642 pUVM->aCpus[0].pUVM = pUVM; 643 pUVM->aCpus[0].vm.s.NativeThreadEMT = RTThreadNativeSelf(); 662 pVM->aCpus[0].pVMR3 = pVM; 663 pVM->aCpus[0].hNativeThread = RTThreadNativeSelf();
|
/vbox/src/VBox/VMM/VMMR0/ |
H A D | GVMMR0.cpp | 807 PGVM pGVM = (PGVM)RTMemAllocZ(RT_UOFFSETOF(GVM, aCpus[cCpus])); 821 const uint32_t cbVM = RT_UOFFSETOF(VM, aCpus[cCpus]); 835 pVM->offVMCPU = RT_UOFFSETOF(VM, aCpus); 838 AssertCompileMemberAlignment(VM, aCpus, PAGE_SIZE); 864 pVM->aCpus[i].pVMR0 = pVM; 865 pVM->aCpus[i].pVMR3 = pVM->pVMR3; 866 pVM->aCpus[i].idHostCpu = NIL_RTCPUID; 867 pVM->aCpus[i].hNativeThreadR0 = NIL_RTNATIVETHREAD; 887 pGVM->aCpus[0].hEMT = hEMT0; 888 pVM->aCpus[ [all...] |
H A D | PGMR0SharedPage.cpp | 51 PVMCPU pVCpu = &pVM->aCpus[idCpu]; 164 CPUMSetChangedFlags(&pVM->aCpus[idCurCpu], CPUM_CHANGED_GLOBAL_TLB_FLUSH);
|
H A D | VMMR0.cpp | 307 PVMCPU pVCpu = &pVM->aCpus[0]; 432 /** @todo I wish to call PGMR0PhysFlushHandyPages(pVM, &pVM->aCpus[idCpu]) 855 PVMCPU pVCpu = &pVM->aCpus[idCpu]; 1324 PVMCPU pVCpu = &pVM->aCpus[idCpu]; 1402 return PGMR0PhysAllocateHandyPages(pVM, &pVM->aCpus[idCpu]); 1407 return PGMR0PhysFlushHandyPages(pVM, &pVM->aCpus[idCpu]); 1412 return PGMR0PhysAllocateLargeHandyPage(pVM, &pVM->aCpus[idCpu]); 1505 PVMCPU pVCpu = &pVM->aCpus[idCpu]; 1747 PVMCPU pVCpu = &pVM->aCpus[idCpu];
|
/vbox/src/VBox/VMM/VMMRZ/ |
H A D | PGMRZDynMap.cpp | 349 PPGMMAPSET pSet = &pVM->aCpus[idCpu].pgm.s.AutoSet; 553 PPGMMAPSET pSet = &pVM->aCpus[idCpu].pgm.s.AutoSet; 2481 PPGMMAPSET pSet = &pVM->aCpus[0].pgm.s.AutoSet; 2503 PGMRZDynMapStartAutoSet(&pVM->aCpus[0]); 2524 PGMR0DynMapMigrateAutoSet(&pVM->aCpus[0]); 2562 PGMR0DynMapMigrateAutoSet(&pVM->aCpus[0]); 2590 PGMR0DynMapMigrateAutoSet(&pVM->aCpus[0]); 2605 PGMR0DynMapMigrateAutoSet(&pVM->aCpus[0]); 2606 PGMRZDynMapReleaseAutoSet(&pVM->aCpus[0]); 2607 PGMRZDynMapStartAutoSet(&pVM->aCpus[ [all...] |