Searched refs:offVirtualSyncGivenUp (Results 1 - 5 of 5) sorted by relevance
/vbox/src/VBox/VMM/VMMAll/ |
H A D | TMAllVirtual.cpp | 331 if (off > u64Sub + pVM->tm.s.offVirtualSyncGivenUp) 334 Log4(("TM: %'RU64/-%'8RU64: sub %RU32 [vsghcul]\n", u64 - off, off - pVM->tm.s.offVirtualSyncGivenUp, u64Sub)); 340 off = pVM->tm.s.offVirtualSyncGivenUp; 397 Log4(("TM: %'RU64/-%'8RU64: exp tmr=>ff [vsghcul]\n", u64, pVM->tm.s.offVirtualSync - pVM->tm.s.offVirtualSyncGivenUp)); 485 Log4(("TM: %'RU64/-%'8RU64: exp tmr=>ff [vsgl]\n", u64, pVM->tm.s.offVirtualSync - pVM->tm.s.offVirtualSyncGivenUp)); 652 uint64_t const offGivenUp = ASMAtomicReadU64(&pVM->tm.s.offVirtualSyncGivenUp); 655 && offGivenUp == ASMAtomicReadU64(&pVM->tm.s.offVirtualSyncGivenUp) 721 Log4(("TM: %'RU64/-%'8RU64: exp tmr=>ff [NoLock]\n", u64, pVM->tm.s.offVirtualSync - pVM->tm.s.offVirtualSyncGivenUp)); 724 Log4(("TM: %'RU64/-%'8RU64: exp tmr [NoLock]\n", u64, pVM->tm.s.offVirtualSync - pVM->tm.s.offVirtualSyncGivenUp)); 832 return pVM->tm.s.offVirtualSync - pVM->tm.s.offVirtualSyncGivenUp; [all...] |
H A D | TMAll.cpp | 883 uint64_t const offGivenUp = ASMAtomicReadU64(&pVM->tm.s.offVirtualSyncGivenUp); 886 && offGivenUp == ASMAtomicReadU64(&pVM->tm.s.offVirtualSyncGivenUp)
|
/vbox/src/VBox/VMM/VMMR3/ |
H A D | TM.cpp | 650 STAM_REL_REG_USED(pVM,(void*)&pVM->tm.s.offVirtualSyncGivenUp, STAMTYPE_U64, "/TM/VirtualSync/GivenUp", STAMUNIT_NS, "Nanoseconds of the 'CurrentOffset' that's been given up and won't ever be attempted caught up with."); 1143 const uint64_t offOld = pVM->tm.s.offVirtualSyncGivenUp; 1146 ASMAtomicWriteU64((uint64_t volatile *)&pVM->tm.s.offVirtualSyncGivenUp, offNew); 1234 SSMR3PutU64(pSSM, pVM->tm.s.offVirtualSyncGivenUp); 1311 pVM->tm.s.offVirtualSyncGivenUp = u64; 2202 uint64_t const offSyncGivenUp = pVM->tm.s.offVirtualSyncGivenUp; 2282 Assert(offSyncGivenUp == pVM->tm.s.offVirtualSyncGivenUp); 2388 uint64_t offLag = offNew - pVM->tm.s.offVirtualSyncGivenUp; 2423 ASMAtomicWriteU64((uint64_t volatile *)&pVM->tm.s.offVirtualSyncGivenUp, offNew); 2448 ASMAtomicWriteU64((uint64_t volatile *)&pVM->tm.s.offVirtualSyncGivenUp, offNe 2714 uint64_t offVirtualSyncGivenUp; local [all...] |
/vbox/src/VBox/VMM/include/ |
H A D | TMInternal.h | 437 * Thus the current lag is offVirtualSync - offVirtualSyncGivenUp. */ 438 uint64_t offVirtualSyncGivenUp; member in struct:TM
|
/vbox/src/VBox/VMM/testcase/ |
H A D | tstVMStruct.h | 1056 GEN_CHECK_OFF(TM, offVirtualSyncGivenUp);
|
Completed in 67 milliseconds