/vbox/include/VBox/vmm/ |
H A D | cpumctx-v1_6.h | 222 uint64_t msrEFER; member in struct:CPUMCTX_VER1_6
|
H A D | cpumctx.h | 391 uint64_t msrEFER; member in struct:CPUMCTX
|
H A D | cpum.h | 1178 return (pCtx->msrEFER & MSR_K6_EFER_LMA) == MSR_K6_EFER_LMA; 1192 if (!(pCtx->msrEFER & MSR_K6_EFER_LMA)) 1222 && !(pCtx->msrEFER & MSR_K6_EFER_LMA));
|
/vbox/src/VBox/VMM/VMMR3/ |
H A D | EMHM.cpp | 522 Log(("HWR%d: %04X:%RGv ESP=%RGv IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pCtx->cs.Sel, (RTGCPTR)pCtx->rip, pCtx->rsp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER)); 524 Log(("HWR%d: %04X:%08X ESP=%08X IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pCtx->cs.Sel, pCtx->eip, pCtx->esp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER)); 531 Log(("HWR%d-CPU%d: %04X:%RGv ESP=%RGv IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pVCpu->idCpu, pCtx->cs.Sel, (RTGCPTR)pCtx->rip, pCtx->rsp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER)); 533 Log(("HWR%d-CPU%d: %04X:%08X ESP=%08X IF=%d IOPL=%d CR0=%x CR4=%x EFER=%x\n", cpl, pVCpu->idCpu, pCtx->cs.Sel, pCtx->eip, pCtx->esp, pCtx->eflags.Bits.u1IF, pCtx->eflags.Bits.u2IOPL, (uint32_t)pCtx->cr0, (uint32_t)pCtx->cr4, (uint32_t)pCtx->msrEFER));
|
H A D | CPUM.cpp | 229 SSMFIELD_ENTRY( CPUMCTX, msrEFER), 374 SSMFIELD_ENTRY( CPUMCTX, msrEFER), 513 SSMFIELD_ENTRY( CPUMCTX, msrEFER), 935 Assert(!pCtx->msrEFER); 1386 PGMNotifyNxeChanged(pVCpu, RT_BOOL(pVCpu->cpum.s.Guest.msrEFER & MSR_K6_EFER_NXE)); 1666 pszPrefix, pCtx->msrEFER,
|
H A D | VMMTests.cpp | 669 PGMChangeMode(pVCpu, pHyperCtx->cr0, pHyperCtx->cr4, pHyperCtx->msrEFER);
|
/vbox/src/VBox/VMM/VMMR0/ |
H A D | CPUMR0UnusedA.asm | 170 test qword [xDX + CPUMCTX.msrEFER], MSR_K6_EFER_LMA 200 test qword [rdx + CPUMCTX.msrEFER], MSR_K6_EFER_LMA 256 test qword [xDX + CPUMCTX.msrEFER], MSR_K6_EFER_LMA 287 test qword [rdx + CPUMCTX.msrEFER], MSR_K6_EFER_LMA
|
H A D | HMR0.cpp | 1992 pCtx->msrEFER,
|
H A D | HMSVMR0.cpp | 1390 pVmcb->guest.u64EFER = pCtx->msrEFER | MSR_K6_EFER_SVME; 1404 if (pCtx->msrEFER & MSR_K6_EFER_LME)
|
H A D | HMVMXR0.cpp | 3245 uint64_t u64GuestEfer = pMixedCtx->msrEFER; 4823 int rc = VMXWriteVmcs64(VMX_VMCS64_GUEST_EFER_FULL, pMixedCtx->msrEFER); 4825 Log4(("Load[%RU32]: VMX_VMCS64_GUEST_EFER_FULL=%#RX64\n", pVCpu->idCpu, pMixedCtx->msrEFER)); 4829 hmR0VmxAddAutoLoadStoreMsr(pVCpu, MSR_K6_EFER, pMixedCtx->msrEFER, false /* fUpdateHostMsr */); 4833 pMixedCtx->msrEFER, pVCpu->hm.s.vmx.cMsrs));
|
/vbox/src/VBox/VMM/include/ |
H A D | EMHandleRCTmpl.h | 165 rc = PGMChangeMode(pVCpu, pCtx->cr0, pCtx->cr4, pCtx->msrEFER);
|
/vbox/src/VBox/VMM/VMMAll/ |
H A D | CPUMAllRegs.cpp | 884 pVCpu->cpum.s.Guest.msrEFER = val; 1150 return pVCpu->cpum.s.Guest.msrEFER; 2207 return !!(pVCpu->cpum.s.Guest.msrEFER & MSR_K6_EFER_NXE); 2305 return (pVCpu->cpum.s.Guest.msrEFER & MSR_K6_EFER_LMA) == MSR_K6_EFER_LMA; 2321 && !(pVCpu->cpum.s.Guest.msrEFER & MSR_K6_EFER_LMA); 2804 else if (!(pVCpu->cpum.s.Guest.msrEFER & MSR_K6_EFER_LMA)) 2826 Assert(!(pVCpu->cpum.s.Guest.msrEFER & MSR_K6_EFER_LMA)); 2832 && (pVCpu->cpum.s.Guest.msrEFER & MSR_K6_EFER_LMA)) 2849 Assert(!(pVCpu->cpum.s.Guest.msrEFER & MSR_K6_EFER_LMA)); 2855 && (pVCpu->cpum.s.Guest.msrEFER [all...] |
H A D | EMAll.cpp | 674 CHECK_FIELD(msrEFER); 1449 uint64_t msrEFER; local 1483 msrEFER = CPUMGetGuestEFER(pVCpu); 1484 if (msrEFER & MSR_K6_EFER_LME) 1502 msrEFER |= MSR_K6_EFER_LMA; 1508 msrEFER &= ~MSR_K6_EFER_LMA; 1511 CPUMSetGuestEFER(pVCpu, msrEFER); 1537 msrEFER = CPUMGetGuestEFER(pVCpu); 1538 if ( (msrEFER & MSR_K6_EFER_LMA)
|
H A D | IEMAllCImpl.cpp.h | 3636 if (!(pCtx->msrEFER & MSR_K6_EFER_SCE)) 3735 if (!(pCtx->msrEFER & MSR_K6_EFER_SCE)) 4827 && (pCtx->msrEFER & MSR_K6_EFER_LME) ) 4856 && (pCtx->msrEFER & MSR_K6_EFER_LME) ) 4858 uint64_t NewEFER = pCtx->msrEFER; 4867 pCtx->msrEFER = NewEFER; 4868 Assert(pCtx->msrEFER == NewEFER); 4883 rcStrict = PGMChangeMode(pVCpu, pCtx->cr0, pCtx->cr4, pCtx->msrEFER); 4924 && (pCtx->msrEFER & MSR_K6_EFER_LME)) 5036 rcStrict = PGMChangeMode(pVCpu, pCtx->cr0, pCtx->cr4, pCtx->msrEFER); [all...] |
H A D | SELMAll.cpp | 543 Assert(!(pCtx->msrEFER & MSR_K6_EFER_LMA));
|
H A D | CPUMAllMsrs.cpp | 1320 *puValue = pVCpu->cpum.s.Guest.msrEFER; 1329 uint64_t const uOldEfer = pVCpu->cpum.s.Guest.msrEFER; 1365 pVCpu->cpum.s.Guest.msrEFER = (uOldEfer & ~fMask) | (uValue & fMask); 1370 != (pVCpu->cpum.s.Guest.msrEFER & (MSR_K6_EFER_NXE | MSR_K6_EFER_LME | MSR_K6_EFER_LMA))) 1377 != (pVCpu->cpum.s.Guest.msrEFER & MSR_K6_EFER_NXE))
|
H A D | IEMAll.cpp | 978 if ((fFlags & X86_PTE_PAE_NX) && (pCtx->msrEFER & MSR_K6_EFER_NXE)) 1128 if ((fFlags & X86_PTE_PAE_NX) && (pCtx->msrEFER & MSR_K6_EFER_NXE)) 3960 else if (pCtx->msrEFER & MSR_K6_EFER_LMA) 4158 && (pIemCpu->CTX_SUFF(pCtx)->msrEFER & MSR_K6_EFER_NXE) ) ) 6250 && (pIemCpu->CTX_SUFF(pCtx)->msrEFER & MSR_K6_EFER_NXE) ) 9626 || (pOrgCtx->msrEFER & MSR_K6_EFER_LMA) 10342 CHECK_FIELD(msrEFER);
|
/vbox/src/recompiler/ |
H A D | VBoxRecompiler.c | 1481 pCtx->msrEFER = env->efer; 1933 pCtx->msrEFER = efer; 2215 pVM->rem.s.Env.efer = pCtx->msrEFER; 2225 if (pCtx->msrEFER & MSR_K6_EFER_LMA) 2715 pCtx->msrEFER = pVM->rem.s.Env.efer; 2969 pCtx->msrEFER = pVM->rem.s.Env.efer;
|
/vbox/src/VBox/VMM/testcase/ |
H A D | tstVMStruct.h | 178 GEN_CHECK_OFF(CPUMCTX, msrEFER);
|