Lines Matching refs:vmcb01
97 kvm_init_shadow_npt_mmu(vcpu, X86_CR0_PG, svm->vmcb01.ptr->save.cr4, in nested_svm_init_mmu_context()
98 svm->vmcb01.ptr->save.efer, in nested_svm_init_mmu_context()
138 h = &svm->vmcb01.ptr->control; in recalc_intercepts()
156 if (!(svm->vmcb01.ptr->save.rflags & X86_EFLAGS_IF)) in recalc_intercepts()
538 svm->nested.vmcb02.ptr->save.g_pat = svm->vmcb01.ptr->save.g_pat; in nested_vmcb02_compute_g_pat()
544 struct vmcb *vmcb01 = svm->vmcb01.ptr; in nested_vmcb02_prepare_save() local
607 } else if (unlikely(vmcb01->control.virt_ext & LBR_CTL_ENABLE_MASK)) { in nested_vmcb02_prepare_save()
608 svm_copy_lbrs(vmcb02, vmcb01); in nested_vmcb02_prepare_save()
644 struct vmcb *vmcb01 = svm->vmcb01.ptr; in nested_vmcb02_prepare_control() local
661 if (vmcb01->control.int_ctl & V_NMI_PENDING_MASK) { in nested_vmcb02_prepare_control()
672 vmcb02->control.nested_ctl = vmcb01->control.nested_ctl; in nested_vmcb02_prepare_control()
673 vmcb02->control.iopm_base_pa = vmcb01->control.iopm_base_pa; in nested_vmcb02_prepare_control()
674 vmcb02->control.msrpm_base_pa = vmcb01->control.msrpm_base_pa; in nested_vmcb02_prepare_control()
698 (vmcb01->control.int_ctl & int_ctl_vmcb01_bits); in nested_vmcb02_prepare_control()
729 vmcb02->control.virt_ext = vmcb01->control.virt_ext & in nested_vmcb02_prepare_control()
753 vmcb02->control.pause_filter_count = vmcb01->control.pause_filter_count; in nested_vmcb02_prepare_control()
754 vmcb02->control.pause_filter_thresh = vmcb01->control.pause_filter_thresh; in nested_vmcb02_prepare_control()
817 nested_svm_copy_common_state(svm->vmcb01.ptr, svm->nested.vmcb02.ptr); in enter_svm_guest_mode()
848 struct vmcb *vmcb01 = svm->vmcb01.ptr; in nested_svm_vmrun() local
899 vmcb01->save.efer = vcpu->arch.efer; in nested_svm_vmrun()
900 vmcb01->save.cr0 = kvm_read_cr0(vcpu); in nested_svm_vmrun()
901 vmcb01->save.cr4 = vcpu->arch.cr4; in nested_svm_vmrun()
902 vmcb01->save.rflags = kvm_get_rflags(vcpu); in nested_svm_vmrun()
903 vmcb01->save.rip = kvm_rip_read(vcpu); in nested_svm_vmrun()
906 vmcb01->save.cr3 = kvm_read_cr3(vcpu); in nested_svm_vmrun()
974 struct vmcb *vmcb01 = svm->vmcb01.ptr; in nested_svm_vmexit() local
1037 vmcb01->control.pause_filter_count = vmcb02->control.pause_filter_count; in nested_svm_vmexit()
1038 vmcb_mark_dirty(vmcb01, VMCB_INTERCEPTS); in nested_svm_vmexit()
1042 nested_svm_copy_common_state(svm->nested.vmcb02.ptr, svm->vmcb01.ptr); in nested_svm_vmexit()
1044 svm_switch_vmcb(svm, &svm->vmcb01); in nested_svm_vmexit()
1072 } else if (unlikely(vmcb01->control.virt_ext & LBR_CTL_ENABLE_MASK)) { in nested_svm_vmexit()
1073 svm_copy_lbrs(vmcb01, vmcb02); in nested_svm_vmexit()
1079 vmcb01->control.int_ctl |= V_NMI_BLOCKING_MASK; in nested_svm_vmexit()
1081 vmcb01->control.int_ctl &= ~V_NMI_BLOCKING_MASK; in nested_svm_vmexit()
1085 vmcb01->control.int_ctl |= V_NMI_PENDING_MASK; in nested_svm_vmexit()
1087 vmcb01->control.int_ctl &= ~V_NMI_PENDING_MASK; in nested_svm_vmexit()
1096 vmcb01->control.exit_int_info = 0; in nested_svm_vmexit()
1099 if (vmcb01->control.tsc_offset != svm->vcpu.arch.tsc_offset) { in nested_svm_vmexit()
1100 vmcb01->control.tsc_offset = svm->vcpu.arch.tsc_offset; in nested_svm_vmexit()
1101 vmcb_mark_dirty(vmcb01, VMCB_INTERCEPTS); in nested_svm_vmexit()
1115 kvm_set_rflags(vcpu, vmcb01->save.rflags); in nested_svm_vmexit()
1116 svm_set_efer(vcpu, vmcb01->save.efer); in nested_svm_vmexit()
1117 svm_set_cr0(vcpu, vmcb01->save.cr0 | X86_CR0_PE); in nested_svm_vmexit()
1118 svm_set_cr4(vcpu, vmcb01->save.cr4); in nested_svm_vmexit()
1119 kvm_rax_write(vcpu, vmcb01->save.rax); in nested_svm_vmexit()
1120 kvm_rsp_write(vcpu, vmcb01->save.rsp); in nested_svm_vmexit()
1121 kvm_rip_write(vcpu, vmcb01->save.rip); in nested_svm_vmexit()
1139 rc = nested_svm_load_cr3(vcpu, vmcb01->save.cr3, false, true); in nested_svm_vmexit()
1157 if (unlikely(vmcb01->save.rflags & X86_EFLAGS_TF)) in nested_svm_vmexit()
1212 if (WARN_ON_ONCE(svm->vmcb != svm->vmcb01.ptr)) in svm_free_nested()
1213 svm_switch_vmcb(svm, &svm->vmcb01); in svm_free_nested()
1243 svm_switch_vmcb(svm, &svm->vmcb01); in svm_leave_nested()
1513 if (svm->vmcb01.ptr->control.intercepts[INTERCEPT_EXCEPTION] & in nested_svm_exit_special()
1642 if (copy_to_user(&user_vmcb->save, &svm->vmcb01.ptr->save, in svm_get_nested_state()
1746 svm->nested.vmcb02.ptr->save = svm->vmcb01.ptr->save; in svm_set_nested_state()
1755 svm_copy_vmrun_state(&svm->vmcb01.ptr->save, save); in svm_set_nested_state()