Home
last modified time | relevance | path

Searched refs:shregs (Results 1 – 13 of 13) sorted by relevance

/linux-6.12.1/arch/powerpc/kvm/
Dbook3s_hv_tm_builtin.c40 newmsr = vcpu->arch.shregs.srr1; in kvmhv_p9_tm_emulation_early()
45 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation_early()
47 vcpu->arch.regs.nip = vcpu->arch.shregs.srr0; in kvmhv_p9_tm_emulation_early()
52 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation_early()
68 vcpu->arch.shregs.msr = msr; in kvmhv_p9_tm_emulation_early()
77 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation_early()
84 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation_early()
90 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation_early()
99 vcpu->arch.shregs.msr = (msr & ~MSR_TS_MASK) | MSR_TS_T; in kvmhv_p9_tm_emulation_early()
115 vcpu->arch.shregs.msr &= ~MSR_TS_MASK; /* go to N state */ in kvmhv_emulate_tm_rollback()
Dbook3s_hv_tm.c19 u64 msr = vcpu->arch.shregs.msr; in emulate_tx_failure()
23 if (MSR_TM_SUSPENDED(vcpu->arch.shregs.msr)) in emulate_tx_failure()
45 u64 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation()
72 newmsr = vcpu->arch.shregs.srr1; in kvmhv_p9_tm_emulation()
78 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation()
80 vcpu->arch.regs.nip = vcpu->arch.shregs.srr0; in kvmhv_p9_tm_emulation()
112 vcpu->arch.shregs.msr = msr; in kvmhv_p9_tm_emulation()
128 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation()
166 vcpu->arch.shregs.msr = msr; in kvmhv_p9_tm_emulation()
205 vcpu->arch.shregs.msr &= ~MSR_TS_MASK; in kvmhv_p9_tm_emulation()
[all …]
Dbook3s_hv_p9_entry.c94 unsigned long guest_msr = vcpu->arch.shregs.msr; in load_vcpu_state()
129 unsigned long guest_msr = vcpu->arch.shregs.msr; in store_vcpu_state()
558 WARN_ON_ONCE(vcpu->arch.shregs.msr & MSR_HV); in kvmhv_vcpu_entry_p9()
559 WARN_ON_ONCE(!(vcpu->arch.shregs.msr & MSR_ME)); in kvmhv_vcpu_entry_p9()
645 mtspr(SPRN_HSRR1, (vcpu->arch.shregs.msr & ~MSR_HV) | MSR_ME); in kvmhv_vcpu_entry_p9()
664 mtspr(SPRN_SPRG0, vcpu->arch.shregs.sprg0); in kvmhv_vcpu_entry_p9()
665 mtspr(SPRN_SPRG1, vcpu->arch.shregs.sprg1); in kvmhv_vcpu_entry_p9()
666 mtspr(SPRN_SPRG2, vcpu->arch.shregs.sprg2); in kvmhv_vcpu_entry_p9()
667 mtspr(SPRN_SPRG3, vcpu->arch.shregs.sprg3); in kvmhv_vcpu_entry_p9()
716 mtspr(SPRN_DAR, vcpu->arch.shregs.dar); in kvmhv_vcpu_entry_p9()
[all …]
Dbook3s_hv_nested.c45 hr->srr0 = vcpu->arch.shregs.srr0; in kvmhv_save_hv_regs()
46 hr->srr1 = vcpu->arch.shregs.srr1; in kvmhv_save_hv_regs()
47 hr->sprg[0] = vcpu->arch.shregs.sprg0; in kvmhv_save_hv_regs()
48 hr->sprg[1] = vcpu->arch.shregs.sprg1; in kvmhv_save_hv_regs()
49 hr->sprg[2] = vcpu->arch.shregs.sprg2; in kvmhv_save_hv_regs()
50 hr->sprg[3] = vcpu->arch.shregs.sprg3; in kvmhv_save_hv_regs()
113 hr->srr0 = vcpu->arch.shregs.srr0; in save_hv_return_state()
114 hr->srr1 = vcpu->arch.shregs.srr1; in save_hv_return_state()
115 hr->sprg[0] = vcpu->arch.shregs.sprg0; in save_hv_return_state()
116 hr->sprg[1] = vcpu->arch.shregs.sprg1; in save_hv_return_state()
[all …]
Dbook3s_hv_nestedv2.c285 vcpu->arch.shregs.srr0); in gs_msg_ops_vcpu_fill_info()
289 vcpu->arch.shregs.srr1); in gs_msg_ops_vcpu_fill_info()
293 vcpu->arch.shregs.sprg0); in gs_msg_ops_vcpu_fill_info()
297 vcpu->arch.shregs.sprg1); in gs_msg_ops_vcpu_fill_info()
301 vcpu->arch.shregs.sprg2); in gs_msg_ops_vcpu_fill_info()
305 vcpu->arch.shregs.sprg3); in gs_msg_ops_vcpu_fill_info()
309 vcpu->arch.shregs.dar); in gs_msg_ops_vcpu_fill_info()
313 vcpu->arch.shregs.dsisr); in gs_msg_ops_vcpu_fill_info()
317 vcpu->arch.shregs.msr); in gs_msg_ops_vcpu_fill_info()
539 vcpu->arch.shregs.srr0 = kvmppc_gse_get_u64(gse); in gs_msg_ops_vcpu_refresh_info()
[all …]
Dbook3s_hv.h58 vcpu->arch.shregs.msr = val; in __kvmppc_set_msr_hv()
65 return vcpu->arch.shregs.msr; in __kvmppc_get_msr_hv()
Dbook3s_hv_ras.c71 unsigned long srr1 = vcpu->arch.shregs.msr; in kvmppc_realmode_mc_power7()
76 unsigned long dsisr = vcpu->arch.shregs.dsisr; in kvmppc_realmode_mc_power7()
Dbook3s_hv.c489 vcpu->arch.regs.nip, vcpu->arch.shregs.msr, vcpu->arch.trap); in kvmppc_dump_regs()
497 vcpu->arch.shregs.srr0, vcpu->arch.shregs.srr1); in kvmppc_dump_regs()
499 vcpu->arch.shregs.sprg0, vcpu->arch.shregs.sprg1); in kvmppc_dump_regs()
501 vcpu->arch.shregs.sprg2, vcpu->arch.shregs.sprg3); in kvmppc_dump_regs()
503 vcpu->arch.regs.ccr, vcpu->arch.regs.xer, vcpu->arch.shregs.dsisr); in kvmppc_dump_regs()
504 pr_err("dar = %.16llx\n", vcpu->arch.shregs.dar); in kvmppc_dump_regs()
810 dt->srr1 = cpu_to_be64(vcpu->arch.shregs.msr); in __kvmppc_create_dtl_entry()
1647 vcpu->arch.shregs.msr); in kvmppc_handle_exit_hv()
2996 vcpu->arch.shared = &vcpu->arch.shregs; in kvmppc_core_vcpu_create_hv()
4301 vcpu->arch.regs.msr = vcpu->arch.shregs.msr; in kvmhv_vcpu_entry_p9_nested()
[all …]
Dbook3s_hv_rm_mmu.c1241 key = (vcpu->arch.shregs.msr & MSR_PR) ? SLB_VSID_KP : SLB_VSID_KS; in kvmppc_hpte_hv_fault()
1258 if (data && (vcpu->arch.shregs.msr & MSR_DR)) { in kvmppc_hpte_hv_fault()
1294 if (data && (vcpu->arch.shregs.msr & MSR_IR)) in kvmppc_hpte_hv_fault()
Dbook3s_hv_builtin.c584 if (vcpu->arch.shregs.msr & MSR_EE) { in kvmppc_guest_entry_inject_int()
Dtrace_hv.h282 __entry->msr = vcpu->arch.shregs.msr;
/linux-6.12.1/arch/powerpc/kernel/
Dasm-offsets.c376 OFFSET(VCPU_MSR, kvm_vcpu, arch.shregs.msr); in main()
377 OFFSET(VCPU_SRR0, kvm_vcpu, arch.shregs.srr0); in main()
378 OFFSET(VCPU_SRR1, kvm_vcpu, arch.shregs.srr1); in main()
379 OFFSET(VCPU_SPRG0, kvm_vcpu, arch.shregs.sprg0); in main()
380 OFFSET(VCPU_SPRG1, kvm_vcpu, arch.shregs.sprg1); in main()
381 OFFSET(VCPU_SPRG2, kvm_vcpu, arch.shregs.sprg2); in main()
382 OFFSET(VCPU_SPRG3, kvm_vcpu, arch.shregs.sprg3); in main()
430 OFFSET(VCPU_DSISR, kvm_vcpu, arch.shregs.dsisr); in main()
431 OFFSET(VCPU_DAR, kvm_vcpu, arch.shregs.dar); in main()
/linux-6.12.1/arch/powerpc/include/asm/
Dkvm_host.h818 struct kvm_vcpu_arch_shared shregs; member