Searched refs:guest_context (Results 1 – 17 of 17) sorted by relevance
/linux-6.12.1/arch/riscv/kvm/ |
D | vcpu_exit.c | 70 old_hstatus = csr_swap(CSR_HSTATUS, vcpu->arch.guest_context.hstatus); in kvm_riscv_vcpu_unpriv_read() 142 if (vcpu->arch.guest_context.sstatus & SR_SPP) in kvm_riscv_vcpu_trap_redirect() 162 vcpu->arch.guest_context.sepc = csr_read(CSR_VSTVEC); in kvm_riscv_vcpu_trap_redirect() 165 vcpu->arch.guest_context.sstatus |= SR_SPP; in kvm_riscv_vcpu_trap_redirect() 190 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) { in kvm_riscv_vcpu_exit() 196 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit() 202 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit() 206 if (vcpu->arch.guest_context.hstatus & HSTATUS_SPV) in kvm_riscv_vcpu_exit() 221 vcpu->arch.guest_context.sepc, in kvm_riscv_vcpu_exit() 222 vcpu->arch.guest_context.sstatus, in kvm_riscv_vcpu_exit() [all …]
|
D | vcpu_insn.c | 162 utrap.sepc = vcpu->arch.guest_context.sepc; in truly_illegal_insn() 178 utrap.sepc = vcpu->arch.guest_context.sepc; in truly_virtual_insn() 212 kvm_vcpu_on_spin(vcpu, vcpu->arch.guest_context.sstatus & SR_SPP); in wrs_insn() 264 SET_RD(insn, &vcpu->arch.guest_context, in kvm_riscv_vcpu_csr_return() 268 vcpu->arch.guest_context.sepc += INSN_LEN(insn); in kvm_riscv_vcpu_csr_return() 278 ulong rs1_val = GET_RS1(insn, &vcpu->arch.guest_context); in csr_insn() 418 vcpu->arch.guest_context.sepc += INSN_LEN(insn); in system_opcode_insn() 447 ct = &vcpu->arch.guest_context; in kvm_riscv_vcpu_virtual_insn() 489 struct kvm_cpu_context *ct = &vcpu->arch.guest_context; in kvm_riscv_vcpu_mmio_load() 615 struct kvm_cpu_context *ct = &vcpu->arch.guest_context; in kvm_riscv_vcpu_mmio_store() [all …]
|
D | vcpu_sbi_replace.c | 20 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_time_handler() 51 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_ipi_handler() 86 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_rfence_handler() 141 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_srst_handler() 183 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_dbcn_handler()
|
D | vcpu_sbi_hsm.c | 18 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_hsm_vcpu_start() 75 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_hsm_vcpu_get_status() 94 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_hsm_handler()
|
D | vcpu_fp.c | 19 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_fp_reset() 81 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_get_reg_fp() 126 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_fp()
|
D | vcpu_vector.c | 22 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_vector_reset() 100 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_vreg_addr() 181 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_vector()
|
D | vcpu.c | 53 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_reset_vcpu() 226 return (vcpu->arch.guest_context.sstatus & SR_SPP) ? true : false; in kvm_arch_vcpu_in_kernel() 591 kvm_riscv_vcpu_guest_fp_restore(&vcpu->arch.guest_context, in kvm_arch_vcpu_load() 594 kvm_riscv_vcpu_guest_vector_restore(&vcpu->arch.guest_context, in kvm_arch_vcpu_load() 612 kvm_riscv_vcpu_guest_fp_save(&vcpu->arch.guest_context, in kvm_arch_vcpu_put() 617 kvm_riscv_vcpu_guest_vector_save(&vcpu->arch.guest_context, in kvm_arch_vcpu_put() 852 trap.sepc = vcpu->arch.guest_context.sepc; in kvm_arch_vcpu_ioctl_run()
|
D | vcpu_sbi.c | 117 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_forward() 157 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_return() 169 vcpu->arch.guest_context.sepc += 4; in kvm_riscv_vcpu_sbi_return() 423 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_riscv_vcpu_sbi_ecall()
|
D | vcpu_sbi_sta.c | 83 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_sta_steal_time_set_shmem() 131 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_sta_handler()
|
D | trace.h | 25 __entry->pc = vcpu->arch.guest_context.sepc;
|
D | vcpu_sbi_base.c | 19 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_base_handler()
|
D | vcpu_sbi_pmu.c | 20 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_pmu_handler()
|
D | vcpu_onereg.c | 359 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_get_reg_core() 392 struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in kvm_riscv_vcpu_set_reg_core() 913 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in num_fp_f_regs() 942 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in num_fp_d_regs() 1086 const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context; in copy_vector_reg_indices()
|
D | vcpu_sbi_v01.c | 24 struct kvm_cpu_context *cp = &vcpu->arch.guest_context; in kvm_sbi_ext_v01_handler()
|
D | aia_imsic.c | 828 vcpu->arch.guest_context.hstatus &= ~HSTATUS_VGEIN; in kvm_riscv_vcpu_aia_imsic_update() 830 vcpu->arch.guest_context.hstatus |= in kvm_riscv_vcpu_aia_imsic_update()
|
/linux-6.12.1/arch/riscv/kernel/ |
D | asm-offsets.c | 135 OFFSET(KVM_ARCH_GUEST_ZERO, kvm_vcpu_arch, guest_context.zero); in asm_offsets() 136 OFFSET(KVM_ARCH_GUEST_RA, kvm_vcpu_arch, guest_context.ra); in asm_offsets() 137 OFFSET(KVM_ARCH_GUEST_SP, kvm_vcpu_arch, guest_context.sp); in asm_offsets() 138 OFFSET(KVM_ARCH_GUEST_GP, kvm_vcpu_arch, guest_context.gp); in asm_offsets() 139 OFFSET(KVM_ARCH_GUEST_TP, kvm_vcpu_arch, guest_context.tp); in asm_offsets() 140 OFFSET(KVM_ARCH_GUEST_T0, kvm_vcpu_arch, guest_context.t0); in asm_offsets() 141 OFFSET(KVM_ARCH_GUEST_T1, kvm_vcpu_arch, guest_context.t1); in asm_offsets() 142 OFFSET(KVM_ARCH_GUEST_T2, kvm_vcpu_arch, guest_context.t2); in asm_offsets() 143 OFFSET(KVM_ARCH_GUEST_S0, kvm_vcpu_arch, guest_context.s0); in asm_offsets() 144 OFFSET(KVM_ARCH_GUEST_S1, kvm_vcpu_arch, guest_context.s1); in asm_offsets() [all …]
|
/linux-6.12.1/arch/riscv/include/asm/ |
D | kvm_host.h | 217 struct kvm_cpu_context guest_context; member
|