Home
last modified time | relevance | path

Searched refs:old_spte (Results 1 – 8 of 8) sorted by relevance

/linux-6.12.1/arch/x86/kvm/mmu/
Dtdp_mmu.c291 u64 old_spte, u64 new_spte, int level,
356 u64 old_spte; in handle_removed_pt() local
368 old_spte = kvm_tdp_mmu_write_spte_atomic(sptep, FROZEN_SPTE); in handle_removed_pt()
369 if (!is_frozen_spte(old_spte)) in handle_removed_pt()
383 old_spte = kvm_tdp_mmu_read_spte(sptep); in handle_removed_pt()
384 if (!is_shadow_present_pte(old_spte)) in handle_removed_pt()
415 old_spte = kvm_tdp_mmu_write_spte(sptep, old_spte, in handle_removed_pt()
419 old_spte, FROZEN_SPTE, level, shared); in handle_removed_pt()
442 u64 old_spte, u64 new_spte, int level, in handle_changed_spte() argument
445 bool was_present = is_shadow_present_pte(old_spte); in handle_changed_spte()
[all …]
Dtdp_iter.h46 static inline bool kvm_tdp_mmu_spte_need_atomic_write(u64 old_spte, int level) in kvm_tdp_mmu_spte_need_atomic_write() argument
48 return is_shadow_present_pte(old_spte) && in kvm_tdp_mmu_spte_need_atomic_write()
49 is_last_spte(old_spte, level) && in kvm_tdp_mmu_spte_need_atomic_write()
50 spte_has_volatile_bits(old_spte); in kvm_tdp_mmu_spte_need_atomic_write()
53 static inline u64 kvm_tdp_mmu_write_spte(tdp_ptep_t sptep, u64 old_spte, in kvm_tdp_mmu_write_spte() argument
56 if (kvm_tdp_mmu_spte_need_atomic_write(old_spte, level)) in kvm_tdp_mmu_write_spte()
60 return old_spte; in kvm_tdp_mmu_write_spte()
63 static inline u64 tdp_mmu_clear_spte_bits(tdp_ptep_t sptep, u64 old_spte, in tdp_mmu_clear_spte_bits() argument
68 if (kvm_tdp_mmu_spte_need_atomic_write(old_spte, level)) { in tdp_mmu_clear_spte_bits()
73 __kvm_tdp_mmu_write_spte(sptep, old_spte & ~mask); in tdp_mmu_clear_spte_bits()
[all …]
Dmmutrace.h258 u64 *sptep, u64 old_spte, int ret),
259 TP_ARGS(vcpu, fault, sptep, old_spte, ret),
266 __field(u64, old_spte)
276 __entry->old_spte = old_spte;
285 __entry->old_spte, __entry->new_spte,
394 TP_PROTO(int as_id, gfn_t gfn, int level, u64 old_spte, u64 new_spte),
395 TP_ARGS(as_id, gfn, level, old_spte, new_spte),
399 __field(u64, old_spte)
409 __entry->old_spte = old_spte;
417 __entry->old_spte, __entry->new_spte
Dtdp_iter.c16 iter->old_spte = kvm_tdp_mmu_read_spte(iter->sptep); in tdp_iter_refresh_sptep()
89 iter->old_spte = kvm_tdp_mmu_read_spte(iter->sptep); in try_step_down()
91 child_pt = spte_to_child_pt(iter->old_spte, iter->level); in try_step_down()
123 iter->old_spte = kvm_tdp_mmu_read_spte(iter->sptep); in try_step_side()
Dmmu.c494 u64 old_spte = *sptep; in mmu_spte_update_no_track() local
499 if (!is_shadow_present_pte(old_spte)) { in mmu_spte_update_no_track()
501 return old_spte; in mmu_spte_update_no_track()
504 if (!spte_has_volatile_bits(old_spte)) in mmu_spte_update_no_track()
507 old_spte = __update_clear_spte_slow(sptep, new_spte); in mmu_spte_update_no_track()
509 WARN_ON_ONCE(spte_to_pfn(old_spte) != spte_to_pfn(new_spte)); in mmu_spte_update_no_track()
511 return old_spte; in mmu_spte_update_no_track()
526 u64 old_spte = mmu_spte_update_no_track(sptep, new_spte); in mmu_spte_update() local
528 if (!is_shadow_present_pte(old_spte)) in mmu_spte_update()
536 if (is_mmu_writable_spte(old_spte) && in mmu_spte_update()
[all …]
Dspte.c160 u64 old_spte, bool prefetch, bool can_unsync, in make_spte() argument
234 if (is_writable_pte(old_spte)) in make_spte()
Dspte.h502 u64 old_spte, bool prefetch, bool can_unsync,
/linux-6.12.1/Documentation/virt/kvm/
Dlocking.rst113 | old_spte = *spte; | |
128 | if (cmpxchg(spte, old_spte, old_spte+W) |
167 | old_spte = *spte; | |
171 | if (old_spte.Accessed == 1 && | |
172 | old_spte.W == 0) | |
186 | old_spte = xchg(spte, 0ull) | |
187 | if (old_spte.Accessed == 1) | |
189 | if (old_spte.Dirty == 1) | |