Searched refs:old_spte (Results 1 – 8 of 8) sorted by relevance
/linux-6.12.1/arch/x86/kvm/mmu/ |
D | tdp_mmu.c | 291 u64 old_spte, u64 new_spte, int level, 356 u64 old_spte; in handle_removed_pt() local 368 old_spte = kvm_tdp_mmu_write_spte_atomic(sptep, FROZEN_SPTE); in handle_removed_pt() 369 if (!is_frozen_spte(old_spte)) in handle_removed_pt() 383 old_spte = kvm_tdp_mmu_read_spte(sptep); in handle_removed_pt() 384 if (!is_shadow_present_pte(old_spte)) in handle_removed_pt() 415 old_spte = kvm_tdp_mmu_write_spte(sptep, old_spte, in handle_removed_pt() 419 old_spte, FROZEN_SPTE, level, shared); in handle_removed_pt() 442 u64 old_spte, u64 new_spte, int level, in handle_changed_spte() argument 445 bool was_present = is_shadow_present_pte(old_spte); in handle_changed_spte() [all …]
|
D | tdp_iter.h | 46 static inline bool kvm_tdp_mmu_spte_need_atomic_write(u64 old_spte, int level) in kvm_tdp_mmu_spte_need_atomic_write() argument 48 return is_shadow_present_pte(old_spte) && in kvm_tdp_mmu_spte_need_atomic_write() 49 is_last_spte(old_spte, level) && in kvm_tdp_mmu_spte_need_atomic_write() 50 spte_has_volatile_bits(old_spte); in kvm_tdp_mmu_spte_need_atomic_write() 53 static inline u64 kvm_tdp_mmu_write_spte(tdp_ptep_t sptep, u64 old_spte, in kvm_tdp_mmu_write_spte() argument 56 if (kvm_tdp_mmu_spte_need_atomic_write(old_spte, level)) in kvm_tdp_mmu_write_spte() 60 return old_spte; in kvm_tdp_mmu_write_spte() 63 static inline u64 tdp_mmu_clear_spte_bits(tdp_ptep_t sptep, u64 old_spte, in tdp_mmu_clear_spte_bits() argument 68 if (kvm_tdp_mmu_spte_need_atomic_write(old_spte, level)) { in tdp_mmu_clear_spte_bits() 73 __kvm_tdp_mmu_write_spte(sptep, old_spte & ~mask); in tdp_mmu_clear_spte_bits() [all …]
|
D | mmutrace.h | 258 u64 *sptep, u64 old_spte, int ret), 259 TP_ARGS(vcpu, fault, sptep, old_spte, ret), 266 __field(u64, old_spte) 276 __entry->old_spte = old_spte; 285 __entry->old_spte, __entry->new_spte, 394 TP_PROTO(int as_id, gfn_t gfn, int level, u64 old_spte, u64 new_spte), 395 TP_ARGS(as_id, gfn, level, old_spte, new_spte), 399 __field(u64, old_spte) 409 __entry->old_spte = old_spte; 417 __entry->old_spte, __entry->new_spte
|
D | tdp_iter.c | 16 iter->old_spte = kvm_tdp_mmu_read_spte(iter->sptep); in tdp_iter_refresh_sptep() 89 iter->old_spte = kvm_tdp_mmu_read_spte(iter->sptep); in try_step_down() 91 child_pt = spte_to_child_pt(iter->old_spte, iter->level); in try_step_down() 123 iter->old_spte = kvm_tdp_mmu_read_spte(iter->sptep); in try_step_side()
|
D | mmu.c | 494 u64 old_spte = *sptep; in mmu_spte_update_no_track() local 499 if (!is_shadow_present_pte(old_spte)) { in mmu_spte_update_no_track() 501 return old_spte; in mmu_spte_update_no_track() 504 if (!spte_has_volatile_bits(old_spte)) in mmu_spte_update_no_track() 507 old_spte = __update_clear_spte_slow(sptep, new_spte); in mmu_spte_update_no_track() 509 WARN_ON_ONCE(spte_to_pfn(old_spte) != spte_to_pfn(new_spte)); in mmu_spte_update_no_track() 511 return old_spte; in mmu_spte_update_no_track() 526 u64 old_spte = mmu_spte_update_no_track(sptep, new_spte); in mmu_spte_update() local 528 if (!is_shadow_present_pte(old_spte)) in mmu_spte_update() 536 if (is_mmu_writable_spte(old_spte) && in mmu_spte_update() [all …]
|
D | spte.c | 160 u64 old_spte, bool prefetch, bool can_unsync, in make_spte() argument 234 if (is_writable_pte(old_spte)) in make_spte()
|
D | spte.h | 502 u64 old_spte, bool prefetch, bool can_unsync,
|
/linux-6.12.1/Documentation/virt/kvm/ |
D | locking.rst | 113 | old_spte = *spte; | | 128 | if (cmpxchg(spte, old_spte, old_spte+W) | 167 | old_spte = *spte; | | 171 | if (old_spte.Accessed == 1 && | | 172 | old_spte.W == 0) | | 186 | old_spte = xchg(spte, 0ull) | | 187 | if (old_spte.Accessed == 1) | | 189 | if (old_spte.Dirty == 1) | |
|