Home
last modified time | relevance | path

Searched refs:SID_SHIFT (Results 1 – 14 of 14) sorted by relevance

/linux-6.12.1/arch/powerpc/include/asm/book3s/64/
Dmmu-hash.h263 return SID_SHIFT; in segment_shift()
441 mask = (1ul << (SID_SHIFT - VPN_SHIFT)) - 1; in hpt_hash()
442 hash = (vpn >> (SID_SHIFT - VPN_SHIFT)) ^ in hpt_hash()
580 #define ESID_BITS (VA_BITS - (SID_SHIFT + CONTEXT_BITS))
623 #define MAX_USER_CONTEXT_65BIT_VA ((ASM_CONST(1) << (65 - (SID_SHIFT + ESID_BITS))) - 2)
651 #define VSID_BITS_256M (VA_BITS - SID_SHIFT)
652 #define VSID_BITS_65_256M (65 - SID_SHIFT)
665 #define USER_VSID_RANGE (1UL << (ESID_BITS + SID_SHIFT))
795 vsid_bits = va_bits - SID_SHIFT; in get_vsid()
797 ((ea >> SID_SHIFT) & ESID_BITS_MASK); in get_vsid()
/linux-6.12.1/arch/powerpc/kvm/
Dbook3s_32_mmu.c36 #ifndef SID_SHIFT
37 #define SID_SHIFT 28 macro
89 kvmppc_mmu_book3s_32_esid_to_vsid(vcpu, eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_book3s_32_ea_to_vp()
158 eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_book3s_32_xlate_bat()
335 kvmppc_mmu_map_segment(vcpu, srnum << SID_SHIFT); in kvmppc_mmu_book3s_32_mtsrin()
351 ulong ea = esid << SID_SHIFT; in kvmppc_mmu_book3s_32_esid_to_vsid()
Dbook3s_64_mmu.c65 return slbe->tb ? SID_SHIFT_1T : SID_SHIFT; in kvmppc_slb_sid_shift()
409 kvmppc_mmu_map_segment(vcpu, esid << SID_SHIFT); in kvmppc_mmu_book3s_64_slbmte()
568 (mp_ea >> SID_SHIFT) == esid; in segment_contains_magic_page()
575 ulong ea = esid << SID_SHIFT; in kvmppc_mmu_book3s_64_esid_to_vsid()
588 gvsid <<= SID_SHIFT_1T - SID_SHIFT; in kvmppc_mmu_book3s_64_esid_to_vsid()
589 gvsid |= esid & ((1ul << (SID_SHIFT_1T - SID_SHIFT)) - 1); in kvmppc_mmu_book3s_64_esid_to_vsid()
637 unlikely(esid == (mp_ea >> SID_SHIFT)) && in kvmppc_mmu_book3s_64_esid_to_vsid()
Dbook3s_32_mmu_host.c158 vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_map_page()
167 vpn = (vsid << (SID_SHIFT - VPN_SHIFT)) | in kvmppc_mmu_map_page()
304 u32 esid = eaddr >> SID_SHIFT; in kvmppc_mmu_map_segment()
Dbook3s_64_mmu_host.c107 vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_map_page()
219 vcpu->arch.mmu.esid_to_vsid(vcpu, pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_unmap_page()
315 u64 esid = eaddr >> SID_SHIFT; in kvmppc_mmu_map_segment()
Dbook3s_pr.c715 pte.vpage |= ((u64)VSID_REAL << (SID_SHIFT - 12)); in kvmppc_handle_pagefault()
724 vcpu->arch.mmu.esid_to_vsid(vcpu, eaddr >> SID_SHIFT, &vsid); in kvmppc_handle_pagefault()
727 pte.vpage |= ((u64)VSID_REAL_DR << (SID_SHIFT - 12)); in kvmppc_handle_pagefault()
729 pte.vpage |= ((u64)VSID_REAL_IR << (SID_SHIFT - 12)); in kvmppc_handle_pagefault()
1170 sr = svcpu->sr[kvmppc_get_pc(vcpu) >> SID_SHIFT]; in kvmppc_handle_exit_pr()
1216 sr = svcpu->sr[dar >> SID_SHIFT]; in kvmppc_handle_exit_pr()
/linux-6.12.1/arch/powerpc/include/asm/
Dpage_64.h27 #define SID_SHIFT 28 macro
30 #define GET_ESID(x) (((x) >> SID_SHIFT) & SID_MASK)
Dkvm_book3s_32.h31 #define SID_SHIFT 28 macro
Dkvm_book3s_64.h336 va_low ^= v >> (SID_SHIFT - 16); in compute_tlbie_rb()
/linux-6.12.1/arch/powerpc/mm/book3s64/
Dslb.c59 ea &= ~((1UL << SID_SHIFT) - 1); in assert_slb_presence()
308 esid = ea >> SID_SHIFT; in preload_add()
424 slbie_data <<= SID_SHIFT; in slb_cache_slbie_kernel()
437 slbie_data <<= SID_SHIFT; in slb_cache_slbie_user()
524 ea = (unsigned long)ti->slb_preload_esid[idx] << SID_SHIFT; in switch_slb()
614 local_paca->slb_cache[slb_cache_index++] = esid_data >> SID_SHIFT; in slb_cache_update()
724 local_paca->slb_cache[slb_cache_index++] = esid_data >> SID_SHIFT; in slb_insert_entry()
Dhash_native.c705 *vpn = vsid << (SID_SHIFT - VPN_SHIFT) | seg_off >> VPN_SHIFT; in hpte_decode()
/linux-6.12.1/drivers/misc/cxl/
Dfault.c43 hash = (slb->esid >> SID_SHIFT) & mask; in find_free_sste()
/linux-6.12.1/arch/powerpc/kernel/
Dsetup_64.c708 return 1UL << SID_SHIFT; in ppc64_bolted_size()
/linux-6.12.1/arch/powerpc/platforms/pseries/
Dlpar.c1942 vsid_modulus = ((1UL << (va_bits - SID_SHIFT)) - 1); in vsid_unscramble()