/linux-6.12.1/arch/sh/kernel/cpu/ |
D | init.c | 128 waysize = current_cpu_data.dcache.sets; in cache_init() 139 waysize <<= current_cpu_data.dcache.entry_shift; in cache_init() 147 ways = current_cpu_data.dcache.ways; in cache_init() 155 addr += current_cpu_data.dcache.linesz) in cache_init() 158 addrstart += current_cpu_data.dcache.way_incr; in cache_init() 170 if (current_cpu_data.dcache.ways > 1) in cache_init() 204 l1d_cache_shape = CACHE_DESC_SHAPE(current_cpu_data.dcache); in detect_cache_shape() 206 if (current_cpu_data.dcache.flags & SH_CACHE_COMBINED) in detect_cache_shape() 209 l1i_cache_shape = CACHE_DESC_SHAPE(current_cpu_data.icache); in detect_cache_shape() 211 if (current_cpu_data.flags & CPU_HAS_L2_CACHE) in detect_cache_shape() [all …]
|
/linux-6.12.1/arch/sh/mm/ |
D | cache-sh7705.c | 35 ways = current_cpu_data.dcache.ways; in cache_wback_all() 36 waysize = current_cpu_data.dcache.sets; in cache_wback_all() 37 waysize <<= current_cpu_data.dcache.entry_shift; in cache_wback_all() 46 addr += current_cpu_data.dcache.linesz) { in cache_wback_all() 57 addrstart += current_cpu_data.dcache.way_incr; in cache_wback_all() 103 ways = current_cpu_data.dcache.ways; in __flush_dcache_page() 104 waysize = current_cpu_data.dcache.sets; in __flush_dcache_page() 105 waysize <<= current_cpu_data.dcache.entry_shift; in __flush_dcache_page() 114 addr += current_cpu_data.dcache.linesz) { in __flush_dcache_page() 124 addrstart += current_cpu_data.dcache.way_incr; in __flush_dcache_page()
|
D | cache-sh3.c | 44 for (j = 0; j < current_cpu_data.dcache.ways; j++) { in sh3__flush_wback_region() 48 addr = addrstart | (v & current_cpu_data.dcache.entry_mask); in sh3__flush_wback_region() 60 addrstart += current_cpu_data.dcache.way_incr; in sh3__flush_wback_region() 85 (v & current_cpu_data.dcache.entry_mask) | SH_CACHE_ASSOC; in sh3__flush_purge_region()
|
D | cache-sh4.c | 164 (current_cpu_data.dcache.sets << in flush_dcache_all() 165 current_cpu_data.dcache.entry_shift) * in flush_dcache_all() 166 current_cpu_data.dcache.ways; in flush_dcache_all() 168 entry_offset = 1 << current_cpu_data.dcache.entry_shift; in flush_dcache_all() 249 map_coherent = (current_cpu_data.dcache.n_aliases && in sh4_flush_cache_page()
|
D | cache-sh2a.c | 60 nr_ways = current_cpu_data.dcache.ways; in sh2a__flush_wback_region() 68 end = begin + (nr_ways * current_cpu_data.dcache.way_size); in sh2a__flush_wback_region() 107 int nr_ways = current_cpu_data.dcache.ways; in sh2a__flush_purge_region()
|
D | cache-debugfs.c | 49 cache = ¤t_cpu_data.dcache; in cache_debugfs_show() 52 cache = ¤t_cpu_data.icache; in cache_debugfs_show()
|
/linux-6.12.1/arch/mips/mm/ |
D | tlb-r4k.c | 74 if (current_cpu_data.tlbsizevtlb) { in local_flush_tlb_all() 79 ftlbhighset = current_cpu_data.tlbsizevtlb + in local_flush_tlb_all() 80 current_cpu_data.tlbsizeftlbsets; in local_flush_tlb_all() 81 for (entry = current_cpu_data.tlbsizevtlb; in local_flush_tlb_all() 89 while (entry < current_cpu_data.tlbsize) { in local_flush_tlb_all() 119 if (size <= (current_cpu_data.tlbsizeftlbsets ? in local_flush_tlb_range() 120 current_cpu_data.tlbsize / 8 : in local_flush_tlb_range() 121 current_cpu_data.tlbsize / 2)) { in local_flush_tlb_range() 173 if (size <= (current_cpu_data.tlbsizeftlbsets ? in local_flush_tlb_kernel_range() 174 current_cpu_data.tlbsize / 8 : in local_flush_tlb_kernel_range() [all …]
|
D | tlb-r3k.c | 44 old_ctx = read_c0_entryhi() & cpu_asid_mask(¤t_cpu_data); in local_flush_tlb_from() 46 while (entry < current_cpu_data.tlbsize) { in local_flush_tlb_from() 70 unsigned long asid_mask = cpu_asid_mask(¤t_cpu_data); in local_flush_tlb_range() 83 if (size <= current_cpu_data.tlbsize) { in local_flush_tlb_range() 120 if (size <= current_cpu_data.tlbsize) { in local_flush_tlb_kernel_range() 149 unsigned long asid_mask = cpu_asid_mask(¤t_cpu_data); in local_flush_tlb_page() 181 unsigned long asid_mask = cpu_asid_mask(¤t_cpu_data); in __update_tlb() 220 unsigned long asid_mask = cpu_asid_mask(¤t_cpu_data); in add_wired_entry()
|
D | c-r4k.c | 231 unsigned long end = start + current_cpu_data.icache.waysize; in tx49_blast_icache32() 232 unsigned long ws_inc = 1UL << current_cpu_data.icache.waybit; in tx49_blast_icache32() 233 unsigned long ws_end = current_cpu_data.icache.ways << in tx49_blast_icache32() 234 current_cpu_data.icache.waybit; in tx49_blast_icache32() 996 struct cpuinfo_mips *c = ¤t_cpu_data; in probe_pcache() 1356 struct cpuinfo_mips *c = ¤t_cpu_data; in probe_vcache() 1390 struct cpuinfo_mips *c = ¤t_cpu_data; in probe_scache() 1443 struct cpuinfo_mips *c = ¤t_cpu_data; in loongson2_sc_init() 1459 struct cpuinfo_mips *c = ¤t_cpu_data; in loongson3_sc_init() 1490 struct cpuinfo_mips *c = ¤t_cpu_data; in setup_scache() [all …]
|
/linux-6.12.1/arch/mips/include/asm/ |
D | r4kcache.h | 205 unsigned long end = start + current_cpu_data.desc.waysize; \ 206 unsigned long ws_inc = 1UL << current_cpu_data.desc.waybit; \ 207 unsigned long ws_end = current_cpu_data.desc.ways << \ 208 current_cpu_data.desc.waybit; \ 230 unsigned long indexmask = current_cpu_data.desc.waysize - 1; \ 233 unsigned long ws_inc = 1UL << current_cpu_data.desc.waybit; \ 234 unsigned long ws_end = current_cpu_data.desc.ways << \ 235 current_cpu_data.desc.waybit; \ 321 unsigned long end = start + current_cpu_data.desc.waysize; \ 322 unsigned long ws_inc = 1UL << current_cpu_data.desc.waybit; \ [all …]
|
/linux-6.12.1/arch/loongarch/mm/ |
D | cache.c | 48 struct cache_desc *cdesc = current_cpu_data.cache_leaves + leaf; in flush_cache_leaf() 69 struct cache_desc *cdesc = current_cpu_data.cache_leaves; in __flush_cache_all() 70 unsigned int cache_present = current_cpu_data.cache_leaves_present; in __flush_cache_all() 119 struct cache_desc *cdesc = current_cpu_data.cache_leaves; in cpu_cache_init() 157 current_cpu_data.cache_leaves_present = leaf; in cpu_cache_init() 158 current_cpu_data.options |= LOONGARCH_CPU_PREFETCH; in cpu_cache_init()
|
D | tlb.c | 70 if (size <= (current_cpu_data.tlbsizestlbsets ? in local_flush_tlb_range() 71 current_cpu_data.tlbsize / 8 : in local_flush_tlb_range() 72 current_cpu_data.tlbsize / 2)) { in local_flush_tlb_range() 95 if (size <= (current_cpu_data.tlbsizestlbsets ? in local_flush_tlb_kernel_range() 96 current_cpu_data.tlbsize / 8 : in local_flush_tlb_kernel_range() 97 current_cpu_data.tlbsize / 2)) { in local_flush_tlb_kernel_range()
|
/linux-6.12.1/arch/loongarch/kernel/ |
D | cacheinfo.c | 14 int cache_present = current_cpu_data.cache_leaves_present; in init_cache_level() 18 current_cpu_data.cache_leaves[cache_present - 1].level; in init_cache_level() 64 int i, cache_present = current_cpu_data.cache_leaves_present; in populate_cache_leaves() 67 struct cache_desc *cd, *cdesc = current_cpu_data.cache_leaves; in populate_cache_leaves()
|
/linux-6.12.1/arch/mips/kernel/ |
D | watch.c | 24 switch (current_cpu_data.watch_reg_use_cnt) { in mips_install_watch_registers() 56 switch (current_cpu_data.watch_reg_use_cnt) { in mips_read_watch_registers() 71 if (current_cpu_data.watch_reg_use_cnt == 1 && in mips_read_watch_registers() 89 switch (current_cpu_data.watch_reg_count) { in mips_clear_watch_registers()
|
D | smp-cps.c | 281 atomic_set(&mips_cps_core_bootcfg[cpu_core(¤t_cpu_data)].vpe_mask, in cps_prepare_cpus() 282 1 << cpu_vpe_id(¤t_cpu_data)); in cps_prepare_cpus() 379 unsigned core = cpu_core(¤t_cpu_data); in remote_vpe_boot() 382 mips_cps_boot_vpes(core_cfg, cpu_vpe_id(¤t_cpu_data)); in remote_vpe_boot() 450 int core = cpu_core(¤t_cpu_data); in cps_init_secondary() 552 core_cfg = &mips_cps_core_bootcfg[cpu_core(¤t_cpu_data)]; in cps_cpu_disable() 553 atomic_sub(1 << cpu_vpe_id(¤t_cpu_data), &core_cfg->vpe_mask); in cps_cpu_disable()
|
D | mips-cpc.c | 99 curr_core = cpu_core(¤t_cpu_data); in mips_cpc_lock_other() 119 curr_core = cpu_core(¤t_cpu_data); in mips_cpc_unlock_other()
|
D | cpu-r3k-probe.c | 70 struct cpuinfo_mips *c = ¤t_cpu_data; in cpu_probe() 145 struct cpuinfo_mips *c = ¤t_cpu_data; in cpu_report()
|
D | cacheinfo.c | 22 struct cpuinfo_mips *c = ¤t_cpu_data; in init_cache_level() 79 struct cpuinfo_mips *c = ¤t_cpu_data; in populate_cache_leaves()
|
D | mips-cm.c | 339 curr_core = cpu_core(¤t_cpu_data); in mips_cm_lock_other() 360 curr_core = cpu_core(¤t_cpu_data); in mips_cm_unlock_other() 521 local_cl = cpu_cluster(¤t_cpu_data); in mips_cps_first_online_in_cluster()
|
/linux-6.12.1/arch/mips/kvm/ |
D | tlb.c | 61 BUG_ON(idx >= current_cpu_data.tlbsize); in _kvm_mips_host_tlb_inv() 274 for (entry = 0; entry < current_cpu_data.tlbsize; entry++) { in kvm_vz_local_flush_roottlb_all_guests() 340 for (entry = 0; entry < current_cpu_data.guest.tlbsize; entry++) { in kvm_vz_local_flush_guesttlb_all() 512 for (i = current_cpu_data.tlbsizevtlb; in kvm_loongson_clear_guest_ftlb() 513 i < (current_cpu_data.tlbsizevtlb + in kvm_loongson_clear_guest_ftlb() 514 current_cpu_data.tlbsizeftlbsets); in kvm_loongson_clear_guest_ftlb()
|
/linux-6.12.1/arch/mips/lib/ |
D | r3k_dump_tlb.c | 30 asid_mask = cpu_asid_mask(¤t_cpu_data); in dump_tlb() 70 dump_tlb(0, current_cpu_data.tlbsize - 1); in dump_tlb_all()
|
D | dump_tlb.c | 72 unsigned long asidmask = cpu_asid_mask(¤t_cpu_data); in dump_tlb() 190 dump_tlb(0, current_cpu_data.tlbsize - 1); in dump_tlb_all()
|
/linux-6.12.1/arch/mips/bcm47xx/ |
D | prom.c | 63 struct cpuinfo_mips *c = ¤t_cpu_data; in prom_init_mem() 134 temp_tlb_entry = current_cpu_data.tlbsize - 1; in early_tlb_init()
|
/linux-6.12.1/arch/sh/kernel/ |
D | setup.c | 360 current_cpu_data.loops_per_jiffy = loops_per_jiffy; in arch_cpu_finalize_init() 362 switch (current_cpu_data.family) { in arch_cpu_finalize_init() 398 pr_info("CPU: %s\n", get_cpu_subtype(¤t_cpu_data)); in arch_cpu_finalize_init()
|
/linux-6.12.1/arch/loongarch/lib/ |
D | dump_tlb.c | 35 unsigned long asidmask = cpu_asid_mask(¤t_cpu_data); in dump_tlb() 110 dump_tlb(0, current_cpu_data.tlbsize - 1); in dump_tlb_all()
|