Home
last modified time | relevance | path

Searched refs:i915_mmio_reg_offset (Results 1 – 25 of 50) sorted by relevance

12

/linux-6.12.1/drivers/gpu/drm/xe/compat-i915-headers/
Dintel_uncore.h30 struct xe_reg reg = XE_REG(i915_mmio_reg_offset(i915_reg)); in intel_uncore_read()
38 struct xe_reg reg = XE_REG(i915_mmio_reg_offset(i915_reg)); in intel_uncore_read8()
46 struct xe_reg reg = XE_REG(i915_mmio_reg_offset(i915_reg)); in intel_uncore_read16()
55 struct xe_reg lower_reg = XE_REG(i915_mmio_reg_offset(i915_lower_reg)); in intel_uncore_read64_2x32()
56 struct xe_reg upper_reg = XE_REG(i915_mmio_reg_offset(i915_upper_reg)); in intel_uncore_read64_2x32()
73 struct xe_reg reg = XE_REG(i915_mmio_reg_offset(i915_reg)); in intel_uncore_posting_read()
81 struct xe_reg reg = XE_REG(i915_mmio_reg_offset(i915_reg)); in intel_uncore_write()
89 struct xe_reg reg = XE_REG(i915_mmio_reg_offset(i915_reg)); in intel_uncore_rmw()
98 struct xe_reg reg = XE_REG(i915_mmio_reg_offset(i915_reg)); in intel_wait_for_register()
108 struct xe_reg reg = XE_REG(i915_mmio_reg_offset(i915_reg)); in intel_wait_for_register_fw()
[all …]
/linux-6.12.1/drivers/gpu/drm/i915/gvt/
Dinterrupt.c173 if (i915_mmio_reg_offset(irq->info[i]->reg_base) == reg) in regbase_to_irq_info()
351 regbase_to_iir(i915_mmio_reg_offset(info->reg_base))) in update_upstream_irq()
353 regbase_to_ier(i915_mmio_reg_offset(info->reg_base))); in update_upstream_irq()
380 u32 isr = i915_mmio_reg_offset(up_irq_info->reg_base); in update_upstream_irq()
386 i915_mmio_reg_offset(up_irq_info->reg_base)); in update_upstream_irq()
388 i915_mmio_reg_offset(up_irq_info->reg_base)); in update_upstream_irq()
469 reg_base = i915_mmio_reg_offset(info->reg_base); in propagate_event()
527 if (!(vgpu_vreg(vgpu, i915_mmio_reg_offset(GEN8_MASTER_IRQ)) & in gen8_check_pending_irq()
538 reg_base = i915_mmio_reg_offset(info->reg_base); in gen8_check_pending_irq()
544 if (vgpu_vreg(vgpu, i915_mmio_reg_offset(GEN8_MASTER_IRQ)) in gen8_check_pending_irq()
Dedid.c384 if (offset == i915_mmio_reg_offset(PCH_GMBUS2)) in intel_gvt_i2c_handle_gmbus_read()
386 else if (offset == i915_mmio_reg_offset(PCH_GMBUS3)) in intel_gvt_i2c_handle_gmbus_read()
414 if (offset == i915_mmio_reg_offset(PCH_GMBUS0)) in intel_gvt_i2c_handle_gmbus_write()
416 else if (offset == i915_mmio_reg_offset(PCH_GMBUS1)) in intel_gvt_i2c_handle_gmbus_write()
418 else if (offset == i915_mmio_reg_offset(PCH_GMBUS2)) in intel_gvt_i2c_handle_gmbus_write()
420 else if (offset == i915_mmio_reg_offset(PCH_GMBUS3)) in intel_gvt_i2c_handle_gmbus_write()
Dhandlers.c173 ((offset - i915_mmio_reg_offset(FENCE_REG_GEN6_LO(0))) >> 3)
176 (num * 8 + i915_mmio_reg_offset(FENCE_REG_GEN6_LO(0)))
790 reg_nonpriv != i915_mmio_reg_offset(RING_NOPID(engine->mmio_base))) { in force_nonpriv_write()
808 if (offset == i915_mmio_reg_offset(DDI_BUF_CTL(PORT_E))) in ddi_buf_ctl_mmio_write()
887 u32 start = i915_mmio_reg_offset(_start); in calc_index()
888 u32 next = i915_mmio_reg_offset(_next); in calc_index()
889 u32 end = i915_mmio_reg_offset(_end); in calc_index()
1086 if (reg == i915_mmio_reg_offset(DP_AUX_CH_CTL(AUX_CH_A))) in trigger_aux_channel_interrupt()
1088 else if (reg == i915_mmio_reg_offset(PCH_DP_AUX_CH_CTL(AUX_CH_B)) || in trigger_aux_channel_interrupt()
1089 reg == i915_mmio_reg_offset(DP_AUX_CH_CTL(AUX_CH_B))) in trigger_aux_channel_interrupt()
[all …]
Dscheduler.c92 i915_mmio_reg_offset(EU_PERF_CNTL0), in sr_oa_regs()
93 i915_mmio_reg_offset(EU_PERF_CNTL1), in sr_oa_regs()
94 i915_mmio_reg_offset(EU_PERF_CNTL2), in sr_oa_regs()
95 i915_mmio_reg_offset(EU_PERF_CNTL3), in sr_oa_regs()
96 i915_mmio_reg_offset(EU_PERF_CNTL4), in sr_oa_regs()
97 i915_mmio_reg_offset(EU_PERF_CNTL5), in sr_oa_regs()
98 i915_mmio_reg_offset(EU_PERF_CNTL6), in sr_oa_regs()
114 i915_mmio_reg_offset(GEN8_OACTXCONTROL); in sr_oa_regs()
274 vgpu_vreg(vgpu, i915_mmio_reg_offset(reg)) = in save_ring_hw_state()
278 vgpu_vreg(vgpu, i915_mmio_reg_offset(reg)) = in save_ring_hw_state()
[all …]
Dmmio_context.c236 *cs++ = i915_mmio_reg_offset(mmio->reg); in restore_context_mmio_for_inhibit()
266 *cs++ = i915_mmio_reg_offset(GEN9_GFX_MOCS(index)); in restore_render_mocs_control_for_inhibit()
293 *cs++ = i915_mmio_reg_offset(GEN9_LNCFCMOCS(index)); in restore_render_mocs_l3cc_for_inhibit()
549 i915_mmio_reg_offset(mmio->reg), in switch_mmio()
/linux-6.12.1/drivers/gpu/drm/i915/gt/
Dselftest_lrc.c304 i915_mmio_reg_offset(RING_START(engine->mmio_base)), in live_lrc_fixed()
309 i915_mmio_reg_offset(RING_CTL(engine->mmio_base)), in live_lrc_fixed()
314 i915_mmio_reg_offset(RING_HEAD(engine->mmio_base)), in live_lrc_fixed()
319 i915_mmio_reg_offset(RING_TAIL(engine->mmio_base)), in live_lrc_fixed()
324 i915_mmio_reg_offset(RING_MI_MODE(engine->mmio_base)), in live_lrc_fixed()
329 i915_mmio_reg_offset(RING_BBSTATE(engine->mmio_base)), in live_lrc_fixed()
334 i915_mmio_reg_offset(RING_BB_PER_CTX_PTR(engine->mmio_base)), in live_lrc_fixed()
339 i915_mmio_reg_offset(RING_INDIRECT_CTX(engine->mmio_base)), in live_lrc_fixed()
344 i915_mmio_reg_offset(RING_INDIRECT_CTX_OFFSET(engine->mmio_base)), in live_lrc_fixed()
349 i915_mmio_reg_offset(RING_CTX_TIMESTAMP(engine->mmio_base)), in live_lrc_fixed()
[all …]
Dintel_lrc.c1263 *cs++ = i915_mmio_reg_offset(GEN8_RING_CS_GPR(0, 0)); in gen12_emit_timestamp_wa()
1271 *cs++ = i915_mmio_reg_offset(GEN8_RING_CS_GPR(0, 0)); in gen12_emit_timestamp_wa()
1272 *cs++ = i915_mmio_reg_offset(RING_CTX_TIMESTAMP(0)); in gen12_emit_timestamp_wa()
1277 *cs++ = i915_mmio_reg_offset(GEN8_RING_CS_GPR(0, 0)); in gen12_emit_timestamp_wa()
1278 *cs++ = i915_mmio_reg_offset(RING_CTX_TIMESTAMP(0)); in gen12_emit_timestamp_wa()
1291 *cs++ = i915_mmio_reg_offset(GEN8_RING_CS_GPR(0, 0)); in gen12_emit_restore_scratch()
1307 *cs++ = i915_mmio_reg_offset(GEN8_RING_CS_GPR(0, 0)); in gen12_emit_cmd_buf_wa()
1315 *cs++ = i915_mmio_reg_offset(GEN8_RING_CS_GPR(0, 0)); in gen12_emit_cmd_buf_wa()
1316 *cs++ = i915_mmio_reg_offset(RING_CMD_BUF_CCTL(0)); in gen12_emit_cmd_buf_wa()
1332 *cs++ = i915_mmio_reg_offset(DRAW_WATERMARK); in dg2_emit_draw_watermark_setting()
[all …]
Dselftest_workarounds.c157 *cs++ = i915_mmio_reg_offset(RING_FORCE_TO_NONPRIV(base, i)); in read_nonprivs()
184 return i915_mmio_reg_offset(reg); in get_whitelist_reg()
465 u32 reg = i915_mmio_reg_offset(engine->whitelist.list[i].reg); in whitelist_writable_count()
520 u32 reg = i915_mmio_reg_offset(engine->whitelist.list[i].reg); in check_dirty_whitelist()
870 u32 reg = i915_mmio_reg_offset(engine->whitelist.list[i].reg); in read_whitelisted_registers()
906 u32 reg = i915_mmio_reg_offset(engine->whitelist.list[i].reg); in scrub_whitelisted_registers()
961 u32 offset = i915_mmio_reg_offset(reg); in find_reg()
965 i915_mmio_reg_offset(tbl->reg) == offset) in find_reg()
989 i915_mmio_reg_offset(reg), a, b); in result_eq()
1011 i915_mmio_reg_offset(reg), a); in result_neq()
[all …]
Dintel_workarounds.c148 unsigned int addr = i915_mmio_reg_offset(wa->reg); in _wa_add()
177 if (i915_mmio_reg_offset(wal->list[mid].reg) < addr) { in _wa_add()
179 } else if (i915_mmio_reg_offset(wal->list[mid].reg) > addr) { in _wa_add()
187 i915_mmio_reg_offset(wa_->reg), in _wa_add()
206 GEM_BUG_ON(i915_mmio_reg_offset(wa_[0].reg) == in _wa_add()
207 i915_mmio_reg_offset(wa_[1].reg)); in _wa_add()
208 if (i915_mmio_reg_offset(wa_[1].reg) > in _wa_add()
209 i915_mmio_reg_offset(wa_[0].reg)) in _wa_add()
1006 *cs++ = i915_mmio_reg_offset(wa->reg); in intel_engine_emit_ctx_wa()
1716 name, from, i915_mmio_reg_offset(wa->reg), in wa_verify()
[all …]
Dintel_ring_submission.c668 *cs++ = i915_mmio_reg_offset(RING_PP_DIR_DCLV(engine->mmio_base)); in load_pd_dir()
672 *cs++ = i915_mmio_reg_offset(RING_PP_DIR_BASE(engine->mmio_base)); in load_pd_dir()
677 *cs++ = i915_mmio_reg_offset(RING_PP_DIR_BASE(engine->mmio_base)); in load_pd_dir()
682 *cs++ = i915_mmio_reg_offset(RING_INSTPM(engine->mmio_base)); in load_pd_dir()
730 *cs++ = i915_mmio_reg_offset( in mi_set_context()
785 *cs++ = i915_mmio_reg_offset(last_reg); in mi_set_context()
792 *cs++ = i915_mmio_reg_offset(last_reg); in mi_set_context()
827 *cs++ = i915_mmio_reg_offset(GEN7_L3LOG(slice, i)); in remap_l3_slice()
Dselftest_rps.c103 *cs++ = i915_mmio_reg_offset(CS_GPR(i)); in create_spin_counter()
105 *cs++ = i915_mmio_reg_offset(CS_GPR(i)) + 4; in create_spin_counter()
110 *cs++ = i915_mmio_reg_offset(CS_GPR(INC)); in create_spin_counter()
125 *cs++ = i915_mmio_reg_offset(CS_GPR(COUNT)); in create_spin_counter()
208 i915_mmio_reg_offset(BXT_RP_STATE_CAP), in show_pstate_limits()
213 i915_mmio_reg_offset(GEN9_RP_STATE_LIMITS), in show_pstate_limits()
Dselftest_mocs.c152 u32 addr = i915_mmio_reg_offset(GEN9_LNCFCMOCS(0)); in read_l3cc_table()
197 u32 reg = i915_mmio_reg_offset(GEN9_LNCFCMOCS(0)); in check_l3cc_table()
Dgen8_engine_cs.c208 *cs++ = i915_mmio_reg_offset(inv_reg) + gsi_offset; in gen12_emit_aux_table_inv()
216 *cs++ = i915_mmio_reg_offset(inv_reg) + gsi_offset; in gen12_emit_aux_table_inv()
501 *cs++ = i915_mmio_reg_offset(RING_PREDICATE_RESULT(0)); in __xehp_emit_bb_start()
Dgen7_renderclear.c400 batch_add(&cmds, i915_mmio_reg_offset(CACHE_MODE_0_GEN7)); in emit_batch()
405 batch_add(&cmds, i915_mmio_reg_offset(CACHE_MODE_1)); in emit_batch()
/linux-6.12.1/drivers/gpu/drm/i915/
Di915_reg_defs.h282 #define i915_mmio_reg_offset(r) \ macro
284 #define i915_mmio_reg_equal(a, b) (i915_mmio_reg_offset(a) == i915_mmio_reg_offset(b))
Dintel_uncore.h329 u32 offset = i915_mmio_reg_offset(reg); \
339 u32 offset = i915_mmio_reg_offset(reg); \
517 readl(base + i915_mmio_reg_offset(reg))
519 writel(value, base + i915_mmio_reg_offset(reg))
Di915_perf.c1333 *cs++ = i915_mmio_reg_offset(reg); in __store_reg_to_mem()
1496 offset = oa_context_image_offset(ce, i915_mmio_reg_offset(reg)); in set_oa_ctx_ctrl_offset()
1924 *cs++ = i915_mmio_reg_offset(reg) + 4 * d; in save_restore_register()
2021 *cs++ = i915_mmio_reg_offset(CS_GPR(START_TS)) + 4; in alloc_noa_wait()
2024 *cs++ = i915_mmio_reg_offset(RING_TIMESTAMP(base)); in alloc_noa_wait()
2025 *cs++ = i915_mmio_reg_offset(CS_GPR(START_TS)); in alloc_noa_wait()
2039 *cs++ = i915_mmio_reg_offset(CS_GPR(NOW_TS)) + 4; in alloc_noa_wait()
2042 *cs++ = i915_mmio_reg_offset(RING_TIMESTAMP(base)); in alloc_noa_wait()
2043 *cs++ = i915_mmio_reg_offset(CS_GPR(NOW_TS)); in alloc_noa_wait()
2062 *cs++ = i915_mmio_reg_offset(CS_GPR(JUMP_PREDICATE)); in alloc_noa_wait()
[all …]
Di915_ioctl.c55 u32 entry_offset = i915_mmio_reg_offset(entry->offset_ldw); in i915_reg_read_ioctl()
Dintel_device_info.c339 ip_ver_read(i915, i915_mmio_reg_offset(GMD_ID_GRAPHICS), in intel_ipver_early_init()
347 ip_ver_read(i915, i915_mmio_reg_offset(GMD_ID_MEDIA), in intel_ipver_early_init()
Dintel_uncore.c1763 i915_mmio_reg_offset(reg))) in __unclaimed_reg_debug()
1777 i915_mmio_reg_offset(reg)); in __unclaimed_previous_reg_debug()
1860 u32 offset = i915_mmio_reg_offset(reg); \
1917 return __fwtable_reg_read_fw_domains(uncore, i915_mmio_reg_offset(reg)); in fwtable_reg_read_fw_domains()
1966 u32 offset = i915_mmio_reg_offset(reg); \
2007 return __fwtable_reg_write_fw_domains(uncore, i915_mmio_reg_offset(reg)); in fwtable_reg_write_fw_domains()
2077 d->reg_set = uncore->regs + i915_mmio_reg_offset(reg_set) + uncore->gsi_offset; in __fw_domain_init()
2078 d->reg_ack = uncore->regs + i915_mmio_reg_offset(reg_ack) + uncore->gsi_offset; in __fw_domain_init()
/linux-6.12.1/drivers/gpu/drm/i915/gt/uc/
Dintel_guc_ads.c340 i915_mmio_reg_offset(reg), \
366 return guc_mmio_reg_add(gt, regset, i915_mmio_reg_offset(reg), flags); in guc_mcr_reg_add()
421 ret |= GUC_MCR_REG_ADD(gt, regset, MCR_REG(i915_mmio_reg_offset(EU_PERF_CNTL0)), false); in guc_mmio_regset_init()
422 ret |= GUC_MCR_REG_ADD(gt, regset, MCR_REG(i915_mmio_reg_offset(EU_PERF_CNTL1)), false); in guc_mmio_regset_init()
423 ret |= GUC_MCR_REG_ADD(gt, regset, MCR_REG(i915_mmio_reg_offset(EU_PERF_CNTL2)), false); in guc_mmio_regset_init()
424 ret |= GUC_MCR_REG_ADD(gt, regset, MCR_REG(i915_mmio_reg_offset(EU_PERF_CNTL3)), false); in guc_mmio_regset_init()
425 ret |= GUC_MCR_REG_ADD(gt, regset, MCR_REG(i915_mmio_reg_offset(EU_PERF_CNTL4)), false); in guc_mmio_regset_init()
426 ret |= GUC_MCR_REG_ADD(gt, regset, MCR_REG(i915_mmio_reg_offset(EU_PERF_CNTL5)), false); in guc_mmio_regset_init()
427 ret |= GUC_MCR_REG_ADD(gt, regset, MCR_REG(i915_mmio_reg_offset(EU_PERF_CNTL6)), false); in guc_mmio_regset_init()
/linux-6.12.1/drivers/gpu/drm/xe/display/ext/
Di915_irq.c38 i915_mmio_reg_offset(reg), val); in gen3_assert_iir_is_zero()
/linux-6.12.1/drivers/gpu/drm/i915/display/
Dintel_dmc.c510 u32 offset = i915_mmio_reg_offset(reg); in is_dmc_evt_ctl_reg()
511 u32 start = i915_mmio_reg_offset(DMC_EVT_CTL(i915, dmc_id, 0)); in is_dmc_evt_ctl_reg()
512 u32 end = i915_mmio_reg_offset(DMC_EVT_CTL(i915, dmc_id, DMC_EVENT_HANDLER_COUNT_GEN12)); in is_dmc_evt_ctl_reg()
520 u32 offset = i915_mmio_reg_offset(reg); in is_dmc_evt_htp_reg()
521 u32 start = i915_mmio_reg_offset(DMC_EVT_HTP(i915, dmc_id, 0)); in is_dmc_evt_htp_reg()
522 u32 end = i915_mmio_reg_offset(DMC_EVT_HTP(i915, dmc_id, DMC_EVENT_HANDLER_COUNT_GEN12)); in is_dmc_evt_htp_reg()
Dintel_dsb.c241 return prev_opcode == opcode && prev_reg == i915_mmio_reg_offset(reg); in intel_dsb_prev_ins_is_write()
295 i915_mmio_reg_offset(reg)); in intel_dsb_reg_write()
309 i915_mmio_reg_offset(reg)); in intel_dsb_reg_write()
341 i915_mmio_reg_offset(reg)); in intel_dsb_reg_write_masked()

12