Home
last modified time | relevance | path

Searched refs:xgmi (Results 1 – 25 of 36) sorted by relevance

12

/linux-6.12.1/drivers/gpu/drm/amd/amdgpu/
Damdgpu_xgmi.c371 return sysfs_emit(buf, "%llu\n", adev->gmc.xgmi.node_id); in amdgpu_xgmi_show_device_id()
382 return sysfs_emit(buf, "%u\n", adev->gmc.xgmi.physical_node_id); in amdgpu_xgmi_show_physical_id()
431 if (top->nodes[i].node_id == adev->gmc.xgmi.node_id) { in amdgpu_xgmi_show_connected_port_num()
599 if (!adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
610 if (hive->hive_id == adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
662 hive->hive_id = adev->gmc.xgmi.hive_id; in amdgpu_get_xgmi_hive()
733 request_adev->gmc.xgmi.node_id, in amdgpu_xgmi_set_pstate()
734 request_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_set_pstate()
765 adev->gmc.xgmi.node_id, in amdgpu_xgmi_update_topology()
766 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_update_topology()
[all …]
Dgfxhub_v1_1.c88 if (max_region || adev->gmc.xgmi.connected_to_cpu) { in gfxhub_v1_1_get_xgmi_info()
89 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v1_1_get_xgmi_info()
91 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes) in gfxhub_v1_1_get_xgmi_info()
95 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_1_get_xgmi_info()
99 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_1_get_xgmi_info()
104 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id) in gfxhub_v1_1_get_xgmi_info()
107 adev->gmc.xgmi.node_segment_size = seg_size; in gfxhub_v1_1_get_xgmi_info()
Damdgpu_xgmi.h74 adev->gmc.xgmi.hive_id && in amdgpu_xgmi_same_hive()
75 adev->gmc.xgmi.hive_id == bo_adev->gmc.xgmi.hive_id); in amdgpu_xgmi_same_hive()
Dgmc_v9_0.c754 !adev->gmc.xgmi.connected_to_cpu && in gmc_v9_0_set_irq_funcs()
1158 adev->gmc.xgmi.connected_to_cpu) in gmc_v9_0_get_coherence_flags()
1450 if (!adev->gmc.xgmi.connected_to_cpu) in gmc_v9_0_set_umc_funcs()
1466 if (!adev->gmc.xgmi.connected_to_cpu && !adev->gmc.is_app_apu) in gmc_v9_0_set_umc_funcs()
1534 if (!adev->gmc.xgmi.connected_to_cpu) { in gmc_v9_0_set_mca_ras_funcs()
1547 if (!adev->gmc.xgmi.connected_to_cpu) in gmc_v9_0_set_xgmi_ras_funcs()
1548 adev->gmc.xgmi.ras = &xgmi_ras; in gmc_v9_0_set_xgmi_ras_funcs()
1563 adev->gmc.xgmi.supported = true; in gmc_v9_0_early_init()
1566 adev->gmc.xgmi.supported = true; in gmc_v9_0_early_init()
1567 adev->gmc.xgmi.connected_to_cpu = in gmc_v9_0_early_init()
[all …]
Damdgpu_gmc.c53 u64 vram_size = adev->gmc.xgmi.node_segment_size * adev->gmc.xgmi.num_physical_nodes; in amdgpu_gmc_pdb0_alloc()
223 if (mc->xgmi.num_physical_nodes == 0) { in amdgpu_gmc_vram_location()
251 u64 hive_vram_end = mc->xgmi.node_segment_size * mc->xgmi.num_physical_nodes - 1; in amdgpu_gmc_sysvm_location()
252 mc->vram_start = mc->xgmi.node_segment_size * mc->xgmi.physical_node_id; in amdgpu_gmc_sysvm_location()
253 mc->vram_end = mc->vram_start + mc->xgmi.node_segment_size - 1; in amdgpu_gmc_sysvm_location()
1010 u64 vram_size = adev->gmc.xgmi.node_segment_size * adev->gmc.xgmi.num_physical_nodes; in amdgpu_gmc_init_pdb0()
1013 adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in amdgpu_gmc_init_pdb0()
Dgfxhub_v1_2.c258 if (adev->gmc.xgmi.connected_to_cpu || adev->gmc.is_app_apu) { in gfxhub_v1_2_xcc_init_cache_regs()
606 if (max_region || adev->gmc.xgmi.connected_to_cpu) { in gfxhub_v1_2_get_xgmi_info()
607 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v1_2_get_xgmi_info()
609 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes) in gfxhub_v1_2_get_xgmi_info()
612 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_2_get_xgmi_info()
616 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id) in gfxhub_v1_2_get_xgmi_info()
619 adev->gmc.xgmi.node_segment_size = seg_size; in gfxhub_v1_2_get_xgmi_info()
Daldebaran.c39 adev->gmc.xgmi.connected_to_cpu)) in aldebaran_is_mode2_default()
173 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in aldebaran_mode2_perform_reset()
190 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in aldebaran_mode2_perform_reset()
385 tmp_adev->gmc.xgmi.num_physical_nodes > 1) in aldebaran_mode2_restore_hwcontext()
Damdgpu_device.c1658 if (adev->gmc.xgmi.pending_reset) in amdgpu_device_need_post()
2905 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_device_ip_init()
2938 if (!adev->gmc.xgmi.pending_reset) { in amdgpu_device_ip_init()
3168 ((adev->asic_type == CHIP_ARCTURUS && adev->gmc.xgmi.num_physical_nodes > 1) || in amdgpu_device_ip_late_init()
3172 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_device_ip_late_init()
3188 if (mgpu_info.num_dgpu == adev->gmc.xgmi.num_physical_nodes) { in amdgpu_device_ip_late_init()
3302 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_fini()
3452 if (adev->gmc.xgmi.pending_reset && in amdgpu_device_ip_suspend_phase2()
4218 if (adev->gmc.xgmi.supported) { in amdgpu_device_init()
4266 if (adev->gmc.xgmi.num_physical_nodes) { in amdgpu_device_init()
[all …]
Dgfxhub_v2_1.c527 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v2_1_get_xgmi_info()
528 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes) in gfxhub_v2_1_get_xgmi_info()
531 adev->gmc.xgmi.physical_node_id = in gfxhub_v2_1_get_xgmi_info()
533 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id) in gfxhub_v2_1_get_xgmi_info()
536 adev->gmc.xgmi.node_segment_size = REG_GET_FIELD( in gfxhub_v2_1_get_xgmi_info()
Damdgpu_amdkfd.c567 adev->gmc.xgmi.physical_node_id, in amdgpu_amdkfd_get_xgmi_hops_count()
568 peer_adev->gmc.xgmi.physical_node_id, ret); in amdgpu_amdkfd_get_xgmi_hops_count()
591 adev->gmc.xgmi.physical_node_id, in amdgpu_amdkfd_get_xgmi_bandwidth_mbytes()
592 peer_adev->gmc.xgmi.physical_node_id, num_links); in amdgpu_amdkfd_get_xgmi_bandwidth_mbytes()
Dgmc_v10_0.c678 base += adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v10_0_vram_gtt_location()
691 adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v10_0_vram_gtt_location()
Damdgpu_ras.c1504 if (adev->gmc.xgmi.num_physical_nodes > 1 && in amdgpu_ras_error_inject()
2524 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) in amdgpu_ras_set_fed_all()
2575 gmc.xgmi.head) in amdgpu_ras_do_recovery()
2584 if (hive && adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_ras_do_recovery()
2588 list_add_tail(&adev->gmc.xgmi.head, &device_list); in amdgpu_ras_do_recovery()
2594 device_list_handle, gmc.xgmi.head) { in amdgpu_ras_do_recovery()
3188 if (adev->gmc.xgmi.pending_reset) in amdgpu_ras_recovery_init()
3228 (adev->gmc.xgmi.connected_to_cpu)) in amdgpu_ras_recovery_init()
3384 if (!adev->gmc.xgmi.num_physical_nodes) in amdgpu_ras_query_ras_capablity_from_vbios()
3402 if (adev->gmc.xgmi.connected_to_cpu || in amdgpu_ras_query_poison_mode()
[all …]
Damdgpu_job.c67 if (!amdgpu_sriov_vf(adev) && (adev->gmc.xgmi.num_physical_nodes > 1) && hive) { in amdgpu_job_core_dump()
68 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) in amdgpu_job_core_dump()
Damdgpu_psp.c440 !adev->gmc.xgmi.connected_to_cpu && in psp_sw_init()
1260 adev->gmc.xgmi.connected_to_cpu)) in psp_xgmi_terminate()
1377 uint64_t src_node_id = psp->adev->gmc.xgmi.node_id; in psp_xgmi_reflect_topology_info()
1386 list_for_each_entry(mirror_adev, &hive->device_list, gmc.xgmi.head) { in psp_xgmi_reflect_topology_info()
1390 if (mirror_adev->gmc.xgmi.node_id != dst_node_id) in psp_xgmi_reflect_topology_info()
1805 if (!adev->gmc.xgmi.connected_to_cpu && !adev->gmc.is_app_apu) in psp_ras_initialize()
2909 if (adev->gmc.xgmi.num_physical_nodes > 1) { in psp_load_fw()
3002 if (adev->gmc.xgmi.num_physical_nodes > 1) in psp_hw_fini()
3020 if (adev->gmc.xgmi.num_physical_nodes > 1 && in psp_suspend()
3115 if (adev->gmc.xgmi.num_physical_nodes > 1) { in psp_resume()
[all …]
Damdgpu_ib.c400 } else if (adev->gmc.xgmi.hive_id) { in amdgpu_ib_ring_tests()
Damdgpu_gmc.h309 struct amdgpu_xgmi xgmi; member
Damdgpu_ucode.h150 struct psp_fw_legacy_bin_desc xgmi; member
Damdgpu_umc.c213 if (adev->gmc.xgmi.connected_to_cpu || in amdgpu_umc_pasid_poison_handler()
Damdgpu_object.c960 if (!adev->gmc.xgmi.connected_to_cpu && !adev->gmc.is_app_apu) { in amdgpu_bo_init()
996 if (!adev->gmc.xgmi.connected_to_cpu && !adev->gmc.is_app_apu) { in amdgpu_bo_fini()
Dgfxhub_v1_0.c208 if (adev->gmc.xgmi.connected_to_cpu) { in gfxhub_v1_0_init_cache_regs()
Dsoc15.c510 if (adev->gmc.xgmi.supported && adev->gmc.xgmi.connected_to_cpu) in soc15_asic_reset_method()
/linux-6.12.1/Documentation/gpu/amdgpu/
Dindex.rst14 xgmi
/linux-6.12.1/drivers/gpu/drm/amd/amdkfd/
Dkfd_migrate.c702 if (adev->gmc.xgmi.connected_to_cpu) in svm_migrate_vma_to_ram()
1036 if (adev->gmc.xgmi.connected_to_cpu) { in kgd2kfd_init_zone_device()
Dkfd_topology.c1287 int num_xgmi_nodes = adev->gmc.xgmi.num_physical_nodes; in kfd_set_recommended_sdma_engines()
1294 int src_socket_id = adev->gmc.xgmi.physical_node_id; in kfd_set_recommended_sdma_engines()
1295 int dst_socket_id = to_dev->gpu->adev->gmc.xgmi.physical_node_id; in kfd_set_recommended_sdma_engines()
2171 dev->gpu->adev->gmc.xgmi.connected_to_cpu) in kfd_topology_add_device()
Dkfd_device.c801 kfd->hive_id = kfd->adev->gmc.xgmi.hive_id; in kgd2kfd_device_init()
1415 if (!node->adev->gmc.xgmi.supported) in kfd_get_num_sdma_engines()

12