Lines Matching refs:xgmi

371 	return sysfs_emit(buf, "%llu\n", adev->gmc.xgmi.node_id);  in amdgpu_xgmi_show_device_id()
382 return sysfs_emit(buf, "%u\n", adev->gmc.xgmi.physical_node_id); in amdgpu_xgmi_show_physical_id()
431 if (top->nodes[i].node_id == adev->gmc.xgmi.node_id) { in amdgpu_xgmi_show_connected_port_num()
599 if (!adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
610 if (hive->hive_id == adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
662 hive->hive_id = adev->gmc.xgmi.hive_id; in amdgpu_get_xgmi_hive()
733 request_adev->gmc.xgmi.node_id, in amdgpu_xgmi_set_pstate()
734 request_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_set_pstate()
765 adev->gmc.xgmi.node_id, in amdgpu_xgmi_update_topology()
766 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_update_topology()
786 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_hops_count()
798 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_num_links()
815 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_initialize_hive_get_data_partition()
836 if (peer_info->nodes[i].node_id == adev->gmc.xgmi.node_id) { in amdgpu_xgmi_fill_topology_info()
838 if (top_info->nodes[j].node_id == peer_adev->gmc.xgmi.node_id) { in amdgpu_xgmi_fill_topology_info()
860 if (!adev->gmc.xgmi.supported) in amdgpu_xgmi_add_device()
863 if (!adev->gmc.xgmi.pending_reset && in amdgpu_xgmi_add_device()
872 ret = psp_xgmi_get_hive_id(&adev->psp, &adev->gmc.xgmi.hive_id); in amdgpu_xgmi_add_device()
879 ret = psp_xgmi_get_node_id(&adev->psp, &adev->gmc.xgmi.node_id); in amdgpu_xgmi_add_device()
886 adev->gmc.xgmi.hive_id = 16; in amdgpu_xgmi_add_device()
887 adev->gmc.xgmi.node_id = adev->gmc.xgmi.physical_node_id + 16; in amdgpu_xgmi_add_device()
895 adev->gmc.xgmi.node_id, adev->gmc.xgmi.hive_id); in amdgpu_xgmi_add_device()
902 list_add_tail(&adev->gmc.xgmi.head, &hive->device_list); in amdgpu_xgmi_add_device()
910 if (!adev->gmc.xgmi.pending_reset && in amdgpu_xgmi_add_device()
912 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_add_device()
917 adev->gmc.xgmi.node_id; in amdgpu_xgmi_add_device()
933 adev->gmc.xgmi.node_id, in amdgpu_xgmi_add_device()
934 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_add_device()
940 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_add_device()
945 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_add_device()
951 tmp_adev->gmc.xgmi.node_id, in amdgpu_xgmi_add_device()
952 tmp_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_add_device()
968 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_xgmi_add_device()
974 tmp_adev->gmc.xgmi.node_id, in amdgpu_xgmi_add_device()
975 tmp_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_add_device()
988 if (!ret && !adev->gmc.xgmi.pending_reset) in amdgpu_xgmi_add_device()
997 adev->gmc.xgmi.physical_node_id, adev->gmc.xgmi.hive_id); in amdgpu_xgmi_add_device()
1001 adev->gmc.xgmi.physical_node_id, adev->gmc.xgmi.hive_id, in amdgpu_xgmi_add_device()
1012 if (!adev->gmc.xgmi.supported) in amdgpu_xgmi_remove_device()
1023 list_del(&adev->gmc.xgmi.head); in amdgpu_xgmi_remove_device()
1096 if (!adev->gmc.xgmi.supported || in amdgpu_xgmi_ras_late_init()
1097 adev->gmc.xgmi.num_physical_nodes == 0) in amdgpu_xgmi_ras_late_init()
1128 struct amdgpu_xgmi *xgmi = &adev->gmc.xgmi; in amdgpu_xgmi_get_relative_phy_addr() local
1129 return (addr + xgmi->physical_node_id * xgmi->node_segment_size); in amdgpu_xgmi_get_relative_phy_addr()
1486 if (!adev->gmc.xgmi.ras) in amdgpu_xgmi_ras_sw_init()
1489 ras = adev->gmc.xgmi.ras; in amdgpu_xgmi_ras_sw_init()
1499 adev->gmc.xgmi.ras_if = &ras->ras_block.ras_comm; in amdgpu_xgmi_ras_sw_init()