Lines Matching refs:err_data

32 				    struct ras_err_data *err_data, uint64_t err_addr,  in amdgpu_umc_convert_error_address()  argument
38 err_data, err_addr, ch_inst, umc_inst); in amdgpu_umc_convert_error_address()
52 struct ras_err_data err_data; in amdgpu_umc_page_retirement_mca() local
55 ret = amdgpu_ras_error_data_init(&err_data); in amdgpu_umc_page_retirement_mca()
59 err_data.err_addr = in amdgpu_umc_page_retirement_mca()
62 if (!err_data.err_addr) { in amdgpu_umc_page_retirement_mca()
69 err_data.err_addr_len = adev->umc.max_ras_err_cnt_per_query; in amdgpu_umc_page_retirement_mca()
74 ret = amdgpu_umc_convert_error_address(adev, &err_data, err_addr, in amdgpu_umc_page_retirement_mca()
80 amdgpu_ras_add_bad_pages(adev, err_data.err_addr, in amdgpu_umc_page_retirement_mca()
81 err_data.err_addr_cnt); in amdgpu_umc_page_retirement_mca()
86 kfree(err_data.err_addr); in amdgpu_umc_page_retirement_mca()
89 amdgpu_ras_error_data_fini(&err_data); in amdgpu_umc_page_retirement_mca()
97 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in amdgpu_umc_handle_bad_pages() local
116 err_data->err_addr = in amdgpu_umc_handle_bad_pages()
123 if(!err_data->err_addr) in amdgpu_umc_handle_bad_pages()
127 err_data->err_addr_len = adev->umc.max_ras_err_cnt_per_query; in amdgpu_umc_handle_bad_pages()
143 err_data->err_addr = in amdgpu_umc_handle_bad_pages()
150 if(!err_data->err_addr) in amdgpu_umc_handle_bad_pages()
154 err_data->err_addr_len = adev->umc.max_ras_err_cnt_per_query; in amdgpu_umc_handle_bad_pages()
164 if (err_data->ue_count || err_data->de_count) { in amdgpu_umc_handle_bad_pages()
165 err_count = err_data->ue_count + err_data->de_count; in amdgpu_umc_handle_bad_pages()
167 err_data->err_addr_cnt) { in amdgpu_umc_handle_bad_pages()
168 amdgpu_ras_add_bad_pages(adev, err_data->err_addr, in amdgpu_umc_handle_bad_pages()
169 err_data->err_addr_cnt); in amdgpu_umc_handle_bad_pages()
181 kfree(err_data->err_addr); in amdgpu_umc_handle_bad_pages()
182 err_data->err_addr = NULL; in amdgpu_umc_handle_bad_pages()
192 struct ras_err_data *err_data = (struct ras_err_data *)ras_error_status; in amdgpu_umc_do_page_retirement() local
198 if ((err_data->ue_count || err_data->de_count) && in amdgpu_umc_do_page_retirement()
227 struct ras_err_data err_data; in amdgpu_umc_pasid_poison_handler() local
233 ret = amdgpu_ras_error_data_init(&err_data); in amdgpu_umc_pasid_poison_handler()
237 ret = amdgpu_umc_do_page_retirement(adev, &err_data, NULL, reset); in amdgpu_umc_pasid_poison_handler()
240 obj->err_data.ue_count += err_data.ue_count; in amdgpu_umc_pasid_poison_handler()
241 obj->err_data.ce_count += err_data.ce_count; in amdgpu_umc_pasid_poison_handler()
242 obj->err_data.de_count += err_data.de_count; in amdgpu_umc_pasid_poison_handler()
245 amdgpu_ras_error_data_fini(&err_data); in amdgpu_umc_pasid_poison_handler()
357 int amdgpu_umc_fill_error_record(struct ras_err_data *err_data, in amdgpu_umc_fill_error_record() argument
365 if (!err_data || in amdgpu_umc_fill_error_record()
366 !err_data->err_addr || in amdgpu_umc_fill_error_record()
367 (err_data->err_addr_cnt >= err_data->err_addr_len)) in amdgpu_umc_fill_error_record()
370 err_rec = &err_data->err_addr[err_data->err_addr_cnt]; in amdgpu_umc_fill_error_record()
381 err_data->err_addr_cnt++; in amdgpu_umc_fill_error_record()