/linux-6.12.1/drivers/net/ethernet/intel/iavf/ |
D | iavf_adminq.c | 16 enum iavf_status ret_code; in iavf_alloc_adminq_asq_ring() local 18 ret_code = iavf_allocate_dma_mem(hw, &hw->aq.asq.desc_buf, in iavf_alloc_adminq_asq_ring() 23 if (ret_code) in iavf_alloc_adminq_asq_ring() 24 return ret_code; in iavf_alloc_adminq_asq_ring() 26 ret_code = iavf_allocate_virt_mem(hw, &hw->aq.asq.cmd_buf, in iavf_alloc_adminq_asq_ring() 29 if (ret_code) { in iavf_alloc_adminq_asq_ring() 31 return ret_code; in iavf_alloc_adminq_asq_ring() 34 return ret_code; in iavf_alloc_adminq_asq_ring() 43 enum iavf_status ret_code; in iavf_alloc_adminq_arq_ring() local 45 ret_code = iavf_allocate_dma_mem(hw, &hw->aq.arq.desc_buf, in iavf_alloc_adminq_arq_ring() [all …]
|
/linux-6.12.1/drivers/net/ethernet/intel/i40e/ |
D | i40e_nvm.c | 22 int ret_code = 0; in i40e_init_nvm() local 42 ret_code = -EIO; in i40e_init_nvm() 46 return ret_code; in i40e_init_nvm() 62 int ret_code = 0; in i40e_acquire_nvm() local 67 ret_code = i40e_aq_request_resource(hw, I40E_NVM_RESOURCE_ID, access, in i40e_acquire_nvm() 75 if (ret_code) in i40e_acquire_nvm() 78 access, time_left, ret_code, hw->aq.asq_last_status); in i40e_acquire_nvm() 80 if (ret_code && time_left) { in i40e_acquire_nvm() 86 ret_code = i40e_aq_request_resource(hw, in i40e_acquire_nvm() 90 if (!ret_code) { in i40e_acquire_nvm() [all …]
|
D | i40e_hmc.c | 26 int ret_code = 0; in i40e_add_sd_table_entry() local 30 ret_code = -EINVAL; in i40e_add_sd_table_entry() 36 ret_code = -EINVAL; in i40e_add_sd_table_entry() 49 ret_code = i40e_allocate_dma_mem(hw, &mem, alloc_len, in i40e_add_sd_table_entry() 51 if (ret_code) in i40e_add_sd_table_entry() 55 ret_code = i40e_allocate_virt_mem(hw, in i40e_add_sd_table_entry() 58 if (ret_code) in i40e_add_sd_table_entry() 78 if (ret_code) in i40e_add_sd_table_entry() 82 return ret_code; in i40e_add_sd_table_entry() 112 int ret_code = 0; in i40e_add_pd_table_entry() local [all …]
|
D | i40e_adminq.c | 17 int ret_code; in i40e_alloc_adminq_asq_ring() local 19 ret_code = i40e_allocate_dma_mem(hw, &hw->aq.asq.desc_buf, in i40e_alloc_adminq_asq_ring() 23 if (ret_code) in i40e_alloc_adminq_asq_ring() 24 return ret_code; in i40e_alloc_adminq_asq_ring() 26 ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.cmd_buf, in i40e_alloc_adminq_asq_ring() 29 if (ret_code) { in i40e_alloc_adminq_asq_ring() 31 return ret_code; in i40e_alloc_adminq_asq_ring() 34 return ret_code; in i40e_alloc_adminq_asq_ring() 43 int ret_code; in i40e_alloc_adminq_arq_ring() local 45 ret_code = i40e_allocate_dma_mem(hw, &hw->aq.arq.desc_buf, in i40e_alloc_adminq_arq_ring() [all …]
|
D | i40e_lan_hmc.c | 79 int ret_code = 0; in i40e_init_lan_hmc() local 87 ret_code = i40e_allocate_virt_mem(hw, &hw->hmc.hmc_obj_virt_mem, in i40e_init_lan_hmc() 89 if (ret_code) in i40e_init_lan_hmc() 111 ret_code = -EINVAL; in i40e_init_lan_hmc() 113 txq_num, obj->max_cnt, ret_code); in i40e_init_lan_hmc() 134 ret_code = -EINVAL; in i40e_init_lan_hmc() 136 rxq_num, obj->max_cnt, ret_code); in i40e_init_lan_hmc() 157 ret_code = -EINVAL; in i40e_init_lan_hmc() 159 fcoe_cntx_num, obj->max_cnt, ret_code); in i40e_init_lan_hmc() 180 ret_code = -EINVAL; in i40e_init_lan_hmc() [all …]
|
D | i40e_diag.c | 79 int ret_code = 0; in i40e_diag_reg_test() local 85 !ret_code; i++) { in i40e_diag_reg_test() 102 for (j = 0; j < elements && !ret_code; j++) { in i40e_diag_reg_test() 105 ret_code = i40e_diag_reg_pattern_test(hw, reg, mask); in i40e_diag_reg_test() 109 return ret_code; in i40e_diag_reg_test() 120 int ret_code; in i40e_diag_eeprom_test() local 124 ret_code = i40e_read_nvm_word(hw, I40E_SR_NVM_CONTROL_WORD, ®_val); in i40e_diag_eeprom_test() 125 if (!ret_code && in i40e_diag_eeprom_test()
|
/linux-6.12.1/drivers/platform/x86/ |
D | inspur_platform_profile.c | 92 u8 ret_code[4] = {0, 0, 0, 0}; in inspur_platform_profile_set() local 97 ret_code[0] = INSPUR_TMP_PROFILE_BALANCE; in inspur_platform_profile_set() 100 ret_code[0] = INSPUR_TMP_PROFILE_PERFORMANCE; in inspur_platform_profile_set() 103 ret_code[0] = INSPUR_TMP_PROFILE_POWERSAVE; in inspur_platform_profile_set() 110 ret_code, sizeof(ret_code), in inspur_platform_profile_set() 111 sizeof(ret_code)); in inspur_platform_profile_set() 116 if (ret_code[0]) in inspur_platform_profile_set() 140 u8 ret_code[4] = {0, 0, 0, 0}; in inspur_platform_profile_get() local 144 &ret_code, sizeof(ret_code), in inspur_platform_profile_get() 145 sizeof(ret_code)); in inspur_platform_profile_get() [all …]
|
/linux-6.12.1/drivers/s390/cio/ |
D | vfio_ccw_fsm.c | 189 private->io_region->ret_code = -EIO; in fsm_io_error() 195 private->io_region->ret_code = -EBUSY; in fsm_io_busy() 201 private->io_region->ret_code = -EAGAIN; in fsm_io_retry() 213 cmd_region->ret_code = -EIO; in fsm_async_error() 219 private->cmd_region->ret_code = -EAGAIN; in fsm_async_retry() 260 io_region->ret_code = -EOPNOTSUPP; in fsm_io_request() 268 io_region->ret_code = cp_init(&private->cp, orb); in fsm_io_request() 269 if (io_region->ret_code) { in fsm_io_request() 274 io_region->ret_code); in fsm_io_request() 279 io_region->ret_code = cp_prefetch(&private->cp); in fsm_io_request() [all …]
|
D | vfio_ccw_async.c | 59 ret = region->ret_code ? region->ret_code : count; in vfio_ccw_async_region_write()
|
/linux-6.12.1/drivers/infiniband/hw/irdma/ |
D | pble.c | 88 int ret_code = 0; in add_sd_direct() local 96 ret_code = irdma_add_sd_table_entry(dev->hw, hmc_info, in add_sd_direct() 100 if (ret_code) in add_sd_direct() 101 return ret_code; in add_sd_direct() 213 int ret_code = 0; in add_pble_prm() local 255 ret_code = add_sd_direct(pble_rsrc, &info); in add_pble_prm() 257 if (ret_code) in add_pble_prm() 263 ret_code = add_bp_pages(pble_rsrc, &info); in add_pble_prm() 264 if (ret_code) in add_pble_prm() 270 ret_code = irdma_prm_add_pble_mem(&pble_rsrc->pinfo, chunk); in add_pble_prm() [all …]
|
D | hmc.c | 153 int ret_code = 0; in irdma_hmc_sd_grp() local 173 ret_code = dev->cqp->process_cqp_sds(dev, &sdinfo); in irdma_hmc_sd_grp() 174 if (ret_code) { in irdma_hmc_sd_grp() 177 ret_code); in irdma_hmc_sd_grp() 178 return ret_code; in irdma_hmc_sd_grp() 185 ret_code = dev->cqp->process_cqp_sds(dev, &sdinfo); in irdma_hmc_sd_grp() 187 return ret_code; in irdma_hmc_sd_grp() 229 int ret_code = 0; in irdma_sc_create_hmc_obj() local 256 ret_code = irdma_add_sd_table_entry(dev->hw, info->hmc_info, j, in irdma_sc_create_hmc_obj() 259 if (ret_code) in irdma_sc_create_hmc_obj() [all …]
|
D | ctrl.c | 377 int ret_code; in irdma_sc_qp_init() local 399 ret_code = irdma_uk_qp_init(&qp->qp_uk, &info->qp_uk_init_info); in irdma_sc_qp_init() 400 if (ret_code) in irdma_sc_qp_init() 401 return ret_code; in irdma_sc_qp_init() 419 ret_code = irdma_fragcnt_to_wqesize_rq(qp->qp_uk.max_rq_frag_cnt, in irdma_sc_qp_init() 421 if (ret_code) in irdma_sc_qp_init() 422 return ret_code; in irdma_sc_qp_init() 2523 int ret_code = 0; in irdma_sc_cq_create() local 2534 ret_code = irdma_sc_add_cq_ctx(ceq, cq); in irdma_sc_cq_create() 2536 if (ret_code) in irdma_sc_cq_create() [all …]
|
D | uk.c | 209 int ret_code; in irdma_qp_get_next_recv_wqe() local 214 IRDMA_ATOMIC_RING_MOVE_HEAD(qp->rq_ring, *wqe_idx, ret_code); in irdma_qp_get_next_recv_wqe() 215 if (ret_code) in irdma_qp_get_next_recv_wqe() 240 int ret_code; in irdma_uk_rdma_write() local 259 ret_code = irdma_fragcnt_to_quanta_sq(frag_cnt, &quanta); in irdma_uk_rdma_write() 260 if (ret_code) in irdma_uk_rdma_write() 261 return ret_code; in irdma_uk_rdma_write() 331 int ret_code; in irdma_uk_rdma_read() local 347 ret_code = irdma_fragcnt_to_quanta_sq(op_info->num_lo_sges, &quanta); in irdma_uk_rdma_read() 348 if (ret_code) in irdma_uk_rdma_read() [all …]
|
/linux-6.12.1/drivers/net/ethernet/intel/ice/ |
D | ice_controlq.c | 350 int ret_code; in ice_init_sq() local 354 ret_code = -EBUSY; in ice_init_sq() 360 ret_code = -EIO; in ice_init_sq() 368 ret_code = ice_alloc_ctrlq_sq_ring(hw, cq); in ice_init_sq() 369 if (ret_code) in ice_init_sq() 373 ret_code = ice_alloc_sq_bufs(hw, cq); in ice_init_sq() 374 if (ret_code) in ice_init_sq() 378 ret_code = ice_cfg_sq_regs(hw, cq); in ice_init_sq() 379 if (ret_code) in ice_init_sq() 391 return ret_code; in ice_init_sq() [all …]
|
/linux-6.12.1/include/uapi/linux/ |
D | vfio_ccw.h | 23 __u32 ret_code; member 34 __u32 ret_code; member
|
/linux-6.12.1/drivers/crypto/intel/qat/qat_common/ |
D | adf_heartbeat_dbgfs.c | 50 int ret_code; in adf_hb_status_read() local 57 ret_code = HB_OK; in adf_hb_status_read() 62 ret_code = HB_ERROR; in adf_hb_status_read() 64 len = scnprintf(ret_str, sizeof(ret_str), "%d\n", ret_code); in adf_hb_status_read()
|
/linux-6.12.1/sound/soc/intel/atom/sst/ |
D | sst_pvt.c | 89 if (block->ret_code < 0) { in sst_wait_interruptible() 91 "stream failed %d\n", block->ret_code); in sst_wait_interruptible() 133 block->ret_code); in sst_wait_timeout() 134 retval = -block->ret_code; in sst_wait_timeout()
|
/linux-6.12.1/drivers/usb/typec/ucsi/ |
D | ucsi_glink.c | 36 u32 ret_code; member 47 u32 ret_code; member 215 if (resp->ret_code) in pmic_glink_ucsi_read_ack() 226 if (resp->ret_code) in pmic_glink_ucsi_write_ack()
|
/linux-6.12.1/include/linux/ |
D | drbd_genl_api.h | 28 __s32 ret_code; member
|
/linux-6.12.1/drivers/mtd/ |
D | mtdcore.c | 1702 int ret_code; in mtd_read_oob() local 1706 ret_code = mtd_check_oob_ops(mtd, from, ops); in mtd_read_oob() 1707 if (ret_code) in mtd_read_oob() 1708 return ret_code; in mtd_read_oob() 1720 ret_code = mtd_io_emulated_slc(mtd, from, true, ops); in mtd_read_oob() 1722 ret_code = mtd_read_oob_std(mtd, from, ops); in mtd_read_oob() 1732 if (unlikely(ret_code < 0)) in mtd_read_oob() 1733 return ret_code; in mtd_read_oob() 1737 ops->stats->max_bitflips = ret_code; in mtd_read_oob() 1738 return ret_code >= mtd->bitflip_threshold ? -EUCLEAN : 0; in mtd_read_oob()
|
/linux-6.12.1/drivers/scsi/bnx2i/ |
D | bnx2i_iscsi.c | 1414 int ret_code; in bnx2i_conn_bind() local 1424 ret_code = -EIO; in bnx2i_conn_bind() 1432 ret_code = -EINVAL; in bnx2i_conn_bind() 1437 ret_code = -EINVAL; in bnx2i_conn_bind() 1450 ret_code = -EEXIST; in bnx2i_conn_bind() 1458 ret_code = bnx2i_bind_conn_to_iscsi_cid(hba, bnx2i_conn, in bnx2i_conn_bind() 1470 return ret_code; in bnx2i_conn_bind()
|
/linux-6.12.1/arch/s390/include/uapi/asm/ |
D | zcrypt.h | 170 __u32 ret_code; member
|
/linux-6.12.1/drivers/gpu/drm/amd/amdgpu/ |
D | vcn_v1_0.c | 1255 int ret_code; in vcn_v1_0_pause_dpg_mode() local 1271 ret_code = 0; in vcn_v1_0_pause_dpg_mode() 1274 ret_code = SOC15_WAIT_ON_RREG(UVD, 0, mmUVD_POWER_STATUS, in vcn_v1_0_pause_dpg_mode() 1278 if (!ret_code) { in vcn_v1_0_pause_dpg_mode() 1327 ret_code = 0; in vcn_v1_0_pause_dpg_mode() 1330 ret_code = SOC15_WAIT_ON_RREG(UVD, 0, mmUVD_POWER_STATUS, in vcn_v1_0_pause_dpg_mode() 1334 if (!ret_code) { in vcn_v1_0_pause_dpg_mode()
|
/linux-6.12.1/drivers/block/drbd/ |
D | drbd_nl.c | 188 adm_ctx->reply_dh->ret_code = NO_ERROR; in drbd_adm_prepare() 322 adm_ctx->reply_dh->ret_code = retcode; in drbd_adm_finish() 3268 dh->ret_code = NO_ERROR; in drbd_adm_dump_resources() 3390 dh->ret_code = retcode; in drbd_adm_dump_devices() 3516 dh->ret_code = retcode; in drbd_adm_dump_connections() 3658 dh->ret_code = retcode; in drbd_adm_dump_peer_devices() 3926 dh->ret_code = NO_ERROR; in get_one_status() 3944 dh->ret_code = NO_ERROR; in get_one_status() 4525 d_out->ret_code = NO_ERROR; in drbd_bcast_event() 4580 dh->ret_code = NO_ERROR; in notify_resource_state() [all …]
|
/linux-6.12.1/drivers/net/ethernet/ti/ |
D | davinci_emac.c | 953 int ret_code; in emac_dev_xmit() local 963 ret_code = skb_put_padto(skb, EMAC_DEF_MIN_ETHPKTSIZE); in emac_dev_xmit() 964 if (unlikely(ret_code < 0)) { in emac_dev_xmit() 972 ret_code = cpdma_chan_submit(priv->txchan, skb, skb->data, skb->len, in emac_dev_xmit() 974 if (unlikely(ret_code != 0)) { in emac_dev_xmit()
|