/linux-6.12.1/drivers/infiniband/sw/rxe/ |
D | rxe_req.c | 18 struct rxe_send_wqe *wqe, int npsn) in retry_first_write_send() 40 struct rxe_send_wqe *wqe; in req_retry() local 121 struct rxe_send_wqe *wqe; in req_check_sq_drain_done() local 177 struct rxe_send_wqe *wqe; in req_next_wqe() local 206 static int rxe_wqe_is_fenced(struct rxe_qp *qp, struct rxe_send_wqe *wqe) in rxe_wqe_is_fenced() 354 static int next_opcode(struct rxe_qp *qp, struct rxe_send_wqe *wqe, in next_opcode() 384 static inline int check_init_depth(struct rxe_qp *qp, struct rxe_send_wqe *wqe) in check_init_depth() 416 struct rxe_send_wqe *wqe, in init_req_packet() 500 struct rxe_send_wqe *wqe, struct rxe_pkt_info *pkt, in finish_packet() 543 struct rxe_send_wqe *wqe, in update_wqe_state() [all …]
|
D | rxe_comp.c | 141 struct rxe_send_wqe *wqe; in get_wqe() local 174 struct rxe_send_wqe *wqe) in check_psn() 217 struct rxe_send_wqe *wqe) in check_ack() 356 struct rxe_send_wqe *wqe) in do_read() 376 struct rxe_send_wqe *wqe) in do_atomic() 393 static void make_send_cqe(struct rxe_qp *qp, struct rxe_send_wqe *wqe, in make_send_cqe() 440 static void do_complete(struct rxe_qp *qp, struct rxe_send_wqe *wqe) in do_complete() 501 struct rxe_send_wqe *wqe) in complete_ack() 525 struct rxe_send_wqe *wqe) in complete_wqe() 557 static int flush_send_wqe(struct rxe_qp *qp, struct rxe_send_wqe *wqe) in flush_send_wqe() [all …]
|
D | rxe_mw.c | 50 static int rxe_check_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe, in rxe_check_bind_mw() 135 static void rxe_do_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe, in rxe_do_bind_mw() 164 int rxe_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe) in rxe_bind_mw()
|
D | rxe_resp.c | 263 struct rxe_recv_wqe *wqe; in get_srq_wqe() local 1060 struct rxe_recv_wqe *wqe = qp->resp.wqe; in do_complete() local 1429 static int flush_recv_wqe(struct rxe_qp *qp, struct rxe_recv_wqe *wqe) in flush_recv_wqe() 1460 struct rxe_recv_wqe *wqe; in flush_recv_queue() local
|
/linux-6.12.1/drivers/infiniband/hw/irdma/ |
D | ctrl.c | 186 __le64 *wqe; in irdma_sc_add_arp_cache_entry() local 222 __le64 *wqe; in irdma_sc_del_arp_cache_entry() local 256 __le64 *wqe; in irdma_sc_manage_apbvt_entry() local 304 __le64 *wqe; in irdma_sc_manage_qhash_table_entry() local 452 __le64 *wqe; in irdma_sc_qp_create() local 502 __le64 *wqe; in irdma_sc_qp_modify() local 572 __le64 *wqe; in irdma_sc_qp_destroy() local 764 __le64 *wqe; in irdma_sc_alloc_local_mac_entry() local 799 __le64 *wqe; in irdma_sc_add_local_mac_entry() local 837 __le64 *wqe; in irdma_sc_del_local_mac_entry() local [all …]
|
D | uk.c | 15 static void irdma_set_fragment(__le64 *wqe, u32 offset, struct ib_sge *sge, in irdma_set_fragment() 39 static void irdma_set_fragment_gen_1(__le64 *wqe, u32 offset, in irdma_set_fragment_gen_1() 61 __le64 *wqe; in irdma_nop_1() local 158 __le64 *wqe; in irdma_qp_get_next_send_wqe() local 208 __le64 *wqe; in irdma_qp_get_next_recv_wqe() local 236 __le64 *wqe; in irdma_uk_rdma_write() local 335 __le64 *wqe; in irdma_uk_rdma_read() local 410 __le64 *wqe; in irdma_uk_send() local 501 static void irdma_set_mw_bind_wqe_gen_1(__le64 *wqe, in irdma_set_mw_bind_wqe_gen_1() 518 static void irdma_copy_inline_data_gen_1(u8 *wqe, struct ib_sge *sge_list, in irdma_copy_inline_data_gen_1() [all …]
|
D | uda.c | 23 __le64 *wqe; in irdma_sc_access_ah() local 119 __le64 *wqe; in irdma_access_mcast_grp() local
|
D | puda.c | 82 __le64 *wqe; in irdma_puda_post_recvbuf() local 427 __le64 *wqe; in irdma_puda_send() local 608 __le64 *wqe; in irdma_puda_qp_wqe() local 725 __le64 *wqe; in irdma_puda_cq_wqe() local 1123 __le64 *wqe; in irdma_ilq_putback_rcvbuf() local
|
/linux-6.12.1/drivers/infiniband/hw/qib/ |
D | qib_rc.c | 42 static u32 restart_sge(struct rvt_sge_state *ss, struct rvt_swqe *wqe, in restart_sge() 222 struct rvt_swqe *wqe; in qib_make_rc_req() local 739 struct rvt_swqe *wqe = rvt_get_swqe_ptr(qp, n); in reset_psn() local 823 struct rvt_swqe *wqe = rvt_get_swqe_ptr(qp, qp->s_acked); in qib_restart_rc() local 859 struct rvt_swqe *wqe; in reset_sending_psn() local 885 struct rvt_swqe *wqe; in qib_rc_send_complete() local 952 struct rvt_swqe *wqe, in do_rc_completion() 1012 struct rvt_swqe *wqe; in do_rc_ack() local 1227 struct rvt_swqe *wqe; in rdma_seq_err() local 1277 struct rvt_swqe *wqe; in qib_rc_rcv_resp() local
|
/linux-6.12.1/drivers/infiniband/sw/siw/ |
D | siw_qp_tx.c | 51 struct siw_wqe *wqe = &c_tx->wqe_active; in siw_try_1seg() local 119 struct siw_wqe *wqe = &c_tx->wqe_active; in siw_qp_prepare_tx() local 434 struct siw_wqe *wqe = &c_tx->wqe_active; in siw_tx_hdt() local 702 static void siw_prepare_fpdu(struct siw_qp *qp, struct siw_wqe *wqe) in siw_prepare_fpdu() 760 static int siw_check_sgl_tx(struct ib_pd *pd, struct siw_wqe *wqe, in siw_check_sgl_tx() 790 static int siw_qp_sq_proc_tx(struct siw_qp *qp, struct siw_wqe *wqe) in siw_qp_sq_proc_tx() 975 static int siw_qp_sq_proc_local(struct siw_qp *qp, struct siw_wqe *wqe) in siw_qp_sq_proc_local() 1022 struct siw_wqe *wqe = tx_wqe(qp); in siw_qp_sq_process() local
|
D | siw_qp_rx.c | 169 struct siw_wqe *wqe = &frx->wqe_active; in siw_rresp_check_ntoh() local 281 struct siw_wqe *wqe = &frx->wqe_active; in siw_send_check_ntoh() local 334 struct siw_wqe *wqe = NULL; in siw_rqe_get() local 439 struct siw_wqe *wqe; in siw_proc_send() local 565 struct siw_wqe *wqe = rx_wqe(frx); in siw_proc_write() local 742 struct siw_wqe *wqe = NULL; in siw_orqe_start_rx() local 786 struct siw_wqe *wqe = rx_wqe(frx); in siw_proc_rresp() local 1202 struct siw_wqe *wqe = rx_wqe(qp->rx_fpdu); in siw_rdmap_complete() local
|
D | siw_qp.c | 265 struct siw_wqe *wqe = tx_wqe(qp); in siw_qp_mpa_rts() local 465 struct siw_wqe *wqe = tx_wqe(qp); in siw_send_terminate() local 887 struct siw_wqe *wqe = tx_wqe(qp); in siw_activate_tx_from_sq() local 983 struct siw_wqe *wqe = tx_wqe(qp); in siw_activate_tx() local 1194 struct siw_wqe *wqe = tx_wqe(qp); in siw_sq_flush() local 1269 struct siw_wqe *wqe = &qp->rx_untagged.wqe_active; in siw_rq_flush() local
|
/linux-6.12.1/drivers/infiniband/hw/cxgb4/ |
D | qp.c | 489 static int build_rdma_send(struct t4_sq *sq, union t4_wr *wqe, in build_rdma_send() 556 static int build_rdma_write(struct t4_sq *sq, union t4_wr *wqe, in build_rdma_write() 660 static int build_rdma_read(union t4_wr *wqe, const struct ib_send_wr *wr, in build_rdma_read() 697 union t4_wr *wqe; in post_write_cmpl() local 759 static int build_rdma_recv(struct c4iw_qp *qhp, union t4_recv_wr *wqe, in build_rdma_recv() 774 static int build_srq_recv(union t4_recv_wr *wqe, const struct ib_recv_wr *wr, in build_srq_recv() 820 static int build_memreg(struct t4_sq *sq, union t4_wr *wqe, in build_memreg() 884 static int build_inv_stag(union t4_wr *wqe, const struct ib_send_wr *wr, in build_inv_stag() 1086 union t4_wr *wqe = NULL; in c4iw_post_send() local 1266 union t4_recv_wr *wqe = NULL; in c4iw_post_receive() local [all …]
|
/linux-6.12.1/drivers/infiniband/hw/mlx5/ |
D | umr.c | 260 struct mlx5r_umr_wqe *wqe, bool with_data) in mlx5r_umr_post_send() 321 struct mlx5r_umr_wqe *wqe, bool with_data) in mlx5r_umr_post_send_wait() 393 struct mlx5r_umr_wqe wqe = {}; in mlx5r_umr_revoke_mr() local 433 struct mlx5r_umr_wqe wqe = {}; in mlx5r_umr_rereg_pd_access() local 606 struct mlx5r_umr_wqe *wqe, in mlx5r_umr_final_update_xlt() 644 struct mlx5r_umr_wqe wqe = {}; in _mlx5r_umr_update_mr_pas() local 763 struct mlx5r_umr_wqe wqe = {}; in mlx5r_umr_update_xlt() local
|
/linux-6.12.1/drivers/infiniband/hw/hfi1/ |
D | rc.c | 394 struct rvt_swqe *wqe; in hfi1_make_rc_req() local 1418 struct rvt_swqe *wqe) in update_num_rd_atomic() 1461 struct rvt_swqe *wqe = rvt_get_swqe_ptr(qp, n); in reset_psn() local 1567 struct rvt_swqe *wqe = rvt_get_swqe_ptr(qp, qp->s_acked); in hfi1_restart_rc() local 1637 struct rvt_swqe *wqe; in reset_sending_psn() local 1697 struct rvt_swqe *wqe; in hfi1_rc_send_complete() local 1825 struct rvt_swqe *wqe, in do_rc_completion() 1970 struct rvt_swqe *wqe; in do_rc_ack() local 2266 struct rvt_swqe *wqe; in rdma_seq_err() local 2310 struct rvt_swqe *wqe; in rc_rcv_resp() local
|
D | tid_rdma.c | 378 struct rvt_swqe *wqe = rvt_get_swqe_ptr(qp, i); in hfi1_qp_priv_init() local 416 struct rvt_swqe *wqe; in hfi1_qp_priv_tid_free() local 1621 void __trdma_clean_swqe(struct rvt_qp *qp, struct rvt_swqe *wqe) in __trdma_clean_swqe() 1703 u32 hfi1_build_tid_rdma_read_packet(struct rvt_swqe *wqe, in hfi1_build_tid_rdma_read_packet() 1787 u32 hfi1_build_tid_rdma_read_req(struct rvt_qp *qp, struct rvt_swqe *wqe, in hfi1_build_tid_rdma_read_req() 2418 struct rvt_swqe *wqe; in find_tid_request() local 2582 struct rvt_swqe *wqe; in hfi1_kern_read_tid_flow_free() local 2630 struct rvt_qp *qp, struct rvt_swqe *wqe) in restart_tid_rdma_read_req() 2663 struct rvt_swqe *wqe; in handle_read_kdeth_eflags() local 3046 void hfi1_tid_rdma_restart_req(struct rvt_qp *qp, struct rvt_swqe *wqe, in hfi1_tid_rdma_restart_req() [all …]
|
D | ud.c | 224 static void hfi1_make_bth_deth(struct rvt_qp *qp, struct rvt_swqe *wqe, in hfi1_make_bth_deth() 262 struct rvt_swqe *wqe) in hfi1_make_ud_req_9B() 333 struct rvt_swqe *wqe) in hfi1_make_ud_req_16B() 441 struct rvt_swqe *wqe; in hfi1_make_ud_req() local
|
/linux-6.12.1/drivers/infiniband/sw/rdmavt/ |
D | qp.c | 591 struct rvt_swqe *wqe = rvt_get_swqe_ptr(qp, qp->s_last); in rvt_clear_mr_refs() local 621 static bool rvt_swqe_has_lkey(struct rvt_swqe *wqe, u32 lkey) in rvt_swqe_has_lkey() 644 struct rvt_swqe *wqe = rvt_get_swqe_ptr(qp, s_last); in rvt_qp_sends_has_lkey() local 979 struct rvt_swqe *wqe; in free_ud_wq_attr() local 999 struct rvt_swqe *wqe; in alloc_ud_wq_attr() local 1797 struct rvt_rwqe *wqe; in rvt_post_recv() local 1964 struct rvt_swqe *wqe; in rvt_post_one_wr() local 2212 struct rvt_rwqe *wqe; in rvt_post_srq_recv() local 2267 static int init_sge(struct rvt_qp *qp, struct rvt_rwqe *wqe) in init_sge() 2350 struct rvt_rwqe *wqe; in rvt_get_rwqe() local [all …]
|
/linux-6.12.1/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_cmdq.c | 54 #define CMDQ_WQE_HEADER(wqe) ((struct hinic_cmdq_header *)(wqe)) argument 177 static void cmdq_prepare_wqe_ctrl(struct hinic_cmdq_wqe *wqe, int wrapped, in cmdq_prepare_wqe_ctrl() 234 static void cmdq_set_direct_wqe_data(struct hinic_cmdq_direct_wqe *wqe, in cmdq_set_direct_wqe_data() 243 static void cmdq_set_lcmd_wqe(struct hinic_cmdq_wqe *wqe, in cmdq_set_lcmd_wqe() 271 static void cmdq_set_direct_wqe(struct hinic_cmdq_wqe *wqe, in cmdq_set_direct_wqe() 529 struct hinic_cmdq_wqe *wqe) in clear_wqe_complete_bit() 563 struct hinic_cmdq_wqe *wqe) in cmdq_arm_ceq_handler()
|
/linux-6.12.1/drivers/infiniband/hw/bnxt_re/ |
D | ib_verbs.c | 411 struct bnxt_qplib_swqe *wqe = &fence->bind_wqe; in bnxt_re_create_fence_wqe() local 443 struct bnxt_qplib_swqe wqe; in bnxt_re_bind_fence_mw() local 1930 struct bnxt_qplib_swqe wqe; in bnxt_re_post_srq_recv() local 2294 struct bnxt_qplib_swqe *wqe, in bnxt_re_build_qp1_send_v2() 2468 struct bnxt_qplib_swqe *wqe, in bnxt_re_build_qp1_shadow_qp_recv() 2517 struct bnxt_qplib_swqe *wqe) in bnxt_re_build_send_wqe() 2555 struct bnxt_qplib_swqe *wqe) in bnxt_re_build_rdma_wqe() 2587 struct bnxt_qplib_swqe *wqe) in bnxt_re_build_atomic_wqe() 2614 struct bnxt_qplib_swqe *wqe) in bnxt_re_build_inv_wqe() 2628 struct bnxt_qplib_swqe *wqe) in bnxt_re_build_reg_wqe() [all …]
|
/linux-6.12.1/drivers/infiniband/hw/mthca/ |
D | mthca_srq.c | 92 static inline int *wqe_to_link(void *wqe) in wqe_to_link() 158 void *wqe; in mthca_alloc_srq_buf() local 495 void *wqe; in mthca_tavor_post_srq_recv() local 588 void *wqe; in mthca_arbel_post_srq_recv() local
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
D | ktls_txrx.c | 74 mlx5e_ktls_build_static_params(struct mlx5e_set_tls_static_params_wqe *wqe, in mlx5e_ktls_build_static_params() 117 mlx5e_ktls_build_progress_params(struct mlx5e_set_tls_progress_params_wqe *wqe, in mlx5e_ktls_build_progress_params()
|
/linux-6.12.1/drivers/scsi/lpfc/ |
D | lpfc_nvme.c | 399 union lpfc_wqe128 *wqe; in lpfc_nvme_gen_req() local 711 struct lpfc_iocbq *wqe, *next_wqe; in __lpfc_nvme_ls_abort() local 831 union lpfc_wqe128 *wqe; in lpfc_nvme_adj_fcp_sgls() local 1216 union lpfc_wqe128 *wqe = &pwqeq->wqe; in lpfc_nvme_prep_io_cmd() local 1342 union lpfc_wqe128 *wqe = &lpfc_ncmd->cur_iocbq.wqe; in lpfc_nvme_prep_io_dma() local 2043 union lpfc_wqe128 *wqe; in lpfc_get_nvme_buf() local
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
D | rx.c | 183 struct mlx5e_rx_wqe_cyc *wqe; in mlx5e_xsk_alloc_rx_wqes_batched() local 206 struct mlx5e_rx_wqe_cyc *wqe; in mlx5e_xsk_alloc_rx_wqes() local
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/ |
D | en_tx.c | 357 struct mlx5e_tx_wqe *wqe; in mlx5e_tx_flush() local 429 struct mlx5e_tx_wqe *wqe, u16 pi, bool xmit_more) in mlx5e_sq_xmit_wqe() 527 struct mlx5e_tx_wqe *wqe; in mlx5e_tx_mpwqe_session_start() local 681 struct mlx5e_tx_wqe *wqe; in mlx5e_xmit() local 982 struct mlx5i_tx_wqe *wqe; in mlx5i_sq_xmit() local
|