/linux-6.12.1/drivers/infiniband/hw/mlx5/ |
D | cq.c | 87 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64; in get_sw_cqe() 470 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64; in mlx5_poll_one() 901 cq->mcq.cqe_sz = cqe_size; in create_cq_kernel() 1012 MLX5_SET(cqc, cqc, cqe_sz, in mlx5_ib_create_cq() 1106 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64; in __mlx5_ib_cq_clean() 1113 dest64 = (cq->mcq.cqe_sz == 64) ? dest : dest + 64; in __mlx5_ib_cq_clean() 1115 memcpy(dest, cqe, cq->mcq.cqe_sz); in __mlx5_ib_cq_clean() 1363 MLX5_SET(cqc, cqc, cqe_sz, in mlx5_ib_resize_cq()
|
/linux-6.12.1/include/uapi/rdma/ |
D | bnxt_re-abi.h | 80 __u32 cqe_sz; member
|
/linux-6.12.1/include/linux/mlx5/ |
D | cq.h | 41 int cqe_sz; member
|
D | mlx5_ifc.h | 4595 u8 cqe_sz[0x3]; member
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/ |
D | wc.c | 55 mcq->cqe_sz = 64; in mlx5_wc_create_cqwq() 121 MLX5_SET(cqc, cqc, cqe_sz, CQE_STRIDE_128_PAD); in mlx5_wc_create_cq()
|
D | wq.c | 164 u8 log_wq_stride = MLX5_GET(cqc, cqc, cqe_sz) == CQE_STRIDE_64 ? 6 : 7; in mlx5_cqwq_create()
|
D | en_main.c | 2104 mcq->cqe_sz = 64; in mlx5e_alloc_cq_common()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
D | aso.c | 60 mcq->cqe_sz = 64; in mlx5_aso_alloc_cq() 134 MLX5_SET(cqc, cqc_data, cqe_sz, CQE_STRIDE_128_PAD); in mlx5_aso_create_cq()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx4/ |
D | mlx4_en.h | 662 static inline struct mlx4_cqe *mlx4_en_get_cqe(void *buf, int idx, int cqe_sz) in mlx4_en_get_cqe() argument 664 return buf + idx * cqe_sz; in mlx4_en_get_cqe()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
D | mlx5hws_debug.c | 282 cq->mcq.cqe_sz, in hws_debug_dump_context_send_engine()
|
D | mlx5hws_send.c | 776 mcq->cqe_sz = 64; in hws_send_ring_alloc_cq() 842 MLX5_SET(cqc, cqc_data, cqe_sz, queue->num_entries); in hws_send_ring_open_cq()
|
/linux-6.12.1/drivers/infiniband/hw/hns/ |
D | hns_roce_cq.c | 335 hr_cq->cqe_size = hr_dev->caps.cqe_sz; in set_cqe_size()
|
D | hns_roce_device.h | 774 u32 cqe_sz; member
|
D | hns_roce_main.c | 422 resp.cqe_size = hr_dev->caps.cqe_sz; in hns_roce_alloc_ucontext()
|
D | hns_roce_hw_v2.h | 1178 u8 cqe_sz; member
|
D | hns_roce_hw_v2.c | 2055 calc_pg_sz(caps->max_cqes, caps->cqe_sz, caps->cqe_hop_num, in set_hem_page_size() 2115 caps->cqe_sz = HNS_ROCE_V3_CQE_SIZE; in apply_func_caps() 2270 caps->cqe_sz = resp_a->cqe_sz; in hns_roce_query_caps()
|
/linux-6.12.1/drivers/vfio/pci/mlx5/ |
D | cmd.c | 1112 cq->mcq.cqe_sz = cqe_size; in mlx5vf_create_cq() 1674 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64; in get_sw_cqe()
|
/linux-6.12.1/drivers/crypto/hisilicon/ |
D | qm.c | 245 #define QM_MK_CQC_DW3_V1(hop_num, pg_sz, buf_sz, cqe_sz) \ argument 249 ((cqe_sz) << QM_CQ_CQE_SIZE_SHIFT)) 251 #define QM_MK_CQC_DW3_V2(cqe_sz, cq_depth) \ argument 252 ((((u32)cq_depth) - 1) | ((cqe_sz) << QM_CQ_CQE_SIZE_SHIFT))
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | params.c | 823 MLX5_SET(cqc, cqc, cqe_sz, CQE_STRIDE_128_PAD); in mlx5e_build_common_cq_param()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/fpga/ |
D | conn.c | 471 conn->cq.mcq.cqe_sz = 64; in mlx5_fpga_conn_create_cq()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
D | dr_send.c | 1125 cq->mcq.cqe_sz = 64; in dr_create_cq()
|
/linux-6.12.1/drivers/vdpa/mlx5/net/ |
D | mlx5_vnet.c | 575 vcq->mcq.cqe_sz = 64; in cq_create()
|
/linux-6.12.1/drivers/infiniband/hw/bnxt_re/ |
D | ib_verbs.c | 4249 resp.cqe_sz = sizeof(struct cq_base); in bnxt_re_alloc_ucontext()
|