/linux-6.12.1/net/9p/ |
D | trans_rdma.c | 152 struct p9_trans_rdma *rdma = clnt->trans; in p9_rdma_show_options() local 154 if (rdma->port != P9_PORT) in p9_rdma_show_options() 155 seq_printf(m, ",port=%u", rdma->port); in p9_rdma_show_options() 156 if (rdma->sq_depth != P9_RDMA_SQ_DEPTH) in p9_rdma_show_options() 157 seq_printf(m, ",sq=%u", rdma->sq_depth); in p9_rdma_show_options() 158 if (rdma->rq_depth != P9_RDMA_RQ_DEPTH) in p9_rdma_show_options() 159 seq_printf(m, ",rq=%u", rdma->rq_depth); in p9_rdma_show_options() 160 if (rdma->timeout != P9_RDMA_TIMEOUT) in p9_rdma_show_options() 161 seq_printf(m, ",timeout=%lu", rdma->timeout); in p9_rdma_show_options() 162 if (rdma->privport) in p9_rdma_show_options() [all …]
|
/linux-6.12.1/drivers/gpu/drm/mediatek/ |
D | mtk_disp_rdma.c | 52 #define RDMA_FIFO_SIZE(rdma) ((rdma)->data->fifo_size) argument 109 struct mtk_disp_rdma *rdma = dev_get_drvdata(dev); in rdma_update_bits() local 110 unsigned int tmp = readl(rdma->regs + reg); in rdma_update_bits() 113 writel(tmp, rdma->regs + reg); in rdma_update_bits() 120 struct mtk_disp_rdma *rdma = dev_get_drvdata(dev); in mtk_rdma_register_vblank_cb() local 122 rdma->vblank_cb = vblank_cb; in mtk_rdma_register_vblank_cb() 123 rdma->vblank_cb_data = vblank_cb_data; in mtk_rdma_register_vblank_cb() 128 struct mtk_disp_rdma *rdma = dev_get_drvdata(dev); in mtk_rdma_unregister_vblank_cb() local 130 rdma->vblank_cb = NULL; in mtk_rdma_unregister_vblank_cb() 131 rdma->vblank_cb_data = NULL; in mtk_rdma_unregister_vblank_cb() [all …]
|
D | mtk_disp_ovl_adaptor.c | 101 static const struct mtk_ddp_comp_funcs rdma = { variable 110 [OVL_ADAPTOR_MDP_RDMA0] = { OVL_ADAPTOR_TYPE_MDP_RDMA, DDP_COMPONENT_MDP_RDMA0, 0, &rdma }, 111 [OVL_ADAPTOR_MDP_RDMA1] = { OVL_ADAPTOR_TYPE_MDP_RDMA, DDP_COMPONENT_MDP_RDMA1, 1, &rdma }, 112 [OVL_ADAPTOR_MDP_RDMA2] = { OVL_ADAPTOR_TYPE_MDP_RDMA, DDP_COMPONENT_MDP_RDMA2, 2, &rdma }, 113 [OVL_ADAPTOR_MDP_RDMA3] = { OVL_ADAPTOR_TYPE_MDP_RDMA, DDP_COMPONENT_MDP_RDMA3, 3, &rdma }, 114 [OVL_ADAPTOR_MDP_RDMA4] = { OVL_ADAPTOR_TYPE_MDP_RDMA, DDP_COMPONENT_MDP_RDMA4, 4, &rdma }, 115 [OVL_ADAPTOR_MDP_RDMA5] = { OVL_ADAPTOR_TYPE_MDP_RDMA, DDP_COMPONENT_MDP_RDMA5, 5, &rdma }, 116 [OVL_ADAPTOR_MDP_RDMA6] = { OVL_ADAPTOR_TYPE_MDP_RDMA, DDP_COMPONENT_MDP_RDMA6, 6, &rdma }, 117 [OVL_ADAPTOR_MDP_RDMA7] = { OVL_ADAPTOR_TYPE_MDP_RDMA, DDP_COMPONENT_MDP_RDMA7, 7, &rdma },
|
/linux-6.12.1/drivers/gpu/drm/meson/ |
D | meson_rdma.c | 26 if (!priv->rdma.addr) { in meson_rdma_init() 28 priv->rdma.addr = in meson_rdma_init() 30 &priv->rdma.addr_dma, in meson_rdma_init() 32 if (!priv->rdma.addr) in meson_rdma_init() 36 priv->rdma.offset = 0; in meson_rdma_init() 50 if (!priv->rdma.addr && !priv->rdma.addr_dma) in meson_rdma_free() 56 priv->rdma.addr, priv->rdma.addr_dma); in meson_rdma_free() 58 priv->rdma.addr = NULL; in meson_rdma_free() 59 priv->rdma.addr_dma = (dma_addr_t)0; in meson_rdma_free() 88 priv->rdma.offset = 0; in meson_rdma_reset() [all …]
|
/linux-6.12.1/net/sunrpc/xprtrdma/ |
D | svc_rdma_sendto.c | 117 svc_rdma_send_ctxt_alloc(struct svcxprt_rdma *rdma) in svc_rdma_send_ctxt_alloc() argument 119 int node = ibdev_to_node(rdma->sc_cm_id->device); in svc_rdma_send_ctxt_alloc() 125 ctxt = kzalloc_node(struct_size(ctxt, sc_sges, rdma->sc_max_send_sges), in svc_rdma_send_ctxt_alloc() 129 buffer = kmalloc_node(rdma->sc_max_req_size, GFP_KERNEL, node); in svc_rdma_send_ctxt_alloc() 132 addr = ib_dma_map_single(rdma->sc_pd->device, buffer, in svc_rdma_send_ctxt_alloc() 133 rdma->sc_max_req_size, DMA_TO_DEVICE); in svc_rdma_send_ctxt_alloc() 134 if (ib_dma_mapping_error(rdma->sc_pd->device, addr)) in svc_rdma_send_ctxt_alloc() 137 svc_rdma_send_cid_init(rdma, &ctxt->sc_cid); in svc_rdma_send_ctxt_alloc() 139 ctxt->sc_rdma = rdma; in svc_rdma_send_ctxt_alloc() 147 rdma->sc_max_req_size); in svc_rdma_send_ctxt_alloc() [all …]
|
D | svc_rdma_recvfrom.c | 119 svc_rdma_recv_ctxt_alloc(struct svcxprt_rdma *rdma) in svc_rdma_recv_ctxt_alloc() argument 121 int node = ibdev_to_node(rdma->sc_cm_id->device); in svc_rdma_recv_ctxt_alloc() 129 buffer = kmalloc_node(rdma->sc_max_req_size, GFP_KERNEL, node); in svc_rdma_recv_ctxt_alloc() 132 addr = ib_dma_map_single(rdma->sc_pd->device, buffer, in svc_rdma_recv_ctxt_alloc() 133 rdma->sc_max_req_size, DMA_FROM_DEVICE); in svc_rdma_recv_ctxt_alloc() 134 if (ib_dma_mapping_error(rdma->sc_pd->device, addr)) in svc_rdma_recv_ctxt_alloc() 137 svc_rdma_recv_cid_init(rdma, &ctxt->rc_cid); in svc_rdma_recv_ctxt_alloc() 149 ctxt->rc_recv_sge.length = rdma->sc_max_req_size; in svc_rdma_recv_ctxt_alloc() 150 ctxt->rc_recv_sge.lkey = rdma->sc_pd->local_dma_lkey; in svc_rdma_recv_ctxt_alloc() 152 svc_rdma_cc_init(rdma, &ctxt->rc_cc); in svc_rdma_recv_ctxt_alloc() [all …]
|
D | svc_rdma_rw.c | 55 svc_rdma_get_rw_ctxt(struct svcxprt_rdma *rdma, unsigned int sges) in svc_rdma_get_rw_ctxt() argument 57 struct ib_device *dev = rdma->sc_cm_id->device; in svc_rdma_get_rw_ctxt() 62 spin_lock(&rdma->sc_rw_ctxt_lock); in svc_rdma_get_rw_ctxt() 63 node = llist_del_first(&rdma->sc_rw_ctxts); in svc_rdma_get_rw_ctxt() 64 spin_unlock(&rdma->sc_rw_ctxt_lock); in svc_rdma_get_rw_ctxt() 87 trace_svcrdma_rwctx_empty(rdma, sges); in svc_rdma_get_rw_ctxt() 98 static void svc_rdma_put_rw_ctxt(struct svcxprt_rdma *rdma, in svc_rdma_put_rw_ctxt() argument 101 __svc_rdma_put_rw_ctxt(ctxt, &rdma->sc_rw_ctxts); in svc_rdma_put_rw_ctxt() 109 void svc_rdma_destroy_rw_ctxts(struct svcxprt_rdma *rdma) in svc_rdma_destroy_rw_ctxts() argument 114 while ((node = llist_del_first(&rdma->sc_rw_ctxts)) != NULL) { in svc_rdma_destroy_rw_ctxts() [all …]
|
D | svc_rdma_transport.c | 330 struct svcxprt_rdma *rdma = cma_id->context; in svc_rdma_cma_handler() local 331 struct svc_xprt *xprt = &rdma->sc_xprt; in svc_rdma_cma_handler() 335 clear_bit(RDMAXPRT_CONN_PENDING, &rdma->sc_flags); in svc_rdma_cma_handler() 388 struct svcxprt_rdma *rdma = container_of(rn, struct svcxprt_rdma, in svc_rdma_xprt_done() local 390 struct rdma_cm_id *id = rdma->sc_cm_id; in svc_rdma_xprt_done() 393 svc_xprt_close(&rdma->sc_xprt); in svc_rdma_xprt_done() 585 struct svcxprt_rdma *rdma = in svc_rdma_detach() local 588 rdma_disconnect(rdma->sc_cm_id); in svc_rdma_detach() 593 struct svcxprt_rdma *rdma = in __svc_rdma_free() local 595 struct ib_device *device = rdma->sc_cm_id->device; in __svc_rdma_free() [all …]
|
D | svc_rdma_backchannel.c | 75 static int svc_rdma_bc_sendto(struct svcxprt_rdma *rdma, in svc_rdma_bc_sendto() argument 83 ret = svc_rdma_map_reply_msg(rdma, sctxt, &empty_pcl, &empty_pcl, in svc_rdma_bc_sendto() 93 return svc_rdma_post_send(rdma, sctxt); in svc_rdma_bc_sendto() 136 rpcrdma_bc_send_request(struct svcxprt_rdma *rdma, struct rpc_rqst *rqst) in rpcrdma_bc_send_request() argument 144 ctxt = svc_rdma_send_ctxt_get(rdma); in rpcrdma_bc_send_request() 160 rc = svc_rdma_bc_sendto(rdma, rqst, ctxt); in rpcrdma_bc_send_request() 166 svc_rdma_send_ctxt_put(rdma, ctxt); in rpcrdma_bc_send_request() 183 struct svcxprt_rdma *rdma = in xprt_rdma_bc_send_request() local 190 ret = rpcrdma_bc_send_request(rdma, rqst); in xprt_rdma_bc_send_request()
|
/linux-6.12.1/include/linux/sunrpc/ |
D | svc_rdma.h | 145 static inline void svc_rdma_recv_cid_init(struct svcxprt_rdma *rdma, in svc_rdma_recv_cid_init() argument 148 cid->ci_queue_id = rdma->sc_rq_cq->res.id; in svc_rdma_recv_cid_init() 149 cid->ci_completion_id = atomic_inc_return(&rdma->sc_completion_ids); in svc_rdma_recv_cid_init() 157 static inline void svc_rdma_send_cid_init(struct svcxprt_rdma *rdma, in svc_rdma_send_cid_init() argument 160 cid->ci_queue_id = rdma->sc_sq_cq->res.id; in svc_rdma_send_cid_init() 161 cid->ci_completion_id = atomic_inc_return(&rdma->sc_completion_ids); in svc_rdma_send_cid_init() 256 extern void svc_rdma_recv_ctxts_destroy(struct svcxprt_rdma *rdma); 257 extern bool svc_rdma_post_recvs(struct svcxprt_rdma *rdma); 259 svc_rdma_recv_ctxt_get(struct svcxprt_rdma *rdma); 260 extern void svc_rdma_recv_ctxt_put(struct svcxprt_rdma *rdma, [all …]
|
/linux-6.12.1/Documentation/admin-guide/cgroup-v1/ |
D | rdma.rst | 28 Currently user space applications can easily take away all the rdma verb 31 rdma resources. This can lead to service unavailability. 34 of processes can be limited. Through this controller different rdma 43 by rdma cgroup, which can be extended later if required. 52 of the child processes which shares the address space, rdma resources are 56 rdma resources. Linking resources around css also ensures that cgroups can be 60 Whenever RDMA resource charging occurs, owner rdma cgroup is returned to 61 the caller. Same rdma cgroup should be passed while uncharging the resource. 81 IB stack honors limits enforced by the rdma controller. When application 86 Following resources can be accounted by rdma controller. [all …]
|
/linux-6.12.1/Documentation/ABI/stable/ |
D | sysfs-driver-ib_srp | 4 Contact: linux-rdma@vger.kernel.org 77 Contact: linux-rdma@vger.kernel.org 83 Contact: linux-rdma@vger.kernel.org 89 Contact: linux-rdma@vger.kernel.org 97 Contact: linux-rdma@vger.kernel.org 104 Contact: linux-rdma@vger.kernel.org 111 Contact: linux-rdma@vger.kernel.org 117 Contact: linux-rdma@vger.kernel.org 124 Contact: linux-rdma@vger.kernel.org 131 Contact: linux-rdma@vger.kernel.org [all …]
|
D | sysfs-class-infiniband | 9 Contact: linux-rdma@vger.kernel.org 24 Contact: linux-rdma@vger.kernel.org 34 Contact: linux-rdma@vger.kernel.org 49 Contact: linux-rdma@vger.kernel.org 80 Contact: linux-rdma@vger.kernel.org 108 Contact: linux-rdma@vger.kernel.org 203 Contact: linux-rdma@vger.kernel.org 222 Contact: linux-rdma@vger.kernel.org 229 Contact: linux-rdma@vger.kernel.org 241 Contact: linux-rdma@vger.kernel.org [all …]
|
D | sysfs-transport-srp | 4 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 11 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 20 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 36 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org 50 Contact: linux-scsi@vger.kernel.org, linux-rdma@vger.kernel.org
|
/linux-6.12.1/drivers/net/ethernet/seeq/ |
D | sgiseeq.c | 69 volatile struct hpc_dma_desc rdma; member 213 sp->rx_desc[i].rdma.pbuf = dma_addr; in seeq_init_ring() 215 sp->rx_desc[i].rdma.cntinfo = RCNTINFO_INIT; in seeq_init_ring() 218 sp->rx_desc[i - 1].rdma.cntinfo |= HPCDMA_EOR; in seeq_init_ring() 263 i, (&r[i]), r[i].rdma.pbuf, r[i].rdma.cntinfo, in sgiseeq_dump_rings() 264 r[i].rdma.pnext); in sgiseeq_dump_rings() 267 i, (&r[i]), r[i].rdma.pbuf, r[i].rdma.cntinfo, in sgiseeq_dump_rings() 268 r[i].rdma.pnext); in sgiseeq_dump_rings() 355 while (!(rd->rdma.cntinfo & HPCDMA_OWN)) { in sgiseeq_rx() 356 len = PKT_BUF_SZ - (rd->rdma.cntinfo & HPCDMA_BCNT) - 3; in sgiseeq_rx() [all …]
|
/linux-6.12.1/Documentation/devicetree/bindings/media/ |
D | mediatek-mdp.txt | 12 "mediatek,mt8173-mdp-rdma" - read DMA 24 "mediatek,mt8173-mdp-rdma" 32 mdp_rdma0: rdma@14001000 { 33 compatible = "mediatek,mt8173-mdp-rdma"; 43 mdp_rdma1: rdma@14002000 { 44 compatible = "mediatek,mt8173-mdp-rdma";
|
/linux-6.12.1/drivers/infiniband/ulp/rtrs/ |
D | rtrs-clt-stats.c | 35 this_cpu_inc(stats->pcpu_stats->rdma.failover_cnt); in rtrs_clt_inc_failover_cnt() 90 r = &per_cpu_ptr(stats->pcpu_stats, cpu)->rdma; in rtrs_clt_stats_rdma_to_str() 120 memset(&s->rdma, 0, sizeof(s->rdma)); in rtrs_clt_reset_rdma_stats() 168 this_cpu_inc(stats->pcpu_stats->rdma.dir[d].cnt); in rtrs_clt_update_rdma_stats() 169 this_cpu_add(stats->pcpu_stats->rdma.dir[d].size_total, size); in rtrs_clt_update_rdma_stats()
|
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb3/ |
D | cxgb3_offload.c | 281 struct rdma_info *rdma = data; in cxgb_rdma_ctl() local 284 rdma->udbell_physbase = pci_resource_start(pdev, 2); in cxgb_rdma_ctl() 285 rdma->udbell_len = pci_resource_len(pdev, 2); in cxgb_rdma_ctl() 286 rdma->tpt_base = in cxgb_rdma_ctl() 288 rdma->tpt_top = t3_read_reg(adapter, A_ULPTX_TPT_ULIMIT); in cxgb_rdma_ctl() 289 rdma->pbl_base = in cxgb_rdma_ctl() 291 rdma->pbl_top = t3_read_reg(adapter, A_ULPTX_PBL_ULIMIT); in cxgb_rdma_ctl() 292 rdma->rqt_base = t3_read_reg(adapter, A_ULPRX_RQ_LLIMIT); in cxgb_rdma_ctl() 293 rdma->rqt_top = t3_read_reg(adapter, A_ULPRX_RQ_ULIMIT); in cxgb_rdma_ctl() 294 rdma->kdb_addr = adapter->regs + A_SG_KDOORBELL; in cxgb_rdma_ctl() [all …]
|
/linux-6.12.1/drivers/media/platform/mediatek/mdp3/ |
D | mtk-mdp3-comp.c | 116 reg = CFG_COMP(MT8183, ctx->param, rdma.src_ctrl); in config_rdma_frame() 118 reg = CFG_COMP(MT8195, ctx->param, rdma.src_ctrl); in config_rdma_frame() 126 reg = CFG_COMP(MT8183, ctx->param, rdma.ufo_dec_y); in config_rdma_frame() 128 reg = CFG_COMP(MT8195, ctx->param, rdma.ufo_dec_y); in config_rdma_frame() 134 reg = CFG_COMP(MT8183, ctx->param, rdma.ufo_dec_c); in config_rdma_frame() 136 reg = CFG_COMP(MT8195, ctx->param, rdma.ufo_dec_c); in config_rdma_frame() 144 reg = CFG_COMP(MT8183, ctx->param, rdma.mf_bkgd_in_pxl); in config_rdma_frame() 146 reg = CFG_COMP(MT8195, ctx->param, rdma.mf_bkgd_in_pxl); in config_rdma_frame() 154 reg = CFG_COMP(MT8183, ctx->param, rdma.control); in config_rdma_frame() 157 reg = CFG_COMP(MT8195, ctx->param, rdma.control); in config_rdma_frame() [all …]
|
/linux-6.12.1/drivers/nvme/target/ |
D | Makefile | 7 obj-$(CONFIG_NVME_TARGET_RDMA) += nvmet-rdma.o 19 nvmet-rdma-y += rdma.o
|
/linux-6.12.1/drivers/nvme/host/ |
D | Makefile | 8 obj-$(CONFIG_NVME_RDMA) += nvme-rdma.o 26 nvme-rdma-y += rdma.o
|
/linux-6.12.1/drivers/infiniband/ |
D | Kconfig | 25 need libibumad from rdma-core 26 <https://github.com/linux-rdma/rdma-core>. 38 rdma-core <https://github.com/linux-rdma/rdma-core>.
|
/linux-6.12.1/drivers/macintosh/ |
D | rack-meter.c | 146 struct rackmeter_dma *rdma = rm->dma_buf_v; in rackmeter_do_pause() local 155 memset(rdma->buf1, 0, sizeof(rdma->buf1)); in rackmeter_do_pause() 156 memset(rdma->buf2, 0, sizeof(rdma->buf2)); in rackmeter_do_pause() 371 struct resource ri2s, rdma; in rackmeter_probe() local 431 of_address_to_resource(i2s, 1, &rdma)) { in rackmeter_probe() 441 pr_debug(" dma @0x%08x\n", (unsigned int)rdma.start); in rackmeter_probe() 475 rm->dma_regs = ioremap(rdma.start, 0x100); in rackmeter_probe()
|
/linux-6.12.1/fs/smb/client/ |
D | smbdirect.c | 51 struct smb_extract_to_rdma *rdma); 2412 static bool smb_set_sge(struct smb_extract_to_rdma *rdma, in smb_set_sge() argument 2415 struct ib_sge *sge = &rdma->sge[rdma->nr_sge]; in smb_set_sge() 2418 addr = ib_dma_map_page(rdma->device, lowest_page, in smb_set_sge() 2419 off, len, rdma->direction); in smb_set_sge() 2420 if (ib_dma_mapping_error(rdma->device, addr)) in smb_set_sge() 2425 sge->lkey = rdma->local_dma_lkey; in smb_set_sge() 2426 rdma->nr_sge++; in smb_set_sge() 2435 struct smb_extract_to_rdma *rdma, in smb_extract_bvec_to_rdma() argument 2455 if (!smb_set_sge(rdma, bv[i].bv_page, off, len)) in smb_extract_bvec_to_rdma() [all …]
|
/linux-6.12.1/arch/arm/boot/dts/mediatek/ |
D | mt7623n.dtsi | 148 rdma0: rdma@14008000 { 149 compatible = "mediatek,mt7623-disp-rdma", 150 "mediatek,mt2701-disp-rdma"; 207 rdma1: rdma@14012000 { 208 compatible = "mediatek,mt7623-disp-rdma", 209 "mediatek,mt2701-disp-rdma";
|