/linux-6.12.1/drivers/net/ethernet/marvell/octeon_ep/ |
D | octep_ctrl_mbox.c | 135 u8 __iomem *qbuf; in octep_write_mbox_data() local 139 qbuf = (q->hw_q + *pi); in octep_write_mbox_data() 142 memcpy_toio(qbuf, buf, w_sz); in octep_write_mbox_data() 147 memcpy_toio(qbuf, buf, cp_sz); in octep_write_mbox_data() 153 qbuf = (q->hw_q + *pi); in octep_write_mbox_data() 154 memcpy_toio(qbuf, buf, w_sz); in octep_write_mbox_data() 200 u8 __iomem *qbuf; in octep_read_mbox_data() local 204 qbuf = (q->hw_q + *ci); in octep_read_mbox_data() 207 memcpy_fromio(buf, qbuf, r_sz); in octep_read_mbox_data() 212 memcpy_fromio(buf, qbuf, cp_sz); in octep_read_mbox_data() [all …]
|
/linux-6.12.1/drivers/infiniband/hw/erdma/ |
D | erdma_cmdq.c | 96 sq->qbuf = dma_alloc_coherent(&dev->pdev->dev, sq->depth << SQEBB_SHIFT, in erdma_cmdq_sq_init() 98 if (!sq->qbuf) in erdma_cmdq_sq_init() 118 sq->qbuf, sq->qbuf_dma_addr); in erdma_cmdq_sq_init() 129 cq->qbuf = dma_alloc_coherent(&dev->pdev->dev, cq->depth << CQE_SHIFT, in erdma_cmdq_cq_init() 131 if (!cq->qbuf) in erdma_cmdq_cq_init() 151 dma_free_coherent(&dev->pdev->dev, cq->depth << CQE_SHIFT, cq->qbuf, in erdma_cmdq_cq_init() 211 cmdq->cq.qbuf, cmdq->cq.qbuf_dma_addr); in erdma_cmdq_init() 217 cmdq->sq.qbuf, cmdq->sq.qbuf_dma_addr); in erdma_cmdq_init() 240 cmdq->sq.qbuf, cmdq->sq.qbuf_dma_addr); in erdma_cmdq_destroy() 245 cmdq->cq.qbuf, cmdq->cq.qbuf_dma_addr); in erdma_cmdq_destroy() [all …]
|
D | erdma.h | 22 void *qbuf; member 41 void *qbuf; member 57 void *qbuf; member 220 static inline void *get_queue_entry(void *qbuf, u32 idx, u32 depth, u32 shift) in get_queue_entry() argument 224 return qbuf + (idx << shift); in get_queue_entry()
|
D | erdma_eq.c | 24 u64 *eqe = get_queue_entry(eq->qbuf, eq->ci, eq->depth, EQE_SHIFT); in get_next_valid_eqe() 87 eq->qbuf = dma_alloc_coherent(&dev->pdev->dev, buf_size, in erdma_eq_common_init() 89 if (!eq->qbuf) in erdma_eq_common_init() 105 dma_free_coherent(&dev->pdev->dev, buf_size, eq->qbuf, in erdma_eq_common_init() 114 dma_free_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT, eq->qbuf, in erdma_eq_destroy()
|
D | erdma_cq.c | 11 __be32 *cqe = get_queue_entry(cq->kern_cq.qbuf, cq->kern_cq.ci, in get_next_valid_cqe()
|
D | erdma_verbs.c | 1253 cq->kern_cq.qbuf, cq->kern_cq.qbuf_dma_addr); in erdma_destroy_cq() 1629 cq->kern_cq.qbuf = in erdma_init_kernel_cq() 1632 if (!cq->kern_cq.qbuf) in erdma_init_kernel_cq() 1648 cq->kern_cq.qbuf, cq->kern_cq.qbuf_dma_addr); in erdma_init_kernel_cq() 1716 cq->kern_cq.qbuf, cq->kern_cq.qbuf_dma_addr); in erdma_create_cq()
|
D | erdma_verbs.h | 244 void *qbuf; member
|
/linux-6.12.1/drivers/net/ethernet/intel/ice/ |
D | ice_lag.c | 402 ice_lag_qbuf_recfg(struct ice_hw *hw, struct ice_aqc_cfg_txqs_buf *qbuf, in ice_lag_qbuf_recfg() argument 430 qbuf->queue_info[count].q_handle = cpu_to_le16(qid); in ice_lag_qbuf_recfg() 431 qbuf->queue_info[count].tc = tc; in ice_lag_qbuf_recfg() 432 qbuf->queue_info[count].q_teid = cpu_to_le32(q_ctx->q_teid); in ice_lag_qbuf_recfg() 524 struct ice_aqc_cfg_txqs_buf *qbuf; in ice_lag_move_vf_node_tc() local 565 qbuf_size = struct_size(qbuf, queue_info, numq); in ice_lag_move_vf_node_tc() 566 qbuf = kzalloc(qbuf_size, GFP_KERNEL); in ice_lag_move_vf_node_tc() 567 if (!qbuf) { in ice_lag_move_vf_node_tc() 573 valq = ice_lag_qbuf_recfg(&lag->pf->hw, qbuf, vsi_num, numq, tc); in ice_lag_move_vf_node_tc() 579 if (ice_aq_cfg_lan_txq(&lag->pf->hw, qbuf, qbuf_size, valq, oldport, in ice_lag_move_vf_node_tc() [all …]
|
/linux-6.12.1/Documentation/userspace-api/media/dvb/ |
D | dmx_fcalls.rst | 30 dmx-qbuf
|
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | t4vf_hw.c | 1300 u16 qbuf[3]; in t4vf_config_rss_range() local 1301 u16 *qbp = qbuf; in t4vf_config_rss_range() 1305 qbuf[0] = qbuf[1] = qbuf[2] = 0; in t4vf_config_rss_range() 1312 *qp++ = cpu_to_be32(FW_RSS_IND_TBL_CMD_IQ0_V(qbuf[0]) | in t4vf_config_rss_range() 1313 FW_RSS_IND_TBL_CMD_IQ1_V(qbuf[1]) | in t4vf_config_rss_range() 1314 FW_RSS_IND_TBL_CMD_IQ2_V(qbuf[2])); in t4vf_config_rss_range()
|
/linux-6.12.1/drivers/net/wireless/mediatek/mt76/ |
D | dma.c | 646 struct mt76_queue_buf qbuf = {}; in mt76_dma_rx_fill() local 663 qbuf.addr = addr + q->buf_offset; in mt76_dma_rx_fill() 665 qbuf.len = len - q->buf_offset; in mt76_dma_rx_fill() 666 qbuf.skip_unmap = false; in mt76_dma_rx_fill() 667 if (mt76_dma_add_rx_buf(dev, q, &qbuf, buf) < 0) { in mt76_dma_rx_fill()
|
/linux-6.12.1/Documentation/userspace-api/media/v4l/ |
D | user-func.rst | 58 vidioc-qbuf
|
/linux-6.12.1/include/uapi/linux/ |
D | i2o-dev.h | 91 void __user *qbuf; /* Pointer to HTTP query string */ member
|