/linux-6.12.1/drivers/net/ethernet/pensando/ionic/ |
D | ionic_txrx.c | 115 static void *ionic_rx_buf_va(struct ionic_buf_info *buf_info) in ionic_rx_buf_va() argument 117 return page_address(buf_info->page) + buf_info->page_offset; in ionic_rx_buf_va() 120 static dma_addr_t ionic_rx_buf_pa(struct ionic_buf_info *buf_info) in ionic_rx_buf_pa() argument 122 return page_pool_get_dma_addr(buf_info->page) + buf_info->page_offset; in ionic_rx_buf_pa() 126 struct ionic_buf_info *buf_info, in __ionic_rx_put_buf() argument 129 if (!buf_info->page) in __ionic_rx_put_buf() 132 page_pool_put_full_page(q->page_pool, buf_info->page, recycle_direct); in __ionic_rx_put_buf() 133 buf_info->page = NULL; in __ionic_rx_put_buf() 134 buf_info->len = 0; in __ionic_rx_put_buf() 135 buf_info->page_offset = 0; in __ionic_rx_put_buf() [all …]
|
/linux-6.12.1/drivers/pci/endpoint/functions/ |
D | pci-epf-mhi.c | 26 struct mhi_ep_buf_info buf_info; member 243 struct mhi_ep_buf_info *buf_info) in pci_epf_mhi_iatu_read() argument 246 size_t offset = get_align_offset(epf_mhi, buf_info->host_addr); in pci_epf_mhi_iatu_read() 253 ret = __pci_epf_mhi_alloc_map(mhi_cntrl, buf_info->host_addr, &tre_phys, in pci_epf_mhi_iatu_read() 254 &tre_buf, offset, buf_info->size); in pci_epf_mhi_iatu_read() 260 memcpy_fromio(buf_info->dev_addr, tre_buf, buf_info->size); in pci_epf_mhi_iatu_read() 262 __pci_epf_mhi_unmap_free(mhi_cntrl, buf_info->host_addr, tre_phys, in pci_epf_mhi_iatu_read() 263 tre_buf, offset, buf_info->size); in pci_epf_mhi_iatu_read() 267 if (buf_info->cb) in pci_epf_mhi_iatu_read() 268 buf_info->cb(buf_info); in pci_epf_mhi_iatu_read() [all …]
|
/linux-6.12.1/drivers/bus/mhi/ep/ |
D | ring.c | 33 struct mhi_ep_buf_info buf_info = {}; in __mhi_ep_cache_ring() local 47 buf_info.size = (end - start) * sizeof(struct mhi_ring_element); in __mhi_ep_cache_ring() 48 buf_info.host_addr = ring->rbase + (start * sizeof(struct mhi_ring_element)); in __mhi_ep_cache_ring() 49 buf_info.dev_addr = &ring->ring_cache[start]; in __mhi_ep_cache_ring() 51 ret = mhi_cntrl->read_sync(mhi_cntrl, &buf_info); in __mhi_ep_cache_ring() 55 buf_info.size = (ring->ring_size - start) * sizeof(struct mhi_ring_element); in __mhi_ep_cache_ring() 56 buf_info.host_addr = ring->rbase + (start * sizeof(struct mhi_ring_element)); in __mhi_ep_cache_ring() 57 buf_info.dev_addr = &ring->ring_cache[start]; in __mhi_ep_cache_ring() 59 ret = mhi_cntrl->read_sync(mhi_cntrl, &buf_info); in __mhi_ep_cache_ring() 64 buf_info.host_addr = ring->rbase; in __mhi_ep_cache_ring() [all …]
|
D | main.c | 337 static void mhi_ep_read_completion(struct mhi_ep_buf_info *buf_info) in mhi_ep_read_completion() argument 339 struct mhi_ep_device *mhi_dev = buf_info->mhi_dev; in mhi_ep_read_completion() 348 result.buf_addr = buf_info->cb_buf; in mhi_ep_read_completion() 350 result.bytes_xferd = buf_info->size; in mhi_ep_read_completion() 360 if (buf_info->code != MHI_EV_CC_OVERFLOW) { in mhi_ep_read_completion() 398 kmem_cache_free(mhi_cntrl->tre_buf_cache, buf_info->cb_buf); in mhi_ep_read_completion() 407 struct mhi_ep_buf_info buf_info = {}; in mhi_ep_read_channel() local 445 buf_info.host_addr = mhi_chan->tre_loc + read_offset; in mhi_ep_read_channel() 446 buf_info.dev_addr = buf_addr + write_offset; in mhi_ep_read_channel() 447 buf_info.size = tr_len; in mhi_ep_read_channel() [all …]
|
/linux-6.12.1/drivers/net/caif/ |
D | caif_virtio.c | 134 struct buf_info { struct 148 static void free_buf_info(struct cfv_info *cfv, struct buf_info *buf_info) in free_buf_info() argument 150 if (!buf_info) in free_buf_info() 152 gen_pool_free(cfv->genpool, (unsigned long) buf_info->vaddr, in free_buf_info() 153 buf_info->size); in free_buf_info() 154 kfree(buf_info); in free_buf_info() 169 struct buf_info *buf_info; in cfv_release_used_buf() local 173 buf_info = virtqueue_get_buf(vq_tx, &len); in cfv_release_used_buf() 177 if (!buf_info) in cfv_release_used_buf() 180 free_buf_info(cfv, buf_info); in cfv_release_used_buf() [all …]
|
/linux-6.12.1/drivers/bus/mhi/host/ |
D | main.c | 185 struct mhi_buf_info *buf_info) in mhi_map_single_no_bb() argument 187 buf_info->p_addr = dma_map_single(mhi_cntrl->cntrl_dev, in mhi_map_single_no_bb() 188 buf_info->v_addr, buf_info->len, in mhi_map_single_no_bb() 189 buf_info->dir); in mhi_map_single_no_bb() 190 if (dma_mapping_error(mhi_cntrl->cntrl_dev, buf_info->p_addr)) in mhi_map_single_no_bb() 197 struct mhi_buf_info *buf_info) in mhi_map_single_use_bb() argument 199 void *buf = dma_alloc_coherent(mhi_cntrl->cntrl_dev, buf_info->len, in mhi_map_single_use_bb() 200 &buf_info->p_addr, GFP_ATOMIC); in mhi_map_single_use_bb() 205 if (buf_info->dir == DMA_TO_DEVICE) in mhi_map_single_use_bb() 206 memcpy(buf, buf_info->v_addr, buf_info->len); in mhi_map_single_use_bb() [all …]
|
D | internal.h | 415 struct mhi_buf_info *buf_info); 417 struct mhi_buf_info *buf_info); 419 struct mhi_buf_info *buf_info); 421 struct mhi_buf_info *buf_info);
|
/linux-6.12.1/drivers/gpu/drm/exynos/ |
D | exynos_drm_g2d.c | 220 struct g2d_buf_info buf_info; member 276 struct g2d_buf_info *buf_info; in g2d_init_cmdlist() local 303 buf_info = &node[nr].buf_info; in g2d_init_cmdlist() 305 buf_info->reg_types[i] = REG_TYPE_NONE; in g2d_init_cmdlist() 687 struct g2d_buf_info *buf_info = &node->buf_info; in g2d_map_cmdlist_gem() local 692 for (i = 0; i < buf_info->map_nr; i++) { in g2d_map_cmdlist_gem() 710 buf_desc = &buf_info->descs[reg_type]; in g2d_map_cmdlist_gem() 712 if (buf_info->types[reg_type] == BUF_TYPE_GEM) { in g2d_map_cmdlist_gem() 729 buf_info->obj[reg_type] = exynos_gem; in g2d_map_cmdlist_gem() 750 &buf_info->obj[reg_type]); in g2d_map_cmdlist_gem() [all …]
|
/linux-6.12.1/include/linux/ |
D | mhi_ep.h | 69 void (*cb)(struct mhi_ep_buf_info *buf_info); 163 int (*read_sync)(struct mhi_ep_cntrl *mhi_cntrl, struct mhi_ep_buf_info *buf_info); 164 int (*write_sync)(struct mhi_ep_cntrl *mhi_cntrl, struct mhi_ep_buf_info *buf_info); 165 int (*read_async)(struct mhi_ep_cntrl *mhi_cntrl, struct mhi_ep_buf_info *buf_info); 166 int (*write_async)(struct mhi_ep_cntrl *mhi_cntrl, struct mhi_ep_buf_info *buf_info);
|
/linux-6.12.1/drivers/net/ethernet/meta/fbnic/ |
D | fbnic_fw.c | 106 mbx->buf_info[tail].msg = msg; in fbnic_mbx_map_msg() 107 mbx->buf_info[tail].addr = addr; in fbnic_mbx_map_msg() 128 if (!mbx->buf_info[desc_idx].msg) in fbnic_mbx_unmap_and_free_msg() 133 dma_unmap_single(fbd->dev, mbx->buf_info[desc_idx].addr, in fbnic_mbx_unmap_and_free_msg() 136 free_page((unsigned long)mbx->buf_info[desc_idx].msg); in fbnic_mbx_unmap_and_free_msg() 137 mbx->buf_info[desc_idx].msg = NULL; in fbnic_mbx_unmap_and_free_msg() 678 dma_unmap_single(fbd->dev, rx_mbx->buf_info[head].addr, in fbnic_mbx_process_rx_msgs() 681 msg = rx_mbx->buf_info[head].msg; in fbnic_mbx_process_rx_msgs() 713 free_page((unsigned long)rx_mbx->buf_info[head].msg); in fbnic_mbx_process_rx_msgs() 714 rx_mbx->buf_info[head].msg = NULL; in fbnic_mbx_process_rx_msgs()
|
D | fbnic_fw.h | 18 } buf_info[FBNIC_IPC_MBX_DESC_LEN]; member
|
/linux-6.12.1/drivers/acpi/ |
D | pfr_update.c | 442 struct pfru_com_buf_info buf_info; in pfru_write() local 449 ret = query_buffer(&buf_info, pfru_dev); in pfru_write() 453 if (len > buf_info.buf_size) in pfru_write() 461 phy_addr = (phys_addr_t)((buf_info.addr_hi << 32) | buf_info.addr_lo); in pfru_write() 462 buf_ptr = memremap(phy_addr, buf_info.buf_size, MEMREMAP_WB); in pfru_write()
|
/linux-6.12.1/drivers/net/vmxnet3/ |
D | vmxnet3_drv.c | 393 BUG_ON(tq->buf_info[eop_idx].sop_idx != tq->tx_ring.next2comp); in vmxnet3_unmap_pkt() 396 tbi = &tq->buf_info[eop_idx]; in vmxnet3_unmap_pkt() 402 vmxnet3_unmap_tx_buf(tq->buf_info + tq->tx_ring.next2comp, in vmxnet3_unmap_pkt() 482 tbi = tq->buf_info + tq->tx_ring.next2comp; in vmxnet3_tq_cleanup() 501 BUG_ON(tq->buf_info[i].map_type != VMXNET3_MAP_NONE); in vmxnet3_tq_cleanup() 539 kfree(tq->buf_info); in vmxnet3_tq_destroy() 540 tq->buf_info = NULL; in vmxnet3_tq_destroy() 581 memset(tq->buf_info, 0, sizeof(tq->buf_info[0]) * tq->tx_ring.size); in vmxnet3_tq_init() 583 tq->buf_info[i].map_type = VMXNET3_MAP_NONE; in vmxnet3_tq_init() 594 tq->comp_ring.base || tq->buf_info); in vmxnet3_tq_create() [all …]
|
D | vmxnet3_int.h | 261 struct vmxnet3_tx_buf_info *buf_info; member 338 struct vmxnet3_rx_buf_info *buf_info[2]; member
|
D | vmxnet3_xdp.c | 133 tbi = tq->buf_info + tq->tx_ring.next2fill; in vmxnet3_xdp_xmit_frame()
|
/linux-6.12.1/drivers/scsi/bfa/ |
D | bfad_bsg.c | 3269 struct bfad_buf_info *buf_base, *buf_info; in bfad_fcxp_map_sg() local 3283 buf_info = buf_base; in bfad_fcxp_map_sg() 3284 buf_info->size = payload_len; in bfad_fcxp_map_sg() 3285 buf_info->virt = dma_alloc_coherent(&bfad->pcidev->dev, in bfad_fcxp_map_sg() 3286 buf_info->size, &buf_info->phys, in bfad_fcxp_map_sg() 3288 if (!buf_info->virt) in bfad_fcxp_map_sg() 3292 memcpy(buf_info->virt, payload_kbuf, buf_info->size); in bfad_fcxp_map_sg() 3297 sg_table->sg_len = buf_info->size; in bfad_fcxp_map_sg() 3298 sg_table->sg_addr = (void *)(size_t) buf_info->phys; in bfad_fcxp_map_sg() 3314 struct bfad_buf_info *buf_info = buf_base; in bfad_fcxp_free_mem() local [all …]
|
/linux-6.12.1/kernel/bpf/ |
D | verifier.c | 6179 const char *buf_info, in __check_buffer_access() argument 6186 regno, buf_info, off, size); in __check_buffer_access() 6224 const char *buf_info = type_is_rdonly_mem(reg->type) ? "rdonly" : "rdwr"; in check_buffer_access() local 6227 err = __check_buffer_access(env, buf_info, reg, regno, off, size); in check_buffer_access()
|