/linux-6.12.1/lib/ |
D | iov_iter.c | 94 n -= fault_in_readable(i->ubuf + i->iov_offset, n); in fault_in_iov_iter_readable() 102 for (p = iter_iov(i), skip = i->iov_offset; count; p++, skip = 0) { in fault_in_iov_iter_readable() 137 n -= fault_in_safe_writeable(i->ubuf + i->iov_offset, n); in fault_in_iov_iter_writeable() 145 for (p = iter_iov(i), skip = i->iov_offset; count; p++, skip = 0) { in fault_in_iov_iter_writeable() 173 .iov_offset = 0, in iov_iter_init() 501 size += i->iov_offset; in iov_iter_bvec_advance() 508 i->iov_offset = size; in iov_iter_bvec_advance() 521 size += i->iov_offset; // from beginning of current segment in iov_iter_iovec_advance() 527 i->iov_offset = size; in iov_iter_iovec_advance() 546 size += i->iov_offset; /* From beginning of current segment. */ in iov_iter_folioq_advance() [all …]
|
D | scatterlist.c | 1170 unsigned long start = iter->iov_offset; in extract_bvec_to_sg() 1216 unsigned long start = iter->iov_offset; in extract_kvec_to_sg() 1279 size_t offset = iter->iov_offset; in extract_folioq_to_sg() 1321 iter->iov_offset = offset; in extract_folioq_to_sg() 1339 loff_t start = iter->xarray_start + iter->iov_offset; in extract_xarray_to_sg()
|
D | kunit_iov_iter.c | 454 KUNIT_EXPECT_EQ(test, iter.iov_offset, pr->to % PAGE_SIZE); in iov_kunit_copy_to_folioq() 516 KUNIT_EXPECT_EQ(test, iter.iov_offset, pr->to % PAGE_SIZE); in iov_kunit_copy_from_folioq() 615 KUNIT_EXPECT_EQ(test, iter.iov_offset, size); in iov_kunit_copy_to_xarray() 673 KUNIT_EXPECT_EQ(test, iter.iov_offset, size); in iov_kunit_copy_from_xarray() 1008 KUNIT_EXPECT_EQ(test, iter.iov_offset, pr->to - pr->from); in iov_kunit_extract_pages_xarray()
|
/linux-6.12.1/include/linux/ |
D | iov_iter.h | 30 remain = step(base + iter->iov_offset, 0, len, priv, priv2); in iterate_ubuf() 32 iter->iov_offset += progress; in iterate_ubuf() 45 size_t progress = 0, skip = iter->iov_offset; in iterate_iovec() 66 iter->iov_offset = skip; in iterate_iovec() 79 size_t progress = 0, skip = iter->iov_offset; in iterate_kvec() 100 iter->iov_offset = skip; in iterate_kvec() 113 size_t progress = 0, skip = iter->iov_offset; in iterate_bvec() 139 iter->iov_offset = skip; in iterate_bvec() 153 size_t progress = 0, skip = iter->iov_offset; in iterate_folioq() 193 iter->iov_offset = skip; in iterate_folioq() [all …]
|
D | uio.h | 38 size_t iov_offset; member 47 size_t iov_offset; member 92 #define iter_iov_addr(iter) (iter_iov(iter)->iov_base + (iter)->iov_offset) 93 #define iter_iov_len(iter) (iter_iov(iter)->iov_len - (iter)->iov_offset) 103 state->iov_offset = iter->iov_offset; in iov_iter_save_state()
|
/linux-6.12.1/drivers/net/ethernet/google/gve/ |
D | gve_tx.c | 105 iov[0].iov_offset = fifo->head; in gve_tx_alloc_fifo() 116 iov[1].iov_offset = 0; /* Start of fifo*/ in gve_tx_alloc_fifo() 556 u64 iov_offset, u64 iov_len) in gve_dma_sync_for_device() argument 558 u64 last_page = (iov_offset + iov_len - 1) / PAGE_SIZE; in gve_dma_sync_for_device() 559 u64 first_page = iov_offset / PAGE_SIZE; in gve_dma_sync_for_device() 603 info->iov[hdr_nfrags - 1].iov_offset, skb->len); in gve_tx_add_skb_copy() 606 tx->tx_fifo.base + info->iov[hdr_nfrags - 1].iov_offset, in gve_tx_add_skb_copy() 609 info->iov[hdr_nfrags - 1].iov_offset, in gve_tx_add_skb_copy() 626 info->iov[i].iov_offset); in gve_tx_add_skb_copy() 629 tx->tx_fifo.base + info->iov[i].iov_offset, in gve_tx_add_skb_copy() [all …]
|
D | gve.h | 325 u32 iov_offset; /* offset into this segment */ member
|
/linux-6.12.1/fs/netfs/ |
D | iterator.c | 115 size_t skip = iter->iov_offset + start_offset; in netfs_limit_bvec() 156 loff_t pos = iter->xarray_start + iter->iov_offset; in netfs_limit_xarray() 215 start_offset += iter->iov_offset; in netfs_limit_folioq()
|
D | read_pgpriv2.c | 112 wreq->io_iter.iov_offset = cache->submit_off; in netfs_pgpriv2_copy_folio() 125 wreq->io_iter.iov_offset = 0; in netfs_pgpriv2_copy_folio()
|
/linux-6.12.1/fs/smb/client/ |
D | compress.c | 166 loff_t start = iter->xarray_start + iter->iov_offset; in collect_sample()
|
D | smbdirect.c | 2439 unsigned long start = iter->iov_offset; in smb_extract_bvec_to_rdma() 2480 unsigned long start = iter->iov_offset; in smb_extract_kvec_to_rdma() 2539 size_t offset = iter->iov_offset; in smb_extract_folioq_to_rdma() 2580 iter->iov_offset = offset; in smb_extract_folioq_to_rdma()
|
/linux-6.12.1/io_uring/ |
D | rsrc.c | 1132 iter->iov_offset = offset; in io_import_fixed() 1143 iter->iov_offset = offset & ((1UL << imu->folio_shift) - 1); in io_import_fixed()
|
D | net.c | 1315 bi.bi_bvec_done = from->iov_offset; in io_sg_from_iter() 1334 from->iov_offset = bi.bi_bvec_done; in io_sg_from_iter()
|
D | rw.c | 656 addr = iter->ubuf + iter->iov_offset; in loop_rw_iter()
|
/linux-6.12.1/net/9p/ |
D | trans_virtio.c | 350 p = data->kvec->iov_base + data->iov_offset; in p9_get_mapped_pages()
|
/linux-6.12.1/include/trace/events/ |
D | afs.h | 856 __entry->offset = msg->msg_iter.xarray_start + msg->msg_iter.iov_offset; 880 __entry->offset = msg->msg_iter.xarray_start + msg->msg_iter.iov_offset;
|
/linux-6.12.1/drivers/nvme/host/ |
D | tcp.c | 302 return req->iter.bvec->bv_offset + req->iter.iov_offset; in nvme_tcp_req_cur_offset() 352 req->iter.iov_offset = offset; in nvme_tcp_init_iter()
|
/linux-6.12.1/block/ |
D | bio.c | 1222 bio->bi_iter.bi_bvec_done = iter->iov_offset; in bio_iov_bvec_set()
|
/linux-6.12.1/net/vmw_vsock/ |
D | virtio_transport_common.c | 51 if (iov_iter->iov_offset) in virtio_transport_can_zcopy()
|
/linux-6.12.1/drivers/block/ |
D | loop.c | 459 iter.iov_offset = offset; in lo_rw_aio()
|
/linux-6.12.1/net/ceph/ |
D | messenger_v2.c | 156 it->bvec->bv_len - it->iov_offset), in do_try_sendpage() 157 it->bvec->bv_offset + it->iov_offset); in do_try_sendpage()
|
/linux-6.12.1/fs/afs/ |
D | dir.c | 200 req->iter->iov_offset, iov_iter_count(req->iter)); in afs_dir_dump()
|
/linux-6.12.1/fs/fuse/ |
D | file.c | 1460 return (unsigned long)iter_iov(ii)->iov_base + ii->iov_offset; in fuse_get_user_addr()
|