Home
last modified time | relevance | path

Searched refs:blk_rq_bytes (Results 1 – 25 of 32) sorted by relevance

12

/linux-6.12.1/block/
Dblk-crypto-internal.h39 return bio_crypt_ctx_mergeable(req->crypt_ctx, blk_rq_bytes(req), in bio_crypt_ctx_back_mergeable()
53 return bio_crypt_ctx_mergeable(req->crypt_ctx, blk_rq_bytes(req), in bio_crypt_ctx_merge_rq()
Dblk-mq.c761 rq->bio, rq->biotail, blk_rq_bytes(rq)); in blk_dump_rq_flags()
797 int total_bytes = blk_rq_bytes(req); in blk_complete_request()
882 if (blk_crypto_rq_has_keyslot(req) && nr_bytes >= blk_rq_bytes(req)) in blk_update_request()
963 if (blk_rq_bytes(req) < blk_rq_cur_bytes(req)) { in blk_update_request()
1051 if (blk_update_request(rq, error, blk_rq_bytes(rq))) in blk_mq_end_request()
1289 blk_rq_bytes(last) >= BLK_PLUG_FLUSH_SIZE)) { in blk_add_rq_to_plug()
3096 if (q->disk && should_fail_request(q->disk->part0, blk_rq_bytes(rq))) in blk_insert_cloned_request()
3184 rq->__data_len = blk_rq_bytes(rq_src); in blk_rq_prep_clone()
Dbsg-lib.c223 buf->payload_len = blk_rq_bytes(req); in bsg_map_buffer()
Dblk-merge.c903 req->__data_len += blk_rq_bytes(next); in attempt_merge()
Dbfq-cgroup.c350 blkg_rwstat_add(&bfqg->stats.bytes, rq->cmd_flags, blk_rq_bytes(rq)); in bfqg_stats_update_legacy_io()
/linux-6.12.1/drivers/scsi/
Dscsi_lib.c732 return blk_rq_bytes(rq); in scsi_rq_err_bytes()
748 BUG_ON(blk_rq_bytes(rq) && !bytes); in scsi_rq_err_bytes()
978 } else if (blk_rq_bytes(req) == 0 && sense_current) { in scsi_io_completion_nz_result()
1066 if (likely(blk_rq_bytes(req) > 0 || blk_stat == BLK_STS_OK)) { in scsi_io_completion()
1073 if (scsi_end_request(req, blk_stat, blk_rq_bytes(req))) in scsi_io_completion()
1142 if (blk_rq_bytes(rq) & rq->q->limits.dma_pad_mask) { in scsi_alloc_sgtables()
1144 (rq->q->limits.dma_pad_mask & ~blk_rq_bytes(rq)) + 1; in scsi_alloc_sgtables()
1272 BUG_ON(blk_rq_bytes(req)); in scsi_setup_scsi_cmnd()
1277 cmd->transfersize = blk_rq_bytes(req); in scsi_setup_scsi_cmnd()
1659 if (blk_rq_bytes(req)) in scsi_prepare_cmd()
Dsd.c2298 good_bytes = blk_rq_bytes(req); in sd_done()
2302 scsi_set_resid(SCpnt, blk_rq_bytes(req)); in sd_done()
/linux-6.12.1/include/linux/
Dblk-mq.h1067 static inline unsigned int blk_rq_bytes(const struct request *rq) in blk_rq_bytes() function
1083 return blk_rq_bytes(rq) >> SECTOR_SHIFT; in blk_rq_sectors()
1106 return blk_rq_bytes(rq); in blk_rq_payload_bytes()
/linux-6.12.1/kernel/trace/
Dblktrace.c850 blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_INSERT, in blk_add_trace_rq_insert()
856 blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_ISSUE, in blk_add_trace_rq_issue()
862 blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_BACKMERGE, in blk_add_trace_rq_merge()
868 blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_REQUEUE, in blk_add_trace_rq_requeue()
1056 __blk_add_trace(bt, blk_rq_pos(rq), blk_rq_bytes(rq), in blk_add_trace_rq_remap()
1083 __blk_add_trace(bt, blk_rq_trace_sector(rq), blk_rq_bytes(rq), 0, in blk_add_driver_data()
/linux-6.12.1/include/scsi/
Dscsi_cmnd.h239 return blk_rq_bytes(scsi_cmd_to_rq(scmd)) >> shift; in scsi_logical_block_count()
/linux-6.12.1/include/trace/events/
Dblock.h211 __entry->bytes = blk_rq_bytes(rq);
/linux-6.12.1/drivers/mtd/ubi/
Dblock.c185 int to_read = blk_rq_bytes(req); in ubiblock_read()
/linux-6.12.1/drivers/block/
Dloop.c332 ret = file->f_op->fallocate(file, mode, pos, blk_rq_bytes(rq)); in lo_fallocate()
360 if (!cmd->use_aio || cmd->ret < 0 || cmd->ret == blk_rq_bytes(rq) || in lo_complete_rq()
458 iov_iter_bvec(&iter, rw, bvec, nr_bvec, blk_rq_bytes(rq)); in lo_rw_aio()
Dublk_drv.c868 const unsigned int rq_bytes = blk_rq_bytes(req); in ublk_map_io()
892 const unsigned int rq_bytes = blk_rq_bytes(req); in ublk_unmap_io()
1156 if (unlikely(mapped_bytes != blk_rq_bytes(req))) { in __ublk_rq_task_work()
1829 if (offset > blk_rq_bytes(req)) in __ublk_check_and_get_req()
Dnbd.c526 blk_rq_bytes(req), (req->timeout / HZ) * cmd->retries); in nbd_xmit_timeout()
681 request.len = htonl(blk_rq_bytes(req)); in nbd_send_cmd()
690 (unsigned long long)blk_rq_pos(req) << 9, blk_rq_bytes(req)); in nbd_send_cmd()
Dataflop.c464 blk_rq_bytes(fd_request))); in fd_end_request_cur()
/linux-6.12.1/drivers/s390/block/
Dscm_blk.c190 aidaw = scm_aidaw_fetch(scmrq, blk_rq_bytes(req)); in scm_request_prepare()
/linux-6.12.1/drivers/memstick/core/
Dmspro_block.c651 count = blk_rq_bytes(msb->block_req); in mspro_block_issue_req()
695 t_len = blk_rq_bytes(msb->block_req); in mspro_block_complete_req()
Dms_block.c1916 blk_rq_bytes(req), &len); in msb_io_work()
1919 blk_rq_bytes(req), &len); in msb_io_work()
/linux-6.12.1/drivers/nvme/host/
Dcore.c343 blk_rq_bytes(req) >> ns->head->lba_shift, in nvme_log_error()
923 cpu_to_le16((blk_rq_bytes(req) >> ns->head->lba_shift) - 1); in nvme_setup_write_zeroes()
954 if (blk_rq_bytes(req) > queue_atomic_write_unit_max_bytes(q)) in nvme_valid_atomic_write()
960 u64 end = start + blk_rq_bytes(req) - 1; in nvme_valid_atomic_write()
963 if (blk_rq_bytes(req) > boundary_bytes) in nvme_valid_atomic_write()
1000 cpu_to_le16((blk_rq_bytes(req) >> ns->head->lba_shift) - 1); in nvme_setup_rw()
Dmultipath.c158 blk_rq_bytes(rq) >> SECTOR_SHIFT, in nvme_mpath_end_request()
/linux-6.12.1/arch/um/drivers/
Dubd_kern.c1207 io_req->io_desc[0].length = blk_rq_bytes(req); in ubd_map_req()
/linux-6.12.1/drivers/block/rnbd/
Drnbd-clt.c1004 msg.bi_size = cpu_to_le32(blk_rq_bytes(rq)); in rnbd_client_xfer_request()
/linux-6.12.1/drivers/block/null_blk/
Dmain.c1279 if (atomic_long_sub_return(blk_rq_bytes(rq), &nullb->cur_bytes) < 0) { in null_handle_throttled()
/linux-6.12.1/drivers/md/
Ddm-mpath.c512 size_t nr_bytes = blk_rq_bytes(rq); in multipath_clone_and_map()

12