/linux-6.12.1/block/ |
D | bsg-lib.c | 49 job = blk_mq_rq_to_pdu(rq); in bsg_transport_sg_io_fn() 207 struct bsg_job *job = blk_mq_rq_to_pdu(rq); in bsg_complete() 234 struct bsg_job *job = blk_mq_rq_to_pdu(req); in bsg_prepare_job() 291 ret = bset->job_fn(blk_mq_rq_to_pdu(req)); in bsg_queue_rq() 304 struct bsg_job *job = blk_mq_rq_to_pdu(req); in bsg_init_rq() 315 struct bsg_job *job = blk_mq_rq_to_pdu(req); in bsg_exit_rq()
|
/linux-6.12.1/rust/helpers/ |
D | blk.c | 8 return blk_mq_rq_to_pdu(rq); in rust_helper_blk_mq_rq_to_pdu()
|
/linux-6.12.1/include/scsi/ |
D | scsi_tcq.h | 39 return blk_mq_rq_to_pdu(req); in scsi_host_find_tag()
|
/linux-6.12.1/drivers/nvme/host/ |
D | apple.c | 337 struct apple_nvme_iod *iod = blk_mq_rq_to_pdu(req); in apple_nvme_iod_list() 345 struct apple_nvme_iod *iod = blk_mq_rq_to_pdu(req); in apple_nvme_free_prps() 360 struct apple_nvme_iod *iod = blk_mq_rq_to_pdu(req); in apple_nvme_unmap_data() 397 struct apple_nvme_iod *iod = blk_mq_rq_to_pdu(req); in apple_nvme_setup_prps() 492 struct apple_nvme_iod *iod = blk_mq_rq_to_pdu(req); in apple_nvme_setup_prp_simple() 511 struct apple_nvme_iod *iod = blk_mq_rq_to_pdu(req); in apple_nvme_map_data() 551 struct apple_nvme_iod *iod = blk_mq_rq_to_pdu(req); in apple_nvme_unmap_rq() 739 struct apple_nvme_iod *iod = blk_mq_rq_to_pdu(req); in apple_nvme_queue_rq() 788 struct apple_nvme_iod *iod = blk_mq_rq_to_pdu(req); in apple_nvme_init_request() 879 struct apple_nvme_iod *iod = blk_mq_rq_to_pdu(req); in apple_nvme_timeout()
|
D | rdma.c | 288 struct nvme_rdma_request *req = blk_mq_rq_to_pdu(rq); in nvme_rdma_exit_request() 298 struct nvme_rdma_request *req = blk_mq_rq_to_pdu(rq); in nvme_rdma_init_request() 1223 struct nvme_rdma_request *req = blk_mq_rq_to_pdu(rq); in nvme_rdma_dma_unmap_req() 1240 struct nvme_rdma_request *req = blk_mq_rq_to_pdu(rq); in nvme_rdma_unmap_data() 1475 struct nvme_rdma_request *req = blk_mq_rq_to_pdu(rq); in nvme_rdma_dma_map_req() 1535 struct nvme_rdma_request *req = blk_mq_rq_to_pdu(rq); in nvme_rdma_map_data() 1711 req = blk_mq_rq_to_pdu(rq); in nvme_rdma_process_nvme_rsp() 1953 struct nvme_rdma_request *req = blk_mq_rq_to_pdu(rq); in nvme_rdma_complete_timed_out() 1962 struct nvme_rdma_request *req = blk_mq_rq_to_pdu(rq); in nvme_rdma_timeout() 2005 struct nvme_rdma_request *req = blk_mq_rq_to_pdu(rq); in nvme_rdma_queue_rq() [all …]
|
D | tcp.c | 522 struct nvme_tcp_request *req = blk_mq_rq_to_pdu(rq); in nvme_tcp_exit_request() 532 struct nvme_tcp_request *req = blk_mq_rq_to_pdu(rq); in nvme_tcp_init_request() 613 req = blk_mq_rq_to_pdu(rq); in nvme_tcp_process_nvme_cqe() 729 req = blk_mq_rq_to_pdu(rq); in nvme_tcp_handle_r2t() 827 struct nvme_tcp_request *req = blk_mq_rq_to_pdu(rq); in nvme_tcp_recv_data() 914 struct nvme_tcp_request *req = blk_mq_rq_to_pdu(rq); in nvme_tcp_recv_ddgst() 927 struct nvme_tcp_request *req = blk_mq_rq_to_pdu(rq); in nvme_tcp_recv_ddgst() 2464 struct nvme_tcp_request *req = blk_mq_rq_to_pdu(rq); in nvme_tcp_complete_timed_out() 2473 struct nvme_tcp_request *req = blk_mq_rq_to_pdu(rq); in nvme_tcp_timeout() 2513 struct nvme_tcp_request *req = blk_mq_rq_to_pdu(rq); in nvme_tcp_map_data() [all …]
|
D | pci.c | 423 struct nvme_iod *iod = blk_mq_rq_to_pdu(req); in nvme_pci_init_request() 527 struct nvme_iod *iod = blk_mq_rq_to_pdu(req); in nvme_free_prps() 542 struct nvme_iod *iod = blk_mq_rq_to_pdu(req); in nvme_unmap_data() 582 struct nvme_iod *iod = blk_mq_rq_to_pdu(req); in nvme_pci_setup_prps() 688 struct nvme_iod *iod = blk_mq_rq_to_pdu(req); in nvme_pci_setup_sgls() 734 struct nvme_iod *iod = blk_mq_rq_to_pdu(req); in nvme_setup_prp_simple() 755 struct nvme_iod *iod = blk_mq_rq_to_pdu(req); in nvme_setup_sgl_simple() 772 struct nvme_iod *iod = blk_mq_rq_to_pdu(req); in nvme_map_data() 828 struct nvme_iod *iod = blk_mq_rq_to_pdu(req); in nvme_map_metadata() 840 struct nvme_iod *iod = blk_mq_rq_to_pdu(req); in nvme_prep_rq() [all …]
|
/linux-6.12.1/drivers/scsi/ |
D | scsi_lib.c | 314 scmd = blk_mq_rq_to_pdu(req); in scsi_execute_cmd() 627 struct scsi_cmnd *cmd = blk_mq_rq_to_pdu(req); in scsi_end_request() 1207 struct scsi_cmnd *cmd = blk_mq_rq_to_pdu(rq); in scsi_initialize_rq() 1236 scsi_mq_uninit_cmd(blk_mq_rq_to_pdu(rq)); in scsi_cleanup_rq() 1259 struct scsi_cmnd *cmd = blk_mq_rq_to_pdu(req); in scsi_setup_scsi_cmnd() 1506 struct scsi_cmnd *cmd = blk_mq_rq_to_pdu(rq); in scsi_complete() 1629 struct scsi_cmnd *cmd = blk_mq_rq_to_pdu(req); in scsi_prepare_cmd() 1778 struct scsi_cmnd *cmd = blk_mq_rq_to_pdu(req); in scsi_mq_set_rq_budget_token() 1785 struct scsi_cmnd *cmd = blk_mq_rq_to_pdu(req); in scsi_mq_get_rq_budget_token() 1797 struct scsi_cmnd *cmd = blk_mq_rq_to_pdu(req); in scsi_queue_rq() [all …]
|
D | scsi_debugfs.c | 56 struct scsi_cmnd *cmd = blk_mq_rq_to_pdu(rq); in scsi_show_rq()
|
D | hosts.c | 594 struct scsi_cmnd *cmd = blk_mq_rq_to_pdu(rq); in scsi_host_check_in_flight() 688 struct scsi_cmnd *scmd = blk_mq_rq_to_pdu(rq); in complete_all_cmds_iter() 724 struct scsi_cmnd *sc = blk_mq_rq_to_pdu(req); in __scsi_host_busy_iter_fn()
|
D | scsi_ioctl.c | 351 struct scsi_cmnd *scmd = blk_mq_rq_to_pdu(rq); in scsi_fill_sghdr_rq() 375 struct scsi_cmnd *scmd = blk_mq_rq_to_pdu(rq); in scsi_complete_sghdr_rq() 445 scmd = blk_mq_rq_to_pdu(rq); in sg_io() 541 scmd = blk_mq_rq_to_pdu(rq); in sg_scsi_ioctl()
|
D | scsi_bsg.c | 34 scmd = blk_mq_rq_to_pdu(rq); in scsi_bsg_sg_io_fn()
|
/linux-6.12.1/drivers/md/ |
D | dm-rq.c | 124 return blk_mq_rq_to_pdu(rq); in tio_from_request() 461 struct dm_rq_target_io *tio = blk_mq_rq_to_pdu(rq); in dm_mq_init_request() 481 struct dm_rq_target_io *tio = blk_mq_rq_to_pdu(rq); in dm_mq_queue_rq()
|
/linux-6.12.1/drivers/nvme/target/ |
D | loop.c | 76 struct nvme_loop_iod *iod = blk_mq_rq_to_pdu(req); in nvme_loop_complete_rq() 137 struct nvme_loop_iod *iod = blk_mq_rq_to_pdu(req); in nvme_loop_queue_rq() 208 struct nvme_loop_iod *iod = blk_mq_rq_to_pdu(req); in nvme_loop_init_request() 212 return nvme_loop_init_iod(ctrl, blk_mq_rq_to_pdu(req), in nvme_loop_init_request()
|
/linux-6.12.1/drivers/mmc/core/ |
D | queue.h | 27 return blk_mq_rq_to_pdu(rq); in req_to_mmc_queue_req()
|
/linux-6.12.1/drivers/block/ |
D | virtio_blk.c | 337 struct virtblk_req *vbr = blk_mq_rq_to_pdu(req); in virtblk_request_done() 432 struct virtblk_req *vbr = blk_mq_rq_to_pdu(req); in virtio_queue_rq() 469 struct virtblk_req *vbr = blk_mq_rq_to_pdu(req); in virtblk_prep_rq_batch() 485 struct virtblk_req *vbr = blk_mq_rq_to_pdu(req); in virtblk_add_req_batch() 575 vbr = blk_mq_rq_to_pdu(req); in virtblk_submit_zone_report() 821 vbr = blk_mq_rq_to_pdu(req); in virtblk_get_id() 1198 virtblk_unmap_data(req, blk_mq_rq_to_pdu(req)); in virtblk_complete_batch()
|
D | nbd.c | 396 struct nbd_cmd *cmd = blk_mq_rq_to_pdu(req); in nbd_complete_rq() 462 struct nbd_cmd *cmd = blk_mq_rq_to_pdu(req); in nbd_xmit_timeout() 831 cmd = blk_mq_rq_to_pdu(req); in nbd_handle_reply() 968 struct nbd_cmd *cmd = blk_mq_rq_to_pdu(req); in nbd_clear_req() 1123 struct nbd_cmd *cmd = blk_mq_rq_to_pdu(bd->rq); in nbd_queue_rq() 1809 struct nbd_cmd *cmd = blk_mq_rq_to_pdu(rq); in nbd_init_request()
|
/linux-6.12.1/rust/kernel/block/mq/ |
D | request.rs | 138 unsafe { bindings::blk_mq_rq_to_pdu(request_ptr).cast::<RequestDataWrapper>() }; in wrapper_ptr()
|
D | operations.rs | 210 let pdu = unsafe { bindings::blk_mq_rq_to_pdu(rq) }.cast::<RequestDataWrapper>(); in exit_request_callback()
|
/linux-6.12.1/drivers/mtd/ubi/ |
D | block.c | 182 struct ubiblock_pdu *pdu = blk_mq_rq_to_pdu(req); in ubiblock_read() 317 struct ubiblock_pdu *pdu = blk_mq_rq_to_pdu(req); in ubiblock_init_request()
|
/linux-6.12.1/drivers/block/mtip32xx/ |
D | mtip32xx.c | 155 return blk_mq_rq_to_pdu(blk_mq_tag_to_rq(dd->tags.tags[0], tag)); in mtip_cmd_from_tag() 999 int_cmd = blk_mq_rq_to_pdu(rq); in mtip_exec_internal_command() 2416 struct mtip_cmd *cmd = blk_mq_rq_to_pdu(rq); in mtip_softirq_done_fn() 2431 struct mtip_cmd *cmd = blk_mq_rq_to_pdu(req); in mtip_abort_cmd() 3238 struct mtip_cmd *cmd = blk_mq_rq_to_pdu(rq); in mtip_check_unal_depth() 3262 struct mtip_cmd *cmd = blk_mq_rq_to_pdu(rq); in mtip_issue_reserved_cmd() 3300 struct mtip_cmd *cmd = blk_mq_rq_to_pdu(rq); in mtip_queue_rq() 3321 struct mtip_cmd *cmd = blk_mq_rq_to_pdu(rq); in mtip_free_cmd() 3334 struct mtip_cmd *cmd = blk_mq_rq_to_pdu(rq); in mtip_init_cmd() 3350 struct mtip_cmd *cmd = blk_mq_rq_to_pdu(req); in mtip_cmd_timeout()
|
/linux-6.12.1/drivers/block/null_blk/ |
D | main.c | 827 struct nullb_cmd *cmd = blk_mq_rq_to_pdu(rq); in null_complete_rq() 1431 struct nullb_cmd *cmd = blk_mq_rq_to_pdu(rq); in should_timeout_request() 1439 struct nullb_cmd *cmd = blk_mq_rq_to_pdu(rq); in should_requeue_request() 1541 cmd = blk_mq_rq_to_pdu(req); in null_poll() 1556 struct nullb_cmd *cmd = blk_mq_rq_to_pdu(rq); in null_timeout_rq() 1590 struct nullb_cmd *cmd = blk_mq_rq_to_pdu(rq); in null_queue_rq()
|
/linux-6.12.1/drivers/s390/block/ |
D | scm_blk.c | 257 error = blk_mq_rq_to_pdu(scmrq->request[i]); in scm_request_finish() 420 blk_status_t *error = blk_mq_rq_to_pdu(req); in scm_blk_request_done()
|
D | dasd_fba.c | 348 blk_mq_rq_to_pdu(req)); in dasd_fba_build_cp_discard() 483 blk_mq_rq_to_pdu(req)); in dasd_fba_build_cp_regular()
|
/linux-6.12.1/drivers/target/ |
D | target_core_pscsi.c | 958 scmd = blk_mq_rq_to_pdu(req); in pscsi_execute_cmd() 1009 struct scsi_cmnd *scmd = blk_mq_rq_to_pdu(req); in pscsi_req_done()
|