/linux-6.12.1/drivers/vfio/pci/mlx5/ |
D | cmd.c | 548 struct mlx5vf_async_data *async_data) in mlx5vf_save_callback_complete() argument 550 kvfree(async_data->out); in mlx5vf_save_callback_complete() 557 struct mlx5vf_async_data *async_data = container_of(_work, in mlx5vf_mig_file_cleanup_cb() local 559 struct mlx5_vf_migration_file *migf = container_of(async_data, in mlx5vf_mig_file_cleanup_cb() 560 struct mlx5_vf_migration_file, async_data); in mlx5vf_mig_file_cleanup_cb() 563 if (async_data->status) { in mlx5vf_mig_file_cleanup_cb() 564 mlx5vf_put_data_buffer(async_data->buf); in mlx5vf_mig_file_cleanup_cb() 565 if (async_data->header_buf) in mlx5vf_mig_file_cleanup_cb() 566 mlx5vf_put_data_buffer(async_data->header_buf); in mlx5vf_mig_file_cleanup_cb() 567 if (!async_data->stop_copy_chunk && in mlx5vf_mig_file_cleanup_cb() [all …]
|
D | cmd.h | 116 struct mlx5vf_async_data async_data; member
|
D | main.c | 662 INIT_WORK(&migf->async_data.work, mlx5vf_mig_file_cleanup_cb); in mlx5vf_pci_save_device_data() 1072 cancel_work_sync(&mvdev->saving_migf->async_data.work); in mlx5vf_disable_fds()
|
/linux-6.12.1/io_uring/ |
D | futex.c | 49 req->async_data = NULL; in __io_futex_complete() 56 struct io_futex_data *ifd = req->async_data; in io_futex_complete() 68 struct futex_vector *futexv = req->async_data; in io_futexv_complete() 80 kfree(req->async_data); in io_futexv_complete() 97 struct io_futex_data *ifd = req->async_data; in __io_futex_cancel() 237 req->async_data = futexv; in io_futexv_prep() 268 struct futex_vector *futexv = req->async_data; in io_futexv_wait() 284 req->async_data = NULL; in io_futexv_wait() 340 req->async_data = ifd; in io_futex_wait()
|
D | uring_cmd.c | 27 req->async_data = cache; in io_uring_async_get() 31 return req->async_data; in io_uring_async_get() 38 struct uring_cache *cache = req->async_data; in io_req_uring_cleanup() 44 req->async_data = NULL; in io_req_uring_cleanup() 194 memcpy(req->async_data, sqe, uring_sqe_size(req->ctx)); in io_uring_cmd_prep_setup() 195 ioucmd->sqe = req->async_data; in io_uring_cmd_prep_setup() 255 struct uring_cache *cache = req->async_data; in io_uring_cmd()
|
D | waitid.c | 37 struct io_waitid_async *iwa = req->async_data; in io_waitid_free() 40 kfree(req->async_data); in io_waitid_free() 41 req->async_data = NULL; in io_waitid_free() 140 struct io_waitid_async *iwa = req->async_data; in __io_waitid_cancel() 210 struct io_waitid_async *iwa = req->async_data; in io_waitid_drop_issue_ref() 227 struct io_waitid_async *iwa = req->async_data; in io_waitid_cb() 309 iwa = req->async_data; in io_waitid()
|
D | timeout.c | 41 struct io_timeout_data *data = req->async_data; in io_is_timeout_noseq() 71 struct io_timeout_data *data = req->async_data; in io_timeout_complete() 92 struct io_timeout_data *io = req->async_data; in io_kill_timeout() 220 struct io_timeout_data *io = link->async_data; in __io_disarm_linked_timeout() 276 io = req->async_data; in io_timeout_extract() 389 io = req->async_data; in io_linked_timeout_update() 411 data = req->async_data; in io_timeout_update() 530 data = req->async_data; in __io_timeout_prep() 570 struct io_timeout_data *data = req->async_data; in io_timeout() 628 struct io_timeout_data *data = req->async_data; in io_queue_linked_timeout()
|
D | rw.c | 160 struct io_async_rw *rw = req->async_data; in io_rw_recycle() 171 req->async_data = NULL; in io_rw_recycle() 224 req->async_data = rw; in io_rw_alloc_async() 229 rw = req->async_data; in io_rw_alloc_async() 251 rw = req->async_data; in io_prep_rw_setup() 347 io = req->async_data; in io_prep_rw_fixed() 389 io_rw_iovec_free(req->async_data); in io_readv_writev_cleanup() 412 struct io_async_rw *io = req->async_data; in io_resubmit_prep() 497 struct io_async_rw *io = req->async_data; in io_fixup_rw_res() 735 struct io_async_rw *io = req->async_data; in io_rw_should_retry() [all …]
|
D | net.c | 140 struct io_async_msghdr *hdr = req->async_data; in io_netmsg_recycle() 154 req->async_data = NULL; in io_netmsg_recycle() 172 req->async_data = hdr; in io_msg_alloc_async() 177 hdr = req->async_data; in io_msg_alloc_async() 354 struct io_async_msghdr *io = req->async_data; in io_sendmsg_recvmsg_cleanup() 362 struct io_async_msghdr *kmsg = req->async_data; in io_send_setup() 528 struct io_async_msghdr *kmsg = req->async_data; in io_sendmsg() 578 struct io_async_msghdr *kmsg = req->async_data; in io_send() 977 struct io_async_msghdr *kmsg = req->async_data; in io_recvmsg() 1130 struct io_async_msghdr *kmsg = req->async_data; in io_recv() [all …]
|
D | poll.c | 110 return req->async_data; in io_poll_get_double() 557 (struct io_poll **) &pt->req->async_data); in io_poll_queue_proc()
|
D | io_uring.c | 416 kfree(req->async_data); in io_clean_op() 417 req->async_data = NULL; in io_clean_op() 950 req->async_data = NULL; in io_preinit_req() 1650 req->async_data = kmalloc(def->async_size, GFP_KERNEL); in io_alloc_async_data() 1651 if (req->async_data) { in io_alloc_async_data()
|
/linux-6.12.1/drivers/infiniband/hw/mlx5/ |
D | devx.c | 1747 struct devx_async_data *async_data = in devx_query_callback() local 1749 struct devx_async_cmd_event_file *ev_file = async_data->ev_file; in devx_query_callback() 1759 list_add_tail(&async_data->list, &ev_queue->event_list); in devx_query_callback() 1783 struct devx_async_data *async_data; in UVERBS_HANDLER() local 1817 async_data = kvzalloc(struct_size(async_data, hdr.out_data, in UVERBS_HANDLER() 1819 if (!async_data) { in UVERBS_HANDLER() 1824 err = uverbs_copy_from(&async_data->hdr.wr_id, attrs, in UVERBS_HANDLER() 1829 async_data->cmd_out_len = cmd_out_len; in UVERBS_HANDLER() 1830 async_data->mdev = mdev; in UVERBS_HANDLER() 1831 async_data->ev_file = ev_file; in UVERBS_HANDLER() [all …]
|
/linux-6.12.1/drivers/net/ethernet/qlogic/qed/ |
D | qed_iwarp.c | 739 union async_output *async_data; in qed_iwarp_mpa_received() local 745 async_data = &ep->ep_buffer_virt->async_output; in qed_iwarp_mpa_received() 747 mpa_rev = async_data->mpa_request.mpa_handshake_mode; in qed_iwarp_mpa_received() 750 async_data->mpa_request.ulp_data_len, in qed_iwarp_mpa_received() 801 async_data->mpa_request.ulp_data_len, mpa_hdr_size); in qed_iwarp_mpa_received() 806 ulp_data_len = le16_to_cpu(async_data->mpa_request.ulp_data_len); in qed_iwarp_mpa_received() 944 union async_output *async_data; in qed_iwarp_parse_private_data() local 960 async_data = &ep->ep_buffer_virt->async_output; in qed_iwarp_parse_private_data() 963 ulp_data_len = le16_to_cpu(async_data->mpa_response.ulp_data_len); in qed_iwarp_parse_private_data()
|
/linux-6.12.1/drivers/scsi/be2iscsi/ |
D | be_main.c | 1632 pasync_ctx->async_data.buffer_size) in beiscsi_hdl_fwd_pdu() 1744 pasync_sge = pasync_ctx->async_data.ring_base; in beiscsi_hdq_post_handles() 1745 pi = pasync_ctx->async_data.pi; in beiscsi_hdq_post_handles() 1771 pasync_ctx->async_data.pi = pi; in beiscsi_hdq_post_handles() 2832 pasync_ctx->async_data.ring_base = in hwi_init_async_pdu_ctx() 2845 pasync_ctx->async_data.handle_base = in hwi_init_async_pdu_ctx() 2853 pasync_ctx->async_data.handle_base; in hwi_init_async_pdu_ctx() 2873 pasync_ctx->async_data.pi = 0; in hwi_init_async_pdu_ctx() 2874 pasync_ctx->async_data.buffer_size = p->defpdu_data_sz; in hwi_init_async_pdu_ctx() 2875 pasync_ctx->async_data.va_base = in hwi_init_async_pdu_ctx() [all …]
|
D | be_main.h | 591 struct hd_async_buf_context async_data; member
|
/linux-6.12.1/include/linux/ |
D | io_uring_types.h | 651 void *async_data; member
|