Searched refs:sqes (Results 1 – 8 of 8) sorted by relevance
/linux-6.12.1/tools/include/io_uring/ |
D | mini_liburing.h | 35 struct io_uring_sqe *sqes; member 89 sq->sqes = mmap(0, size, PROT_READ | PROT_WRITE, in io_uring_mmap() 91 if (sq->sqes == MAP_FAILED) { in io_uring_mmap() 103 munmap(sq->sqes, p->sq_entries * sizeof(struct io_uring_sqe)); in io_uring_mmap() 158 return &sq->sqes[sq->sqe_tail++ & *sq->kring_mask]; in io_uring_get_sqe() 224 munmap(sq->sqes, *sq->kring_entries * sizeof(struct io_uring_sqe)); in io_uring_queue_exit()
|
/linux-6.12.1/io_uring/ |
D | uring_cmd.h | 4 struct io_uring_sqe sqes[2]; member
|
/linux-6.12.1/tools/testing/selftests/x86/ |
D | lam.c | 97 struct io_uring_sqe *sqes; member 429 s->sq_ring.queue.sqes = mmap(0, p.sq_entries * sizeof(struct io_uring_sqe), in mmap_io_uring() 432 if (s->sq_ring.queue.sqes == MAP_FAILED) { in mmap_io_uring() 540 sqe = &ring->sq_ring.queue.sqes[index]; in handle_uring_sq()
|
/linux-6.12.1/drivers/nvme/target/ |
D | passthru.c | 133 id->sqes = min_t(__u8, ((0x6 << 4) | 0x6), id->sqes); in nvmet_passthru_override_id_ctrl()
|
D | admin-cmd.c | 427 id->sqes = (0x6 << 4) | 0x6; in nvmet_execute_identify_ctrl()
|
/linux-6.12.1/drivers/nvme/host/ |
D | apple.c | 134 struct nvme_command *sqes; member 301 memcpy(&q->sqes[tag], cmd, sizeof(*cmd)); in apple_nvme_submit_cmd() 1306 q->sqes = dmam_alloc_coherent(anv->dev, in apple_nvme_queue_alloc() 1309 if (!q->sqes) in apple_nvme_queue_alloc()
|
D | pci.c | 35 #define SQ_SIZE(q) ((q)->q_depth << (q)->sqes) 206 u8 sqes; member 491 memcpy(nvmeq->sq_cmds + (nvmeq->sq_tail << nvmeq->sqes), in nvme_sq_copy_cmd() 1571 nvmeq->sqes = qid ? dev->io_sqes : NVME_ADM_SQES; in nvme_alloc_queue()
|
/linux-6.12.1/include/linux/ |
D | nvme.h | 336 __u8 sqes; member
|