/linux-6.12.1/include/asm-generic/ |
D | qrwlock.h | 40 * queued_read_trylock - try to acquire read lock of a queued rwlock 41 * @lock : Pointer to queued rwlock structure 59 * queued_write_trylock - try to acquire write lock of a queued rwlock 60 * @lock : Pointer to queued rwlock structure 75 * queued_read_lock - acquire read lock of a queued rwlock 76 * @lock: Pointer to queued rwlock structure 91 * queued_write_lock - acquire write lock of a queued rwlock 92 * @lock : Pointer to queued rwlock structure 105 * queued_read_unlock - release read lock of a queued rwlock 106 * @lock : Pointer to queued rwlock structure [all …]
|
D | qspinlock.h | 3 * Queued spinlock 48 * @lock: Pointer to queued spinlock structure 63 * @lock: queued spinlock structure 78 * @lock : Pointer to queued spinlock structure 86 * queued_spin_trylock - try to acquire the queued spinlock 87 * @lock : Pointer to queued spinlock structure 104 * queued_spin_lock - acquire a queued spinlock 105 * @lock: Pointer to queued spinlock structure 120 * queued_spin_unlock - release a queued spinlock 121 * @lock : Pointer to queued spinlock structure [all …]
|
/linux-6.12.1/security/integrity/ima/ |
D | ima_queue_keys.c | 18 * right away or should be queued for processing later. 29 * If custom IMA policy is not loaded then keys queued up 40 * queued up in case custom IMA policy was not loaded. 49 * This function sets up a worker to free queued keys in case 107 bool queued = false; in ima_queue_key() local 117 queued = true; in ima_queue_key() 121 if (!queued) in ima_queue_key() 124 return queued; in ima_queue_key() 128 * ima_process_queued_keys() - process keys queued for measurement 130 * This function sets ima_process_keys to true and processes queued keys. [all …]
|
/linux-6.12.1/block/ |
D | blk-throttle.h | 14 * To avoid such starvation, dispatched bios are queued separately 19 * throtl_qnode is used to keep the queued bios separated by their sources. 20 * Bios are queued to throtl_qnode which in turn is queued to 24 * belongs to a throtl_grp and gets queued on itself or the parent, so 26 * queued and decrementing when dequeued is enough to keep the whole blkg 30 struct list_head node; /* service_queue->queued[] */ 31 struct bio_list bios; /* queued bios */ 39 * Bios queued directly to this service_queue or dispatched from 42 struct list_head queued[2]; /* throtl_qnode [READ/WRITE] */ member 43 unsigned int nr_queued[2]; /* number of queued bios */ [all …]
|
D | blk-throttle.c | 41 /* Total Number of queued bios on READ and WRITE lists */ 154 * @queued: the service_queue->queued[] list @qn belongs to 156 * Add @bio to @qn and put @qn on @queued if it's not already on. 161 struct list_head *queued) in throtl_qnode_add_bio() argument 165 list_add_tail(&qn->node, queued); in throtl_qnode_add_bio() 172 * @queued: the qnode list to peek 174 static struct bio *throtl_peek_queued(struct list_head *queued) in throtl_peek_queued() argument 179 if (list_empty(queued)) in throtl_peek_queued() 182 qn = list_first_entry(queued, struct throtl_qnode, node); in throtl_peek_queued() 190 * @queued: the qnode list to pop a bio from [all …]
|
/linux-6.12.1/drivers/net/wireless/mediatek/mt76/ |
D | debugfs.c | 59 seq_puts(s, " queue | hw-queued | head | tail |\n"); in mt76_queues_read() 67 i, q->queued, q->head, q->tail); in mt76_queues_read() 77 int i, queued; in mt76_rx_queues_read() local 79 seq_puts(s, " queue | hw-queued | head | tail |\n"); in mt76_rx_queues_read() 83 queued = mt76_is_usb(dev) ? q->ndesc - q->queued : q->queued; in mt76_rx_queues_read() 85 i, queued, q->head, q->tail); in mt76_rx_queues_read()
|
/linux-6.12.1/drivers/gpu/drm/ |
D | drm_flip_work.c | 51 list_add_tail(&task->node, &work->queued); in drm_flip_work_queue_task() 79 * drm_flip_work_commit - commit queued work 81 * @wq: the work-queue to run the queued work on 83 * Trigger work previously queued by drm_flip_work_queue() to run 86 * prior), and then from vblank irq commit the queued work. 94 list_splice_tail(&work->queued, &work->commited); in drm_flip_work_commit() 95 INIT_LIST_HEAD(&work->queued); in drm_flip_work_commit() 138 INIT_LIST_HEAD(&work->queued); in drm_flip_work_init() 155 WARN_ON(!list_empty(&work->queued) || !list_empty(&work->commited)); in drm_flip_work_cleanup()
|
/linux-6.12.1/include/linux/ |
D | dynamic_queue_limits.h | 11 * 1) Objects are queued up to some limit specified as number of objects. 14 * 3) Starvation occurs when limit has been reached, all queued data has 17 * 4) Minimizing the amount of queued data is desirable. 24 * dql_avail - returns how many objects are available to be queued based 49 unsigned int num_queued; /* Total ever queued */ 96 * bit trackes time the packet was queued. The dql->history buffer in dql_queue_stall() 116 /* Populate the history with an entry (bit) per queued */ in dql_queue_stall() 122 * Record number of objects queued. Assumes that caller has already checked 146 /* Returns how many objects can be queued, < 0 indicates over limit. */
|
/linux-6.12.1/Documentation/userspace-api/media/mediactl/ |
D | media-request-ioc-queue.rst | 34 If the request was successfully queued, then the file descriptor can be 37 If the request was already queued before, then ``EBUSY`` is returned. 42 Once a request is queued, then the driver is required to gracefully handle 49 queued directly and you next try to queue a request, or vice versa. 62 The request was already queued or the application queued the first
|
D | request-api.rst | 24 buffer queues since in practice only one buffer would be queued at a time. 59 instead of being immediately applied, and buffers queued to a request do not 60 enter the regular buffer queue until the request itself is queued. 66 queued by calling :ref:`MEDIA_REQUEST_IOC_QUEUE` on the request file descriptor. 68 A queued request cannot be modified anymore. 86 a buffer was queued via a request or vice versa will result in an ``EBUSY`` 109 request that has been queued but not yet completed will return ``EBUSY`` 121 longer in use by the kernel. That is, if the request is queued and then the 165 Once the request is fully prepared, it can be queued to the driver: 245 Once the request is fully prepared, it can be queued to the driver:
|
/linux-6.12.1/net/x25/ |
D | x25_in.c | 210 int queued = 0; in x25_state3_machine() local 277 queued = 1; in x25_state3_machine() 315 queued = !sock_queue_rcv_skb(sk, skb); in x25_state3_machine() 319 queued = 1; in x25_state3_machine() 330 return queued; in x25_state3_machine() 418 int queued = 0, frametype, ns, nr, q, d, m; in x25_process_rx_frame() local 427 queued = x25_state1_machine(sk, skb, frametype); in x25_process_rx_frame() 430 queued = x25_state2_machine(sk, skb, frametype); in x25_process_rx_frame() 433 queued = x25_state3_machine(sk, skb, frametype, ns, nr, q, d, m); in x25_process_rx_frame() 436 queued = x25_state4_machine(sk, skb, frametype); in x25_process_rx_frame() [all …]
|
/linux-6.12.1/net/rose/ |
D | rose_in.c | 105 int queued = 0; in rose_state3_machine() local 168 queued = 1; in rose_state3_machine() 205 return queued; in rose_state3_machine() 266 int queued = 0, frametype, ns, nr, q, d, m; in rose_process_rx_frame() local 275 queued = rose_state1_machine(sk, skb, frametype); in rose_process_rx_frame() 278 queued = rose_state2_machine(sk, skb, frametype); in rose_process_rx_frame() 281 queued = rose_state3_machine(sk, skb, frametype, ns, nr, q, d, m); in rose_process_rx_frame() 284 queued = rose_state4_machine(sk, skb, frametype); in rose_process_rx_frame() 287 queued = rose_state5_machine(sk, skb, frametype); in rose_process_rx_frame() 293 return queued; in rose_process_rx_frame()
|
/linux-6.12.1/tools/testing/selftests/net/ |
D | fq_band_pktlimit.sh | 7 # 2. send 20 pkts on band A: verify that 10 are queued, 10 dropped 8 # 3. send 20 pkts on band A: verify that 0 are queued, 20 dropped 9 # 4. send 20 pkts on band B: verify that 10 are queued, 10 dropped 12 # packets are still queued when later ones are sent. 44 # queued in FQ. Sleep for at least the delay period and see that
|
/linux-6.12.1/net/ax25/ |
D | ax25_std_in.c | 143 int queued = 0; in ax25_std_state3_machine() local 225 queued = ax25_rx_iframe(ax25, skb); in ax25_std_state3_machine() 258 return queued; in ax25_std_state3_machine() 268 int queued = 0; in ax25_std_state4_machine() local 380 queued = ax25_rx_iframe(ax25, skb); in ax25_std_state4_machine() 413 return queued; in ax25_std_state4_machine() 421 int queued = 0, frametype, ns, nr, pf; in ax25_std_frame_in() local 427 queued = ax25_std_state1_machine(ax25, skb, frametype, pf, type); in ax25_std_frame_in() 430 queued = ax25_std_state2_machine(ax25, skb, frametype, pf, type); in ax25_std_frame_in() 433 queued = ax25_std_state3_machine(ax25, skb, frametype, ns, nr, pf, type); in ax25_std_frame_in() [all …]
|
D | ax25_ds_in.c | 147 int queued = 0; in ax25_ds_state3_machine() local 240 queued = ax25_rx_iframe(ax25, skb); in ax25_ds_state3_machine() 273 return queued; in ax25_ds_state3_machine() 281 int queued = 0, frametype, ns, nr, pf; in ax25_ds_frame_in() local 287 queued = ax25_ds_state1_machine(ax25, skb, frametype, pf, type); in ax25_ds_frame_in() 290 queued = ax25_ds_state2_machine(ax25, skb, frametype, pf, type); in ax25_ds_frame_in() 293 queued = ax25_ds_state3_machine(ax25, skb, frametype, ns, nr, pf, type); in ax25_ds_frame_in() 297 return queued; in ax25_ds_frame_in()
|
/linux-6.12.1/include/drm/ |
D | drm_flip_work.h | 46 * @val: value queued via drm_flip_work_queue() 58 * @queued: queued tasks 60 * @lock: lock to access queued and commited lists 66 struct list_head queued; member
|
/linux-6.12.1/kernel/locking/ |
D | qrwlock.c | 3 * Queued read/write locks 18 * queued_read_lock_slowpath - acquire read lock of a queued rwlock 19 * @lock: Pointer to queued rwlock structure 63 * queued_write_lock_slowpath - acquire write lock of a queued rwlock 64 * @lock : Pointer to queued rwlock structure
|
D | qspinlock.c | 3 * Queued spinlock 28 * Include queued spinlock statistics code 41 * This queued spinlock implementation is based on the MCS lock, however to 145 * @lock: Pointer to queued spinlock structure 156 * @lock: Pointer to queued spinlock structure 169 * @lock : Pointer to queued spinlock structure 191 * @lock: Pointer to queued spinlock structure 202 * @lock: Pointer to queued spinlock structure 213 * @lock : Pointer to queued spinlock structure 241 * @lock : Pointer to queued spinlock structure [all …]
|
/linux-6.12.1/drivers/media/platform/renesas/vsp1/ |
D | vsp1_dl.c | 208 * @lock: protects the free, active, queued, and pending lists 211 * @queued: list queued to the hardware (written to the DL registers) 212 * @pending: list waiting to be queued to the hardware 224 struct vsp1_dl_list *queued; member 791 * loop on the same list until a new one is queued. In singleshot mode in vsp1_dl_list_fill_header() 841 if (!dlm->queued) in vsp1_dl_list_hw_update_pending() 871 * If a previous display list has been queued to the hardware but not in vsp1_dl_list_commit_continuous() 873 * case we can't replace the queued list by the new one, as we could in vsp1_dl_list_commit_continuous() 875 * be queued up to the hardware by the frame end interrupt handler. in vsp1_dl_list_commit_continuous() 894 * Pass the new display list to the hardware and mark it as queued. It in vsp1_dl_list_commit_continuous() [all …]
|
/linux-6.12.1/Documentation/userspace-api/media/v4l/ |
D | vidioc-streamon.rst | 51 If ``VIDIOC_STREAMON`` fails then any already queued buffers will remain 52 queued. 63 If buffers have been queued with :ref:`VIDIOC_QBUF` and 65 ``VIDIOC_STREAMON``, then those queued buffers will also be removed from 77 but ``VIDIOC_STREAMOFF`` will return queued buffers to their starting
|
D | dev-encoder.rst | 309 fast raw frames are queued on the ``OUTPUT`` queue. 468 * a buffer queued to ``OUTPUT`` may result in more than one buffer produced on 473 * a buffer queued to ``OUTPUT`` may result in a buffer being produced on 479 buffers queued to ``OUTPUT`` (e.g. during drain or ``EOS``), because of the 480 ``OUTPUT`` buffers queued in the past whose encoding results are only 483 * buffers queued to ``OUTPUT`` may not become available to dequeue instantly 549 sequence to avoid losing the already queued/encoded frames. 559 To ensure that all the queued ``OUTPUT`` buffers have been processed and the 562 received all encoded frames for all ``OUTPUT`` buffers queued before the 586 2. Any ``OUTPUT`` buffers queued by the client before the [all …]
|
/linux-6.12.1/net/dccp/ |
D | input.c | 45 int queued = 0; in dccp_rcv_close() local 76 queued = 1; in dccp_rcv_close() 86 return queued; in dccp_rcv_close() 91 int queued = 0; in dccp_rcv_closereq() local 101 return queued; in dccp_rcv_closereq() 113 queued = 1; in dccp_rcv_closereq() 120 return queued; in dccp_rcv_closereq() 524 int queued = 0; in dccp_rcv_respond_partopen_state_process() local 562 queued = 1; /* packet was queued in dccp_rcv_respond_partopen_state_process() 568 return queued; in dccp_rcv_respond_partopen_state_process() [all …]
|
/linux-6.12.1/net/netrom/ |
D | nr_in.c | 153 int queued = 0; in nr_state3_machine() local 225 queued = 1; in nr_state3_machine() 272 return queued; in nr_state3_machine() 279 int queued = 0, frametype; in nr_process_rx_frame() local 288 queued = nr_state1_machine(sk, skb, frametype); in nr_process_rx_frame() 291 queued = nr_state2_machine(sk, skb, frametype); in nr_process_rx_frame() 294 queued = nr_state3_machine(sk, skb, frametype); in nr_process_rx_frame() 300 return queued; in nr_process_rx_frame()
|
/linux-6.12.1/tools/testing/selftests/net/mptcp/ |
D | mptcp_inq.c | 208 int nsd, ret, queued = -1; in wait_for_ack() local 211 ret = ioctl(fd, TIOCOUTQ, &queued); in wait_for_ack() 219 if ((size_t)queued > total) in wait_for_ack() 220 xerror("TIOCOUTQ %u, but only %zu expected\n", queued, total); in wait_for_ack() 221 assert(nsd <= queued); in wait_for_ack() 223 if (queued == 0) in wait_for_ack() 232 xerror("still tx data queued after %u ms\n", timeout); in wait_for_ack() 356 unsigned int queued; in process_one_client() local 358 ret = ioctl(fd, FIONREAD, &queued); in process_one_client() 361 if (queued > expect_len) in process_one_client() [all …]
|
/linux-6.12.1/drivers/md/ |
D | dm-cache-background-tracker.c | 21 struct list_head queued; member 42 INIT_LIST_HEAD(&b->queued); in btracker_create() 55 list_for_each_entry_safe (w, tmp, &b->queued, list) { in btracker_destroy() 201 list_add(&w->list, &b->queued); in btracker_queue() 215 if (list_empty(&b->queued)) in btracker_issue() 218 w = list_first_entry(&b->queued, struct bt_work, list); in btracker_issue()
|