/linux-6.12.1/net/ieee802154/6lowpan/ |
D | rx.c | 54 static inline bool lowpan_is_frag1(u8 dispatch) in lowpan_is_frag1() argument 56 return (dispatch & LOWPAN_DISPATCH_FRAG_MASK) == LOWPAN_DISPATCH_FRAG1; in lowpan_is_frag1() 59 static inline bool lowpan_is_fragn(u8 dispatch) in lowpan_is_fragn() argument 61 return (dispatch & LOWPAN_DISPATCH_FRAG_MASK) == LOWPAN_DISPATCH_FRAGN; in lowpan_is_fragn() 122 static inline bool lowpan_is_esc(u8 dispatch) in lowpan_is_esc() argument 124 return dispatch == LOWPAN_DISPATCH_ESC; in lowpan_is_esc() 138 static inline bool lowpan_is_hc1(u8 dispatch) in lowpan_is_hc1() argument 140 return dispatch == LOWPAN_DISPATCH_HC1; in lowpan_is_hc1() 154 static inline bool lowpan_is_dff(u8 dispatch) in lowpan_is_dff() argument 156 return dispatch == LOWPAN_DISPATCH_DFF; in lowpan_is_dff() [all …]
|
/linux-6.12.1/drivers/net/ethernet/cavium/liquidio/ |
D | octeon_device.c | 714 oct->dispatch.dlist = (struct octeon_dispatch *) in octeon_allocate_device_mem() 1027 oct->dispatch.count = 0; in octeon_init_dispatch_list() 1030 oct->dispatch.dlist[i].opcode = 0; in octeon_init_dispatch_list() 1031 INIT_LIST_HEAD(&oct->dispatch.dlist[i].list); in octeon_init_dispatch_list() 1037 spin_lock_init(&oct->dispatch.lock); in octeon_init_dispatch_list() 1050 spin_lock_bh(&oct->dispatch.lock); in octeon_delete_dispatch_list() 1053 struct list_head *dispatch; in octeon_delete_dispatch_list() local 1055 dispatch = &oct->dispatch.dlist[i].list; in octeon_delete_dispatch_list() 1056 while (dispatch->next != dispatch) { in octeon_delete_dispatch_list() 1057 temp = dispatch->next; in octeon_delete_dispatch_list() [all …]
|
D | octeon_droq.c | 54 struct list_head *dispatch; in octeon_get_dispatch_arg() local 60 spin_lock_bh(&octeon_dev->dispatch.lock); in octeon_get_dispatch_arg() 62 if (octeon_dev->dispatch.count == 0) { in octeon_get_dispatch_arg() 63 spin_unlock_bh(&octeon_dev->dispatch.lock); in octeon_get_dispatch_arg() 67 if (octeon_dev->dispatch.dlist[idx].opcode == combined_opcode) { in octeon_get_dispatch_arg() 68 fn_arg = octeon_dev->dispatch.dlist[idx].arg; in octeon_get_dispatch_arg() 70 list_for_each(dispatch, in octeon_get_dispatch_arg() 71 &octeon_dev->dispatch.dlist[idx].list) { in octeon_get_dispatch_arg() 72 if (((struct octeon_dispatch *)dispatch)->opcode == in octeon_get_dispatch_arg() 75 dispatch)->arg; in octeon_get_dispatch_arg() [all …]
|
/linux-6.12.1/drivers/scsi/sym53c8xx_2/ |
D | sym_fw1.h | 62 u32 dispatch [ 28]; member 417 PADDR_A (dispatch), 434 PADDR_A (dispatch), 443 PADDR_A (dispatch), 464 PADDR_A (dispatch), 483 PADDR_A (dispatch), 499 PADDR_A (dispatch), 524 PADDR_A (dispatch), 543 PADDR_A (dispatch), 557 PADDR_A (dispatch), [all …]
|
D | sym_fw2.h | 63 u32 dispatch [ 28]; member 402 PADDR_A (dispatch), 419 PADDR_A (dispatch), 428 PADDR_A (dispatch), 449 PADDR_A (dispatch), 467 PADDR_A (dispatch), 483 PADDR_A (dispatch), 508 PADDR_A (dispatch), 526 PADDR_A (dispatch), 540 PADDR_A (dispatch), [all …]
|
/linux-6.12.1/tools/testing/selftests/sched_ext/ |
D | select_cpu_dispatch.bpf.c | 22 goto dispatch; in BPF_STRUCT_OPS() 26 goto dispatch; in BPF_STRUCT_OPS() 31 dispatch: in BPF_STRUCT_OPS()
|
D | maybe_null_fail_dsp.bpf.c | 22 .dispatch = (void *) maybe_null_fail_dispatch,
|
D | maybe_null.bpf.c | 32 .dispatch = (void *) maybe_null_success_dispatch,
|
D | dsp_local_on.bpf.c | 61 .dispatch = (void *) dsp_local_on_dispatch,
|
D | exit.bpf.c | 79 .dispatch = (void *) exit_dispatch,
|
/linux-6.12.1/include/net/ |
D | 6lowpan.h | 88 static inline bool lowpan_is_ipv6(u8 dispatch) in lowpan_is_ipv6() argument 90 return dispatch == LOWPAN_DISPATCH_IPV6; in lowpan_is_ipv6() 93 static inline bool lowpan_is_iphc(u8 dispatch) in lowpan_is_iphc() argument 95 return (dispatch & LOWPAN_DISPATCH_IPHC_MASK) == LOWPAN_DISPATCH_IPHC; in lowpan_is_iphc()
|
/linux-6.12.1/block/ |
D | blk-mq-sched.c | 67 goto dispatch; in blk_mq_dispatch_hctx_list() 73 dispatch: in blk_mq_dispatch_hctx_list() 109 if (!list_empty_careful(&hctx->dispatch)) { in __blk_mq_do_dispatch_sched() 226 if (!list_empty_careful(&hctx->dispatch)) { in blk_mq_do_dispatch_ctx() 279 if (!list_empty_careful(&hctx->dispatch)) { in __blk_mq_sched_dispatch_requests() 281 if (!list_empty(&hctx->dispatch)) in __blk_mq_sched_dispatch_requests() 282 list_splice_init(&hctx->dispatch, &rq_list); in __blk_mq_sched_dispatch_requests()
|
D | mq-deadline.c | 74 struct list_head dispatch; member 328 if (!list_empty(&per_prio->dispatch)) { in __dd_dispatch_request() 329 rq = list_first_entry(&per_prio->dispatch, struct request, in __dd_dispatch_request() 591 INIT_LIST_HEAD(&per_prio->dispatch); in dd_init_sched() 699 list_add(&rq->queuelist, &per_prio->dispatch); in dd_insert_request() 773 return !list_empty_careful(&per_prio->dispatch) || in dd_has_work_for_prio() 1001 return seq_list_start(&per_prio->dispatch, *pos); \ 1011 return seq_list_next(v, &per_prio->dispatch, pos); \
|
/linux-6.12.1/drivers/acpi/acpica/ |
D | evgpeutil.c | 300 ACPI_FREE(gpe_event_info->dispatch.handler); in acpi_ev_delete_gpe_handlers() 301 gpe_event_info->dispatch.handler = NULL; in acpi_ev_delete_gpe_handlers() 309 notify = gpe_event_info->dispatch.notify_list; in acpi_ev_delete_gpe_handlers() 316 gpe_event_info->dispatch.notify_list = NULL; in acpi_ev_delete_gpe_handlers()
|
D | evgpe.c | 479 notify = gpe_event_info->dispatch.notify_list; in acpi_ev_asynch_execute_gpe_method() 503 gpe_event_info->dispatch.method_node; in acpi_ev_asynch_execute_gpe_method() 514 dispatch. in acpi_ev_asynch_execute_gpe_method() 703 gpe_handler_info = gpe_event_info->dispatch.handler; in acpi_ev_detect_gpe() 804 gpe_event_info->dispatch.handler->address(gpe_device, in acpi_ev_gpe_dispatch() 807 dispatch.handler-> in acpi_ev_gpe_dispatch()
|
D | evxface.c | 770 handler->method_node = gpe_event_info->dispatch.method_node; in ACPI_EXPORT_SYMBOL() 798 gpe_event_info->dispatch.handler = handler; in ACPI_EXPORT_SYMBOL() 948 if (gpe_event_info->dispatch.handler->address != address) { in ACPI_EXPORT_SYMBOL() 955 handler = gpe_event_info->dispatch.handler; in ACPI_EXPORT_SYMBOL() 956 gpe_event_info->dispatch.handler = NULL; in ACPI_EXPORT_SYMBOL() 960 gpe_event_info->dispatch.method_node = handler->method_node; in ACPI_EXPORT_SYMBOL()
|
/linux-6.12.1/Documentation/arch/powerpc/ |
D | vcpudispatch_stats.rst | 9 chips (representing the "home" node) and tries to always dispatch vcpus 15 related to the vcpu dispatch behavior. Writing '1' to this file enables 29 The next 4 numbers represent vcpu dispatch dispersions: 71 its last dispatch.
|
/linux-6.12.1/Documentation/scheduler/ |
D | sched-ext.rst | 133 * then dispatch the task directly to SCX_DSQ_LOCAL and skip the 157 * Do a direct dispatch of a task to the global DSQ. This ops.enqueue() 158 * callback will only be invoked if we failed to find a core to dispatch 199 sched_ext uses DSQs (dispatch queues) which can operate as both a FIFO and a 211 global DSQ. If that doesn't yield a runnable task either, ``ops.dispatch()`` 246 * Immediately dispatch the task to either the global or local DSQ by 250 * Immediately dispatch the task to a custom DSQ by calling 257 run, ``ops.dispatch()`` is invoked which can use the following two 273 4. After ``ops.dispatch()`` returns, if there are tasks in the local DSQ, 278 * If ``ops.dispatch()`` has dispatched any tasks, retry #3. [all …]
|
/linux-6.12.1/Documentation/arch/x86/x86_64/ |
D | fred.rst | 44 vector, FRED requires the software to dispatch an event to its handler 45 based on both the event's type and vector. Therefore, an event dispatch 47 dispatch process. The FRED event dispatch framework takes control 48 once an event is delivered, and employs a two-level dispatch.
|
/linux-6.12.1/Documentation/block/ |
D | blk-mq.rst | 49 blk-mq has two group of queues: software staging queues and hardware dispatch 102 Hardware dispatch queues 110 dispatch to the hardware. 113 added to a linked list (``hctx->dispatch``) of requests. Then, 115 ``dispatch`` list first, to ensure a fairness dispatch with those 132 identified by an integer, ranging from 0 to the dispatch queue size. This tag
|
D | deadline-iosched.rst | 50 device dispatch queue, we always give a preference to reads. However, we 53 done writes_starved number of times, we dispatch some writes based on the
|
/linux-6.12.1/tools/testing/selftests/hid/tests/ |
D | base.py | 267 self.uhdev.dispatch(1) 281 self.uhdev.dispatch(1) 319 self.uhdev.dispatch(1)
|
D | test_sony.py | 233 while self.uhdev.dispatch(10): 245 self.uhdev.dispatch(10) 249 self.uhdev.dispatch(10)
|
/linux-6.12.1/arch/xtensa/kernel/ |
D | vectors.S | 74 xsr a3, excsave1 # save a3 and get dispatch table 82 xsr a3, excsave1 # restore a3 and dispatch table 101 xsr a3, excsave1 # save a3, and get dispatch table 109 xsr a3, excsave1 # restore a3 and dispatch table
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/engine/ce/fuc/ |
D | com.fuc | 174 call #dispatch 284 dispatch: 293 // lookup method in the dispatch table, ILLEGAL_MTHD if not found 321 // depending on dispatch flags: execute method, or save data as state 364 // $r4: dispatch table entry 379 // $r4: dispatch table entry 399 // $r4: dispatch table entry 777 // $r4: dispatch table entry 851 // $r4: dispatch table entry
|