/linux-6.12.1/fs/bcachefs/ |
D | rcu_pending.c | 176 static noinline void __process_finished_items(struct rcu_pending *pending, in __process_finished_items() argument 185 __poll_state_synchronize_rcu(pending->srcu, p->objs.data[0].seq)) { in __process_finished_items() 197 switch ((ulong) pending->process) { in __process_finished_items() 251 pending->process(pending, *genradix_ptr(&objs.objs, i)); in __process_finished_items() 261 pending->process(pending, obj); in __process_finished_items() 267 static bool process_finished_items(struct rcu_pending *pending, in process_finished_items() argument 276 if ((p->objs.nr && __poll_state_synchronize_rcu(pending->srcu, p->objs.data[0].seq)) || in process_finished_items() 277 (p->lists[0].head && __poll_state_synchronize_rcu(pending->srcu, p->lists[0].seq)) || in process_finished_items() 278 (p->lists[1].head && __poll_state_synchronize_rcu(pending->srcu, p->lists[1].seq)) || in process_finished_items() 280 __process_finished_items(pending, p, flags); in process_finished_items() [all …]
|
D | rcu_pending.h | 18 void rcu_pending_enqueue(struct rcu_pending *pending, struct rcu_head *obj); 19 struct rcu_head *rcu_pending_dequeue(struct rcu_pending *pending); 20 struct rcu_head *rcu_pending_dequeue_from_all(struct rcu_pending *pending); 22 void rcu_pending_exit(struct rcu_pending *pending); 23 int rcu_pending_init(struct rcu_pending *pending,
|
/linux-6.12.1/tools/lib/ |
D | list_sort.c | 176 struct list_head *list = head->next, *pending = NULL; in list_sort() local 205 struct list_head **tail = &pending; in list_sort() 221 list->prev = pending; in list_sort() 222 pending = list; in list_sort() 224 pending->next = NULL; in list_sort() 229 list = pending; in list_sort() 230 pending = pending->prev; in list_sort() 232 struct list_head *next = pending->prev; in list_sort() 236 list = merge(priv, cmp, pending, list); in list_sort() 237 pending = next; in list_sort() [all …]
|
/linux-6.12.1/lib/ |
D | list_sort.c | 187 struct list_head *list = head->next, *pending = NULL; in list_sort() local 216 struct list_head **tail = &pending; in list_sort() 232 list->prev = pending; in list_sort() 233 pending = list; in list_sort() 235 pending->next = NULL; in list_sort() 240 list = pending; in list_sort() 241 pending = pending->prev; in list_sort() 243 struct list_head *next = pending->prev; in list_sort() 247 list = merge(priv, cmp, pending, list); in list_sort() 248 pending = next; in list_sort() [all …]
|
/linux-6.12.1/arch/mips/sni/ |
D | pcit.c | 183 u32 pending = *(volatile u32 *)SNI_PCIT_INT_REG; in pcit_hwint1() local 187 irq = ffs((pending >> 16) & 0x7f); in pcit_hwint1() 196 u32 pending = *(volatile u32 *)SNI_PCIT_INT_REG; in pcit_hwint0() local 200 irq = ffs((pending >> 16) & 0x3f); in pcit_hwint0() 209 u32 pending = read_c0_cause() & read_c0_status(); in sni_pcit_hwint() local 211 if (pending & C_IRQ1) in sni_pcit_hwint() 213 else if (pending & C_IRQ2) in sni_pcit_hwint() 215 else if (pending & C_IRQ3) in sni_pcit_hwint() 217 else if (pending & C_IRQ5) in sni_pcit_hwint() 223 u32 pending = read_c0_cause() & read_c0_status(); in sni_pcit_hwint_cplus() local [all …]
|
/linux-6.12.1/drivers/gpu/drm/mediatek/ |
D | mtk_plane.c | 48 state->pending.format = DRM_FORMAT_RGB565; in mtk_plane_reset() 49 state->pending.modifier = DRM_FORMAT_MOD_LINEAR; in mtk_plane_reset() 65 state->pending = old_state->pending; in mtk_plane_duplicate_state() 200 mtk_plane_state->pending.enable = true; in mtk_plane_update_new_state() 201 mtk_plane_state->pending.pitch = pitch; in mtk_plane_update_new_state() 202 mtk_plane_state->pending.hdr_pitch = hdr_pitch; in mtk_plane_update_new_state() 203 mtk_plane_state->pending.format = format; in mtk_plane_update_new_state() 204 mtk_plane_state->pending.modifier = modifier; in mtk_plane_update_new_state() 205 mtk_plane_state->pending.addr = addr; in mtk_plane_update_new_state() 206 mtk_plane_state->pending.hdr_addr = hdr_addr; in mtk_plane_update_new_state() [all …]
|
D | mtk_disp_ovl_adaptor.c | 137 struct mtk_plane_pending_state *pending = &state->pending; in mtk_ovl_adaptor_layer_config() local 143 const struct drm_format_info *fmt_info = drm_format_info(pending->format); in mtk_ovl_adaptor_layer_config() 150 pending->enable, pending->format); in mtk_ovl_adaptor_layer_config() 152 &pending->addr, (pending->pitch / fmt_info->cpp[0]), in mtk_ovl_adaptor_layer_config() 153 pending->x, pending->y, pending->width, pending->height); in mtk_ovl_adaptor_layer_config() 160 if (!pending->enable || !pending->width || !pending->height) { in mtk_ovl_adaptor_layer_config() 169 align_width = ALIGN_DOWN(pending->width, 2); in mtk_ovl_adaptor_layer_config() 175 l_w = (align_width / 2) + ((pending->width / 2) % 2); in mtk_ovl_adaptor_layer_config() 180 mtk_merge_advance_config(merge, l_w, r_w, pending->height, 0, 0, cmdq_pkt); in mtk_ovl_adaptor_layer_config() 182 pending->height, cmdq_pkt); in mtk_ovl_adaptor_layer_config() [all …]
|
/linux-6.12.1/arch/mips/cobalt/ |
D | irq.c | 25 unsigned pending = read_c0_status() & read_c0_cause() & ST0_IM; in plat_irq_dispatch() local 28 if (pending & CAUSEF_IP2) in plat_irq_dispatch() 30 else if (pending & CAUSEF_IP6) { in plat_irq_dispatch() 36 } else if (pending & CAUSEF_IP3) in plat_irq_dispatch() 38 else if (pending & CAUSEF_IP4) in plat_irq_dispatch() 40 else if (pending & CAUSEF_IP5) in plat_irq_dispatch() 42 else if (pending & CAUSEF_IP7) in plat_irq_dispatch()
|
/linux-6.12.1/drivers/platform/surface/aggregator/ |
D | ssh_request_layer.c | 132 spin_lock(&rtl->pending.lock); in ssh_rtl_pending_remove() 135 spin_unlock(&rtl->pending.lock); in ssh_rtl_pending_remove() 139 atomic_dec(&rtl->pending.count); in ssh_rtl_pending_remove() 142 spin_unlock(&rtl->pending.lock); in ssh_rtl_pending_remove() 151 spin_lock(&rtl->pending.lock); in ssh_rtl_tx_pending_push() 154 spin_unlock(&rtl->pending.lock); in ssh_rtl_tx_pending_push() 159 spin_unlock(&rtl->pending.lock); in ssh_rtl_tx_pending_push() 163 atomic_inc(&rtl->pending.count); in ssh_rtl_tx_pending_push() 164 list_add_tail(&ssh_request_get(rqst)->node, &rtl->pending.head); in ssh_rtl_tx_pending_push() 166 spin_unlock(&rtl->pending.lock); in ssh_rtl_tx_pending_push() [all …]
|
D | ssh_packet_layer.c | 804 spin_lock(&ptl->pending.lock); in ssh_ptl_pending_push() 808 spin_unlock(&ptl->pending.lock); in ssh_ptl_pending_push() 821 atomic_inc(&ptl->pending.count); in ssh_ptl_pending_push() 822 list_add_tail(&ssh_packet_get(p)->pending_node, &ptl->pending.head); in ssh_ptl_pending_push() 825 spin_unlock(&ptl->pending.lock); in ssh_ptl_pending_push() 835 spin_lock(&ptl->pending.lock); in ssh_ptl_pending_remove() 838 spin_unlock(&ptl->pending.lock); in ssh_ptl_pending_remove() 843 atomic_dec(&ptl->pending.count); in ssh_ptl_pending_remove() 845 spin_unlock(&ptl->pending.lock); in ssh_ptl_pending_remove() 888 return !atomic_read(&ptl->pending.count); in ssh_ptl_tx_can_process() [all …]
|
/linux-6.12.1/arch/loongarch/kvm/ |
D | interrupt.c | 83 unsigned long *pending = &vcpu->arch.irq_pending; in kvm_deliver_intr() local 86 if (!(*pending) && !(*pending_clr)) in kvm_deliver_intr() 99 if (*pending) { in kvm_deliver_intr() 100 priority = __ffs(*pending); in kvm_deliver_intr() 103 priority = find_next_bit(pending, in kvm_deliver_intr() 104 BITS_PER_BYTE * sizeof(*pending), in kvm_deliver_intr() 175 unsigned long *pending = &vcpu->arch.exception_pending; in kvm_deliver_exception() local 177 if (*pending) { in kvm_deliver_exception() 178 code = __ffs(*pending); in kvm_deliver_exception() 180 *pending = 0; in kvm_deliver_exception()
|
/linux-6.12.1/arch/mips/ralink/ |
D | irq.c | 99 u32 pending = rt_intc_r32(INTC_REG_STATUS0); in ralink_intc_irq_handler() local 101 if (pending) { in ralink_intc_irq_handler() 103 generic_handle_domain_irq(domain, __ffs(pending)); in ralink_intc_irq_handler() 111 unsigned long pending; in plat_irq_dispatch() local 113 pending = read_c0_status() & read_c0_cause() & ST0_IM; in plat_irq_dispatch() 115 if (pending & STATUSF_IP7) in plat_irq_dispatch() 118 else if (pending & STATUSF_IP5) in plat_irq_dispatch() 121 else if (pending & STATUSF_IP6) in plat_irq_dispatch() 124 else if (pending & STATUSF_IP4) in plat_irq_dispatch() 127 else if (pending & STATUSF_IP2) in plat_irq_dispatch()
|
/linux-6.12.1/drivers/irqchip/ |
D | irq-ath79-cpu.c | 36 unsigned long pending; in plat_irq_dispatch() local 39 pending = read_c0_status() & read_c0_cause() & ST0_IM; in plat_irq_dispatch() 41 if (!pending) { in plat_irq_dispatch() 46 pending >>= CAUSEB_IP; in plat_irq_dispatch() 47 while (pending) { in plat_irq_dispatch() 48 irq = fls(pending) - 1; in plat_irq_dispatch() 52 pending &= ~BIT(irq); in plat_irq_dispatch()
|
D | irq-loongson-htpic.c | 33 uint32_t pending; in htpic_irq_dispatch() local 36 pending = readl(priv->base); in htpic_irq_dispatch() 38 writel(pending, priv->base); in htpic_irq_dispatch() 40 if (!pending) in htpic_irq_dispatch() 43 while (pending) { in htpic_irq_dispatch() 44 int bit = __ffs(pending); in htpic_irq_dispatch() 52 pending &= ~BIT(bit); in htpic_irq_dispatch()
|
/linux-6.12.1/drivers/gpu/drm/qxl/ |
D | qxl_irq.c | 36 uint32_t pending; in qxl_irq_handler() local 38 pending = xchg(&qdev->ram_header->int_pending, 0); in qxl_irq_handler() 40 if (!pending) in qxl_irq_handler() 45 if (pending & QXL_INTERRUPT_DISPLAY) { in qxl_irq_handler() 50 if (pending & QXL_INTERRUPT_CURSOR) { in qxl_irq_handler() 54 if (pending & QXL_INTERRUPT_IO_CMD) { in qxl_irq_handler() 58 if (pending & QXL_INTERRUPT_ERROR) { in qxl_irq_handler() 66 if (pending & QXL_INTERRUPT_CLIENT_MONITORS_CONFIG) { in qxl_irq_handler()
|
/linux-6.12.1/drivers/gpu/drm/i915/ |
D | i915_sw_fence.c | 149 atomic_set_release(&fence->pending, -1); /* 0 -> -1 [done] */ in __i915_sw_fence_wake_up_all() 196 if (!atomic_dec_and_test(&fence->pending)) in __i915_sw_fence_complete() 224 int pending; in i915_sw_fence_await() local 230 pending = atomic_read(&fence->pending); in i915_sw_fence_await() 232 if (pending < 1) in i915_sw_fence_await() 234 } while (!atomic_try_cmpxchg(&fence->pending, &pending, pending + 1)); in i915_sw_fence_await() 257 atomic_set(&fence->pending, 1); in i915_sw_fence_reinit() 344 unsigned int pending; in __i915_sw_fence_await_sw_fence() local 361 pending = I915_SW_FENCE_FLAG_FENCE; in __i915_sw_fence_await_sw_fence() 373 pending |= I915_SW_FENCE_FLAG_ALLOC; in __i915_sw_fence_await_sw_fence() [all …]
|
/linux-6.12.1/arch/mips/sibyte/sb1250/ |
D | irq.c | 131 u64 pending; in ack_sb1250_irq() local 139 pending = __raw_readq(IOADDR(A_IMR_REGISTER(sb1250_irq_owner[irq], in ack_sb1250_irq() 141 pending &= ((u64)1 << (irq)); in ack_sb1250_irq() 142 if (pending) { in ack_sb1250_irq() 155 __raw_writeq(pending, in ack_sb1250_irq() 290 unsigned int pending; in plat_irq_dispatch() local 302 pending = read_c0_cause() & read_c0_status() & ST0_IM; in plat_irq_dispatch() 304 if (pending & CAUSEF_IP7) /* CPU performance counter interrupt */ in plat_irq_dispatch() 306 else if (pending & CAUSEF_IP4) in plat_irq_dispatch() 310 else if (pending & CAUSEF_IP3) in plat_irq_dispatch() [all …]
|
/linux-6.12.1/tools/testing/selftests/bpf/progs/ |
D | cgroup_hierarchical_stats.c | 21 __u64 pending; member 56 static int create_attach_counter(__u64 cg_id, __u64 state, __u64 pending) in create_attach_counter() argument 58 struct attach_counter init = {.state = state, .pending = pending}; in create_attach_counter() 109 if (total_counter->pending) { in BPF_PROG() 110 delta += total_counter->pending; in BPF_PROG() 111 total_counter->pending = 0; in BPF_PROG() 126 parent_counter->pending += delta; in BPF_PROG()
|
/linux-6.12.1/arch/mips/loongson2ef/fuloong-2e/ |
D | irq.c | 24 asmlinkage void mach_irq_dispatch(unsigned int pending) in mach_irq_dispatch() argument 26 if (pending & CAUSEF_IP7) in mach_irq_dispatch() 28 else if (pending & CAUSEF_IP6) /* perf counter loverflow */ in mach_irq_dispatch() 30 else if (pending & CAUSEF_IP5) in mach_irq_dispatch() 32 else if (pending & CAUSEF_IP2) in mach_irq_dispatch()
|
/linux-6.12.1/net/vmw_vsock/ |
D | vmci_transport.c | 43 struct sock *pending, 484 struct sock *pending; in vmci_transport_get_pending() local 495 pending = sk_vsock(vpending); in vmci_transport_get_pending() 496 sock_hold(pending); in vmci_transport_get_pending() 501 pending = NULL; in vmci_transport_get_pending() 503 return pending; in vmci_transport_get_pending() 507 static void vmci_transport_release_pending(struct sock *pending) in vmci_transport_release_pending() argument 509 sock_put(pending); in vmci_transport_release_pending() 943 struct sock *pending; in vmci_transport_recv_listen() local 957 pending = vmci_transport_get_pending(sk, pkt); in vmci_transport_recv_listen() [all …]
|
/linux-6.12.1/arch/mips/kvm/ |
D | interrupt.c | 26 unsigned long *pending = &vcpu->arch.pending_exceptions; in kvm_mips_deliver_interrupts() local 30 if (!(*pending) && !(*pending_clr)) in kvm_mips_deliver_interrupts() 42 priority = __ffs(*pending); in kvm_mips_deliver_interrupts() 46 priority = find_next_bit(pending, in kvm_mips_deliver_interrupts() 47 BITS_PER_BYTE * sizeof(*pending), in kvm_mips_deliver_interrupts()
|
/linux-6.12.1/arch/mips/sibyte/bcm1480/ |
D | irq.c | 143 u64 pending; in ack_bcm1480_irq() local 157 pending = __raw_readq(IOADDR(A_BCM1480_IMR_REGISTER(bcm1480_irq_owner[irq], in ack_bcm1480_irq() 159 pending &= ((u64)1 << (irq_dirty)); in ack_bcm1480_irq() 160 if (pending) { in ack_bcm1480_irq() 168 __raw_writeq(pending, IOADDR(A_BCM1480_IMR_REGISTER(cpu_logical_map(i), in ack_bcm1480_irq() 172 …__raw_writeq(pending, IOADDR(A_BCM1480_IMR_REGISTER(0, R_BCM1480_IMR_LDT_INTERRUPT_CLR_H + (k*BCM1… in ack_bcm1480_irq() 335 unsigned int pending; in plat_irq_dispatch() local 337 pending = read_c0_cause() & read_c0_status(); in plat_irq_dispatch() 339 if (pending & CAUSEF_IP4) in plat_irq_dispatch() 342 else if (pending & CAUSEF_IP3) in plat_irq_dispatch() [all …]
|
/linux-6.12.1/kernel/irq/ |
D | irq_sim.c | 17 unsigned long *pending; member 62 *state = test_bit(hwirq, irq_ctx->work_ctx->pending); in irq_sim_get_irqchip_state() 80 assign_bit(hwirq, irq_ctx->work_ctx->pending, state); in irq_sim_set_irqchip_state() 136 while (!bitmap_empty(work_ctx->pending, work_ctx->irq_count)) { in irq_sim_handle_irq() 137 offset = find_next_bit(work_ctx->pending, in irq_sim_handle_irq() 139 clear_bit(offset, work_ctx->pending); in irq_sim_handle_irq() 210 unsigned long *pending __free(bitmap) = bitmap_zalloc(num_irqs, GFP_KERNEL); in irq_domain_create_sim_full() 211 if (!pending) in irq_domain_create_sim_full() 222 work_ctx->pending = no_free_ptr(pending); in irq_domain_create_sim_full() 243 bitmap_free(work_ctx->pending); in irq_domain_remove_sim()
|
/linux-6.12.1/drivers/hid/ |
D | hid-sensor-hub.c | 308 memset(&hsdev->pending, 0, sizeof(hsdev->pending)); in sensor_hub_input_attr_get_raw_value() 309 init_completion(&hsdev->pending.ready); in sensor_hub_input_attr_get_raw_value() 310 hsdev->pending.usage_id = usage_id; in sensor_hub_input_attr_get_raw_value() 311 hsdev->pending.attr_usage_id = attr_usage_id; in sensor_hub_input_attr_get_raw_value() 312 hsdev->pending.raw_size = 0; in sensor_hub_input_attr_get_raw_value() 315 hsdev->pending.status = true; in sensor_hub_input_attr_get_raw_value() 323 &hsdev->pending.ready, HZ*5); in sensor_hub_input_attr_get_raw_value() 324 switch (hsdev->pending.raw_size) { in sensor_hub_input_attr_get_raw_value() 327 ret_val = *(s8 *)hsdev->pending.raw_data; in sensor_hub_input_attr_get_raw_value() 329 ret_val = *(u8 *)hsdev->pending.raw_data; in sensor_hub_input_attr_get_raw_value() [all …]
|
/linux-6.12.1/fs/btrfs/ |
D | async-thread.c | 36 atomic_t pending; member 71 return atomic_read(&wq->pending) > wq->thresh * 2; in btrfs_workqueue_normal_congested() 78 atomic_set(&wq->pending, 0); in btrfs_init_workqueue() 159 atomic_inc(&wq->pending); in thresh_queue_hook() 170 long pending; in thresh_exec_hook() local 176 atomic_dec(&wq->pending); in thresh_exec_hook() 192 pending = atomic_read(&wq->pending); in thresh_exec_hook() 193 if (pending > wq->thresh) in thresh_exec_hook() 195 if (pending < wq->thresh / 2) in thresh_exec_hook()
|