/linux-6.12.1/kernel/sched/ |
D | loadavg.c | 80 long nr_active, delta = 0; in calc_load_fold_active() local 82 nr_active = this_rq->nr_running - adjust; in calc_load_fold_active() 83 nr_active += (int)this_rq->nr_uninterruptible; in calc_load_fold_active() 85 if (nr_active != this_rq->calc_load_active) { in calc_load_fold_active() 86 delta = nr_active - this_rq->calc_load_active; in calc_load_fold_active() 87 this_rq->calc_load_active = nr_active; in calc_load_fold_active()
|
/linux-6.12.1/include/trace/events/ |
D | vmscan.h | 403 unsigned long nr_active, unsigned long nr_deactivated, 406 TP_ARGS(nid, nr_taken, nr_active, nr_deactivated, nr_referenced, priority, file), 411 __field(unsigned long, nr_active) 421 __entry->nr_active = nr_active; 431 __entry->nr_active, __entry->nr_deactivated, __entry->nr_referenced,
|
/linux-6.12.1/mm/ |
D | dmapool.c | 53 size_t nr_active; member 83 pool->name, pool->nr_active, in pools_show() 187 pool->nr_active++; in pool_block_pop() 377 if (pool->nr_active) { in dma_pool_destroy() 458 pool->nr_active--; in dma_pool_free()
|
/linux-6.12.1/arch/powerpc/platforms/cell/spufs/ |
D | sched.c | 680 cbe_spu_info[node].nr_active--; in find_victim() 709 cbe_spu_info[node].nr_active++; in __spu_schedule() 750 cbe_spu_info[node].nr_active--; in spu_unschedule() 951 int nr_active = 0, node; in count_active_contexts() local 954 nr_active += cbe_spu_info[node].nr_active; in count_active_contexts() 955 nr_active += spu_prio->nr_waiting; in count_active_contexts() 957 return nr_active; in count_active_contexts()
|
/linux-6.12.1/tools/sched_ext/ |
D | scx_flatcg.bpf.c | 201 if (!cgc->nr_active) { in cgrp_refresh_hweight() 238 is_active = cgc->nr_active; in cgrp_refresh_hweight() 480 if (!cgc->nr_active++) { in update_active_weight_sums() 488 if (!--cgc->nr_active) { in update_active_weight_sums() 604 if (pcgc && cgc->nr_active) in BPF_STRUCT_OPS()
|
D | scx_flatcg.h | 40 u32 nr_active; member
|
/linux-6.12.1/block/ |
D | blk-mq.h | 282 atomic_add(val, &hctx->nr_active); in __blk_mq_add_active_requests() 296 atomic_sub(val, &hctx->nr_active); in __blk_mq_sub_active_requests() 334 return atomic_read(&hctx->nr_active); in __blk_mq_active_requests()
|
D | blk-mq.c | 3759 atomic_set(&hctx->nr_active, 0); in blk_mq_alloc_hctx()
|
/linux-6.12.1/tools/testing/selftests/bpf/prog_tests/ |
D | setget_sockopt.c | 64 ASSERT_EQ(bss->nr_active, 1, "nr_active"); in test_tcp() 144 ASSERT_EQ(bss->nr_active, 1, "nr_active"); in test_ktls()
|
/linux-6.12.1/fs/ceph/ |
D | mdsmap.c | 426 int i, nr_active = 0; in ceph_mdsmap_is_cluster_available() local 435 nr_active++; in ceph_mdsmap_is_cluster_available() 437 return nr_active > 0; in ceph_mdsmap_is_cluster_available()
|
/linux-6.12.1/tools/testing/selftests/bpf/progs/ |
D | setget_sockopt.c | 18 int nr_active; variable 401 nr_active += !(bpf_test_sockopt(skops, sk) || in skops_sockopt()
|
/linux-6.12.1/drivers/nvme/host/ |
D | multipath.c | 135 atomic_inc(&ns->ctrl->nr_active); in nvme_mpath_start_request() 153 atomic_dec_if_positive(&ns->ctrl->nr_active); in nvme_mpath_end_request() 363 depth = atomic_read(&ns->ctrl->nr_active); in nvme_queue_depth_path() 1042 atomic_set(&ctrl->nr_active, 0); in nvme_mpath_init_identify()
|
D | nvme.h | 369 atomic_t nr_active; member
|
/linux-6.12.1/kernel/ |
D | workqueue.c | 282 int nr_active; /* L: nr of active works */ member 1670 return !pwq->nr_active && list_empty(&pwq->inactive_works); in pwq_is_empty() 1721 obtained = pwq->nr_active < READ_ONCE(wq->max_active); in pwq_tryinc_nr_active() 1773 pwq->nr_active++; in pwq_tryinc_nr_active() 1902 pwq->nr_active++; in node_activate_pending_pwq() 1941 pwq->nr_active--; in pwq_dec_nr_active() 3498 if (pwq->nr_active && need_to_create_worker(pool)) { in rescuer_thread() 6214 pwq->nr_active, pwq->refcnt, in show_pwq() 6821 WARN_ON_ONCE(pwq->nr_active < 0); in freeze_workqueues_busy() 6822 if (pwq->nr_active) { in freeze_workqueues_busy()
|
/linux-6.12.1/arch/powerpc/include/asm/ |
D | spu.h | 183 int nr_active; member
|
/linux-6.12.1/include/linux/ |
D | blk-mq.h | 429 atomic_t nr_active; member
|
/linux-6.12.1/net/mptcp/ |
D | protocol.c | 1431 int i, nr_active = 0; in mptcp_subflow_get_send() local 1451 nr_active += !backup; in mptcp_subflow_get_send() 1470 if (!nr_active) in mptcp_subflow_get_send()
|