/linux-6.12.1/tools/bpf/bpftool/ |
D | pids.c | 34 struct obj_refs *refs; in add_ref() local 40 refs = entry->pvalue; in add_ref() 42 for (i = 0; i < refs->ref_cnt; i++) { in add_ref() 43 if (refs->refs[i].pid == e->pid) in add_ref() 47 tmp = realloc(refs->refs, (refs->ref_cnt + 1) * sizeof(*ref)); in add_ref() 53 refs->refs = tmp; in add_ref() 54 ref = &refs->refs[refs->ref_cnt]; in add_ref() 57 refs->ref_cnt++; in add_ref() 63 refs = calloc(1, sizeof(*refs)); in add_ref() 64 if (!refs) { in add_ref() [all …]
|
/linux-6.12.1/drivers/media/v4l2-core/ |
D | v4l2-h264.c | 60 b->refs[i].longterm = true; in v4l2_h264_init_reflist_builder() 68 if (!b->refs[i].longterm && dpb[i].frame_num > cur_frame_num) in v4l2_h264_init_reflist_builder() 69 b->refs[i].frame_num = (int)dpb[i].frame_num - in v4l2_h264_init_reflist_builder() 72 b->refs[i].frame_num = dpb[i].frame_num; in v4l2_h264_init_reflist_builder() 74 b->refs[i].top_field_order_cnt = dpb[i].top_field_order_cnt; in v4l2_h264_init_reflist_builder() 75 b->refs[i].bottom_field_order_cnt = dpb[i].bottom_field_order_cnt; in v4l2_h264_init_reflist_builder() 113 return min(b->refs[ref->index].top_field_order_cnt, in v4l2_h264_get_poc() 114 b->refs[ref->index].bottom_field_order_cnt); in v4l2_h264_get_poc() 116 return b->refs[ref->index].top_field_order_cnt; in v4l2_h264_get_poc() 118 return b->refs[ref->index].bottom_field_order_cnt; in v4l2_h264_get_poc() [all …]
|
/linux-6.12.1/drivers/xen/ |
D | gntdev-dmabuf.c | 44 grant_ref_t *refs; member 437 int count, u32 domid, u32 *refs, u32 *fd) in dmabuf_exp_from_refs() argument 449 map->grants[i].ref = refs[i]; in dmabuf_exp_from_refs() 488 dmabuf_imp_grant_foreign_access(unsigned long *gfns, u32 *refs, in dmabuf_imp_grant_foreign_access() argument 512 refs[i] = cur_ref; in dmabuf_imp_grant_foreign_access() 522 static void dmabuf_imp_end_foreign_access(u32 *refs, int count) in dmabuf_imp_end_foreign_access() argument 527 if (refs[i] != INVALID_GRANT_REF) in dmabuf_imp_end_foreign_access() 528 gnttab_end_foreign_access(refs[i], NULL); in dmabuf_imp_end_foreign_access() 533 kfree(gntdev_dmabuf->u.imp.refs); in dmabuf_imp_free_storage() 546 gntdev_dmabuf->u.imp.refs = kcalloc(count, in dmabuf_imp_alloc_storage() [all …]
|
/linux-6.12.1/io_uring/ |
D | refs.h | 12 ((unsigned int) atomic_read(&(req->refs)) + 127u <= 127u) 17 return atomic_inc_not_zero(&req->refs); in req_ref_inc_not_zero() 26 return atomic_dec_and_test(&req->refs); in req_ref_put_and_test() 33 atomic_inc(&req->refs); in req_ref_get() 40 atomic_dec(&req->refs); in req_ref_put() 47 atomic_set(&req->refs, nr); in __io_req_set_refcount()
|
D | eventfd.c | 18 refcount_t refs; member 40 if (refcount_dec_and_test(&ev_fd->refs)) in io_eventfd_do_signal() 66 if (!refcount_inc_not_zero(&ev_fd->refs)) in io_eventfd_signal() 80 if (refcount_dec_and_test(&ev_fd->refs)) in io_eventfd_signal() 140 refcount_set(&ev_fd->refs, 1); in io_eventfd_register() 155 if (refcount_dec_and_test(&ev_fd->refs)) in io_eventfd_unregister()
|
D | waitid.c | 29 atomic_t refs; member 124 WARN_ON_ONCE(!(atomic_read(&iw->refs) & IO_WAITID_REF_MASK)); in io_waitid_complete() 146 atomic_or(IO_WAITID_CANCEL_FLAG, &iw->refs); in __io_waitid_cancel() 149 if (atomic_fetch_inc(&iw->refs) & IO_WAITID_REF_MASK) in __io_waitid_cancel() 212 if (!atomic_sub_return(1, &iw->refs)) in io_waitid_drop_issue_ref() 246 if (!(atomic_read(&iw->refs) & IO_WAITID_CANCEL_FLAG)) { in io_waitid_cb() 276 if (atomic_fetch_inc(&iw->refs) & IO_WAITID_REF_MASK) in io_waitid_wait() 322 atomic_set(&iw->refs, 1); in io_waitid()
|
D | rsrc.h | 36 int refs; member 48 refcount_t refs; member 91 if (node && !--node->refs) in io_put_rsrc_node() 98 node->refs++; in io_charge_rsrc_node()
|
/linux-6.12.1/include/media/ |
D | v4l2-h264.h | 18 * @refs.top_field_order_cnt: top field order count 19 * @refs.bottom_field_order_cnt: bottom field order count 20 * @refs.frame_num: reference frame number 21 * @refs.longterm: set to true for a long term reference 22 * @refs: array of references 27 * @num_valid: number of valid references in the refs array 39 } refs[V4L2_H264_NUM_DPB_ENTRIES]; member
|
/linux-6.12.1/fs/smb/client/ |
D | dfs_cache.c | 251 static inline void dump_refs(const struct dfs_info3_param *refs, int numrefs) in dump_refs() argument 257 const struct dfs_info3_param *ref = &refs[i]; in dump_refs() 383 static int copy_ref_data(const struct dfs_info3_param *refs, int numrefs, in copy_ref_data() argument 389 ce->ttl = max_t(int, refs[0].ttl, CACHE_MIN_TTL); in copy_ref_data() 391 ce->srvtype = refs[0].server_type; in copy_ref_data() 392 ce->hdr_flags = refs[0].flags; in copy_ref_data() 393 ce->ref_flags = refs[0].ref_flag; in copy_ref_data() 394 ce->path_consumed = refs[0].path_consumed; in copy_ref_data() 399 t = alloc_target(refs[i].node_name, refs[i].path_consumed); in copy_ref_data() 421 static struct cache_entry *alloc_cache_entry(struct dfs_info3_param *refs, int numrefs) in alloc_cache_entry() argument [all …]
|
/linux-6.12.1/fs/btrfs/ |
D | delayed-ref.h | 87 refcount_t refs; member 90 * how many refs is this entry adding or deleting. For 91 * head refs, this may be a negative number because it is keeping 93 * For individual refs, this will always be a positive number 117 * the head refs are used to hold a lock on a given extent, which allows us 118 * to make sure that only one process is running the delayed refs 132 * the mutex is held while running the refs, and it is also 137 refcount_t refs; member 148 * This is used to track the final ref_mod from all the refs associated 149 * with this head ref, this is not adjusted as delayed refs are run, [all …]
|
D | extent-tree.c | 96 * the delayed refs are not processed. 100 u64 offset, int metadata, u64 *refs, u64 *flags, in btrfs_lookup_extent_info() argument 188 refcount_inc(&head->refs); in btrfs_lookup_extent_info() 213 if (refs) in btrfs_lookup_extent_info() 214 *refs = num_refs; in btrfs_lookup_extent_info() 225 * Back reference rules. Back refs have three main goals: 238 * There are two kinds of back refs. The implicit back refs is optimized 240 * back refs of this kind provide information about the block's owner tree 242 * b-tree searching. The full back refs is for pointers in tree blocks not 244 * in the back refs. Actually the full back refs is generic, and can be [all …]
|
D | backref.c | 218 * Return 0 when both refs are for the same block (and can be merged). 271 * Add @newref to the @root rbtree, merging identical refs. 300 /* Identical refs, merge them and free @newref */ in prelim_ref_insert() 357 * delayed refs 370 * on disk refs (inline or keyed) 418 /* direct refs use root == 0, key == NULL */ 428 /* indirect refs use parent == 0 */ 606 * adding new delayed refs. To deal with this we need to look in cache in resolve_indirect_ref() 717 * We maintain three separate rbtrees: one for direct refs, one for 718 * indirect refs which have a key, and one for indirect refs which do not [all …]
|
D | delayed-inode.c | 59 refcount_set(&delayed_node->refs, 0); in btrfs_init_delayed_node() 76 refcount_inc(&node->refs); in btrfs_get_delayed_node() 85 refcount_inc(&node->refs); /* can be accessed */ in btrfs_get_delayed_node() 107 if (refcount_inc_not_zero(&node->refs)) { in btrfs_get_delayed_node() 108 refcount_inc(&node->refs); in btrfs_get_delayed_node() 143 refcount_set(&node->refs, 2); in btrfs_get_or_create_delayed_node() 188 refcount_inc(&node->refs); /* inserted into list */ in btrfs_queue_delayed_node() 202 refcount_dec(&node->refs); /* not in the list */ in btrfs_dequeue_delayed_node() 223 refcount_inc(&node->refs); in btrfs_first_delayed_node() 250 refcount_inc(&next->refs); in btrfs_next_delayed_node() [all …]
|
/linux-6.12.1/tools/include/linux/ |
D | refcount.h | 53 atomic_t refs; member 56 #define REFCOUNT_INIT(n) { .refs = ATOMIC_INIT(n), } 60 atomic_set(&r->refs, n); in refcount_set() 65 return atomic_read(&r->refs); in refcount_read() 78 unsigned int old, new, val = atomic_read(&r->refs); in refcount_inc_not_zero() 89 old = atomic_cmpxchg_relaxed(&r->refs, val, new); in refcount_inc_not_zero() 123 unsigned int old, new, val = atomic_read(&r->refs); in refcount_sub_and_test() 135 old = atomic_cmpxchg_release(&r->refs, val, new); in refcount_sub_and_test()
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/core/ |
D | event.c | 34 if (--event->refs[index * event->types_nr + type] == 0) { in nvkm_event_put() 51 if (++event->refs[index * event->types_nr + type] == 1) { in nvkm_event_get() 175 if (!event->refs || WARN_ON(id >= event->index_nr)) in nvkm_event_ntfy() 194 if (event->refs) { in nvkm_event_fini() 195 kfree(event->refs); in nvkm_event_fini() 196 event->refs = NULL; in nvkm_event_fini() 204 event->refs = kzalloc(array3_size(index_nr, types_nr, sizeof(*event->refs)), GFP_KERNEL); in __nvkm_event_init() 205 if (!event->refs) in __nvkm_event_init()
|
/linux-6.12.1/tools/testing/selftests/net/ |
D | nl_netdev.py | 43 refs = sum([pp["inflight"] for pp in pp_list]) 44 ksft_eq(refs, 0) 55 refs = sum([pp["inflight"] for pp in pp_list if pp.get("ifindex") == nsim.ifindex]) 56 ksft_ge(refs, 1) 62 refs = sum([pp["inflight"] for pp in pp_list]) 63 ksft_eq(refs, 1)
|
/linux-6.12.1/tools/net/ynl/samples/ |
D | page-pool.c | 18 size_t refs, bytes; member 52 s->live[l].refs += pp->inflight; in count() 126 printf("\t\trefs: %zu bytes: %zu (refs: %zu bytes: %zu)\n", in main() 127 s->live[1].refs, s->live[1].bytes, in main() 128 s->live[0].refs, s->live[0].bytes); in main()
|
/linux-6.12.1/drivers/acpi/acpica/ |
D | acpredef.h | 225 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 229 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 233 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 237 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 241 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 245 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 249 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 253 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 257 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 261 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ [all …]
|
/linux-6.12.1/drivers/android/ |
D | binder_internal.h | 187 * @refs: list of references on this node 193 * @local_weak_refs: weak user refs from local process 196 * @local_strong_refs: strong user refs from local process 199 * @tmp_refs: temporary kernel refs 243 struct hlist_head refs; member 314 * @node_entry: list entry for node->refs list in target node 332 /* node => refs + procs (proc exit) */ 351 * @refs_by_desc: rbtree of refs ordered by ref->desc 353 * @refs_by_node: rbtree of refs ordered by ref->node 568 * @fbo: binder object (nodes and refs)
|
/linux-6.12.1/tools/testing/selftests/bpf/progs/ |
D | map_kptr.c | 318 if (p_st->cnt.refs.counter != ref) { in test_map_kptr_ref_pre() 328 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_pre() 336 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_pre() 348 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_pre() 365 if (!p_st || p_st->cnt.refs.counter != ref) in test_map_kptr_ref_post() 371 if (p_st->cnt.refs.counter != ref) { in test_map_kptr_ref_post() 381 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_post() 481 if (p->cnt.refs.counter != ref) { in test_map_kptr_ref3()
|
/linux-6.12.1/drivers/thermal/ |
D | k3_j72xx_bandgap.c | 74 * @refs: reference error values for -40C, 30C, 125C & 150C 78 int refs[4]; member 90 idx1 = err_vals->refs[seg]; in create_table_segments() 92 idx2 = err_vals->refs[seg + 1]; in create_table_segments() 95 ref1 = err_vals->refs[seg]; in create_table_segments() 96 ref2 = err_vals->refs[seg + 1]; in create_table_segments() 490 err_vals.refs[0] = MINUS40CREF; in k3_j72xx_bandgap_probe() 491 err_vals.refs[1] = PLUS30CREF; in k3_j72xx_bandgap_probe() 492 err_vals.refs[2] = PLUS125CREF; in k3_j72xx_bandgap_probe() 493 err_vals.refs[3] = PLUS150CREF; in k3_j72xx_bandgap_probe()
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
D | cgrp.c | 39 if (refcount_dec_and_test(&ectx->refs)) { in nvkm_cgrp_ectx_put() 65 refcount_inc(&ectx->refs); in nvkm_cgrp_ectx_get() 76 refcount_set(&ectx->refs, 1); in nvkm_cgrp_ectx_get() 100 if (refcount_dec_and_test(&vctx->refs)) { in nvkm_cgrp_vctx_put() 130 refcount_inc(&vctx->refs); in nvkm_cgrp_vctx_get() 151 refcount_set(&vctx->refs, 1); in nvkm_cgrp_vctx_get()
|
/linux-6.12.1/mm/ |
D | gup.c | 73 static inline struct folio *try_get_folio(struct page *page, int refs) in try_get_folio() argument 81 if (unlikely(!folio_ref_try_add(folio, refs))) in try_get_folio() 94 if (!put_devmap_managed_folio_refs(folio, refs)) in try_get_folio() 95 folio_put_refs(folio, refs); in try_get_folio() 102 static void gup_put_folio(struct folio *folio, int refs, unsigned int flags) in gup_put_folio() argument 107 node_stat_mod_folio(folio, NR_FOLL_PIN_RELEASED, refs); in gup_put_folio() 109 atomic_sub(refs, &folio->_pincount); in gup_put_folio() 111 refs *= GUP_PIN_COUNTING_BIAS; in gup_put_folio() 114 if (!put_devmap_managed_folio_refs(folio, refs)) in gup_put_folio() 115 folio_put_refs(folio, refs); in gup_put_folio() [all …]
|
/linux-6.12.1/include/uapi/xen/ |
D | gntdev.h | 70 struct ioctl_gntdev_grant_ref refs[1]; member 226 * Create a dma-buf [1] from grant references @refs of count @count provided 247 /* Number of grant references in @refs array. */ 256 __u32 refs[1]; member 281 * to the pages of that dma-buf into array @refs of size @count. 290 /* Number of grant references in @refs array. */ 298 __u32 refs[1]; member
|
/linux-6.12.1/fs/btrfs/tests/ |
D | qgroup-tests.c | 75 u64 refs; in add_tree_ref() local 99 refs = btrfs_extent_refs(path->nodes[0], item); in add_tree_ref() 100 btrfs_set_extent_refs(path->nodes[0], item, refs + 1); in add_tree_ref() 157 u64 refs; in remove_extent_ref() local 181 refs = btrfs_extent_refs(path->nodes[0], item); in remove_extent_ref() 182 btrfs_set_extent_refs(path->nodes[0], item, refs - 1); in remove_extent_ref() 229 * Since the test trans doesn't have the complicated delayed refs, in test_no_shared_qgroup() 329 test_msg("running qgroup multiple refs test"); in test_multiple_refs()
|