Home
last modified time | relevance | path

Searched full:refs (Results 1 – 25 of 331) sorted by relevance

12345678910>>...14

/linux-6.12.1/tools/bpf/bpftool/
Dpids.c34 struct obj_refs *refs; in add_ref() local
40 refs = entry->pvalue; in add_ref()
42 for (i = 0; i < refs->ref_cnt; i++) { in add_ref()
43 if (refs->refs[i].pid == e->pid) in add_ref()
47 tmp = realloc(refs->refs, (refs->ref_cnt + 1) * sizeof(*ref)); in add_ref()
53 refs->refs = tmp; in add_ref()
54 ref = &refs->refs[refs->ref_cnt]; in add_ref()
57 refs->ref_cnt++; in add_ref()
63 refs = calloc(1, sizeof(*refs)); in add_ref()
64 if (!refs) { in add_ref()
[all …]
/linux-6.12.1/drivers/media/v4l2-core/
Dv4l2-h264.c60 b->refs[i].longterm = true; in v4l2_h264_init_reflist_builder()
68 if (!b->refs[i].longterm && dpb[i].frame_num > cur_frame_num) in v4l2_h264_init_reflist_builder()
69 b->refs[i].frame_num = (int)dpb[i].frame_num - in v4l2_h264_init_reflist_builder()
72 b->refs[i].frame_num = dpb[i].frame_num; in v4l2_h264_init_reflist_builder()
74 b->refs[i].top_field_order_cnt = dpb[i].top_field_order_cnt; in v4l2_h264_init_reflist_builder()
75 b->refs[i].bottom_field_order_cnt = dpb[i].bottom_field_order_cnt; in v4l2_h264_init_reflist_builder()
113 return min(b->refs[ref->index].top_field_order_cnt, in v4l2_h264_get_poc()
114 b->refs[ref->index].bottom_field_order_cnt); in v4l2_h264_get_poc()
116 return b->refs[ref->index].top_field_order_cnt; in v4l2_h264_get_poc()
118 return b->refs[ref->index].bottom_field_order_cnt; in v4l2_h264_get_poc()
[all …]
/linux-6.12.1/drivers/xen/
Dgntdev-dmabuf.c44 grant_ref_t *refs; member
437 int count, u32 domid, u32 *refs, u32 *fd) in dmabuf_exp_from_refs() argument
449 map->grants[i].ref = refs[i]; in dmabuf_exp_from_refs()
488 dmabuf_imp_grant_foreign_access(unsigned long *gfns, u32 *refs, in dmabuf_imp_grant_foreign_access() argument
512 refs[i] = cur_ref; in dmabuf_imp_grant_foreign_access()
522 static void dmabuf_imp_end_foreign_access(u32 *refs, int count) in dmabuf_imp_end_foreign_access() argument
527 if (refs[i] != INVALID_GRANT_REF) in dmabuf_imp_end_foreign_access()
528 gnttab_end_foreign_access(refs[i], NULL); in dmabuf_imp_end_foreign_access()
533 kfree(gntdev_dmabuf->u.imp.refs); in dmabuf_imp_free_storage()
546 gntdev_dmabuf->u.imp.refs = kcalloc(count, in dmabuf_imp_alloc_storage()
[all …]
/linux-6.12.1/io_uring/
Drefs.h12 ((unsigned int) atomic_read(&(req->refs)) + 127u <= 127u)
17 return atomic_inc_not_zero(&req->refs); in req_ref_inc_not_zero()
26 return atomic_dec_and_test(&req->refs); in req_ref_put_and_test()
33 atomic_inc(&req->refs); in req_ref_get()
40 atomic_dec(&req->refs); in req_ref_put()
47 atomic_set(&req->refs, nr); in __io_req_set_refcount()
Deventfd.c18 refcount_t refs; member
40 if (refcount_dec_and_test(&ev_fd->refs)) in io_eventfd_do_signal()
66 if (!refcount_inc_not_zero(&ev_fd->refs)) in io_eventfd_signal()
80 if (refcount_dec_and_test(&ev_fd->refs)) in io_eventfd_signal()
140 refcount_set(&ev_fd->refs, 1); in io_eventfd_register()
155 if (refcount_dec_and_test(&ev_fd->refs)) in io_eventfd_unregister()
Dwaitid.c29 atomic_t refs; member
124 WARN_ON_ONCE(!(atomic_read(&iw->refs) & IO_WAITID_REF_MASK)); in io_waitid_complete()
146 atomic_or(IO_WAITID_CANCEL_FLAG, &iw->refs); in __io_waitid_cancel()
149 if (atomic_fetch_inc(&iw->refs) & IO_WAITID_REF_MASK) in __io_waitid_cancel()
212 if (!atomic_sub_return(1, &iw->refs)) in io_waitid_drop_issue_ref()
246 if (!(atomic_read(&iw->refs) & IO_WAITID_CANCEL_FLAG)) { in io_waitid_cb()
276 if (atomic_fetch_inc(&iw->refs) & IO_WAITID_REF_MASK) in io_waitid_wait()
322 atomic_set(&iw->refs, 1); in io_waitid()
Drsrc.h36 int refs; member
48 refcount_t refs; member
91 if (node && !--node->refs) in io_put_rsrc_node()
98 node->refs++; in io_charge_rsrc_node()
/linux-6.12.1/include/media/
Dv4l2-h264.h18 * @refs.top_field_order_cnt: top field order count
19 * @refs.bottom_field_order_cnt: bottom field order count
20 * @refs.frame_num: reference frame number
21 * @refs.longterm: set to true for a long term reference
22 * @refs: array of references
27 * @num_valid: number of valid references in the refs array
39 } refs[V4L2_H264_NUM_DPB_ENTRIES]; member
/linux-6.12.1/fs/smb/client/
Ddfs_cache.c251 static inline void dump_refs(const struct dfs_info3_param *refs, int numrefs) in dump_refs() argument
257 const struct dfs_info3_param *ref = &refs[i]; in dump_refs()
383 static int copy_ref_data(const struct dfs_info3_param *refs, int numrefs, in copy_ref_data() argument
389 ce->ttl = max_t(int, refs[0].ttl, CACHE_MIN_TTL); in copy_ref_data()
391 ce->srvtype = refs[0].server_type; in copy_ref_data()
392 ce->hdr_flags = refs[0].flags; in copy_ref_data()
393 ce->ref_flags = refs[0].ref_flag; in copy_ref_data()
394 ce->path_consumed = refs[0].path_consumed; in copy_ref_data()
399 t = alloc_target(refs[i].node_name, refs[i].path_consumed); in copy_ref_data()
421 static struct cache_entry *alloc_cache_entry(struct dfs_info3_param *refs, int numrefs) in alloc_cache_entry() argument
[all …]
/linux-6.12.1/fs/btrfs/
Ddelayed-ref.h87 refcount_t refs; member
90 * how many refs is this entry adding or deleting. For
91 * head refs, this may be a negative number because it is keeping
93 * For individual refs, this will always be a positive number
117 * the head refs are used to hold a lock on a given extent, which allows us
118 * to make sure that only one process is running the delayed refs
132 * the mutex is held while running the refs, and it is also
137 refcount_t refs; member
148 * This is used to track the final ref_mod from all the refs associated
149 * with this head ref, this is not adjusted as delayed refs are run,
[all …]
Dextent-tree.c96 * the delayed refs are not processed.
100 u64 offset, int metadata, u64 *refs, u64 *flags, in btrfs_lookup_extent_info() argument
188 refcount_inc(&head->refs); in btrfs_lookup_extent_info()
213 if (refs) in btrfs_lookup_extent_info()
214 *refs = num_refs; in btrfs_lookup_extent_info()
225 * Back reference rules. Back refs have three main goals:
238 * There are two kinds of back refs. The implicit back refs is optimized
240 * back refs of this kind provide information about the block's owner tree
242 * b-tree searching. The full back refs is for pointers in tree blocks not
244 * in the back refs. Actually the full back refs is generic, and can be
[all …]
Dbackref.c218 * Return 0 when both refs are for the same block (and can be merged).
271 * Add @newref to the @root rbtree, merging identical refs.
300 /* Identical refs, merge them and free @newref */ in prelim_ref_insert()
357 * delayed refs
370 * on disk refs (inline or keyed)
418 /* direct refs use root == 0, key == NULL */
428 /* indirect refs use parent == 0 */
606 * adding new delayed refs. To deal with this we need to look in cache in resolve_indirect_ref()
717 * We maintain three separate rbtrees: one for direct refs, one for
718 * indirect refs which have a key, and one for indirect refs which do not
[all …]
Ddelayed-inode.c59 refcount_set(&delayed_node->refs, 0); in btrfs_init_delayed_node()
76 refcount_inc(&node->refs); in btrfs_get_delayed_node()
85 refcount_inc(&node->refs); /* can be accessed */ in btrfs_get_delayed_node()
107 if (refcount_inc_not_zero(&node->refs)) { in btrfs_get_delayed_node()
108 refcount_inc(&node->refs); in btrfs_get_delayed_node()
143 refcount_set(&node->refs, 2); in btrfs_get_or_create_delayed_node()
188 refcount_inc(&node->refs); /* inserted into list */ in btrfs_queue_delayed_node()
202 refcount_dec(&node->refs); /* not in the list */ in btrfs_dequeue_delayed_node()
223 refcount_inc(&node->refs); in btrfs_first_delayed_node()
250 refcount_inc(&next->refs); in btrfs_next_delayed_node()
[all …]
/linux-6.12.1/tools/include/linux/
Drefcount.h53 atomic_t refs; member
56 #define REFCOUNT_INIT(n) { .refs = ATOMIC_INIT(n), }
60 atomic_set(&r->refs, n); in refcount_set()
65 return atomic_read(&r->refs); in refcount_read()
78 unsigned int old, new, val = atomic_read(&r->refs); in refcount_inc_not_zero()
89 old = atomic_cmpxchg_relaxed(&r->refs, val, new); in refcount_inc_not_zero()
123 unsigned int old, new, val = atomic_read(&r->refs); in refcount_sub_and_test()
135 old = atomic_cmpxchg_release(&r->refs, val, new); in refcount_sub_and_test()
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/core/
Devent.c34 if (--event->refs[index * event->types_nr + type] == 0) { in nvkm_event_put()
51 if (++event->refs[index * event->types_nr + type] == 1) { in nvkm_event_get()
175 if (!event->refs || WARN_ON(id >= event->index_nr)) in nvkm_event_ntfy()
194 if (event->refs) { in nvkm_event_fini()
195 kfree(event->refs); in nvkm_event_fini()
196 event->refs = NULL; in nvkm_event_fini()
204 event->refs = kzalloc(array3_size(index_nr, types_nr, sizeof(*event->refs)), GFP_KERNEL); in __nvkm_event_init()
205 if (!event->refs) in __nvkm_event_init()
/linux-6.12.1/tools/testing/selftests/net/
Dnl_netdev.py43 refs = sum([pp["inflight"] for pp in pp_list])
44 ksft_eq(refs, 0)
55 refs = sum([pp["inflight"] for pp in pp_list if pp.get("ifindex") == nsim.ifindex])
56 ksft_ge(refs, 1)
62 refs = sum([pp["inflight"] for pp in pp_list])
63 ksft_eq(refs, 1)
/linux-6.12.1/tools/net/ynl/samples/
Dpage-pool.c18 size_t refs, bytes; member
52 s->live[l].refs += pp->inflight; in count()
126 printf("\t\trefs: %zu bytes: %zu (refs: %zu bytes: %zu)\n", in main()
127 s->live[1].refs, s->live[1].bytes, in main()
128 s->live[0].refs, s->live[0].bytes); in main()
/linux-6.12.1/drivers/acpi/acpica/
Dacpredef.h225 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */
229 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */
233 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */
237 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */
241 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */
245 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */
249 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */
253 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */
257 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */
261 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */
[all …]
/linux-6.12.1/drivers/android/
Dbinder_internal.h187 * @refs: list of references on this node
193 * @local_weak_refs: weak user refs from local process
196 * @local_strong_refs: strong user refs from local process
199 * @tmp_refs: temporary kernel refs
243 struct hlist_head refs; member
314 * @node_entry: list entry for node->refs list in target node
332 /* node => refs + procs (proc exit) */
351 * @refs_by_desc: rbtree of refs ordered by ref->desc
353 * @refs_by_node: rbtree of refs ordered by ref->node
568 * @fbo: binder object (nodes and refs)
/linux-6.12.1/tools/testing/selftests/bpf/progs/
Dmap_kptr.c318 if (p_st->cnt.refs.counter != ref) { in test_map_kptr_ref_pre()
328 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_pre()
336 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_pre()
348 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_pre()
365 if (!p_st || p_st->cnt.refs.counter != ref) in test_map_kptr_ref_post()
371 if (p_st->cnt.refs.counter != ref) { in test_map_kptr_ref_post()
381 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_post()
481 if (p->cnt.refs.counter != ref) { in test_map_kptr_ref3()
/linux-6.12.1/drivers/thermal/
Dk3_j72xx_bandgap.c74 * @refs: reference error values for -40C, 30C, 125C & 150C
78 int refs[4]; member
90 idx1 = err_vals->refs[seg]; in create_table_segments()
92 idx2 = err_vals->refs[seg + 1]; in create_table_segments()
95 ref1 = err_vals->refs[seg]; in create_table_segments()
96 ref2 = err_vals->refs[seg + 1]; in create_table_segments()
490 err_vals.refs[0] = MINUS40CREF; in k3_j72xx_bandgap_probe()
491 err_vals.refs[1] = PLUS30CREF; in k3_j72xx_bandgap_probe()
492 err_vals.refs[2] = PLUS125CREF; in k3_j72xx_bandgap_probe()
493 err_vals.refs[3] = PLUS150CREF; in k3_j72xx_bandgap_probe()
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
Dcgrp.c39 if (refcount_dec_and_test(&ectx->refs)) { in nvkm_cgrp_ectx_put()
65 refcount_inc(&ectx->refs); in nvkm_cgrp_ectx_get()
76 refcount_set(&ectx->refs, 1); in nvkm_cgrp_ectx_get()
100 if (refcount_dec_and_test(&vctx->refs)) { in nvkm_cgrp_vctx_put()
130 refcount_inc(&vctx->refs); in nvkm_cgrp_vctx_get()
151 refcount_set(&vctx->refs, 1); in nvkm_cgrp_vctx_get()
/linux-6.12.1/mm/
Dgup.c73 static inline struct folio *try_get_folio(struct page *page, int refs) in try_get_folio() argument
81 if (unlikely(!folio_ref_try_add(folio, refs))) in try_get_folio()
94 if (!put_devmap_managed_folio_refs(folio, refs)) in try_get_folio()
95 folio_put_refs(folio, refs); in try_get_folio()
102 static void gup_put_folio(struct folio *folio, int refs, unsigned int flags) in gup_put_folio() argument
107 node_stat_mod_folio(folio, NR_FOLL_PIN_RELEASED, refs); in gup_put_folio()
109 atomic_sub(refs, &folio->_pincount); in gup_put_folio()
111 refs *= GUP_PIN_COUNTING_BIAS; in gup_put_folio()
114 if (!put_devmap_managed_folio_refs(folio, refs)) in gup_put_folio()
115 folio_put_refs(folio, refs); in gup_put_folio()
[all …]
/linux-6.12.1/include/uapi/xen/
Dgntdev.h70 struct ioctl_gntdev_grant_ref refs[1]; member
226 * Create a dma-buf [1] from grant references @refs of count @count provided
247 /* Number of grant references in @refs array. */
256 __u32 refs[1]; member
281 * to the pages of that dma-buf into array @refs of size @count.
290 /* Number of grant references in @refs array. */
298 __u32 refs[1]; member
/linux-6.12.1/fs/btrfs/tests/
Dqgroup-tests.c75 u64 refs; in add_tree_ref() local
99 refs = btrfs_extent_refs(path->nodes[0], item); in add_tree_ref()
100 btrfs_set_extent_refs(path->nodes[0], item, refs + 1); in add_tree_ref()
157 u64 refs; in remove_extent_ref() local
181 refs = btrfs_extent_refs(path->nodes[0], item); in remove_extent_ref()
182 btrfs_set_extent_refs(path->nodes[0], item, refs - 1); in remove_extent_ref()
229 * Since the test trans doesn't have the complicated delayed refs, in test_no_shared_qgroup()
329 test_msg("running qgroup multiple refs test"); in test_multiple_refs()

12345678910>>...14