Home
last modified time | relevance | path

Searched refs:alloc (Results 1 – 25 of 582) sorted by relevance

12345678910>>...24

/linux-6.12.1/drivers/android/
Dbinder_alloc.c60 static size_t binder_alloc_buffer_size(struct binder_alloc *alloc, in binder_alloc_buffer_size() argument
63 if (list_is_last(&buffer->entry, &alloc->buffers)) in binder_alloc_buffer_size()
64 return alloc->buffer + alloc->buffer_size - buffer->user_data; in binder_alloc_buffer_size()
68 static void binder_insert_free_buffer(struct binder_alloc *alloc, in binder_insert_free_buffer() argument
71 struct rb_node **p = &alloc->free_buffers.rb_node; in binder_insert_free_buffer()
79 new_buffer_size = binder_alloc_buffer_size(alloc, new_buffer); in binder_insert_free_buffer()
83 alloc->pid, new_buffer_size, new_buffer); in binder_insert_free_buffer()
90 buffer_size = binder_alloc_buffer_size(alloc, buffer); in binder_insert_free_buffer()
98 rb_insert_color(&new_buffer->rb_node, &alloc->free_buffers); in binder_insert_free_buffer()
102 struct binder_alloc *alloc, struct binder_buffer *new_buffer) in binder_insert_allocated_buffer_locked() argument
[all …]
Dbinder_alloc_selftest.c96 static bool check_buffer_pages_allocated(struct binder_alloc *alloc, in check_buffer_pages_allocated() argument
107 page_index = (page_addr - alloc->buffer) / PAGE_SIZE; in check_buffer_pages_allocated()
108 if (!alloc->pages[page_index].page_ptr || in check_buffer_pages_allocated()
109 !list_empty(&alloc->pages[page_index].lru)) { in check_buffer_pages_allocated()
111 alloc->pages[page_index].page_ptr ? in check_buffer_pages_allocated()
119 static void binder_selftest_alloc_buf(struct binder_alloc *alloc, in binder_selftest_alloc_buf() argument
126 buffers[i] = binder_alloc_new_buf(alloc, sizes[i], 0, 0, 0); in binder_selftest_alloc_buf()
128 !check_buffer_pages_allocated(alloc, buffers[i], in binder_selftest_alloc_buf()
136 static void binder_selftest_free_buf(struct binder_alloc *alloc, in binder_selftest_free_buf() argument
143 binder_alloc_free_buf(alloc, buffers[seq[i]]); in binder_selftest_free_buf()
[all …]
Dbinder_alloc.h70 struct binder_alloc *alloc; member
115 void binder_selftest_alloc(struct binder_alloc *alloc);
117 static inline void binder_selftest_alloc(struct binder_alloc *alloc) {} in binder_selftest_alloc() argument
122 struct binder_buffer *binder_alloc_new_buf(struct binder_alloc *alloc,
127 void binder_alloc_init(struct binder_alloc *alloc);
130 void binder_alloc_vma_close(struct binder_alloc *alloc);
132 binder_alloc_prepare_to_free(struct binder_alloc *alloc,
134 void binder_alloc_free_buf(struct binder_alloc *alloc,
136 int binder_alloc_mmap_handler(struct binder_alloc *alloc,
138 void binder_alloc_deferred_release(struct binder_alloc *alloc);
[all …]
Dbinder_trace.h319 TP_PROTO(struct binder_alloc *alloc, bool allocate,
321 TP_ARGS(alloc, allocate, start, end),
329 __entry->proc = alloc->pid;
331 __entry->offset = start - alloc->buffer;
340 TP_PROTO(const struct binder_alloc *alloc, size_t page_index),
341 TP_ARGS(alloc, page_index),
347 __entry->proc = alloc->pid;
355 TP_PROTO(const struct binder_alloc *alloc, size_t page_index),
356 TP_ARGS(alloc, page_index));
359 TP_PROTO(const struct binder_alloc *alloc, size_t page_index),
[all …]
/linux-6.12.1/drivers/infiniband/hw/cxgb4/
Did_table.c44 u32 c4iw_id_alloc(struct c4iw_id_table *alloc) in c4iw_id_alloc() argument
49 spin_lock_irqsave(&alloc->lock, flags); in c4iw_id_alloc()
51 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in c4iw_id_alloc()
52 if (obj >= alloc->max) in c4iw_id_alloc()
53 obj = find_first_zero_bit(alloc->table, alloc->max); in c4iw_id_alloc()
55 if (obj < alloc->max) { in c4iw_id_alloc()
56 if (alloc->flags & C4IW_ID_TABLE_F_RANDOM) in c4iw_id_alloc()
57 alloc->last += get_random_u32_below(RANDOM_SKIP); in c4iw_id_alloc()
59 alloc->last = obj + 1; in c4iw_id_alloc()
60 if (alloc->last >= alloc->max) in c4iw_id_alloc()
[all …]
/linux-6.12.1/sound/isa/gus/
Dgus_mem.c18 void snd_gf1_mem_lock(struct snd_gf1_mem * alloc, int xup) in snd_gf1_mem_lock() argument
21 mutex_lock(&alloc->memory_mutex); in snd_gf1_mem_lock()
23 mutex_unlock(&alloc->memory_mutex); in snd_gf1_mem_lock()
28 snd_gf1_mem_xalloc(struct snd_gf1_mem *alloc, struct snd_gf1_mem_block *block, in snd_gf1_mem_xalloc() argument
43 pblock = alloc->first; in snd_gf1_mem_xalloc()
49 if (pblock == alloc->first) in snd_gf1_mem_xalloc()
50 alloc->first = nblock; in snd_gf1_mem_xalloc()
53 mutex_unlock(&alloc->memory_mutex); in snd_gf1_mem_xalloc()
59 if (alloc->last == NULL) { in snd_gf1_mem_xalloc()
61 alloc->first = alloc->last = nblock; in snd_gf1_mem_xalloc()
[all …]
/linux-6.12.1/drivers/infiniband/hw/mthca/
Dmthca_allocator.c40 u32 mthca_alloc(struct mthca_alloc *alloc) in mthca_alloc() argument
45 spin_lock_irqsave(&alloc->lock, flags); in mthca_alloc()
47 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in mthca_alloc()
48 if (obj >= alloc->max) { in mthca_alloc()
49 alloc->top = (alloc->top + alloc->max) & alloc->mask; in mthca_alloc()
50 obj = find_first_zero_bit(alloc->table, alloc->max); in mthca_alloc()
53 if (obj < alloc->max) { in mthca_alloc()
54 __set_bit(obj, alloc->table); in mthca_alloc()
55 obj |= alloc->top; in mthca_alloc()
59 spin_unlock_irqrestore(&alloc->lock, flags); in mthca_alloc()
[all …]
Dmthca_uar.c40 uar->index = mthca_alloc(&dev->uar_table.alloc); in mthca_uar_alloc()
51 mthca_free(&dev->uar_table.alloc, uar->index); in mthca_uar_free()
58 ret = mthca_alloc_init(&dev->uar_table.alloc, in mthca_init_uar_table()
67 mthca_alloc_cleanup(&dev->uar_table.alloc); in mthca_init_uar_table()
77 mthca_alloc_cleanup(&dev->uar_table.alloc); in mthca_cleanup_uar_table()
Dmthca_pd.c46 pd->pd_num = mthca_alloc(&dev->pd_table.alloc); in mthca_pd_alloc()
56 mthca_free(&dev->pd_table.alloc, pd->pd_num); in mthca_pd_alloc()
66 mthca_free(&dev->pd_table.alloc, pd->pd_num); in mthca_pd_free()
71 return mthca_alloc_init(&dev->pd_table.alloc, in mthca_init_pd_table()
80 mthca_alloc_cleanup(&dev->pd_table.alloc); in mthca_cleanup_pd_table()
/linux-6.12.1/fs/ocfs2/
Dlocalalloc.c35 static u32 ocfs2_local_alloc_count_bits(struct ocfs2_dinode *alloc);
38 struct ocfs2_dinode *alloc,
42 static void ocfs2_clear_local_alloc(struct ocfs2_dinode *alloc);
46 struct ocfs2_dinode *alloc,
272 struct ocfs2_dinode *alloc = NULL; in ocfs2_load_local_alloc() local
306 alloc = (struct ocfs2_dinode *) alloc_bh->b_data; in ocfs2_load_local_alloc()
307 la = OCFS2_LOCAL_ALLOC(alloc); in ocfs2_load_local_alloc()
309 if (!(le32_to_cpu(alloc->i_flags) & in ocfs2_load_local_alloc()
326 num_used = ocfs2_local_alloc_count_bits(alloc); in ocfs2_load_local_alloc()
331 || alloc->id1.bitmap1.i_used in ocfs2_load_local_alloc()
[all …]
/linux-6.12.1/fs/xfs/libxfs/
Dxfs_alloc_btree.c139 key->alloc.ar_startblock = rec->alloc.ar_startblock; in xfs_allocbt_init_key_from_rec()
140 key->alloc.ar_blockcount = rec->alloc.ar_blockcount; in xfs_allocbt_init_key_from_rec()
150 x = be32_to_cpu(rec->alloc.ar_startblock); in xfs_bnobt_init_high_key_from_rec()
151 x += be32_to_cpu(rec->alloc.ar_blockcount) - 1; in xfs_bnobt_init_high_key_from_rec()
152 key->alloc.ar_startblock = cpu_to_be32(x); in xfs_bnobt_init_high_key_from_rec()
153 key->alloc.ar_blockcount = 0; in xfs_bnobt_init_high_key_from_rec()
161 key->alloc.ar_blockcount = rec->alloc.ar_blockcount; in xfs_cntbt_init_high_key_from_rec()
162 key->alloc.ar_startblock = 0; in xfs_cntbt_init_high_key_from_rec()
170 rec->alloc.ar_startblock = cpu_to_be32(cur->bc_rec.a.ar_startblock); in xfs_allocbt_init_rec_from_cur()
171 rec->alloc.ar_blockcount = cpu_to_be32(cur->bc_rec.a.ar_blockcount); in xfs_allocbt_init_rec_from_cur()
[all …]
/linux-6.12.1/lib/zstd/compress/
Dzstd_cwksp.h247 void* const alloc = (BYTE*)ws->allocStart - bytes; in ZSTD_cwksp_reserve_internal_buffer_space() local
250 alloc, bytes, ZSTD_cwksp_available_space(ws) - bytes); in ZSTD_cwksp_reserve_internal_buffer_space()
252 assert(alloc >= bottom); in ZSTD_cwksp_reserve_internal_buffer_space()
253 if (alloc < bottom) { in ZSTD_cwksp_reserve_internal_buffer_space()
260 if (alloc < ws->tableValidEnd) { in ZSTD_cwksp_reserve_internal_buffer_space()
261 ws->tableValidEnd = alloc; in ZSTD_cwksp_reserve_internal_buffer_space()
263 ws->allocStart = alloc; in ZSTD_cwksp_reserve_internal_buffer_space()
264 return alloc; in ZSTD_cwksp_reserve_internal_buffer_space()
295 void* const alloc = ws->objectEnd; in ZSTD_cwksp_internal_advance_phase() local
296 … size_t const bytesToAlign = ZSTD_cwksp_bytes_to_align_ptr(alloc, ZSTD_CWKSP_ALIGNMENT_BYTES); in ZSTD_cwksp_internal_advance_phase()
[all …]
/linux-6.12.1/tools/perf/util/
Dstrbuf.c22 sb->alloc = sb->len = 0; in strbuf_init()
31 if (sb->alloc) { in strbuf_release()
39 char *res = sb->alloc ? sb->buf : NULL; in strbuf_detach()
51 if (nr < sb->alloc) in strbuf_grow()
57 if (alloc_nr(sb->alloc) > nr) in strbuf_grow()
58 nr = alloc_nr(sb->alloc); in strbuf_grow()
64 buf = realloc(sb->alloc ? sb->buf : NULL, nr * sizeof(*buf)); in strbuf_grow()
69 sb->alloc = nr; in strbuf_grow()
106 len = vsnprintf(sb->buf + sb->len, sb->alloc - sb->len, fmt, ap); in strbuf_addv()
117 len = vsnprintf(sb->buf + sb->len, sb->alloc - sb->len, fmt, ap_saved); in strbuf_addv()
[all …]
Dstrbuf.h51 size_t alloc; member
65 return sb->alloc ? sb->alloc - sb->len - 1 : 0; in strbuf_avail()
71 if (!sb->alloc) { in strbuf_setlen()
76 assert(len < sb->alloc); in strbuf_setlen()
Dhelp-unknown-cmd.c37 if (nr > cmds->alloc) { in add_cmd_list()
39 if (alloc_nr(cmds->alloc) < nr) in add_cmd_list()
40 cmds->alloc = nr; in add_cmd_list()
42 cmds->alloc = alloc_nr(cmds->alloc); in add_cmd_list()
43 tmp = realloc(cmds->names, cmds->alloc * sizeof(*cmds->names)); in add_cmd_list()
/linux-6.12.1/tools/lib/subcmd/
Dsubcmd-util.h38 #define ALLOC_GROW(x, nr, alloc) \ argument
40 if ((nr) > alloc) { \
41 if (alloc_nr(alloc) < (nr)) \
42 alloc = (nr); \
44 alloc = alloc_nr(alloc); \
45 x = xrealloc((x), alloc * sizeof(*(x))); \
/linux-6.12.1/tools/testing/selftests/mm/
Ddroppable.c22 void *alloc; in main() local
28 alloc = mmap(0, alloc_size, PROT_READ | PROT_WRITE, MAP_ANONYMOUS | MAP_DROPPABLE, -1, 0); in main()
29 assert(alloc != MAP_FAILED); in main()
30 memset(alloc, 'A', alloc_size); in main()
32 assert(*(uint8_t *)(alloc + i)); in main()
43 if (!*(uint8_t *)(alloc + i)) { in main()
/linux-6.12.1/arch/s390/mm/
Dpgalloc.c298 unsigned long end, int alloc) in base_page_walk() argument
302 if (!alloc) in base_page_walk()
314 unsigned long end, int alloc) in base_segment_walk() argument
324 if (!alloc) in base_segment_walk()
332 rc = base_page_walk(table, addr, next, alloc); in base_segment_walk()
335 if (!alloc) in base_segment_walk()
343 unsigned long end, int alloc) in base_region3_walk() argument
353 if (!alloc) in base_region3_walk()
361 rc = base_segment_walk(table, addr, next, alloc); in base_region3_walk()
364 if (!alloc) in base_region3_walk()
[all …]
/linux-6.12.1/tools/testing/selftests/bpf/progs/
Dverifier_kfunc_prog_types.c126 struct bpf_cpumask *alloc, *ref; in cpumask_kfunc_load_test() local
128 alloc = bpf_cpumask_create(); in cpumask_kfunc_load_test()
129 if (!alloc) in cpumask_kfunc_load_test()
132 ref = bpf_cpumask_acquire(alloc); in cpumask_kfunc_load_test()
133 bpf_cpumask_set_cpu(0, alloc); in cpumask_kfunc_load_test()
137 bpf_cpumask_release(alloc); in cpumask_kfunc_load_test()
/linux-6.12.1/net/core/
Dpage_pool.c399 pool->alloc.cache[pool->alloc.count++] = netmem; in page_pool_refill_alloc_cache()
411 } while (pool->alloc.count < PP_ALLOC_CACHE_REFILL); in page_pool_refill_alloc_cache()
414 if (likely(pool->alloc.count > 0)) { in page_pool_refill_alloc_cache()
415 netmem = pool->alloc.cache[--pool->alloc.count]; in page_pool_refill_alloc_cache()
428 if (likely(pool->alloc.count)) { in __page_pool_get_cached()
430 netmem = pool->alloc.cache[--pool->alloc.count]; in __page_pool_get_cached()
532 if (unlikely(pool->alloc.count > 0)) in __page_pool_alloc_pages_slow()
533 return pool->alloc.cache[--pool->alloc.count]; in __page_pool_alloc_pages_slow()
536 memset(&pool->alloc.cache, 0, sizeof(void *) * bulk); in __page_pool_alloc_pages_slow()
540 (struct page **)pool->alloc.cache); in __page_pool_alloc_pages_slow()
[all …]
/linux-6.12.1/arch/xtensa/variants/csp/include/variant/
Dtie-asm.h76 .macro xchal_ncp_store ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0
84 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0
96 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
116 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
142 .macro xchal_ncp_load ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0
150 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0
162 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
182 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
/linux-6.12.1/arch/xtensa/variants/dc233c/include/variant/
Dtie-asm.h77 .macro xchal_ncp_store ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0
85 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0
97 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
115 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
141 .macro xchal_ncp_load ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0
149 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0
161 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
179 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
/linux-6.12.1/fs/nfs/
Dnfs3acl.c260 struct posix_acl *orig = acl, *dfacl = NULL, *alloc; in nfs3_set_acl() local
267 alloc = get_inode_acl(inode, ACL_TYPE_DEFAULT); in nfs3_set_acl()
268 if (IS_ERR(alloc)) in nfs3_set_acl()
270 dfacl = alloc; in nfs3_set_acl()
274 alloc = get_inode_acl(inode, ACL_TYPE_ACCESS); in nfs3_set_acl()
275 if (IS_ERR(alloc)) in nfs3_set_acl()
278 acl = alloc; in nfs3_set_acl()
284 alloc = posix_acl_from_mode(inode->i_mode, GFP_KERNEL); in nfs3_set_acl()
285 if (IS_ERR(alloc)) in nfs3_set_acl()
287 acl = alloc; in nfs3_set_acl()
[all …]
/linux-6.12.1/mm/
Dmempool.c93 if (pool->alloc == mempool_kmalloc) { in poison_element()
95 } else if (pool->alloc == mempool_alloc_slab) { in poison_element()
97 } else if (pool->alloc == mempool_alloc_pages) { in poison_element()
117 if (pool->alloc == mempool_alloc_slab || pool->alloc == mempool_kmalloc) in kasan_poison_element()
119 else if (pool->alloc == mempool_alloc_pages) in kasan_poison_element()
127 if (pool->alloc == mempool_kmalloc) in kasan_unpoison_element()
129 else if (pool->alloc == mempool_alloc_slab) in kasan_unpoison_element()
132 else if (pool->alloc == mempool_alloc_pages) in kasan_unpoison_element()
202 pool->alloc = alloc_fn; in mempool_init_node()
217 element = pool->alloc(gfp_mask, pool->pool_data); in mempool_init_node()
[all …]
/linux-6.12.1/scripts/gdb/linux/
Dslab.py70 def get_track(cache, object_pointer, alloc): argument
72 p += (alloc * track_type.sizeof)
153 def slabtrace(alloc, cache_name): argument
164 def process_slab(loc_track, slab_list, alloc, cache): argument
172 p = get_track(cache, object_pointer, alloc)
174 if alloc == track_alloc:
202 process_slab(loc_track, cache_node['partial'], alloc, target_cache)
203 process_slab(loc_track, cache_node['full'], alloc, target_cache)
262 alloc = track_alloc # default show alloc_traces
266 alloc = track_alloc
[all …]

12345678910>>...24