Searched refs:last_alloc (Results 1 – 5 of 5) sorted by relevance
/linux-6.12.1/mm/ |
D | percpu-stats.c | 56 int i, last_alloc, as_len, start, end; in chunk_map_stats() local 69 last_alloc = find_last_bit(chunk->alloc_map, in chunk_map_stats() 72 last_alloc = test_bit(last_alloc, chunk->alloc_map) ? in chunk_map_stats() 73 last_alloc + 1 : 0; in chunk_map_stats() 87 while (start < last_alloc) { in chunk_map_stats() 89 end = find_next_bit(chunk->bound_map, last_alloc, in chunk_map_stats() 93 end = find_next_bit(chunk->alloc_map, last_alloc, in chunk_map_stats()
|
/linux-6.12.1/fs/nilfs2/ |
D | sufile.c | 315 __u64 segnum, maxsegnum, last_alloc; in nilfs_sufile_alloc() local 327 last_alloc = le64_to_cpu(header->sh_last_alloc); in nilfs_sufile_alloc() 332 segnum = last_alloc + 1; in nilfs_sufile_alloc() 345 maxsegnum = last_alloc; in nilfs_sufile_alloc()
|
/linux-6.12.1/fs/btrfs/ |
D | zoned.c | 1553 u64 last_alloc = 0; in btrfs_load_block_group_zone_info() local 1603 ret = calculate_alloc_pointer(cache, &last_alloc, new); in btrfs_load_block_group_zone_info() 1610 cache->alloc_offset = last_alloc; in btrfs_load_block_group_zone_info() 1680 if (!ret && num_conventional && last_alloc > cache->alloc_offset) { in btrfs_load_block_group_zone_info() 1683 logical, last_alloc, cache->alloc_offset); in btrfs_load_block_group_zone_info()
|
D | inode.c | 8774 u64 last_alloc = (u64)-1; in __btrfs_prealloc_file_range() local 8790 cur_bytes = min(cur_bytes, last_alloc); in __btrfs_prealloc_file_range() 8805 last_alloc = ins.offset; in __btrfs_prealloc_file_range()
|
/linux-6.12.1/tools/perf/ |
D | builtin-kmem.c | 66 u64 last_alloc; member 127 data->last_alloc = bytes_alloc; in insert_alloc_stat() 255 total_freed += s_alloc->last_alloc; in evsel__process_free_event()
|