/linux-6.12.1/drivers/gpu/drm/ |
D | drm_mm.c | 336 u64 hole_start; in find_hole_addr() local 342 hole_start = __drm_mm_hole_node_start(node); in find_hole_addr() 344 if (addr < hole_start) in find_hole_addr() 346 else if (addr > hole_start + node->hole_size) in find_hole_addr() 453 u64 hole_start, hole_end; in drm_mm_reserve_node() local 466 adj_start = hole_start = __drm_mm_hole_node_start(hole); in drm_mm_reserve_node() 467 adj_end = hole_end = hole_start + hole->hole_size; in drm_mm_reserve_node() 483 if (node->start > hole_start) in drm_mm_reserve_node() 543 u64 hole_start = __drm_mm_hole_node_start(hole); in drm_mm_insert_node_in_range() local 544 u64 hole_end = hole_start + hole->hole_size; in drm_mm_insert_node_in_range() [all …]
|
/linux-6.12.1/drivers/gpu/drm/i915/selftests/ |
D | i915_gem_gtt.c | 245 u64 hole_start, u64 hole_end, in lowlevel_hole() argument 259 for (size = 12; (hole_end - hole_start) >> size; size++) { in lowlevel_hole() 266 hole_size = (hole_end - hole_start) >> aligned_size; in lowlevel_hole() 272 __func__, hole_start, hole_end, size, hole_size); in lowlevel_hole() 288 GEM_BUG_ON(hole_start + count * BIT_ULL(aligned_size) > hole_end); in lowlevel_hole() 311 u64 addr = hole_start + order[n] * BIT_ULL(aligned_size); in lowlevel_hole() 319 hole_end = hole_start; /* quit */ in lowlevel_hole() 370 u64 addr = hole_start + order[n] * BIT_ULL(aligned_size); in lowlevel_hole() 409 u64 hole_start, u64 hole_end, in fill_hole() argument 412 const u64 hole_size = hole_end - hole_start; in fill_hole() [all …]
|
/linux-6.12.1/drivers/gpu/drm/xe/ |
D | xe_ggtt.c | 713 u64 hole_start, hole_end, hole_size; in xe_ggtt_largest_hole() local 718 drm_mm_for_each_hole(entry, mm, hole_start, hole_end) { in xe_ggtt_largest_hole() 719 hole_start = max(hole_start, hole_min_start); in xe_ggtt_largest_hole() 720 hole_start = ALIGN(hole_start, alignment); in xe_ggtt_largest_hole() 722 if (hole_start >= hole_end) in xe_ggtt_largest_hole() 724 hole_size = hole_end - hole_start; in xe_ggtt_largest_hole() 813 u64 hole_start, hole_end, hole_size; in xe_ggtt_print_holes() local 819 drm_mm_for_each_hole(entry, mm, hole_start, hole_end) { in xe_ggtt_print_holes() 820 hole_start = max(hole_start, hole_min_start); in xe_ggtt_print_holes() 821 hole_start = ALIGN(hole_start, alignment); in xe_ggtt_print_holes() [all …]
|
/linux-6.12.1/drivers/gpu/drm/tests/ |
D | drm_mm_test.c | 41 u64 hole_start, __always_unused hole_end; in assert_no_holes() local 45 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) in assert_no_holes() 66 u64 hole_start, hole_end; in assert_one_hole() local 74 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) { in assert_one_hole() 75 if (start != hole_start || end != hole_end) { in assert_one_hole() 79 hole_start, hole_end, start, end); in assert_one_hole()
|
/linux-6.12.1/arch/sparc/kernel/ |
D | adi_64.c | 163 unsigned long end_addr, hole_start, hole_end; in alloc_tag_store() local 167 hole_start = 0; in alloc_tag_store() 198 (tag_desc->end > hole_start)) in alloc_tag_store() 199 hole_start = tag_desc->end; in alloc_tag_store() 273 if (tmp_addr < hole_start) { in alloc_tag_store() 277 tmp_addr = hole_start + 1; in alloc_tag_store()
|
/linux-6.12.1/include/drm/ |
D | drm_mm.h | 394 #define drm_mm_for_each_hole(pos, mm, hole_start, hole_end) \ argument 398 hole_start = drm_mm_hole_node_start(pos), \ 399 hole_end = hole_start + pos->hole_size, \
|
/linux-6.12.1/kernel/ |
D | kexec_core.c | 409 unsigned long hole_start, hole_end, size; in kimage_alloc_crash_control_pages() local 414 hole_start = ALIGN(image->control_page, size); in kimage_alloc_crash_control_pages() 415 hole_end = hole_start + size - 1; in kimage_alloc_crash_control_pages() 429 if ((hole_end >= mstart) && (hole_start <= mend)) { in kimage_alloc_crash_control_pages() 431 hole_start = ALIGN(mend, size); in kimage_alloc_crash_control_pages() 432 hole_end = hole_start + size - 1; in kimage_alloc_crash_control_pages() 438 pages = pfn_to_page(hole_start >> PAGE_SHIFT); in kimage_alloc_crash_control_pages()
|
/linux-6.12.1/fs/bcachefs/ |
D | fs-io.c | 596 u64 hole_start, hole_end; in __bchfs_fallocate() local 615 hole_start = iter.pos.offset; in __bchfs_fallocate() 638 &hole_start, in __bchfs_fallocate() 643 &hole_start, in __bchfs_fallocate() 649 bch2_btree_iter_set_pos(&iter, POS(iter.pos.inode, hole_start)); in __bchfs_fallocate() 654 if (hole_start == hole_end) in __bchfs_fallocate() 658 sectors = hole_end - hole_start; in __bchfs_fallocate() 675 if (bch2_mark_pagecache_reserved(inode, &hole_start, in __bchfs_fallocate() 678 bch2_mark_pagecache_reserved(inode, &hole_start, in __bchfs_fallocate()
|
D | fs-io-pagecache.c | 797 u64 *hole_start, in bch2_clamp_data_hole() argument 805 *hole_start << 9, *hole_end << 9, min_replicas, nonblock) >> 9; in bch2_clamp_data_hole() 809 *hole_start = ret; in bch2_clamp_data_hole() 811 if (*hole_start == *hole_end) in bch2_clamp_data_hole() 815 *hole_start << 9, *hole_end << 9, min_replicas, nonblock) >> 9; in bch2_clamp_data_hole()
|
D | alloc_background.c | 1104 struct bpos hole_start = bkey_start_pos(k.k); in bch2_get_key_or_real_bucket_hole() local 1106 if (!*ca || !bucket_valid(*ca, hole_start.offset)) { in bch2_get_key_or_real_bucket_hole() 1107 if (!next_bucket(c, ca, &hole_start)) in bch2_get_key_or_real_bucket_hole() 1110 bch2_btree_iter_set_pos(iter, hole_start); in bch2_get_key_or_real_bucket_hole() 1115 bch2_key_resize(hole, (*ca)->mi.nbuckets - hole_start.offset); in bch2_get_key_or_real_bucket_hole()
|
/linux-6.12.1/fs/btrfs/ |
D | zoned.h | 69 u64 btrfs_find_allocatable_zones(struct btrfs_device *device, u64 hole_start, 164 u64 hole_start, u64 hole_end, in btrfs_find_allocatable_zones() argument 167 return hole_start; in btrfs_find_allocatable_zones()
|
D | volumes.c | 1492 u64 *hole_start, u64 *hole_size, in dev_extent_hole_check_zoned() argument 1500 ASSERT(IS_ALIGNED(*hole_start, zone_size)); in dev_extent_hole_check_zoned() 1503 pos = btrfs_find_allocatable_zones(device, *hole_start, in dev_extent_hole_check_zoned() 1504 *hole_start + *hole_size, in dev_extent_hole_check_zoned() 1506 if (pos != *hole_start) { in dev_extent_hole_check_zoned() 1507 *hole_size = *hole_start + *hole_size - pos; in dev_extent_hole_check_zoned() 1508 *hole_start = pos; in dev_extent_hole_check_zoned() 1522 *hole_start += *hole_size; in dev_extent_hole_check_zoned() 1527 *hole_start += zone_size; in dev_extent_hole_check_zoned() 1546 static bool dev_extent_hole_check(struct btrfs_device *device, u64 *hole_start, in dev_extent_hole_check() argument [all …]
|
D | zoned.c | 1030 u64 btrfs_find_allocatable_zones(struct btrfs_device *device, u64 hole_start, in btrfs_find_allocatable_zones() argument 1036 u64 pos = hole_start; in btrfs_find_allocatable_zones() 1041 ASSERT(IS_ALIGNED(hole_start, zinfo->zone_size)); in btrfs_find_allocatable_zones()
|
D | inode.c | 4926 u64 hole_start = ALIGN(oldsize, fs_info->sectorsize); in btrfs_cont_expand() local 4942 if (size <= hole_start) in btrfs_cont_expand() 4945 btrfs_lock_and_flush_ordered_range(inode, hole_start, block_end - 1, in btrfs_cont_expand() 4947 cur_offset = hole_start; in btrfs_cont_expand() 5003 unlock_extent(io_tree, hole_start, block_end - 1, &cached_state); in btrfs_cont_expand()
|
/linux-6.12.1/fs/hugetlbfs/ |
D | inode.c | 752 loff_t hole_start, hole_end; in hugetlbfs_punch_hole() local 757 hole_start = round_up(offset, hpage_size); in hugetlbfs_punch_hole() 771 if (offset < hole_start) in hugetlbfs_punch_hole() 773 offset, min(offset + len, hole_start)); in hugetlbfs_punch_hole() 776 if (hole_end > hole_start) { in hugetlbfs_punch_hole() 779 hole_start >> PAGE_SHIFT, in hugetlbfs_punch_hole() 784 if ((offset + len) > hole_end && (offset + len) > hole_start) in hugetlbfs_punch_hole() 791 if (hole_end > hole_start) in hugetlbfs_punch_hole() 792 remove_inode_hugepages(inode, hole_start, hole_end); in hugetlbfs_punch_hole()
|
/linux-6.12.1/drivers/gpu/drm/i915/gt/ |
D | intel_ggtt.c | 827 unsigned long hole_start, hole_end; in init_ggtt() local 898 drm_mm_for_each_hole(entry, &ggtt->vm.mm, hole_start, hole_end) { in init_ggtt() 901 hole_start, hole_end); in init_ggtt() 902 ggtt->vm.clear_range(&ggtt->vm, hole_start, in init_ggtt() 903 hole_end - hole_start); in init_ggtt()
|
/linux-6.12.1/fs/ext4/ |
D | extents.c | 4113 ext4_lblk_t hole_start, len; in ext4_ext_determine_insert_hole() local 4116 hole_start = lblk; in ext4_ext_determine_insert_hole() 4117 len = ext4_ext_find_hole(inode, path, &hole_start); in ext4_ext_determine_insert_hole() 4119 ext4_es_find_extent_range(inode, &ext4_es_is_delayed, hole_start, in ext4_ext_determine_insert_hole() 4120 hole_start + len - 1, &es); in ext4_ext_determine_insert_hole() 4133 len -= lblk - hole_start; in ext4_ext_determine_insert_hole() 4134 hole_start = lblk; in ext4_ext_determine_insert_hole() 4154 len = min(es.es_lblk - hole_start, len); in ext4_ext_determine_insert_hole() 4159 ext_debug(inode, " -> %u:%u\n", hole_start, len); in ext4_ext_determine_insert_hole() 4160 ext4_es_insert_extent(inode, hole_start, len, ~0, in ext4_ext_determine_insert_hole() [all …]
|