/linux-6.12.1/arch/arm/mm/ |
D | cache-feroceon-l2.c | 141 unsigned long range_end; in calc_range_end() local 149 range_end = end; in calc_range_end() 156 if (range_end > start + MAX_RANGE_SIZE) in calc_range_end() 157 range_end = start + MAX_RANGE_SIZE; in calc_range_end() 162 if (range_end > (start | (PAGE_SIZE - 1)) + 1) in calc_range_end() 163 range_end = (start | (PAGE_SIZE - 1)) + 1; in calc_range_end() 165 return range_end; in calc_range_end() 190 unsigned long range_end = calc_range_end(start, end); in feroceon_l2_inv_range() local 191 l2_inv_pa_range(start, range_end - CACHE_LINE_SIZE); in feroceon_l2_inv_range() 192 start = range_end; in feroceon_l2_inv_range() [all …]
|
D | pageattr.c | 29 unsigned long range_start, unsigned long range_end) in range_in_range() argument 31 return start >= range_start && start < range_end && in range_in_range() 32 size <= range_end - start; in range_in_range()
|
D | init.c | 244 phys_addr_t range_start, range_end; in free_highpages() local 249 &range_start, &range_end, NULL) { in free_highpages() 251 unsigned long end = PFN_DOWN(range_end); in free_highpages()
|
/linux-6.12.1/tools/testing/selftests/kvm/include/ |
D | sparsebit.h | 82 #define sparsebit_for_each_set_range(s, range_begin, range_end) \ argument 84 range_end = sparsebit_next_clear(s, range_begin) - 1; \ 85 range_begin && range_end; \ 86 range_begin = sparsebit_next_set(s, range_end), \ 87 range_end = sparsebit_next_clear(s, range_begin) - 1)
|
/linux-6.12.1/fs/btrfs/ |
D | fiemap.c | 179 const u64 range_end = offset + len; in emit_fiemap_extent() local 209 if (range_end <= cache_end) in emit_fiemap_extent() 216 len = range_end - cache_end; in emit_fiemap_extent() 643 u64 range_end; in extent_fiemap() local 661 range_end = round_up(start + len, sectorsize); in extent_fiemap() 664 lock_extent(&inode->io_tree, range_start, range_end, &cached_state); in extent_fiemap() 684 while (prev_extent_end < range_end) { in extent_fiemap() 714 const u64 hole_end = min(key.offset, range_end) - 1; in extent_fiemap() 729 if (key.offset >= range_end) { in extent_fiemap() 813 if (!stopped && prev_extent_end < range_end) { in extent_fiemap() [all …]
|
D | compression.h | 86 static inline u32 btrfs_calc_input_length(u64 range_end, u64 cur) in btrfs_calc_input_length() argument 90 return min(range_end, page_end) - cur; in btrfs_calc_input_length()
|
/linux-6.12.1/drivers/firmware/efi/ |
D | unaccepted_memory.c | 36 unsigned long range_start, range_end; in accept_memory() local 127 for_each_set_bitrange_from(range_start, range_end, unaccepted->bitmap, in accept_memory() 130 unsigned long len = range_end - range_start; in accept_memory() 133 phys_end = range_end * unit_size + unaccepted->phys_base; in accept_memory()
|
/linux-6.12.1/net/bridge/ |
D | br_vlan_options.c | 33 const struct net_bridge_vlan *range_end) in __vlan_tun_can_enter_range() argument 35 return (!v_curr->tinfo.tunnel_dst && !range_end->tinfo.tunnel_dst) || in __vlan_tun_can_enter_range() 36 vlan_tunid_inrange(v_curr, range_end); in __vlan_tun_can_enter_range() 41 const struct net_bridge_vlan *range_end) in br_vlan_opts_eq_range() argument 43 u8 range_mc_rtr = br_vlan_multicast_router(range_end); in br_vlan_opts_eq_range() 46 return v_curr->state == range_end->state && in br_vlan_opts_eq_range() 47 __vlan_tun_can_enter_range(v_curr, range_end) && in br_vlan_opts_eq_range() 266 struct net_bridge_vlan *range_end, in br_vlan_process_options() argument 284 if (!range_end || !br_vlan_should_use(range_end)) { in br_vlan_process_options() 290 for (vid = range_start->vid; vid <= range_end->vid; vid++) { in br_vlan_process_options()
|
D | br_vlan.c | 1944 const struct net_bridge_vlan *range_end) in br_vlan_can_enter_range() argument 1946 return v_curr->vid - range_end->vid == 1 && in br_vlan_can_enter_range() 1947 range_end->flags == v_curr->flags && in br_vlan_can_enter_range() 1948 br_vlan_opts_eq_range(v_curr, range_end); in br_vlan_can_enter_range() 1956 struct net_bridge_vlan *v, *range_start = NULL, *range_end = NULL; in br_vlan_dump_dev() local 2011 range_end = v; in br_vlan_dump_dev() 2016 if (br_vlan_global_opts_can_enter_range(v, range_end)) in br_vlan_dump_dev() 2019 range_end->vid, in br_vlan_dump_dev() 2025 idx += range_end->vid - range_start->vid + 1; in br_vlan_dump_dev() 2029 !br_vlan_can_enter_range(v, range_end)) { in br_vlan_dump_dev() [all …]
|
/linux-6.12.1/drivers/firmware/efi/libstub/ |
D | unaccepted_memory.c | 182 unsigned long range_start, range_end; in accept_memory() local 212 for_each_set_bitrange_from(range_start, range_end, in accept_memory() 217 phys_end = range_end * unit_size + unaccepted_table->phys_base; in accept_memory() 221 range_start, range_end - range_start); in accept_memory()
|
/linux-6.12.1/drivers/infiniband/hw/hfi1/ |
D | fault.c | 114 unsigned long range_start, range_end, i; in fault_opcodes_write() local 132 if (kstrtoul(token, 0, &range_end)) in fault_opcodes_write() 135 range_end = range_start; in fault_opcodes_write() 137 if (range_start == range_end && range_start == -1UL) { in fault_opcodes_write() 143 if (range_start >= bound || range_end >= bound) in fault_opcodes_write() 146 for (i = range_start; i <= range_end; i++) { in fault_opcodes_write()
|
/linux-6.12.1/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_doorbell.c | 213 int range_end = dev->shared_resources.non_cp_doorbells_end; in init_doorbell_bitmap() local 219 pr_debug("reserved doorbell 0x%03x - 0x%03x\n", range_start, range_end); in init_doorbell_bitmap() 222 range_end + KFD_QUEUE_DOORBELL_MIRROR_OFFSET); in init_doorbell_bitmap() 225 if (i >= range_start && i <= range_end) { in init_doorbell_bitmap()
|
D | kfd_device.c | 1278 (*mem_obj)->range_end = found; in kfd_gtt_sa_allocate() 1301 (*mem_obj)->range_end = in kfd_gtt_sa_allocate() 1309 if ((*mem_obj)->range_end != found) { in kfd_gtt_sa_allocate() 1329 (*mem_obj)->range_start, (*mem_obj)->range_end); in kfd_gtt_sa_allocate() 1333 (*mem_obj)->range_end - (*mem_obj)->range_start + 1); in kfd_gtt_sa_allocate() 1355 mem_obj, mem_obj->range_start, mem_obj->range_end); in kfd_gtt_sa_free() 1361 mem_obj->range_end - mem_obj->range_start + 1); in kfd_gtt_sa_free()
|
/linux-6.12.1/drivers/gpu/drm/xe/ |
D | xe_reg_whitelist.c | 122 u32 range_start, range_end; in xe_reg_whitelist_print_entry() local 140 range_end = range_start | REG_GENMASK(range_bit, 0); in xe_reg_whitelist_print_entry() 155 range_start, range_end, in xe_reg_whitelist_print_entry()
|
/linux-6.12.1/drivers/gpu/drm/ |
D | drm_mm.c | 518 u64 range_start, u64 range_end, in drm_mm_insert_node_in_range() argument 525 DRM_MM_BUG_ON(range_start > range_end); in drm_mm_insert_node_in_range() 527 if (unlikely(size == 0 || range_end - range_start < size)) in drm_mm_insert_node_in_range() 540 for (hole = first_hole(mm, range_start, range_end, size, mode); in drm_mm_insert_node_in_range() 548 if (mode == DRM_MM_INSERT_LOW && hole_start >= range_end) in drm_mm_insert_node_in_range() 560 adj_end = min(col_end, range_end); in drm_mm_insert_node_in_range() 581 min(col_end, range_end) - adj_start < size) in drm_mm_insert_node_in_range() 727 scan->range_end = end; in drm_mm_scan_init_with_range() 778 adj_end = min(col_end, scan->range_end); in drm_mm_scan_add_block() 797 min(col_end, scan->range_end) - adj_start < scan->size) in drm_mm_scan_add_block()
|
/linux-6.12.1/arch/xtensa/mm/ |
D | init.c | 86 phys_addr_t range_start, range_end; in free_highpages() local 91 &range_start, &range_end, NULL) { in free_highpages() 93 unsigned long end = PFN_DOWN(range_end); in free_highpages()
|
/linux-6.12.1/arch/arm64/kvm/hyp/nvhe/ |
D | page_alloc.c | 45 if (addr < pool->range_start || addr >= pool->range_end) in __find_buddy_nocheck() 103 if (phys < pool->range_start || phys >= pool->range_end) in __hyp_attach_page() 236 pool->range_end = phys + (nr_pages << PAGE_SHIFT); in hyp_pool_init()
|
/linux-6.12.1/arch/powerpc/platforms/powernv/ |
D | opal-prd.c | 64 uint64_t range_addr, range_size, range_end; in opal_prd_range_is_valid() local 73 range_end = range_addr + range_size; in opal_prd_range_is_valid() 81 if (range_end <= range_addr) in opal_prd_range_is_valid() 84 if (addr >= range_addr && addr + size <= range_end) { in opal_prd_range_is_valid()
|
/linux-6.12.1/drivers/gpu/drm/amd/display/dc/dml2/dml21/src/dml2_top/ |
D | dml_top_mcache.c | 29 int range_end; in calculate_first_second_splitting() local 42 range_end = mcache_boundaries[left_cache_id] - shift - 1; in calculate_first_second_splitting() 44 if (range_start <= pipe_h_vp_start && pipe_h_vp_start <= range_end) in calculate_first_second_splitting() 47 range_start = range_end + 1; in calculate_first_second_splitting() 50 range_end = MAX_VP; in calculate_first_second_splitting() 57 if (range_start <= pipe_h_vp_end && pipe_h_vp_end <= range_end) { in calculate_first_second_splitting() 60 range_end = range_start - 1; in calculate_first_second_splitting()
|
/linux-6.12.1/arch/s390/boot/ |
D | physmem_info.c | 252 unsigned long range_start, range_end; in __physmem_alloc_range() local 258 __get_physmem_range(nranges - 1, &range_start, &range_end, false); in __physmem_alloc_range() 259 pos = min(range_end, pos); in __physmem_alloc_range()
|
/linux-6.12.1/drivers/gpu/drm/msm/ |
D | msm_gem_vma.c | 125 u64 range_start, u64 range_end) in msm_gem_vma_init() argument 139 range_start, range_end, 0); in msm_gem_vma_init()
|
D | msm_gem.c | 481 u64 range_start, u64 range_end) in get_vma_locked() argument 497 range_start, range_end); in get_vma_locked() 504 GEM_WARN_ON((vma->iova + obj->size) > range_end); in get_vma_locked() 571 u64 range_start, u64 range_end) in get_and_pin_iova_range_locked() argument 578 vma = get_vma_locked(obj, aspace, range_start, range_end); in get_and_pin_iova_range_locked() 597 u64 range_start, u64 range_end) in msm_gem_get_and_pin_iova_range() argument 602 ret = get_and_pin_iova_range_locked(obj, aspace, iova, range_start, range_end); in msm_gem_get_and_pin_iova_range()
|
D | msm_gem.h | 71 u64 range_start, u64 range_end); 137 u64 range_start, u64 range_end);
|
/linux-6.12.1/arch/arm64/kvm/hyp/include/nvhe/ |
D | gfp.h | 21 phys_addr_t range_end; member
|
/linux-6.12.1/fs/ocfs2/ |
D | file.c | 861 u64 *range_start, u64 *range_end) in ocfs2_zero_extend_get_range() argument 889 *range_end = 0; in ocfs2_zero_extend_get_range() 921 *range_end = ocfs2_clusters_to_bytes(inode->i_sb, in ocfs2_zero_extend_get_range() 933 u64 range_end, struct buffer_head *di_bh) in ocfs2_zero_extend_range() argument 942 (unsigned long long)range_end); in ocfs2_zero_extend_range() 943 BUG_ON(range_start >= range_end); in ocfs2_zero_extend_range() 945 while (zero_pos < range_end) { in ocfs2_zero_extend_range() 947 if (next_pos > range_end) in ocfs2_zero_extend_range() 948 next_pos = range_end; in ocfs2_zero_extend_range() 970 u64 zero_start, range_start = 0, range_end = 0; in ocfs2_zero_extend() local [all …]
|