Home
last modified time | relevance | path

Searched refs:alloc_flags (Results 1 – 25 of 28) sorted by relevance

12

/linux-6.12.1/mm/
Dpage_alloc.c1562 unsigned int alloc_flags) in prep_new_page() argument
1575 if (alloc_flags & ALLOC_NO_WATERMARKS) in prep_new_page()
1914 unsigned int alloc_flags, bool whole_block) in steal_suitable_fallback() argument
1945 if (boost_watermark(zone) && (alloc_flags & ALLOC_KSWAPD)) in steal_suitable_fallback()
2183 unsigned int alloc_flags) in __rmqueue_fallback() argument
2197 if (order < pageblock_order && alloc_flags & ALLOC_NOFRAGMENT) in __rmqueue_fallback()
2250 start_migratetype, alloc_flags, can_steal); in __rmqueue_fallback()
2264 unsigned int alloc_flags) in __rmqueue() argument
2274 if (alloc_flags & ALLOC_CMA && in __rmqueue()
2285 if (alloc_flags & ALLOC_CMA) in __rmqueue()
[all …]
Dcompaction.c2453 int alloc_flags) in compaction_zonelist_suitable() argument
2491 int highest_zoneidx, unsigned int alloc_flags) in compaction_suit_allocation_order() argument
2495 watermark = wmark_pages(zone, alloc_flags & ALLOC_WMARK_MASK); in compaction_suit_allocation_order()
2497 alloc_flags)) in compaction_suit_allocation_order()
2535 cc->alloc_flags); in compact_zone()
2745 unsigned int alloc_flags, int highest_zoneidx, in compact_zone_order() argument
2756 .alloc_flags = alloc_flags, in compact_zone_order()
2809 unsigned int alloc_flags, const struct alloc_context *ac, in try_to_compact_pages() argument
2827 (alloc_flags & ALLOC_CPUSET) && in try_to_compact_pages()
2838 alloc_flags, ac->highest_zoneidx, capture); in try_to_compact_pages()
/linux-6.12.1/drivers/base/regmap/
Dregcache-maple.c77 map->alloc_flags); in regcache_maple_write()
95 ret = mas_store_gfp(&mas, entry, map->alloc_flags); in regcache_maple_write()
138 map->alloc_flags); in regcache_maple_drop()
151 map->alloc_flags); in regcache_maple_drop()
165 ret = mas_store_gfp(&mas, lower, map->alloc_flags); in regcache_maple_drop()
173 ret = mas_store_gfp(&mas, upper, map->alloc_flags); in regcache_maple_drop()
207 buf = kmalloc(val_bytes * (max - min), map->alloc_flags); in regcache_maple_sync_block()
323 entry = kcalloc(last - first + 1, sizeof(unsigned long), map->alloc_flags); in regcache_maple_insert_block()
334 ret = mas_store_gfp(&mas, entry, map->alloc_flags); in regcache_maple_insert_block()
351 mt = kmalloc(sizeof(*mt), map->alloc_flags); in regcache_maple_init()
Dregcache-rbtree.c190 map->cache = kmalloc(sizeof *rbtree_ctx, map->alloc_flags); in regcache_rbtree_init()
280 map->alloc_flags); in regcache_rbtree_insert_to_block()
289 map->alloc_flags); in regcache_rbtree_insert_to_block()
323 rbnode = kzalloc(sizeof(*rbnode), map->alloc_flags); in regcache_rbtree_node_alloc()
349 map->alloc_flags); in regcache_rbtree_node_alloc()
355 map->alloc_flags); in regcache_rbtree_node_alloc()
Dregcache-flat.c30 + 1, sizeof(unsigned int), map->alloc_flags); in regcache_flat_init()
Dinternal.h65 gfp_t alloc_flags; member
Dregmap.c755 map->alloc_flags = GFP_ATOMIC; in __regmap_init()
757 map->alloc_flags = GFP_KERNEL; in __regmap_init()
2351 wval = kmemdup_array(val, val_count, val_bytes, map->alloc_flags); in regmap_bulk_write()
/linux-6.12.1/fs/xfs/libxfs/
Dxfs_alloc.c1677 uint32_t alloc_flags) in xfs_alloc_ag_vextent_near() argument
1697 alloc_flags |= XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_near()
1762 acur.busy_gen, alloc_flags); in xfs_alloc_ag_vextent_near()
1766 alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_near()
1792 uint32_t alloc_flags) in xfs_alloc_ag_vextent_size() argument
1807 alloc_flags |= XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_size()
1878 busy_gen, alloc_flags); in xfs_alloc_ag_vextent_size()
1882 alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_size()
1977 busy_gen, alloc_flags); in xfs_alloc_ag_vextent_size()
1981 alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH; in xfs_alloc_ag_vextent_size()
[all …]
Dxfs_alloc.h199 int xfs_alloc_fix_freelist(struct xfs_alloc_arg *args, uint32_t alloc_flags);
/linux-6.12.1/drivers/gpu/drm/amd/amdgpu/
Damdgpu_amdkfd_gpuvm.c296 u32 alloc_flags = bo->kfd_bo->alloc_flags; in amdgpu_amdkfd_release_notify() local
299 amdgpu_amdkfd_unreserve_mem_limit(adev, size, alloc_flags, in amdgpu_amdkfd_release_notify()
326 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) in create_dmamap_sg_bo()
510 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE) in get_pte_flags()
512 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_EXECUTABLE) in get_pte_flags()
555 mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmamap_userptr()
657 mmio = (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP); in kfd_mem_dmamap_sg_bo()
663 dir = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmamap_sg_bo()
722 mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmaunmap_userptr()
784 dir = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmaunmap_sg_bo()
[all …]
Damdgpu_amdkfd.h81 uint32_t alloc_flags; member
/linux-6.12.1/include/linux/
Dcompaction.h88 unsigned int order, unsigned int alloc_flags,
99 int alloc_flags);
Dmmzone.h1457 int highest_zoneidx, unsigned int alloc_flags,
1461 unsigned int alloc_flags);
/linux-6.12.1/lib/
Dstackdepot.c586 gfp_t alloc_flags, in stack_depot_save_flags() argument
627 page = alloc_pages(gfp_nested_mask(alloc_flags), in stack_depot_save_flags()
676 gfp_t alloc_flags) in stack_depot_save() argument
678 return stack_depot_save_flags(entries, nr_entries, alloc_flags, in stack_depot_save()
/linux-6.12.1/drivers/md/
Ddm-zoned-reclaim.c286 int alloc_flags = DMZ_ALLOC_SEQ; in dmz_reclaim_rnd_data() local
292 alloc_flags | DMZ_ALLOC_RECLAIM); in dmz_reclaim_rnd_data()
293 if (!szone && alloc_flags == DMZ_ALLOC_SEQ && dmz_nr_cache_zones(zmd)) { in dmz_reclaim_rnd_data()
294 alloc_flags = DMZ_ALLOC_RND; in dmz_reclaim_rnd_data()
Ddm-zoned-metadata.c2056 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; in dmz_get_chunk_mapping() local
2071 dzone = dmz_alloc_zone(zmd, 0, alloc_flags); in dmz_get_chunk_mapping()
2168 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; in dmz_get_chunk_buffer() local
2177 bzone = dmz_alloc_zone(zmd, 0, alloc_flags); in dmz_get_chunk_buffer()
/linux-6.12.1/drivers/gpu/drm/amd/amdkfd/
Dkfd_chardev.c1937 bo_bucket->alloc_flags = (uint32_t)kgd_mem->alloc_flags; in criu_checkpoint_bos()
1940 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) { in criu_checkpoint_bos()
1948 if (bo_bucket->alloc_flags in criu_checkpoint_bos()
1951 bo_bucket->alloc_flags & in criu_checkpoint_bos()
1960 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL) in criu_checkpoint_bos()
1963 else if (bo_bucket->alloc_flags & in criu_checkpoint_bos()
1981 bo_bucket->alloc_flags, in criu_checkpoint_bos()
2296 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL) { in criu_restore_memory_of_gpu()
2304 } else if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP) { in criu_restore_memory_of_gpu()
2315 } else if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) { in criu_restore_memory_of_gpu()
[all …]
/linux-6.12.1/fs/xfs/
Dxfs_extent_busy.c602 uint32_t alloc_flags) in xfs_extent_busy_flush() argument
613 if (alloc_flags & XFS_ALLOC_FLAG_TRYFLUSH) in xfs_extent_busy_flush()
619 if (alloc_flags & XFS_ALLOC_FLAG_FREEING) in xfs_extent_busy_flush()
Dxfs_extent_busy.h74 unsigned busy_gen, uint32_t alloc_flags);
/linux-6.12.1/fs/xfs/scrub/
Drepair.h54 int xrep_fix_freelist(struct xfs_scrub *sc, int alloc_flags);
Drepair.c411 int alloc_flags) in xrep_fix_freelist() argument
421 return xfs_alloc_fix_freelist(&args, alloc_flags); in xrep_fix_freelist()
/linux-6.12.1/fs/btrfs/
Dblock-group.c2940 u64 alloc_flags; in btrfs_inc_block_group_ro() local
2988 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro()
2989 if (alloc_flags != cache->flags) { in btrfs_inc_block_group_ro()
2990 ret = btrfs_chunk_alloc(trans, alloc_flags, in btrfs_inc_block_group_ro()
3018 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->space_info->flags); in btrfs_inc_block_group_ro()
3019 ret = btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_inc_block_group_ro()
3035 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro()
3037 check_system_chunk(trans, alloc_flags); in btrfs_inc_block_group_ro()
3875 u64 alloc_flags = btrfs_get_alloc_profile(trans->fs_info, type); in btrfs_force_chunk_alloc() local
3877 return btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_force_chunk_alloc()
/linux-6.12.1/drivers/iommu/
Ddma-iommu.c915 gfp_t alloc_flags = gfp; in __iommu_dma_alloc_pages() local
919 alloc_flags |= __GFP_NORETRY; in __iommu_dma_alloc_pages()
920 page = alloc_pages_node(nid, alloc_flags, order); in __iommu_dma_alloc_pages()
/linux-6.12.1/drivers/gpu/drm/i915/gem/
Di915_gem_object.h49 unsigned alloc_flags);
/linux-6.12.1/include/uapi/linux/
Dkfd_ioctl.h715 __u32 alloc_flags; member

12