/linux-6.12.1/tools/testing/vma/linux/ |
D | mmzone.h | 21 #define MAX_PAGE_ORDER 10 macro 22 #define MAX_ORDER_NR_PAGES (1 << MAX_PAGE_ORDER) 24 #define pageblock_order MAX_PAGE_ORDER
|
/linux-6.12.1/tools/testing/memblock/linux/ |
D | mmzone.h | 21 #define MAX_PAGE_ORDER 10 macro 22 #define MAX_ORDER_NR_PAGES (1 << MAX_PAGE_ORDER) 24 #define pageblock_order MAX_PAGE_ORDER
|
/linux-6.12.1/include/linux/ |
D | pageblock-flags.h | 44 #define pageblock_order MIN_T(unsigned int, HUGETLB_PAGE_ORDER, MAX_PAGE_ORDER) 50 #define pageblock_order MIN_T(unsigned int, HPAGE_PMD_ORDER, MAX_PAGE_ORDER) 55 #define pageblock_order MAX_PAGE_ORDER
|
D | mmzone.h | 30 #define MAX_PAGE_ORDER 10 macro 32 #define MAX_PAGE_ORDER CONFIG_ARCH_FORCE_MAX_ORDER macro 34 #define MAX_ORDER_NR_PAGES (1 << MAX_PAGE_ORDER) 38 #define NR_PAGE_ORDERS (MAX_PAGE_ORDER + 1) 1777 #if (MAX_PAGE_ORDER + PAGE_SHIFT) > SECTION_SIZE_BITS 1778 #error Allocator MAX_PAGE_ORDER exceeds SECTION_SIZE
|
/linux-6.12.1/mm/kmsan/ |
D | init.c | 147 .order = MAX_PAGE_ORDER, 217 collect.order = MAX_PAGE_ORDER; in kmsan_memblock_discard() 218 for (int i = MAX_PAGE_ORDER; i >= 0; i--) { in kmsan_memblock_discard()
|
/linux-6.12.1/drivers/gpu/drm/ttm/tests/ |
D | ttm_pool_test.c | 111 .order = MAX_PAGE_ORDER + 1, 120 .order = MAX_PAGE_ORDER + 1, 167 if (params->order <= MAX_PAGE_ORDER) { in ttm_pool_alloc_basic() 184 min_t(unsigned int, MAX_PAGE_ORDER, in ttm_pool_alloc_basic()
|
/linux-6.12.1/mm/ |
D | shuffle.h | 7 #define SHUFFLE_ORDER MAX_PAGE_ORDER
|
D | debug_page_alloc.c | 25 if (kstrtoul(buf, 10, &res) < 0 || res > MAX_PAGE_ORDER / 2) { in debug_guardpage_minorder_setup()
|
D | page_reporting.c | 23 return param_set_uint_minmax(val, kp, 0, MAX_PAGE_ORDER); in page_order_update_notify() 373 if (prdev->order > 0 && prdev->order <= MAX_PAGE_ORDER) in page_reporting_register()
|
D | page_owner.c | 456 if (freepage_order <= MAX_PAGE_ORDER) in pagetypeinfo_showmixedcount_print() 691 if (freepage_order <= MAX_PAGE_ORDER) in read_page_owner() 799 if (order > 0 && order <= MAX_PAGE_ORDER) in init_pages_in_zone()
|
D | page_alloc.c | 737 if (order >= MAX_PAGE_ORDER - 1) in buddy_merge_likely() 791 while (order < MAX_PAGE_ORDER) { in __free_one_page() 1314 if (order == MAX_PAGE_ORDER && __free_unaccepted(page)) in __free_pages_core() 1745 if (++order > MAX_PAGE_ORDER) in find_large_buddy() 1788 if (pageblock_order == MAX_PAGE_ORDER) in move_freepages_block_isolate() 2205 for (current_order = MAX_PAGE_ORDER; current_order >= min_order; in __rmqueue_fallback() 2243 VM_BUG_ON(current_order > MAX_PAGE_ORDER); in __rmqueue_fallback() 4727 if (WARN_ON_ONCE_GFP(order > MAX_PAGE_ORDER, gfp)) in __alloc_pages_noprof() 6885 return order <= MAX_PAGE_ORDER; in is_free_buddy_page() 7046 accept_memory(page_to_phys(page), PAGE_SIZE << MAX_PAGE_ORDER); in __accept_page() [all …]
|
D | page_isolation.c | 228 if (order >= pageblock_order && order < MAX_PAGE_ORDER) { in unset_migratetype_isolate()
|
/linux-6.12.1/drivers/base/regmap/ |
D | regmap-debugfs.c | 229 if (count > (PAGE_SIZE << MAX_PAGE_ORDER)) in regmap_read_debugfs() 230 count = PAGE_SIZE << MAX_PAGE_ORDER; in regmap_read_debugfs() 376 if (count > (PAGE_SIZE << MAX_PAGE_ORDER)) in regmap_reg_ranges_read_file() 377 count = PAGE_SIZE << MAX_PAGE_ORDER; in regmap_reg_ranges_read_file()
|
/linux-6.12.1/kernel/events/ |
D | ring_buffer.c | 619 if (order > MAX_PAGE_ORDER) in rb_alloc_aux_page() 620 order = MAX_PAGE_ORDER; in rb_alloc_aux_page() 718 if (get_order((unsigned long)nr_pages * sizeof(void *)) > MAX_PAGE_ORDER) in rb_alloc_aux() 835 if (order_base_2(size) > PAGE_SHIFT+MAX_PAGE_ORDER) in rb_alloc()
|
/linux-6.12.1/drivers/crypto/hisilicon/ |
D | sgl.c | 76 block_size = 1 << (PAGE_SHIFT + MAX_PAGE_ORDER < 32 ? in hisi_acc_create_sgl_pool() 77 PAGE_SHIFT + MAX_PAGE_ORDER : 31); in hisi_acc_create_sgl_pool()
|
/linux-6.12.1/drivers/gpu/drm/i915/gem/ |
D | i915_gem_internal.c | 39 int max_order = MAX_PAGE_ORDER; in i915_gem_object_get_pages_internal()
|
/linux-6.12.1/arch/sparc/mm/ |
D | tsb.c | 405 if (max_tsb_size > PAGE_SIZE << MAX_PAGE_ORDER) in tsb_grow() 406 max_tsb_size = PAGE_SIZE << MAX_PAGE_ORDER; in tsb_grow()
|
/linux-6.12.1/fs/ramfs/ |
D | file-nommu.c | 73 if (unlikely(order > MAX_PAGE_ORDER)) in ramfs_nommu_expand_for_mapping()
|
/linux-6.12.1/arch/arm64/kvm/hyp/nvhe/ |
D | page_alloc.c | 231 pool->max_order = min(MAX_PAGE_ORDER, in hyp_pool_init()
|
/linux-6.12.1/kernel/dma/ |
D | pool.c | 88 order = min(get_order(pool_size), MAX_PAGE_ORDER); in atomic_pool_expand()
|
/linux-6.12.1/io_uring/ |
D | memmap.c | 23 if (order > MAX_PAGE_ORDER) in io_mem_alloc_compound()
|
/linux-6.12.1/Documentation/networking/ |
D | packet_mmap.rst | 267 region allocated by __get_free_pages is determined by the MAX_PAGE_ORDER macro. 270 PAGE_SIZE << MAX_PAGE_ORDER 273 In a 2.4/i386 kernel MAX_PAGE_ORDER is 10 274 In a 2.6/i386 kernel MAX_PAGE_ORDER is 11 280 /usr/include/linux/mmzone.h to get PAGE_SIZE MAX_PAGE_ORDER declarations. 328 <max-order> is the value defined with MAX_PAGE_ORDER
|
/linux-6.12.1/arch/powerpc/mm/book3s64/ |
D | iommu_api.c | 100 chunk = (1UL << (PAGE_SHIFT + MAX_PAGE_ORDER)) / in mm_iommu_do_alloc()
|
/linux-6.12.1/arch/um/kernel/ |
D | um_arch.c | 377 max_physmem &= ~((1 << (PAGE_SHIFT + MAX_PAGE_ORDER)) - 1); in linux_main()
|
/linux-6.12.1/drivers/misc/genwqe/ |
D | card_utils.c | 213 if (get_order(size) > MAX_PAGE_ORDER) in __genwqe_alloc_consistent() 311 if (get_order(sgl->sgl_size) > MAX_PAGE_ORDER) { in genwqe_alloc_sync_sgl()
|