/linux-6.12.1/include/linux/ |
D | huge_mm.h | 186 static inline int highest_order(unsigned long orders) in highest_order() argument 188 return fls_long(orders) - 1; in highest_order() 191 static inline int next_order(unsigned long *orders, int prev) in next_order() argument 193 *orders &= ~BIT(prev); in next_order() 194 return highest_order(*orders); in next_order() 234 unsigned long addr, unsigned long orders) in thp_vma_suitable_orders() argument 245 order = highest_order(orders); in thp_vma_suitable_orders() 247 while (orders) { in thp_vma_suitable_orders() 250 order = next_order(&orders, order); in thp_vma_suitable_orders() 253 return orders; in thp_vma_suitable_orders() [all …]
|
D | folio_queue.h | 31 u8 orders[PAGEVEC_SIZE]; /* Order of each folio */ member 238 folioq->orders[slot] = __folio_order(folio); in folioq_append() 260 folioq->orders[slot] = __folio_order(folio); in folioq_append_mark() 291 return folioq->orders[slot]; in folioq_folio_order()
|
/linux-6.12.1/tools/testing/selftests/mm/ |
D | thp_settings.c | 203 unsigned long orders = thp_supported_orders(); in thp_read_settings() local 230 if (!((1 << i) & orders)) { in thp_read_settings() 255 unsigned long orders = thp_supported_orders(); in thp_write_settings() local 281 if (!((1 << i) & orders)) in thp_write_settings() 352 unsigned long orders = 0; in __thp_supported_orders() local 369 orders |= 1UL << i; in __thp_supported_orders() 372 return orders; in __thp_supported_orders()
|
D | cow.c | 52 unsigned long orders; in detect_thp_sizes() local 60 orders = 1UL << sz2ord(pmdsize); in detect_thp_sizes() 61 orders |= thp_supported_orders(); in detect_thp_sizes() 63 for (i = 0; orders && count < max; i++) { in detect_thp_sizes() 64 if (!(orders & (1UL << i))) in detect_thp_sizes() 66 orders &= ~(1UL << i); in detect_thp_sizes()
|
/linux-6.12.1/drivers/gpu/drm/ttm/tests/ |
D | ttm_pool_test.c | 250 pt = &pool->caching[caching].orders[order]; in ttm_pool_alloc_order_caching_match() 279 pt_pool = &pool->caching[pool_caching].orders[order]; in ttm_pool_alloc_caching_mismatch() 280 pt_tt = &pool->caching[tt_caching].orders[order]; in ttm_pool_alloc_caching_mismatch() 313 pt_pool = &pool->caching[caching].orders[order]; in ttm_pool_alloc_order_mismatch() 314 pt_tt = &pool->caching[caching].orders[0]; in ttm_pool_alloc_order_mismatch() 354 pt = &pool->caching[caching].orders[order]; in ttm_pool_free_dma_alloc() 385 pt = &pool->caching[caching].orders[order]; in ttm_pool_free_no_dma_alloc() 405 pt = &pool->caching[caching].orders[order]; in ttm_pool_fini_basic()
|
D | ttm_device_test.c | 179 pt = pool->caching[i].orders[j]; in ttm_device_init_pools()
|
/linux-6.12.1/drivers/dma-buf/heaps/ |
D | system_heap.c | 54 static const unsigned int orders[] = {8, 4, 0}; variable 55 #define NUM_ORDERS ARRAY_SIZE(orders) 321 if (size < (PAGE_SIZE << orders[i])) in alloc_largest_available() 323 if (max_order < orders[i]) in alloc_largest_available() 326 page = alloc_pages(order_flags[i], orders[i]); in alloc_largest_available() 342 unsigned int max_order = orders[0]; in system_heap_allocate()
|
/linux-6.12.1/drivers/gpu/drm/ttm/ |
D | ttm_pool.c | 292 return &pool->caching[caching].orders[order]; in ttm_pool_select_type() 298 return &pool->caching[caching].orders[order]; in ttm_pool_select_type() 306 return &pool->caching[caching].orders[order]; in ttm_pool_select_type() 581 if (pt != &pool->caching[i].orders[j]) in ttm_pool_init() 619 if (pt != &pool->caching[i].orders[j]) in ttm_pool_fini() 755 ttm_pool_debugfs_orders(pool->caching[i].orders, m); in ttm_pool_debugfs()
|
/linux-6.12.1/fs/netfs/ |
D | read_collect.c | 121 folioq->orders[0], folioq->orders[1], in netfs_consume_read_data() 122 folioq->orders[2], folioq->orders[3]); in netfs_consume_read_data() 226 subreq->curr_folio_order = folioq->orders[slot]; in netfs_consume_read_data()
|
D | read_retry.c | 20 subreq->curr_folio_order = subreq->curr_folioq->orders[subreq->curr_folioq_slot]; in netfs_reissue_read()
|
D | buffered_read.c | 84 folioq->orders[i] = order; in netfs_load_buffer_from_ra() 169 subreq->curr_folio_order = subreq->curr_folioq->orders[subreq->curr_folioq_slot]; in netfs_prepare_read_iterator()
|
/linux-6.12.1/include/drm/ttm/ |
D | ttm_pool.h | 78 struct ttm_pool_type orders[NR_PAGE_ORDERS]; member
|
/linux-6.12.1/Documentation/userspace-api/media/v4l/ |
D | pixfmt-bayer.rst | 15 orders. See also `the Wikipedia article on Bayer filter
|
D | field-order.rst | 80 If multiple field orders are possible the 81 driver must choose one of the possible field orders during
|
/linux-6.12.1/mm/ |
D | memory.c | 4085 unsigned long orders) in thp_swap_suitable_orders() argument 4089 order = highest_order(orders); in thp_swap_suitable_orders() 4096 while (orders) { in thp_swap_suitable_orders() 4100 order = next_order(&orders, order); in thp_swap_suitable_orders() 4103 return orders; in thp_swap_suitable_orders() 4109 unsigned long orders; in alloc_swap_folio() local 4138 orders = thp_vma_allowable_orders(vma, vma->vm_flags, in alloc_swap_folio() 4140 orders = thp_vma_suitable_orders(vma, vmf->address, orders); in alloc_swap_folio() 4141 orders = thp_swap_suitable_orders(swp_offset(entry), in alloc_swap_folio() 4142 vmf->address, orders); in alloc_swap_folio() [all …]
|
D | huge_memory.c | 90 unsigned long orders) in __thp_vma_allowable_orders() argument 105 orders &= supported_orders; in __thp_vma_allowable_orders() 106 if (!orders) in __thp_vma_allowable_orders() 117 return in_pf ? orders : 0; in __thp_vma_allowable_orders() 135 int order = highest_order(orders); in __thp_vma_allowable_orders() 138 while (orders) { in __thp_vma_allowable_orders() 142 order = next_order(&orders, order); in __thp_vma_allowable_orders() 145 if (!orders) in __thp_vma_allowable_orders() 174 return orders; in __thp_vma_allowable_orders() 177 return orders; in __thp_vma_allowable_orders() [all …]
|
D | shmem.c | 1718 unsigned long orders) in shmem_suitable_orders() argument 1726 orders = thp_vma_suitable_orders(vma, vmf->address, orders); in shmem_suitable_orders() 1727 if (!orders) in shmem_suitable_orders() 1732 order = highest_order(orders); in shmem_suitable_orders() 1733 while (orders) { in shmem_suitable_orders() 1747 order = next_order(&orders, order); in shmem_suitable_orders() 1750 return orders; in shmem_suitable_orders() 1755 unsigned long orders) in shmem_suitable_orders() argument 1777 struct mm_struct *fault_mm, unsigned long orders) in shmem_alloc_and_add_folio() argument 1787 orders = 0; in shmem_alloc_and_add_folio() [all …]
|
/linux-6.12.1/tools/memory-model/Documentation/ |
D | glossary.txt | 29 a special operation that includes a load and which orders that 117 Fully Ordered: An operation such as smp_mb() that orders all of 120 that orders all of its CPU's prior accesses, itself, and 167 a special operation that includes a store and which orders that
|
D | recipes.txt | 232 The smp_store_release() macro orders any prior accesses against the 233 store, while the smp_load_acquire macro orders the load against any 273 smp_store_release(), but the rcu_dereference() macro orders the load only 310 The smp_wmb() macro orders prior stores against later stores, and the 311 smp_rmb() macro orders prior loads against later loads. Therefore, if
|
D | ordering.txt | 11 1. Barriers (also known as "fences"). A barrier orders some or 67 First, the smp_mb() full memory barrier orders all of the CPU's prior 115 synchronize_srcu() and so on. However, these primitives have orders
|
/linux-6.12.1/Documentation/ |
D | atomic_t.txt | 194 smp_mb__before_atomic() orders all earlier accesses against the RMW op 195 itself and all accesses following it, and smp_mb__after_atomic() orders all 226 a RELEASE because it orders preceding instructions against both the read
|
/linux-6.12.1/arch/arm/boot/dts/marvell/ |
D | armada-385-synology-ds116.dts | 53 * and takes single-character orders :
|
/linux-6.12.1/tools/virtio/virtio-trace/ |
D | README | 9 - controlled by start/stop orders from a Host
|
/linux-6.12.1/tools/memory-model/ |
D | lock.cat | 8 * Generate coherence orders and handle lock operations
|
/linux-6.12.1/Documentation/core-api/ |
D | genalloc.rst | 49 that state, so one of the first orders of business is usually to add memory
|