/linux-6.12.1/drivers/mtd/tests/ |
D | pagetest.c | 34 static int pgsize; variable 66 for (j = 0; j < pgcnt - 1; ++j, addr += pgsize) { in verify_eraseblock() 78 if (memcmp(twopages, writebuf + (j * pgsize), bufsize)) { in verify_eraseblock() 85 if (addr <= addrn - pgsize - pgsize && !bbt[ebnum + 1]) { in verify_eraseblock() 99 memcpy(boundary, writebuf + mtd->erasesize - pgsize, pgsize); in verify_eraseblock() 100 prandom_bytes_state(&rnd_state, boundary + pgsize, pgsize); in verify_eraseblock() 118 pp1 = kcalloc(pgsize, 4, GFP_KERNEL); in crosstest() 121 pp2 = pp1 + pgsize; in crosstest() 122 pp3 = pp2 + pgsize; in crosstest() 123 pp4 = pp3 + pgsize; in crosstest() [all …]
|
D | torturetest.c | 70 static int pgsize; variable 97 addr = (loff_t)(ebnum + 1) * mtd->erasesize - pgcnt * pgsize; in check_eraseblock() 98 len = pgcnt * pgsize; in check_eraseblock() 151 addr = (loff_t)(ebnum + 1) * mtd->erasesize - pgcnt * pgsize; in write_pattern() 152 len = pgcnt * pgsize; in write_pattern() 203 pgsize = 512; in tort_init() 205 pgsize = mtd->writesize; in tort_init() 207 if (pgcnt && (pgcnt > mtd->erasesize / pgsize || pgcnt < 0)) { in tort_init() 235 for (i = 0; i < mtd->erasesize / pgsize; i++) { in tort_init() 237 memset(patt_5A5 + i * pgsize, 0x55, pgsize); in tort_init() [all …]
|
D | speedtest.c | 37 static int pgsize; variable 77 err = mtdtest_write(mtd, addr, pgsize, buf); in write_eraseblock_by_page() 80 addr += pgsize; in write_eraseblock_by_page() 81 buf += pgsize; in write_eraseblock_by_page() 89 size_t sz = pgsize * 2; in write_eraseblock_by_2pages() 102 err = mtdtest_write(mtd, addr, pgsize, buf); in write_eraseblock_by_2pages() 121 err = mtdtest_read(mtd, addr, pgsize, buf); in read_eraseblock_by_page() 124 addr += pgsize; in read_eraseblock_by_page() 125 buf += pgsize; in read_eraseblock_by_page() 133 size_t sz = pgsize * 2; in read_eraseblock_by_2pages() [all …]
|
D | readtest.c | 31 static int pgsize; variable 43 memset(buf, 0 , pgsize); in read_eraseblock_by_page() 44 ret = mtdtest_read(mtd, addr, pgsize, buf); in read_eraseblock_by_page() 72 addr += pgsize; in read_eraseblock_by_page() 73 buf += pgsize; in read_eraseblock_by_page() 138 pgsize = 512; in mtd_readtest_init() 140 pgsize = mtd->writesize; in mtd_readtest_init() 145 pgcnt = mtd->erasesize / pgsize; in mtd_readtest_init() 151 pgsize, ebcnt, pgcnt, mtd->oobsize); in mtd_readtest_init()
|
D | stresstest.c | 38 static int pgsize; variable 95 len = ((len + pgsize - 1) / pgsize) * pgsize; in do_write() 154 pgsize = 512; in mtd_stresstest_init() 156 pgsize = mtd->writesize; in mtd_stresstest_init() 161 pgcnt = mtd->erasesize / pgsize; in mtd_stresstest_init() 167 pgsize, ebcnt, pgcnt, mtd->oobsize); in mtd_stresstest_init()
|
/linux-6.12.1/arch/arm64/mm/ |
D | hugetlbpage.c | 83 pte_t *ptep, size_t *pgsize) in find_num_contig() argument 90 *pgsize = PAGE_SIZE; in find_num_contig() 95 *pgsize = PMD_SIZE; in find_num_contig() 101 static inline int num_contig_ptes(unsigned long size, size_t *pgsize) in num_contig_ptes() argument 105 *pgsize = size; in num_contig_ptes() 118 *pgsize = PMD_SIZE; in num_contig_ptes() 122 *pgsize = PAGE_SIZE; in num_contig_ptes() 133 size_t pgsize; in huge_ptep_get() local 139 ncontig = num_contig_ptes(page_size(pte_page(orig_pte)), &pgsize); in huge_ptep_get() 163 unsigned long pgsize, in get_clear_contig() argument [all …]
|
/linux-6.12.1/tools/testing/selftests/powerpc/mm/ |
D | pkey_exec_prot.c | 27 static unsigned long pgsize, numinsns; variable 62 if (mprotect(insns, pgsize, PROT_READ | PROT_WRITE)) { in segv_handler() 84 if (mprotect(insns, pgsize, PROT_EXEC)) { in segv_handler() 129 pgsize = getpagesize(); in test() 130 numinsns = pgsize / sizeof(unsigned int); in test() 131 insns = (unsigned int *) mmap(NULL, pgsize, PROT_READ | PROT_WRITE, in test() 179 FAIL_IF(sys_pkey_mprotect(insns, pgsize, PROT_EXEC, pkey) != 0); in test() 196 FAIL_IF(sys_pkey_mprotect(insns, pgsize, PROT_EXEC, pkey) != 0); in test() 216 FAIL_IF(sys_pkey_mprotect(insns, pgsize, PROT_EXEC, pkey) != 0); in test() 233 FAIL_IF(sys_pkey_mprotect(insns, pgsize, PROT_EXEC, pkey) != 0); in test() [all …]
|
D | exec_prot.c | 29 static unsigned long pgsize, numinsns; variable 73 if (mprotect(insns, pgsize, PROT_READ | PROT_WRITE | PROT_EXEC)) { in segv_handler() 98 FAIL_IF(mprotect(insns, pgsize, rights) != 0); in check_exec_fault() 136 pgsize = getpagesize(); in test() 137 numinsns = pgsize / sizeof(unsigned int); in test() 138 insns = (unsigned int *)mmap(NULL, pgsize, PROT_READ | PROT_WRITE, in test() 173 FAIL_IF(mprotect(insns, pgsize, PROT_EXEC) != 0); in test() 185 FAIL_IF(mprotect(insns, pgsize, PROT_EXEC) != 0); in test() 223 FAIL_IF(munmap((void *)insns, pgsize)); in test()
|
D | pkey_siginfo.c | 38 size_t pgsize; in segv_handler() local 67 pgsize = getpagesize(); in segv_handler() 68 pgstart = (void *) ((unsigned long) fault_addr & ~(pgsize - 1)); in segv_handler() 83 mprotect(pgstart, pgsize, PROT_EXEC)) in segv_handler()
|
/linux-6.12.1/tools/testing/selftests/powerpc/tm/ |
D | tm-vmxcopy.c | 39 unsigned long pgsize = getpagesize(); in test_vmxcopy() local 42 int size = pgsize*16; in test_vmxcopy() 44 char buf[pgsize]; in test_vmxcopy() 55 memset(buf, 0, pgsize); in test_vmxcopy() 56 for (i = 0; i < size; i += pgsize) in test_vmxcopy() 57 assert(write(fd, buf, pgsize) == pgsize); in test_vmxcopy()
|
/linux-6.12.1/drivers/iommu/amd/ |
D | io_pgtable.c | 329 phys_addr_t paddr, size_t pgsize, size_t pgcount, in iommu_v1_map_pages() argument 337 size_t size = pgcount << __ffs(pgsize); in iommu_v1_map_pages() 340 BUG_ON(!IS_ALIGNED(iova, pgsize)); in iommu_v1_map_pages() 341 BUG_ON(!IS_ALIGNED(paddr, pgsize)); in iommu_v1_map_pages() 348 count = PAGE_SIZE_PTE_COUNT(pgsize); in iommu_v1_map_pages() 349 pte = alloc_pte(pgtable, iova, pgsize, NULL, gfp, &updated); in iommu_v1_map_pages() 362 __pte = PAGE_SIZE_PTE(__sme_set(paddr), pgsize); in iommu_v1_map_pages() 375 iova += pgsize; in iommu_v1_map_pages() 376 paddr += pgsize; in iommu_v1_map_pages() 379 *mapped += pgsize; in iommu_v1_map_pages() [all …]
|
D | io_pgtable_v2.c | 233 phys_addr_t paddr, size_t pgsize, size_t pgcount, in iommu_v2_map_pages() argument 242 size_t size = pgcount << __ffs(pgsize); in iommu_v2_map_pages() 246 if (WARN_ON(!pgsize || (pgsize & cfg->pgsize_bitmap) != pgsize) || !pgcount) in iommu_v2_map_pages() 253 map_size = get_alloc_page_size(pgsize); in iommu_v2_map_pages() 283 size_t pgsize, size_t pgcount, in iommu_v2_unmap_pages() argument 290 size_t size = pgcount << __ffs(pgsize); in iommu_v2_unmap_pages() 293 if (WARN_ON(!pgsize || (pgsize & cfg->pgsize_bitmap) != pgsize || !pgcount)) in iommu_v2_unmap_pages()
|
/linux-6.12.1/drivers/gpu/drm/msm/ |
D | msm_iommu.c | 42 size_t offset, pgsize, pgsize_next; in calc_pgsize() local 57 pgsize = BIT(pgsize_idx); in calc_pgsize() 59 return pgsize; in calc_pgsize() 88 return pgsize; in calc_pgsize() 98 size_t unmapped, pgsize, count; in msm_iommu_pagetable_unmap() local 100 pgsize = calc_pgsize(pagetable, iova, iova, size, &count); in msm_iommu_pagetable_unmap() 102 unmapped = ops->unmap_pages(ops, iova, pgsize, count, NULL); in msm_iommu_pagetable_unmap() 129 size_t pgsize, count, mapped = 0; in msm_iommu_pagetable_map() local 132 pgsize = calc_pgsize(pagetable, addr, phys, size, &count); in msm_iommu_pagetable_map() 134 ret = ops->map_pages(ops, addr, phys, pgsize, count, in msm_iommu_pagetable_map()
|
/linux-6.12.1/drivers/gpu/drm/etnaviv/ |
D | etnaviv_mmu.c | 20 size_t pgsize = SZ_4K; in etnaviv_context_unmap() local 22 if (!IS_ALIGNED(iova | size, pgsize)) { in etnaviv_context_unmap() 24 iova, size, pgsize); in etnaviv_context_unmap() 30 pgsize); in etnaviv_context_unmap() 44 size_t pgsize = SZ_4K; in etnaviv_context_map() local 48 if (!IS_ALIGNED(iova | paddr | size, pgsize)) { in etnaviv_context_map() 50 iova, &paddr, size, pgsize); in etnaviv_context_map() 55 ret = context->global->ops->map(context, iova, paddr, pgsize, in etnaviv_context_map() 60 iova += pgsize; in etnaviv_context_map() 61 paddr += pgsize; in etnaviv_context_map() [all …]
|
/linux-6.12.1/drivers/media/pci/intel/ipu6/ |
D | ipu6-mmu.c | 626 size_t pgsize; in ipu6_mmu_pgsize() local 638 pgsize = (1UL << (pgsize_idx + 1)) - 1; in ipu6_mmu_pgsize() 639 pgsize &= pgsize_bitmap; in ipu6_mmu_pgsize() 641 WARN_ON(!pgsize); in ipu6_mmu_pgsize() 644 pgsize_idx = __fls(pgsize); in ipu6_mmu_pgsize() 645 pgsize = 1UL << pgsize_idx; in ipu6_mmu_pgsize() 647 return pgsize; in ipu6_mmu_pgsize() 675 size_t pgsize = ipu6_mmu_pgsize(mmu_info->pgsize_bitmap, in ipu6_mmu_unmap() local 678 unmapped_page = __ipu6_mmu_unmap(mmu_info, iova, pgsize); in ipu6_mmu_unmap() 722 size_t pgsize = ipu6_mmu_pgsize(mmu_info->pgsize_bitmap, in ipu6_mmu_map() local [all …]
|
/linux-6.12.1/drivers/iommu/ |
D | io-pgtable-dart.c | 232 phys_addr_t paddr, size_t pgsize, size_t pgcount, in dart_map_pages() argument 242 if (WARN_ON(pgsize != cfg->pgsize_bitmap)) in dart_map_pages() 281 *mapped += num_entries * pgsize; in dart_map_pages() 293 size_t pgsize, size_t pgcount, in dart_unmap_pages() argument 301 if (WARN_ON(pgsize != cfg->pgsize_bitmap || !pgcount)) in dart_unmap_pages() 326 iova + i * pgsize, pgsize); in dart_unmap_pages() 332 return i * pgsize; in dart_unmap_pages()
|
D | apple-dart.c | 217 u32 pgsize; member 530 phys_addr_t paddr, size_t pgsize, in apple_dart_map_pages() argument 540 return ops->map_pages(ops, iova, paddr, pgsize, pgcount, prot, gfp, in apple_dart_map_pages() 545 unsigned long iova, size_t pgsize, in apple_dart_unmap_pages() argument 552 return ops->unmap_pages(ops, iova, pgsize, pgcount, gather); in apple_dart_unmap_pages() 581 if (dart->pgsize > PAGE_SIZE) in apple_dart_finalize_domain() 597 .pgsize_bitmap = dart->pgsize, in apple_dart_finalize_domain() 805 if (cfg_dart->pgsize != dart->pgsize) in apple_dart_of_xlate() 946 if (cfg->stream_maps[0].dart->pgsize > PAGE_SIZE) in apple_dart_def_domain_type() 1119 dart->pgsize = 1 << FIELD_GET(DART_PARAMS1_PAGE_SHIFT, dart_params[0]); in apple_dart_probe() [all …]
|
D | io-pgtable-arm.c | 500 phys_addr_t paddr, size_t pgsize, size_t pgcount, in arm_lpae_map_pages() argument 510 if (WARN_ON(!pgsize || (pgsize & cfg->pgsize_bitmap) != pgsize)) in arm_lpae_map_pages() 522 ret = __arm_lpae_map(data, iova, paddr, pgsize, pgcount, prot, lvl, in arm_lpae_map_pages() 695 size_t pgsize, size_t pgcount, in arm_lpae_unmap_pages() argument 703 if (WARN_ON(!pgsize || (pgsize & cfg->pgsize_bitmap) != pgsize || !pgcount)) in arm_lpae_unmap_pages() 711 return __arm_lpae_unmap(data, gather, iova, pgsize, pgcount, in arm_lpae_unmap_pages() 1394 static const unsigned long pgsize[] __initconst = { in arm_lpae_do_selftests() local 1416 for (i = 0; i < ARRAY_SIZE(pgsize); ++i) { in arm_lpae_do_selftests() 1418 cfg.pgsize_bitmap = pgsize[i]; in arm_lpae_do_selftests() 1421 pgsize[i], ias[j]); in arm_lpae_do_selftests()
|
D | io-pgtable-arm-v7s.c | 545 phys_addr_t paddr, size_t pgsize, size_t pgcount, in arm_v7s_map_pages() argument 559 ret = __arm_v7s_map(data, iova, paddr, pgsize, prot, 1, data->pgd, in arm_v7s_map_pages() 564 iova += pgsize; in arm_v7s_map_pages() 565 paddr += pgsize; in arm_v7s_map_pages() 566 *mapped += pgsize; in arm_v7s_map_pages() 738 size_t pgsize, size_t pgcount, in arm_v7s_unmap_pages() argument 748 ret = __arm_v7s_unmap(data, gather, iova, pgsize, 1, data->pgd); in arm_v7s_unmap_pages() 752 unmapped += pgsize; in arm_v7s_unmap_pages() 753 iova += pgsize; in arm_v7s_unmap_pages()
|
/linux-6.12.1/arch/riscv/mm/ |
D | hugetlbpage.c | 203 unsigned long pgsize, in clear_flush() argument 209 for (i = 0; i < ncontig; i++, addr += pgsize, ptep++) in clear_flush() 229 unsigned long hugepage_shift, pgsize; in set_huge_pte_at() local 244 pgsize = 1 << hugepage_shift; in set_huge_pte_at() 247 for (i = 0; i < pte_num; i++, ptep++, addr += pgsize) in set_huge_pte_at() 257 clear_flush(mm, addr, ptep, pgsize, pte_num); in set_huge_pte_at() 259 for (i = 0; i < pte_num; i++, ptep++, addr += pgsize) in set_huge_pte_at()
|
/linux-6.12.1/drivers/vfio/ |
D | vfio_iommu_type1.c | 234 static int vfio_dma_bitmap_alloc(struct vfio_dma *dma, size_t pgsize) in vfio_dma_bitmap_alloc() argument 236 uint64_t npages = dma->size / pgsize; in vfio_dma_bitmap_alloc() 260 static void vfio_dma_populate_bitmap(struct vfio_dma *dma, size_t pgsize) in vfio_dma_populate_bitmap() argument 263 unsigned long pgshift = __ffs(pgsize); in vfio_dma_populate_bitmap() 284 static int vfio_dma_bitmap_alloc_all(struct vfio_iommu *iommu, size_t pgsize) in vfio_dma_bitmap_alloc_all() argument 292 ret = vfio_dma_bitmap_alloc(dma, pgsize); in vfio_dma_bitmap_alloc_all() 304 vfio_dma_populate_bitmap(dma, pgsize); in vfio_dma_bitmap_alloc_all() 1144 size_t pgsize) in update_user_bitmap() argument 1146 unsigned long pgshift = __ffs(pgsize); in update_user_bitmap() 1180 dma_addr_t iova, size_t size, size_t pgsize) in vfio_iova_dirty_bitmap() argument [all …]
|
/linux-6.12.1/drivers/iommu/iommufd/ |
D | selftest.c | 255 unsigned long pgsize = MOCK_IO_PAGE_SIZE; in mock_domain_read_and_clear_dirty() local 260 iova += pgsize; in mock_domain_read_and_clear_dirty() 265 pgsize = MOCK_HUGE_PAGE_SIZE; in mock_domain_read_and_clear_dirty() 266 head = iova & ~(pgsize - 1); in mock_domain_read_and_clear_dirty() 269 if (mock_test_and_clear_dirty(mock, head, pgsize, flags)) in mock_domain_read_and_clear_dirty() 270 iommu_dirty_bitmap_record(dirty, iova, pgsize); in mock_domain_read_and_clear_dirty() 271 iova += pgsize; in mock_domain_read_and_clear_dirty() 378 size_t pgsize, size_t pgcount, int prot, in mock_domain_map_pages() argument 394 WARN_ON(pgsize % MOCK_IO_PAGE_SIZE); in mock_domain_map_pages() 398 for (cur = 0; cur != pgsize; cur += MOCK_IO_PAGE_SIZE) { in mock_domain_map_pages() [all …]
|
/linux-6.12.1/drivers/infiniband/hw/bnxt_re/ |
D | qplib_res.c | 97 rdma_umem_for_each_dma_block(sginfo->umem, &biter, sginfo->pgsize) { in bnxt_qplib_fill_user_dma_pages() 117 pages = ib_umem_num_dma_blocks(sginfo->umem, sginfo->pgsize); in __alloc_pbl() 132 pbl->pg_size = sginfo->pgsize; in __alloc_pbl() 196 pg_size = hwq_attr->sginfo->pgsize; in bnxt_qplib_alloc_init_hwq() 219 hwq_attr->sginfo->pgsize); in bnxt_qplib_alloc_init_hwq() 244 sginfo.pgsize = npde * pg_size; in bnxt_qplib_alloc_init_hwq() 252 sginfo.pgsize = PAGE_SIZE; in bnxt_qplib_alloc_init_hwq() 298 sginfo.pgsize = PAGE_SIZE; in bnxt_qplib_alloc_init_hwq() 382 sginfo.pgsize = PAGE_SIZE; in bnxt_qplib_alloc_tqm_rings() 499 sginfo.pgsize = PAGE_SIZE; in bnxt_qplib_alloc_ctx()
|
/linux-6.12.1/drivers/gpu/drm/panfrost/ |
D | panfrost_mmu.c | 306 size_t pgsize = get_pgsize(iova | paddr, len, &pgcount); in mmu_map_sg() local 308 ops->map_pages(ops, iova, paddr, pgsize, pgcount, prot, in mmu_map_sg() 311 mapped = max(mapped, pgsize); in mmu_map_sg() 367 size_t pgsize = get_pgsize(iova, len - unmapped_len, &pgcount); in panfrost_mmu_unmap() local 372 unmapped_page = ops->unmap_pages(ops, iova, pgsize, pgcount, NULL); in panfrost_mmu_unmap() 373 WARN_ON(unmapped_page != pgsize * pgcount); in panfrost_mmu_unmap() 375 iova += pgsize * pgcount; in panfrost_mmu_unmap() 376 unmapped_len += pgsize * pgcount; in panfrost_mmu_unmap()
|
/linux-6.12.1/arch/powerpc/kvm/ |
D | book3s_64_mmu.c | 206 int pgsize; in kvmppc_mmu_book3s_64_xlate() local 242 pgsize = slbe->large ? MMU_PAGE_16M : MMU_PAGE_4K; in kvmppc_mmu_book3s_64_xlate() 271 pgsize = decode_pagesize(slbe, pte1); in kvmppc_mmu_book3s_64_xlate() 272 if (pgsize < 0) in kvmppc_mmu_book3s_64_xlate() 296 eaddr_mask = (1ull << mmu_pagesize(pgsize)) - 1; in kvmppc_mmu_book3s_64_xlate() 298 gpte->page_size = pgsize; in kvmppc_mmu_book3s_64_xlate()
|