/linux-6.12.1/fs/coda/ |
D | file.c | 37 struct vm_operations_struct vm_ops; member 126 container_of(vma->vm_ops, struct coda_vm_ops, vm_ops); in coda_vm_open() 138 container_of(vma->vm_ops, struct coda_vm_ops, vm_ops); in coda_vm_close() 144 vma->vm_ops = cvm_ops->host_vm_ops; in coda_vm_close() 212 cvm_ops->host_vm_ops = vma->vm_ops; in coda_file_mmap() 213 if (vma->vm_ops) in coda_file_mmap() 214 cvm_ops->vm_ops = *vma->vm_ops; in coda_file_mmap() 216 cvm_ops->vm_ops.open = coda_vm_open; in coda_file_mmap() 217 cvm_ops->vm_ops.close = coda_vm_close; in coda_file_mmap() 221 vma->vm_ops = &cvm_ops->vm_ops; in coda_file_mmap()
|
/linux-6.12.1/tools/testing/vma/ |
D | vma.c | 390 const struct vm_operations_struct vm_ops = { in test_merge_new() local 436 vma_a->vm_ops = &vm_ops; /* This should have no impact. */ in test_merge_new() 473 vma_d->vm_ops = &vm_ops; /* This should have no impact. */ in test_merge_new() 491 vma_d->vm_ops = NULL; /* This would otherwise degrade the merge. */ in test_merge_new() 646 const struct vm_operations_struct vm_ops = { in test_vma_merge_with_close() local 721 vma_next->vm_ops = &vm_ops; in test_vma_merge_with_close() 746 vma->vm_ops = &vm_ops; in test_vma_merge_with_close() 774 vma->vm_ops = &vm_ops; in test_vma_merge_with_close() 803 vma->vm_ops = &vm_ops; in test_vma_merge_with_close() 829 vma_next->vm_ops = &vm_ops; in test_vma_merge_with_close() [all …]
|
D | vma_internal.h | 259 const struct vm_operations_struct *vm_ops; member 429 vma->vm_ops = &vma_dummy_vm_ops; in vma_init() 566 return !vma->vm_ops; in vma_is_anonymous()
|
/linux-6.12.1/fs/kernfs/ |
D | file.c | 357 if (!of->vm_ops) in kernfs_vma_open() 363 if (of->vm_ops->open) in kernfs_vma_open() 364 of->vm_ops->open(vma); in kernfs_vma_open() 375 if (!of->vm_ops) in kernfs_vma_fault() 382 if (of->vm_ops->fault) in kernfs_vma_fault() 383 ret = of->vm_ops->fault(vmf); in kernfs_vma_fault() 395 if (!of->vm_ops) in kernfs_vma_page_mkwrite() 402 if (of->vm_ops->page_mkwrite) in kernfs_vma_page_mkwrite() 403 ret = of->vm_ops->page_mkwrite(vmf); in kernfs_vma_page_mkwrite() 418 if (!of->vm_ops) in kernfs_vma_access() [all …]
|
/linux-6.12.1/arch/um/kernel/ |
D | tlb.c | 18 struct vm_ops { struct 65 struct vm_ops *ops) in update_pte_range() 102 struct vm_ops *ops) in update_pmd_range() 125 struct vm_ops *ops) in update_pud_range() 148 struct vm_ops *ops) in update_p4d_range() 172 struct vm_ops ops; in um_tlb_sync()
|
/linux-6.12.1/ipc/ |
D | shm.c | 88 const struct vm_operations_struct *vm_ops; member 309 if (sfd->vm_ops->open) in shm_open() 310 sfd->vm_ops->open(vma); in shm_open() 402 if (sfd->vm_ops->close) in shm_close() 403 sfd->vm_ops->close(vma); in shm_close() 540 return sfd->vm_ops->fault(vmf); in shm_fault() 548 if (sfd->vm_ops->may_split) in shm_may_split() 549 return sfd->vm_ops->may_split(vma, addr); in shm_may_split() 559 if (sfd->vm_ops->pagesize) in shm_pagesize() 560 return sfd->vm_ops->pagesize(vma); in shm_pagesize() [all …]
|
/linux-6.12.1/mm/ |
D | vma.c | 377 if (vma->vm_ops && vma->vm_ops->may_split) { in __split_vma() 378 err = vma->vm_ops->may_split(vma, addr); in __split_vma() 410 if (new->vm_ops && new->vm_ops->open) in __split_vma() 411 new->vm_ops->open(new); in __split_vma() 627 return !vma->vm_ops || !vma->vm_ops->close; in can_merge_remove_vma() 1675 if (new_vma->vm_ops && new_vma->vm_ops->open) in copy_vma() 1676 new_vma->vm_ops->open(new_vma); in copy_vma() 1797 static bool vm_ops_needs_writenotify(const struct vm_operations_struct *vm_ops) in vm_ops_needs_writenotify() argument 1799 return vm_ops && (vm_ops->page_mkwrite || vm_ops->pfn_mkwrite); in vm_ops_needs_writenotify() 1829 if (vm_ops_needs_writenotify(vma->vm_ops)) in vma_needs_dirty_tracking() [all …]
|
D | mremap.c | 686 if (vma->vm_ops && vma->vm_ops->may_split) { in move_vma() 688 err = vma->vm_ops->may_split(vma, old_addr); in move_vma() 690 err = vma->vm_ops->may_split(vma, old_addr + old_len); in move_vma() 726 } else if (vma->vm_ops && vma->vm_ops->mremap) { in move_vma() 727 err = vma->vm_ops->mremap(new_vma); in move_vma()
|
D | memory.c | 541 vma->vm_ops ? vma->vm_ops->fault : NULL, in print_bad_pte() 601 if (vma->vm_ops && vma->vm_ops->find_special_page) in vm_normal_page() 602 return vma->vm_ops->find_special_page(vma, addr); in vm_normal_page() 2021 return vma->vm_ops && vma->vm_ops->pfn_mkwrite && in vm_mixed_zeropage_allowed() 3162 ret = vmf->vma->vm_ops->page_mkwrite(vmf); in do_page_mkwrite() 3190 bool page_mkwrite = vma->vm_ops && vma->vm_ops->page_mkwrite; in fault_dirty_shared_page() 3275 if (vma->vm_ops->map_pages || !(vmf->flags & FAULT_FLAG_VMA_LOCK)) in vmf_can_call_fault() 3527 if (vma->vm_ops && vma->vm_ops->pfn_mkwrite) { in wp_pfn_shared() 3536 ret = vma->vm_ops->pfn_mkwrite(vmf); in wp_pfn_shared() 3553 if (vma->vm_ops && vma->vm_ops->page_mkwrite) { in wp_page_shared() [all …]
|
D | secretmem.c | 134 vma->vm_ops = &secretmem_vm_ops; in secretmem_mmap() 141 return vma->vm_ops == &secretmem_vm_ops; in vma_is_secretmem()
|
D | internal.h | 133 vma->vm_ops = &vma_dummy_vm_ops; in mmap_file() 145 if (vma->vm_ops && vma->vm_ops->close) { in vma_close() 146 vma->vm_ops->close(vma); in vma_close() 152 vma->vm_ops = &vma_dummy_vm_ops; in vma_close()
|
/linux-6.12.1/drivers/gpu/drm/xen/ |
D | xen_drm_front_gem.c | 66 vma->vm_ops = gem_obj->funcs->vm_ops; in xen_drm_front_gem_object_mmap() 111 .vm_ops = &xen_drm_drv_vm_ops,
|
/linux-6.12.1/drivers/gpu/drm/mediatek/ |
D | mtk_gem.c | 20 static const struct vm_operations_struct vm_ops = { variable 31 .vm_ops = &vm_ops,
|
/linux-6.12.1/arch/hexagon/kernel/ |
D | Makefile | 14 obj-y += vm_entry.o vm_events.o vm_switch.o vm_ops.o vm_init_segtable.o
|
/linux-6.12.1/include/linux/ |
D | agp_backend.h | 56 const struct vm_operations_struct *vm_ops; member
|
/linux-6.12.1/drivers/pci/ |
D | mmap.c | 47 vma->vm_ops = &pci_phys_vm_ops; in pci_mmap_resource_range()
|
/linux-6.12.1/drivers/gpu/drm/i915/gem/ |
D | i915_gem_mman.c | 1026 vma->vm_ops = obj->ops->mmap_ops; in i915_gem_object_mmap() 1037 vma->vm_ops = &vm_ops_cpu; in i915_gem_object_mmap() 1045 vma->vm_ops = &vm_ops_cpu; in i915_gem_object_mmap() 1051 vma->vm_ops = &vm_ops_cpu; in i915_gem_object_mmap() 1057 vma->vm_ops = &vm_ops_gtt; in i915_gem_object_mmap()
|
/linux-6.12.1/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_gem.c | 121 vma->vm_ops = NULL; in vmw_gem_mmap() 154 .vm_ops = &vmw_vm_ops,
|
/linux-6.12.1/arch/x86/kernel/cpu/sgx/ |
D | encl.h | 95 if (!result || result->vm_ops != &sgx_vm_ops) in sgx_encl_find()
|
/linux-6.12.1/drivers/gpu/drm/ttm/ |
D | ttm_bo_vm.c | 471 if (!vma->vm_ops) in ttm_bo_mmap_obj() 472 vma->vm_ops = &ttm_bo_vm_ops; in ttm_bo_mmap_obj()
|
/linux-6.12.1/drivers/media/common/videobuf2/ |
D | videobuf2-vmalloc.c | 195 vma->vm_ops = &vb2_common_vm_ops; in vb2_vmalloc_mmap() 197 vma->vm_ops->open(vma); in vb2_vmalloc_mmap()
|
/linux-6.12.1/drivers/gpu/drm/etnaviv/ |
D | etnaviv_gem.c | 530 static const struct vm_operations_struct vm_ops = { variable 543 .vm_ops = &vm_ops,
|
/linux-6.12.1/drivers/infiniband/sw/rxe/ |
D | rxe_mmap.c | 105 vma->vm_ops = &rxe_vm_ops; in rxe_mmap()
|
/linux-6.12.1/fs/nilfs2/ |
D | file.c | 131 vma->vm_ops = &nilfs_file_vm_ops; in nilfs_file_mmap()
|
/linux-6.12.1/fs/ocfs2/ |
D | mmap.c | 174 vma->vm_ops = &ocfs2_file_vm_ops; in ocfs2_mmap()
|