Home
last modified time | relevance | path

Searched refs:xe_vm (Results 1 – 25 of 26) sorted by relevance

12

/linux-6.12.1/drivers/gpu/drm/xe/
Dxe_vm.h27 struct xe_vm *xe_vm_create(struct xe_device *xe, u32 flags);
29 struct xe_vm *xe_vm_lookup(struct xe_file *xef, u32 id);
32 static inline struct xe_vm *xe_vm_get(struct xe_vm *vm) in xe_vm_get()
38 static inline void xe_vm_put(struct xe_vm *vm) in xe_vm_put()
43 int xe_vm_lock(struct xe_vm *vm, bool intr);
45 void xe_vm_unlock(struct xe_vm *vm);
47 static inline bool xe_vm_is_closed(struct xe_vm *vm) in xe_vm_is_closed()
53 static inline bool xe_vm_is_banned(struct xe_vm *vm) in xe_vm_is_banned()
58 static inline bool xe_vm_is_closed_or_banned(struct xe_vm *vm) in xe_vm_is_closed_or_banned()
65 xe_vm_find_overlapping_vma(struct xe_vm *vm, u64 start, u64 range);
[all …]
Dxe_trace_bo.h31 __field(struct xe_vm *, vm)
175 DECLARE_EVENT_CLASS(xe_vm,
176 TP_PROTO(struct xe_vm *vm),
181 __field(struct xe_vm *, vm)
195 DEFINE_EVENT(xe_vm, xe_vm_kill,
196 TP_PROTO(struct xe_vm *vm),
200 DEFINE_EVENT(xe_vm, xe_vm_create,
201 TP_PROTO(struct xe_vm *vm),
205 DEFINE_EVENT(xe_vm, xe_vm_free,
206 TP_PROTO(struct xe_vm *vm),
[all …]
Dxe_exec_queue.h17 struct xe_exec_queue *xe_exec_queue_create(struct xe_device *xe, struct xe_vm *vm,
22 struct xe_vm *vm,
76 void xe_exec_queue_last_fence_put(struct xe_exec_queue *e, struct xe_vm *vm);
79 struct xe_vm *vm);
81 struct xe_vm *vm);
82 void xe_exec_queue_last_fence_set(struct xe_exec_queue *e, struct xe_vm *vm,
85 struct xe_vm *vm);
Dxe_vm.c43 static struct drm_gem_object *xe_vm_obj(struct xe_vm *vm) in xe_vm_obj()
70 struct xe_vm *vm = xe_vma_vm(vma); in xe_vma_userptr_pin_pages()
79 static bool preempt_fences_waiting(struct xe_vm *vm) in preempt_fences_waiting()
105 static int alloc_preempt_fences(struct xe_vm *vm, struct list_head *list, in alloc_preempt_fences()
126 static int wait_for_existing_preempt_fences(struct xe_vm *vm) in wait_for_existing_preempt_fences()
148 static bool xe_vm_is_idle(struct xe_vm *vm) in xe_vm_is_idle()
161 static void arm_preempt_fences(struct xe_vm *vm, struct list_head *list) in arm_preempt_fences()
180 static int add_preempt_fences(struct xe_vm *vm, struct xe_bo *bo) in add_preempt_fences()
204 static void resume_and_reinstall_preempt_fences(struct xe_vm *vm, in resume_and_reinstall_preempt_fences()
220 int xe_vm_add_compute_exec_queue(struct xe_vm *vm, struct xe_exec_queue *q) in xe_vm_add_compute_exec_queue()
[all …]
Dxe_pt.h18 struct xe_vm;
30 struct xe_pt *xe_pt_create(struct xe_vm *vm, struct xe_tile *tile,
33 void xe_pt_populate_empty(struct xe_tile *tile, struct xe_vm *vm,
Dxe_bo.h78 struct xe_tile *tile, struct xe_vm *vm,
82 struct xe_vm *vm, size_t size,
85 struct xe_vm *vm, size_t size,
88 struct xe_vm *vm, size_t size,
92 struct xe_vm *vm, size_t size,
95 struct xe_vm *vm, size_t size, u64 offset,
162 int xe_bo_validate(struct xe_bo *bo, struct xe_vm *vm, bool allow_res_evict);
Dxe_exec_queue.c49 struct xe_vm *vm, in __xe_exec_queue_alloc()
110 struct xe_vm *vm = q->vm; in __xe_exec_queue_init()
145 struct xe_exec_queue *xe_exec_queue_create(struct xe_device *xe, struct xe_vm *vm, in xe_exec_queue_create()
170 struct xe_vm *vm, in xe_exec_queue_create_class()
217 struct xe_vm *migrate_vm; in xe_exec_queue_create_bind()
540 struct xe_vm *vm; in xe_exec_queue_create_ioctl()
845 struct xe_vm *vm) in xe_exec_queue_last_fence_lockdep_assert()
860 void xe_exec_queue_last_fence_put(struct xe_exec_queue *q, struct xe_vm *vm) in xe_exec_queue_last_fence_put()
891 struct xe_vm *vm) in xe_exec_queue_last_fence_get()
918 struct xe_vm *vm) in xe_exec_queue_last_fence_get_for_resume()
[all …]
Dxe_bo_types.h19 struct xe_vm;
35 struct xe_vm *vm;
Dxe_sync.h15 struct xe_vm;
32 struct xe_exec_queue *q, struct xe_vm *vm);
Dxe_vm_types.h23 struct xe_vm;
138 struct xe_vm { struct
363 struct xe_vm *vm;
Dxe_migrate.h23 struct xe_vm;
114 struct xe_vm *xe_migrate_get_vm(struct xe_migrate *m);
Dxe_sched_job.h12 struct xe_vm;
61 int xe_sched_job_last_fence_add_dep(struct xe_sched_job *job, struct xe_vm *vm);
Dxe_gt_pagefault.c83 static struct xe_vma *lookup_vma(struct xe_vm *vm, u64 page_addr) in lookup_vma()
101 struct xe_vm *vm = xe_vma_vm(vma); in xe_pf_begin()
131 struct xe_vm *vm = xe_vma_vm(vma); in handle_vma_pagefault()
188 static struct xe_vm *asid_to_vm(struct xe_device *xe, u32 asid) in asid_to_vm()
190 struct xe_vm *vm; in asid_to_vm()
207 struct xe_vm *vm; in handle_pagefault()
538 static struct xe_vma *get_acc_vma(struct xe_vm *vm, struct acc *acc) in get_acc_vma()
551 struct xe_vm *vm; in handle_acc()
Dxe_exec_queue_types.h22 struct xe_vm;
55 struct xe_vm *vm;
Dxe_pt.c56 static u64 __xe_pt_empty_pte(struct xe_tile *tile, struct xe_vm *vm, in __xe_pt_empty_pte()
97 struct xe_pt *xe_pt_create(struct xe_vm *vm, struct xe_tile *tile, in xe_pt_create()
150 void xe_pt_populate_empty(struct xe_tile *tile, struct xe_vm *vm, in xe_pt_populate_empty()
258 struct xe_vm *vm;
486 struct xe_vm *vm = xe_walk->vm; in xe_pt_stage_bind_entry()
877 struct xe_vm *vm = xe_vma_vm(vma); in xe_pt_commit_locks_assert()
1066 static int op_add_deps(struct xe_vm *vm, struct xe_vma_op *op, in op_add_deps()
1097 struct xe_vm *vm, in xe_pt_vm_dependencies()
1170 struct xe_vm *vm = vops->vm; in xe_pt_pre_commit()
1203 static int vma_check_userptr(struct xe_vm *vm, struct xe_vma *vma, in vma_check_userptr()
[all …]
Dxe_lrc.h21 struct xe_vm;
25 struct xe_lrc *xe_lrc_create(struct xe_hw_engine *hwe, struct xe_vm *vm,
Dxe_exec.c99 struct xe_vm *vm = container_of(vm_exec->vm, struct xe_vm, gpuvm); in xe_exec_fn()
119 struct xe_vm *vm; in xe_exec_ioctl()
Dxe_hmm.c129 struct xe_vm *vm = xe_vma_vm(vma); in xe_hmm_userptr_free_sg()
174 struct xe_vm *vm = xe_vma_vm(vma); in xe_hmm_userptr_populate_range()
Dxe_bo.c503 struct xe_vm *vm = gpuvm_to_vm(vm_bo->vm); in xe_bo_trigger_rebind()
1432 struct xe_tile *tile, struct xe_vm *vm, in __xe_bo_create_locked()
1501 struct xe_tile *tile, struct xe_vm *vm, in xe_bo_create_locked_range()
1509 struct xe_vm *vm, size_t size, in xe_bo_create_locked()
1516 struct xe_vm *vm, size_t size, in xe_bo_create_user()
1530 struct xe_vm *vm, size_t size, in xe_bo_create()
1542 struct xe_vm *vm, in xe_bo_create_pin_map_at()
1581 struct xe_vm *vm, size_t size, in xe_bo_create_pin_map()
1851 int xe_bo_validate(struct xe_bo *bo, struct xe_vm *vm, bool allow_res_evict) in xe_bo_validate()
1969 struct xe_vm *vm = NULL; in xe_gem_create_ioctl()
Dxe_sched_job.c244 struct xe_vm *vm = q->vm; in xe_sched_job_arm()
303 int xe_sched_job_last_fence_add_dep(struct xe_sched_job *job, struct xe_vm *vm) in xe_sched_job_last_fence_add_dep()
Dxe_migrate.c139 static void xe_migrate_program_identity(struct xe_device *xe, struct xe_vm *vm, struct xe_bo *bo, in xe_migrate_program_identity()
185 struct xe_vm *vm) in xe_migrate_prepare_vm()
400 struct xe_vm *vm; in xe_migrate_init()
565 struct xe_vm *vm = m->q->vm; in emit_pte()
1241 struct xe_vm *xe_migrate_get_vm(struct xe_migrate *m) in xe_migrate_get_vm()
1264 struct xe_vm *vm = pt_update->vops->vm; in xe_migrate_update_pgtables_cpu()
1399 struct xe_vm *vm = pt_update->vops->vm; in __xe_migrate_update_pgtables()
Dxe_lrc.c875 static void xe_lrc_set_ppgtt(struct xe_lrc *lrc, struct xe_vm *vm) in xe_lrc_set_ppgtt()
896 struct xe_vm *vm, u32 ring_size) in xe_lrc_init()
1031 struct xe_lrc *xe_lrc_create(struct xe_hw_engine *hwe, struct xe_vm *vm, in xe_lrc_create()
1674 struct xe_vm *vm; in xe_lrc_snapshot_capture_delayed()
1753 struct xe_vm *vm; in xe_lrc_snapshot_free()
Dxe_sync.c275 struct xe_exec_queue *q, struct xe_vm *vm) in xe_sync_in_fence_get()
DMakefile109 xe_vm.o \
/linux-6.12.1/drivers/gpu/drm/xe/tests/
Dxe_bo.c198 struct xe_vm *vm = xe_migrate_get_vm(xe_device_get_root_tile(xe)->migrate); in evict_test_run_tile()

12