Home
last modified time | relevance | path

Searched refs:bo (Results 1 – 25 of 408) sorted by relevance

12345678910>>...17

/linux-6.12.1/drivers/staging/media/atomisp/pci/hmm/
Dhmm_bo.c47 static int __bo_init(struct hmm_bo_device *bdev, struct hmm_buffer_object *bo, in __bo_init() argument
59 memset(bo, 0, sizeof(*bo)); in __bo_init()
60 mutex_init(&bo->mutex); in __bo_init()
63 INIT_LIST_HEAD(&bo->list); in __bo_init()
65 bo->bdev = bdev; in __bo_init()
66 bo->vmap_addr = NULL; in __bo_init()
67 bo->status = HMM_BO_FREE; in __bo_init()
68 bo->start = bdev->start; in __bo_init()
69 bo->pgnr = pgnr; in __bo_init()
70 bo->end = bo->start + pgnr_to_size(pgnr); in __bo_init()
[all …]
Dhmm.c52 struct hmm_buffer_object *bo; in bo_show() local
67 list_for_each_entry(bo, bo_list, list) { in bo_show()
68 if ((active && (bo->status & HMM_BO_ALLOCED)) || in bo_show()
69 (!active && !(bo->status & HMM_BO_ALLOCED))) { in bo_show()
72 hmm_bo_type_string[bo->type], bo->pgnr); in bo_show()
74 total[bo->type] += bo->pgnr; in bo_show()
75 count[bo->type]++; in bo_show()
173 struct hmm_buffer_object *bo; in __hmm_alloc() local
187 bo = hmm_bo_alloc(&bo_device, pgnr); in __hmm_alloc()
188 if (!bo) { in __hmm_alloc()
[all …]
/linux-6.12.1/drivers/gpu/drm/ttm/
Dttm_bo.c49 static void ttm_bo_mem_space_debug(struct ttm_buffer_object *bo, in ttm_bo_mem_space_debug() argument
60 man = ttm_manager_type(bo->bdev, mem_type); in ttm_bo_mem_space_debug()
74 void ttm_bo_move_to_lru_tail(struct ttm_buffer_object *bo) in ttm_bo_move_to_lru_tail() argument
76 dma_resv_assert_held(bo->base.resv); in ttm_bo_move_to_lru_tail()
78 if (bo->resource) in ttm_bo_move_to_lru_tail()
79 ttm_resource_move_to_lru_tail(bo->resource); in ttm_bo_move_to_lru_tail()
97 void ttm_bo_set_bulk_move(struct ttm_buffer_object *bo, in ttm_bo_set_bulk_move() argument
100 dma_resv_assert_held(bo->base.resv); in ttm_bo_set_bulk_move()
102 if (bo->bulk_move == bulk) in ttm_bo_set_bulk_move()
105 spin_lock(&bo->bdev->lru_lock); in ttm_bo_set_bulk_move()
[all …]
Dttm_bo_util.c42 struct ttm_buffer_object *bo; member
142 int ttm_bo_move_memcpy(struct ttm_buffer_object *bo, in ttm_bo_move_memcpy() argument
146 struct ttm_device *bdev = bo->bdev; in ttm_bo_move_memcpy()
148 ttm_manager_type(bo->bdev, dst_mem->mem_type); in ttm_bo_move_memcpy()
149 struct ttm_tt *ttm = bo->ttm; in ttm_bo_move_memcpy()
150 struct ttm_resource *src_mem = bo->resource; in ttm_bo_move_memcpy()
173 dst_iter = ttm_kmap_iter_tt_init(&_dst_iter.tt, bo->ttm); in ttm_bo_move_memcpy()
179 src_iter = ttm_kmap_iter_tt_init(&_src_iter.tt, bo->ttm); in ttm_bo_move_memcpy()
191 ttm_bo_move_sync_cleanup(bo, dst_mem); in ttm_bo_move_memcpy()
201 static void ttm_transfered_destroy(struct ttm_buffer_object *bo) in ttm_transfered_destroy() argument
[all …]
Dttm_bo_vm.c41 static vm_fault_t ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo, in ttm_bo_vm_fault_idle() argument
49 if (dma_resv_test_signaled(bo->base.resv, DMA_RESV_USAGE_KERNEL)) in ttm_bo_vm_fault_idle()
61 ttm_bo_get(bo); in ttm_bo_vm_fault_idle()
63 (void)dma_resv_wait_timeout(bo->base.resv, in ttm_bo_vm_fault_idle()
66 dma_resv_unlock(bo->base.resv); in ttm_bo_vm_fault_idle()
67 ttm_bo_put(bo); in ttm_bo_vm_fault_idle()
74 err = dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_KERNEL, true, in ttm_bo_vm_fault_idle()
84 static unsigned long ttm_bo_io_mem_pfn(struct ttm_buffer_object *bo, in ttm_bo_io_mem_pfn() argument
87 struct ttm_device *bdev = bo->bdev; in ttm_bo_io_mem_pfn()
90 return bdev->funcs->io_mem_pfn(bo, page_offset); in ttm_bo_io_mem_pfn()
[all …]
/linux-6.12.1/drivers/accel/ivpu/
Divpu_gem.c25 static inline void ivpu_dbg_bo(struct ivpu_device *vdev, struct ivpu_bo *bo, const char *action) in ivpu_dbg_bo() argument
29 action, bo, bo->vpu_addr, ivpu_bo_size(bo), bo->ctx ? bo->ctx->id : 0, in ivpu_dbg_bo()
30 (bool)bo->base.pages, (bool)bo->base.sgt, bo->mmu_mapped, bo->base.map_wc, in ivpu_dbg_bo()
31 (bool)bo->base.base.import_attach); in ivpu_dbg_bo()
41 int __must_check ivpu_bo_pin(struct ivpu_bo *bo) in ivpu_bo_pin() argument
43 struct ivpu_device *vdev = ivpu_bo_to_vdev(bo); in ivpu_bo_pin()
46 mutex_lock(&bo->lock); in ivpu_bo_pin()
48 ivpu_dbg_bo(vdev, bo, "pin"); in ivpu_bo_pin()
49 drm_WARN_ON(&vdev->drm, !bo->ctx); in ivpu_bo_pin()
51 if (!bo->mmu_mapped) { in ivpu_bo_pin()
[all …]
Divpu_gem.h27 int ivpu_bo_pin(struct ivpu_bo *bo);
34 void ivpu_bo_free(struct ivpu_bo *bo);
48 static inline void *ivpu_bo_vaddr(struct ivpu_bo *bo) in ivpu_bo_vaddr() argument
50 return bo->base.vaddr; in ivpu_bo_vaddr()
53 static inline size_t ivpu_bo_size(struct ivpu_bo *bo) in ivpu_bo_size() argument
55 return bo->base.base.size; in ivpu_bo_size()
58 static inline u32 ivpu_bo_cache_mode(struct ivpu_bo *bo) in ivpu_bo_cache_mode() argument
60 return bo->flags & DRM_IVPU_BO_CACHE_MASK; in ivpu_bo_cache_mode()
63 static inline struct ivpu_device *ivpu_bo_to_vdev(struct ivpu_bo *bo) in ivpu_bo_to_vdev() argument
65 return to_ivpu_device(bo->base.base.dev); in ivpu_bo_to_vdev()
[all …]
/linux-6.12.1/drivers/gpu/drm/qxl/
Dqxl_object.c34 struct qxl_bo *bo; in qxl_ttm_bo_destroy() local
37 bo = to_qxl_bo(tbo); in qxl_ttm_bo_destroy()
38 qdev = to_qxl(bo->tbo.base.dev); in qxl_ttm_bo_destroy()
40 qxl_surface_evict(qdev, bo, false); in qxl_ttm_bo_destroy()
41 WARN_ON_ONCE(bo->map_count > 0); in qxl_ttm_bo_destroy()
43 list_del_init(&bo->list); in qxl_ttm_bo_destroy()
45 drm_gem_object_release(&bo->tbo.base); in qxl_ttm_bo_destroy()
46 kfree(bo); in qxl_ttm_bo_destroy()
49 bool qxl_ttm_bo_is_qxl_bo(struct ttm_buffer_object *bo) in qxl_ttm_bo_is_qxl_bo() argument
51 if (bo->destroy == &qxl_ttm_bo_destroy) in qxl_ttm_bo_is_qxl_bo()
[all …]
/linux-6.12.1/drivers/gpu/drm/radeon/
Dradeon_object.c45 static void radeon_bo_clear_surface_reg(struct radeon_bo *bo);
54 struct radeon_bo *bo; in radeon_ttm_bo_destroy() local
56 bo = container_of(tbo, struct radeon_bo, tbo); in radeon_ttm_bo_destroy()
58 mutex_lock(&bo->rdev->gem.mutex); in radeon_ttm_bo_destroy()
59 list_del_init(&bo->list); in radeon_ttm_bo_destroy()
60 mutex_unlock(&bo->rdev->gem.mutex); in radeon_ttm_bo_destroy()
61 radeon_bo_clear_surface_reg(bo); in radeon_ttm_bo_destroy()
62 WARN_ON_ONCE(!list_empty(&bo->va)); in radeon_ttm_bo_destroy()
63 if (bo->tbo.base.import_attach) in radeon_ttm_bo_destroy()
64 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); in radeon_ttm_bo_destroy()
[all …]
Dradeon_object.h64 static inline int radeon_bo_reserve(struct radeon_bo *bo, bool no_intr) in radeon_bo_reserve() argument
68 r = ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL); in radeon_bo_reserve()
71 dev_err(bo->rdev->dev, "%p reserve failed\n", bo); in radeon_bo_reserve()
77 static inline void radeon_bo_unreserve(struct radeon_bo *bo) in radeon_bo_unreserve() argument
79 ttm_bo_unreserve(&bo->tbo); in radeon_bo_unreserve()
91 static inline u64 radeon_bo_gpu_offset(struct radeon_bo *bo) in radeon_bo_gpu_offset() argument
96 rdev = radeon_get_rdev(bo->tbo.bdev); in radeon_bo_gpu_offset()
98 switch (bo->tbo.resource->mem_type) { in radeon_bo_gpu_offset()
107 return (bo->tbo.resource->start << PAGE_SHIFT) + start; in radeon_bo_gpu_offset()
110 static inline unsigned long radeon_bo_size(struct radeon_bo *bo) in radeon_bo_size() argument
[all …]
/linux-6.12.1/drivers/gpu/drm/xe/
Dxe_bo.h69 void xe_bo_free(struct xe_bo *bo);
71 struct xe_bo *___xe_bo_create_locked(struct xe_device *xe, struct xe_bo *bo,
106 int xe_bo_placement_for_flags(struct xe_device *xe, struct xe_bo *bo,
109 static inline struct xe_bo *ttm_to_xe_bo(const struct ttm_buffer_object *bo) in ttm_to_xe_bo() argument
111 return container_of(bo, struct xe_bo, ttm); in ttm_to_xe_bo()
119 #define xe_bo_device(bo) ttm_to_xe_device((bo)->ttm.bdev) argument
121 static inline struct xe_bo *xe_bo_get(struct xe_bo *bo) in xe_bo_get() argument
123 if (bo) in xe_bo_get()
124 drm_gem_object_get(&bo->ttm.base); in xe_bo_get()
126 return bo; in xe_bo_get()
[all …]
Dxe_bo.c87 bool xe_bo_is_vram(struct xe_bo *bo) in xe_bo_is_vram() argument
89 return resource_is_vram(bo->ttm.resource) || in xe_bo_is_vram()
90 resource_is_stolen_vram(xe_bo_device(bo), bo->ttm.resource); in xe_bo_is_vram()
93 bool xe_bo_is_stolen(struct xe_bo *bo) in xe_bo_is_stolen() argument
95 return bo->ttm.resource->mem_type == XE_PL_STOLEN; in xe_bo_is_stolen()
107 bool xe_bo_has_single_placement(struct xe_bo *bo) in xe_bo_has_single_placement() argument
109 return bo->placement.num_placement == 1; in xe_bo_has_single_placement()
121 bool xe_bo_is_stolen_devmem(struct xe_bo *bo) in xe_bo_is_stolen_devmem() argument
123 return xe_bo_is_stolen(bo) && in xe_bo_is_stolen_devmem()
124 GRAPHICS_VERx100(xe_bo_device(bo)) >= 1270; in xe_bo_is_stolen_devmem()
[all …]
Dxe_bo_evict.c30 struct xe_bo *bo; in xe_bo_evict_all() local
64 bo = list_first_entry_or_null(&xe->pinned.external_vram, in xe_bo_evict_all()
65 typeof(*bo), pinned_link); in xe_bo_evict_all()
66 if (!bo) in xe_bo_evict_all()
68 xe_bo_get(bo); in xe_bo_evict_all()
69 list_move_tail(&bo->pinned_link, &still_in_list); in xe_bo_evict_all()
72 xe_bo_lock(bo, false); in xe_bo_evict_all()
73 ret = xe_bo_evict_pinned(bo); in xe_bo_evict_all()
74 xe_bo_unlock(bo); in xe_bo_evict_all()
75 xe_bo_put(bo); in xe_bo_evict_all()
[all …]
/linux-6.12.1/drivers/gpu/drm/tegra/
Dgem.c52 static void tegra_bo_put(struct host1x_bo *bo) in tegra_bo_put() argument
54 struct tegra_bo *obj = host1x_to_tegra_bo(bo); in tegra_bo_put()
59 static struct host1x_bo_mapping *tegra_bo_pin(struct device *dev, struct host1x_bo *bo, in tegra_bo_pin() argument
62 struct tegra_bo *obj = host1x_to_tegra_bo(bo); in tegra_bo_pin()
72 map->bo = host1x_bo_get(bo); in tegra_bo_pin()
173 host1x_bo_put(map->bo); in tegra_bo_unpin()
177 static void *tegra_bo_mmap(struct host1x_bo *bo) in tegra_bo_mmap() argument
179 struct tegra_bo *obj = host1x_to_tegra_bo(bo); in tegra_bo_mmap()
203 static void tegra_bo_munmap(struct host1x_bo *bo, void *addr) in tegra_bo_munmap() argument
205 struct tegra_bo *obj = host1x_to_tegra_bo(bo); in tegra_bo_munmap()
[all …]
/linux-6.12.1/drivers/gpu/drm/amd/amdgpu/
Damdgpu_object.c59 struct amdgpu_bo *bo = ttm_to_amdgpu_bo(tbo); in amdgpu_bo_destroy() local
61 amdgpu_bo_kunmap(bo); in amdgpu_bo_destroy()
63 if (bo->tbo.base.import_attach) in amdgpu_bo_destroy()
64 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); in amdgpu_bo_destroy()
65 drm_gem_object_release(&bo->tbo.base); in amdgpu_bo_destroy()
66 amdgpu_bo_unref(&bo->parent); in amdgpu_bo_destroy()
67 kvfree(bo); in amdgpu_bo_destroy()
72 struct amdgpu_bo *bo = ttm_to_amdgpu_bo(tbo); in amdgpu_bo_user_destroy() local
75 ubo = to_amdgpu_bo_user(bo); in amdgpu_bo_user_destroy()
90 bool amdgpu_bo_is_amdgpu_bo(struct ttm_buffer_object *bo) in amdgpu_bo_is_amdgpu_bo() argument
[all …]
Damdgpu_dma_buf.c57 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); in amdgpu_dma_buf_attach() local
58 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_dma_buf_attach()
76 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); in amdgpu_dma_buf_pin() local
79 return amdgpu_bo_pin(bo, AMDGPU_GEM_DOMAIN_GTT); in amdgpu_dma_buf_pin()
92 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); in amdgpu_dma_buf_unpin() local
94 amdgpu_bo_unpin(bo); in amdgpu_dma_buf_unpin()
115 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); in amdgpu_dma_buf_map() local
116 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_dma_buf_map()
120 if (!bo->tbo.pin_count) { in amdgpu_dma_buf_map()
125 if (bo->preferred_domains & AMDGPU_GEM_DOMAIN_VRAM && in amdgpu_dma_buf_map()
[all …]
Damdgpu_object.h45 #define to_amdgpu_bo_user(abo) container_of((abo), struct amdgpu_bo_user, bo)
46 #define to_amdgpu_bo_vm(abo) container_of((abo), struct amdgpu_bo_vm, bo)
58 void (*destroy)(struct ttm_buffer_object *bo);
129 struct amdgpu_bo bo; member
138 struct amdgpu_bo bo; member
212 static inline int amdgpu_bo_reserve(struct amdgpu_bo *bo, bool no_intr) in amdgpu_bo_reserve() argument
214 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_bo_reserve()
217 r = ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL); in amdgpu_bo_reserve()
220 dev_err(adev->dev, "%p reserve failed\n", bo); in amdgpu_bo_reserve()
226 static inline void amdgpu_bo_unreserve(struct amdgpu_bo *bo) in amdgpu_bo_unreserve() argument
[all …]
/linux-6.12.1/drivers/gpu/drm/vc4/
Dvc4_bo.c134 struct vc4_bo *bo = to_vc4_bo(gem_obj); in vc4_bo_set_label() local
144 vc4->bo_labels[bo->label].num_allocated--; in vc4_bo_set_label()
145 vc4->bo_labels[bo->label].size_allocated -= gem_obj->size; in vc4_bo_set_label()
147 if (vc4->bo_labels[bo->label].num_allocated == 0 && in vc4_bo_set_label()
148 is_user_label(bo->label)) { in vc4_bo_set_label()
154 kfree(vc4->bo_labels[bo->label].name); in vc4_bo_set_label()
155 vc4->bo_labels[bo->label].name = NULL; in vc4_bo_set_label()
158 bo->label = label; in vc4_bo_set_label()
166 static void vc4_bo_destroy(struct vc4_bo *bo) in vc4_bo_destroy() argument
168 struct drm_gem_object *obj = &bo->base.base; in vc4_bo_destroy()
[all …]
/linux-6.12.1/include/drm/ttm/
Dttm_bo.h164 struct ttm_buffer_object *bo; member
208 s64 (*process_bo)(struct ttm_lru_walk *walk, struct ttm_buffer_object *bo);
233 static inline void ttm_bo_get(struct ttm_buffer_object *bo) in ttm_bo_get() argument
235 kref_get(&bo->kref); in ttm_bo_get()
249 ttm_bo_get_unless_zero(struct ttm_buffer_object *bo) in ttm_bo_get_unless_zero() argument
251 if (!kref_get_unless_zero(&bo->kref)) in ttm_bo_get_unless_zero()
253 return bo; in ttm_bo_get_unless_zero()
278 static inline int ttm_bo_reserve(struct ttm_buffer_object *bo, in ttm_bo_reserve() argument
290 success = dma_resv_trylock(bo->base.resv); in ttm_bo_reserve()
295 ret = dma_resv_lock_interruptible(bo->base.resv, ticket); in ttm_bo_reserve()
[all …]
/linux-6.12.1/drivers/gpu/drm/lima/
Dlima_gem.c21 int lima_heap_alloc(struct lima_bo *bo, struct lima_vm *vm) in lima_heap_alloc() argument
24 struct address_space *mapping = bo->base.base.filp->f_mapping; in lima_heap_alloc()
25 struct device *dev = bo->base.base.dev->dev; in lima_heap_alloc()
26 size_t old_size = bo->heap_size; in lima_heap_alloc()
27 size_t new_size = bo->heap_size ? bo->heap_size * 2 : in lima_heap_alloc()
32 if (bo->heap_size >= bo->base.base.size) in lima_heap_alloc()
35 new_size = min(new_size, bo->base.base.size); in lima_heap_alloc()
37 dma_resv_lock(bo->base.base.resv, NULL); in lima_heap_alloc()
39 if (bo->base.pages) { in lima_heap_alloc()
40 pages = bo->base.pages; in lima_heap_alloc()
[all …]
/linux-6.12.1/drivers/gpu/drm/ttm/tests/
Dttm_bo_test.c62 struct ttm_buffer_object *bo; in ttm_bo_reserve_optimistic_no_ticket() local
65 bo = ttm_bo_kunit_init(test, test->priv, BO_SIZE, NULL); in ttm_bo_reserve_optimistic_no_ticket()
67 err = ttm_bo_reserve(bo, params->interruptible, params->no_wait, NULL); in ttm_bo_reserve_optimistic_no_ticket()
70 dma_resv_unlock(bo->base.resv); in ttm_bo_reserve_optimistic_no_ticket()
75 struct ttm_buffer_object *bo; in ttm_bo_reserve_locked_no_sleep() local
80 bo = ttm_bo_kunit_init(test, test->priv, BO_SIZE, NULL); in ttm_bo_reserve_locked_no_sleep()
83 dma_resv_lock(bo->base.resv, NULL); in ttm_bo_reserve_locked_no_sleep()
85 err = ttm_bo_reserve(bo, interruptible, no_wait, NULL); in ttm_bo_reserve_locked_no_sleep()
86 dma_resv_unlock(bo->base.resv); in ttm_bo_reserve_locked_no_sleep()
93 struct ttm_buffer_object *bo; in ttm_bo_reserve_no_wait_ticket() local
[all …]
Dttm_tt_test.c42 struct ttm_buffer_object *bo; in ttm_tt_init_basic() local
53 bo = ttm_bo_kunit_init(test, test->priv, params->size, NULL); in ttm_tt_init_basic()
55 err = ttm_tt_init(tt, bo, page_flags, caching, extra_pages); in ttm_tt_init_basic()
69 struct ttm_buffer_object *bo; in ttm_tt_init_misaligned() local
79 bo = ttm_bo_kunit_init(test, test->priv, size, NULL); in ttm_tt_init_misaligned()
82 bo->base.size += 1; in ttm_tt_init_misaligned()
84 err = ttm_tt_init(tt, bo, 0, caching, 0); in ttm_tt_init_misaligned()
92 struct ttm_buffer_object *bo; in ttm_tt_fini_basic() local
100 bo = ttm_bo_kunit_init(test, test->priv, BO_SIZE, NULL); in ttm_tt_fini_basic()
102 err = ttm_tt_init(tt, bo, 0, caching, 0); in ttm_tt_fini_basic()
[all …]
/linux-6.12.1/drivers/gpu/drm/panfrost/
Dpanfrost_gem.c19 struct panfrost_gem_object *bo = to_panfrost_bo(obj); in panfrost_gem_free_object() local
29 list_del_init(&bo->base.madv_list); in panfrost_gem_free_object()
36 WARN_ON_ONCE(!list_empty(&bo->mappings.list)); in panfrost_gem_free_object()
38 if (bo->sgts) { in panfrost_gem_free_object()
40 int n_sgt = bo->base.base.size / SZ_2M; in panfrost_gem_free_object()
43 if (bo->sgts[i].sgl) { in panfrost_gem_free_object()
44 dma_unmap_sgtable(pfdev->dev, &bo->sgts[i], in panfrost_gem_free_object()
46 sg_free_table(&bo->sgts[i]); in panfrost_gem_free_object()
49 kvfree(bo->sgts); in panfrost_gem_free_object()
52 drm_gem_shmem_free(&bo->base); in panfrost_gem_free_object()
[all …]
/linux-6.12.1/drivers/gpu/drm/panthor/
Dpanthor_gem.c18 struct panthor_gem_object *bo = to_panthor_bo(obj); in panthor_gem_free_object() local
19 struct drm_gem_object *vm_root_gem = bo->exclusive_vm_root_gem; in panthor_gem_free_object()
21 drm_gem_free_mmap_offset(&bo->base.base); in panthor_gem_free_object()
22 mutex_destroy(&bo->gpuva_list_lock); in panthor_gem_free_object()
23 drm_gem_shmem_free(&bo->base); in panthor_gem_free_object()
32 void panthor_kernel_bo_destroy(struct panthor_kernel_bo *bo) in panthor_kernel_bo_destroy() argument
37 if (IS_ERR_OR_NULL(bo)) in panthor_kernel_bo_destroy()
40 vm = bo->vm; in panthor_kernel_bo_destroy()
41 panthor_kernel_bo_vunmap(bo); in panthor_kernel_bo_destroy()
43 if (drm_WARN_ON(bo->obj->dev, in panthor_kernel_bo_destroy()
[all …]
/linux-6.12.1/drivers/gpu/drm/xe/display/
Dintel_fb_bo.c14 void intel_fb_bo_framebuffer_fini(struct xe_bo *bo) in intel_fb_bo_framebuffer_fini() argument
16 if (bo->flags & XE_BO_FLAG_PINNED) { in intel_fb_bo_framebuffer_fini()
18 xe_bo_lock(bo, false); in intel_fb_bo_framebuffer_fini()
19 xe_bo_unpin(bo); in intel_fb_bo_framebuffer_fini()
20 xe_bo_unlock(bo); in intel_fb_bo_framebuffer_fini()
22 xe_bo_put(bo); in intel_fb_bo_framebuffer_fini()
26 struct xe_bo *bo, in intel_fb_bo_framebuffer_init() argument
29 struct xe_device *xe = to_xe_device(bo->ttm.base.dev); in intel_fb_bo_framebuffer_init()
37 !(bo->flags & XE_BO_FLAG_NEEDS_64K))) in intel_fb_bo_framebuffer_init()
40 xe_bo_get(bo); in intel_fb_bo_framebuffer_init()
[all …]

12345678910>>...17