/linux-6.12.1/fs/ocfs2/ |
D | reservations.c | 40 struct ocfs2_alloc_reservation *resv) in ocfs2_resv_window_bits() argument 45 if (!(resv->r_flags & OCFS2_RESV_FLAG_DIR)) { in ocfs2_resv_window_bits() 54 static inline unsigned int ocfs2_resv_end(struct ocfs2_alloc_reservation *resv) in ocfs2_resv_end() argument 56 if (resv->r_len) in ocfs2_resv_end() 57 return resv->r_start + resv->r_len - 1; in ocfs2_resv_end() 58 return resv->r_start; in ocfs2_resv_end() 61 static inline int ocfs2_resv_empty(struct ocfs2_alloc_reservation *resv) in ocfs2_resv_empty() argument 63 return !!(resv->r_len == 0); in ocfs2_resv_empty() 77 struct ocfs2_alloc_reservation *resv; in ocfs2_dump_resv() local 85 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_dump_resv() [all …]
|
D | reservations.h | 54 void ocfs2_resv_init_once(struct ocfs2_alloc_reservation *resv); 57 void ocfs2_resv_set_type(struct ocfs2_alloc_reservation *resv, 71 struct ocfs2_alloc_reservation *resv); 122 struct ocfs2_alloc_reservation *resv, 141 struct ocfs2_alloc_reservation *resv,
|
/linux-6.12.1/drivers/dma-buf/ |
D | st-dma-resv.c | 40 struct dma_resv resv; in sanitycheck() local 53 dma_resv_init(&resv); in sanitycheck() 54 r = dma_resv_lock(&resv, NULL); in sanitycheck() 58 dma_resv_unlock(&resv); in sanitycheck() 59 dma_resv_fini(&resv); in sanitycheck() 66 struct dma_resv resv; in test_signaling() local 76 dma_resv_init(&resv); in test_signaling() 77 r = dma_resv_lock(&resv, NULL); in test_signaling() 83 r = dma_resv_reserve_fences(&resv, 1); in test_signaling() 89 dma_resv_add_fence(&resv, f, usage); in test_signaling() [all …]
|
D | dma-buf.c | 104 if (dmabuf->resv == (struct dma_resv *)&dmabuf[1]) in dma_buf_release() 105 dma_resv_fini(dmabuf->resv); in dma_buf_release() 234 static bool dma_buf_poll_add_cb(struct dma_resv *resv, bool write, in dma_buf_poll_add_cb() argument 241 dma_resv_for_each_fence(&cursor, resv, dma_resv_usage_rw(write), in dma_buf_poll_add_cb() 256 struct dma_resv *resv; in dma_buf_poll() local 260 if (!dmabuf || !dmabuf->resv) in dma_buf_poll() 263 resv = dmabuf->resv; in dma_buf_poll() 271 dma_resv_lock(resv, NULL); in dma_buf_poll() 288 if (!dma_buf_poll_add_cb(resv, true, dcb)) in dma_buf_poll() 311 if (!dma_buf_poll_add_cb(resv, false, dcb)) in dma_buf_poll() [all …]
|
/linux-6.12.1/fs/xfs/scrub/ |
D | newbt.c | 89 enum xfs_ag_resv_type resv) in xrep_newbt_init_ag() argument 95 xnr->resv = resv; in xrep_newbt_init_ag() 147 struct xrep_newbt_resv *resv; in xrep_newbt_add_blocks() local 150 resv = kmalloc(sizeof(struct xrep_newbt_resv), XCHK_GFP_FLAGS); in xrep_newbt_add_blocks() 151 if (!resv) in xrep_newbt_add_blocks() 154 INIT_LIST_HEAD(&resv->list); in xrep_newbt_add_blocks() 155 resv->agbno = XFS_FSB_TO_AGBNO(mp, args->fsbno); in xrep_newbt_add_blocks() 156 resv->len = args->len; in xrep_newbt_add_blocks() 157 resv->used = 0; in xrep_newbt_add_blocks() 158 resv->pag = xfs_perag_hold(pag); in xrep_newbt_add_blocks() [all …]
|
D | alloc_repair.c | 534 struct xrep_newbt_resv *resv) in xrep_abt_dispose_one() argument 538 xfs_agblock_t free_agbno = resv->agbno + resv->used; in xrep_abt_dispose_one() 539 xfs_extlen_t free_aglen = resv->len - resv->used; in xrep_abt_dispose_one() 542 ASSERT(pag == resv->pag); in xrep_abt_dispose_one() 545 if (resv->used > 0) in xrep_abt_dispose_one() 546 xfs_rmap_alloc_extent(sc->tp, pag->pag_agno, resv->agbno, in xrep_abt_dispose_one() 547 resv->used, XFS_RMAP_OWN_AG); in xrep_abt_dispose_one() 557 trace_xrep_newbt_free_blocks(sc->mp, resv->pag->pag_agno, free_agbno, in xrep_abt_dispose_one() 560 error = __xfs_free_extent(sc->tp, resv->pag, free_agbno, free_aglen, in xrep_abt_dispose_one() 579 struct xrep_newbt_resv *resv, *n; in xrep_abt_dispose_reservations() local [all …]
|
/linux-6.12.1/fs/xfs/libxfs/ |
D | xfs_log_rlimit.c | 84 struct xfs_trans_resv *resv) in xfs_log_calc_trans_resv_for_minlogblocks() argument 93 xfs_trans_resv_calc(mp, resv); in xfs_log_calc_trans_resv_for_minlogblocks() 108 xfs_trans_resv_calc(mp, resv); in xfs_log_calc_trans_resv_for_minlogblocks() 115 resv->tr_write.tr_logcount = XFS_WRITE_LOG_COUNT_REFLINK; in xfs_log_calc_trans_resv_for_minlogblocks() 116 resv->tr_itruncate.tr_logcount = in xfs_log_calc_trans_resv_for_minlogblocks() 118 resv->tr_qm_dqalloc.tr_logcount = XFS_WRITE_LOG_COUNT_REFLINK; in xfs_log_calc_trans_resv_for_minlogblocks() 124 resv->tr_write.tr_logcount = XFS_WRITE_LOG_COUNT; in xfs_log_calc_trans_resv_for_minlogblocks() 125 resv->tr_itruncate.tr_logcount = XFS_ITRUNCATE_LOG_COUNT; in xfs_log_calc_trans_resv_for_minlogblocks() 126 resv->tr_qm_dqalloc.tr_logcount = XFS_WRITE_LOG_COUNT; in xfs_log_calc_trans_resv_for_minlogblocks() 134 resv->tr_write.tr_logres = in xfs_log_calc_trans_resv_for_minlogblocks() [all …]
|
D | xfs_ag_resv.c | 134 struct xfs_ag_resv *resv; in __xfs_ag_resv_free() local 139 resv = xfs_perag_resv(pag, type); in __xfs_ag_resv_free() 141 pag->pag_mount->m_ag_max_usable += resv->ar_asked; in __xfs_ag_resv_free() 148 oldresv = resv->ar_orig_reserved; in __xfs_ag_resv_free() 150 oldresv = resv->ar_reserved; in __xfs_ag_resv_free() 152 resv->ar_reserved = 0; in __xfs_ag_resv_free() 153 resv->ar_asked = 0; in __xfs_ag_resv_free() 154 resv->ar_orig_reserved = 0; in __xfs_ag_resv_free() 174 struct xfs_ag_resv *resv; in __xfs_ag_resv_init() local 226 resv = xfs_perag_resv(pag, type); in __xfs_ag_resv_init() [all …]
|
/linux-6.12.1/drivers/gpu/drm/ttm/tests/ |
D | ttm_bo_test.c | 70 dma_resv_unlock(bo->base.resv); in ttm_bo_reserve_optimistic_no_ticket() 83 dma_resv_lock(bo->base.resv, NULL); in ttm_bo_reserve_locked_no_sleep() 86 dma_resv_unlock(bo->base.resv); in ttm_bo_reserve_locked_no_sleep() 126 dma_resv_unlock(bo->base.resv); in ttm_bo_reserve_double_resv() 151 ww_mutex_base_lock(&bo2->base.resv->lock.base); in ttm_bo_reserve_deadlock() 154 lock_release(&bo2->base.resv->lock.base.dep_map, 1); in ttm_bo_reserve_deadlock() 156 bo2->base.resv->lock.ctx = &ctx2; in ttm_bo_reserve_deadlock() 166 dma_resv_unlock(bo1->base.resv); in ttm_bo_reserve_deadlock() 225 mutex_lock(&bo->base.resv->lock.base); in ttm_bo_reserve_interrupted() 231 mutex_unlock(&bo->base.resv->lock.base); in ttm_bo_reserve_interrupted() [all …]
|
D | ttm_bo_validate_test.c | 66 struct dma_resv *resv, in dma_resv_kunit_active_fence_init() argument 74 dma_resv_lock(resv, NULL); in dma_resv_kunit_active_fence_init() 75 dma_resv_reserve_fences(resv, 1); in dma_resv_kunit_active_fence_init() 76 dma_resv_add_fence(resv, fence, usage); in dma_resv_kunit_active_fence_init() 77 dma_resv_unlock(resv); in dma_resv_kunit_active_fence_init() 127 dma_resv_unlock(bo->base.resv); in ttm_bo_init_reserved_sys_man() 139 KUNIT_EXPECT_NOT_NULL(test, (void *)bo->base.resv->fences); in ttm_bo_init_reserved_sys_man() 176 dma_resv_unlock(bo->base.resv); in ttm_bo_init_reserved_mock_man() 202 struct dma_resv resv; in ttm_bo_init_reserved_resv() local 212 dma_resv_init(&resv); in ttm_bo_init_reserved_resv() [all …]
|
/linux-6.12.1/drivers/gpu/drm/ttm/ |
D | ttm_bo.c | 76 dma_resv_assert_held(bo->base.resv); in ttm_bo_move_to_lru_tail() 100 dma_resv_assert_held(bo->base.resv); in ttm_bo_set_bulk_move() 148 ret = dma_resv_reserve_fences(bo->base.resv, 1); in ttm_bo_handle_move_mem() 190 if (bo->base.resv == &bo->base._resv) in ttm_bo_individualize_resv() 195 r = dma_resv_copy_fences(&bo->base._resv, bo->base.resv); in ttm_bo_individualize_resv() 206 bo->base.resv = &bo->base._resv; in ttm_bo_individualize_resv() 215 struct dma_resv *resv = &bo->base._resv; in ttm_bo_flush_all_fences() local 219 dma_resv_iter_begin(&cursor, resv, DMA_RESV_USAGE_BOOKKEEP); in ttm_bo_flush_all_fences() 237 dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP, false, in ttm_bo_delayed_delete() 239 dma_resv_lock(bo->base.resv, NULL); in ttm_bo_delayed_delete() [all …]
|
D | ttm_execbuf_util.c | 38 dma_resv_unlock(bo->base.resv); in ttm_eu_backoff_reservation_reverse() 54 dma_resv_unlock(bo->base.resv); in ttm_eu_backoff_reservation() 102 ret = dma_resv_reserve_fences(bo->base.resv, in ttm_eu_reserve_buffers() 119 ret = dma_resv_reserve_fences(bo->base.resv, in ttm_eu_reserve_buffers() 153 dma_resv_add_fence(bo->base.resv, fence, entry->num_shared ? in ttm_eu_fence_buffer_objects() 156 dma_resv_unlock(bo->base.resv); in ttm_eu_fence_buffer_objects()
|
D | ttm_bo_util.c | 250 fbo->base.base.resv = &fbo->base.base._resv; in ttm_buffer_object_transfer() 474 dma_resv_assert_held(bo->base.resv); in ttm_bo_vmap() 542 dma_resv_assert_held(bo->base.resv); in ttm_bo_vunmap() 562 ret = dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP, in ttm_bo_wait_free_node() 662 dma_resv_add_fence(bo->base.resv, fence, DMA_RESV_USAGE_KERNEL); in ttm_bo_move_accel_cleanup() 721 if (dma_resv_test_signaled(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP)) { in ttm_bo_pipeline_gutting() 755 ret = dma_resv_copy_fences(&ghost->base._resv, bo->base.resv); in ttm_bo_pipeline_gutting() 758 dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP, in ttm_bo_pipeline_gutting() 780 if (dma_resv_trylock(bo->base.resv)) { in ttm_lru_walk_trylock() 785 if (bo->base.resv == ctx->resv && ctx->allow_res_evict) { in ttm_lru_walk_trylock() [all …]
|
/linux-6.12.1/drivers/gpu/drm/ |
D | drm_gem_shmem_helper.c | 144 dma_resv_lock(shmem->base.resv, NULL); in drm_gem_shmem_free() 159 dma_resv_unlock(shmem->base.resv); in drm_gem_shmem_free() 172 dma_resv_assert_held(shmem->base.resv); in drm_gem_shmem_get_pages() 210 dma_resv_assert_held(shmem->base.resv); in drm_gem_shmem_put_pages() 234 dma_resv_assert_held(shmem->base.resv); in drm_gem_shmem_pin_locked() 246 dma_resv_assert_held(shmem->base.resv); in drm_gem_shmem_unpin_locked() 269 ret = dma_resv_lock_interruptible(shmem->base.resv, NULL); in drm_gem_shmem_pin() 273 dma_resv_unlock(shmem->base.resv); in drm_gem_shmem_pin() 292 dma_resv_lock(shmem->base.resv, NULL); in drm_gem_shmem_unpin() 294 dma_resv_unlock(shmem->base.resv); in drm_gem_shmem_unpin() [all …]
|
D | drm_exec.c | 60 dma_resv_unlock(obj->resv); in drm_exec_unlock_all() 173 ret = dma_resv_lock_slow_interruptible(obj->resv, in drm_exec_lock_contended() 178 dma_resv_lock_slow(obj->resv, &exec->ticket); in drm_exec_lock_contended() 189 dma_resv_unlock(obj->resv); in drm_exec_lock_contended() 222 ret = dma_resv_lock_interruptible(obj->resv, &exec->ticket); in drm_exec_lock_obj() 224 ret = dma_resv_lock(obj->resv, &exec->ticket); in drm_exec_lock_obj() 246 dma_resv_unlock(obj->resv); in drm_exec_lock_obj() 266 dma_resv_unlock(obj->resv); in drm_exec_unlock_obj() 298 ret = dma_resv_reserve_fences(obj->resv, num_fences); in drm_exec_prepare_obj()
|
D | drm_gem.c | 164 if (!obj->resv) in drm_gem_private_object_init() 165 obj->resv = &obj->_resv; in drm_gem_private_object_init() 779 ret = dma_resv_wait_timeout(obj->resv, dma_resv_usage_rw(wait_all), in drm_gem_dma_resv_wait() 1179 dma_resv_lock(obj->resv, NULL); in drm_gem_pin() 1181 dma_resv_unlock(obj->resv); in drm_gem_pin() 1188 dma_resv_lock(obj->resv, NULL); in drm_gem_unpin() 1190 dma_resv_unlock(obj->resv); in drm_gem_unpin() 1197 dma_resv_assert_held(obj->resv); in drm_gem_vmap() 1214 dma_resv_assert_held(obj->resv); in drm_gem_vunmap() 1229 dma_resv_lock(obj->resv, NULL); in drm_gem_lock() [all …]
|
/linux-6.12.1/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_dma_buf.c | 289 struct dma_resv *resv = dma_buf->resv; in amdgpu_dma_buf_create_obj() local 296 dma_resv_lock(resv, NULL); in amdgpu_dma_buf_create_obj() 309 ttm_bo_type_sg, resv, &gobj, 0); in amdgpu_dma_buf_create_obj() 317 dma_resv_unlock(resv); in amdgpu_dma_buf_create_obj() 321 dma_resv_unlock(resv); in amdgpu_dma_buf_create_obj() 337 struct ww_acquire_ctx *ticket = dma_resv_locking_ctx(obj->resv); in amdgpu_dma_buf_move_notify() 360 struct dma_resv *resv = vm->root.bo->tbo.base.resv; in amdgpu_dma_buf_move_notify() local 367 r = dma_resv_lock(resv, ticket); in amdgpu_dma_buf_move_notify() 376 if (!dma_resv_trylock(resv)) in amdgpu_dma_buf_move_notify() 381 r = dma_resv_reserve_fences(resv, 2); in amdgpu_dma_buf_move_notify() [all …]
|
/linux-6.12.1/include/uapi/linux/ |
D | io_uring.h | 533 __u32 resv[3]; member 631 __u32 resv; member 651 __u32 resv; member 657 __u32 resv; member 671 __u8 resv; member 679 __u16 resv; member 691 __u8 resv; member 714 __u16 resv; member 761 __u64 resv[3]; member 768 __u32 resv[8]; member [all …]
|
/linux-6.12.1/drivers/gpu/drm/vgem/ |
D | vgem_fence.c | 131 struct dma_resv *resv; in vgem_fence_attach_ioctl() local 154 resv = obj->resv; in vgem_fence_attach_ioctl() 156 if (!dma_resv_test_signaled(resv, usage)) { in vgem_fence_attach_ioctl() 162 dma_resv_lock(resv, NULL); in vgem_fence_attach_ioctl() 163 ret = dma_resv_reserve_fences(resv, 1); in vgem_fence_attach_ioctl() 165 dma_resv_add_fence(resv, fence, arg->flags & VGEM_FENCE_WRITE ? in vgem_fence_attach_ioctl() 167 dma_resv_unlock(resv); in vgem_fence_attach_ioctl()
|
/linux-6.12.1/include/drm/ttm/ |
D | ttm_bo.h | 189 struct dma_resv *resv; member 290 success = dma_resv_trylock(bo->base.resv); in ttm_bo_reserve() 295 ret = dma_resv_lock_interruptible(bo->base.resv, ticket); in ttm_bo_reserve() 297 ret = dma_resv_lock(bo->base.resv, ticket); in ttm_bo_reserve() 318 int ret = dma_resv_lock_slow_interruptible(bo->base.resv, in ttm_bo_reserve_slowpath() 324 dma_resv_lock_slow(bo->base.resv, ticket); in ttm_bo_reserve_slowpath() 369 dma_resv_unlock(bo->base.resv); in ttm_bo_unreserve() 403 struct sg_table *sg, struct dma_resv *resv, 408 struct sg_table *sg, struct dma_resv *resv,
|
/linux-6.12.1/drivers/gpu/drm/radeon/ |
D | radeon_prime.c | 49 struct dma_resv *resv = attach->dmabuf->resv; in radeon_gem_prime_import_sg_table() local 54 dma_resv_lock(resv, NULL); in radeon_gem_prime_import_sg_table() 56 RADEON_GEM_DOMAIN_GTT, 0, sg, resv, &bo); in radeon_gem_prime_import_sg_table() 57 dma_resv_unlock(resv); in radeon_gem_prime_import_sg_table()
|
/linux-6.12.1/drivers/gpu/drm/i915/gem/ |
D | i915_gem_wait.c | 36 i915_gem_object_boost(struct dma_resv *resv, unsigned int flags) in i915_gem_object_boost() argument 57 dma_resv_iter_begin(&cursor, resv, in i915_gem_object_boost() 67 i915_gem_object_wait_reservation(struct dma_resv *resv, in i915_gem_object_wait_reservation() argument 75 i915_gem_object_boost(resv, flags); in i915_gem_object_wait_reservation() 77 dma_resv_iter_begin(&cursor, resv, in i915_gem_object_wait_reservation() 155 dma_resv_iter_begin(&cursor, obj->base.resv, in i915_gem_object_wait_priority() 177 timeout = i915_gem_object_wait_reservation(obj->base.resv, in i915_gem_object_wait()
|
/linux-6.12.1/tools/include/uapi/linux/ |
D | io_uring.h | 493 __u32 resv[3]; member 577 __u32 resv; member 597 __u32 resv; member 603 __u32 resv; member 617 __u8 resv; member 625 __u16 resv; member 637 __u8 resv; member 645 __u16 resv; member 684 __u64 resv[3]; member 733 __u64 resv; member
|
/linux-6.12.1/kernel/irq/ |
D | affinity.c | 113 unsigned int resv = affd->pre_vectors + affd->post_vectors; in irq_calc_affinity_vectors() local 116 if (resv > minvec) in irq_calc_affinity_vectors() 120 set_vecs = maxvec - resv; in irq_calc_affinity_vectors() 127 return resv + min(set_vecs, maxvec - resv); in irq_calc_affinity_vectors()
|
/linux-6.12.1/drivers/gpu/drm/loongson/ |
D | lsdc_gem.c | 22 dma_resv_assert_held(obj->resv); in lsdc_gem_prime_pin() 35 dma_resv_assert_held(obj->resv); in lsdc_gem_prime_unpin() 145 struct dma_resv *resv) in lsdc_gem_object_create() argument 152 lbo = lsdc_bo_create(ddev, domain, size, kerenl, sg, resv); in lsdc_gem_object_create() 179 struct dma_resv *resv = attach->dmabuf->resv; in lsdc_prime_import_sg_table() local 184 dma_resv_lock(resv, NULL); in lsdc_prime_import_sg_table() 186 sg, resv); in lsdc_prime_import_sg_table() 187 dma_resv_unlock(resv); in lsdc_prime_import_sg_table()
|