Home
last modified time | relevance | path

Searched refs:obj_size (Results 1 – 25 of 124) sorted by relevance

12345

/linux-6.12.1/include/linux/
Dgeneric-radix-tree.h158 static inline size_t __idx_to_offset(size_t idx, size_t obj_size) in __idx_to_offset() argument
160 if (__builtin_constant_p(obj_size)) in __idx_to_offset()
161 BUILD_BUG_ON(obj_size > GENRADIX_NODE_SIZE); in __idx_to_offset()
163 BUG_ON(obj_size > GENRADIX_NODE_SIZE); in __idx_to_offset()
165 if (!is_power_of_2(obj_size)) { in __idx_to_offset()
166 size_t objs_per_page = GENRADIX_NODE_SIZE / obj_size; in __idx_to_offset()
169 (idx % objs_per_page) * obj_size; in __idx_to_offset()
171 return idx * obj_size; in __idx_to_offset()
313 size_t obj_size) in __genradix_iter_advance() argument
315 if (iter->offset + obj_size < iter->offset) { in __genradix_iter_advance()
[all …]
Dobjagg.h8 size_t obj_size; member
/linux-6.12.1/fs/xfs/scrub/
Dxfarray.c50 return div_u64((xfarray_idx_t)pos, array->obj_size); in xfarray_idx()
59 return idx * array->obj_size; in xfarray_pos()
73 size_t obj_size, in xfarray_create() argument
80 ASSERT(obj_size < PAGE_SIZE); in xfarray_create()
87 array = kzalloc(sizeof(struct xfarray) + obj_size, XCHK_GFP_FLAGS); in xfarray_create()
92 array->obj_size = obj_size; in xfarray_create()
94 if (is_power_of_2(obj_size)) in xfarray_create()
95 array->obj_size_log = ilog2(obj_size); in xfarray_create()
139 return xfile_load(array->xfile, ptr, array->obj_size, in xfarray_load()
155 error = xfile_load(array->xfile, temp, array->obj_size, pos); in xfarray_is_unset()
[all …]
Dxfarray.h34 size_t obj_size; member
41 size_t obj_size, struct xfarray **arrayp);
64 memset(rec, 0, array->obj_size); in xfarray_load_sparse()
/linux-6.12.1/drivers/gpu/drm/imagination/
Dpvr_drv.c1163 pvr_get_uobj(u64 usr_ptr, u32 usr_stride, u32 min_stride, u32 obj_size, void *out) in pvr_get_uobj() argument
1168 return copy_struct_from_user(out, obj_size, u64_to_user_ptr(usr_ptr), usr_stride); in pvr_get_uobj()
1172 pvr_set_uobj(u64 usr_ptr, u32 usr_stride, u32 min_stride, u32 obj_size, const void *in) in pvr_set_uobj() argument
1177 if (copy_to_user(u64_to_user_ptr(usr_ptr), in, min_t(u32, usr_stride, obj_size))) in pvr_set_uobj()
1180 if (usr_stride > obj_size && in pvr_set_uobj()
1181 clear_user(u64_to_user_ptr(usr_ptr + obj_size), usr_stride - obj_size)) { in pvr_set_uobj()
1189 pvr_get_uobj_array(const struct drm_pvr_obj_array *in, u32 min_stride, u32 obj_size, void **out) in pvr_get_uobj_array() argument
1200 out_alloc = kvmalloc_array(in->count, obj_size, GFP_KERNEL); in pvr_get_uobj_array()
1204 if (obj_size == in->stride) { in pvr_get_uobj_array()
1206 (unsigned long)obj_size * in->count)) in pvr_get_uobj_array()
[all …]
Dpvr_drv.h22 int pvr_get_uobj(u64 usr_ptr, u32 usr_size, u32 min_size, u32 obj_size, void *out);
23 int pvr_set_uobj(u64 usr_ptr, u32 usr_size, u32 min_size, u32 obj_size, const void *in);
24 int pvr_get_uobj_array(const struct drm_pvr_obj_array *in, u32 min_stride, u32 obj_size,
26 int pvr_set_uobj_array(const struct drm_pvr_obj_array *out, u32 min_stride, u32 obj_size,
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx4/
Dicm.c261 (MLX4_TABLE_CHUNK_SIZE / table->obj_size); in mlx4_table_get()
299 i = (obj & (table->num_obj - 1)) / (MLX4_TABLE_CHUNK_SIZE / table->obj_size); in mlx4_table_put()
328 idx = (u64) (obj & (table->num_obj - 1)) * table->obj_size; in mlx4_table_find()
385 int inc = MLX4_TABLE_CHUNK_SIZE / table->obj_size; in mlx4_table_get_range()
411 for (i = start; i <= end; i += MLX4_TABLE_CHUNK_SIZE / table->obj_size) in mlx4_table_put_range()
416 u64 virt, int obj_size, u32 nobj, int reserved, in mlx4_init_icm_table() argument
425 obj_per_chunk = MLX4_TABLE_CHUNK_SIZE / obj_size; in mlx4_init_icm_table()
436 table->obj_size = obj_size; in mlx4_init_icm_table()
441 size = (u64) nobj * obj_size; in mlx4_init_icm_table()
442 for (i = 0; i * MLX4_TABLE_CHUNK_SIZE < reserved * obj_size; ++i) { in mlx4_init_icm_table()
/linux-6.12.1/drivers/net/can/spi/mcp251xfd/
Dmcp251xfd-ring.c349 priv->rx_obj_num_coalesce_irq, rx_ring->obj_size, in mcp251xfd_ring_init()
350 priv->rx_obj_num_coalesce_irq * rx_ring->obj_size); in mcp251xfd_ring_init()
360 rx_ring->obj_size, in mcp251xfd_ring_init()
362 rx_ring->obj_size); in mcp251xfd_ring_init()
368 rx_ring->obj_num, rx_ring->obj_size, in mcp251xfd_ring_init()
369 rx_ring->obj_num * rx_ring->obj_size); in mcp251xfd_ring_init()
377 priv->tx->obj_num, priv->tx->obj_size, in mcp251xfd_ring_init()
378 priv->tx->obj_num * priv->tx->obj_size); in mcp251xfd_ring_init()
515 tx_ring->obj_size = tx_obj_size; in mcp251xfd_ring_alloc()
539 rx_ring->obj_size = rx_obj_size; in mcp251xfd_ring_alloc()
Dmcp251xfd-rx.c43 *rx_tail = fifo_ua / ring->obj_size; in mcp251xfd_rx_tail_get_from_chip()
223 len * ring->obj_size / val_bytes); in mcp251xfd_rx_obj_read()
284 i * ring->obj_size); in mcp251xfd_handle_rxif_ring()
Dmcp251xfd.h547 u8 obj_size; member
565 u8 obj_size; member
850 return ring->base + ring->obj_size * n; in mcp251xfd_get_tx_obj_addr()
856 return ring->base + ring->obj_size * n; in mcp251xfd_get_rx_obj_addr()
916 tx_ring->obj_size; in mcp251xfd_get_tx_nr_by_addr()
Dmcp251xfd-dump.c176 .val = rx->obj_size, in mcp251xfd_dump_rx_ring_one()
219 .val = tx->obj_size, in mcp251xfd_dump_tx_ring()
/linux-6.12.1/drivers/infiniband/hw/mthca/
Dmthca_memfree.c224 int i = (obj & (table->num_obj - 1)) * table->obj_size / MTHCA_TABLE_CHUNK_SIZE; in mthca_table_get()
264 i = (obj & (table->num_obj - 1)) * table->obj_size / MTHCA_TABLE_CHUNK_SIZE; in mthca_table_put()
290 idx = (obj & (table->num_obj - 1)) * table->obj_size; in mthca_table_find()
324 int inc = MTHCA_TABLE_CHUNK_SIZE / table->obj_size; in mthca_table_get_range()
352 for (i = start; i <= end; i += MTHCA_TABLE_CHUNK_SIZE / table->obj_size) in mthca_table_put_range()
357 u64 virt, int obj_size, in mthca_alloc_icm_table() argument
367 obj_per_chunk = MTHCA_TABLE_CHUNK_SIZE / obj_size; in mthca_alloc_icm_table()
377 table->obj_size = obj_size; in mthca_alloc_icm_table()
385 for (i = 0; i * MTHCA_TABLE_CHUNK_SIZE < reserved * obj_size; ++i) { in mthca_alloc_icm_table()
387 if ((i + 1) * MTHCA_TABLE_CHUNK_SIZE > nobj * obj_size) in mthca_alloc_icm_table()
[all …]
Dmthca_memfree.h67 int obj_size; member
87 u64 virt, int obj_size,
/linux-6.12.1/include/rdma/
Duverbs_types.h80 size_t obj_size; member
163 .obj_size = (_obj_size) + \
175 .obj_size = (_size) + \
/linux-6.12.1/lib/
Dobjpool.c37 obj = obj + pool->obj_size; in objpool_init_percpu_slot()
69 pool->obj_size * nodes; in objpool_init_percpu_slots()
140 pool->obj_size = object_size; in objpool_init()
Dobjagg.c372 objagg_obj = kzalloc(sizeof(*objagg_obj) + objagg->ops->obj_size, in objagg_obj_create()
377 memcpy(objagg_obj->obj, obj, objagg->ops->obj_size); in objagg_obj_create()
539 objagg->ht_params.key_len = ops->obj_size; in objagg_create()
653 struct objagg_obj *objagg_obj, size_t obj_size, in objagg_hints_node_create() argument
660 hnode = kzalloc(sizeof(*hnode) + obj_size, GFP_KERNEL); in objagg_hints_node_create()
663 memcpy(hnode->obj, &objagg_obj->obj, obj_size); in objagg_hints_node_create()
864 objagg->ops->obj_size, in objagg_opt_simple_greedy_fillup_hints()
880 objagg->ops->obj_size, in objagg_opt_simple_greedy_fillup_hints()
942 objagg_hints->ht_params.key_len = objagg->ops->obj_size; in objagg_hints_get()
/linux-6.12.1/net/smc/
Dsmc_inet.c28 .obj_size = sizeof(struct smc_sock),
75 .obj_size = sizeof(struct smc6_sock),
/linux-6.12.1/drivers/infiniband/hw/hns/
Dhns_roce_hem.c226 table_idx = *obj / (chunk_size / table->obj_size); in hns_roce_calc_hem_mhop()
540 i = obj / (table->table_chunk_size / table->obj_size); in hns_roce_table_get()
667 i = obj / (table->table_chunk_size / table->obj_size); in hns_roce_table_put()
701 obj_per_chunk = table->table_chunk_size / table->obj_size; in hns_roce_table_find()
704 dma_offset = offset = idx_offset * table->obj_size; in hns_roce_table_find()
738 unsigned long obj_size, unsigned long nobj) in hns_roce_init_hem_table() argument
745 obj_per_chunk = table->table_chunk_size / obj_size; in hns_roce_init_hem_table()
767 obj_per_chunk = buf_chunk_size / obj_size; in hns_roce_init_hem_table()
814 table->obj_size = obj_size; in hns_roce_init_hem_table()
853 obj = i * buf_chunk_size / table->obj_size; in hns_roce_cleanup_mhop_hem_table()
[all …]
Dhns_roce_hem.h98 unsigned long obj_size, unsigned long nobj);
/linux-6.12.1/net/dccp/
Dccid.c79 static __printf(3, 4) struct kmem_cache *ccid_kmem_cache_create(int obj_size, char *slab_name_fmt, … in ccid_kmem_cache_create() argument
88 slab = kmem_cache_create(slab_name_fmt, sizeof(struct ccid) + obj_size, 0, in ccid_kmem_cache_create()
/linux-6.12.1/drivers/crypto/intel/qat/qat_common/
Dqat_uclo.c1690 unsigned int *obj_size) in qat_uclo_seek_obj_inside_mof() argument
1699 *obj_size = obj_hdr[i].obj_size; in qat_uclo_seek_obj_inside_mof()
1725 mobj_hdr->obj_size = (unsigned int)obj_chunkhdr->size; in qat_uclo_map_obj_from_mof()
1842 char **obj_ptr, unsigned int *obj_size) in qat_uclo_map_mof_obj() argument
1854 if (obj_size) in qat_uclo_map_mof_obj()
1855 *obj_size = mof_size; in qat_uclo_map_mof_obj()
1888 obj_ptr, obj_size); in qat_uclo_map_mof_obj()
1895 u32 obj_size; in qat_uclo_map_obj() local
1906 &obj_addr, &obj_size); in qat_uclo_map_obj()
1911 obj_size = mem_size; in qat_uclo_map_obj()
[all …]
/linux-6.12.1/net/netfilter/
Dnf_conntrack_proto_generic.c75 .obj_size = sizeof(unsigned int),
Dnf_conntrack_proto_udp.c297 .obj_size = sizeof(unsigned int) * CTA_TIMEOUT_UDP_MAX,
319 .obj_size = sizeof(unsigned int) * CTA_TIMEOUT_UDP_MAX,
/linux-6.12.1/drivers/gpu/drm/panthor/
Dpanthor_drv.c85 u32 obj_size) in panthor_get_uobj_array() argument
99 out_alloc = kvmalloc_array(in->count, obj_size, GFP_KERNEL); in panthor_get_uobj_array()
103 if (obj_size == in->stride) { in panthor_get_uobj_array()
106 (unsigned long)obj_size * in->count)) in panthor_get_uobj_array()
114 ret = copy_struct_from_user(out_ptr, obj_size, in_ptr, in->stride); in panthor_get_uobj_array()
118 out_ptr += obj_size; in panthor_get_uobj_array()
/linux-6.12.1/net/ipv4/
Dudplite.c69 .obj_size = sizeof(struct udp_sock),

12345