/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | rqt.c | 145 rss_rqns = kvmalloc_array(indir->actual_table_size, sizeof(*rss_rqns), GFP_KERNEL); in mlx5e_rqt_init_indir() 150 rss_vhca_ids = kvmalloc_array(indir->actual_table_size, sizeof(*rss_vhca_ids), in mlx5e_rqt_init_indir() 243 rss_rqns = kvmalloc_array(indir->actual_table_size, sizeof(*rss_rqns), GFP_KERNEL); in mlx5e_rqt_redirect_indir() 248 rss_vhca_ids = kvmalloc_array(indir->actual_table_size, sizeof(*rss_vhca_ids), in mlx5e_rqt_redirect_indir()
|
/linux-6.12.1/drivers/gpu/drm/xe/ |
D | xe_hmm.c | 89 pages = kvmalloc_array(npages, sizeof(*pages), GFP_KERNEL); in xe_build_sg() 197 pfns = kvmalloc_array(npages, sizeof(*pfns), GFP_KERNEL); in xe_hmm_userptr_populate_range()
|
/linux-6.12.1/scripts/coccinelle/api/ |
D | kfree_mismatch.cocci | 58 kvmalloc_array\)(...) 124 * kvmalloc_array\)(...)@a 134 kvmalloc_array\)(...)
|
/linux-6.12.1/lib/ |
D | bucket_locks.c | 34 tlocks = kvmalloc_array(size, sizeof(spinlock_t), gfp); in __alloc_bucket_spinlocks()
|
/linux-6.12.1/io_uring/ |
D | alloc_cache.h | 37 cache->entries = kvmalloc_array(max_nr, sizeof(void *), GFP_KERNEL); in io_alloc_cache_init()
|
D | memmap.c | 68 pages = kvmalloc_array(nr_pages, sizeof(struct page *), gfp); in io_pages_map() 144 pages = kvmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL); in io_pin_pages()
|
/linux-6.12.1/drivers/gpu/drm/etnaviv/ |
D | etnaviv_gem_submit.c | 471 bos = kvmalloc_array(args->nr_bos, sizeof(*bos), GFP_KERNEL); in etnaviv_ioctl_gem_submit() 472 relocs = kvmalloc_array(args->nr_relocs, sizeof(*relocs), GFP_KERNEL); in etnaviv_ioctl_gem_submit() 473 pmrs = kvmalloc_array(args->nr_pmrs, sizeof(*pmrs), GFP_KERNEL); in etnaviv_ioctl_gem_submit() 474 stream = kvmalloc_array(1, args->stream_size, GFP_KERNEL); in etnaviv_ioctl_gem_submit()
|
D | etnaviv_gem_prime.c | 129 etnaviv_obj->pages = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in etnaviv_gem_prime_import_sg_table()
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
D | mem.c | 194 if (!(mem->mem = kvmalloc_array(size, sizeof(*mem->mem), GFP_KERNEL))) in nvkm_mem_new_host() 196 if (!(mem->dma = kvmalloc_array(size, sizeof(*mem->dma), GFP_KERNEL))) in nvkm_mem_new_host()
|
/linux-6.12.1/drivers/gpu/drm/v3d/ |
D | v3d_submit.c | 334 kvmalloc_array(count, in v3d_get_multisync_post_deps() 477 query_info->queries = kvmalloc_array(timestamp.count, in v3d_get_cpu_timestamp_query_params() 543 query_info->queries = kvmalloc_array(reset.count, in v3d_get_cpu_reset_timestamp_params() 607 query_info->queries = kvmalloc_array(copy.count, in v3d_get_cpu_copy_query_results_params() 681 kvmalloc_array(nperfmons, in v3d_copy_query_info() 741 kvmalloc_array(reset.count, in v3d_get_cpu_reset_performance_params() 790 kvmalloc_array(copy.count, in v3d_get_cpu_copy_performance_query_params()
|
/linux-6.12.1/kernel/dma/ |
D | remap.c | 48 pages = kvmalloc_array(count, sizeof(struct page *), GFP_KERNEL); in dma_common_contiguous_remap()
|
/linux-6.12.1/drivers/gpu/drm/nouveau/dispnv50/ |
D | lut.c | 41 in = kvmalloc_array(1024, sizeof(*in), GFP_KERNEL); in nv50_lut_load()
|
/linux-6.12.1/drivers/xen/ |
D | gntdev.c | 139 add->grants = kvmalloc_array(count, sizeof(add->grants[0]), in gntdev_alloc_map() 141 add->map_ops = kvmalloc_array(count, sizeof(add->map_ops[0]), in gntdev_alloc_map() 143 add->unmap_ops = kvmalloc_array(count, sizeof(add->unmap_ops[0]), in gntdev_alloc_map() 155 add->kmap_ops = kvmalloc_array(count, sizeof(add->kmap_ops[0]), in gntdev_alloc_map() 157 add->kunmap_ops = kvmalloc_array(count, sizeof(add->kunmap_ops[0]), in gntdev_alloc_map()
|
/linux-6.12.1/tools/virtio/ringtest/ |
D | ptr_ring.c | 61 #define kvmalloc_array kmalloc_array macro
|
/linux-6.12.1/drivers/net/ethernet/wangxun/txgbe/ |
D | txgbe_hw.c | 89 eeprom_ptrs = kvmalloc_array(TXGBE_EEPROM_LAST_WORD, sizeof(u16), in txgbe_calc_eeprom_checksum()
|
/linux-6.12.1/drivers/net/ethernet/wangxun/ngbe/ |
D | ngbe_ethtool.c | 84 temp_ring = kvmalloc_array(i, sizeof(struct wx_ring), GFP_KERNEL); in ngbe_set_ringparam()
|
/linux-6.12.1/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_hmm.c | 184 pfns = kvmalloc_array(npages, sizeof(*pfns), GFP_KERNEL); in amdgpu_hmm_range_get_pages()
|
/linux-6.12.1/drivers/gpu/drm/i915/gt/ |
D | shmem_utils.c | 65 pages = kvmalloc_array(n_pages, sizeof(*pages), GFP_KERNEL); in shmem_pin_map()
|
/linux-6.12.1/drivers/gpu/drm/ |
D | drm_exec.c | 85 exec->objects = kvmalloc_array(nr, sizeof(void *), GFP_KERNEL); in drm_exec_init()
|
D | drm_gem.c | 559 pages = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in drm_gem_get_pages() 707 objs = kvmalloc_array(count, sizeof(struct drm_gem_object *), in drm_gem_objects_lookup() 714 handles = kvmalloc_array(count, sizeof(u32), GFP_KERNEL); in drm_gem_objects_lookup()
|
/linux-6.12.1/drivers/gpu/drm/i915/gem/ |
D | i915_gem_pages.c | 313 pages = kvmalloc_array(n_pages, sizeof(*pages), GFP_KERNEL); in i915_gem_object_map_page() 343 pfns = kvmalloc_array(n_pfn, sizeof(*pfns), GFP_KERNEL); in i915_gem_object_map_pfn()
|
/linux-6.12.1/drivers/dma-buf/ |
D | st-dma-fence-chain.c | 119 fc->chains = kvmalloc_array(count, sizeof(*fc->chains), in fence_chains_init() 124 fc->fences = kvmalloc_array(count, sizeof(*fc->fences), in fence_chains_init()
|
/linux-6.12.1/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_blit.c | 589 src_pages = kvmalloc_array(src->ttm->num_pages, in vmw_bo_cpu_blit() 599 dst_pages = kvmalloc_array(dst->ttm->num_pages, in vmw_bo_cpu_blit()
|
/linux-6.12.1/drivers/gpu/drm/xen/ |
D | xen_drm_front_gem.c | 49 xen_obj->pages = kvmalloc_array(xen_obj->num_pages, in gem_alloc_pages_array()
|
/linux-6.12.1/drivers/gpu/drm/virtio/ |
D | virtgpu_object.c | 154 *ents = kvmalloc_array(*nents, in virtio_gpu_object_shmem_init()
|