/linux-6.12.1/drivers/crypto/cavium/cpt/ |
D | cptvf_main.c | 42 if (cptvf->nr_queues) { in init_worker_threads() 44 cptvf->nr_queues); in init_worker_threads() 47 for (i = 0; i < cptvf->nr_queues; i++) { in init_worker_threads() 69 if (cptvf->nr_queues) { in cleanup_worker_threads() 71 cptvf->nr_queues); in cleanup_worker_threads() 74 for (i = 0; i < cptvf->nr_queues; i++) in cleanup_worker_threads() 100 pqinfo->nr_queues = 0; in free_pending_queues() 104 u32 nr_queues) in alloc_pending_queues() argument 110 pqinfo->nr_queues = nr_queues; in alloc_pending_queues() 136 static int init_pending_queues(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues) in init_pending_queues() argument [all …]
|
D | cptvf.h | 85 u32 nr_queues; /* Number of queues supported */ member 91 for (i = 0, q = &qinfo->queue[i]; i < qinfo->nr_queues; i++, \ 110 u32 nr_queues; member
|
D | cptvf_reqmanager.c | 233 if (unlikely(qno >= cptvf->nr_queues)) { in send_cpt_command() 235 qno, cptvf->nr_queues); in send_cpt_command() 545 if (unlikely(qno > cptvf->nr_queues)) { in vq_post_process()
|
/linux-6.12.1/drivers/crypto/cavium/nitrox/ |
D | nitrox_sriov.c | 58 int nr_queues = 0; in vf_mode_to_nr_queues() local 62 nr_queues = MAX_PF_QUEUES; in vf_mode_to_nr_queues() 65 nr_queues = 8; in vf_mode_to_nr_queues() 68 nr_queues = 4; in vf_mode_to_nr_queues() 71 nr_queues = 2; in vf_mode_to_nr_queues() 74 nr_queues = 1; in vf_mode_to_nr_queues() 78 return nr_queues; in vf_mode_to_nr_queues()
|
D | nitrox_lib.c | 91 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_aqm_queues() 102 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_aqm_queues() 142 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_pktin_queues() 155 ndev->pkt_inq = kcalloc_node(ndev->nr_queues, in nitrox_alloc_pktin_queues() 161 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_pktin_queues()
|
D | nitrox_dev.h | 160 int nr_queues; member 250 u16 nr_queues; member
|
D | nitrox_mbx.c | 68 vfdev->nr_queues = vfdev->msg.data; in pf2vf_send_response() 76 vfdev->nr_queues = 0; in pf2vf_send_response()
|
D | nitrox_hal.c | 124 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_config_pkt_input_rings() 240 for (i = 0; i < ndev->nr_queues; i++) in nitrox_config_pkt_solicit_ports() 356 for (ring = 0; ring < ndev->nr_queues; ring++) { in nitrox_config_aqm_rings()
|
/linux-6.12.1/drivers/ufs/core/ |
D | ufs-mcq.c | 205 hba->nr_queues[HCTX_TYPE_DEFAULT] = rw_queues; in ufshcd_mcq_config_nr_queues() 206 rem -= hba->nr_queues[HCTX_TYPE_DEFAULT]; in ufshcd_mcq_config_nr_queues() 212 hba->nr_queues[HCTX_TYPE_POLL] = poll_queues; in ufshcd_mcq_config_nr_queues() 213 rem -= hba->nr_queues[HCTX_TYPE_POLL]; in ufshcd_mcq_config_nr_queues() 217 hba->nr_queues[HCTX_TYPE_READ] = read_queues; in ufshcd_mcq_config_nr_queues() 218 rem -= hba->nr_queues[HCTX_TYPE_READ]; in ufshcd_mcq_config_nr_queues() 221 if (!hba->nr_queues[HCTX_TYPE_DEFAULT]) in ufshcd_mcq_config_nr_queues() 222 hba->nr_queues[HCTX_TYPE_DEFAULT] = min3(rem, rw_queues, in ufshcd_mcq_config_nr_queues() 226 host->nr_hw_queues += hba->nr_queues[i]; in ufshcd_mcq_config_nr_queues() 402 if (i < hba->nr_hw_queues - hba->nr_queues[HCTX_TYPE_POLL]) in ufshcd_mcq_make_queues_operational()
|
/linux-6.12.1/block/ |
D | blk-mq-pci.c | 31 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_pci_map_queues() 43 WARN_ON_ONCE(qmap->nr_queues > 1); in blk_mq_pci_map_queues()
|
D | blk-mq-cpumap.c | 23 masks = group_cpus_evenly(qmap->nr_queues); in blk_mq_map_queues() 30 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_map_queues()
|
D | blk-mq-virtio.c | 32 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_virtio_map_queues()
|
/linux-6.12.1/drivers/media/pci/intel/ipu6/ |
D | ipu6-isys-video.h | 61 int nr_queues; /* Number of capture queues */ member 116 int nr_queues); 122 struct media_entity **source_entity, int *nr_queues);
|
D | ipu6-isys-queue.c | 350 if (stream->nr_streaming != stream->nr_queues) { in buf_queue() 516 int nr_queues, ret; in start_streaming() local 522 ret = ipu6_isys_setup_video(av, &source_entity, &nr_queues); in start_streaming() 544 nr_queues); in start_streaming() 551 stream->nr_queues); in start_streaming() 557 if (stream->nr_streaming != stream->nr_queues) in start_streaming() 605 if (stream->nr_streaming == stream->nr_queues && stream->streaming) in stop_streaming()
|
D | ipu6-isys-video.c | 715 int nr_queues) in ipu6_isys_video_prepare_stream() argument 723 stream->nr_queues = nr_queues; in ipu6_isys_video_prepare_stream() 1169 struct media_entity **source_entity, int *nr_queues) in ipu6_isys_setup_video() argument 1184 *nr_queues = 0; in ipu6_isys_setup_video() 1205 (*nr_queues)++; in ipu6_isys_setup_video()
|
/linux-6.12.1/tools/perf/util/ |
D | auxtrace.c | 200 static struct auxtrace_queue *auxtrace_alloc_queue_array(unsigned int nr_queues) in auxtrace_alloc_queue_array() argument 206 if (nr_queues > max_nr_queues) in auxtrace_alloc_queue_array() 209 queue_array = calloc(nr_queues, sizeof(struct auxtrace_queue)); in auxtrace_alloc_queue_array() 213 for (i = 0; i < nr_queues; i++) { in auxtrace_alloc_queue_array() 221 int auxtrace_queues__init_nr(struct auxtrace_queues *queues, int nr_queues) in auxtrace_queues__init_nr() argument 223 queues->nr_queues = nr_queues; in auxtrace_queues__init_nr() 224 queues->queue_array = auxtrace_alloc_queue_array(queues->nr_queues); in auxtrace_queues__init_nr() 238 unsigned int nr_queues = queues->nr_queues; in auxtrace_queues__grow() local 242 if (!nr_queues) in auxtrace_queues__grow() 243 nr_queues = AUXTRACE_INIT_NR_QUEUES; in auxtrace_queues__grow() [all …]
|
D | intel-bts.c | 211 for (i = 0; i < bts->queues.nr_queues; i++) { in intel_bts_setup_queues() 542 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_process_tid_exit() 713 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_free_events()
|
D | s390-cpumsf.c | 203 if (!sf->use_logfile || sf->queues.nr_queues <= sample->cpu) in s390_cpumcf_dumpctr() 825 for (i = 0; i < sf->queues.nr_queues; i++) { in s390_cpumsf_setup_queues() 1014 for (i = 0; i < queues->nr_queues; i++) { in s390_cpumsf_free_queues()
|
D | cs-etm.c | 310 for (unsigned int i = 0; i < etm->queues.nr_queues; ++i) { in cs_etm__map_trace_id_v0() 368 for (unsigned int i = 0; i < etm->queues.nr_queues; ++i) { in cs_etm__process_trace_id_v0_1() 989 for (i = 0; i < queues->nr_queues; i++) { in cs_etm__free_events() 2591 for (i = 0; i < queues->nr_queues; i++) { in cs_etm__process_timeless_queues() 2629 for (i = 0; i < etm->queues.nr_queues; i++) { in cs_etm__process_timestamped_queues() 2856 for (i = 0; i < etm->queues.nr_queues; ++i) in dump_queued_data() 3317 for (unsigned int i = 0; i < queues->nr_queues; i++) { in cs_etm__create_decoders() 3429 for (unsigned int j = 0; j < etm->queues.nr_queues; ++j) { in cs_etm__process_auxtrace_info_full()
|
D | auxtrace.h | 296 unsigned int nr_queues; member 523 int auxtrace_queues__init_nr(struct auxtrace_queues *queues, int nr_queues);
|
/linux-6.12.1/drivers/block/rnbd/ |
D | rnbd-clt.c | 1171 set->map[HCTX_TYPE_DEFAULT].nr_queues = num_online_cpus(); in rnbd_rdma_map_queues() 1173 set->map[HCTX_TYPE_READ].nr_queues = num_online_cpus(); in rnbd_rdma_map_queues() 1180 set->map[HCTX_TYPE_POLL].nr_queues = sess->nr_poll_queues; in rnbd_rdma_map_queues() 1182 set->map[HCTX_TYPE_READ].nr_queues; in rnbd_rdma_map_queues() 1186 set->map[HCTX_TYPE_DEFAULT].nr_queues, in rnbd_rdma_map_queues() 1187 set->map[HCTX_TYPE_READ].nr_queues, in rnbd_rdma_map_queues() 1188 set->map[HCTX_TYPE_POLL].nr_queues); in rnbd_rdma_map_queues() 1192 set->map[HCTX_TYPE_DEFAULT].nr_queues, in rnbd_rdma_map_queues() 1193 set->map[HCTX_TYPE_READ].nr_queues); in rnbd_rdma_map_queues()
|
/linux-6.12.1/drivers/nvme/host/ |
D | fabrics.c | 1135 set->map[HCTX_TYPE_DEFAULT].nr_queues = in nvmf_map_queues() 1138 set->map[HCTX_TYPE_READ].nr_queues = in nvmf_map_queues() 1144 set->map[HCTX_TYPE_DEFAULT].nr_queues = in nvmf_map_queues() 1147 set->map[HCTX_TYPE_READ].nr_queues = in nvmf_map_queues() 1156 set->map[HCTX_TYPE_POLL].nr_queues = io_queues[HCTX_TYPE_POLL]; in nvmf_map_queues()
|
D | pci.c | 448 map->nr_queues = dev->io_queues[i]; in nvme_pci_map_queues() 449 if (!map->nr_queues) { in nvme_pci_map_queues() 463 qoff += map->nr_queues; in nvme_pci_map_queues() 464 offset += map->nr_queues; in nvme_pci_map_queues() 2468 int nr_queues = dev->online_queues - 1, sent = 0; in __nvme_delete_io_queues() local 2473 while (nr_queues > 0) { in __nvme_delete_io_queues() 2474 if (nvme_delete_queue(&dev->queues[nr_queues], opcode)) in __nvme_delete_io_queues() 2476 nr_queues--; in __nvme_delete_io_queues() 2480 struct nvme_queue *nvmeq = &dev->queues[nr_queues + sent]; in __nvme_delete_io_queues() 2488 if (nr_queues) in __nvme_delete_io_queues()
|
/linux-6.12.1/drivers/s390/cio/ |
D | qdio_setup.c | 100 static int __qdio_allocate_qs(struct qdio_q **irq_ptr_qs, int nr_queues) in __qdio_allocate_qs() argument 105 for (i = 0; i < nr_queues; i++) { in __qdio_allocate_qs()
|
/linux-6.12.1/drivers/char/ |
D | virtio_console.c | 1808 u32 i, j, nr_ports, nr_queues; in init_vqs() local 1812 nr_queues = use_multiport(portdev) ? (nr_ports + 1) * 2 : 2; in init_vqs() 1814 vqs = kmalloc_array(nr_queues, sizeof(struct virtqueue *), GFP_KERNEL); in init_vqs() 1815 vqs_info = kcalloc(nr_queues, sizeof(*vqs_info), GFP_KERNEL); in init_vqs() 1851 err = virtio_find_vqs(portdev->vdev, nr_queues, vqs, vqs_info, NULL); in init_vqs()
|