Lines Matching refs:kernel_queues

50 	q = &hdev->kernel_queues[0];  in hl_hw_queue_update_ci()
242 struct hl_hw_queue *q = &hdev->kernel_queues[hw_queue_id]; in hl_hw_queue_send_cb_no_cmpl()
282 struct hl_hw_queue *q = &hdev->kernel_queues[job->hw_queue_id]; in ext_queue_schedule_job()
347 struct hl_hw_queue *q = &hdev->kernel_queues[job->hw_queue_id]; in int_queue_schedule_job()
383 struct hl_hw_queue *q = &hdev->kernel_queues[job->hw_queue_id]; in hw_queue_schedule_job()
424 prop = &hdev->kernel_queues[q_idx].sync_stream_prop; in init_signal_cs()
483 prop = &hdev->kernel_queues[q_idx].sync_stream_prop; in init_wait_cs()
663 q = &hdev->kernel_queues[0]; in hl_hw_queue_schedule_cs()
794 q = &hdev->kernel_queues[0]; in hl_hw_queue_schedule_cs()
819 struct hl_hw_queue *q = &hdev->kernel_queues[hw_queue_id]; in hl_hw_queue_inc_ci_kernel()
920 sync_stream_prop = &hdev->kernel_queues[q_idx].sync_stream_prop; in sync_stream_queue_init()
927 if (hdev->kernel_queues[q_idx].collective_mode == in sync_stream_queue_init()
940 } else if (hdev->kernel_queues[q_idx].collective_mode == in sync_stream_queue_init()
949 if (!hdev->kernel_queues[q_idx].supports_sync_stream) in sync_stream_queue_init()
975 &hdev->kernel_queues[q_idx].sync_stream_prop; in sync_stream_queue_reset()
1085 hdev->kernel_queues = kcalloc(asic->max_queues, in hl_hw_queues_create()
1086 sizeof(*hdev->kernel_queues), GFP_KERNEL); in hl_hw_queues_create()
1088 if (!hdev->kernel_queues) { in hl_hw_queues_create()
1094 for (i = 0, q_ready_cnt = 0, q = hdev->kernel_queues; in hl_hw_queues_create()
1118 for (i = 0, q = hdev->kernel_queues ; i < q_ready_cnt ; i++, q++) in hl_hw_queues_create()
1121 kfree(hdev->kernel_queues); in hl_hw_queues_create()
1132 for (i = 0, q = hdev->kernel_queues ; i < max_queues ; i++, q++) in hl_hw_queues_destroy()
1135 kfree(hdev->kernel_queues); in hl_hw_queues_destroy()
1144 for (i = 0, q = hdev->kernel_queues ; i < max_queues ; i++, q++) { in hl_hw_queue_reset()