1 /* SPDX-License-Identifier: MIT */
2 /*
3  * Copyright © 2023 Intel Corporation
4  */
5 
6 #ifndef _XE_GPU_SCHEDULER_H_
7 #define _XE_GPU_SCHEDULER_H_
8 
9 #include "xe_gpu_scheduler_types.h"
10 #include "xe_sched_job_types.h"
11 
12 int xe_sched_init(struct xe_gpu_scheduler *sched,
13 		  const struct drm_sched_backend_ops *ops,
14 		  const struct xe_sched_backend_ops *xe_ops,
15 		  struct workqueue_struct *submit_wq,
16 		  uint32_t hw_submission, unsigned hang_limit,
17 		  long timeout, struct workqueue_struct *timeout_wq,
18 		  atomic_t *score, const char *name,
19 		  struct device *dev);
20 void xe_sched_fini(struct xe_gpu_scheduler *sched);
21 
22 void xe_sched_submission_start(struct xe_gpu_scheduler *sched);
23 void xe_sched_submission_stop(struct xe_gpu_scheduler *sched);
24 
25 void xe_sched_submission_resume_tdr(struct xe_gpu_scheduler *sched);
26 
27 void xe_sched_add_msg(struct xe_gpu_scheduler *sched,
28 		      struct xe_sched_msg *msg);
29 void xe_sched_add_msg_locked(struct xe_gpu_scheduler *sched,
30 			     struct xe_sched_msg *msg);
31 
xe_sched_msg_lock(struct xe_gpu_scheduler * sched)32 static inline void xe_sched_msg_lock(struct xe_gpu_scheduler *sched)
33 {
34 	spin_lock(&sched->base.job_list_lock);
35 }
36 
xe_sched_msg_unlock(struct xe_gpu_scheduler * sched)37 static inline void xe_sched_msg_unlock(struct xe_gpu_scheduler *sched)
38 {
39 	spin_unlock(&sched->base.job_list_lock);
40 }
41 
xe_sched_stop(struct xe_gpu_scheduler * sched)42 static inline void xe_sched_stop(struct xe_gpu_scheduler *sched)
43 {
44 	drm_sched_stop(&sched->base, NULL);
45 }
46 
xe_sched_tdr_queue_imm(struct xe_gpu_scheduler * sched)47 static inline void xe_sched_tdr_queue_imm(struct xe_gpu_scheduler *sched)
48 {
49 	drm_sched_tdr_queue_imm(&sched->base);
50 }
51 
xe_sched_resubmit_jobs(struct xe_gpu_scheduler * sched)52 static inline void xe_sched_resubmit_jobs(struct xe_gpu_scheduler *sched)
53 {
54 	drm_sched_resubmit_jobs(&sched->base);
55 }
56 
57 static inline bool
xe_sched_invalidate_job(struct xe_sched_job * job,int threshold)58 xe_sched_invalidate_job(struct xe_sched_job *job, int threshold)
59 {
60 	return drm_sched_invalidate_job(&job->drm, threshold);
61 }
62 
xe_sched_add_pending_job(struct xe_gpu_scheduler * sched,struct xe_sched_job * job)63 static inline void xe_sched_add_pending_job(struct xe_gpu_scheduler *sched,
64 					    struct xe_sched_job *job)
65 {
66 	spin_lock(&sched->base.job_list_lock);
67 	list_add(&job->drm.list, &sched->base.pending_list);
68 	spin_unlock(&sched->base.job_list_lock);
69 }
70 
71 static inline
xe_sched_first_pending_job(struct xe_gpu_scheduler * sched)72 struct xe_sched_job *xe_sched_first_pending_job(struct xe_gpu_scheduler *sched)
73 {
74 	return list_first_entry_or_null(&sched->base.pending_list,
75 					struct xe_sched_job, drm.list);
76 }
77 
78 static inline int
xe_sched_entity_init(struct xe_sched_entity * entity,struct xe_gpu_scheduler * sched)79 xe_sched_entity_init(struct xe_sched_entity *entity,
80 		     struct xe_gpu_scheduler *sched)
81 {
82 	return drm_sched_entity_init(entity, 0,
83 				     (struct drm_gpu_scheduler **)&sched,
84 				     1, NULL);
85 }
86 
87 #define xe_sched_entity_fini drm_sched_entity_fini
88 
89 #endif
90