Lines Matching refs:__ksym
37 s32 scx_bpf_create_dsq(u64 dsq_id, s32 node) __ksym;
38 … scx_bpf_select_cpu_dfl(struct task_struct *p, s32 prev_cpu, u64 wake_flags, bool *is_idle) __ksym;
39 void scx_bpf_dispatch(struct task_struct *p, u64 dsq_id, u64 slice, u64 enq_flags) __ksym;
40 …_bpf_dispatch_vtime(struct task_struct *p, u64 dsq_id, u64 slice, u64 vtime, u64 enq_flags) __ksym;
41 u32 scx_bpf_dispatch_nr_slots(void) __ksym;
42 void scx_bpf_dispatch_cancel(void) __ksym;
43 bool scx_bpf_consume(u64 dsq_id) __ksym;
44 void scx_bpf_dispatch_from_dsq_set_slice(struct bpf_iter_scx_dsq *it__iter, u64 slice) __ksym __wea…
45 void scx_bpf_dispatch_from_dsq_set_vtime(struct bpf_iter_scx_dsq *it__iter, u64 vtime) __ksym __wea…
46 …struct bpf_iter_scx_dsq *it__iter, struct task_struct *p, u64 dsq_id, u64 enq_flags) __ksym __weak;
47 …struct bpf_iter_scx_dsq *it__iter, struct task_struct *p, u64 dsq_id, u64 enq_flags) __ksym __weak;
48 u32 scx_bpf_reenqueue_local(void) __ksym;
49 void scx_bpf_kick_cpu(s32 cpu, u64 flags) __ksym;
50 s32 scx_bpf_dsq_nr_queued(u64 dsq_id) __ksym;
51 void scx_bpf_destroy_dsq(u64 dsq_id) __ksym;
52 int bpf_iter_scx_dsq_new(struct bpf_iter_scx_dsq *it, u64 dsq_id, u64 flags) __ksym __weak;
53 struct task_struct *bpf_iter_scx_dsq_next(struct bpf_iter_scx_dsq *it) __ksym __weak;
54 void bpf_iter_scx_dsq_destroy(struct bpf_iter_scx_dsq *it) __ksym __weak;
55 void scx_bpf_exit_bstr(s64 exit_code, char *fmt, unsigned long long *data, u32 data__sz) __ksym __w…
56 void scx_bpf_error_bstr(char *fmt, unsigned long long *data, u32 data_len) __ksym;
57 void scx_bpf_dump_bstr(char *fmt, unsigned long long *data, u32 data_len) __ksym __weak;
58 u32 scx_bpf_cpuperf_cap(s32 cpu) __ksym __weak;
59 u32 scx_bpf_cpuperf_cur(s32 cpu) __ksym __weak;
60 void scx_bpf_cpuperf_set(s32 cpu, u32 perf) __ksym __weak;
61 u32 scx_bpf_nr_cpu_ids(void) __ksym __weak;
62 const struct cpumask *scx_bpf_get_possible_cpumask(void) __ksym __weak;
63 const struct cpumask *scx_bpf_get_online_cpumask(void) __ksym __weak;
64 void scx_bpf_put_cpumask(const struct cpumask *cpumask) __ksym __weak;
65 const struct cpumask *scx_bpf_get_idle_cpumask(void) __ksym;
66 const struct cpumask *scx_bpf_get_idle_smtmask(void) __ksym;
67 void scx_bpf_put_idle_cpumask(const struct cpumask *cpumask) __ksym;
68 bool scx_bpf_test_and_clear_cpu_idle(s32 cpu) __ksym;
69 s32 scx_bpf_pick_idle_cpu(const cpumask_t *cpus_allowed, u64 flags) __ksym;
70 s32 scx_bpf_pick_any_cpu(const cpumask_t *cpus_allowed, u64 flags) __ksym;
71 bool scx_bpf_task_running(const struct task_struct *p) __ksym;
72 s32 scx_bpf_task_cpu(const struct task_struct *p) __ksym;
73 struct rq *scx_bpf_cpu_rq(s32 cpu) __ksym;
74 struct cgroup *scx_bpf_task_cgroup(struct task_struct *p) __ksym __weak;
248 void *bpf_obj_new_impl(__u64 local_type_id, void *meta) __ksym;
249 void bpf_obj_drop_impl(void *kptr, void *meta) __ksym;
254 void bpf_list_push_front(struct bpf_list_head *head, struct bpf_list_node *node) __ksym;
255 void bpf_list_push_back(struct bpf_list_head *head, struct bpf_list_node *node) __ksym;
256 struct bpf_list_node *bpf_list_pop_front(struct bpf_list_head *head) __ksym;
257 struct bpf_list_node *bpf_list_pop_back(struct bpf_list_head *head) __ksym;
259 struct bpf_rb_node *node) __ksym;
262 void *meta, __u64 off) __ksym;
265 struct bpf_rb_node *bpf_rbtree_first(struct bpf_rb_root *root) __ksym;
267 void *bpf_refcount_acquire_impl(void *kptr, void *meta) __ksym;
271 struct task_struct *bpf_task_from_pid(s32 pid) __ksym;
272 struct task_struct *bpf_task_acquire(struct task_struct *p) __ksym;
273 void bpf_task_release(struct task_struct *p) __ksym;
276 struct cgroup *bpf_cgroup_ancestor(struct cgroup *cgrp, int level) __ksym;
277 void bpf_cgroup_release(struct cgroup *cgrp) __ksym;
278 struct cgroup *bpf_cgroup_from_id(u64 cgid) __ksym;
285 unsigned int flags) __weak __ksym;
287 bpf_iter_css_next(struct bpf_iter_css *it) __weak __ksym;
288 extern void bpf_iter_css_destroy(struct bpf_iter_css *it) __weak __ksym;
291 struct bpf_cpumask *bpf_cpumask_create(void) __ksym;
292 struct bpf_cpumask *bpf_cpumask_acquire(struct bpf_cpumask *cpumask) __ksym;
293 void bpf_cpumask_release(struct bpf_cpumask *cpumask) __ksym;
294 u32 bpf_cpumask_first(const struct cpumask *cpumask) __ksym;
295 u32 bpf_cpumask_first_zero(const struct cpumask *cpumask) __ksym;
296 void bpf_cpumask_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym;
297 void bpf_cpumask_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym;
298 bool bpf_cpumask_test_cpu(u32 cpu, const struct cpumask *cpumask) __ksym;
299 bool bpf_cpumask_test_and_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym;
300 bool bpf_cpumask_test_and_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym;
301 void bpf_cpumask_setall(struct bpf_cpumask *cpumask) __ksym;
302 void bpf_cpumask_clear(struct bpf_cpumask *cpumask) __ksym;
304 const struct cpumask *src2) __ksym;
306 const struct cpumask *src2) __ksym;
308 const struct cpumask *src2) __ksym;
309 bool bpf_cpumask_equal(const struct cpumask *src1, const struct cpumask *src2) __ksym;
310 bool bpf_cpumask_intersects(const struct cpumask *src1, const struct cpumask *src2) __ksym;
311 bool bpf_cpumask_subset(const struct cpumask *src1, const struct cpumask *src2) __ksym;
312 bool bpf_cpumask_empty(const struct cpumask *cpumask) __ksym;
313 bool bpf_cpumask_full(const struct cpumask *cpumask) __ksym;
314 void bpf_cpumask_copy(struct bpf_cpumask *dst, const struct cpumask *src) __ksym;
315 u32 bpf_cpumask_any_distribute(const struct cpumask *cpumask) __ksym;
317 const struct cpumask *src2) __ksym;
318 u32 bpf_cpumask_weight(const struct cpumask *cpumask) __ksym;
329 void bpf_rcu_read_lock(void) __ksym;
330 void bpf_rcu_read_unlock(void) __ksym;