Lines Matching +full:no +full:- +full:tick +full:- +full:in +full:- +full:suspend
1 // SPDX-License-Identifier: GPL-2.0-only
7 * tasks which are handled in sched/fair.c )
14 * sched_idle_set_state - Record idle state for the current CPU.
29 cpu_idle_force_poll--; in cpu_idle_poll_ctrl()
104 * default_idle_call - Default CPU idle routine.
132 return -EBUSY; in call_cpuidle_s2idle()
142 * update no idle residency and return. in call_cpuidle()
145 dev->last_residency_ns = 0; in call_cpuidle()
147 return -EBUSY; in call_cpuidle()
153 * care of re-enabling the local interrupts in call_cpuidle()
159 * cpuidle_idle_call - the main idle function
161 * NOTE: no locks or semaphores should be used here
175 * case, exit the function after re-enabling the local IRQ. in cpuidle_idle_call()
190 * Suspend-to-idle ("s2idle") is a system state in which all user space in cpuidle_idle_call()
192 * activity happens here and in interrupts (if any). In that case bypass in cpuidle_idle_call()
194 * available. Possibly also suspend the local tick and the entire in cpuidle_idle_call()
199 if (idle_should_enter_s2idle() || dev->forced_idle_latency_limit_ns) { in cpuidle_idle_call()
210 max_latency_ns = dev->forced_idle_latency_limit_ns; in cpuidle_idle_call()
241 * It is up to the idle functions to re-enable local interrupts in cpuidle_idle_call()
264 * Our polling bit is clear if we're not scheduled (i.e. if rq->curr != in do_idle()
265 * rq->idle). This means that, if rq->idle has the polling bit set, in do_idle()
277 * Interrupts shouldn't be re-enabled from that point on until in do_idle()
285 * - SLEEP-UNTIL-PENDING-INTERRUPT based instructions such as in do_idle()
289 * - sti;mwait() couple is fine because the interrupts are in do_idle()
290 * re-enabled only upon the execution of mwait, leaving no gap in do_idle()
291 * in-between. in do_idle()
293 * - ROLLBACK based idle handlers with the sleeping instruction in do_idle()
294 * called with interrupts enabled are NOT fine. In this scheme in do_idle()
297 * need_resched() check before re-executing the sleeping in do_idle()
300 * appropriate CPU sleeping instruction, then a FAST-FORWARD in do_idle()
304 * again to reprogram the tick. in do_idle()
317 * In poll mode we re-enable interrupts and spin. Also if we in do_idle()
318 * detected in the wakeup from idle path that the tick in do_idle()
350 * RCU relies on this call to be done outside of an RCU read-side in do_idle()
375 WRITE_ONCE(it->done, 1); in idle_inject_timer_fn()
386 * Only FIFO tasks can disable the tick since they don't need the forced in play_idle_precise()
389 WARN_ON_ONCE(current->policy != SCHED_FIFO); in play_idle_precise()
390 WARN_ON_ONCE(current->nr_cpus_allowed != 1); in play_idle_precise()
391 WARN_ON_ONCE(!(current->flags & PF_KTHREAD)); in play_idle_precise()
392 WARN_ON_ONCE(!(current->flags & PF_NO_SETAFFINITY)); in play_idle_precise()
394 WARN_ON_ONCE(current->mm); in play_idle_precise()
398 current->flags |= PF_IDLE; in play_idle_precise()
411 current->flags &= ~PF_IDLE; in play_idle_precise()
420 current->flags |= PF_IDLE; in cpu_startup_entry()
428 * idle-task scheduling class.
463 schedstat_inc(rq->sched_goidle); in set_next_task_idle()
464 next->se.exec_start = rq_clock_task(rq); in set_next_task_idle()
469 return rq->idle; in pick_task_idle()
473 * It is not legal to sleep in the idle task - print a warning
487 * scheduler tick hitting a task of our scheduling class.
489 * NOTE: This function can be called remotely by the tick offload that
490 * goes along full dynticks. Therefore no local assumption can be made
491 * and everything must be accessed through the @rq and @curr passed in
514 * Simple, special scheduling class for the per-CPU idle tasks:
518 /* no enqueue/yield_task for idle tasks */