1  /* SPDX-License-Identifier: GPL-2.0-only */
2  /*
3   * linux/include/linux/cpufreq.h
4   *
5   * Copyright (C) 2001 Russell King
6   *           (C) 2002 - 2003 Dominik Brodowski <linux@brodo.de>
7   */
8  #ifndef _LINUX_CPUFREQ_H
9  #define _LINUX_CPUFREQ_H
10  
11  #include <linux/clk.h>
12  #include <linux/cpu.h>
13  #include <linux/cpumask.h>
14  #include <linux/completion.h>
15  #include <linux/kobject.h>
16  #include <linux/notifier.h>
17  #include <linux/of.h>
18  #include <linux/pm_opp.h>
19  #include <linux/pm_qos.h>
20  #include <linux/spinlock.h>
21  #include <linux/sysfs.h>
22  #include <linux/minmax.h>
23  
24  /*********************************************************************
25   *                        CPUFREQ INTERFACE                          *
26   *********************************************************************/
27  /*
28   * Frequency values here are CPU kHz
29   *
30   * Maximum transition latency is in nanoseconds - if it's unknown,
31   * CPUFREQ_ETERNAL shall be used.
32   */
33  
34  #define CPUFREQ_ETERNAL			(-1)
35  #define CPUFREQ_NAME_LEN		16
36  /* Print length for names. Extra 1 space for accommodating '\n' in prints */
37  #define CPUFREQ_NAME_PLEN		(CPUFREQ_NAME_LEN + 1)
38  
39  struct cpufreq_governor;
40  
41  enum cpufreq_table_sorting {
42  	CPUFREQ_TABLE_UNSORTED,
43  	CPUFREQ_TABLE_SORTED_ASCENDING,
44  	CPUFREQ_TABLE_SORTED_DESCENDING
45  };
46  
47  struct cpufreq_cpuinfo {
48  	unsigned int		max_freq;
49  	unsigned int		min_freq;
50  
51  	/* in 10^(-9) s = nanoseconds */
52  	unsigned int		transition_latency;
53  };
54  
55  struct cpufreq_policy {
56  	/* CPUs sharing clock, require sw coordination */
57  	cpumask_var_t		cpus;	/* Online CPUs only */
58  	cpumask_var_t		related_cpus; /* Online + Offline CPUs */
59  	cpumask_var_t		real_cpus; /* Related and present */
60  
61  	unsigned int		shared_type; /* ACPI: ANY or ALL affected CPUs
62  						should set cpufreq */
63  	unsigned int		cpu;    /* cpu managing this policy, must be online */
64  
65  	struct clk		*clk;
66  	struct cpufreq_cpuinfo	cpuinfo;/* see above */
67  
68  	unsigned int		min;    /* in kHz */
69  	unsigned int		max;    /* in kHz */
70  	unsigned int		cur;    /* in kHz, only needed if cpufreq
71  					 * governors are used */
72  	unsigned int		suspend_freq; /* freq to set during suspend */
73  
74  	unsigned int		policy; /* see above */
75  	unsigned int		last_policy; /* policy before unplug */
76  	struct cpufreq_governor	*governor; /* see below */
77  	void			*governor_data;
78  	char			last_governor[CPUFREQ_NAME_LEN]; /* last governor used */
79  
80  	struct work_struct	update; /* if update_policy() needs to be
81  					 * called, but you're in IRQ context */
82  
83  	struct freq_constraints	constraints;
84  	struct freq_qos_request	*min_freq_req;
85  	struct freq_qos_request	*max_freq_req;
86  
87  	struct cpufreq_frequency_table	*freq_table;
88  	enum cpufreq_table_sorting freq_table_sorted;
89  
90  	struct list_head        policy_list;
91  	struct kobject		kobj;
92  	struct completion	kobj_unregister;
93  
94  	/*
95  	 * The rules for this semaphore:
96  	 * - Any routine that wants to read from the policy structure will
97  	 *   do a down_read on this semaphore.
98  	 * - Any routine that will write to the policy structure and/or may take away
99  	 *   the policy altogether (eg. CPU hotplug), will hold this lock in write
100  	 *   mode before doing so.
101  	 */
102  	struct rw_semaphore	rwsem;
103  
104  	/*
105  	 * Fast switch flags:
106  	 * - fast_switch_possible should be set by the driver if it can
107  	 *   guarantee that frequency can be changed on any CPU sharing the
108  	 *   policy and that the change will affect all of the policy CPUs then.
109  	 * - fast_switch_enabled is to be set by governors that support fast
110  	 *   frequency switching with the help of cpufreq_enable_fast_switch().
111  	 */
112  	bool			fast_switch_possible;
113  	bool			fast_switch_enabled;
114  
115  	/*
116  	 * Set if the CPUFREQ_GOV_STRICT_TARGET flag is set for the current
117  	 * governor.
118  	 */
119  	bool			strict_target;
120  
121  	/*
122  	 * Set if inefficient frequencies were found in the frequency table.
123  	 * This indicates if the relation flag CPUFREQ_RELATION_E can be
124  	 * honored.
125  	 */
126  	bool			efficiencies_available;
127  
128  	/*
129  	 * Preferred average time interval between consecutive invocations of
130  	 * the driver to set the frequency for this policy.  To be set by the
131  	 * scaling driver (0, which is the default, means no preference).
132  	 */
133  	unsigned int		transition_delay_us;
134  
135  	/*
136  	 * Remote DVFS flag (Not added to the driver structure as we don't want
137  	 * to access another structure from scheduler hotpath).
138  	 *
139  	 * Should be set if CPUs can do DVFS on behalf of other CPUs from
140  	 * different cpufreq policies.
141  	 */
142  	bool			dvfs_possible_from_any_cpu;
143  
144  	/* Per policy boost enabled flag. */
145  	bool			boost_enabled;
146  
147  	 /* Cached frequency lookup from cpufreq_driver_resolve_freq. */
148  	unsigned int cached_target_freq;
149  	unsigned int cached_resolved_idx;
150  
151  	/* Synchronization for frequency transitions */
152  	bool			transition_ongoing; /* Tracks transition status */
153  	spinlock_t		transition_lock;
154  	wait_queue_head_t	transition_wait;
155  	struct task_struct	*transition_task; /* Task which is doing the transition */
156  
157  	/* cpufreq-stats */
158  	struct cpufreq_stats	*stats;
159  
160  	/* For cpufreq driver's internal use */
161  	void			*driver_data;
162  
163  	/* Pointer to the cooling device if used for thermal mitigation */
164  	struct thermal_cooling_device *cdev;
165  
166  	struct notifier_block nb_min;
167  	struct notifier_block nb_max;
168  };
169  
170  /*
171   * Used for passing new cpufreq policy data to the cpufreq driver's ->verify()
172   * callback for sanitization.  That callback is only expected to modify the min
173   * and max values, if necessary, and specifically it must not update the
174   * frequency table.
175   */
176  struct cpufreq_policy_data {
177  	struct cpufreq_cpuinfo		cpuinfo;
178  	struct cpufreq_frequency_table	*freq_table;
179  	unsigned int			cpu;
180  	unsigned int			min;    /* in kHz */
181  	unsigned int			max;    /* in kHz */
182  };
183  
184  struct cpufreq_freqs {
185  	struct cpufreq_policy *policy;
186  	unsigned int old;
187  	unsigned int new;
188  	u8 flags;		/* flags of cpufreq_driver, see below. */
189  };
190  
191  /* Only for ACPI */
192  #define CPUFREQ_SHARED_TYPE_NONE (0) /* None */
193  #define CPUFREQ_SHARED_TYPE_HW	 (1) /* HW does needed coordination */
194  #define CPUFREQ_SHARED_TYPE_ALL	 (2) /* All dependent CPUs should set freq */
195  #define CPUFREQ_SHARED_TYPE_ANY	 (3) /* Freq can be set from any dependent CPU*/
196  
197  #ifdef CONFIG_CPU_FREQ
198  struct cpufreq_policy *cpufreq_cpu_get_raw(unsigned int cpu);
199  struct cpufreq_policy *cpufreq_cpu_get(unsigned int cpu);
200  void cpufreq_cpu_put(struct cpufreq_policy *policy);
201  #else
cpufreq_cpu_get_raw(unsigned int cpu)202  static inline struct cpufreq_policy *cpufreq_cpu_get_raw(unsigned int cpu)
203  {
204  	return NULL;
205  }
cpufreq_cpu_get(unsigned int cpu)206  static inline struct cpufreq_policy *cpufreq_cpu_get(unsigned int cpu)
207  {
208  	return NULL;
209  }
cpufreq_cpu_put(struct cpufreq_policy * policy)210  static inline void cpufreq_cpu_put(struct cpufreq_policy *policy) { }
211  #endif
212  
policy_is_inactive(struct cpufreq_policy * policy)213  static inline bool policy_is_inactive(struct cpufreq_policy *policy)
214  {
215  	return cpumask_empty(policy->cpus);
216  }
217  
policy_is_shared(struct cpufreq_policy * policy)218  static inline bool policy_is_shared(struct cpufreq_policy *policy)
219  {
220  	return cpumask_weight(policy->cpus) > 1;
221  }
222  
223  #ifdef CONFIG_CPU_FREQ
224  unsigned int cpufreq_get(unsigned int cpu);
225  unsigned int cpufreq_quick_get(unsigned int cpu);
226  unsigned int cpufreq_quick_get_max(unsigned int cpu);
227  unsigned int cpufreq_get_hw_max_freq(unsigned int cpu);
228  void disable_cpufreq(void);
229  
230  u64 get_cpu_idle_time(unsigned int cpu, u64 *wall, int io_busy);
231  
232  struct cpufreq_policy *cpufreq_cpu_acquire(unsigned int cpu);
233  void cpufreq_cpu_release(struct cpufreq_policy *policy);
234  int cpufreq_get_policy(struct cpufreq_policy *policy, unsigned int cpu);
235  void refresh_frequency_limits(struct cpufreq_policy *policy);
236  void cpufreq_update_policy(unsigned int cpu);
237  void cpufreq_update_limits(unsigned int cpu);
238  bool have_governor_per_policy(void);
239  bool cpufreq_supports_freq_invariance(void);
240  struct kobject *get_governor_parent_kobj(struct cpufreq_policy *policy);
241  void cpufreq_enable_fast_switch(struct cpufreq_policy *policy);
242  void cpufreq_disable_fast_switch(struct cpufreq_policy *policy);
243  bool has_target_index(void);
244  
245  DECLARE_PER_CPU(unsigned long, cpufreq_pressure);
cpufreq_get_pressure(int cpu)246  static inline unsigned long cpufreq_get_pressure(int cpu)
247  {
248  	return READ_ONCE(per_cpu(cpufreq_pressure, cpu));
249  }
250  #else
cpufreq_get(unsigned int cpu)251  static inline unsigned int cpufreq_get(unsigned int cpu)
252  {
253  	return 0;
254  }
cpufreq_quick_get(unsigned int cpu)255  static inline unsigned int cpufreq_quick_get(unsigned int cpu)
256  {
257  	return 0;
258  }
cpufreq_quick_get_max(unsigned int cpu)259  static inline unsigned int cpufreq_quick_get_max(unsigned int cpu)
260  {
261  	return 0;
262  }
cpufreq_get_hw_max_freq(unsigned int cpu)263  static inline unsigned int cpufreq_get_hw_max_freq(unsigned int cpu)
264  {
265  	return 0;
266  }
cpufreq_supports_freq_invariance(void)267  static inline bool cpufreq_supports_freq_invariance(void)
268  {
269  	return false;
270  }
disable_cpufreq(void)271  static inline void disable_cpufreq(void) { }
cpufreq_update_limits(unsigned int cpu)272  static inline void cpufreq_update_limits(unsigned int cpu) { }
cpufreq_get_pressure(int cpu)273  static inline unsigned long cpufreq_get_pressure(int cpu)
274  {
275  	return 0;
276  }
277  #endif
278  
279  #ifdef CONFIG_CPU_FREQ_STAT
280  void cpufreq_stats_create_table(struct cpufreq_policy *policy);
281  void cpufreq_stats_free_table(struct cpufreq_policy *policy);
282  void cpufreq_stats_record_transition(struct cpufreq_policy *policy,
283  				     unsigned int new_freq);
284  #else
cpufreq_stats_create_table(struct cpufreq_policy * policy)285  static inline void cpufreq_stats_create_table(struct cpufreq_policy *policy) { }
cpufreq_stats_free_table(struct cpufreq_policy * policy)286  static inline void cpufreq_stats_free_table(struct cpufreq_policy *policy) { }
cpufreq_stats_record_transition(struct cpufreq_policy * policy,unsigned int new_freq)287  static inline void cpufreq_stats_record_transition(struct cpufreq_policy *policy,
288  						   unsigned int new_freq) { }
289  #endif /* CONFIG_CPU_FREQ_STAT */
290  
291  /*********************************************************************
292   *                      CPUFREQ DRIVER INTERFACE                     *
293   *********************************************************************/
294  
295  #define CPUFREQ_RELATION_L 0  /* lowest frequency at or above target */
296  #define CPUFREQ_RELATION_H 1  /* highest frequency below or at target */
297  #define CPUFREQ_RELATION_C 2  /* closest frequency to target */
298  /* relation flags */
299  #define CPUFREQ_RELATION_E BIT(2) /* Get if possible an efficient frequency */
300  
301  #define CPUFREQ_RELATION_LE (CPUFREQ_RELATION_L | CPUFREQ_RELATION_E)
302  #define CPUFREQ_RELATION_HE (CPUFREQ_RELATION_H | CPUFREQ_RELATION_E)
303  #define CPUFREQ_RELATION_CE (CPUFREQ_RELATION_C | CPUFREQ_RELATION_E)
304  
305  struct freq_attr {
306  	struct attribute attr;
307  	ssize_t (*show)(struct cpufreq_policy *, char *);
308  	ssize_t (*store)(struct cpufreq_policy *, const char *, size_t count);
309  };
310  
311  #define cpufreq_freq_attr_ro(_name)		\
312  static struct freq_attr _name =			\
313  __ATTR(_name, 0444, show_##_name, NULL)
314  
315  #define cpufreq_freq_attr_ro_perm(_name, _perm)	\
316  static struct freq_attr _name =			\
317  __ATTR(_name, _perm, show_##_name, NULL)
318  
319  #define cpufreq_freq_attr_rw(_name)		\
320  static struct freq_attr _name =			\
321  __ATTR(_name, 0644, show_##_name, store_##_name)
322  
323  #define cpufreq_freq_attr_wo(_name)		\
324  static struct freq_attr _name =			\
325  __ATTR(_name, 0200, NULL, store_##_name)
326  
327  #define define_one_global_ro(_name)		\
328  static struct kobj_attribute _name =		\
329  __ATTR(_name, 0444, show_##_name, NULL)
330  
331  #define define_one_global_rw(_name)		\
332  static struct kobj_attribute _name =		\
333  __ATTR(_name, 0644, show_##_name, store_##_name)
334  
335  
336  struct cpufreq_driver {
337  	char		name[CPUFREQ_NAME_LEN];
338  	u16		flags;
339  	void		*driver_data;
340  
341  	/* needed by all drivers */
342  	int		(*init)(struct cpufreq_policy *policy);
343  	int		(*verify)(struct cpufreq_policy_data *policy);
344  
345  	/* define one out of two */
346  	int		(*setpolicy)(struct cpufreq_policy *policy);
347  
348  	int		(*target)(struct cpufreq_policy *policy,
349  				  unsigned int target_freq,
350  				  unsigned int relation);	/* Deprecated */
351  	int		(*target_index)(struct cpufreq_policy *policy,
352  					unsigned int index);
353  	unsigned int	(*fast_switch)(struct cpufreq_policy *policy,
354  				       unsigned int target_freq);
355  	/*
356  	 * ->fast_switch() replacement for drivers that use an internal
357  	 * representation of performance levels and can pass hints other than
358  	 * the target performance level to the hardware. This can only be set
359  	 * if ->fast_switch is set too, because in those cases (under specific
360  	 * conditions) scale invariance can be disabled, which causes the
361  	 * schedutil governor to fall back to the latter.
362  	 */
363  	void		(*adjust_perf)(unsigned int cpu,
364  				       unsigned long min_perf,
365  				       unsigned long target_perf,
366  				       unsigned long capacity);
367  
368  	/*
369  	 * Only for drivers with target_index() and CPUFREQ_ASYNC_NOTIFICATION
370  	 * unset.
371  	 *
372  	 * get_intermediate should return a stable intermediate frequency
373  	 * platform wants to switch to and target_intermediate() should set CPU
374  	 * to that frequency, before jumping to the frequency corresponding
375  	 * to 'index'. Core will take care of sending notifications and driver
376  	 * doesn't have to handle them in target_intermediate() or
377  	 * target_index().
378  	 *
379  	 * Drivers can return '0' from get_intermediate() in case they don't
380  	 * wish to switch to intermediate frequency for some target frequency.
381  	 * In that case core will directly call ->target_index().
382  	 */
383  	unsigned int	(*get_intermediate)(struct cpufreq_policy *policy,
384  					    unsigned int index);
385  	int		(*target_intermediate)(struct cpufreq_policy *policy,
386  					       unsigned int index);
387  
388  	/* should be defined, if possible, return 0 on error */
389  	unsigned int	(*get)(unsigned int cpu);
390  
391  	/* Called to update policy limits on firmware notifications. */
392  	void		(*update_limits)(unsigned int cpu);
393  
394  	/* optional */
395  	int		(*bios_limit)(int cpu, unsigned int *limit);
396  
397  	int		(*online)(struct cpufreq_policy *policy);
398  	int		(*offline)(struct cpufreq_policy *policy);
399  	void		(*exit)(struct cpufreq_policy *policy);
400  	int		(*suspend)(struct cpufreq_policy *policy);
401  	int		(*resume)(struct cpufreq_policy *policy);
402  
403  	/* Will be called after the driver is fully initialized */
404  	void		(*ready)(struct cpufreq_policy *policy);
405  
406  	struct freq_attr **attr;
407  
408  	/* platform specific boost support code */
409  	bool		boost_enabled;
410  	int		(*set_boost)(struct cpufreq_policy *policy, int state);
411  
412  	/*
413  	 * Set by drivers that want to register with the energy model after the
414  	 * policy is properly initialized, but before the governor is started.
415  	 */
416  	void		(*register_em)(struct cpufreq_policy *policy);
417  };
418  
419  /* flags */
420  
421  /*
422   * Set by drivers that need to update internal upper and lower boundaries along
423   * with the target frequency and so the core and governors should also invoke
424   * the diver if the target frequency does not change, but the policy min or max
425   * may have changed.
426   */
427  #define CPUFREQ_NEED_UPDATE_LIMITS		BIT(0)
428  
429  /* loops_per_jiffy or other kernel "constants" aren't affected by frequency transitions */
430  #define CPUFREQ_CONST_LOOPS			BIT(1)
431  
432  /*
433   * Set by drivers that want the core to automatically register the cpufreq
434   * driver as a thermal cooling device.
435   */
436  #define CPUFREQ_IS_COOLING_DEV			BIT(2)
437  
438  /*
439   * This should be set by platforms having multiple clock-domains, i.e.
440   * supporting multiple policies. With this sysfs directories of governor would
441   * be created in cpu/cpu<num>/cpufreq/ directory and so they can use the same
442   * governor with different tunables for different clusters.
443   */
444  #define CPUFREQ_HAVE_GOVERNOR_PER_POLICY	BIT(3)
445  
446  /*
447   * Driver will do POSTCHANGE notifications from outside of their ->target()
448   * routine and so must set cpufreq_driver->flags with this flag, so that core
449   * can handle them specially.
450   */
451  #define CPUFREQ_ASYNC_NOTIFICATION		BIT(4)
452  
453  /*
454   * Set by drivers which want cpufreq core to check if CPU is running at a
455   * frequency present in freq-table exposed by the driver. For these drivers if
456   * CPU is found running at an out of table freq, we will try to set it to a freq
457   * from the table. And if that fails, we will stop further boot process by
458   * issuing a BUG_ON().
459   */
460  #define CPUFREQ_NEED_INITIAL_FREQ_CHECK	BIT(5)
461  
462  /*
463   * Set by drivers to disallow use of governors with "dynamic_switching" flag
464   * set.
465   */
466  #define CPUFREQ_NO_AUTO_DYNAMIC_SWITCHING	BIT(6)
467  
468  int cpufreq_register_driver(struct cpufreq_driver *driver_data);
469  void cpufreq_unregister_driver(struct cpufreq_driver *driver_data);
470  
471  bool cpufreq_driver_test_flags(u16 flags);
472  const char *cpufreq_get_current_driver(void);
473  void *cpufreq_get_driver_data(void);
474  
cpufreq_thermal_control_enabled(struct cpufreq_driver * drv)475  static inline int cpufreq_thermal_control_enabled(struct cpufreq_driver *drv)
476  {
477  	return IS_ENABLED(CONFIG_CPU_THERMAL) &&
478  		(drv->flags & CPUFREQ_IS_COOLING_DEV);
479  }
480  
cpufreq_verify_within_limits(struct cpufreq_policy_data * policy,unsigned int min,unsigned int max)481  static inline void cpufreq_verify_within_limits(struct cpufreq_policy_data *policy,
482  						unsigned int min,
483  						unsigned int max)
484  {
485  	policy->max = clamp(policy->max, min, max);
486  	policy->min = clamp(policy->min, min, policy->max);
487  }
488  
489  static inline void
cpufreq_verify_within_cpu_limits(struct cpufreq_policy_data * policy)490  cpufreq_verify_within_cpu_limits(struct cpufreq_policy_data *policy)
491  {
492  	cpufreq_verify_within_limits(policy, policy->cpuinfo.min_freq,
493  				     policy->cpuinfo.max_freq);
494  }
495  
496  #ifdef CONFIG_CPU_FREQ
497  void cpufreq_suspend(void);
498  void cpufreq_resume(void);
499  int cpufreq_generic_suspend(struct cpufreq_policy *policy);
500  #else
cpufreq_suspend(void)501  static inline void cpufreq_suspend(void) {}
cpufreq_resume(void)502  static inline void cpufreq_resume(void) {}
503  #endif
504  
505  /*********************************************************************
506   *                     CPUFREQ NOTIFIER INTERFACE                    *
507   *********************************************************************/
508  
509  #define CPUFREQ_TRANSITION_NOTIFIER	(0)
510  #define CPUFREQ_POLICY_NOTIFIER		(1)
511  
512  /* Transition notifiers */
513  #define CPUFREQ_PRECHANGE		(0)
514  #define CPUFREQ_POSTCHANGE		(1)
515  
516  /* Policy Notifiers  */
517  #define CPUFREQ_CREATE_POLICY		(0)
518  #define CPUFREQ_REMOVE_POLICY		(1)
519  
520  #ifdef CONFIG_CPU_FREQ
521  int cpufreq_register_notifier(struct notifier_block *nb, unsigned int list);
522  int cpufreq_unregister_notifier(struct notifier_block *nb, unsigned int list);
523  
524  void cpufreq_freq_transition_begin(struct cpufreq_policy *policy,
525  		struct cpufreq_freqs *freqs);
526  void cpufreq_freq_transition_end(struct cpufreq_policy *policy,
527  		struct cpufreq_freqs *freqs, int transition_failed);
528  
529  #else /* CONFIG_CPU_FREQ */
cpufreq_register_notifier(struct notifier_block * nb,unsigned int list)530  static inline int cpufreq_register_notifier(struct notifier_block *nb,
531  						unsigned int list)
532  {
533  	return 0;
534  }
cpufreq_unregister_notifier(struct notifier_block * nb,unsigned int list)535  static inline int cpufreq_unregister_notifier(struct notifier_block *nb,
536  						unsigned int list)
537  {
538  	return 0;
539  }
540  #endif /* !CONFIG_CPU_FREQ */
541  
542  /**
543   * cpufreq_scale - "old * mult / div" calculation for large values (32-bit-arch
544   * safe)
545   * @old:   old value
546   * @div:   divisor
547   * @mult:  multiplier
548   *
549   *
550   * new = old * mult / div
551   */
cpufreq_scale(unsigned long old,u_int div,u_int mult)552  static inline unsigned long cpufreq_scale(unsigned long old, u_int div,
553  		u_int mult)
554  {
555  #if BITS_PER_LONG == 32
556  	u64 result = ((u64) old) * ((u64) mult);
557  	do_div(result, div);
558  	return (unsigned long) result;
559  
560  #elif BITS_PER_LONG == 64
561  	unsigned long result = old * ((u64) mult);
562  	result /= div;
563  	return result;
564  #endif
565  }
566  
567  /*********************************************************************
568   *                          CPUFREQ GOVERNORS                        *
569   *********************************************************************/
570  
571  #define CPUFREQ_POLICY_UNKNOWN		(0)
572  /*
573   * If (cpufreq_driver->target) exists, the ->governor decides what frequency
574   * within the limits is used. If (cpufreq_driver->setpolicy> exists, these
575   * two generic policies are available:
576   */
577  #define CPUFREQ_POLICY_POWERSAVE	(1)
578  #define CPUFREQ_POLICY_PERFORMANCE	(2)
579  
580  struct cpufreq_governor {
581  	char	name[CPUFREQ_NAME_LEN];
582  	int	(*init)(struct cpufreq_policy *policy);
583  	void	(*exit)(struct cpufreq_policy *policy);
584  	int	(*start)(struct cpufreq_policy *policy);
585  	void	(*stop)(struct cpufreq_policy *policy);
586  	void	(*limits)(struct cpufreq_policy *policy);
587  	ssize_t	(*show_setspeed)	(struct cpufreq_policy *policy,
588  					 char *buf);
589  	int	(*store_setspeed)	(struct cpufreq_policy *policy,
590  					 unsigned int freq);
591  	struct list_head	governor_list;
592  	struct module		*owner;
593  	u8			flags;
594  };
595  
596  /* Governor flags */
597  
598  /* For governors which change frequency dynamically by themselves */
599  #define CPUFREQ_GOV_DYNAMIC_SWITCHING	BIT(0)
600  
601  /* For governors wanting the target frequency to be set exactly */
602  #define CPUFREQ_GOV_STRICT_TARGET	BIT(1)
603  
604  
605  /* Pass a target to the cpufreq driver */
606  unsigned int cpufreq_driver_fast_switch(struct cpufreq_policy *policy,
607  					unsigned int target_freq);
608  void cpufreq_driver_adjust_perf(unsigned int cpu,
609  				unsigned long min_perf,
610  				unsigned long target_perf,
611  				unsigned long capacity);
612  bool cpufreq_driver_has_adjust_perf(void);
613  int cpufreq_driver_target(struct cpufreq_policy *policy,
614  				 unsigned int target_freq,
615  				 unsigned int relation);
616  int __cpufreq_driver_target(struct cpufreq_policy *policy,
617  				   unsigned int target_freq,
618  				   unsigned int relation);
619  unsigned int cpufreq_driver_resolve_freq(struct cpufreq_policy *policy,
620  					 unsigned int target_freq);
621  unsigned int cpufreq_policy_transition_delay_us(struct cpufreq_policy *policy);
622  int cpufreq_register_governor(struct cpufreq_governor *governor);
623  void cpufreq_unregister_governor(struct cpufreq_governor *governor);
624  int cpufreq_start_governor(struct cpufreq_policy *policy);
625  void cpufreq_stop_governor(struct cpufreq_policy *policy);
626  
627  #define cpufreq_governor_init(__governor)			\
628  static int __init __governor##_init(void)			\
629  {								\
630  	return cpufreq_register_governor(&__governor);	\
631  }								\
632  core_initcall(__governor##_init)
633  
634  #define cpufreq_governor_exit(__governor)			\
635  static void __exit __governor##_exit(void)			\
636  {								\
637  	return cpufreq_unregister_governor(&__governor);	\
638  }								\
639  module_exit(__governor##_exit)
640  
641  struct cpufreq_governor *cpufreq_default_governor(void);
642  struct cpufreq_governor *cpufreq_fallback_governor(void);
643  
cpufreq_policy_apply_limits(struct cpufreq_policy * policy)644  static inline void cpufreq_policy_apply_limits(struct cpufreq_policy *policy)
645  {
646  	if (policy->max < policy->cur)
647  		__cpufreq_driver_target(policy, policy->max,
648  					CPUFREQ_RELATION_HE);
649  	else if (policy->min > policy->cur)
650  		__cpufreq_driver_target(policy, policy->min,
651  					CPUFREQ_RELATION_LE);
652  }
653  
654  /* Governor attribute set */
655  struct gov_attr_set {
656  	struct kobject kobj;
657  	struct list_head policy_list;
658  	struct mutex update_lock;
659  	int usage_count;
660  };
661  
662  /* sysfs ops for cpufreq governors */
663  extern const struct sysfs_ops governor_sysfs_ops;
664  
to_gov_attr_set(struct kobject * kobj)665  static inline struct gov_attr_set *to_gov_attr_set(struct kobject *kobj)
666  {
667  	return container_of(kobj, struct gov_attr_set, kobj);
668  }
669  
670  void gov_attr_set_init(struct gov_attr_set *attr_set, struct list_head *list_node);
671  void gov_attr_set_get(struct gov_attr_set *attr_set, struct list_head *list_node);
672  unsigned int gov_attr_set_put(struct gov_attr_set *attr_set, struct list_head *list_node);
673  
674  /* Governor sysfs attribute */
675  struct governor_attr {
676  	struct attribute attr;
677  	ssize_t (*show)(struct gov_attr_set *attr_set, char *buf);
678  	ssize_t (*store)(struct gov_attr_set *attr_set, const char *buf,
679  			 size_t count);
680  };
681  
682  /*********************************************************************
683   *                     FREQUENCY TABLE HELPERS                       *
684   *********************************************************************/
685  
686  /* Special Values of .frequency field */
687  #define CPUFREQ_ENTRY_INVALID		~0u
688  #define CPUFREQ_TABLE_END		~1u
689  /* Special Values of .flags field */
690  #define CPUFREQ_BOOST_FREQ		(1 << 0)
691  #define CPUFREQ_INEFFICIENT_FREQ	(1 << 1)
692  
693  struct cpufreq_frequency_table {
694  	unsigned int	flags;
695  	unsigned int	driver_data; /* driver specific data, not used by core */
696  	unsigned int	frequency; /* kHz - doesn't need to be in ascending
697  				    * order */
698  };
699  
700  /*
701   * cpufreq_for_each_entry -	iterate over a cpufreq_frequency_table
702   * @pos:	the cpufreq_frequency_table * to use as a loop cursor.
703   * @table:	the cpufreq_frequency_table * to iterate over.
704   */
705  
706  #define cpufreq_for_each_entry(pos, table)	\
707  	for (pos = table; pos->frequency != CPUFREQ_TABLE_END; pos++)
708  
709  /*
710   * cpufreq_for_each_entry_idx -	iterate over a cpufreq_frequency_table
711   *	with index
712   * @pos:	the cpufreq_frequency_table * to use as a loop cursor.
713   * @table:	the cpufreq_frequency_table * to iterate over.
714   * @idx:	the table entry currently being processed
715   */
716  
717  #define cpufreq_for_each_entry_idx(pos, table, idx)	\
718  	for (pos = table, idx = 0; pos->frequency != CPUFREQ_TABLE_END; \
719  		pos++, idx++)
720  
721  /*
722   * cpufreq_for_each_valid_entry -     iterate over a cpufreq_frequency_table
723   *	excluding CPUFREQ_ENTRY_INVALID frequencies.
724   * @pos:        the cpufreq_frequency_table * to use as a loop cursor.
725   * @table:      the cpufreq_frequency_table * to iterate over.
726   */
727  
728  #define cpufreq_for_each_valid_entry(pos, table)			\
729  	for (pos = table; pos->frequency != CPUFREQ_TABLE_END; pos++)	\
730  		if (pos->frequency == CPUFREQ_ENTRY_INVALID)		\
731  			continue;					\
732  		else
733  
734  /*
735   * cpufreq_for_each_valid_entry_idx -     iterate with index over a cpufreq
736   *	frequency_table excluding CPUFREQ_ENTRY_INVALID frequencies.
737   * @pos:	the cpufreq_frequency_table * to use as a loop cursor.
738   * @table:	the cpufreq_frequency_table * to iterate over.
739   * @idx:	the table entry currently being processed
740   */
741  
742  #define cpufreq_for_each_valid_entry_idx(pos, table, idx)		\
743  	cpufreq_for_each_entry_idx(pos, table, idx)			\
744  		if (pos->frequency == CPUFREQ_ENTRY_INVALID)		\
745  			continue;					\
746  		else
747  
748  /**
749   * cpufreq_for_each_efficient_entry_idx - iterate with index over a cpufreq
750   *	frequency_table excluding CPUFREQ_ENTRY_INVALID and
751   *	CPUFREQ_INEFFICIENT_FREQ frequencies.
752   * @pos: the &struct cpufreq_frequency_table to use as a loop cursor.
753   * @table: the &struct cpufreq_frequency_table to iterate over.
754   * @idx: the table entry currently being processed.
755   * @efficiencies: set to true to only iterate over efficient frequencies.
756   */
757  
758  #define cpufreq_for_each_efficient_entry_idx(pos, table, idx, efficiencies)	\
759  	cpufreq_for_each_valid_entry_idx(pos, table, idx)			\
760  		if (efficiencies && (pos->flags & CPUFREQ_INEFFICIENT_FREQ))	\
761  			continue;						\
762  		else
763  
764  
765  int cpufreq_frequency_table_cpuinfo(struct cpufreq_policy *policy,
766  				    struct cpufreq_frequency_table *table);
767  
768  int cpufreq_frequency_table_verify(struct cpufreq_policy_data *policy,
769  				   struct cpufreq_frequency_table *table);
770  int cpufreq_generic_frequency_table_verify(struct cpufreq_policy_data *policy);
771  
772  int cpufreq_table_index_unsorted(struct cpufreq_policy *policy,
773  				 unsigned int target_freq,
774  				 unsigned int relation);
775  int cpufreq_frequency_table_get_index(struct cpufreq_policy *policy,
776  		unsigned int freq);
777  
778  ssize_t cpufreq_show_cpus(const struct cpumask *mask, char *buf);
779  
780  #ifdef CONFIG_CPU_FREQ
781  int cpufreq_boost_trigger_state(int state);
782  bool cpufreq_boost_enabled(void);
783  int cpufreq_enable_boost_support(void);
784  bool policy_has_boost_freq(struct cpufreq_policy *policy);
785  
786  /* Find lowest freq at or above target in a table in ascending order */
cpufreq_table_find_index_al(struct cpufreq_policy * policy,unsigned int target_freq,bool efficiencies)787  static inline int cpufreq_table_find_index_al(struct cpufreq_policy *policy,
788  					      unsigned int target_freq,
789  					      bool efficiencies)
790  {
791  	struct cpufreq_frequency_table *table = policy->freq_table;
792  	struct cpufreq_frequency_table *pos;
793  	unsigned int freq;
794  	int idx, best = -1;
795  
796  	cpufreq_for_each_efficient_entry_idx(pos, table, idx, efficiencies) {
797  		freq = pos->frequency;
798  
799  		if (freq >= target_freq)
800  			return idx;
801  
802  		best = idx;
803  	}
804  
805  	return best;
806  }
807  
808  /* Find lowest freq at or above target in a table in descending order */
cpufreq_table_find_index_dl(struct cpufreq_policy * policy,unsigned int target_freq,bool efficiencies)809  static inline int cpufreq_table_find_index_dl(struct cpufreq_policy *policy,
810  					      unsigned int target_freq,
811  					      bool efficiencies)
812  {
813  	struct cpufreq_frequency_table *table = policy->freq_table;
814  	struct cpufreq_frequency_table *pos;
815  	unsigned int freq;
816  	int idx, best = -1;
817  
818  	cpufreq_for_each_efficient_entry_idx(pos, table, idx, efficiencies) {
819  		freq = pos->frequency;
820  
821  		if (freq == target_freq)
822  			return idx;
823  
824  		if (freq > target_freq) {
825  			best = idx;
826  			continue;
827  		}
828  
829  		/* No freq found above target_freq */
830  		if (best == -1)
831  			return idx;
832  
833  		return best;
834  	}
835  
836  	return best;
837  }
838  
839  /* Works only on sorted freq-tables */
cpufreq_table_find_index_l(struct cpufreq_policy * policy,unsigned int target_freq,bool efficiencies)840  static inline int cpufreq_table_find_index_l(struct cpufreq_policy *policy,
841  					     unsigned int target_freq,
842  					     bool efficiencies)
843  {
844  	target_freq = clamp_val(target_freq, policy->min, policy->max);
845  
846  	if (policy->freq_table_sorted == CPUFREQ_TABLE_SORTED_ASCENDING)
847  		return cpufreq_table_find_index_al(policy, target_freq,
848  						   efficiencies);
849  	else
850  		return cpufreq_table_find_index_dl(policy, target_freq,
851  						   efficiencies);
852  }
853  
854  /* Find highest freq at or below target in a table in ascending order */
cpufreq_table_find_index_ah(struct cpufreq_policy * policy,unsigned int target_freq,bool efficiencies)855  static inline int cpufreq_table_find_index_ah(struct cpufreq_policy *policy,
856  					      unsigned int target_freq,
857  					      bool efficiencies)
858  {
859  	struct cpufreq_frequency_table *table = policy->freq_table;
860  	struct cpufreq_frequency_table *pos;
861  	unsigned int freq;
862  	int idx, best = -1;
863  
864  	cpufreq_for_each_efficient_entry_idx(pos, table, idx, efficiencies) {
865  		freq = pos->frequency;
866  
867  		if (freq == target_freq)
868  			return idx;
869  
870  		if (freq < target_freq) {
871  			best = idx;
872  			continue;
873  		}
874  
875  		/* No freq found below target_freq */
876  		if (best == -1)
877  			return idx;
878  
879  		return best;
880  	}
881  
882  	return best;
883  }
884  
885  /* Find highest freq at or below target in a table in descending order */
cpufreq_table_find_index_dh(struct cpufreq_policy * policy,unsigned int target_freq,bool efficiencies)886  static inline int cpufreq_table_find_index_dh(struct cpufreq_policy *policy,
887  					      unsigned int target_freq,
888  					      bool efficiencies)
889  {
890  	struct cpufreq_frequency_table *table = policy->freq_table;
891  	struct cpufreq_frequency_table *pos;
892  	unsigned int freq;
893  	int idx, best = -1;
894  
895  	cpufreq_for_each_efficient_entry_idx(pos, table, idx, efficiencies) {
896  		freq = pos->frequency;
897  
898  		if (freq <= target_freq)
899  			return idx;
900  
901  		best = idx;
902  	}
903  
904  	return best;
905  }
906  
907  /* Works only on sorted freq-tables */
cpufreq_table_find_index_h(struct cpufreq_policy * policy,unsigned int target_freq,bool efficiencies)908  static inline int cpufreq_table_find_index_h(struct cpufreq_policy *policy,
909  					     unsigned int target_freq,
910  					     bool efficiencies)
911  {
912  	target_freq = clamp_val(target_freq, policy->min, policy->max);
913  
914  	if (policy->freq_table_sorted == CPUFREQ_TABLE_SORTED_ASCENDING)
915  		return cpufreq_table_find_index_ah(policy, target_freq,
916  						   efficiencies);
917  	else
918  		return cpufreq_table_find_index_dh(policy, target_freq,
919  						   efficiencies);
920  }
921  
922  /* Find closest freq to target in a table in ascending order */
cpufreq_table_find_index_ac(struct cpufreq_policy * policy,unsigned int target_freq,bool efficiencies)923  static inline int cpufreq_table_find_index_ac(struct cpufreq_policy *policy,
924  					      unsigned int target_freq,
925  					      bool efficiencies)
926  {
927  	struct cpufreq_frequency_table *table = policy->freq_table;
928  	struct cpufreq_frequency_table *pos;
929  	unsigned int freq;
930  	int idx, best = -1;
931  
932  	cpufreq_for_each_efficient_entry_idx(pos, table, idx, efficiencies) {
933  		freq = pos->frequency;
934  
935  		if (freq == target_freq)
936  			return idx;
937  
938  		if (freq < target_freq) {
939  			best = idx;
940  			continue;
941  		}
942  
943  		/* No freq found below target_freq */
944  		if (best == -1)
945  			return idx;
946  
947  		/* Choose the closest freq */
948  		if (target_freq - table[best].frequency > freq - target_freq)
949  			return idx;
950  
951  		return best;
952  	}
953  
954  	return best;
955  }
956  
957  /* Find closest freq to target in a table in descending order */
cpufreq_table_find_index_dc(struct cpufreq_policy * policy,unsigned int target_freq,bool efficiencies)958  static inline int cpufreq_table_find_index_dc(struct cpufreq_policy *policy,
959  					      unsigned int target_freq,
960  					      bool efficiencies)
961  {
962  	struct cpufreq_frequency_table *table = policy->freq_table;
963  	struct cpufreq_frequency_table *pos;
964  	unsigned int freq;
965  	int idx, best = -1;
966  
967  	cpufreq_for_each_efficient_entry_idx(pos, table, idx, efficiencies) {
968  		freq = pos->frequency;
969  
970  		if (freq == target_freq)
971  			return idx;
972  
973  		if (freq > target_freq) {
974  			best = idx;
975  			continue;
976  		}
977  
978  		/* No freq found above target_freq */
979  		if (best == -1)
980  			return idx;
981  
982  		/* Choose the closest freq */
983  		if (table[best].frequency - target_freq > target_freq - freq)
984  			return idx;
985  
986  		return best;
987  	}
988  
989  	return best;
990  }
991  
992  /* Works only on sorted freq-tables */
cpufreq_table_find_index_c(struct cpufreq_policy * policy,unsigned int target_freq,bool efficiencies)993  static inline int cpufreq_table_find_index_c(struct cpufreq_policy *policy,
994  					     unsigned int target_freq,
995  					     bool efficiencies)
996  {
997  	target_freq = clamp_val(target_freq, policy->min, policy->max);
998  
999  	if (policy->freq_table_sorted == CPUFREQ_TABLE_SORTED_ASCENDING)
1000  		return cpufreq_table_find_index_ac(policy, target_freq,
1001  						   efficiencies);
1002  	else
1003  		return cpufreq_table_find_index_dc(policy, target_freq,
1004  						   efficiencies);
1005  }
1006  
cpufreq_is_in_limits(struct cpufreq_policy * policy,int idx)1007  static inline bool cpufreq_is_in_limits(struct cpufreq_policy *policy, int idx)
1008  {
1009  	unsigned int freq;
1010  
1011  	if (idx < 0)
1012  		return false;
1013  
1014  	freq = policy->freq_table[idx].frequency;
1015  
1016  	return freq == clamp_val(freq, policy->min, policy->max);
1017  }
1018  
cpufreq_frequency_table_target(struct cpufreq_policy * policy,unsigned int target_freq,unsigned int relation)1019  static inline int cpufreq_frequency_table_target(struct cpufreq_policy *policy,
1020  						 unsigned int target_freq,
1021  						 unsigned int relation)
1022  {
1023  	bool efficiencies = policy->efficiencies_available &&
1024  			    (relation & CPUFREQ_RELATION_E);
1025  	int idx;
1026  
1027  	/* cpufreq_table_index_unsorted() has no use for this flag anyway */
1028  	relation &= ~CPUFREQ_RELATION_E;
1029  
1030  	if (unlikely(policy->freq_table_sorted == CPUFREQ_TABLE_UNSORTED))
1031  		return cpufreq_table_index_unsorted(policy, target_freq,
1032  						    relation);
1033  retry:
1034  	switch (relation) {
1035  	case CPUFREQ_RELATION_L:
1036  		idx = cpufreq_table_find_index_l(policy, target_freq,
1037  						 efficiencies);
1038  		break;
1039  	case CPUFREQ_RELATION_H:
1040  		idx = cpufreq_table_find_index_h(policy, target_freq,
1041  						 efficiencies);
1042  		break;
1043  	case CPUFREQ_RELATION_C:
1044  		idx = cpufreq_table_find_index_c(policy, target_freq,
1045  						 efficiencies);
1046  		break;
1047  	default:
1048  		WARN_ON_ONCE(1);
1049  		return 0;
1050  	}
1051  
1052  	/* Limit frequency index to honor policy->min/max */
1053  	if (!cpufreq_is_in_limits(policy, idx) && efficiencies) {
1054  		efficiencies = false;
1055  		goto retry;
1056  	}
1057  
1058  	return idx;
1059  }
1060  
cpufreq_table_count_valid_entries(const struct cpufreq_policy * policy)1061  static inline int cpufreq_table_count_valid_entries(const struct cpufreq_policy *policy)
1062  {
1063  	struct cpufreq_frequency_table *pos;
1064  	int count = 0;
1065  
1066  	if (unlikely(!policy->freq_table))
1067  		return 0;
1068  
1069  	cpufreq_for_each_valid_entry(pos, policy->freq_table)
1070  		count++;
1071  
1072  	return count;
1073  }
1074  
1075  /**
1076   * cpufreq_table_set_inefficient() - Mark a frequency as inefficient
1077   * @policy:	the &struct cpufreq_policy containing the inefficient frequency
1078   * @frequency:	the inefficient frequency
1079   *
1080   * The &struct cpufreq_policy must use a sorted frequency table
1081   *
1082   * Return:	%0 on success or a negative errno code
1083   */
1084  
1085  static inline int
cpufreq_table_set_inefficient(struct cpufreq_policy * policy,unsigned int frequency)1086  cpufreq_table_set_inefficient(struct cpufreq_policy *policy,
1087  			      unsigned int frequency)
1088  {
1089  	struct cpufreq_frequency_table *pos;
1090  
1091  	/* Not supported */
1092  	if (policy->freq_table_sorted == CPUFREQ_TABLE_UNSORTED)
1093  		return -EINVAL;
1094  
1095  	cpufreq_for_each_valid_entry(pos, policy->freq_table) {
1096  		if (pos->frequency == frequency) {
1097  			pos->flags |= CPUFREQ_INEFFICIENT_FREQ;
1098  			policy->efficiencies_available = true;
1099  			return 0;
1100  		}
1101  	}
1102  
1103  	return -EINVAL;
1104  }
1105  
parse_perf_domain(int cpu,const char * list_name,const char * cell_name,struct of_phandle_args * args)1106  static inline int parse_perf_domain(int cpu, const char *list_name,
1107  				    const char *cell_name,
1108  				    struct of_phandle_args *args)
1109  {
1110  	int ret;
1111  
1112  	struct device_node *cpu_np __free(device_node) = of_cpu_device_node_get(cpu);
1113  	if (!cpu_np)
1114  		return -ENODEV;
1115  
1116  	ret = of_parse_phandle_with_args(cpu_np, list_name, cell_name, 0,
1117  					 args);
1118  	if (ret < 0)
1119  		return ret;
1120  	return 0;
1121  }
1122  
of_perf_domain_get_sharing_cpumask(int pcpu,const char * list_name,const char * cell_name,struct cpumask * cpumask,struct of_phandle_args * pargs)1123  static inline int of_perf_domain_get_sharing_cpumask(int pcpu, const char *list_name,
1124  						     const char *cell_name, struct cpumask *cpumask,
1125  						     struct of_phandle_args *pargs)
1126  {
1127  	int cpu, ret;
1128  	struct of_phandle_args args;
1129  
1130  	ret = parse_perf_domain(pcpu, list_name, cell_name, pargs);
1131  	if (ret < 0)
1132  		return ret;
1133  
1134  	cpumask_set_cpu(pcpu, cpumask);
1135  
1136  	for_each_possible_cpu(cpu) {
1137  		if (cpu == pcpu)
1138  			continue;
1139  
1140  		ret = parse_perf_domain(cpu, list_name, cell_name, &args);
1141  		if (ret < 0)
1142  			continue;
1143  
1144  		if (of_phandle_args_equal(pargs, &args))
1145  			cpumask_set_cpu(cpu, cpumask);
1146  
1147  		of_node_put(args.np);
1148  	}
1149  
1150  	return 0;
1151  }
1152  #else
cpufreq_boost_trigger_state(int state)1153  static inline int cpufreq_boost_trigger_state(int state)
1154  {
1155  	return 0;
1156  }
cpufreq_boost_enabled(void)1157  static inline bool cpufreq_boost_enabled(void)
1158  {
1159  	return false;
1160  }
1161  
cpufreq_enable_boost_support(void)1162  static inline int cpufreq_enable_boost_support(void)
1163  {
1164  	return -EINVAL;
1165  }
1166  
policy_has_boost_freq(struct cpufreq_policy * policy)1167  static inline bool policy_has_boost_freq(struct cpufreq_policy *policy)
1168  {
1169  	return false;
1170  }
1171  
1172  static inline int
cpufreq_table_set_inefficient(struct cpufreq_policy * policy,unsigned int frequency)1173  cpufreq_table_set_inefficient(struct cpufreq_policy *policy,
1174  			      unsigned int frequency)
1175  {
1176  	return -EINVAL;
1177  }
1178  
of_perf_domain_get_sharing_cpumask(int pcpu,const char * list_name,const char * cell_name,struct cpumask * cpumask,struct of_phandle_args * pargs)1179  static inline int of_perf_domain_get_sharing_cpumask(int pcpu, const char *list_name,
1180  						     const char *cell_name, struct cpumask *cpumask,
1181  						     struct of_phandle_args *pargs)
1182  {
1183  	return -EOPNOTSUPP;
1184  }
1185  #endif
1186  
1187  extern unsigned int arch_freq_get_on_cpu(int cpu);
1188  
1189  #ifndef arch_set_freq_scale
1190  static __always_inline
arch_set_freq_scale(const struct cpumask * cpus,unsigned long cur_freq,unsigned long max_freq)1191  void arch_set_freq_scale(const struct cpumask *cpus,
1192  			 unsigned long cur_freq,
1193  			 unsigned long max_freq)
1194  {
1195  }
1196  #endif
1197  
1198  /* the following are really really optional */
1199  extern struct freq_attr cpufreq_freq_attr_scaling_available_freqs;
1200  extern struct freq_attr cpufreq_freq_attr_scaling_boost_freqs;
1201  extern struct freq_attr *cpufreq_generic_attr[];
1202  int cpufreq_table_validate_and_sort(struct cpufreq_policy *policy);
1203  
1204  unsigned int cpufreq_generic_get(unsigned int cpu);
1205  void cpufreq_generic_init(struct cpufreq_policy *policy,
1206  		struct cpufreq_frequency_table *table,
1207  		unsigned int transition_latency);
1208  
cpufreq_register_em_with_opp(struct cpufreq_policy * policy)1209  static inline void cpufreq_register_em_with_opp(struct cpufreq_policy *policy)
1210  {
1211  	dev_pm_opp_of_register_em(get_cpu_device(policy->cpu),
1212  				  policy->related_cpus);
1213  }
1214  #endif /* _LINUX_CPUFREQ_H */
1215