1  /* SPDX-License-Identifier: GPL-2.0 */
2  #ifndef _LINUX_PAGE_COUNTER_H
3  #define _LINUX_PAGE_COUNTER_H
4  
5  #include <linux/atomic.h>
6  #include <linux/cache.h>
7  #include <linux/limits.h>
8  #include <asm/page.h>
9  
10  struct page_counter {
11  	/*
12  	 * Make sure 'usage' does not share cacheline with any other field. The
13  	 * memcg->memory.usage is a hot member of struct mem_cgroup.
14  	 */
15  	atomic_long_t usage;
16  	CACHELINE_PADDING(_pad1_);
17  
18  	/* effective memory.min and memory.min usage tracking */
19  	unsigned long emin;
20  	atomic_long_t min_usage;
21  	atomic_long_t children_min_usage;
22  
23  	/* effective memory.low and memory.low usage tracking */
24  	unsigned long elow;
25  	atomic_long_t low_usage;
26  	atomic_long_t children_low_usage;
27  
28  	unsigned long watermark;
29  	/* Latest cg2 reset watermark */
30  	unsigned long local_watermark;
31  	unsigned long failcnt;
32  
33  	/* Keep all the read most fields in a separete cacheline. */
34  	CACHELINE_PADDING(_pad2_);
35  
36  	bool protection_support;
37  	unsigned long min;
38  	unsigned long low;
39  	unsigned long high;
40  	unsigned long max;
41  	struct page_counter *parent;
42  } ____cacheline_internodealigned_in_smp;
43  
44  #if BITS_PER_LONG == 32
45  #define PAGE_COUNTER_MAX LONG_MAX
46  #else
47  #define PAGE_COUNTER_MAX (LONG_MAX / PAGE_SIZE)
48  #endif
49  
50  /*
51   * Protection is supported only for the first counter (with id 0).
52   */
page_counter_init(struct page_counter * counter,struct page_counter * parent,bool protection_support)53  static inline void page_counter_init(struct page_counter *counter,
54  				     struct page_counter *parent,
55  				     bool protection_support)
56  {
57  	counter->usage = (atomic_long_t)ATOMIC_LONG_INIT(0);
58  	counter->max = PAGE_COUNTER_MAX;
59  	counter->parent = parent;
60  	counter->protection_support = protection_support;
61  }
62  
page_counter_read(struct page_counter * counter)63  static inline unsigned long page_counter_read(struct page_counter *counter)
64  {
65  	return atomic_long_read(&counter->usage);
66  }
67  
68  void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages);
69  void page_counter_charge(struct page_counter *counter, unsigned long nr_pages);
70  bool page_counter_try_charge(struct page_counter *counter,
71  			     unsigned long nr_pages,
72  			     struct page_counter **fail);
73  void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages);
74  void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages);
75  void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages);
76  
page_counter_set_high(struct page_counter * counter,unsigned long nr_pages)77  static inline void page_counter_set_high(struct page_counter *counter,
78  					 unsigned long nr_pages)
79  {
80  	WRITE_ONCE(counter->high, nr_pages);
81  }
82  
83  int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages);
84  int page_counter_memparse(const char *buf, const char *max,
85  			  unsigned long *nr_pages);
86  
page_counter_reset_watermark(struct page_counter * counter)87  static inline void page_counter_reset_watermark(struct page_counter *counter)
88  {
89  	unsigned long usage = page_counter_read(counter);
90  
91  	/*
92  	 * Update local_watermark first, so it's always <= watermark
93  	 * (modulo CPU/compiler re-ordering)
94  	 */
95  	counter->local_watermark = usage;
96  	counter->watermark = usage;
97  }
98  
99  #ifdef CONFIG_MEMCG
100  void page_counter_calculate_protection(struct page_counter *root,
101  				       struct page_counter *counter,
102  				       bool recursive_protection);
103  #else
page_counter_calculate_protection(struct page_counter * root,struct page_counter * counter,bool recursive_protection)104  static inline void page_counter_calculate_protection(struct page_counter *root,
105  						     struct page_counter *counter,
106  						     bool recursive_protection) {}
107  #endif
108  
109  #endif /* _LINUX_PAGE_COUNTER_H */
110