1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef _LINUX_MM_TYPES_TASK_H
3 #define _LINUX_MM_TYPES_TASK_H
4 
5 /*
6  * Here are the definitions of the MM data types that are embedded in 'struct task_struct'.
7  *
8  * (These are defined separately to decouple sched.h from mm_types.h as much as possible.)
9  */
10 
11 #include <linux/types.h>
12 
13 #include <asm/page.h>
14 
15 #ifdef CONFIG_ARCH_WANT_BATCHED_UNMAP_TLB_FLUSH
16 #include <asm/tlbbatch.h>
17 #endif
18 
19 #define ALLOC_SPLIT_PTLOCKS	(SPINLOCK_SIZE > BITS_PER_LONG/8)
20 
21 /*
22  * When updating this, please also update struct resident_page_types[] in
23  * kernel/fork.c
24  */
25 enum {
26 	MM_FILEPAGES,	/* Resident file mapping pages */
27 	MM_ANONPAGES,	/* Resident anonymous pages */
28 	MM_SWAPENTS,	/* Anonymous swap entries */
29 	MM_SHMEMPAGES,	/* Resident shared memory pages */
30 	NR_MM_COUNTERS
31 };
32 
33 struct page;
34 
35 struct page_frag {
36 	struct page *page;
37 #if (BITS_PER_LONG > 32) || (PAGE_SIZE >= 65536)
38 	__u32 offset;
39 	__u32 size;
40 #else
41 	__u16 offset;
42 	__u16 size;
43 #endif
44 };
45 
46 /* Track pages that require TLB flushes */
47 struct tlbflush_unmap_batch {
48 #ifdef CONFIG_ARCH_WANT_BATCHED_UNMAP_TLB_FLUSH
49 	/*
50 	 * The arch code makes the following promise: generic code can modify a
51 	 * PTE, then call arch_tlbbatch_add_pending() (which internally provides
52 	 * all needed barriers), then call arch_tlbbatch_flush(), and the entries
53 	 * will be flushed on all CPUs by the time that arch_tlbbatch_flush()
54 	 * returns.
55 	 */
56 	struct arch_tlbflush_unmap_batch arch;
57 
58 	/* True if a flush is needed. */
59 	bool flush_required;
60 
61 	/*
62 	 * If true then the PTE was dirty when unmapped. The entry must be
63 	 * flushed before IO is initiated or a stale TLB entry potentially
64 	 * allows an update without redirtying the page.
65 	 */
66 	bool writable;
67 #endif
68 };
69 
70 #endif /* _LINUX_MM_TYPES_TASK_H */
71