| 1 | /* SPDX-License-Identifier: GPL-2.0 */ | 
|---|
| 2 | #ifndef _LINUX_PAGE_COUNTER_H | 
|---|
| 3 | #define _LINUX_PAGE_COUNTER_H | 
|---|
| 4 |  | 
|---|
| 5 | #include <linux/atomic.h> | 
|---|
| 6 | #include <linux/cache.h> | 
|---|
| 7 | #include <linux/limits.h> | 
|---|
| 8 | #include <asm/page.h> | 
|---|
| 9 |  | 
|---|
| 10 | struct page_counter { | 
|---|
| 11 | /* | 
|---|
| 12 | * Make sure 'usage' does not share cacheline with any other field in | 
|---|
| 13 | * v2. The memcg->memory.usage is a hot member of struct mem_cgroup. | 
|---|
| 14 | */ | 
|---|
| 15 | atomic_long_t usage; | 
|---|
| 16 | unsigned long failcnt; /* v1-only field */ | 
|---|
| 17 |  | 
|---|
| 18 | CACHELINE_PADDING(_pad1_); | 
|---|
| 19 |  | 
|---|
| 20 | /* effective memory.min and memory.min usage tracking */ | 
|---|
| 21 | unsigned long emin; | 
|---|
| 22 | atomic_long_t min_usage; | 
|---|
| 23 | atomic_long_t children_min_usage; | 
|---|
| 24 |  | 
|---|
| 25 | /* effective memory.low and memory.low usage tracking */ | 
|---|
| 26 | unsigned long elow; | 
|---|
| 27 | atomic_long_t low_usage; | 
|---|
| 28 | atomic_long_t children_low_usage; | 
|---|
| 29 |  | 
|---|
| 30 | unsigned long watermark; | 
|---|
| 31 | /* Latest cg2 reset watermark */ | 
|---|
| 32 | unsigned long local_watermark; | 
|---|
| 33 |  | 
|---|
| 34 | /* Keep all the read most fields in a separete cacheline. */ | 
|---|
| 35 | CACHELINE_PADDING(_pad2_); | 
|---|
| 36 |  | 
|---|
| 37 | bool protection_support; | 
|---|
| 38 | bool track_failcnt; | 
|---|
| 39 | unsigned long min; | 
|---|
| 40 | unsigned long low; | 
|---|
| 41 | unsigned long high; | 
|---|
| 42 | unsigned long max; | 
|---|
| 43 | struct page_counter *parent; | 
|---|
| 44 | } ____cacheline_internodealigned_in_smp; | 
|---|
| 45 |  | 
|---|
| 46 | #if BITS_PER_LONG == 32 | 
|---|
| 47 | #define PAGE_COUNTER_MAX LONG_MAX | 
|---|
| 48 | #else | 
|---|
| 49 | #define PAGE_COUNTER_MAX (LONG_MAX / PAGE_SIZE) | 
|---|
| 50 | #endif | 
|---|
| 51 |  | 
|---|
| 52 | /* | 
|---|
| 53 | * Protection is supported only for the first counter (with id 0). | 
|---|
| 54 | */ | 
|---|
| 55 | static inline void page_counter_init(struct page_counter *counter, | 
|---|
| 56 | struct page_counter *parent, | 
|---|
| 57 | bool protection_support) | 
|---|
| 58 | { | 
|---|
| 59 | counter->usage = (atomic_long_t)ATOMIC_LONG_INIT(0); | 
|---|
| 60 | counter->max = PAGE_COUNTER_MAX; | 
|---|
| 61 | counter->parent = parent; | 
|---|
| 62 | counter->protection_support = protection_support; | 
|---|
| 63 | counter->track_failcnt = false; | 
|---|
| 64 | } | 
|---|
| 65 |  | 
|---|
| 66 | static inline unsigned long page_counter_read(struct page_counter *counter) | 
|---|
| 67 | { | 
|---|
| 68 | return atomic_long_read(v: &counter->usage); | 
|---|
| 69 | } | 
|---|
| 70 |  | 
|---|
| 71 | void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages); | 
|---|
| 72 | void page_counter_charge(struct page_counter *counter, unsigned long nr_pages); | 
|---|
| 73 | bool page_counter_try_charge(struct page_counter *counter, | 
|---|
| 74 | unsigned long nr_pages, | 
|---|
| 75 | struct page_counter **fail); | 
|---|
| 76 | void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages); | 
|---|
| 77 | void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages); | 
|---|
| 78 | void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages); | 
|---|
| 79 |  | 
|---|
| 80 | static inline void page_counter_set_high(struct page_counter *counter, | 
|---|
| 81 | unsigned long nr_pages) | 
|---|
| 82 | { | 
|---|
| 83 | WRITE_ONCE(counter->high, nr_pages); | 
|---|
| 84 | } | 
|---|
| 85 |  | 
|---|
| 86 | int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages); | 
|---|
| 87 | int page_counter_memparse(const char *buf, const char *max, | 
|---|
| 88 | unsigned long *nr_pages); | 
|---|
| 89 |  | 
|---|
| 90 | static inline void page_counter_reset_watermark(struct page_counter *counter) | 
|---|
| 91 | { | 
|---|
| 92 | unsigned long usage = page_counter_read(counter); | 
|---|
| 93 |  | 
|---|
| 94 | /* | 
|---|
| 95 | * Update local_watermark first, so it's always <= watermark | 
|---|
| 96 | * (modulo CPU/compiler re-ordering) | 
|---|
| 97 | */ | 
|---|
| 98 | counter->local_watermark = usage; | 
|---|
| 99 | counter->watermark = usage; | 
|---|
| 100 | } | 
|---|
| 101 |  | 
|---|
| 102 | #if IS_ENABLED(CONFIG_MEMCG) || IS_ENABLED(CONFIG_CGROUP_DMEM) | 
|---|
| 103 | void page_counter_calculate_protection(struct page_counter *root, | 
|---|
| 104 | struct page_counter *counter, | 
|---|
| 105 | bool recursive_protection); | 
|---|
| 106 | #else | 
|---|
| 107 | static inline void page_counter_calculate_protection(struct page_counter *root, | 
|---|
| 108 | struct page_counter *counter, | 
|---|
| 109 | bool recursive_protection) {} | 
|---|
| 110 | #endif | 
|---|
| 111 |  | 
|---|
| 112 | #endif /* _LINUX_PAGE_COUNTER_H */ | 
|---|
| 113 |  | 
|---|