| 1 | /* SPDX-License-Identifier: GPL-2.0 */ | 
|---|
| 2 | #ifndef _ASM_X86_TLB_H | 
|---|
| 3 | #define _ASM_X86_TLB_H | 
|---|
| 4 |  | 
|---|
| 5 | #define tlb_flush tlb_flush | 
|---|
| 6 | static inline void tlb_flush(struct mmu_gather *tlb); | 
|---|
| 7 |  | 
|---|
| 8 | #include <asm-generic/tlb.h> | 
|---|
| 9 | #include <linux/kernel.h> | 
|---|
| 10 | #include <vdso/bits.h> | 
|---|
| 11 | #include <vdso/page.h> | 
|---|
| 12 |  | 
|---|
| 13 | static inline void tlb_flush(struct mmu_gather *tlb) | 
|---|
| 14 | { | 
|---|
| 15 | unsigned long start = 0UL, end = TLB_FLUSH_ALL; | 
|---|
| 16 | unsigned int stride_shift = tlb_get_unmap_shift(tlb); | 
|---|
| 17 |  | 
|---|
| 18 | if (!tlb->fullmm && !tlb->need_flush_all) { | 
|---|
| 19 | start = tlb->start; | 
|---|
| 20 | end = tlb->end; | 
|---|
| 21 | } | 
|---|
| 22 |  | 
|---|
| 23 | flush_tlb_mm_range(mm: tlb->mm, start, end, stride_shift, freed_tables: tlb->freed_tables); | 
|---|
| 24 | } | 
|---|
| 25 |  | 
|---|
| 26 | static inline void invlpg(unsigned long addr) | 
|---|
| 27 | { | 
|---|
| 28 | asm volatile( "invlpg (%0)":: "r"(addr) : "memory"); | 
|---|
| 29 | } | 
|---|
| 30 |  | 
|---|
| 31 | enum addr_stride { | 
|---|
| 32 | PTE_STRIDE = 0, | 
|---|
| 33 | PMD_STRIDE = 1 | 
|---|
| 34 | }; | 
|---|
| 35 |  | 
|---|
| 36 | /* | 
|---|
| 37 | * INVLPGB can be targeted by virtual address, PCID, ASID, or any combination | 
|---|
| 38 | * of the three. For example: | 
|---|
| 39 | * - FLAG_VA | FLAG_INCLUDE_GLOBAL: invalidate all TLB entries at the address | 
|---|
| 40 | * - FLAG_PCID:			    invalidate all TLB entries matching the PCID | 
|---|
| 41 | * | 
|---|
| 42 | * The first is used to invalidate (kernel) mappings at a particular | 
|---|
| 43 | * address across all processes. | 
|---|
| 44 | * | 
|---|
| 45 | * The latter invalidates all TLB entries matching a PCID. | 
|---|
| 46 | */ | 
|---|
| 47 | #define INVLPGB_FLAG_VA			BIT(0) | 
|---|
| 48 | #define INVLPGB_FLAG_PCID		BIT(1) | 
|---|
| 49 | #define INVLPGB_FLAG_ASID		BIT(2) | 
|---|
| 50 | #define INVLPGB_FLAG_INCLUDE_GLOBAL	BIT(3) | 
|---|
| 51 | #define INVLPGB_FLAG_FINAL_ONLY		BIT(4) | 
|---|
| 52 | #define INVLPGB_FLAG_INCLUDE_NESTED	BIT(5) | 
|---|
| 53 |  | 
|---|
| 54 | /* The implied mode when all bits are clear: */ | 
|---|
| 55 | #define INVLPGB_MODE_ALL_NONGLOBALS	0UL | 
|---|
| 56 |  | 
|---|
| 57 | #ifdef CONFIG_BROADCAST_TLB_FLUSH | 
|---|
| 58 | /* | 
|---|
| 59 | * INVLPGB does broadcast TLB invalidation across all the CPUs in the system. | 
|---|
| 60 | * | 
|---|
| 61 | * The INVLPGB instruction is weakly ordered, and a batch of invalidations can | 
|---|
| 62 | * be done in a parallel fashion. | 
|---|
| 63 | * | 
|---|
| 64 | * The instruction takes the number of extra pages to invalidate, beyond the | 
|---|
| 65 | * first page, while __invlpgb gets the more human readable number of pages to | 
|---|
| 66 | * invalidate. | 
|---|
| 67 | * | 
|---|
| 68 | * The bits in rax[0:2] determine respectively which components of the address | 
|---|
| 69 | * (VA, PCID, ASID) get compared when flushing. If neither bits are set, *any* | 
|---|
| 70 | * address in the specified range matches. | 
|---|
| 71 | * | 
|---|
| 72 | * Since it is desired to only flush TLB entries for the ASID that is executing | 
|---|
| 73 | * the instruction (a host/hypervisor or a guest), the ASID valid bit should | 
|---|
| 74 | * always be set. On a host/hypervisor, the hardware will use the ASID value | 
|---|
| 75 | * specified in EDX[15:0] (which should be 0). On a guest, the hardware will | 
|---|
| 76 | * use the actual ASID value of the guest. | 
|---|
| 77 | * | 
|---|
| 78 | * TLBSYNC is used to ensure that pending INVLPGB invalidations initiated from | 
|---|
| 79 | * this CPU have completed. | 
|---|
| 80 | */ | 
|---|
| 81 | static inline void __invlpgb(unsigned long asid, unsigned long pcid, | 
|---|
| 82 | unsigned long addr, u16 nr_pages, | 
|---|
| 83 | enum addr_stride stride, u8 flags) | 
|---|
| 84 | { | 
|---|
| 85 | u64 rax = addr | flags | INVLPGB_FLAG_ASID; | 
|---|
| 86 | u32 ecx = (stride << 31) | (nr_pages - 1); | 
|---|
| 87 | u32 edx = (pcid << 16) | asid; | 
|---|
| 88 |  | 
|---|
| 89 | /* The low bits in rax are for flags. Verify addr is clean. */ | 
|---|
| 90 | VM_WARN_ON_ONCE(addr & ~PAGE_MASK); | 
|---|
| 91 |  | 
|---|
| 92 | /* INVLPGB; supported in binutils >= 2.36. */ | 
|---|
| 93 | asm volatile( ".byte 0x0f, 0x01, 0xfe":: "a"(rax), "c"(ecx), "d"(edx)); | 
|---|
| 94 | } | 
|---|
| 95 |  | 
|---|
| 96 | static inline void __invlpgb_all(unsigned long asid, unsigned long pcid, u8 flags) | 
|---|
| 97 | { | 
|---|
| 98 | __invlpgb(asid, pcid, addr: 0, nr_pages: 1, stride: 0, flags); | 
|---|
| 99 | } | 
|---|
| 100 |  | 
|---|
| 101 | static inline void __tlbsync(void) | 
|---|
| 102 | { | 
|---|
| 103 | /* | 
|---|
| 104 | * TLBSYNC waits for INVLPGB instructions originating on the same CPU | 
|---|
| 105 | * to have completed. Print a warning if the task has been migrated, | 
|---|
| 106 | * and might not be waiting on all the INVLPGBs issued during this TLB | 
|---|
| 107 | * invalidation sequence. | 
|---|
| 108 | */ | 
|---|
| 109 | cant_migrate(); | 
|---|
| 110 |  | 
|---|
| 111 | /* TLBSYNC: supported in binutils >= 0.36. */ | 
|---|
| 112 | asm volatile( ".byte 0x0f, 0x01, 0xff"::: "memory"); | 
|---|
| 113 | } | 
|---|
| 114 | #else | 
|---|
| 115 | /* Some compilers (I'm looking at you clang!) simply can't do DCE */ | 
|---|
| 116 | static inline void __invlpgb(unsigned long asid, unsigned long pcid, | 
|---|
| 117 | unsigned long addr, u16 nr_pages, | 
|---|
| 118 | enum addr_stride s, u8 flags) { } | 
|---|
| 119 | static inline void __invlpgb_all(unsigned long asid, unsigned long pcid, u8 flags) { } | 
|---|
| 120 | static inline void __tlbsync(void) { } | 
|---|
| 121 | #endif | 
|---|
| 122 |  | 
|---|
| 123 | static inline void invlpgb_flush_user_nr_nosync(unsigned long pcid, | 
|---|
| 124 | unsigned long addr, | 
|---|
| 125 | u16 nr, bool stride) | 
|---|
| 126 | { | 
|---|
| 127 | enum addr_stride str = stride ? PMD_STRIDE : PTE_STRIDE; | 
|---|
| 128 | u8 flags = INVLPGB_FLAG_PCID | INVLPGB_FLAG_VA; | 
|---|
| 129 |  | 
|---|
| 130 | __invlpgb(asid: 0, pcid, addr, nr_pages: nr, stride: str, flags); | 
|---|
| 131 | } | 
|---|
| 132 |  | 
|---|
| 133 | /* Flush all mappings for a given PCID, not including globals. */ | 
|---|
| 134 | static inline void invlpgb_flush_single_pcid_nosync(unsigned long pcid) | 
|---|
| 135 | { | 
|---|
| 136 | __invlpgb_all(asid: 0, pcid, INVLPGB_FLAG_PCID); | 
|---|
| 137 | } | 
|---|
| 138 |  | 
|---|
| 139 | /* Flush all mappings, including globals, for all PCIDs. */ | 
|---|
| 140 | static inline void invlpgb_flush_all(void) | 
|---|
| 141 | { | 
|---|
| 142 | /* | 
|---|
| 143 | * TLBSYNC at the end needs to make sure all flushes done on the | 
|---|
| 144 | * current CPU have been executed system-wide. Therefore, make | 
|---|
| 145 | * sure nothing gets migrated in-between but disable preemption | 
|---|
| 146 | * as it is cheaper. | 
|---|
| 147 | */ | 
|---|
| 148 | guard(preempt)(); | 
|---|
| 149 | __invlpgb_all(asid: 0, pcid: 0, INVLPGB_FLAG_INCLUDE_GLOBAL); | 
|---|
| 150 | __tlbsync(); | 
|---|
| 151 | } | 
|---|
| 152 |  | 
|---|
| 153 | /* Flush addr, including globals, for all PCIDs. */ | 
|---|
| 154 | static inline void invlpgb_flush_addr_nosync(unsigned long addr, u16 nr) | 
|---|
| 155 | { | 
|---|
| 156 | __invlpgb(asid: 0, pcid: 0, addr, nr_pages: nr, stride: PTE_STRIDE, INVLPGB_FLAG_INCLUDE_GLOBAL); | 
|---|
| 157 | } | 
|---|
| 158 |  | 
|---|
| 159 | /* Flush all mappings for all PCIDs except globals. */ | 
|---|
| 160 | static inline void invlpgb_flush_all_nonglobals(void) | 
|---|
| 161 | { | 
|---|
| 162 | guard(preempt)(); | 
|---|
| 163 | __invlpgb_all(asid: 0, pcid: 0, INVLPGB_MODE_ALL_NONGLOBALS); | 
|---|
| 164 | __tlbsync(); | 
|---|
| 165 | } | 
|---|
| 166 | #endif /* _ASM_X86_TLB_H */ | 
|---|
| 167 |  | 
|---|