1/* SPDX-License-Identifier: GPL-2.0 */
2#ifndef _ASM_X86_TLB_H
3#define _ASM_X86_TLB_H
4
5#define tlb_flush tlb_flush
6static inline void tlb_flush(struct mmu_gather *tlb);
7
8#include <asm-generic/tlb.h>
9#include <linux/kernel.h>
10#include <vdso/bits.h>
11#include <vdso/page.h>
12
13static inline void tlb_flush(struct mmu_gather *tlb)
14{
15 unsigned long start = 0UL, end = TLB_FLUSH_ALL;
16 unsigned int stride_shift = tlb_get_unmap_shift(tlb);
17
18 if (!tlb->fullmm && !tlb->need_flush_all) {
19 start = tlb->start;
20 end = tlb->end;
21 }
22
23 flush_tlb_mm_range(mm: tlb->mm, start, end, stride_shift, freed_tables: tlb->freed_tables);
24}
25
26static inline void invlpg(unsigned long addr)
27{
28 asm volatile("invlpg (%0)" ::"r" (addr) : "memory");
29}
30
31enum addr_stride {
32 PTE_STRIDE = 0,
33 PMD_STRIDE = 1
34};
35
36/*
37 * INVLPGB can be targeted by virtual address, PCID, ASID, or any combination
38 * of the three. For example:
39 * - FLAG_VA | FLAG_INCLUDE_GLOBAL: invalidate all TLB entries at the address
40 * - FLAG_PCID: invalidate all TLB entries matching the PCID
41 *
42 * The first is used to invalidate (kernel) mappings at a particular
43 * address across all processes.
44 *
45 * The latter invalidates all TLB entries matching a PCID.
46 */
47#define INVLPGB_FLAG_VA BIT(0)
48#define INVLPGB_FLAG_PCID BIT(1)
49#define INVLPGB_FLAG_ASID BIT(2)
50#define INVLPGB_FLAG_INCLUDE_GLOBAL BIT(3)
51#define INVLPGB_FLAG_FINAL_ONLY BIT(4)
52#define INVLPGB_FLAG_INCLUDE_NESTED BIT(5)
53
54/* The implied mode when all bits are clear: */
55#define INVLPGB_MODE_ALL_NONGLOBALS 0UL
56
57#ifdef CONFIG_BROADCAST_TLB_FLUSH
58/*
59 * INVLPGB does broadcast TLB invalidation across all the CPUs in the system.
60 *
61 * The INVLPGB instruction is weakly ordered, and a batch of invalidations can
62 * be done in a parallel fashion.
63 *
64 * The instruction takes the number of extra pages to invalidate, beyond the
65 * first page, while __invlpgb gets the more human readable number of pages to
66 * invalidate.
67 *
68 * The bits in rax[0:2] determine respectively which components of the address
69 * (VA, PCID, ASID) get compared when flushing. If neither bits are set, *any*
70 * address in the specified range matches.
71 *
72 * Since it is desired to only flush TLB entries for the ASID that is executing
73 * the instruction (a host/hypervisor or a guest), the ASID valid bit should
74 * always be set. On a host/hypervisor, the hardware will use the ASID value
75 * specified in EDX[15:0] (which should be 0). On a guest, the hardware will
76 * use the actual ASID value of the guest.
77 *
78 * TLBSYNC is used to ensure that pending INVLPGB invalidations initiated from
79 * this CPU have completed.
80 */
81static inline void __invlpgb(unsigned long asid, unsigned long pcid,
82 unsigned long addr, u16 nr_pages,
83 enum addr_stride stride, u8 flags)
84{
85 u64 rax = addr | flags | INVLPGB_FLAG_ASID;
86 u32 ecx = (stride << 31) | (nr_pages - 1);
87 u32 edx = (pcid << 16) | asid;
88
89 /* The low bits in rax are for flags. Verify addr is clean. */
90 VM_WARN_ON_ONCE(addr & ~PAGE_MASK);
91
92 /* INVLPGB; supported in binutils >= 2.36. */
93 asm volatile(".byte 0x0f, 0x01, 0xfe" :: "a" (rax), "c" (ecx), "d" (edx));
94}
95
96static inline void __invlpgb_all(unsigned long asid, unsigned long pcid, u8 flags)
97{
98 __invlpgb(asid, pcid, addr: 0, nr_pages: 1, stride: 0, flags);
99}
100
101static inline void __tlbsync(void)
102{
103 /*
104 * TLBSYNC waits for INVLPGB instructions originating on the same CPU
105 * to have completed. Print a warning if the task has been migrated,
106 * and might not be waiting on all the INVLPGBs issued during this TLB
107 * invalidation sequence.
108 */
109 cant_migrate();
110
111 /* TLBSYNC: supported in binutils >= 0.36. */
112 asm volatile(".byte 0x0f, 0x01, 0xff" ::: "memory");
113}
114#else
115/* Some compilers (I'm looking at you clang!) simply can't do DCE */
116static inline void __invlpgb(unsigned long asid, unsigned long pcid,
117 unsigned long addr, u16 nr_pages,
118 enum addr_stride s, u8 flags) { }
119static inline void __invlpgb_all(unsigned long asid, unsigned long pcid, u8 flags) { }
120static inline void __tlbsync(void) { }
121#endif
122
123static inline void invlpgb_flush_user_nr_nosync(unsigned long pcid,
124 unsigned long addr,
125 u16 nr, bool stride)
126{
127 enum addr_stride str = stride ? PMD_STRIDE : PTE_STRIDE;
128 u8 flags = INVLPGB_FLAG_PCID | INVLPGB_FLAG_VA;
129
130 __invlpgb(asid: 0, pcid, addr, nr_pages: nr, stride: str, flags);
131}
132
133/* Flush all mappings for a given PCID, not including globals. */
134static inline void invlpgb_flush_single_pcid_nosync(unsigned long pcid)
135{
136 __invlpgb_all(asid: 0, pcid, INVLPGB_FLAG_PCID);
137}
138
139/* Flush all mappings, including globals, for all PCIDs. */
140static inline void invlpgb_flush_all(void)
141{
142 /*
143 * TLBSYNC at the end needs to make sure all flushes done on the
144 * current CPU have been executed system-wide. Therefore, make
145 * sure nothing gets migrated in-between but disable preemption
146 * as it is cheaper.
147 */
148 guard(preempt)();
149 __invlpgb_all(asid: 0, pcid: 0, INVLPGB_FLAG_INCLUDE_GLOBAL);
150 __tlbsync();
151}
152
153/* Flush addr, including globals, for all PCIDs. */
154static inline void invlpgb_flush_addr_nosync(unsigned long addr, u16 nr)
155{
156 __invlpgb(asid: 0, pcid: 0, addr, nr_pages: nr, stride: PTE_STRIDE, INVLPGB_FLAG_INCLUDE_GLOBAL);
157}
158
159/* Flush all mappings for all PCIDs except globals. */
160static inline void invlpgb_flush_all_nonglobals(void)
161{
162 guard(preempt)();
163 __invlpgb_all(asid: 0, pcid: 0, INVLPGB_MODE_ALL_NONGLOBALS);
164 __tlbsync();
165}
166#endif /* _ASM_X86_TLB_H */
167