| 1 | /* | 
|---|
| 2 | * VMware Detection code. | 
|---|
| 3 | * | 
|---|
| 4 | * Copyright (C) 2008, VMware, Inc. | 
|---|
| 5 | * Author : Alok N Kataria <akataria@vmware.com> | 
|---|
| 6 | * | 
|---|
| 7 | * This program is free software; you can redistribute it and/or modify | 
|---|
| 8 | * it under the terms of the GNU General Public License as published by | 
|---|
| 9 | * the Free Software Foundation; either version 2 of the License, or | 
|---|
| 10 | * (at your option) any later version. | 
|---|
| 11 | * | 
|---|
| 12 | * This program is distributed in the hope that it will be useful, but | 
|---|
| 13 | * WITHOUT ANY WARRANTY; without even the implied warranty of | 
|---|
| 14 | * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or | 
|---|
| 15 | * NON INFRINGEMENT.  See the GNU General Public License for more | 
|---|
| 16 | * details. | 
|---|
| 17 | * | 
|---|
| 18 | * You should have received a copy of the GNU General Public License | 
|---|
| 19 | * along with this program; if not, write to the Free Software | 
|---|
| 20 | * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. | 
|---|
| 21 | * | 
|---|
| 22 | */ | 
|---|
| 23 |  | 
|---|
| 24 | #include <linux/dmi.h> | 
|---|
| 25 | #include <linux/init.h> | 
|---|
| 26 | #include <linux/export.h> | 
|---|
| 27 | #include <linux/clocksource.h> | 
|---|
| 28 | #include <linux/cpu.h> | 
|---|
| 29 | #include <linux/efi.h> | 
|---|
| 30 | #include <linux/reboot.h> | 
|---|
| 31 | #include <linux/static_call.h> | 
|---|
| 32 | #include <asm/div64.h> | 
|---|
| 33 | #include <asm/x86_init.h> | 
|---|
| 34 | #include <asm/hypervisor.h> | 
|---|
| 35 | #include <asm/timer.h> | 
|---|
| 36 | #include <asm/apic.h> | 
|---|
| 37 | #include <asm/vmware.h> | 
|---|
| 38 | #include <asm/svm.h> | 
|---|
| 39 |  | 
|---|
| 40 | #undef pr_fmt | 
|---|
| 41 | #define pr_fmt(fmt)	"vmware: " fmt | 
|---|
| 42 |  | 
|---|
| 43 | #define CPUID_VMWARE_INFO_LEAF               0x40000000 | 
|---|
| 44 | #define CPUID_VMWARE_FEATURES_LEAF           0x40000010 | 
|---|
| 45 |  | 
|---|
| 46 | #define GETVCPU_INFO_LEGACY_X2APIC           BIT(3) | 
|---|
| 47 | #define GETVCPU_INFO_VCPU_RESERVED           BIT(31) | 
|---|
| 48 |  | 
|---|
| 49 | #define STEALCLOCK_NOT_AVAILABLE (-1) | 
|---|
| 50 | #define STEALCLOCK_DISABLED        0 | 
|---|
| 51 | #define STEALCLOCK_ENABLED         1 | 
|---|
| 52 |  | 
|---|
| 53 | struct vmware_steal_time { | 
|---|
| 54 | union { | 
|---|
| 55 | u64 clock;	/* stolen time counter in units of vtsc */ | 
|---|
| 56 | struct { | 
|---|
| 57 | /* only for little-endian */ | 
|---|
| 58 | u32 clock_low; | 
|---|
| 59 | u32 clock_high; | 
|---|
| 60 | }; | 
|---|
| 61 | }; | 
|---|
| 62 | u64 reserved[7]; | 
|---|
| 63 | }; | 
|---|
| 64 |  | 
|---|
| 65 | static unsigned long vmware_tsc_khz __ro_after_init; | 
|---|
| 66 | static u8 vmware_hypercall_mode     __ro_after_init; | 
|---|
| 67 |  | 
|---|
| 68 | unsigned long vmware_hypercall_slow(unsigned long cmd, | 
|---|
| 69 | unsigned long in1, unsigned long in3, | 
|---|
| 70 | unsigned long in4, unsigned long in5, | 
|---|
| 71 | u32 *out1, u32 *out2, u32 *out3, | 
|---|
| 72 | u32 *out4, u32 *out5) | 
|---|
| 73 | { | 
|---|
| 74 | unsigned long out0, rbx, rcx, rdx, rsi, rdi; | 
|---|
| 75 |  | 
|---|
| 76 | switch (vmware_hypercall_mode) { | 
|---|
| 77 | case CPUID_VMWARE_FEATURES_ECX_VMCALL: | 
|---|
| 78 | asm_inline volatile ( "vmcall" | 
|---|
| 79 | : "=a"(out0), "=b"(rbx), "=c"(rcx), | 
|---|
| 80 | "=d"(rdx), "=S"(rsi), "=D"(rdi) | 
|---|
| 81 | : "a"(VMWARE_HYPERVISOR_MAGIC), | 
|---|
| 82 | "b"(in1), | 
|---|
| 83 | "c"(cmd), | 
|---|
| 84 | "d"(in3), | 
|---|
| 85 | "S"(in4), | 
|---|
| 86 | "D"(in5) | 
|---|
| 87 | : "cc", "memory"); | 
|---|
| 88 | break; | 
|---|
| 89 | case CPUID_VMWARE_FEATURES_ECX_VMMCALL: | 
|---|
| 90 | asm_inline volatile ( "vmmcall" | 
|---|
| 91 | : "=a"(out0), "=b"(rbx), "=c"(rcx), | 
|---|
| 92 | "=d"(rdx), "=S"(rsi), "=D"(rdi) | 
|---|
| 93 | : "a"(VMWARE_HYPERVISOR_MAGIC), | 
|---|
| 94 | "b"(in1), | 
|---|
| 95 | "c"(cmd), | 
|---|
| 96 | "d"(in3), | 
|---|
| 97 | "S"(in4), | 
|---|
| 98 | "D"(in5) | 
|---|
| 99 | : "cc", "memory"); | 
|---|
| 100 | break; | 
|---|
| 101 | default: | 
|---|
| 102 | asm_inline volatile ( "movw %[port], %%dx; inl (%%dx), %%eax" | 
|---|
| 103 | : "=a"(out0), "=b"(rbx), "=c"(rcx), | 
|---|
| 104 | "=d"(rdx), "=S"(rsi), "=D"(rdi) | 
|---|
| 105 | : [port] "i"(VMWARE_HYPERVISOR_PORT), | 
|---|
| 106 | "a"(VMWARE_HYPERVISOR_MAGIC), | 
|---|
| 107 | "b"(in1), | 
|---|
| 108 | "c"(cmd), | 
|---|
| 109 | "d"(in3), | 
|---|
| 110 | "S"(in4), | 
|---|
| 111 | "D"(in5) | 
|---|
| 112 | : "cc", "memory"); | 
|---|
| 113 | break; | 
|---|
| 114 | } | 
|---|
| 115 |  | 
|---|
| 116 | if (out1) | 
|---|
| 117 | *out1 = rbx; | 
|---|
| 118 | if (out2) | 
|---|
| 119 | *out2 = rcx; | 
|---|
| 120 | if (out3) | 
|---|
| 121 | *out3 = rdx; | 
|---|
| 122 | if (out4) | 
|---|
| 123 | *out4 = rsi; | 
|---|
| 124 | if (out5) | 
|---|
| 125 | *out5 = rdi; | 
|---|
| 126 |  | 
|---|
| 127 | return out0; | 
|---|
| 128 | } | 
|---|
| 129 |  | 
|---|
| 130 | static inline int __vmware_platform(void) | 
|---|
| 131 | { | 
|---|
| 132 | u32 eax, ebx, ecx; | 
|---|
| 133 |  | 
|---|
| 134 | eax = vmware_hypercall3(VMWARE_CMD_GETVERSION, in1: 0, out1: &ebx, out2: &ecx); | 
|---|
| 135 | return eax != UINT_MAX && ebx == VMWARE_HYPERVISOR_MAGIC; | 
|---|
| 136 | } | 
|---|
| 137 |  | 
|---|
| 138 | static unsigned long vmware_get_tsc_khz(void) | 
|---|
| 139 | { | 
|---|
| 140 | return vmware_tsc_khz; | 
|---|
| 141 | } | 
|---|
| 142 |  | 
|---|
| 143 | #ifdef CONFIG_PARAVIRT | 
|---|
| 144 | static struct cyc2ns_data vmware_cyc2ns __ro_after_init; | 
|---|
| 145 | static bool vmw_sched_clock __initdata = true; | 
|---|
| 146 | static DEFINE_PER_CPU_DECRYPTED(struct vmware_steal_time, vmw_steal_time) __aligned(64); | 
|---|
| 147 | static bool has_steal_clock; | 
|---|
| 148 | static bool steal_acc __initdata = true; /* steal time accounting */ | 
|---|
| 149 |  | 
|---|
| 150 | static __init int setup_vmw_sched_clock(char *s) | 
|---|
| 151 | { | 
|---|
| 152 | vmw_sched_clock = false; | 
|---|
| 153 | return 0; | 
|---|
| 154 | } | 
|---|
| 155 | early_param( "no-vmw-sched-clock", setup_vmw_sched_clock); | 
|---|
| 156 |  | 
|---|
| 157 | static __init int parse_no_stealacc(char *arg) | 
|---|
| 158 | { | 
|---|
| 159 | steal_acc = false; | 
|---|
| 160 | return 0; | 
|---|
| 161 | } | 
|---|
| 162 | early_param( "no-steal-acc", parse_no_stealacc); | 
|---|
| 163 |  | 
|---|
| 164 | static noinstr u64 vmware_sched_clock(void) | 
|---|
| 165 | { | 
|---|
| 166 | unsigned long long ns; | 
|---|
| 167 |  | 
|---|
| 168 | ns = mul_u64_u32_shr(a: rdtsc(), mul: vmware_cyc2ns.cyc2ns_mul, | 
|---|
| 169 | shift: vmware_cyc2ns.cyc2ns_shift); | 
|---|
| 170 | ns -= vmware_cyc2ns.cyc2ns_offset; | 
|---|
| 171 | return ns; | 
|---|
| 172 | } | 
|---|
| 173 |  | 
|---|
| 174 | static void __init vmware_cyc2ns_setup(void) | 
|---|
| 175 | { | 
|---|
| 176 | struct cyc2ns_data *d = &vmware_cyc2ns; | 
|---|
| 177 | unsigned long long tsc_now = rdtsc(); | 
|---|
| 178 |  | 
|---|
| 179 | clocks_calc_mult_shift(mult: &d->cyc2ns_mul, shift: &d->cyc2ns_shift, | 
|---|
| 180 | from: vmware_tsc_khz, NSEC_PER_MSEC, minsec: 0); | 
|---|
| 181 | d->cyc2ns_offset = mul_u64_u32_shr(a: tsc_now, mul: d->cyc2ns_mul, | 
|---|
| 182 | shift: d->cyc2ns_shift); | 
|---|
| 183 |  | 
|---|
| 184 | pr_info( "using clock offset of %llu ns\n", d->cyc2ns_offset); | 
|---|
| 185 | } | 
|---|
| 186 |  | 
|---|
| 187 | static int vmware_cmd_stealclock(u32 addr_hi, u32 addr_lo) | 
|---|
| 188 | { | 
|---|
| 189 | u32 info; | 
|---|
| 190 |  | 
|---|
| 191 | return vmware_hypercall5(VMWARE_CMD_STEALCLOCK, in1: 0, in3: 0, in4: addr_hi, in5: addr_lo, | 
|---|
| 192 | out2: &info); | 
|---|
| 193 | } | 
|---|
| 194 |  | 
|---|
| 195 | static bool stealclock_enable(phys_addr_t pa) | 
|---|
| 196 | { | 
|---|
| 197 | return vmware_cmd_stealclock(upper_32_bits(pa), | 
|---|
| 198 | lower_32_bits(pa)) == STEALCLOCK_ENABLED; | 
|---|
| 199 | } | 
|---|
| 200 |  | 
|---|
| 201 | static int __stealclock_disable(void) | 
|---|
| 202 | { | 
|---|
| 203 | return vmware_cmd_stealclock(addr_hi: 0, addr_lo: 1); | 
|---|
| 204 | } | 
|---|
| 205 |  | 
|---|
| 206 | static void stealclock_disable(void) | 
|---|
| 207 | { | 
|---|
| 208 | __stealclock_disable(); | 
|---|
| 209 | } | 
|---|
| 210 |  | 
|---|
| 211 | static bool vmware_is_stealclock_available(void) | 
|---|
| 212 | { | 
|---|
| 213 | return __stealclock_disable() != STEALCLOCK_NOT_AVAILABLE; | 
|---|
| 214 | } | 
|---|
| 215 |  | 
|---|
| 216 | /** | 
|---|
| 217 | * vmware_steal_clock() - read the per-cpu steal clock | 
|---|
| 218 | * @cpu:            the cpu number whose steal clock we want to read | 
|---|
| 219 | * | 
|---|
| 220 | * The function reads the steal clock if we are on a 64-bit system, otherwise | 
|---|
| 221 | * reads it in parts, checking that the high part didn't change in the | 
|---|
| 222 | * meantime. | 
|---|
| 223 | * | 
|---|
| 224 | * Return: | 
|---|
| 225 | *      The steal clock reading in ns. | 
|---|
| 226 | */ | 
|---|
| 227 | static u64 vmware_steal_clock(int cpu) | 
|---|
| 228 | { | 
|---|
| 229 | struct vmware_steal_time *steal = &per_cpu(vmw_steal_time, cpu); | 
|---|
| 230 | u64 clock; | 
|---|
| 231 |  | 
|---|
| 232 | if (IS_ENABLED(CONFIG_64BIT)) | 
|---|
| 233 | clock = READ_ONCE(steal->clock); | 
|---|
| 234 | else { | 
|---|
| 235 | u32 initial_high, low, high; | 
|---|
| 236 |  | 
|---|
| 237 | do { | 
|---|
| 238 | initial_high = READ_ONCE(steal->clock_high); | 
|---|
| 239 | /* Do not reorder initial_high and high readings */ | 
|---|
| 240 | virt_rmb(); | 
|---|
| 241 | low = READ_ONCE(steal->clock_low); | 
|---|
| 242 | /* Keep low reading in between */ | 
|---|
| 243 | virt_rmb(); | 
|---|
| 244 | high = READ_ONCE(steal->clock_high); | 
|---|
| 245 | } while (initial_high != high); | 
|---|
| 246 |  | 
|---|
| 247 | clock = ((u64)high << 32) | low; | 
|---|
| 248 | } | 
|---|
| 249 |  | 
|---|
| 250 | return mul_u64_u32_shr(a: clock, mul: vmware_cyc2ns.cyc2ns_mul, | 
|---|
| 251 | shift: vmware_cyc2ns.cyc2ns_shift); | 
|---|
| 252 | } | 
|---|
| 253 |  | 
|---|
| 254 | static void vmware_register_steal_time(void) | 
|---|
| 255 | { | 
|---|
| 256 | int cpu = smp_processor_id(); | 
|---|
| 257 | struct vmware_steal_time *st = &per_cpu(vmw_steal_time, cpu); | 
|---|
| 258 |  | 
|---|
| 259 | if (!has_steal_clock) | 
|---|
| 260 | return; | 
|---|
| 261 |  | 
|---|
| 262 | if (!stealclock_enable(pa: slow_virt_to_phys(address: st))) { | 
|---|
| 263 | has_steal_clock = false; | 
|---|
| 264 | return; | 
|---|
| 265 | } | 
|---|
| 266 |  | 
|---|
| 267 | pr_info( "vmware-stealtime: cpu %d, pa %llx\n", | 
|---|
| 268 | cpu, (unsigned long long) slow_virt_to_phys(st)); | 
|---|
| 269 | } | 
|---|
| 270 |  | 
|---|
| 271 | static void vmware_disable_steal_time(void) | 
|---|
| 272 | { | 
|---|
| 273 | if (!has_steal_clock) | 
|---|
| 274 | return; | 
|---|
| 275 |  | 
|---|
| 276 | stealclock_disable(); | 
|---|
| 277 | } | 
|---|
| 278 |  | 
|---|
| 279 | static void vmware_guest_cpu_init(void) | 
|---|
| 280 | { | 
|---|
| 281 | if (has_steal_clock) | 
|---|
| 282 | vmware_register_steal_time(); | 
|---|
| 283 | } | 
|---|
| 284 |  | 
|---|
| 285 | static void vmware_pv_guest_cpu_reboot(void *unused) | 
|---|
| 286 | { | 
|---|
| 287 | vmware_disable_steal_time(); | 
|---|
| 288 | } | 
|---|
| 289 |  | 
|---|
| 290 | static int vmware_pv_reboot_notify(struct notifier_block *nb, | 
|---|
| 291 | unsigned long code, void *unused) | 
|---|
| 292 | { | 
|---|
| 293 | if (code == SYS_RESTART) | 
|---|
| 294 | on_each_cpu(func: vmware_pv_guest_cpu_reboot, NULL, wait: 1); | 
|---|
| 295 | return NOTIFY_DONE; | 
|---|
| 296 | } | 
|---|
| 297 |  | 
|---|
| 298 | static struct notifier_block vmware_pv_reboot_nb = { | 
|---|
| 299 | .notifier_call = vmware_pv_reboot_notify, | 
|---|
| 300 | }; | 
|---|
| 301 |  | 
|---|
| 302 | #ifdef CONFIG_SMP | 
|---|
| 303 | static void __init vmware_smp_prepare_boot_cpu(void) | 
|---|
| 304 | { | 
|---|
| 305 | vmware_guest_cpu_init(); | 
|---|
| 306 | native_smp_prepare_boot_cpu(); | 
|---|
| 307 | } | 
|---|
| 308 |  | 
|---|
| 309 | static int vmware_cpu_online(unsigned int cpu) | 
|---|
| 310 | { | 
|---|
| 311 | local_irq_disable(); | 
|---|
| 312 | vmware_guest_cpu_init(); | 
|---|
| 313 | local_irq_enable(); | 
|---|
| 314 | return 0; | 
|---|
| 315 | } | 
|---|
| 316 |  | 
|---|
| 317 | static int vmware_cpu_down_prepare(unsigned int cpu) | 
|---|
| 318 | { | 
|---|
| 319 | local_irq_disable(); | 
|---|
| 320 | vmware_disable_steal_time(); | 
|---|
| 321 | local_irq_enable(); | 
|---|
| 322 | return 0; | 
|---|
| 323 | } | 
|---|
| 324 | #endif | 
|---|
| 325 |  | 
|---|
| 326 | static __init int activate_jump_labels(void) | 
|---|
| 327 | { | 
|---|
| 328 | if (has_steal_clock) { | 
|---|
| 329 | static_key_slow_inc(key: ¶virt_steal_enabled); | 
|---|
| 330 | if (steal_acc) | 
|---|
| 331 | static_key_slow_inc(key: ¶virt_steal_rq_enabled); | 
|---|
| 332 | } | 
|---|
| 333 |  | 
|---|
| 334 | return 0; | 
|---|
| 335 | } | 
|---|
| 336 | arch_initcall(activate_jump_labels); | 
|---|
| 337 |  | 
|---|
| 338 | static void __init vmware_paravirt_ops_setup(void) | 
|---|
| 339 | { | 
|---|
| 340 | pv_info.name = "VMware hypervisor"; | 
|---|
| 341 | pv_ops.cpu.io_delay = paravirt_nop; | 
|---|
| 342 |  | 
|---|
| 343 | if (vmware_tsc_khz == 0) | 
|---|
| 344 | return; | 
|---|
| 345 |  | 
|---|
| 346 | vmware_cyc2ns_setup(); | 
|---|
| 347 |  | 
|---|
| 348 | if (vmw_sched_clock) | 
|---|
| 349 | paravirt_set_sched_clock(func: vmware_sched_clock); | 
|---|
| 350 |  | 
|---|
| 351 | if (vmware_is_stealclock_available()) { | 
|---|
| 352 | has_steal_clock = true; | 
|---|
| 353 | static_call_update(pv_steal_clock, vmware_steal_clock); | 
|---|
| 354 |  | 
|---|
| 355 | /* We use reboot notifier only to disable steal clock */ | 
|---|
| 356 | register_reboot_notifier(&vmware_pv_reboot_nb); | 
|---|
| 357 |  | 
|---|
| 358 | #ifdef CONFIG_SMP | 
|---|
| 359 | smp_ops.smp_prepare_boot_cpu = | 
|---|
| 360 | vmware_smp_prepare_boot_cpu; | 
|---|
| 361 | if (cpuhp_setup_state_nocalls(state: CPUHP_AP_ONLINE_DYN, | 
|---|
| 362 | name: "x86/vmware:online", | 
|---|
| 363 | startup: vmware_cpu_online, | 
|---|
| 364 | teardown: vmware_cpu_down_prepare) < 0) | 
|---|
| 365 | pr_err( "vmware_guest: Failed to install cpu hotplug callbacks\n"); | 
|---|
| 366 | #else | 
|---|
| 367 | vmware_guest_cpu_init(); | 
|---|
| 368 | #endif | 
|---|
| 369 | } | 
|---|
| 370 | } | 
|---|
| 371 | #else | 
|---|
| 372 | #define vmware_paravirt_ops_setup() do {} while (0) | 
|---|
| 373 | #endif | 
|---|
| 374 |  | 
|---|
| 375 | /* | 
|---|
| 376 | * VMware hypervisor takes care of exporting a reliable TSC to the guest. | 
|---|
| 377 | * Still, due to timing difference when running on virtual cpus, the TSC can | 
|---|
| 378 | * be marked as unstable in some cases. For example, the TSC sync check at | 
|---|
| 379 | * bootup can fail due to a marginal offset between vcpus' TSCs (though the | 
|---|
| 380 | * TSCs do not drift from each other).  Also, the ACPI PM timer clocksource | 
|---|
| 381 | * is not suitable as a watchdog when running on a hypervisor because the | 
|---|
| 382 | * kernel may miss a wrap of the counter if the vcpu is descheduled for a | 
|---|
| 383 | * long time. To skip these checks at runtime we set these capability bits, | 
|---|
| 384 | * so that the kernel could just trust the hypervisor with providing a | 
|---|
| 385 | * reliable virtual TSC that is suitable for timekeeping. | 
|---|
| 386 | */ | 
|---|
| 387 | static void __init vmware_set_capabilities(void) | 
|---|
| 388 | { | 
|---|
| 389 | setup_force_cpu_cap(X86_FEATURE_CONSTANT_TSC); | 
|---|
| 390 | setup_force_cpu_cap(X86_FEATURE_TSC_RELIABLE); | 
|---|
| 391 | if (vmware_tsc_khz) | 
|---|
| 392 | setup_force_cpu_cap(X86_FEATURE_TSC_KNOWN_FREQ); | 
|---|
| 393 | if (vmware_hypercall_mode == CPUID_VMWARE_FEATURES_ECX_VMCALL) | 
|---|
| 394 | setup_force_cpu_cap(X86_FEATURE_VMCALL); | 
|---|
| 395 | else if (vmware_hypercall_mode == CPUID_VMWARE_FEATURES_ECX_VMMCALL) | 
|---|
| 396 | setup_force_cpu_cap(X86_FEATURE_VMW_VMMCALL); | 
|---|
| 397 | } | 
|---|
| 398 |  | 
|---|
| 399 | static void __init vmware_platform_setup(void) | 
|---|
| 400 | { | 
|---|
| 401 | u32 eax, ebx, ecx; | 
|---|
| 402 | u64 lpj, tsc_khz; | 
|---|
| 403 |  | 
|---|
| 404 | eax = vmware_hypercall3(VMWARE_CMD_GETHZ, UINT_MAX, out1: &ebx, out2: &ecx); | 
|---|
| 405 |  | 
|---|
| 406 | if (ebx != UINT_MAX) { | 
|---|
| 407 | lpj = tsc_khz = eax | (((u64)ebx) << 32); | 
|---|
| 408 | do_div(tsc_khz, 1000); | 
|---|
| 409 | WARN_ON(tsc_khz >> 32); | 
|---|
| 410 | pr_info( "TSC freq read from hypervisor : %lu.%03lu MHz\n", | 
|---|
| 411 | (unsigned long) tsc_khz / 1000, | 
|---|
| 412 | (unsigned long) tsc_khz % 1000); | 
|---|
| 413 |  | 
|---|
| 414 | if (!preset_lpj) { | 
|---|
| 415 | do_div(lpj, HZ); | 
|---|
| 416 | preset_lpj = lpj; | 
|---|
| 417 | } | 
|---|
| 418 |  | 
|---|
| 419 | vmware_tsc_khz = tsc_khz; | 
|---|
| 420 | x86_platform.calibrate_tsc = vmware_get_tsc_khz; | 
|---|
| 421 | x86_platform.calibrate_cpu = vmware_get_tsc_khz; | 
|---|
| 422 |  | 
|---|
| 423 | #ifdef CONFIG_X86_LOCAL_APIC | 
|---|
| 424 | /* Skip lapic calibration since we know the bus frequency. */ | 
|---|
| 425 | lapic_timer_period = ecx / HZ; | 
|---|
| 426 | pr_info( "Host bus clock speed read from hypervisor : %u Hz\n", | 
|---|
| 427 | ecx); | 
|---|
| 428 | #endif | 
|---|
| 429 | } else { | 
|---|
| 430 | pr_warn( "Failed to get TSC freq from the hypervisor\n"); | 
|---|
| 431 | } | 
|---|
| 432 |  | 
|---|
| 433 | if (cc_platform_has(attr: CC_ATTR_GUEST_SEV_SNP) && !efi_enabled(EFI_BOOT)) | 
|---|
| 434 | x86_init.mpparse.find_mptable = mpparse_find_mptable; | 
|---|
| 435 |  | 
|---|
| 436 | vmware_paravirt_ops_setup(); | 
|---|
| 437 |  | 
|---|
| 438 | #ifdef CONFIG_X86_IO_APIC | 
|---|
| 439 | no_timer_check = 1; | 
|---|
| 440 | #endif | 
|---|
| 441 |  | 
|---|
| 442 | vmware_set_capabilities(); | 
|---|
| 443 | } | 
|---|
| 444 |  | 
|---|
| 445 | static u8 __init vmware_select_hypercall(void) | 
|---|
| 446 | { | 
|---|
| 447 | int eax, ebx, ecx, edx; | 
|---|
| 448 |  | 
|---|
| 449 | cpuid(CPUID_VMWARE_FEATURES_LEAF, eax: &eax, ebx: &ebx, ecx: &ecx, edx: &edx); | 
|---|
| 450 | return (ecx & (CPUID_VMWARE_FEATURES_ECX_VMMCALL | | 
|---|
| 451 | CPUID_VMWARE_FEATURES_ECX_VMCALL)); | 
|---|
| 452 | } | 
|---|
| 453 |  | 
|---|
| 454 | /* | 
|---|
| 455 | * While checking the dmi string information, just checking the product | 
|---|
| 456 | * serial key should be enough, as this will always have a VMware | 
|---|
| 457 | * specific string when running under VMware hypervisor. | 
|---|
| 458 | * If !boot_cpu_has(X86_FEATURE_HYPERVISOR), vmware_hypercall_mode | 
|---|
| 459 | * intentionally defaults to 0. | 
|---|
| 460 | */ | 
|---|
| 461 | static u32 __init vmware_platform(void) | 
|---|
| 462 | { | 
|---|
| 463 | if (boot_cpu_has(X86_FEATURE_HYPERVISOR)) { | 
|---|
| 464 | unsigned int eax; | 
|---|
| 465 | unsigned int hyper_vendor_id[3]; | 
|---|
| 466 |  | 
|---|
| 467 | cpuid(CPUID_VMWARE_INFO_LEAF, eax: &eax, ebx: &hyper_vendor_id[0], | 
|---|
| 468 | ecx: &hyper_vendor_id[1], edx: &hyper_vendor_id[2]); | 
|---|
| 469 | if (!memcmp(hyper_vendor_id, "VMwareVMware", 12)) { | 
|---|
| 470 | if (eax >= CPUID_VMWARE_FEATURES_LEAF) | 
|---|
| 471 | vmware_hypercall_mode = | 
|---|
| 472 | vmware_select_hypercall(); | 
|---|
| 473 |  | 
|---|
| 474 | pr_info( "hypercall mode: 0x%02x\n", | 
|---|
| 475 | (unsigned int) vmware_hypercall_mode); | 
|---|
| 476 |  | 
|---|
| 477 | return CPUID_VMWARE_INFO_LEAF; | 
|---|
| 478 | } | 
|---|
| 479 | } else if (dmi_available && dmi_name_in_serial(str: "VMware") && | 
|---|
| 480 | __vmware_platform()) | 
|---|
| 481 | return 1; | 
|---|
| 482 |  | 
|---|
| 483 | return 0; | 
|---|
| 484 | } | 
|---|
| 485 |  | 
|---|
| 486 | /* Checks if hypervisor supports x2apic without VT-D interrupt remapping. */ | 
|---|
| 487 | static bool __init vmware_legacy_x2apic_available(void) | 
|---|
| 488 | { | 
|---|
| 489 | u32 eax; | 
|---|
| 490 |  | 
|---|
| 491 | eax = vmware_hypercall1(VMWARE_CMD_GETVCPU_INFO, in1: 0); | 
|---|
| 492 | return !(eax & GETVCPU_INFO_VCPU_RESERVED) && | 
|---|
| 493 | (eax & GETVCPU_INFO_LEGACY_X2APIC); | 
|---|
| 494 | } | 
|---|
| 495 |  | 
|---|
| 496 | #ifdef CONFIG_INTEL_TDX_GUEST | 
|---|
| 497 | /* | 
|---|
| 498 | * TDCALL[TDG.VP.VMCALL] uses %rax (arg0) and %rcx (arg2). Therefore, | 
|---|
| 499 | * we remap those registers to %r12 and %r13, respectively. | 
|---|
| 500 | */ | 
|---|
| 501 | unsigned long vmware_tdx_hypercall(unsigned long cmd, | 
|---|
| 502 | unsigned long in1, unsigned long in3, | 
|---|
| 503 | unsigned long in4, unsigned long in5, | 
|---|
| 504 | u32 *out1, u32 *out2, u32 *out3, | 
|---|
| 505 | u32 *out4, u32 *out5) | 
|---|
| 506 | { | 
|---|
| 507 | struct tdx_module_args args = {}; | 
|---|
| 508 |  | 
|---|
| 509 | if (!hypervisor_is_type(X86_HYPER_VMWARE)) { | 
|---|
| 510 | pr_warn_once( "Incorrect usage\n"); | 
|---|
| 511 | return ULONG_MAX; | 
|---|
| 512 | } | 
|---|
| 513 |  | 
|---|
| 514 | if (cmd & ~VMWARE_CMD_MASK) { | 
|---|
| 515 | pr_warn_once( "Out of range command %lx\n", cmd); | 
|---|
| 516 | return ULONG_MAX; | 
|---|
| 517 | } | 
|---|
| 518 |  | 
|---|
| 519 | args.rbx = in1; | 
|---|
| 520 | args.rdx = in3; | 
|---|
| 521 | args.rsi = in4; | 
|---|
| 522 | args.rdi = in5; | 
|---|
| 523 | args.r10 = VMWARE_TDX_VENDOR_LEAF; | 
|---|
| 524 | args.r11 = VMWARE_TDX_HCALL_FUNC; | 
|---|
| 525 | args.r12 = VMWARE_HYPERVISOR_MAGIC; | 
|---|
| 526 | args.r13 = cmd; | 
|---|
| 527 | /* CPL */ | 
|---|
| 528 | args.r15 = 0; | 
|---|
| 529 |  | 
|---|
| 530 | __tdx_hypercall(&args); | 
|---|
| 531 |  | 
|---|
| 532 | if (out1) | 
|---|
| 533 | *out1 = args.rbx; | 
|---|
| 534 | if (out2) | 
|---|
| 535 | *out2 = args.r13; | 
|---|
| 536 | if (out3) | 
|---|
| 537 | *out3 = args.rdx; | 
|---|
| 538 | if (out4) | 
|---|
| 539 | *out4 = args.rsi; | 
|---|
| 540 | if (out5) | 
|---|
| 541 | *out5 = args.rdi; | 
|---|
| 542 |  | 
|---|
| 543 | return args.r12; | 
|---|
| 544 | } | 
|---|
| 545 | EXPORT_SYMBOL_GPL(vmware_tdx_hypercall); | 
|---|
| 546 | #endif | 
|---|
| 547 |  | 
|---|
| 548 | #ifdef CONFIG_AMD_MEM_ENCRYPT | 
|---|
| 549 | static void vmware_sev_es_hcall_prepare(struct ghcb *ghcb, | 
|---|
| 550 | struct pt_regs *regs) | 
|---|
| 551 | { | 
|---|
| 552 | /* Copy VMWARE specific Hypercall parameters to the GHCB */ | 
|---|
| 553 | ghcb_set_rip(ghcb, regs->ip); | 
|---|
| 554 | ghcb_set_rbx(ghcb, regs->bx); | 
|---|
| 555 | ghcb_set_rcx(ghcb, regs->cx); | 
|---|
| 556 | ghcb_set_rdx(ghcb, regs->dx); | 
|---|
| 557 | ghcb_set_rsi(ghcb, regs->si); | 
|---|
| 558 | ghcb_set_rdi(ghcb, regs->di); | 
|---|
| 559 | ghcb_set_rbp(ghcb, regs->bp); | 
|---|
| 560 | } | 
|---|
| 561 |  | 
|---|
| 562 | static bool vmware_sev_es_hcall_finish(struct ghcb *ghcb, struct pt_regs *regs) | 
|---|
| 563 | { | 
|---|
| 564 | if (!(ghcb_rbx_is_valid(ghcb) && | 
|---|
| 565 | ghcb_rcx_is_valid(ghcb) && | 
|---|
| 566 | ghcb_rdx_is_valid(ghcb) && | 
|---|
| 567 | ghcb_rsi_is_valid(ghcb) && | 
|---|
| 568 | ghcb_rdi_is_valid(ghcb) && | 
|---|
| 569 | ghcb_rbp_is_valid(ghcb))) | 
|---|
| 570 | return false; | 
|---|
| 571 |  | 
|---|
| 572 | regs->bx = ghcb_get_rbx(ghcb); | 
|---|
| 573 | regs->cx = ghcb_get_rcx(ghcb); | 
|---|
| 574 | regs->dx = ghcb_get_rdx(ghcb); | 
|---|
| 575 | regs->si = ghcb_get_rsi(ghcb); | 
|---|
| 576 | regs->di = ghcb_get_rdi(ghcb); | 
|---|
| 577 | regs->bp = ghcb_get_rbp(ghcb); | 
|---|
| 578 |  | 
|---|
| 579 | return true; | 
|---|
| 580 | } | 
|---|
| 581 | #endif | 
|---|
| 582 |  | 
|---|
| 583 | const __initconst struct hypervisor_x86 x86_hyper_vmware = { | 
|---|
| 584 | .name				= "VMware", | 
|---|
| 585 | .detect				= vmware_platform, | 
|---|
| 586 | .type				= X86_HYPER_VMWARE, | 
|---|
| 587 | .init.init_platform		= vmware_platform_setup, | 
|---|
| 588 | .init.x2apic_available		= vmware_legacy_x2apic_available, | 
|---|
| 589 | #ifdef CONFIG_AMD_MEM_ENCRYPT | 
|---|
| 590 | .runtime.sev_es_hcall_prepare	= vmware_sev_es_hcall_prepare, | 
|---|
| 591 | .runtime.sev_es_hcall_finish	= vmware_sev_es_hcall_finish, | 
|---|
| 592 | #endif | 
|---|
| 593 | }; | 
|---|
| 594 |  | 
|---|