Commit | Line | Data |
---|---|---|
b2441318 | 1 | // SPDX-License-Identifier: GPL-2.0 |
1da177e4 | 2 | |
1da177e4 | 3 | #include <linux/bitops.h> |
186f4360 | 4 | #include <linux/init.h> |
dec7fdc0 | 5 | #include <linux/kernel.h> |
cb5f4c76 | 6 | #include <linux/minmax.h> |
dec7fdc0 AD |
7 | #include <linux/smp.h> |
8 | #include <linux/string.h> | |
b5969494 | 9 | #include <linux/types.h> |
dec7fdc0 AD |
10 | |
11 | #ifdef CONFIG_X86_64 | |
12 | #include <linux/topology.h> | |
13 | #endif | |
1da177e4 | 14 | |
73bdb73f | 15 | #include <asm/bugs.h> |
dec7fdc0 AD |
16 | #include <asm/cpu_device_id.h> |
17 | #include <asm/cpufeature.h> | |
1f442d70 | 18 | #include <asm/cpu.h> |
968e3000 | 19 | #include <asm/cpuid/api.h> |
dec7fdc0 | 20 | #include <asm/hwcap2.h> |
08e237fa | 21 | #include <asm/intel-family.h> |
82ad097b | 22 | #include <asm/microcode.h> |
dec7fdc0 | 23 | #include <asm/msr.h> |
0cd39f46 | 24 | #include <asm/numa.h> |
dec7fdc0 | 25 | #include <asm/resctrl.h> |
9223d0dc | 26 | #include <asm/thermal.h> |
dec7fdc0 | 27 | #include <asm/uaccess.h> |
185f3b9d | 28 | |
1da177e4 LT |
29 | #include "cpu.h" |
30 | ||
1e03bff3 RN |
31 | /* |
32 | * Processors which have self-snooping capability can handle conflicting | |
33 | * memory type across CPUs by snooping its own cache. However, there exists | |
34 | * CPU models in which having conflicting memory types still leads to | |
35 | * unpredictable behavior, machine check errors, or hangs. Clear this | |
36 | * feature to prevent its use on machines with known erratas. | |
37 | */ | |
38 | static void check_memory_type_self_snoop_errata(struct cpuinfo_x86 *c) | |
39 | { | |
6568fc18 TL |
40 | switch (c->x86_vfm) { |
41 | case INTEL_CORE_YONAH: | |
42 | case INTEL_CORE2_MEROM: | |
43 | case INTEL_CORE2_MEROM_L: | |
44 | case INTEL_CORE2_PENRYN: | |
45 | case INTEL_CORE2_DUNNINGTON: | |
46 | case INTEL_NEHALEM: | |
47 | case INTEL_NEHALEM_G: | |
48 | case INTEL_NEHALEM_EP: | |
49 | case INTEL_NEHALEM_EX: | |
50 | case INTEL_WESTMERE: | |
51 | case INTEL_WESTMERE_EP: | |
52 | case INTEL_SANDYBRIDGE: | |
1e03bff3 RN |
53 | setup_clear_cpu_cap(X86_FEATURE_SELFSNOOP); |
54 | } | |
55 | } | |
56 | ||
e16fd002 GA |
57 | static bool ring3mwait_disabled __read_mostly; |
58 | ||
59 | static int __init ring3mwait_disable(char *__unused) | |
60 | { | |
61 | ring3mwait_disabled = true; | |
12441ccd | 62 | return 1; |
e16fd002 GA |
63 | } |
64 | __setup("ring3mwait=disable", ring3mwait_disable); | |
65 | ||
66 | static void probe_xeon_phi_r3mwait(struct cpuinfo_x86 *c) | |
67 | { | |
68 | /* | |
69 | * Ring 3 MONITOR/MWAIT feature cannot be detected without | |
70 | * cpu model and family comparison. | |
71 | */ | |
4d8bb006 | 72 | if (c->x86 != 6) |
e16fd002 | 73 | return; |
6568fc18 TL |
74 | switch (c->x86_vfm) { |
75 | case INTEL_XEON_PHI_KNL: | |
76 | case INTEL_XEON_PHI_KNM: | |
4d8bb006 PL |
77 | break; |
78 | default: | |
79 | return; | |
80 | } | |
e16fd002 | 81 | |
e9ea1e7f | 82 | if (ring3mwait_disabled) |
e16fd002 | 83 | return; |
e16fd002 GA |
84 | |
85 | set_cpu_cap(c, X86_FEATURE_RING3MWAIT); | |
e9ea1e7f KH |
86 | this_cpu_or(msr_misc_features_shadow, |
87 | 1UL << MSR_MISC_FEATURES_ENABLES_RING3MWAIT_BIT); | |
e16fd002 GA |
88 | |
89 | if (c == &boot_cpu_data) | |
90 | ELF_HWCAP2 |= HWCAP2_RING3MWAIT; | |
91 | } | |
92 | ||
a5b29663 DW |
93 | /* |
94 | * Early microcode releases for the Spectre v2 mitigation were broken. | |
95 | * Information taken from; | |
e3b3121f | 96 | * - https://newsroom.intel.com/wp-content/uploads/sites/11/2018/03/microcode-update-guidance.pdf |
a5b29663 DW |
97 | * - https://kb.vmware.com/s/article/52345 |
98 | * - Microcode revisions observed in the wild | |
99 | * - Release note from 20180108 microcode release | |
100 | */ | |
101 | struct sku_microcode { | |
6568fc18 | 102 | u32 vfm; |
a5b29663 DW |
103 | u8 stepping; |
104 | u32 microcode; | |
105 | }; | |
106 | static const struct sku_microcode spectre_bad_microcodes[] = { | |
6568fc18 TL |
107 | { INTEL_KABYLAKE, 0x0B, 0x80 }, |
108 | { INTEL_KABYLAKE, 0x0A, 0x80 }, | |
109 | { INTEL_KABYLAKE, 0x09, 0x80 }, | |
110 | { INTEL_KABYLAKE_L, 0x0A, 0x80 }, | |
111 | { INTEL_KABYLAKE_L, 0x09, 0x80 }, | |
112 | { INTEL_SKYLAKE_X, 0x03, 0x0100013e }, | |
113 | { INTEL_SKYLAKE_X, 0x04, 0x0200003c }, | |
114 | { INTEL_BROADWELL, 0x04, 0x28 }, | |
115 | { INTEL_BROADWELL_G, 0x01, 0x1b }, | |
116 | { INTEL_BROADWELL_D, 0x02, 0x14 }, | |
117 | { INTEL_BROADWELL_D, 0x03, 0x07000011 }, | |
118 | { INTEL_BROADWELL_X, 0x01, 0x0b000025 }, | |
119 | { INTEL_HASWELL_L, 0x01, 0x21 }, | |
120 | { INTEL_HASWELL_G, 0x01, 0x18 }, | |
121 | { INTEL_HASWELL, 0x03, 0x23 }, | |
122 | { INTEL_HASWELL_X, 0x02, 0x3b }, | |
123 | { INTEL_HASWELL_X, 0x04, 0x10 }, | |
124 | { INTEL_IVYBRIDGE_X, 0x04, 0x42a }, | |
a5b29663 | 125 | /* Observed in the wild */ |
6568fc18 TL |
126 | { INTEL_SANDYBRIDGE_X, 0x06, 0x61b }, |
127 | { INTEL_SANDYBRIDGE_X, 0x07, 0x712 }, | |
a5b29663 DW |
128 | }; |
129 | ||
130 | static bool bad_spectre_microcode(struct cpuinfo_x86 *c) | |
131 | { | |
132 | int i; | |
133 | ||
36268223 KRW |
134 | /* |
135 | * We know that the hypervisor lie to us on the microcode version so | |
136 | * we may as well hope that it is running the correct version. | |
137 | */ | |
138 | if (cpu_has(c, X86_FEATURE_HYPERVISOR)) | |
139 | return false; | |
140 | ||
a5b29663 | 141 | for (i = 0; i < ARRAY_SIZE(spectre_bad_microcodes); i++) { |
6568fc18 | 142 | if (c->x86_vfm == spectre_bad_microcodes[i].vfm && |
b399151c | 143 | c->x86_stepping == spectre_bad_microcodes[i].stepping) |
a5b29663 DW |
144 | return (c->microcode <= spectre_bad_microcodes[i].microcode); |
145 | } | |
146 | return false; | |
147 | } | |
148 | ||
6890cb1a PB |
149 | #define MSR_IA32_TME_ACTIVATE 0x982 |
150 | ||
151 | /* Helpers to access TME_ACTIVATE MSR */ | |
152 | #define TME_ACTIVATE_LOCKED(x) (x & 0x1) | |
153 | #define TME_ACTIVATE_ENABLED(x) (x & 0x2) | |
154 | ||
6890cb1a PB |
155 | #define TME_ACTIVATE_KEYID_BITS(x) ((x >> 32) & 0xf) /* Bits 35:32 */ |
156 | ||
6890cb1a PB |
157 | static void detect_tme_early(struct cpuinfo_x86 *c) |
158 | { | |
98b83cf0 AS |
159 | u64 tme_activate; |
160 | int keyid_bits; | |
6890cb1a | 161 | |
c435e608 | 162 | rdmsrq(MSR_IA32_TME_ACTIVATE, tme_activate); |
6890cb1a | 163 | |
6890cb1a PB |
164 | if (!TME_ACTIVATE_LOCKED(tme_activate) || !TME_ACTIVATE_ENABLED(tme_activate)) { |
165 | pr_info_once("x86/tme: not enabled by BIOS\n"); | |
cd2236c2 | 166 | clear_cpu_cap(c, X86_FEATURE_TME); |
6890cb1a PB |
167 | return; |
168 | } | |
98b83cf0 | 169 | pr_info_once("x86/tme: enabled by BIOS\n"); |
6890cb1a | 170 | keyid_bits = TME_ACTIVATE_KEYID_BITS(tme_activate); |
98b83cf0 AS |
171 | if (!keyid_bits) |
172 | return; | |
6890cb1a PB |
173 | |
174 | /* | |
98b83cf0 AS |
175 | * KeyID bits are set by BIOS and can be present regardless |
176 | * of whether the kernel is using them. They effectively lower | |
177 | * the number of physical address bits. | |
178 | * | |
179 | * Update cpuinfo_x86::x86_phys_bits accordingly. | |
6890cb1a PB |
180 | */ |
181 | c->x86_phys_bits -= keyid_bits; | |
98b83cf0 AS |
182 | pr_info_once("x86/mktme: BIOS enabled: x86_phys_bits reduced by %d\n", |
183 | keyid_bits); | |
6890cb1a PB |
184 | } |
185 | ||
0c2f6d04 TG |
186 | void intel_unlock_cpuid_leafs(struct cpuinfo_x86 *c) |
187 | { | |
188 | if (boot_cpu_data.x86_vendor != X86_VENDOR_INTEL) | |
189 | return; | |
190 | ||
eaa472f7 | 191 | if (c->x86_vfm < INTEL_PENTIUM_M_DOTHAN) |
0c2f6d04 TG |
192 | return; |
193 | ||
194 | /* | |
195 | * The BIOS can have limited CPUID to leaf 2, which breaks feature | |
196 | * enumeration. Unlock it and update the maximum leaf info. | |
197 | */ | |
198 | if (msr_clear_bit(MSR_IA32_MISC_ENABLE, MSR_IA32_MISC_ENABLE_LIMIT_CPUID_BIT) > 0) | |
199 | c->cpuid_level = cpuid_eax(0); | |
200 | } | |
201 | ||
148f9bb8 | 202 | static void early_init_intel(struct cpuinfo_x86 *c) |
1da177e4 | 203 | { |
161ec53c FY |
204 | u64 misc_enable; |
205 | ||
4167709b BP |
206 | if (c->x86 >= 6 && !cpu_has(c, X86_FEATURE_IA64)) |
207 | c->microcode = intel_get_microcode_revision(); | |
506ed6b5 | 208 | |
2961298e | 209 | /* Now if any of them are set, check the blacklist and clear the lot */ |
7fcae111 DW |
210 | if ((cpu_has(c, X86_FEATURE_SPEC_CTRL) || |
211 | cpu_has(c, X86_FEATURE_INTEL_STIBP) || | |
212 | cpu_has(c, X86_FEATURE_IBRS) || cpu_has(c, X86_FEATURE_IBPB) || | |
2961298e DW |
213 | cpu_has(c, X86_FEATURE_STIBP)) && bad_spectre_microcode(c)) { |
214 | pr_warn("Intel Spectre v2 broken microcode detected; disabling Speculation Control\n"); | |
7fcae111 DW |
215 | setup_clear_cpu_cap(X86_FEATURE_IBRS); |
216 | setup_clear_cpu_cap(X86_FEATURE_IBPB); | |
217 | setup_clear_cpu_cap(X86_FEATURE_STIBP); | |
218 | setup_clear_cpu_cap(X86_FEATURE_SPEC_CTRL); | |
7eb8956a | 219 | setup_clear_cpu_cap(X86_FEATURE_MSR_SPEC_CTRL); |
7fcae111 | 220 | setup_clear_cpu_cap(X86_FEATURE_INTEL_STIBP); |
9f65fb29 | 221 | setup_clear_cpu_cap(X86_FEATURE_SSBD); |
52817587 | 222 | setup_clear_cpu_cap(X86_FEATURE_SPEC_CTRL_SSBD); |
a5b29663 DW |
223 | } |
224 | ||
7a0fc404 PA |
225 | /* |
226 | * Atom erratum AAE44/AAF40/AAG38/AAH41: | |
227 | * | |
228 | * A race condition between speculative fetches and invalidating | |
229 | * a large page. This is worked around in microcode, but we | |
230 | * need the microcode to have already been loaded... so if it is | |
231 | * not, recommend a BIOS update and disable large pages. | |
232 | */ | |
6568fc18 | 233 | if (c->x86_vfm == INTEL_ATOM_BONNELL && c->x86_stepping <= 2 && |
30963c0a | 234 | c->microcode < 0x20e) { |
1b74dde7 | 235 | pr_warn("Atom PSE erratum detected, BIOS microcode update recommended\n"); |
30963c0a | 236 | clear_cpu_cap(c, X86_FEATURE_PSE); |
7a0fc404 PA |
237 | } |
238 | ||
185f3b9d YL |
239 | #ifdef CONFIG_X86_64 |
240 | set_cpu_cap(c, X86_FEATURE_SYSENTER32); | |
241 | #else | |
242 | /* Netburst reports 64 bytes clflush size, but does IO in 128 bytes */ | |
243 | if (c->x86 == 15 && c->x86_cache_alignment == 64) | |
244 | c->x86_cache_alignment = 128; | |
245 | #endif | |
40fb1715 | 246 | |
13c6c532 | 247 | /* CPUID workaround for 0F33/0F34 CPU */ |
fc866f24 SM |
248 | if (c->x86_vfm == INTEL_P4_PRESCOTT && |
249 | (c->x86_stepping == 0x3 || c->x86_stepping == 0x4)) | |
13c6c532 JB |
250 | c->x86_phys_bits = 36; |
251 | ||
40fb1715 VP |
252 | /* |
253 | * c->x86_power is 8000_0007 edx. Bit 8 is TSC runs at constant rate | |
83ce4009 IM |
254 | * with P/T states and does not stop in deep C-states. |
255 | * | |
256 | * It is also reliable across cores and sockets. (but not across | |
257 | * cabinets - we turn it off in that case explicitly.) | |
fadb6f56 SM |
258 | * |
259 | * Use a model-specific check for some older CPUs that have invariant | |
260 | * TSC but may not report it architecturally via 8000_0007. | |
40fb1715 VP |
261 | */ |
262 | if (c->x86_power & (1 << 8)) { | |
263 | set_cpu_cap(c, X86_FEATURE_CONSTANT_TSC); | |
264 | set_cpu_cap(c, X86_FEATURE_NONSTOP_TSC); | |
fadb6f56 SM |
265 | } else if ((c->x86_vfm >= INTEL_P4_PRESCOTT && c->x86_vfm <= INTEL_P4_WILLAMETTE) || |
266 | (c->x86_vfm >= INTEL_CORE_YONAH && c->x86_vfm <= INTEL_IVYBRIDGE)) { | |
267 | set_cpu_cap(c, X86_FEATURE_CONSTANT_TSC); | |
40fb1715 VP |
268 | } |
269 | ||
c54fdbb2 | 270 | /* Penwell and Cloverview have the TSC which doesn't sleep on S3 */ |
34b3fc55 AC |
271 | switch (c->x86_vfm) { |
272 | case INTEL_ATOM_SALTWELL_MID: | |
273 | case INTEL_ATOM_SALTWELL_TABLET: | |
274 | case INTEL_ATOM_SILVERMONT_MID: | |
275 | case INTEL_ATOM_AIRMONT_NP: | |
276 | set_cpu_cap(c, X86_FEATURE_NONSTOP_TSC_S3); | |
277 | break; | |
c54fdbb2 FT |
278 | } |
279 | ||
75a04811 | 280 | /* |
fd82221a DH |
281 | * PAT is broken on early family 6 CPUs, the last of which |
282 | * is "Yonah" where the erratum is named "AN7": | |
75a04811 | 283 | * |
fd82221a DH |
284 | * Page with PAT (Page Attribute Table) Set to USWC |
285 | * (Uncacheable Speculative Write Combine) While | |
286 | * Associated MTRR (Memory Type Range Register) Is UC | |
287 | * (Uncacheable) May Consolidate to UC | |
288 | * | |
289 | * Disable PAT and fall back to MTRR on these CPUs. | |
75a04811 | 290 | */ |
fd82221a DH |
291 | if (c->x86_vfm >= INTEL_PENTIUM_PRO && |
292 | c->x86_vfm <= INTEL_CORE_YONAH) | |
75a04811 | 293 | clear_cpu_cap(c, X86_FEATURE_PAT); |
f8561296 | 294 | |
161ec53c | 295 | /* |
15b7ddcf SM |
296 | * Modern CPUs are generally expected to have a sane fast string |
297 | * implementation. However, BIOSes typically have a knob to tweak | |
298 | * the architectural MISC_ENABLE.FAST_STRING enable bit. | |
299 | * | |
300 | * Adhere to the preference and program the Linux-defined fast | |
301 | * string flag and enhanced fast string capabilities accordingly. | |
161ec53c | 302 | */ |
15b7ddcf | 303 | if (c->x86_vfm >= INTEL_PENTIUM_M_DOTHAN) { |
c435e608 | 304 | rdmsrq(MSR_IA32_MISC_ENABLE, misc_enable); |
15b7ddcf SM |
305 | if (misc_enable & MSR_IA32_MISC_ENABLE_FAST_STRING) { |
306 | /* X86_FEATURE_ERMS is set based on CPUID */ | |
307 | set_cpu_cap(c, X86_FEATURE_REP_GOOD); | |
308 | } else { | |
1b74dde7 | 309 | pr_info("Disabled fast string operations\n"); |
161ec53c FY |
310 | setup_clear_cpu_cap(X86_FEATURE_REP_GOOD); |
311 | setup_clear_cpu_cap(X86_FEATURE_ERMS); | |
312 | } | |
313 | } | |
ee1b5b16 BD |
314 | |
315 | /* | |
316 | * Intel Quark Core DevMan_001.pdf section 6.4.11 | |
317 | * "The operating system also is required to invalidate (i.e., flush) | |
318 | * the TLB when any changes are made to any of the page table entries. | |
319 | * The operating system must reload CR3 to cause the TLB to be flushed" | |
320 | * | |
c109bf95 | 321 | * As a result, boot_cpu_has(X86_FEATURE_PGE) in arch/x86/include/asm/tlbflush.h |
163b0991 | 322 | * should be false so that __flush_tlb_all() causes CR3 instead of CR4.PGE |
c109bf95 | 323 | * to be modified. |
ee1b5b16 | 324 | */ |
6568fc18 | 325 | if (c->x86_vfm == INTEL_QUARK_X1000) { |
ee1b5b16 BD |
326 | pr_info("Disabling PGE capability bit\n"); |
327 | setup_clear_cpu_cap(X86_FEATURE_PGE); | |
328 | } | |
1f12e32f | 329 | |
1e03bff3 | 330 | check_memory_type_self_snoop_errata(c); |
1910ad56 | 331 | |
6890cb1a PB |
332 | /* |
333 | * Adjust the number of physical bits early because it affects the | |
334 | * valid bits of the MTRR mask registers. | |
335 | */ | |
336 | if (cpu_has(c, X86_FEATURE_TME)) | |
337 | detect_tme_early(c); | |
1da177e4 LT |
338 | } |
339 | ||
923f3a2b RC |
340 | static void bsp_init_intel(struct cpuinfo_x86 *c) |
341 | { | |
342 | resctrl_cpu_detect(c); | |
343 | } | |
344 | ||
185f3b9d | 345 | #ifdef CONFIG_X86_32 |
1da177e4 LT |
346 | /* |
347 | * Early probe support logic for ppro memory erratum #50 | |
348 | * | |
349 | * This is called before we do cpu ident work | |
350 | */ | |
65eb6b43 | 351 | |
148f9bb8 | 352 | int ppro_with_ram_bug(void) |
1da177e4 LT |
353 | { |
354 | /* Uses data from early_cpu_detect now */ | |
eaa472f7 | 355 | if (boot_cpu_data.x86_vfm == INTEL_PENTIUM_PRO && |
b399151c | 356 | boot_cpu_data.x86_stepping < 8) { |
1b74dde7 | 357 | pr_info("Pentium Pro with Errata#50 detected. Taking evasive action.\n"); |
1da177e4 LT |
358 | return 1; |
359 | } | |
360 | return 0; | |
361 | } | |
65eb6b43 | 362 | |
148f9bb8 | 363 | static void intel_smp_check(struct cpuinfo_x86 *c) |
1f442d70 | 364 | { |
1f442d70 | 365 | /* calling is from identify_secondary_cpu() ? */ |
f6e9456c | 366 | if (!c->cpu_index) |
1f442d70 YL |
367 | return; |
368 | ||
369 | /* | |
370 | * Mask B, Pentium, but not Pentium MMX | |
371 | */ | |
eb1ac333 SM |
372 | if (c->x86_vfm >= INTEL_FAM5_START && c->x86_vfm < INTEL_PENTIUM_MMX && |
373 | c->x86_stepping >= 1 && c->x86_stepping <= 4) { | |
1f442d70 YL |
374 | /* |
375 | * Remember we have B step Pentia with bugs | |
376 | */ | |
377 | WARN_ONCE(1, "WARNING: SMP operation may be unreliable" | |
378 | "with B stepping processors.\n"); | |
379 | } | |
1f442d70 YL |
380 | } |
381 | ||
69f2366c CB |
382 | static int forcepae; |
383 | static int __init forcepae_setup(char *__unused) | |
384 | { | |
385 | forcepae = 1; | |
386 | return 1; | |
387 | } | |
388 | __setup("forcepae", forcepae_setup); | |
389 | ||
148f9bb8 | 390 | static void intel_workarounds(struct cpuinfo_x86 *c) |
1da177e4 | 391 | { |
4052704d YL |
392 | #ifdef CONFIG_X86_F00F_BUG |
393 | /* | |
d4e1a0af | 394 | * All models of Pentium and Pentium with MMX technology CPUs |
8bdbd962 | 395 | * have the F0 0F bug, which lets nonprivileged users lock up the |
4eefbe79 | 396 | * system. Announce that the fault handler will be checking for it. |
d4e1a0af | 397 | * The Quark is also family 5, but does not have the same bug. |
4052704d | 398 | */ |
e2604b49 | 399 | clear_cpu_bug(c, X86_BUG_F00F); |
eb1ac333 | 400 | if (c->x86_vfm >= INTEL_FAM5_START && c->x86_vfm < INTEL_QUARK_X1000) { |
4052704d YL |
401 | static int f00f_workaround_enabled; |
402 | ||
e2604b49 | 403 | set_cpu_bug(c, X86_BUG_F00F); |
4052704d | 404 | if (!f00f_workaround_enabled) { |
1b74dde7 | 405 | pr_notice("Intel Pentium with F0 0F bug - workaround enabled.\n"); |
4052704d YL |
406 | f00f_workaround_enabled = 1; |
407 | } | |
408 | } | |
409 | #endif | |
410 | ||
411 | /* | |
412 | * SEP CPUID bug: Pentium Pro reports SEP but doesn't have it until | |
413 | * model 3 mask 3 | |
414 | */ | |
eaa472f7 SM |
415 | if ((c->x86_vfm == INTEL_PENTIUM_II_KLAMATH && c->x86_stepping < 3) || |
416 | c->x86_vfm < INTEL_PENTIUM_II_KLAMATH) | |
4052704d YL |
417 | clear_cpu_cap(c, X86_FEATURE_SEP); |
418 | ||
69f2366c CB |
419 | /* |
420 | * PAE CPUID issue: many Pentium M report no PAE but may have a | |
421 | * functionally usable PAE implementation. | |
422 | * Forcefully enable PAE if kernel parameter "forcepae" is present. | |
423 | */ | |
424 | if (forcepae) { | |
1b74dde7 | 425 | pr_warn("PAE forced!\n"); |
69f2366c CB |
426 | set_cpu_cap(c, X86_FEATURE_PAE); |
427 | add_taint(TAINT_CPU_OUT_OF_SPEC, LOCKDEP_NOW_UNRELIABLE); | |
428 | } | |
429 | ||
4052704d | 430 | /* |
f0133acc | 431 | * P4 Xeon erratum 037 workaround. |
4052704d YL |
432 | * Hardware prefetcher may cause stale data to be loaded into the cache. |
433 | */ | |
fc866f24 | 434 | if (c->x86_vfm == INTEL_P4_WILLAMETTE && c->x86_stepping == 1) { |
0b131be8 | 435 | if (msr_set_bit(MSR_IA32_MISC_ENABLE, |
f0133acc | 436 | MSR_IA32_MISC_ENABLE_PREFETCH_DISABLE_BIT) > 0) { |
c0a639ad | 437 | pr_info("CPU: C0 stepping P4 Xeon detected.\n"); |
f0133acc | 438 | pr_info("CPU: Disabling hardware prefetching (Erratum 037)\n"); |
1da177e4 LT |
439 | } |
440 | } | |
1da177e4 | 441 | |
4052704d YL |
442 | /* |
443 | * See if we have a good local APIC by checking for buggy Pentia, | |
444 | * i.e. all B steppings and the C2 stepping of P54C when using their | |
445 | * integrated APIC (see 11AP erratum in "Pentium Processor | |
446 | * Specification Update"). | |
447 | */ | |
eb1ac333 | 448 | if (boot_cpu_has(X86_FEATURE_APIC) && c->x86_vfm == INTEL_PENTIUM_75 && |
b399151c | 449 | (c->x86_stepping < 0x6 || c->x86_stepping == 0xb)) |
9b13a93d | 450 | set_cpu_bug(c, X86_BUG_11AP); |
185f3b9d | 451 | |
4052704d | 452 | #ifdef CONFIG_X86_INTEL_USERCOPY |
185f3b9d | 453 | /* |
7e67f361 SM |
454 | * MOVSL bulk memory moves can be slow when source and dest are not |
455 | * both 8-byte aligned. PII/PIII only like MOVSL with 8-byte alignment. | |
456 | * | |
457 | * Set the preferred alignment for Pentium Pro and newer processors, as | |
458 | * it has only been tested on these. | |
185f3b9d | 459 | */ |
7e67f361 | 460 | if (c->x86_vfm >= INTEL_PENTIUM_PRO) |
4052704d | 461 | movsl_mask.mask = 7; |
185f3b9d | 462 | #endif |
4052704d | 463 | |
1f442d70 | 464 | intel_smp_check(c); |
4052704d YL |
465 | } |
466 | #else | |
148f9bb8 | 467 | static void intel_workarounds(struct cpuinfo_x86 *c) |
4052704d YL |
468 | { |
469 | } | |
185f3b9d YL |
470 | #endif |
471 | ||
148f9bb8 | 472 | static void srat_detect_node(struct cpuinfo_x86 *c) |
185f3b9d | 473 | { |
645a7919 | 474 | #ifdef CONFIG_NUMA |
185f3b9d YL |
475 | unsigned node; |
476 | int cpu = smp_processor_id(); | |
185f3b9d YL |
477 | |
478 | /* Don't do the funky fallback heuristics the AMD version employs | |
479 | for now. */ | |
bbc9e2f4 | 480 | node = numa_cpu_node(cpu); |
50f2d7f6 | 481 | if (node == NUMA_NO_NODE || !node_online(node)) { |
d9c2d5ac YL |
482 | /* reuse the value from init_cpu_to_node() */ |
483 | node = cpu_to_node(cpu); | |
484 | } | |
185f3b9d | 485 | numa_set_node(cpu, node); |
185f3b9d YL |
486 | #endif |
487 | } | |
488 | ||
90218ac7 KH |
489 | static void init_cpuid_fault(struct cpuinfo_x86 *c) |
490 | { | |
491 | u64 msr; | |
492 | ||
6fe22aba | 493 | if (!rdmsrq_safe(MSR_PLATFORM_INFO, &msr)) { |
90218ac7 KH |
494 | if (msr & MSR_PLATFORM_INFO_CPUID_FAULT) |
495 | set_cpu_cap(c, X86_FEATURE_CPUID_FAULT); | |
496 | } | |
497 | } | |
498 | ||
499 | static void init_intel_misc_features(struct cpuinfo_x86 *c) | |
500 | { | |
501 | u64 msr; | |
502 | ||
6fe22aba | 503 | if (rdmsrq_safe(MSR_MISC_FEATURES_ENABLES, &msr)) |
90218ac7 KH |
504 | return; |
505 | ||
e9ea1e7f KH |
506 | /* Clear all MISC features */ |
507 | this_cpu_write(msr_misc_features_shadow, 0); | |
508 | ||
509 | /* Check features and update capabilities and shadow control bits */ | |
90218ac7 KH |
510 | init_cpuid_fault(c); |
511 | probe_xeon_phi_r3mwait(c); | |
e9ea1e7f KH |
512 | |
513 | msr = this_cpu_read(msr_misc_features_shadow); | |
78255eb2 | 514 | wrmsrq(MSR_MISC_FEATURES_ENABLES, msr); |
90218ac7 KH |
515 | } |
516 | ||
968e9bc4 EB |
517 | /* |
518 | * This is a list of Intel CPUs that are known to suffer from downclocking when | |
519 | * ZMM registers (512-bit vectors) are used. On these CPUs, when the kernel | |
520 | * executes SIMD-optimized code such as cryptography functions or CRCs, it | |
521 | * should prefer 256-bit (YMM) code to 512-bit (ZMM) code. | |
522 | */ | |
523 | static const struct x86_cpu_id zmm_exclusion_list[] = { | |
524 | X86_MATCH_VFM(INTEL_SKYLAKE_X, 0), | |
525 | X86_MATCH_VFM(INTEL_ICELAKE_X, 0), | |
526 | X86_MATCH_VFM(INTEL_ICELAKE_D, 0), | |
527 | X86_MATCH_VFM(INTEL_ICELAKE, 0), | |
528 | X86_MATCH_VFM(INTEL_ICELAKE_L, 0), | |
529 | X86_MATCH_VFM(INTEL_ICELAKE_NNPI, 0), | |
530 | X86_MATCH_VFM(INTEL_TIGERLAKE_L, 0), | |
531 | X86_MATCH_VFM(INTEL_TIGERLAKE, 0), | |
532 | /* Allow Rocket Lake and later, and Sapphire Rapids and later. */ | |
533 | {}, | |
534 | }; | |
535 | ||
148f9bb8 | 536 | static void init_intel(struct cpuinfo_x86 *c) |
1da177e4 | 537 | { |
2b16a235 AK |
538 | early_init_intel(c); |
539 | ||
4052704d | 540 | intel_workarounds(c); |
1da177e4 | 541 | |
807e9bc8 | 542 | init_intel_cacheinfo(c); |
aece118e | 543 | |
65eb6b43 | 544 | if (c->cpuid_level > 9) { |
0080e667 VP |
545 | unsigned eax = cpuid_eax(10); |
546 | /* Check for version and the number of counters */ | |
547 | if ((eax & 0xff) && (((eax>>8) & 0xff) > 1)) | |
d0e95ebd | 548 | set_cpu_cap(c, X86_FEATURE_ARCH_PERFMON); |
0080e667 | 549 | } |
1da177e4 | 550 | |
054efb64 | 551 | if (cpu_has(c, X86_FEATURE_XMM2)) |
4052704d | 552 | set_cpu_cap(c, X86_FEATURE_LFENCE_RDTSC); |
362f924b BP |
553 | |
554 | if (boot_cpu_has(X86_FEATURE_DS)) { | |
807e9bc8 DW |
555 | unsigned int l1, l2; |
556 | ||
4052704d | 557 | rdmsr(MSR_IA32_MISC_ENABLE, l1, l2); |
3f2adf00 | 558 | if (!(l1 & MSR_IA32_MISC_ENABLE_BTS_UNAVAIL)) |
4052704d | 559 | set_cpu_cap(c, X86_FEATURE_BTS); |
3f2adf00 | 560 | if (!(l1 & MSR_IA32_MISC_ENABLE_PEBS_UNAVAIL)) |
4052704d | 561 | set_cpu_cap(c, X86_FEATURE_PEBS); |
4052704d | 562 | } |
1da177e4 | 563 | |
6568fc18 TL |
564 | if (boot_cpu_has(X86_FEATURE_CLFLUSH) && |
565 | (c->x86_vfm == INTEL_CORE2_DUNNINGTON || | |
566 | c->x86_vfm == INTEL_NEHALEM_EX || | |
567 | c->x86_vfm == INTEL_WESTMERE_EX)) | |
9b13a93d | 568 | set_cpu_bug(c, X86_BUG_CLFLUSH_MONITOR); |
e736ad54 | 569 | |
c9a4b554 LB |
570 | if (boot_cpu_has(X86_FEATURE_MWAIT) && |
571 | (c->x86_vfm == INTEL_ATOM_GOLDMONT || | |
572 | c->x86_vfm == INTEL_LUNARLAKE_M)) | |
08e237fa PZ |
573 | set_cpu_bug(c, X86_BUG_MONITOR); |
574 | ||
4052704d YL |
575 | #ifdef CONFIG_X86_64 |
576 | if (c->x86 == 15) | |
577 | c->x86_cache_alignment = c->x86_clflush_size * 2; | |
4052704d | 578 | #else |
65eb6b43 PC |
579 | /* |
580 | * Names for the Pentium II/Celeron processors | |
581 | * detectable only by also checking the cache size. | |
582 | * Dixon is NOT a Celeron. | |
583 | */ | |
1da177e4 | 584 | if (c->x86 == 6) { |
807e9bc8 | 585 | unsigned int l2 = c->x86_cache_size; |
4052704d YL |
586 | char *p = NULL; |
587 | ||
1da177e4 LT |
588 | switch (c->x86_model) { |
589 | case 5: | |
865be7a8 OZ |
590 | if (l2 == 0) |
591 | p = "Celeron (Covington)"; | |
592 | else if (l2 == 256) | |
593 | p = "Mobile Pentium II (Dixon)"; | |
1da177e4 | 594 | break; |
65eb6b43 | 595 | |
1da177e4 LT |
596 | case 6: |
597 | if (l2 == 128) | |
598 | p = "Celeron (Mendocino)"; | |
b399151c | 599 | else if (c->x86_stepping == 0 || c->x86_stepping == 5) |
1da177e4 LT |
600 | p = "Celeron-A"; |
601 | break; | |
65eb6b43 | 602 | |
1da177e4 LT |
603 | case 8: |
604 | if (l2 == 128) | |
605 | p = "Celeron (Coppermine)"; | |
606 | break; | |
607 | } | |
1da177e4 | 608 | |
4052704d YL |
609 | if (p) |
610 | strcpy(c->x86_model_id, p); | |
1da177e4 | 611 | } |
f4166c54 | 612 | #endif |
185f3b9d | 613 | |
968e9bc4 EB |
614 | if (x86_match_cpu(zmm_exclusion_list)) |
615 | set_cpu_cap(c, X86_FEATURE_PREFER_YMM); | |
616 | ||
185f3b9d | 617 | /* Work around errata */ |
2759c328 | 618 | srat_detect_node(c); |
e38e05a8 | 619 | |
1db2a6e1 SC |
620 | init_ia32_feat_ctl(c); |
621 | ||
90218ac7 | 622 | init_intel_misc_features(c); |
95c5824f | 623 | |
6650cdd9 | 624 | split_lock_init(); |
9223d0dc BP |
625 | |
626 | intel_init_thermal(c); | |
42ed458a | 627 | } |
1da177e4 | 628 | |
185f3b9d | 629 | #ifdef CONFIG_X86_32 |
148f9bb8 | 630 | static unsigned int intel_size_cache(struct cpuinfo_x86 *c, unsigned int size) |
1da177e4 | 631 | { |
65eb6b43 PC |
632 | /* |
633 | * Intel PIII Tualatin. This comes in two flavours. | |
1da177e4 LT |
634 | * One has 256kb of cache, the other 512. We have no way |
635 | * to determine which, so we use a boottime override | |
636 | * for the 512kb model, and assume 256 otherwise. | |
637 | */ | |
eaa472f7 | 638 | if (c->x86_vfm == INTEL_PENTIUM_III_TUALATIN && size == 0) |
1da177e4 | 639 | size = 256; |
aece118e BD |
640 | |
641 | /* | |
642 | * Intel Quark SoC X1000 contains a 4-way set associative | |
643 | * 16K cache with a 16 byte cache line and 256 lines per tag | |
644 | */ | |
eb1ac333 | 645 | if (c->x86_vfm == INTEL_QUARK_X1000) |
aece118e | 646 | size = 16; |
1da177e4 LT |
647 | return size; |
648 | } | |
185f3b9d | 649 | #endif |
1da177e4 | 650 | |
119deb95 | 651 | static void intel_tlb_lookup(const struct leaf_0x2_table *desc) |
e0ba94f1 | 652 | { |
119deb95 | 653 | short entries = desc->entries; |
e0ba94f1 | 654 | |
119deb95 | 655 | switch (desc->t_type) { |
e0ba94f1 | 656 | case STLB_4K: |
8b7e54b5 AD |
657 | tlb_lli_4k = max(tlb_lli_4k, entries); |
658 | tlb_lld_4k = max(tlb_lld_4k, entries); | |
e0ba94f1 | 659 | break; |
dd360393 | 660 | case STLB_4K_2M: |
8b7e54b5 AD |
661 | tlb_lli_4k = max(tlb_lli_4k, entries); |
662 | tlb_lld_4k = max(tlb_lld_4k, entries); | |
663 | tlb_lli_2m = max(tlb_lli_2m, entries); | |
664 | tlb_lld_2m = max(tlb_lld_2m, entries); | |
665 | tlb_lli_4m = max(tlb_lli_4m, entries); | |
666 | tlb_lld_4m = max(tlb_lld_4m, entries); | |
dd360393 | 667 | break; |
e0ba94f1 | 668 | case TLB_INST_ALL: |
8b7e54b5 AD |
669 | tlb_lli_4k = max(tlb_lli_4k, entries); |
670 | tlb_lli_2m = max(tlb_lli_2m, entries); | |
671 | tlb_lli_4m = max(tlb_lli_4m, entries); | |
e0ba94f1 AS |
672 | break; |
673 | case TLB_INST_4K: | |
8b7e54b5 | 674 | tlb_lli_4k = max(tlb_lli_4k, entries); |
e0ba94f1 AS |
675 | break; |
676 | case TLB_INST_4M: | |
8b7e54b5 | 677 | tlb_lli_4m = max(tlb_lli_4m, entries); |
e0ba94f1 AS |
678 | break; |
679 | case TLB_INST_2M_4M: | |
8b7e54b5 AD |
680 | tlb_lli_2m = max(tlb_lli_2m, entries); |
681 | tlb_lli_4m = max(tlb_lli_4m, entries); | |
e0ba94f1 AS |
682 | break; |
683 | case TLB_DATA_4K: | |
684 | case TLB_DATA0_4K: | |
8b7e54b5 | 685 | tlb_lld_4k = max(tlb_lld_4k, entries); |
e0ba94f1 AS |
686 | break; |
687 | case TLB_DATA_4M: | |
688 | case TLB_DATA0_4M: | |
8b7e54b5 | 689 | tlb_lld_4m = max(tlb_lld_4m, entries); |
e0ba94f1 AS |
690 | break; |
691 | case TLB_DATA_2M_4M: | |
692 | case TLB_DATA0_2M_4M: | |
8b7e54b5 AD |
693 | tlb_lld_2m = max(tlb_lld_2m, entries); |
694 | tlb_lld_4m = max(tlb_lld_4m, entries); | |
e0ba94f1 AS |
695 | break; |
696 | case TLB_DATA_4K_4M: | |
8b7e54b5 AD |
697 | tlb_lld_4k = max(tlb_lld_4k, entries); |
698 | tlb_lld_4m = max(tlb_lld_4m, entries); | |
e0ba94f1 | 699 | break; |
f6bdaab7 | 700 | case TLB_DATA_1G_2M_4M: |
8b7e54b5 AD |
701 | tlb_lld_2m = max(tlb_lld_2m, TLB_0x63_2M_4M_ENTRIES); |
702 | tlb_lld_4m = max(tlb_lld_4m, TLB_0x63_2M_4M_ENTRIES); | |
f6bdaab7 | 703 | fallthrough; |
dd360393 | 704 | case TLB_DATA_1G: |
8b7e54b5 | 705 | tlb_lld_1g = max(tlb_lld_1g, entries); |
e0ba94f1 AS |
706 | break; |
707 | } | |
708 | } | |
709 | ||
148f9bb8 | 710 | static void intel_detect_tlb(struct cpuinfo_x86 *c) |
e0ba94f1 | 711 | { |
119deb95 | 712 | const struct leaf_0x2_table *desc; |
fe78079e | 713 | union leaf_0x2_regs regs; |
4772304e | 714 | u8 *ptr; |
5b556332 BP |
715 | |
716 | if (c->cpuid_level < 2) | |
717 | return; | |
718 | ||
e7df7289 | 719 | cpuid_leaf_0x2(®s); |
119deb95 AD |
720 | for_each_cpuid_0x2_desc(regs, ptr, desc) |
721 | intel_tlb_lookup(desc); | |
e0ba94f1 AS |
722 | } |
723 | ||
148f9bb8 | 724 | static const struct cpu_dev intel_cpu_dev = { |
1da177e4 | 725 | .c_vendor = "Intel", |
65eb6b43 | 726 | .c_ident = { "GenuineIntel" }, |
185f3b9d | 727 | #ifdef CONFIG_X86_32 |
09dc68d9 JB |
728 | .legacy_models = { |
729 | { .family = 4, .model_names = | |
65eb6b43 PC |
730 | { |
731 | [0] = "486 DX-25/33", | |
732 | [1] = "486 DX-50", | |
733 | [2] = "486 SX", | |
734 | [3] = "486 DX/2", | |
735 | [4] = "486 SL", | |
736 | [5] = "486 SX/2", | |
737 | [7] = "486 DX/2-WB", | |
738 | [8] = "486 DX/4", | |
1da177e4 LT |
739 | [9] = "486 DX/4-WB" |
740 | } | |
741 | }, | |
09dc68d9 | 742 | { .family = 5, .model_names = |
65eb6b43 PC |
743 | { |
744 | [0] = "Pentium 60/66 A-step", | |
745 | [1] = "Pentium 60/66", | |
1da177e4 | 746 | [2] = "Pentium 75 - 200", |
65eb6b43 | 747 | [3] = "OverDrive PODP5V83", |
1da177e4 | 748 | [4] = "Pentium MMX", |
65eb6b43 | 749 | [7] = "Mobile Pentium 75 - 200", |
aece118e BD |
750 | [8] = "Mobile Pentium MMX", |
751 | [9] = "Quark SoC X1000", | |
1da177e4 LT |
752 | } |
753 | }, | |
09dc68d9 | 754 | { .family = 6, .model_names = |
65eb6b43 | 755 | { |
1da177e4 | 756 | [0] = "Pentium Pro A-step", |
65eb6b43 PC |
757 | [1] = "Pentium Pro", |
758 | [3] = "Pentium II (Klamath)", | |
759 | [4] = "Pentium II (Deschutes)", | |
760 | [5] = "Pentium II (Deschutes)", | |
1da177e4 | 761 | [6] = "Mobile Pentium II", |
65eb6b43 PC |
762 | [7] = "Pentium III (Katmai)", |
763 | [8] = "Pentium III (Coppermine)", | |
1da177e4 LT |
764 | [10] = "Pentium III (Cascades)", |
765 | [11] = "Pentium III (Tualatin)", | |
766 | } | |
767 | }, | |
09dc68d9 | 768 | { .family = 15, .model_names = |
1da177e4 LT |
769 | { |
770 | [0] = "Pentium 4 (Unknown)", | |
771 | [1] = "Pentium 4 (Willamette)", | |
772 | [2] = "Pentium 4 (Northwood)", | |
773 | [4] = "Pentium 4 (Foster)", | |
774 | [5] = "Pentium 4 (Foster)", | |
775 | } | |
776 | }, | |
777 | }, | |
09dc68d9 | 778 | .legacy_cache_size = intel_size_cache, |
185f3b9d | 779 | #endif |
e0ba94f1 | 780 | .c_detect_tlb = intel_detect_tlb, |
03ae5768 | 781 | .c_early_init = early_init_intel, |
923f3a2b | 782 | .c_bsp_init = bsp_init_intel, |
1da177e4 | 783 | .c_init = init_intel, |
10a434fc | 784 | .c_x86_vendor = X86_VENDOR_INTEL, |
1da177e4 LT |
785 | }; |
786 | ||
10a434fc | 787 | cpu_dev_register(intel_cpu_dev); |