xref: /openbmc/linux/arch/x86/include/asm/microcode.h (revision 9b799b78)
1 #ifndef _ASM_X86_MICROCODE_H
2 #define _ASM_X86_MICROCODE_H
3 
4 #define native_rdmsr(msr, val1, val2)			\
5 do {							\
6 	u64 __val = native_read_msr((msr));		\
7 	(void)((val1) = (u32)__val);			\
8 	(void)((val2) = (u32)(__val >> 32));		\
9 } while (0)
10 
11 #define native_wrmsr(msr, low, high)			\
12 	native_write_msr(msr, low, high)
13 
14 #define native_wrmsrl(msr, val)				\
15 	native_write_msr((msr),				\
16 			 (u32)((u64)(val)),		\
17 			 (u32)((u64)(val) >> 32))
18 
19 struct cpu_signature {
20 	unsigned int sig;
21 	unsigned int pf;
22 	unsigned int rev;
23 };
24 
25 struct device;
26 
27 enum ucode_state { UCODE_ERROR, UCODE_OK, UCODE_NFOUND };
28 extern bool dis_ucode_ldr;
29 
30 struct microcode_ops {
31 	enum ucode_state (*request_microcode_user) (int cpu,
32 				const void __user *buf, size_t size);
33 
34 	enum ucode_state (*request_microcode_fw) (int cpu, struct device *,
35 						  bool refresh_fw);
36 
37 	void (*microcode_fini_cpu) (int cpu);
38 
39 	/*
40 	 * The generic 'microcode_core' part guarantees that
41 	 * the callbacks below run on a target cpu when they
42 	 * are being called.
43 	 * See also the "Synchronization" section in microcode_core.c.
44 	 */
45 	int (*apply_microcode) (int cpu);
46 	int (*collect_cpu_info) (int cpu, struct cpu_signature *csig);
47 };
48 
49 struct ucode_cpu_info {
50 	struct cpu_signature	cpu_sig;
51 	int			valid;
52 	void			*mc;
53 };
54 extern struct ucode_cpu_info ucode_cpu_info[];
55 
56 #ifdef CONFIG_MICROCODE_INTEL
57 extern struct microcode_ops * __init init_intel_microcode(void);
58 #else
59 static inline struct microcode_ops * __init init_intel_microcode(void)
60 {
61 	return NULL;
62 }
63 #endif /* CONFIG_MICROCODE_INTEL */
64 
65 #ifdef CONFIG_MICROCODE_AMD
66 extern struct microcode_ops * __init init_amd_microcode(void);
67 extern void __exit exit_amd_microcode(void);
68 #else
69 static inline struct microcode_ops * __init init_amd_microcode(void)
70 {
71 	return NULL;
72 }
73 static inline void __exit exit_amd_microcode(void) {}
74 #endif
75 
76 #ifdef CONFIG_MICROCODE_EARLY
77 #define MAX_UCODE_COUNT 128
78 
79 #define QCHAR(a, b, c, d) ((a) + ((b) << 8) + ((c) << 16) + ((d) << 24))
80 #define CPUID_INTEL1 QCHAR('G', 'e', 'n', 'u')
81 #define CPUID_INTEL2 QCHAR('i', 'n', 'e', 'I')
82 #define CPUID_INTEL3 QCHAR('n', 't', 'e', 'l')
83 #define CPUID_AMD1 QCHAR('A', 'u', 't', 'h')
84 #define CPUID_AMD2 QCHAR('e', 'n', 't', 'i')
85 #define CPUID_AMD3 QCHAR('c', 'A', 'M', 'D')
86 
87 #define CPUID_IS(a, b, c, ebx, ecx, edx)	\
88 		(!((ebx ^ (a))|(edx ^ (b))|(ecx ^ (c))))
89 
90 /*
91  * In early loading microcode phase on BSP, boot_cpu_data is not set up yet.
92  * x86_vendor() gets vendor id for BSP.
93  *
94  * In 32 bit AP case, accessing boot_cpu_data needs linear address. To simplify
95  * coding, we still use x86_vendor() to get vendor id for AP.
96  *
97  * x86_vendor() gets vendor information directly from CPUID.
98  */
99 static inline int x86_vendor(void)
100 {
101 	u32 eax = 0x00000000;
102 	u32 ebx, ecx = 0, edx;
103 
104 	native_cpuid(&eax, &ebx, &ecx, &edx);
105 
106 	if (CPUID_IS(CPUID_INTEL1, CPUID_INTEL2, CPUID_INTEL3, ebx, ecx, edx))
107 		return X86_VENDOR_INTEL;
108 
109 	if (CPUID_IS(CPUID_AMD1, CPUID_AMD2, CPUID_AMD3, ebx, ecx, edx))
110 		return X86_VENDOR_AMD;
111 
112 	return X86_VENDOR_UNKNOWN;
113 }
114 
115 static inline unsigned int __x86_family(unsigned int sig)
116 {
117 	unsigned int x86;
118 
119 	x86 = (sig >> 8) & 0xf;
120 
121 	if (x86 == 0xf)
122 		x86 += (sig >> 20) & 0xff;
123 
124 	return x86;
125 }
126 
127 static inline unsigned int x86_family(void)
128 {
129 	u32 eax = 0x00000001;
130 	u32 ebx, ecx = 0, edx;
131 
132 	native_cpuid(&eax, &ebx, &ecx, &edx);
133 
134 	return __x86_family(eax);
135 }
136 
137 static inline unsigned int x86_model(unsigned int sig)
138 {
139 	unsigned int x86, model;
140 
141 	x86 = __x86_family(sig);
142 
143 	model = (sig >> 4) & 0xf;
144 
145 	if (x86 == 0x6 || x86 == 0xf)
146 		model += ((sig >> 16) & 0xf) << 4;
147 
148 	return model;
149 }
150 
151 extern void __init load_ucode_bsp(void);
152 extern void load_ucode_ap(void);
153 extern int __init save_microcode_in_initrd(void);
154 void reload_early_microcode(void);
155 #else
156 static inline void __init load_ucode_bsp(void) {}
157 static inline void load_ucode_ap(void) {}
158 static inline int __init save_microcode_in_initrd(void)
159 {
160 	return 0;
161 }
162 static inline void reload_early_microcode(void) {}
163 #endif
164 
165 #endif /* _ASM_X86_MICROCODE_H */
166