xref: /openbmc/linux/arch/mips/include/asm/hazards.h (revision dc6a81c3)
1 /*
2  * This file is subject to the terms and conditions of the GNU General Public
3  * License.  See the file "COPYING" in the main directory of this archive
4  * for more details.
5  *
6  * Copyright (C) 2003, 04, 07 Ralf Baechle <ralf@linux-mips.org>
7  * Copyright (C) MIPS Technologies, Inc.
8  *   written by Ralf Baechle <ralf@linux-mips.org>
9  */
10 #ifndef _ASM_HAZARDS_H
11 #define _ASM_HAZARDS_H
12 
13 #include <linux/stringify.h>
14 #include <asm/compiler.h>
15 
16 #define ___ssnop							\
17 	sll	$0, $0, 1
18 
19 #define ___ehb								\
20 	sll	$0, $0, 3
21 
22 /*
23  * TLB hazards
24  */
25 #if (defined(CONFIG_CPU_MIPSR2) || defined(CONFIG_CPU_MIPSR6)) && \
26 	!defined(CONFIG_CPU_CAVIUM_OCTEON) && !defined(CONFIG_CPU_LOONGSON64)
27 
28 /*
29  * MIPSR2 defines ehb for hazard avoidance
30  */
31 
32 #define __mtc0_tlbw_hazard						\
33 	___ehb
34 
35 #define __mtc0_tlbr_hazard						\
36 	___ehb
37 
38 #define __tlbw_use_hazard						\
39 	___ehb
40 
41 #define __tlb_read_hazard						\
42 	___ehb
43 
44 #define __tlb_probe_hazard						\
45 	___ehb
46 
47 #define __irq_enable_hazard						\
48 	___ehb
49 
50 #define __irq_disable_hazard						\
51 	___ehb
52 
53 #define __back_to_back_c0_hazard					\
54 	___ehb
55 
56 /*
57  * gcc has a tradition of misscompiling the previous construct using the
58  * address of a label as argument to inline assembler.	Gas otoh has the
59  * annoying difference between la and dla which are only usable for 32-bit
60  * rsp. 64-bit code, so can't be used without conditional compilation.
61  * The alternative is switching the assembler to 64-bit code which happens
62  * to work right even for 32-bit code...
63  */
64 #define instruction_hazard()						\
65 do {									\
66 	unsigned long tmp;						\
67 									\
68 	__asm__ __volatile__(						\
69 	"	.set	push					\n"	\
70 	"	.set "MIPS_ISA_LEVEL"				\n"	\
71 	"	dla	%0, 1f					\n"	\
72 	"	jr.hb	%0					\n"	\
73 	"	.set	pop					\n"	\
74 	"1:							\n"	\
75 	: "=r" (tmp));							\
76 } while (0)
77 
78 #elif (defined(CONFIG_CPU_MIPSR1) && !defined(CONFIG_MIPS_ALCHEMY)) || \
79 	defined(CONFIG_CPU_BMIPS)
80 
81 /*
82  * These are slightly complicated by the fact that we guarantee R1 kernels to
83  * run fine on R2 processors.
84  */
85 
86 #define __mtc0_tlbw_hazard						\
87 	___ssnop;							\
88 	___ssnop;							\
89 	___ehb
90 
91 #define __mtc0_tlbr_hazard						\
92 	___ssnop;							\
93 	___ssnop;							\
94 	___ehb
95 
96 #define __tlbw_use_hazard						\
97 	___ssnop;							\
98 	___ssnop;							\
99 	___ssnop;							\
100 	___ehb
101 
102 #define __tlb_read_hazard						\
103 	___ssnop;							\
104 	___ssnop;							\
105 	___ssnop;							\
106 	___ehb
107 
108 #define __tlb_probe_hazard						\
109 	___ssnop;							\
110 	___ssnop;							\
111 	___ssnop;							\
112 	___ehb
113 
114 #define __irq_enable_hazard						\
115 	___ssnop;							\
116 	___ssnop;							\
117 	___ssnop;							\
118 	___ehb
119 
120 #define __irq_disable_hazard						\
121 	___ssnop;							\
122 	___ssnop;							\
123 	___ssnop;							\
124 	___ehb
125 
126 #define __back_to_back_c0_hazard					\
127 	___ssnop;							\
128 	___ssnop;							\
129 	___ssnop;							\
130 	___ehb
131 
132 /*
133  * gcc has a tradition of misscompiling the previous construct using the
134  * address of a label as argument to inline assembler.	Gas otoh has the
135  * annoying difference between la and dla which are only usable for 32-bit
136  * rsp. 64-bit code, so can't be used without conditional compilation.
137  * The alternative is switching the assembler to 64-bit code which happens
138  * to work right even for 32-bit code...
139  */
140 #define __instruction_hazard()						\
141 do {									\
142 	unsigned long tmp;						\
143 									\
144 	__asm__ __volatile__(						\
145 	"	.set	push					\n"	\
146 	"	.set	mips64r2				\n"	\
147 	"	dla	%0, 1f					\n"	\
148 	"	jr.hb	%0					\n"	\
149 	"	.set	pop					\n"	\
150 	"1:							\n"	\
151 	: "=r" (tmp));							\
152 } while (0)
153 
154 #define instruction_hazard()						\
155 do {									\
156 	if (cpu_has_mips_r2_r6)						\
157 		__instruction_hazard();					\
158 } while (0)
159 
160 #elif defined(CONFIG_MIPS_ALCHEMY) || defined(CONFIG_CPU_CAVIUM_OCTEON) || \
161 	defined(CONFIG_CPU_LOONGSON2EF) || defined(CONFIG_CPU_LOONGSON64) || \
162 	defined(CONFIG_CPU_R10000) || defined(CONFIG_CPU_R5500) || defined(CONFIG_CPU_XLR)
163 
164 /*
165  * R10000 rocks - all hazards handled in hardware, so this becomes a nobrainer.
166  */
167 
168 #define __mtc0_tlbw_hazard
169 
170 #define __mtc0_tlbr_hazard
171 
172 #define __tlbw_use_hazard
173 
174 #define __tlb_read_hazard
175 
176 #define __tlb_probe_hazard
177 
178 #define __irq_enable_hazard
179 
180 #define __irq_disable_hazard
181 
182 #define __back_to_back_c0_hazard
183 
184 #define instruction_hazard() do { } while (0)
185 
186 #elif defined(CONFIG_CPU_SB1)
187 
188 /*
189  * Mostly like R4000 for historic reasons
190  */
191 #define __mtc0_tlbw_hazard
192 
193 #define __mtc0_tlbr_hazard
194 
195 #define __tlbw_use_hazard
196 
197 #define __tlb_read_hazard
198 
199 #define __tlb_probe_hazard
200 
201 #define __irq_enable_hazard
202 
203 #define __irq_disable_hazard						\
204 	___ssnop;							\
205 	___ssnop;							\
206 	___ssnop
207 
208 #define __back_to_back_c0_hazard
209 
210 #define instruction_hazard() do { } while (0)
211 
212 #else
213 
214 /*
215  * Finally the catchall case for all other processors including R4000, R4400,
216  * R4600, R4700, R5000, RM7000, NEC VR41xx etc.
217  *
218  * The taken branch will result in a two cycle penalty for the two killed
219  * instructions on R4000 / R4400.  Other processors only have a single cycle
220  * hazard so this is nice trick to have an optimal code for a range of
221  * processors.
222  */
223 #define __mtc0_tlbw_hazard						\
224 	nop;								\
225 	nop
226 
227 #define __mtc0_tlbr_hazard						\
228 	nop;								\
229 	nop
230 
231 #define __tlbw_use_hazard						\
232 	nop;								\
233 	nop;								\
234 	nop
235 
236 #define __tlb_read_hazard						\
237 	nop;								\
238 	nop;								\
239 	nop
240 
241 #define __tlb_probe_hazard						\
242 	nop;								\
243 	nop;								\
244 	nop
245 
246 #define __irq_enable_hazard						\
247 	___ssnop;							\
248 	___ssnop;							\
249 	___ssnop
250 
251 #define __irq_disable_hazard						\
252 	nop;								\
253 	nop;								\
254 	nop
255 
256 #define __back_to_back_c0_hazard					\
257 	___ssnop;							\
258 	___ssnop;							\
259 	___ssnop
260 
261 #define instruction_hazard() do { } while (0)
262 
263 #endif
264 
265 
266 /* FPU hazards */
267 
268 #if defined(CONFIG_CPU_SB1)
269 
270 #define __enable_fpu_hazard						\
271 	.set	push;							\
272 	.set	mips64;							\
273 	.set	noreorder;						\
274 	___ssnop;							\
275 	bnezl	$0, .+4;						\
276 	___ssnop;							\
277 	.set	pop
278 
279 #define __disable_fpu_hazard
280 
281 #elif defined(CONFIG_CPU_MIPSR2) || defined(CONFIG_CPU_MIPSR6)
282 
283 #define __enable_fpu_hazard						\
284 	___ehb
285 
286 #define __disable_fpu_hazard						\
287 	___ehb
288 
289 #else
290 
291 #define __enable_fpu_hazard						\
292 	nop;								\
293 	nop;								\
294 	nop;								\
295 	nop
296 
297 #define __disable_fpu_hazard						\
298 	___ehb
299 
300 #endif
301 
302 #ifdef __ASSEMBLY__
303 
304 #define _ssnop ___ssnop
305 #define	_ehb ___ehb
306 #define mtc0_tlbw_hazard __mtc0_tlbw_hazard
307 #define mtc0_tlbr_hazard __mtc0_tlbr_hazard
308 #define tlbw_use_hazard __tlbw_use_hazard
309 #define tlb_read_hazard __tlb_read_hazard
310 #define tlb_probe_hazard __tlb_probe_hazard
311 #define irq_enable_hazard __irq_enable_hazard
312 #define irq_disable_hazard __irq_disable_hazard
313 #define back_to_back_c0_hazard __back_to_back_c0_hazard
314 #define enable_fpu_hazard __enable_fpu_hazard
315 #define disable_fpu_hazard __disable_fpu_hazard
316 
317 #else
318 
319 #define _ssnop()							\
320 do {									\
321 	__asm__ __volatile__(						\
322 	__stringify(___ssnop)						\
323 	);								\
324 } while (0)
325 
326 #define	_ehb()								\
327 do {									\
328 	__asm__ __volatile__(						\
329 	__stringify(___ehb)						\
330 	);								\
331 } while (0)
332 
333 
334 #define mtc0_tlbw_hazard()						\
335 do {									\
336 	__asm__ __volatile__(						\
337 	__stringify(__mtc0_tlbw_hazard)					\
338 	);								\
339 } while (0)
340 
341 
342 #define mtc0_tlbr_hazard()						\
343 do {									\
344 	__asm__ __volatile__(						\
345 	__stringify(__mtc0_tlbr_hazard)					\
346 	);								\
347 } while (0)
348 
349 
350 #define tlbw_use_hazard()						\
351 do {									\
352 	__asm__ __volatile__(						\
353 	__stringify(__tlbw_use_hazard)					\
354 	);								\
355 } while (0)
356 
357 
358 #define tlb_read_hazard()						\
359 do {									\
360 	__asm__ __volatile__(						\
361 	__stringify(__tlb_read_hazard)					\
362 	);								\
363 } while (0)
364 
365 
366 #define tlb_probe_hazard()						\
367 do {									\
368 	__asm__ __volatile__(						\
369 	__stringify(__tlb_probe_hazard)					\
370 	);								\
371 } while (0)
372 
373 
374 #define irq_enable_hazard()						\
375 do {									\
376 	__asm__ __volatile__(						\
377 	__stringify(__irq_enable_hazard)				\
378 	);								\
379 } while (0)
380 
381 
382 #define irq_disable_hazard()						\
383 do {									\
384 	__asm__ __volatile__(						\
385 	__stringify(__irq_disable_hazard)				\
386 	);								\
387 } while (0)
388 
389 
390 #define back_to_back_c0_hazard() 					\
391 do {									\
392 	__asm__ __volatile__(						\
393 	__stringify(__back_to_back_c0_hazard)				\
394 	);								\
395 } while (0)
396 
397 
398 #define enable_fpu_hazard()						\
399 do {									\
400 	__asm__ __volatile__(						\
401 	__stringify(__enable_fpu_hazard)				\
402 	);								\
403 } while (0)
404 
405 
406 #define disable_fpu_hazard()						\
407 do {									\
408 	__asm__ __volatile__(						\
409 	__stringify(__disable_fpu_hazard)				\
410 	);								\
411 } while (0)
412 
413 /*
414  * MIPS R2 instruction hazard barrier.   Needs to be called as a subroutine.
415  */
416 extern void mips_ihb(void);
417 
418 #endif /* __ASSEMBLY__  */
419 
420 #endif /* _ASM_HAZARDS_H */
421