xref: /openbmc/linux/arch/powerpc/include/asm/cache.h (revision d7cceda96badc1bd444cff27ab9c375a1277c1e3)
1b2441318SGreg Kroah-Hartman /* SPDX-License-Identifier: GPL-2.0 */
2b8b572e1SStephen Rothwell #ifndef _ASM_POWERPC_CACHE_H
3b8b572e1SStephen Rothwell #define _ASM_POWERPC_CACHE_H
4b8b572e1SStephen Rothwell 
5b8b572e1SStephen Rothwell #ifdef __KERNEL__
6b8b572e1SStephen Rothwell 
7b8b572e1SStephen Rothwell 
8b8b572e1SStephen Rothwell /* bytes per L1 cache line */
9968159c0SChristophe Leroy #if defined(CONFIG_PPC_8xx) || defined(CONFIG_403GCX)
10b8b572e1SStephen Rothwell #define L1_CACHE_SHIFT		4
11b8b572e1SStephen Rothwell #define MAX_COPY_PREFETCH	1
121128bb78SChristophe Leroy #define IFETCH_ALIGN_SHIFT	2
13b8b572e1SStephen Rothwell #elif defined(CONFIG_PPC_E500MC)
14b8b572e1SStephen Rothwell #define L1_CACHE_SHIFT		6
15b8b572e1SStephen Rothwell #define MAX_COPY_PREFETCH	4
161128bb78SChristophe Leroy #define IFETCH_ALIGN_SHIFT	3
17b8b572e1SStephen Rothwell #elif defined(CONFIG_PPC32)
18b8b572e1SStephen Rothwell #define MAX_COPY_PREFETCH	4
191128bb78SChristophe Leroy #define IFETCH_ALIGN_SHIFT	3	/* 603 fetches 2 insn at a time */
20e7f75ad0SDave Kleikamp #if defined(CONFIG_PPC_47x)
21e7f75ad0SDave Kleikamp #define L1_CACHE_SHIFT		7
22e7f75ad0SDave Kleikamp #else
23e7f75ad0SDave Kleikamp #define L1_CACHE_SHIFT		5
24e7f75ad0SDave Kleikamp #endif
25b8b572e1SStephen Rothwell #else /* CONFIG_PPC64 */
26b8b572e1SStephen Rothwell #define L1_CACHE_SHIFT		7
27f4329f2eSNicholas Piggin #define IFETCH_ALIGN_SHIFT	4 /* POWER8,9 */
28b8b572e1SStephen Rothwell #endif
29b8b572e1SStephen Rothwell 
30b8b572e1SStephen Rothwell #define	L1_CACHE_BYTES		(1 << L1_CACHE_SHIFT)
31b8b572e1SStephen Rothwell 
32b8b572e1SStephen Rothwell #define	SMP_CACHE_BYTES		L1_CACHE_BYTES
33b8b572e1SStephen Rothwell 
34f4329f2eSNicholas Piggin #define IFETCH_ALIGN_BYTES	(1 << IFETCH_ALIGN_SHIFT)
35f4329f2eSNicholas Piggin 
36b8b572e1SStephen Rothwell #if defined(__powerpc64__) && !defined(__ASSEMBLY__)
37e2827fe5SBenjamin Herrenschmidt 
38e2827fe5SBenjamin Herrenschmidt struct ppc_cache_info {
39e2827fe5SBenjamin Herrenschmidt 	u32 size;
40e2827fe5SBenjamin Herrenschmidt 	u32 line_size;
41e2827fe5SBenjamin Herrenschmidt 	u32 block_size;	/* L1 only */
42e2827fe5SBenjamin Herrenschmidt 	u32 log_block_size;
43e2827fe5SBenjamin Herrenschmidt 	u32 blocks_per_page;
44e2827fe5SBenjamin Herrenschmidt 	u32 sets;
4598a5f361SBenjamin Herrenschmidt 	u32 assoc;
46e2827fe5SBenjamin Herrenschmidt };
47e2827fe5SBenjamin Herrenschmidt 
48b8b572e1SStephen Rothwell struct ppc64_caches {
49e2827fe5SBenjamin Herrenschmidt 	struct ppc_cache_info l1d;
50e2827fe5SBenjamin Herrenschmidt 	struct ppc_cache_info l1i;
5165e01f38SBenjamin Herrenschmidt 	struct ppc_cache_info l2;
5265e01f38SBenjamin Herrenschmidt 	struct ppc_cache_info l3;
53b8b572e1SStephen Rothwell };
54b8b572e1SStephen Rothwell 
55b8b572e1SStephen Rothwell extern struct ppc64_caches ppc64_caches;
56b8b572e1SStephen Rothwell #endif /* __powerpc64__ && ! __ASSEMBLY__ */
57b8b572e1SStephen Rothwell 
580ce63670SKevin Hao #if defined(__ASSEMBLY__)
590ce63670SKevin Hao /*
600ce63670SKevin Hao  * For a snooping icache, we still need a dummy icbi to purge all the
610ce63670SKevin Hao  * prefetched instructions from the ifetch buffers. We also need a sync
620ce63670SKevin Hao  * before the icbi to order the the actual stores to memory that might
630ce63670SKevin Hao  * have modified instructions with the icbi.
640ce63670SKevin Hao  */
650ce63670SKevin Hao #define PURGE_PREFETCHED_INS	\
660ce63670SKevin Hao 	sync;			\
670ce63670SKevin Hao 	icbi	0,r3;		\
680ce63670SKevin Hao 	sync;			\
690ce63670SKevin Hao 	isync
70ae3a197eSDavid Howells 
710ce63670SKevin Hao #else
7254cb27a7SDenys Vlasenko #define __read_mostly __attribute__((__section__(".data..read_mostly")))
73ae3a197eSDavid Howells 
74*d7cceda9SChristophe Leroy #ifdef CONFIG_PPC_BOOK3S_32
75ae3a197eSDavid Howells extern long _get_L2CR(void);
76ae3a197eSDavid Howells extern long _get_L3CR(void);
77ae3a197eSDavid Howells extern void _set_L2CR(unsigned long);
78ae3a197eSDavid Howells extern void _set_L3CR(unsigned long);
79ae3a197eSDavid Howells #else
80ae3a197eSDavid Howells #define _get_L2CR()	0L
81ae3a197eSDavid Howells #define _get_L3CR()	0L
82ae3a197eSDavid Howells #define _set_L2CR(val)	do { } while(0)
83ae3a197eSDavid Howells #define _set_L3CR(val)	do { } while(0)
84b8b572e1SStephen Rothwell #endif
85b8b572e1SStephen Rothwell 
86d6bfa02fSChristophe Leroy static inline void dcbz(void *addr)
87d6bfa02fSChristophe Leroy {
88d6bfa02fSChristophe Leroy 	__asm__ __volatile__ ("dcbz 0, %0" : : "r"(addr) : "memory");
89d6bfa02fSChristophe Leroy }
90d6bfa02fSChristophe Leroy 
91d6bfa02fSChristophe Leroy static inline void dcbi(void *addr)
92d6bfa02fSChristophe Leroy {
93d6bfa02fSChristophe Leroy 	__asm__ __volatile__ ("dcbi 0, %0" : : "r"(addr) : "memory");
94d6bfa02fSChristophe Leroy }
95d6bfa02fSChristophe Leroy 
96d6bfa02fSChristophe Leroy static inline void dcbf(void *addr)
97d6bfa02fSChristophe Leroy {
98d6bfa02fSChristophe Leroy 	__asm__ __volatile__ ("dcbf 0, %0" : : "r"(addr) : "memory");
99d6bfa02fSChristophe Leroy }
100d6bfa02fSChristophe Leroy 
101d6bfa02fSChristophe Leroy static inline void dcbst(void *addr)
102d6bfa02fSChristophe Leroy {
103d6bfa02fSChristophe Leroy 	__asm__ __volatile__ ("dcbst 0, %0" : : "r"(addr) : "memory");
104d6bfa02fSChristophe Leroy }
105ae3a197eSDavid Howells #endif /* !__ASSEMBLY__ */
106b8b572e1SStephen Rothwell #endif /* __KERNEL__ */
107b8b572e1SStephen Rothwell #endif /* _ASM_POWERPC_CACHE_H */
108