1 /* SPDX-License-Identifier: GPL-2.0 */
2 /*
3  * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
4  */
5 #ifndef _ASM_PROCESSOR_H
6 #define _ASM_PROCESSOR_H
7 
8 #include <linux/atomic.h>
9 #include <linux/cpumask.h>
10 #include <linux/sizes.h>
11 
12 #include <asm/cpu.h>
13 #include <asm/cpu-info.h>
14 #include <asm/loongarch.h>
15 #include <asm/vdso/processor.h>
16 #include <uapi/asm/ptrace.h>
17 #include <uapi/asm/sigcontext.h>
18 
19 #ifdef CONFIG_32BIT
20 
21 #define TASK_SIZE	0x80000000UL
22 #define TASK_SIZE_MIN	TASK_SIZE
23 #define STACK_TOP_MAX	TASK_SIZE
24 
25 #define TASK_IS_32BIT_ADDR 1
26 
27 #endif
28 
29 #ifdef CONFIG_64BIT
30 
31 #define TASK_SIZE32	0x100000000UL
32 #define TASK_SIZE64     (0x1UL << ((cpu_vabits > VA_BITS) ? VA_BITS : cpu_vabits))
33 
34 #define TASK_SIZE	(test_thread_flag(TIF_32BIT_ADDR) ? TASK_SIZE32 : TASK_SIZE64)
35 #define TASK_SIZE_MIN	TASK_SIZE32
36 #define STACK_TOP_MAX	TASK_SIZE64
37 
38 #define TASK_SIZE_OF(tsk)						\
39 	(test_tsk_thread_flag(tsk, TIF_32BIT_ADDR) ? TASK_SIZE32 : TASK_SIZE64)
40 
41 #define TASK_IS_32BIT_ADDR test_thread_flag(TIF_32BIT_ADDR)
42 
43 #endif
44 
45 #define VDSO_RANDOMIZE_SIZE	(TASK_IS_32BIT_ADDR ? SZ_1M : SZ_64M)
46 
47 unsigned long stack_top(void);
48 #define STACK_TOP stack_top()
49 
50 /*
51  * This decides where the kernel will search for a free chunk of vm
52  * space during mmap's.
53  */
54 #define TASK_UNMAPPED_BASE PAGE_ALIGN(TASK_SIZE / 3)
55 
56 #define FPU_REG_WIDTH		256
57 #define FPU_ALIGN		__attribute__((aligned(32)))
58 
59 union fpureg {
60 	__u32	val32[FPU_REG_WIDTH / 32];
61 	__u64	val64[FPU_REG_WIDTH / 64];
62 };
63 
64 #define FPR_IDX(width, idx)	(idx)
65 
66 #define BUILD_FPR_ACCESS(width) \
67 static inline u##width get_fpr##width(union fpureg *fpr, unsigned idx)	\
68 {									\
69 	return fpr->val##width[FPR_IDX(width, idx)];			\
70 }									\
71 									\
72 static inline void set_fpr##width(union fpureg *fpr, unsigned int idx,	\
73 				  u##width val)				\
74 {									\
75 	fpr->val##width[FPR_IDX(width, idx)] = val;			\
76 }
77 
78 BUILD_FPR_ACCESS(32)
79 BUILD_FPR_ACCESS(64)
80 
81 struct loongarch_fpu {
82 	unsigned int	fcsr;
83 	unsigned int	vcsr;
84 	uint64_t	fcc;	/* 8x8 */
85 	union fpureg	fpr[NUM_FPU_REGS];
86 };
87 
88 #define INIT_CPUMASK { \
89 	{0,} \
90 }
91 
92 #define ARCH_MIN_TASKALIGN	32
93 
94 struct loongarch_vdso_info;
95 
96 /*
97  * If you change thread_struct remember to change the #defines below too!
98  */
99 struct thread_struct {
100 	/* Main processor registers. */
101 	unsigned long reg01, reg03, reg22; /* ra sp fp */
102 	unsigned long reg23, reg24, reg25, reg26; /* s0-s3 */
103 	unsigned long reg27, reg28, reg29, reg30, reg31; /* s4-s8 */
104 
105 	/* CSR registers */
106 	unsigned long csr_prmd;
107 	unsigned long csr_crmd;
108 	unsigned long csr_euen;
109 	unsigned long csr_ecfg;
110 	unsigned long csr_badvaddr;	/* Last user fault */
111 
112 	/* Scratch registers */
113 	unsigned long scr0;
114 	unsigned long scr1;
115 	unsigned long scr2;
116 	unsigned long scr3;
117 
118 	/* Eflags register */
119 	unsigned long eflags;
120 
121 	/* Other stuff associated with the thread. */
122 	unsigned long trap_nr;
123 	unsigned long error_code;
124 	struct loongarch_vdso_info *vdso;
125 
126 	/*
127 	 * FPU & vector registers, must be at last because
128 	 * they are conditionally copied at fork().
129 	 */
130 	struct loongarch_fpu fpu FPU_ALIGN;
131 };
132 
133 #define INIT_THREAD  {						\
134 	/*							\
135 	 * Main processor registers				\
136 	 */							\
137 	.reg01			= 0,				\
138 	.reg03			= 0,				\
139 	.reg22			= 0,				\
140 	.reg23			= 0,				\
141 	.reg24			= 0,				\
142 	.reg25			= 0,				\
143 	.reg26			= 0,				\
144 	.reg27			= 0,				\
145 	.reg28			= 0,				\
146 	.reg29			= 0,				\
147 	.reg30			= 0,				\
148 	.reg31			= 0,				\
149 	.csr_crmd		= 0,				\
150 	.csr_prmd		= 0,				\
151 	.csr_euen		= 0,				\
152 	.csr_ecfg		= 0,				\
153 	.csr_badvaddr		= 0,				\
154 	/*							\
155 	 * Other stuff associated with the process		\
156 	 */							\
157 	.trap_nr		= 0,				\
158 	.error_code		= 0,				\
159 	/*							\
160 	 * FPU & vector registers				\
161 	 */							\
162 	.fpu			= {				\
163 		.fcsr		= 0,				\
164 		.vcsr		= 0,				\
165 		.fcc		= 0,				\
166 		.fpr		= {{{0,},},},			\
167 	},							\
168 }
169 
170 struct task_struct;
171 
172 /* Free all resources held by a thread. */
173 #define release_thread(thread) do { } while (0)
174 
175 enum idle_boot_override {IDLE_NO_OVERRIDE = 0, IDLE_HALT, IDLE_NOMWAIT, IDLE_POLL};
176 
177 extern unsigned long		boot_option_idle_override;
178 /*
179  * Do necessary setup to start up a newly executed thread.
180  */
181 extern void start_thread(struct pt_regs *regs, unsigned long pc, unsigned long sp);
182 
183 static inline void flush_thread(void)
184 {
185 }
186 
187 unsigned long __get_wchan(struct task_struct *p);
188 
189 #define __KSTK_TOS(tsk) ((unsigned long)task_stack_page(tsk) + \
190 			 THREAD_SIZE - 32 - sizeof(struct pt_regs))
191 #define task_pt_regs(tsk) ((struct pt_regs *)__KSTK_TOS(tsk))
192 #define KSTK_EIP(tsk) (task_pt_regs(tsk)->csr_era)
193 #define KSTK_ESP(tsk) (task_pt_regs(tsk)->regs[3])
194 #define KSTK_EUEN(tsk) (task_pt_regs(tsk)->csr_euen)
195 #define KSTK_ECFG(tsk) (task_pt_regs(tsk)->csr_ecfg)
196 
197 #define return_address() ({__asm__ __volatile__("":::"$1"); __builtin_return_address(0);})
198 
199 #ifdef CONFIG_CPU_HAS_PREFETCH
200 
201 #define ARCH_HAS_PREFETCH
202 #define prefetch(x) __builtin_prefetch((x), 0, 1)
203 
204 #define ARCH_HAS_PREFETCHW
205 #define prefetchw(x) __builtin_prefetch((x), 1, 1)
206 
207 #endif
208 
209 #endif /* _ASM_PROCESSOR_H */
210