1 /* SPDX-License-Identifier: GPL-2.0 */ 2 /* 3 * Copyright (C) 2020-2022 Loongson Technology Corporation Limited 4 */ 5 #ifndef _ASM_PROCESSOR_H 6 #define _ASM_PROCESSOR_H 7 8 #include <linux/atomic.h> 9 #include <linux/cpumask.h> 10 #include <linux/sizes.h> 11 12 #include <asm/cpu.h> 13 #include <asm/cpu-info.h> 14 #include <asm/loongarch.h> 15 #include <asm/vdso/processor.h> 16 #include <uapi/asm/ptrace.h> 17 #include <uapi/asm/sigcontext.h> 18 19 #ifdef CONFIG_32BIT 20 21 #define TASK_SIZE 0x80000000UL 22 #define TASK_SIZE_MIN TASK_SIZE 23 #define STACK_TOP_MAX TASK_SIZE 24 25 #define TASK_IS_32BIT_ADDR 1 26 27 #endif 28 29 #ifdef CONFIG_64BIT 30 31 #define TASK_SIZE32 0x100000000UL 32 #define TASK_SIZE64 (0x1UL << ((cpu_vabits > VA_BITS) ? VA_BITS : cpu_vabits)) 33 34 #define TASK_SIZE (test_thread_flag(TIF_32BIT_ADDR) ? TASK_SIZE32 : TASK_SIZE64) 35 #define TASK_SIZE_MIN TASK_SIZE32 36 #define STACK_TOP_MAX TASK_SIZE64 37 38 #define TASK_SIZE_OF(tsk) \ 39 (test_tsk_thread_flag(tsk, TIF_32BIT_ADDR) ? TASK_SIZE32 : TASK_SIZE64) 40 41 #define TASK_IS_32BIT_ADDR test_thread_flag(TIF_32BIT_ADDR) 42 43 #endif 44 45 #define VDSO_RANDOMIZE_SIZE (TASK_IS_32BIT_ADDR ? SZ_1M : SZ_64M) 46 47 unsigned long stack_top(void); 48 #define STACK_TOP stack_top() 49 50 /* 51 * This decides where the kernel will search for a free chunk of vm 52 * space during mmap's. 53 */ 54 #define TASK_UNMAPPED_BASE PAGE_ALIGN(TASK_SIZE / 3) 55 56 #define FPU_REG_WIDTH 256 57 #define FPU_ALIGN __attribute__((aligned(32))) 58 59 union fpureg { 60 __u32 val32[FPU_REG_WIDTH / 32]; 61 __u64 val64[FPU_REG_WIDTH / 64]; 62 }; 63 64 #define FPR_IDX(width, idx) (idx) 65 66 #define BUILD_FPR_ACCESS(width) \ 67 static inline u##width get_fpr##width(union fpureg *fpr, unsigned idx) \ 68 { \ 69 return fpr->val##width[FPR_IDX(width, idx)]; \ 70 } \ 71 \ 72 static inline void set_fpr##width(union fpureg *fpr, unsigned int idx, \ 73 u##width val) \ 74 { \ 75 fpr->val##width[FPR_IDX(width, idx)] = val; \ 76 } 77 78 BUILD_FPR_ACCESS(32) 79 BUILD_FPR_ACCESS(64) 80 81 struct loongarch_fpu { 82 unsigned int fcsr; 83 unsigned int vcsr; 84 uint64_t fcc; /* 8x8 */ 85 union fpureg fpr[NUM_FPU_REGS]; 86 }; 87 88 #define INIT_CPUMASK { \ 89 {0,} \ 90 } 91 92 #define ARCH_MIN_TASKALIGN 32 93 94 struct loongarch_vdso_info; 95 96 /* 97 * If you change thread_struct remember to change the #defines below too! 98 */ 99 struct thread_struct { 100 /* Main processor registers. */ 101 unsigned long reg01, reg03, reg22; /* ra sp fp */ 102 unsigned long reg23, reg24, reg25, reg26; /* s0-s3 */ 103 unsigned long reg27, reg28, reg29, reg30, reg31; /* s4-s8 */ 104 105 /* CSR registers */ 106 unsigned long csr_prmd; 107 unsigned long csr_crmd; 108 unsigned long csr_euen; 109 unsigned long csr_ecfg; 110 unsigned long csr_badvaddr; /* Last user fault */ 111 112 /* Scratch registers */ 113 unsigned long scr0; 114 unsigned long scr1; 115 unsigned long scr2; 116 unsigned long scr3; 117 118 /* Eflags register */ 119 unsigned long eflags; 120 121 /* Other stuff associated with the thread. */ 122 unsigned long trap_nr; 123 unsigned long error_code; 124 struct loongarch_vdso_info *vdso; 125 126 /* 127 * FPU & vector registers, must be at last because 128 * they are conditionally copied at fork(). 129 */ 130 struct loongarch_fpu fpu FPU_ALIGN; 131 }; 132 133 #define INIT_THREAD { \ 134 /* \ 135 * Main processor registers \ 136 */ \ 137 .reg01 = 0, \ 138 .reg03 = 0, \ 139 .reg22 = 0, \ 140 .reg23 = 0, \ 141 .reg24 = 0, \ 142 .reg25 = 0, \ 143 .reg26 = 0, \ 144 .reg27 = 0, \ 145 .reg28 = 0, \ 146 .reg29 = 0, \ 147 .reg30 = 0, \ 148 .reg31 = 0, \ 149 .csr_crmd = 0, \ 150 .csr_prmd = 0, \ 151 .csr_euen = 0, \ 152 .csr_ecfg = 0, \ 153 .csr_badvaddr = 0, \ 154 /* \ 155 * Other stuff associated with the process \ 156 */ \ 157 .trap_nr = 0, \ 158 .error_code = 0, \ 159 /* \ 160 * FPU & vector registers \ 161 */ \ 162 .fpu = { \ 163 .fcsr = 0, \ 164 .vcsr = 0, \ 165 .fcc = 0, \ 166 .fpr = {{{0,},},}, \ 167 }, \ 168 } 169 170 struct task_struct; 171 172 /* Free all resources held by a thread. */ 173 #define release_thread(thread) do { } while (0) 174 175 enum idle_boot_override {IDLE_NO_OVERRIDE = 0, IDLE_HALT, IDLE_NOMWAIT, IDLE_POLL}; 176 177 extern unsigned long boot_option_idle_override; 178 /* 179 * Do necessary setup to start up a newly executed thread. 180 */ 181 extern void start_thread(struct pt_regs *regs, unsigned long pc, unsigned long sp); 182 183 static inline void flush_thread(void) 184 { 185 } 186 187 unsigned long __get_wchan(struct task_struct *p); 188 189 #define __KSTK_TOS(tsk) ((unsigned long)task_stack_page(tsk) + \ 190 THREAD_SIZE - 32 - sizeof(struct pt_regs)) 191 #define task_pt_regs(tsk) ((struct pt_regs *)__KSTK_TOS(tsk)) 192 #define KSTK_EIP(tsk) (task_pt_regs(tsk)->csr_era) 193 #define KSTK_ESP(tsk) (task_pt_regs(tsk)->regs[3]) 194 #define KSTK_EUEN(tsk) (task_pt_regs(tsk)->csr_euen) 195 #define KSTK_ECFG(tsk) (task_pt_regs(tsk)->csr_ecfg) 196 197 #define return_address() ({__asm__ __volatile__("":::"$1"); __builtin_return_address(0);}) 198 199 #ifdef CONFIG_CPU_HAS_PREFETCH 200 201 #define ARCH_HAS_PREFETCH 202 #define prefetch(x) __builtin_prefetch((x), 0, 1) 203 204 #define ARCH_HAS_PREFETCHW 205 #define prefetchw(x) __builtin_prefetch((x), 1, 1) 206 207 #endif 208 209 #endif /* _ASM_PROCESSOR_H */ 210