1 #include <common.h> 2 #include <exports.h> 3 #include <linux/compiler.h> 4 5 #define FO(x) offsetof(struct jt_funcs, x) 6 7 #if defined(CONFIG_X86) 8 /* 9 * x86 does not have a dedicated register to store the pointer to 10 * the global_data. Thus the jump table address is stored in a 11 * global variable, but such approach does not allow for execution 12 * from flash memory. The global_data address is passed as argv[-1] 13 * to the application program. 14 */ 15 static struct jt_funcs *jt; 16 gd_t *global_data; 17 18 #define EXPORT_FUNC(f, a, x, ...) \ 19 asm volatile ( \ 20 " .globl " #x "\n" \ 21 #x ":\n" \ 22 " movl %0, %%eax\n" \ 23 " movl jt, %%ecx\n" \ 24 " jmp *(%%ecx, %%eax)\n" \ 25 : : "i"(FO(x)) : "eax", "ecx"); 26 #elif defined(CONFIG_PPC) 27 /* 28 * r2 holds the pointer to the global_data, r11 is a call-clobbered 29 * register 30 */ 31 #define EXPORT_FUNC(f, a, x, ...) \ 32 asm volatile ( \ 33 " .globl " #x "\n" \ 34 #x ":\n" \ 35 " lwz %%r11, %0(%%r2)\n" \ 36 " lwz %%r11, %1(%%r11)\n" \ 37 " mtctr %%r11\n" \ 38 " bctr\n" \ 39 : : "i"(offsetof(gd_t, jt)), "i"(FO(x)) : "r11"); 40 #elif defined(CONFIG_ARM) 41 #ifdef CONFIG_ARM64 42 /* 43 * x18 holds the pointer to the global_data, x9 is a call-clobbered 44 * register 45 */ 46 #define EXPORT_FUNC(f, a, x, ...) \ 47 asm volatile ( \ 48 " .globl " #x "\n" \ 49 #x ":\n" \ 50 " ldr x9, [x18, %0]\n" \ 51 " ldr x9, [x9, %1]\n" \ 52 " br x9\n" \ 53 : : "i"(offsetof(gd_t, jt)), "i"(FO(x)) : "x9"); 54 #else 55 /* 56 * r9 holds the pointer to the global_data, ip is a call-clobbered 57 * register 58 */ 59 #define EXPORT_FUNC(f, a, x, ...) \ 60 asm volatile ( \ 61 " .globl " #x "\n" \ 62 #x ":\n" \ 63 " ldr ip, [r9, %0]\n" \ 64 " ldr pc, [ip, %1]\n" \ 65 : : "i"(offsetof(gd_t, jt)), "i"(FO(x)) : "ip"); 66 #endif 67 #elif defined(CONFIG_MIPS) 68 #ifdef CONFIG_CPU_MIPS64 69 /* 70 * k0 ($26) holds the pointer to the global_data; t9 ($25) is a call- 71 * clobbered register that is also used to set gp ($26). Note that the 72 * jr instruction also executes the instruction immediately following 73 * it; however, GCC/mips generates an additional `nop' after each asm 74 * statement 75 */ 76 #define EXPORT_FUNC(f, a, x, ...) \ 77 asm volatile ( \ 78 " .globl " #x "\n" \ 79 #x ":\n" \ 80 " ld $25, %0($26)\n" \ 81 " ld $25, %1($25)\n" \ 82 " jr $25\n" \ 83 : : "i"(offsetof(gd_t, jt)), "i"(FO(x)) : "t9"); 84 #else 85 /* 86 * k0 ($26) holds the pointer to the global_data; t9 ($25) is a call- 87 * clobbered register that is also used to set gp ($26). Note that the 88 * jr instruction also executes the instruction immediately following 89 * it; however, GCC/mips generates an additional `nop' after each asm 90 * statement 91 */ 92 #define EXPORT_FUNC(f, a, x, ...) \ 93 asm volatile ( \ 94 " .globl " #x "\n" \ 95 #x ":\n" \ 96 " lw $25, %0($26)\n" \ 97 " lw $25, %1($25)\n" \ 98 " jr $25\n" \ 99 : : "i"(offsetof(gd_t, jt)), "i"(FO(x)) : "t9"); 100 #endif 101 #elif defined(CONFIG_NIOS2) 102 /* 103 * gp holds the pointer to the global_data, r8 is call-clobbered 104 */ 105 #define EXPORT_FUNC(f, a, x, ...) \ 106 asm volatile ( \ 107 " .globl " #x "\n" \ 108 #x ":\n" \ 109 " movhi r8, %%hi(%0)\n" \ 110 " ori r8, r0, %%lo(%0)\n" \ 111 " add r8, r8, gp\n" \ 112 " ldw r8, 0(r8)\n" \ 113 " ldw r8, %1(r8)\n" \ 114 " jmp r8\n" \ 115 : : "i"(offsetof(gd_t, jt)), "i"(FO(x)) : "gp"); 116 #elif defined(CONFIG_M68K) 117 /* 118 * d7 holds the pointer to the global_data, a0 is a call-clobbered 119 * register 120 */ 121 #define EXPORT_FUNC(f, a, x, ...) \ 122 asm volatile ( \ 123 " .globl " #x "\n" \ 124 #x ":\n" \ 125 " move.l %%d7, %%a0\n" \ 126 " adda.l %0, %%a0\n" \ 127 " move.l (%%a0), %%a0\n" \ 128 " adda.l %1, %%a0\n" \ 129 " move.l (%%a0), %%a0\n" \ 130 " jmp (%%a0)\n" \ 131 : : "i"(offsetof(gd_t, jt)), "i"(FO(x)) : "a0"); 132 #elif defined(CONFIG_MICROBLAZE) 133 /* 134 * r31 holds the pointer to the global_data. r5 is a call-clobbered. 135 */ 136 #define EXPORT_FUNC(f, a, x, ...) \ 137 asm volatile ( \ 138 " .globl " #x "\n" \ 139 #x ":\n" \ 140 " lwi r5, r31, %0\n" \ 141 " lwi r5, r5, %1\n" \ 142 " bra r5\n" \ 143 : : "i"(offsetof(gd_t, jt)), "i"(FO(x)) : "r5"); 144 #elif defined(CONFIG_AVR32) 145 /* 146 * r6 holds the pointer to the global_data. r8 is call clobbered. 147 */ 148 #define EXPORT_FUNC(f, a, x, ...) \ 149 asm volatile( \ 150 " .globl\t" #x "\n" \ 151 #x ":\n" \ 152 " ld.w r8, r6[%0]\n" \ 153 " ld.w pc, r8[%1]\n" \ 154 : \ 155 : "i"(offsetof(gd_t, jt)), "i"(FO(x)) \ 156 : "r8"); 157 #elif defined(CONFIG_SH) 158 /* 159 * r13 holds the pointer to the global_data. r1 is a call clobbered. 160 */ 161 #define EXPORT_FUNC(f, a, x, ...) \ 162 asm volatile ( \ 163 " .align 2\n" \ 164 " .globl " #x "\n" \ 165 #x ":\n" \ 166 " mov r13, r1\n" \ 167 " add %0, r1\n" \ 168 " mov.l @r1, r2\n" \ 169 " add %1, r2\n" \ 170 " mov.l @r2, r1\n" \ 171 " jmp @r1\n" \ 172 " nop\n" \ 173 " nop\n" \ 174 : : "i"(offsetof(gd_t, jt)), "i"(FO(x)) : "r1", "r2"); 175 #elif defined(CONFIG_NDS32) 176 /* 177 * r16 holds the pointer to the global_data. gp is call clobbered. 178 * not support reduced register (16 GPR). 179 */ 180 #define EXPORT_FUNC(f, a, x, ...) \ 181 asm volatile ( \ 182 " .globl " #x "\n" \ 183 #x ":\n" \ 184 " lwi $r16, [$gp + (%0)]\n" \ 185 " lwi $r16, [$r16 + (%1)]\n" \ 186 " jr $r16\n" \ 187 : : "i"(offsetof(gd_t, jt)), "i"(FO(x)) : "$r16"); 188 #elif defined(CONFIG_ARC) 189 /* 190 * r25 holds the pointer to the global_data. r10 is call clobbered. 191 */ 192 #define EXPORT_FUNC(f, a, x, ...) \ 193 asm volatile( \ 194 " .align 4\n" \ 195 " .globl " #x "\n" \ 196 #x ":\n" \ 197 " ld r10, [r25, %0]\n" \ 198 " ld r10, [r10, %1]\n" \ 199 " j [r10]\n" \ 200 : : "i"(offsetof(gd_t, jt)), "i"(FO(x)) : "r10"); 201 #elif defined(CONFIG_XTENSA) 202 /* 203 * Global data ptr is in global_data, jump table ptr is in jt. 204 * Windowed ABI: Jump just past 'entry' in target and adjust stack frame 205 * (extract stack frame size from target 'entry' instruction). 206 */ 207 208 static void **jt; 209 210 #if defined(__XTENSA_CALL0_ABI__) 211 #define EXPORT_FUNC(f, a, x, ...) \ 212 asm volatile ( \ 213 " .extern jt\n" \ 214 " .globl " #x "\n" \ 215 " .align 4\n" \ 216 #x ":\n" \ 217 " l32i a8, %0, 0\n" \ 218 " l32i a8, a8, %1\n" \ 219 " jx a8\n" \ 220 : : "r"(jt), "i" (FO(x)) : "a8"); 221 #elif defined(__XTENSA_WINDOWED_ABI__) 222 #if XCHAL_HAVE_BE 223 # define SFT "8" 224 #else 225 # define SFT "12" 226 #endif 227 #define EXPORT_FUNC(f, a, x, ...) \ 228 asm volatile ( \ 229 " .extern jt\n" \ 230 " .globl " #x "\n" \ 231 " .align 4\n" \ 232 #x ":\n" \ 233 " entry sp, 16\n" \ 234 " l32i a8, %0, 0\n" \ 235 " l32i a8, a8, %1\n" \ 236 " l32i a9, a8, 0\n" \ 237 " extui a9, a9, " SFT ", 12\n" \ 238 " subx8 a9, a9, sp\n" \ 239 " movi a10, 16\n" \ 240 " sub a9, a10, a9\n" \ 241 " movsp sp, a9\n" \ 242 " addi a8, a8, 3\n" \ 243 " jx a8\n" \ 244 : : "r"(jt), "i" (FO(x)) : "a8", "a9", "a10"); 245 #else 246 #error Unsupported Xtensa ABI 247 #endif 248 #else 249 /*" addi $sp, $sp, -24\n" \ 250 " br $r16\n" \*/ 251 252 #error stubs definition missing for this architecture 253 #endif 254 255 /* This function is necessary to prevent the compiler from 256 * generating prologue/epilogue, preparing stack frame etc. 257 * The stub functions are special, they do not use the stack 258 * frame passed to them, but pass it intact to the actual 259 * implementation. On the other hand, asm() statements with 260 * arguments can be used only inside the functions (gcc limitation) 261 */ 262 #if GCC_VERSION < 30400 263 static 264 #endif /* GCC_VERSION */ 265 void __attribute__((unused)) dummy(void) 266 { 267 #include <_exports.h> 268 } 269 270 #include <asm/sections.h> 271 272 void app_startup(char * const *argv) 273 { 274 char *cp = __bss_start; 275 276 /* Zero out BSS */ 277 while (cp < _end) 278 *cp++ = 0; 279 280 #if defined(CONFIG_X86) 281 /* x86 does not have a dedicated register for passing global_data */ 282 global_data = (gd_t *)argv[-1]; 283 jt = global_data->jt; 284 #endif 285 } 286 287 #undef EXPORT_FUNC 288