1/* SPDX-License-Identifier: GPL-2.0-only */ 2/* 3 * Early kernel startup code for Hexagon 4 * 5 * Copyright (c) 2010-2013, The Linux Foundation. All rights reserved. 6 */ 7 8#include <linux/linkage.h> 9#include <linux/init.h> 10#include <asm/asm-offsets.h> 11#include <asm/mem-layout.h> 12#include <asm/vm_mmu.h> 13#include <asm/page.h> 14#include <asm/hexagon_vm.h> 15 16#define SEGTABLE_ENTRIES #0x0e0 17 18 __INIT 19ENTRY(stext) 20 /* 21 * VMM will already have set up true vector page, MMU, etc. 22 * To set up initial kernel identity map, we have to pass 23 * the VMM a pointer to some canonical page tables. In 24 * this implementation, we're assuming that we've got 25 * them precompiled. Generate value in R24, as we'll need 26 * it again shortly. 27 */ 28 r24.L = #LO(swapper_pg_dir) 29 r24.H = #HI(swapper_pg_dir) 30 31 /* 32 * Symbol is kernel segment address, but we need 33 * the logical/physical address. 34 */ 35 r25 = pc; 36 r2.h = #0xffc0; 37 r2.l = #0x0000; 38 r25 = and(r2,r25); /* R25 holds PHYS_OFFSET now */ 39 r1.h = #HI(PAGE_OFFSET); 40 r1.l = #LO(PAGE_OFFSET); 41 r24 = sub(r24,r1); /* swapper_pg_dir - PAGE_OFFSET */ 42 r24 = add(r24,r25); /* + PHYS_OFFSET */ 43 44 r0 = r24; /* aka __pa(swapper_pg_dir) */ 45 46 /* 47 * Initialize page dir to make the virtual and physical 48 * addresses where the kernel was loaded be identical. 49 * Done in 4MB chunks. 50 */ 51#define PTE_BITS ( __HVM_PTE_R | __HVM_PTE_W | __HVM_PTE_X \ 52 | __HEXAGON_C_WB_L2 << 6 \ 53 | __HVM_PDE_S_4MB) 54 55 /* 56 * Get number of VA=PA entries; only really needed for jump 57 * to hyperspace; gets blown away immediately after 58 */ 59 60 { 61 r1.l = #LO(_end); 62 r2.l = #LO(stext); 63 r3 = #1; 64 } 65 { 66 r1.h = #HI(_end); 67 r2.h = #HI(stext); 68 r3 = asl(r3, #22); 69 } 70 { 71 r1 = sub(r1, r2); 72 r3 = add(r3, #-1); 73 } /* r1 = _end - stext */ 74 r1 = add(r1, r3); /* + (4M-1) */ 75 r26 = lsr(r1, #22); /* / 4M = # of entries */ 76 77 r1 = r25; 78 r2.h = #0xffc0; 79 r2.l = #0x0000; /* round back down to 4MB boundary */ 80 r1 = and(r1,r2); 81 r2 = lsr(r1, #22) /* 4MB page number */ 82 r2 = asl(r2, #2) /* times sizeof(PTE) (4bytes) */ 83 r0 = add(r0,r2) /* r0 = address of correct PTE */ 84 r2 = #PTE_BITS 85 r1 = add(r1,r2) /* r1 = 4MB PTE for the first entry */ 86 r2.h = #0x0040 87 r2.l = #0x0000 /* 4MB increments */ 88 loop0(1f,r26); 891: 90 memw(r0 ++ #4) = r1 91 { r1 = add(r1, r2); } :endloop0 92 93 /* Also need to overwrite the initial 0xc0000000 entries */ 94 /* PAGE_OFFSET >> (4MB shift - 4 bytes per entry shift) */ 95 R1.H = #HI(PAGE_OFFSET >> (22 - 2)) 96 R1.L = #LO(PAGE_OFFSET >> (22 - 2)) 97 98 r0 = add(r1, r24); /* advance to 0xc0000000 entry */ 99 r1 = r25; 100 r2.h = #0xffc0; 101 r2.l = #0x0000; /* round back down to 4MB boundary */ 102 r1 = and(r1,r2); /* for huge page */ 103 r2 = #PTE_BITS 104 r1 = add(r1,r2); 105 r2.h = #0x0040 106 r2.l = #0x0000 /* 4MB increments */ 107 108 loop0(1f,SEGTABLE_ENTRIES); 1091: 110 memw(r0 ++ #4) = r1; 111 { r1 = add(r1,r2); } :endloop0 112 113 r0 = r24; 114 115 /* 116 * The subroutine wrapper around the virtual instruction touches 117 * no memory, so we should be able to use it even here. 118 * Note that in this version, R1 and R2 get "clobbered"; see 119 * vm_ops.S 120 */ 121 r1 = #VM_TRANS_TYPE_TABLE 122 call __vmnewmap; 123 124 /* Jump into virtual address range. */ 125 126 r31.h = #hi(__head_s_vaddr_target) 127 r31.l = #lo(__head_s_vaddr_target) 128 jumpr r31 129 130 /* Insert trippy space effects. */ 131 132__head_s_vaddr_target: 133 /* 134 * Tear down VA=PA translation now that we are running 135 * in kernel virtual space. 136 */ 137 r0 = #__HVM_PDE_S_INVALID 138 139 r1.h = #0xffc0; 140 r1.l = #0x0000; 141 r2 = r25; /* phys_offset */ 142 r2 = and(r1,r2); 143 144 r1.l = #lo(swapper_pg_dir) 145 r1.h = #hi(swapper_pg_dir) 146 r2 = lsr(r2, #22) /* 4MB page number */ 147 r2 = asl(r2, #2) /* times sizeof(PTE) (4bytes) */ 148 r1 = add(r1,r2); 149 loop0(1f,r26) 150 1511: 152 { 153 memw(R1 ++ #4) = R0 154 }:endloop0 155 156 r0 = r24 157 r1 = #VM_TRANS_TYPE_TABLE 158 call __vmnewmap 159 160 /* Go ahead and install the trap0 return so angel calls work */ 161 r0.h = #hi(_K_provisional_vec) 162 r0.l = #lo(_K_provisional_vec) 163 call __vmsetvec 164 165 /* 166 * OK, at this point we should start to be much more careful, 167 * we're going to enter C code and start touching memory 168 * in all sorts of places. 169 * This means: 170 * SGP needs to be OK 171 * Need to lock shared resources 172 * A bunch of other things that will cause 173 * all kinds of painful bugs 174 */ 175 176 /* 177 * Stack pointer should be pointed at the init task's 178 * thread stack, which should have been declared in arch/init_task.c. 179 * So uhhhhh... 180 * It's accessible via the init_thread_union, which is a union 181 * of a thread_info struct and a stack; of course, the top 182 * of the stack is not for you. The end of the stack 183 * is simply init_thread_union + THREAD_SIZE. 184 */ 185 186 {r29.H = #HI(init_thread_union); r0.H = #HI(_THREAD_SIZE); } 187 {r29.L = #LO(init_thread_union); r0.L = #LO(_THREAD_SIZE); } 188 189 /* initialize the register used to point to current_thread_info */ 190 /* Fixme: THREADINFO_REG can't be R2 because of that memset thing. */ 191 {r29 = add(r29,r0); THREADINFO_REG = r29; } 192 193 /* Hack: zero bss; */ 194 { r0.L = #LO(__bss_start); r1 = #0; r2.l = #LO(__bss_stop); } 195 { r0.H = #HI(__bss_start); r2.h = #HI(__bss_stop); } 196 197 r2 = sub(r2,r0); 198 call memset; 199 200 /* Set PHYS_OFFSET; should be in R25 */ 201#ifdef CONFIG_HEXAGON_PHYS_OFFSET 202 r0.l = #LO(__phys_offset); 203 r0.h = #HI(__phys_offset); 204 memw(r0) = r25; 205#endif 206 207 /* Time to make the doughnuts. */ 208 call start_kernel 209 210 /* 211 * Should not reach here. 212 */ 2131: 214 jump 1b 215 216.p2align PAGE_SHIFT 217ENTRY(external_cmdline_buffer) 218 .fill _PAGE_SIZE,1,0 219 220.data 221.p2align PAGE_SHIFT 222ENTRY(empty_zero_page) 223 .fill _PAGE_SIZE,1,0 224