Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 | /* SPDX-License-Identifier: GPL-2.0 */ #include <asm/core.h> #include <asm/regs.h> #include <asm/asmmacro.h> #include <asm/cacheasm.h> #include <asm/processor.h> /* * RB-Data: RedBoot data/bss * P: Boot-Parameters * L: Kernel-Loader * * The Linux-Kernel image including the loader must be loaded * to a position so that the kernel and the boot parameters * can fit in the space before the load address. * ______________________________________________________ * |_RB-Data_|_P_|__________|_L_|___Linux-Kernel___|______| * ^ * ^ Load address * ______________________________________________________ * |___Linux-Kernel___|_P_|_L_|___________________________| * * The loader copies the parameter to the position that will * be the end of the kernel and itself to the end of the * parameter list. */ /* Make sure we have enough space for the 'uncompressor' */ #define STACK_SIZE 32768 #define HEAP_SIZE (131072*4) # a2: Parameter list # a3: Size of parameter list .section .start, "ax" .globl __start /* this must be the first byte of the loader! */ __start: abi_entry(32) # we do not intend to return _call0 _start __start_a0: .align 4 .section .text, "ax" .literal_position .begin literal_prefix .text /* put literals in here! */ .globl _start _start: /* 'reset' window registers */ movi a4, 1 wsr a4, ps rsync #if XCHAL_HAVE_WINDOWED rsr a5, windowbase ssl a5 sll a4, a4 wsr a4, windowstart rsync #endif movi a4, KERNEL_PS_WOE_MASK wsr a4, ps rsync KABI_C0 mov abi_saved0, abi_arg0 /* copy the loader to its address * Note: The loader itself is a very small piece, so we assume we * don't partially overlap. We also assume (even more important) * that the kernel image is out of the way. Usually, when the * load address of this image is not at an arbitrary address, * but aligned to some 10K's we shouldn't overlap. */ /* Note: The assembler cannot relax "addi a0, a0, ..." to an l32r, so we load to a4 first. */ # addi a4, a0, __start - __start_a0 # mov a0, a4 movi a4, __start movi a5, __start_a0 add a4, a0, a4 sub a0, a4, a5 movi a4, __start movi a5, __reloc_end # a0: address where this code has been loaded # a4: compiled address of __start # a5: compiled end address mov.n a7, a0 mov.n a8, a4 1: l32i a10, a7, 0 l32i a11, a7, 4 s32i a10, a8, 0 s32i a11, a8, 4 l32i a10, a7, 8 l32i a11, a7, 12 s32i a10, a8, 8 s32i a11, a8, 12 addi a8, a8, 16 addi a7, a7, 16 blt a8, a5, 1b /* We have to flush and invalidate the caches here before we jump. */ #if XCHAL_DCACHE_IS_WRITEBACK ___flush_dcache_all a5 a6 #endif ___invalidate_icache_all a5 a6 isync movi a11, _reloc jx a11 .globl _reloc _reloc: /* RedBoot is now at the end of the memory, so we don't have * to copy the parameter list. Keep the code around; in case * we need it again. */ #if 0 # a0: load address # a2: start address of parameter list # a3: length of parameter list # a4: __start /* copy the parameter list out of the way */ movi a6, _param_start add a3, a2, a3 2: l32i a8, a2, 0 s32i a8, a6, 0 addi a2, a2, 4 addi a6, a6, 4 blt a2, a3, 2b #endif /* clear BSS section */ movi a6, __bss_start movi a7, __bss_end movi.n a5, 0 3: s32i a5, a6, 0 addi a6, a6, 4 blt a6, a7, 3b movi a5, -16 movi a1, _stack + STACK_SIZE and a1, a1, a5 /* Uncompress the kernel */ # a0: load address # a2: boot parameter # a4: __start movi a3, __image_load sub a4, a3, a4 add abi_arg2, a0, a4 # a1 Stack # a8(a4) Load address of the image movi abi_arg0, _image_start movi abi_arg4, _image_end movi abi_arg1, 0x1000000 sub abi_tmp0, abi_arg4, abi_arg0 movi abi_arg3, complen s32i abi_tmp0, abi_arg3, 0 movi a0, 0 # abi_arg0 destination # abi_arg1 maximum size of destination # abi_arg2 source # abi_arg3 ptr to length .extern gunzip movi abi_tmp0, gunzip beqz abi_tmp0, 1f abi_callx abi_tmp0 j 2f # abi_arg0 destination start # abi_arg1 maximum size of destination # abi_arg2 source start # abi_arg3 ptr to length # abi_arg4 destination end 1: l32i abi_tmp0, abi_arg2, 0 l32i abi_tmp1, abi_arg2, 4 s32i abi_tmp0, abi_arg0, 0 s32i abi_tmp1, abi_arg0, 4 l32i abi_tmp0, abi_arg2, 8 l32i abi_tmp1, abi_arg2, 12 s32i abi_tmp0, abi_arg0, 8 s32i abi_tmp1, abi_arg0, 12 addi abi_arg0, abi_arg0, 16 addi abi_arg2, abi_arg2, 16 blt abi_arg0, abi_arg4, 1b /* jump to the kernel */ 2: #if XCHAL_DCACHE_IS_WRITEBACK ___flush_dcache_all a5 a6 #endif ___invalidate_icache_all a5 a6 isync # a2 Boot parameter list KABI_C0 mov abi_arg0, abi_saved0 movi a0, _image_start jx a0 .align 16 .data .globl avail_ram avail_ram: .long _heap .globl end_avail end_avail: .long _heap + HEAP_SIZE .comm _stack, STACK_SIZE .comm _heap, HEAP_SIZE .globl end_avail .comm complen, 4 .end literal_prefix |