1 /*- 2 * Copyright 1996-1998 John D. Polstra. 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 15 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 16 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 17 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 18 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 19 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 20 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 21 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 22 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 23 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 24 */ 25 26 #include <sys/cdefs.h> 27 __FBSDID("$FreeBSD$"); 28 29 #include <sys/param.h> 30 #include <sys/kernel.h> 31 #include <sys/systm.h> 32 #include <sys/exec.h> 33 #include <sys/imgact.h> 34 #include <sys/linker.h> 35 #include <sys/sysent.h> 36 #include <sys/imgact_elf.h> 37 #include <sys/proc.h> 38 #include <sys/syscall.h> 39 #include <sys/signalvar.h> 40 #include <sys/vnode.h> 41 42 #include <vm/vm.h> 43 #include <vm/pmap.h> 44 #include <vm/vm_param.h> 45 46 #include <machine/elf.h> 47 #include <machine/md_var.h> 48 49 static boolean_t elf32_arm_abi_supported(struct image_params *); 50 51 struct sysentvec elf32_freebsd_sysvec = { 52 .sv_size = SYS_MAXSYSCALL, 53 .sv_table = sysent, 54 .sv_mask = 0, 55 .sv_errsize = 0, 56 .sv_errtbl = NULL, 57 .sv_transtrap = NULL, 58 .sv_fixup = __elfN(freebsd_fixup), 59 .sv_sendsig = sendsig, 60 .sv_sigcode = sigcode, 61 .sv_szsigcode = &szsigcode, 62 .sv_name = "FreeBSD ELF32", 63 .sv_coredump = __elfN(coredump), 64 .sv_imgact_try = NULL, 65 .sv_minsigstksz = MINSIGSTKSZ, 66 .sv_pagesize = PAGE_SIZE, 67 .sv_minuser = VM_MIN_ADDRESS, 68 .sv_maxuser = VM_MAXUSER_ADDRESS, 69 .sv_usrstack = USRSTACK, 70 .sv_psstrings = PS_STRINGS, 71 .sv_stackprot = VM_PROT_ALL, 72 .sv_copyout_strings = exec_copyout_strings, 73 .sv_setregs = exec_setregs, 74 .sv_fixlimit = NULL, 75 .sv_maxssiz = NULL, 76 .sv_flags = 77 #if __ARM_ARCH >= 6 78 SV_SHP | SV_TIMEKEEP | 79 #endif 80 SV_ABI_FREEBSD | SV_ILP32, 81 .sv_set_syscall_retval = cpu_set_syscall_retval, 82 .sv_fetch_syscall_args = cpu_fetch_syscall_args, 83 .sv_syscallnames = syscallnames, 84 .sv_shared_page_base = SHAREDPAGE, 85 .sv_shared_page_len = PAGE_SIZE, 86 .sv_schedtail = NULL, 87 .sv_thread_detach = NULL, 88 .sv_trap = NULL, 89 }; 90 INIT_SYSENTVEC(elf32_sysvec, &elf32_freebsd_sysvec); 91 92 static Elf32_Brandinfo freebsd_brand_info = { 93 .brand = ELFOSABI_FREEBSD, 94 .machine = EM_ARM, 95 .compat_3_brand = "FreeBSD", 96 .emul_path = NULL, 97 .interp_path = "/libexec/ld-elf.so.1", 98 .sysvec = &elf32_freebsd_sysvec, 99 .interp_newpath = NULL, 100 .brand_note = &elf32_freebsd_brandnote, 101 .flags = BI_CAN_EXEC_DYN | BI_BRAND_NOTE, 102 .header_supported= elf32_arm_abi_supported, 103 }; 104 105 SYSINIT(elf32, SI_SUB_EXEC, SI_ORDER_FIRST, 106 (sysinit_cfunc_t) elf32_insert_brand_entry, 107 &freebsd_brand_info); 108 109 static boolean_t 110 elf32_arm_abi_supported(struct image_params *imgp) 111 { 112 const Elf_Ehdr *hdr = (const Elf_Ehdr *)imgp->image_header; 113 114 /* 115 * When configured for EABI, FreeBSD supports EABI vesions 4 and 5. 116 */ 117 if (EF_ARM_EABI_VERSION(hdr->e_flags) < EF_ARM_EABI_FREEBSD_MIN) { 118 if (bootverbose) 119 uprintf("Attempting to execute non EABI binary (rev %d) image %s", 120 EF_ARM_EABI_VERSION(hdr->e_flags), imgp->args->fname); 121 return (FALSE); 122 } 123 return (TRUE); 124 } 125 126 void 127 elf32_dump_thread(struct thread *td __unused, void *dst __unused, 128 size_t *off __unused) 129 { 130 } 131 132 /* 133 * It is possible for the compiler to emit relocations for unaligned data. 134 * We handle this situation with these inlines. 135 */ 136 #define RELOC_ALIGNED_P(x) \ 137 (((uintptr_t)(x) & (sizeof(void *) - 1)) == 0) 138 139 static __inline Elf_Addr 140 load_ptr(Elf_Addr *where) 141 { 142 Elf_Addr res; 143 144 if (RELOC_ALIGNED_P(where)) 145 return *where; 146 memcpy(&res, where, sizeof(res)); 147 return (res); 148 } 149 150 static __inline void 151 store_ptr(Elf_Addr *where, Elf_Addr val) 152 { 153 if (RELOC_ALIGNED_P(where)) 154 *where = val; 155 else 156 memcpy(where, &val, sizeof(val)); 157 } 158 #undef RELOC_ALIGNED_P 159 160 161 /* Process one elf relocation with addend. */ 162 static int 163 elf_reloc_internal(linker_file_t lf, Elf_Addr relocbase, const void *data, 164 int type, int local, elf_lookup_fn lookup) 165 { 166 Elf_Addr *where; 167 Elf_Addr addr; 168 Elf_Addr addend; 169 Elf_Word rtype, symidx; 170 const Elf_Rel *rel; 171 const Elf_Rela *rela; 172 int error; 173 174 switch (type) { 175 case ELF_RELOC_REL: 176 rel = (const Elf_Rel *)data; 177 where = (Elf_Addr *) (relocbase + rel->r_offset); 178 addend = load_ptr(where); 179 rtype = ELF_R_TYPE(rel->r_info); 180 symidx = ELF_R_SYM(rel->r_info); 181 break; 182 case ELF_RELOC_RELA: 183 rela = (const Elf_Rela *)data; 184 where = (Elf_Addr *) (relocbase + rela->r_offset); 185 addend = rela->r_addend; 186 rtype = ELF_R_TYPE(rela->r_info); 187 symidx = ELF_R_SYM(rela->r_info); 188 break; 189 default: 190 panic("unknown reloc type %d\n", type); 191 } 192 193 if (local) { 194 if (rtype == R_ARM_RELATIVE) { /* A + B */ 195 addr = elf_relocaddr(lf, relocbase + addend); 196 if (load_ptr(where) != addr) 197 store_ptr(where, addr); 198 } 199 return (0); 200 } 201 202 switch (rtype) { 203 204 case R_ARM_NONE: /* none */ 205 break; 206 207 case R_ARM_ABS32: 208 error = lookup(lf, symidx, 1, &addr); 209 if (error != 0) 210 return -1; 211 store_ptr(where, addr + load_ptr(where)); 212 break; 213 214 case R_ARM_COPY: /* none */ 215 /* 216 * There shouldn't be copy relocations in kernel 217 * objects. 218 */ 219 printf("kldload: unexpected R_COPY relocation\n"); 220 return -1; 221 break; 222 223 case R_ARM_JUMP_SLOT: 224 error = lookup(lf, symidx, 1, &addr); 225 if (error == 0) { 226 store_ptr(where, addr); 227 return (0); 228 } 229 return (-1); 230 case R_ARM_RELATIVE: 231 break; 232 233 default: 234 printf("kldload: unexpected relocation type %d\n", 235 rtype); 236 return -1; 237 } 238 return(0); 239 } 240 241 int 242 elf_reloc(linker_file_t lf, Elf_Addr relocbase, const void *data, int type, 243 elf_lookup_fn lookup) 244 { 245 246 return (elf_reloc_internal(lf, relocbase, data, type, 0, lookup)); 247 } 248 249 int 250 elf_reloc_local(linker_file_t lf, Elf_Addr relocbase, const void *data, 251 int type, elf_lookup_fn lookup) 252 { 253 254 return (elf_reloc_internal(lf, relocbase, data, type, 1, lookup)); 255 } 256 257 int 258 elf_cpu_load_file(linker_file_t lf) 259 { 260 261 /* 262 * The pmap code does not do an icache sync upon establishing executable 263 * mappings in the kernel pmap. It's an optimization based on the fact 264 * that kernel memory allocations always have EXECUTABLE protection even 265 * when the memory isn't going to hold executable code. The only time 266 * kernel memory holding instructions does need a sync is after loading 267 * a kernel module, and that's when this function gets called. 268 * 269 * This syncs data and instruction caches after loading a module. We 270 * don't worry about the kernel itself (lf->id is 1) as locore.S did 271 * that on entry. Even if data cache maintenance was done by IO code, 272 * the relocation fixup process creates dirty cache entries that we must 273 * write back before doing icache sync. The instruction cache sync also 274 * invalidates the branch predictor cache on platforms that have one. 275 */ 276 if (lf->id == 1) 277 return (0); 278 #if __ARM_ARCH >= 6 279 dcache_wb_pou((vm_offset_t)lf->address, (vm_size_t)lf->size); 280 icache_inv_all(); 281 #else 282 cpu_dcache_wb_range((vm_offset_t)lf->address, (vm_size_t)lf->size); 283 cpu_l2cache_wb_range((vm_offset_t)lf->address, (vm_size_t)lf->size); 284 cpu_icache_sync_range((vm_offset_t)lf->address, (vm_size_t)lf->size); 285 #endif 286 return (0); 287 } 288 289 int 290 elf_cpu_unload_file(linker_file_t lf __unused) 291 { 292 293 return (0); 294 } 295