1 /* $NetBSD: exec_elf.c,v 1.27 2010/10/07 02:14:02 chs Exp $ */ 2 3 /*- 4 * Copyright (c) 1994, 2000, 2005 The NetBSD Foundation, Inc. 5 * All rights reserved. 6 * 7 * This code is derived from software contributed to The NetBSD Foundation 8 * by Christos Zoulas. 9 * 10 * Redistribution and use in source and binary forms, with or without 11 * modification, are permitted provided that the following conditions 12 * are met: 13 * 1. Redistributions of source code must retain the above copyright 14 * notice, this list of conditions and the following disclaimer. 15 * 2. Redistributions in binary form must reproduce the above copyright 16 * notice, this list of conditions and the following disclaimer in the 17 * documentation and/or other materials provided with the distribution. 18 * 19 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS 20 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED 21 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 22 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS 23 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 24 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 25 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 26 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 27 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 28 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 29 * POSSIBILITY OF SUCH DAMAGE. 30 */ 31 32 /* 33 * Copyright (c) 1996 Christopher G. Demetriou 34 * All rights reserved. 35 * 36 * Redistribution and use in source and binary forms, with or without 37 * modification, are permitted provided that the following conditions 38 * are met: 39 * 1. Redistributions of source code must retain the above copyright 40 * notice, this list of conditions and the following disclaimer. 41 * 2. Redistributions in binary form must reproduce the above copyright 42 * notice, this list of conditions and the following disclaimer in the 43 * documentation and/or other materials provided with the distribution. 44 * 3. The name of the author may not be used to endorse or promote products 45 * derived from this software without specific prior written permission 46 * 47 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 48 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 49 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 50 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 51 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 52 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 53 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 54 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 55 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 56 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 57 */ 58 59 #include <sys/cdefs.h> 60 __KERNEL_RCSID(1, "$NetBSD: exec_elf.c,v 1.27 2010/10/07 02:14:02 chs Exp $"); 61 62 #ifdef _KERNEL_OPT 63 #include "opt_pax.h" 64 #endif /* _KERNEL_OPT */ 65 66 #include <sys/param.h> 67 #include <sys/proc.h> 68 #include <sys/malloc.h> 69 #include <sys/kmem.h> 70 #include <sys/namei.h> 71 #include <sys/vnode.h> 72 #include <sys/exec.h> 73 #include <sys/exec_elf.h> 74 #include <sys/syscall.h> 75 #include <sys/signalvar.h> 76 #include <sys/mount.h> 77 #include <sys/stat.h> 78 #include <sys/kauth.h> 79 #include <sys/bitops.h> 80 81 #include <sys/cpu.h> 82 #include <machine/reg.h> 83 84 #include <compat/common/compat_util.h> 85 86 #include <sys/pax.h> 87 88 extern struct emul emul_netbsd; 89 90 #define elf_check_header ELFNAME(check_header) 91 #define elf_copyargs ELFNAME(copyargs) 92 #define elf_load_file ELFNAME(load_file) 93 #define elf_load_psection ELFNAME(load_psection) 94 #define exec_elf_makecmds ELFNAME2(exec,makecmds) 95 #define netbsd_elf_signature ELFNAME2(netbsd,signature) 96 #define netbsd_elf_probe ELFNAME2(netbsd,probe) 97 #define coredump ELFNAMEEND(coredump) 98 99 int elf_load_file(struct lwp *, struct exec_package *, char *, 100 struct exec_vmcmd_set *, u_long *, struct elf_args *, Elf_Addr *); 101 void elf_load_psection(struct exec_vmcmd_set *, struct vnode *, 102 const Elf_Phdr *, Elf_Addr *, u_long *, int *, int); 103 104 int netbsd_elf_signature(struct lwp *, struct exec_package *, Elf_Ehdr *); 105 int netbsd_elf_probe(struct lwp *, struct exec_package *, void *, char *, 106 vaddr_t *); 107 108 /* round up and down to page boundaries. */ 109 #define ELF_ROUND(a, b) (((a) + (b) - 1) & ~((b) - 1)) 110 #define ELF_TRUNC(a, b) ((a) & ~((b) - 1)) 111 112 /* 113 * Arbitrary limits to avoid DoS for excessive memory allocation. 114 */ 115 #define MAXPHNUM 128 116 #define MAXSHNUM 32768 117 #define MAXNOTESIZE 1024 118 119 static void 120 elf_placedynexec(struct lwp *l, struct exec_package *epp, Elf_Ehdr *eh, 121 Elf_Phdr *ph) 122 { 123 Elf_Addr align, offset; 124 int i; 125 126 for (align = i = 0; i < eh->e_phnum; i++) 127 if (ph[i].p_type == PT_LOAD && ph[i].p_align > align) 128 align = ph[i].p_align; 129 130 #ifdef PAX_ASLR 131 if (pax_aslr_active(l)) { 132 size_t pax_align, l2, delta; 133 uint32_t r; 134 135 pax_align = align; 136 137 r = arc4random(); 138 139 if (pax_align == 0) 140 pax_align = PGSHIFT; 141 l2 = ilog2(pax_align); 142 delta = PAX_ASLR_DELTA(r, l2, PAX_ASLR_DELTA_EXEC_LEN); 143 offset = ELF_TRUNC(delta, pax_align) + PAGE_SIZE; 144 #ifdef PAX_ASLR_DEBUG 145 uprintf("r=0x%x l2=0x%zx PGSHIFT=0x%x Delta=0x%zx\n", r, l2, 146 PGSHIFT, delta); 147 uprintf("pax offset=0x%llx entry=0x%llx\n", 148 (unsigned long long)offset, 149 (unsigned long long)eh->e_entry); 150 #endif /* PAX_ASLR_DEBUG */ 151 } else 152 #endif /* PAX_ASLR */ 153 offset = MAX(align, PAGE_SIZE); 154 155 for (i = 0; i < eh->e_phnum; i++) 156 ph[i].p_vaddr += offset; 157 eh->e_entry += offset; 158 } 159 160 /* 161 * Copy arguments onto the stack in the normal way, but add some 162 * extra information in case of dynamic binding. 163 */ 164 int 165 elf_copyargs(struct lwp *l, struct exec_package *pack, 166 struct ps_strings *arginfo, char **stackp, void *argp) 167 { 168 size_t len, vlen; 169 AuxInfo ai[ELF_AUX_ENTRIES], *a, *execname; 170 struct elf_args *ap; 171 int error; 172 173 if ((error = copyargs(l, pack, arginfo, stackp, argp)) != 0) 174 return error; 175 176 a = ai; 177 execname = NULL; 178 179 /* 180 * Push extra arguments on the stack needed by dynamically 181 * linked binaries 182 */ 183 if ((ap = (struct elf_args *)pack->ep_emul_arg)) { 184 struct vattr *vap = pack->ep_vap; 185 186 a->a_type = AT_PHDR; 187 a->a_v = ap->arg_phaddr; 188 a++; 189 190 a->a_type = AT_PHENT; 191 a->a_v = ap->arg_phentsize; 192 a++; 193 194 a->a_type = AT_PHNUM; 195 a->a_v = ap->arg_phnum; 196 a++; 197 198 a->a_type = AT_PAGESZ; 199 a->a_v = PAGE_SIZE; 200 a++; 201 202 a->a_type = AT_BASE; 203 a->a_v = ap->arg_interp; 204 a++; 205 206 a->a_type = AT_FLAGS; 207 a->a_v = 0; 208 a++; 209 210 a->a_type = AT_ENTRY; 211 a->a_v = ap->arg_entry; 212 a++; 213 214 a->a_type = AT_EUID; 215 if (vap->va_mode & S_ISUID) 216 a->a_v = vap->va_uid; 217 else 218 a->a_v = kauth_cred_geteuid(l->l_cred); 219 a++; 220 221 a->a_type = AT_RUID; 222 a->a_v = kauth_cred_getuid(l->l_cred); 223 a++; 224 225 a->a_type = AT_EGID; 226 if (vap->va_mode & S_ISGID) 227 a->a_v = vap->va_gid; 228 else 229 a->a_v = kauth_cred_getegid(l->l_cred); 230 a++; 231 232 a->a_type = AT_RGID; 233 a->a_v = kauth_cred_getgid(l->l_cred); 234 a++; 235 236 if (pack->ep_path) { 237 execname = a; 238 a->a_type = AT_SUN_EXECNAME; 239 a++; 240 } 241 242 free(ap, M_TEMP); 243 pack->ep_emul_arg = NULL; 244 } 245 246 a->a_type = AT_NULL; 247 a->a_v = 0; 248 a++; 249 250 vlen = (a - ai) * sizeof(AuxInfo); 251 252 if (execname) { 253 char *path = pack->ep_path; 254 execname->a_v = (uintptr_t)(*stackp + vlen); 255 len = strlen(path) + 1; 256 if ((error = copyout(path, (*stackp + vlen), len)) != 0) 257 return error; 258 len = ALIGN(len); 259 } else 260 len = 0; 261 262 if ((error = copyout(ai, *stackp, vlen)) != 0) 263 return error; 264 *stackp += vlen + len; 265 266 return 0; 267 } 268 269 /* 270 * elf_check_header(): 271 * 272 * Check header for validity; return 0 of ok ENOEXEC if error 273 */ 274 int 275 elf_check_header(Elf_Ehdr *eh, int type) 276 { 277 278 if (memcmp(eh->e_ident, ELFMAG, SELFMAG) != 0 || 279 eh->e_ident[EI_CLASS] != ELFCLASS) 280 return ENOEXEC; 281 282 switch (eh->e_machine) { 283 284 ELFDEFNNAME(MACHDEP_ID_CASES) 285 286 default: 287 return ENOEXEC; 288 } 289 290 if (ELF_EHDR_FLAGS_OK(eh) == 0) 291 return ENOEXEC; 292 293 if (eh->e_type != type) 294 return ENOEXEC; 295 296 if (eh->e_shnum > MAXSHNUM || eh->e_phnum > MAXPHNUM) 297 return ENOEXEC; 298 299 return 0; 300 } 301 302 /* 303 * elf_load_psection(): 304 * 305 * Load a psection at the appropriate address 306 */ 307 void 308 elf_load_psection(struct exec_vmcmd_set *vcset, struct vnode *vp, 309 const Elf_Phdr *ph, Elf_Addr *addr, u_long *size, int *prot, int flags) 310 { 311 u_long msize, psize, rm, rf; 312 long diff, offset; 313 314 /* 315 * If the user specified an address, then we load there. 316 */ 317 if (*addr == ELFDEFNNAME(NO_ADDR)) 318 *addr = ph->p_vaddr; 319 320 if (ph->p_align > 1) { 321 /* 322 * Make sure we are virtually aligned as we are supposed to be. 323 */ 324 diff = ph->p_vaddr - ELF_TRUNC(ph->p_vaddr, ph->p_align); 325 KASSERT(*addr - diff == ELF_TRUNC(*addr, ph->p_align)); 326 /* 327 * But make sure to not map any pages before the start of the 328 * psection by limiting the difference to within a page. 329 */ 330 diff &= PAGE_MASK; 331 } else 332 diff = 0; 333 334 *prot |= (ph->p_flags & PF_R) ? VM_PROT_READ : 0; 335 *prot |= (ph->p_flags & PF_W) ? VM_PROT_WRITE : 0; 336 *prot |= (ph->p_flags & PF_X) ? VM_PROT_EXECUTE : 0; 337 338 /* 339 * Adjust everything so it all starts on a page boundary. 340 */ 341 *addr -= diff; 342 offset = ph->p_offset - diff; 343 *size = ph->p_filesz + diff; 344 msize = ph->p_memsz + diff; 345 346 if (ph->p_align >= PAGE_SIZE) { 347 if ((ph->p_flags & PF_W) != 0) { 348 /* 349 * Because the pagedvn pager can't handle zero fill 350 * of the last data page if it's not page aligned we 351 * map the last page readvn. 352 */ 353 psize = trunc_page(*size); 354 } else { 355 psize = round_page(*size); 356 } 357 } else { 358 psize = *size; 359 } 360 361 if (psize > 0) { 362 NEW_VMCMD2(vcset, ph->p_align < PAGE_SIZE ? 363 vmcmd_map_readvn : vmcmd_map_pagedvn, psize, *addr, vp, 364 offset, *prot, flags); 365 flags &= VMCMD_RELATIVE; 366 } 367 if (psize < *size) { 368 NEW_VMCMD2(vcset, vmcmd_map_readvn, *size - psize, 369 *addr + psize, vp, offset + psize, *prot, flags); 370 } 371 372 /* 373 * Check if we need to extend the size of the segment (does 374 * bss extend page the next page boundary)? 375 */ 376 rm = round_page(*addr + msize); 377 rf = round_page(*addr + *size); 378 379 if (rm != rf) { 380 NEW_VMCMD2(vcset, vmcmd_map_zero, rm - rf, rf, NULLVP, 381 0, *prot, flags & VMCMD_RELATIVE); 382 *size = msize; 383 } 384 } 385 386 /* 387 * elf_load_file(): 388 * 389 * Load a file (interpreter/library) pointed to by path 390 * [stolen from coff_load_shlib()]. Made slightly generic 391 * so it might be used externally. 392 */ 393 int 394 elf_load_file(struct lwp *l, struct exec_package *epp, char *path, 395 struct exec_vmcmd_set *vcset, u_long *entryoff, struct elf_args *ap, 396 Elf_Addr *last) 397 { 398 int error, i; 399 struct vnode *vp; 400 struct vattr attr; 401 Elf_Ehdr eh; 402 Elf_Phdr *ph = NULL; 403 const Elf_Phdr *ph0; 404 const Elf_Phdr *base_ph; 405 const Elf_Phdr *last_ph; 406 u_long phsize; 407 Elf_Addr addr = *last; 408 struct proc *p; 409 410 p = l->l_proc; 411 412 /* 413 * 1. open file 414 * 2. read filehdr 415 * 3. map text, data, and bss out of it using VM_* 416 */ 417 vp = epp->ep_interp; 418 if (vp == NULL) { 419 error = emul_find_interp(l, epp, path); 420 if (error != 0) 421 return error; 422 vp = epp->ep_interp; 423 } 424 /* We'll tidy this ourselves - otherwise we have locking issues */ 425 epp->ep_interp = NULL; 426 vn_lock(vp, LK_EXCLUSIVE | LK_RETRY); 427 428 /* 429 * Similarly, if it's not marked as executable, or it's not a regular 430 * file, we don't allow it to be used. 431 */ 432 if (vp->v_type != VREG) { 433 error = EACCES; 434 goto badunlock; 435 } 436 if ((error = VOP_ACCESS(vp, VEXEC, l->l_cred)) != 0) 437 goto badunlock; 438 439 /* get attributes */ 440 if ((error = VOP_GETATTR(vp, &attr, l->l_cred)) != 0) 441 goto badunlock; 442 443 /* 444 * Check mount point. Though we're not trying to exec this binary, 445 * we will be executing code from it, so if the mount point 446 * disallows execution or set-id-ness, we punt or kill the set-id. 447 */ 448 if (vp->v_mount->mnt_flag & MNT_NOEXEC) { 449 error = EACCES; 450 goto badunlock; 451 } 452 if (vp->v_mount->mnt_flag & MNT_NOSUID) 453 epp->ep_vap->va_mode &= ~(S_ISUID | S_ISGID); 454 455 #ifdef notyet /* XXX cgd 960926 */ 456 XXX cgd 960926: (maybe) VOP_OPEN it (and VOP_CLOSE in copyargs?) 457 #endif 458 459 error = vn_marktext(vp); 460 if (error) 461 goto badunlock; 462 463 VOP_UNLOCK(vp); 464 465 if ((error = exec_read_from(l, vp, 0, &eh, sizeof(eh))) != 0) 466 goto bad; 467 468 if ((error = elf_check_header(&eh, ET_DYN)) != 0) 469 goto bad; 470 471 if (eh.e_phnum > MAXPHNUM || eh.e_phnum == 0) { 472 error = ENOEXEC; 473 goto bad; 474 } 475 476 phsize = eh.e_phnum * sizeof(Elf_Phdr); 477 ph = kmem_alloc(phsize, KM_SLEEP); 478 479 if ((error = exec_read_from(l, vp, eh.e_phoff, ph, phsize)) != 0) 480 goto bad; 481 482 #ifdef ELF_INTERP_NON_RELOCATABLE 483 /* 484 * Evil hack: Only MIPS should be non-relocatable, and the 485 * psections should have a high address (typically 0x5ffe0000). 486 * If it's now relocatable, it should be linked at 0 and the 487 * psections should have zeros in the upper part of the address. 488 * Otherwise, force the load at the linked address. 489 */ 490 if (*last == ELF_LINK_ADDR && (ph->p_vaddr & 0xffff0000) == 0) 491 *last = ELFDEFNNAME(NO_ADDR); 492 #endif 493 494 /* 495 * If no position to load the interpreter was set by a probe 496 * function, pick the same address that a non-fixed mmap(0, ..) 497 * would (i.e. something safely out of the way). 498 */ 499 if (*last == ELFDEFNNAME(NO_ADDR)) { 500 u_long limit = 0; 501 /* 502 * Find the start and ending addresses of the psections to 503 * be loaded. This will give us the size. 504 */ 505 for (i = 0, ph0 = ph, base_ph = NULL; i < eh.e_phnum; 506 i++, ph0++) { 507 if (ph0->p_type == PT_LOAD) { 508 u_long psize = ph0->p_vaddr + ph0->p_memsz; 509 if (base_ph == NULL) 510 base_ph = ph0; 511 if (psize > limit) 512 limit = psize; 513 } 514 } 515 516 if (base_ph == NULL) { 517 error = ENOEXEC; 518 goto bad; 519 } 520 521 /* 522 * Now compute the size and load address. 523 */ 524 addr = (*epp->ep_esch->es_emul->e_vm_default_addr)(p, 525 epp->ep_daddr, 526 round_page(limit) - trunc_page(base_ph->p_vaddr)); 527 } else 528 addr = *last; /* may be ELF_LINK_ADDR */ 529 530 /* 531 * Load all the necessary sections 532 */ 533 for (i = 0, ph0 = ph, base_ph = NULL, last_ph = NULL; 534 i < eh.e_phnum; i++, ph0++) { 535 switch (ph0->p_type) { 536 case PT_LOAD: { 537 u_long size; 538 int prot = 0; 539 int flags; 540 541 if (base_ph == NULL) { 542 /* 543 * First encountered psection is always the 544 * base psection. Make sure it's aligned 545 * properly (align down for topdown and align 546 * upwards for not topdown). 547 */ 548 base_ph = ph0; 549 flags = VMCMD_BASE; 550 if (addr == ELF_LINK_ADDR) 551 addr = ph0->p_vaddr; 552 if (p->p_vmspace->vm_map.flags & VM_MAP_TOPDOWN) 553 addr = ELF_TRUNC(addr, ph0->p_align); 554 else 555 addr = ELF_ROUND(addr, ph0->p_align); 556 } else { 557 u_long limit = round_page(last_ph->p_vaddr 558 + last_ph->p_memsz); 559 u_long base = trunc_page(ph0->p_vaddr); 560 561 /* 562 * If there is a gap in between the psections, 563 * map it as inaccessible so nothing else 564 * mmap'ed will be placed there. 565 */ 566 if (limit != base) { 567 NEW_VMCMD2(vcset, vmcmd_map_zero, 568 base - limit, 569 limit - base_ph->p_vaddr, NULLVP, 570 0, VM_PROT_NONE, VMCMD_RELATIVE); 571 } 572 573 addr = ph0->p_vaddr - base_ph->p_vaddr; 574 flags = VMCMD_RELATIVE; 575 } 576 last_ph = ph0; 577 elf_load_psection(vcset, vp, &ph[i], &addr, 578 &size, &prot, flags); 579 /* 580 * If entry is within this psection then this 581 * must contain the .text section. *entryoff is 582 * relative to the base psection. 583 */ 584 if (eh.e_entry >= ph0->p_vaddr && 585 eh.e_entry < (ph0->p_vaddr + size)) { 586 *entryoff = eh.e_entry - base_ph->p_vaddr; 587 } 588 addr += size; 589 break; 590 } 591 592 case PT_DYNAMIC: 593 case PT_PHDR: 594 break; 595 596 case PT_NOTE: 597 break; 598 599 default: 600 break; 601 } 602 } 603 604 kmem_free(ph, phsize); 605 /* 606 * This value is ignored if TOPDOWN. 607 */ 608 *last = addr; 609 vrele(vp); 610 return 0; 611 612 badunlock: 613 VOP_UNLOCK(vp); 614 615 bad: 616 if (ph != NULL) 617 kmem_free(ph, phsize); 618 #ifdef notyet /* XXX cgd 960926 */ 619 (maybe) VOP_CLOSE it 620 #endif 621 vrele(vp); 622 return error; 623 } 624 625 /* 626 * exec_elf_makecmds(): Prepare an Elf binary's exec package 627 * 628 * First, set of the various offsets/lengths in the exec package. 629 * 630 * Then, mark the text image busy (so it can be demand paged) or error 631 * out if this is not possible. Finally, set up vmcmds for the 632 * text, data, bss, and stack segments. 633 */ 634 int 635 exec_elf_makecmds(struct lwp *l, struct exec_package *epp) 636 { 637 Elf_Ehdr *eh = epp->ep_hdr; 638 Elf_Phdr *ph, *pp; 639 Elf_Addr phdr = 0, computed_phdr = 0, pos = 0, end_text = 0; 640 int error, i, nload; 641 char *interp = NULL; 642 u_long phsize; 643 struct proc *p; 644 struct elf_args *ap = NULL; 645 bool is_dyn; 646 647 if (epp->ep_hdrvalid < sizeof(Elf_Ehdr)) 648 return ENOEXEC; 649 650 is_dyn = elf_check_header(eh, ET_DYN) == 0; 651 /* 652 * XXX allow for executing shared objects. It seems silly 653 * but other ELF-based systems allow it as well. 654 */ 655 if (elf_check_header(eh, ET_EXEC) != 0 && !is_dyn) 656 return ENOEXEC; 657 658 if (eh->e_phnum > MAXPHNUM || eh->e_phnum == 0) 659 return ENOEXEC; 660 661 error = vn_marktext(epp->ep_vp); 662 if (error) 663 return error; 664 665 /* 666 * Allocate space to hold all the program headers, and read them 667 * from the file 668 */ 669 p = l->l_proc; 670 phsize = eh->e_phnum * sizeof(Elf_Phdr); 671 ph = kmem_alloc(phsize, KM_SLEEP); 672 673 if ((error = exec_read_from(l, epp->ep_vp, eh->e_phoff, ph, phsize)) != 674 0) 675 goto bad; 676 677 epp->ep_taddr = epp->ep_tsize = ELFDEFNNAME(NO_ADDR); 678 epp->ep_daddr = epp->ep_dsize = ELFDEFNNAME(NO_ADDR); 679 680 for (i = 0; i < eh->e_phnum; i++) { 681 pp = &ph[i]; 682 if (pp->p_type == PT_INTERP) { 683 if (pp->p_filesz >= MAXPATHLEN) { 684 error = ENOEXEC; 685 goto bad; 686 } 687 interp = PNBUF_GET(); 688 interp[0] = '\0'; 689 if ((error = exec_read_from(l, epp->ep_vp, 690 pp->p_offset, interp, pp->p_filesz)) != 0) 691 goto bad; 692 break; 693 } 694 } 695 696 /* 697 * On the same architecture, we may be emulating different systems. 698 * See which one will accept this executable. 699 * 700 * Probe functions would normally see if the interpreter (if any) 701 * exists. Emulation packages may possibly replace the interpreter in 702 * interp[] with a changed path (/emul/xxx/<path>). 703 */ 704 pos = ELFDEFNNAME(NO_ADDR); 705 if (epp->ep_esch->u.elf_probe_func) { 706 vaddr_t startp = (vaddr_t)pos; 707 708 error = (*epp->ep_esch->u.elf_probe_func)(l, epp, eh, interp, 709 &startp); 710 if (error) 711 goto bad; 712 pos = (Elf_Addr)startp; 713 } 714 715 #if defined(PAX_MPROTECT) || defined(PAX_SEGVGUARD) || defined(PAX_ASLR) 716 p->p_pax = epp->ep_pax_flags; 717 #endif /* PAX_MPROTECT || PAX_SEGVGUARD || PAX_ASLR */ 718 719 if (is_dyn) 720 elf_placedynexec(l, epp, eh, ph); 721 722 /* 723 * Load all the necessary sections 724 */ 725 for (i = nload = 0; i < eh->e_phnum; i++) { 726 Elf_Addr addr = ELFDEFNNAME(NO_ADDR); 727 u_long size = 0; 728 int prot = 0; 729 730 pp = &ph[i]; 731 732 switch (ph[i].p_type) { 733 case PT_LOAD: 734 elf_load_psection(&epp->ep_vmcmds, epp->ep_vp, 735 &ph[i], &addr, &size, &prot, VMCMD_FIXED); 736 737 /* 738 * Consider this as text segment, if it is executable. 739 * If there is more than one text segment, pick the 740 * largest. 741 */ 742 if (ph[i].p_flags & PF_X) { 743 if (epp->ep_taddr == ELFDEFNNAME(NO_ADDR) || 744 size > epp->ep_tsize) { 745 epp->ep_taddr = addr; 746 epp->ep_tsize = size; 747 } 748 end_text = addr + size; 749 } else { 750 epp->ep_daddr = addr; 751 epp->ep_dsize = size; 752 } 753 if (ph[i].p_offset == 0) { 754 computed_phdr = ph[i].p_vaddr + eh->e_phoff; 755 } 756 break; 757 758 case PT_SHLIB: 759 /* SCO has these sections. */ 760 case PT_INTERP: 761 /* Already did this one. */ 762 case PT_DYNAMIC: 763 break; 764 case PT_NOTE: 765 break; 766 case PT_PHDR: 767 /* Note address of program headers (in text segment) */ 768 phdr = pp->p_vaddr; 769 break; 770 771 default: 772 /* 773 * Not fatal; we don't need to understand everything. 774 */ 775 break; 776 } 777 } 778 if (interp || (epp->ep_flags & EXEC_FORCEAUX) != 0) { 779 ap = malloc(sizeof(struct elf_args), M_TEMP, M_WAITOK); 780 ap->arg_interp = (vaddr_t)NULL; 781 } 782 783 if (epp->ep_daddr == ELFDEFNNAME(NO_ADDR)) { 784 epp->ep_daddr = round_page(end_text); 785 epp->ep_dsize = 0; 786 } 787 788 /* 789 * Check if we found a dynamically linked binary and arrange to load 790 * its interpreter 791 */ 792 if (interp) { 793 int j = epp->ep_vmcmds.evs_used; 794 u_long interp_offset; 795 796 if ((error = elf_load_file(l, epp, interp, 797 &epp->ep_vmcmds, &interp_offset, ap, &pos)) != 0) { 798 goto bad; 799 } 800 ap->arg_interp = epp->ep_vmcmds.evs_cmds[j].ev_addr; 801 epp->ep_entry = ap->arg_interp + interp_offset; 802 PNBUF_PUT(interp); 803 } else 804 epp->ep_entry = eh->e_entry; 805 806 if (ap) { 807 ap->arg_phaddr = phdr ? phdr : computed_phdr; 808 ap->arg_phentsize = eh->e_phentsize; 809 ap->arg_phnum = eh->e_phnum; 810 ap->arg_entry = eh->e_entry; 811 epp->ep_emul_arg = ap; 812 } 813 814 #ifdef ELF_MAP_PAGE_ZERO 815 /* Dell SVR4 maps page zero, yeuch! */ 816 NEW_VMCMD(&epp->ep_vmcmds, vmcmd_map_readvn, PAGE_SIZE, 0, 817 epp->ep_vp, 0, VM_PROT_READ); 818 #endif 819 kmem_free(ph, phsize); 820 return (*epp->ep_esch->es_setup_stack)(l, epp); 821 822 bad: 823 if (interp) 824 PNBUF_PUT(interp); 825 if (ap) 826 free(ap, M_TEMP); 827 kmem_free(ph, phsize); 828 kill_vmcmds(&epp->ep_vmcmds); 829 return error; 830 } 831 832 int 833 netbsd_elf_signature(struct lwp *l, struct exec_package *epp, 834 Elf_Ehdr *eh) 835 { 836 size_t i; 837 Elf_Shdr *sh; 838 Elf_Nhdr *np; 839 size_t shsize; 840 int error; 841 int isnetbsd = 0; 842 char *ndata; 843 844 epp->ep_pax_flags = 0; 845 if (eh->e_shnum > MAXSHNUM || eh->e_shnum == 0) 846 return ENOEXEC; 847 848 shsize = eh->e_shnum * sizeof(Elf_Shdr); 849 sh = kmem_alloc(shsize, KM_SLEEP); 850 error = exec_read_from(l, epp->ep_vp, eh->e_shoff, sh, shsize); 851 if (error) 852 goto out; 853 854 np = kmem_alloc(MAXNOTESIZE, KM_SLEEP); 855 for (i = 0; i < eh->e_shnum; i++) { 856 Elf_Shdr *shp = &sh[i]; 857 858 if (shp->sh_type != SHT_NOTE || 859 shp->sh_size > MAXNOTESIZE || 860 shp->sh_size < sizeof(Elf_Nhdr) + ELF_NOTE_NETBSD_NAMESZ) 861 continue; 862 863 error = exec_read_from(l, epp->ep_vp, shp->sh_offset, np, 864 shp->sh_size); 865 if (error) 866 continue; 867 868 ndata = (char *)(np + 1); 869 switch (np->n_type) { 870 case ELF_NOTE_TYPE_NETBSD_TAG: 871 if (np->n_namesz != ELF_NOTE_NETBSD_NAMESZ || 872 np->n_descsz != ELF_NOTE_NETBSD_DESCSZ || 873 memcmp(ndata, ELF_NOTE_NETBSD_NAME, 874 ELF_NOTE_NETBSD_NAMESZ)) 875 goto bad; 876 isnetbsd = 1; 877 break; 878 879 case ELF_NOTE_TYPE_PAX_TAG: 880 if (np->n_namesz != ELF_NOTE_PAX_NAMESZ || 881 np->n_descsz != ELF_NOTE_PAX_DESCSZ || 882 memcmp(ndata, ELF_NOTE_PAX_NAME, 883 ELF_NOTE_PAX_NAMESZ)) { 884 bad: 885 #ifdef DIAGNOSTIC 886 printf("%s: bad tag %d: " 887 "[%d %d, %d %d, %*.*s %*.*s]\n", 888 epp->ep_kname, 889 np->n_type, 890 np->n_namesz, ELF_NOTE_PAX_NAMESZ, 891 np->n_descsz, ELF_NOTE_PAX_DESCSZ, 892 ELF_NOTE_PAX_NAMESZ, 893 ELF_NOTE_PAX_NAMESZ, 894 ndata, 895 ELF_NOTE_PAX_NAMESZ, 896 ELF_NOTE_PAX_NAMESZ, 897 ELF_NOTE_PAX_NAME); 898 #endif 899 continue; 900 } 901 (void)memcpy(&epp->ep_pax_flags, 902 ndata + ELF_NOTE_PAX_NAMESZ, 903 sizeof(epp->ep_pax_flags)); 904 break; 905 906 default: 907 #ifdef DIAGNOSTIC 908 printf("%s: unknown note type %d\n", epp->ep_kname, 909 np->n_type); 910 #endif 911 break; 912 } 913 } 914 kmem_free(np, MAXNOTESIZE); 915 916 error = isnetbsd ? 0 : ENOEXEC; 917 out: 918 kmem_free(sh, shsize); 919 return error; 920 } 921 922 int 923 netbsd_elf_probe(struct lwp *l, struct exec_package *epp, void *eh, char *itp, 924 vaddr_t *pos) 925 { 926 int error; 927 928 if ((error = netbsd_elf_signature(l, epp, eh)) != 0) 929 return error; 930 #ifdef ELF_MD_PROBE_FUNC 931 if ((error = ELF_MD_PROBE_FUNC(l, epp, eh, itp, pos)) != 0) 932 return error; 933 #elif defined(ELF_INTERP_NON_RELOCATABLE) 934 *pos = ELF_LINK_ADDR; 935 #endif 936 return 0; 937 } 938