Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 | #ifndef _ASM_IA64_ELF_H #define _ASM_IA64_ELF_H /* * ELF-specific definitions. * * Copyright (C) 1998-1999, 2002-2004 Hewlett-Packard Co * David Mosberger-Tang <davidm@hpl.hp.com> */ #include <linux/config.h> #include <asm/fpu.h> #include <asm/page.h> #include <asm/auxvec.h> /* * This is used to ensure we don't load something for the wrong architecture. */ #define elf_check_arch(x) ((x)->e_machine == EM_IA_64) /* * These are used to set parameters in the core dumps. */ #define ELF_CLASS ELFCLASS64 #define ELF_DATA ELFDATA2LSB #define ELF_ARCH EM_IA_64 #define USE_ELF_CORE_DUMP /* Least-significant four bits of ELF header's e_flags are OS-specific. The bits are interpreted as follows by Linux: */ #define EF_IA_64_LINUX_EXECUTABLE_STACK 0x1 /* is stack (& heap) executable by default? */ #define ELF_EXEC_PAGESIZE PAGE_SIZE /* * This is the location that an ET_DYN program is loaded if exec'ed. * Typical use of this is to invoke "./ld.so someprog" to test out a * new version of the loader. We need to make sure that it is out of * the way of the program that it will "exec", and that there is * sufficient room for the brk. */ #define ELF_ET_DYN_BASE (TASK_UNMAPPED_BASE + 0x800000000UL) #define PT_IA_64_UNWIND 0x70000001 /* IA-64 relocations: */ #define R_IA64_NONE 0x00 /* none */ #define R_IA64_IMM14 0x21 /* symbol + addend, add imm14 */ #define R_IA64_IMM22 0x22 /* symbol + addend, add imm22 */ #define R_IA64_IMM64 0x23 /* symbol + addend, mov imm64 */ #define R_IA64_DIR32MSB 0x24 /* symbol + addend, data4 MSB */ #define R_IA64_DIR32LSB 0x25 /* symbol + addend, data4 LSB */ #define R_IA64_DIR64MSB 0x26 /* symbol + addend, data8 MSB */ #define R_IA64_DIR64LSB 0x27 /* symbol + addend, data8 LSB */ #define R_IA64_GPREL22 0x2a /* @gprel(sym+add), add imm22 */ #define R_IA64_GPREL64I 0x2b /* @gprel(sym+add), mov imm64 */ #define R_IA64_GPREL32MSB 0x2c /* @gprel(sym+add), data4 MSB */ #define R_IA64_GPREL32LSB 0x2d /* @gprel(sym+add), data4 LSB */ #define R_IA64_GPREL64MSB 0x2e /* @gprel(sym+add), data8 MSB */ #define R_IA64_GPREL64LSB 0x2f /* @gprel(sym+add), data8 LSB */ #define R_IA64_LTOFF22 0x32 /* @ltoff(sym+add), add imm22 */ #define R_IA64_LTOFF64I 0x33 /* @ltoff(sym+add), mov imm64 */ #define R_IA64_PLTOFF22 0x3a /* @pltoff(sym+add), add imm22 */ #define R_IA64_PLTOFF64I 0x3b /* @pltoff(sym+add), mov imm64 */ #define R_IA64_PLTOFF64MSB 0x3e /* @pltoff(sym+add), data8 MSB */ #define R_IA64_PLTOFF64LSB 0x3f /* @pltoff(sym+add), data8 LSB */ #define R_IA64_FPTR64I 0x43 /* @fptr(sym+add), mov imm64 */ #define R_IA64_FPTR32MSB 0x44 /* @fptr(sym+add), data4 MSB */ #define R_IA64_FPTR32LSB 0x45 /* @fptr(sym+add), data4 LSB */ #define R_IA64_FPTR64MSB 0x46 /* @fptr(sym+add), data8 MSB */ #define R_IA64_FPTR64LSB 0x47 /* @fptr(sym+add), data8 LSB */ #define R_IA64_PCREL60B 0x48 /* @pcrel(sym+add), brl */ #define R_IA64_PCREL21B 0x49 /* @pcrel(sym+add), ptb, call */ #define R_IA64_PCREL21M 0x4a /* @pcrel(sym+add), chk.s */ #define R_IA64_PCREL21F 0x4b /* @pcrel(sym+add), fchkf */ #define R_IA64_PCREL32MSB 0x4c /* @pcrel(sym+add), data4 MSB */ #define R_IA64_PCREL32LSB 0x4d /* @pcrel(sym+add), data4 LSB */ #define R_IA64_PCREL64MSB 0x4e /* @pcrel(sym+add), data8 MSB */ #define R_IA64_PCREL64LSB 0x4f /* @pcrel(sym+add), data8 LSB */ #define R_IA64_LTOFF_FPTR22 0x52 /* @ltoff(@fptr(s+a)), imm22 */ #define R_IA64_LTOFF_FPTR64I 0x53 /* @ltoff(@fptr(s+a)), imm64 */ #define R_IA64_LTOFF_FPTR32MSB 0x54 /* @ltoff(@fptr(s+a)), 4 MSB */ #define R_IA64_LTOFF_FPTR32LSB 0x55 /* @ltoff(@fptr(s+a)), 4 LSB */ #define R_IA64_LTOFF_FPTR64MSB 0x56 /* @ltoff(@fptr(s+a)), 8 MSB */ #define R_IA64_LTOFF_FPTR64LSB 0x57 /* @ltoff(@fptr(s+a)), 8 LSB */ #define R_IA64_SEGREL32MSB 0x5c /* @segrel(sym+add), data4 MSB */ #define R_IA64_SEGREL32LSB 0x5d /* @segrel(sym+add), data4 LSB */ #define R_IA64_SEGREL64MSB 0x5e /* @segrel(sym+add), data8 MSB */ #define R_IA64_SEGREL64LSB 0x5f /* @segrel(sym+add), data8 LSB */ #define R_IA64_SECREL32MSB 0x64 /* @secrel(sym+add), data4 MSB */ #define R_IA64_SECREL32LSB 0x65 /* @secrel(sym+add), data4 LSB */ #define R_IA64_SECREL64MSB 0x66 /* @secrel(sym+add), data8 MSB */ #define R_IA64_SECREL64LSB 0x67 /* @secrel(sym+add), data8 LSB */ #define R_IA64_REL32MSB 0x6c /* data 4 + REL */ #define R_IA64_REL32LSB 0x6d /* data 4 + REL */ #define R_IA64_REL64MSB 0x6e /* data 8 + REL */ #define R_IA64_REL64LSB 0x6f /* data 8 + REL */ #define R_IA64_LTV32MSB 0x74 /* symbol + addend, data4 MSB */ #define R_IA64_LTV32LSB 0x75 /* symbol + addend, data4 LSB */ #define R_IA64_LTV64MSB 0x76 /* symbol + addend, data8 MSB */ #define R_IA64_LTV64LSB 0x77 /* symbol + addend, data8 LSB */ #define R_IA64_PCREL21BI 0x79 /* @pcrel(sym+add), ptb, call */ #define R_IA64_PCREL22 0x7a /* @pcrel(sym+add), imm22 */ #define R_IA64_PCREL64I 0x7b /* @pcrel(sym+add), imm64 */ #define R_IA64_IPLTMSB 0x80 /* dynamic reloc, imported PLT, MSB */ #define R_IA64_IPLTLSB 0x81 /* dynamic reloc, imported PLT, LSB */ #define R_IA64_COPY 0x84 /* dynamic reloc, data copy */ #define R_IA64_SUB 0x85 /* -symbol + addend, add imm22 */ #define R_IA64_LTOFF22X 0x86 /* LTOFF22, relaxable. */ #define R_IA64_LDXMOV 0x87 /* Use of LTOFF22X. */ #define R_IA64_TPREL14 0x91 /* @tprel(sym+add), add imm14 */ #define R_IA64_TPREL22 0x92 /* @tprel(sym+add), add imm22 */ #define R_IA64_TPREL64I 0x93 /* @tprel(sym+add), add imm64 */ #define R_IA64_TPREL64MSB 0x96 /* @tprel(sym+add), data8 MSB */ #define R_IA64_TPREL64LSB 0x97 /* @tprel(sym+add), data8 LSB */ #define R_IA64_LTOFF_TPREL22 0x9a /* @ltoff(@tprel(s+a)), add imm22 */ #define R_IA64_DTPMOD64MSB 0xa6 /* @dtpmod(sym+add), data8 MSB */ #define R_IA64_DTPMOD64LSB 0xa7 /* @dtpmod(sym+add), data8 LSB */ #define R_IA64_LTOFF_DTPMOD22 0xaa /* @ltoff(@dtpmod(s+a)), imm22 */ #define R_IA64_DTPREL14 0xb1 /* @dtprel(sym+add), imm14 */ #define R_IA64_DTPREL22 0xb2 /* @dtprel(sym+add), imm22 */ #define R_IA64_DTPREL64I 0xb3 /* @dtprel(sym+add), imm64 */ #define R_IA64_DTPREL32MSB 0xb4 /* @dtprel(sym+add), data4 MSB */ #define R_IA64_DTPREL32LSB 0xb5 /* @dtprel(sym+add), data4 LSB */ #define R_IA64_DTPREL64MSB 0xb6 /* @dtprel(sym+add), data8 MSB */ #define R_IA64_DTPREL64LSB 0xb7 /* @dtprel(sym+add), data8 LSB */ #define R_IA64_LTOFF_DTPREL22 0xba /* @ltoff(@dtprel(s+a)), imm22 */ /* IA-64 specific section flags: */ #define SHF_IA_64_SHORT 0x10000000 /* section near gp */ /* * We use (abuse?) this macro to insert the (empty) vm_area that is * used to map the register backing store. I don't see any better * place to do this, but we should discuss this with Linus once we can * talk to him... */ extern void ia64_init_addr_space (void); #define ELF_PLAT_INIT(_r, load_addr) ia64_init_addr_space() /* ELF register definitions. This is needed for core dump support. */ /* * elf_gregset_t contains the application-level state in the following order: * r0-r31 * NaT bits (for r0-r31; bit N == 1 iff rN is a NaT) * predicate registers (p0-p63) * b0-b7 * ip cfm psr * ar.rsc ar.bsp ar.bspstore ar.rnat * ar.ccv ar.unat ar.fpsr ar.pfs ar.lc ar.ec ar.csd ar.ssd */ #define ELF_NGREG 128 /* we really need just 72 but let's leave some headroom... */ #define ELF_NFPREG 128 /* f0 and f1 could be omitted, but so what... */ typedef unsigned long elf_fpxregset_t; typedef unsigned long elf_greg_t; typedef elf_greg_t elf_gregset_t[ELF_NGREG]; typedef struct ia64_fpreg elf_fpreg_t; typedef elf_fpreg_t elf_fpregset_t[ELF_NFPREG]; struct pt_regs; /* forward declaration... */ extern void ia64_elf_core_copy_regs (struct pt_regs *src, elf_gregset_t dst); #define ELF_CORE_COPY_REGS(_dest,_regs) ia64_elf_core_copy_regs(_regs, _dest); /* This macro yields a bitmask that programs can use to figure out what instruction set this CPU supports. */ #define ELF_HWCAP 0 /* This macro yields a string that ld.so will use to load implementation specific libraries for optimization. Not terribly relevant until we have real hardware to play with... */ #define ELF_PLATFORM NULL #ifdef __KERNEL__ #define SET_PERSONALITY(ex, ibcs2) set_personality(PER_LINUX) #define elf_read_implies_exec(ex, executable_stack) \ ((executable_stack!=EXSTACK_DISABLE_X) && ((ex).e_flags & EF_IA_64_LINUX_EXECUTABLE_STACK) != 0) struct task_struct; extern int dump_task_regs(struct task_struct *, elf_gregset_t *); extern int dump_task_fpu (struct task_struct *, elf_fpregset_t *); #define ELF_CORE_COPY_TASK_REGS(tsk, elf_gregs) dump_task_regs(tsk, elf_gregs) #define ELF_CORE_COPY_FPREGS(tsk, elf_fpregs) dump_task_fpu(tsk, elf_fpregs) #define GATE_EHDR ((const struct elfhdr *) GATE_ADDR) #define ARCH_DLINFO \ do { \ extern char __kernel_syscall_via_epc[]; \ NEW_AUX_ENT(AT_SYSINFO, (unsigned long) __kernel_syscall_via_epc); \ NEW_AUX_ENT(AT_SYSINFO_EHDR, (unsigned long) GATE_EHDR); \ } while (0) /* * These macros parameterize elf_core_dump in fs/binfmt_elf.c to write out * extra segments containing the gate DSO contents. Dumping its * contents makes post-mortem fully interpretable later without matching up * the same kernel and hardware config to see what PC values meant. * Dumping its extra ELF program headers includes all the other information * a debugger needs to easily find how the gate DSO was being used. */ #define ELF_CORE_EXTRA_PHDRS (GATE_EHDR->e_phnum) #define ELF_CORE_WRITE_EXTRA_PHDRS \ do { \ const struct elf_phdr *const gate_phdrs = \ (const struct elf_phdr *) (GATE_ADDR + GATE_EHDR->e_phoff); \ int i; \ Elf64_Off ofs = 0; \ for (i = 0; i < GATE_EHDR->e_phnum; ++i) { \ struct elf_phdr phdr = gate_phdrs[i]; \ if (phdr.p_type == PT_LOAD) { \ phdr.p_memsz = PAGE_ALIGN(phdr.p_memsz); \ phdr.p_filesz = phdr.p_memsz; \ if (ofs == 0) { \ ofs = phdr.p_offset = offset; \ offset += phdr.p_filesz; \ } \ else \ phdr.p_offset = ofs; \ } \ else \ phdr.p_offset += ofs; \ phdr.p_paddr = 0; /* match other core phdrs */ \ DUMP_WRITE(&phdr, sizeof(phdr)); \ } \ } while (0) #define ELF_CORE_WRITE_EXTRA_DATA \ do { \ const struct elf_phdr *const gate_phdrs = \ (const struct elf_phdr *) (GATE_ADDR + GATE_EHDR->e_phoff); \ int i; \ for (i = 0; i < GATE_EHDR->e_phnum; ++i) { \ if (gate_phdrs[i].p_type == PT_LOAD) { \ DUMP_WRITE((void *) gate_phdrs[i].p_vaddr, \ PAGE_ALIGN(gate_phdrs[i].p_memsz)); \ break; \ } \ } \ } while (0) #endif /* __KERNEL__ */ #endif /* _ASM_IA64_ELF_H */ |