xref: /openbmc/linux/arch/ia64/include/asm/elf.h (revision a09d2831)
1 #ifndef _ASM_IA64_ELF_H
2 #define _ASM_IA64_ELF_H
3 
4 /*
5  * ELF-specific definitions.
6  *
7  * Copyright (C) 1998-1999, 2002-2004 Hewlett-Packard Co
8  *	David Mosberger-Tang <davidm@hpl.hp.com>
9  */
10 
11 
12 #include <asm/fpu.h>
13 #include <asm/page.h>
14 #include <asm/auxvec.h>
15 
16 /*
17  * This is used to ensure we don't load something for the wrong architecture.
18  */
19 #define elf_check_arch(x) ((x)->e_machine == EM_IA_64)
20 
21 /*
22  * These are used to set parameters in the core dumps.
23  */
24 #define ELF_CLASS	ELFCLASS64
25 #define ELF_DATA	ELFDATA2LSB
26 #define ELF_ARCH	EM_IA_64
27 
28 #define CORE_DUMP_USE_REGSET
29 
30 /* Least-significant four bits of ELF header's e_flags are OS-specific.  The bits are
31    interpreted as follows by Linux: */
32 #define EF_IA_64_LINUX_EXECUTABLE_STACK	0x1	/* is stack (& heap) executable by default? */
33 
34 #define ELF_EXEC_PAGESIZE	PAGE_SIZE
35 
36 /*
37  * This is the location that an ET_DYN program is loaded if exec'ed.
38  * Typical use of this is to invoke "./ld.so someprog" to test out a
39  * new version of the loader.  We need to make sure that it is out of
40  * the way of the program that it will "exec", and that there is
41  * sufficient room for the brk.
42  */
43 #define ELF_ET_DYN_BASE		(TASK_UNMAPPED_BASE + 0x800000000UL)
44 
45 #define PT_IA_64_UNWIND		0x70000001
46 
47 /* IA-64 relocations: */
48 #define R_IA64_NONE		0x00	/* none */
49 #define R_IA64_IMM14		0x21	/* symbol + addend, add imm14 */
50 #define R_IA64_IMM22		0x22	/* symbol + addend, add imm22 */
51 #define R_IA64_IMM64		0x23	/* symbol + addend, mov imm64 */
52 #define R_IA64_DIR32MSB		0x24	/* symbol + addend, data4 MSB */
53 #define R_IA64_DIR32LSB		0x25	/* symbol + addend, data4 LSB */
54 #define R_IA64_DIR64MSB		0x26	/* symbol + addend, data8 MSB */
55 #define R_IA64_DIR64LSB		0x27	/* symbol + addend, data8 LSB */
56 #define R_IA64_GPREL22		0x2a	/* @gprel(sym+add), add imm22 */
57 #define R_IA64_GPREL64I		0x2b	/* @gprel(sym+add), mov imm64 */
58 #define R_IA64_GPREL32MSB	0x2c	/* @gprel(sym+add), data4 MSB */
59 #define R_IA64_GPREL32LSB	0x2d	/* @gprel(sym+add), data4 LSB */
60 #define R_IA64_GPREL64MSB	0x2e	/* @gprel(sym+add), data8 MSB */
61 #define R_IA64_GPREL64LSB	0x2f	/* @gprel(sym+add), data8 LSB */
62 #define R_IA64_LTOFF22		0x32	/* @ltoff(sym+add), add imm22 */
63 #define R_IA64_LTOFF64I		0x33	/* @ltoff(sym+add), mov imm64 */
64 #define R_IA64_PLTOFF22		0x3a	/* @pltoff(sym+add), add imm22 */
65 #define R_IA64_PLTOFF64I	0x3b	/* @pltoff(sym+add), mov imm64 */
66 #define R_IA64_PLTOFF64MSB	0x3e	/* @pltoff(sym+add), data8 MSB */
67 #define R_IA64_PLTOFF64LSB	0x3f	/* @pltoff(sym+add), data8 LSB */
68 #define R_IA64_FPTR64I		0x43	/* @fptr(sym+add), mov imm64 */
69 #define R_IA64_FPTR32MSB	0x44	/* @fptr(sym+add), data4 MSB */
70 #define R_IA64_FPTR32LSB	0x45	/* @fptr(sym+add), data4 LSB */
71 #define R_IA64_FPTR64MSB	0x46	/* @fptr(sym+add), data8 MSB */
72 #define R_IA64_FPTR64LSB	0x47	/* @fptr(sym+add), data8 LSB */
73 #define R_IA64_PCREL60B		0x48	/* @pcrel(sym+add), brl */
74 #define R_IA64_PCREL21B		0x49	/* @pcrel(sym+add), ptb, call */
75 #define R_IA64_PCREL21M		0x4a	/* @pcrel(sym+add), chk.s */
76 #define R_IA64_PCREL21F		0x4b	/* @pcrel(sym+add), fchkf */
77 #define R_IA64_PCREL32MSB	0x4c	/* @pcrel(sym+add), data4 MSB */
78 #define R_IA64_PCREL32LSB	0x4d	/* @pcrel(sym+add), data4 LSB */
79 #define R_IA64_PCREL64MSB	0x4e	/* @pcrel(sym+add), data8 MSB */
80 #define R_IA64_PCREL64LSB	0x4f	/* @pcrel(sym+add), data8 LSB */
81 #define R_IA64_LTOFF_FPTR22	0x52	/* @ltoff(@fptr(s+a)), imm22 */
82 #define R_IA64_LTOFF_FPTR64I	0x53	/* @ltoff(@fptr(s+a)), imm64 */
83 #define R_IA64_LTOFF_FPTR32MSB	0x54	/* @ltoff(@fptr(s+a)), 4 MSB */
84 #define R_IA64_LTOFF_FPTR32LSB	0x55	/* @ltoff(@fptr(s+a)), 4 LSB */
85 #define R_IA64_LTOFF_FPTR64MSB	0x56	/* @ltoff(@fptr(s+a)), 8 MSB */
86 #define R_IA64_LTOFF_FPTR64LSB	0x57	/* @ltoff(@fptr(s+a)), 8 LSB */
87 #define R_IA64_SEGREL32MSB	0x5c	/* @segrel(sym+add), data4 MSB */
88 #define R_IA64_SEGREL32LSB	0x5d	/* @segrel(sym+add), data4 LSB */
89 #define R_IA64_SEGREL64MSB	0x5e	/* @segrel(sym+add), data8 MSB */
90 #define R_IA64_SEGREL64LSB	0x5f	/* @segrel(sym+add), data8 LSB */
91 #define R_IA64_SECREL32MSB	0x64	/* @secrel(sym+add), data4 MSB */
92 #define R_IA64_SECREL32LSB	0x65	/* @secrel(sym+add), data4 LSB */
93 #define R_IA64_SECREL64MSB	0x66	/* @secrel(sym+add), data8 MSB */
94 #define R_IA64_SECREL64LSB	0x67	/* @secrel(sym+add), data8 LSB */
95 #define R_IA64_REL32MSB		0x6c	/* data 4 + REL */
96 #define R_IA64_REL32LSB		0x6d	/* data 4 + REL */
97 #define R_IA64_REL64MSB		0x6e	/* data 8 + REL */
98 #define R_IA64_REL64LSB		0x6f	/* data 8 + REL */
99 #define R_IA64_LTV32MSB		0x74	/* symbol + addend, data4 MSB */
100 #define R_IA64_LTV32LSB		0x75	/* symbol + addend, data4 LSB */
101 #define R_IA64_LTV64MSB		0x76	/* symbol + addend, data8 MSB */
102 #define R_IA64_LTV64LSB		0x77	/* symbol + addend, data8 LSB */
103 #define R_IA64_PCREL21BI	0x79	/* @pcrel(sym+add), ptb, call */
104 #define R_IA64_PCREL22		0x7a	/* @pcrel(sym+add), imm22 */
105 #define R_IA64_PCREL64I		0x7b	/* @pcrel(sym+add), imm64 */
106 #define R_IA64_IPLTMSB		0x80	/* dynamic reloc, imported PLT, MSB */
107 #define R_IA64_IPLTLSB		0x81	/* dynamic reloc, imported PLT, LSB */
108 #define R_IA64_COPY		0x84	/* dynamic reloc, data copy */
109 #define R_IA64_SUB		0x85	/* -symbol + addend, add imm22 */
110 #define R_IA64_LTOFF22X		0x86	/* LTOFF22, relaxable.  */
111 #define R_IA64_LDXMOV		0x87	/* Use of LTOFF22X.  */
112 #define R_IA64_TPREL14		0x91	/* @tprel(sym+add), add imm14 */
113 #define R_IA64_TPREL22		0x92	/* @tprel(sym+add), add imm22 */
114 #define R_IA64_TPREL64I		0x93	/* @tprel(sym+add), add imm64 */
115 #define R_IA64_TPREL64MSB	0x96	/* @tprel(sym+add), data8 MSB */
116 #define R_IA64_TPREL64LSB	0x97	/* @tprel(sym+add), data8 LSB */
117 #define R_IA64_LTOFF_TPREL22	0x9a	/* @ltoff(@tprel(s+a)), add imm22 */
118 #define R_IA64_DTPMOD64MSB	0xa6	/* @dtpmod(sym+add), data8 MSB */
119 #define R_IA64_DTPMOD64LSB	0xa7	/* @dtpmod(sym+add), data8 LSB */
120 #define R_IA64_LTOFF_DTPMOD22	0xaa	/* @ltoff(@dtpmod(s+a)), imm22 */
121 #define R_IA64_DTPREL14		0xb1	/* @dtprel(sym+add), imm14 */
122 #define R_IA64_DTPREL22		0xb2	/* @dtprel(sym+add), imm22 */
123 #define R_IA64_DTPREL64I	0xb3	/* @dtprel(sym+add), imm64 */
124 #define R_IA64_DTPREL32MSB	0xb4	/* @dtprel(sym+add), data4 MSB */
125 #define R_IA64_DTPREL32LSB	0xb5	/* @dtprel(sym+add), data4 LSB */
126 #define R_IA64_DTPREL64MSB	0xb6	/* @dtprel(sym+add), data8 MSB */
127 #define R_IA64_DTPREL64LSB	0xb7	/* @dtprel(sym+add), data8 LSB */
128 #define R_IA64_LTOFF_DTPREL22	0xba	/* @ltoff(@dtprel(s+a)), imm22 */
129 
130 /* IA-64 specific section flags: */
131 #define SHF_IA_64_SHORT		0x10000000	/* section near gp */
132 
133 /*
134  * We use (abuse?) this macro to insert the (empty) vm_area that is
135  * used to map the register backing store.  I don't see any better
136  * place to do this, but we should discuss this with Linus once we can
137  * talk to him...
138  */
139 extern void ia64_init_addr_space (void);
140 #define ELF_PLAT_INIT(_r, load_addr)	ia64_init_addr_space()
141 
142 /* ELF register definitions.  This is needed for core dump support.  */
143 
144 /*
145  * elf_gregset_t contains the application-level state in the following order:
146  *	r0-r31
147  *	NaT bits (for r0-r31; bit N == 1 iff rN is a NaT)
148  *	predicate registers (p0-p63)
149  *	b0-b7
150  *	ip cfm psr
151  *	ar.rsc ar.bsp ar.bspstore ar.rnat
152  *	ar.ccv ar.unat ar.fpsr ar.pfs ar.lc ar.ec ar.csd ar.ssd
153  */
154 #define ELF_NGREG	128	/* we really need just 72 but let's leave some headroom... */
155 #define ELF_NFPREG	128	/* f0 and f1 could be omitted, but so what... */
156 
157 /* elf_gregset_t register offsets */
158 #define ELF_GR_0_OFFSET     0
159 #define ELF_NAT_OFFSET     (32 * sizeof(elf_greg_t))
160 #define ELF_PR_OFFSET      (33 * sizeof(elf_greg_t))
161 #define ELF_BR_0_OFFSET    (34 * sizeof(elf_greg_t))
162 #define ELF_CR_IIP_OFFSET  (42 * sizeof(elf_greg_t))
163 #define ELF_CFM_OFFSET     (43 * sizeof(elf_greg_t))
164 #define ELF_CR_IPSR_OFFSET (44 * sizeof(elf_greg_t))
165 #define ELF_GR_OFFSET(i)   (ELF_GR_0_OFFSET + i * sizeof(elf_greg_t))
166 #define ELF_BR_OFFSET(i)   (ELF_BR_0_OFFSET + i * sizeof(elf_greg_t))
167 #define ELF_AR_RSC_OFFSET  (45 * sizeof(elf_greg_t))
168 #define ELF_AR_BSP_OFFSET  (46 * sizeof(elf_greg_t))
169 #define ELF_AR_BSPSTORE_OFFSET (47 * sizeof(elf_greg_t))
170 #define ELF_AR_RNAT_OFFSET (48 * sizeof(elf_greg_t))
171 #define ELF_AR_CCV_OFFSET  (49 * sizeof(elf_greg_t))
172 #define ELF_AR_UNAT_OFFSET (50 * sizeof(elf_greg_t))
173 #define ELF_AR_FPSR_OFFSET (51 * sizeof(elf_greg_t))
174 #define ELF_AR_PFS_OFFSET  (52 * sizeof(elf_greg_t))
175 #define ELF_AR_LC_OFFSET   (53 * sizeof(elf_greg_t))
176 #define ELF_AR_EC_OFFSET   (54 * sizeof(elf_greg_t))
177 #define ELF_AR_CSD_OFFSET  (55 * sizeof(elf_greg_t))
178 #define ELF_AR_SSD_OFFSET  (56 * sizeof(elf_greg_t))
179 #define ELF_AR_END_OFFSET  (57 * sizeof(elf_greg_t))
180 
181 typedef unsigned long elf_fpxregset_t;
182 
183 typedef unsigned long elf_greg_t;
184 typedef elf_greg_t elf_gregset_t[ELF_NGREG];
185 
186 typedef struct ia64_fpreg elf_fpreg_t;
187 typedef elf_fpreg_t elf_fpregset_t[ELF_NFPREG];
188 
189 
190 
191 struct pt_regs;	/* forward declaration... */
192 extern void ia64_elf_core_copy_regs (struct pt_regs *src, elf_gregset_t dst);
193 #define ELF_CORE_COPY_REGS(_dest,_regs)	ia64_elf_core_copy_regs(_regs, _dest);
194 
195 /* This macro yields a bitmask that programs can use to figure out
196    what instruction set this CPU supports.  */
197 #define ELF_HWCAP 	0
198 
199 /* This macro yields a string that ld.so will use to load
200    implementation specific libraries for optimization.  Not terribly
201    relevant until we have real hardware to play with... */
202 #define ELF_PLATFORM	NULL
203 
204 #define SET_PERSONALITY(ex)	set_personality(PER_LINUX)
205 #define elf_read_implies_exec(ex, executable_stack)					\
206 	((executable_stack!=EXSTACK_DISABLE_X) && ((ex).e_flags & EF_IA_64_LINUX_EXECUTABLE_STACK) != 0)
207 
208 struct task_struct;
209 
210 #define GATE_EHDR	((const struct elfhdr *) GATE_ADDR)
211 
212 /* update AT_VECTOR_SIZE_ARCH if the number of NEW_AUX_ENT entries changes */
213 #define ARCH_DLINFO								\
214 do {										\
215 	extern char __kernel_syscall_via_epc[];					\
216 	NEW_AUX_ENT(AT_SYSINFO, (unsigned long) __kernel_syscall_via_epc);	\
217 	NEW_AUX_ENT(AT_SYSINFO_EHDR, (unsigned long) GATE_EHDR);		\
218 } while (0)
219 
220 
221 /*
222  * These macros parameterize elf_core_dump in fs/binfmt_elf.c to write out
223  * extra segments containing the gate DSO contents.  Dumping its
224  * contents makes post-mortem fully interpretable later without matching up
225  * the same kernel and hardware config to see what PC values meant.
226  * Dumping its extra ELF program headers includes all the other information
227  * a debugger needs to easily find how the gate DSO was being used.
228  */
229 #define ELF_CORE_EXTRA_PHDRS		(GATE_EHDR->e_phnum)
230 #define ELF_CORE_WRITE_EXTRA_PHDRS						\
231 do {										\
232 	const struct elf_phdr *const gate_phdrs =			      \
233 		(const struct elf_phdr *) (GATE_ADDR + GATE_EHDR->e_phoff);   \
234 	int i;									\
235 	Elf64_Off ofs = 0;						      \
236 	for (i = 0; i < GATE_EHDR->e_phnum; ++i) {				\
237 		struct elf_phdr phdr = gate_phdrs[i];			      \
238 		if (phdr.p_type == PT_LOAD) {					\
239 			phdr.p_memsz = PAGE_ALIGN(phdr.p_memsz);	      \
240 			phdr.p_filesz = phdr.p_memsz;			      \
241 			if (ofs == 0) {					      \
242 				ofs = phdr.p_offset = offset;		      \
243 			offset += phdr.p_filesz;				\
244 		}							      \
245 		else							      \
246 				phdr.p_offset = ofs;			      \
247 		}							      \
248 		else							      \
249 			phdr.p_offset += ofs;					\
250 		phdr.p_paddr = 0; /* match other core phdrs */			\
251 		DUMP_WRITE(&phdr, sizeof(phdr));				\
252 	}									\
253 } while (0)
254 #define ELF_CORE_WRITE_EXTRA_DATA					\
255 do {									\
256 	const struct elf_phdr *const gate_phdrs =			      \
257 		(const struct elf_phdr *) (GATE_ADDR + GATE_EHDR->e_phoff);   \
258 	int i;								\
259 	for (i = 0; i < GATE_EHDR->e_phnum; ++i) {			\
260 		if (gate_phdrs[i].p_type == PT_LOAD) {			      \
261 			DUMP_WRITE((void *) gate_phdrs[i].p_vaddr,	      \
262 				   PAGE_ALIGN(gate_phdrs[i].p_memsz));	      \
263 			break;						      \
264 		}							      \
265 	}								\
266 } while (0)
267 
268 /*
269  * format for entries in the Global Offset Table
270  */
271 struct got_entry {
272 	uint64_t val;
273 };
274 
275 /*
276  * Layout of the Function Descriptor
277  */
278 struct fdesc {
279 	uint64_t ip;
280 	uint64_t gp;
281 };
282 
283 #endif /* _ASM_IA64_ELF_H */
284