1 /* AArch64-specific support for NN-bit ELF.
2 Copyright (C) 2009-2021 Free Software Foundation, Inc.
3 Contributed by ARM Ltd.
4
5 This file is part of BFD, the Binary File Descriptor library.
6
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 3 of the License, or
10 (at your option) any later version.
11
12 This program is distributed in the hope that it will be useful,
13 but WITHOUT ANY WARRANTY; without even the implied warranty of
14 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 GNU General Public License for more details.
16
17 You should have received a copy of the GNU General Public License
18 along with this program; see the file COPYING3. If not,
19 see <http://www.gnu.org/licenses/>. */
20
21 /* Notes on implementation:
22
23 Thread Local Store (TLS)
24
25 Overview:
26
27 The implementation currently supports both traditional TLS and TLS
28 descriptors, but only general dynamic (GD).
29
30 For traditional TLS the assembler will present us with code
31 fragments of the form:
32
33 adrp x0, :tlsgd:foo
34 R_AARCH64_TLSGD_ADR_PAGE21(foo)
35 add x0, :tlsgd_lo12:foo
36 R_AARCH64_TLSGD_ADD_LO12_NC(foo)
37 bl __tls_get_addr
38 nop
39
40 For TLS descriptors the assembler will present us with code
41 fragments of the form:
42
43 adrp x0, :tlsdesc:foo R_AARCH64_TLSDESC_ADR_PAGE21(foo)
44 ldr x1, [x0, #:tlsdesc_lo12:foo] R_AARCH64_TLSDESC_LD64_LO12(foo)
45 add x0, x0, #:tlsdesc_lo12:foo R_AARCH64_TLSDESC_ADD_LO12(foo)
46 .tlsdesccall foo
47 blr x1 R_AARCH64_TLSDESC_CALL(foo)
48
49 The relocations R_AARCH64_TLSGD_{ADR_PREL21,ADD_LO12_NC} against foo
50 indicate that foo is thread local and should be accessed via the
51 traditional TLS mechanims.
52
53 The relocations R_AARCH64_TLSDESC_{ADR_PAGE21,LD64_LO12_NC,ADD_LO12_NC}
54 against foo indicate that 'foo' is thread local and should be accessed
55 via a TLS descriptor mechanism.
56
57 The precise instruction sequence is only relevant from the
58 perspective of linker relaxation which is currently not implemented.
59
60 The static linker must detect that 'foo' is a TLS object and
61 allocate a double GOT entry. The GOT entry must be created for both
62 global and local TLS symbols. Note that this is different to none
63 TLS local objects which do not need a GOT entry.
64
65 In the traditional TLS mechanism, the double GOT entry is used to
66 provide the tls_index structure, containing module and offset
67 entries. The static linker places the relocation R_AARCH64_TLS_DTPMOD
68 on the module entry. The loader will subsequently fixup this
69 relocation with the module identity.
70
71 For global traditional TLS symbols the static linker places an
72 R_AARCH64_TLS_DTPREL relocation on the offset entry. The loader
73 will subsequently fixup the offset. For local TLS symbols the static
74 linker fixes up offset.
75
76 In the TLS descriptor mechanism the double GOT entry is used to
77 provide the descriptor. The static linker places the relocation
78 R_AARCH64_TLSDESC on the first GOT slot. The loader will
79 subsequently fix this up.
80
81 Implementation:
82
83 The handling of TLS symbols is implemented across a number of
84 different backend functions. The following is a top level view of
85 what processing is performed where.
86
87 The TLS implementation maintains state information for each TLS
88 symbol. The state information for local and global symbols is kept
89 in different places. Global symbols use generic BFD structures while
90 local symbols use backend specific structures that are allocated and
91 maintained entirely by the backend.
92
93 The flow:
94
95 elfNN_aarch64_check_relocs()
96
97 This function is invoked for each relocation.
98
99 The TLS relocations R_AARCH64_TLSGD_{ADR_PREL21,ADD_LO12_NC} and
100 R_AARCH64_TLSDESC_{ADR_PAGE21,LD64_LO12_NC,ADD_LO12_NC} are
101 spotted. One time creation of local symbol data structures are
102 created when the first local symbol is seen.
103
104 The reference count for a symbol is incremented. The GOT type for
105 each symbol is marked as general dynamic.
106
107 elfNN_aarch64_allocate_dynrelocs ()
108
109 For each global with positive reference count we allocate a double
110 GOT slot. For a traditional TLS symbol we allocate space for two
111 relocation entries on the GOT, for a TLS descriptor symbol we
112 allocate space for one relocation on the slot. Record the GOT offset
113 for this symbol.
114
115 elfNN_aarch64_size_dynamic_sections ()
116
117 Iterate all input BFDS, look for in the local symbol data structure
118 constructed earlier for local TLS symbols and allocate them double
119 GOT slots along with space for a single GOT relocation. Update the
120 local symbol structure to record the GOT offset allocated.
121
122 elfNN_aarch64_relocate_section ()
123
124 Calls elfNN_aarch64_final_link_relocate ()
125
126 Emit the relevant TLS relocations against the GOT for each TLS
127 symbol. For local TLS symbols emit the GOT offset directly. The GOT
128 relocations are emitted once the first time a TLS symbol is
129 encountered. The implementation uses the LSB of the GOT offset to
130 flag that the relevant GOT relocations for a symbol have been
131 emitted. All of the TLS code that uses the GOT offset needs to take
132 care to mask out this flag bit before using the offset.
133
134 elfNN_aarch64_final_link_relocate ()
135
136 Fixup the R_AARCH64_TLSGD_{ADR_PREL21, ADD_LO12_NC} relocations. */
137
138 #include "sysdep.h"
139 #include "bfd.h"
140 #include "libiberty.h"
141 #include "libbfd.h"
142 #include "elf-bfd.h"
143 #include "bfdlink.h"
144 #include "objalloc.h"
145 #include "elf/aarch64.h"
146 #include "elfxx-aarch64.h"
147 #include "cpu-aarch64.h"
148
149 #define ARCH_SIZE NN
150
151 #if ARCH_SIZE == 64
152 #define AARCH64_R(NAME) R_AARCH64_ ## NAME
153 #define AARCH64_R_STR(NAME) "R_AARCH64_" #NAME
154 #define HOWTO64(...) HOWTO (__VA_ARGS__)
155 #define HOWTO32(...) EMPTY_HOWTO (0)
156 #define LOG_FILE_ALIGN 3
157 #define BFD_RELOC_AARCH64_TLSDESC_LD64_LO12_NC BFD_RELOC_AARCH64_TLSDESC_LD64_LO12
158 #endif
159
160 #if ARCH_SIZE == 32
161 #define AARCH64_R(NAME) R_AARCH64_P32_ ## NAME
162 #define AARCH64_R_STR(NAME) "R_AARCH64_P32_" #NAME
163 #define HOWTO64(...) EMPTY_HOWTO (0)
164 #define HOWTO32(...) HOWTO (__VA_ARGS__)
165 #define LOG_FILE_ALIGN 2
166 #define BFD_RELOC_AARCH64_TLSDESC_LD32_LO12 BFD_RELOC_AARCH64_TLSDESC_LD32_LO12_NC
167 #define R_AARCH64_P32_TLSDESC_ADD_LO12 R_AARCH64_P32_TLSDESC_ADD_LO12_NC
168 #endif
169
170 #define IS_AARCH64_TLS_RELOC(R_TYPE) \
171 ((R_TYPE) == BFD_RELOC_AARCH64_TLSGD_ADD_LO12_NC \
172 || (R_TYPE) == BFD_RELOC_AARCH64_TLSGD_ADR_PAGE21 \
173 || (R_TYPE) == BFD_RELOC_AARCH64_TLSGD_ADR_PREL21 \
174 || (R_TYPE) == BFD_RELOC_AARCH64_TLSGD_MOVW_G0_NC \
175 || (R_TYPE) == BFD_RELOC_AARCH64_TLSGD_MOVW_G1 \
176 || (R_TYPE) == BFD_RELOC_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21 \
177 || (R_TYPE) == BFD_RELOC_AARCH64_TLSIE_LD32_GOTTPREL_LO12_NC \
178 || (R_TYPE) == BFD_RELOC_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC \
179 || (R_TYPE) == BFD_RELOC_AARCH64_TLSIE_LD_GOTTPREL_PREL19 \
180 || (R_TYPE) == BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G0_NC \
181 || (R_TYPE) == BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G1 \
182 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_ADD_DTPREL_HI12 \
183 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_ADD_DTPREL_LO12 \
184 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_ADD_DTPREL_LO12_NC \
185 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_ADD_LO12_NC \
186 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_ADR_PAGE21 \
187 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_ADR_PREL21 \
188 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_LDST16_DTPREL_LO12 \
189 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_LDST16_DTPREL_LO12_NC \
190 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_LDST32_DTPREL_LO12 \
191 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_LDST32_DTPREL_LO12_NC \
192 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_LDST64_DTPREL_LO12 \
193 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_LDST64_DTPREL_LO12_NC \
194 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_LDST8_DTPREL_LO12 \
195 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_LDST8_DTPREL_LO12_NC \
196 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_MOVW_DTPREL_G0 \
197 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_MOVW_DTPREL_G0_NC \
198 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_MOVW_DTPREL_G1 \
199 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_MOVW_DTPREL_G1_NC \
200 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_MOVW_DTPREL_G2 \
201 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_ADD_TPREL_HI12 \
202 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_ADD_TPREL_LO12 \
203 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_ADD_TPREL_LO12_NC \
204 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_LDST16_TPREL_LO12 \
205 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_LDST16_TPREL_LO12_NC \
206 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_LDST32_TPREL_LO12 \
207 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_LDST32_TPREL_LO12_NC \
208 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_LDST64_TPREL_LO12 \
209 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_LDST64_TPREL_LO12_NC \
210 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_LDST8_TPREL_LO12 \
211 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_LDST8_TPREL_LO12_NC \
212 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G0 \
213 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G0_NC \
214 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G1 \
215 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G1_NC \
216 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G2 \
217 || (R_TYPE) == BFD_RELOC_AARCH64_TLS_DTPMOD \
218 || (R_TYPE) == BFD_RELOC_AARCH64_TLS_DTPREL \
219 || (R_TYPE) == BFD_RELOC_AARCH64_TLS_TPREL \
220 || IS_AARCH64_TLSDESC_RELOC ((R_TYPE)))
221
222 #define IS_AARCH64_TLS_RELAX_RELOC(R_TYPE) \
223 ((R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_ADD \
224 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_ADD_LO12 \
225 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_ADR_PAGE21 \
226 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_ADR_PREL21 \
227 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_CALL \
228 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_LD_PREL19 \
229 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_LDNN_LO12_NC \
230 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_LDR \
231 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_OFF_G0_NC \
232 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_OFF_G1 \
233 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_LDR \
234 || (R_TYPE) == BFD_RELOC_AARCH64_TLSGD_ADR_PAGE21 \
235 || (R_TYPE) == BFD_RELOC_AARCH64_TLSGD_ADR_PREL21 \
236 || (R_TYPE) == BFD_RELOC_AARCH64_TLSGD_ADD_LO12_NC \
237 || (R_TYPE) == BFD_RELOC_AARCH64_TLSGD_MOVW_G0_NC \
238 || (R_TYPE) == BFD_RELOC_AARCH64_TLSGD_MOVW_G1 \
239 || (R_TYPE) == BFD_RELOC_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21 \
240 || (R_TYPE) == BFD_RELOC_AARCH64_TLSIE_LD_GOTTPREL_PREL19 \
241 || (R_TYPE) == BFD_RELOC_AARCH64_TLSIE_LDNN_GOTTPREL_LO12_NC \
242 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_ADD_LO12_NC \
243 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_ADR_PAGE21 \
244 || (R_TYPE) == BFD_RELOC_AARCH64_TLSLD_ADR_PREL21)
245
246 #define IS_AARCH64_TLSDESC_RELOC(R_TYPE) \
247 ((R_TYPE) == BFD_RELOC_AARCH64_TLSDESC \
248 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_ADD \
249 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_ADD_LO12 \
250 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_ADR_PAGE21 \
251 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_ADR_PREL21 \
252 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_CALL \
253 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_LD32_LO12_NC \
254 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_LD64_LO12 \
255 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_LDR \
256 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_LD_PREL19 \
257 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_OFF_G0_NC \
258 || (R_TYPE) == BFD_RELOC_AARCH64_TLSDESC_OFF_G1)
259
260 #define ELIMINATE_COPY_RELOCS 1
261
262 /* Return size of a relocation entry. HTAB is the bfd's
263 elf_aarch64_link_hash_entry. */
264 #define RELOC_SIZE(HTAB) (sizeof (ElfNN_External_Rela))
265
266 /* GOT Entry size - 8 bytes in ELF64 and 4 bytes in ELF32. */
267 #define GOT_ENTRY_SIZE (ARCH_SIZE / 8)
268 #define PLT_ENTRY_SIZE (32)
269 #define PLT_SMALL_ENTRY_SIZE (16)
270 #define PLT_TLSDESC_ENTRY_SIZE (32)
271 /* PLT sizes with BTI insn. */
272 #define PLT_BTI_SMALL_ENTRY_SIZE (24)
273 /* PLT sizes with PAC insn. */
274 #define PLT_PAC_SMALL_ENTRY_SIZE (24)
275 /* PLT sizes with BTI and PAC insn. */
276 #define PLT_BTI_PAC_SMALL_ENTRY_SIZE (24)
277
278 /* Encoding of the nop instruction. */
279 #define INSN_NOP 0xd503201f
280
281 #define aarch64_compute_jump_table_size(htab) \
282 (((htab)->root.srelplt == NULL) ? 0 \
283 : (htab)->root.srelplt->reloc_count * GOT_ENTRY_SIZE)
284
285 /* The first entry in a procedure linkage table looks like this
286 if the distance between the PLTGOT and the PLT is < 4GB use
287 these PLT entries. Note that the dynamic linker gets &PLTGOT[2]
288 in x16 and needs to work out PLTGOT[1] by using an address of
289 [x16,#-GOT_ENTRY_SIZE]. */
290 static const bfd_byte elfNN_aarch64_small_plt0_entry[PLT_ENTRY_SIZE] =
291 {
292 0xf0, 0x7b, 0xbf, 0xa9, /* stp x16, x30, [sp, #-16]! */
293 0x10, 0x00, 0x00, 0x90, /* adrp x16, (GOT+16) */
294 #if ARCH_SIZE == 64
295 0x11, 0x0A, 0x40, 0xf9, /* ldr x17, [x16, #PLT_GOT+0x10] */
296 0x10, 0x42, 0x00, 0x91, /* add x16, x16,#PLT_GOT+0x10 */
297 #else
298 0x11, 0x0A, 0x40, 0xb9, /* ldr w17, [x16, #PLT_GOT+0x8] */
299 0x10, 0x22, 0x00, 0x11, /* add w16, w16,#PLT_GOT+0x8 */
300 #endif
301 0x20, 0x02, 0x1f, 0xd6, /* br x17 */
302 0x1f, 0x20, 0x03, 0xd5, /* nop */
303 0x1f, 0x20, 0x03, 0xd5, /* nop */
304 0x1f, 0x20, 0x03, 0xd5, /* nop */
305 };
306
307 static const bfd_byte elfNN_aarch64_small_plt0_bti_entry[PLT_ENTRY_SIZE] =
308 {
309 0x5f, 0x24, 0x03, 0xd5, /* bti c. */
310 0xf0, 0x7b, 0xbf, 0xa9, /* stp x16, x30, [sp, #-16]! */
311 0x10, 0x00, 0x00, 0x90, /* adrp x16, (GOT+16) */
312 #if ARCH_SIZE == 64
313 0x11, 0x0A, 0x40, 0xf9, /* ldr x17, [x16, #PLT_GOT+0x10] */
314 0x10, 0x42, 0x00, 0x91, /* add x16, x16,#PLT_GOT+0x10 */
315 #else
316 0x11, 0x0A, 0x40, 0xb9, /* ldr w17, [x16, #PLT_GOT+0x8] */
317 0x10, 0x22, 0x00, 0x11, /* add w16, w16,#PLT_GOT+0x8 */
318 #endif
319 0x20, 0x02, 0x1f, 0xd6, /* br x17 */
320 0x1f, 0x20, 0x03, 0xd5, /* nop */
321 0x1f, 0x20, 0x03, 0xd5, /* nop */
322 };
323
324 /* Per function entry in a procedure linkage table looks like this
325 if the distance between the PLTGOT and the PLT is < 4GB use
326 these PLT entries. Use BTI versions of the PLTs when enabled. */
327 static const bfd_byte elfNN_aarch64_small_plt_entry[PLT_SMALL_ENTRY_SIZE] =
328 {
329 0x10, 0x00, 0x00, 0x90, /* adrp x16, PLTGOT + n * 8 */
330 #if ARCH_SIZE == 64
331 0x11, 0x02, 0x40, 0xf9, /* ldr x17, [x16, PLTGOT + n * 8] */
332 0x10, 0x02, 0x00, 0x91, /* add x16, x16, :lo12:PLTGOT + n * 8 */
333 #else
334 0x11, 0x02, 0x40, 0xb9, /* ldr w17, [x16, PLTGOT + n * 4] */
335 0x10, 0x02, 0x00, 0x11, /* add w16, w16, :lo12:PLTGOT + n * 4 */
336 #endif
337 0x20, 0x02, 0x1f, 0xd6, /* br x17. */
338 };
339
340 static const bfd_byte
341 elfNN_aarch64_small_plt_bti_entry[PLT_BTI_SMALL_ENTRY_SIZE] =
342 {
343 0x5f, 0x24, 0x03, 0xd5, /* bti c. */
344 0x10, 0x00, 0x00, 0x90, /* adrp x16, PLTGOT + n * 8 */
345 #if ARCH_SIZE == 64
346 0x11, 0x02, 0x40, 0xf9, /* ldr x17, [x16, PLTGOT + n * 8] */
347 0x10, 0x02, 0x00, 0x91, /* add x16, x16, :lo12:PLTGOT + n * 8 */
348 #else
349 0x11, 0x02, 0x40, 0xb9, /* ldr w17, [x16, PLTGOT + n * 4] */
350 0x10, 0x02, 0x00, 0x11, /* add w16, w16, :lo12:PLTGOT + n * 4 */
351 #endif
352 0x20, 0x02, 0x1f, 0xd6, /* br x17. */
353 0x1f, 0x20, 0x03, 0xd5, /* nop */
354 };
355
356 static const bfd_byte
357 elfNN_aarch64_small_plt_pac_entry[PLT_PAC_SMALL_ENTRY_SIZE] =
358 {
359 0x10, 0x00, 0x00, 0x90, /* adrp x16, PLTGOT + n * 8 */
360 #if ARCH_SIZE == 64
361 0x11, 0x02, 0x40, 0xf9, /* ldr x17, [x16, PLTGOT + n * 8] */
362 0x10, 0x02, 0x00, 0x91, /* add x16, x16, :lo12:PLTGOT + n * 8 */
363 #else
364 0x11, 0x02, 0x40, 0xb9, /* ldr w17, [x16, PLTGOT + n * 4] */
365 0x10, 0x02, 0x00, 0x11, /* add w16, w16, :lo12:PLTGOT + n * 4 */
366 #endif
367 0x9f, 0x21, 0x03, 0xd5, /* autia1716 */
368 0x20, 0x02, 0x1f, 0xd6, /* br x17. */
369 0x1f, 0x20, 0x03, 0xd5, /* nop */
370 };
371
372 static const bfd_byte
373 elfNN_aarch64_small_plt_bti_pac_entry[PLT_BTI_PAC_SMALL_ENTRY_SIZE] =
374 {
375 0x5f, 0x24, 0x03, 0xd5, /* bti c. */
376 0x10, 0x00, 0x00, 0x90, /* adrp x16, PLTGOT + n * 8 */
377 #if ARCH_SIZE == 64
378 0x11, 0x02, 0x40, 0xf9, /* ldr x17, [x16, PLTGOT + n * 8] */
379 0x10, 0x02, 0x00, 0x91, /* add x16, x16, :lo12:PLTGOT + n * 8 */
380 #else
381 0x11, 0x02, 0x40, 0xb9, /* ldr w17, [x16, PLTGOT + n * 4] */
382 0x10, 0x02, 0x00, 0x11, /* add w16, w16, :lo12:PLTGOT + n * 4 */
383 #endif
384 0x9f, 0x21, 0x03, 0xd5, /* autia1716 */
385 0x20, 0x02, 0x1f, 0xd6, /* br x17. */
386 };
387
388 static const bfd_byte
389 elfNN_aarch64_tlsdesc_small_plt_entry[PLT_TLSDESC_ENTRY_SIZE] =
390 {
391 0xe2, 0x0f, 0xbf, 0xa9, /* stp x2, x3, [sp, #-16]! */
392 0x02, 0x00, 0x00, 0x90, /* adrp x2, 0 */
393 0x03, 0x00, 0x00, 0x90, /* adrp x3, 0 */
394 #if ARCH_SIZE == 64
395 0x42, 0x00, 0x40, 0xf9, /* ldr x2, [x2, #0] */
396 0x63, 0x00, 0x00, 0x91, /* add x3, x3, 0 */
397 #else
398 0x42, 0x00, 0x40, 0xb9, /* ldr w2, [x2, #0] */
399 0x63, 0x00, 0x00, 0x11, /* add w3, w3, 0 */
400 #endif
401 0x40, 0x00, 0x1f, 0xd6, /* br x2 */
402 0x1f, 0x20, 0x03, 0xd5, /* nop */
403 0x1f, 0x20, 0x03, 0xd5, /* nop */
404 };
405
406 static const bfd_byte
407 elfNN_aarch64_tlsdesc_small_plt_bti_entry[PLT_TLSDESC_ENTRY_SIZE] =
408 {
409 0x5f, 0x24, 0x03, 0xd5, /* bti c. */
410 0xe2, 0x0f, 0xbf, 0xa9, /* stp x2, x3, [sp, #-16]! */
411 0x02, 0x00, 0x00, 0x90, /* adrp x2, 0 */
412 0x03, 0x00, 0x00, 0x90, /* adrp x3, 0 */
413 #if ARCH_SIZE == 64
414 0x42, 0x00, 0x40, 0xf9, /* ldr x2, [x2, #0] */
415 0x63, 0x00, 0x00, 0x91, /* add x3, x3, 0 */
416 #else
417 0x42, 0x00, 0x40, 0xb9, /* ldr w2, [x2, #0] */
418 0x63, 0x00, 0x00, 0x11, /* add w3, w3, 0 */
419 #endif
420 0x40, 0x00, 0x1f, 0xd6, /* br x2 */
421 0x1f, 0x20, 0x03, 0xd5, /* nop */
422 };
423
424 #define elf_info_to_howto elfNN_aarch64_info_to_howto
425 #define elf_info_to_howto_rel elfNN_aarch64_info_to_howto
426
427 #define AARCH64_ELF_ABI_VERSION 0
428
429 /* In case we're on a 32-bit machine, construct a 64-bit "-1" value. */
430 #define ALL_ONES (~ (bfd_vma) 0)
431
432 /* Indexed by the bfd interal reloc enumerators.
433 Therefore, the table needs to be synced with BFD_RELOC_AARCH64_*
434 in reloc.c. */
435
436 static reloc_howto_type elfNN_aarch64_howto_table[] =
437 {
438 EMPTY_HOWTO (0),
439
440 /* Basic data relocations. */
441
442 /* Deprecated, but retained for backwards compatibility. */
443 HOWTO64 (R_AARCH64_NULL, /* type */
444 0, /* rightshift */
445 3, /* size (0 = byte, 1 = short, 2 = long) */
446 0, /* bitsize */
447 false, /* pc_relative */
448 0, /* bitpos */
449 complain_overflow_dont, /* complain_on_overflow */
450 bfd_elf_generic_reloc, /* special_function */
451 "R_AARCH64_NULL", /* name */
452 false, /* partial_inplace */
453 0, /* src_mask */
454 0, /* dst_mask */
455 false), /* pcrel_offset */
456 HOWTO (R_AARCH64_NONE, /* type */
457 0, /* rightshift */
458 3, /* size (0 = byte, 1 = short, 2 = long) */
459 0, /* bitsize */
460 false, /* pc_relative */
461 0, /* bitpos */
462 complain_overflow_dont, /* complain_on_overflow */
463 bfd_elf_generic_reloc, /* special_function */
464 "R_AARCH64_NONE", /* name */
465 false, /* partial_inplace */
466 0, /* src_mask */
467 0, /* dst_mask */
468 false), /* pcrel_offset */
469
470 /* .xword: (S+A) */
471 HOWTO64 (AARCH64_R (ABS64), /* type */
472 0, /* rightshift */
473 4, /* size (4 = long long) */
474 64, /* bitsize */
475 false, /* pc_relative */
476 0, /* bitpos */
477 complain_overflow_unsigned, /* complain_on_overflow */
478 bfd_elf_generic_reloc, /* special_function */
479 AARCH64_R_STR (ABS64), /* name */
480 false, /* partial_inplace */
481 ALL_ONES, /* src_mask */
482 ALL_ONES, /* dst_mask */
483 false), /* pcrel_offset */
484
485 /* .word: (S+A) */
486 HOWTO (AARCH64_R (ABS32), /* type */
487 0, /* rightshift */
488 2, /* size (0 = byte, 1 = short, 2 = long) */
489 32, /* bitsize */
490 false, /* pc_relative */
491 0, /* bitpos */
492 complain_overflow_unsigned, /* complain_on_overflow */
493 bfd_elf_generic_reloc, /* special_function */
494 AARCH64_R_STR (ABS32), /* name */
495 false, /* partial_inplace */
496 0xffffffff, /* src_mask */
497 0xffffffff, /* dst_mask */
498 false), /* pcrel_offset */
499
500 /* .half: (S+A) */
501 HOWTO (AARCH64_R (ABS16), /* type */
502 0, /* rightshift */
503 1, /* size (0 = byte, 1 = short, 2 = long) */
504 16, /* bitsize */
505 false, /* pc_relative */
506 0, /* bitpos */
507 complain_overflow_unsigned, /* complain_on_overflow */
508 bfd_elf_generic_reloc, /* special_function */
509 AARCH64_R_STR (ABS16), /* name */
510 false, /* partial_inplace */
511 0xffff, /* src_mask */
512 0xffff, /* dst_mask */
513 false), /* pcrel_offset */
514
515 /* .xword: (S+A-P) */
516 HOWTO64 (AARCH64_R (PREL64), /* type */
517 0, /* rightshift */
518 4, /* size (4 = long long) */
519 64, /* bitsize */
520 true, /* pc_relative */
521 0, /* bitpos */
522 complain_overflow_signed, /* complain_on_overflow */
523 bfd_elf_generic_reloc, /* special_function */
524 AARCH64_R_STR (PREL64), /* name */
525 false, /* partial_inplace */
526 ALL_ONES, /* src_mask */
527 ALL_ONES, /* dst_mask */
528 true), /* pcrel_offset */
529
530 /* .word: (S+A-P) */
531 HOWTO (AARCH64_R (PREL32), /* type */
532 0, /* rightshift */
533 2, /* size (0 = byte, 1 = short, 2 = long) */
534 32, /* bitsize */
535 true, /* pc_relative */
536 0, /* bitpos */
537 complain_overflow_signed, /* complain_on_overflow */
538 bfd_elf_generic_reloc, /* special_function */
539 AARCH64_R_STR (PREL32), /* name */
540 false, /* partial_inplace */
541 0xffffffff, /* src_mask */
542 0xffffffff, /* dst_mask */
543 true), /* pcrel_offset */
544
545 /* .half: (S+A-P) */
546 HOWTO (AARCH64_R (PREL16), /* type */
547 0, /* rightshift */
548 1, /* size (0 = byte, 1 = short, 2 = long) */
549 16, /* bitsize */
550 true, /* pc_relative */
551 0, /* bitpos */
552 complain_overflow_signed, /* complain_on_overflow */
553 bfd_elf_generic_reloc, /* special_function */
554 AARCH64_R_STR (PREL16), /* name */
555 false, /* partial_inplace */
556 0xffff, /* src_mask */
557 0xffff, /* dst_mask */
558 true), /* pcrel_offset */
559
560 /* Group relocations to create a 16, 32, 48 or 64 bit
561 unsigned data or abs address inline. */
562
563 /* MOVZ: ((S+A) >> 0) & 0xffff */
564 HOWTO (AARCH64_R (MOVW_UABS_G0), /* type */
565 0, /* rightshift */
566 2, /* size (0 = byte, 1 = short, 2 = long) */
567 16, /* bitsize */
568 false, /* pc_relative */
569 0, /* bitpos */
570 complain_overflow_unsigned, /* complain_on_overflow */
571 bfd_elf_generic_reloc, /* special_function */
572 AARCH64_R_STR (MOVW_UABS_G0), /* name */
573 false, /* partial_inplace */
574 0xffff, /* src_mask */
575 0xffff, /* dst_mask */
576 false), /* pcrel_offset */
577
578 /* MOVK: ((S+A) >> 0) & 0xffff [no overflow check] */
579 HOWTO (AARCH64_R (MOVW_UABS_G0_NC), /* type */
580 0, /* rightshift */
581 2, /* size (0 = byte, 1 = short, 2 = long) */
582 16, /* bitsize */
583 false, /* pc_relative */
584 0, /* bitpos */
585 complain_overflow_dont, /* complain_on_overflow */
586 bfd_elf_generic_reloc, /* special_function */
587 AARCH64_R_STR (MOVW_UABS_G0_NC), /* name */
588 false, /* partial_inplace */
589 0xffff, /* src_mask */
590 0xffff, /* dst_mask */
591 false), /* pcrel_offset */
592
593 /* MOVZ: ((S+A) >> 16) & 0xffff */
594 HOWTO (AARCH64_R (MOVW_UABS_G1), /* type */
595 16, /* rightshift */
596 2, /* size (0 = byte, 1 = short, 2 = long) */
597 16, /* bitsize */
598 false, /* pc_relative */
599 0, /* bitpos */
600 complain_overflow_unsigned, /* complain_on_overflow */
601 bfd_elf_generic_reloc, /* special_function */
602 AARCH64_R_STR (MOVW_UABS_G1), /* name */
603 false, /* partial_inplace */
604 0xffff, /* src_mask */
605 0xffff, /* dst_mask */
606 false), /* pcrel_offset */
607
608 /* MOVK: ((S+A) >> 16) & 0xffff [no overflow check] */
609 HOWTO64 (AARCH64_R (MOVW_UABS_G1_NC), /* type */
610 16, /* rightshift */
611 2, /* size (0 = byte, 1 = short, 2 = long) */
612 16, /* bitsize */
613 false, /* pc_relative */
614 0, /* bitpos */
615 complain_overflow_dont, /* complain_on_overflow */
616 bfd_elf_generic_reloc, /* special_function */
617 AARCH64_R_STR (MOVW_UABS_G1_NC), /* name */
618 false, /* partial_inplace */
619 0xffff, /* src_mask */
620 0xffff, /* dst_mask */
621 false), /* pcrel_offset */
622
623 /* MOVZ: ((S+A) >> 32) & 0xffff */
624 HOWTO64 (AARCH64_R (MOVW_UABS_G2), /* type */
625 32, /* rightshift */
626 2, /* size (0 = byte, 1 = short, 2 = long) */
627 16, /* bitsize */
628 false, /* pc_relative */
629 0, /* bitpos */
630 complain_overflow_unsigned, /* complain_on_overflow */
631 bfd_elf_generic_reloc, /* special_function */
632 AARCH64_R_STR (MOVW_UABS_G2), /* name */
633 false, /* partial_inplace */
634 0xffff, /* src_mask */
635 0xffff, /* dst_mask */
636 false), /* pcrel_offset */
637
638 /* MOVK: ((S+A) >> 32) & 0xffff [no overflow check] */
639 HOWTO64 (AARCH64_R (MOVW_UABS_G2_NC), /* type */
640 32, /* rightshift */
641 2, /* size (0 = byte, 1 = short, 2 = long) */
642 16, /* bitsize */
643 false, /* pc_relative */
644 0, /* bitpos */
645 complain_overflow_dont, /* complain_on_overflow */
646 bfd_elf_generic_reloc, /* special_function */
647 AARCH64_R_STR (MOVW_UABS_G2_NC), /* name */
648 false, /* partial_inplace */
649 0xffff, /* src_mask */
650 0xffff, /* dst_mask */
651 false), /* pcrel_offset */
652
653 /* MOVZ: ((S+A) >> 48) & 0xffff */
654 HOWTO64 (AARCH64_R (MOVW_UABS_G3), /* type */
655 48, /* rightshift */
656 2, /* size (0 = byte, 1 = short, 2 = long) */
657 16, /* bitsize */
658 false, /* pc_relative */
659 0, /* bitpos */
660 complain_overflow_unsigned, /* complain_on_overflow */
661 bfd_elf_generic_reloc, /* special_function */
662 AARCH64_R_STR (MOVW_UABS_G3), /* name */
663 false, /* partial_inplace */
664 0xffff, /* src_mask */
665 0xffff, /* dst_mask */
666 false), /* pcrel_offset */
667
668 /* Group relocations to create high part of a 16, 32, 48 or 64 bit
669 signed data or abs address inline. Will change instruction
670 to MOVN or MOVZ depending on sign of calculated value. */
671
672 /* MOV[ZN]: ((S+A) >> 0) & 0xffff */
673 HOWTO (AARCH64_R (MOVW_SABS_G0), /* type */
674 0, /* rightshift */
675 2, /* size (0 = byte, 1 = short, 2 = long) */
676 17, /* bitsize */
677 false, /* pc_relative */
678 0, /* bitpos */
679 complain_overflow_signed, /* complain_on_overflow */
680 bfd_elf_generic_reloc, /* special_function */
681 AARCH64_R_STR (MOVW_SABS_G0), /* name */
682 false, /* partial_inplace */
683 0xffff, /* src_mask */
684 0xffff, /* dst_mask */
685 false), /* pcrel_offset */
686
687 /* MOV[ZN]: ((S+A) >> 16) & 0xffff */
688 HOWTO64 (AARCH64_R (MOVW_SABS_G1), /* type */
689 16, /* rightshift */
690 2, /* size (0 = byte, 1 = short, 2 = long) */
691 17, /* bitsize */
692 false, /* pc_relative */
693 0, /* bitpos */
694 complain_overflow_signed, /* complain_on_overflow */
695 bfd_elf_generic_reloc, /* special_function */
696 AARCH64_R_STR (MOVW_SABS_G1), /* name */
697 false, /* partial_inplace */
698 0xffff, /* src_mask */
699 0xffff, /* dst_mask */
700 false), /* pcrel_offset */
701
702 /* MOV[ZN]: ((S+A) >> 32) & 0xffff */
703 HOWTO64 (AARCH64_R (MOVW_SABS_G2), /* type */
704 32, /* rightshift */
705 2, /* size (0 = byte, 1 = short, 2 = long) */
706 17, /* bitsize */
707 false, /* pc_relative */
708 0, /* bitpos */
709 complain_overflow_signed, /* complain_on_overflow */
710 bfd_elf_generic_reloc, /* special_function */
711 AARCH64_R_STR (MOVW_SABS_G2), /* name */
712 false, /* partial_inplace */
713 0xffff, /* src_mask */
714 0xffff, /* dst_mask */
715 false), /* pcrel_offset */
716
717 /* Group relocations to create a 16, 32, 48 or 64 bit
718 PC relative address inline. */
719
720 /* MOV[NZ]: ((S+A-P) >> 0) & 0xffff */
721 HOWTO (AARCH64_R (MOVW_PREL_G0), /* type */
722 0, /* rightshift */
723 2, /* size (0 = byte, 1 = short, 2 = long) */
724 17, /* bitsize */
725 true, /* pc_relative */
726 0, /* bitpos */
727 complain_overflow_signed, /* complain_on_overflow */
728 bfd_elf_generic_reloc, /* special_function */
729 AARCH64_R_STR (MOVW_PREL_G0), /* name */
730 false, /* partial_inplace */
731 0xffff, /* src_mask */
732 0xffff, /* dst_mask */
733 true), /* pcrel_offset */
734
735 /* MOVK: ((S+A-P) >> 0) & 0xffff [no overflow check] */
736 HOWTO (AARCH64_R (MOVW_PREL_G0_NC), /* type */
737 0, /* rightshift */
738 2, /* size (0 = byte, 1 = short, 2 = long) */
739 16, /* bitsize */
740 true, /* pc_relative */
741 0, /* bitpos */
742 complain_overflow_dont, /* complain_on_overflow */
743 bfd_elf_generic_reloc, /* special_function */
744 AARCH64_R_STR (MOVW_PREL_G0_NC), /* name */
745 false, /* partial_inplace */
746 0xffff, /* src_mask */
747 0xffff, /* dst_mask */
748 true), /* pcrel_offset */
749
750 /* MOV[NZ]: ((S+A-P) >> 16) & 0xffff */
751 HOWTO (AARCH64_R (MOVW_PREL_G1), /* type */
752 16, /* rightshift */
753 2, /* size (0 = byte, 1 = short, 2 = long) */
754 17, /* bitsize */
755 true, /* pc_relative */
756 0, /* bitpos */
757 complain_overflow_signed, /* complain_on_overflow */
758 bfd_elf_generic_reloc, /* special_function */
759 AARCH64_R_STR (MOVW_PREL_G1), /* name */
760 false, /* partial_inplace */
761 0xffff, /* src_mask */
762 0xffff, /* dst_mask */
763 true), /* pcrel_offset */
764
765 /* MOVK: ((S+A-P) >> 16) & 0xffff [no overflow check] */
766 HOWTO64 (AARCH64_R (MOVW_PREL_G1_NC), /* type */
767 16, /* rightshift */
768 2, /* size (0 = byte, 1 = short, 2 = long) */
769 16, /* bitsize */
770 true, /* pc_relative */
771 0, /* bitpos */
772 complain_overflow_dont, /* complain_on_overflow */
773 bfd_elf_generic_reloc, /* special_function */
774 AARCH64_R_STR (MOVW_PREL_G1_NC), /* name */
775 false, /* partial_inplace */
776 0xffff, /* src_mask */
777 0xffff, /* dst_mask */
778 true), /* pcrel_offset */
779
780 /* MOV[NZ]: ((S+A-P) >> 32) & 0xffff */
781 HOWTO64 (AARCH64_R (MOVW_PREL_G2), /* type */
782 32, /* rightshift */
783 2, /* size (0 = byte, 1 = short, 2 = long) */
784 17, /* bitsize */
785 true, /* pc_relative */
786 0, /* bitpos */
787 complain_overflow_signed, /* complain_on_overflow */
788 bfd_elf_generic_reloc, /* special_function */
789 AARCH64_R_STR (MOVW_PREL_G2), /* name */
790 false, /* partial_inplace */
791 0xffff, /* src_mask */
792 0xffff, /* dst_mask */
793 true), /* pcrel_offset */
794
795 /* MOVK: ((S+A-P) >> 32) & 0xffff [no overflow check] */
796 HOWTO64 (AARCH64_R (MOVW_PREL_G2_NC), /* type */
797 32, /* rightshift */
798 2, /* size (0 = byte, 1 = short, 2 = long) */
799 16, /* bitsize */
800 true, /* pc_relative */
801 0, /* bitpos */
802 complain_overflow_dont, /* complain_on_overflow */
803 bfd_elf_generic_reloc, /* special_function */
804 AARCH64_R_STR (MOVW_PREL_G2_NC), /* name */
805 false, /* partial_inplace */
806 0xffff, /* src_mask */
807 0xffff, /* dst_mask */
808 true), /* pcrel_offset */
809
810 /* MOV[NZ]: ((S+A-P) >> 48) & 0xffff */
811 HOWTO64 (AARCH64_R (MOVW_PREL_G3), /* type */
812 48, /* rightshift */
813 2, /* size (0 = byte, 1 = short, 2 = long) */
814 16, /* bitsize */
815 true, /* pc_relative */
816 0, /* bitpos */
817 complain_overflow_dont, /* complain_on_overflow */
818 bfd_elf_generic_reloc, /* special_function */
819 AARCH64_R_STR (MOVW_PREL_G3), /* name */
820 false, /* partial_inplace */
821 0xffff, /* src_mask */
822 0xffff, /* dst_mask */
823 true), /* pcrel_offset */
824
825 /* Relocations to generate 19, 21 and 33 bit PC-relative load/store
826 addresses: PG(x) is (x & ~0xfff). */
827
828 /* LD-lit: ((S+A-P) >> 2) & 0x7ffff */
829 HOWTO (AARCH64_R (LD_PREL_LO19), /* type */
830 2, /* rightshift */
831 2, /* size (0 = byte, 1 = short, 2 = long) */
832 19, /* bitsize */
833 true, /* pc_relative */
834 0, /* bitpos */
835 complain_overflow_signed, /* complain_on_overflow */
836 bfd_elf_generic_reloc, /* special_function */
837 AARCH64_R_STR (LD_PREL_LO19), /* name */
838 false, /* partial_inplace */
839 0x7ffff, /* src_mask */
840 0x7ffff, /* dst_mask */
841 true), /* pcrel_offset */
842
843 /* ADR: (S+A-P) & 0x1fffff */
844 HOWTO (AARCH64_R (ADR_PREL_LO21), /* type */
845 0, /* rightshift */
846 2, /* size (0 = byte, 1 = short, 2 = long) */
847 21, /* bitsize */
848 true, /* pc_relative */
849 0, /* bitpos */
850 complain_overflow_signed, /* complain_on_overflow */
851 bfd_elf_generic_reloc, /* special_function */
852 AARCH64_R_STR (ADR_PREL_LO21), /* name */
853 false, /* partial_inplace */
854 0x1fffff, /* src_mask */
855 0x1fffff, /* dst_mask */
856 true), /* pcrel_offset */
857
858 /* ADRP: ((PG(S+A)-PG(P)) >> 12) & 0x1fffff */
859 HOWTO (AARCH64_R (ADR_PREL_PG_HI21), /* type */
860 12, /* rightshift */
861 2, /* size (0 = byte, 1 = short, 2 = long) */
862 21, /* bitsize */
863 true, /* pc_relative */
864 0, /* bitpos */
865 complain_overflow_signed, /* complain_on_overflow */
866 bfd_elf_generic_reloc, /* special_function */
867 AARCH64_R_STR (ADR_PREL_PG_HI21), /* name */
868 false, /* partial_inplace */
869 0x1fffff, /* src_mask */
870 0x1fffff, /* dst_mask */
871 true), /* pcrel_offset */
872
873 /* ADRP: ((PG(S+A)-PG(P)) >> 12) & 0x1fffff [no overflow check] */
874 HOWTO64 (AARCH64_R (ADR_PREL_PG_HI21_NC), /* type */
875 12, /* rightshift */
876 2, /* size (0 = byte, 1 = short, 2 = long) */
877 21, /* bitsize */
878 true, /* pc_relative */
879 0, /* bitpos */
880 complain_overflow_dont, /* complain_on_overflow */
881 bfd_elf_generic_reloc, /* special_function */
882 AARCH64_R_STR (ADR_PREL_PG_HI21_NC), /* name */
883 false, /* partial_inplace */
884 0x1fffff, /* src_mask */
885 0x1fffff, /* dst_mask */
886 true), /* pcrel_offset */
887
888 /* ADD: (S+A) & 0xfff [no overflow check] */
889 HOWTO (AARCH64_R (ADD_ABS_LO12_NC), /* type */
890 0, /* rightshift */
891 2, /* size (0 = byte, 1 = short, 2 = long) */
892 12, /* bitsize */
893 false, /* pc_relative */
894 10, /* bitpos */
895 complain_overflow_dont, /* complain_on_overflow */
896 bfd_elf_generic_reloc, /* special_function */
897 AARCH64_R_STR (ADD_ABS_LO12_NC), /* name */
898 false, /* partial_inplace */
899 0x3ffc00, /* src_mask */
900 0x3ffc00, /* dst_mask */
901 false), /* pcrel_offset */
902
903 /* LD/ST8: (S+A) & 0xfff */
904 HOWTO (AARCH64_R (LDST8_ABS_LO12_NC), /* type */
905 0, /* rightshift */
906 2, /* size (0 = byte, 1 = short, 2 = long) */
907 12, /* bitsize */
908 false, /* pc_relative */
909 0, /* bitpos */
910 complain_overflow_dont, /* complain_on_overflow */
911 bfd_elf_generic_reloc, /* special_function */
912 AARCH64_R_STR (LDST8_ABS_LO12_NC), /* name */
913 false, /* partial_inplace */
914 0xfff, /* src_mask */
915 0xfff, /* dst_mask */
916 false), /* pcrel_offset */
917
918 /* Relocations for control-flow instructions. */
919
920 /* TBZ/NZ: ((S+A-P) >> 2) & 0x3fff */
921 HOWTO (AARCH64_R (TSTBR14), /* type */
922 2, /* rightshift */
923 2, /* size (0 = byte, 1 = short, 2 = long) */
924 14, /* bitsize */
925 true, /* pc_relative */
926 0, /* bitpos */
927 complain_overflow_signed, /* complain_on_overflow */
928 bfd_elf_generic_reloc, /* special_function */
929 AARCH64_R_STR (TSTBR14), /* name */
930 false, /* partial_inplace */
931 0x3fff, /* src_mask */
932 0x3fff, /* dst_mask */
933 true), /* pcrel_offset */
934
935 /* B.cond: ((S+A-P) >> 2) & 0x7ffff */
936 HOWTO (AARCH64_R (CONDBR19), /* type */
937 2, /* rightshift */
938 2, /* size (0 = byte, 1 = short, 2 = long) */
939 19, /* bitsize */
940 true, /* pc_relative */
941 0, /* bitpos */
942 complain_overflow_signed, /* complain_on_overflow */
943 bfd_elf_generic_reloc, /* special_function */
944 AARCH64_R_STR (CONDBR19), /* name */
945 false, /* partial_inplace */
946 0x7ffff, /* src_mask */
947 0x7ffff, /* dst_mask */
948 true), /* pcrel_offset */
949
950 /* B: ((S+A-P) >> 2) & 0x3ffffff */
951 HOWTO (AARCH64_R (JUMP26), /* type */
952 2, /* rightshift */
953 2, /* size (0 = byte, 1 = short, 2 = long) */
954 26, /* bitsize */
955 true, /* pc_relative */
956 0, /* bitpos */
957 complain_overflow_signed, /* complain_on_overflow */
958 bfd_elf_generic_reloc, /* special_function */
959 AARCH64_R_STR (JUMP26), /* name */
960 false, /* partial_inplace */
961 0x3ffffff, /* src_mask */
962 0x3ffffff, /* dst_mask */
963 true), /* pcrel_offset */
964
965 /* BL: ((S+A-P) >> 2) & 0x3ffffff */
966 HOWTO (AARCH64_R (CALL26), /* type */
967 2, /* rightshift */
968 2, /* size (0 = byte, 1 = short, 2 = long) */
969 26, /* bitsize */
970 true, /* pc_relative */
971 0, /* bitpos */
972 complain_overflow_signed, /* complain_on_overflow */
973 bfd_elf_generic_reloc, /* special_function */
974 AARCH64_R_STR (CALL26), /* name */
975 false, /* partial_inplace */
976 0x3ffffff, /* src_mask */
977 0x3ffffff, /* dst_mask */
978 true), /* pcrel_offset */
979
980 /* LD/ST16: (S+A) & 0xffe */
981 HOWTO (AARCH64_R (LDST16_ABS_LO12_NC), /* type */
982 1, /* rightshift */
983 2, /* size (0 = byte, 1 = short, 2 = long) */
984 12, /* bitsize */
985 false, /* pc_relative */
986 0, /* bitpos */
987 complain_overflow_dont, /* complain_on_overflow */
988 bfd_elf_generic_reloc, /* special_function */
989 AARCH64_R_STR (LDST16_ABS_LO12_NC), /* name */
990 false, /* partial_inplace */
991 0xffe, /* src_mask */
992 0xffe, /* dst_mask */
993 false), /* pcrel_offset */
994
995 /* LD/ST32: (S+A) & 0xffc */
996 HOWTO (AARCH64_R (LDST32_ABS_LO12_NC), /* type */
997 2, /* rightshift */
998 2, /* size (0 = byte, 1 = short, 2 = long) */
999 12, /* bitsize */
1000 false, /* pc_relative */
1001 0, /* bitpos */
1002 complain_overflow_dont, /* complain_on_overflow */
1003 bfd_elf_generic_reloc, /* special_function */
1004 AARCH64_R_STR (LDST32_ABS_LO12_NC), /* name */
1005 false, /* partial_inplace */
1006 0xffc, /* src_mask */
1007 0xffc, /* dst_mask */
1008 false), /* pcrel_offset */
1009
1010 /* LD/ST64: (S+A) & 0xff8 */
1011 HOWTO (AARCH64_R (LDST64_ABS_LO12_NC), /* type */
1012 3, /* rightshift */
1013 2, /* size (0 = byte, 1 = short, 2 = long) */
1014 12, /* bitsize */
1015 false, /* pc_relative */
1016 0, /* bitpos */
1017 complain_overflow_dont, /* complain_on_overflow */
1018 bfd_elf_generic_reloc, /* special_function */
1019 AARCH64_R_STR (LDST64_ABS_LO12_NC), /* name */
1020 false, /* partial_inplace */
1021 0xff8, /* src_mask */
1022 0xff8, /* dst_mask */
1023 false), /* pcrel_offset */
1024
1025 /* LD/ST128: (S+A) & 0xff0 */
1026 HOWTO (AARCH64_R (LDST128_ABS_LO12_NC), /* type */
1027 4, /* rightshift */
1028 2, /* size (0 = byte, 1 = short, 2 = long) */
1029 12, /* bitsize */
1030 false, /* pc_relative */
1031 0, /* bitpos */
1032 complain_overflow_dont, /* complain_on_overflow */
1033 bfd_elf_generic_reloc, /* special_function */
1034 AARCH64_R_STR (LDST128_ABS_LO12_NC), /* name */
1035 false, /* partial_inplace */
1036 0xff0, /* src_mask */
1037 0xff0, /* dst_mask */
1038 false), /* pcrel_offset */
1039
1040 /* Set a load-literal immediate field to bits
1041 0x1FFFFC of G(S)-P */
1042 HOWTO (AARCH64_R (GOT_LD_PREL19), /* type */
1043 2, /* rightshift */
1044 2, /* size (0 = byte,1 = short,2 = long) */
1045 19, /* bitsize */
1046 true, /* pc_relative */
1047 0, /* bitpos */
1048 complain_overflow_signed, /* complain_on_overflow */
1049 bfd_elf_generic_reloc, /* special_function */
1050 AARCH64_R_STR (GOT_LD_PREL19), /* name */
1051 false, /* partial_inplace */
1052 0xffffe0, /* src_mask */
1053 0xffffe0, /* dst_mask */
1054 true), /* pcrel_offset */
1055
1056 /* Get to the page for the GOT entry for the symbol
1057 (G(S) - P) using an ADRP instruction. */
1058 HOWTO (AARCH64_R (ADR_GOT_PAGE), /* type */
1059 12, /* rightshift */
1060 2, /* size (0 = byte, 1 = short, 2 = long) */
1061 21, /* bitsize */
1062 true, /* pc_relative */
1063 0, /* bitpos */
1064 complain_overflow_dont, /* complain_on_overflow */
1065 bfd_elf_generic_reloc, /* special_function */
1066 AARCH64_R_STR (ADR_GOT_PAGE), /* name */
1067 false, /* partial_inplace */
1068 0x1fffff, /* src_mask */
1069 0x1fffff, /* dst_mask */
1070 true), /* pcrel_offset */
1071
1072 /* LD64: GOT offset G(S) & 0xff8 */
1073 HOWTO64 (AARCH64_R (LD64_GOT_LO12_NC), /* type */
1074 3, /* rightshift */
1075 2, /* size (0 = byte, 1 = short, 2 = long) */
1076 12, /* bitsize */
1077 false, /* pc_relative */
1078 0, /* bitpos */
1079 complain_overflow_dont, /* complain_on_overflow */
1080 bfd_elf_generic_reloc, /* special_function */
1081 AARCH64_R_STR (LD64_GOT_LO12_NC), /* name */
1082 false, /* partial_inplace */
1083 0xff8, /* src_mask */
1084 0xff8, /* dst_mask */
1085 false), /* pcrel_offset */
1086
1087 /* LD32: GOT offset G(S) & 0xffc */
1088 HOWTO32 (AARCH64_R (LD32_GOT_LO12_NC), /* type */
1089 2, /* rightshift */
1090 2, /* size (0 = byte, 1 = short, 2 = long) */
1091 12, /* bitsize */
1092 false, /* pc_relative */
1093 0, /* bitpos */
1094 complain_overflow_dont, /* complain_on_overflow */
1095 bfd_elf_generic_reloc, /* special_function */
1096 AARCH64_R_STR (LD32_GOT_LO12_NC), /* name */
1097 false, /* partial_inplace */
1098 0xffc, /* src_mask */
1099 0xffc, /* dst_mask */
1100 false), /* pcrel_offset */
1101
1102 /* Lower 16 bits of GOT offset for the symbol. */
1103 HOWTO64 (AARCH64_R (MOVW_GOTOFF_G0_NC), /* type */
1104 0, /* rightshift */
1105 2, /* size (0 = byte, 1 = short, 2 = long) */
1106 16, /* bitsize */
1107 false, /* pc_relative */
1108 0, /* bitpos */
1109 complain_overflow_dont, /* complain_on_overflow */
1110 bfd_elf_generic_reloc, /* special_function */
1111 AARCH64_R_STR (MOVW_GOTOFF_G0_NC), /* name */
1112 false, /* partial_inplace */
1113 0xffff, /* src_mask */
1114 0xffff, /* dst_mask */
1115 false), /* pcrel_offset */
1116
1117 /* Higher 16 bits of GOT offset for the symbol. */
1118 HOWTO64 (AARCH64_R (MOVW_GOTOFF_G1), /* type */
1119 16, /* rightshift */
1120 2, /* size (0 = byte, 1 = short, 2 = long) */
1121 16, /* bitsize */
1122 false, /* pc_relative */
1123 0, /* bitpos */
1124 complain_overflow_unsigned, /* complain_on_overflow */
1125 bfd_elf_generic_reloc, /* special_function */
1126 AARCH64_R_STR (MOVW_GOTOFF_G1), /* name */
1127 false, /* partial_inplace */
1128 0xffff, /* src_mask */
1129 0xffff, /* dst_mask */
1130 false), /* pcrel_offset */
1131
1132 /* LD64: GOT offset for the symbol. */
1133 HOWTO64 (AARCH64_R (LD64_GOTOFF_LO15), /* type */
1134 3, /* rightshift */
1135 2, /* size (0 = byte, 1 = short, 2 = long) */
1136 12, /* bitsize */
1137 false, /* pc_relative */
1138 0, /* bitpos */
1139 complain_overflow_unsigned, /* complain_on_overflow */
1140 bfd_elf_generic_reloc, /* special_function */
1141 AARCH64_R_STR (LD64_GOTOFF_LO15), /* name */
1142 false, /* partial_inplace */
1143 0x7ff8, /* src_mask */
1144 0x7ff8, /* dst_mask */
1145 false), /* pcrel_offset */
1146
1147 /* LD32: GOT offset to the page address of GOT table.
1148 (G(S) - PAGE (_GLOBAL_OFFSET_TABLE_)) & 0x5ffc. */
1149 HOWTO32 (AARCH64_R (LD32_GOTPAGE_LO14), /* type */
1150 2, /* rightshift */
1151 2, /* size (0 = byte, 1 = short, 2 = long) */
1152 12, /* bitsize */
1153 false, /* pc_relative */
1154 0, /* bitpos */
1155 complain_overflow_unsigned, /* complain_on_overflow */
1156 bfd_elf_generic_reloc, /* special_function */
1157 AARCH64_R_STR (LD32_GOTPAGE_LO14), /* name */
1158 false, /* partial_inplace */
1159 0x5ffc, /* src_mask */
1160 0x5ffc, /* dst_mask */
1161 false), /* pcrel_offset */
1162
1163 /* LD64: GOT offset to the page address of GOT table.
1164 (G(S) - PAGE (_GLOBAL_OFFSET_TABLE_)) & 0x7ff8. */
1165 HOWTO64 (AARCH64_R (LD64_GOTPAGE_LO15), /* type */
1166 3, /* rightshift */
1167 2, /* size (0 = byte, 1 = short, 2 = long) */
1168 12, /* bitsize */
1169 false, /* pc_relative */
1170 0, /* bitpos */
1171 complain_overflow_unsigned, /* complain_on_overflow */
1172 bfd_elf_generic_reloc, /* special_function */
1173 AARCH64_R_STR (LD64_GOTPAGE_LO15), /* name */
1174 false, /* partial_inplace */
1175 0x7ff8, /* src_mask */
1176 0x7ff8, /* dst_mask */
1177 false), /* pcrel_offset */
1178
1179 /* Get to the page for the GOT entry for the symbol
1180 (G(S) - P) using an ADRP instruction. */
1181 HOWTO (AARCH64_R (TLSGD_ADR_PAGE21), /* type */
1182 12, /* rightshift */
1183 2, /* size (0 = byte, 1 = short, 2 = long) */
1184 21, /* bitsize */
1185 true, /* pc_relative */
1186 0, /* bitpos */
1187 complain_overflow_dont, /* complain_on_overflow */
1188 bfd_elf_generic_reloc, /* special_function */
1189 AARCH64_R_STR (TLSGD_ADR_PAGE21), /* name */
1190 false, /* partial_inplace */
1191 0x1fffff, /* src_mask */
1192 0x1fffff, /* dst_mask */
1193 true), /* pcrel_offset */
1194
1195 HOWTO (AARCH64_R (TLSGD_ADR_PREL21), /* type */
1196 0, /* rightshift */
1197 2, /* size (0 = byte, 1 = short, 2 = long) */
1198 21, /* bitsize */
1199 true, /* pc_relative */
1200 0, /* bitpos */
1201 complain_overflow_dont, /* complain_on_overflow */
1202 bfd_elf_generic_reloc, /* special_function */
1203 AARCH64_R_STR (TLSGD_ADR_PREL21), /* name */
1204 false, /* partial_inplace */
1205 0x1fffff, /* src_mask */
1206 0x1fffff, /* dst_mask */
1207 true), /* pcrel_offset */
1208
1209 /* ADD: GOT offset G(S) & 0xff8 [no overflow check] */
1210 HOWTO (AARCH64_R (TLSGD_ADD_LO12_NC), /* type */
1211 0, /* rightshift */
1212 2, /* size (0 = byte, 1 = short, 2 = long) */
1213 12, /* bitsize */
1214 false, /* pc_relative */
1215 0, /* bitpos */
1216 complain_overflow_dont, /* complain_on_overflow */
1217 bfd_elf_generic_reloc, /* special_function */
1218 AARCH64_R_STR (TLSGD_ADD_LO12_NC), /* name */
1219 false, /* partial_inplace */
1220 0xfff, /* src_mask */
1221 0xfff, /* dst_mask */
1222 false), /* pcrel_offset */
1223
1224 /* Lower 16 bits of GOT offset to tls_index. */
1225 HOWTO64 (AARCH64_R (TLSGD_MOVW_G0_NC), /* type */
1226 0, /* rightshift */
1227 2, /* size (0 = byte, 1 = short, 2 = long) */
1228 16, /* bitsize */
1229 false, /* pc_relative */
1230 0, /* bitpos */
1231 complain_overflow_dont, /* complain_on_overflow */
1232 bfd_elf_generic_reloc, /* special_function */
1233 AARCH64_R_STR (TLSGD_MOVW_G0_NC), /* name */
1234 false, /* partial_inplace */
1235 0xffff, /* src_mask */
1236 0xffff, /* dst_mask */
1237 false), /* pcrel_offset */
1238
1239 /* Higher 16 bits of GOT offset to tls_index. */
1240 HOWTO64 (AARCH64_R (TLSGD_MOVW_G1), /* type */
1241 16, /* rightshift */
1242 2, /* size (0 = byte, 1 = short, 2 = long) */
1243 16, /* bitsize */
1244 false, /* pc_relative */
1245 0, /* bitpos */
1246 complain_overflow_unsigned, /* complain_on_overflow */
1247 bfd_elf_generic_reloc, /* special_function */
1248 AARCH64_R_STR (TLSGD_MOVW_G1), /* name */
1249 false, /* partial_inplace */
1250 0xffff, /* src_mask */
1251 0xffff, /* dst_mask */
1252 false), /* pcrel_offset */
1253
1254 HOWTO (AARCH64_R (TLSIE_ADR_GOTTPREL_PAGE21), /* type */
1255 12, /* rightshift */
1256 2, /* size (0 = byte, 1 = short, 2 = long) */
1257 21, /* bitsize */
1258 false, /* pc_relative */
1259 0, /* bitpos */
1260 complain_overflow_dont, /* complain_on_overflow */
1261 bfd_elf_generic_reloc, /* special_function */
1262 AARCH64_R_STR (TLSIE_ADR_GOTTPREL_PAGE21), /* name */
1263 false, /* partial_inplace */
1264 0x1fffff, /* src_mask */
1265 0x1fffff, /* dst_mask */
1266 false), /* pcrel_offset */
1267
1268 HOWTO64 (AARCH64_R (TLSIE_LD64_GOTTPREL_LO12_NC), /* type */
1269 3, /* rightshift */
1270 2, /* size (0 = byte, 1 = short, 2 = long) */
1271 12, /* bitsize */
1272 false, /* pc_relative */
1273 0, /* bitpos */
1274 complain_overflow_dont, /* complain_on_overflow */
1275 bfd_elf_generic_reloc, /* special_function */
1276 AARCH64_R_STR (TLSIE_LD64_GOTTPREL_LO12_NC), /* name */
1277 false, /* partial_inplace */
1278 0xff8, /* src_mask */
1279 0xff8, /* dst_mask */
1280 false), /* pcrel_offset */
1281
1282 HOWTO32 (AARCH64_R (TLSIE_LD32_GOTTPREL_LO12_NC), /* type */
1283 2, /* rightshift */
1284 2, /* size (0 = byte, 1 = short, 2 = long) */
1285 12, /* bitsize */
1286 false, /* pc_relative */
1287 0, /* bitpos */
1288 complain_overflow_dont, /* complain_on_overflow */
1289 bfd_elf_generic_reloc, /* special_function */
1290 AARCH64_R_STR (TLSIE_LD32_GOTTPREL_LO12_NC), /* name */
1291 false, /* partial_inplace */
1292 0xffc, /* src_mask */
1293 0xffc, /* dst_mask */
1294 false), /* pcrel_offset */
1295
1296 HOWTO (AARCH64_R (TLSIE_LD_GOTTPREL_PREL19), /* type */
1297 2, /* rightshift */
1298 2, /* size (0 = byte, 1 = short, 2 = long) */
1299 19, /* bitsize */
1300 false, /* pc_relative */
1301 0, /* bitpos */
1302 complain_overflow_dont, /* complain_on_overflow */
1303 bfd_elf_generic_reloc, /* special_function */
1304 AARCH64_R_STR (TLSIE_LD_GOTTPREL_PREL19), /* name */
1305 false, /* partial_inplace */
1306 0x1ffffc, /* src_mask */
1307 0x1ffffc, /* dst_mask */
1308 false), /* pcrel_offset */
1309
1310 HOWTO64 (AARCH64_R (TLSIE_MOVW_GOTTPREL_G0_NC), /* type */
1311 0, /* rightshift */
1312 2, /* size (0 = byte, 1 = short, 2 = long) */
1313 16, /* bitsize */
1314 false, /* pc_relative */
1315 0, /* bitpos */
1316 complain_overflow_dont, /* complain_on_overflow */
1317 bfd_elf_generic_reloc, /* special_function */
1318 AARCH64_R_STR (TLSIE_MOVW_GOTTPREL_G0_NC), /* name */
1319 false, /* partial_inplace */
1320 0xffff, /* src_mask */
1321 0xffff, /* dst_mask */
1322 false), /* pcrel_offset */
1323
1324 HOWTO64 (AARCH64_R (TLSIE_MOVW_GOTTPREL_G1), /* type */
1325 16, /* rightshift */
1326 2, /* size (0 = byte, 1 = short, 2 = long) */
1327 16, /* bitsize */
1328 false, /* pc_relative */
1329 0, /* bitpos */
1330 complain_overflow_unsigned, /* complain_on_overflow */
1331 bfd_elf_generic_reloc, /* special_function */
1332 AARCH64_R_STR (TLSIE_MOVW_GOTTPREL_G1), /* name */
1333 false, /* partial_inplace */
1334 0xffff, /* src_mask */
1335 0xffff, /* dst_mask */
1336 false), /* pcrel_offset */
1337
1338 /* ADD: bit[23:12] of byte offset to module TLS base address. */
1339 HOWTO (AARCH64_R (TLSLD_ADD_DTPREL_HI12), /* type */
1340 12, /* rightshift */
1341 2, /* size (0 = byte, 1 = short, 2 = long) */
1342 12, /* bitsize */
1343 false, /* pc_relative */
1344 0, /* bitpos */
1345 complain_overflow_unsigned, /* complain_on_overflow */
1346 bfd_elf_generic_reloc, /* special_function */
1347 AARCH64_R_STR (TLSLD_ADD_DTPREL_HI12), /* name */
1348 false, /* partial_inplace */
1349 0xfff, /* src_mask */
1350 0xfff, /* dst_mask */
1351 false), /* pcrel_offset */
1352
1353 /* Unsigned 12 bit byte offset to module TLS base address. */
1354 HOWTO (AARCH64_R (TLSLD_ADD_DTPREL_LO12), /* type */
1355 0, /* rightshift */
1356 2, /* size (0 = byte, 1 = short, 2 = long) */
1357 12, /* bitsize */
1358 false, /* pc_relative */
1359 0, /* bitpos */
1360 complain_overflow_unsigned, /* complain_on_overflow */
1361 bfd_elf_generic_reloc, /* special_function */
1362 AARCH64_R_STR (TLSLD_ADD_DTPREL_LO12), /* name */
1363 false, /* partial_inplace */
1364 0xfff, /* src_mask */
1365 0xfff, /* dst_mask */
1366 false), /* pcrel_offset */
1367
1368 /* No overflow check version of BFD_RELOC_AARCH64_TLSLD_ADD_DTPREL_LO12. */
1369 HOWTO (AARCH64_R (TLSLD_ADD_DTPREL_LO12_NC), /* type */
1370 0, /* rightshift */
1371 2, /* size (0 = byte, 1 = short, 2 = long) */
1372 12, /* bitsize */
1373 false, /* pc_relative */
1374 0, /* bitpos */
1375 complain_overflow_dont, /* complain_on_overflow */
1376 bfd_elf_generic_reloc, /* special_function */
1377 AARCH64_R_STR (TLSLD_ADD_DTPREL_LO12_NC), /* name */
1378 false, /* partial_inplace */
1379 0xfff, /* src_mask */
1380 0xfff, /* dst_mask */
1381 false), /* pcrel_offset */
1382
1383 /* ADD: GOT offset G(S) & 0xff8 [no overflow check] */
1384 HOWTO (AARCH64_R (TLSLD_ADD_LO12_NC), /* type */
1385 0, /* rightshift */
1386 2, /* size (0 = byte, 1 = short, 2 = long) */
1387 12, /* bitsize */
1388 false, /* pc_relative */
1389 0, /* bitpos */
1390 complain_overflow_dont, /* complain_on_overflow */
1391 bfd_elf_generic_reloc, /* special_function */
1392 AARCH64_R_STR (TLSLD_ADD_LO12_NC), /* name */
1393 false, /* partial_inplace */
1394 0xfff, /* src_mask */
1395 0xfff, /* dst_mask */
1396 false), /* pcrel_offset */
1397
1398 /* Get to the page for the GOT entry for the symbol
1399 (G(S) - P) using an ADRP instruction. */
1400 HOWTO (AARCH64_R (TLSLD_ADR_PAGE21), /* type */
1401 12, /* rightshift */
1402 2, /* size (0 = byte, 1 = short, 2 = long) */
1403 21, /* bitsize */
1404 true, /* pc_relative */
1405 0, /* bitpos */
1406 complain_overflow_signed, /* complain_on_overflow */
1407 bfd_elf_generic_reloc, /* special_function */
1408 AARCH64_R_STR (TLSLD_ADR_PAGE21), /* name */
1409 false, /* partial_inplace */
1410 0x1fffff, /* src_mask */
1411 0x1fffff, /* dst_mask */
1412 true), /* pcrel_offset */
1413
1414 HOWTO (AARCH64_R (TLSLD_ADR_PREL21), /* type */
1415 0, /* rightshift */
1416 2, /* size (0 = byte, 1 = short, 2 = long) */
1417 21, /* bitsize */
1418 true, /* pc_relative */
1419 0, /* bitpos */
1420 complain_overflow_signed, /* complain_on_overflow */
1421 bfd_elf_generic_reloc, /* special_function */
1422 AARCH64_R_STR (TLSLD_ADR_PREL21), /* name */
1423 false, /* partial_inplace */
1424 0x1fffff, /* src_mask */
1425 0x1fffff, /* dst_mask */
1426 true), /* pcrel_offset */
1427
1428 /* LD/ST16: bit[11:1] of byte offset to module TLS base address. */
1429 HOWTO64 (AARCH64_R (TLSLD_LDST16_DTPREL_LO12), /* type */
1430 1, /* rightshift */
1431 2, /* size (0 = byte, 1 = short, 2 = long) */
1432 11, /* bitsize */
1433 false, /* pc_relative */
1434 10, /* bitpos */
1435 complain_overflow_unsigned, /* complain_on_overflow */
1436 bfd_elf_generic_reloc, /* special_function */
1437 AARCH64_R_STR (TLSLD_LDST16_DTPREL_LO12), /* name */
1438 false, /* partial_inplace */
1439 0x1ffc00, /* src_mask */
1440 0x1ffc00, /* dst_mask */
1441 false), /* pcrel_offset */
1442
1443 /* Same as BFD_RELOC_AARCH64_TLSLD_LDST16_DTPREL_LO12, but no overflow check. */
1444 HOWTO64 (AARCH64_R (TLSLD_LDST16_DTPREL_LO12_NC), /* type */
1445 1, /* rightshift */
1446 2, /* size (0 = byte, 1 = short, 2 = long) */
1447 11, /* bitsize */
1448 false, /* pc_relative */
1449 10, /* bitpos */
1450 complain_overflow_dont, /* complain_on_overflow */
1451 bfd_elf_generic_reloc, /* special_function */
1452 AARCH64_R_STR (TLSLD_LDST16_DTPREL_LO12_NC), /* name */
1453 false, /* partial_inplace */
1454 0x1ffc00, /* src_mask */
1455 0x1ffc00, /* dst_mask */
1456 false), /* pcrel_offset */
1457
1458 /* LD/ST32: bit[11:2] of byte offset to module TLS base address. */
1459 HOWTO64 (AARCH64_R (TLSLD_LDST32_DTPREL_LO12), /* type */
1460 2, /* rightshift */
1461 2, /* size (0 = byte, 1 = short, 2 = long) */
1462 10, /* bitsize */
1463 false, /* pc_relative */
1464 10, /* bitpos */
1465 complain_overflow_unsigned, /* complain_on_overflow */
1466 bfd_elf_generic_reloc, /* special_function */
1467 AARCH64_R_STR (TLSLD_LDST32_DTPREL_LO12), /* name */
1468 false, /* partial_inplace */
1469 0x3ffc00, /* src_mask */
1470 0x3ffc00, /* dst_mask */
1471 false), /* pcrel_offset */
1472
1473 /* Same as BFD_RELOC_AARCH64_TLSLD_LDST32_DTPREL_LO12, but no overflow check. */
1474 HOWTO64 (AARCH64_R (TLSLD_LDST32_DTPREL_LO12_NC), /* type */
1475 2, /* rightshift */
1476 2, /* size (0 = byte, 1 = short, 2 = long) */
1477 10, /* bitsize */
1478 false, /* pc_relative */
1479 10, /* bitpos */
1480 complain_overflow_dont, /* complain_on_overflow */
1481 bfd_elf_generic_reloc, /* special_function */
1482 AARCH64_R_STR (TLSLD_LDST32_DTPREL_LO12_NC), /* name */
1483 false, /* partial_inplace */
1484 0xffc00, /* src_mask */
1485 0xffc00, /* dst_mask */
1486 false), /* pcrel_offset */
1487
1488 /* LD/ST64: bit[11:3] of byte offset to module TLS base address. */
1489 HOWTO64 (AARCH64_R (TLSLD_LDST64_DTPREL_LO12), /* type */
1490 3, /* rightshift */
1491 2, /* size (0 = byte, 1 = short, 2 = long) */
1492 9, /* bitsize */
1493 false, /* pc_relative */
1494 10, /* bitpos */
1495 complain_overflow_unsigned, /* complain_on_overflow */
1496 bfd_elf_generic_reloc, /* special_function */
1497 AARCH64_R_STR (TLSLD_LDST64_DTPREL_LO12), /* name */
1498 false, /* partial_inplace */
1499 0x3ffc00, /* src_mask */
1500 0x3ffc00, /* dst_mask */
1501 false), /* pcrel_offset */
1502
1503 /* Same as BFD_RELOC_AARCH64_TLSLD_LDST64_DTPREL_LO12, but no overflow check. */
1504 HOWTO64 (AARCH64_R (TLSLD_LDST64_DTPREL_LO12_NC), /* type */
1505 3, /* rightshift */
1506 2, /* size (0 = byte, 1 = short, 2 = long) */
1507 9, /* bitsize */
1508 false, /* pc_relative */
1509 10, /* bitpos */
1510 complain_overflow_dont, /* complain_on_overflow */
1511 bfd_elf_generic_reloc, /* special_function */
1512 AARCH64_R_STR (TLSLD_LDST64_DTPREL_LO12_NC), /* name */
1513 false, /* partial_inplace */
1514 0x7fc00, /* src_mask */
1515 0x7fc00, /* dst_mask */
1516 false), /* pcrel_offset */
1517
1518 /* LD/ST8: bit[11:0] of byte offset to module TLS base address. */
1519 HOWTO64 (AARCH64_R (TLSLD_LDST8_DTPREL_LO12), /* type */
1520 0, /* rightshift */
1521 2, /* size (0 = byte, 1 = short, 2 = long) */
1522 12, /* bitsize */
1523 false, /* pc_relative */
1524 10, /* bitpos */
1525 complain_overflow_unsigned, /* complain_on_overflow */
1526 bfd_elf_generic_reloc, /* special_function */
1527 AARCH64_R_STR (TLSLD_LDST8_DTPREL_LO12), /* name */
1528 false, /* partial_inplace */
1529 0x3ffc00, /* src_mask */
1530 0x3ffc00, /* dst_mask */
1531 false), /* pcrel_offset */
1532
1533 /* Same as BFD_RELOC_AARCH64_TLSLD_LDST8_DTPREL_LO12, but no overflow check. */
1534 HOWTO64 (AARCH64_R (TLSLD_LDST8_DTPREL_LO12_NC), /* type */
1535 0, /* rightshift */
1536 2, /* size (0 = byte, 1 = short, 2 = long) */
1537 12, /* bitsize */
1538 false, /* pc_relative */
1539 10, /* bitpos */
1540 complain_overflow_dont, /* complain_on_overflow */
1541 bfd_elf_generic_reloc, /* special_function */
1542 AARCH64_R_STR (TLSLD_LDST8_DTPREL_LO12_NC), /* name */
1543 false, /* partial_inplace */
1544 0x3ffc00, /* src_mask */
1545 0x3ffc00, /* dst_mask */
1546 false), /* pcrel_offset */
1547
1548 /* MOVZ: bit[15:0] of byte offset to module TLS base address. */
1549 HOWTO (AARCH64_R (TLSLD_MOVW_DTPREL_G0), /* type */
1550 0, /* rightshift */
1551 2, /* size (0 = byte, 1 = short, 2 = long) */
1552 16, /* bitsize */
1553 false, /* pc_relative */
1554 0, /* bitpos */
1555 complain_overflow_unsigned, /* complain_on_overflow */
1556 bfd_elf_generic_reloc, /* special_function */
1557 AARCH64_R_STR (TLSLD_MOVW_DTPREL_G0), /* name */
1558 false, /* partial_inplace */
1559 0xffff, /* src_mask */
1560 0xffff, /* dst_mask */
1561 false), /* pcrel_offset */
1562
1563 /* No overflow check version of BFD_RELOC_AARCH64_TLSLD_MOVW_DTPREL_G0. */
1564 HOWTO (AARCH64_R (TLSLD_MOVW_DTPREL_G0_NC), /* type */
1565 0, /* rightshift */
1566 2, /* size (0 = byte, 1 = short, 2 = long) */
1567 16, /* bitsize */
1568 false, /* pc_relative */
1569 0, /* bitpos */
1570 complain_overflow_dont, /* complain_on_overflow */
1571 bfd_elf_generic_reloc, /* special_function */
1572 AARCH64_R_STR (TLSLD_MOVW_DTPREL_G0_NC), /* name */
1573 false, /* partial_inplace */
1574 0xffff, /* src_mask */
1575 0xffff, /* dst_mask */
1576 false), /* pcrel_offset */
1577
1578 /* MOVZ: bit[31:16] of byte offset to module TLS base address. */
1579 HOWTO (AARCH64_R (TLSLD_MOVW_DTPREL_G1), /* type */
1580 16, /* rightshift */
1581 2, /* size (0 = byte, 1 = short, 2 = long) */
1582 16, /* bitsize */
1583 false, /* pc_relative */
1584 0, /* bitpos */
1585 complain_overflow_unsigned, /* complain_on_overflow */
1586 bfd_elf_generic_reloc, /* special_function */
1587 AARCH64_R_STR (TLSLD_MOVW_DTPREL_G1), /* name */
1588 false, /* partial_inplace */
1589 0xffff, /* src_mask */
1590 0xffff, /* dst_mask */
1591 false), /* pcrel_offset */
1592
1593 /* No overflow check version of BFD_RELOC_AARCH64_TLSLD_MOVW_DTPREL_G1. */
1594 HOWTO64 (AARCH64_R (TLSLD_MOVW_DTPREL_G1_NC), /* type */
1595 16, /* rightshift */
1596 2, /* size (0 = byte, 1 = short, 2 = long) */
1597 16, /* bitsize */
1598 false, /* pc_relative */
1599 0, /* bitpos */
1600 complain_overflow_dont, /* complain_on_overflow */
1601 bfd_elf_generic_reloc, /* special_function */
1602 AARCH64_R_STR (TLSLD_MOVW_DTPREL_G1_NC), /* name */
1603 false, /* partial_inplace */
1604 0xffff, /* src_mask */
1605 0xffff, /* dst_mask */
1606 false), /* pcrel_offset */
1607
1608 /* MOVZ: bit[47:32] of byte offset to module TLS base address. */
1609 HOWTO64 (AARCH64_R (TLSLD_MOVW_DTPREL_G2), /* type */
1610 32, /* rightshift */
1611 2, /* size (0 = byte, 1 = short, 2 = long) */
1612 16, /* bitsize */
1613 false, /* pc_relative */
1614 0, /* bitpos */
1615 complain_overflow_unsigned, /* complain_on_overflow */
1616 bfd_elf_generic_reloc, /* special_function */
1617 AARCH64_R_STR (TLSLD_MOVW_DTPREL_G2), /* name */
1618 false, /* partial_inplace */
1619 0xffff, /* src_mask */
1620 0xffff, /* dst_mask */
1621 false), /* pcrel_offset */
1622
1623 HOWTO64 (AARCH64_R (TLSLE_MOVW_TPREL_G2), /* type */
1624 32, /* rightshift */
1625 2, /* size (0 = byte, 1 = short, 2 = long) */
1626 16, /* bitsize */
1627 false, /* pc_relative */
1628 0, /* bitpos */
1629 complain_overflow_unsigned, /* complain_on_overflow */
1630 bfd_elf_generic_reloc, /* special_function */
1631 AARCH64_R_STR (TLSLE_MOVW_TPREL_G2), /* name */
1632 false, /* partial_inplace */
1633 0xffff, /* src_mask */
1634 0xffff, /* dst_mask */
1635 false), /* pcrel_offset */
1636
1637 HOWTO (AARCH64_R (TLSLE_MOVW_TPREL_G1), /* type */
1638 16, /* rightshift */
1639 2, /* size (0 = byte, 1 = short, 2 = long) */
1640 16, /* bitsize */
1641 false, /* pc_relative */
1642 0, /* bitpos */
1643 complain_overflow_dont, /* complain_on_overflow */
1644 bfd_elf_generic_reloc, /* special_function */
1645 AARCH64_R_STR (TLSLE_MOVW_TPREL_G1), /* name */
1646 false, /* partial_inplace */
1647 0xffff, /* src_mask */
1648 0xffff, /* dst_mask */
1649 false), /* pcrel_offset */
1650
1651 HOWTO64 (AARCH64_R (TLSLE_MOVW_TPREL_G1_NC), /* type */
1652 16, /* rightshift */
1653 2, /* size (0 = byte, 1 = short, 2 = long) */
1654 16, /* bitsize */
1655 false, /* pc_relative */
1656 0, /* bitpos */
1657 complain_overflow_dont, /* complain_on_overflow */
1658 bfd_elf_generic_reloc, /* special_function */
1659 AARCH64_R_STR (TLSLE_MOVW_TPREL_G1_NC), /* name */
1660 false, /* partial_inplace */
1661 0xffff, /* src_mask */
1662 0xffff, /* dst_mask */
1663 false), /* pcrel_offset */
1664
1665 HOWTO (AARCH64_R (TLSLE_MOVW_TPREL_G0), /* type */
1666 0, /* rightshift */
1667 2, /* size (0 = byte, 1 = short, 2 = long) */
1668 16, /* bitsize */
1669 false, /* pc_relative */
1670 0, /* bitpos */
1671 complain_overflow_dont, /* complain_on_overflow */
1672 bfd_elf_generic_reloc, /* special_function */
1673 AARCH64_R_STR (TLSLE_MOVW_TPREL_G0), /* name */
1674 false, /* partial_inplace */
1675 0xffff, /* src_mask */
1676 0xffff, /* dst_mask */
1677 false), /* pcrel_offset */
1678
1679 HOWTO (AARCH64_R (TLSLE_MOVW_TPREL_G0_NC), /* type */
1680 0, /* rightshift */
1681 2, /* size (0 = byte, 1 = short, 2 = long) */
1682 16, /* bitsize */
1683 false, /* pc_relative */
1684 0, /* bitpos */
1685 complain_overflow_dont, /* complain_on_overflow */
1686 bfd_elf_generic_reloc, /* special_function */
1687 AARCH64_R_STR (TLSLE_MOVW_TPREL_G0_NC), /* name */
1688 false, /* partial_inplace */
1689 0xffff, /* src_mask */
1690 0xffff, /* dst_mask */
1691 false), /* pcrel_offset */
1692
1693 HOWTO (AARCH64_R (TLSLE_ADD_TPREL_HI12), /* type */
1694 12, /* rightshift */
1695 2, /* size (0 = byte, 1 = short, 2 = long) */
1696 12, /* bitsize */
1697 false, /* pc_relative */
1698 0, /* bitpos */
1699 complain_overflow_unsigned, /* complain_on_overflow */
1700 bfd_elf_generic_reloc, /* special_function */
1701 AARCH64_R_STR (TLSLE_ADD_TPREL_HI12), /* name */
1702 false, /* partial_inplace */
1703 0xfff, /* src_mask */
1704 0xfff, /* dst_mask */
1705 false), /* pcrel_offset */
1706
1707 HOWTO (AARCH64_R (TLSLE_ADD_TPREL_LO12), /* type */
1708 0, /* rightshift */
1709 2, /* size (0 = byte, 1 = short, 2 = long) */
1710 12, /* bitsize */
1711 false, /* pc_relative */
1712 0, /* bitpos */
1713 complain_overflow_unsigned, /* complain_on_overflow */
1714 bfd_elf_generic_reloc, /* special_function */
1715 AARCH64_R_STR (TLSLE_ADD_TPREL_LO12), /* name */
1716 false, /* partial_inplace */
1717 0xfff, /* src_mask */
1718 0xfff, /* dst_mask */
1719 false), /* pcrel_offset */
1720
1721 HOWTO (AARCH64_R (TLSLE_ADD_TPREL_LO12_NC), /* type */
1722 0, /* rightshift */
1723 2, /* size (0 = byte, 1 = short, 2 = long) */
1724 12, /* bitsize */
1725 false, /* pc_relative */
1726 0, /* bitpos */
1727 complain_overflow_dont, /* complain_on_overflow */
1728 bfd_elf_generic_reloc, /* special_function */
1729 AARCH64_R_STR (TLSLE_ADD_TPREL_LO12_NC), /* name */
1730 false, /* partial_inplace */
1731 0xfff, /* src_mask */
1732 0xfff, /* dst_mask */
1733 false), /* pcrel_offset */
1734
1735 /* LD/ST16: bit[11:1] of byte offset to module TLS base address. */
1736 HOWTO (AARCH64_R (TLSLE_LDST16_TPREL_LO12), /* type */
1737 1, /* rightshift */
1738 2, /* size (0 = byte, 1 = short, 2 = long) */
1739 11, /* bitsize */
1740 false, /* pc_relative */
1741 10, /* bitpos */
1742 complain_overflow_unsigned, /* complain_on_overflow */
1743 bfd_elf_generic_reloc, /* special_function */
1744 AARCH64_R_STR (TLSLE_LDST16_TPREL_LO12), /* name */
1745 false, /* partial_inplace */
1746 0x1ffc00, /* src_mask */
1747 0x1ffc00, /* dst_mask */
1748 false), /* pcrel_offset */
1749
1750 /* Same as BFD_RELOC_AARCH64_TLSLE_LDST16_TPREL_LO12, but no overflow check. */
1751 HOWTO (AARCH64_R (TLSLE_LDST16_TPREL_LO12_NC), /* type */
1752 1, /* rightshift */
1753 2, /* size (0 = byte, 1 = short, 2 = long) */
1754 11, /* bitsize */
1755 false, /* pc_relative */
1756 10, /* bitpos */
1757 complain_overflow_dont, /* complain_on_overflow */
1758 bfd_elf_generic_reloc, /* special_function */
1759 AARCH64_R_STR (TLSLE_LDST16_TPREL_LO12_NC), /* name */
1760 false, /* partial_inplace */
1761 0x1ffc00, /* src_mask */
1762 0x1ffc00, /* dst_mask */
1763 false), /* pcrel_offset */
1764
1765 /* LD/ST32: bit[11:2] of byte offset to module TLS base address. */
1766 HOWTO (AARCH64_R (TLSLE_LDST32_TPREL_LO12), /* type */
1767 2, /* rightshift */
1768 2, /* size (0 = byte, 1 = short, 2 = long) */
1769 10, /* bitsize */
1770 false, /* pc_relative */
1771 10, /* bitpos */
1772 complain_overflow_unsigned, /* complain_on_overflow */
1773 bfd_elf_generic_reloc, /* special_function */
1774 AARCH64_R_STR (TLSLE_LDST32_TPREL_LO12), /* name */
1775 false, /* partial_inplace */
1776 0xffc00, /* src_mask */
1777 0xffc00, /* dst_mask */
1778 false), /* pcrel_offset */
1779
1780 /* Same as BFD_RELOC_AARCH64_TLSLE_LDST32_TPREL_LO12, but no overflow check. */
1781 HOWTO (AARCH64_R (TLSLE_LDST32_TPREL_LO12_NC), /* type */
1782 2, /* rightshift */
1783 2, /* size (0 = byte, 1 = short, 2 = long) */
1784 10, /* bitsize */
1785 false, /* pc_relative */
1786 10, /* bitpos */
1787 complain_overflow_dont, /* complain_on_overflow */
1788 bfd_elf_generic_reloc, /* special_function */
1789 AARCH64_R_STR (TLSLE_LDST32_TPREL_LO12_NC), /* name */
1790 false, /* partial_inplace */
1791 0xffc00, /* src_mask */
1792 0xffc00, /* dst_mask */
1793 false), /* pcrel_offset */
1794
1795 /* LD/ST64: bit[11:3] of byte offset to module TLS base address. */
1796 HOWTO (AARCH64_R (TLSLE_LDST64_TPREL_LO12), /* type */
1797 3, /* rightshift */
1798 2, /* size (0 = byte, 1 = short, 2 = long) */
1799 9, /* bitsize */
1800 false, /* pc_relative */
1801 10, /* bitpos */
1802 complain_overflow_unsigned, /* complain_on_overflow */
1803 bfd_elf_generic_reloc, /* special_function */
1804 AARCH64_R_STR (TLSLE_LDST64_TPREL_LO12), /* name */
1805 false, /* partial_inplace */
1806 0x7fc00, /* src_mask */
1807 0x7fc00, /* dst_mask */
1808 false), /* pcrel_offset */
1809
1810 /* Same as BFD_RELOC_AARCH64_TLSLE_LDST64_TPREL_LO12, but no overflow check. */
1811 HOWTO (AARCH64_R (TLSLE_LDST64_TPREL_LO12_NC), /* type */
1812 3, /* rightshift */
1813 2, /* size (0 = byte, 1 = short, 2 = long) */
1814 9, /* bitsize */
1815 false, /* pc_relative */
1816 10, /* bitpos */
1817 complain_overflow_dont, /* complain_on_overflow */
1818 bfd_elf_generic_reloc, /* special_function */
1819 AARCH64_R_STR (TLSLE_LDST64_TPREL_LO12_NC), /* name */
1820 false, /* partial_inplace */
1821 0x7fc00, /* src_mask */
1822 0x7fc00, /* dst_mask */
1823 false), /* pcrel_offset */
1824
1825 /* LD/ST8: bit[11:0] of byte offset to module TLS base address. */
1826 HOWTO (AARCH64_R (TLSLE_LDST8_TPREL_LO12), /* type */
1827 0, /* rightshift */
1828 2, /* size (0 = byte, 1 = short, 2 = long) */
1829 12, /* bitsize */
1830 false, /* pc_relative */
1831 10, /* bitpos */
1832 complain_overflow_unsigned, /* complain_on_overflow */
1833 bfd_elf_generic_reloc, /* special_function */
1834 AARCH64_R_STR (TLSLE_LDST8_TPREL_LO12), /* name */
1835 false, /* partial_inplace */
1836 0x3ffc00, /* src_mask */
1837 0x3ffc00, /* dst_mask */
1838 false), /* pcrel_offset */
1839
1840 /* Same as BFD_RELOC_AARCH64_TLSLE_LDST8_TPREL_LO12, but no overflow check. */
1841 HOWTO (AARCH64_R (TLSLE_LDST8_TPREL_LO12_NC), /* type */
1842 0, /* rightshift */
1843 2, /* size (0 = byte, 1 = short, 2 = long) */
1844 12, /* bitsize */
1845 false, /* pc_relative */
1846 10, /* bitpos */
1847 complain_overflow_dont, /* complain_on_overflow */
1848 bfd_elf_generic_reloc, /* special_function */
1849 AARCH64_R_STR (TLSLE_LDST8_TPREL_LO12_NC), /* name */
1850 false, /* partial_inplace */
1851 0x3ffc00, /* src_mask */
1852 0x3ffc00, /* dst_mask */
1853 false), /* pcrel_offset */
1854
1855 HOWTO (AARCH64_R (TLSDESC_LD_PREL19), /* type */
1856 2, /* rightshift */
1857 2, /* size (0 = byte, 1 = short, 2 = long) */
1858 19, /* bitsize */
1859 true, /* pc_relative */
1860 0, /* bitpos */
1861 complain_overflow_dont, /* complain_on_overflow */
1862 bfd_elf_generic_reloc, /* special_function */
1863 AARCH64_R_STR (TLSDESC_LD_PREL19), /* name */
1864 false, /* partial_inplace */
1865 0x0ffffe0, /* src_mask */
1866 0x0ffffe0, /* dst_mask */
1867 true), /* pcrel_offset */
1868
1869 HOWTO (AARCH64_R (TLSDESC_ADR_PREL21), /* type */
1870 0, /* rightshift */
1871 2, /* size (0 = byte, 1 = short, 2 = long) */
1872 21, /* bitsize */
1873 true, /* pc_relative */
1874 0, /* bitpos */
1875 complain_overflow_dont, /* complain_on_overflow */
1876 bfd_elf_generic_reloc, /* special_function */
1877 AARCH64_R_STR (TLSDESC_ADR_PREL21), /* name */
1878 false, /* partial_inplace */
1879 0x1fffff, /* src_mask */
1880 0x1fffff, /* dst_mask */
1881 true), /* pcrel_offset */
1882
1883 /* Get to the page for the GOT entry for the symbol
1884 (G(S) - P) using an ADRP instruction. */
1885 HOWTO (AARCH64_R (TLSDESC_ADR_PAGE21), /* type */
1886 12, /* rightshift */
1887 2, /* size (0 = byte, 1 = short, 2 = long) */
1888 21, /* bitsize */
1889 true, /* pc_relative */
1890 0, /* bitpos */
1891 complain_overflow_dont, /* complain_on_overflow */
1892 bfd_elf_generic_reloc, /* special_function */
1893 AARCH64_R_STR (TLSDESC_ADR_PAGE21), /* name */
1894 false, /* partial_inplace */
1895 0x1fffff, /* src_mask */
1896 0x1fffff, /* dst_mask */
1897 true), /* pcrel_offset */
1898
1899 /* LD64: GOT offset G(S) & 0xff8. */
1900 HOWTO64 (AARCH64_R (TLSDESC_LD64_LO12), /* type */
1901 3, /* rightshift */
1902 2, /* size (0 = byte, 1 = short, 2 = long) */
1903 12, /* bitsize */
1904 false, /* pc_relative */
1905 0, /* bitpos */
1906 complain_overflow_dont, /* complain_on_overflow */
1907 bfd_elf_generic_reloc, /* special_function */
1908 AARCH64_R_STR (TLSDESC_LD64_LO12), /* name */
1909 false, /* partial_inplace */
1910 0xff8, /* src_mask */
1911 0xff8, /* dst_mask */
1912 false), /* pcrel_offset */
1913
1914 /* LD32: GOT offset G(S) & 0xffc. */
1915 HOWTO32 (AARCH64_R (TLSDESC_LD32_LO12_NC), /* type */
1916 2, /* rightshift */
1917 2, /* size (0 = byte, 1 = short, 2 = long) */
1918 12, /* bitsize */
1919 false, /* pc_relative */
1920 0, /* bitpos */
1921 complain_overflow_dont, /* complain_on_overflow */
1922 bfd_elf_generic_reloc, /* special_function */
1923 AARCH64_R_STR (TLSDESC_LD32_LO12_NC), /* name */
1924 false, /* partial_inplace */
1925 0xffc, /* src_mask */
1926 0xffc, /* dst_mask */
1927 false), /* pcrel_offset */
1928
1929 /* ADD: GOT offset G(S) & 0xfff. */
1930 HOWTO (AARCH64_R (TLSDESC_ADD_LO12), /* type */
1931 0, /* rightshift */
1932 2, /* size (0 = byte, 1 = short, 2 = long) */
1933 12, /* bitsize */
1934 false, /* pc_relative */
1935 0, /* bitpos */
1936 complain_overflow_dont,/* complain_on_overflow */
1937 bfd_elf_generic_reloc, /* special_function */
1938 AARCH64_R_STR (TLSDESC_ADD_LO12), /* name */
1939 false, /* partial_inplace */
1940 0xfff, /* src_mask */
1941 0xfff, /* dst_mask */
1942 false), /* pcrel_offset */
1943
1944 HOWTO64 (AARCH64_R (TLSDESC_OFF_G1), /* type */
1945 16, /* rightshift */
1946 2, /* size (0 = byte, 1 = short, 2 = long) */
1947 12, /* bitsize */
1948 false, /* pc_relative */
1949 0, /* bitpos */
1950 complain_overflow_unsigned, /* complain_on_overflow */
1951 bfd_elf_generic_reloc, /* special_function */
1952 AARCH64_R_STR (TLSDESC_OFF_G1), /* name */
1953 false, /* partial_inplace */
1954 0xffff, /* src_mask */
1955 0xffff, /* dst_mask */
1956 false), /* pcrel_offset */
1957
1958 HOWTO64 (AARCH64_R (TLSDESC_OFF_G0_NC), /* type */
1959 0, /* rightshift */
1960 2, /* size (0 = byte, 1 = short, 2 = long) */
1961 12, /* bitsize */
1962 false, /* pc_relative */
1963 0, /* bitpos */
1964 complain_overflow_dont, /* complain_on_overflow */
1965 bfd_elf_generic_reloc, /* special_function */
1966 AARCH64_R_STR (TLSDESC_OFF_G0_NC), /* name */
1967 false, /* partial_inplace */
1968 0xffff, /* src_mask */
1969 0xffff, /* dst_mask */
1970 false), /* pcrel_offset */
1971
1972 HOWTO64 (AARCH64_R (TLSDESC_LDR), /* type */
1973 0, /* rightshift */
1974 2, /* size (0 = byte, 1 = short, 2 = long) */
1975 12, /* bitsize */
1976 false, /* pc_relative */
1977 0, /* bitpos */
1978 complain_overflow_dont, /* complain_on_overflow */
1979 bfd_elf_generic_reloc, /* special_function */
1980 AARCH64_R_STR (TLSDESC_LDR), /* name */
1981 false, /* partial_inplace */
1982 0x0, /* src_mask */
1983 0x0, /* dst_mask */
1984 false), /* pcrel_offset */
1985
1986 HOWTO64 (AARCH64_R (TLSDESC_ADD), /* type */
1987 0, /* rightshift */
1988 2, /* size (0 = byte, 1 = short, 2 = long) */
1989 12, /* bitsize */
1990 false, /* pc_relative */
1991 0, /* bitpos */
1992 complain_overflow_dont, /* complain_on_overflow */
1993 bfd_elf_generic_reloc, /* special_function */
1994 AARCH64_R_STR (TLSDESC_ADD), /* name */
1995 false, /* partial_inplace */
1996 0x0, /* src_mask */
1997 0x0, /* dst_mask */
1998 false), /* pcrel_offset */
1999
2000 HOWTO (AARCH64_R (TLSDESC_CALL), /* type */
2001 0, /* rightshift */
2002 2, /* size (0 = byte, 1 = short, 2 = long) */
2003 0, /* bitsize */
2004 false, /* pc_relative */
2005 0, /* bitpos */
2006 complain_overflow_dont, /* complain_on_overflow */
2007 bfd_elf_generic_reloc, /* special_function */
2008 AARCH64_R_STR (TLSDESC_CALL), /* name */
2009 false, /* partial_inplace */
2010 0x0, /* src_mask */
2011 0x0, /* dst_mask */
2012 false), /* pcrel_offset */
2013
2014 HOWTO (AARCH64_R (COPY), /* type */
2015 0, /* rightshift */
2016 2, /* size (0 = byte, 1 = short, 2 = long) */
2017 64, /* bitsize */
2018 false, /* pc_relative */
2019 0, /* bitpos */
2020 complain_overflow_bitfield, /* complain_on_overflow */
2021 bfd_elf_generic_reloc, /* special_function */
2022 AARCH64_R_STR (COPY), /* name */
2023 true, /* partial_inplace */
2024 0xffffffff, /* src_mask */
2025 0xffffffff, /* dst_mask */
2026 false), /* pcrel_offset */
2027
2028 HOWTO (AARCH64_R (GLOB_DAT), /* type */
2029 0, /* rightshift */
2030 2, /* size (0 = byte, 1 = short, 2 = long) */
2031 64, /* bitsize */
2032 false, /* pc_relative */
2033 0, /* bitpos */
2034 complain_overflow_bitfield, /* complain_on_overflow */
2035 bfd_elf_generic_reloc, /* special_function */
2036 AARCH64_R_STR (GLOB_DAT), /* name */
2037 true, /* partial_inplace */
2038 0xffffffff, /* src_mask */
2039 0xffffffff, /* dst_mask */
2040 false), /* pcrel_offset */
2041
2042 HOWTO (AARCH64_R (JUMP_SLOT), /* type */
2043 0, /* rightshift */
2044 2, /* size (0 = byte, 1 = short, 2 = long) */
2045 64, /* bitsize */
2046 false, /* pc_relative */
2047 0, /* bitpos */
2048 complain_overflow_bitfield, /* complain_on_overflow */
2049 bfd_elf_generic_reloc, /* special_function */
2050 AARCH64_R_STR (JUMP_SLOT), /* name */
2051 true, /* partial_inplace */
2052 0xffffffff, /* src_mask */
2053 0xffffffff, /* dst_mask */
2054 false), /* pcrel_offset */
2055
2056 HOWTO (AARCH64_R (RELATIVE), /* type */
2057 0, /* rightshift */
2058 2, /* size (0 = byte, 1 = short, 2 = long) */
2059 64, /* bitsize */
2060 false, /* pc_relative */
2061 0, /* bitpos */
2062 complain_overflow_bitfield, /* complain_on_overflow */
2063 bfd_elf_generic_reloc, /* special_function */
2064 AARCH64_R_STR (RELATIVE), /* name */
2065 true, /* partial_inplace */
2066 ALL_ONES, /* src_mask */
2067 ALL_ONES, /* dst_mask */
2068 false), /* pcrel_offset */
2069
2070 HOWTO (AARCH64_R (TLS_DTPMOD), /* type */
2071 0, /* rightshift */
2072 2, /* size (0 = byte, 1 = short, 2 = long) */
2073 64, /* bitsize */
2074 false, /* pc_relative */
2075 0, /* bitpos */
2076 complain_overflow_dont, /* complain_on_overflow */
2077 bfd_elf_generic_reloc, /* special_function */
2078 #if ARCH_SIZE == 64
2079 AARCH64_R_STR (TLS_DTPMOD64), /* name */
2080 #else
2081 AARCH64_R_STR (TLS_DTPMOD), /* name */
2082 #endif
2083 false, /* partial_inplace */
2084 0, /* src_mask */
2085 ALL_ONES, /* dst_mask */
2086 false), /* pc_reloffset */
2087
2088 HOWTO (AARCH64_R (TLS_DTPREL), /* type */
2089 0, /* rightshift */
2090 2, /* size (0 = byte, 1 = short, 2 = long) */
2091 64, /* bitsize */
2092 false, /* pc_relative */
2093 0, /* bitpos */
2094 complain_overflow_dont, /* complain_on_overflow */
2095 bfd_elf_generic_reloc, /* special_function */
2096 #if ARCH_SIZE == 64
2097 AARCH64_R_STR (TLS_DTPREL64), /* name */
2098 #else
2099 AARCH64_R_STR (TLS_DTPREL), /* name */
2100 #endif
2101 false, /* partial_inplace */
2102 0, /* src_mask */
2103 ALL_ONES, /* dst_mask */
2104 false), /* pcrel_offset */
2105
2106 HOWTO (AARCH64_R (TLS_TPREL), /* type */
2107 0, /* rightshift */
2108 2, /* size (0 = byte, 1 = short, 2 = long) */
2109 64, /* bitsize */
2110 false, /* pc_relative */
2111 0, /* bitpos */
2112 complain_overflow_dont, /* complain_on_overflow */
2113 bfd_elf_generic_reloc, /* special_function */
2114 #if ARCH_SIZE == 64
2115 AARCH64_R_STR (TLS_TPREL64), /* name */
2116 #else
2117 AARCH64_R_STR (TLS_TPREL), /* name */
2118 #endif
2119 false, /* partial_inplace */
2120 0, /* src_mask */
2121 ALL_ONES, /* dst_mask */
2122 false), /* pcrel_offset */
2123
2124 HOWTO (AARCH64_R (TLSDESC), /* type */
2125 0, /* rightshift */
2126 2, /* size (0 = byte, 1 = short, 2 = long) */
2127 64, /* bitsize */
2128 false, /* pc_relative */
2129 0, /* bitpos */
2130 complain_overflow_dont, /* complain_on_overflow */
2131 bfd_elf_generic_reloc, /* special_function */
2132 AARCH64_R_STR (TLSDESC), /* name */
2133 false, /* partial_inplace */
2134 0, /* src_mask */
2135 ALL_ONES, /* dst_mask */
2136 false), /* pcrel_offset */
2137
2138 HOWTO (AARCH64_R (IRELATIVE), /* type */
2139 0, /* rightshift */
2140 2, /* size (0 = byte, 1 = short, 2 = long) */
2141 64, /* bitsize */
2142 false, /* pc_relative */
2143 0, /* bitpos */
2144 complain_overflow_bitfield, /* complain_on_overflow */
2145 bfd_elf_generic_reloc, /* special_function */
2146 AARCH64_R_STR (IRELATIVE), /* name */
2147 false, /* partial_inplace */
2148 0, /* src_mask */
2149 ALL_ONES, /* dst_mask */
2150 false), /* pcrel_offset */
2151
2152 EMPTY_HOWTO (0),
2153 };
2154
2155 static reloc_howto_type elfNN_aarch64_howto_none =
2156 HOWTO (R_AARCH64_NONE, /* type */
2157 0, /* rightshift */
2158 3, /* size (0 = byte, 1 = short, 2 = long) */
2159 0, /* bitsize */
2160 false, /* pc_relative */
2161 0, /* bitpos */
2162 complain_overflow_dont,/* complain_on_overflow */
2163 bfd_elf_generic_reloc, /* special_function */
2164 "R_AARCH64_NONE", /* name */
2165 false, /* partial_inplace */
2166 0, /* src_mask */
2167 0, /* dst_mask */
2168 false); /* pcrel_offset */
2169
2170 /* Given HOWTO, return the bfd internal relocation enumerator. */
2171
2172 static bfd_reloc_code_real_type
elfNN_aarch64_bfd_reloc_from_howto(reloc_howto_type * howto)2173 elfNN_aarch64_bfd_reloc_from_howto (reloc_howto_type *howto)
2174 {
2175 const int size
2176 = (int) ARRAY_SIZE (elfNN_aarch64_howto_table);
2177 const ptrdiff_t offset
2178 = howto - elfNN_aarch64_howto_table;
2179
2180 if (offset > 0 && offset < size - 1)
2181 return BFD_RELOC_AARCH64_RELOC_START + offset;
2182
2183 if (howto == &elfNN_aarch64_howto_none)
2184 return BFD_RELOC_AARCH64_NONE;
2185
2186 return BFD_RELOC_AARCH64_RELOC_START;
2187 }
2188
2189 /* Given R_TYPE, return the bfd internal relocation enumerator. */
2190
2191 static bfd_reloc_code_real_type
elfNN_aarch64_bfd_reloc_from_type(bfd * abfd,unsigned int r_type)2192 elfNN_aarch64_bfd_reloc_from_type (bfd *abfd, unsigned int r_type)
2193 {
2194 static bool initialized_p = false;
2195 /* Indexed by R_TYPE, values are offsets in the howto_table. */
2196 static unsigned int offsets[R_AARCH64_end];
2197
2198 if (!initialized_p)
2199 {
2200 unsigned int i;
2201
2202 for (i = 1; i < ARRAY_SIZE (elfNN_aarch64_howto_table) - 1; ++i)
2203 if (elfNN_aarch64_howto_table[i].type != 0)
2204 offsets[elfNN_aarch64_howto_table[i].type] = i;
2205
2206 initialized_p = true;
2207 }
2208
2209 if (r_type == R_AARCH64_NONE || r_type == R_AARCH64_NULL)
2210 return BFD_RELOC_AARCH64_NONE;
2211
2212 /* PR 17512: file: b371e70a. */
2213 if (r_type >= R_AARCH64_end)
2214 {
2215 _bfd_error_handler (_("%pB: unsupported relocation type %#x"),
2216 abfd, r_type);
2217 bfd_set_error (bfd_error_bad_value);
2218 return BFD_RELOC_AARCH64_NONE;
2219 }
2220
2221 return BFD_RELOC_AARCH64_RELOC_START + offsets[r_type];
2222 }
2223
2224 struct elf_aarch64_reloc_map
2225 {
2226 bfd_reloc_code_real_type from;
2227 bfd_reloc_code_real_type to;
2228 };
2229
2230 /* Map bfd generic reloc to AArch64-specific reloc. */
2231 static const struct elf_aarch64_reloc_map elf_aarch64_reloc_map[] =
2232 {
2233 {BFD_RELOC_NONE, BFD_RELOC_AARCH64_NONE},
2234
2235 /* Basic data relocations. */
2236 {BFD_RELOC_CTOR, BFD_RELOC_AARCH64_NN},
2237 {BFD_RELOC_64, BFD_RELOC_AARCH64_64},
2238 {BFD_RELOC_32, BFD_RELOC_AARCH64_32},
2239 {BFD_RELOC_16, BFD_RELOC_AARCH64_16},
2240 {BFD_RELOC_64_PCREL, BFD_RELOC_AARCH64_64_PCREL},
2241 {BFD_RELOC_32_PCREL, BFD_RELOC_AARCH64_32_PCREL},
2242 {BFD_RELOC_16_PCREL, BFD_RELOC_AARCH64_16_PCREL},
2243 };
2244
2245 /* Given the bfd internal relocation enumerator in CODE, return the
2246 corresponding howto entry. */
2247
2248 static reloc_howto_type *
elfNN_aarch64_howto_from_bfd_reloc(bfd_reloc_code_real_type code)2249 elfNN_aarch64_howto_from_bfd_reloc (bfd_reloc_code_real_type code)
2250 {
2251 unsigned int i;
2252
2253 /* Convert bfd generic reloc to AArch64-specific reloc. */
2254 if (code < BFD_RELOC_AARCH64_RELOC_START
2255 || code > BFD_RELOC_AARCH64_RELOC_END)
2256 for (i = 0; i < ARRAY_SIZE (elf_aarch64_reloc_map); i++)
2257 if (elf_aarch64_reloc_map[i].from == code)
2258 {
2259 code = elf_aarch64_reloc_map[i].to;
2260 break;
2261 }
2262
2263 if (code > BFD_RELOC_AARCH64_RELOC_START
2264 && code < BFD_RELOC_AARCH64_RELOC_END)
2265 if (elfNN_aarch64_howto_table[code - BFD_RELOC_AARCH64_RELOC_START].type)
2266 return &elfNN_aarch64_howto_table[code - BFD_RELOC_AARCH64_RELOC_START];
2267
2268 if (code == BFD_RELOC_AARCH64_NONE)
2269 return &elfNN_aarch64_howto_none;
2270
2271 return NULL;
2272 }
2273
2274 static reloc_howto_type *
elfNN_aarch64_howto_from_type(bfd * abfd,unsigned int r_type)2275 elfNN_aarch64_howto_from_type (bfd *abfd, unsigned int r_type)
2276 {
2277 bfd_reloc_code_real_type val;
2278 reloc_howto_type *howto;
2279
2280 #if ARCH_SIZE == 32
2281 if (r_type > 256)
2282 {
2283 bfd_set_error (bfd_error_bad_value);
2284 return NULL;
2285 }
2286 #endif
2287
2288 if (r_type == R_AARCH64_NONE)
2289 return &elfNN_aarch64_howto_none;
2290
2291 val = elfNN_aarch64_bfd_reloc_from_type (abfd, r_type);
2292 howto = elfNN_aarch64_howto_from_bfd_reloc (val);
2293
2294 if (howto != NULL)
2295 return howto;
2296
2297 bfd_set_error (bfd_error_bad_value);
2298 return NULL;
2299 }
2300
2301 static bool
elfNN_aarch64_info_to_howto(bfd * abfd,arelent * bfd_reloc,Elf_Internal_Rela * elf_reloc)2302 elfNN_aarch64_info_to_howto (bfd *abfd, arelent *bfd_reloc,
2303 Elf_Internal_Rela *elf_reloc)
2304 {
2305 unsigned int r_type;
2306
2307 r_type = ELFNN_R_TYPE (elf_reloc->r_info);
2308 bfd_reloc->howto = elfNN_aarch64_howto_from_type (abfd, r_type);
2309
2310 if (bfd_reloc->howto == NULL)
2311 {
2312 /* xgettext:c-format */
2313 _bfd_error_handler (_("%pB: unsupported relocation type %#x"), abfd, r_type);
2314 return false;
2315 }
2316 return true;
2317 }
2318
2319 static reloc_howto_type *
elfNN_aarch64_reloc_type_lookup(bfd * abfd ATTRIBUTE_UNUSED,bfd_reloc_code_real_type code)2320 elfNN_aarch64_reloc_type_lookup (bfd *abfd ATTRIBUTE_UNUSED,
2321 bfd_reloc_code_real_type code)
2322 {
2323 reloc_howto_type *howto = elfNN_aarch64_howto_from_bfd_reloc (code);
2324
2325 if (howto != NULL)
2326 return howto;
2327
2328 bfd_set_error (bfd_error_bad_value);
2329 return NULL;
2330 }
2331
2332 static reloc_howto_type *
elfNN_aarch64_reloc_name_lookup(bfd * abfd ATTRIBUTE_UNUSED,const char * r_name)2333 elfNN_aarch64_reloc_name_lookup (bfd *abfd ATTRIBUTE_UNUSED,
2334 const char *r_name)
2335 {
2336 unsigned int i;
2337
2338 for (i = 1; i < ARRAY_SIZE (elfNN_aarch64_howto_table) - 1; ++i)
2339 if (elfNN_aarch64_howto_table[i].name != NULL
2340 && strcasecmp (elfNN_aarch64_howto_table[i].name, r_name) == 0)
2341 return &elfNN_aarch64_howto_table[i];
2342
2343 return NULL;
2344 }
2345
2346 #define TARGET_LITTLE_SYM aarch64_elfNN_le_vec
2347 #define TARGET_LITTLE_NAME "elfNN-littleaarch64"
2348 #define TARGET_BIG_SYM aarch64_elfNN_be_vec
2349 #define TARGET_BIG_NAME "elfNN-bigaarch64"
2350
2351 /* The linker script knows the section names for placement.
2352 The entry_names are used to do simple name mangling on the stubs.
2353 Given a function name, and its type, the stub can be found. The
2354 name can be changed. The only requirement is the %s be present. */
2355 #define STUB_ENTRY_NAME "__%s_veneer"
2356
2357 /* The name of the dynamic interpreter. This is put in the .interp
2358 section. */
2359 #define ELF_DYNAMIC_INTERPRETER "/lib/ld.so.1"
2360
2361 #define AARCH64_MAX_FWD_BRANCH_OFFSET \
2362 (((1 << 25) - 1) << 2)
2363 #define AARCH64_MAX_BWD_BRANCH_OFFSET \
2364 (-((1 << 25) << 2))
2365
2366 #define AARCH64_MAX_ADRP_IMM ((1 << 20) - 1)
2367 #define AARCH64_MIN_ADRP_IMM (-(1 << 20))
2368
2369 static int
aarch64_valid_for_adrp_p(bfd_vma value,bfd_vma place)2370 aarch64_valid_for_adrp_p (bfd_vma value, bfd_vma place)
2371 {
2372 bfd_signed_vma offset = (bfd_signed_vma) (PG (value) - PG (place)) >> 12;
2373 return offset <= AARCH64_MAX_ADRP_IMM && offset >= AARCH64_MIN_ADRP_IMM;
2374 }
2375
2376 static int
aarch64_valid_branch_p(bfd_vma value,bfd_vma place)2377 aarch64_valid_branch_p (bfd_vma value, bfd_vma place)
2378 {
2379 bfd_signed_vma offset = (bfd_signed_vma) (value - place);
2380 return (offset <= AARCH64_MAX_FWD_BRANCH_OFFSET
2381 && offset >= AARCH64_MAX_BWD_BRANCH_OFFSET);
2382 }
2383
2384 static const uint32_t aarch64_adrp_branch_stub [] =
2385 {
2386 0x90000010, /* adrp ip0, X */
2387 /* R_AARCH64_ADR_HI21_PCREL(X) */
2388 0x91000210, /* add ip0, ip0, :lo12:X */
2389 /* R_AARCH64_ADD_ABS_LO12_NC(X) */
2390 0xd61f0200, /* br ip0 */
2391 };
2392
2393 static const uint32_t aarch64_long_branch_stub[] =
2394 {
2395 #if ARCH_SIZE == 64
2396 0x58000090, /* ldr ip0, 1f */
2397 #else
2398 0x18000090, /* ldr wip0, 1f */
2399 #endif
2400 0x10000011, /* adr ip1, #0 */
2401 0x8b110210, /* add ip0, ip0, ip1 */
2402 0xd61f0200, /* br ip0 */
2403 0x00000000, /* 1: .xword or .word
2404 R_AARCH64_PRELNN(X) + 12
2405 */
2406 0x00000000,
2407 };
2408
2409 static const uint32_t aarch64_erratum_835769_stub[] =
2410 {
2411 0x00000000, /* Placeholder for multiply accumulate. */
2412 0x14000000, /* b <label> */
2413 };
2414
2415 static const uint32_t aarch64_erratum_843419_stub[] =
2416 {
2417 0x00000000, /* Placeholder for LDR instruction. */
2418 0x14000000, /* b <label> */
2419 };
2420
2421 /* Section name for stubs is the associated section name plus this
2422 string. */
2423 #define STUB_SUFFIX ".stub"
2424
2425 enum elf_aarch64_stub_type
2426 {
2427 aarch64_stub_none,
2428 aarch64_stub_adrp_branch,
2429 aarch64_stub_long_branch,
2430 aarch64_stub_erratum_835769_veneer,
2431 aarch64_stub_erratum_843419_veneer,
2432 };
2433
2434 struct elf_aarch64_stub_hash_entry
2435 {
2436 /* Base hash table entry structure. */
2437 struct bfd_hash_entry root;
2438
2439 /* The stub section. */
2440 asection *stub_sec;
2441
2442 /* Offset within stub_sec of the beginning of this stub. */
2443 bfd_vma stub_offset;
2444
2445 /* Given the symbol's value and its section we can determine its final
2446 value when building the stubs (so the stub knows where to jump). */
2447 bfd_vma target_value;
2448 asection *target_section;
2449
2450 enum elf_aarch64_stub_type stub_type;
2451
2452 /* The symbol table entry, if any, that this was derived from. */
2453 struct elf_aarch64_link_hash_entry *h;
2454
2455 /* Destination symbol type */
2456 unsigned char st_type;
2457
2458 /* Where this stub is being called from, or, in the case of combined
2459 stub sections, the first input section in the group. */
2460 asection *id_sec;
2461
2462 /* The name for the local symbol at the start of this stub. The
2463 stub name in the hash table has to be unique; this does not, so
2464 it can be friendlier. */
2465 char *output_name;
2466
2467 /* The instruction which caused this stub to be generated (only valid for
2468 erratum 835769 workaround stubs at present). */
2469 uint32_t veneered_insn;
2470
2471 /* In an erratum 843419 workaround stub, the ADRP instruction offset. */
2472 bfd_vma adrp_offset;
2473 };
2474
2475 /* Used to build a map of a section. This is required for mixed-endian
2476 code/data. */
2477
2478 typedef struct elf_elf_section_map
2479 {
2480 bfd_vma vma;
2481 char type;
2482 }
2483 elf_aarch64_section_map;
2484
2485
2486 typedef struct _aarch64_elf_section_data
2487 {
2488 struct bfd_elf_section_data elf;
2489 unsigned int mapcount;
2490 unsigned int mapsize;
2491 elf_aarch64_section_map *map;
2492 }
2493 _aarch64_elf_section_data;
2494
2495 #define elf_aarch64_section_data(sec) \
2496 ((_aarch64_elf_section_data *) elf_section_data (sec))
2497
2498 /* The size of the thread control block which is defined to be two pointers. */
2499 #define TCB_SIZE (ARCH_SIZE/8)*2
2500
2501 struct elf_aarch64_local_symbol
2502 {
2503 unsigned int got_type;
2504 bfd_signed_vma got_refcount;
2505 bfd_vma got_offset;
2506
2507 /* Offset of the GOTPLT entry reserved for the TLS descriptor. The
2508 offset is from the end of the jump table and reserved entries
2509 within the PLTGOT.
2510
2511 The magic value (bfd_vma) -1 indicates that an offset has not be
2512 allocated. */
2513 bfd_vma tlsdesc_got_jump_table_offset;
2514 };
2515
2516 struct elf_aarch64_obj_tdata
2517 {
2518 struct elf_obj_tdata root;
2519
2520 /* local symbol descriptors */
2521 struct elf_aarch64_local_symbol *locals;
2522
2523 /* Zero to warn when linking objects with incompatible enum sizes. */
2524 int no_enum_size_warning;
2525
2526 /* Zero to warn when linking objects with incompatible wchar_t sizes. */
2527 int no_wchar_size_warning;
2528
2529 /* All GNU_PROPERTY_AARCH64_FEATURE_1_AND properties. */
2530 uint32_t gnu_and_prop;
2531
2532 /* Zero to warn when linking objects with incompatible
2533 GNU_PROPERTY_AARCH64_FEATURE_1_BTI. */
2534 int no_bti_warn;
2535
2536 /* PLT type based on security. */
2537 aarch64_plt_type plt_type;
2538 };
2539
2540 #define elf_aarch64_tdata(bfd) \
2541 ((struct elf_aarch64_obj_tdata *) (bfd)->tdata.any)
2542
2543 #define elf_aarch64_locals(bfd) (elf_aarch64_tdata (bfd)->locals)
2544
2545 #define is_aarch64_elf(bfd) \
2546 (bfd_get_flavour (bfd) == bfd_target_elf_flavour \
2547 && elf_tdata (bfd) != NULL \
2548 && elf_object_id (bfd) == AARCH64_ELF_DATA)
2549
2550 static bool
elfNN_aarch64_mkobject(bfd * abfd)2551 elfNN_aarch64_mkobject (bfd *abfd)
2552 {
2553 return bfd_elf_allocate_object (abfd, sizeof (struct elf_aarch64_obj_tdata),
2554 AARCH64_ELF_DATA);
2555 }
2556
2557 #define elf_aarch64_hash_entry(ent) \
2558 ((struct elf_aarch64_link_hash_entry *)(ent))
2559
2560 #define GOT_UNKNOWN 0
2561 #define GOT_NORMAL 1
2562 #define GOT_TLS_GD 2
2563 #define GOT_TLS_IE 4
2564 #define GOT_TLSDESC_GD 8
2565
2566 #define GOT_TLS_GD_ANY_P(type) ((type & GOT_TLS_GD) || (type & GOT_TLSDESC_GD))
2567
2568 /* AArch64 ELF linker hash entry. */
2569 struct elf_aarch64_link_hash_entry
2570 {
2571 struct elf_link_hash_entry root;
2572
2573 /* Since PLT entries have variable size, we need to record the
2574 index into .got.plt instead of recomputing it from the PLT
2575 offset. */
2576 bfd_signed_vma plt_got_offset;
2577
2578 /* Bit mask representing the type of GOT entry(s) if any required by
2579 this symbol. */
2580 unsigned int got_type;
2581
2582 /* A pointer to the most recently used stub hash entry against this
2583 symbol. */
2584 struct elf_aarch64_stub_hash_entry *stub_cache;
2585
2586 /* Offset of the GOTPLT entry reserved for the TLS descriptor. The offset
2587 is from the end of the jump table and reserved entries within the PLTGOT.
2588
2589 The magic value (bfd_vma) -1 indicates that an offset has not
2590 be allocated. */
2591 bfd_vma tlsdesc_got_jump_table_offset;
2592 };
2593
2594 static unsigned int
elfNN_aarch64_symbol_got_type(struct elf_link_hash_entry * h,bfd * abfd,unsigned long r_symndx)2595 elfNN_aarch64_symbol_got_type (struct elf_link_hash_entry *h,
2596 bfd *abfd,
2597 unsigned long r_symndx)
2598 {
2599 if (h)
2600 return elf_aarch64_hash_entry (h)->got_type;
2601
2602 if (! elf_aarch64_locals (abfd))
2603 return GOT_UNKNOWN;
2604
2605 return elf_aarch64_locals (abfd)[r_symndx].got_type;
2606 }
2607
2608 /* Get the AArch64 elf linker hash table from a link_info structure. */
2609 #define elf_aarch64_hash_table(info) \
2610 ((struct elf_aarch64_link_hash_table *) ((info)->hash))
2611
2612 #define aarch64_stub_hash_lookup(table, string, create, copy) \
2613 ((struct elf_aarch64_stub_hash_entry *) \
2614 bfd_hash_lookup ((table), (string), (create), (copy)))
2615
2616 /* AArch64 ELF linker hash table. */
2617 struct elf_aarch64_link_hash_table
2618 {
2619 /* The main hash table. */
2620 struct elf_link_hash_table root;
2621
2622 /* Nonzero to force PIC branch veneers. */
2623 int pic_veneer;
2624
2625 /* Fix erratum 835769. */
2626 int fix_erratum_835769;
2627
2628 /* Fix erratum 843419. */
2629 erratum_84319_opts fix_erratum_843419;
2630
2631 /* Don't apply link-time values for dynamic relocations. */
2632 int no_apply_dynamic_relocs;
2633
2634 /* The number of bytes in the initial entry in the PLT. */
2635 bfd_size_type plt_header_size;
2636
2637 /* The bytes of the initial PLT entry. */
2638 const bfd_byte *plt0_entry;
2639
2640 /* The number of bytes in the subsequent PLT entries. */
2641 bfd_size_type plt_entry_size;
2642
2643 /* The bytes of the subsequent PLT entry. */
2644 const bfd_byte *plt_entry;
2645
2646 /* For convenience in allocate_dynrelocs. */
2647 bfd *obfd;
2648
2649 /* The amount of space used by the reserved portion of the sgotplt
2650 section, plus whatever space is used by the jump slots. */
2651 bfd_vma sgotplt_jump_table_size;
2652
2653 /* The stub hash table. */
2654 struct bfd_hash_table stub_hash_table;
2655
2656 /* Linker stub bfd. */
2657 bfd *stub_bfd;
2658
2659 /* Linker call-backs. */
2660 asection *(*add_stub_section) (const char *, asection *);
2661 void (*layout_sections_again) (void);
2662
2663 /* Array to keep track of which stub sections have been created, and
2664 information on stub grouping. */
2665 struct map_stub
2666 {
2667 /* This is the section to which stubs in the group will be
2668 attached. */
2669 asection *link_sec;
2670 /* The stub section. */
2671 asection *stub_sec;
2672 } *stub_group;
2673
2674 /* Assorted information used by elfNN_aarch64_size_stubs. */
2675 unsigned int bfd_count;
2676 unsigned int top_index;
2677 asection **input_list;
2678
2679 /* JUMP_SLOT relocs for variant PCS symbols may be present. */
2680 int variant_pcs;
2681
2682 /* The number of bytes in the PLT enty for the TLS descriptor. */
2683 bfd_size_type tlsdesc_plt_entry_size;
2684
2685 /* Used by local STT_GNU_IFUNC symbols. */
2686 htab_t loc_hash_table;
2687 void * loc_hash_memory;
2688 };
2689
2690 /* Create an entry in an AArch64 ELF linker hash table. */
2691
2692 static struct bfd_hash_entry *
elfNN_aarch64_link_hash_newfunc(struct bfd_hash_entry * entry,struct bfd_hash_table * table,const char * string)2693 elfNN_aarch64_link_hash_newfunc (struct bfd_hash_entry *entry,
2694 struct bfd_hash_table *table,
2695 const char *string)
2696 {
2697 struct elf_aarch64_link_hash_entry *ret =
2698 (struct elf_aarch64_link_hash_entry *) entry;
2699
2700 /* Allocate the structure if it has not already been allocated by a
2701 subclass. */
2702 if (ret == NULL)
2703 ret = bfd_hash_allocate (table,
2704 sizeof (struct elf_aarch64_link_hash_entry));
2705 if (ret == NULL)
2706 return (struct bfd_hash_entry *) ret;
2707
2708 /* Call the allocation method of the superclass. */
2709 ret = ((struct elf_aarch64_link_hash_entry *)
2710 _bfd_elf_link_hash_newfunc ((struct bfd_hash_entry *) ret,
2711 table, string));
2712 if (ret != NULL)
2713 {
2714 ret->got_type = GOT_UNKNOWN;
2715 ret->plt_got_offset = (bfd_vma) - 1;
2716 ret->stub_cache = NULL;
2717 ret->tlsdesc_got_jump_table_offset = (bfd_vma) - 1;
2718 }
2719
2720 return (struct bfd_hash_entry *) ret;
2721 }
2722
2723 /* Initialize an entry in the stub hash table. */
2724
2725 static struct bfd_hash_entry *
stub_hash_newfunc(struct bfd_hash_entry * entry,struct bfd_hash_table * table,const char * string)2726 stub_hash_newfunc (struct bfd_hash_entry *entry,
2727 struct bfd_hash_table *table, const char *string)
2728 {
2729 /* Allocate the structure if it has not already been allocated by a
2730 subclass. */
2731 if (entry == NULL)
2732 {
2733 entry = bfd_hash_allocate (table,
2734 sizeof (struct
2735 elf_aarch64_stub_hash_entry));
2736 if (entry == NULL)
2737 return entry;
2738 }
2739
2740 /* Call the allocation method of the superclass. */
2741 entry = bfd_hash_newfunc (entry, table, string);
2742 if (entry != NULL)
2743 {
2744 struct elf_aarch64_stub_hash_entry *eh;
2745
2746 /* Initialize the local fields. */
2747 eh = (struct elf_aarch64_stub_hash_entry *) entry;
2748 eh->adrp_offset = 0;
2749 eh->stub_sec = NULL;
2750 eh->stub_offset = 0;
2751 eh->target_value = 0;
2752 eh->target_section = NULL;
2753 eh->stub_type = aarch64_stub_none;
2754 eh->h = NULL;
2755 eh->id_sec = NULL;
2756 }
2757
2758 return entry;
2759 }
2760
2761 /* Compute a hash of a local hash entry. We use elf_link_hash_entry
2762 for local symbol so that we can handle local STT_GNU_IFUNC symbols
2763 as global symbol. We reuse indx and dynstr_index for local symbol
2764 hash since they aren't used by global symbols in this backend. */
2765
2766 static hashval_t
elfNN_aarch64_local_htab_hash(const void * ptr)2767 elfNN_aarch64_local_htab_hash (const void *ptr)
2768 {
2769 struct elf_link_hash_entry *h
2770 = (struct elf_link_hash_entry *) ptr;
2771 return ELF_LOCAL_SYMBOL_HASH (h->indx, h->dynstr_index);
2772 }
2773
2774 /* Compare local hash entries. */
2775
2776 static int
elfNN_aarch64_local_htab_eq(const void * ptr1,const void * ptr2)2777 elfNN_aarch64_local_htab_eq (const void *ptr1, const void *ptr2)
2778 {
2779 struct elf_link_hash_entry *h1
2780 = (struct elf_link_hash_entry *) ptr1;
2781 struct elf_link_hash_entry *h2
2782 = (struct elf_link_hash_entry *) ptr2;
2783
2784 return h1->indx == h2->indx && h1->dynstr_index == h2->dynstr_index;
2785 }
2786
2787 /* Find and/or create a hash entry for local symbol. */
2788
2789 static struct elf_link_hash_entry *
elfNN_aarch64_get_local_sym_hash(struct elf_aarch64_link_hash_table * htab,bfd * abfd,const Elf_Internal_Rela * rel,bool create)2790 elfNN_aarch64_get_local_sym_hash (struct elf_aarch64_link_hash_table *htab,
2791 bfd *abfd, const Elf_Internal_Rela *rel,
2792 bool create)
2793 {
2794 struct elf_aarch64_link_hash_entry e, *ret;
2795 asection *sec = abfd->sections;
2796 hashval_t h = ELF_LOCAL_SYMBOL_HASH (sec->id,
2797 ELFNN_R_SYM (rel->r_info));
2798 void **slot;
2799
2800 e.root.indx = sec->id;
2801 e.root.dynstr_index = ELFNN_R_SYM (rel->r_info);
2802 slot = htab_find_slot_with_hash (htab->loc_hash_table, &e, h,
2803 create ? INSERT : NO_INSERT);
2804
2805 if (!slot)
2806 return NULL;
2807
2808 if (*slot)
2809 {
2810 ret = (struct elf_aarch64_link_hash_entry *) *slot;
2811 return &ret->root;
2812 }
2813
2814 ret = (struct elf_aarch64_link_hash_entry *)
2815 objalloc_alloc ((struct objalloc *) htab->loc_hash_memory,
2816 sizeof (struct elf_aarch64_link_hash_entry));
2817 if (ret)
2818 {
2819 memset (ret, 0, sizeof (*ret));
2820 ret->root.indx = sec->id;
2821 ret->root.dynstr_index = ELFNN_R_SYM (rel->r_info);
2822 ret->root.dynindx = -1;
2823 *slot = ret;
2824 }
2825 return &ret->root;
2826 }
2827
2828 /* Copy the extra info we tack onto an elf_link_hash_entry. */
2829
2830 static void
elfNN_aarch64_copy_indirect_symbol(struct bfd_link_info * info,struct elf_link_hash_entry * dir,struct elf_link_hash_entry * ind)2831 elfNN_aarch64_copy_indirect_symbol (struct bfd_link_info *info,
2832 struct elf_link_hash_entry *dir,
2833 struct elf_link_hash_entry *ind)
2834 {
2835 struct elf_aarch64_link_hash_entry *edir, *eind;
2836
2837 edir = (struct elf_aarch64_link_hash_entry *) dir;
2838 eind = (struct elf_aarch64_link_hash_entry *) ind;
2839
2840 if (ind->root.type == bfd_link_hash_indirect)
2841 {
2842 /* Copy over PLT info. */
2843 if (dir->got.refcount <= 0)
2844 {
2845 edir->got_type = eind->got_type;
2846 eind->got_type = GOT_UNKNOWN;
2847 }
2848 }
2849
2850 _bfd_elf_link_hash_copy_indirect (info, dir, ind);
2851 }
2852
2853 /* Merge non-visibility st_other attributes. */
2854
2855 static void
elfNN_aarch64_merge_symbol_attribute(struct elf_link_hash_entry * h,unsigned int st_other,bool definition ATTRIBUTE_UNUSED,bool dynamic ATTRIBUTE_UNUSED)2856 elfNN_aarch64_merge_symbol_attribute (struct elf_link_hash_entry *h,
2857 unsigned int st_other,
2858 bool definition ATTRIBUTE_UNUSED,
2859 bool dynamic ATTRIBUTE_UNUSED)
2860 {
2861 unsigned int isym_sto = st_other & ~ELF_ST_VISIBILITY (-1);
2862 unsigned int h_sto = h->other & ~ELF_ST_VISIBILITY (-1);
2863
2864 if (isym_sto == h_sto)
2865 return;
2866
2867 if (isym_sto & ~STO_AARCH64_VARIANT_PCS)
2868 /* Not fatal, this callback cannot fail. */
2869 _bfd_error_handler (_("unknown attribute for symbol `%s': 0x%02x"),
2870 h->root.root.string, isym_sto);
2871
2872 /* Note: Ideally we would warn about any attribute mismatch, but
2873 this api does not allow that without substantial changes. */
2874 if (isym_sto & STO_AARCH64_VARIANT_PCS)
2875 h->other |= STO_AARCH64_VARIANT_PCS;
2876 }
2877
2878 /* Destroy an AArch64 elf linker hash table. */
2879
2880 static void
elfNN_aarch64_link_hash_table_free(bfd * obfd)2881 elfNN_aarch64_link_hash_table_free (bfd *obfd)
2882 {
2883 struct elf_aarch64_link_hash_table *ret
2884 = (struct elf_aarch64_link_hash_table *) obfd->link.hash;
2885
2886 if (ret->loc_hash_table)
2887 htab_delete (ret->loc_hash_table);
2888 if (ret->loc_hash_memory)
2889 objalloc_free ((struct objalloc *) ret->loc_hash_memory);
2890
2891 bfd_hash_table_free (&ret->stub_hash_table);
2892 _bfd_elf_link_hash_table_free (obfd);
2893 }
2894
2895 /* Create an AArch64 elf linker hash table. */
2896
2897 static struct bfd_link_hash_table *
elfNN_aarch64_link_hash_table_create(bfd * abfd)2898 elfNN_aarch64_link_hash_table_create (bfd *abfd)
2899 {
2900 struct elf_aarch64_link_hash_table *ret;
2901 size_t amt = sizeof (struct elf_aarch64_link_hash_table);
2902
2903 ret = bfd_zmalloc (amt);
2904 if (ret == NULL)
2905 return NULL;
2906
2907 if (!_bfd_elf_link_hash_table_init
2908 (&ret->root, abfd, elfNN_aarch64_link_hash_newfunc,
2909 sizeof (struct elf_aarch64_link_hash_entry), AARCH64_ELF_DATA))
2910 {
2911 free (ret);
2912 return NULL;
2913 }
2914
2915 ret->plt_header_size = PLT_ENTRY_SIZE;
2916 ret->plt0_entry = elfNN_aarch64_small_plt0_entry;
2917 ret->plt_entry_size = PLT_SMALL_ENTRY_SIZE;
2918 ret->plt_entry = elfNN_aarch64_small_plt_entry;
2919 ret->tlsdesc_plt_entry_size = PLT_TLSDESC_ENTRY_SIZE;
2920 ret->obfd = abfd;
2921 ret->root.tlsdesc_got = (bfd_vma) - 1;
2922
2923 if (!bfd_hash_table_init (&ret->stub_hash_table, stub_hash_newfunc,
2924 sizeof (struct elf_aarch64_stub_hash_entry)))
2925 {
2926 _bfd_elf_link_hash_table_free (abfd);
2927 return NULL;
2928 }
2929
2930 ret->loc_hash_table = htab_try_create (1024,
2931 elfNN_aarch64_local_htab_hash,
2932 elfNN_aarch64_local_htab_eq,
2933 NULL);
2934 ret->loc_hash_memory = objalloc_create ();
2935 if (!ret->loc_hash_table || !ret->loc_hash_memory)
2936 {
2937 elfNN_aarch64_link_hash_table_free (abfd);
2938 return NULL;
2939 }
2940 ret->root.root.hash_table_free = elfNN_aarch64_link_hash_table_free;
2941
2942 return &ret->root.root;
2943 }
2944
2945 /* Perform relocation R_TYPE. Returns TRUE upon success, FALSE otherwise. */
2946
2947 static bool
aarch64_relocate(unsigned int r_type,bfd * input_bfd,asection * input_section,bfd_vma offset,bfd_vma value)2948 aarch64_relocate (unsigned int r_type, bfd *input_bfd, asection *input_section,
2949 bfd_vma offset, bfd_vma value)
2950 {
2951 reloc_howto_type *howto;
2952 bfd_vma place;
2953
2954 howto = elfNN_aarch64_howto_from_type (input_bfd, r_type);
2955 place = (input_section->output_section->vma + input_section->output_offset
2956 + offset);
2957
2958 r_type = elfNN_aarch64_bfd_reloc_from_type (input_bfd, r_type);
2959 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, r_type, place,
2960 value, 0, false);
2961 return _bfd_aarch64_elf_put_addend (input_bfd,
2962 input_section->contents + offset, r_type,
2963 howto, value) == bfd_reloc_ok;
2964 }
2965
2966 static enum elf_aarch64_stub_type
aarch64_select_branch_stub(bfd_vma value,bfd_vma place)2967 aarch64_select_branch_stub (bfd_vma value, bfd_vma place)
2968 {
2969 if (aarch64_valid_for_adrp_p (value, place))
2970 return aarch64_stub_adrp_branch;
2971 return aarch64_stub_long_branch;
2972 }
2973
2974 /* Determine the type of stub needed, if any, for a call. */
2975
2976 static enum elf_aarch64_stub_type
aarch64_type_of_stub(asection * input_sec,const Elf_Internal_Rela * rel,asection * sym_sec,unsigned char st_type,bfd_vma destination)2977 aarch64_type_of_stub (asection *input_sec,
2978 const Elf_Internal_Rela *rel,
2979 asection *sym_sec,
2980 unsigned char st_type,
2981 bfd_vma destination)
2982 {
2983 bfd_vma location;
2984 bfd_signed_vma branch_offset;
2985 unsigned int r_type;
2986 enum elf_aarch64_stub_type stub_type = aarch64_stub_none;
2987
2988 if (st_type != STT_FUNC
2989 && (sym_sec == input_sec))
2990 return stub_type;
2991
2992 /* Determine where the call point is. */
2993 location = (input_sec->output_offset
2994 + input_sec->output_section->vma + rel->r_offset);
2995
2996 branch_offset = (bfd_signed_vma) (destination - location);
2997
2998 r_type = ELFNN_R_TYPE (rel->r_info);
2999
3000 /* We don't want to redirect any old unconditional jump in this way,
3001 only one which is being used for a sibcall, where it is
3002 acceptable for the IP0 and IP1 registers to be clobbered. */
3003 if ((r_type == AARCH64_R (CALL26) || r_type == AARCH64_R (JUMP26))
3004 && (branch_offset > AARCH64_MAX_FWD_BRANCH_OFFSET
3005 || branch_offset < AARCH64_MAX_BWD_BRANCH_OFFSET))
3006 {
3007 stub_type = aarch64_stub_long_branch;
3008 }
3009
3010 return stub_type;
3011 }
3012
3013 /* Build a name for an entry in the stub hash table. */
3014
3015 static char *
elfNN_aarch64_stub_name(const asection * input_section,const asection * sym_sec,const struct elf_aarch64_link_hash_entry * hash,const Elf_Internal_Rela * rel)3016 elfNN_aarch64_stub_name (const asection *input_section,
3017 const asection *sym_sec,
3018 const struct elf_aarch64_link_hash_entry *hash,
3019 const Elf_Internal_Rela *rel)
3020 {
3021 char *stub_name;
3022 bfd_size_type len;
3023
3024 if (hash)
3025 {
3026 len = 8 + 1 + strlen (hash->root.root.root.string) + 1 + 16 + 1;
3027 stub_name = bfd_malloc (len);
3028 if (stub_name != NULL)
3029 snprintf (stub_name, len, "%08x_%s+%" BFD_VMA_FMT "x",
3030 (unsigned int) input_section->id,
3031 hash->root.root.root.string,
3032 rel->r_addend);
3033 }
3034 else
3035 {
3036 len = 8 + 1 + 8 + 1 + 8 + 1 + 16 + 1;
3037 stub_name = bfd_malloc (len);
3038 if (stub_name != NULL)
3039 snprintf (stub_name, len, "%08x_%x:%x+%" BFD_VMA_FMT "x",
3040 (unsigned int) input_section->id,
3041 (unsigned int) sym_sec->id,
3042 (unsigned int) ELFNN_R_SYM (rel->r_info),
3043 rel->r_addend);
3044 }
3045
3046 return stub_name;
3047 }
3048
3049 /* Return TRUE if symbol H should be hashed in the `.gnu.hash' section. For
3050 executable PLT slots where the executable never takes the address of those
3051 functions, the function symbols are not added to the hash table. */
3052
3053 static bool
elf_aarch64_hash_symbol(struct elf_link_hash_entry * h)3054 elf_aarch64_hash_symbol (struct elf_link_hash_entry *h)
3055 {
3056 if (h->plt.offset != (bfd_vma) -1
3057 && !h->def_regular
3058 && !h->pointer_equality_needed)
3059 return false;
3060
3061 return _bfd_elf_hash_symbol (h);
3062 }
3063
3064
3065 /* Look up an entry in the stub hash. Stub entries are cached because
3066 creating the stub name takes a bit of time. */
3067
3068 static struct elf_aarch64_stub_hash_entry *
elfNN_aarch64_get_stub_entry(const asection * input_section,const asection * sym_sec,struct elf_link_hash_entry * hash,const Elf_Internal_Rela * rel,struct elf_aarch64_link_hash_table * htab)3069 elfNN_aarch64_get_stub_entry (const asection *input_section,
3070 const asection *sym_sec,
3071 struct elf_link_hash_entry *hash,
3072 const Elf_Internal_Rela *rel,
3073 struct elf_aarch64_link_hash_table *htab)
3074 {
3075 struct elf_aarch64_stub_hash_entry *stub_entry;
3076 struct elf_aarch64_link_hash_entry *h =
3077 (struct elf_aarch64_link_hash_entry *) hash;
3078 const asection *id_sec;
3079
3080 if ((input_section->flags & SEC_CODE) == 0)
3081 return NULL;
3082
3083 /* If this input section is part of a group of sections sharing one
3084 stub section, then use the id of the first section in the group.
3085 Stub names need to include a section id, as there may well be
3086 more than one stub used to reach say, printf, and we need to
3087 distinguish between them. */
3088 id_sec = htab->stub_group[input_section->id].link_sec;
3089
3090 if (h != NULL && h->stub_cache != NULL
3091 && h->stub_cache->h == h && h->stub_cache->id_sec == id_sec)
3092 {
3093 stub_entry = h->stub_cache;
3094 }
3095 else
3096 {
3097 char *stub_name;
3098
3099 stub_name = elfNN_aarch64_stub_name (id_sec, sym_sec, h, rel);
3100 if (stub_name == NULL)
3101 return NULL;
3102
3103 stub_entry = aarch64_stub_hash_lookup (&htab->stub_hash_table,
3104 stub_name, false, false);
3105 if (h != NULL)
3106 h->stub_cache = stub_entry;
3107
3108 free (stub_name);
3109 }
3110
3111 return stub_entry;
3112 }
3113
3114
3115 /* Create a stub section. */
3116
3117 static asection *
_bfd_aarch64_create_stub_section(asection * section,struct elf_aarch64_link_hash_table * htab)3118 _bfd_aarch64_create_stub_section (asection *section,
3119 struct elf_aarch64_link_hash_table *htab)
3120 {
3121 size_t namelen;
3122 bfd_size_type len;
3123 char *s_name;
3124
3125 namelen = strlen (section->name);
3126 len = namelen + sizeof (STUB_SUFFIX);
3127 s_name = bfd_alloc (htab->stub_bfd, len);
3128 if (s_name == NULL)
3129 return NULL;
3130
3131 memcpy (s_name, section->name, namelen);
3132 memcpy (s_name + namelen, STUB_SUFFIX, sizeof (STUB_SUFFIX));
3133 return (*htab->add_stub_section) (s_name, section);
3134 }
3135
3136
3137 /* Find or create a stub section for a link section.
3138
3139 Fix or create the stub section used to collect stubs attached to
3140 the specified link section. */
3141
3142 static asection *
_bfd_aarch64_get_stub_for_link_section(asection * link_section,struct elf_aarch64_link_hash_table * htab)3143 _bfd_aarch64_get_stub_for_link_section (asection *link_section,
3144 struct elf_aarch64_link_hash_table *htab)
3145 {
3146 if (htab->stub_group[link_section->id].stub_sec == NULL)
3147 htab->stub_group[link_section->id].stub_sec
3148 = _bfd_aarch64_create_stub_section (link_section, htab);
3149 return htab->stub_group[link_section->id].stub_sec;
3150 }
3151
3152
3153 /* Find or create a stub section in the stub group for an input
3154 section. */
3155
3156 static asection *
_bfd_aarch64_create_or_find_stub_sec(asection * section,struct elf_aarch64_link_hash_table * htab)3157 _bfd_aarch64_create_or_find_stub_sec (asection *section,
3158 struct elf_aarch64_link_hash_table *htab)
3159 {
3160 asection *link_sec = htab->stub_group[section->id].link_sec;
3161 return _bfd_aarch64_get_stub_for_link_section (link_sec, htab);
3162 }
3163
3164
3165 /* Add a new stub entry in the stub group associated with an input
3166 section to the stub hash. Not all fields of the new stub entry are
3167 initialised. */
3168
3169 static struct elf_aarch64_stub_hash_entry *
_bfd_aarch64_add_stub_entry_in_group(const char * stub_name,asection * section,struct elf_aarch64_link_hash_table * htab)3170 _bfd_aarch64_add_stub_entry_in_group (const char *stub_name,
3171 asection *section,
3172 struct elf_aarch64_link_hash_table *htab)
3173 {
3174 asection *link_sec;
3175 asection *stub_sec;
3176 struct elf_aarch64_stub_hash_entry *stub_entry;
3177
3178 link_sec = htab->stub_group[section->id].link_sec;
3179 stub_sec = _bfd_aarch64_create_or_find_stub_sec (section, htab);
3180
3181 /* Enter this entry into the linker stub hash table. */
3182 stub_entry = aarch64_stub_hash_lookup (&htab->stub_hash_table, stub_name,
3183 true, false);
3184 if (stub_entry == NULL)
3185 {
3186 /* xgettext:c-format */
3187 _bfd_error_handler (_("%pB: cannot create stub entry %s"),
3188 section->owner, stub_name);
3189 return NULL;
3190 }
3191
3192 stub_entry->stub_sec = stub_sec;
3193 stub_entry->stub_offset = 0;
3194 stub_entry->id_sec = link_sec;
3195
3196 return stub_entry;
3197 }
3198
3199 /* Add a new stub entry in the final stub section to the stub hash.
3200 Not all fields of the new stub entry are initialised. */
3201
3202 static struct elf_aarch64_stub_hash_entry *
_bfd_aarch64_add_stub_entry_after(const char * stub_name,asection * link_section,struct elf_aarch64_link_hash_table * htab)3203 _bfd_aarch64_add_stub_entry_after (const char *stub_name,
3204 asection *link_section,
3205 struct elf_aarch64_link_hash_table *htab)
3206 {
3207 asection *stub_sec;
3208 struct elf_aarch64_stub_hash_entry *stub_entry;
3209
3210 stub_sec = NULL;
3211 /* Only create the actual stub if we will end up needing it. */
3212 if (htab->fix_erratum_843419 & ERRAT_ADRP)
3213 stub_sec = _bfd_aarch64_get_stub_for_link_section (link_section, htab);
3214 stub_entry = aarch64_stub_hash_lookup (&htab->stub_hash_table, stub_name,
3215 true, false);
3216 if (stub_entry == NULL)
3217 {
3218 _bfd_error_handler (_("cannot create stub entry %s"), stub_name);
3219 return NULL;
3220 }
3221
3222 stub_entry->stub_sec = stub_sec;
3223 stub_entry->stub_offset = 0;
3224 stub_entry->id_sec = link_section;
3225
3226 return stub_entry;
3227 }
3228
3229
3230 static bool
aarch64_build_one_stub(struct bfd_hash_entry * gen_entry,void * in_arg)3231 aarch64_build_one_stub (struct bfd_hash_entry *gen_entry,
3232 void *in_arg)
3233 {
3234 struct elf_aarch64_stub_hash_entry *stub_entry;
3235 asection *stub_sec;
3236 bfd *stub_bfd;
3237 bfd_byte *loc;
3238 bfd_vma sym_value;
3239 bfd_vma veneered_insn_loc;
3240 bfd_vma veneer_entry_loc;
3241 bfd_signed_vma branch_offset = 0;
3242 unsigned int template_size;
3243 const uint32_t *template;
3244 unsigned int i;
3245 struct bfd_link_info *info;
3246
3247 /* Massage our args to the form they really have. */
3248 stub_entry = (struct elf_aarch64_stub_hash_entry *) gen_entry;
3249
3250 info = (struct bfd_link_info *) in_arg;
3251
3252 /* Fail if the target section could not be assigned to an output
3253 section. The user should fix his linker script. */
3254 if (stub_entry->target_section->output_section == NULL
3255 && info->non_contiguous_regions)
3256 info->callbacks->einfo (_("%F%P: Could not assign '%pA' to an output section. "
3257 "Retry without "
3258 "--enable-non-contiguous-regions.\n"),
3259 stub_entry->target_section);
3260
3261 stub_sec = stub_entry->stub_sec;
3262
3263 /* Make a note of the offset within the stubs for this entry. */
3264 stub_entry->stub_offset = stub_sec->size;
3265 loc = stub_sec->contents + stub_entry->stub_offset;
3266
3267 stub_bfd = stub_sec->owner;
3268
3269 /* This is the address of the stub destination. */
3270 sym_value = (stub_entry->target_value
3271 + stub_entry->target_section->output_offset
3272 + stub_entry->target_section->output_section->vma);
3273
3274 if (stub_entry->stub_type == aarch64_stub_long_branch)
3275 {
3276 bfd_vma place = (stub_entry->stub_offset + stub_sec->output_section->vma
3277 + stub_sec->output_offset);
3278
3279 /* See if we can relax the stub. */
3280 if (aarch64_valid_for_adrp_p (sym_value, place))
3281 stub_entry->stub_type = aarch64_select_branch_stub (sym_value, place);
3282 }
3283
3284 switch (stub_entry->stub_type)
3285 {
3286 case aarch64_stub_adrp_branch:
3287 template = aarch64_adrp_branch_stub;
3288 template_size = sizeof (aarch64_adrp_branch_stub);
3289 break;
3290 case aarch64_stub_long_branch:
3291 template = aarch64_long_branch_stub;
3292 template_size = sizeof (aarch64_long_branch_stub);
3293 break;
3294 case aarch64_stub_erratum_835769_veneer:
3295 template = aarch64_erratum_835769_stub;
3296 template_size = sizeof (aarch64_erratum_835769_stub);
3297 break;
3298 case aarch64_stub_erratum_843419_veneer:
3299 template = aarch64_erratum_843419_stub;
3300 template_size = sizeof (aarch64_erratum_843419_stub);
3301 break;
3302 default:
3303 abort ();
3304 }
3305
3306 for (i = 0; i < (template_size / sizeof template[0]); i++)
3307 {
3308 bfd_putl32 (template[i], loc);
3309 loc += 4;
3310 }
3311
3312 template_size = (template_size + 7) & ~7;
3313 stub_sec->size += template_size;
3314
3315 switch (stub_entry->stub_type)
3316 {
3317 case aarch64_stub_adrp_branch:
3318 if (!aarch64_relocate (AARCH64_R (ADR_PREL_PG_HI21), stub_bfd, stub_sec,
3319 stub_entry->stub_offset, sym_value))
3320 /* The stub would not have been relaxed if the offset was out
3321 of range. */
3322 BFD_FAIL ();
3323
3324 if (!aarch64_relocate (AARCH64_R (ADD_ABS_LO12_NC), stub_bfd, stub_sec,
3325 stub_entry->stub_offset + 4, sym_value))
3326 BFD_FAIL ();
3327 break;
3328
3329 case aarch64_stub_long_branch:
3330 /* We want the value relative to the address 12 bytes back from the
3331 value itself. */
3332 if (!aarch64_relocate (AARCH64_R (PRELNN), stub_bfd, stub_sec,
3333 stub_entry->stub_offset + 16, sym_value + 12))
3334 BFD_FAIL ();
3335 break;
3336
3337 case aarch64_stub_erratum_835769_veneer:
3338 veneered_insn_loc = stub_entry->target_section->output_section->vma
3339 + stub_entry->target_section->output_offset
3340 + stub_entry->target_value;
3341 veneer_entry_loc = stub_entry->stub_sec->output_section->vma
3342 + stub_entry->stub_sec->output_offset
3343 + stub_entry->stub_offset;
3344 branch_offset = veneered_insn_loc - veneer_entry_loc;
3345 branch_offset >>= 2;
3346 branch_offset &= 0x3ffffff;
3347 bfd_putl32 (stub_entry->veneered_insn,
3348 stub_sec->contents + stub_entry->stub_offset);
3349 bfd_putl32 (template[1] | branch_offset,
3350 stub_sec->contents + stub_entry->stub_offset + 4);
3351 break;
3352
3353 case aarch64_stub_erratum_843419_veneer:
3354 if (!aarch64_relocate (AARCH64_R (JUMP26), stub_bfd, stub_sec,
3355 stub_entry->stub_offset + 4, sym_value + 4))
3356 BFD_FAIL ();
3357 break;
3358
3359 default:
3360 abort ();
3361 }
3362
3363 return true;
3364 }
3365
3366 /* As above, but don't actually build the stub. Just bump offset so
3367 we know stub section sizes. */
3368
3369 static bool
aarch64_size_one_stub(struct bfd_hash_entry * gen_entry,void * in_arg)3370 aarch64_size_one_stub (struct bfd_hash_entry *gen_entry, void *in_arg)
3371 {
3372 struct elf_aarch64_stub_hash_entry *stub_entry;
3373 struct elf_aarch64_link_hash_table *htab;
3374 int size;
3375
3376 /* Massage our args to the form they really have. */
3377 stub_entry = (struct elf_aarch64_stub_hash_entry *) gen_entry;
3378 htab = (struct elf_aarch64_link_hash_table *) in_arg;
3379
3380 switch (stub_entry->stub_type)
3381 {
3382 case aarch64_stub_adrp_branch:
3383 size = sizeof (aarch64_adrp_branch_stub);
3384 break;
3385 case aarch64_stub_long_branch:
3386 size = sizeof (aarch64_long_branch_stub);
3387 break;
3388 case aarch64_stub_erratum_835769_veneer:
3389 size = sizeof (aarch64_erratum_835769_stub);
3390 break;
3391 case aarch64_stub_erratum_843419_veneer:
3392 {
3393 if (htab->fix_erratum_843419 == ERRAT_ADR)
3394 return true;
3395 size = sizeof (aarch64_erratum_843419_stub);
3396 }
3397 break;
3398 default:
3399 abort ();
3400 }
3401
3402 size = (size + 7) & ~7;
3403 stub_entry->stub_sec->size += size;
3404 return true;
3405 }
3406
3407 /* External entry points for sizing and building linker stubs. */
3408
3409 /* Set up various things so that we can make a list of input sections
3410 for each output section included in the link. Returns -1 on error,
3411 0 when no stubs will be needed, and 1 on success. */
3412
3413 int
elfNN_aarch64_setup_section_lists(bfd * output_bfd,struct bfd_link_info * info)3414 elfNN_aarch64_setup_section_lists (bfd *output_bfd,
3415 struct bfd_link_info *info)
3416 {
3417 bfd *input_bfd;
3418 unsigned int bfd_count;
3419 unsigned int top_id, top_index;
3420 asection *section;
3421 asection **input_list, **list;
3422 size_t amt;
3423 struct elf_aarch64_link_hash_table *htab =
3424 elf_aarch64_hash_table (info);
3425
3426 if (!is_elf_hash_table (&htab->root.root))
3427 return 0;
3428
3429 /* Count the number of input BFDs and find the top input section id. */
3430 for (input_bfd = info->input_bfds, bfd_count = 0, top_id = 0;
3431 input_bfd != NULL; input_bfd = input_bfd->link.next)
3432 {
3433 bfd_count += 1;
3434 for (section = input_bfd->sections;
3435 section != NULL; section = section->next)
3436 {
3437 if (top_id < section->id)
3438 top_id = section->id;
3439 }
3440 }
3441 htab->bfd_count = bfd_count;
3442
3443 amt = sizeof (struct map_stub) * (top_id + 1);
3444 htab->stub_group = bfd_zmalloc (amt);
3445 if (htab->stub_group == NULL)
3446 return -1;
3447
3448 /* We can't use output_bfd->section_count here to find the top output
3449 section index as some sections may have been removed, and
3450 _bfd_strip_section_from_output doesn't renumber the indices. */
3451 for (section = output_bfd->sections, top_index = 0;
3452 section != NULL; section = section->next)
3453 {
3454 if (top_index < section->index)
3455 top_index = section->index;
3456 }
3457
3458 htab->top_index = top_index;
3459 amt = sizeof (asection *) * (top_index + 1);
3460 input_list = bfd_malloc (amt);
3461 htab->input_list = input_list;
3462 if (input_list == NULL)
3463 return -1;
3464
3465 /* For sections we aren't interested in, mark their entries with a
3466 value we can check later. */
3467 list = input_list + top_index;
3468 do
3469 *list = bfd_abs_section_ptr;
3470 while (list-- != input_list);
3471
3472 for (section = output_bfd->sections;
3473 section != NULL; section = section->next)
3474 {
3475 if ((section->flags & SEC_CODE) != 0)
3476 input_list[section->index] = NULL;
3477 }
3478
3479 return 1;
3480 }
3481
3482 /* Used by elfNN_aarch64_next_input_section and group_sections. */
3483 #define PREV_SEC(sec) (htab->stub_group[(sec)->id].link_sec)
3484
3485 /* The linker repeatedly calls this function for each input section,
3486 in the order that input sections are linked into output sections.
3487 Build lists of input sections to determine groupings between which
3488 we may insert linker stubs. */
3489
3490 void
elfNN_aarch64_next_input_section(struct bfd_link_info * info,asection * isec)3491 elfNN_aarch64_next_input_section (struct bfd_link_info *info, asection *isec)
3492 {
3493 struct elf_aarch64_link_hash_table *htab =
3494 elf_aarch64_hash_table (info);
3495
3496 if (isec->output_section->index <= htab->top_index)
3497 {
3498 asection **list = htab->input_list + isec->output_section->index;
3499
3500 if (*list != bfd_abs_section_ptr && (isec->flags & SEC_CODE) != 0)
3501 {
3502 /* Steal the link_sec pointer for our list. */
3503 /* This happens to make the list in reverse order,
3504 which is what we want. */
3505 PREV_SEC (isec) = *list;
3506 *list = isec;
3507 }
3508 }
3509 }
3510
3511 /* See whether we can group stub sections together. Grouping stub
3512 sections may result in fewer stubs. More importantly, we need to
3513 put all .init* and .fini* stubs at the beginning of the .init or
3514 .fini output sections respectively, because glibc splits the
3515 _init and _fini functions into multiple parts. Putting a stub in
3516 the middle of a function is not a good idea. */
3517
3518 static void
group_sections(struct elf_aarch64_link_hash_table * htab,bfd_size_type stub_group_size,bool stubs_always_after_branch)3519 group_sections (struct elf_aarch64_link_hash_table *htab,
3520 bfd_size_type stub_group_size,
3521 bool stubs_always_after_branch)
3522 {
3523 asection **list = htab->input_list;
3524
3525 do
3526 {
3527 asection *tail = *list;
3528 asection *head;
3529
3530 if (tail == bfd_abs_section_ptr)
3531 continue;
3532
3533 /* Reverse the list: we must avoid placing stubs at the
3534 beginning of the section because the beginning of the text
3535 section may be required for an interrupt vector in bare metal
3536 code. */
3537 #define NEXT_SEC PREV_SEC
3538 head = NULL;
3539 while (tail != NULL)
3540 {
3541 /* Pop from tail. */
3542 asection *item = tail;
3543 tail = PREV_SEC (item);
3544
3545 /* Push on head. */
3546 NEXT_SEC (item) = head;
3547 head = item;
3548 }
3549
3550 while (head != NULL)
3551 {
3552 asection *curr;
3553 asection *next;
3554 bfd_vma stub_group_start = head->output_offset;
3555 bfd_vma end_of_next;
3556
3557 curr = head;
3558 while (NEXT_SEC (curr) != NULL)
3559 {
3560 next = NEXT_SEC (curr);
3561 end_of_next = next->output_offset + next->size;
3562 if (end_of_next - stub_group_start >= stub_group_size)
3563 /* End of NEXT is too far from start, so stop. */
3564 break;
3565 /* Add NEXT to the group. */
3566 curr = next;
3567 }
3568
3569 /* OK, the size from the start to the start of CURR is less
3570 than stub_group_size and thus can be handled by one stub
3571 section. (Or the head section is itself larger than
3572 stub_group_size, in which case we may be toast.)
3573 We should really be keeping track of the total size of
3574 stubs added here, as stubs contribute to the final output
3575 section size. */
3576 do
3577 {
3578 next = NEXT_SEC (head);
3579 /* Set up this stub group. */
3580 htab->stub_group[head->id].link_sec = curr;
3581 }
3582 while (head != curr && (head = next) != NULL);
3583
3584 /* But wait, there's more! Input sections up to stub_group_size
3585 bytes after the stub section can be handled by it too. */
3586 if (!stubs_always_after_branch)
3587 {
3588 stub_group_start = curr->output_offset + curr->size;
3589
3590 while (next != NULL)
3591 {
3592 end_of_next = next->output_offset + next->size;
3593 if (end_of_next - stub_group_start >= stub_group_size)
3594 /* End of NEXT is too far from stubs, so stop. */
3595 break;
3596 /* Add NEXT to the stub group. */
3597 head = next;
3598 next = NEXT_SEC (head);
3599 htab->stub_group[head->id].link_sec = curr;
3600 }
3601 }
3602 head = next;
3603 }
3604 }
3605 while (list++ != htab->input_list + htab->top_index);
3606
3607 free (htab->input_list);
3608 }
3609
3610 #undef PREV_SEC
3611 #undef PREV_SEC
3612
3613 #define AARCH64_BITS(x, pos, n) (((x) >> (pos)) & ((1 << (n)) - 1))
3614
3615 #define AARCH64_RT(insn) AARCH64_BITS (insn, 0, 5)
3616 #define AARCH64_RT2(insn) AARCH64_BITS (insn, 10, 5)
3617 #define AARCH64_RA(insn) AARCH64_BITS (insn, 10, 5)
3618 #define AARCH64_RD(insn) AARCH64_BITS (insn, 0, 5)
3619 #define AARCH64_RN(insn) AARCH64_BITS (insn, 5, 5)
3620 #define AARCH64_RM(insn) AARCH64_BITS (insn, 16, 5)
3621
3622 #define AARCH64_MAC(insn) (((insn) & 0xff000000) == 0x9b000000)
3623 #define AARCH64_BIT(insn, n) AARCH64_BITS (insn, n, 1)
3624 #define AARCH64_OP31(insn) AARCH64_BITS (insn, 21, 3)
3625 #define AARCH64_ZR 0x1f
3626
3627 /* All ld/st ops. See C4-182 of the ARM ARM. The encoding space for
3628 LD_PCREL, LDST_RO, LDST_UI and LDST_UIMM cover prefetch ops. */
3629
3630 #define AARCH64_LD(insn) (AARCH64_BIT (insn, 22) == 1)
3631 #define AARCH64_LDST(insn) (((insn) & 0x0a000000) == 0x08000000)
3632 #define AARCH64_LDST_EX(insn) (((insn) & 0x3f000000) == 0x08000000)
3633 #define AARCH64_LDST_PCREL(insn) (((insn) & 0x3b000000) == 0x18000000)
3634 #define AARCH64_LDST_NAP(insn) (((insn) & 0x3b800000) == 0x28000000)
3635 #define AARCH64_LDSTP_PI(insn) (((insn) & 0x3b800000) == 0x28800000)
3636 #define AARCH64_LDSTP_O(insn) (((insn) & 0x3b800000) == 0x29000000)
3637 #define AARCH64_LDSTP_PRE(insn) (((insn) & 0x3b800000) == 0x29800000)
3638 #define AARCH64_LDST_UI(insn) (((insn) & 0x3b200c00) == 0x38000000)
3639 #define AARCH64_LDST_PIIMM(insn) (((insn) & 0x3b200c00) == 0x38000400)
3640 #define AARCH64_LDST_U(insn) (((insn) & 0x3b200c00) == 0x38000800)
3641 #define AARCH64_LDST_PREIMM(insn) (((insn) & 0x3b200c00) == 0x38000c00)
3642 #define AARCH64_LDST_RO(insn) (((insn) & 0x3b200c00) == 0x38200800)
3643 #define AARCH64_LDST_UIMM(insn) (((insn) & 0x3b000000) == 0x39000000)
3644 #define AARCH64_LDST_SIMD_M(insn) (((insn) & 0xbfbf0000) == 0x0c000000)
3645 #define AARCH64_LDST_SIMD_M_PI(insn) (((insn) & 0xbfa00000) == 0x0c800000)
3646 #define AARCH64_LDST_SIMD_S(insn) (((insn) & 0xbf9f0000) == 0x0d000000)
3647 #define AARCH64_LDST_SIMD_S_PI(insn) (((insn) & 0xbf800000) == 0x0d800000)
3648
3649 /* Classify an INSN if it is indeed a load/store.
3650
3651 Return TRUE if INSN is a LD/ST instruction otherwise return FALSE.
3652
3653 For scalar LD/ST instructions PAIR is FALSE, RT is returned and RT2
3654 is set equal to RT.
3655
3656 For LD/ST pair instructions PAIR is TRUE, RT and RT2 are returned. */
3657
3658 static bool
aarch64_mem_op_p(uint32_t insn,unsigned int * rt,unsigned int * rt2,bool * pair,bool * load)3659 aarch64_mem_op_p (uint32_t insn, unsigned int *rt, unsigned int *rt2,
3660 bool *pair, bool *load)
3661 {
3662 uint32_t opcode;
3663 unsigned int r;
3664 uint32_t opc = 0;
3665 uint32_t v = 0;
3666 uint32_t opc_v = 0;
3667
3668 /* Bail out quickly if INSN doesn't fall into the load-store
3669 encoding space. */
3670 if (!AARCH64_LDST (insn))
3671 return false;
3672
3673 *pair = false;
3674 *load = false;
3675 if (AARCH64_LDST_EX (insn))
3676 {
3677 *rt = AARCH64_RT (insn);
3678 *rt2 = *rt;
3679 if (AARCH64_BIT (insn, 21) == 1)
3680 {
3681 *pair = true;
3682 *rt2 = AARCH64_RT2 (insn);
3683 }
3684 *load = AARCH64_LD (insn);
3685 return true;
3686 }
3687 else if (AARCH64_LDST_NAP (insn)
3688 || AARCH64_LDSTP_PI (insn)
3689 || AARCH64_LDSTP_O (insn)
3690 || AARCH64_LDSTP_PRE (insn))
3691 {
3692 *pair = true;
3693 *rt = AARCH64_RT (insn);
3694 *rt2 = AARCH64_RT2 (insn);
3695 *load = AARCH64_LD (insn);
3696 return true;
3697 }
3698 else if (AARCH64_LDST_PCREL (insn)
3699 || AARCH64_LDST_UI (insn)
3700 || AARCH64_LDST_PIIMM (insn)
3701 || AARCH64_LDST_U (insn)
3702 || AARCH64_LDST_PREIMM (insn)
3703 || AARCH64_LDST_RO (insn)
3704 || AARCH64_LDST_UIMM (insn))
3705 {
3706 *rt = AARCH64_RT (insn);
3707 *rt2 = *rt;
3708 if (AARCH64_LDST_PCREL (insn))
3709 *load = true;
3710 opc = AARCH64_BITS (insn, 22, 2);
3711 v = AARCH64_BIT (insn, 26);
3712 opc_v = opc | (v << 2);
3713 *load = (opc_v == 1 || opc_v == 2 || opc_v == 3
3714 || opc_v == 5 || opc_v == 7);
3715 return true;
3716 }
3717 else if (AARCH64_LDST_SIMD_M (insn)
3718 || AARCH64_LDST_SIMD_M_PI (insn))
3719 {
3720 *rt = AARCH64_RT (insn);
3721 *load = AARCH64_BIT (insn, 22);
3722 opcode = (insn >> 12) & 0xf;
3723 switch (opcode)
3724 {
3725 case 0:
3726 case 2:
3727 *rt2 = *rt + 3;
3728 break;
3729
3730 case 4:
3731 case 6:
3732 *rt2 = *rt + 2;
3733 break;
3734
3735 case 7:
3736 *rt2 = *rt;
3737 break;
3738
3739 case 8:
3740 case 10:
3741 *rt2 = *rt + 1;
3742 break;
3743
3744 default:
3745 return false;
3746 }
3747 return true;
3748 }
3749 else if (AARCH64_LDST_SIMD_S (insn)
3750 || AARCH64_LDST_SIMD_S_PI (insn))
3751 {
3752 *rt = AARCH64_RT (insn);
3753 r = (insn >> 21) & 1;
3754 *load = AARCH64_BIT (insn, 22);
3755 opcode = (insn >> 13) & 0x7;
3756 switch (opcode)
3757 {
3758 case 0:
3759 case 2:
3760 case 4:
3761 *rt2 = *rt + r;
3762 break;
3763
3764 case 1:
3765 case 3:
3766 case 5:
3767 *rt2 = *rt + (r == 0 ? 2 : 3);
3768 break;
3769
3770 case 6:
3771 *rt2 = *rt + r;
3772 break;
3773
3774 case 7:
3775 *rt2 = *rt + (r == 0 ? 2 : 3);
3776 break;
3777
3778 default:
3779 return false;
3780 }
3781 return true;
3782 }
3783
3784 return false;
3785 }
3786
3787 /* Return TRUE if INSN is multiply-accumulate. */
3788
3789 static bool
aarch64_mlxl_p(uint32_t insn)3790 aarch64_mlxl_p (uint32_t insn)
3791 {
3792 uint32_t op31 = AARCH64_OP31 (insn);
3793
3794 if (AARCH64_MAC (insn)
3795 && (op31 == 0 || op31 == 1 || op31 == 5)
3796 /* Exclude MUL instructions which are encoded as a multiple accumulate
3797 with RA = XZR. */
3798 && AARCH64_RA (insn) != AARCH64_ZR)
3799 return true;
3800
3801 return false;
3802 }
3803
3804 /* Some early revisions of the Cortex-A53 have an erratum (835769) whereby
3805 it is possible for a 64-bit multiply-accumulate instruction to generate an
3806 incorrect result. The details are quite complex and hard to
3807 determine statically, since branches in the code may exist in some
3808 circumstances, but all cases end with a memory (load, store, or
3809 prefetch) instruction followed immediately by the multiply-accumulate
3810 operation. We employ a linker patching technique, by moving the potentially
3811 affected multiply-accumulate instruction into a patch region and replacing
3812 the original instruction with a branch to the patch. This function checks
3813 if INSN_1 is the memory operation followed by a multiply-accumulate
3814 operation (INSN_2). Return TRUE if an erratum sequence is found, FALSE
3815 if INSN_1 and INSN_2 are safe. */
3816
3817 static bool
aarch64_erratum_sequence(uint32_t insn_1,uint32_t insn_2)3818 aarch64_erratum_sequence (uint32_t insn_1, uint32_t insn_2)
3819 {
3820 uint32_t rt;
3821 uint32_t rt2;
3822 uint32_t rn;
3823 uint32_t rm;
3824 uint32_t ra;
3825 bool pair;
3826 bool load;
3827
3828 if (aarch64_mlxl_p (insn_2)
3829 && aarch64_mem_op_p (insn_1, &rt, &rt2, &pair, &load))
3830 {
3831 /* Any SIMD memory op is independent of the subsequent MLA
3832 by definition of the erratum. */
3833 if (AARCH64_BIT (insn_1, 26))
3834 return true;
3835
3836 /* If not SIMD, check for integer memory ops and MLA relationship. */
3837 rn = AARCH64_RN (insn_2);
3838 ra = AARCH64_RA (insn_2);
3839 rm = AARCH64_RM (insn_2);
3840
3841 /* If this is a load and there's a true(RAW) dependency, we are safe
3842 and this is not an erratum sequence. */
3843 if (load &&
3844 (rt == rn || rt == rm || rt == ra
3845 || (pair && (rt2 == rn || rt2 == rm || rt2 == ra))))
3846 return false;
3847
3848 /* We conservatively put out stubs for all other cases (including
3849 writebacks). */
3850 return true;
3851 }
3852
3853 return false;
3854 }
3855
3856 /* Used to order a list of mapping symbols by address. */
3857
3858 static int
elf_aarch64_compare_mapping(const void * a,const void * b)3859 elf_aarch64_compare_mapping (const void *a, const void *b)
3860 {
3861 const elf_aarch64_section_map *amap = (const elf_aarch64_section_map *) a;
3862 const elf_aarch64_section_map *bmap = (const elf_aarch64_section_map *) b;
3863
3864 if (amap->vma > bmap->vma)
3865 return 1;
3866 else if (amap->vma < bmap->vma)
3867 return -1;
3868 else if (amap->type > bmap->type)
3869 /* Ensure results do not depend on the host qsort for objects with
3870 multiple mapping symbols at the same address by sorting on type
3871 after vma. */
3872 return 1;
3873 else if (amap->type < bmap->type)
3874 return -1;
3875 else
3876 return 0;
3877 }
3878
3879
3880 static char *
_bfd_aarch64_erratum_835769_stub_name(unsigned num_fixes)3881 _bfd_aarch64_erratum_835769_stub_name (unsigned num_fixes)
3882 {
3883 char *stub_name = (char *) bfd_malloc
3884 (strlen ("__erratum_835769_veneer_") + 16);
3885 if (stub_name != NULL)
3886 sprintf (stub_name,"__erratum_835769_veneer_%d", num_fixes);
3887 return stub_name;
3888 }
3889
3890 /* Scan for Cortex-A53 erratum 835769 sequence.
3891
3892 Return TRUE else FALSE on abnormal termination. */
3893
3894 static bool
_bfd_aarch64_erratum_835769_scan(bfd * input_bfd,struct bfd_link_info * info,unsigned int * num_fixes_p)3895 _bfd_aarch64_erratum_835769_scan (bfd *input_bfd,
3896 struct bfd_link_info *info,
3897 unsigned int *num_fixes_p)
3898 {
3899 asection *section;
3900 struct elf_aarch64_link_hash_table *htab = elf_aarch64_hash_table (info);
3901 unsigned int num_fixes = *num_fixes_p;
3902
3903 if (htab == NULL)
3904 return true;
3905
3906 for (section = input_bfd->sections;
3907 section != NULL;
3908 section = section->next)
3909 {
3910 bfd_byte *contents = NULL;
3911 struct _aarch64_elf_section_data *sec_data;
3912 unsigned int span;
3913
3914 if (elf_section_type (section) != SHT_PROGBITS
3915 || (elf_section_flags (section) & SHF_EXECINSTR) == 0
3916 || (section->flags & SEC_EXCLUDE) != 0
3917 || (section->sec_info_type == SEC_INFO_TYPE_JUST_SYMS)
3918 || (section->output_section == bfd_abs_section_ptr))
3919 continue;
3920
3921 if (elf_section_data (section)->this_hdr.contents != NULL)
3922 contents = elf_section_data (section)->this_hdr.contents;
3923 else if (! bfd_malloc_and_get_section (input_bfd, section, &contents))
3924 return false;
3925
3926 sec_data = elf_aarch64_section_data (section);
3927
3928 if (sec_data->mapcount)
3929 qsort (sec_data->map, sec_data->mapcount,
3930 sizeof (elf_aarch64_section_map), elf_aarch64_compare_mapping);
3931
3932 for (span = 0; span < sec_data->mapcount; span++)
3933 {
3934 unsigned int span_start = sec_data->map[span].vma;
3935 unsigned int span_end = ((span == sec_data->mapcount - 1)
3936 ? sec_data->map[0].vma + section->size
3937 : sec_data->map[span + 1].vma);
3938 unsigned int i;
3939 char span_type = sec_data->map[span].type;
3940
3941 if (span_type == 'd')
3942 continue;
3943
3944 for (i = span_start; i + 4 < span_end; i += 4)
3945 {
3946 uint32_t insn_1 = bfd_getl32 (contents + i);
3947 uint32_t insn_2 = bfd_getl32 (contents + i + 4);
3948
3949 if (aarch64_erratum_sequence (insn_1, insn_2))
3950 {
3951 struct elf_aarch64_stub_hash_entry *stub_entry;
3952 char *stub_name = _bfd_aarch64_erratum_835769_stub_name (num_fixes);
3953 if (! stub_name)
3954 return false;
3955
3956 stub_entry = _bfd_aarch64_add_stub_entry_in_group (stub_name,
3957 section,
3958 htab);
3959 if (! stub_entry)
3960 return false;
3961
3962 stub_entry->stub_type = aarch64_stub_erratum_835769_veneer;
3963 stub_entry->target_section = section;
3964 stub_entry->target_value = i + 4;
3965 stub_entry->veneered_insn = insn_2;
3966 stub_entry->output_name = stub_name;
3967 num_fixes++;
3968 }
3969 }
3970 }
3971 if (elf_section_data (section)->this_hdr.contents == NULL)
3972 free (contents);
3973 }
3974
3975 *num_fixes_p = num_fixes;
3976
3977 return true;
3978 }
3979
3980
3981 /* Test if instruction INSN is ADRP. */
3982
3983 static bool
_bfd_aarch64_adrp_p(uint32_t insn)3984 _bfd_aarch64_adrp_p (uint32_t insn)
3985 {
3986 return ((insn & AARCH64_ADRP_OP_MASK) == AARCH64_ADRP_OP);
3987 }
3988
3989
3990 /* Helper predicate to look for cortex-a53 erratum 843419 sequence 1. */
3991
3992 static bool
_bfd_aarch64_erratum_843419_sequence_p(uint32_t insn_1,uint32_t insn_2,uint32_t insn_3)3993 _bfd_aarch64_erratum_843419_sequence_p (uint32_t insn_1, uint32_t insn_2,
3994 uint32_t insn_3)
3995 {
3996 uint32_t rt;
3997 uint32_t rt2;
3998 bool pair;
3999 bool load;
4000
4001 return (aarch64_mem_op_p (insn_2, &rt, &rt2, &pair, &load)
4002 && (!pair
4003 || (pair && !load))
4004 && AARCH64_LDST_UIMM (insn_3)
4005 && AARCH64_RN (insn_3) == AARCH64_RD (insn_1));
4006 }
4007
4008
4009 /* Test for the presence of Cortex-A53 erratum 843419 instruction sequence.
4010
4011 Return TRUE if section CONTENTS at offset I contains one of the
4012 erratum 843419 sequences, otherwise return FALSE. If a sequence is
4013 seen set P_VENEER_I to the offset of the final LOAD/STORE
4014 instruction in the sequence.
4015 */
4016
4017 static bool
_bfd_aarch64_erratum_843419_p(bfd_byte * contents,bfd_vma vma,bfd_vma i,bfd_vma span_end,bfd_vma * p_veneer_i)4018 _bfd_aarch64_erratum_843419_p (bfd_byte *contents, bfd_vma vma,
4019 bfd_vma i, bfd_vma span_end,
4020 bfd_vma *p_veneer_i)
4021 {
4022 uint32_t insn_1 = bfd_getl32 (contents + i);
4023
4024 if (!_bfd_aarch64_adrp_p (insn_1))
4025 return false;
4026
4027 if (span_end < i + 12)
4028 return false;
4029
4030 uint32_t insn_2 = bfd_getl32 (contents + i + 4);
4031 uint32_t insn_3 = bfd_getl32 (contents + i + 8);
4032
4033 if ((vma & 0xfff) != 0xff8 && (vma & 0xfff) != 0xffc)
4034 return false;
4035
4036 if (_bfd_aarch64_erratum_843419_sequence_p (insn_1, insn_2, insn_3))
4037 {
4038 *p_veneer_i = i + 8;
4039 return true;
4040 }
4041
4042 if (span_end < i + 16)
4043 return false;
4044
4045 uint32_t insn_4 = bfd_getl32 (contents + i + 12);
4046
4047 if (_bfd_aarch64_erratum_843419_sequence_p (insn_1, insn_2, insn_4))
4048 {
4049 *p_veneer_i = i + 12;
4050 return true;
4051 }
4052
4053 return false;
4054 }
4055
4056
4057 /* Resize all stub sections. */
4058
4059 static void
_bfd_aarch64_resize_stubs(struct elf_aarch64_link_hash_table * htab)4060 _bfd_aarch64_resize_stubs (struct elf_aarch64_link_hash_table *htab)
4061 {
4062 asection *section;
4063
4064 /* OK, we've added some stubs. Find out the new size of the
4065 stub sections. */
4066 for (section = htab->stub_bfd->sections;
4067 section != NULL; section = section->next)
4068 {
4069 /* Ignore non-stub sections. */
4070 if (!strstr (section->name, STUB_SUFFIX))
4071 continue;
4072 section->size = 0;
4073 }
4074
4075 bfd_hash_traverse (&htab->stub_hash_table, aarch64_size_one_stub, htab);
4076
4077 for (section = htab->stub_bfd->sections;
4078 section != NULL; section = section->next)
4079 {
4080 if (!strstr (section->name, STUB_SUFFIX))
4081 continue;
4082
4083 /* Add space for a branch. Add 8 bytes to keep section 8 byte aligned,
4084 as long branch stubs contain a 64-bit address. */
4085 if (section->size)
4086 section->size += 8;
4087
4088 /* Ensure all stub sections have a size which is a multiple of
4089 4096. This is important in order to ensure that the insertion
4090 of stub sections does not in itself move existing code around
4091 in such a way that new errata sequences are created. We only do this
4092 when the ADRP workaround is enabled. If only the ADR workaround is
4093 enabled then the stubs workaround won't ever be used. */
4094 if (htab->fix_erratum_843419 & ERRAT_ADRP)
4095 if (section->size)
4096 section->size = BFD_ALIGN (section->size, 0x1000);
4097 }
4098 }
4099
4100 /* Construct an erratum 843419 workaround stub name. */
4101
4102 static char *
_bfd_aarch64_erratum_843419_stub_name(asection * input_section,bfd_vma offset)4103 _bfd_aarch64_erratum_843419_stub_name (asection *input_section,
4104 bfd_vma offset)
4105 {
4106 const bfd_size_type len = 8 + 4 + 1 + 8 + 1 + 16 + 1;
4107 char *stub_name = bfd_malloc (len);
4108
4109 if (stub_name != NULL)
4110 snprintf (stub_name, len, "e843419@%04x_%08x_%" BFD_VMA_FMT "x",
4111 input_section->owner->id,
4112 input_section->id,
4113 offset);
4114 return stub_name;
4115 }
4116
4117 /* Build a stub_entry structure describing an 843419 fixup.
4118
4119 The stub_entry constructed is populated with the bit pattern INSN
4120 of the instruction located at OFFSET within input SECTION.
4121
4122 Returns TRUE on success. */
4123
4124 static bool
_bfd_aarch64_erratum_843419_fixup(uint32_t insn,bfd_vma adrp_offset,bfd_vma ldst_offset,asection * section,struct bfd_link_info * info)4125 _bfd_aarch64_erratum_843419_fixup (uint32_t insn,
4126 bfd_vma adrp_offset,
4127 bfd_vma ldst_offset,
4128 asection *section,
4129 struct bfd_link_info *info)
4130 {
4131 struct elf_aarch64_link_hash_table *htab = elf_aarch64_hash_table (info);
4132 char *stub_name;
4133 struct elf_aarch64_stub_hash_entry *stub_entry;
4134
4135 stub_name = _bfd_aarch64_erratum_843419_stub_name (section, ldst_offset);
4136 if (stub_name == NULL)
4137 return false;
4138 stub_entry = aarch64_stub_hash_lookup (&htab->stub_hash_table, stub_name,
4139 false, false);
4140 if (stub_entry)
4141 {
4142 free (stub_name);
4143 return true;
4144 }
4145
4146 /* We always place an 843419 workaround veneer in the stub section
4147 attached to the input section in which an erratum sequence has
4148 been found. This ensures that later in the link process (in
4149 elfNN_aarch64_write_section) when we copy the veneered
4150 instruction from the input section into the stub section the
4151 copied instruction will have had any relocations applied to it.
4152 If we placed workaround veneers in any other stub section then we
4153 could not assume that all relocations have been processed on the
4154 corresponding input section at the point we output the stub
4155 section. */
4156
4157 stub_entry = _bfd_aarch64_add_stub_entry_after (stub_name, section, htab);
4158 if (stub_entry == NULL)
4159 {
4160 free (stub_name);
4161 return false;
4162 }
4163
4164 stub_entry->adrp_offset = adrp_offset;
4165 stub_entry->target_value = ldst_offset;
4166 stub_entry->target_section = section;
4167 stub_entry->stub_type = aarch64_stub_erratum_843419_veneer;
4168 stub_entry->veneered_insn = insn;
4169 stub_entry->output_name = stub_name;
4170
4171 return true;
4172 }
4173
4174
4175 /* Scan an input section looking for the signature of erratum 843419.
4176
4177 Scans input SECTION in INPUT_BFD looking for erratum 843419
4178 signatures, for each signature found a stub_entry is created
4179 describing the location of the erratum for subsequent fixup.
4180
4181 Return TRUE on successful scan, FALSE on failure to scan.
4182 */
4183
4184 static bool
_bfd_aarch64_erratum_843419_scan(bfd * input_bfd,asection * section,struct bfd_link_info * info)4185 _bfd_aarch64_erratum_843419_scan (bfd *input_bfd, asection *section,
4186 struct bfd_link_info *info)
4187 {
4188 struct elf_aarch64_link_hash_table *htab = elf_aarch64_hash_table (info);
4189
4190 if (htab == NULL)
4191 return true;
4192
4193 if (elf_section_type (section) != SHT_PROGBITS
4194 || (elf_section_flags (section) & SHF_EXECINSTR) == 0
4195 || (section->flags & SEC_EXCLUDE) != 0
4196 || (section->sec_info_type == SEC_INFO_TYPE_JUST_SYMS)
4197 || (section->output_section == bfd_abs_section_ptr))
4198 return true;
4199
4200 do
4201 {
4202 bfd_byte *contents = NULL;
4203 struct _aarch64_elf_section_data *sec_data;
4204 unsigned int span;
4205
4206 if (elf_section_data (section)->this_hdr.contents != NULL)
4207 contents = elf_section_data (section)->this_hdr.contents;
4208 else if (! bfd_malloc_and_get_section (input_bfd, section, &contents))
4209 return false;
4210
4211 sec_data = elf_aarch64_section_data (section);
4212
4213 if (sec_data->mapcount)
4214 qsort (sec_data->map, sec_data->mapcount,
4215 sizeof (elf_aarch64_section_map), elf_aarch64_compare_mapping);
4216
4217 for (span = 0; span < sec_data->mapcount; span++)
4218 {
4219 unsigned int span_start = sec_data->map[span].vma;
4220 unsigned int span_end = ((span == sec_data->mapcount - 1)
4221 ? sec_data->map[0].vma + section->size
4222 : sec_data->map[span + 1].vma);
4223 unsigned int i;
4224 char span_type = sec_data->map[span].type;
4225
4226 if (span_type == 'd')
4227 continue;
4228
4229 for (i = span_start; i + 8 < span_end; i += 4)
4230 {
4231 bfd_vma vma = (section->output_section->vma
4232 + section->output_offset
4233 + i);
4234 bfd_vma veneer_i;
4235
4236 if (_bfd_aarch64_erratum_843419_p
4237 (contents, vma, i, span_end, &veneer_i))
4238 {
4239 uint32_t insn = bfd_getl32 (contents + veneer_i);
4240
4241 if (!_bfd_aarch64_erratum_843419_fixup (insn, i, veneer_i,
4242 section, info))
4243 return false;
4244 }
4245 }
4246 }
4247
4248 if (elf_section_data (section)->this_hdr.contents == NULL)
4249 free (contents);
4250 }
4251 while (0);
4252
4253 return true;
4254 }
4255
4256
4257 /* Determine and set the size of the stub section for a final link.
4258
4259 The basic idea here is to examine all the relocations looking for
4260 PC-relative calls to a target that is unreachable with a "bl"
4261 instruction. */
4262
4263 bool
elfNN_aarch64_size_stubs(bfd * output_bfd,bfd * stub_bfd,struct bfd_link_info * info,bfd_signed_vma group_size,asection * (* add_stub_section)(const char *,asection *),void (* layout_sections_again)(void))4264 elfNN_aarch64_size_stubs (bfd *output_bfd,
4265 bfd *stub_bfd,
4266 struct bfd_link_info *info,
4267 bfd_signed_vma group_size,
4268 asection * (*add_stub_section) (const char *,
4269 asection *),
4270 void (*layout_sections_again) (void))
4271 {
4272 bfd_size_type stub_group_size;
4273 bool stubs_always_before_branch;
4274 bool stub_changed = false;
4275 struct elf_aarch64_link_hash_table *htab = elf_aarch64_hash_table (info);
4276 unsigned int num_erratum_835769_fixes = 0;
4277
4278 /* Propagate mach to stub bfd, because it may not have been
4279 finalized when we created stub_bfd. */
4280 bfd_set_arch_mach (stub_bfd, bfd_get_arch (output_bfd),
4281 bfd_get_mach (output_bfd));
4282
4283 /* Stash our params away. */
4284 htab->stub_bfd = stub_bfd;
4285 htab->add_stub_section = add_stub_section;
4286 htab->layout_sections_again = layout_sections_again;
4287 stubs_always_before_branch = group_size < 0;
4288 if (group_size < 0)
4289 stub_group_size = -group_size;
4290 else
4291 stub_group_size = group_size;
4292
4293 if (stub_group_size == 1)
4294 {
4295 /* Default values. */
4296 /* AArch64 branch range is +-128MB. The value used is 1MB less. */
4297 stub_group_size = 127 * 1024 * 1024;
4298 }
4299
4300 group_sections (htab, stub_group_size, stubs_always_before_branch);
4301
4302 (*htab->layout_sections_again) ();
4303
4304 if (htab->fix_erratum_835769)
4305 {
4306 bfd *input_bfd;
4307
4308 for (input_bfd = info->input_bfds;
4309 input_bfd != NULL; input_bfd = input_bfd->link.next)
4310 {
4311 if (!is_aarch64_elf (input_bfd)
4312 || (input_bfd->flags & BFD_LINKER_CREATED) != 0)
4313 continue;
4314
4315 if (!_bfd_aarch64_erratum_835769_scan (input_bfd, info,
4316 &num_erratum_835769_fixes))
4317 return false;
4318 }
4319
4320 _bfd_aarch64_resize_stubs (htab);
4321 (*htab->layout_sections_again) ();
4322 }
4323
4324 if (htab->fix_erratum_843419 != ERRAT_NONE)
4325 {
4326 bfd *input_bfd;
4327
4328 for (input_bfd = info->input_bfds;
4329 input_bfd != NULL;
4330 input_bfd = input_bfd->link.next)
4331 {
4332 asection *section;
4333
4334 if (!is_aarch64_elf (input_bfd)
4335 || (input_bfd->flags & BFD_LINKER_CREATED) != 0)
4336 continue;
4337
4338 for (section = input_bfd->sections;
4339 section != NULL;
4340 section = section->next)
4341 if (!_bfd_aarch64_erratum_843419_scan (input_bfd, section, info))
4342 return false;
4343 }
4344
4345 _bfd_aarch64_resize_stubs (htab);
4346 (*htab->layout_sections_again) ();
4347 }
4348
4349 while (1)
4350 {
4351 bfd *input_bfd;
4352
4353 for (input_bfd = info->input_bfds;
4354 input_bfd != NULL; input_bfd = input_bfd->link.next)
4355 {
4356 Elf_Internal_Shdr *symtab_hdr;
4357 asection *section;
4358 Elf_Internal_Sym *local_syms = NULL;
4359
4360 if (!is_aarch64_elf (input_bfd)
4361 || (input_bfd->flags & BFD_LINKER_CREATED) != 0)
4362 continue;
4363
4364 /* We'll need the symbol table in a second. */
4365 symtab_hdr = &elf_tdata (input_bfd)->symtab_hdr;
4366 if (symtab_hdr->sh_info == 0)
4367 continue;
4368
4369 /* Walk over each section attached to the input bfd. */
4370 for (section = input_bfd->sections;
4371 section != NULL; section = section->next)
4372 {
4373 Elf_Internal_Rela *internal_relocs, *irelaend, *irela;
4374
4375 /* If there aren't any relocs, then there's nothing more
4376 to do. */
4377 if ((section->flags & SEC_RELOC) == 0
4378 || section->reloc_count == 0
4379 || (section->flags & SEC_CODE) == 0)
4380 continue;
4381
4382 /* If this section is a link-once section that will be
4383 discarded, then don't create any stubs. */
4384 if (section->output_section == NULL
4385 || section->output_section->owner != output_bfd)
4386 continue;
4387
4388 /* Get the relocs. */
4389 internal_relocs
4390 = _bfd_elf_link_read_relocs (input_bfd, section, NULL,
4391 NULL, info->keep_memory);
4392 if (internal_relocs == NULL)
4393 goto error_ret_free_local;
4394
4395 /* Now examine each relocation. */
4396 irela = internal_relocs;
4397 irelaend = irela + section->reloc_count;
4398 for (; irela < irelaend; irela++)
4399 {
4400 unsigned int r_type, r_indx;
4401 enum elf_aarch64_stub_type stub_type;
4402 struct elf_aarch64_stub_hash_entry *stub_entry;
4403 asection *sym_sec;
4404 bfd_vma sym_value;
4405 bfd_vma destination;
4406 struct elf_aarch64_link_hash_entry *hash;
4407 const char *sym_name;
4408 char *stub_name;
4409 const asection *id_sec;
4410 unsigned char st_type;
4411 bfd_size_type len;
4412
4413 r_type = ELFNN_R_TYPE (irela->r_info);
4414 r_indx = ELFNN_R_SYM (irela->r_info);
4415
4416 if (r_type >= (unsigned int) R_AARCH64_end)
4417 {
4418 bfd_set_error (bfd_error_bad_value);
4419 error_ret_free_internal:
4420 if (elf_section_data (section)->relocs == NULL)
4421 free (internal_relocs);
4422 goto error_ret_free_local;
4423 }
4424
4425 /* Only look for stubs on unconditional branch and
4426 branch and link instructions. */
4427 if (r_type != (unsigned int) AARCH64_R (CALL26)
4428 && r_type != (unsigned int) AARCH64_R (JUMP26))
4429 continue;
4430
4431 /* Now determine the call target, its name, value,
4432 section. */
4433 sym_sec = NULL;
4434 sym_value = 0;
4435 destination = 0;
4436 hash = NULL;
4437 sym_name = NULL;
4438 if (r_indx < symtab_hdr->sh_info)
4439 {
4440 /* It's a local symbol. */
4441 Elf_Internal_Sym *sym;
4442 Elf_Internal_Shdr *hdr;
4443
4444 if (local_syms == NULL)
4445 {
4446 local_syms
4447 = (Elf_Internal_Sym *) symtab_hdr->contents;
4448 if (local_syms == NULL)
4449 local_syms
4450 = bfd_elf_get_elf_syms (input_bfd, symtab_hdr,
4451 symtab_hdr->sh_info, 0,
4452 NULL, NULL, NULL);
4453 if (local_syms == NULL)
4454 goto error_ret_free_internal;
4455 }
4456
4457 sym = local_syms + r_indx;
4458 hdr = elf_elfsections (input_bfd)[sym->st_shndx];
4459 sym_sec = hdr->bfd_section;
4460 if (!sym_sec)
4461 /* This is an undefined symbol. It can never
4462 be resolved. */
4463 continue;
4464
4465 if (ELF_ST_TYPE (sym->st_info) != STT_SECTION)
4466 sym_value = sym->st_value;
4467 destination = (sym_value + irela->r_addend
4468 + sym_sec->output_offset
4469 + sym_sec->output_section->vma);
4470 st_type = ELF_ST_TYPE (sym->st_info);
4471 sym_name
4472 = bfd_elf_string_from_elf_section (input_bfd,
4473 symtab_hdr->sh_link,
4474 sym->st_name);
4475 }
4476 else
4477 {
4478 int e_indx;
4479
4480 e_indx = r_indx - symtab_hdr->sh_info;
4481 hash = ((struct elf_aarch64_link_hash_entry *)
4482 elf_sym_hashes (input_bfd)[e_indx]);
4483
4484 while (hash->root.root.type == bfd_link_hash_indirect
4485 || hash->root.root.type == bfd_link_hash_warning)
4486 hash = ((struct elf_aarch64_link_hash_entry *)
4487 hash->root.root.u.i.link);
4488
4489 if (hash->root.root.type == bfd_link_hash_defined
4490 || hash->root.root.type == bfd_link_hash_defweak)
4491 {
4492 struct elf_aarch64_link_hash_table *globals =
4493 elf_aarch64_hash_table (info);
4494 sym_sec = hash->root.root.u.def.section;
4495 sym_value = hash->root.root.u.def.value;
4496 /* For a destination in a shared library,
4497 use the PLT stub as target address to
4498 decide whether a branch stub is
4499 needed. */
4500 if (globals->root.splt != NULL && hash != NULL
4501 && hash->root.plt.offset != (bfd_vma) - 1)
4502 {
4503 sym_sec = globals->root.splt;
4504 sym_value = hash->root.plt.offset;
4505 if (sym_sec->output_section != NULL)
4506 destination = (sym_value
4507 + sym_sec->output_offset
4508 +
4509 sym_sec->output_section->vma);
4510 }
4511 else if (sym_sec->output_section != NULL)
4512 destination = (sym_value + irela->r_addend
4513 + sym_sec->output_offset
4514 + sym_sec->output_section->vma);
4515 }
4516 else if (hash->root.root.type == bfd_link_hash_undefined
4517 || (hash->root.root.type
4518 == bfd_link_hash_undefweak))
4519 {
4520 /* For a shared library, use the PLT stub as
4521 target address to decide whether a long
4522 branch stub is needed.
4523 For absolute code, they cannot be handled. */
4524 struct elf_aarch64_link_hash_table *globals =
4525 elf_aarch64_hash_table (info);
4526
4527 if (globals->root.splt != NULL && hash != NULL
4528 && hash->root.plt.offset != (bfd_vma) - 1)
4529 {
4530 sym_sec = globals->root.splt;
4531 sym_value = hash->root.plt.offset;
4532 if (sym_sec->output_section != NULL)
4533 destination = (sym_value
4534 + sym_sec->output_offset
4535 +
4536 sym_sec->output_section->vma);
4537 }
4538 else
4539 continue;
4540 }
4541 else
4542 {
4543 bfd_set_error (bfd_error_bad_value);
4544 goto error_ret_free_internal;
4545 }
4546 st_type = ELF_ST_TYPE (hash->root.type);
4547 sym_name = hash->root.root.root.string;
4548 }
4549
4550 /* Determine what (if any) linker stub is needed. */
4551 stub_type = aarch64_type_of_stub (section, irela, sym_sec,
4552 st_type, destination);
4553 if (stub_type == aarch64_stub_none)
4554 continue;
4555
4556 /* Support for grouping stub sections. */
4557 id_sec = htab->stub_group[section->id].link_sec;
4558
4559 /* Get the name of this stub. */
4560 stub_name = elfNN_aarch64_stub_name (id_sec, sym_sec, hash,
4561 irela);
4562 if (!stub_name)
4563 goto error_ret_free_internal;
4564
4565 stub_entry =
4566 aarch64_stub_hash_lookup (&htab->stub_hash_table,
4567 stub_name, false, false);
4568 if (stub_entry != NULL)
4569 {
4570 /* The proper stub has already been created. */
4571 free (stub_name);
4572 /* Always update this stub's target since it may have
4573 changed after layout. */
4574 stub_entry->target_value = sym_value + irela->r_addend;
4575 continue;
4576 }
4577
4578 stub_entry = _bfd_aarch64_add_stub_entry_in_group
4579 (stub_name, section, htab);
4580 if (stub_entry == NULL)
4581 {
4582 free (stub_name);
4583 goto error_ret_free_internal;
4584 }
4585
4586 stub_entry->target_value = sym_value + irela->r_addend;
4587 stub_entry->target_section = sym_sec;
4588 stub_entry->stub_type = stub_type;
4589 stub_entry->h = hash;
4590 stub_entry->st_type = st_type;
4591
4592 if (sym_name == NULL)
4593 sym_name = "unnamed";
4594 len = sizeof (STUB_ENTRY_NAME) + strlen (sym_name);
4595 stub_entry->output_name = bfd_alloc (htab->stub_bfd, len);
4596 if (stub_entry->output_name == NULL)
4597 {
4598 free (stub_name);
4599 goto error_ret_free_internal;
4600 }
4601
4602 snprintf (stub_entry->output_name, len, STUB_ENTRY_NAME,
4603 sym_name);
4604
4605 stub_changed = true;
4606 }
4607
4608 /* We're done with the internal relocs, free them. */
4609 if (elf_section_data (section)->relocs == NULL)
4610 free (internal_relocs);
4611 }
4612 }
4613
4614 if (!stub_changed)
4615 break;
4616
4617 _bfd_aarch64_resize_stubs (htab);
4618
4619 /* Ask the linker to do its stuff. */
4620 (*htab->layout_sections_again) ();
4621 stub_changed = false;
4622 }
4623
4624 return true;
4625
4626 error_ret_free_local:
4627 return false;
4628 }
4629
4630 /* Build all the stubs associated with the current output file. The
4631 stubs are kept in a hash table attached to the main linker hash
4632 table. We also set up the .plt entries for statically linked PIC
4633 functions here. This function is called via aarch64_elf_finish in the
4634 linker. */
4635
4636 bool
elfNN_aarch64_build_stubs(struct bfd_link_info * info)4637 elfNN_aarch64_build_stubs (struct bfd_link_info *info)
4638 {
4639 asection *stub_sec;
4640 struct bfd_hash_table *table;
4641 struct elf_aarch64_link_hash_table *htab;
4642
4643 htab = elf_aarch64_hash_table (info);
4644
4645 for (stub_sec = htab->stub_bfd->sections;
4646 stub_sec != NULL; stub_sec = stub_sec->next)
4647 {
4648 bfd_size_type size;
4649
4650 /* Ignore non-stub sections. */
4651 if (!strstr (stub_sec->name, STUB_SUFFIX))
4652 continue;
4653
4654 /* Allocate memory to hold the linker stubs. */
4655 size = stub_sec->size;
4656 stub_sec->contents = bfd_zalloc (htab->stub_bfd, size);
4657 if (stub_sec->contents == NULL && size != 0)
4658 return false;
4659 stub_sec->size = 0;
4660
4661 /* Add a branch around the stub section, and a nop, to keep it 8 byte
4662 aligned, as long branch stubs contain a 64-bit address. */
4663 bfd_putl32 (0x14000000 | (size >> 2), stub_sec->contents);
4664 bfd_putl32 (INSN_NOP, stub_sec->contents + 4);
4665 stub_sec->size += 8;
4666 }
4667
4668 /* Build the stubs as directed by the stub hash table. */
4669 table = &htab->stub_hash_table;
4670 bfd_hash_traverse (table, aarch64_build_one_stub, info);
4671
4672 return true;
4673 }
4674
4675
4676 /* Add an entry to the code/data map for section SEC. */
4677
4678 static void
elfNN_aarch64_section_map_add(asection * sec,char type,bfd_vma vma)4679 elfNN_aarch64_section_map_add (asection *sec, char type, bfd_vma vma)
4680 {
4681 struct _aarch64_elf_section_data *sec_data =
4682 elf_aarch64_section_data (sec);
4683 unsigned int newidx;
4684
4685 if (sec_data->map == NULL)
4686 {
4687 sec_data->map = bfd_malloc (sizeof (elf_aarch64_section_map));
4688 sec_data->mapcount = 0;
4689 sec_data->mapsize = 1;
4690 }
4691
4692 newidx = sec_data->mapcount++;
4693
4694 if (sec_data->mapcount > sec_data->mapsize)
4695 {
4696 sec_data->mapsize *= 2;
4697 sec_data->map = bfd_realloc_or_free
4698 (sec_data->map, sec_data->mapsize * sizeof (elf_aarch64_section_map));
4699 }
4700
4701 if (sec_data->map)
4702 {
4703 sec_data->map[newidx].vma = vma;
4704 sec_data->map[newidx].type = type;
4705 }
4706 }
4707
4708
4709 /* Initialise maps of insn/data for input BFDs. */
4710 void
bfd_elfNN_aarch64_init_maps(bfd * abfd)4711 bfd_elfNN_aarch64_init_maps (bfd *abfd)
4712 {
4713 Elf_Internal_Sym *isymbuf;
4714 Elf_Internal_Shdr *hdr;
4715 unsigned int i, localsyms;
4716
4717 /* Make sure that we are dealing with an AArch64 elf binary. */
4718 if (!is_aarch64_elf (abfd))
4719 return;
4720
4721 if ((abfd->flags & DYNAMIC) != 0)
4722 return;
4723
4724 hdr = &elf_symtab_hdr (abfd);
4725 localsyms = hdr->sh_info;
4726
4727 /* Obtain a buffer full of symbols for this BFD. The hdr->sh_info field
4728 should contain the number of local symbols, which should come before any
4729 global symbols. Mapping symbols are always local. */
4730 isymbuf = bfd_elf_get_elf_syms (abfd, hdr, localsyms, 0, NULL, NULL, NULL);
4731
4732 /* No internal symbols read? Skip this BFD. */
4733 if (isymbuf == NULL)
4734 return;
4735
4736 for (i = 0; i < localsyms; i++)
4737 {
4738 Elf_Internal_Sym *isym = &isymbuf[i];
4739 asection *sec = bfd_section_from_elf_index (abfd, isym->st_shndx);
4740 const char *name;
4741
4742 if (sec != NULL && ELF_ST_BIND (isym->st_info) == STB_LOCAL)
4743 {
4744 name = bfd_elf_string_from_elf_section (abfd,
4745 hdr->sh_link,
4746 isym->st_name);
4747
4748 if (bfd_is_aarch64_special_symbol_name
4749 (name, BFD_AARCH64_SPECIAL_SYM_TYPE_MAP))
4750 elfNN_aarch64_section_map_add (sec, name[1], isym->st_value);
4751 }
4752 }
4753 }
4754
4755 static void
setup_plt_values(struct bfd_link_info * link_info,aarch64_plt_type plt_type)4756 setup_plt_values (struct bfd_link_info *link_info,
4757 aarch64_plt_type plt_type)
4758 {
4759 struct elf_aarch64_link_hash_table *globals;
4760 globals = elf_aarch64_hash_table (link_info);
4761
4762 if (plt_type == PLT_BTI_PAC)
4763 {
4764 globals->plt0_entry = elfNN_aarch64_small_plt0_bti_entry;
4765
4766 /* Only in ET_EXEC we need PLTn with BTI. */
4767 if (bfd_link_pde (link_info))
4768 {
4769 globals->plt_entry_size = PLT_BTI_PAC_SMALL_ENTRY_SIZE;
4770 globals->plt_entry = elfNN_aarch64_small_plt_bti_pac_entry;
4771 }
4772 else
4773 {
4774 globals->plt_entry_size = PLT_PAC_SMALL_ENTRY_SIZE;
4775 globals->plt_entry = elfNN_aarch64_small_plt_pac_entry;
4776 }
4777 }
4778 else if (plt_type == PLT_BTI)
4779 {
4780 globals->plt0_entry = elfNN_aarch64_small_plt0_bti_entry;
4781
4782 /* Only in ET_EXEC we need PLTn with BTI. */
4783 if (bfd_link_pde (link_info))
4784 {
4785 globals->plt_entry_size = PLT_BTI_SMALL_ENTRY_SIZE;
4786 globals->plt_entry = elfNN_aarch64_small_plt_bti_entry;
4787 }
4788 }
4789 else if (plt_type == PLT_PAC)
4790 {
4791 globals->plt_entry_size = PLT_PAC_SMALL_ENTRY_SIZE;
4792 globals->plt_entry = elfNN_aarch64_small_plt_pac_entry;
4793 }
4794 }
4795
4796 /* Set option values needed during linking. */
4797 void
bfd_elfNN_aarch64_set_options(struct bfd * output_bfd,struct bfd_link_info * link_info,int no_enum_warn,int no_wchar_warn,int pic_veneer,int fix_erratum_835769,erratum_84319_opts fix_erratum_843419,int no_apply_dynamic_relocs,aarch64_bti_pac_info bp_info)4798 bfd_elfNN_aarch64_set_options (struct bfd *output_bfd,
4799 struct bfd_link_info *link_info,
4800 int no_enum_warn,
4801 int no_wchar_warn, int pic_veneer,
4802 int fix_erratum_835769,
4803 erratum_84319_opts fix_erratum_843419,
4804 int no_apply_dynamic_relocs,
4805 aarch64_bti_pac_info bp_info)
4806 {
4807 struct elf_aarch64_link_hash_table *globals;
4808
4809 globals = elf_aarch64_hash_table (link_info);
4810 globals->pic_veneer = pic_veneer;
4811 globals->fix_erratum_835769 = fix_erratum_835769;
4812 /* If the default options are used, then ERRAT_ADR will be set by default
4813 which will enable the ADRP->ADR workaround for the erratum 843419
4814 workaround. */
4815 globals->fix_erratum_843419 = fix_erratum_843419;
4816 globals->no_apply_dynamic_relocs = no_apply_dynamic_relocs;
4817
4818 BFD_ASSERT (is_aarch64_elf (output_bfd));
4819 elf_aarch64_tdata (output_bfd)->no_enum_size_warning = no_enum_warn;
4820 elf_aarch64_tdata (output_bfd)->no_wchar_size_warning = no_wchar_warn;
4821
4822 switch (bp_info.bti_type)
4823 {
4824 case BTI_WARN:
4825 elf_aarch64_tdata (output_bfd)->no_bti_warn = 0;
4826 elf_aarch64_tdata (output_bfd)->gnu_and_prop
4827 |= GNU_PROPERTY_AARCH64_FEATURE_1_BTI;
4828 break;
4829
4830 default:
4831 break;
4832 }
4833 elf_aarch64_tdata (output_bfd)->plt_type = bp_info.plt_type;
4834 setup_plt_values (link_info, bp_info.plt_type);
4835 }
4836
4837 static bfd_vma
aarch64_calculate_got_entry_vma(struct elf_link_hash_entry * h,struct elf_aarch64_link_hash_table * globals,struct bfd_link_info * info,bfd_vma value,bfd * output_bfd,bool * unresolved_reloc_p)4838 aarch64_calculate_got_entry_vma (struct elf_link_hash_entry *h,
4839 struct elf_aarch64_link_hash_table
4840 *globals, struct bfd_link_info *info,
4841 bfd_vma value, bfd *output_bfd,
4842 bool *unresolved_reloc_p)
4843 {
4844 bfd_vma off = (bfd_vma) - 1;
4845 asection *basegot = globals->root.sgot;
4846 bool dyn = globals->root.dynamic_sections_created;
4847
4848 if (h != NULL)
4849 {
4850 BFD_ASSERT (basegot != NULL);
4851 off = h->got.offset;
4852 BFD_ASSERT (off != (bfd_vma) - 1);
4853 if (!WILL_CALL_FINISH_DYNAMIC_SYMBOL (dyn, bfd_link_pic (info), h)
4854 || ((bfd_link_pic (info)
4855 || (!bfd_link_relocatable (info)
4856 && (info->export_dynamic || info->dynamic)))
4857 && SYMBOL_REFERENCES_LOCAL (info, h))
4858 || (ELF_ST_VISIBILITY (h->other)
4859 && h->root.type == bfd_link_hash_undefweak))
4860 {
4861 /* This is actually a static link, or it is a -Bsymbolic link
4862 and the symbol is defined locally. We must initialize this
4863 entry in the global offset table. Since the offset must
4864 always be a multiple of 8 (4 in the case of ILP32), we use
4865 the least significant bit to record whether we have
4866 initialized it already.
4867 When doing a dynamic link, we create a .rel(a).got relocation
4868 entry to initialize the value. This is done in the
4869 finish_dynamic_symbol routine. */
4870 if ((off & 1) != 0)
4871 off &= ~1;
4872 else
4873 {
4874 bfd_put_NN (output_bfd, value, basegot->contents + off);
4875 h->got.offset |= 1;
4876 }
4877 }
4878 else
4879 *unresolved_reloc_p = false;
4880
4881 off = off + basegot->output_section->vma + basegot->output_offset;
4882 }
4883
4884 return off;
4885 }
4886
4887 /* Change R_TYPE to a more efficient access model where possible,
4888 return the new reloc type. */
4889
4890 static bfd_reloc_code_real_type
aarch64_tls_transition_without_check(bfd_reloc_code_real_type r_type,struct elf_link_hash_entry * h)4891 aarch64_tls_transition_without_check (bfd_reloc_code_real_type r_type,
4892 struct elf_link_hash_entry *h)
4893 {
4894 bool is_local = h == NULL;
4895
4896 switch (r_type)
4897 {
4898 case BFD_RELOC_AARCH64_TLSDESC_ADR_PAGE21:
4899 case BFD_RELOC_AARCH64_TLSGD_ADR_PAGE21:
4900 return (is_local
4901 ? BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G1
4902 : BFD_RELOC_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21);
4903
4904 case BFD_RELOC_AARCH64_TLSDESC_ADR_PREL21:
4905 return (is_local
4906 ? BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G0_NC
4907 : r_type);
4908
4909 case BFD_RELOC_AARCH64_TLSDESC_LD_PREL19:
4910 return (is_local
4911 ? BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G1
4912 : BFD_RELOC_AARCH64_TLSIE_LD_GOTTPREL_PREL19);
4913
4914 case BFD_RELOC_AARCH64_TLSDESC_LDR:
4915 return (is_local
4916 ? BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G0_NC
4917 : BFD_RELOC_AARCH64_NONE);
4918
4919 case BFD_RELOC_AARCH64_TLSDESC_OFF_G0_NC:
4920 return (is_local
4921 ? BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G1_NC
4922 : BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G0_NC);
4923
4924 case BFD_RELOC_AARCH64_TLSDESC_OFF_G1:
4925 return (is_local
4926 ? BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G2
4927 : BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G1);
4928
4929 case BFD_RELOC_AARCH64_TLSDESC_LDNN_LO12_NC:
4930 case BFD_RELOC_AARCH64_TLSGD_ADD_LO12_NC:
4931 return (is_local
4932 ? BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G0_NC
4933 : BFD_RELOC_AARCH64_TLSIE_LDNN_GOTTPREL_LO12_NC);
4934
4935 case BFD_RELOC_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21:
4936 return is_local ? BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G1 : r_type;
4937
4938 case BFD_RELOC_AARCH64_TLSIE_LDNN_GOTTPREL_LO12_NC:
4939 return is_local ? BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G0_NC : r_type;
4940
4941 case BFD_RELOC_AARCH64_TLSIE_LD_GOTTPREL_PREL19:
4942 return r_type;
4943
4944 case BFD_RELOC_AARCH64_TLSGD_ADR_PREL21:
4945 return (is_local
4946 ? BFD_RELOC_AARCH64_TLSLE_ADD_TPREL_HI12
4947 : BFD_RELOC_AARCH64_TLSIE_LD_GOTTPREL_PREL19);
4948
4949 case BFD_RELOC_AARCH64_TLSDESC_ADD:
4950 case BFD_RELOC_AARCH64_TLSDESC_ADD_LO12:
4951 case BFD_RELOC_AARCH64_TLSDESC_CALL:
4952 /* Instructions with these relocations will become NOPs. */
4953 return BFD_RELOC_AARCH64_NONE;
4954
4955 case BFD_RELOC_AARCH64_TLSLD_ADD_LO12_NC:
4956 case BFD_RELOC_AARCH64_TLSLD_ADR_PAGE21:
4957 case BFD_RELOC_AARCH64_TLSLD_ADR_PREL21:
4958 return is_local ? BFD_RELOC_AARCH64_NONE : r_type;
4959
4960 #if ARCH_SIZE == 64
4961 case BFD_RELOC_AARCH64_TLSGD_MOVW_G0_NC:
4962 return is_local
4963 ? BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G1_NC
4964 : BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G0_NC;
4965
4966 case BFD_RELOC_AARCH64_TLSGD_MOVW_G1:
4967 return is_local
4968 ? BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G2
4969 : BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G1;
4970 #endif
4971
4972 default:
4973 break;
4974 }
4975
4976 return r_type;
4977 }
4978
4979 static unsigned int
aarch64_reloc_got_type(bfd_reloc_code_real_type r_type)4980 aarch64_reloc_got_type (bfd_reloc_code_real_type r_type)
4981 {
4982 switch (r_type)
4983 {
4984 case BFD_RELOC_AARCH64_ADR_GOT_PAGE:
4985 case BFD_RELOC_AARCH64_GOT_LD_PREL19:
4986 case BFD_RELOC_AARCH64_LD32_GOTPAGE_LO14:
4987 case BFD_RELOC_AARCH64_LD32_GOT_LO12_NC:
4988 case BFD_RELOC_AARCH64_LD64_GOTOFF_LO15:
4989 case BFD_RELOC_AARCH64_LD64_GOTPAGE_LO15:
4990 case BFD_RELOC_AARCH64_LD64_GOT_LO12_NC:
4991 case BFD_RELOC_AARCH64_MOVW_GOTOFF_G0_NC:
4992 case BFD_RELOC_AARCH64_MOVW_GOTOFF_G1:
4993 return GOT_NORMAL;
4994
4995 case BFD_RELOC_AARCH64_TLSGD_ADD_LO12_NC:
4996 case BFD_RELOC_AARCH64_TLSGD_ADR_PAGE21:
4997 case BFD_RELOC_AARCH64_TLSGD_ADR_PREL21:
4998 case BFD_RELOC_AARCH64_TLSGD_MOVW_G0_NC:
4999 case BFD_RELOC_AARCH64_TLSGD_MOVW_G1:
5000 case BFD_RELOC_AARCH64_TLSLD_ADD_LO12_NC:
5001 case BFD_RELOC_AARCH64_TLSLD_ADR_PAGE21:
5002 case BFD_RELOC_AARCH64_TLSLD_ADR_PREL21:
5003 return GOT_TLS_GD;
5004
5005 case BFD_RELOC_AARCH64_TLSDESC_ADD:
5006 case BFD_RELOC_AARCH64_TLSDESC_ADD_LO12:
5007 case BFD_RELOC_AARCH64_TLSDESC_ADR_PAGE21:
5008 case BFD_RELOC_AARCH64_TLSDESC_ADR_PREL21:
5009 case BFD_RELOC_AARCH64_TLSDESC_CALL:
5010 case BFD_RELOC_AARCH64_TLSDESC_LD32_LO12_NC:
5011 case BFD_RELOC_AARCH64_TLSDESC_LD64_LO12:
5012 case BFD_RELOC_AARCH64_TLSDESC_LD_PREL19:
5013 case BFD_RELOC_AARCH64_TLSDESC_LDR:
5014 case BFD_RELOC_AARCH64_TLSDESC_OFF_G0_NC:
5015 case BFD_RELOC_AARCH64_TLSDESC_OFF_G1:
5016 return GOT_TLSDESC_GD;
5017
5018 case BFD_RELOC_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21:
5019 case BFD_RELOC_AARCH64_TLSIE_LD32_GOTTPREL_LO12_NC:
5020 case BFD_RELOC_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC:
5021 case BFD_RELOC_AARCH64_TLSIE_LD_GOTTPREL_PREL19:
5022 case BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G0_NC:
5023 case BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G1:
5024 return GOT_TLS_IE;
5025
5026 default:
5027 break;
5028 }
5029 return GOT_UNKNOWN;
5030 }
5031
5032 static bool
aarch64_can_relax_tls(bfd * input_bfd,struct bfd_link_info * info,bfd_reloc_code_real_type r_type,struct elf_link_hash_entry * h,unsigned long r_symndx)5033 aarch64_can_relax_tls (bfd *input_bfd,
5034 struct bfd_link_info *info,
5035 bfd_reloc_code_real_type r_type,
5036 struct elf_link_hash_entry *h,
5037 unsigned long r_symndx)
5038 {
5039 unsigned int symbol_got_type;
5040 unsigned int reloc_got_type;
5041
5042 if (! IS_AARCH64_TLS_RELAX_RELOC (r_type))
5043 return false;
5044
5045 symbol_got_type = elfNN_aarch64_symbol_got_type (h, input_bfd, r_symndx);
5046 reloc_got_type = aarch64_reloc_got_type (r_type);
5047
5048 if (symbol_got_type == GOT_TLS_IE && GOT_TLS_GD_ANY_P (reloc_got_type))
5049 return true;
5050
5051 if (!bfd_link_executable (info))
5052 return false;
5053
5054 if (h && h->root.type == bfd_link_hash_undefweak)
5055 return false;
5056
5057 return true;
5058 }
5059
5060 /* Given the relocation code R_TYPE, return the relaxed bfd reloc
5061 enumerator. */
5062
5063 static bfd_reloc_code_real_type
aarch64_tls_transition(bfd * input_bfd,struct bfd_link_info * info,unsigned int r_type,struct elf_link_hash_entry * h,unsigned long r_symndx)5064 aarch64_tls_transition (bfd *input_bfd,
5065 struct bfd_link_info *info,
5066 unsigned int r_type,
5067 struct elf_link_hash_entry *h,
5068 unsigned long r_symndx)
5069 {
5070 bfd_reloc_code_real_type bfd_r_type
5071 = elfNN_aarch64_bfd_reloc_from_type (input_bfd, r_type);
5072
5073 if (! aarch64_can_relax_tls (input_bfd, info, bfd_r_type, h, r_symndx))
5074 return bfd_r_type;
5075
5076 return aarch64_tls_transition_without_check (bfd_r_type, h);
5077 }
5078
5079 /* Return the base VMA address which should be subtracted from real addresses
5080 when resolving R_AARCH64_TLS_DTPREL relocation. */
5081
5082 static bfd_vma
dtpoff_base(struct bfd_link_info * info)5083 dtpoff_base (struct bfd_link_info *info)
5084 {
5085 /* If tls_sec is NULL, we should have signalled an error already. */
5086 BFD_ASSERT (elf_hash_table (info)->tls_sec != NULL);
5087 return elf_hash_table (info)->tls_sec->vma;
5088 }
5089
5090 /* Return the base VMA address which should be subtracted from real addresses
5091 when resolving R_AARCH64_TLS_GOTTPREL64 relocations. */
5092
5093 static bfd_vma
tpoff_base(struct bfd_link_info * info)5094 tpoff_base (struct bfd_link_info *info)
5095 {
5096 struct elf_link_hash_table *htab = elf_hash_table (info);
5097
5098 /* If tls_sec is NULL, we should have signalled an error already. */
5099 BFD_ASSERT (htab->tls_sec != NULL);
5100
5101 bfd_vma base = align_power ((bfd_vma) TCB_SIZE,
5102 htab->tls_sec->alignment_power);
5103 return htab->tls_sec->vma - base;
5104 }
5105
5106 static bfd_vma *
symbol_got_offset_ref(bfd * input_bfd,struct elf_link_hash_entry * h,unsigned long r_symndx)5107 symbol_got_offset_ref (bfd *input_bfd, struct elf_link_hash_entry *h,
5108 unsigned long r_symndx)
5109 {
5110 /* Calculate the address of the GOT entry for symbol
5111 referred to in h. */
5112 if (h != NULL)
5113 return &h->got.offset;
5114 else
5115 {
5116 /* local symbol */
5117 struct elf_aarch64_local_symbol *l;
5118
5119 l = elf_aarch64_locals (input_bfd);
5120 return &l[r_symndx].got_offset;
5121 }
5122 }
5123
5124 static void
symbol_got_offset_mark(bfd * input_bfd,struct elf_link_hash_entry * h,unsigned long r_symndx)5125 symbol_got_offset_mark (bfd *input_bfd, struct elf_link_hash_entry *h,
5126 unsigned long r_symndx)
5127 {
5128 bfd_vma *p;
5129 p = symbol_got_offset_ref (input_bfd, h, r_symndx);
5130 *p |= 1;
5131 }
5132
5133 static int
symbol_got_offset_mark_p(bfd * input_bfd,struct elf_link_hash_entry * h,unsigned long r_symndx)5134 symbol_got_offset_mark_p (bfd *input_bfd, struct elf_link_hash_entry *h,
5135 unsigned long r_symndx)
5136 {
5137 bfd_vma value;
5138 value = * symbol_got_offset_ref (input_bfd, h, r_symndx);
5139 return value & 1;
5140 }
5141
5142 static bfd_vma
symbol_got_offset(bfd * input_bfd,struct elf_link_hash_entry * h,unsigned long r_symndx)5143 symbol_got_offset (bfd *input_bfd, struct elf_link_hash_entry *h,
5144 unsigned long r_symndx)
5145 {
5146 bfd_vma value;
5147 value = * symbol_got_offset_ref (input_bfd, h, r_symndx);
5148 value &= ~1;
5149 return value;
5150 }
5151
5152 static bfd_vma *
symbol_tlsdesc_got_offset_ref(bfd * input_bfd,struct elf_link_hash_entry * h,unsigned long r_symndx)5153 symbol_tlsdesc_got_offset_ref (bfd *input_bfd, struct elf_link_hash_entry *h,
5154 unsigned long r_symndx)
5155 {
5156 /* Calculate the address of the GOT entry for symbol
5157 referred to in h. */
5158 if (h != NULL)
5159 {
5160 struct elf_aarch64_link_hash_entry *eh;
5161 eh = (struct elf_aarch64_link_hash_entry *) h;
5162 return &eh->tlsdesc_got_jump_table_offset;
5163 }
5164 else
5165 {
5166 /* local symbol */
5167 struct elf_aarch64_local_symbol *l;
5168
5169 l = elf_aarch64_locals (input_bfd);
5170 return &l[r_symndx].tlsdesc_got_jump_table_offset;
5171 }
5172 }
5173
5174 static void
symbol_tlsdesc_got_offset_mark(bfd * input_bfd,struct elf_link_hash_entry * h,unsigned long r_symndx)5175 symbol_tlsdesc_got_offset_mark (bfd *input_bfd, struct elf_link_hash_entry *h,
5176 unsigned long r_symndx)
5177 {
5178 bfd_vma *p;
5179 p = symbol_tlsdesc_got_offset_ref (input_bfd, h, r_symndx);
5180 *p |= 1;
5181 }
5182
5183 static int
symbol_tlsdesc_got_offset_mark_p(bfd * input_bfd,struct elf_link_hash_entry * h,unsigned long r_symndx)5184 symbol_tlsdesc_got_offset_mark_p (bfd *input_bfd,
5185 struct elf_link_hash_entry *h,
5186 unsigned long r_symndx)
5187 {
5188 bfd_vma value;
5189 value = * symbol_tlsdesc_got_offset_ref (input_bfd, h, r_symndx);
5190 return value & 1;
5191 }
5192
5193 static bfd_vma
symbol_tlsdesc_got_offset(bfd * input_bfd,struct elf_link_hash_entry * h,unsigned long r_symndx)5194 symbol_tlsdesc_got_offset (bfd *input_bfd, struct elf_link_hash_entry *h,
5195 unsigned long r_symndx)
5196 {
5197 bfd_vma value;
5198 value = * symbol_tlsdesc_got_offset_ref (input_bfd, h, r_symndx);
5199 value &= ~1;
5200 return value;
5201 }
5202
5203 /* Data for make_branch_to_erratum_835769_stub(). */
5204
5205 struct erratum_835769_branch_to_stub_data
5206 {
5207 struct bfd_link_info *info;
5208 asection *output_section;
5209 bfd_byte *contents;
5210 };
5211
5212 /* Helper to insert branches to erratum 835769 stubs in the right
5213 places for a particular section. */
5214
5215 static bool
make_branch_to_erratum_835769_stub(struct bfd_hash_entry * gen_entry,void * in_arg)5216 make_branch_to_erratum_835769_stub (struct bfd_hash_entry *gen_entry,
5217 void *in_arg)
5218 {
5219 struct elf_aarch64_stub_hash_entry *stub_entry;
5220 struct erratum_835769_branch_to_stub_data *data;
5221 bfd_byte *contents;
5222 unsigned long branch_insn = 0;
5223 bfd_vma veneered_insn_loc, veneer_entry_loc;
5224 bfd_signed_vma branch_offset;
5225 unsigned int target;
5226 bfd *abfd;
5227
5228 stub_entry = (struct elf_aarch64_stub_hash_entry *) gen_entry;
5229 data = (struct erratum_835769_branch_to_stub_data *) in_arg;
5230
5231 if (stub_entry->target_section != data->output_section
5232 || stub_entry->stub_type != aarch64_stub_erratum_835769_veneer)
5233 return true;
5234
5235 contents = data->contents;
5236 veneered_insn_loc = stub_entry->target_section->output_section->vma
5237 + stub_entry->target_section->output_offset
5238 + stub_entry->target_value;
5239 veneer_entry_loc = stub_entry->stub_sec->output_section->vma
5240 + stub_entry->stub_sec->output_offset
5241 + stub_entry->stub_offset;
5242 branch_offset = veneer_entry_loc - veneered_insn_loc;
5243
5244 abfd = stub_entry->target_section->owner;
5245 if (!aarch64_valid_branch_p (veneer_entry_loc, veneered_insn_loc))
5246 _bfd_error_handler
5247 (_("%pB: error: erratum 835769 stub out "
5248 "of range (input file too large)"), abfd);
5249
5250 target = stub_entry->target_value;
5251 branch_insn = 0x14000000;
5252 branch_offset >>= 2;
5253 branch_offset &= 0x3ffffff;
5254 branch_insn |= branch_offset;
5255 bfd_putl32 (branch_insn, &contents[target]);
5256
5257 return true;
5258 }
5259
5260
5261 static bool
_bfd_aarch64_erratum_843419_branch_to_stub(struct bfd_hash_entry * gen_entry,void * in_arg)5262 _bfd_aarch64_erratum_843419_branch_to_stub (struct bfd_hash_entry *gen_entry,
5263 void *in_arg)
5264 {
5265 struct elf_aarch64_stub_hash_entry *stub_entry
5266 = (struct elf_aarch64_stub_hash_entry *) gen_entry;
5267 struct erratum_835769_branch_to_stub_data *data
5268 = (struct erratum_835769_branch_to_stub_data *) in_arg;
5269 struct bfd_link_info *info;
5270 struct elf_aarch64_link_hash_table *htab;
5271 bfd_byte *contents;
5272 asection *section;
5273 bfd *abfd;
5274 bfd_vma place;
5275 uint32_t insn;
5276
5277 info = data->info;
5278 contents = data->contents;
5279 section = data->output_section;
5280
5281 htab = elf_aarch64_hash_table (info);
5282
5283 if (stub_entry->target_section != section
5284 || stub_entry->stub_type != aarch64_stub_erratum_843419_veneer)
5285 return true;
5286
5287 BFD_ASSERT (((htab->fix_erratum_843419 & ERRAT_ADRP) && stub_entry->stub_sec)
5288 || (htab->fix_erratum_843419 & ERRAT_ADR));
5289
5290 /* Only update the stub section if we have one. We should always have one if
5291 we're allowed to use the ADRP errata workaround, otherwise it is not
5292 required. */
5293 if (stub_entry->stub_sec)
5294 {
5295 insn = bfd_getl32 (contents + stub_entry->target_value);
5296 bfd_putl32 (insn,
5297 stub_entry->stub_sec->contents + stub_entry->stub_offset);
5298 }
5299
5300 place = (section->output_section->vma + section->output_offset
5301 + stub_entry->adrp_offset);
5302 insn = bfd_getl32 (contents + stub_entry->adrp_offset);
5303
5304 if (!_bfd_aarch64_adrp_p (insn))
5305 abort ();
5306
5307 bfd_signed_vma imm =
5308 (_bfd_aarch64_sign_extend
5309 ((bfd_vma) _bfd_aarch64_decode_adrp_imm (insn) << 12, 33)
5310 - (place & 0xfff));
5311
5312 if ((htab->fix_erratum_843419 & ERRAT_ADR)
5313 && (imm >= AARCH64_MIN_ADRP_IMM && imm <= AARCH64_MAX_ADRP_IMM))
5314 {
5315 insn = (_bfd_aarch64_reencode_adr_imm (AARCH64_ADR_OP, imm)
5316 | AARCH64_RT (insn));
5317 bfd_putl32 (insn, contents + stub_entry->adrp_offset);
5318 /* Stub is not needed, don't map it out. */
5319 stub_entry->stub_type = aarch64_stub_none;
5320 }
5321 else if (htab->fix_erratum_843419 & ERRAT_ADRP)
5322 {
5323 bfd_vma veneered_insn_loc;
5324 bfd_vma veneer_entry_loc;
5325 bfd_signed_vma branch_offset;
5326 uint32_t branch_insn;
5327
5328 veneered_insn_loc = stub_entry->target_section->output_section->vma
5329 + stub_entry->target_section->output_offset
5330 + stub_entry->target_value;
5331 veneer_entry_loc = stub_entry->stub_sec->output_section->vma
5332 + stub_entry->stub_sec->output_offset
5333 + stub_entry->stub_offset;
5334 branch_offset = veneer_entry_loc - veneered_insn_loc;
5335
5336 abfd = stub_entry->target_section->owner;
5337 if (!aarch64_valid_branch_p (veneer_entry_loc, veneered_insn_loc))
5338 _bfd_error_handler
5339 (_("%pB: error: erratum 843419 stub out "
5340 "of range (input file too large)"), abfd);
5341
5342 branch_insn = 0x14000000;
5343 branch_offset >>= 2;
5344 branch_offset &= 0x3ffffff;
5345 branch_insn |= branch_offset;
5346 bfd_putl32 (branch_insn, contents + stub_entry->target_value);
5347 }
5348 else
5349 {
5350 abfd = stub_entry->target_section->owner;
5351 _bfd_error_handler
5352 (_("%pB: error: erratum 843419 immediate 0x%" BFD_VMA_FMT "x "
5353 "out of range for ADR (input file too large) and "
5354 "--fix-cortex-a53-843419=adr used. Run the linker with "
5355 "--fix-cortex-a53-843419=full instead"), abfd, imm);
5356 bfd_set_error (bfd_error_bad_value);
5357 /* This function is called inside a hashtable traversal and the error
5358 handlers called above turn into non-fatal errors. Which means this
5359 case ld returns an exit code 0 and also produces a broken object file.
5360 To prevent this, issue a hard abort. */
5361 BFD_FAIL ();
5362 }
5363 return true;
5364 }
5365
5366
5367 static bool
elfNN_aarch64_write_section(bfd * output_bfd ATTRIBUTE_UNUSED,struct bfd_link_info * link_info,asection * sec,bfd_byte * contents)5368 elfNN_aarch64_write_section (bfd *output_bfd ATTRIBUTE_UNUSED,
5369 struct bfd_link_info *link_info,
5370 asection *sec,
5371 bfd_byte *contents)
5372
5373 {
5374 struct elf_aarch64_link_hash_table *globals =
5375 elf_aarch64_hash_table (link_info);
5376
5377 if (globals == NULL)
5378 return false;
5379
5380 /* Fix code to point to erratum 835769 stubs. */
5381 if (globals->fix_erratum_835769)
5382 {
5383 struct erratum_835769_branch_to_stub_data data;
5384
5385 data.info = link_info;
5386 data.output_section = sec;
5387 data.contents = contents;
5388 bfd_hash_traverse (&globals->stub_hash_table,
5389 make_branch_to_erratum_835769_stub, &data);
5390 }
5391
5392 if (globals->fix_erratum_843419)
5393 {
5394 struct erratum_835769_branch_to_stub_data data;
5395
5396 data.info = link_info;
5397 data.output_section = sec;
5398 data.contents = contents;
5399 bfd_hash_traverse (&globals->stub_hash_table,
5400 _bfd_aarch64_erratum_843419_branch_to_stub, &data);
5401 }
5402
5403 return false;
5404 }
5405
5406 /* Return TRUE if RELOC is a relocation against the base of GOT table. */
5407
5408 static bool
aarch64_relocation_aginst_gp_p(bfd_reloc_code_real_type reloc)5409 aarch64_relocation_aginst_gp_p (bfd_reloc_code_real_type reloc)
5410 {
5411 return (reloc == BFD_RELOC_AARCH64_LD32_GOTPAGE_LO14
5412 || reloc == BFD_RELOC_AARCH64_LD64_GOTPAGE_LO15
5413 || reloc == BFD_RELOC_AARCH64_LD64_GOTOFF_LO15
5414 || reloc == BFD_RELOC_AARCH64_MOVW_GOTOFF_G0_NC
5415 || reloc == BFD_RELOC_AARCH64_MOVW_GOTOFF_G1);
5416 }
5417
5418 /* Perform a relocation as part of a final link. The input relocation type
5419 should be TLS relaxed. */
5420
5421 static bfd_reloc_status_type
elfNN_aarch64_final_link_relocate(reloc_howto_type * howto,bfd * input_bfd,bfd * output_bfd,asection * input_section,bfd_byte * contents,Elf_Internal_Rela * rel,bfd_vma value,struct bfd_link_info * info,asection * sym_sec,struct elf_link_hash_entry * h,bool * unresolved_reloc_p,bool save_addend,bfd_vma * saved_addend,Elf_Internal_Sym * sym)5422 elfNN_aarch64_final_link_relocate (reloc_howto_type *howto,
5423 bfd *input_bfd,
5424 bfd *output_bfd,
5425 asection *input_section,
5426 bfd_byte *contents,
5427 Elf_Internal_Rela *rel,
5428 bfd_vma value,
5429 struct bfd_link_info *info,
5430 asection *sym_sec,
5431 struct elf_link_hash_entry *h,
5432 bool *unresolved_reloc_p,
5433 bool save_addend,
5434 bfd_vma *saved_addend,
5435 Elf_Internal_Sym *sym)
5436 {
5437 Elf_Internal_Shdr *symtab_hdr;
5438 unsigned int r_type = howto->type;
5439 bfd_reloc_code_real_type bfd_r_type
5440 = elfNN_aarch64_bfd_reloc_from_howto (howto);
5441 unsigned long r_symndx;
5442 bfd_byte *hit_data = contents + rel->r_offset;
5443 bfd_vma place, off, got_entry_addr = 0;
5444 bfd_signed_vma signed_addend;
5445 struct elf_aarch64_link_hash_table *globals;
5446 bool weak_undef_p;
5447 bool relative_reloc;
5448 asection *base_got;
5449 bfd_vma orig_value = value;
5450 bool resolved_to_zero;
5451 bool abs_symbol_p;
5452
5453 globals = elf_aarch64_hash_table (info);
5454
5455 symtab_hdr = &elf_symtab_hdr (input_bfd);
5456
5457 BFD_ASSERT (is_aarch64_elf (input_bfd));
5458
5459 r_symndx = ELFNN_R_SYM (rel->r_info);
5460
5461 place = input_section->output_section->vma
5462 + input_section->output_offset + rel->r_offset;
5463
5464 /* Get addend, accumulating the addend for consecutive relocs
5465 which refer to the same offset. */
5466 signed_addend = saved_addend ? *saved_addend : 0;
5467 signed_addend += rel->r_addend;
5468
5469 weak_undef_p = (h ? h->root.type == bfd_link_hash_undefweak
5470 : bfd_is_und_section (sym_sec));
5471 abs_symbol_p = h != NULL && bfd_is_abs_symbol (&h->root);
5472
5473
5474 /* Since STT_GNU_IFUNC symbol must go through PLT, we handle
5475 it here if it is defined in a non-shared object. */
5476 if (h != NULL
5477 && h->type == STT_GNU_IFUNC
5478 && h->def_regular)
5479 {
5480 asection *plt;
5481 const char *name;
5482 bfd_vma addend = 0;
5483
5484 if ((input_section->flags & SEC_ALLOC) == 0)
5485 {
5486 /* If this is a SHT_NOTE section without SHF_ALLOC, treat
5487 STT_GNU_IFUNC symbol as STT_FUNC. */
5488 if (elf_section_type (input_section) == SHT_NOTE)
5489 goto skip_ifunc;
5490
5491 /* Dynamic relocs are not propagated for SEC_DEBUGGING
5492 sections because such sections are not SEC_ALLOC and
5493 thus ld.so will not process them. */
5494 if ((input_section->flags & SEC_DEBUGGING) != 0)
5495 return bfd_reloc_ok;
5496
5497 if (h->root.root.string)
5498 name = h->root.root.string;
5499 else
5500 name = bfd_elf_sym_name (input_bfd, symtab_hdr, sym, NULL);
5501 _bfd_error_handler
5502 /* xgettext:c-format */
5503 (_("%pB(%pA+%#" PRIx64 "): "
5504 "unresolvable %s relocation against symbol `%s'"),
5505 input_bfd, input_section, (uint64_t) rel->r_offset,
5506 howto->name, name);
5507 bfd_set_error (bfd_error_bad_value);
5508 return bfd_reloc_notsupported;
5509 }
5510 else if (h->plt.offset == (bfd_vma) -1)
5511 goto bad_ifunc_reloc;
5512
5513 /* STT_GNU_IFUNC symbol must go through PLT. */
5514 plt = globals->root.splt ? globals->root.splt : globals->root.iplt;
5515 value = (plt->output_section->vma + plt->output_offset + h->plt.offset);
5516
5517 switch (bfd_r_type)
5518 {
5519 default:
5520 bad_ifunc_reloc:
5521 if (h->root.root.string)
5522 name = h->root.root.string;
5523 else
5524 name = bfd_elf_sym_name (input_bfd, symtab_hdr, sym,
5525 NULL);
5526 _bfd_error_handler
5527 /* xgettext:c-format */
5528 (_("%pB: relocation %s against STT_GNU_IFUNC "
5529 "symbol `%s' isn't handled by %s"), input_bfd,
5530 howto->name, name, __FUNCTION__);
5531 bfd_set_error (bfd_error_bad_value);
5532 return bfd_reloc_notsupported;
5533
5534 case BFD_RELOC_AARCH64_NN:
5535 if (rel->r_addend != 0)
5536 {
5537 if (h->root.root.string)
5538 name = h->root.root.string;
5539 else
5540 name = bfd_elf_sym_name (input_bfd, symtab_hdr,
5541 sym, NULL);
5542 _bfd_error_handler
5543 /* xgettext:c-format */
5544 (_("%pB: relocation %s against STT_GNU_IFUNC "
5545 "symbol `%s' has non-zero addend: %" PRId64),
5546 input_bfd, howto->name, name, (int64_t) rel->r_addend);
5547 bfd_set_error (bfd_error_bad_value);
5548 return bfd_reloc_notsupported;
5549 }
5550
5551 /* Generate dynamic relocation only when there is a
5552 non-GOT reference in a shared object. */
5553 if (bfd_link_pic (info) && h->non_got_ref)
5554 {
5555 Elf_Internal_Rela outrel;
5556 asection *sreloc;
5557
5558 /* Need a dynamic relocation to get the real function
5559 address. */
5560 outrel.r_offset = _bfd_elf_section_offset (output_bfd,
5561 info,
5562 input_section,
5563 rel->r_offset);
5564 if (outrel.r_offset == (bfd_vma) -1
5565 || outrel.r_offset == (bfd_vma) -2)
5566 abort ();
5567
5568 outrel.r_offset += (input_section->output_section->vma
5569 + input_section->output_offset);
5570
5571 if (h->dynindx == -1
5572 || h->forced_local
5573 || bfd_link_executable (info))
5574 {
5575 /* This symbol is resolved locally. */
5576 outrel.r_info = ELFNN_R_INFO (0, AARCH64_R (IRELATIVE));
5577 outrel.r_addend = (h->root.u.def.value
5578 + h->root.u.def.section->output_section->vma
5579 + h->root.u.def.section->output_offset);
5580 }
5581 else
5582 {
5583 outrel.r_info = ELFNN_R_INFO (h->dynindx, r_type);
5584 outrel.r_addend = 0;
5585 }
5586
5587 sreloc = globals->root.irelifunc;
5588 elf_append_rela (output_bfd, sreloc, &outrel);
5589
5590 /* If this reloc is against an external symbol, we
5591 do not want to fiddle with the addend. Otherwise,
5592 we need to include the symbol value so that it
5593 becomes an addend for the dynamic reloc. For an
5594 internal symbol, we have updated addend. */
5595 return bfd_reloc_ok;
5596 }
5597 /* FALLTHROUGH */
5598 case BFD_RELOC_AARCH64_CALL26:
5599 case BFD_RELOC_AARCH64_JUMP26:
5600 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, bfd_r_type,
5601 place, value,
5602 signed_addend,
5603 weak_undef_p);
5604 return _bfd_aarch64_elf_put_addend (input_bfd, hit_data, bfd_r_type,
5605 howto, value);
5606 case BFD_RELOC_AARCH64_ADR_GOT_PAGE:
5607 case BFD_RELOC_AARCH64_GOT_LD_PREL19:
5608 case BFD_RELOC_AARCH64_LD32_GOTPAGE_LO14:
5609 case BFD_RELOC_AARCH64_LD32_GOT_LO12_NC:
5610 case BFD_RELOC_AARCH64_LD64_GOTPAGE_LO15:
5611 case BFD_RELOC_AARCH64_MOVW_GOTOFF_G0_NC:
5612 case BFD_RELOC_AARCH64_MOVW_GOTOFF_G1:
5613 case BFD_RELOC_AARCH64_LD64_GOTOFF_LO15:
5614 case BFD_RELOC_AARCH64_LD64_GOT_LO12_NC:
5615 base_got = globals->root.sgot;
5616 off = h->got.offset;
5617
5618 if (base_got == NULL)
5619 abort ();
5620
5621 if (off == (bfd_vma) -1)
5622 {
5623 bfd_vma plt_index;
5624
5625 /* We can't use h->got.offset here to save state, or
5626 even just remember the offset, as finish_dynamic_symbol
5627 would use that as offset into .got. */
5628
5629 if (globals->root.splt != NULL)
5630 {
5631 plt_index = ((h->plt.offset - globals->plt_header_size) /
5632 globals->plt_entry_size);
5633 off = (plt_index + 3) * GOT_ENTRY_SIZE;
5634 base_got = globals->root.sgotplt;
5635 }
5636 else
5637 {
5638 plt_index = h->plt.offset / globals->plt_entry_size;
5639 off = plt_index * GOT_ENTRY_SIZE;
5640 base_got = globals->root.igotplt;
5641 }
5642
5643 if (h->dynindx == -1
5644 || h->forced_local
5645 || info->symbolic)
5646 {
5647 /* This references the local definition. We must
5648 initialize this entry in the global offset table.
5649 Since the offset must always be a multiple of 8,
5650 we use the least significant bit to record
5651 whether we have initialized it already.
5652
5653 When doing a dynamic link, we create a .rela.got
5654 relocation entry to initialize the value. This
5655 is done in the finish_dynamic_symbol routine. */
5656 if ((off & 1) != 0)
5657 off &= ~1;
5658 else
5659 {
5660 bfd_put_NN (output_bfd, value,
5661 base_got->contents + off);
5662 /* Note that this is harmless as -1 | 1 still is -1. */
5663 h->got.offset |= 1;
5664 }
5665 }
5666 value = (base_got->output_section->vma
5667 + base_got->output_offset + off);
5668 }
5669 else
5670 value = aarch64_calculate_got_entry_vma (h, globals, info,
5671 value, output_bfd,
5672 unresolved_reloc_p);
5673
5674 if (aarch64_relocation_aginst_gp_p (bfd_r_type))
5675 addend = (globals->root.sgot->output_section->vma
5676 + globals->root.sgot->output_offset);
5677
5678 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, bfd_r_type,
5679 place, value,
5680 addend, weak_undef_p);
5681 return _bfd_aarch64_elf_put_addend (input_bfd, hit_data, bfd_r_type, howto, value);
5682 case BFD_RELOC_AARCH64_ADD_LO12:
5683 case BFD_RELOC_AARCH64_ADR_HI21_PCREL:
5684 break;
5685 }
5686 }
5687
5688 skip_ifunc:
5689 resolved_to_zero = (h != NULL
5690 && UNDEFWEAK_NO_DYNAMIC_RELOC (info, h));
5691
5692 switch (bfd_r_type)
5693 {
5694 case BFD_RELOC_AARCH64_NONE:
5695 case BFD_RELOC_AARCH64_TLSDESC_ADD:
5696 case BFD_RELOC_AARCH64_TLSDESC_CALL:
5697 case BFD_RELOC_AARCH64_TLSDESC_LDR:
5698 *unresolved_reloc_p = false;
5699 return bfd_reloc_ok;
5700
5701 case BFD_RELOC_AARCH64_NN:
5702
5703 /* When generating a shared object or relocatable executable, these
5704 relocations are copied into the output file to be resolved at
5705 run time. */
5706 if (((bfd_link_pic (info)
5707 || globals->root.is_relocatable_executable)
5708 && (input_section->flags & SEC_ALLOC)
5709 && (h == NULL
5710 || (ELF_ST_VISIBILITY (h->other) == STV_DEFAULT
5711 && !resolved_to_zero)
5712 || h->root.type != bfd_link_hash_undefweak))
5713 /* Or we are creating an executable, we may need to keep relocations
5714 for symbols satisfied by a dynamic library if we manage to avoid
5715 copy relocs for the symbol. */
5716 || (ELIMINATE_COPY_RELOCS
5717 && !bfd_link_pic (info)
5718 && h != NULL
5719 && (input_section->flags & SEC_ALLOC)
5720 && h->dynindx != -1
5721 && !h->non_got_ref
5722 && ((h->def_dynamic
5723 && !h->def_regular)
5724 || h->root.type == bfd_link_hash_undefweak
5725 || h->root.type == bfd_link_hash_undefined)))
5726 {
5727 Elf_Internal_Rela outrel;
5728 bfd_byte *loc;
5729 bool skip, relocate;
5730 asection *sreloc;
5731
5732 *unresolved_reloc_p = false;
5733
5734 skip = false;
5735 relocate = false;
5736
5737 outrel.r_addend = signed_addend;
5738 outrel.r_offset =
5739 _bfd_elf_section_offset (output_bfd, info, input_section,
5740 rel->r_offset);
5741 if (outrel.r_offset == (bfd_vma) - 1)
5742 skip = true;
5743 else if (outrel.r_offset == (bfd_vma) - 2)
5744 {
5745 skip = true;
5746 relocate = true;
5747 }
5748 else if (abs_symbol_p)
5749 {
5750 /* Local absolute symbol. */
5751 skip = (h->forced_local || (h->dynindx == -1));
5752 relocate = skip;
5753 }
5754
5755 outrel.r_offset += (input_section->output_section->vma
5756 + input_section->output_offset);
5757
5758 if (skip)
5759 memset (&outrel, 0, sizeof outrel);
5760 else if (h != NULL
5761 && h->dynindx != -1
5762 && (!bfd_link_pic (info)
5763 || !(bfd_link_pie (info) || SYMBOLIC_BIND (info, h))
5764 || !h->def_regular))
5765 outrel.r_info = ELFNN_R_INFO (h->dynindx, r_type);
5766 else
5767 {
5768 int symbol;
5769
5770 /* On SVR4-ish systems, the dynamic loader cannot
5771 relocate the text and data segments independently,
5772 so the symbol does not matter. */
5773 symbol = 0;
5774 relocate = !globals->no_apply_dynamic_relocs;
5775 outrel.r_info = ELFNN_R_INFO (symbol, AARCH64_R (RELATIVE));
5776 outrel.r_addend += value;
5777 }
5778
5779 sreloc = elf_section_data (input_section)->sreloc;
5780 if (sreloc == NULL || sreloc->contents == NULL)
5781 return bfd_reloc_notsupported;
5782
5783 loc = sreloc->contents + sreloc->reloc_count++ * RELOC_SIZE (globals);
5784 bfd_elfNN_swap_reloca_out (output_bfd, &outrel, loc);
5785
5786 if (sreloc->reloc_count * RELOC_SIZE (globals) > sreloc->size)
5787 {
5788 /* Sanity to check that we have previously allocated
5789 sufficient space in the relocation section for the
5790 number of relocations we actually want to emit. */
5791 abort ();
5792 }
5793
5794 /* If this reloc is against an external symbol, we do not want to
5795 fiddle with the addend. Otherwise, we need to include the symbol
5796 value so that it becomes an addend for the dynamic reloc. */
5797 if (!relocate)
5798 return bfd_reloc_ok;
5799
5800 return _bfd_final_link_relocate (howto, input_bfd, input_section,
5801 contents, rel->r_offset, value,
5802 signed_addend);
5803 }
5804 else
5805 value += signed_addend;
5806 break;
5807
5808 case BFD_RELOC_AARCH64_CALL26:
5809 case BFD_RELOC_AARCH64_JUMP26:
5810 {
5811 asection *splt = globals->root.splt;
5812 bool via_plt_p =
5813 splt != NULL && h != NULL && h->plt.offset != (bfd_vma) - 1;
5814
5815 /* A call to an undefined weak symbol is converted to a jump to
5816 the next instruction unless a PLT entry will be created.
5817 The jump to the next instruction is optimized as a NOP.
5818 Do the same for local undefined symbols. */
5819 if (weak_undef_p && ! via_plt_p)
5820 {
5821 bfd_putl32 (INSN_NOP, hit_data);
5822 return bfd_reloc_ok;
5823 }
5824
5825 /* If the call goes through a PLT entry, make sure to
5826 check distance to the right destination address. */
5827 if (via_plt_p)
5828 value = (splt->output_section->vma
5829 + splt->output_offset + h->plt.offset);
5830
5831 /* Check if a stub has to be inserted because the destination
5832 is too far away. */
5833 struct elf_aarch64_stub_hash_entry *stub_entry = NULL;
5834
5835 /* If the branch destination is directed to plt stub, "value" will be
5836 the final destination, otherwise we should plus signed_addend, it may
5837 contain non-zero value, for example call to local function symbol
5838 which are turned into "sec_sym + sec_off", and sec_off is kept in
5839 signed_addend. */
5840 if (! aarch64_valid_branch_p (via_plt_p ? value : value + signed_addend,
5841 place))
5842 /* The target is out of reach, so redirect the branch to
5843 the local stub for this function. */
5844 stub_entry = elfNN_aarch64_get_stub_entry (input_section, sym_sec, h,
5845 rel, globals);
5846 if (stub_entry != NULL)
5847 {
5848 value = (stub_entry->stub_offset
5849 + stub_entry->stub_sec->output_offset
5850 + stub_entry->stub_sec->output_section->vma);
5851
5852 /* We have redirected the destination to stub entry address,
5853 so ignore any addend record in the original rela entry. */
5854 signed_addend = 0;
5855 }
5856 }
5857 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, bfd_r_type,
5858 place, value,
5859 signed_addend, weak_undef_p);
5860 *unresolved_reloc_p = false;
5861 break;
5862
5863 case BFD_RELOC_AARCH64_16_PCREL:
5864 case BFD_RELOC_AARCH64_32_PCREL:
5865 case BFD_RELOC_AARCH64_64_PCREL:
5866 case BFD_RELOC_AARCH64_ADR_HI21_NC_PCREL:
5867 case BFD_RELOC_AARCH64_ADR_HI21_PCREL:
5868 case BFD_RELOC_AARCH64_ADR_LO21_PCREL:
5869 case BFD_RELOC_AARCH64_LD_LO19_PCREL:
5870 case BFD_RELOC_AARCH64_MOVW_PREL_G0:
5871 case BFD_RELOC_AARCH64_MOVW_PREL_G0_NC:
5872 case BFD_RELOC_AARCH64_MOVW_PREL_G1:
5873 case BFD_RELOC_AARCH64_MOVW_PREL_G1_NC:
5874 case BFD_RELOC_AARCH64_MOVW_PREL_G2:
5875 case BFD_RELOC_AARCH64_MOVW_PREL_G2_NC:
5876 case BFD_RELOC_AARCH64_MOVW_PREL_G3:
5877 if (bfd_link_pic (info)
5878 && (input_section->flags & SEC_ALLOC) != 0
5879 && (input_section->flags & SEC_READONLY) != 0
5880 && !SYMBOL_REFERENCES_LOCAL (info, h))
5881 {
5882 int howto_index = bfd_r_type - BFD_RELOC_AARCH64_RELOC_START;
5883
5884 _bfd_error_handler
5885 /* xgettext:c-format */
5886 (_("%pB: relocation %s against symbol `%s' which may bind "
5887 "externally can not be used when making a shared object; "
5888 "recompile with -fPIC"),
5889 input_bfd, elfNN_aarch64_howto_table[howto_index].name,
5890 h->root.root.string);
5891 bfd_set_error (bfd_error_bad_value);
5892 return bfd_reloc_notsupported;
5893 }
5894 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, bfd_r_type,
5895 place, value,
5896 signed_addend,
5897 weak_undef_p);
5898 break;
5899
5900 case BFD_RELOC_AARCH64_BRANCH19:
5901 case BFD_RELOC_AARCH64_TSTBR14:
5902 if (h && h->root.type == bfd_link_hash_undefined)
5903 {
5904 _bfd_error_handler
5905 /* xgettext:c-format */
5906 (_("%pB: conditional branch to undefined symbol `%s' "
5907 "not allowed"), input_bfd, h->root.root.string);
5908 bfd_set_error (bfd_error_bad_value);
5909 return bfd_reloc_notsupported;
5910 }
5911 /* Fall through. */
5912
5913 case BFD_RELOC_AARCH64_16:
5914 #if ARCH_SIZE == 64
5915 case BFD_RELOC_AARCH64_32:
5916 #endif
5917 case BFD_RELOC_AARCH64_ADD_LO12:
5918 case BFD_RELOC_AARCH64_LDST128_LO12:
5919 case BFD_RELOC_AARCH64_LDST16_LO12:
5920 case BFD_RELOC_AARCH64_LDST32_LO12:
5921 case BFD_RELOC_AARCH64_LDST64_LO12:
5922 case BFD_RELOC_AARCH64_LDST8_LO12:
5923 case BFD_RELOC_AARCH64_MOVW_G0:
5924 case BFD_RELOC_AARCH64_MOVW_G0_NC:
5925 case BFD_RELOC_AARCH64_MOVW_G0_S:
5926 case BFD_RELOC_AARCH64_MOVW_G1:
5927 case BFD_RELOC_AARCH64_MOVW_G1_NC:
5928 case BFD_RELOC_AARCH64_MOVW_G1_S:
5929 case BFD_RELOC_AARCH64_MOVW_G2:
5930 case BFD_RELOC_AARCH64_MOVW_G2_NC:
5931 case BFD_RELOC_AARCH64_MOVW_G2_S:
5932 case BFD_RELOC_AARCH64_MOVW_G3:
5933 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, bfd_r_type,
5934 place, value,
5935 signed_addend, weak_undef_p);
5936 break;
5937
5938 case BFD_RELOC_AARCH64_ADR_GOT_PAGE:
5939 case BFD_RELOC_AARCH64_GOT_LD_PREL19:
5940 case BFD_RELOC_AARCH64_LD32_GOTPAGE_LO14:
5941 case BFD_RELOC_AARCH64_LD32_GOT_LO12_NC:
5942 case BFD_RELOC_AARCH64_LD64_GOTPAGE_LO15:
5943 case BFD_RELOC_AARCH64_LD64_GOT_LO12_NC:
5944 case BFD_RELOC_AARCH64_LD64_GOTOFF_LO15:
5945 case BFD_RELOC_AARCH64_MOVW_GOTOFF_G0_NC:
5946 case BFD_RELOC_AARCH64_MOVW_GOTOFF_G1:
5947 if (globals->root.sgot == NULL)
5948 BFD_ASSERT (h != NULL);
5949
5950 relative_reloc = false;
5951 if (h != NULL)
5952 {
5953 bfd_vma addend = 0;
5954
5955 /* If a symbol is not dynamic and is not undefined weak, bind it
5956 locally and generate a RELATIVE relocation under PIC mode.
5957
5958 NOTE: one symbol may be referenced by several relocations, we
5959 should only generate one RELATIVE relocation for that symbol.
5960 Therefore, check GOT offset mark first. */
5961 if (h->dynindx == -1
5962 && !h->forced_local
5963 && h->root.type != bfd_link_hash_undefweak
5964 && bfd_link_pic (info)
5965 && !symbol_got_offset_mark_p (input_bfd, h, r_symndx))
5966 relative_reloc = true;
5967
5968 value = aarch64_calculate_got_entry_vma (h, globals, info, value,
5969 output_bfd,
5970 unresolved_reloc_p);
5971 /* Record the GOT entry address which will be used when generating
5972 RELATIVE relocation. */
5973 if (relative_reloc)
5974 got_entry_addr = value;
5975
5976 if (aarch64_relocation_aginst_gp_p (bfd_r_type))
5977 addend = (globals->root.sgot->output_section->vma
5978 + globals->root.sgot->output_offset);
5979 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, bfd_r_type,
5980 place, value,
5981 addend, weak_undef_p);
5982 }
5983 else
5984 {
5985 bfd_vma addend = 0;
5986 struct elf_aarch64_local_symbol *locals
5987 = elf_aarch64_locals (input_bfd);
5988
5989 if (locals == NULL)
5990 {
5991 int howto_index = bfd_r_type - BFD_RELOC_AARCH64_RELOC_START;
5992 _bfd_error_handler
5993 /* xgettext:c-format */
5994 (_("%pB: local symbol descriptor table be NULL when applying "
5995 "relocation %s against local symbol"),
5996 input_bfd, elfNN_aarch64_howto_table[howto_index].name);
5997 abort ();
5998 }
5999
6000 off = symbol_got_offset (input_bfd, h, r_symndx);
6001 base_got = globals->root.sgot;
6002 got_entry_addr = (base_got->output_section->vma
6003 + base_got->output_offset + off);
6004
6005 if (!symbol_got_offset_mark_p (input_bfd, h, r_symndx))
6006 {
6007 bfd_put_64 (output_bfd, value, base_got->contents + off);
6008
6009 /* For local symbol, we have done absolute relocation in static
6010 linking stage. While for shared library, we need to update the
6011 content of GOT entry according to the shared object's runtime
6012 base address. So, we need to generate a R_AARCH64_RELATIVE reloc
6013 for dynamic linker. */
6014 if (bfd_link_pic (info))
6015 relative_reloc = true;
6016
6017 symbol_got_offset_mark (input_bfd, h, r_symndx);
6018 }
6019
6020 /* Update the relocation value to GOT entry addr as we have transformed
6021 the direct data access into indirect data access through GOT. */
6022 value = got_entry_addr;
6023
6024 if (aarch64_relocation_aginst_gp_p (bfd_r_type))
6025 addend = base_got->output_section->vma + base_got->output_offset;
6026
6027 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, bfd_r_type,
6028 place, value,
6029 addend, weak_undef_p);
6030 }
6031
6032 if (relative_reloc)
6033 {
6034 asection *s;
6035 Elf_Internal_Rela outrel;
6036
6037 s = globals->root.srelgot;
6038 if (s == NULL)
6039 abort ();
6040
6041 outrel.r_offset = got_entry_addr;
6042 outrel.r_info = ELFNN_R_INFO (0, AARCH64_R (RELATIVE));
6043 outrel.r_addend = orig_value;
6044 elf_append_rela (output_bfd, s, &outrel);
6045 }
6046 break;
6047
6048 case BFD_RELOC_AARCH64_TLSGD_ADD_LO12_NC:
6049 case BFD_RELOC_AARCH64_TLSGD_ADR_PAGE21:
6050 case BFD_RELOC_AARCH64_TLSGD_ADR_PREL21:
6051 case BFD_RELOC_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21:
6052 case BFD_RELOC_AARCH64_TLSIE_LD32_GOTTPREL_LO12_NC:
6053 case BFD_RELOC_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC:
6054 case BFD_RELOC_AARCH64_TLSIE_LD_GOTTPREL_PREL19:
6055 case BFD_RELOC_AARCH64_TLSLD_ADD_LO12_NC:
6056 case BFD_RELOC_AARCH64_TLSLD_ADR_PAGE21:
6057 case BFD_RELOC_AARCH64_TLSLD_ADR_PREL21:
6058 if (globals->root.sgot == NULL)
6059 return bfd_reloc_notsupported;
6060
6061 value = (symbol_got_offset (input_bfd, h, r_symndx)
6062 + globals->root.sgot->output_section->vma
6063 + globals->root.sgot->output_offset);
6064
6065 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, bfd_r_type,
6066 place, value,
6067 0, weak_undef_p);
6068 *unresolved_reloc_p = false;
6069 break;
6070
6071 case BFD_RELOC_AARCH64_TLSGD_MOVW_G0_NC:
6072 case BFD_RELOC_AARCH64_TLSGD_MOVW_G1:
6073 case BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G0_NC:
6074 case BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G1:
6075 if (globals->root.sgot == NULL)
6076 return bfd_reloc_notsupported;
6077
6078 value = symbol_got_offset (input_bfd, h, r_symndx);
6079 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, bfd_r_type,
6080 place, value,
6081 0, weak_undef_p);
6082 *unresolved_reloc_p = false;
6083 break;
6084
6085 case BFD_RELOC_AARCH64_TLSLD_ADD_DTPREL_HI12:
6086 case BFD_RELOC_AARCH64_TLSLD_ADD_DTPREL_LO12:
6087 case BFD_RELOC_AARCH64_TLSLD_ADD_DTPREL_LO12_NC:
6088 case BFD_RELOC_AARCH64_TLSLD_LDST16_DTPREL_LO12:
6089 case BFD_RELOC_AARCH64_TLSLD_LDST16_DTPREL_LO12_NC:
6090 case BFD_RELOC_AARCH64_TLSLD_LDST32_DTPREL_LO12:
6091 case BFD_RELOC_AARCH64_TLSLD_LDST32_DTPREL_LO12_NC:
6092 case BFD_RELOC_AARCH64_TLSLD_LDST64_DTPREL_LO12:
6093 case BFD_RELOC_AARCH64_TLSLD_LDST64_DTPREL_LO12_NC:
6094 case BFD_RELOC_AARCH64_TLSLD_LDST8_DTPREL_LO12:
6095 case BFD_RELOC_AARCH64_TLSLD_LDST8_DTPREL_LO12_NC:
6096 case BFD_RELOC_AARCH64_TLSLD_MOVW_DTPREL_G0:
6097 case BFD_RELOC_AARCH64_TLSLD_MOVW_DTPREL_G0_NC:
6098 case BFD_RELOC_AARCH64_TLSLD_MOVW_DTPREL_G1:
6099 case BFD_RELOC_AARCH64_TLSLD_MOVW_DTPREL_G1_NC:
6100 case BFD_RELOC_AARCH64_TLSLD_MOVW_DTPREL_G2:
6101 {
6102 if (!(weak_undef_p || elf_hash_table (info)->tls_sec))
6103 {
6104 int howto_index = bfd_r_type - BFD_RELOC_AARCH64_RELOC_START;
6105 _bfd_error_handler
6106 /* xgettext:c-format */
6107 (_("%pB: TLS relocation %s against undefined symbol `%s'"),
6108 input_bfd, elfNN_aarch64_howto_table[howto_index].name,
6109 h->root.root.string);
6110 bfd_set_error (bfd_error_bad_value);
6111 return bfd_reloc_notsupported;
6112 }
6113
6114 bfd_vma def_value
6115 = weak_undef_p ? 0 : signed_addend - dtpoff_base (info);
6116 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, bfd_r_type,
6117 place, value,
6118 def_value, weak_undef_p);
6119 break;
6120 }
6121
6122 case BFD_RELOC_AARCH64_TLSLE_ADD_TPREL_HI12:
6123 case BFD_RELOC_AARCH64_TLSLE_ADD_TPREL_LO12:
6124 case BFD_RELOC_AARCH64_TLSLE_ADD_TPREL_LO12_NC:
6125 case BFD_RELOC_AARCH64_TLSLE_LDST16_TPREL_LO12:
6126 case BFD_RELOC_AARCH64_TLSLE_LDST16_TPREL_LO12_NC:
6127 case BFD_RELOC_AARCH64_TLSLE_LDST32_TPREL_LO12:
6128 case BFD_RELOC_AARCH64_TLSLE_LDST32_TPREL_LO12_NC:
6129 case BFD_RELOC_AARCH64_TLSLE_LDST64_TPREL_LO12:
6130 case BFD_RELOC_AARCH64_TLSLE_LDST64_TPREL_LO12_NC:
6131 case BFD_RELOC_AARCH64_TLSLE_LDST8_TPREL_LO12:
6132 case BFD_RELOC_AARCH64_TLSLE_LDST8_TPREL_LO12_NC:
6133 case BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G0:
6134 case BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G0_NC:
6135 case BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G1:
6136 case BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G1_NC:
6137 case BFD_RELOC_AARCH64_TLSLE_MOVW_TPREL_G2:
6138 {
6139 if (!(weak_undef_p || elf_hash_table (info)->tls_sec))
6140 {
6141 int howto_index = bfd_r_type - BFD_RELOC_AARCH64_RELOC_START;
6142 _bfd_error_handler
6143 /* xgettext:c-format */
6144 (_("%pB: TLS relocation %s against undefined symbol `%s'"),
6145 input_bfd, elfNN_aarch64_howto_table[howto_index].name,
6146 h->root.root.string);
6147 bfd_set_error (bfd_error_bad_value);
6148 return bfd_reloc_notsupported;
6149 }
6150
6151 bfd_vma def_value
6152 = weak_undef_p ? 0 : signed_addend - tpoff_base (info);
6153 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, bfd_r_type,
6154 place, value,
6155 def_value, weak_undef_p);
6156 *unresolved_reloc_p = false;
6157 break;
6158 }
6159
6160 case BFD_RELOC_AARCH64_TLSDESC_ADD_LO12:
6161 case BFD_RELOC_AARCH64_TLSDESC_ADR_PAGE21:
6162 case BFD_RELOC_AARCH64_TLSDESC_ADR_PREL21:
6163 case BFD_RELOC_AARCH64_TLSDESC_LD32_LO12_NC:
6164 case BFD_RELOC_AARCH64_TLSDESC_LD64_LO12:
6165 case BFD_RELOC_AARCH64_TLSDESC_LD_PREL19:
6166 if (globals->root.sgot == NULL)
6167 return bfd_reloc_notsupported;
6168 value = (symbol_tlsdesc_got_offset (input_bfd, h, r_symndx)
6169 + globals->root.sgotplt->output_section->vma
6170 + globals->root.sgotplt->output_offset
6171 + globals->sgotplt_jump_table_size);
6172
6173 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, bfd_r_type,
6174 place, value,
6175 0, weak_undef_p);
6176 *unresolved_reloc_p = false;
6177 break;
6178
6179 case BFD_RELOC_AARCH64_TLSDESC_OFF_G0_NC:
6180 case BFD_RELOC_AARCH64_TLSDESC_OFF_G1:
6181 if (globals->root.sgot == NULL)
6182 return bfd_reloc_notsupported;
6183
6184 value = (symbol_tlsdesc_got_offset (input_bfd, h, r_symndx)
6185 + globals->root.sgotplt->output_section->vma
6186 + globals->root.sgotplt->output_offset
6187 + globals->sgotplt_jump_table_size);
6188
6189 value -= (globals->root.sgot->output_section->vma
6190 + globals->root.sgot->output_offset);
6191
6192 value = _bfd_aarch64_elf_resolve_relocation (input_bfd, bfd_r_type,
6193 place, value,
6194 0, weak_undef_p);
6195 *unresolved_reloc_p = false;
6196 break;
6197
6198 default:
6199 return bfd_reloc_notsupported;
6200 }
6201
6202 if (saved_addend)
6203 *saved_addend = value;
6204
6205 /* Only apply the final relocation in a sequence. */
6206 if (save_addend)
6207 return bfd_reloc_continue;
6208
6209 return _bfd_aarch64_elf_put_addend (input_bfd, hit_data, bfd_r_type,
6210 howto, value);
6211 }
6212
6213 /* LP64 and ILP32 operates on x- and w-registers respectively.
6214 Next definitions take into account the difference between
6215 corresponding machine codes. R means x-register if the target
6216 arch is LP64, and w-register if the target is ILP32. */
6217
6218 #if ARCH_SIZE == 64
6219 # define add_R0_R0 (0x91000000)
6220 # define add_R0_R0_R1 (0x8b000020)
6221 # define add_R0_R1 (0x91400020)
6222 # define ldr_R0 (0x58000000)
6223 # define ldr_R0_mask(i) (i & 0xffffffe0)
6224 # define ldr_R0_x0 (0xf9400000)
6225 # define ldr_hw_R0 (0xf2a00000)
6226 # define movk_R0 (0xf2800000)
6227 # define movz_R0 (0xd2a00000)
6228 # define movz_hw_R0 (0xd2c00000)
6229 #else /*ARCH_SIZE == 32 */
6230 # define add_R0_R0 (0x11000000)
6231 # define add_R0_R0_R1 (0x0b000020)
6232 # define add_R0_R1 (0x11400020)
6233 # define ldr_R0 (0x18000000)
6234 # define ldr_R0_mask(i) (i & 0xbfffffe0)
6235 # define ldr_R0_x0 (0xb9400000)
6236 # define ldr_hw_R0 (0x72a00000)
6237 # define movk_R0 (0x72800000)
6238 # define movz_R0 (0x52a00000)
6239 # define movz_hw_R0 (0x52c00000)
6240 #endif
6241
6242 /* Structure to hold payload for _bfd_aarch64_erratum_843419_clear_stub,
6243 it is used to identify the stub information to reset. */
6244
6245 struct erratum_843419_branch_to_stub_clear_data
6246 {
6247 bfd_vma adrp_offset;
6248 asection *output_section;
6249 };
6250
6251 /* Clear the erratum information for GEN_ENTRY if the ADRP_OFFSET and
6252 section inside IN_ARG matches. The clearing is done by setting the
6253 stub_type to none. */
6254
6255 static bool
_bfd_aarch64_erratum_843419_clear_stub(struct bfd_hash_entry * gen_entry,void * in_arg)6256 _bfd_aarch64_erratum_843419_clear_stub (struct bfd_hash_entry *gen_entry,
6257 void *in_arg)
6258 {
6259 struct elf_aarch64_stub_hash_entry *stub_entry
6260 = (struct elf_aarch64_stub_hash_entry *) gen_entry;
6261 struct erratum_843419_branch_to_stub_clear_data *data
6262 = (struct erratum_843419_branch_to_stub_clear_data *) in_arg;
6263
6264 if (stub_entry->target_section != data->output_section
6265 || stub_entry->stub_type != aarch64_stub_erratum_843419_veneer
6266 || stub_entry->adrp_offset != data->adrp_offset)
6267 return true;
6268
6269 /* Change the stub type instead of removing the entry, removing from the hash
6270 table would be slower and we have already reserved the memory for the entry
6271 so there wouldn't be much gain. Changing the stub also keeps around a
6272 record of what was there before. */
6273 stub_entry->stub_type = aarch64_stub_none;
6274
6275 /* We're done and there could have been only one matching stub at that
6276 particular offset, so abort further traversal. */
6277 return false;
6278 }
6279
6280 /* TLS Relaxations may relax an adrp sequence that matches the erratum 843419
6281 sequence. In this case the erratum no longer applies and we need to remove
6282 the entry from the pending stub generation. This clears matching adrp insn
6283 at ADRP_OFFSET in INPUT_SECTION in the stub table defined in GLOBALS. */
6284
6285 static void
clear_erratum_843419_entry(struct elf_aarch64_link_hash_table * globals,bfd_vma adrp_offset,asection * input_section)6286 clear_erratum_843419_entry (struct elf_aarch64_link_hash_table *globals,
6287 bfd_vma adrp_offset, asection *input_section)
6288 {
6289 if (globals->fix_erratum_843419 & ERRAT_ADRP)
6290 {
6291 struct erratum_843419_branch_to_stub_clear_data data;
6292 data.adrp_offset = adrp_offset;
6293 data.output_section = input_section;
6294
6295 bfd_hash_traverse (&globals->stub_hash_table,
6296 _bfd_aarch64_erratum_843419_clear_stub, &data);
6297 }
6298 }
6299
6300 /* Handle TLS relaxations. Relaxing is possible for symbols that use
6301 R_AARCH64_TLSDESC_ADR_{PAGE, LD64_LO12_NC, ADD_LO12_NC} during a static
6302 link.
6303
6304 Return bfd_reloc_ok if we're done, bfd_reloc_continue if the caller
6305 is to then call final_link_relocate. Return other values in the
6306 case of error. */
6307
6308 static bfd_reloc_status_type
elfNN_aarch64_tls_relax(struct elf_aarch64_link_hash_table * globals,bfd * input_bfd,asection * input_section,bfd_byte * contents,Elf_Internal_Rela * rel,struct elf_link_hash_entry * h)6309 elfNN_aarch64_tls_relax (struct elf_aarch64_link_hash_table *globals,
6310 bfd *input_bfd, asection *input_section,
6311 bfd_byte *contents, Elf_Internal_Rela *rel,
6312 struct elf_link_hash_entry *h)
6313 {
6314 bool is_local = h == NULL;
6315 unsigned int r_type = ELFNN_R_TYPE (rel->r_info);
6316 unsigned long insn;
6317
6318 BFD_ASSERT (globals && input_bfd && contents && rel);
6319
6320 switch (elfNN_aarch64_bfd_reloc_from_type (input_bfd, r_type))
6321 {
6322 case BFD_RELOC_AARCH64_TLSDESC_ADR_PAGE21:
6323 case BFD_RELOC_AARCH64_TLSGD_ADR_PAGE21:
6324 if (is_local)
6325 {
6326 /* GD->LE relaxation:
6327 adrp x0, :tlsgd:var => movz R0, :tprel_g1:var
6328 or
6329 adrp x0, :tlsdesc:var => movz R0, :tprel_g1:var
6330
6331 Where R is x for LP64, and w for ILP32. */
6332 bfd_putl32 (movz_R0, contents + rel->r_offset);
6333 /* We have relaxed the adrp into a mov, we may have to clear any
6334 pending erratum fixes. */
6335 clear_erratum_843419_entry (globals, rel->r_offset, input_section);
6336 return bfd_reloc_continue;
6337 }
6338 else
6339 {
6340 /* GD->IE relaxation:
6341 adrp x0, :tlsgd:var => adrp x0, :gottprel:var
6342 or
6343 adrp x0, :tlsdesc:var => adrp x0, :gottprel:var
6344 */
6345 return bfd_reloc_continue;
6346 }
6347
6348 case BFD_RELOC_AARCH64_TLSDESC_ADR_PREL21:
6349 BFD_ASSERT (0);
6350 break;
6351
6352 case BFD_RELOC_AARCH64_TLSDESC_LD_PREL19:
6353 if (is_local)
6354 {
6355 /* Tiny TLSDESC->LE relaxation:
6356 ldr x1, :tlsdesc:var => movz R0, #:tprel_g1:var
6357 adr x0, :tlsdesc:var => movk R0, #:tprel_g0_nc:var
6358 .tlsdesccall var
6359 blr x1 => nop
6360
6361 Where R is x for LP64, and w for ILP32. */
6362 BFD_ASSERT (ELFNN_R_TYPE (rel[1].r_info) == AARCH64_R (TLSDESC_ADR_PREL21));
6363 BFD_ASSERT (ELFNN_R_TYPE (rel[2].r_info) == AARCH64_R (TLSDESC_CALL));
6364
6365 rel[1].r_info = ELFNN_R_INFO (ELFNN_R_SYM (rel->r_info),
6366 AARCH64_R (TLSLE_MOVW_TPREL_G0_NC));
6367 rel[2].r_info = ELFNN_R_INFO (STN_UNDEF, R_AARCH64_NONE);
6368
6369 bfd_putl32 (movz_R0, contents + rel->r_offset);
6370 bfd_putl32 (movk_R0, contents + rel->r_offset + 4);
6371 bfd_putl32 (INSN_NOP, contents + rel->r_offset + 8);
6372 return bfd_reloc_continue;
6373 }
6374 else
6375 {
6376 /* Tiny TLSDESC->IE relaxation:
6377 ldr x1, :tlsdesc:var => ldr x0, :gottprel:var
6378 adr x0, :tlsdesc:var => nop
6379 .tlsdesccall var
6380 blr x1 => nop
6381 */
6382 BFD_ASSERT (ELFNN_R_TYPE (rel[1].r_info) == AARCH64_R (TLSDESC_ADR_PREL21));
6383 BFD_ASSERT (ELFNN_R_TYPE (rel[2].r_info) == AARCH64_R (TLSDESC_CALL));
6384
6385 rel[1].r_info = ELFNN_R_INFO (STN_UNDEF, R_AARCH64_NONE);
6386 rel[2].r_info = ELFNN_R_INFO (STN_UNDEF, R_AARCH64_NONE);
6387
6388 bfd_putl32 (ldr_R0, contents + rel->r_offset);
6389 bfd_putl32 (INSN_NOP, contents + rel->r_offset + 4);
6390 bfd_putl32 (INSN_NOP, contents + rel->r_offset + 8);
6391 return bfd_reloc_continue;
6392 }
6393
6394 case BFD_RELOC_AARCH64_TLSGD_ADR_PREL21:
6395 if (is_local)
6396 {
6397 /* Tiny GD->LE relaxation:
6398 adr x0, :tlsgd:var => mrs x1, tpidr_el0
6399 bl __tls_get_addr => add R0, R1, #:tprel_hi12:x, lsl #12
6400 nop => add R0, R0, #:tprel_lo12_nc:x
6401
6402 Where R is x for LP64, and x for Ilp32. */
6403
6404 /* First kill the tls_get_addr reloc on the bl instruction. */
6405 BFD_ASSERT (rel->r_offset + 4 == rel[1].r_offset);
6406
6407 bfd_putl32 (0xd53bd041, contents + rel->r_offset + 0);
6408 bfd_putl32 (add_R0_R1, contents + rel->r_offset + 4);
6409 bfd_putl32 (add_R0_R0, contents + rel->r_offset + 8);
6410
6411 rel[1].r_info = ELFNN_R_INFO (ELFNN_R_SYM (rel->r_info),
6412 AARCH64_R (TLSLE_ADD_TPREL_LO12_NC));
6413 rel[1].r_offset = rel->r_offset + 8;
6414
6415 /* Move the current relocation to the second instruction in
6416 the sequence. */
6417 rel->r_offset += 4;
6418 rel->r_info = ELFNN_R_INFO (ELFNN_R_SYM (rel->r_info),
6419 AARCH64_R (TLSLE_ADD_TPREL_HI12));
6420 return bfd_reloc_continue;
6421 }
6422 else
6423 {
6424 /* Tiny GD->IE relaxation:
6425 adr x0, :tlsgd:var => ldr R0, :gottprel:var
6426 bl __tls_get_addr => mrs x1, tpidr_el0
6427 nop => add R0, R0, R1
6428
6429 Where R is x for LP64, and w for Ilp32. */
6430
6431 /* First kill the tls_get_addr reloc on the bl instruction. */
6432 BFD_ASSERT (rel->r_offset + 4 == rel[1].r_offset);
6433 rel[1].r_info = ELFNN_R_INFO (STN_UNDEF, R_AARCH64_NONE);
6434
6435 bfd_putl32 (ldr_R0, contents + rel->r_offset);
6436 bfd_putl32 (0xd53bd041, contents + rel->r_offset + 4);
6437 bfd_putl32 (add_R0_R0_R1, contents + rel->r_offset + 8);
6438 return bfd_reloc_continue;
6439 }
6440
6441 #if ARCH_SIZE == 64
6442 case BFD_RELOC_AARCH64_TLSGD_MOVW_G1:
6443 BFD_ASSERT (ELFNN_R_TYPE (rel[1].r_info) == AARCH64_R (TLSGD_MOVW_G0_NC));
6444 BFD_ASSERT (rel->r_offset + 12 == rel[2].r_offset);
6445 BFD_ASSERT (ELFNN_R_TYPE (rel[2].r_info) == AARCH64_R (CALL26));
6446
6447 if (is_local)
6448 {
6449 /* Large GD->LE relaxation:
6450 movz x0, #:tlsgd_g1:var => movz x0, #:tprel_g2:var, lsl #32
6451 movk x0, #:tlsgd_g0_nc:var => movk x0, #:tprel_g1_nc:var, lsl #16
6452 add x0, gp, x0 => movk x0, #:tprel_g0_nc:var
6453 bl __tls_get_addr => mrs x1, tpidr_el0
6454 nop => add x0, x0, x1
6455 */
6456 rel[2].r_info = ELFNN_R_INFO (ELFNN_R_SYM (rel->r_info),
6457 AARCH64_R (TLSLE_MOVW_TPREL_G0_NC));
6458 rel[2].r_offset = rel->r_offset + 8;
6459
6460 bfd_putl32 (movz_hw_R0, contents + rel->r_offset + 0);
6461 bfd_putl32 (ldr_hw_R0, contents + rel->r_offset + 4);
6462 bfd_putl32 (movk_R0, contents + rel->r_offset + 8);
6463 bfd_putl32 (0xd53bd041, contents + rel->r_offset + 12);
6464 bfd_putl32 (add_R0_R0_R1, contents + rel->r_offset + 16);
6465 }
6466 else
6467 {
6468 /* Large GD->IE relaxation:
6469 movz x0, #:tlsgd_g1:var => movz x0, #:gottprel_g1:var, lsl #16
6470 movk x0, #:tlsgd_g0_nc:var => movk x0, #:gottprel_g0_nc:var
6471 add x0, gp, x0 => ldr x0, [gp, x0]
6472 bl __tls_get_addr => mrs x1, tpidr_el0
6473 nop => add x0, x0, x1
6474 */
6475 rel[2].r_info = ELFNN_R_INFO (STN_UNDEF, R_AARCH64_NONE);
6476 bfd_putl32 (0xd2a80000, contents + rel->r_offset + 0);
6477 bfd_putl32 (ldr_R0, contents + rel->r_offset + 8);
6478 bfd_putl32 (0xd53bd041, contents + rel->r_offset + 12);
6479 bfd_putl32 (add_R0_R0_R1, contents + rel->r_offset + 16);
6480 }
6481 return bfd_reloc_continue;
6482
6483 case BFD_RELOC_AARCH64_TLSGD_MOVW_G0_NC:
6484 return bfd_reloc_continue;
6485 #endif
6486
6487 case BFD_RELOC_AARCH64_TLSIE_LD_GOTTPREL_PREL19:
6488 return bfd_reloc_continue;
6489
6490 case BFD_RELOC_AARCH64_TLSDESC_LDNN_LO12_NC:
6491 if (is_local)
6492 {
6493 /* GD->LE relaxation:
6494 ldr xd, [x0, #:tlsdesc_lo12:var] => movk x0, :tprel_g0_nc:var
6495
6496 Where R is x for lp64 mode, and w for ILP32 mode. */
6497 bfd_putl32 (movk_R0, contents + rel->r_offset);
6498 return bfd_reloc_continue;
6499 }
6500 else
6501 {
6502 /* GD->IE relaxation:
6503 ldr xd, [x0, #:tlsdesc_lo12:var] => ldr R0, [x0, #:gottprel_lo12:var]
6504
6505 Where R is x for lp64 mode, and w for ILP32 mode. */
6506 insn = bfd_getl32 (contents + rel->r_offset);
6507 bfd_putl32 (ldr_R0_mask (insn), contents + rel->r_offset);
6508 return bfd_reloc_continue;
6509 }
6510
6511 case BFD_RELOC_AARCH64_TLSGD_ADD_LO12_NC:
6512 if (is_local)
6513 {
6514 /* GD->LE relaxation
6515 add x0, #:tlsgd_lo12:var => movk R0, :tprel_g0_nc:var
6516 bl __tls_get_addr => mrs x1, tpidr_el0
6517 nop => add R0, R1, R0
6518
6519 Where R is x for lp64 mode, and w for ILP32 mode. */
6520
6521 /* First kill the tls_get_addr reloc on the bl instruction. */
6522 BFD_ASSERT (rel->r_offset + 4 == rel[1].r_offset);
6523 rel[1].r_info = ELFNN_R_INFO (STN_UNDEF, R_AARCH64_NONE);
6524
6525 bfd_putl32 (movk_R0, contents + rel->r_offset);
6526 bfd_putl32 (0xd53bd041, contents + rel->r_offset + 4);
6527 bfd_putl32 (add_R0_R0_R1, contents + rel->r_offset + 8);
6528 return bfd_reloc_continue;
6529 }
6530 else
6531 {
6532 /* GD->IE relaxation
6533 ADD x0, #:tlsgd_lo12:var => ldr R0, [x0, #:gottprel_lo12:var]
6534 BL __tls_get_addr => mrs x1, tpidr_el0
6535 R_AARCH64_CALL26
6536 NOP => add R0, R1, R0
6537
6538 Where R is x for lp64 mode, and w for ilp32 mode. */
6539
6540 BFD_ASSERT (ELFNN_R_TYPE (rel[1].r_info) == AARCH64_R (CALL26));
6541
6542 /* Remove the relocation on the BL instruction. */
6543 rel[1].r_info = ELFNN_R_INFO (STN_UNDEF, R_AARCH64_NONE);
6544
6545 /* We choose to fixup the BL and NOP instructions using the
6546 offset from the second relocation to allow flexibility in
6547 scheduling instructions between the ADD and BL. */
6548 bfd_putl32 (ldr_R0_x0, contents + rel->r_offset);
6549 bfd_putl32 (0xd53bd041, contents + rel[1].r_offset);
6550 bfd_putl32 (add_R0_R0_R1, contents + rel[1].r_offset + 4);
6551 return bfd_reloc_continue;
6552 }
6553
6554 case BFD_RELOC_AARCH64_TLSDESC_ADD:
6555 case BFD_RELOC_AARCH64_TLSDESC_ADD_LO12:
6556 case BFD_RELOC_AARCH64_TLSDESC_CALL:
6557 /* GD->IE/LE relaxation:
6558 add x0, x0, #:tlsdesc_lo12:var => nop
6559 blr xd => nop
6560 */
6561 bfd_putl32 (INSN_NOP, contents + rel->r_offset);
6562 return bfd_reloc_ok;
6563
6564 case BFD_RELOC_AARCH64_TLSDESC_LDR:
6565 if (is_local)
6566 {
6567 /* GD->LE relaxation:
6568 ldr xd, [gp, xn] => movk R0, #:tprel_g0_nc:var
6569
6570 Where R is x for lp64 mode, and w for ILP32 mode. */
6571 bfd_putl32 (movk_R0, contents + rel->r_offset);
6572 return bfd_reloc_continue;
6573 }
6574 else
6575 {
6576 /* GD->IE relaxation:
6577 ldr xd, [gp, xn] => ldr R0, [gp, xn]
6578
6579 Where R is x for lp64 mode, and w for ILP32 mode. */
6580 insn = bfd_getl32 (contents + rel->r_offset);
6581 bfd_putl32 (ldr_R0_mask (insn), contents + rel->r_offset);
6582 return bfd_reloc_ok;
6583 }
6584
6585 case BFD_RELOC_AARCH64_TLSDESC_OFF_G0_NC:
6586 /* GD->LE relaxation:
6587 movk xd, #:tlsdesc_off_g0_nc:var => movk R0, #:tprel_g1_nc:var, lsl #16
6588 GD->IE relaxation:
6589 movk xd, #:tlsdesc_off_g0_nc:var => movk Rd, #:gottprel_g0_nc:var
6590
6591 Where R is x for lp64 mode, and w for ILP32 mode. */
6592 if (is_local)
6593 bfd_putl32 (ldr_hw_R0, contents + rel->r_offset);
6594 return bfd_reloc_continue;
6595
6596 case BFD_RELOC_AARCH64_TLSDESC_OFF_G1:
6597 if (is_local)
6598 {
6599 /* GD->LE relaxation:
6600 movz xd, #:tlsdesc_off_g1:var => movz R0, #:tprel_g2:var, lsl #32
6601
6602 Where R is x for lp64 mode, and w for ILP32 mode. */
6603 bfd_putl32 (movz_hw_R0, contents + rel->r_offset);
6604 return bfd_reloc_continue;
6605 }
6606 else
6607 {
6608 /* GD->IE relaxation:
6609 movz xd, #:tlsdesc_off_g1:var => movz Rd, #:gottprel_g1:var, lsl #16
6610
6611 Where R is x for lp64 mode, and w for ILP32 mode. */
6612 insn = bfd_getl32 (contents + rel->r_offset);
6613 bfd_putl32 (movz_R0 | (insn & 0x1f), contents + rel->r_offset);
6614 return bfd_reloc_continue;
6615 }
6616
6617 case BFD_RELOC_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21:
6618 /* IE->LE relaxation:
6619 adrp xd, :gottprel:var => movz Rd, :tprel_g1:var
6620
6621 Where R is x for lp64 mode, and w for ILP32 mode. */
6622 if (is_local)
6623 {
6624 insn = bfd_getl32 (contents + rel->r_offset);
6625 bfd_putl32 (movz_R0 | (insn & 0x1f), contents + rel->r_offset);
6626 /* We have relaxed the adrp into a mov, we may have to clear any
6627 pending erratum fixes. */
6628 clear_erratum_843419_entry (globals, rel->r_offset, input_section);
6629 }
6630 return bfd_reloc_continue;
6631
6632 case BFD_RELOC_AARCH64_TLSIE_LDNN_GOTTPREL_LO12_NC:
6633 /* IE->LE relaxation:
6634 ldr xd, [xm, #:gottprel_lo12:var] => movk Rd, :tprel_g0_nc:var
6635
6636 Where R is x for lp64 mode, and w for ILP32 mode. */
6637 if (is_local)
6638 {
6639 insn = bfd_getl32 (contents + rel->r_offset);
6640 bfd_putl32 (movk_R0 | (insn & 0x1f), contents + rel->r_offset);
6641 }
6642 return bfd_reloc_continue;
6643
6644 case BFD_RELOC_AARCH64_TLSLD_ADR_PREL21:
6645 /* LD->LE relaxation (tiny):
6646 adr x0, :tlsldm:x => mrs x0, tpidr_el0
6647 bl __tls_get_addr => add R0, R0, TCB_SIZE
6648
6649 Where R is x for lp64 mode, and w for ilp32 mode. */
6650 if (is_local)
6651 {
6652 BFD_ASSERT (rel->r_offset + 4 == rel[1].r_offset);
6653 BFD_ASSERT (ELFNN_R_TYPE (rel[1].r_info) == AARCH64_R (CALL26));
6654 /* No need of CALL26 relocation for tls_get_addr. */
6655 rel[1].r_info = ELFNN_R_INFO (STN_UNDEF, R_AARCH64_NONE);
6656 bfd_putl32 (0xd53bd040, contents + rel->r_offset + 0);
6657 bfd_putl32 (add_R0_R0 | (TCB_SIZE << 10),
6658 contents + rel->r_offset + 4);
6659 return bfd_reloc_ok;
6660 }
6661 return bfd_reloc_continue;
6662
6663 case BFD_RELOC_AARCH64_TLSLD_ADR_PAGE21:
6664 /* LD->LE relaxation (small):
6665 adrp x0, :tlsldm:x => mrs x0, tpidr_el0
6666 */
6667 if (is_local)
6668 {
6669 bfd_putl32 (0xd53bd040, contents + rel->r_offset);
6670 return bfd_reloc_ok;
6671 }
6672 return bfd_reloc_continue;
6673
6674 case BFD_RELOC_AARCH64_TLSLD_ADD_LO12_NC:
6675 /* LD->LE relaxation (small):
6676 add x0, #:tlsldm_lo12:x => add R0, R0, TCB_SIZE
6677 bl __tls_get_addr => nop
6678
6679 Where R is x for lp64 mode, and w for ilp32 mode. */
6680 if (is_local)
6681 {
6682 BFD_ASSERT (rel->r_offset + 4 == rel[1].r_offset);
6683 BFD_ASSERT (ELFNN_R_TYPE (rel[1].r_info) == AARCH64_R (CALL26));
6684 /* No need of CALL26 relocation for tls_get_addr. */
6685 rel[1].r_info = ELFNN_R_INFO (STN_UNDEF, R_AARCH64_NONE);
6686 bfd_putl32 (add_R0_R0 | (TCB_SIZE << 10),
6687 contents + rel->r_offset + 0);
6688 bfd_putl32 (INSN_NOP, contents + rel->r_offset + 4);
6689 return bfd_reloc_ok;
6690 }
6691 return bfd_reloc_continue;
6692
6693 default:
6694 return bfd_reloc_continue;
6695 }
6696
6697 return bfd_reloc_ok;
6698 }
6699
6700 /* Relocate an AArch64 ELF section. */
6701
6702 static int
elfNN_aarch64_relocate_section(bfd * output_bfd,struct bfd_link_info * info,bfd * input_bfd,asection * input_section,bfd_byte * contents,Elf_Internal_Rela * relocs,Elf_Internal_Sym * local_syms,asection ** local_sections)6703 elfNN_aarch64_relocate_section (bfd *output_bfd,
6704 struct bfd_link_info *info,
6705 bfd *input_bfd,
6706 asection *input_section,
6707 bfd_byte *contents,
6708 Elf_Internal_Rela *relocs,
6709 Elf_Internal_Sym *local_syms,
6710 asection **local_sections)
6711 {
6712 Elf_Internal_Shdr *symtab_hdr;
6713 struct elf_link_hash_entry **sym_hashes;
6714 Elf_Internal_Rela *rel;
6715 Elf_Internal_Rela *relend;
6716 const char *name;
6717 struct elf_aarch64_link_hash_table *globals;
6718 bool save_addend = false;
6719 bfd_vma addend = 0;
6720
6721 globals = elf_aarch64_hash_table (info);
6722
6723 symtab_hdr = &elf_symtab_hdr (input_bfd);
6724 sym_hashes = elf_sym_hashes (input_bfd);
6725
6726 rel = relocs;
6727 relend = relocs + input_section->reloc_count;
6728 for (; rel < relend; rel++)
6729 {
6730 unsigned int r_type;
6731 bfd_reloc_code_real_type bfd_r_type;
6732 bfd_reloc_code_real_type relaxed_bfd_r_type;
6733 reloc_howto_type *howto;
6734 unsigned long r_symndx;
6735 Elf_Internal_Sym *sym;
6736 asection *sec;
6737 struct elf_link_hash_entry *h;
6738 bfd_vma relocation;
6739 bfd_reloc_status_type r;
6740 arelent bfd_reloc;
6741 char sym_type;
6742 bool unresolved_reloc = false;
6743 char *error_message = NULL;
6744
6745 r_symndx = ELFNN_R_SYM (rel->r_info);
6746 r_type = ELFNN_R_TYPE (rel->r_info);
6747
6748 bfd_reloc.howto = elfNN_aarch64_howto_from_type (input_bfd, r_type);
6749 howto = bfd_reloc.howto;
6750
6751 if (howto == NULL)
6752 return _bfd_unrecognized_reloc (input_bfd, input_section, r_type);
6753
6754 bfd_r_type = elfNN_aarch64_bfd_reloc_from_howto (howto);
6755
6756 h = NULL;
6757 sym = NULL;
6758 sec = NULL;
6759
6760 if (r_symndx < symtab_hdr->sh_info)
6761 {
6762 sym = local_syms + r_symndx;
6763 sym_type = ELFNN_ST_TYPE (sym->st_info);
6764 sec = local_sections[r_symndx];
6765
6766 /* An object file might have a reference to a local
6767 undefined symbol. This is a daft object file, but we
6768 should at least do something about it. */
6769 if (r_type != R_AARCH64_NONE && r_type != R_AARCH64_NULL
6770 && bfd_is_und_section (sec)
6771 && ELF_ST_BIND (sym->st_info) != STB_WEAK)
6772 (*info->callbacks->undefined_symbol)
6773 (info, bfd_elf_string_from_elf_section
6774 (input_bfd, symtab_hdr->sh_link, sym->st_name),
6775 input_bfd, input_section, rel->r_offset, true);
6776
6777 relocation = _bfd_elf_rela_local_sym (output_bfd, sym, &sec, rel);
6778
6779 /* Relocate against local STT_GNU_IFUNC symbol. */
6780 if (!bfd_link_relocatable (info)
6781 && ELF_ST_TYPE (sym->st_info) == STT_GNU_IFUNC)
6782 {
6783 h = elfNN_aarch64_get_local_sym_hash (globals, input_bfd,
6784 rel, false);
6785 if (h == NULL)
6786 abort ();
6787
6788 /* Set STT_GNU_IFUNC symbol value. */
6789 h->root.u.def.value = sym->st_value;
6790 h->root.u.def.section = sec;
6791 }
6792 }
6793 else
6794 {
6795 bool warned, ignored;
6796
6797 RELOC_FOR_GLOBAL_SYMBOL (info, input_bfd, input_section, rel,
6798 r_symndx, symtab_hdr, sym_hashes,
6799 h, sec, relocation,
6800 unresolved_reloc, warned, ignored);
6801
6802 sym_type = h->type;
6803 }
6804
6805 if (sec != NULL && discarded_section (sec))
6806 RELOC_AGAINST_DISCARDED_SECTION (info, input_bfd, input_section,
6807 rel, 1, relend, howto, 0, contents);
6808
6809 if (bfd_link_relocatable (info))
6810 continue;
6811
6812 if (h != NULL)
6813 name = h->root.root.string;
6814 else
6815 {
6816 name = (bfd_elf_string_from_elf_section
6817 (input_bfd, symtab_hdr->sh_link, sym->st_name));
6818 if (name == NULL || *name == '\0')
6819 name = bfd_section_name (sec);
6820 }
6821
6822 if (r_symndx != 0
6823 && r_type != R_AARCH64_NONE
6824 && r_type != R_AARCH64_NULL
6825 && (h == NULL
6826 || h->root.type == bfd_link_hash_defined
6827 || h->root.type == bfd_link_hash_defweak)
6828 && IS_AARCH64_TLS_RELOC (bfd_r_type) != (sym_type == STT_TLS))
6829 {
6830 _bfd_error_handler
6831 ((sym_type == STT_TLS
6832 /* xgettext:c-format */
6833 ? _("%pB(%pA+%#" PRIx64 "): %s used with TLS symbol %s")
6834 /* xgettext:c-format */
6835 : _("%pB(%pA+%#" PRIx64 "): %s used with non-TLS symbol %s")),
6836 input_bfd,
6837 input_section, (uint64_t) rel->r_offset, howto->name, name);
6838 }
6839
6840 /* We relax only if we can see that there can be a valid transition
6841 from a reloc type to another.
6842 We call elfNN_aarch64_final_link_relocate unless we're completely
6843 done, i.e., the relaxation produced the final output we want. */
6844
6845 relaxed_bfd_r_type = aarch64_tls_transition (input_bfd, info, r_type,
6846 h, r_symndx);
6847 if (relaxed_bfd_r_type != bfd_r_type)
6848 {
6849 bfd_r_type = relaxed_bfd_r_type;
6850 howto = elfNN_aarch64_howto_from_bfd_reloc (bfd_r_type);
6851 BFD_ASSERT (howto != NULL);
6852 r_type = howto->type;
6853 r = elfNN_aarch64_tls_relax (globals, input_bfd, input_section,
6854 contents, rel, h);
6855 unresolved_reloc = 0;
6856 }
6857 else
6858 r = bfd_reloc_continue;
6859
6860 /* There may be multiple consecutive relocations for the
6861 same offset. In that case we are supposed to treat the
6862 output of each relocation as the addend for the next. */
6863 if (rel + 1 < relend
6864 && rel->r_offset == rel[1].r_offset
6865 && ELFNN_R_TYPE (rel[1].r_info) != R_AARCH64_NONE
6866 && ELFNN_R_TYPE (rel[1].r_info) != R_AARCH64_NULL)
6867 save_addend = true;
6868 else
6869 save_addend = false;
6870
6871 if (r == bfd_reloc_continue)
6872 r = elfNN_aarch64_final_link_relocate (howto, input_bfd, output_bfd,
6873 input_section, contents, rel,
6874 relocation, info, sec,
6875 h, &unresolved_reloc,
6876 save_addend, &addend, sym);
6877
6878 switch (elfNN_aarch64_bfd_reloc_from_type (input_bfd, r_type))
6879 {
6880 case BFD_RELOC_AARCH64_TLSGD_ADD_LO12_NC:
6881 case BFD_RELOC_AARCH64_TLSGD_ADR_PAGE21:
6882 case BFD_RELOC_AARCH64_TLSGD_ADR_PREL21:
6883 case BFD_RELOC_AARCH64_TLSGD_MOVW_G0_NC:
6884 case BFD_RELOC_AARCH64_TLSGD_MOVW_G1:
6885 case BFD_RELOC_AARCH64_TLSLD_ADD_LO12_NC:
6886 case BFD_RELOC_AARCH64_TLSLD_ADR_PAGE21:
6887 case BFD_RELOC_AARCH64_TLSLD_ADR_PREL21:
6888 if (! symbol_got_offset_mark_p (input_bfd, h, r_symndx))
6889 {
6890 bool need_relocs = false;
6891 bfd_byte *loc;
6892 int indx;
6893 bfd_vma off;
6894
6895 off = symbol_got_offset (input_bfd, h, r_symndx);
6896 indx = h && h->dynindx != -1 ? h->dynindx : 0;
6897
6898 need_relocs =
6899 (!bfd_link_executable (info) || indx != 0) &&
6900 (h == NULL
6901 || ELF_ST_VISIBILITY (h->other) == STV_DEFAULT
6902 || h->root.type != bfd_link_hash_undefweak);
6903
6904 BFD_ASSERT (globals->root.srelgot != NULL);
6905
6906 if (need_relocs)
6907 {
6908 Elf_Internal_Rela rela;
6909 rela.r_info = ELFNN_R_INFO (indx, AARCH64_R (TLS_DTPMOD));
6910 rela.r_addend = 0;
6911 rela.r_offset = globals->root.sgot->output_section->vma +
6912 globals->root.sgot->output_offset + off;
6913
6914
6915 loc = globals->root.srelgot->contents;
6916 loc += globals->root.srelgot->reloc_count++
6917 * RELOC_SIZE (htab);
6918 bfd_elfNN_swap_reloca_out (output_bfd, &rela, loc);
6919
6920 bfd_reloc_code_real_type real_type =
6921 elfNN_aarch64_bfd_reloc_from_type (input_bfd, r_type);
6922
6923 if (real_type == BFD_RELOC_AARCH64_TLSLD_ADR_PREL21
6924 || real_type == BFD_RELOC_AARCH64_TLSLD_ADR_PAGE21
6925 || real_type == BFD_RELOC_AARCH64_TLSLD_ADD_LO12_NC)
6926 {
6927 /* For local dynamic, don't generate DTPREL in any case.
6928 Initialize the DTPREL slot into zero, so we get module
6929 base address when invoke runtime TLS resolver. */
6930 bfd_put_NN (output_bfd, 0,
6931 globals->root.sgot->contents + off
6932 + GOT_ENTRY_SIZE);
6933 }
6934 else if (indx == 0)
6935 {
6936 bfd_put_NN (output_bfd,
6937 relocation - dtpoff_base (info),
6938 globals->root.sgot->contents + off
6939 + GOT_ENTRY_SIZE);
6940 }
6941 else
6942 {
6943 /* This TLS symbol is global. We emit a
6944 relocation to fixup the tls offset at load
6945 time. */
6946 rela.r_info =
6947 ELFNN_R_INFO (indx, AARCH64_R (TLS_DTPREL));
6948 rela.r_addend = 0;
6949 rela.r_offset =
6950 (globals->root.sgot->output_section->vma
6951 + globals->root.sgot->output_offset + off
6952 + GOT_ENTRY_SIZE);
6953
6954 loc = globals->root.srelgot->contents;
6955 loc += globals->root.srelgot->reloc_count++
6956 * RELOC_SIZE (globals);
6957 bfd_elfNN_swap_reloca_out (output_bfd, &rela, loc);
6958 bfd_put_NN (output_bfd, (bfd_vma) 0,
6959 globals->root.sgot->contents + off
6960 + GOT_ENTRY_SIZE);
6961 }
6962 }
6963 else
6964 {
6965 bfd_put_NN (output_bfd, (bfd_vma) 1,
6966 globals->root.sgot->contents + off);
6967 bfd_put_NN (output_bfd,
6968 relocation - dtpoff_base (info),
6969 globals->root.sgot->contents + off
6970 + GOT_ENTRY_SIZE);
6971 }
6972
6973 symbol_got_offset_mark (input_bfd, h, r_symndx);
6974 }
6975 break;
6976
6977 case BFD_RELOC_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21:
6978 case BFD_RELOC_AARCH64_TLSIE_LDNN_GOTTPREL_LO12_NC:
6979 case BFD_RELOC_AARCH64_TLSIE_LD_GOTTPREL_PREL19:
6980 case BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G0_NC:
6981 case BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G1:
6982 if (! symbol_got_offset_mark_p (input_bfd, h, r_symndx))
6983 {
6984 bool need_relocs = false;
6985 bfd_byte *loc;
6986 int indx;
6987 bfd_vma off;
6988
6989 off = symbol_got_offset (input_bfd, h, r_symndx);
6990
6991 indx = h && h->dynindx != -1 ? h->dynindx : 0;
6992
6993 need_relocs =
6994 (!bfd_link_executable (info) || indx != 0) &&
6995 (h == NULL
6996 || ELF_ST_VISIBILITY (h->other) == STV_DEFAULT
6997 || h->root.type != bfd_link_hash_undefweak);
6998
6999 BFD_ASSERT (globals->root.srelgot != NULL);
7000
7001 if (need_relocs)
7002 {
7003 Elf_Internal_Rela rela;
7004
7005 if (indx == 0)
7006 rela.r_addend = relocation - dtpoff_base (info);
7007 else
7008 rela.r_addend = 0;
7009
7010 rela.r_info = ELFNN_R_INFO (indx, AARCH64_R (TLS_TPREL));
7011 rela.r_offset = globals->root.sgot->output_section->vma +
7012 globals->root.sgot->output_offset + off;
7013
7014 loc = globals->root.srelgot->contents;
7015 loc += globals->root.srelgot->reloc_count++
7016 * RELOC_SIZE (htab);
7017
7018 bfd_elfNN_swap_reloca_out (output_bfd, &rela, loc);
7019
7020 bfd_put_NN (output_bfd, rela.r_addend,
7021 globals->root.sgot->contents + off);
7022 }
7023 else
7024 bfd_put_NN (output_bfd, relocation - tpoff_base (info),
7025 globals->root.sgot->contents + off);
7026
7027 symbol_got_offset_mark (input_bfd, h, r_symndx);
7028 }
7029 break;
7030
7031 case BFD_RELOC_AARCH64_TLSDESC_ADD_LO12:
7032 case BFD_RELOC_AARCH64_TLSDESC_ADR_PAGE21:
7033 case BFD_RELOC_AARCH64_TLSDESC_ADR_PREL21:
7034 case BFD_RELOC_AARCH64_TLSDESC_LDNN_LO12_NC:
7035 case BFD_RELOC_AARCH64_TLSDESC_LD_PREL19:
7036 case BFD_RELOC_AARCH64_TLSDESC_OFF_G0_NC:
7037 case BFD_RELOC_AARCH64_TLSDESC_OFF_G1:
7038 if (! symbol_tlsdesc_got_offset_mark_p (input_bfd, h, r_symndx))
7039 {
7040 bool need_relocs = false;
7041 int indx = h && h->dynindx != -1 ? h->dynindx : 0;
7042 bfd_vma off = symbol_tlsdesc_got_offset (input_bfd, h, r_symndx);
7043
7044 need_relocs = (h == NULL
7045 || ELF_ST_VISIBILITY (h->other) == STV_DEFAULT
7046 || h->root.type != bfd_link_hash_undefweak);
7047
7048 BFD_ASSERT (globals->root.srelgot != NULL);
7049 BFD_ASSERT (globals->root.sgot != NULL);
7050
7051 if (need_relocs)
7052 {
7053 bfd_byte *loc;
7054 Elf_Internal_Rela rela;
7055 rela.r_info = ELFNN_R_INFO (indx, AARCH64_R (TLSDESC));
7056
7057 rela.r_addend = 0;
7058 rela.r_offset = (globals->root.sgotplt->output_section->vma
7059 + globals->root.sgotplt->output_offset
7060 + off + globals->sgotplt_jump_table_size);
7061
7062 if (indx == 0)
7063 rela.r_addend = relocation - dtpoff_base (info);
7064
7065 /* Allocate the next available slot in the PLT reloc
7066 section to hold our R_AARCH64_TLSDESC, the next
7067 available slot is determined from reloc_count,
7068 which we step. But note, reloc_count was
7069 artifically moved down while allocating slots for
7070 real PLT relocs such that all of the PLT relocs
7071 will fit above the initial reloc_count and the
7072 extra stuff will fit below. */
7073 loc = globals->root.srelplt->contents;
7074 loc += globals->root.srelplt->reloc_count++
7075 * RELOC_SIZE (globals);
7076
7077 bfd_elfNN_swap_reloca_out (output_bfd, &rela, loc);
7078
7079 bfd_put_NN (output_bfd, (bfd_vma) 0,
7080 globals->root.sgotplt->contents + off +
7081 globals->sgotplt_jump_table_size);
7082 bfd_put_NN (output_bfd, (bfd_vma) 0,
7083 globals->root.sgotplt->contents + off +
7084 globals->sgotplt_jump_table_size +
7085 GOT_ENTRY_SIZE);
7086 }
7087
7088 symbol_tlsdesc_got_offset_mark (input_bfd, h, r_symndx);
7089 }
7090 break;
7091 default:
7092 break;
7093 }
7094
7095 /* Dynamic relocs are not propagated for SEC_DEBUGGING sections
7096 because such sections are not SEC_ALLOC and thus ld.so will
7097 not process them. */
7098 if (unresolved_reloc
7099 && !((input_section->flags & SEC_DEBUGGING) != 0
7100 && h->def_dynamic)
7101 && _bfd_elf_section_offset (output_bfd, info, input_section,
7102 +rel->r_offset) != (bfd_vma) - 1)
7103 {
7104 _bfd_error_handler
7105 /* xgettext:c-format */
7106 (_("%pB(%pA+%#" PRIx64 "): "
7107 "unresolvable %s relocation against symbol `%s'"),
7108 input_bfd, input_section, (uint64_t) rel->r_offset, howto->name,
7109 h->root.root.string);
7110 return false;
7111 }
7112
7113 if (r != bfd_reloc_ok && r != bfd_reloc_continue)
7114 {
7115 bfd_reloc_code_real_type real_r_type
7116 = elfNN_aarch64_bfd_reloc_from_type (input_bfd, r_type);
7117
7118 switch (r)
7119 {
7120 case bfd_reloc_overflow:
7121 (*info->callbacks->reloc_overflow)
7122 (info, (h ? &h->root : NULL), name, howto->name, (bfd_vma) 0,
7123 input_bfd, input_section, rel->r_offset);
7124 if (real_r_type == BFD_RELOC_AARCH64_LD64_GOTPAGE_LO15
7125 || real_r_type == BFD_RELOC_AARCH64_LD32_GOTPAGE_LO14)
7126 {
7127 (*info->callbacks->warning)
7128 (info,
7129 _("too many GOT entries for -fpic, "
7130 "please recompile with -fPIC"),
7131 name, input_bfd, input_section, rel->r_offset);
7132 return false;
7133 }
7134 /* Overflow can occur when a variable is referenced with a type
7135 that has a larger alignment than the type with which it was
7136 declared. eg:
7137 file1.c: extern int foo; int a (void) { return foo; }
7138 file2.c: char bar, foo, baz;
7139 If the variable is placed into a data section at an offset
7140 that is incompatible with the larger alignment requirement
7141 overflow will occur. (Strictly speaking this is not overflow
7142 but rather an alignment problem, but the bfd_reloc_ error
7143 enum does not have a value to cover that situation).
7144
7145 Try to catch this situation here and provide a more helpful
7146 error message to the user. */
7147 if (addend & (((bfd_vma) 1 << howto->rightshift) - 1)
7148 /* FIXME: Are we testing all of the appropriate reloc
7149 types here ? */
7150 && (real_r_type == BFD_RELOC_AARCH64_LD_LO19_PCREL
7151 || real_r_type == BFD_RELOC_AARCH64_LDST16_LO12
7152 || real_r_type == BFD_RELOC_AARCH64_LDST32_LO12
7153 || real_r_type == BFD_RELOC_AARCH64_LDST64_LO12
7154 || real_r_type == BFD_RELOC_AARCH64_LDST128_LO12))
7155 {
7156 info->callbacks->warning
7157 (info, _("one possible cause of this error is that the \
7158 symbol is being referenced in the indicated code as if it had a larger \
7159 alignment than was declared where it was defined"),
7160 name, input_bfd, input_section, rel->r_offset);
7161 }
7162 break;
7163
7164 case bfd_reloc_undefined:
7165 (*info->callbacks->undefined_symbol)
7166 (info, name, input_bfd, input_section, rel->r_offset, true);
7167 break;
7168
7169 case bfd_reloc_outofrange:
7170 error_message = _("out of range");
7171 goto common_error;
7172
7173 case bfd_reloc_notsupported:
7174 error_message = _("unsupported relocation");
7175 goto common_error;
7176
7177 case bfd_reloc_dangerous:
7178 /* error_message should already be set. */
7179 goto common_error;
7180
7181 default:
7182 error_message = _("unknown error");
7183 /* Fall through. */
7184
7185 common_error:
7186 BFD_ASSERT (error_message != NULL);
7187 (*info->callbacks->reloc_dangerous)
7188 (info, error_message, input_bfd, input_section, rel->r_offset);
7189 break;
7190 }
7191 }
7192
7193 if (!save_addend)
7194 addend = 0;
7195 }
7196
7197 return true;
7198 }
7199
7200 /* Set the right machine number. */
7201
7202 static bool
elfNN_aarch64_object_p(bfd * abfd)7203 elfNN_aarch64_object_p (bfd *abfd)
7204 {
7205 #if ARCH_SIZE == 32
7206 bfd_default_set_arch_mach (abfd, bfd_arch_aarch64, bfd_mach_aarch64_ilp32);
7207 #else
7208 bfd_default_set_arch_mach (abfd, bfd_arch_aarch64, bfd_mach_aarch64);
7209 #endif
7210 return true;
7211 }
7212
7213 /* Function to keep AArch64 specific flags in the ELF header. */
7214
7215 static bool
elfNN_aarch64_set_private_flags(bfd * abfd,flagword flags)7216 elfNN_aarch64_set_private_flags (bfd *abfd, flagword flags)
7217 {
7218 if (elf_flags_init (abfd) && elf_elfheader (abfd)->e_flags != flags)
7219 {
7220 }
7221 else
7222 {
7223 elf_elfheader (abfd)->e_flags = flags;
7224 elf_flags_init (abfd) = true;
7225 }
7226
7227 return true;
7228 }
7229
7230 /* Merge backend specific data from an object file to the output
7231 object file when linking. */
7232
7233 static bool
elfNN_aarch64_merge_private_bfd_data(bfd * ibfd,struct bfd_link_info * info)7234 elfNN_aarch64_merge_private_bfd_data (bfd *ibfd, struct bfd_link_info *info)
7235 {
7236 bfd *obfd = info->output_bfd;
7237 flagword out_flags;
7238 flagword in_flags;
7239 bool flags_compatible = true;
7240 asection *sec;
7241
7242 /* Check if we have the same endianess. */
7243 if (!_bfd_generic_verify_endian_match (ibfd, info))
7244 return false;
7245
7246 if (!is_aarch64_elf (ibfd) || !is_aarch64_elf (obfd))
7247 return true;
7248
7249 /* The input BFD must have had its flags initialised. */
7250 /* The following seems bogus to me -- The flags are initialized in
7251 the assembler but I don't think an elf_flags_init field is
7252 written into the object. */
7253 /* BFD_ASSERT (elf_flags_init (ibfd)); */
7254
7255 in_flags = elf_elfheader (ibfd)->e_flags;
7256 out_flags = elf_elfheader (obfd)->e_flags;
7257
7258 if (!elf_flags_init (obfd))
7259 {
7260 /* If the input is the default architecture and had the default
7261 flags then do not bother setting the flags for the output
7262 architecture, instead allow future merges to do this. If no
7263 future merges ever set these flags then they will retain their
7264 uninitialised values, which surprise surprise, correspond
7265 to the default values. */
7266 if (bfd_get_arch_info (ibfd)->the_default
7267 && elf_elfheader (ibfd)->e_flags == 0)
7268 return true;
7269
7270 elf_flags_init (obfd) = true;
7271 elf_elfheader (obfd)->e_flags = in_flags;
7272
7273 if (bfd_get_arch (obfd) == bfd_get_arch (ibfd)
7274 && bfd_get_arch_info (obfd)->the_default)
7275 return bfd_set_arch_mach (obfd, bfd_get_arch (ibfd),
7276 bfd_get_mach (ibfd));
7277
7278 return true;
7279 }
7280
7281 /* Identical flags must be compatible. */
7282 if (in_flags == out_flags)
7283 return true;
7284
7285 /* Check to see if the input BFD actually contains any sections. If
7286 not, its flags may not have been initialised either, but it
7287 cannot actually cause any incompatiblity. Do not short-circuit
7288 dynamic objects; their section list may be emptied by
7289 elf_link_add_object_symbols.
7290
7291 Also check to see if there are no code sections in the input.
7292 In this case there is no need to check for code specific flags.
7293 XXX - do we need to worry about floating-point format compatability
7294 in data sections ? */
7295 if (!(ibfd->flags & DYNAMIC))
7296 {
7297 bool null_input_bfd = true;
7298 bool only_data_sections = true;
7299
7300 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
7301 {
7302 if ((bfd_section_flags (sec)
7303 & (SEC_LOAD | SEC_CODE | SEC_HAS_CONTENTS))
7304 == (SEC_LOAD | SEC_CODE | SEC_HAS_CONTENTS))
7305 only_data_sections = false;
7306
7307 null_input_bfd = false;
7308 break;
7309 }
7310
7311 if (null_input_bfd || only_data_sections)
7312 return true;
7313 }
7314
7315 return flags_compatible;
7316 }
7317
7318 /* Display the flags field. */
7319
7320 static bool
elfNN_aarch64_print_private_bfd_data(bfd * abfd,void * ptr)7321 elfNN_aarch64_print_private_bfd_data (bfd *abfd, void *ptr)
7322 {
7323 FILE *file = (FILE *) ptr;
7324 unsigned long flags;
7325
7326 BFD_ASSERT (abfd != NULL && ptr != NULL);
7327
7328 /* Print normal ELF private data. */
7329 _bfd_elf_print_private_bfd_data (abfd, ptr);
7330
7331 flags = elf_elfheader (abfd)->e_flags;
7332 /* Ignore init flag - it may not be set, despite the flags field
7333 containing valid data. */
7334
7335 /* xgettext:c-format */
7336 fprintf (file, _("private flags = 0x%lx:"), elf_elfheader (abfd)->e_flags);
7337
7338 if (flags)
7339 fprintf (file, _(" <Unrecognised flag bits set>"));
7340
7341 fputc ('\n', file);
7342
7343 return true;
7344 }
7345
7346 /* Return true if we need copy relocation against EH. */
7347
7348 static bool
need_copy_relocation_p(struct elf_aarch64_link_hash_entry * eh)7349 need_copy_relocation_p (struct elf_aarch64_link_hash_entry *eh)
7350 {
7351 struct elf_dyn_relocs *p;
7352 asection *s;
7353
7354 for (p = eh->root.dyn_relocs; p != NULL; p = p->next)
7355 {
7356 /* If there is any pc-relative reference, we need to keep copy relocation
7357 to avoid propagating the relocation into runtime that current glibc
7358 does not support. */
7359 if (p->pc_count)
7360 return true;
7361
7362 s = p->sec->output_section;
7363 /* Need copy relocation if it's against read-only section. */
7364 if (s != NULL && (s->flags & SEC_READONLY) != 0)
7365 return true;
7366 }
7367
7368 return false;
7369 }
7370
7371 /* Adjust a symbol defined by a dynamic object and referenced by a
7372 regular object. The current definition is in some section of the
7373 dynamic object, but we're not including those sections. We have to
7374 change the definition to something the rest of the link can
7375 understand. */
7376
7377 static bool
elfNN_aarch64_adjust_dynamic_symbol(struct bfd_link_info * info,struct elf_link_hash_entry * h)7378 elfNN_aarch64_adjust_dynamic_symbol (struct bfd_link_info *info,
7379 struct elf_link_hash_entry *h)
7380 {
7381 struct elf_aarch64_link_hash_table *htab;
7382 asection *s, *srel;
7383
7384 /* If this is a function, put it in the procedure linkage table. We
7385 will fill in the contents of the procedure linkage table later,
7386 when we know the address of the .got section. */
7387 if (h->type == STT_FUNC || h->type == STT_GNU_IFUNC || h->needs_plt)
7388 {
7389 if (h->plt.refcount <= 0
7390 || (h->type != STT_GNU_IFUNC
7391 && (SYMBOL_CALLS_LOCAL (info, h)
7392 || (ELF_ST_VISIBILITY (h->other) != STV_DEFAULT
7393 && h->root.type == bfd_link_hash_undefweak))))
7394 {
7395 /* This case can occur if we saw a CALL26 reloc in
7396 an input file, but the symbol wasn't referred to
7397 by a dynamic object or all references were
7398 garbage collected. In which case we can end up
7399 resolving. */
7400 h->plt.offset = (bfd_vma) - 1;
7401 h->needs_plt = 0;
7402 }
7403
7404 return true;
7405 }
7406 else
7407 /* Otherwise, reset to -1. */
7408 h->plt.offset = (bfd_vma) - 1;
7409
7410
7411 /* If this is a weak symbol, and there is a real definition, the
7412 processor independent code will have arranged for us to see the
7413 real definition first, and we can just use the same value. */
7414 if (h->is_weakalias)
7415 {
7416 struct elf_link_hash_entry *def = weakdef (h);
7417 BFD_ASSERT (def->root.type == bfd_link_hash_defined);
7418 h->root.u.def.section = def->root.u.def.section;
7419 h->root.u.def.value = def->root.u.def.value;
7420 if (ELIMINATE_COPY_RELOCS || info->nocopyreloc)
7421 h->non_got_ref = def->non_got_ref;
7422 return true;
7423 }
7424
7425 /* If we are creating a shared library, we must presume that the
7426 only references to the symbol are via the global offset table.
7427 For such cases we need not do anything here; the relocations will
7428 be handled correctly by relocate_section. */
7429 if (bfd_link_pic (info))
7430 return true;
7431
7432 /* If there are no references to this symbol that do not use the
7433 GOT, we don't need to generate a copy reloc. */
7434 if (!h->non_got_ref)
7435 return true;
7436
7437 /* If -z nocopyreloc was given, we won't generate them either. */
7438 if (info->nocopyreloc)
7439 {
7440 h->non_got_ref = 0;
7441 return true;
7442 }
7443
7444 if (ELIMINATE_COPY_RELOCS)
7445 {
7446 struct elf_aarch64_link_hash_entry *eh;
7447 /* If we don't find any dynamic relocs in read-only sections, then
7448 we'll be keeping the dynamic relocs and avoiding the copy reloc. */
7449 eh = (struct elf_aarch64_link_hash_entry *) h;
7450 if (!need_copy_relocation_p (eh))
7451 {
7452 h->non_got_ref = 0;
7453 return true;
7454 }
7455 }
7456
7457 /* We must allocate the symbol in our .dynbss section, which will
7458 become part of the .bss section of the executable. There will be
7459 an entry for this symbol in the .dynsym section. The dynamic
7460 object will contain position independent code, so all references
7461 from the dynamic object to this symbol will go through the global
7462 offset table. The dynamic linker will use the .dynsym entry to
7463 determine the address it must put in the global offset table, so
7464 both the dynamic object and the regular object will refer to the
7465 same memory location for the variable. */
7466
7467 htab = elf_aarch64_hash_table (info);
7468
7469 /* We must generate a R_AARCH64_COPY reloc to tell the dynamic linker
7470 to copy the initial value out of the dynamic object and into the
7471 runtime process image. */
7472 if ((h->root.u.def.section->flags & SEC_READONLY) != 0)
7473 {
7474 s = htab->root.sdynrelro;
7475 srel = htab->root.sreldynrelro;
7476 }
7477 else
7478 {
7479 s = htab->root.sdynbss;
7480 srel = htab->root.srelbss;
7481 }
7482 if ((h->root.u.def.section->flags & SEC_ALLOC) != 0 && h->size != 0)
7483 {
7484 srel->size += RELOC_SIZE (htab);
7485 h->needs_copy = 1;
7486 }
7487
7488 return _bfd_elf_adjust_dynamic_copy (info, h, s);
7489
7490 }
7491
7492 static bool
elfNN_aarch64_allocate_local_symbols(bfd * abfd,unsigned number)7493 elfNN_aarch64_allocate_local_symbols (bfd *abfd, unsigned number)
7494 {
7495 struct elf_aarch64_local_symbol *locals;
7496 locals = elf_aarch64_locals (abfd);
7497 if (locals == NULL)
7498 {
7499 locals = (struct elf_aarch64_local_symbol *)
7500 bfd_zalloc (abfd, number * sizeof (struct elf_aarch64_local_symbol));
7501 if (locals == NULL)
7502 return false;
7503 elf_aarch64_locals (abfd) = locals;
7504 }
7505 return true;
7506 }
7507
7508 /* Create the .got section to hold the global offset table. */
7509
7510 static bool
aarch64_elf_create_got_section(bfd * abfd,struct bfd_link_info * info)7511 aarch64_elf_create_got_section (bfd *abfd, struct bfd_link_info *info)
7512 {
7513 const struct elf_backend_data *bed = get_elf_backend_data (abfd);
7514 flagword flags;
7515 asection *s;
7516 struct elf_link_hash_entry *h;
7517 struct elf_link_hash_table *htab = elf_hash_table (info);
7518
7519 /* This function may be called more than once. */
7520 if (htab->sgot != NULL)
7521 return true;
7522
7523 flags = bed->dynamic_sec_flags;
7524
7525 s = bfd_make_section_anyway_with_flags (abfd,
7526 (bed->rela_plts_and_copies_p
7527 ? ".rela.got" : ".rel.got"),
7528 (bed->dynamic_sec_flags
7529 | SEC_READONLY));
7530 if (s == NULL
7531 || !bfd_set_section_alignment (s, bed->s->log_file_align))
7532 return false;
7533 htab->srelgot = s;
7534
7535 s = bfd_make_section_anyway_with_flags (abfd, ".got", flags);
7536 if (s == NULL
7537 || !bfd_set_section_alignment (s, bed->s->log_file_align))
7538 return false;
7539 htab->sgot = s;
7540 htab->sgot->size += GOT_ENTRY_SIZE;
7541
7542 if (bed->want_got_sym)
7543 {
7544 /* Define the symbol _GLOBAL_OFFSET_TABLE_ at the start of the .got
7545 (or .got.plt) section. We don't do this in the linker script
7546 because we don't want to define the symbol if we are not creating
7547 a global offset table. */
7548 h = _bfd_elf_define_linkage_sym (abfd, info, s,
7549 "_GLOBAL_OFFSET_TABLE_");
7550 elf_hash_table (info)->hgot = h;
7551 if (h == NULL)
7552 return false;
7553 }
7554
7555 if (bed->want_got_plt)
7556 {
7557 s = bfd_make_section_anyway_with_flags (abfd, ".got.plt", flags);
7558 if (s == NULL
7559 || !bfd_set_section_alignment (s, bed->s->log_file_align))
7560 return false;
7561 htab->sgotplt = s;
7562 }
7563
7564 /* The first bit of the global offset table is the header. */
7565 s->size += bed->got_header_size;
7566
7567 return true;
7568 }
7569
7570 /* Look through the relocs for a section during the first phase. */
7571
7572 static bool
elfNN_aarch64_check_relocs(bfd * abfd,struct bfd_link_info * info,asection * sec,const Elf_Internal_Rela * relocs)7573 elfNN_aarch64_check_relocs (bfd *abfd, struct bfd_link_info *info,
7574 asection *sec, const Elf_Internal_Rela *relocs)
7575 {
7576 Elf_Internal_Shdr *symtab_hdr;
7577 struct elf_link_hash_entry **sym_hashes;
7578 const Elf_Internal_Rela *rel;
7579 const Elf_Internal_Rela *rel_end;
7580 asection *sreloc;
7581
7582 struct elf_aarch64_link_hash_table *htab;
7583
7584 if (bfd_link_relocatable (info))
7585 return true;
7586
7587 BFD_ASSERT (is_aarch64_elf (abfd));
7588
7589 htab = elf_aarch64_hash_table (info);
7590 sreloc = NULL;
7591
7592 symtab_hdr = &elf_symtab_hdr (abfd);
7593 sym_hashes = elf_sym_hashes (abfd);
7594
7595 rel_end = relocs + sec->reloc_count;
7596 for (rel = relocs; rel < rel_end; rel++)
7597 {
7598 struct elf_link_hash_entry *h;
7599 unsigned int r_symndx;
7600 unsigned int r_type;
7601 bfd_reloc_code_real_type bfd_r_type;
7602 Elf_Internal_Sym *isym;
7603
7604 r_symndx = ELFNN_R_SYM (rel->r_info);
7605 r_type = ELFNN_R_TYPE (rel->r_info);
7606
7607 if (r_symndx >= NUM_SHDR_ENTRIES (symtab_hdr))
7608 {
7609 /* xgettext:c-format */
7610 _bfd_error_handler (_("%pB: bad symbol index: %d"), abfd, r_symndx);
7611 return false;
7612 }
7613
7614 if (r_symndx < symtab_hdr->sh_info)
7615 {
7616 /* A local symbol. */
7617 isym = bfd_sym_from_r_symndx (&htab->root.sym_cache,
7618 abfd, r_symndx);
7619 if (isym == NULL)
7620 return false;
7621
7622 /* Check relocation against local STT_GNU_IFUNC symbol. */
7623 if (ELF_ST_TYPE (isym->st_info) == STT_GNU_IFUNC)
7624 {
7625 h = elfNN_aarch64_get_local_sym_hash (htab, abfd, rel,
7626 true);
7627 if (h == NULL)
7628 return false;
7629
7630 /* Fake a STT_GNU_IFUNC symbol. */
7631 h->type = STT_GNU_IFUNC;
7632 h->def_regular = 1;
7633 h->ref_regular = 1;
7634 h->forced_local = 1;
7635 h->root.type = bfd_link_hash_defined;
7636 }
7637 else
7638 h = NULL;
7639 }
7640 else
7641 {
7642 h = sym_hashes[r_symndx - symtab_hdr->sh_info];
7643 while (h->root.type == bfd_link_hash_indirect
7644 || h->root.type == bfd_link_hash_warning)
7645 h = (struct elf_link_hash_entry *) h->root.u.i.link;
7646 }
7647
7648 /* Could be done earlier, if h were already available. */
7649 bfd_r_type = aarch64_tls_transition (abfd, info, r_type, h, r_symndx);
7650
7651 if (h != NULL)
7652 {
7653 /* If a relocation refers to _GLOBAL_OFFSET_TABLE_, create the .got.
7654 This shows up in particular in an R_AARCH64_PREL64 in large model
7655 when calculating the pc-relative address to .got section which is
7656 used to initialize the gp register. */
7657 if (h->root.root.string
7658 && strcmp (h->root.root.string, "_GLOBAL_OFFSET_TABLE_") == 0)
7659 {
7660 if (htab->root.dynobj == NULL)
7661 htab->root.dynobj = abfd;
7662
7663 if (! aarch64_elf_create_got_section (htab->root.dynobj, info))
7664 return false;
7665
7666 BFD_ASSERT (h == htab->root.hgot);
7667 }
7668
7669 /* Create the ifunc sections for static executables. If we
7670 never see an indirect function symbol nor we are building
7671 a static executable, those sections will be empty and
7672 won't appear in output. */
7673 switch (bfd_r_type)
7674 {
7675 default:
7676 break;
7677
7678 case BFD_RELOC_AARCH64_ADD_LO12:
7679 case BFD_RELOC_AARCH64_ADR_GOT_PAGE:
7680 case BFD_RELOC_AARCH64_ADR_HI21_PCREL:
7681 case BFD_RELOC_AARCH64_CALL26:
7682 case BFD_RELOC_AARCH64_GOT_LD_PREL19:
7683 case BFD_RELOC_AARCH64_JUMP26:
7684 case BFD_RELOC_AARCH64_LD32_GOTPAGE_LO14:
7685 case BFD_RELOC_AARCH64_LD32_GOT_LO12_NC:
7686 case BFD_RELOC_AARCH64_LD64_GOTOFF_LO15:
7687 case BFD_RELOC_AARCH64_LD64_GOTPAGE_LO15:
7688 case BFD_RELOC_AARCH64_LD64_GOT_LO12_NC:
7689 case BFD_RELOC_AARCH64_MOVW_GOTOFF_G0_NC:
7690 case BFD_RELOC_AARCH64_MOVW_GOTOFF_G1:
7691 case BFD_RELOC_AARCH64_NN:
7692 if (htab->root.dynobj == NULL)
7693 htab->root.dynobj = abfd;
7694 if (!_bfd_elf_create_ifunc_sections (htab->root.dynobj, info))
7695 return false;
7696 break;
7697 }
7698
7699 /* It is referenced by a non-shared object. */
7700 h->ref_regular = 1;
7701 }
7702
7703 switch (bfd_r_type)
7704 {
7705 case BFD_RELOC_AARCH64_16:
7706 #if ARCH_SIZE == 64
7707 case BFD_RELOC_AARCH64_32:
7708 #endif
7709 if (bfd_link_pic (info) && (sec->flags & SEC_ALLOC) != 0)
7710 {
7711 if (h != NULL
7712 /* This is an absolute symbol. It represents a value instead
7713 of an address. */
7714 && (bfd_is_abs_symbol (&h->root)
7715 /* This is an undefined symbol. */
7716 || h->root.type == bfd_link_hash_undefined))
7717 break;
7718
7719 /* For local symbols, defined global symbols in a non-ABS section,
7720 it is assumed that the value is an address. */
7721 int howto_index = bfd_r_type - BFD_RELOC_AARCH64_RELOC_START;
7722 _bfd_error_handler
7723 /* xgettext:c-format */
7724 (_("%pB: relocation %s against `%s' can not be used when making "
7725 "a shared object"),
7726 abfd, elfNN_aarch64_howto_table[howto_index].name,
7727 (h) ? h->root.root.string : "a local symbol");
7728 bfd_set_error (bfd_error_bad_value);
7729 return false;
7730 }
7731 else
7732 break;
7733
7734 case BFD_RELOC_AARCH64_MOVW_G0_NC:
7735 case BFD_RELOC_AARCH64_MOVW_G1_NC:
7736 case BFD_RELOC_AARCH64_MOVW_G2_NC:
7737 case BFD_RELOC_AARCH64_MOVW_G3:
7738 if (bfd_link_pic (info))
7739 {
7740 int howto_index = bfd_r_type - BFD_RELOC_AARCH64_RELOC_START;
7741 _bfd_error_handler
7742 /* xgettext:c-format */
7743 (_("%pB: relocation %s against `%s' can not be used when making "
7744 "a shared object; recompile with -fPIC"),
7745 abfd, elfNN_aarch64_howto_table[howto_index].name,
7746 (h) ? h->root.root.string : "a local symbol");
7747 bfd_set_error (bfd_error_bad_value);
7748 return false;
7749 }
7750 /* Fall through. */
7751
7752 case BFD_RELOC_AARCH64_16_PCREL:
7753 case BFD_RELOC_AARCH64_32_PCREL:
7754 case BFD_RELOC_AARCH64_64_PCREL:
7755 case BFD_RELOC_AARCH64_ADD_LO12:
7756 case BFD_RELOC_AARCH64_ADR_HI21_NC_PCREL:
7757 case BFD_RELOC_AARCH64_ADR_HI21_PCREL:
7758 case BFD_RELOC_AARCH64_ADR_LO21_PCREL:
7759 case BFD_RELOC_AARCH64_LDST128_LO12:
7760 case BFD_RELOC_AARCH64_LDST16_LO12:
7761 case BFD_RELOC_AARCH64_LDST32_LO12:
7762 case BFD_RELOC_AARCH64_LDST64_LO12:
7763 case BFD_RELOC_AARCH64_LDST8_LO12:
7764 case BFD_RELOC_AARCH64_LD_LO19_PCREL:
7765 if (h == NULL || bfd_link_pic (info))
7766 break;
7767 /* Fall through. */
7768
7769 case BFD_RELOC_AARCH64_NN:
7770
7771 /* We don't need to handle relocs into sections not going into
7772 the "real" output. */
7773 if ((sec->flags & SEC_ALLOC) == 0)
7774 break;
7775
7776 if (h != NULL)
7777 {
7778 if (!bfd_link_pic (info))
7779 h->non_got_ref = 1;
7780
7781 h->plt.refcount += 1;
7782 h->pointer_equality_needed = 1;
7783 }
7784
7785 /* No need to do anything if we're not creating a shared
7786 object. */
7787 if (!(bfd_link_pic (info)
7788 /* If on the other hand, we are creating an executable, we
7789 may need to keep relocations for symbols satisfied by a
7790 dynamic library if we manage to avoid copy relocs for the
7791 symbol.
7792
7793 NOTE: Currently, there is no support of copy relocs
7794 elimination on pc-relative relocation types, because there is
7795 no dynamic relocation support for them in glibc. We still
7796 record the dynamic symbol reference for them. This is
7797 because one symbol may be referenced by both absolute
7798 relocation (for example, BFD_RELOC_AARCH64_NN) and
7799 pc-relative relocation. We need full symbol reference
7800 information to make correct decision later in
7801 elfNN_aarch64_adjust_dynamic_symbol. */
7802 || (ELIMINATE_COPY_RELOCS
7803 && !bfd_link_pic (info)
7804 && h != NULL
7805 && (h->root.type == bfd_link_hash_defweak
7806 || !h->def_regular))))
7807 break;
7808
7809 {
7810 struct elf_dyn_relocs *p;
7811 struct elf_dyn_relocs **head;
7812 int howto_index = bfd_r_type - BFD_RELOC_AARCH64_RELOC_START;
7813
7814 /* We must copy these reloc types into the output file.
7815 Create a reloc section in dynobj and make room for
7816 this reloc. */
7817 if (sreloc == NULL)
7818 {
7819 if (htab->root.dynobj == NULL)
7820 htab->root.dynobj = abfd;
7821
7822 sreloc = _bfd_elf_make_dynamic_reloc_section
7823 (sec, htab->root.dynobj, LOG_FILE_ALIGN, abfd, /*rela? */ true);
7824
7825 if (sreloc == NULL)
7826 return false;
7827 }
7828
7829 /* If this is a global symbol, we count the number of
7830 relocations we need for this symbol. */
7831 if (h != NULL)
7832 {
7833 head = &h->dyn_relocs;
7834 }
7835 else
7836 {
7837 /* Track dynamic relocs needed for local syms too.
7838 We really need local syms available to do this
7839 easily. Oh well. */
7840
7841 asection *s;
7842 void **vpp;
7843
7844 isym = bfd_sym_from_r_symndx (&htab->root.sym_cache,
7845 abfd, r_symndx);
7846 if (isym == NULL)
7847 return false;
7848
7849 s = bfd_section_from_elf_index (abfd, isym->st_shndx);
7850 if (s == NULL)
7851 s = sec;
7852
7853 /* Beware of type punned pointers vs strict aliasing
7854 rules. */
7855 vpp = &(elf_section_data (s)->local_dynrel);
7856 head = (struct elf_dyn_relocs **) vpp;
7857 }
7858
7859 p = *head;
7860 if (p == NULL || p->sec != sec)
7861 {
7862 size_t amt = sizeof *p;
7863 p = ((struct elf_dyn_relocs *)
7864 bfd_zalloc (htab->root.dynobj, amt));
7865 if (p == NULL)
7866 return false;
7867 p->next = *head;
7868 *head = p;
7869 p->sec = sec;
7870 }
7871
7872 p->count += 1;
7873
7874 if (elfNN_aarch64_howto_table[howto_index].pc_relative)
7875 p->pc_count += 1;
7876 }
7877 break;
7878
7879 /* RR: We probably want to keep a consistency check that
7880 there are no dangling GOT_PAGE relocs. */
7881 case BFD_RELOC_AARCH64_ADR_GOT_PAGE:
7882 case BFD_RELOC_AARCH64_GOT_LD_PREL19:
7883 case BFD_RELOC_AARCH64_LD32_GOTPAGE_LO14:
7884 case BFD_RELOC_AARCH64_LD32_GOT_LO12_NC:
7885 case BFD_RELOC_AARCH64_LD64_GOTOFF_LO15:
7886 case BFD_RELOC_AARCH64_LD64_GOTPAGE_LO15:
7887 case BFD_RELOC_AARCH64_LD64_GOT_LO12_NC:
7888 case BFD_RELOC_AARCH64_MOVW_GOTOFF_G0_NC:
7889 case BFD_RELOC_AARCH64_MOVW_GOTOFF_G1:
7890 case BFD_RELOC_AARCH64_TLSDESC_ADD_LO12:
7891 case BFD_RELOC_AARCH64_TLSDESC_ADR_PAGE21:
7892 case BFD_RELOC_AARCH64_TLSDESC_ADR_PREL21:
7893 case BFD_RELOC_AARCH64_TLSDESC_LD32_LO12_NC:
7894 case BFD_RELOC_AARCH64_TLSDESC_LD64_LO12:
7895 case BFD_RELOC_AARCH64_TLSDESC_LD_PREL19:
7896 case BFD_RELOC_AARCH64_TLSDESC_OFF_G0_NC:
7897 case BFD_RELOC_AARCH64_TLSDESC_OFF_G1:
7898 case BFD_RELOC_AARCH64_TLSGD_ADD_LO12_NC:
7899 case BFD_RELOC_AARCH64_TLSGD_ADR_PAGE21:
7900 case BFD_RELOC_AARCH64_TLSGD_ADR_PREL21:
7901 case BFD_RELOC_AARCH64_TLSGD_MOVW_G0_NC:
7902 case BFD_RELOC_AARCH64_TLSGD_MOVW_G1:
7903 case BFD_RELOC_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21:
7904 case BFD_RELOC_AARCH64_TLSIE_LD32_GOTTPREL_LO12_NC:
7905 case BFD_RELOC_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC:
7906 case BFD_RELOC_AARCH64_TLSIE_LD_GOTTPREL_PREL19:
7907 case BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G0_NC:
7908 case BFD_RELOC_AARCH64_TLSIE_MOVW_GOTTPREL_G1:
7909 case BFD_RELOC_AARCH64_TLSLD_ADD_LO12_NC:
7910 case BFD_RELOC_AARCH64_TLSLD_ADR_PAGE21:
7911 case BFD_RELOC_AARCH64_TLSLD_ADR_PREL21:
7912 {
7913 unsigned got_type;
7914 unsigned old_got_type;
7915
7916 got_type = aarch64_reloc_got_type (bfd_r_type);
7917
7918 if (h)
7919 {
7920 h->got.refcount += 1;
7921 old_got_type = elf_aarch64_hash_entry (h)->got_type;
7922 }
7923 else
7924 {
7925 struct elf_aarch64_local_symbol *locals;
7926
7927 if (!elfNN_aarch64_allocate_local_symbols
7928 (abfd, symtab_hdr->sh_info))
7929 return false;
7930
7931 locals = elf_aarch64_locals (abfd);
7932 BFD_ASSERT (r_symndx < symtab_hdr->sh_info);
7933 locals[r_symndx].got_refcount += 1;
7934 old_got_type = locals[r_symndx].got_type;
7935 }
7936
7937 /* If a variable is accessed with both general dynamic TLS
7938 methods, two slots may be created. */
7939 if (GOT_TLS_GD_ANY_P (old_got_type) && GOT_TLS_GD_ANY_P (got_type))
7940 got_type |= old_got_type;
7941
7942 /* We will already have issued an error message if there
7943 is a TLS/non-TLS mismatch, based on the symbol type.
7944 So just combine any TLS types needed. */
7945 if (old_got_type != GOT_UNKNOWN && old_got_type != GOT_NORMAL
7946 && got_type != GOT_NORMAL)
7947 got_type |= old_got_type;
7948
7949 /* If the symbol is accessed by both IE and GD methods, we
7950 are able to relax. Turn off the GD flag, without
7951 messing up with any other kind of TLS types that may be
7952 involved. */
7953 if ((got_type & GOT_TLS_IE) && GOT_TLS_GD_ANY_P (got_type))
7954 got_type &= ~ (GOT_TLSDESC_GD | GOT_TLS_GD);
7955
7956 if (old_got_type != got_type)
7957 {
7958 if (h != NULL)
7959 elf_aarch64_hash_entry (h)->got_type = got_type;
7960 else
7961 {
7962 struct elf_aarch64_local_symbol *locals;
7963 locals = elf_aarch64_locals (abfd);
7964 BFD_ASSERT (r_symndx < symtab_hdr->sh_info);
7965 locals[r_symndx].got_type = got_type;
7966 }
7967 }
7968
7969 if (htab->root.dynobj == NULL)
7970 htab->root.dynobj = abfd;
7971 if (! aarch64_elf_create_got_section (htab->root.dynobj, info))
7972 return false;
7973 break;
7974 }
7975
7976 case BFD_RELOC_AARCH64_CALL26:
7977 case BFD_RELOC_AARCH64_JUMP26:
7978 /* If this is a local symbol then we resolve it
7979 directly without creating a PLT entry. */
7980 if (h == NULL)
7981 continue;
7982
7983 h->needs_plt = 1;
7984 if (h->plt.refcount <= 0)
7985 h->plt.refcount = 1;
7986 else
7987 h->plt.refcount += 1;
7988 break;
7989
7990 default:
7991 break;
7992 }
7993 }
7994
7995 return true;
7996 }
7997
7998 /* Treat mapping symbols as special target symbols. */
7999
8000 static bool
elfNN_aarch64_is_target_special_symbol(bfd * abfd ATTRIBUTE_UNUSED,asymbol * sym)8001 elfNN_aarch64_is_target_special_symbol (bfd *abfd ATTRIBUTE_UNUSED,
8002 asymbol *sym)
8003 {
8004 return bfd_is_aarch64_special_symbol_name (sym->name,
8005 BFD_AARCH64_SPECIAL_SYM_TYPE_ANY);
8006 }
8007
8008 /* If the ELF symbol SYM might be a function in SEC, return the
8009 function size and set *CODE_OFF to the function's entry point,
8010 otherwise return zero. */
8011
8012 static bfd_size_type
elfNN_aarch64_maybe_function_sym(const asymbol * sym,asection * sec,bfd_vma * code_off)8013 elfNN_aarch64_maybe_function_sym (const asymbol *sym, asection *sec,
8014 bfd_vma *code_off)
8015 {
8016 bfd_size_type size;
8017 elf_symbol_type * elf_sym = (elf_symbol_type *) sym;
8018
8019 if ((sym->flags & (BSF_SECTION_SYM | BSF_FILE | BSF_OBJECT
8020 | BSF_THREAD_LOCAL | BSF_RELC | BSF_SRELC)) != 0
8021 || sym->section != sec)
8022 return 0;
8023
8024 size = (sym->flags & BSF_SYNTHETIC) ? 0 : elf_sym->internal_elf_sym.st_size;
8025
8026 if (!(sym->flags & BSF_SYNTHETIC))
8027 switch (ELF_ST_TYPE (elf_sym->internal_elf_sym.st_info))
8028 {
8029 case STT_NOTYPE:
8030 /* Ignore symbols created by the annobin plugin for gcc and clang.
8031 These symbols are hidden, local, notype and have a size of 0. */
8032 if (size == 0
8033 && sym->flags & BSF_LOCAL
8034 && ELF_ST_VISIBILITY (elf_sym->internal_elf_sym.st_other) == STV_HIDDEN)
8035 return 0;
8036 /* Fall through. */
8037 case STT_FUNC:
8038 /* FIXME: Allow STT_GNU_IFUNC as well ? */
8039 break;
8040 default:
8041 return 0;
8042 }
8043
8044 if ((sym->flags & BSF_LOCAL)
8045 && bfd_is_aarch64_special_symbol_name (sym->name,
8046 BFD_AARCH64_SPECIAL_SYM_TYPE_ANY))
8047 return 0;
8048
8049 *code_off = sym->value;
8050
8051 /* Do not return 0 for the function's size. */
8052 return size ? size : 1;
8053 }
8054
8055 static bool
elfNN_aarch64_find_inliner_info(bfd * abfd,const char ** filename_ptr,const char ** functionname_ptr,unsigned int * line_ptr)8056 elfNN_aarch64_find_inliner_info (bfd *abfd,
8057 const char **filename_ptr,
8058 const char **functionname_ptr,
8059 unsigned int *line_ptr)
8060 {
8061 bool found;
8062 found = _bfd_dwarf2_find_inliner_info
8063 (abfd, filename_ptr,
8064 functionname_ptr, line_ptr, &elf_tdata (abfd)->dwarf2_find_line_info);
8065 return found;
8066 }
8067
8068
8069 static bool
elfNN_aarch64_init_file_header(bfd * abfd,struct bfd_link_info * link_info)8070 elfNN_aarch64_init_file_header (bfd *abfd, struct bfd_link_info *link_info)
8071 {
8072 Elf_Internal_Ehdr *i_ehdrp; /* ELF file header, internal form. */
8073
8074 if (!_bfd_elf_init_file_header (abfd, link_info))
8075 return false;
8076
8077 i_ehdrp = elf_elfheader (abfd);
8078 i_ehdrp->e_ident[EI_ABIVERSION] = AARCH64_ELF_ABI_VERSION;
8079 return true;
8080 }
8081
8082 static enum elf_reloc_type_class
elfNN_aarch64_reloc_type_class(const struct bfd_link_info * info ATTRIBUTE_UNUSED,const asection * rel_sec ATTRIBUTE_UNUSED,const Elf_Internal_Rela * rela)8083 elfNN_aarch64_reloc_type_class (const struct bfd_link_info *info ATTRIBUTE_UNUSED,
8084 const asection *rel_sec ATTRIBUTE_UNUSED,
8085 const Elf_Internal_Rela *rela)
8086 {
8087 struct elf_aarch64_link_hash_table *htab = elf_aarch64_hash_table (info);
8088
8089 if (htab->root.dynsym != NULL
8090 && htab->root.dynsym->contents != NULL)
8091 {
8092 /* Check relocation against STT_GNU_IFUNC symbol if there are
8093 dynamic symbols. */
8094 bfd *abfd = info->output_bfd;
8095 const struct elf_backend_data *bed = get_elf_backend_data (abfd);
8096 unsigned long r_symndx = ELFNN_R_SYM (rela->r_info);
8097 if (r_symndx != STN_UNDEF)
8098 {
8099 Elf_Internal_Sym sym;
8100 if (!bed->s->swap_symbol_in (abfd,
8101 (htab->root.dynsym->contents
8102 + r_symndx * bed->s->sizeof_sym),
8103 0, &sym))
8104 {
8105 /* xgettext:c-format */
8106 _bfd_error_handler (_("%pB symbol number %lu references"
8107 " nonexistent SHT_SYMTAB_SHNDX section"),
8108 abfd, r_symndx);
8109 /* Ideally an error class should be returned here. */
8110 }
8111 else if (ELF_ST_TYPE (sym.st_info) == STT_GNU_IFUNC)
8112 return reloc_class_ifunc;
8113 }
8114 }
8115
8116 switch ((int) ELFNN_R_TYPE (rela->r_info))
8117 {
8118 case AARCH64_R (IRELATIVE):
8119 return reloc_class_ifunc;
8120 case AARCH64_R (RELATIVE):
8121 return reloc_class_relative;
8122 case AARCH64_R (JUMP_SLOT):
8123 return reloc_class_plt;
8124 case AARCH64_R (COPY):
8125 return reloc_class_copy;
8126 default:
8127 return reloc_class_normal;
8128 }
8129 }
8130
8131 /* Handle an AArch64 specific section when reading an object file. This is
8132 called when bfd_section_from_shdr finds a section with an unknown
8133 type. */
8134
8135 static bool
elfNN_aarch64_section_from_shdr(bfd * abfd,Elf_Internal_Shdr * hdr,const char * name,int shindex)8136 elfNN_aarch64_section_from_shdr (bfd *abfd,
8137 Elf_Internal_Shdr *hdr,
8138 const char *name, int shindex)
8139 {
8140 /* There ought to be a place to keep ELF backend specific flags, but
8141 at the moment there isn't one. We just keep track of the
8142 sections by their name, instead. Fortunately, the ABI gives
8143 names for all the AArch64 specific sections, so we will probably get
8144 away with this. */
8145 switch (hdr->sh_type)
8146 {
8147 case SHT_AARCH64_ATTRIBUTES:
8148 break;
8149
8150 default:
8151 return false;
8152 }
8153
8154 if (!_bfd_elf_make_section_from_shdr (abfd, hdr, name, shindex))
8155 return false;
8156
8157 return true;
8158 }
8159
8160 /* A structure used to record a list of sections, independently
8161 of the next and prev fields in the asection structure. */
8162 typedef struct section_list
8163 {
8164 asection *sec;
8165 struct section_list *next;
8166 struct section_list *prev;
8167 }
8168 section_list;
8169
8170 /* Unfortunately we need to keep a list of sections for which
8171 an _aarch64_elf_section_data structure has been allocated. This
8172 is because it is possible for functions like elfNN_aarch64_write_section
8173 to be called on a section which has had an elf_data_structure
8174 allocated for it (and so the used_by_bfd field is valid) but
8175 for which the AArch64 extended version of this structure - the
8176 _aarch64_elf_section_data structure - has not been allocated. */
8177 static section_list *sections_with_aarch64_elf_section_data = NULL;
8178
8179 static void
record_section_with_aarch64_elf_section_data(asection * sec)8180 record_section_with_aarch64_elf_section_data (asection *sec)
8181 {
8182 struct section_list *entry;
8183
8184 entry = bfd_malloc (sizeof (*entry));
8185 if (entry == NULL)
8186 return;
8187 entry->sec = sec;
8188 entry->next = sections_with_aarch64_elf_section_data;
8189 entry->prev = NULL;
8190 if (entry->next != NULL)
8191 entry->next->prev = entry;
8192 sections_with_aarch64_elf_section_data = entry;
8193 }
8194
8195 static struct section_list *
find_aarch64_elf_section_entry(asection * sec)8196 find_aarch64_elf_section_entry (asection *sec)
8197 {
8198 struct section_list *entry;
8199 static struct section_list *last_entry = NULL;
8200
8201 /* This is a short cut for the typical case where the sections are added
8202 to the sections_with_aarch64_elf_section_data list in forward order and
8203 then looked up here in backwards order. This makes a real difference
8204 to the ld-srec/sec64k.exp linker test. */
8205 entry = sections_with_aarch64_elf_section_data;
8206 if (last_entry != NULL)
8207 {
8208 if (last_entry->sec == sec)
8209 entry = last_entry;
8210 else if (last_entry->next != NULL && last_entry->next->sec == sec)
8211 entry = last_entry->next;
8212 }
8213
8214 for (; entry; entry = entry->next)
8215 if (entry->sec == sec)
8216 break;
8217
8218 if (entry)
8219 /* Record the entry prior to this one - it is the entry we are
8220 most likely to want to locate next time. Also this way if we
8221 have been called from
8222 unrecord_section_with_aarch64_elf_section_data () we will not
8223 be caching a pointer that is about to be freed. */
8224 last_entry = entry->prev;
8225
8226 return entry;
8227 }
8228
8229 static void
unrecord_section_with_aarch64_elf_section_data(asection * sec)8230 unrecord_section_with_aarch64_elf_section_data (asection *sec)
8231 {
8232 struct section_list *entry;
8233
8234 entry = find_aarch64_elf_section_entry (sec);
8235
8236 if (entry)
8237 {
8238 if (entry->prev != NULL)
8239 entry->prev->next = entry->next;
8240 if (entry->next != NULL)
8241 entry->next->prev = entry->prev;
8242 if (entry == sections_with_aarch64_elf_section_data)
8243 sections_with_aarch64_elf_section_data = entry->next;
8244 free (entry);
8245 }
8246 }
8247
8248
8249 typedef struct
8250 {
8251 void *finfo;
8252 struct bfd_link_info *info;
8253 asection *sec;
8254 int sec_shndx;
8255 int (*func) (void *, const char *, Elf_Internal_Sym *,
8256 asection *, struct elf_link_hash_entry *);
8257 } output_arch_syminfo;
8258
8259 enum map_symbol_type
8260 {
8261 AARCH64_MAP_INSN,
8262 AARCH64_MAP_DATA
8263 };
8264
8265
8266 /* Output a single mapping symbol. */
8267
8268 static bool
elfNN_aarch64_output_map_sym(output_arch_syminfo * osi,enum map_symbol_type type,bfd_vma offset)8269 elfNN_aarch64_output_map_sym (output_arch_syminfo *osi,
8270 enum map_symbol_type type, bfd_vma offset)
8271 {
8272 static const char *names[2] = { "$x", "$d" };
8273 Elf_Internal_Sym sym;
8274
8275 sym.st_value = (osi->sec->output_section->vma
8276 + osi->sec->output_offset + offset);
8277 sym.st_size = 0;
8278 sym.st_other = 0;
8279 sym.st_info = ELF_ST_INFO (STB_LOCAL, STT_NOTYPE);
8280 sym.st_shndx = osi->sec_shndx;
8281 return osi->func (osi->finfo, names[type], &sym, osi->sec, NULL) == 1;
8282 }
8283
8284 /* Output a single local symbol for a generated stub. */
8285
8286 static bool
elfNN_aarch64_output_stub_sym(output_arch_syminfo * osi,const char * name,bfd_vma offset,bfd_vma size)8287 elfNN_aarch64_output_stub_sym (output_arch_syminfo *osi, const char *name,
8288 bfd_vma offset, bfd_vma size)
8289 {
8290 Elf_Internal_Sym sym;
8291
8292 sym.st_value = (osi->sec->output_section->vma
8293 + osi->sec->output_offset + offset);
8294 sym.st_size = size;
8295 sym.st_other = 0;
8296 sym.st_info = ELF_ST_INFO (STB_LOCAL, STT_FUNC);
8297 sym.st_shndx = osi->sec_shndx;
8298 return osi->func (osi->finfo, name, &sym, osi->sec, NULL) == 1;
8299 }
8300
8301 static bool
aarch64_map_one_stub(struct bfd_hash_entry * gen_entry,void * in_arg)8302 aarch64_map_one_stub (struct bfd_hash_entry *gen_entry, void *in_arg)
8303 {
8304 struct elf_aarch64_stub_hash_entry *stub_entry;
8305 asection *stub_sec;
8306 bfd_vma addr;
8307 char *stub_name;
8308 output_arch_syminfo *osi;
8309
8310 /* Massage our args to the form they really have. */
8311 stub_entry = (struct elf_aarch64_stub_hash_entry *) gen_entry;
8312 osi = (output_arch_syminfo *) in_arg;
8313
8314 stub_sec = stub_entry->stub_sec;
8315
8316 /* Ensure this stub is attached to the current section being
8317 processed. */
8318 if (stub_sec != osi->sec)
8319 return true;
8320
8321 addr = (bfd_vma) stub_entry->stub_offset;
8322
8323 stub_name = stub_entry->output_name;
8324
8325 switch (stub_entry->stub_type)
8326 {
8327 case aarch64_stub_adrp_branch:
8328 if (!elfNN_aarch64_output_stub_sym (osi, stub_name, addr,
8329 sizeof (aarch64_adrp_branch_stub)))
8330 return false;
8331 if (!elfNN_aarch64_output_map_sym (osi, AARCH64_MAP_INSN, addr))
8332 return false;
8333 break;
8334 case aarch64_stub_long_branch:
8335 if (!elfNN_aarch64_output_stub_sym
8336 (osi, stub_name, addr, sizeof (aarch64_long_branch_stub)))
8337 return false;
8338 if (!elfNN_aarch64_output_map_sym (osi, AARCH64_MAP_INSN, addr))
8339 return false;
8340 if (!elfNN_aarch64_output_map_sym (osi, AARCH64_MAP_DATA, addr + 16))
8341 return false;
8342 break;
8343 case aarch64_stub_erratum_835769_veneer:
8344 if (!elfNN_aarch64_output_stub_sym (osi, stub_name, addr,
8345 sizeof (aarch64_erratum_835769_stub)))
8346 return false;
8347 if (!elfNN_aarch64_output_map_sym (osi, AARCH64_MAP_INSN, addr))
8348 return false;
8349 break;
8350 case aarch64_stub_erratum_843419_veneer:
8351 if (!elfNN_aarch64_output_stub_sym (osi, stub_name, addr,
8352 sizeof (aarch64_erratum_843419_stub)))
8353 return false;
8354 if (!elfNN_aarch64_output_map_sym (osi, AARCH64_MAP_INSN, addr))
8355 return false;
8356 break;
8357 case aarch64_stub_none:
8358 break;
8359
8360 default:
8361 abort ();
8362 }
8363
8364 return true;
8365 }
8366
8367 /* Output mapping symbols for linker generated sections. */
8368
8369 static bool
elfNN_aarch64_output_arch_local_syms(bfd * output_bfd,struct bfd_link_info * info,void * finfo,int (* func)(void *,const char *,Elf_Internal_Sym *,asection *,struct elf_link_hash_entry *))8370 elfNN_aarch64_output_arch_local_syms (bfd *output_bfd,
8371 struct bfd_link_info *info,
8372 void *finfo,
8373 int (*func) (void *, const char *,
8374 Elf_Internal_Sym *,
8375 asection *,
8376 struct elf_link_hash_entry
8377 *))
8378 {
8379 output_arch_syminfo osi;
8380 struct elf_aarch64_link_hash_table *htab;
8381
8382 htab = elf_aarch64_hash_table (info);
8383
8384 osi.finfo = finfo;
8385 osi.info = info;
8386 osi.func = func;
8387
8388 /* Long calls stubs. */
8389 if (htab->stub_bfd && htab->stub_bfd->sections)
8390 {
8391 asection *stub_sec;
8392
8393 for (stub_sec = htab->stub_bfd->sections;
8394 stub_sec != NULL; stub_sec = stub_sec->next)
8395 {
8396 /* Ignore non-stub sections. */
8397 if (!strstr (stub_sec->name, STUB_SUFFIX))
8398 continue;
8399
8400 osi.sec = stub_sec;
8401
8402 osi.sec_shndx = _bfd_elf_section_from_bfd_section
8403 (output_bfd, osi.sec->output_section);
8404
8405 /* The first instruction in a stub is always a branch. */
8406 if (!elfNN_aarch64_output_map_sym (&osi, AARCH64_MAP_INSN, 0))
8407 return false;
8408
8409 bfd_hash_traverse (&htab->stub_hash_table, aarch64_map_one_stub,
8410 &osi);
8411 }
8412 }
8413
8414 /* Finally, output mapping symbols for the PLT. */
8415 if (!htab->root.splt || htab->root.splt->size == 0)
8416 return true;
8417
8418 osi.sec_shndx = _bfd_elf_section_from_bfd_section
8419 (output_bfd, htab->root.splt->output_section);
8420 osi.sec = htab->root.splt;
8421
8422 elfNN_aarch64_output_map_sym (&osi, AARCH64_MAP_INSN, 0);
8423
8424 return true;
8425
8426 }
8427
8428 /* Allocate target specific section data. */
8429
8430 static bool
elfNN_aarch64_new_section_hook(bfd * abfd,asection * sec)8431 elfNN_aarch64_new_section_hook (bfd *abfd, asection *sec)
8432 {
8433 if (!sec->used_by_bfd)
8434 {
8435 _aarch64_elf_section_data *sdata;
8436 size_t amt = sizeof (*sdata);
8437
8438 sdata = bfd_zalloc (abfd, amt);
8439 if (sdata == NULL)
8440 return false;
8441 sec->used_by_bfd = sdata;
8442 }
8443
8444 record_section_with_aarch64_elf_section_data (sec);
8445
8446 return _bfd_elf_new_section_hook (abfd, sec);
8447 }
8448
8449
8450 static void
unrecord_section_via_map_over_sections(bfd * abfd ATTRIBUTE_UNUSED,asection * sec,void * ignore ATTRIBUTE_UNUSED)8451 unrecord_section_via_map_over_sections (bfd *abfd ATTRIBUTE_UNUSED,
8452 asection *sec,
8453 void *ignore ATTRIBUTE_UNUSED)
8454 {
8455 unrecord_section_with_aarch64_elf_section_data (sec);
8456 }
8457
8458 static bool
elfNN_aarch64_close_and_cleanup(bfd * abfd)8459 elfNN_aarch64_close_and_cleanup (bfd *abfd)
8460 {
8461 if (abfd->sections)
8462 bfd_map_over_sections (abfd,
8463 unrecord_section_via_map_over_sections, NULL);
8464
8465 return _bfd_elf_close_and_cleanup (abfd);
8466 }
8467
8468 static bool
elfNN_aarch64_bfd_free_cached_info(bfd * abfd)8469 elfNN_aarch64_bfd_free_cached_info (bfd *abfd)
8470 {
8471 if (abfd->sections)
8472 bfd_map_over_sections (abfd,
8473 unrecord_section_via_map_over_sections, NULL);
8474
8475 return _bfd_free_cached_info (abfd);
8476 }
8477
8478 /* Create dynamic sections. This is different from the ARM backend in that
8479 the got, plt, gotplt and their relocation sections are all created in the
8480 standard part of the bfd elf backend. */
8481
8482 static bool
elfNN_aarch64_create_dynamic_sections(bfd * dynobj,struct bfd_link_info * info)8483 elfNN_aarch64_create_dynamic_sections (bfd *dynobj,
8484 struct bfd_link_info *info)
8485 {
8486 /* We need to create .got section. */
8487 if (!aarch64_elf_create_got_section (dynobj, info))
8488 return false;
8489
8490 return _bfd_elf_create_dynamic_sections (dynobj, info);
8491 }
8492
8493
8494 /* Allocate space in .plt, .got and associated reloc sections for
8495 dynamic relocs. */
8496
8497 static bool
elfNN_aarch64_allocate_dynrelocs(struct elf_link_hash_entry * h,void * inf)8498 elfNN_aarch64_allocate_dynrelocs (struct elf_link_hash_entry *h, void *inf)
8499 {
8500 struct bfd_link_info *info;
8501 struct elf_aarch64_link_hash_table *htab;
8502 struct elf_aarch64_link_hash_entry *eh;
8503 struct elf_dyn_relocs *p;
8504
8505 /* An example of a bfd_link_hash_indirect symbol is versioned
8506 symbol. For example: __gxx_personality_v0(bfd_link_hash_indirect)
8507 -> __gxx_personality_v0(bfd_link_hash_defined)
8508
8509 There is no need to process bfd_link_hash_indirect symbols here
8510 because we will also be presented with the concrete instance of
8511 the symbol and elfNN_aarch64_copy_indirect_symbol () will have been
8512 called to copy all relevant data from the generic to the concrete
8513 symbol instance. */
8514 if (h->root.type == bfd_link_hash_indirect)
8515 return true;
8516
8517 if (h->root.type == bfd_link_hash_warning)
8518 h = (struct elf_link_hash_entry *) h->root.u.i.link;
8519
8520 info = (struct bfd_link_info *) inf;
8521 htab = elf_aarch64_hash_table (info);
8522
8523 /* Since STT_GNU_IFUNC symbol must go through PLT, we handle it
8524 here if it is defined and referenced in a non-shared object. */
8525 if (h->type == STT_GNU_IFUNC
8526 && h->def_regular)
8527 return true;
8528 else if (htab->root.dynamic_sections_created && h->plt.refcount > 0)
8529 {
8530 /* Make sure this symbol is output as a dynamic symbol.
8531 Undefined weak syms won't yet be marked as dynamic. */
8532 if (h->dynindx == -1 && !h->forced_local
8533 && h->root.type == bfd_link_hash_undefweak)
8534 {
8535 if (!bfd_elf_link_record_dynamic_symbol (info, h))
8536 return false;
8537 }
8538
8539 if (bfd_link_pic (info) || WILL_CALL_FINISH_DYNAMIC_SYMBOL (1, 0, h))
8540 {
8541 asection *s = htab->root.splt;
8542
8543 /* If this is the first .plt entry, make room for the special
8544 first entry. */
8545 if (s->size == 0)
8546 s->size += htab->plt_header_size;
8547
8548 h->plt.offset = s->size;
8549
8550 /* If this symbol is not defined in a regular file, and we are
8551 not generating a shared library, then set the symbol to this
8552 location in the .plt. This is required to make function
8553 pointers compare as equal between the normal executable and
8554 the shared library. */
8555 if (!bfd_link_pic (info) && !h->def_regular)
8556 {
8557 h->root.u.def.section = s;
8558 h->root.u.def.value = h->plt.offset;
8559 }
8560
8561 /* Make room for this entry. For now we only create the
8562 small model PLT entries. We later need to find a way
8563 of relaxing into these from the large model PLT entries. */
8564 s->size += htab->plt_entry_size;
8565
8566 /* We also need to make an entry in the .got.plt section, which
8567 will be placed in the .got section by the linker script. */
8568 htab->root.sgotplt->size += GOT_ENTRY_SIZE;
8569
8570 /* We also need to make an entry in the .rela.plt section. */
8571 htab->root.srelplt->size += RELOC_SIZE (htab);
8572
8573 /* We need to ensure that all GOT entries that serve the PLT
8574 are consecutive with the special GOT slots [0] [1] and
8575 [2]. Any addtional relocations, such as
8576 R_AARCH64_TLSDESC, must be placed after the PLT related
8577 entries. We abuse the reloc_count such that during
8578 sizing we adjust reloc_count to indicate the number of
8579 PLT related reserved entries. In subsequent phases when
8580 filling in the contents of the reloc entries, PLT related
8581 entries are placed by computing their PLT index (0
8582 .. reloc_count). While other none PLT relocs are placed
8583 at the slot indicated by reloc_count and reloc_count is
8584 updated. */
8585
8586 htab->root.srelplt->reloc_count++;
8587
8588 /* Mark the DSO in case R_<CLS>_JUMP_SLOT relocs against
8589 variant PCS symbols are present. */
8590 if (h->other & STO_AARCH64_VARIANT_PCS)
8591 htab->variant_pcs = 1;
8592
8593 }
8594 else
8595 {
8596 h->plt.offset = (bfd_vma) - 1;
8597 h->needs_plt = 0;
8598 }
8599 }
8600 else
8601 {
8602 h->plt.offset = (bfd_vma) - 1;
8603 h->needs_plt = 0;
8604 }
8605
8606 eh = (struct elf_aarch64_link_hash_entry *) h;
8607 eh->tlsdesc_got_jump_table_offset = (bfd_vma) - 1;
8608
8609 if (h->got.refcount > 0)
8610 {
8611 bool dyn;
8612 unsigned got_type = elf_aarch64_hash_entry (h)->got_type;
8613
8614 h->got.offset = (bfd_vma) - 1;
8615
8616 dyn = htab->root.dynamic_sections_created;
8617
8618 /* Make sure this symbol is output as a dynamic symbol.
8619 Undefined weak syms won't yet be marked as dynamic. */
8620 if (dyn && h->dynindx == -1 && !h->forced_local
8621 && h->root.type == bfd_link_hash_undefweak)
8622 {
8623 if (!bfd_elf_link_record_dynamic_symbol (info, h))
8624 return false;
8625 }
8626
8627 if (got_type == GOT_UNKNOWN)
8628 {
8629 }
8630 else if (got_type == GOT_NORMAL)
8631 {
8632 h->got.offset = htab->root.sgot->size;
8633 htab->root.sgot->size += GOT_ENTRY_SIZE;
8634 if ((ELF_ST_VISIBILITY (h->other) == STV_DEFAULT
8635 || h->root.type != bfd_link_hash_undefweak)
8636 && (bfd_link_pic (info)
8637 || WILL_CALL_FINISH_DYNAMIC_SYMBOL (dyn, 0, h))
8638 /* Undefined weak symbol in static PIE resolves to 0 without
8639 any dynamic relocations. */
8640 && !UNDEFWEAK_NO_DYNAMIC_RELOC (info, h))
8641 {
8642 htab->root.srelgot->size += RELOC_SIZE (htab);
8643 }
8644 }
8645 else
8646 {
8647 int indx;
8648 if (got_type & GOT_TLSDESC_GD)
8649 {
8650 eh->tlsdesc_got_jump_table_offset =
8651 (htab->root.sgotplt->size
8652 - aarch64_compute_jump_table_size (htab));
8653 htab->root.sgotplt->size += GOT_ENTRY_SIZE * 2;
8654 h->got.offset = (bfd_vma) - 2;
8655 }
8656
8657 if (got_type & GOT_TLS_GD)
8658 {
8659 h->got.offset = htab->root.sgot->size;
8660 htab->root.sgot->size += GOT_ENTRY_SIZE * 2;
8661 }
8662
8663 if (got_type & GOT_TLS_IE)
8664 {
8665 h->got.offset = htab->root.sgot->size;
8666 htab->root.sgot->size += GOT_ENTRY_SIZE;
8667 }
8668
8669 indx = h && h->dynindx != -1 ? h->dynindx : 0;
8670 if ((ELF_ST_VISIBILITY (h->other) == STV_DEFAULT
8671 || h->root.type != bfd_link_hash_undefweak)
8672 && (!bfd_link_executable (info)
8673 || indx != 0
8674 || WILL_CALL_FINISH_DYNAMIC_SYMBOL (dyn, 0, h)))
8675 {
8676 if (got_type & GOT_TLSDESC_GD)
8677 {
8678 htab->root.srelplt->size += RELOC_SIZE (htab);
8679 /* Note reloc_count not incremented here! We have
8680 already adjusted reloc_count for this relocation
8681 type. */
8682
8683 /* TLSDESC PLT is now needed, but not yet determined. */
8684 htab->root.tlsdesc_plt = (bfd_vma) - 1;
8685 }
8686
8687 if (got_type & GOT_TLS_GD)
8688 htab->root.srelgot->size += RELOC_SIZE (htab) * 2;
8689
8690 if (got_type & GOT_TLS_IE)
8691 htab->root.srelgot->size += RELOC_SIZE (htab);
8692 }
8693 }
8694 }
8695 else
8696 {
8697 h->got.offset = (bfd_vma) - 1;
8698 }
8699
8700 if (h->dyn_relocs == NULL)
8701 return true;
8702
8703 /* In the shared -Bsymbolic case, discard space allocated for
8704 dynamic pc-relative relocs against symbols which turn out to be
8705 defined in regular objects. For the normal shared case, discard
8706 space for pc-relative relocs that have become local due to symbol
8707 visibility changes. */
8708
8709 if (bfd_link_pic (info))
8710 {
8711 /* Relocs that use pc_count are those that appear on a call
8712 insn, or certain REL relocs that can generated via assembly.
8713 We want calls to protected symbols to resolve directly to the
8714 function rather than going via the plt. If people want
8715 function pointer comparisons to work as expected then they
8716 should avoid writing weird assembly. */
8717 if (SYMBOL_CALLS_LOCAL (info, h))
8718 {
8719 struct elf_dyn_relocs **pp;
8720
8721 for (pp = &h->dyn_relocs; (p = *pp) != NULL;)
8722 {
8723 p->count -= p->pc_count;
8724 p->pc_count = 0;
8725 if (p->count == 0)
8726 *pp = p->next;
8727 else
8728 pp = &p->next;
8729 }
8730 }
8731
8732 /* Also discard relocs on undefined weak syms with non-default
8733 visibility. */
8734 if (h->dyn_relocs != NULL && h->root.type == bfd_link_hash_undefweak)
8735 {
8736 if (ELF_ST_VISIBILITY (h->other) != STV_DEFAULT
8737 || UNDEFWEAK_NO_DYNAMIC_RELOC (info, h))
8738 h->dyn_relocs = NULL;
8739
8740 /* Make sure undefined weak symbols are output as a dynamic
8741 symbol in PIEs. */
8742 else if (h->dynindx == -1
8743 && !h->forced_local
8744 && h->root.type == bfd_link_hash_undefweak
8745 && !bfd_elf_link_record_dynamic_symbol (info, h))
8746 return false;
8747 }
8748
8749 }
8750 else if (ELIMINATE_COPY_RELOCS)
8751 {
8752 /* For the non-shared case, discard space for relocs against
8753 symbols which turn out to need copy relocs or are not
8754 dynamic. */
8755
8756 if (!h->non_got_ref
8757 && ((h->def_dynamic
8758 && !h->def_regular)
8759 || (htab->root.dynamic_sections_created
8760 && (h->root.type == bfd_link_hash_undefweak
8761 || h->root.type == bfd_link_hash_undefined))))
8762 {
8763 /* Make sure this symbol is output as a dynamic symbol.
8764 Undefined weak syms won't yet be marked as dynamic. */
8765 if (h->dynindx == -1
8766 && !h->forced_local
8767 && h->root.type == bfd_link_hash_undefweak
8768 && !bfd_elf_link_record_dynamic_symbol (info, h))
8769 return false;
8770
8771 /* If that succeeded, we know we'll be keeping all the
8772 relocs. */
8773 if (h->dynindx != -1)
8774 goto keep;
8775 }
8776
8777 h->dyn_relocs = NULL;
8778
8779 keep:;
8780 }
8781
8782 /* Finally, allocate space. */
8783 for (p = h->dyn_relocs; p != NULL; p = p->next)
8784 {
8785 asection *sreloc;
8786
8787 sreloc = elf_section_data (p->sec)->sreloc;
8788
8789 BFD_ASSERT (sreloc != NULL);
8790
8791 sreloc->size += p->count * RELOC_SIZE (htab);
8792 }
8793
8794 return true;
8795 }
8796
8797 /* Allocate space in .plt, .got and associated reloc sections for
8798 ifunc dynamic relocs. */
8799
8800 static bool
elfNN_aarch64_allocate_ifunc_dynrelocs(struct elf_link_hash_entry * h,void * inf)8801 elfNN_aarch64_allocate_ifunc_dynrelocs (struct elf_link_hash_entry *h,
8802 void *inf)
8803 {
8804 struct bfd_link_info *info;
8805 struct elf_aarch64_link_hash_table *htab;
8806
8807 /* An example of a bfd_link_hash_indirect symbol is versioned
8808 symbol. For example: __gxx_personality_v0(bfd_link_hash_indirect)
8809 -> __gxx_personality_v0(bfd_link_hash_defined)
8810
8811 There is no need to process bfd_link_hash_indirect symbols here
8812 because we will also be presented with the concrete instance of
8813 the symbol and elfNN_aarch64_copy_indirect_symbol () will have been
8814 called to copy all relevant data from the generic to the concrete
8815 symbol instance. */
8816 if (h->root.type == bfd_link_hash_indirect)
8817 return true;
8818
8819 if (h->root.type == bfd_link_hash_warning)
8820 h = (struct elf_link_hash_entry *) h->root.u.i.link;
8821
8822 info = (struct bfd_link_info *) inf;
8823 htab = elf_aarch64_hash_table (info);
8824
8825 /* Since STT_GNU_IFUNC symbol must go through PLT, we handle it
8826 here if it is defined and referenced in a non-shared object. */
8827 if (h->type == STT_GNU_IFUNC
8828 && h->def_regular)
8829 return _bfd_elf_allocate_ifunc_dyn_relocs (info, h,
8830 &h->dyn_relocs,
8831 htab->plt_entry_size,
8832 htab->plt_header_size,
8833 GOT_ENTRY_SIZE,
8834 false);
8835 return true;
8836 }
8837
8838 /* Allocate space in .plt, .got and associated reloc sections for
8839 local ifunc dynamic relocs. */
8840
8841 static int
elfNN_aarch64_allocate_local_ifunc_dynrelocs(void ** slot,void * inf)8842 elfNN_aarch64_allocate_local_ifunc_dynrelocs (void **slot, void *inf)
8843 {
8844 struct elf_link_hash_entry *h
8845 = (struct elf_link_hash_entry *) *slot;
8846
8847 if (h->type != STT_GNU_IFUNC
8848 || !h->def_regular
8849 || !h->ref_regular
8850 || !h->forced_local
8851 || h->root.type != bfd_link_hash_defined)
8852 abort ();
8853
8854 return elfNN_aarch64_allocate_ifunc_dynrelocs (h, inf);
8855 }
8856
8857 /* This is the most important function of all . Innocuosly named
8858 though ! */
8859
8860 static bool
elfNN_aarch64_size_dynamic_sections(bfd * output_bfd ATTRIBUTE_UNUSED,struct bfd_link_info * info)8861 elfNN_aarch64_size_dynamic_sections (bfd *output_bfd ATTRIBUTE_UNUSED,
8862 struct bfd_link_info *info)
8863 {
8864 struct elf_aarch64_link_hash_table *htab;
8865 bfd *dynobj;
8866 asection *s;
8867 bool relocs;
8868 bfd *ibfd;
8869
8870 htab = elf_aarch64_hash_table ((info));
8871 dynobj = htab->root.dynobj;
8872
8873 BFD_ASSERT (dynobj != NULL);
8874
8875 if (htab->root.dynamic_sections_created)
8876 {
8877 if (bfd_link_executable (info) && !info->nointerp)
8878 {
8879 s = bfd_get_linker_section (dynobj, ".interp");
8880 if (s == NULL)
8881 abort ();
8882 s->size = sizeof ELF_DYNAMIC_INTERPRETER;
8883 s->contents = (unsigned char *) ELF_DYNAMIC_INTERPRETER;
8884 }
8885 }
8886
8887 /* Set up .got offsets for local syms, and space for local dynamic
8888 relocs. */
8889 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link.next)
8890 {
8891 struct elf_aarch64_local_symbol *locals = NULL;
8892 Elf_Internal_Shdr *symtab_hdr;
8893 asection *srel;
8894 unsigned int i;
8895
8896 if (!is_aarch64_elf (ibfd))
8897 continue;
8898
8899 for (s = ibfd->sections; s != NULL; s = s->next)
8900 {
8901 struct elf_dyn_relocs *p;
8902
8903 for (p = (struct elf_dyn_relocs *)
8904 (elf_section_data (s)->local_dynrel); p != NULL; p = p->next)
8905 {
8906 if (!bfd_is_abs_section (p->sec)
8907 && bfd_is_abs_section (p->sec->output_section))
8908 {
8909 /* Input section has been discarded, either because
8910 it is a copy of a linkonce section or due to
8911 linker script /DISCARD/, so we'll be discarding
8912 the relocs too. */
8913 }
8914 else if (p->count != 0)
8915 {
8916 srel = elf_section_data (p->sec)->sreloc;
8917 srel->size += p->count * RELOC_SIZE (htab);
8918 if ((p->sec->output_section->flags & SEC_READONLY) != 0)
8919 info->flags |= DF_TEXTREL;
8920 }
8921 }
8922 }
8923
8924 locals = elf_aarch64_locals (ibfd);
8925 if (!locals)
8926 continue;
8927
8928 symtab_hdr = &elf_symtab_hdr (ibfd);
8929 srel = htab->root.srelgot;
8930 for (i = 0; i < symtab_hdr->sh_info; i++)
8931 {
8932 locals[i].got_offset = (bfd_vma) - 1;
8933 locals[i].tlsdesc_got_jump_table_offset = (bfd_vma) - 1;
8934 if (locals[i].got_refcount > 0)
8935 {
8936 unsigned got_type = locals[i].got_type;
8937 if (got_type & GOT_TLSDESC_GD)
8938 {
8939 locals[i].tlsdesc_got_jump_table_offset =
8940 (htab->root.sgotplt->size
8941 - aarch64_compute_jump_table_size (htab));
8942 htab->root.sgotplt->size += GOT_ENTRY_SIZE * 2;
8943 locals[i].got_offset = (bfd_vma) - 2;
8944 }
8945
8946 if (got_type & GOT_TLS_GD)
8947 {
8948 locals[i].got_offset = htab->root.sgot->size;
8949 htab->root.sgot->size += GOT_ENTRY_SIZE * 2;
8950 }
8951
8952 if (got_type & GOT_TLS_IE
8953 || got_type & GOT_NORMAL)
8954 {
8955 locals[i].got_offset = htab->root.sgot->size;
8956 htab->root.sgot->size += GOT_ENTRY_SIZE;
8957 }
8958
8959 if (got_type == GOT_UNKNOWN)
8960 {
8961 }
8962
8963 if (bfd_link_pic (info))
8964 {
8965 if (got_type & GOT_TLSDESC_GD)
8966 {
8967 htab->root.srelplt->size += RELOC_SIZE (htab);
8968 /* Note RELOC_COUNT not incremented here! */
8969 htab->root.tlsdesc_plt = (bfd_vma) - 1;
8970 }
8971
8972 if (got_type & GOT_TLS_GD)
8973 htab->root.srelgot->size += RELOC_SIZE (htab) * 2;
8974
8975 if (got_type & GOT_TLS_IE
8976 || got_type & GOT_NORMAL)
8977 htab->root.srelgot->size += RELOC_SIZE (htab);
8978 }
8979 }
8980 else
8981 {
8982 locals[i].got_refcount = (bfd_vma) - 1;
8983 }
8984 }
8985 }
8986
8987
8988 /* Allocate global sym .plt and .got entries, and space for global
8989 sym dynamic relocs. */
8990 elf_link_hash_traverse (&htab->root, elfNN_aarch64_allocate_dynrelocs,
8991 info);
8992
8993 /* Allocate global ifunc sym .plt and .got entries, and space for global
8994 ifunc sym dynamic relocs. */
8995 elf_link_hash_traverse (&htab->root, elfNN_aarch64_allocate_ifunc_dynrelocs,
8996 info);
8997
8998 /* Allocate .plt and .got entries, and space for local ifunc symbols. */
8999 htab_traverse (htab->loc_hash_table,
9000 elfNN_aarch64_allocate_local_ifunc_dynrelocs,
9001 info);
9002
9003 /* For every jump slot reserved in the sgotplt, reloc_count is
9004 incremented. However, when we reserve space for TLS descriptors,
9005 it's not incremented, so in order to compute the space reserved
9006 for them, it suffices to multiply the reloc count by the jump
9007 slot size. */
9008
9009 if (htab->root.srelplt)
9010 htab->sgotplt_jump_table_size = aarch64_compute_jump_table_size (htab);
9011
9012 if (htab->root.tlsdesc_plt)
9013 {
9014 if (htab->root.splt->size == 0)
9015 htab->root.splt->size += htab->plt_header_size;
9016
9017 /* If we're not using lazy TLS relocations, don't generate the
9018 GOT and PLT entry required. */
9019 if ((info->flags & DF_BIND_NOW))
9020 htab->root.tlsdesc_plt = 0;
9021 else
9022 {
9023 htab->root.tlsdesc_plt = htab->root.splt->size;
9024 htab->root.splt->size += htab->tlsdesc_plt_entry_size;
9025
9026 htab->root.tlsdesc_got = htab->root.sgot->size;
9027 htab->root.sgot->size += GOT_ENTRY_SIZE;
9028 }
9029 }
9030
9031 /* Init mapping symbols information to use later to distingush between
9032 code and data while scanning for errata. */
9033 if (htab->fix_erratum_835769 || htab->fix_erratum_843419)
9034 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link.next)
9035 {
9036 if (!is_aarch64_elf (ibfd))
9037 continue;
9038 bfd_elfNN_aarch64_init_maps (ibfd);
9039 }
9040
9041 /* We now have determined the sizes of the various dynamic sections.
9042 Allocate memory for them. */
9043 relocs = false;
9044 for (s = dynobj->sections; s != NULL; s = s->next)
9045 {
9046 if ((s->flags & SEC_LINKER_CREATED) == 0)
9047 continue;
9048
9049 if (s == htab->root.splt
9050 || s == htab->root.sgot
9051 || s == htab->root.sgotplt
9052 || s == htab->root.iplt
9053 || s == htab->root.igotplt
9054 || s == htab->root.sdynbss
9055 || s == htab->root.sdynrelro)
9056 {
9057 /* Strip this section if we don't need it; see the
9058 comment below. */
9059 }
9060 else if (startswith (bfd_section_name (s), ".rela"))
9061 {
9062 if (s->size != 0 && s != htab->root.srelplt)
9063 relocs = true;
9064
9065 /* We use the reloc_count field as a counter if we need
9066 to copy relocs into the output file. */
9067 if (s != htab->root.srelplt)
9068 s->reloc_count = 0;
9069 }
9070 else
9071 {
9072 /* It's not one of our sections, so don't allocate space. */
9073 continue;
9074 }
9075
9076 if (s->size == 0)
9077 {
9078 /* If we don't need this section, strip it from the
9079 output file. This is mostly to handle .rela.bss and
9080 .rela.plt. We must create both sections in
9081 create_dynamic_sections, because they must be created
9082 before the linker maps input sections to output
9083 sections. The linker does that before
9084 adjust_dynamic_symbol is called, and it is that
9085 function which decides whether anything needs to go
9086 into these sections. */
9087 s->flags |= SEC_EXCLUDE;
9088 continue;
9089 }
9090
9091 if ((s->flags & SEC_HAS_CONTENTS) == 0)
9092 continue;
9093
9094 /* Allocate memory for the section contents. We use bfd_zalloc
9095 here in case unused entries are not reclaimed before the
9096 section's contents are written out. This should not happen,
9097 but this way if it does, we get a R_AARCH64_NONE reloc instead
9098 of garbage. */
9099 s->contents = (bfd_byte *) bfd_zalloc (dynobj, s->size);
9100 if (s->contents == NULL)
9101 return false;
9102 }
9103
9104 if (htab->root.dynamic_sections_created)
9105 {
9106 /* Add some entries to the .dynamic section. We fill in the
9107 values later, in elfNN_aarch64_finish_dynamic_sections, but we
9108 must add the entries now so that we get the correct size for
9109 the .dynamic section. The DT_DEBUG entry is filled in by the
9110 dynamic linker and used by the debugger. */
9111 #define add_dynamic_entry(TAG, VAL) \
9112 _bfd_elf_add_dynamic_entry (info, TAG, VAL)
9113
9114 if (!_bfd_elf_add_dynamic_tags (output_bfd, info, relocs))
9115 return false;
9116
9117 if (htab->root.splt->size != 0)
9118 {
9119 if (htab->variant_pcs
9120 && !add_dynamic_entry (DT_AARCH64_VARIANT_PCS, 0))
9121 return false;
9122
9123 if ((elf_aarch64_tdata (output_bfd)->plt_type == PLT_BTI_PAC)
9124 && (!add_dynamic_entry (DT_AARCH64_BTI_PLT, 0)
9125 || !add_dynamic_entry (DT_AARCH64_PAC_PLT, 0)))
9126 return false;
9127
9128 else if ((elf_aarch64_tdata (output_bfd)->plt_type == PLT_BTI)
9129 && !add_dynamic_entry (DT_AARCH64_BTI_PLT, 0))
9130 return false;
9131
9132 else if ((elf_aarch64_tdata (output_bfd)->plt_type == PLT_PAC)
9133 && !add_dynamic_entry (DT_AARCH64_PAC_PLT, 0))
9134 return false;
9135 }
9136 }
9137 #undef add_dynamic_entry
9138
9139 return true;
9140 }
9141
9142 static inline void
elf_aarch64_update_plt_entry(bfd * output_bfd,bfd_reloc_code_real_type r_type,bfd_byte * plt_entry,bfd_vma value)9143 elf_aarch64_update_plt_entry (bfd *output_bfd,
9144 bfd_reloc_code_real_type r_type,
9145 bfd_byte *plt_entry, bfd_vma value)
9146 {
9147 reloc_howto_type *howto = elfNN_aarch64_howto_from_bfd_reloc (r_type);
9148
9149 /* FIXME: We should check the return value from this function call. */
9150 (void) _bfd_aarch64_elf_put_addend (output_bfd, plt_entry, r_type, howto, value);
9151 }
9152
9153 static void
elfNN_aarch64_create_small_pltn_entry(struct elf_link_hash_entry * h,struct elf_aarch64_link_hash_table * htab,bfd * output_bfd,struct bfd_link_info * info)9154 elfNN_aarch64_create_small_pltn_entry (struct elf_link_hash_entry *h,
9155 struct elf_aarch64_link_hash_table
9156 *htab, bfd *output_bfd,
9157 struct bfd_link_info *info)
9158 {
9159 bfd_byte *plt_entry;
9160 bfd_vma plt_index;
9161 bfd_vma got_offset;
9162 bfd_vma gotplt_entry_address;
9163 bfd_vma plt_entry_address;
9164 Elf_Internal_Rela rela;
9165 bfd_byte *loc;
9166 asection *plt, *gotplt, *relplt;
9167
9168 /* When building a static executable, use .iplt, .igot.plt and
9169 .rela.iplt sections for STT_GNU_IFUNC symbols. */
9170 if (htab->root.splt != NULL)
9171 {
9172 plt = htab->root.splt;
9173 gotplt = htab->root.sgotplt;
9174 relplt = htab->root.srelplt;
9175 }
9176 else
9177 {
9178 plt = htab->root.iplt;
9179 gotplt = htab->root.igotplt;
9180 relplt = htab->root.irelplt;
9181 }
9182
9183 /* Get the index in the procedure linkage table which
9184 corresponds to this symbol. This is the index of this symbol
9185 in all the symbols for which we are making plt entries. The
9186 first entry in the procedure linkage table is reserved.
9187
9188 Get the offset into the .got table of the entry that
9189 corresponds to this function. Each .got entry is GOT_ENTRY_SIZE
9190 bytes. The first three are reserved for the dynamic linker.
9191
9192 For static executables, we don't reserve anything. */
9193
9194 if (plt == htab->root.splt)
9195 {
9196 plt_index = (h->plt.offset - htab->plt_header_size) / htab->plt_entry_size;
9197 got_offset = (plt_index + 3) * GOT_ENTRY_SIZE;
9198 }
9199 else
9200 {
9201 plt_index = h->plt.offset / htab->plt_entry_size;
9202 got_offset = plt_index * GOT_ENTRY_SIZE;
9203 }
9204
9205 plt_entry = plt->contents + h->plt.offset;
9206 plt_entry_address = plt->output_section->vma
9207 + plt->output_offset + h->plt.offset;
9208 gotplt_entry_address = gotplt->output_section->vma +
9209 gotplt->output_offset + got_offset;
9210
9211 /* Copy in the boiler-plate for the PLTn entry. */
9212 memcpy (plt_entry, htab->plt_entry, htab->plt_entry_size);
9213
9214 /* First instruction in BTI enabled PLT stub is a BTI
9215 instruction so skip it. */
9216 if (elf_aarch64_tdata (output_bfd)->plt_type & PLT_BTI
9217 && elf_elfheader (output_bfd)->e_type == ET_EXEC)
9218 plt_entry = plt_entry + 4;
9219
9220 /* Fill in the top 21 bits for this: ADRP x16, PLT_GOT + n * 8.
9221 ADRP: ((PG(S+A)-PG(P)) >> 12) & 0x1fffff */
9222 elf_aarch64_update_plt_entry (output_bfd, BFD_RELOC_AARCH64_ADR_HI21_PCREL,
9223 plt_entry,
9224 PG (gotplt_entry_address) -
9225 PG (plt_entry_address));
9226
9227 /* Fill in the lo12 bits for the load from the pltgot. */
9228 elf_aarch64_update_plt_entry (output_bfd, BFD_RELOC_AARCH64_LDSTNN_LO12,
9229 plt_entry + 4,
9230 PG_OFFSET (gotplt_entry_address));
9231
9232 /* Fill in the lo12 bits for the add from the pltgot entry. */
9233 elf_aarch64_update_plt_entry (output_bfd, BFD_RELOC_AARCH64_ADD_LO12,
9234 plt_entry + 8,
9235 PG_OFFSET (gotplt_entry_address));
9236
9237 /* All the GOTPLT Entries are essentially initialized to PLT0. */
9238 bfd_put_NN (output_bfd,
9239 plt->output_section->vma + plt->output_offset,
9240 gotplt->contents + got_offset);
9241
9242 rela.r_offset = gotplt_entry_address;
9243
9244 if (h->dynindx == -1
9245 || ((bfd_link_executable (info)
9246 || ELF_ST_VISIBILITY (h->other) != STV_DEFAULT)
9247 && h->def_regular
9248 && h->type == STT_GNU_IFUNC))
9249 {
9250 /* If an STT_GNU_IFUNC symbol is locally defined, generate
9251 R_AARCH64_IRELATIVE instead of R_AARCH64_JUMP_SLOT. */
9252 rela.r_info = ELFNN_R_INFO (0, AARCH64_R (IRELATIVE));
9253 rela.r_addend = (h->root.u.def.value
9254 + h->root.u.def.section->output_section->vma
9255 + h->root.u.def.section->output_offset);
9256 }
9257 else
9258 {
9259 /* Fill in the entry in the .rela.plt section. */
9260 rela.r_info = ELFNN_R_INFO (h->dynindx, AARCH64_R (JUMP_SLOT));
9261 rela.r_addend = 0;
9262 }
9263
9264 /* Compute the relocation entry to used based on PLT index and do
9265 not adjust reloc_count. The reloc_count has already been adjusted
9266 to account for this entry. */
9267 loc = relplt->contents + plt_index * RELOC_SIZE (htab);
9268 bfd_elfNN_swap_reloca_out (output_bfd, &rela, loc);
9269 }
9270
9271 /* Size sections even though they're not dynamic. We use it to setup
9272 _TLS_MODULE_BASE_, if needed. */
9273
9274 static bool
elfNN_aarch64_always_size_sections(bfd * output_bfd,struct bfd_link_info * info)9275 elfNN_aarch64_always_size_sections (bfd *output_bfd,
9276 struct bfd_link_info *info)
9277 {
9278 asection *tls_sec;
9279
9280 if (bfd_link_relocatable (info))
9281 return true;
9282
9283 tls_sec = elf_hash_table (info)->tls_sec;
9284
9285 if (tls_sec)
9286 {
9287 struct elf_link_hash_entry *tlsbase;
9288
9289 tlsbase = elf_link_hash_lookup (elf_hash_table (info),
9290 "_TLS_MODULE_BASE_", true, true, false);
9291
9292 if (tlsbase)
9293 {
9294 struct bfd_link_hash_entry *h = NULL;
9295 const struct elf_backend_data *bed =
9296 get_elf_backend_data (output_bfd);
9297
9298 if (!(_bfd_generic_link_add_one_symbol
9299 (info, output_bfd, "_TLS_MODULE_BASE_", BSF_LOCAL,
9300 tls_sec, 0, NULL, false, bed->collect, &h)))
9301 return false;
9302
9303 tlsbase->type = STT_TLS;
9304 tlsbase = (struct elf_link_hash_entry *) h;
9305 tlsbase->def_regular = 1;
9306 tlsbase->other = STV_HIDDEN;
9307 (*bed->elf_backend_hide_symbol) (info, tlsbase, true);
9308 }
9309 }
9310
9311 return true;
9312 }
9313
9314 /* Finish up dynamic symbol handling. We set the contents of various
9315 dynamic sections here. */
9316
9317 static bool
elfNN_aarch64_finish_dynamic_symbol(bfd * output_bfd,struct bfd_link_info * info,struct elf_link_hash_entry * h,Elf_Internal_Sym * sym)9318 elfNN_aarch64_finish_dynamic_symbol (bfd *output_bfd,
9319 struct bfd_link_info *info,
9320 struct elf_link_hash_entry *h,
9321 Elf_Internal_Sym *sym)
9322 {
9323 struct elf_aarch64_link_hash_table *htab;
9324 htab = elf_aarch64_hash_table (info);
9325
9326 if (h->plt.offset != (bfd_vma) - 1)
9327 {
9328 asection *plt, *gotplt, *relplt;
9329
9330 /* This symbol has an entry in the procedure linkage table. Set
9331 it up. */
9332
9333 /* When building a static executable, use .iplt, .igot.plt and
9334 .rela.iplt sections for STT_GNU_IFUNC symbols. */
9335 if (htab->root.splt != NULL)
9336 {
9337 plt = htab->root.splt;
9338 gotplt = htab->root.sgotplt;
9339 relplt = htab->root.srelplt;
9340 }
9341 else
9342 {
9343 plt = htab->root.iplt;
9344 gotplt = htab->root.igotplt;
9345 relplt = htab->root.irelplt;
9346 }
9347
9348 /* This symbol has an entry in the procedure linkage table. Set
9349 it up. */
9350 if ((h->dynindx == -1
9351 && !((h->forced_local || bfd_link_executable (info))
9352 && h->def_regular
9353 && h->type == STT_GNU_IFUNC))
9354 || plt == NULL
9355 || gotplt == NULL
9356 || relplt == NULL)
9357 return false;
9358
9359 elfNN_aarch64_create_small_pltn_entry (h, htab, output_bfd, info);
9360 if (!h->def_regular)
9361 {
9362 /* Mark the symbol as undefined, rather than as defined in
9363 the .plt section. */
9364 sym->st_shndx = SHN_UNDEF;
9365 /* If the symbol is weak we need to clear the value.
9366 Otherwise, the PLT entry would provide a definition for
9367 the symbol even if the symbol wasn't defined anywhere,
9368 and so the symbol would never be NULL. Leave the value if
9369 there were any relocations where pointer equality matters
9370 (this is a clue for the dynamic linker, to make function
9371 pointer comparisons work between an application and shared
9372 library). */
9373 if (!h->ref_regular_nonweak || !h->pointer_equality_needed)
9374 sym->st_value = 0;
9375 }
9376 }
9377
9378 if (h->got.offset != (bfd_vma) - 1
9379 && elf_aarch64_hash_entry (h)->got_type == GOT_NORMAL
9380 /* Undefined weak symbol in static PIE resolves to 0 without
9381 any dynamic relocations. */
9382 && !UNDEFWEAK_NO_DYNAMIC_RELOC (info, h))
9383 {
9384 Elf_Internal_Rela rela;
9385 bfd_byte *loc;
9386
9387 /* This symbol has an entry in the global offset table. Set it
9388 up. */
9389 if (htab->root.sgot == NULL || htab->root.srelgot == NULL)
9390 abort ();
9391
9392 rela.r_offset = (htab->root.sgot->output_section->vma
9393 + htab->root.sgot->output_offset
9394 + (h->got.offset & ~(bfd_vma) 1));
9395
9396 if (h->def_regular
9397 && h->type == STT_GNU_IFUNC)
9398 {
9399 if (bfd_link_pic (info))
9400 {
9401 /* Generate R_AARCH64_GLOB_DAT. */
9402 goto do_glob_dat;
9403 }
9404 else
9405 {
9406 asection *plt;
9407
9408 if (!h->pointer_equality_needed)
9409 abort ();
9410
9411 /* For non-shared object, we can't use .got.plt, which
9412 contains the real function address if we need pointer
9413 equality. We load the GOT entry with the PLT entry. */
9414 plt = htab->root.splt ? htab->root.splt : htab->root.iplt;
9415 bfd_put_NN (output_bfd, (plt->output_section->vma
9416 + plt->output_offset
9417 + h->plt.offset),
9418 htab->root.sgot->contents
9419 + (h->got.offset & ~(bfd_vma) 1));
9420 return true;
9421 }
9422 }
9423 else if (bfd_link_pic (info) && SYMBOL_REFERENCES_LOCAL (info, h))
9424 {
9425 if (!(h->def_regular || ELF_COMMON_DEF_P (h)))
9426 return false;
9427
9428 BFD_ASSERT ((h->got.offset & 1) != 0);
9429 rela.r_info = ELFNN_R_INFO (0, AARCH64_R (RELATIVE));
9430 rela.r_addend = (h->root.u.def.value
9431 + h->root.u.def.section->output_section->vma
9432 + h->root.u.def.section->output_offset);
9433 }
9434 else if (!bfd_link_relocatable (info)
9435 && (info->export_dynamic || info->dynamic)
9436 && SYMBOL_REFERENCES_LOCAL (info, h))
9437 {
9438 BFD_ASSERT ((h->got.offset & 1) != 0);
9439 rela.r_info = ELFNN_R_INFO (h->dynindx, AARCH64_R (GLOB_DAT));
9440 rela.r_addend = 0;
9441 }
9442 else
9443 {
9444 do_glob_dat:
9445 BFD_ASSERT ((h->got.offset & 1) == 0);
9446 bfd_put_NN (output_bfd, (bfd_vma) 0,
9447 htab->root.sgot->contents + h->got.offset);
9448 rela.r_info = ELFNN_R_INFO (h->dynindx, AARCH64_R (GLOB_DAT));
9449 rela.r_addend = 0;
9450 }
9451
9452 loc = htab->root.srelgot->contents;
9453 loc += htab->root.srelgot->reloc_count++ * RELOC_SIZE (htab);
9454 bfd_elfNN_swap_reloca_out (output_bfd, &rela, loc);
9455 }
9456
9457 if (h->needs_copy)
9458 {
9459 Elf_Internal_Rela rela;
9460 asection *s;
9461 bfd_byte *loc;
9462
9463 /* This symbol needs a copy reloc. Set it up. */
9464 if (h->dynindx == -1
9465 || (h->root.type != bfd_link_hash_defined
9466 && h->root.type != bfd_link_hash_defweak)
9467 || htab->root.srelbss == NULL)
9468 abort ();
9469
9470 rela.r_offset = (h->root.u.def.value
9471 + h->root.u.def.section->output_section->vma
9472 + h->root.u.def.section->output_offset);
9473 rela.r_info = ELFNN_R_INFO (h->dynindx, AARCH64_R (COPY));
9474 rela.r_addend = 0;
9475 if (h->root.u.def.section == htab->root.sdynrelro)
9476 s = htab->root.sreldynrelro;
9477 else
9478 s = htab->root.srelbss;
9479 loc = s->contents + s->reloc_count++ * RELOC_SIZE (htab);
9480 bfd_elfNN_swap_reloca_out (output_bfd, &rela, loc);
9481 }
9482
9483 /* Mark _DYNAMIC and _GLOBAL_OFFSET_TABLE_ as absolute. SYM may
9484 be NULL for local symbols. */
9485 if (sym != NULL
9486 && (h == elf_hash_table (info)->hdynamic
9487 || h == elf_hash_table (info)->hgot))
9488 sym->st_shndx = SHN_ABS;
9489
9490 return true;
9491 }
9492
9493 /* Finish up local dynamic symbol handling. We set the contents of
9494 various dynamic sections here. */
9495
9496 static int
elfNN_aarch64_finish_local_dynamic_symbol(void ** slot,void * inf)9497 elfNN_aarch64_finish_local_dynamic_symbol (void **slot, void *inf)
9498 {
9499 struct elf_link_hash_entry *h
9500 = (struct elf_link_hash_entry *) *slot;
9501 struct bfd_link_info *info
9502 = (struct bfd_link_info *) inf;
9503
9504 return elfNN_aarch64_finish_dynamic_symbol (info->output_bfd,
9505 info, h, NULL);
9506 }
9507
9508 static void
elfNN_aarch64_init_small_plt0_entry(bfd * output_bfd ATTRIBUTE_UNUSED,struct elf_aarch64_link_hash_table * htab)9509 elfNN_aarch64_init_small_plt0_entry (bfd *output_bfd ATTRIBUTE_UNUSED,
9510 struct elf_aarch64_link_hash_table
9511 *htab)
9512 {
9513 /* Fill in PLT0. Fixme:RR Note this doesn't distinguish between
9514 small and large plts and at the minute just generates
9515 the small PLT. */
9516
9517 /* PLT0 of the small PLT looks like this in ELF64 -
9518 stp x16, x30, [sp, #-16]! // Save the reloc and lr on stack.
9519 adrp x16, PLT_GOT + 16 // Get the page base of the GOTPLT
9520 ldr x17, [x16, #:lo12:PLT_GOT+16] // Load the address of the
9521 // symbol resolver
9522 add x16, x16, #:lo12:PLT_GOT+16 // Load the lo12 bits of the
9523 // GOTPLT entry for this.
9524 br x17
9525 PLT0 will be slightly different in ELF32 due to different got entry
9526 size. */
9527 bfd_vma plt_got_2nd_ent; /* Address of GOT[2]. */
9528 bfd_vma plt_base;
9529
9530
9531 memcpy (htab->root.splt->contents, htab->plt0_entry,
9532 htab->plt_header_size);
9533
9534 /* PR 26312: Explicitly set the sh_entsize to 0 so that
9535 consumers do not think that the section contains fixed
9536 sized objects. */
9537 elf_section_data (htab->root.splt->output_section)->this_hdr.sh_entsize = 0;
9538
9539 plt_got_2nd_ent = (htab->root.sgotplt->output_section->vma
9540 + htab->root.sgotplt->output_offset
9541 + GOT_ENTRY_SIZE * 2);
9542
9543 plt_base = htab->root.splt->output_section->vma +
9544 htab->root.splt->output_offset;
9545
9546 /* First instruction in BTI enabled PLT stub is a BTI
9547 instruction so skip it. */
9548 bfd_byte *plt0_entry = htab->root.splt->contents;
9549 if (elf_aarch64_tdata (output_bfd)->plt_type & PLT_BTI)
9550 plt0_entry = plt0_entry + 4;
9551
9552 /* Fill in the top 21 bits for this: ADRP x16, PLT_GOT + n * 8.
9553 ADRP: ((PG(S+A)-PG(P)) >> 12) & 0x1fffff */
9554 elf_aarch64_update_plt_entry (output_bfd, BFD_RELOC_AARCH64_ADR_HI21_PCREL,
9555 plt0_entry + 4,
9556 PG (plt_got_2nd_ent) - PG (plt_base + 4));
9557
9558 elf_aarch64_update_plt_entry (output_bfd, BFD_RELOC_AARCH64_LDSTNN_LO12,
9559 plt0_entry + 8,
9560 PG_OFFSET (plt_got_2nd_ent));
9561
9562 elf_aarch64_update_plt_entry (output_bfd, BFD_RELOC_AARCH64_ADD_LO12,
9563 plt0_entry + 12,
9564 PG_OFFSET (plt_got_2nd_ent));
9565 }
9566
9567 static bool
elfNN_aarch64_finish_dynamic_sections(bfd * output_bfd,struct bfd_link_info * info)9568 elfNN_aarch64_finish_dynamic_sections (bfd *output_bfd,
9569 struct bfd_link_info *info)
9570 {
9571 struct elf_aarch64_link_hash_table *htab;
9572 bfd *dynobj;
9573 asection *sdyn;
9574
9575 htab = elf_aarch64_hash_table (info);
9576 dynobj = htab->root.dynobj;
9577 sdyn = bfd_get_linker_section (dynobj, ".dynamic");
9578
9579 if (htab->root.dynamic_sections_created)
9580 {
9581 ElfNN_External_Dyn *dyncon, *dynconend;
9582
9583 if (sdyn == NULL || htab->root.sgot == NULL)
9584 abort ();
9585
9586 dyncon = (ElfNN_External_Dyn *) sdyn->contents;
9587 dynconend = (ElfNN_External_Dyn *) (sdyn->contents + sdyn->size);
9588 for (; dyncon < dynconend; dyncon++)
9589 {
9590 Elf_Internal_Dyn dyn;
9591 asection *s;
9592
9593 bfd_elfNN_swap_dyn_in (dynobj, dyncon, &dyn);
9594
9595 switch (dyn.d_tag)
9596 {
9597 default:
9598 continue;
9599
9600 case DT_PLTGOT:
9601 s = htab->root.sgotplt;
9602 dyn.d_un.d_ptr = s->output_section->vma + s->output_offset;
9603 break;
9604
9605 case DT_JMPREL:
9606 s = htab->root.srelplt;
9607 dyn.d_un.d_ptr = s->output_section->vma + s->output_offset;
9608 break;
9609
9610 case DT_PLTRELSZ:
9611 s = htab->root.srelplt;
9612 dyn.d_un.d_val = s->size;
9613 break;
9614
9615 case DT_TLSDESC_PLT:
9616 s = htab->root.splt;
9617 dyn.d_un.d_ptr = s->output_section->vma + s->output_offset
9618 + htab->root.tlsdesc_plt;
9619 break;
9620
9621 case DT_TLSDESC_GOT:
9622 s = htab->root.sgot;
9623 BFD_ASSERT (htab->root.tlsdesc_got != (bfd_vma)-1);
9624 dyn.d_un.d_ptr = s->output_section->vma + s->output_offset
9625 + htab->root.tlsdesc_got;
9626 break;
9627 }
9628
9629 bfd_elfNN_swap_dyn_out (output_bfd, &dyn, dyncon);
9630 }
9631
9632 }
9633
9634 /* Fill in the special first entry in the procedure linkage table. */
9635 if (htab->root.splt && htab->root.splt->size > 0)
9636 {
9637 elfNN_aarch64_init_small_plt0_entry (output_bfd, htab);
9638
9639 if (htab->root.tlsdesc_plt && !(info->flags & DF_BIND_NOW))
9640 {
9641 BFD_ASSERT (htab->root.tlsdesc_got != (bfd_vma)-1);
9642 bfd_put_NN (output_bfd, (bfd_vma) 0,
9643 htab->root.sgot->contents + htab->root.tlsdesc_got);
9644
9645 const bfd_byte *entry = elfNN_aarch64_tlsdesc_small_plt_entry;
9646 htab->tlsdesc_plt_entry_size = PLT_TLSDESC_ENTRY_SIZE;
9647
9648 aarch64_plt_type type = elf_aarch64_tdata (output_bfd)->plt_type;
9649 if (type == PLT_BTI || type == PLT_BTI_PAC)
9650 {
9651 entry = elfNN_aarch64_tlsdesc_small_plt_bti_entry;
9652 }
9653
9654 memcpy (htab->root.splt->contents + htab->root.tlsdesc_plt,
9655 entry, htab->tlsdesc_plt_entry_size);
9656
9657 {
9658 bfd_vma adrp1_addr =
9659 htab->root.splt->output_section->vma
9660 + htab->root.splt->output_offset
9661 + htab->root.tlsdesc_plt + 4;
9662
9663 bfd_vma adrp2_addr = adrp1_addr + 4;
9664
9665 bfd_vma got_addr =
9666 htab->root.sgot->output_section->vma
9667 + htab->root.sgot->output_offset;
9668
9669 bfd_vma pltgot_addr =
9670 htab->root.sgotplt->output_section->vma
9671 + htab->root.sgotplt->output_offset;
9672
9673 bfd_vma dt_tlsdesc_got = got_addr + htab->root.tlsdesc_got;
9674
9675 bfd_byte *plt_entry =
9676 htab->root.splt->contents + htab->root.tlsdesc_plt;
9677
9678 /* First instruction in BTI enabled PLT stub is a BTI
9679 instruction so skip it. */
9680 if (type & PLT_BTI)
9681 {
9682 plt_entry = plt_entry + 4;
9683 adrp1_addr = adrp1_addr + 4;
9684 adrp2_addr = adrp2_addr + 4;
9685 }
9686
9687 /* adrp x2, DT_TLSDESC_GOT */
9688 elf_aarch64_update_plt_entry (output_bfd,
9689 BFD_RELOC_AARCH64_ADR_HI21_PCREL,
9690 plt_entry + 4,
9691 (PG (dt_tlsdesc_got)
9692 - PG (adrp1_addr)));
9693
9694 /* adrp x3, 0 */
9695 elf_aarch64_update_plt_entry (output_bfd,
9696 BFD_RELOC_AARCH64_ADR_HI21_PCREL,
9697 plt_entry + 8,
9698 (PG (pltgot_addr)
9699 - PG (adrp2_addr)));
9700
9701 /* ldr x2, [x2, #0] */
9702 elf_aarch64_update_plt_entry (output_bfd,
9703 BFD_RELOC_AARCH64_LDSTNN_LO12,
9704 plt_entry + 12,
9705 PG_OFFSET (dt_tlsdesc_got));
9706
9707 /* add x3, x3, 0 */
9708 elf_aarch64_update_plt_entry (output_bfd,
9709 BFD_RELOC_AARCH64_ADD_LO12,
9710 plt_entry + 16,
9711 PG_OFFSET (pltgot_addr));
9712 }
9713 }
9714 }
9715
9716 if (htab->root.sgotplt)
9717 {
9718 if (bfd_is_abs_section (htab->root.sgotplt->output_section))
9719 {
9720 _bfd_error_handler
9721 (_("discarded output section: `%pA'"), htab->root.sgotplt);
9722 return false;
9723 }
9724
9725 /* Fill in the first three entries in the global offset table. */
9726 if (htab->root.sgotplt->size > 0)
9727 {
9728 bfd_put_NN (output_bfd, (bfd_vma) 0, htab->root.sgotplt->contents);
9729
9730 /* Write GOT[1] and GOT[2], needed for the dynamic linker. */
9731 bfd_put_NN (output_bfd,
9732 (bfd_vma) 0,
9733 htab->root.sgotplt->contents + GOT_ENTRY_SIZE);
9734 bfd_put_NN (output_bfd,
9735 (bfd_vma) 0,
9736 htab->root.sgotplt->contents + GOT_ENTRY_SIZE * 2);
9737 }
9738
9739 if (htab->root.sgot)
9740 {
9741 if (htab->root.sgot->size > 0)
9742 {
9743 bfd_vma addr =
9744 sdyn ? sdyn->output_section->vma + sdyn->output_offset : 0;
9745 bfd_put_NN (output_bfd, addr, htab->root.sgot->contents);
9746 }
9747 }
9748
9749 elf_section_data (htab->root.sgotplt->output_section)->
9750 this_hdr.sh_entsize = GOT_ENTRY_SIZE;
9751 }
9752
9753 if (htab->root.sgot && htab->root.sgot->size > 0)
9754 elf_section_data (htab->root.sgot->output_section)->this_hdr.sh_entsize
9755 = GOT_ENTRY_SIZE;
9756
9757 /* Fill PLT and GOT entries for local STT_GNU_IFUNC symbols. */
9758 htab_traverse (htab->loc_hash_table,
9759 elfNN_aarch64_finish_local_dynamic_symbol,
9760 info);
9761
9762 return true;
9763 }
9764
9765 /* Check if BTI enabled PLTs are needed. Returns the type needed. */
9766 static aarch64_plt_type
get_plt_type(bfd * abfd)9767 get_plt_type (bfd *abfd)
9768 {
9769 aarch64_plt_type ret = PLT_NORMAL;
9770 bfd_byte *contents, *extdyn, *extdynend;
9771 asection *sec = bfd_get_section_by_name (abfd, ".dynamic");
9772 if (!sec || !bfd_malloc_and_get_section (abfd, sec, &contents))
9773 return ret;
9774 extdyn = contents;
9775 extdynend = contents + sec->size;
9776 for (; extdyn < extdynend; extdyn += sizeof (ElfNN_External_Dyn))
9777 {
9778 Elf_Internal_Dyn dyn;
9779 bfd_elfNN_swap_dyn_in (abfd, extdyn, &dyn);
9780
9781 /* Let's check the processor specific dynamic array tags. */
9782 bfd_vma tag = dyn.d_tag;
9783 if (tag < DT_LOPROC || tag > DT_HIPROC)
9784 continue;
9785
9786 switch (tag)
9787 {
9788 case DT_AARCH64_BTI_PLT:
9789 ret |= PLT_BTI;
9790 break;
9791
9792 case DT_AARCH64_PAC_PLT:
9793 ret |= PLT_PAC;
9794 break;
9795
9796 default: break;
9797 }
9798 }
9799 free (contents);
9800 return ret;
9801 }
9802
9803 static long
elfNN_aarch64_get_synthetic_symtab(bfd * abfd,long symcount,asymbol ** syms,long dynsymcount,asymbol ** dynsyms,asymbol ** ret)9804 elfNN_aarch64_get_synthetic_symtab (bfd *abfd,
9805 long symcount,
9806 asymbol **syms,
9807 long dynsymcount,
9808 asymbol **dynsyms,
9809 asymbol **ret)
9810 {
9811 elf_aarch64_tdata (abfd)->plt_type = get_plt_type (abfd);
9812 return _bfd_elf_get_synthetic_symtab (abfd, symcount, syms,
9813 dynsymcount, dynsyms, ret);
9814 }
9815
9816 /* Return address for Ith PLT stub in section PLT, for relocation REL
9817 or (bfd_vma) -1 if it should not be included. */
9818
9819 static bfd_vma
elfNN_aarch64_plt_sym_val(bfd_vma i,const asection * plt,const arelent * rel ATTRIBUTE_UNUSED)9820 elfNN_aarch64_plt_sym_val (bfd_vma i, const asection *plt,
9821 const arelent *rel ATTRIBUTE_UNUSED)
9822 {
9823 size_t plt0_size = PLT_ENTRY_SIZE;
9824 size_t pltn_size = PLT_SMALL_ENTRY_SIZE;
9825
9826 if (elf_aarch64_tdata (plt->owner)->plt_type == PLT_BTI_PAC)
9827 {
9828 if (elf_elfheader (plt->owner)->e_type == ET_EXEC)
9829 pltn_size = PLT_BTI_PAC_SMALL_ENTRY_SIZE;
9830 else
9831 pltn_size = PLT_PAC_SMALL_ENTRY_SIZE;
9832 }
9833 else if (elf_aarch64_tdata (plt->owner)->plt_type == PLT_BTI)
9834 {
9835 if (elf_elfheader (plt->owner)->e_type == ET_EXEC)
9836 pltn_size = PLT_BTI_SMALL_ENTRY_SIZE;
9837 }
9838 else if (elf_aarch64_tdata (plt->owner)->plt_type == PLT_PAC)
9839 {
9840 pltn_size = PLT_PAC_SMALL_ENTRY_SIZE;
9841 }
9842
9843 return plt->vma + plt0_size + i * pltn_size;
9844 }
9845
9846 /* Returns TRUE if NAME is an AArch64 mapping symbol.
9847 The ARM ELF standard defines $x (for A64 code) and $d (for data).
9848 It also allows a period initiated suffix to be added to the symbol, ie:
9849 "$[adtx]\.[:sym_char]+". */
9850
9851 static bool
is_aarch64_mapping_symbol(const char * name)9852 is_aarch64_mapping_symbol (const char * name)
9853 {
9854 return name != NULL /* Paranoia. */
9855 && name[0] == '$' /* Note: if objcopy --prefix-symbols has been used then
9856 the mapping symbols could have acquired a prefix.
9857 We do not support this here, since such symbols no
9858 longer conform to the ARM ELF ABI. */
9859 && (name[1] == 'd' || name[1] == 'x')
9860 && (name[2] == 0 || name[2] == '.');
9861 /* FIXME: Strictly speaking the symbol is only a valid mapping symbol if
9862 any characters that follow the period are legal characters for the body
9863 of a symbol's name. For now we just assume that this is the case. */
9864 }
9865
9866 /* Make sure that mapping symbols in object files are not removed via the
9867 "strip --strip-unneeded" tool. These symbols might needed in order to
9868 correctly generate linked files. Once an object file has been linked,
9869 it should be safe to remove them. */
9870
9871 static void
elfNN_aarch64_backend_symbol_processing(bfd * abfd,asymbol * sym)9872 elfNN_aarch64_backend_symbol_processing (bfd *abfd, asymbol *sym)
9873 {
9874 if (((abfd->flags & (EXEC_P | DYNAMIC)) == 0)
9875 && sym->section != bfd_abs_section_ptr
9876 && is_aarch64_mapping_symbol (sym->name))
9877 sym->flags |= BSF_KEEP;
9878 }
9879
9880 /* Implement elf_backend_setup_gnu_properties for AArch64. It serves as a
9881 wrapper function for _bfd_aarch64_elf_link_setup_gnu_properties to account
9882 for the effect of GNU properties of the output_bfd. */
9883 static bfd *
elfNN_aarch64_link_setup_gnu_properties(struct bfd_link_info * info)9884 elfNN_aarch64_link_setup_gnu_properties (struct bfd_link_info *info)
9885 {
9886 uint32_t prop = elf_aarch64_tdata (info->output_bfd)->gnu_and_prop;
9887 bfd *pbfd = _bfd_aarch64_elf_link_setup_gnu_properties (info, &prop);
9888 elf_aarch64_tdata (info->output_bfd)->gnu_and_prop = prop;
9889 elf_aarch64_tdata (info->output_bfd)->plt_type
9890 |= (prop & GNU_PROPERTY_AARCH64_FEATURE_1_BTI) ? PLT_BTI : 0;
9891 setup_plt_values (info, elf_aarch64_tdata (info->output_bfd)->plt_type);
9892 return pbfd;
9893 }
9894
9895 /* Implement elf_backend_merge_gnu_properties for AArch64. It serves as a
9896 wrapper function for _bfd_aarch64_elf_merge_gnu_properties to account
9897 for the effect of GNU properties of the output_bfd. */
9898 static bool
elfNN_aarch64_merge_gnu_properties(struct bfd_link_info * info,bfd * abfd,bfd * bbfd,elf_property * aprop,elf_property * bprop)9899 elfNN_aarch64_merge_gnu_properties (struct bfd_link_info *info,
9900 bfd *abfd, bfd *bbfd,
9901 elf_property *aprop,
9902 elf_property *bprop)
9903 {
9904 uint32_t prop
9905 = elf_aarch64_tdata (info->output_bfd)->gnu_and_prop;
9906
9907 /* If output has been marked with BTI using command line argument, give out
9908 warning if necessary. */
9909 /* Properties are merged per type, hence only check for warnings when merging
9910 GNU_PROPERTY_AARCH64_FEATURE_1_AND. */
9911 if (((aprop && aprop->pr_type == GNU_PROPERTY_AARCH64_FEATURE_1_AND)
9912 || (bprop && bprop->pr_type == GNU_PROPERTY_AARCH64_FEATURE_1_AND))
9913 && (prop & GNU_PROPERTY_AARCH64_FEATURE_1_BTI)
9914 && (!elf_aarch64_tdata (info->output_bfd)->no_bti_warn))
9915 {
9916 if ((aprop && !(aprop->u.number & GNU_PROPERTY_AARCH64_FEATURE_1_BTI))
9917 || !aprop)
9918 {
9919 _bfd_error_handler (_("%pB: warning: BTI turned on by -z force-bti when "
9920 "all inputs do not have BTI in NOTE section."),
9921 abfd);
9922 }
9923 if ((bprop && !(bprop->u.number & GNU_PROPERTY_AARCH64_FEATURE_1_BTI))
9924 || !bprop)
9925 {
9926 _bfd_error_handler (_("%pB: warning: BTI turned on by -z force-bti when "
9927 "all inputs do not have BTI in NOTE section."),
9928 bbfd);
9929 }
9930 }
9931
9932 return _bfd_aarch64_elf_merge_gnu_properties (info, abfd, aprop,
9933 bprop, prop);
9934 }
9935
9936 /* We use this so we can override certain functions
9937 (though currently we don't). */
9938
9939 const struct elf_size_info elfNN_aarch64_size_info =
9940 {
9941 sizeof (ElfNN_External_Ehdr),
9942 sizeof (ElfNN_External_Phdr),
9943 sizeof (ElfNN_External_Shdr),
9944 sizeof (ElfNN_External_Rel),
9945 sizeof (ElfNN_External_Rela),
9946 sizeof (ElfNN_External_Sym),
9947 sizeof (ElfNN_External_Dyn),
9948 sizeof (Elf_External_Note),
9949 4, /* Hash table entry size. */
9950 1, /* Internal relocs per external relocs. */
9951 ARCH_SIZE, /* Arch size. */
9952 LOG_FILE_ALIGN, /* Log_file_align. */
9953 ELFCLASSNN, EV_CURRENT,
9954 bfd_elfNN_write_out_phdrs,
9955 bfd_elfNN_write_shdrs_and_ehdr,
9956 bfd_elfNN_checksum_contents,
9957 bfd_elfNN_write_relocs,
9958 bfd_elfNN_swap_symbol_in,
9959 bfd_elfNN_swap_symbol_out,
9960 bfd_elfNN_slurp_reloc_table,
9961 bfd_elfNN_slurp_symbol_table,
9962 bfd_elfNN_swap_dyn_in,
9963 bfd_elfNN_swap_dyn_out,
9964 bfd_elfNN_swap_reloc_in,
9965 bfd_elfNN_swap_reloc_out,
9966 bfd_elfNN_swap_reloca_in,
9967 bfd_elfNN_swap_reloca_out
9968 };
9969
9970 #define ELF_ARCH bfd_arch_aarch64
9971 #define ELF_MACHINE_CODE EM_AARCH64
9972 #define ELF_MAXPAGESIZE 0x10000
9973 #define ELF_MINPAGESIZE 0x1000
9974 #define ELF_COMMONPAGESIZE 0x1000
9975
9976 #define bfd_elfNN_close_and_cleanup \
9977 elfNN_aarch64_close_and_cleanup
9978
9979 #define bfd_elfNN_bfd_free_cached_info \
9980 elfNN_aarch64_bfd_free_cached_info
9981
9982 #define bfd_elfNN_bfd_is_target_special_symbol \
9983 elfNN_aarch64_is_target_special_symbol
9984
9985 #define bfd_elfNN_bfd_link_hash_table_create \
9986 elfNN_aarch64_link_hash_table_create
9987
9988 #define bfd_elfNN_bfd_merge_private_bfd_data \
9989 elfNN_aarch64_merge_private_bfd_data
9990
9991 #define bfd_elfNN_bfd_print_private_bfd_data \
9992 elfNN_aarch64_print_private_bfd_data
9993
9994 #define bfd_elfNN_bfd_reloc_type_lookup \
9995 elfNN_aarch64_reloc_type_lookup
9996
9997 #define bfd_elfNN_bfd_reloc_name_lookup \
9998 elfNN_aarch64_reloc_name_lookup
9999
10000 #define bfd_elfNN_bfd_set_private_flags \
10001 elfNN_aarch64_set_private_flags
10002
10003 #define bfd_elfNN_find_inliner_info \
10004 elfNN_aarch64_find_inliner_info
10005
10006 #define bfd_elfNN_get_synthetic_symtab \
10007 elfNN_aarch64_get_synthetic_symtab
10008
10009 #define bfd_elfNN_mkobject \
10010 elfNN_aarch64_mkobject
10011
10012 #define bfd_elfNN_new_section_hook \
10013 elfNN_aarch64_new_section_hook
10014
10015 #define elf_backend_adjust_dynamic_symbol \
10016 elfNN_aarch64_adjust_dynamic_symbol
10017
10018 #define elf_backend_always_size_sections \
10019 elfNN_aarch64_always_size_sections
10020
10021 #define elf_backend_check_relocs \
10022 elfNN_aarch64_check_relocs
10023
10024 #define elf_backend_copy_indirect_symbol \
10025 elfNN_aarch64_copy_indirect_symbol
10026
10027 #define elf_backend_merge_symbol_attribute \
10028 elfNN_aarch64_merge_symbol_attribute
10029
10030 /* Create .dynbss, and .rela.bss sections in DYNOBJ, and set up shortcuts
10031 to them in our hash. */
10032 #define elf_backend_create_dynamic_sections \
10033 elfNN_aarch64_create_dynamic_sections
10034
10035 #define elf_backend_init_index_section \
10036 _bfd_elf_init_2_index_sections
10037
10038 #define elf_backend_finish_dynamic_sections \
10039 elfNN_aarch64_finish_dynamic_sections
10040
10041 #define elf_backend_finish_dynamic_symbol \
10042 elfNN_aarch64_finish_dynamic_symbol
10043
10044 #define elf_backend_object_p \
10045 elfNN_aarch64_object_p
10046
10047 #define elf_backend_output_arch_local_syms \
10048 elfNN_aarch64_output_arch_local_syms
10049
10050 #define elf_backend_maybe_function_sym \
10051 elfNN_aarch64_maybe_function_sym
10052
10053 #define elf_backend_plt_sym_val \
10054 elfNN_aarch64_plt_sym_val
10055
10056 #define elf_backend_init_file_header \
10057 elfNN_aarch64_init_file_header
10058
10059 #define elf_backend_relocate_section \
10060 elfNN_aarch64_relocate_section
10061
10062 #define elf_backend_reloc_type_class \
10063 elfNN_aarch64_reloc_type_class
10064
10065 #define elf_backend_section_from_shdr \
10066 elfNN_aarch64_section_from_shdr
10067
10068 #define elf_backend_size_dynamic_sections \
10069 elfNN_aarch64_size_dynamic_sections
10070
10071 #define elf_backend_size_info \
10072 elfNN_aarch64_size_info
10073
10074 #define elf_backend_write_section \
10075 elfNN_aarch64_write_section
10076
10077 #define elf_backend_symbol_processing \
10078 elfNN_aarch64_backend_symbol_processing
10079
10080 #define elf_backend_setup_gnu_properties \
10081 elfNN_aarch64_link_setup_gnu_properties
10082
10083 #define elf_backend_merge_gnu_properties \
10084 elfNN_aarch64_merge_gnu_properties
10085
10086 #define elf_backend_can_refcount 1
10087 #define elf_backend_can_gc_sections 1
10088 #define elf_backend_plt_readonly 1
10089 #define elf_backend_want_got_plt 1
10090 #define elf_backend_want_plt_sym 0
10091 #define elf_backend_want_dynrelro 1
10092 #define elf_backend_may_use_rel_p 0
10093 #define elf_backend_may_use_rela_p 1
10094 #define elf_backend_default_use_rela_p 1
10095 #define elf_backend_rela_normal 1
10096 #define elf_backend_dtrel_excludes_plt 1
10097 #define elf_backend_got_header_size (GOT_ENTRY_SIZE * 3)
10098 #define elf_backend_default_execstack 0
10099 #define elf_backend_extern_protected_data 1
10100 #define elf_backend_hash_symbol elf_aarch64_hash_symbol
10101
10102 #undef elf_backend_obj_attrs_section
10103 #define elf_backend_obj_attrs_section ".ARM.attributes"
10104
10105 #include "elfNN-target.h"
10106
10107 /* CloudABI support. */
10108
10109 #undef TARGET_LITTLE_SYM
10110 #define TARGET_LITTLE_SYM aarch64_elfNN_le_cloudabi_vec
10111 #undef TARGET_LITTLE_NAME
10112 #define TARGET_LITTLE_NAME "elfNN-littleaarch64-cloudabi"
10113 #undef TARGET_BIG_SYM
10114 #define TARGET_BIG_SYM aarch64_elfNN_be_cloudabi_vec
10115 #undef TARGET_BIG_NAME
10116 #define TARGET_BIG_NAME "elfNN-bigaarch64-cloudabi"
10117
10118 #undef ELF_OSABI
10119 #define ELF_OSABI ELFOSABI_CLOUDABI
10120
10121 #undef elfNN_bed
10122 #define elfNN_bed elfNN_aarch64_cloudabi_bed
10123
10124 #include "elfNN-target.h"
10125