| // aarch64.cc -- aarch64 target support for gold. |
| |
| // Copyright (C) 2014-2021 Free Software Foundation, Inc. |
| // Written by Jing Yu <jingyu@google.com> and Han Shen <shenhan@google.com>. |
| |
| // This file is part of gold. |
| |
| // This program is free software; you can redistribute it and/or modify |
| // it under the terms of the GNU General Public License as published by |
| // the Free Software Foundation; either version 3 of the License, or |
| // (at your option) any later version. |
| |
| // This program is distributed in the hope that it will be useful, |
| // but WITHOUT ANY WARRANTY; without even the implied warranty of |
| // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| // GNU General Public License for more details. |
| |
| // You should have received a copy of the GNU General Public License |
| // along with this program; if not, write to the Free Software |
| // Foundation, Inc., 51 Franklin Street - Fifth Floor, Boston, |
| // MA 02110-1301, USA. |
| |
| #include "gold.h" |
| |
| #include <cstring> |
| #include <map> |
| #include <set> |
| |
| #include "elfcpp.h" |
| #include "dwarf.h" |
| #include "parameters.h" |
| #include "reloc.h" |
| #include "aarch64.h" |
| #include "object.h" |
| #include "symtab.h" |
| #include "layout.h" |
| #include "output.h" |
| #include "copy-relocs.h" |
| #include "target.h" |
| #include "target-reloc.h" |
| #include "target-select.h" |
| #include "tls.h" |
| #include "freebsd.h" |
| #include "nacl.h" |
| #include "gc.h" |
| #include "icf.h" |
| #include "aarch64-reloc-property.h" |
| |
| // The first three .got.plt entries are reserved. |
| const int32_t AARCH64_GOTPLT_RESERVE_COUNT = 3; |
| |
| |
| namespace |
| { |
| |
| using namespace gold; |
| |
| template<int size, bool big_endian> |
| class Output_data_plt_aarch64; |
| |
| template<int size, bool big_endian> |
| class Output_data_plt_aarch64_standard; |
| |
| template<int size, bool big_endian> |
| class Target_aarch64; |
| |
| template<int size, bool big_endian> |
| class AArch64_relocate_functions; |
| |
| // Utility class dealing with insns. This is ported from macros in |
| // bfd/elfnn-aarch64.cc, but wrapped inside a class as static members. This |
| // class is used in erratum sequence scanning. |
| |
| template<bool big_endian> |
| class AArch64_insn_utilities |
| { |
| public: |
| typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype; |
| |
| static const int BYTES_PER_INSN; |
| |
| // Zero register encoding - 31. |
| static const unsigned int AARCH64_ZR; |
| |
| static unsigned int |
| aarch64_bit(Insntype insn, int pos) |
| { return ((1 << pos) & insn) >> pos; } |
| |
| static unsigned int |
| aarch64_bits(Insntype insn, int pos, int l) |
| { return (insn >> pos) & ((1 << l) - 1); } |
| |
| // Get the encoding field "op31" of 3-source data processing insns. "op31" is |
| // the name defined in armv8 insn manual C3.5.9. |
| static unsigned int |
| aarch64_op31(Insntype insn) |
| { return aarch64_bits(insn, 21, 3); } |
| |
| // Get the encoding field "ra" of 3-source data processing insns. "ra" is the |
| // third source register. See armv8 insn manual C3.5.9. |
| static unsigned int |
| aarch64_ra(Insntype insn) |
| { return aarch64_bits(insn, 10, 5); } |
| |
| static bool |
| is_adr(const Insntype insn) |
| { return (insn & 0x9F000000) == 0x10000000; } |
| |
| static bool |
| is_adrp(const Insntype insn) |
| { return (insn & 0x9F000000) == 0x90000000; } |
| |
| static bool |
| is_mrs_tpidr_el0(const Insntype insn) |
| { return (insn & 0xFFFFFFE0) == 0xd53bd040; } |
| |
| static unsigned int |
| aarch64_rm(const Insntype insn) |
| { return aarch64_bits(insn, 16, 5); } |
| |
| static unsigned int |
| aarch64_rn(const Insntype insn) |
| { return aarch64_bits(insn, 5, 5); } |
| |
| static unsigned int |
| aarch64_rd(const Insntype insn) |
| { return aarch64_bits(insn, 0, 5); } |
| |
| static unsigned int |
| aarch64_rt(const Insntype insn) |
| { return aarch64_bits(insn, 0, 5); } |
| |
| static unsigned int |
| aarch64_rt2(const Insntype insn) |
| { return aarch64_bits(insn, 10, 5); } |
| |
| // Encode imm21 into adr. Signed imm21 is in the range of [-1M, 1M). |
| static Insntype |
| aarch64_adr_encode_imm(Insntype adr, int imm21) |
| { |
| gold_assert(is_adr(adr)); |
| gold_assert(-(1 << 20) <= imm21 && imm21 < (1 << 20)); |
| const int mask19 = (1 << 19) - 1; |
| const int mask2 = 3; |
| adr &= ~((mask19 << 5) | (mask2 << 29)); |
| adr |= ((imm21 & mask2) << 29) | (((imm21 >> 2) & mask19) << 5); |
| return adr; |
| } |
| |
| // Retrieve encoded adrp 33-bit signed imm value. This value is obtained by |
| // 21-bit signed imm encoded in the insn multiplied by 4k (page size) and |
| // 64-bit sign-extended, resulting in [-4G, 4G) with 12-lsb being 0. |
| static int64_t |
| aarch64_adrp_decode_imm(const Insntype adrp) |
| { |
| const int mask19 = (1 << 19) - 1; |
| const int mask2 = 3; |
| gold_assert(is_adrp(adrp)); |
| // 21-bit imm encoded in adrp. |
| uint64_t imm = ((adrp >> 29) & mask2) | (((adrp >> 5) & mask19) << 2); |
| // Retrieve msb of 21-bit-signed imm for sign extension. |
| uint64_t msbt = (imm >> 20) & 1; |
| // Real value is imm multiplied by 4k. Value now has 33-bit information. |
| int64_t value = imm << 12; |
| // Sign extend to 64-bit by repeating msbt 31 (64-33) times and merge it |
| // with value. |
| return ((((uint64_t)(1) << 32) - msbt) << 33) | value; |
| } |
| |
| static bool |
| aarch64_b(const Insntype insn) |
| { return (insn & 0xFC000000) == 0x14000000; } |
| |
| static bool |
| aarch64_bl(const Insntype insn) |
| { return (insn & 0xFC000000) == 0x94000000; } |
| |
| static bool |
| aarch64_blr(const Insntype insn) |
| { return (insn & 0xFFFFFC1F) == 0xD63F0000; } |
| |
| static bool |
| aarch64_br(const Insntype insn) |
| { return (insn & 0xFFFFFC1F) == 0xD61F0000; } |
| |
| // All ld/st ops. See C4-182 of the ARM ARM. The encoding space for |
| // LD_PCREL, LDST_RO, LDST_UI and LDST_UIMM cover prefetch ops. |
| static bool |
| aarch64_ld(Insntype insn) { return aarch64_bit(insn, 22) == 1; } |
| |
| static bool |
| aarch64_ldst(Insntype insn) |
| { return (insn & 0x0a000000) == 0x08000000; } |
| |
| static bool |
| aarch64_ldst_ex(Insntype insn) |
| { return (insn & 0x3f000000) == 0x08000000; } |
| |
| static bool |
| aarch64_ldst_pcrel(Insntype insn) |
| { return (insn & 0x3b000000) == 0x18000000; } |
| |
| static bool |
| aarch64_ldst_nap(Insntype insn) |
| { return (insn & 0x3b800000) == 0x28000000; } |
| |
| static bool |
| aarch64_ldstp_pi(Insntype insn) |
| { return (insn & 0x3b800000) == 0x28800000; } |
| |
| static bool |
| aarch64_ldstp_o(Insntype insn) |
| { return (insn & 0x3b800000) == 0x29000000; } |
| |
| static bool |
| aarch64_ldstp_pre(Insntype insn) |
| { return (insn & 0x3b800000) == 0x29800000; } |
| |
| static bool |
| aarch64_ldst_ui(Insntype insn) |
| { return (insn & 0x3b200c00) == 0x38000000; } |
| |
| static bool |
| aarch64_ldst_piimm(Insntype insn) |
| { return (insn & 0x3b200c00) == 0x38000400; } |
| |
| static bool |
| aarch64_ldst_u(Insntype insn) |
| { return (insn & 0x3b200c00) == 0x38000800; } |
| |
| static bool |
| aarch64_ldst_preimm(Insntype insn) |
| { return (insn & 0x3b200c00) == 0x38000c00; } |
| |
| static bool |
| aarch64_ldst_ro(Insntype insn) |
| { return (insn & 0x3b200c00) == 0x38200800; } |
| |
| static bool |
| aarch64_ldst_uimm(Insntype insn) |
| { return (insn & 0x3b000000) == 0x39000000; } |
| |
| static bool |
| aarch64_ldst_simd_m(Insntype insn) |
| { return (insn & 0xbfbf0000) == 0x0c000000; } |
| |
| static bool |
| aarch64_ldst_simd_m_pi(Insntype insn) |
| { return (insn & 0xbfa00000) == 0x0c800000; } |
| |
| static bool |
| aarch64_ldst_simd_s(Insntype insn) |
| { return (insn & 0xbf9f0000) == 0x0d000000; } |
| |
| static bool |
| aarch64_ldst_simd_s_pi(Insntype insn) |
| { return (insn & 0xbf800000) == 0x0d800000; } |
| |
| // Classify an INSN if it is indeed a load/store. Return true if INSN is a |
| // LD/ST instruction otherwise return false. For scalar LD/ST instructions |
| // PAIR is FALSE, RT is returned and RT2 is set equal to RT. For LD/ST pair |
| // instructions PAIR is TRUE, RT and RT2 are returned. |
| static bool |
| aarch64_mem_op_p(Insntype insn, unsigned int *rt, unsigned int *rt2, |
| bool *pair, bool *load) |
| { |
| uint32_t opcode; |
| unsigned int r; |
| uint32_t opc = 0; |
| uint32_t v = 0; |
| uint32_t opc_v = 0; |
| |
| /* Bail out quickly if INSN doesn't fall into the load-store |
| encoding space. */ |
| if (!aarch64_ldst (insn)) |
| return false; |
| |
| *pair = false; |
| *load = false; |
| if (aarch64_ldst_ex (insn)) |
| { |
| *rt = aarch64_rt (insn); |
| *rt2 = *rt; |
| if (aarch64_bit (insn, 21) == 1) |
| { |
| *pair = true; |
| *rt2 = aarch64_rt2 (insn); |
| } |
| *load = aarch64_ld (insn); |
| return true; |
| } |
| else if (aarch64_ldst_nap (insn) |
| || aarch64_ldstp_pi (insn) |
| || aarch64_ldstp_o (insn) |
| || aarch64_ldstp_pre (insn)) |
| { |
| *pair = true; |
| *rt = aarch64_rt (insn); |
| *rt2 = aarch64_rt2 (insn); |
| *load = aarch64_ld (insn); |
| return true; |
| } |
| else if (aarch64_ldst_pcrel (insn) |
| || aarch64_ldst_ui (insn) |
| || aarch64_ldst_piimm (insn) |
| || aarch64_ldst_u (insn) |
| || aarch64_ldst_preimm (insn) |
| || aarch64_ldst_ro (insn) |
| || aarch64_ldst_uimm (insn)) |
| { |
| *rt = aarch64_rt (insn); |
| *rt2 = *rt; |
| if (aarch64_ldst_pcrel (insn)) |
| *load = true; |
| opc = aarch64_bits (insn, 22, 2); |
| v = aarch64_bit (insn, 26); |
| opc_v = opc | (v << 2); |
| *load = (opc_v == 1 || opc_v == 2 || opc_v == 3 |
| || opc_v == 5 || opc_v == 7); |
| return true; |
| } |
| else if (aarch64_ldst_simd_m (insn) |
| || aarch64_ldst_simd_m_pi (insn)) |
| { |
| *rt = aarch64_rt (insn); |
| *load = aarch64_bit (insn, 22); |
| opcode = (insn >> 12) & 0xf; |
| switch (opcode) |
| { |
| case 0: |
| case 2: |
| *rt2 = *rt + 3; |
| break; |
| |
| case 4: |
| case 6: |
| *rt2 = *rt + 2; |
| break; |
| |
| case 7: |
| *rt2 = *rt; |
| break; |
| |
| case 8: |
| case 10: |
| *rt2 = *rt + 1; |
| break; |
| |
| default: |
| return false; |
| } |
| return true; |
| } |
| else if (aarch64_ldst_simd_s (insn) |
| || aarch64_ldst_simd_s_pi (insn)) |
| { |
| *rt = aarch64_rt (insn); |
| r = (insn >> 21) & 1; |
| *load = aarch64_bit (insn, 22); |
| opcode = (insn >> 13) & 0x7; |
| switch (opcode) |
| { |
| case 0: |
| case 2: |
| case 4: |
| *rt2 = *rt + r; |
| break; |
| |
| case 1: |
| case 3: |
| case 5: |
| *rt2 = *rt + (r == 0 ? 2 : 3); |
| break; |
| |
| case 6: |
| *rt2 = *rt + r; |
| break; |
| |
| case 7: |
| *rt2 = *rt + (r == 0 ? 2 : 3); |
| break; |
| |
| default: |
| return false; |
| } |
| return true; |
| } |
| return false; |
| } // End of "aarch64_mem_op_p". |
| |
| // Return true if INSN is mac insn. |
| static bool |
| aarch64_mac(Insntype insn) |
| { return (insn & 0xff000000) == 0x9b000000; } |
| |
| // Return true if INSN is multiply-accumulate. |
| // (This is similar to implementaton in elfnn-aarch64.c.) |
| static bool |
| aarch64_mlxl(Insntype insn) |
| { |
| uint32_t op31 = aarch64_op31(insn); |
| if (aarch64_mac(insn) |
| && (op31 == 0 || op31 == 1 || op31 == 5) |
| /* Exclude MUL instructions which are encoded as a multiple-accumulate |
| with RA = XZR. */ |
| && aarch64_ra(insn) != AARCH64_ZR) |
| { |
| return true; |
| } |
| return false; |
| } |
| }; // End of "AArch64_insn_utilities". |
| |
| |
| // Insn length in byte. |
| |
| template<bool big_endian> |
| const int AArch64_insn_utilities<big_endian>::BYTES_PER_INSN = 4; |
| |
| |
| // Zero register encoding - 31. |
| |
| template<bool big_endian> |
| const unsigned int AArch64_insn_utilities<big_endian>::AARCH64_ZR = 0x1f; |
| |
| |
| // Output_data_got_aarch64 class. |
| |
| template<int size, bool big_endian> |
| class Output_data_got_aarch64 : public Output_data_got<size, big_endian> |
| { |
| public: |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr Valtype; |
| Output_data_got_aarch64(Symbol_table* symtab, Layout* layout) |
| : Output_data_got<size, big_endian>(), |
| symbol_table_(symtab), layout_(layout) |
| { } |
| |
| // Add a static entry for the GOT entry at OFFSET. GSYM is a global |
| // symbol and R_TYPE is the code of a dynamic relocation that needs to be |
| // applied in a static link. |
| void |
| add_static_reloc(unsigned int got_offset, unsigned int r_type, Symbol* gsym) |
| { this->static_relocs_.push_back(Static_reloc(got_offset, r_type, gsym)); } |
| |
| |
| // Add a static reloc for the GOT entry at OFFSET. RELOBJ is an object |
| // defining a local symbol with INDEX. R_TYPE is the code of a dynamic |
| // relocation that needs to be applied in a static link. |
| void |
| add_static_reloc(unsigned int got_offset, unsigned int r_type, |
| Sized_relobj_file<size, big_endian>* relobj, |
| unsigned int index) |
| { |
| this->static_relocs_.push_back(Static_reloc(got_offset, r_type, relobj, |
| index)); |
| } |
| |
| |
| protected: |
| // Write out the GOT table. |
| void |
| do_write(Output_file* of) { |
| // The first entry in the GOT is the address of the .dynamic section. |
| gold_assert(this->data_size() >= size / 8); |
| Output_section* dynamic = this->layout_->dynamic_section(); |
| Valtype dynamic_addr = dynamic == NULL ? 0 : dynamic->address(); |
| this->replace_constant(0, dynamic_addr); |
| Output_data_got<size, big_endian>::do_write(of); |
| |
| // Handling static relocs |
| if (this->static_relocs_.empty()) |
| return; |
| |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address; |
| |
| gold_assert(parameters->doing_static_link()); |
| const off_t offset = this->offset(); |
| const section_size_type oview_size = |
| convert_to_section_size_type(this->data_size()); |
| unsigned char* const oview = of->get_output_view(offset, oview_size); |
| |
| Output_segment* tls_segment = this->layout_->tls_segment(); |
| gold_assert(tls_segment != NULL); |
| |
| AArch64_address aligned_tcb_address = |
| align_address(Target_aarch64<size, big_endian>::TCB_SIZE, |
| tls_segment->maximum_alignment()); |
| |
| for (size_t i = 0; i < this->static_relocs_.size(); ++i) |
| { |
| Static_reloc& reloc(this->static_relocs_[i]); |
| AArch64_address value; |
| |
| if (!reloc.symbol_is_global()) |
| { |
| Sized_relobj_file<size, big_endian>* object = reloc.relobj(); |
| const Symbol_value<size>* psymval = |
| reloc.relobj()->local_symbol(reloc.index()); |
| |
| // We are doing static linking. Issue an error and skip this |
| // relocation if the symbol is undefined or in a discarded_section. |
| bool is_ordinary; |
| unsigned int shndx = psymval->input_shndx(&is_ordinary); |
| if ((shndx == elfcpp::SHN_UNDEF) |
| || (is_ordinary |
| && shndx != elfcpp::SHN_UNDEF |
| && !object->is_section_included(shndx) |
| && !this->symbol_table_->is_section_folded(object, shndx))) |
| { |
| gold_error(_("undefined or discarded local symbol %u from " |
| " object %s in GOT"), |
| reloc.index(), reloc.relobj()->name().c_str()); |
| continue; |
| } |
| value = psymval->value(object, 0); |
| } |
| else |
| { |
| const Symbol* gsym = reloc.symbol(); |
| gold_assert(gsym != NULL); |
| if (gsym->is_forwarder()) |
| gsym = this->symbol_table_->resolve_forwards(gsym); |
| |
| // We are doing static linking. Issue an error and skip this |
| // relocation if the symbol is undefined or in a discarded_section |
| // unless it is a weakly_undefined symbol. |
| if ((gsym->is_defined_in_discarded_section() |
| || gsym->is_undefined()) |
| && !gsym->is_weak_undefined()) |
| { |
| gold_error(_("undefined or discarded symbol %s in GOT"), |
| gsym->name()); |
| continue; |
| } |
| |
| if (!gsym->is_weak_undefined()) |
| { |
| const Sized_symbol<size>* sym = |
| static_cast<const Sized_symbol<size>*>(gsym); |
| value = sym->value(); |
| } |
| else |
| value = 0; |
| } |
| |
| unsigned got_offset = reloc.got_offset(); |
| gold_assert(got_offset < oview_size); |
| |
| typedef typename elfcpp::Swap<size, big_endian>::Valtype Valtype; |
| Valtype* wv = reinterpret_cast<Valtype*>(oview + got_offset); |
| Valtype x; |
| switch (reloc.r_type()) |
| { |
| case elfcpp::R_AARCH64_TLS_DTPREL64: |
| x = value; |
| break; |
| case elfcpp::R_AARCH64_TLS_TPREL64: |
| x = value + aligned_tcb_address; |
| break; |
| default: |
| gold_unreachable(); |
| } |
| elfcpp::Swap<size, big_endian>::writeval(wv, x); |
| } |
| |
| of->write_output_view(offset, oview_size, oview); |
| } |
| |
| private: |
| // Symbol table of the output object. |
| Symbol_table* symbol_table_; |
| // A pointer to the Layout class, so that we can find the .dynamic |
| // section when we write out the GOT section. |
| Layout* layout_; |
| |
| // This class represent dynamic relocations that need to be applied by |
| // gold because we are using TLS relocations in a static link. |
| class Static_reloc |
| { |
| public: |
| Static_reloc(unsigned int got_offset, unsigned int r_type, Symbol* gsym) |
| : got_offset_(got_offset), r_type_(r_type), symbol_is_global_(true) |
| { this->u_.global.symbol = gsym; } |
| |
| Static_reloc(unsigned int got_offset, unsigned int r_type, |
| Sized_relobj_file<size, big_endian>* relobj, unsigned int index) |
| : got_offset_(got_offset), r_type_(r_type), symbol_is_global_(false) |
| { |
| this->u_.local.relobj = relobj; |
| this->u_.local.index = index; |
| } |
| |
| // Return the GOT offset. |
| unsigned int |
| got_offset() const |
| { return this->got_offset_; } |
| |
| // Relocation type. |
| unsigned int |
| r_type() const |
| { return this->r_type_; } |
| |
| // Whether the symbol is global or not. |
| bool |
| symbol_is_global() const |
| { return this->symbol_is_global_; } |
| |
| // For a relocation against a global symbol, the global symbol. |
| Symbol* |
| symbol() const |
| { |
| gold_assert(this->symbol_is_global_); |
| return this->u_.global.symbol; |
| } |
| |
| // For a relocation against a local symbol, the defining object. |
| Sized_relobj_file<size, big_endian>* |
| relobj() const |
| { |
| gold_assert(!this->symbol_is_global_); |
| return this->u_.local.relobj; |
| } |
| |
| // For a relocation against a local symbol, the local symbol index. |
| unsigned int |
| index() const |
| { |
| gold_assert(!this->symbol_is_global_); |
| return this->u_.local.index; |
| } |
| |
| private: |
| // GOT offset of the entry to which this relocation is applied. |
| unsigned int got_offset_; |
| // Type of relocation. |
| unsigned int r_type_; |
| // Whether this relocation is against a global symbol. |
| bool symbol_is_global_; |
| // A global or local symbol. |
| union |
| { |
| struct |
| { |
| // For a global symbol, the symbol itself. |
| Symbol* symbol; |
| } global; |
| struct |
| { |
| // For a local symbol, the object defining the symbol. |
| Sized_relobj_file<size, big_endian>* relobj; |
| // For a local symbol, the symbol index. |
| unsigned int index; |
| } local; |
| } u_; |
| }; // End of inner class Static_reloc |
| |
| std::vector<Static_reloc> static_relocs_; |
| }; // End of Output_data_got_aarch64 |
| |
| |
| template<int size, bool big_endian> |
| class AArch64_input_section; |
| |
| |
| template<int size, bool big_endian> |
| class AArch64_output_section; |
| |
| |
| template<int size, bool big_endian> |
| class AArch64_relobj; |
| |
| |
| // Stub type enum constants. |
| |
| enum |
| { |
| ST_NONE = 0, |
| |
| // Using adrp/add pair, 4 insns (including alignment) without mem access, |
| // the fastest stub. This has a limited jump distance, which is tested by |
| // aarch64_valid_for_adrp_p. |
| ST_ADRP_BRANCH = 1, |
| |
| // Using ldr-absolute-address/br-register, 4 insns with 1 mem access, |
| // unlimited in jump distance. |
| ST_LONG_BRANCH_ABS = 2, |
| |
| // Using ldr/calculate-pcrel/jump, 8 insns (including alignment) with 1 |
| // mem access, slowest one. Only used in position independent executables. |
| ST_LONG_BRANCH_PCREL = 3, |
| |
| // Stub for erratum 843419 handling. |
| ST_E_843419 = 4, |
| |
| // Stub for erratum 835769 handling. |
| ST_E_835769 = 5, |
| |
| // Number of total stub types. |
| ST_NUMBER = 6 |
| }; |
| |
| |
| // Struct that wraps insns for a particular stub. All stub templates are |
| // created/initialized as constants by Stub_template_repertoire. |
| |
| template<bool big_endian> |
| struct Stub_template |
| { |
| const typename AArch64_insn_utilities<big_endian>::Insntype* insns; |
| const int insn_num; |
| }; |
| |
| |
| // Simple singleton class that creates/initializes/stores all types of stub |
| // templates. |
| |
| template<bool big_endian> |
| class Stub_template_repertoire |
| { |
| public: |
| typedef typename AArch64_insn_utilities<big_endian>::Insntype Insntype; |
| |
| // Single static method to get stub template for a given stub type. |
| static const Stub_template<big_endian>* |
| get_stub_template(int type) |
| { |
| static Stub_template_repertoire<big_endian> singleton; |
| return singleton.stub_templates_[type]; |
| } |
| |
| private: |
| // Constructor - creates/initializes all stub templates. |
| Stub_template_repertoire(); |
| ~Stub_template_repertoire() |
| { } |
| |
| // Disallowing copy ctor and copy assignment operator. |
| Stub_template_repertoire(Stub_template_repertoire&); |
| Stub_template_repertoire& operator=(Stub_template_repertoire&); |
| |
| // Data that stores all insn templates. |
| const Stub_template<big_endian>* stub_templates_[ST_NUMBER]; |
| }; // End of "class Stub_template_repertoire". |
| |
| |
| // Constructor - creates/initilizes all stub templates. |
| |
| template<bool big_endian> |
| Stub_template_repertoire<big_endian>::Stub_template_repertoire() |
| { |
| // Insn array definitions. |
| const static Insntype ST_NONE_INSNS[] = {}; |
| |
| const static Insntype ST_ADRP_BRANCH_INSNS[] = |
| { |
| 0x90000010, /* adrp ip0, X */ |
| /* ADR_PREL_PG_HI21(X) */ |
| 0x91000210, /* add ip0, ip0, :lo12:X */ |
| /* ADD_ABS_LO12_NC(X) */ |
| 0xd61f0200, /* br ip0 */ |
| 0x00000000, /* alignment padding */ |
| }; |
| |
| const static Insntype ST_LONG_BRANCH_ABS_INSNS[] = |
| { |
| 0x58000050, /* ldr ip0, 0x8 */ |
| 0xd61f0200, /* br ip0 */ |
| 0x00000000, /* address field */ |
| 0x00000000, /* address fields */ |
| }; |
| |
| const static Insntype ST_LONG_BRANCH_PCREL_INSNS[] = |
| { |
| 0x58000090, /* ldr ip0, 0x10 */ |
| 0x10000011, /* adr ip1, #0 */ |
| 0x8b110210, /* add ip0, ip0, ip1 */ |
| 0xd61f0200, /* br ip0 */ |
| 0x00000000, /* address field */ |
| 0x00000000, /* address field */ |
| 0x00000000, /* alignment padding */ |
| 0x00000000, /* alignment padding */ |
| }; |
| |
| const static Insntype ST_E_843419_INSNS[] = |
| { |
| 0x00000000, /* Placeholder for erratum insn. */ |
| 0x14000000, /* b <label> */ |
| }; |
| |
| // ST_E_835769 has the same stub template as ST_E_843419 |
| // but we reproduce the array here so that the sizeof |
| // expressions in install_insn_template will work. |
| const static Insntype ST_E_835769_INSNS[] = |
| { |
| 0x00000000, /* Placeholder for erratum insn. */ |
| 0x14000000, /* b <label> */ |
| }; |
| |
| #define install_insn_template(T) \ |
| const static Stub_template<big_endian> template_##T = { \ |
| T##_INSNS, sizeof(T##_INSNS) / sizeof(T##_INSNS[0]) }; \ |
| this->stub_templates_[T] = &template_##T |
| |
| install_insn_template(ST_NONE); |
| install_insn_template(ST_ADRP_BRANCH); |
| install_insn_template(ST_LONG_BRANCH_ABS); |
| install_insn_template(ST_LONG_BRANCH_PCREL); |
| install_insn_template(ST_E_843419); |
| install_insn_template(ST_E_835769); |
| |
| #undef install_insn_template |
| } |
| |
| |
| // Base class for stubs. |
| |
| template<int size, bool big_endian> |
| class Stub_base |
| { |
| public: |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address; |
| typedef typename AArch64_insn_utilities<big_endian>::Insntype Insntype; |
| |
| static const AArch64_address invalid_address = |
| static_cast<AArch64_address>(-1); |
| |
| static const section_offset_type invalid_offset = |
| static_cast<section_offset_type>(-1); |
| |
| Stub_base(int type) |
| : destination_address_(invalid_address), |
| offset_(invalid_offset), |
| type_(type) |
| {} |
| |
| ~Stub_base() |
| {} |
| |
| // Get stub type. |
| int |
| type() const |
| { return this->type_; } |
| |
| // Get stub template that provides stub insn information. |
| const Stub_template<big_endian>* |
| stub_template() const |
| { |
| return Stub_template_repertoire<big_endian>:: |
| get_stub_template(this->type()); |
| } |
| |
| // Get destination address. |
| AArch64_address |
| destination_address() const |
| { |
| gold_assert(this->destination_address_ != this->invalid_address); |
| return this->destination_address_; |
| } |
| |
| // Set destination address. |
| void |
| set_destination_address(AArch64_address address) |
| { |
| gold_assert(address != this->invalid_address); |
| this->destination_address_ = address; |
| } |
| |
| // Reset the destination address. |
| void |
| reset_destination_address() |
| { this->destination_address_ = this->invalid_address; } |
| |
| // Get offset of code stub. For Reloc_stub, it is the offset from the |
| // beginning of its containing stub table; for Erratum_stub, it is the offset |
| // from the end of reloc_stubs. |
| section_offset_type |
| offset() const |
| { |
| gold_assert(this->offset_ != this->invalid_offset); |
| return this->offset_; |
| } |
| |
| // Set stub offset. |
| void |
| set_offset(section_offset_type offset) |
| { this->offset_ = offset; } |
| |
| // Return the stub insn. |
| const Insntype* |
| insns() const |
| { return this->stub_template()->insns; } |
| |
| // Return num of stub insns. |
| unsigned int |
| insn_num() const |
| { return this->stub_template()->insn_num; } |
| |
| // Get size of the stub. |
| int |
| stub_size() const |
| { |
| return this->insn_num() * |
| AArch64_insn_utilities<big_endian>::BYTES_PER_INSN; |
| } |
| |
| // Write stub to output file. |
| void |
| write(unsigned char* view, section_size_type view_size) |
| { this->do_write(view, view_size); } |
| |
| protected: |
| // Abstract method to be implemented by sub-classes. |
| virtual void |
| do_write(unsigned char*, section_size_type) = 0; |
| |
| private: |
| // The last insn of a stub is a jump to destination insn. This field records |
| // the destination address. |
| AArch64_address destination_address_; |
| // The stub offset. Note this has difference interpretations between an |
| // Reloc_stub and an Erratum_stub. For Reloc_stub this is the offset from the |
| // beginning of the containing stub_table, whereas for Erratum_stub, this is |
| // the offset from the end of reloc_stubs. |
| section_offset_type offset_; |
| // Stub type. |
| const int type_; |
| }; // End of "Stub_base". |
| |
| |
| // Erratum stub class. An erratum stub differs from a reloc stub in that for |
| // each erratum occurrence, we generate an erratum stub. We never share erratum |
| // stubs, whereas for reloc stubs, different branch insns share a single reloc |
| // stub as long as the branch targets are the same. (More to the point, reloc |
| // stubs can be shared because they're used to reach a specific target, whereas |
| // erratum stubs branch back to the original control flow.) |
| |
| template<int size, bool big_endian> |
| class Erratum_stub : public Stub_base<size, big_endian> |
| { |
| public: |
| typedef AArch64_relobj<size, big_endian> The_aarch64_relobj; |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address; |
| typedef AArch64_insn_utilities<big_endian> Insn_utilities; |
| typedef typename AArch64_insn_utilities<big_endian>::Insntype Insntype; |
| |
| static const int STUB_ADDR_ALIGN; |
| |
| static const Insntype invalid_insn = static_cast<Insntype>(-1); |
| |
| Erratum_stub(The_aarch64_relobj* relobj, int type, |
| unsigned shndx, unsigned int sh_offset) |
| : Stub_base<size, big_endian>(type), relobj_(relobj), |
| shndx_(shndx), sh_offset_(sh_offset), |
| erratum_insn_(invalid_insn), |
| erratum_address_(this->invalid_address) |
| {} |
| |
| ~Erratum_stub() {} |
| |
| // Return the object that contains the erratum. |
| The_aarch64_relobj* |
| relobj() |
| { return this->relobj_; } |
| |
| // Get section index of the erratum. |
| unsigned int |
| shndx() const |
| { return this->shndx_; } |
| |
| // Get section offset of the erratum. |
| unsigned int |
| sh_offset() const |
| { return this->sh_offset_; } |
| |
| // Get the erratum insn. This is the insn located at erratum_insn_address. |
| Insntype |
| erratum_insn() const |
| { |
| gold_assert(this->erratum_insn_ != this->invalid_insn); |
| return this->erratum_insn_; |
| } |
| |
| // Set the insn that the erratum happens to. |
| void |
| set_erratum_insn(Insntype insn) |
| { this->erratum_insn_ = insn; } |
| |
| // For 843419, the erratum insn is ld/st xt, [xn, #uimm], which may be a |
| // relocation spot, in this case, the erratum_insn_ recorded at scanning phase |
| // is no longer the one we want to write out to the stub, update erratum_insn_ |
| // with relocated version. Also note that in this case xn must not be "PC", so |
| // it is safe to move the erratum insn from the origin place to the stub. For |
| // 835769, the erratum insn is multiply-accumulate insn, which could not be a |
| // relocation spot (assertion added though). |
| void |
| update_erratum_insn(Insntype insn) |
| { |
| gold_assert(this->erratum_insn_ != this->invalid_insn); |
| switch (this->type()) |
| { |
| case ST_E_843419: |
| gold_assert(Insn_utilities::aarch64_ldst_uimm(insn)); |
| gold_assert(Insn_utilities::aarch64_ldst_uimm(this->erratum_insn())); |
| gold_assert(Insn_utilities::aarch64_rd(insn) == |
| Insn_utilities::aarch64_rd(this->erratum_insn())); |
| gold_assert(Insn_utilities::aarch64_rn(insn) == |
| Insn_utilities::aarch64_rn(this->erratum_insn())); |
| // Update plain ld/st insn with relocated insn. |
| this->erratum_insn_ = insn; |
| break; |
| case ST_E_835769: |
| gold_assert(insn == this->erratum_insn()); |
| break; |
| default: |
| gold_unreachable(); |
| } |
| } |
| |
| |
| // Return the address where an erratum must be done. |
| AArch64_address |
| erratum_address() const |
| { |
| gold_assert(this->erratum_address_ != this->invalid_address); |
| return this->erratum_address_; |
| } |
| |
| // Set the address where an erratum must be done. |
| void |
| set_erratum_address(AArch64_address addr) |
| { this->erratum_address_ = addr; } |
| |
| // Later relaxation passes of may alter the recorded erratum and destination |
| // address. Given an up to date output section address of shidx_ in |
| // relobj_ we can derive the erratum_address and destination address. |
| void |
| update_erratum_address(AArch64_address output_section_addr) |
| { |
| const int BPI = AArch64_insn_utilities<big_endian>::BYTES_PER_INSN; |
| AArch64_address updated_addr = output_section_addr + this->sh_offset_; |
| this->set_erratum_address(updated_addr); |
| this->set_destination_address(updated_addr + BPI); |
| } |
| |
| // Comparator used to group Erratum_stubs in a set by (obj, shndx, |
| // sh_offset). We do not include 'type' in the calculation, because there is |
| // at most one stub type at (obj, shndx, sh_offset). |
| bool |
| operator<(const Erratum_stub<size, big_endian>& k) const |
| { |
| if (this == &k) |
| return false; |
| // We group stubs by relobj. |
| if (this->relobj_ != k.relobj_) |
| return this->relobj_ < k.relobj_; |
| // Then by section index. |
| if (this->shndx_ != k.shndx_) |
| return this->shndx_ < k.shndx_; |
| // Lastly by section offset. |
| return this->sh_offset_ < k.sh_offset_; |
| } |
| |
| void |
| invalidate_erratum_stub() |
| { |
| gold_assert(this->erratum_insn_ != invalid_insn); |
| this->erratum_insn_ = invalid_insn; |
| } |
| |
| bool |
| is_invalidated_erratum_stub() |
| { return this->erratum_insn_ == invalid_insn; } |
| |
| protected: |
| virtual void |
| do_write(unsigned char*, section_size_type); |
| |
| private: |
| // The object that needs to be fixed. |
| The_aarch64_relobj* relobj_; |
| // The shndx in the object that needs to be fixed. |
| const unsigned int shndx_; |
| // The section offset in the obejct that needs to be fixed. |
| const unsigned int sh_offset_; |
| // The insn to be fixed. |
| Insntype erratum_insn_; |
| // The address of the above insn. |
| AArch64_address erratum_address_; |
| }; // End of "Erratum_stub". |
| |
| |
| // Erratum sub class to wrap additional info needed by 843419. In fixing this |
| // erratum, we may choose to replace 'adrp' with 'adr', in this case, we need |
| // adrp's code position (two or three insns before erratum insn itself). |
| |
| template<int size, bool big_endian> |
| class E843419_stub : public Erratum_stub<size, big_endian> |
| { |
| public: |
| typedef typename AArch64_insn_utilities<big_endian>::Insntype Insntype; |
| |
| E843419_stub(AArch64_relobj<size, big_endian>* relobj, |
| unsigned int shndx, unsigned int sh_offset, |
| unsigned int adrp_sh_offset) |
| : Erratum_stub<size, big_endian>(relobj, ST_E_843419, shndx, sh_offset), |
| adrp_sh_offset_(adrp_sh_offset) |
| {} |
| |
| unsigned int |
| adrp_sh_offset() const |
| { return this->adrp_sh_offset_; } |
| |
| private: |
| // Section offset of "adrp". (We do not need a "adrp_shndx_" field, because we |
| // can obtain it from its parent.) |
| const unsigned int adrp_sh_offset_; |
| }; |
| |
| |
| template<int size, bool big_endian> |
| const int Erratum_stub<size, big_endian>::STUB_ADDR_ALIGN = 4; |
| |
| // Comparator used in set definition. |
| template<int size, bool big_endian> |
| struct Erratum_stub_less |
| { |
| bool |
| operator()(const Erratum_stub<size, big_endian>* s1, |
| const Erratum_stub<size, big_endian>* s2) const |
| { return *s1 < *s2; } |
| }; |
| |
| // Erratum_stub implementation for writing stub to output file. |
| |
| template<int size, bool big_endian> |
| void |
| Erratum_stub<size, big_endian>::do_write(unsigned char* view, section_size_type) |
| { |
| typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype; |
| const Insntype* insns = this->insns(); |
| uint32_t num_insns = this->insn_num(); |
| Insntype* ip = reinterpret_cast<Insntype*>(view); |
| // For current implemented erratum 843419 and 835769, the first insn in the |
| // stub is always a copy of the problematic insn (in 843419, the mem access |
| // insn, in 835769, the mac insn), followed by a jump-back. |
| elfcpp::Swap<32, big_endian>::writeval(ip, this->erratum_insn()); |
| for (uint32_t i = 1; i < num_insns; ++i) |
| elfcpp::Swap<32, big_endian>::writeval(ip + i, insns[i]); |
| } |
| |
| |
| // Reloc stub class. |
| |
| template<int size, bool big_endian> |
| class Reloc_stub : public Stub_base<size, big_endian> |
| { |
| public: |
| typedef Reloc_stub<size, big_endian> This; |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address; |
| |
| // Branch range. This is used to calculate the section group size, as well as |
| // determine whether a stub is needed. |
| static const int MAX_BRANCH_OFFSET = ((1 << 25) - 1) << 2; |
| static const int MIN_BRANCH_OFFSET = -((1 << 25) << 2); |
| |
| // Constant used to determine if an offset fits in the adrp instruction |
| // encoding. |
| static const int MAX_ADRP_IMM = (1 << 20) - 1; |
| static const int MIN_ADRP_IMM = -(1 << 20); |
| |
| static const int BYTES_PER_INSN = 4; |
| static const int STUB_ADDR_ALIGN; |
| |
| // Determine whether the offset fits in the jump/branch instruction. |
| static bool |
| aarch64_valid_branch_offset_p(int64_t offset) |
| { return offset >= MIN_BRANCH_OFFSET && offset <= MAX_BRANCH_OFFSET; } |
| |
| // Determine whether the offset fits in the adrp immediate field. |
| static bool |
| aarch64_valid_for_adrp_p(AArch64_address location, AArch64_address dest) |
| { |
| typedef AArch64_relocate_functions<size, big_endian> Reloc; |
| int64_t adrp_imm = (Reloc::Page(dest) - Reloc::Page(location)) >> 12; |
| return adrp_imm >= MIN_ADRP_IMM && adrp_imm <= MAX_ADRP_IMM; |
| } |
| |
| // Determine the stub type for a certain relocation or ST_NONE, if no stub is |
| // needed. |
| static int |
| stub_type_for_reloc(unsigned int r_type, AArch64_address address, |
| AArch64_address target); |
| |
| Reloc_stub(int type) |
| : Stub_base<size, big_endian>(type) |
| { } |
| |
| ~Reloc_stub() |
| { } |
| |
| // The key class used to index the stub instance in the stub table's stub map. |
| class Key |
| { |
| public: |
| Key(int type, const Symbol* symbol, const Relobj* relobj, |
| unsigned int r_sym, int32_t addend) |
| : type_(type), addend_(addend) |
| { |
| if (symbol != NULL) |
| { |
| this->r_sym_ = Reloc_stub::invalid_index; |
| this->u_.symbol = symbol; |
| } |
| else |
| { |
| gold_assert(relobj != NULL && r_sym != invalid_index); |
| this->r_sym_ = r_sym; |
| this->u_.relobj = relobj; |
| } |
| } |
| |
| ~Key() |
| { } |
| |
| // Return stub type. |
| int |
| type() const |
| { return this->type_; } |
| |
| // Return the local symbol index or invalid_index. |
| unsigned int |
| r_sym() const |
| { return this->r_sym_; } |
| |
| // Return the symbol if there is one. |
| const Symbol* |
| symbol() const |
| { return this->r_sym_ == invalid_index ? this->u_.symbol : NULL; } |
| |
| // Return the relobj if there is one. |
| const Relobj* |
| relobj() const |
| { return this->r_sym_ != invalid_index ? this->u_.relobj : NULL; } |
| |
| // Whether this equals to another key k. |
| bool |
| eq(const Key& k) const |
| { |
| return ((this->type_ == k.type_) |
| && (this->r_sym_ == k.r_sym_) |
| && ((this->r_sym_ != Reloc_stub::invalid_index) |
| ? (this->u_.relobj == k.u_.relobj) |
| : (this->u_.symbol == k.u_.symbol)) |
| && (this->addend_ == k.addend_)); |
| } |
| |
| // Return a hash value. |
| size_t |
| hash_value() const |
| { |
| size_t name_hash_value = gold::string_hash<char>( |
| (this->r_sym_ != Reloc_stub::invalid_index) |
| ? this->u_.relobj->name().c_str() |
| : this->u_.symbol->name()); |
| // We only have 4 stub types. |
| size_t stub_type_hash_value = 0x03 & this->type_; |
| return (name_hash_value |
| ^ stub_type_hash_value |
| ^ ((this->r_sym_ & 0x3fff) << 2) |
| ^ ((this->addend_ & 0xffff) << 16)); |
| } |
| |
| // Functors for STL associative containers. |
| struct hash |
| { |
| size_t |
| operator()(const Key& k) const |
| { return k.hash_value(); } |
| }; |
| |
| struct equal_to |
| { |
| bool |
| operator()(const Key& k1, const Key& k2) const |
| { return k1.eq(k2); } |
| }; |
| |
| private: |
| // Stub type. |
| const int type_; |
| // If this is a local symbol, this is the index in the defining object. |
| // Otherwise, it is invalid_index for a global symbol. |
| unsigned int r_sym_; |
| // If r_sym_ is an invalid index, this points to a global symbol. |
| // Otherwise, it points to a relobj. We used the unsized and target |
| // independent Symbol and Relobj classes instead of Sized_symbol<32> and |
| // Arm_relobj, in order to avoid making the stub class a template |
| // as most of the stub machinery is endianness-neutral. However, it |
| // may require a bit of casting done by users of this class. |
| union |
| { |
| const Symbol* symbol; |
| const Relobj* relobj; |
| } u_; |
| // Addend associated with a reloc. |
| int32_t addend_; |
| }; // End of inner class Reloc_stub::Key |
| |
| protected: |
| // This may be overridden in the child class. |
| virtual void |
| do_write(unsigned char*, section_size_type); |
| |
| private: |
| static const unsigned int invalid_index = static_cast<unsigned int>(-1); |
| }; // End of Reloc_stub |
| |
| template<int size, bool big_endian> |
| const int Reloc_stub<size, big_endian>::STUB_ADDR_ALIGN = 4; |
| |
| // Write data to output file. |
| |
| template<int size, bool big_endian> |
| void |
| Reloc_stub<size, big_endian>:: |
| do_write(unsigned char* view, section_size_type) |
| { |
| typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype; |
| const uint32_t* insns = this->insns(); |
| uint32_t num_insns = this->insn_num(); |
| Insntype* ip = reinterpret_cast<Insntype*>(view); |
| for (uint32_t i = 0; i < num_insns; ++i) |
| elfcpp::Swap<32, big_endian>::writeval(ip + i, insns[i]); |
| } |
| |
| |
| // Determine the stub type for a certain relocation or ST_NONE, if no stub is |
| // needed. |
| |
| template<int size, bool big_endian> |
| inline int |
| Reloc_stub<size, big_endian>::stub_type_for_reloc( |
| unsigned int r_type, AArch64_address location, AArch64_address dest) |
| { |
| int64_t branch_offset = 0; |
| switch(r_type) |
| { |
| case elfcpp::R_AARCH64_CALL26: |
| case elfcpp::R_AARCH64_JUMP26: |
| branch_offset = dest - location; |
| break; |
| default: |
| gold_unreachable(); |
| } |
| |
| if (aarch64_valid_branch_offset_p(branch_offset)) |
| return ST_NONE; |
| |
| if (aarch64_valid_for_adrp_p(location, dest)) |
| return ST_ADRP_BRANCH; |
| |
| // Always use PC-relative addressing in case of -shared or -pie. |
| if (parameters->options().output_is_position_independent()) |
| return ST_LONG_BRANCH_PCREL; |
| |
| // This saves 2 insns per stub, compared to ST_LONG_BRANCH_PCREL. |
| // But is only applicable to non-shared or non-pie. |
| return ST_LONG_BRANCH_ABS; |
| } |
| |
| // A class to hold stubs for the ARM target. This contains 2 different types of |
| // stubs - reloc stubs and erratum stubs. |
| |
| template<int size, bool big_endian> |
| class Stub_table : public Output_data |
| { |
| public: |
| typedef Target_aarch64<size, big_endian> The_target_aarch64; |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address; |
| typedef AArch64_relobj<size, big_endian> The_aarch64_relobj; |
| typedef AArch64_input_section<size, big_endian> The_aarch64_input_section; |
| typedef Reloc_stub<size, big_endian> The_reloc_stub; |
| typedef typename The_reloc_stub::Key The_reloc_stub_key; |
| typedef Erratum_stub<size, big_endian> The_erratum_stub; |
| typedef Erratum_stub_less<size, big_endian> The_erratum_stub_less; |
| typedef typename The_reloc_stub_key::hash The_reloc_stub_key_hash; |
| typedef typename The_reloc_stub_key::equal_to The_reloc_stub_key_equal_to; |
| typedef Stub_table<size, big_endian> The_stub_table; |
| typedef Unordered_map<The_reloc_stub_key, The_reloc_stub*, |
| The_reloc_stub_key_hash, The_reloc_stub_key_equal_to> |
| Reloc_stub_map; |
| typedef typename Reloc_stub_map::const_iterator Reloc_stub_map_const_iter; |
| typedef Relocate_info<size, big_endian> The_relocate_info; |
| |
| typedef std::set<The_erratum_stub*, The_erratum_stub_less> Erratum_stub_set; |
| typedef typename Erratum_stub_set::iterator Erratum_stub_set_iter; |
| |
| Stub_table(The_aarch64_input_section* owner) |
| : Output_data(), owner_(owner), reloc_stubs_size_(0), |
| erratum_stubs_size_(0), prev_data_size_(0) |
| { } |
| |
| ~Stub_table() |
| { } |
| |
| The_aarch64_input_section* |
| owner() const |
| { return owner_; } |
| |
| // Whether this stub table is empty. |
| bool |
| empty() const |
| { return reloc_stubs_.empty() && erratum_stubs_.empty(); } |
| |
| // Return the current data size. |
| off_t |
| current_data_size() const |
| { return this->current_data_size_for_child(); } |
| |
| // Add a STUB using KEY. The caller is responsible for avoiding addition |
| // if a STUB with the same key has already been added. |
| void |
| add_reloc_stub(The_reloc_stub* stub, const The_reloc_stub_key& key); |
| |
| // Add an erratum stub into the erratum stub set. The set is ordered by |
| // (relobj, shndx, sh_offset). |
| void |
| add_erratum_stub(The_erratum_stub* stub); |
| |
| // Find if such erratum exists for any given (obj, shndx, sh_offset). |
| The_erratum_stub* |
| find_erratum_stub(The_aarch64_relobj* a64relobj, |
| unsigned int shndx, unsigned int sh_offset); |
| |
| // Find all the erratums for a given input section. The return value is a pair |
| // of iterators [begin, end). |
| std::pair<Erratum_stub_set_iter, Erratum_stub_set_iter> |
| find_erratum_stubs_for_input_section(The_aarch64_relobj* a64relobj, |
| unsigned int shndx); |
| |
| // Compute the erratum stub address. |
| AArch64_address |
| erratum_stub_address(The_erratum_stub* stub) const |
| { |
| AArch64_address r = align_address(this->address() + this->reloc_stubs_size_, |
| The_erratum_stub::STUB_ADDR_ALIGN); |
| r += stub->offset(); |
| return r; |
| } |
| |
| // Finalize stubs. No-op here, just for completeness. |
| void |
| finalize_stubs() |
| { } |
| |
| // Look up a relocation stub using KEY. Return NULL if there is none. |
| The_reloc_stub* |
| find_reloc_stub(The_reloc_stub_key& key) |
| { |
| Reloc_stub_map_const_iter p = this->reloc_stubs_.find(key); |
| return (p != this->reloc_stubs_.end()) ? p->second : NULL; |
| } |
| |
| // Relocate reloc stubs in this stub table. This does not relocate erratum stubs. |
| void |
| relocate_reloc_stubs(const The_relocate_info*, |
| The_target_aarch64*, |
| Output_section*, |
| unsigned char*, |
| AArch64_address, |
| section_size_type); |
| |
| // Relocate an erratum stub. |
| void |
| relocate_erratum_stub(The_erratum_stub*, unsigned char*); |
| |
| // Update data size at the end of a relaxation pass. Return true if data size |
| // is different from that of the previous relaxation pass. |
| bool |
| update_data_size_changed_p() |
| { |
| // No addralign changed here. |
| off_t s = align_address(this->reloc_stubs_size_, |
| The_erratum_stub::STUB_ADDR_ALIGN) |
| + this->erratum_stubs_size_; |
| bool changed = (s != this->prev_data_size_); |
| this->prev_data_size_ = s; |
| return changed; |
| } |
| |
| protected: |
| // Write out section contents. |
| void |
| do_write(Output_file*); |
| |
| // Return the required alignment. |
| uint64_t |
| do_addralign() const |
| { |
| return std::max(The_reloc_stub::STUB_ADDR_ALIGN, |
| The_erratum_stub::STUB_ADDR_ALIGN); |
| } |
| |
| // Reset address and file offset. |
| void |
| do_reset_address_and_file_offset() |
| { this->set_current_data_size_for_child(this->prev_data_size_); } |
| |
| // Set final data size. |
| void |
| set_final_data_size() |
| { this->set_data_size(this->current_data_size()); } |
| |
| private: |
| // Relocate one reloc stub. |
| void |
| relocate_reloc_stub(The_reloc_stub*, |
| const The_relocate_info*, |
| The_target_aarch64*, |
| Output_section*, |
| unsigned char*, |
| AArch64_address, |
| section_size_type); |
| |
| private: |
| // Owner of this stub table. |
| The_aarch64_input_section* owner_; |
| // The relocation stubs. |
| Reloc_stub_map reloc_stubs_; |
| // The erratum stubs. |
| Erratum_stub_set erratum_stubs_; |
| // Size of reloc stubs. |
| off_t reloc_stubs_size_; |
| // Size of erratum stubs. |
| off_t erratum_stubs_size_; |
| // data size of this in the previous pass. |
| off_t prev_data_size_; |
| }; // End of Stub_table |
| |
| |
| // Add an erratum stub into the erratum stub set. The set is ordered by |
| // (relobj, shndx, sh_offset). |
| |
| template<int size, bool big_endian> |
| void |
| Stub_table<size, big_endian>::add_erratum_stub(The_erratum_stub* stub) |
| { |
| std::pair<Erratum_stub_set_iter, bool> ret = |
| this->erratum_stubs_.insert(stub); |
| gold_assert(ret.second); |
| this->erratum_stubs_size_ = align_address( |
| this->erratum_stubs_size_, The_erratum_stub::STUB_ADDR_ALIGN); |
| stub->set_offset(this->erratum_stubs_size_); |
| this->erratum_stubs_size_ += stub->stub_size(); |
| } |
| |
| |
| // Find if such erratum exists for given (obj, shndx, sh_offset). |
| |
| template<int size, bool big_endian> |
| Erratum_stub<size, big_endian>* |
| Stub_table<size, big_endian>::find_erratum_stub( |
| The_aarch64_relobj* a64relobj, unsigned int shndx, unsigned int sh_offset) |
| { |
| // A dummy object used as key to search in the set. |
| The_erratum_stub key(a64relobj, ST_NONE, |
| shndx, sh_offset); |
| Erratum_stub_set_iter i = this->erratum_stubs_.find(&key); |
| if (i != this->erratum_stubs_.end()) |
| { |
| The_erratum_stub* stub(*i); |
| gold_assert(stub->erratum_insn() != 0); |
| return stub; |
| } |
| return NULL; |
| } |
| |
| |
| // Find all the errata for a given input section. The return value is a pair of |
| // iterators [begin, end). |
| |
| template<int size, bool big_endian> |
| std::pair<typename Stub_table<size, big_endian>::Erratum_stub_set_iter, |
| typename Stub_table<size, big_endian>::Erratum_stub_set_iter> |
| Stub_table<size, big_endian>::find_erratum_stubs_for_input_section( |
| The_aarch64_relobj* a64relobj, unsigned int shndx) |
| { |
| typedef std::pair<Erratum_stub_set_iter, Erratum_stub_set_iter> Result_pair; |
| Erratum_stub_set_iter start, end; |
| The_erratum_stub low_key(a64relobj, ST_NONE, shndx, 0); |
| start = this->erratum_stubs_.lower_bound(&low_key); |
| if (start == this->erratum_stubs_.end()) |
| return Result_pair(this->erratum_stubs_.end(), |
| this->erratum_stubs_.end()); |
| end = start; |
| while (end != this->erratum_stubs_.end() && |
| (*end)->relobj() == a64relobj && (*end)->shndx() == shndx) |
| ++end; |
| return Result_pair(start, end); |
| } |
| |
| |
| // Add a STUB using KEY. The caller is responsible for avoiding addition |
| // if a STUB with the same key has already been added. |
| |
| template<int size, bool big_endian> |
| void |
| Stub_table<size, big_endian>::add_reloc_stub( |
| The_reloc_stub* stub, const The_reloc_stub_key& key) |
| { |
| gold_assert(stub->type() == key.type()); |
| this->reloc_stubs_[key] = stub; |
| |
| // Assign stub offset early. We can do this because we never remove |
| // reloc stubs and they are in the beginning of the stub table. |
| this->reloc_stubs_size_ = align_address(this->reloc_stubs_size_, |
| The_reloc_stub::STUB_ADDR_ALIGN); |
| stub->set_offset(this->reloc_stubs_size_); |
| this->reloc_stubs_size_ += stub->stub_size(); |
| } |
| |
| |
| // Relocate an erratum stub. |
| |
| template<int size, bool big_endian> |
| void |
| Stub_table<size, big_endian>:: |
| relocate_erratum_stub(The_erratum_stub* estub, |
| unsigned char* view) |
| { |
| // Just for convenience. |
| const int BPI = AArch64_insn_utilities<big_endian>::BYTES_PER_INSN; |
| |
| gold_assert(!estub->is_invalidated_erratum_stub()); |
| AArch64_address stub_address = this->erratum_stub_address(estub); |
| // The address of "b" in the stub that is to be "relocated". |
| AArch64_address stub_b_insn_address; |
| // Branch offset that is to be filled in "b" insn. |
| int b_offset = 0; |
| switch (estub->type()) |
| { |
| case ST_E_843419: |
| case ST_E_835769: |
| // The 1st insn of the erratum could be a relocation spot, |
| // in this case we need to fix it with |
| // "(*i)->erratum_insn()". |
| elfcpp::Swap<32, big_endian>::writeval( |
| view + (stub_address - this->address()), |
| estub->erratum_insn()); |
| // For the erratum, the 2nd insn is a b-insn to be patched |
| // (relocated). |
| stub_b_insn_address = stub_address + 1 * BPI; |
| b_offset = estub->destination_address() - stub_b_insn_address; |
| AArch64_relocate_functions<size, big_endian>::construct_b( |
| view + (stub_b_insn_address - this->address()), |
| ((unsigned int)(b_offset)) & 0xfffffff); |
| break; |
| default: |
| gold_unreachable(); |
| break; |
| } |
| estub->invalidate_erratum_stub(); |
| } |
| |
| |
| // Relocate only reloc stubs in this stub table. This does not relocate erratum |
| // stubs. |
| |
| template<int size, bool big_endian> |
| void |
| Stub_table<size, big_endian>:: |
| relocate_reloc_stubs(const The_relocate_info* relinfo, |
| The_target_aarch64* target_aarch64, |
| Output_section* output_section, |
| unsigned char* view, |
| AArch64_address address, |
| section_size_type view_size) |
| { |
| // "view_size" is the total size of the stub_table. |
| gold_assert(address == this->address() && |
| view_size == static_cast<section_size_type>(this->data_size())); |
| for(Reloc_stub_map_const_iter p = this->reloc_stubs_.begin(); |
| p != this->reloc_stubs_.end(); ++p) |
| relocate_reloc_stub(p->second, relinfo, target_aarch64, output_section, |
| view, address, view_size); |
| } |
| |
| |
| // Relocate one reloc stub. This is a helper for |
| // Stub_table::relocate_reloc_stubs(). |
| |
| template<int size, bool big_endian> |
| void |
| Stub_table<size, big_endian>:: |
| relocate_reloc_stub(The_reloc_stub* stub, |
| const The_relocate_info* relinfo, |
| The_target_aarch64* target_aarch64, |
| Output_section* output_section, |
| unsigned char* view, |
| AArch64_address address, |
| section_size_type view_size) |
| { |
| // "offset" is the offset from the beginning of the stub_table. |
| section_size_type offset = stub->offset(); |
| section_size_type stub_size = stub->stub_size(); |
| // "view_size" is the total size of the stub_table. |
| gold_assert(offset + stub_size <= view_size); |
| |
| target_aarch64->relocate_reloc_stub(stub, relinfo, output_section, |
| view + offset, address + offset, view_size); |
| } |
| |
| |
| // Write out the stubs to file. |
| |
| template<int size, bool big_endian> |
| void |
| Stub_table<size, big_endian>::do_write(Output_file* of) |
| { |
| off_t offset = this->offset(); |
| const section_size_type oview_size = |
| convert_to_section_size_type(this->data_size()); |
| unsigned char* const oview = of->get_output_view(offset, oview_size); |
| |
| // Write relocation stubs. |
| for (typename Reloc_stub_map::const_iterator p = this->reloc_stubs_.begin(); |
| p != this->reloc_stubs_.end(); ++p) |
| { |
| The_reloc_stub* stub = p->second; |
| AArch64_address address = this->address() + stub->offset(); |
| gold_assert(address == |
| align_address(address, The_reloc_stub::STUB_ADDR_ALIGN)); |
| stub->write(oview + stub->offset(), stub->stub_size()); |
| } |
| |
| // Write erratum stubs. |
| unsigned int erratum_stub_start_offset = |
| align_address(this->reloc_stubs_size_, The_erratum_stub::STUB_ADDR_ALIGN); |
| for (typename Erratum_stub_set::iterator p = this->erratum_stubs_.begin(); |
| p != this->erratum_stubs_.end(); ++p) |
| { |
| The_erratum_stub* stub(*p); |
| stub->write(oview + erratum_stub_start_offset + stub->offset(), |
| stub->stub_size()); |
| } |
| |
| of->write_output_view(this->offset(), oview_size, oview); |
| } |
| |
| |
| // AArch64_relobj class. |
| |
| template<int size, bool big_endian> |
| class AArch64_relobj : public Sized_relobj_file<size, big_endian> |
| { |
| public: |
| typedef AArch64_relobj<size, big_endian> This; |
| typedef Target_aarch64<size, big_endian> The_target_aarch64; |
| typedef AArch64_input_section<size, big_endian> The_aarch64_input_section; |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address; |
| typedef Stub_table<size, big_endian> The_stub_table; |
| typedef Erratum_stub<size, big_endian> The_erratum_stub; |
| typedef typename The_stub_table::Erratum_stub_set_iter Erratum_stub_set_iter; |
| typedef std::vector<The_stub_table*> Stub_table_list; |
| static const AArch64_address invalid_address = |
| static_cast<AArch64_address>(-1); |
| |
| AArch64_relobj(const std::string& name, Input_file* input_file, off_t offset, |
| const typename elfcpp::Ehdr<size, big_endian>& ehdr) |
| : Sized_relobj_file<size, big_endian>(name, input_file, offset, ehdr), |
| stub_tables_() |
| { } |
| |
| ~AArch64_relobj() |
| { } |
| |
| // Return the stub table of the SHNDX-th section if there is one. |
| The_stub_table* |
| stub_table(unsigned int shndx) const |
| { |
| gold_assert(shndx < this->stub_tables_.size()); |
| return this->stub_tables_[shndx]; |
| } |
| |
| // Set STUB_TABLE to be the stub_table of the SHNDX-th section. |
| void |
| set_stub_table(unsigned int shndx, The_stub_table* stub_table) |
| { |
| gold_assert(shndx < this->stub_tables_.size()); |
| this->stub_tables_[shndx] = stub_table; |
| } |
| |
| // Entrance to errata scanning. |
| void |
| scan_errata(unsigned int shndx, |
| const elfcpp::Shdr<size, big_endian>&, |
| Output_section*, const Symbol_table*, |
| The_target_aarch64*); |
| |
| // Scan all relocation sections for stub generation. |
| void |
| scan_sections_for_stubs(The_target_aarch64*, const Symbol_table*, |
| const Layout*); |
| |
| // Whether a section is a scannable text section. |
| bool |
| text_section_is_scannable(const elfcpp::Shdr<size, big_endian>&, unsigned int, |
| const Output_section*, const Symbol_table*); |
| |
| // Convert regular input section with index SHNDX to a relaxed section. |
| void |
| convert_input_section_to_relaxed_section(unsigned shndx) |
| { |
| // The stubs have relocations and we need to process them after writing |
| // out the stubs. So relocation now must follow section write. |
| this->set_section_offset(shndx, -1ULL); |
| this->set_relocs_must_follow_section_writes(); |
| } |
| |
| // Structure for mapping symbol position. |
| struct Mapping_symbol_position |
| { |
| Mapping_symbol_position(unsigned int shndx, AArch64_address offset): |
| shndx_(shndx), offset_(offset) |
| {} |
| |
| // "<" comparator used in ordered_map container. |
| bool |
| operator<(const Mapping_symbol_position& p) const |
| { |
| return (this->shndx_ < p.shndx_ |
| || (this->shndx_ == p.shndx_ && this->offset_ < p.offset_)); |
| } |
| |
| // Section index. |
| unsigned int shndx_; |
| |
| // Section offset. |
| AArch64_address offset_; |
| }; |
| |
| typedef std::map<Mapping_symbol_position, char> Mapping_symbol_info; |
| |
| protected: |
| // Post constructor setup. |
| void |
| do_setup() |
| { |
| // Call parent's setup method. |
| Sized_relobj_file<size, big_endian>::do_setup(); |
| |
| // Initialize look-up tables. |
| this->stub_tables_.resize(this->shnum()); |
| } |
| |
| virtual void |
| do_relocate_sections( |
| const Symbol_table* symtab, const Layout* layout, |
| const unsigned char* pshdrs, Output_file* of, |
| typename Sized_relobj_file<size, big_endian>::Views* pviews); |
| |
| // Count local symbols and (optionally) record mapping info. |
| virtual void |
| do_count_local_symbols(Stringpool_template<char>*, |
| Stringpool_template<char>*); |
| |
| private: |
| // Fix all errata in the object, and for each erratum, relocate corresponding |
| // erratum stub. |
| void |
| fix_errata_and_relocate_erratum_stubs( |
| typename Sized_relobj_file<size, big_endian>::Views* pviews); |
| |
| // Try to fix erratum 843419 in an optimized way. Return true if patch is |
| // applied. |
| bool |
| try_fix_erratum_843419_optimized( |
| The_erratum_stub*, AArch64_address, |
| typename Sized_relobj_file<size, big_endian>::View_size&); |
| |
| // Whether a section needs to be scanned for relocation stubs. |
| bool |
| section_needs_reloc_stub_scanning(const elfcpp::Shdr<size, big_endian>&, |
| const Relobj::Output_sections&, |
| const Symbol_table*, const unsigned char*); |
| |
| // List of stub tables. |
| Stub_table_list stub_tables_; |
| |
| // Mapping symbol information sorted by (section index, section_offset). |
| Mapping_symbol_info mapping_symbol_info_; |
| }; // End of AArch64_relobj |
| |
| |
| // Override to record mapping symbol information. |
| template<int size, bool big_endian> |
| void |
| AArch64_relobj<size, big_endian>::do_count_local_symbols( |
| Stringpool_template<char>* pool, Stringpool_template<char>* dynpool) |
| { |
| Sized_relobj_file<size, big_endian>::do_count_local_symbols(pool, dynpool); |
| |
| // Only erratum-fixing work needs mapping symbols, so skip this time consuming |
| // processing if not fixing erratum. |
| if (!parameters->options().fix_cortex_a53_843419() |
| && !parameters->options().fix_cortex_a53_835769()) |
| return; |
| |
| const unsigned int loccount = this->local_symbol_count(); |
| if (loccount == 0) |
| return; |
| |
| // Read the symbol table section header. |
| const unsigned int symtab_shndx = this->symtab_shndx(); |
| elfcpp::Shdr<size, big_endian> |
| symtabshdr(this, this->elf_file()->section_header(symtab_shndx)); |
| gold_assert(symtabshdr.get_sh_type() == elfcpp::SHT_SYMTAB); |
| |
| // Read the local symbols. |
| const int sym_size =elfcpp::Elf_sizes<size>::sym_size; |
| gold_assert(loccount == symtabshdr.get_sh_info()); |
| off_t locsize = loccount * sym_size; |
| const unsigned char* psyms = this->get_view(symtabshdr.get_sh_offset(), |
| locsize, true, true); |
| |
| // For mapping symbol processing, we need to read the symbol names. |
| unsigned int strtab_shndx = this->adjust_shndx(symtabshdr.get_sh_link()); |
| if (strtab_shndx >= this->shnum()) |
| { |
| this->error(_("invalid symbol table name index: %u"), strtab_shndx); |
| return; |
| } |
| |
| elfcpp::Shdr<size, big_endian> |
| strtabshdr(this, this->elf_file()->section_header(strtab_shndx)); |
| if (strtabshdr.get_sh_type() != elfcpp::SHT_STRTAB) |
| { |
| this->error(_("symbol table name section has wrong type: %u"), |
| static_cast<unsigned int>(strtabshdr.get_sh_type())); |
| return; |
| } |
| |
| const char* pnames = |
| reinterpret_cast<const char*>(this->get_view(strtabshdr.get_sh_offset(), |
| strtabshdr.get_sh_size(), |
| false, false)); |
| |
| // Skip the first dummy symbol. |
| psyms += sym_size; |
| typename Sized_relobj_file<size, big_endian>::Local_values* |
| plocal_values = this->local_values(); |
| for (unsigned int i = 1; i < loccount; ++i, psyms += sym_size) |
| { |
| elfcpp::Sym<size, big_endian> sym(psyms); |
| Symbol_value<size>& lv((*plocal_values)[i]); |
| AArch64_address input_value = lv.input_value(); |
| |
| // Check to see if this is a mapping symbol. AArch64 mapping symbols are |
| // defined in "ELF for the ARM 64-bit Architecture", Table 4-4, Mapping |
| // symbols. |
| // Mapping symbols could be one of the following 4 forms - |
| // a) $x |
| // b) $x.<any...> |
| // c) $d |
| // d) $d.<any...> |
| const char* sym_name = pnames + sym.get_st_name(); |
| if (sym_name[0] == '$' && (sym_name[1] == 'x' || sym_name[1] == 'd') |
| && (sym_name[2] == '\0' || sym_name[2] == '.')) |
| { |
| bool is_ordinary; |
| unsigned int input_shndx = |
| this->adjust_sym_shndx(i, sym.get_st_shndx(), &is_ordinary); |
| gold_assert(is_ordinary); |
| |
| Mapping_symbol_position msp(input_shndx, input_value); |
| // Insert mapping_symbol_info into map whose ordering is defined by |
| // (shndx, offset_within_section). |
| this->mapping_symbol_info_[msp] = sym_name[1]; |
| } |
| } |
| } |
| |
| |
| // Fix all errata in the object and for each erratum, we relocate the |
| // corresponding erratum stub (by calling Stub_table::relocate_erratum_stub). |
| |
| template<int size, bool big_endian> |
| void |
| AArch64_relobj<size, big_endian>::fix_errata_and_relocate_erratum_stubs( |
| typename Sized_relobj_file<size, big_endian>::Views* pviews) |
| { |
| typedef typename elfcpp::Swap<32,big_endian>::Valtype Insntype; |
| unsigned int shnum = this->shnum(); |
| const Relobj::Output_sections& out_sections(this->output_sections()); |
| for (unsigned int i = 1; i < shnum; ++i) |
| { |
| The_stub_table* stub_table = this->stub_table(i); |
| if (!stub_table) |
| continue; |
| std::pair<Erratum_stub_set_iter, Erratum_stub_set_iter> |
| ipair(stub_table->find_erratum_stubs_for_input_section(this, i)); |
| Erratum_stub_set_iter p = ipair.first, end = ipair.second; |
| typename Sized_relobj_file<size, big_endian>::View_size& |
| pview((*pviews)[i]); |
| AArch64_address view_offset = 0; |
| if (pview.is_input_output_view) |
| { |
| // In this case, write_sections has not added the output offset to |
| // the view's address, so we must do so. Currently this only happens |
| // for a relaxed section. |
| unsigned int index = this->adjust_shndx(i); |
| const Output_relaxed_input_section* poris = |
| out_sections[index]->find_relaxed_input_section(this, index); |
| gold_assert(poris != NULL); |
| view_offset = poris->address() - pview.address; |
| } |
| |
| while (p != end) |
| { |
| The_erratum_stub* stub = *p; |
| |
| // Double check data before fix. |
| gold_assert(pview.address + view_offset + stub->sh_offset() |
| == stub->erratum_address()); |
| |
| // Update previously recorded erratum insn with relocated |
| // version. |
| Insntype* ip = |
| reinterpret_cast<Insntype*>( |
| pview.view + view_offset + stub->sh_offset()); |
| Insntype insn_to_fix = ip[0]; |
| stub->update_erratum_insn(insn_to_fix); |
| |
| // First try to see if erratum is 843419 and if it can be fixed |
| // without using branch-to-stub. |
| if (!try_fix_erratum_843419_optimized(stub, view_offset, pview)) |
| { |
| // Replace the erratum insn with a branch-to-stub. |
| AArch64_address stub_address = |
| stub_table->erratum_stub_address(stub); |
| unsigned int b_offset = stub_address - stub->erratum_address(); |
| AArch64_relocate_functions<size, big_endian>::construct_b( |
| pview.view + view_offset + stub->sh_offset(), |
| b_offset & 0xfffffff); |
| } |
| |
| // Erratum fix is done (or skipped), continue to relocate erratum |
| // stub. Note, when erratum fix is skipped (either because we |
| // proactively change the code sequence or the code sequence is |
| // changed by relaxation, etc), we can still safely relocate the |
| // erratum stub, ignoring the fact the erratum could never be |
| // executed. |
| stub_table->relocate_erratum_stub( |
| stub, |
| pview.view + (stub_table->address() - pview.address)); |
| |
| // Next erratum stub. |
| ++p; |
| } |
| } |
| } |
| |
| |
| // This is an optimization for 843419. This erratum requires the sequence begin |
| // with 'adrp', when final value calculated by adrp fits in adr, we can just |
| // replace 'adrp' with 'adr', so we save 2 jumps per occurrence. (Note, however, |
| // in this case, we do not delete the erratum stub (too late to do so), it is |
| // merely generated without ever being called.) |
| |
| template<int size, bool big_endian> |
| bool |
| AArch64_relobj<size, big_endian>::try_fix_erratum_843419_optimized( |
| The_erratum_stub* stub, AArch64_address view_offset, |
| typename Sized_relobj_file<size, big_endian>::View_size& pview) |
| { |
| if (stub->type() != ST_E_843419) |
| return false; |
| |
| typedef AArch64_insn_utilities<big_endian> Insn_utilities; |
| typedef typename elfcpp::Swap<32,big_endian>::Valtype Insntype; |
| E843419_stub<size, big_endian>* e843419_stub = |
| reinterpret_cast<E843419_stub<size, big_endian>*>(stub); |
| AArch64_address pc = |
| pview.address + view_offset + e843419_stub->adrp_sh_offset(); |
| unsigned int adrp_offset = e843419_stub->adrp_sh_offset (); |
| Insntype* adrp_view = |
| reinterpret_cast<Insntype*>(pview.view + view_offset + adrp_offset); |
| Insntype adrp_insn = adrp_view[0]; |
| |
| // If the instruction at adrp_sh_offset is "mrs R, tpidr_el0", it may come |
| // from IE -> LE relaxation etc. This is a side-effect of TLS relaxation that |
| // ADRP has been turned into MRS, there is no erratum risk anymore. |
| // Therefore, we return true to avoid doing unnecessary branch-to-stub. |
| if (Insn_utilities::is_mrs_tpidr_el0(adrp_insn)) |
| return true; |
| |
| // If the instruction at adrp_sh_offset is not ADRP and the instruction before |
| // it is "mrs R, tpidr_el0", it may come from LD -> LE relaxation etc. |
| // Like the above case, there is no erratum risk any more, we can safely |
| // return true. |
| if (!Insn_utilities::is_adrp(adrp_insn) && adrp_offset) |
| { |
| Insntype* prev_view = |
| reinterpret_cast<Insntype*>( |
| pview.view + view_offset + adrp_offset - 4); |
| Insntype prev_insn = prev_view[0]; |
| |
| if (Insn_utilities::is_mrs_tpidr_el0(prev_insn)) |
| return true; |
| } |
| |
| /* If we reach here, the first instruction must be ADRP. */ |
| gold_assert(Insn_utilities::is_adrp(adrp_insn)); |
| // Get adrp 33-bit signed imm value. |
| int64_t adrp_imm = Insn_utilities:: |
| aarch64_adrp_decode_imm(adrp_insn); |
| // adrp - final value transferred to target register is calculated as: |
| // PC[11:0] = Zeros(12) |
| // adrp_dest_value = PC + adrp_imm; |
| int64_t adrp_dest_value = (pc & ~((1 << 12) - 1)) + adrp_imm; |
| // adr -final value transferred to target register is calucalted as: |
| // PC + adr_imm |
| // So we have: |
| // PC + adr_imm = adrp_dest_value |
| // ==> |
| // adr_imm = adrp_dest_value - PC |
| int64_t adr_imm = adrp_dest_value - pc; |
| // Check if imm fits in adr (21-bit signed). |
| if (-(1 << 20) <= adr_imm && adr_imm < (1 << 20)) |
| { |
| // Convert 'adrp' into 'adr'. |
| Insntype adr_insn = adrp_insn & ((1u << 31) - 1); |
| adr_insn = Insn_utilities:: |
| aarch64_adr_encode_imm(adr_insn, adr_imm); |
| elfcpp::Swap<32, big_endian>::writeval(adrp_view, adr_insn); |
| return true; |
| } |
| return false; |
| } |
| |
| |
| // Relocate sections. |
| |
| template<int size, bool big_endian> |
| void |
| AArch64_relobj<size, big_endian>::do_relocate_sections( |
| const Symbol_table* symtab, const Layout* layout, |
| const unsigned char* pshdrs, Output_file* of, |
| typename Sized_relobj_file<size, big_endian>::Views* pviews) |
| { |
| // Relocate the section data. |
| this->relocate_section_range(symtab, layout, pshdrs, of, pviews, |
| 1, this->shnum() - 1); |
| |
| // We do not generate stubs if doing a relocatable link. |
| if (parameters->options().relocatable()) |
| return; |
| |
| // This part only relocates erratum stubs that belong to input sections of this |
| // object file. |
| if (parameters->options().fix_cortex_a53_843419() |
| || parameters->options().fix_cortex_a53_835769()) |
| this->fix_errata_and_relocate_erratum_stubs(pviews); |
| |
| Relocate_info<size, big_endian> relinfo; |
| relinfo.symtab = symtab; |
| relinfo.layout = layout; |
| relinfo.object = this; |
| |
| // This part relocates all reloc stubs that are contained in stub_tables of |
| // this object file. |
| unsigned int shnum = this->shnum(); |
| The_target_aarch64* target = The_target_aarch64::current_target(); |
| |
| for (unsigned int i = 1; i < shnum; ++i) |
| { |
| The_aarch64_input_section* aarch64_input_section = |
| target->find_aarch64_input_section(this, i); |
| if (aarch64_input_section != NULL |
| && aarch64_input_section->is_stub_table_owner() |
| && !aarch64_input_section->stub_table()->empty()) |
| { |
| Output_section* os = this->output_section(i); |
| gold_assert(os != NULL); |
| |
| relinfo.reloc_shndx = elfcpp::SHN_UNDEF; |
| relinfo.reloc_shdr = NULL; |
| relinfo.data_shndx = i; |
| relinfo.data_shdr = pshdrs + i * elfcpp::Elf_sizes<size>::shdr_size; |
| |
| typename Sized_relobj_file<size, big_endian>::View_size& |
| view_struct = (*pviews)[i]; |
| gold_assert(view_struct.view != NULL); |
| |
| The_stub_table* stub_table = aarch64_input_section->stub_table(); |
| off_t offset = stub_table->address() - view_struct.address; |
| unsigned char* view = view_struct.view + offset; |
| AArch64_address address = stub_table->address(); |
| section_size_type view_size = stub_table->data_size(); |
| stub_table->relocate_reloc_stubs(&relinfo, target, os, view, address, |
| view_size); |
| } |
| } |
| } |
| |
| |
| // Determine if an input section is scannable for stub processing. SHDR is |
| // the header of the section and SHNDX is the section index. OS is the output |
| // section for the input section and SYMTAB is the global symbol table used to |
| // look up ICF information. |
| |
| template<int size, bool big_endian> |
| bool |
| AArch64_relobj<size, big_endian>::text_section_is_scannable( |
| const elfcpp::Shdr<size, big_endian>& text_shdr, |
| unsigned int text_shndx, |
| const Output_section* os, |
| const Symbol_table* symtab) |
| { |
| // Skip any empty sections, unallocated sections or sections whose |
| // type are not SHT_PROGBITS. |
| if (text_shdr.get_sh_size() == 0 |
| || (text_shdr.get_sh_flags() & elfcpp::SHF_ALLOC) == 0 |
| || text_shdr.get_sh_type() != elfcpp::SHT_PROGBITS) |
| return false; |
| |
| // Skip any discarded or ICF'ed sections. |
| if (os == NULL || symtab->is_section_folded(this, text_shndx)) |
| return false; |
| |
| // Skip exception frame. |
| if (strcmp(os->name(), ".eh_frame") == 0) |
| return false ; |
| |
| gold_assert(!this->is_output_section_offset_invalid(text_shndx) || |
| os->find_relaxed_input_section(this, text_shndx) != NULL); |
| |
| return true; |
| } |
| |
| |
| // Determine if we want to scan the SHNDX-th section for relocation stubs. |
| // This is a helper for AArch64_relobj::scan_sections_for_stubs(). |
| |
| template<int size, bool big_endian> |
| bool |
| AArch64_relobj<size, big_endian>::section_needs_reloc_stub_scanning( |
| const elfcpp::Shdr<size, big_endian>& shdr, |
| const Relobj::Output_sections& out_sections, |
| const Symbol_table* symtab, |
| const unsigned char* pshdrs) |
| { |
| unsigned int sh_type = shdr.get_sh_type(); |
| if (sh_type != elfcpp::SHT_RELA) |
| return false; |
| |
| // Ignore empty section. |
| off_t sh_size = shdr.get_sh_size(); |
| if (sh_size == 0) |
| return false; |
| |
| // Ignore reloc section with unexpected symbol table. The |
| // error will be reported in the final link. |
| if (this->adjust_shndx(shdr.get_sh_link()) != this->symtab_shndx()) |
| return false; |
| |
| gold_assert(sh_type == elfcpp::SHT_RELA); |
| unsigned int reloc_size = elfcpp::Elf_sizes<size>::rela_size; |
| |
| // Ignore reloc section with unexpected entsize or uneven size. |
| // The error will be reported in the final link. |
| if (reloc_size != shdr.get_sh_entsize() || sh_size % reloc_size != 0) |
| return false; |
| |
| // Ignore reloc section with bad info. This error will be |
| // reported in the final link. |
| unsigned int text_shndx = this->adjust_shndx(shdr.get_sh_info()); |
| if (text_shndx >= this->shnum()) |
| return false; |
| |
| const unsigned int shdr_size = elfcpp::Elf_sizes<size>::shdr_size; |
| const elfcpp::Shdr<size, big_endian> text_shdr(pshdrs + |
| text_shndx * shdr_size); |
| return this->text_section_is_scannable(text_shdr, text_shndx, |
| out_sections[text_shndx], symtab); |
| } |
| |
| |
| // Scan section SHNDX for erratum 843419 and 835769. |
| |
| template<int size, bool big_endian> |
| void |
| AArch64_relobj<size, big_endian>::scan_errata( |
| unsigned int shndx, const elfcpp::Shdr<size, big_endian>& shdr, |
| Output_section* os, const Symbol_table* symtab, |
| The_target_aarch64* target) |
| { |
| if (shdr.get_sh_size() == 0 |
| || (shdr.get_sh_flags() & |
| (elfcpp::SHF_ALLOC | elfcpp::SHF_EXECINSTR)) == 0 |
| || shdr.get_sh_type() != elfcpp::SHT_PROGBITS) |
| return; |
| |
| if (!os || symtab->is_section_folded(this, shndx)) return; |
| |
| AArch64_address output_offset = this->get_output_section_offset(shndx); |
| AArch64_address output_address; |
| if (output_offset != invalid_address) |
| output_address = os->address() + output_offset; |
| else |
| { |
| const Output_relaxed_input_section* poris = |
| os->find_relaxed_input_section(this, shndx); |
| if (!poris) return; |
| output_address = poris->address(); |
| } |
| |
| // Update the addresses in previously generated erratum stubs. Unlike when |
| // we scan relocations for stubs, if section addresses have changed due to |
| // other relaxations we are unlikely to scan the same erratum instances |
| // again. |
| The_stub_table* stub_table = this->stub_table(shndx); |
| if (stub_table) |
| { |
| std::pair<Erratum_stub_set_iter, Erratum_stub_set_iter> |
| ipair(stub_table->find_erratum_stubs_for_input_section(this, shndx)); |
| for (Erratum_stub_set_iter p = ipair.first; p != ipair.second; ++p) |
| (*p)->update_erratum_address(output_address); |
| } |
| |
| section_size_type input_view_size = 0; |
| const unsigned char* input_view = |
| this->section_contents(shndx, &input_view_size, false); |
| |
| Mapping_symbol_position section_start(shndx, 0); |
| // Find the first mapping symbol record within section shndx. |
| typename Mapping_symbol_info::const_iterator p = |
| this->mapping_symbol_info_.lower_bound(section_start); |
| while (p != this->mapping_symbol_info_.end() && |
| p->first.shndx_ == shndx) |
| { |
| typename Mapping_symbol_info::const_iterator prev = p; |
| ++p; |
| if (prev->second == 'x') |
| { |
| section_size_type span_start = |
| convert_to_section_size_type(prev->first.offset_); |
| section_size_type span_end; |
| if (p != this->mapping_symbol_info_.end() |
| && p->first.shndx_ == shndx) |
| span_end = convert_to_section_size_type(p->first.offset_); |
| else |
| span_end = convert_to_section_size_type(shdr.get_sh_size()); |
| |
| // Here we do not share the scanning code of both errata. For 843419, |
| // only the last few insns of each page are examined, which is fast, |
| // whereas, for 835769, every insn pair needs to be checked. |
| |
| if (parameters->options().fix_cortex_a53_843419()) |
| target->scan_erratum_843419_span( |
| this, shndx, span_start, span_end, |
| const_cast<unsigned char*>(input_view), output_address); |
| |
| if (parameters->options().fix_cortex_a53_835769()) |
| target->scan_erratum_835769_span( |
| this, shndx, span_start, span_end, |
| const_cast<unsigned char*>(input_view), output_address); |
| } |
| } |
| } |
| |
| |
| // Scan relocations for stub generation. |
| |
| template<int size, bool big_endian> |
| void |
| AArch64_relobj<size, big_endian>::scan_sections_for_stubs( |
| The_target_aarch64* target, |
| const Symbol_table* symtab, |
| const Layout* layout) |
| { |
| unsigned int shnum = this->shnum(); |
| const unsigned int shdr_size = elfcpp::Elf_sizes<size>::shdr_size; |
| |
| // Read the section headers. |
| const unsigned char* pshdrs = this->get_view(this->elf_file()->shoff(), |
| shnum * shdr_size, |
| true, true); |
| |
| // To speed up processing, we set up hash tables for fast lookup of |
| // input offsets to output addresses. |
| this->initialize_input_to_output_maps(); |
| |
| const Relobj::Output_sections& out_sections(this->output_sections()); |
| |
| Relocate_info<size, big_endian> relinfo; |
| relinfo.symtab = symtab; |
| relinfo.layout = layout; |
| relinfo.object = this; |
| |
| // Do relocation stubs scanning. |
| const unsigned char* p = pshdrs + shdr_size; |
| for (unsigned int i = 1; i < shnum; ++i, p += shdr_size) |
| { |
| const elfcpp::Shdr<size, big_endian> shdr(p); |
| if (parameters->options().fix_cortex_a53_843419() |
| || parameters->options().fix_cortex_a53_835769()) |
| scan_errata(i, shdr, out_sections[i], symtab, target); |
| if (this->section_needs_reloc_stub_scanning(shdr, out_sections, symtab, |
| pshdrs)) |
| { |
| unsigned int index = this->adjust_shndx(shdr.get_sh_info()); |
| AArch64_address output_offset = |
| this->get_output_section_offset(index); |
| AArch64_address output_address; |
| if (output_offset != invalid_address) |
| { |
| output_address = out_sections[index]->address() + output_offset; |
| } |
| else |
| { |
| // Currently this only happens for a relaxed section. |
| const Output_relaxed_input_section* poris = |
| out_sections[index]->find_relaxed_input_section(this, index); |
| gold_assert(poris != NULL); |
| output_address = poris->address(); |
| } |
| |
| // Get the relocations. |
| const unsigned char* prelocs = this->get_view(shdr.get_sh_offset(), |
| shdr.get_sh_size(), |
| true, false); |
| |
| // Get the section contents. |
| section_size_type input_view_size = 0; |
| const unsigned char* input_view = |
| this->section_contents(index, &input_view_size, false); |
| |
| relinfo.reloc_shndx = i; |
| relinfo.data_shndx = index; |
| unsigned int sh_type = shdr.get_sh_type(); |
| unsigned int reloc_size; |
| gold_assert (sh_type == elfcpp::SHT_RELA); |
| reloc_size = elfcpp::Elf_sizes<size>::rela_size; |
| |
| Output_section* os = out_sections[index]; |
| target->scan_section_for_stubs(&relinfo, sh_type, prelocs, |
| shdr.get_sh_size() / reloc_size, |
| os, |
| output_offset == invalid_address, |
| input_view, output_address, |
| input_view_size); |
| } |
| } |
| } |
| |
| |
| // A class to wrap an ordinary input section containing executable code. |
| |
| template<int size, bool big_endian> |
| class AArch64_input_section : public Output_relaxed_input_section |
| { |
| public: |
| typedef Stub_table<size, big_endian> The_stub_table; |
| |
| AArch64_input_section(Relobj* relobj, unsigned int shndx) |
| : Output_relaxed_input_section(relobj, shndx, 1), |
| stub_table_(NULL), |
| original_contents_(NULL), original_size_(0), |
| original_addralign_(1) |
| { } |
| |
| ~AArch64_input_section() |
| { delete[] this->original_contents_; } |
| |
| // Initialize. |
| void |
| init(); |
| |
| // Set the stub_table. |
| void |
| set_stub_table(The_stub_table* st) |
| { this->stub_table_ = st; } |
| |
| // Whether this is a stub table owner. |
| bool |
| is_stub_table_owner() const |
| { return this->stub_table_ != NULL && this->stub_table_->owner() == this; } |
| |
| // Return the original size of the section. |
| uint32_t |
| original_size() const |
| { return this->original_size_; } |
| |
| // Return the stub table. |
| The_stub_table* |
| stub_table() |
| { return stub_table_; } |
| |
| protected: |
| // Write out this input section. |
| void |
| do_write(Output_file*); |
| |
| // Return required alignment of this. |
| uint64_t |
| do_addralign() const |
| { |
| if (this->is_stub_table_owner()) |
| return std::max(this->stub_table_->addralign(), |
| static_cast<uint64_t>(this->original_addralign_)); |
| else |
| return this->original_addralign_; |
| } |
| |
| // Finalize data size. |
| void |
| set_final_data_size(); |
| |
| // Reset address and file offset. |
| void |
| do_reset_address_and_file_offset(); |
| |
| // Output offset. |
| bool |
| do_output_offset(const Relobj* object, unsigned int shndx, |
| section_offset_type offset, |
| section_offset_type* poutput) const |
| { |
| if ((object == this->relobj()) |
| && (shndx == this->shndx()) |
| && (offset >= 0) |
| && (offset <= |
| convert_types<section_offset_type, uint32_t>(this->original_size_))) |
| { |
| *poutput = offset; |
| return true; |
| } |
| else |
| return false; |
| } |
| |
| private: |
| // Copying is not allowed. |
| AArch64_input_section(const AArch64_input_section&); |
| AArch64_input_section& operator=(const AArch64_input_section&); |
| |
| // The relocation stubs. |
| The_stub_table* stub_table_; |
| // Original section contents. We have to make a copy here since the file |
| // containing the original section may not be locked when we need to access |
| // the contents. |
| unsigned char* original_contents_; |
| // Section size of the original input section. |
| uint32_t original_size_; |
| // Address alignment of the original input section. |
| uint32_t original_addralign_; |
| }; // End of AArch64_input_section |
| |
| |
| // Finalize data size. |
| |
| template<int size, bool big_endian> |
| void |
| AArch64_input_section<size, big_endian>::set_final_data_size() |
| { |
| off_t off = convert_types<off_t, uint64_t>(this->original_size_); |
| |
| if (this->is_stub_table_owner()) |
| { |
| this->stub_table_->finalize_data_size(); |
| off = align_address(off, this->stub_table_->addralign()); |
| off += this->stub_table_->data_size(); |
| } |
| this->set_data_size(off); |
| } |
| |
| |
| // Reset address and file offset. |
| |
| template<int size, bool big_endian> |
| void |
| AArch64_input_section<size, big_endian>::do_reset_address_and_file_offset() |
| { |
| // Size of the original input section contents. |
| off_t off = convert_types<off_t, uint64_t>(this->original_size_); |
| |
| // If this is a stub table owner, account for the stub table size. |
| if (this->is_stub_table_owner()) |
| { |
| The_stub_table* stub_table = this->stub_table_; |
| |
| // Reset the stub table's address and file offset. The |
| // current data size for child will be updated after that. |
| stub_table_->reset_address_and_file_offset(); |
| off = align_address(off, stub_table_->addralign()); |
| off += stub_table->current_data_size(); |
| } |
| |
| this->set_current_data_size(off); |
| } |
| |
| |
| // Initialize an Arm_input_section. |
| |
| template<int size, bool big_endian> |
| void |
| AArch64_input_section<size, big_endian>::init() |
| { |
| Relobj* relobj = this->relobj(); |
| unsigned int shndx = this->shndx(); |
| |
| // We have to cache original size, alignment and contents to avoid locking |
| // the original file. |
| this->original_addralign_ = |
| convert_types<uint32_t, uint64_t>(relobj->section_addralign(shndx)); |
| |
| // This is not efficient but we expect only a small number of relaxed |
| // input sections for stubs. |
| section_size_type section_size; |
| const unsigned char* section_contents = |
| relobj->section_contents(shndx, §ion_size, false); |
| this->original_size_ = |
| convert_types<uint32_t, uint64_t>(relobj->section_size(shndx)); |
| |
| gold_assert(this->original_contents_ == NULL); |
| this->original_contents_ = new unsigned char[section_size]; |
| memcpy(this->original_contents_, section_contents, section_size); |
| |
| // We want to make this look like the original input section after |
| // output sections are finalized. |
| Output_section* os = relobj->output_section(shndx); |
| off_t offset = relobj->output_section_offset(shndx); |
| gold_assert(os != NULL && !relobj->is_output_section_offset_invalid(shndx)); |
| this->set_address(os->address() + offset); |
| this->set_file_offset(os->offset() + offset); |
| this->set_current_data_size(this->original_size_); |
| this->finalize_data_size(); |
| } |
| |
| |
| // Write data to output file. |
| |
| template<int size, bool big_endian> |
| void |
| AArch64_input_section<size, big_endian>::do_write(Output_file* of) |
| { |
| // We have to write out the original section content. |
| gold_assert(this->original_contents_ != NULL); |
| of->write(this->offset(), this->original_contents_, |
| this->original_size_); |
| |
| // If this owns a stub table and it is not empty, write it. |
| if (this->is_stub_table_owner() && !this->stub_table_->empty()) |
| this->stub_table_->write(of); |
| } |
| |
| |
| // Arm output section class. This is defined mainly to add a number of stub |
| // generation methods. |
| |
| template<int size, bool big_endian> |
| class AArch64_output_section : public Output_section |
| { |
| public: |
| typedef Target_aarch64<size, big_endian> The_target_aarch64; |
| typedef AArch64_relobj<size, big_endian> The_aarch64_relobj; |
| typedef Stub_table<size, big_endian> The_stub_table; |
| typedef AArch64_input_section<size, big_endian> The_aarch64_input_section; |
| |
| public: |
| AArch64_output_section(const char* name, elfcpp::Elf_Word type, |
| elfcpp::Elf_Xword flags) |
| : Output_section(name, type, flags) |
| { } |
| |
| ~AArch64_output_section() {} |
| |
| // Group input sections for stub generation. |
| void |
| group_sections(section_size_type, bool, Target_aarch64<size, big_endian>*, |
| const Task*); |
| |
| private: |
| typedef Output_section::Input_section Input_section; |
| typedef Output_section::Input_section_list Input_section_list; |
| |
| // Create a stub group. |
| void |
| create_stub_group(Input_section_list::const_iterator, |
| Input_section_list::const_iterator, |
| Input_section_list::const_iterator, |
| The_target_aarch64*, |
| std::vector<Output_relaxed_input_section*>&, |
| const Task*); |
| }; // End of AArch64_output_section |
| |
| |
| // Create a stub group for input sections from FIRST to LAST. OWNER points to |
| // the input section that will be the owner of the stub table. |
| |
| template<int size, bool big_endian> void |
| AArch64_output_section<size, big_endian>::create_stub_group( |
| Input_section_list::const_iterator first, |
| Input_section_list::const_iterator last, |
| Input_section_list::const_iterator owner, |
| The_target_aarch64* target, |
| std::vector<Output_relaxed_input_section*>& new_relaxed_sections, |
| const Task* task) |
| { |
| // Currently we convert ordinary input sections into relaxed sections only |
| // at this point. |
| The_aarch64_input_section* input_section; |
| if (owner->is_relaxed_input_section()) |
| gold_unreachable(); |
| else |
| { |
| gold_assert(owner->is_input_section()); |
| // Create a new relaxed input section. We need to lock the original |
| // file. |
| Task_lock_obj<Object> tl(task, owner->relobj()); |
| input_section = |
| target->new_aarch64_input_section(owner->relobj(), owner->shndx()); |
| new_relaxed_sections.push_back(input_section); |
| } |
| |
| // Create a stub table. |
| The_stub_table* stub_table = |
| target->new_stub_table(input_section); |
| |
| input_section->set_stub_table(stub_table); |
| |
| Input_section_list::const_iterator p = first; |
| // Look for input sections or relaxed input sections in [first ... last]. |
| do |
| { |
| if (p->is_input_section() || p->is_relaxed_input_section()) |
| { |
| // The stub table information for input sections live |
| // in their objects. |
| The_aarch64_relobj* aarch64_relobj = |
| static_cast<The_aarch64_relobj*>(p->relobj()); |
| aarch64_relobj->set_stub_table(p->shndx(), stub_table); |
| } |
| } |
| while (p++ != last); |
| } |
| |
| |
| // Group input sections for stub generation. GROUP_SIZE is roughly the limit of |
| // stub groups. We grow a stub group by adding input section until the size is |
| // just below GROUP_SIZE. The last input section will be converted into a stub |
| // table owner. If STUB_ALWAYS_AFTER_BRANCH is false, we also add input sectiond |
| // after the stub table, effectively doubling the group size. |
| // |
| // This is similar to the group_sections() function in elf32-arm.c but is |
| // implemented differently. |
| |
| template<int size, bool big_endian> |
| void AArch64_output_section<size, big_endian>::group_sections( |
| section_size_type group_size, |
| bool stubs_always_after_branch, |
| Target_aarch64<size, big_endian>* target, |
| const Task* task) |
| { |
| typedef enum |
| { |
| NO_GROUP, |
| FINDING_STUB_SECTION, |
| HAS_STUB_SECTION |
| } State; |
| |
| std::vector<Output_relaxed_input_section*> new_relaxed_sections; |
| |
| State state = NO_GROUP; |
| section_size_type off = 0; |
| section_size_type group_begin_offset = 0; |
| section_size_type group_end_offset = 0; |
| section_size_type stub_table_end_offset = 0; |
| Input_section_list::const_iterator group_begin = |
| this->input_sections().end(); |
| Input_section_list::const_iterator stub_table = |
| this->input_sections().end(); |
| Input_section_list::const_iterator group_end = this->input_sections().end(); |
| for (Input_section_list::const_iterator p = this->input_sections().begin(); |
| p != this->input_sections().end(); |
| ++p) |
| { |
| section_size_type section_begin_offset = |
| align_address(off, p->addralign()); |
| section_size_type section_end_offset = |
| section_begin_offset + p->data_size(); |
| |
| // Check to see if we should group the previously seen sections. |
| switch (state) |
| { |
| case NO_GROUP: |
| break; |
| |
| case FINDING_STUB_SECTION: |
| // Adding this section makes the group larger than GROUP_SIZE. |
| if (section_end_offset - group_begin_offset >= group_size) |
| { |
| if (stubs_always_after_branch) |
| { |
| gold_assert(group_end != this->input_sections().end()); |
| this->create_stub_group(group_begin, group_end, group_end, |
| target, new_relaxed_sections, |
| task); |
| state = NO_GROUP; |
| } |
| else |
| { |
| // Input sections up to stub_group_size bytes after the stub |
| // table can be handled by it too. |
| state = HAS_STUB_SECTION; |
| stub_table = group_end; |
| stub_table_end_offset = group_end_offset; |
| } |
| } |
| break; |
| |
| case HAS_STUB_SECTION: |
| // Adding this section makes the post stub-section group larger |
| // than GROUP_SIZE. |
| gold_unreachable(); |
| // NOT SUPPORTED YET. For completeness only. |
| if (section_end_offset - stub_table_end_offset >= group_size) |
| { |
| gold_assert(group_end != this->input_sections().end()); |
| this->create_stub_group(group_begin, group_end, stub_table, |
| target, new_relaxed_sections, task); |
| state = NO_GROUP; |
| } |
| break; |
| |
| default: |
| gold_unreachable(); |
| } |
| |
| // If we see an input section and currently there is no group, start |
| // a new one. Skip any empty sections. We look at the data size |
| // instead of calling p->relobj()->section_size() to avoid locking. |
| if ((p->is_input_section() || p->is_relaxed_input_section()) |
| && (p->data_size() != 0)) |
| { |
| if (state == NO_GROUP) |
| { |
| state = FINDING_STUB_SECTION; |
| group_begin = p; |
| group_begin_offset = section_begin_offset; |
| } |
| |
| // Keep track of the last input section seen. |
| group_end = p; |
| group_end_offset = section_end_offset; |
| } |
| |
| off = section_end_offset; |
| } |
| |
| // Create a stub group for any ungrouped sections. |
| if (state == FINDING_STUB_SECTION || state == HAS_STUB_SECTION) |
| { |
| gold_assert(group_end != this->input_sections().end()); |
| this->create_stub_group(group_begin, group_end, |
| (state == FINDING_STUB_SECTION |
| ? group_end |
| : stub_table), |
| target, new_relaxed_sections, task); |
| } |
| |
| if (!new_relaxed_sections.empty()) |
| this->convert_input_sections_to_relaxed_sections(new_relaxed_sections); |
| |
| // Update the section offsets |
| for (size_t i = 0; i < new_relaxed_sections.size(); ++i) |
| { |
| The_aarch64_relobj* relobj = static_cast<The_aarch64_relobj*>( |
| new_relaxed_sections[i]->relobj()); |
| unsigned int shndx = new_relaxed_sections[i]->shndx(); |
| // Tell AArch64_relobj that this input section is converted. |
| relobj->convert_input_section_to_relaxed_section(shndx); |
| } |
| } // End of AArch64_output_section::group_sections |
| |
| |
| AArch64_reloc_property_table* aarch64_reloc_property_table = NULL; |
| |
| |
| // The aarch64 target class. |
| // See the ABI at |
| // http://infocenter.arm.com/help/topic/com.arm.doc.ihi0056b/IHI0056B_aaelf64.pdf |
| template<int size, bool big_endian> |
| class Target_aarch64 : public Sized_target<size, big_endian> |
| { |
| public: |
| typedef Target_aarch64<size, big_endian> This; |
| typedef Output_data_reloc<elfcpp::SHT_RELA, true, size, big_endian> |
| Reloc_section; |
| typedef Relocate_info<size, big_endian> The_relocate_info; |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr Address; |
| typedef AArch64_relobj<size, big_endian> The_aarch64_relobj; |
| typedef Reloc_stub<size, big_endian> The_reloc_stub; |
| typedef Erratum_stub<size, big_endian> The_erratum_stub; |
| typedef typename Reloc_stub<size, big_endian>::Key The_reloc_stub_key; |
| typedef Stub_table<size, big_endian> The_stub_table; |
| typedef std::vector<The_stub_table*> Stub_table_list; |
| typedef typename Stub_table_list::iterator Stub_table_iterator; |
| typedef AArch64_input_section<size, big_endian> The_aarch64_input_section; |
| typedef AArch64_output_section<size, big_endian> The_aarch64_output_section; |
| typedef Unordered_map<Section_id, |
| AArch64_input_section<size, big_endian>*, |
| Section_id_hash> AArch64_input_section_map; |
| typedef AArch64_insn_utilities<big_endian> Insn_utilities; |
| const static int TCB_SIZE = size / 8 * 2; |
| |
| Target_aarch64(const Target::Target_info* info = &aarch64_info) |
| : Sized_target<size, big_endian>(info), |
| got_(NULL), plt_(NULL), got_plt_(NULL), got_irelative_(NULL), |
| got_tlsdesc_(NULL), global_offset_table_(NULL), rela_dyn_(NULL), |
| rela_irelative_(NULL), copy_relocs_(elfcpp::R_AARCH64_COPY), |
| got_mod_index_offset_(-1U), |
| tlsdesc_reloc_info_(), tls_base_symbol_defined_(false), |
| stub_tables_(), stub_group_size_(0), aarch64_input_section_map_() |
| { } |
| |
| // Scan the relocations to determine unreferenced sections for |
| // garbage collection. |
| void |
| gc_process_relocs(Symbol_table* symtab, |
| Layout* layout, |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int data_shndx, |
| unsigned int sh_type, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| bool needs_special_offset_handling, |
| size_t local_symbol_count, |
| const unsigned char* plocal_symbols); |
| |
| // Scan the relocations to look for symbol adjustments. |
| void |
| scan_relocs(Symbol_table* symtab, |
| Layout* layout, |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int data_shndx, |
| unsigned int sh_type, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| bool needs_special_offset_handling, |
| size_t local_symbol_count, |
| const unsigned char* plocal_symbols); |
| |
| // Finalize the sections. |
| void |
| do_finalize_sections(Layout*, const Input_objects*, Symbol_table*); |
| |
| // Return the value to use for a dynamic which requires special |
| // treatment. |
| uint64_t |
| do_dynsym_value(const Symbol*) const; |
| |
| // Relocate a section. |
| void |
| relocate_section(const Relocate_info<size, big_endian>*, |
| unsigned int sh_type, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| bool needs_special_offset_handling, |
| unsigned char* view, |
| typename elfcpp::Elf_types<size>::Elf_Addr view_address, |
| section_size_type view_size, |
| const Reloc_symbol_changes*); |
| |
| // Scan the relocs during a relocatable link. |
| void |
| scan_relocatable_relocs(Symbol_table* symtab, |
| Layout* layout, |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int data_shndx, |
| unsigned int sh_type, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| bool needs_special_offset_handling, |
| size_t local_symbol_count, |
| const unsigned char* plocal_symbols, |
| Relocatable_relocs*); |
| |
| // Scan the relocs for --emit-relocs. |
| void |
| emit_relocs_scan(Symbol_table* symtab, |
| Layout* layout, |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int data_shndx, |
| unsigned int sh_type, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| bool needs_special_offset_handling, |
| size_t local_symbol_count, |
| const unsigned char* plocal_syms, |
| Relocatable_relocs* rr); |
| |
| // Relocate a section during a relocatable link. |
| void |
| relocate_relocs( |
| const Relocate_info<size, big_endian>*, |
| unsigned int sh_type, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| typename elfcpp::Elf_types<size>::Elf_Off offset_in_output_section, |
| unsigned char* view, |
| typename elfcpp::Elf_types<size>::Elf_Addr view_address, |
| section_size_type view_size, |
| unsigned char* reloc_view, |
| section_size_type reloc_view_size); |
| |
| // Return the symbol index to use for a target specific relocation. |
| // The only target specific relocation is R_AARCH64_TLSDESC for a |
| // local symbol, which is an absolute reloc. |
| unsigned int |
| do_reloc_symbol_index(void*, unsigned int r_type) const |
| { |
| gold_assert(r_type == elfcpp::R_AARCH64_TLSDESC); |
| return 0; |
| } |
| |
| // Return the addend to use for a target specific relocation. |
| uint64_t |
| do_reloc_addend(void* arg, unsigned int r_type, uint64_t addend) const; |
| |
| // Return the PLT section. |
| uint64_t |
| do_plt_address_for_global(const Symbol* gsym) const |
| { return this->plt_section()->address_for_global(gsym); } |
| |
| uint64_t |
| do_plt_address_for_local(const Relobj* relobj, unsigned int symndx) const |
| { return this->plt_section()->address_for_local(relobj, symndx); } |
| |
| // This function should be defined in targets that can use relocation |
| // types to determine (implemented in local_reloc_may_be_function_pointer |
| // and global_reloc_may_be_function_pointer) |
| // if a function's pointer is taken. ICF uses this in safe mode to only |
| // fold those functions whose pointer is defintely not taken. |
| bool |
| do_can_check_for_function_pointers() const |
| { return true; } |
| |
| // Return the number of entries in the PLT. |
| unsigned int |
| plt_entry_count() const; |
| |
| //Return the offset of the first non-reserved PLT entry. |
| unsigned int |
| first_plt_entry_offset() const; |
| |
| // Return the size of each PLT entry. |
| unsigned int |
| plt_entry_size() const; |
| |
| // Create a stub table. |
| The_stub_table* |
| new_stub_table(The_aarch64_input_section*); |
| |
| // Create an aarch64 input section. |
| The_aarch64_input_section* |
| new_aarch64_input_section(Relobj*, unsigned int); |
| |
| // Find an aarch64 input section instance for a given OBJ and SHNDX. |
| The_aarch64_input_section* |
| find_aarch64_input_section(Relobj*, unsigned int) const; |
| |
| // Return the thread control block size. |
| unsigned int |
| tcb_size() const { return This::TCB_SIZE; } |
| |
| // Scan a section for stub generation. |
| void |
| scan_section_for_stubs(const Relocate_info<size, big_endian>*, unsigned int, |
| const unsigned char*, size_t, Output_section*, |
| bool, const unsigned char*, |
| Address, |
| section_size_type); |
| |
| // Scan a relocation section for stub. |
| template<int sh_type> |
| void |
| scan_reloc_section_for_stubs( |
| const The_relocate_info* relinfo, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| bool needs_special_offset_handling, |
| const unsigned char* view, |
| Address view_address, |
| section_size_type); |
| |
| // Relocate a single reloc stub. |
| void |
| relocate_reloc_stub(The_reloc_stub*, const Relocate_info<size, big_endian>*, |
| Output_section*, unsigned char*, Address, |
| section_size_type); |
| |
| // Get the default AArch64 target. |
| static This* |
| current_target() |
| { |
| gold_assert(parameters->target().machine_code() == elfcpp::EM_AARCH64 |
| && parameters->target().get_size() == size |
| && parameters->target().is_big_endian() == big_endian); |
| return static_cast<This*>(parameters->sized_target<size, big_endian>()); |
| } |
| |
| |
| // Scan erratum 843419 for a part of a section. |
| void |
| scan_erratum_843419_span( |
| AArch64_relobj<size, big_endian>*, |
| unsigned int, |
| const section_size_type, |
| const section_size_type, |
| unsigned char*, |
| Address); |
| |
| // Scan erratum 835769 for a part of a section. |
| void |
| scan_erratum_835769_span( |
| AArch64_relobj<size, big_endian>*, |
| unsigned int, |
| const section_size_type, |
| const section_size_type, |
| unsigned char*, |
| Address); |
| |
| protected: |
| void |
| do_select_as_default_target() |
| { |
| gold_assert(aarch64_reloc_property_table == NULL); |
| aarch64_reloc_property_table = new AArch64_reloc_property_table(); |
| } |
| |
| // Add a new reloc argument, returning the index in the vector. |
| size_t |
| add_tlsdesc_info(Sized_relobj_file<size, big_endian>* object, |
| unsigned int r_sym) |
| { |
| this->tlsdesc_reloc_info_.push_back(Tlsdesc_info(object, r_sym)); |
| return this->tlsdesc_reloc_info_.size() - 1; |
| } |
| |
| virtual Output_data_plt_aarch64<size, big_endian>* |
| do_make_data_plt(Layout* layout, |
| Output_data_got_aarch64<size, big_endian>* got, |
| Output_data_space* got_plt, |
| Output_data_space* got_irelative) |
| { |
| return new Output_data_plt_aarch64_standard<size, big_endian>( |
| layout, got, got_plt, got_irelative); |
| } |
| |
| |
| // do_make_elf_object to override the same function in the base class. |
| Object* |
| do_make_elf_object(const std::string&, Input_file*, off_t, |
| const elfcpp::Ehdr<size, big_endian>&); |
| |
| Output_data_plt_aarch64<size, big_endian>* |
| make_data_plt(Layout* layout, |
| Output_data_got_aarch64<size, big_endian>* got, |
| Output_data_space* got_plt, |
| Output_data_space* got_irelative) |
| { |
| return this->do_make_data_plt(layout, got, got_plt, got_irelative); |
| } |
| |
| // We only need to generate stubs, and hence perform relaxation if we are |
| // not doing relocatable linking. |
| virtual bool |
| do_may_relax() const |
| { return !parameters->options().relocatable(); } |
| |
| // Relaxation hook. This is where we do stub generation. |
| virtual bool |
| do_relax(int, const Input_objects*, Symbol_table*, Layout*, const Task*); |
| |
| void |
| group_sections(Layout* layout, |
| section_size_type group_size, |
| bool stubs_always_after_branch, |
| const Task* task); |
| |
| void |
| scan_reloc_for_stub(const The_relocate_info*, unsigned int, |
| const Sized_symbol<size>*, unsigned int, |
| const Symbol_value<size>*, |
| typename elfcpp::Elf_types<size>::Elf_Swxword, |
| Address Elf_Addr); |
| |
| // Make an output section. |
| Output_section* |
| do_make_output_section(const char* name, elfcpp::Elf_Word type, |
| elfcpp::Elf_Xword flags) |
| { return new The_aarch64_output_section(name, type, flags); } |
| |
| private: |
| // The class which scans relocations. |
| class Scan |
| { |
| public: |
| Scan() |
| : issued_non_pic_error_(false) |
| { } |
| |
| inline void |
| local(Symbol_table* symtab, Layout* layout, Target_aarch64* target, |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int data_shndx, |
| Output_section* output_section, |
| const elfcpp::Rela<size, big_endian>& reloc, unsigned int r_type, |
| const elfcpp::Sym<size, big_endian>& lsym, |
| bool is_discarded); |
| |
| inline void |
| global(Symbol_table* symtab, Layout* layout, Target_aarch64* target, |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int data_shndx, |
| Output_section* output_section, |
| const elfcpp::Rela<size, big_endian>& reloc, unsigned int r_type, |
| Symbol* gsym); |
| |
| inline bool |
| local_reloc_may_be_function_pointer(Symbol_table* , Layout* , |
| Target_aarch64<size, big_endian>* , |
| Sized_relobj_file<size, big_endian>* , |
| unsigned int , |
| Output_section* , |
| const elfcpp::Rela<size, big_endian>& , |
| unsigned int r_type, |
| const elfcpp::Sym<size, big_endian>&); |
| |
| inline bool |
| global_reloc_may_be_function_pointer(Symbol_table* , Layout* , |
| Target_aarch64<size, big_endian>* , |
| Sized_relobj_file<size, big_endian>* , |
| unsigned int , |
| Output_section* , |
| const elfcpp::Rela<size, big_endian>& , |
| unsigned int r_type, |
| Symbol* gsym); |
| |
| private: |
| static void |
| unsupported_reloc_local(Sized_relobj_file<size, big_endian>*, |
| unsigned int r_type); |
| |
| static void |
| unsupported_reloc_global(Sized_relobj_file<size, big_endian>*, |
| unsigned int r_type, Symbol*); |
| |
| inline bool |
| possible_function_pointer_reloc(unsigned int r_type); |
| |
| void |
| check_non_pic(Relobj*, unsigned int r_type); |
| |
| bool |
| reloc_needs_plt_for_ifunc(Sized_relobj_file<size, big_endian>*, |
| unsigned int r_type); |
| |
| // Whether we have issued an error about a non-PIC compilation. |
| bool issued_non_pic_error_; |
| }; |
| |
| // The class which implements relocation. |
| class Relocate |
| { |
| public: |
| Relocate() |
| : skip_call_tls_get_addr_(false) |
| { } |
| |
| ~Relocate() |
| { } |
| |
| // Do a relocation. Return false if the caller should not issue |
| // any warnings about this relocation. |
| inline bool |
| relocate(const Relocate_info<size, big_endian>*, unsigned int, |
| Target_aarch64*, Output_section*, size_t, const unsigned char*, |
| const Sized_symbol<size>*, const Symbol_value<size>*, |
| unsigned char*, typename elfcpp::Elf_types<size>::Elf_Addr, |
| section_size_type); |
| |
| private: |
| inline typename AArch64_relocate_functions<size, big_endian>::Status |
| relocate_tls(const Relocate_info<size, big_endian>*, |
| Target_aarch64<size, big_endian>*, |
| size_t, |
| const elfcpp::Rela<size, big_endian>&, |
| unsigned int r_type, const Sized_symbol<size>*, |
| const Symbol_value<size>*, |
| unsigned char*, |
| typename elfcpp::Elf_types<size>::Elf_Addr); |
| |
| inline typename AArch64_relocate_functions<size, big_endian>::Status |
| tls_gd_to_le( |
| const Relocate_info<size, big_endian>*, |
| Target_aarch64<size, big_endian>*, |
| const elfcpp::Rela<size, big_endian>&, |
| unsigned int, |
| unsigned char*, |
| const Symbol_value<size>*); |
| |
| inline typename AArch64_relocate_functions<size, big_endian>::Status |
| tls_ld_to_le( |
| const Relocate_info<size, big_endian>*, |
| Target_aarch64<size, big_endian>*, |
| const elfcpp::Rela<size, big_endian>&, |
| unsigned int, |
| unsigned char*, |
| const Symbol_value<size>*); |
| |
| inline typename AArch64_relocate_functions<size, big_endian>::Status |
| tls_ie_to_le( |
| const Relocate_info<size, big_endian>*, |
| Target_aarch64<size, big_endian>*, |
| const elfcpp::Rela<size, big_endian>&, |
| unsigned int, |
| unsigned char*, |
| const Symbol_value<size>*); |
| |
| inline typename AArch64_relocate_functions<size, big_endian>::Status |
| tls_desc_gd_to_le( |
| const Relocate_info<size, big_endian>*, |
| Target_aarch64<size, big_endian>*, |
| const elfcpp::Rela<size, big_endian>&, |
| unsigned int, |
| unsigned char*, |
| const Symbol_value<size>*); |
| |
| inline typename AArch64_relocate_functions<size, big_endian>::Status |
| tls_desc_gd_to_ie( |
| const Relocate_info<size, big_endian>*, |
| Target_aarch64<size, big_endian>*, |
| const elfcpp::Rela<size, big_endian>&, |
| unsigned int, |
| unsigned char*, |
| const Symbol_value<size>*, |
| typename elfcpp::Elf_types<size>::Elf_Addr, |
| typename elfcpp::Elf_types<size>::Elf_Addr); |
| |
| bool skip_call_tls_get_addr_; |
| |
| }; // End of class Relocate |
| |
| // Adjust TLS relocation type based on the options and whether this |
| // is a local symbol. |
| static tls::Tls_optimization |
| optimize_tls_reloc(bool is_final, int r_type); |
| |
| // Get the GOT section, creating it if necessary. |
| Output_data_got_aarch64<size, big_endian>* |
| got_section(Symbol_table*, Layout*); |
| |
| // Get the GOT PLT section. |
| Output_data_space* |
| got_plt_section() const |
| { |
| gold_assert(this->got_plt_ != NULL); |
| return this->got_plt_; |
| } |
| |
| // Get the GOT section for TLSDESC entries. |
| Output_data_got<size, big_endian>* |
| got_tlsdesc_section() const |
| { |
| gold_assert(this->got_tlsdesc_ != NULL); |
| return this->got_tlsdesc_; |
| } |
| |
| // Create the PLT section. |
| void |
| make_plt_section(Symbol_table* symtab, Layout* layout); |
| |
| // Create a PLT entry for a global symbol. |
| void |
| make_plt_entry(Symbol_table*, Layout*, Symbol*); |
| |
| // Create a PLT entry for a local STT_GNU_IFUNC symbol. |
| void |
| make_local_ifunc_plt_entry(Symbol_table*, Layout*, |
| Sized_relobj_file<size, big_endian>* relobj, |
| unsigned int local_sym_index); |
| |
| // Define the _TLS_MODULE_BASE_ symbol in the TLS segment. |
| void |
| define_tls_base_symbol(Symbol_table*, Layout*); |
| |
| // Create the reserved PLT and GOT entries for the TLS descriptor resolver. |
| void |
| reserve_tlsdesc_entries(Symbol_table* symtab, Layout* layout); |
| |
| // Create a GOT entry for the TLS module index. |
| unsigned int |
| got_mod_index_entry(Symbol_table* symtab, Layout* layout, |
| Sized_relobj_file<size, big_endian>* object); |
| |
| // Get the PLT section. |
| Output_data_plt_aarch64<size, big_endian>* |
| plt_section() const |
| { |
| gold_assert(this->plt_ != NULL); |
| return this->plt_; |
| } |
| |
| // Helper method to create erratum stubs for ST_E_843419 and ST_E_835769. For |
| // ST_E_843419, we need an additional field for adrp offset. |
| void create_erratum_stub( |
| AArch64_relobj<size, big_endian>* relobj, |
| unsigned int shndx, |
| section_size_type erratum_insn_offset, |
| Address erratum_address, |
| typename Insn_utilities::Insntype erratum_insn, |
| int erratum_type, |
| unsigned int e843419_adrp_offset=0); |
| |
| // Return whether this is a 3-insn erratum sequence. |
| bool is_erratum_843419_sequence( |
| typename elfcpp::Swap<32,big_endian>::Valtype insn1, |
| typename elfcpp::Swap<32,big_endian>::Valtype insn2, |
| typename elfcpp::Swap<32,big_endian>::Valtype insn3); |
| |
| // Return whether this is a 835769 sequence. |
| // (Similarly implemented as in elfnn-aarch64.c.) |
| bool is_erratum_835769_sequence( |
| typename elfcpp::Swap<32,big_endian>::Valtype, |
| typename elfcpp::Swap<32,big_endian>::Valtype); |
| |
| // Get the dynamic reloc section, creating it if necessary. |
| Reloc_section* |
| rela_dyn_section(Layout*); |
| |
| // Get the section to use for TLSDESC relocations. |
| Reloc_section* |
| rela_tlsdesc_section(Layout*) const; |
| |
| // Get the section to use for IRELATIVE relocations. |
| Reloc_section* |
| rela_irelative_section(Layout*); |
| |
| // Add a potential copy relocation. |
| void |
| copy_reloc(Symbol_table* symtab, Layout* layout, |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int shndx, Output_section* output_section, |
| Symbol* sym, const elfcpp::Rela<size, big_endian>& reloc) |
| { |
| unsigned int r_type = elfcpp::elf_r_type<size>(reloc.get_r_info()); |
| this->copy_relocs_.copy_reloc(symtab, layout, |
| symtab->get_sized_symbol<size>(sym), |
| object, shndx, output_section, |
| r_type, reloc.get_r_offset(), |
| reloc.get_r_addend(), |
| this->rela_dyn_section(layout)); |
| } |
| |
| // Information about this specific target which we pass to the |
| // general Target structure. |
| static const Target::Target_info aarch64_info; |
| |
| // The types of GOT entries needed for this platform. |
| // These values are exposed to the ABI in an incremental link. |
| // Do not renumber existing values without changing the version |
| // number of the .gnu_incremental_inputs section. |
| enum Got_type |
| { |
| GOT_TYPE_STANDARD = 0, // GOT entry for a regular symbol |
| GOT_TYPE_TLS_OFFSET = 1, // GOT entry for TLS offset |
| GOT_TYPE_TLS_PAIR = 2, // GOT entry for TLS module/offset pair |
| GOT_TYPE_TLS_DESC = 3 // GOT entry for TLS_DESC pair |
| }; |
| |
| // This type is used as the argument to the target specific |
| // relocation routines. The only target specific reloc is |
| // R_AARCh64_TLSDESC against a local symbol. |
| struct Tlsdesc_info |
| { |
| Tlsdesc_info(Sized_relobj_file<size, big_endian>* a_object, |
| unsigned int a_r_sym) |
| : object(a_object), r_sym(a_r_sym) |
| { } |
| |
| // The object in which the local symbol is defined. |
| Sized_relobj_file<size, big_endian>* object; |
| // The local symbol index in the object. |
| unsigned int r_sym; |
| }; |
| |
| // The GOT section. |
| Output_data_got_aarch64<size, big_endian>* got_; |
| // The PLT section. |
| Output_data_plt_aarch64<size, big_endian>* plt_; |
| // The GOT PLT section. |
| Output_data_space* got_plt_; |
| // The GOT section for IRELATIVE relocations. |
| Output_data_space* got_irelative_; |
| // The GOT section for TLSDESC relocations. |
| Output_data_got<size, big_endian>* got_tlsdesc_; |
| // The _GLOBAL_OFFSET_TABLE_ symbol. |
| Symbol* global_offset_table_; |
| // The dynamic reloc section. |
| Reloc_section* rela_dyn_; |
| // The section to use for IRELATIVE relocs. |
| Reloc_section* rela_irelative_; |
| // Relocs saved to avoid a COPY reloc. |
| Copy_relocs<elfcpp::SHT_RELA, size, big_endian> copy_relocs_; |
| // Offset of the GOT entry for the TLS module index. |
| unsigned int got_mod_index_offset_; |
| // We handle R_AARCH64_TLSDESC against a local symbol as a target |
| // specific relocation. Here we store the object and local symbol |
| // index for the relocation. |
| std::vector<Tlsdesc_info> tlsdesc_reloc_info_; |
| // True if the _TLS_MODULE_BASE_ symbol has been defined. |
| bool tls_base_symbol_defined_; |
| // List of stub_tables |
| Stub_table_list stub_tables_; |
| // Actual stub group size |
| section_size_type stub_group_size_; |
| AArch64_input_section_map aarch64_input_section_map_; |
| }; // End of Target_aarch64 |
| |
| |
| template<> |
| const Target::Target_info Target_aarch64<64, false>::aarch64_info = |
| { |
| 64, // size |
| false, // is_big_endian |
| elfcpp::EM_AARCH64, // machine_code |
| false, // has_make_symbol |
| false, // has_resolve |
| false, // has_code_fill |
| false, // is_default_stack_executable |
| true, // can_icf_inline_merge_sections |
| '\0', // wrap_char |
| "/lib/ld.so.1", // program interpreter |
| 0x400000, // default_text_segment_address |
| 0x10000, // abi_pagesize (overridable by -z max-page-size) |
| 0x1000, // common_pagesize (overridable by -z common-page-size) |
| false, // isolate_execinstr |
| 0, // rosegment_gap |
| elfcpp::SHN_UNDEF, // small_common_shndx |
| elfcpp::SHN_UNDEF, // large_common_shndx |
| 0, // small_common_section_flags |
| 0, // large_common_section_flags |
| NULL, // attributes_section |
| NULL, // attributes_vendor |
| "_start", // entry_symbol_name |
| 32, // hash_entry_size |
| elfcpp::SHT_PROGBITS, // unwind_section_type |
| }; |
| |
| template<> |
| const Target::Target_info Target_aarch64<32, false>::aarch64_info = |
| { |
| 32, // size |
| false, // is_big_endian |
| elfcpp::EM_AARCH64, // machine_code |
| false, // has_make_symbol |
| false, // has_resolve |
| false, // has_code_fill |
| false, // is_default_stack_executable |
| false, // can_icf_inline_merge_sections |
| '\0', // wrap_char |
| "/lib/ld.so.1", // program interpreter |
| 0x400000, // default_text_segment_address |
| 0x10000, // abi_pagesize (overridable by -z max-page-size) |
| 0x1000, // common_pagesize (overridable by -z common-page-size) |
| false, // isolate_execinstr |
| 0, // rosegment_gap |
| elfcpp::SHN_UNDEF, // small_common_shndx |
| elfcpp::SHN_UNDEF, // large_common_shndx |
| 0, // small_common_section_flags |
| 0, // large_common_section_flags |
| NULL, // attributes_section |
| NULL, // attributes_vendor |
| "_start", // entry_symbol_name |
| 32, // hash_entry_size |
| elfcpp::SHT_PROGBITS, // unwind_section_type |
| }; |
| |
| template<> |
| const Target::Target_info Target_aarch64<64, true>::aarch64_info = |
| { |
| 64, // size |
| true, // is_big_endian |
| elfcpp::EM_AARCH64, // machine_code |
| false, // has_make_symbol |
| false, // has_resolve |
| false, // has_code_fill |
| false, // is_default_stack_executable |
| true, // can_icf_inline_merge_sections |
| '\0', // wrap_char |
| "/lib/ld.so.1", // program interpreter |
| 0x400000, // default_text_segment_address |
| 0x10000, // abi_pagesize (overridable by -z max-page-size) |
| 0x1000, // common_pagesize (overridable by -z common-page-size) |
| false, // isolate_execinstr |
| 0, // rosegment_gap |
| elfcpp::SHN_UNDEF, // small_common_shndx |
| elfcpp::SHN_UNDEF, // large_common_shndx |
| 0, // small_common_section_flags |
| 0, // large_common_section_flags |
| NULL, // attributes_section |
| NULL, // attributes_vendor |
| "_start", // entry_symbol_name |
| 32, // hash_entry_size |
| elfcpp::SHT_PROGBITS, // unwind_section_type |
| }; |
| |
| template<> |
| const Target::Target_info Target_aarch64<32, true>::aarch64_info = |
| { |
| 32, // size |
| true, // is_big_endian |
| elfcpp::EM_AARCH64, // machine_code |
| false, // has_make_symbol |
| false, // has_resolve |
| false, // has_code_fill |
| false, // is_default_stack_executable |
| false, // can_icf_inline_merge_sections |
| '\0', // wrap_char |
| "/lib/ld.so.1", // program interpreter |
| 0x400000, // default_text_segment_address |
| 0x10000, // abi_pagesize (overridable by -z max-page-size) |
| 0x1000, // common_pagesize (overridable by -z common-page-size) |
| false, // isolate_execinstr |
| 0, // rosegment_gap |
| elfcpp::SHN_UNDEF, // small_common_shndx |
| elfcpp::SHN_UNDEF, // large_common_shndx |
| 0, // small_common_section_flags |
| 0, // large_common_section_flags |
| NULL, // attributes_section |
| NULL, // attributes_vendor |
| "_start", // entry_symbol_name |
| 32, // hash_entry_size |
| elfcpp::SHT_PROGBITS, // unwind_section_type |
| }; |
| |
| // Get the GOT section, creating it if necessary. |
| |
| template<int size, bool big_endian> |
| Output_data_got_aarch64<size, big_endian>* |
| Target_aarch64<size, big_endian>::got_section(Symbol_table* symtab, |
| Layout* layout) |
| { |
| if (this->got_ == NULL) |
| { |
| gold_assert(symtab != NULL && layout != NULL); |
| |
| // When using -z now, we can treat .got.plt as a relro section. |
| // Without -z now, it is modified after program startup by lazy |
| // PLT relocations. |
| bool is_got_plt_relro = parameters->options().now(); |
| Output_section_order got_order = (is_got_plt_relro |
| ? ORDER_RELRO |
| : ORDER_RELRO_LAST); |
| Output_section_order got_plt_order = (is_got_plt_relro |
| ? ORDER_RELRO |
| : ORDER_NON_RELRO_FIRST); |
| |
| // Layout of .got and .got.plt sections. |
| // .got[0] &_DYNAMIC <-_GLOBAL_OFFSET_TABLE_ |
| // ... |
| // .gotplt[0] reserved for ld.so (&linkmap) <--DT_PLTGOT |
| // .gotplt[1] reserved for ld.so (resolver) |
| // .gotplt[2] reserved |
| |
| // Generate .got section. |
| this->got_ = new Output_data_got_aarch64<size, big_endian>(symtab, |
| layout); |
| layout->add_output_section_data(".got", elfcpp::SHT_PROGBITS, |
| (elfcpp::SHF_ALLOC | elfcpp::SHF_WRITE), |
| this->got_, got_order, true); |
| // The first word of GOT is reserved for the address of .dynamic. |
| // We put 0 here now. The value will be replaced later in |
| // Output_data_got_aarch64::do_write. |
| this->got_->add_constant(0); |
| |
| // Define _GLOBAL_OFFSET_TABLE_ at the start of the PLT. |
| // _GLOBAL_OFFSET_TABLE_ value points to the start of the .got section, |
| // even if there is a .got.plt section. |
| this->global_offset_table_ = |
| symtab->define_in_output_data("_GLOBAL_OFFSET_TABLE_", NULL, |
| Symbol_table::PREDEFINED, |
| this->got_, |
| 0, 0, elfcpp::STT_OBJECT, |
| elfcpp::STB_LOCAL, |
| elfcpp::STV_HIDDEN, 0, |
| false, false); |
| |
| // Generate .got.plt section. |
| this->got_plt_ = new Output_data_space(size / 8, "** GOT PLT"); |
| layout->add_output_section_data(".got.plt", elfcpp::SHT_PROGBITS, |
| (elfcpp::SHF_ALLOC |
| | elfcpp::SHF_WRITE), |
| this->got_plt_, got_plt_order, |
| is_got_plt_relro); |
| |
| // The first three entries are reserved. |
| this->got_plt_->set_current_data_size( |
| AARCH64_GOTPLT_RESERVE_COUNT * (size / 8)); |
| |
| // If there are any IRELATIVE relocations, they get GOT entries |
| // in .got.plt after the jump slot entries. |
| this->got_irelative_ = new Output_data_space(size / 8, |
| "** GOT IRELATIVE PLT"); |
| layout->add_output_section_data(".got.plt", elfcpp::SHT_PROGBITS, |
| (elfcpp::SHF_ALLOC |
| | elfcpp::SHF_WRITE), |
| this->got_irelative_, |
| got_plt_order, |
| is_got_plt_relro); |
| |
| // If there are any TLSDESC relocations, they get GOT entries in |
| // .got.plt after the jump slot and IRELATIVE entries. |
| this->got_tlsdesc_ = new Output_data_got<size, big_endian>(); |
| layout->add_output_section_data(".got.plt", elfcpp::SHT_PROGBITS, |
| (elfcpp::SHF_ALLOC |
| | elfcpp::SHF_WRITE), |
| this->got_tlsdesc_, |
| got_plt_order, |
| is_got_plt_relro); |
| |
| if (!is_got_plt_relro) |
| { |
| // Those bytes can go into the relro segment. |
| layout->increase_relro( |
| AARCH64_GOTPLT_RESERVE_COUNT * (size / 8)); |
| } |
| |
| } |
| return this->got_; |
| } |
| |
| // Get the dynamic reloc section, creating it if necessary. |
| |
| template<int size, bool big_endian> |
| typename Target_aarch64<size, big_endian>::Reloc_section* |
| Target_aarch64<size, big_endian>::rela_dyn_section(Layout* layout) |
| { |
| if (this->rela_dyn_ == NULL) |
| { |
| gold_assert(layout != NULL); |
| this->rela_dyn_ = new Reloc_section(parameters->options().combreloc()); |
| layout->add_output_section_data(".rela.dyn", elfcpp::SHT_RELA, |
| elfcpp::SHF_ALLOC, this->rela_dyn_, |
| ORDER_DYNAMIC_RELOCS, false); |
| } |
| return this->rela_dyn_; |
| } |
| |
| // Get the section to use for IRELATIVE relocs, creating it if |
| // necessary. These go in .rela.dyn, but only after all other dynamic |
| // relocations. They need to follow the other dynamic relocations so |
| // that they can refer to global variables initialized by those |
| // relocs. |
| |
| template<int size, bool big_endian> |
| typename Target_aarch64<size, big_endian>::Reloc_section* |
| Target_aarch64<size, big_endian>::rela_irelative_section(Layout* layout) |
| { |
| if (this->rela_irelative_ == NULL) |
| { |
| // Make sure we have already created the dynamic reloc section. |
| this->rela_dyn_section(layout); |
| this->rela_irelative_ = new Reloc_section(false); |
| layout->add_output_section_data(".rela.dyn", elfcpp::SHT_RELA, |
| elfcpp::SHF_ALLOC, this->rela_irelative_, |
| ORDER_DYNAMIC_RELOCS, false); |
| gold_assert(this->rela_dyn_->output_section() |
| == this->rela_irelative_->output_section()); |
| } |
| return this->rela_irelative_; |
| } |
| |
| |
| // do_make_elf_object to override the same function in the base class. We need |
| // to use a target-specific sub-class of Sized_relobj_file<size, big_endian> to |
| // store backend specific information. Hence we need to have our own ELF object |
| // creation. |
| |
| template<int size, bool big_endian> |
| Object* |
| Target_aarch64<size, big_endian>::do_make_elf_object( |
| const std::string& name, |
| Input_file* input_file, |
| off_t offset, const elfcpp::Ehdr<size, big_endian>& ehdr) |
| { |
| int et = ehdr.get_e_type(); |
| // ET_EXEC files are valid input for --just-symbols/-R, |
| // and we treat them as relocatable objects. |
| if (et == elfcpp::ET_EXEC && input_file->just_symbols()) |
| return Sized_target<size, big_endian>::do_make_elf_object( |
| name, input_file, offset, ehdr); |
| else if (et == elfcpp::ET_REL) |
| { |
| AArch64_relobj<size, big_endian>* obj = |
| new AArch64_relobj<size, big_endian>(name, input_file, offset, ehdr); |
| obj->setup(); |
| return obj; |
| } |
| else if (et == elfcpp::ET_DYN) |
| { |
| // Keep base implementation. |
| Sized_dynobj<size, big_endian>* obj = |
| new Sized_dynobj<size, big_endian>(name, input_file, offset, ehdr); |
| obj->setup(); |
| return obj; |
| } |
| else |
| { |
| gold_error(_("%s: unsupported ELF file type %d"), |
| name.c_str(), et); |
| return NULL; |
| } |
| } |
| |
| |
| // Scan a relocation for stub generation. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::scan_reloc_for_stub( |
| const Relocate_info<size, big_endian>* relinfo, |
| unsigned int r_type, |
| const Sized_symbol<size>* gsym, |
| unsigned int r_sym, |
| const Symbol_value<size>* psymval, |
| typename elfcpp::Elf_types<size>::Elf_Swxword addend, |
| Address address) |
| { |
| const AArch64_relobj<size, big_endian>* aarch64_relobj = |
| static_cast<AArch64_relobj<size, big_endian>*>(relinfo->object); |
| |
| Symbol_value<size> symval; |
| if (gsym != NULL) |
| { |
| const AArch64_reloc_property* arp = aarch64_reloc_property_table-> |
| get_reloc_property(r_type); |
| if (gsym->use_plt_offset(arp->reference_flags())) |
| { |
| // This uses a PLT, change the symbol value. |
| symval.set_output_value(this->plt_address_for_global(gsym)); |
| psymval = &symval; |
| } |
| else if (gsym->is_undefined()) |
| { |
| // There is no need to generate a stub symbol if the original symbol |
| // is undefined. |
| gold_debug(DEBUG_TARGET, |
| "stub: not creating a stub for undefined symbol %s in file %s", |
| gsym->name(), aarch64_relobj->name().c_str()); |
| return; |
| } |
| } |
| |
| // Get the symbol value. |
| typename Symbol_value<size>::Value value = psymval->value(aarch64_relobj, 0); |
| |
| // Owing to pipelining, the PC relative branches below actually skip |
| // two instructions when the branch offset is 0. |
| Address destination = static_cast<Address>(-1); |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_CALL26: |
| case elfcpp::R_AARCH64_JUMP26: |
| destination = value + addend; |
| break; |
| default: |
| gold_unreachable(); |
| } |
| |
| int stub_type = The_reloc_stub:: |
| stub_type_for_reloc(r_type, address, destination); |
| if (stub_type == ST_NONE) |
| return; |
| |
| The_stub_table* stub_table = aarch64_relobj->stub_table(relinfo->data_shndx); |
| gold_assert(stub_table != NULL); |
| |
| The_reloc_stub_key key(stub_type, gsym, aarch64_relobj, r_sym, addend); |
| The_reloc_stub* stub = stub_table->find_reloc_stub(key); |
| if (stub == NULL) |
| { |
| stub = new The_reloc_stub(stub_type); |
| stub_table->add_reloc_stub(stub, key); |
| } |
| stub->set_destination_address(destination); |
| } // End of Target_aarch64::scan_reloc_for_stub |
| |
| |
| // This function scans a relocation section for stub generation. |
| // The template parameter Relocate must be a class type which provides |
| // a single function, relocate(), which implements the machine |
| // specific part of a relocation. |
| |
| // BIG_ENDIAN is the endianness of the data. SH_TYPE is the section type: |
| // SHT_REL or SHT_RELA. |
| |
| // PRELOCS points to the relocation data. RELOC_COUNT is the number |
| // of relocs. OUTPUT_SECTION is the output section. |
| // NEEDS_SPECIAL_OFFSET_HANDLING is true if input offsets need to be |
| // mapped to output offsets. |
| |
| // VIEW is the section data, VIEW_ADDRESS is its memory address, and |
| // VIEW_SIZE is the size. These refer to the input section, unless |
| // NEEDS_SPECIAL_OFFSET_HANDLING is true, in which case they refer to |
| // the output section. |
| |
| template<int size, bool big_endian> |
| template<int sh_type> |
| void inline |
| Target_aarch64<size, big_endian>::scan_reloc_section_for_stubs( |
| const Relocate_info<size, big_endian>* relinfo, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* /*output_section*/, |
| bool /*needs_special_offset_handling*/, |
| const unsigned char* /*view*/, |
| Address view_address, |
| section_size_type) |
| { |
| typedef typename Reloc_types<sh_type,size,big_endian>::Reloc Reltype; |
| |
| const int reloc_size = |
| Reloc_types<sh_type,size,big_endian>::reloc_size; |
| AArch64_relobj<size, big_endian>* object = |
| static_cast<AArch64_relobj<size, big_endian>*>(relinfo->object); |
| unsigned int local_count = object->local_symbol_count(); |
| |
| gold::Default_comdat_behavior default_comdat_behavior; |
| Comdat_behavior comdat_behavior = CB_UNDETERMINED; |
| |
| for (size_t i = 0; i < reloc_count; ++i, prelocs += reloc_size) |
| { |
| Reltype reloc(prelocs); |
| typename elfcpp::Elf_types<size>::Elf_WXword r_info = reloc.get_r_info(); |
| unsigned int r_sym = elfcpp::elf_r_sym<size>(r_info); |
| unsigned int r_type = elfcpp::elf_r_type<size>(r_info); |
| if (r_type != elfcpp::R_AARCH64_CALL26 |
| && r_type != elfcpp::R_AARCH64_JUMP26) |
| continue; |
| |
| section_offset_type offset = |
| convert_to_section_size_type(reloc.get_r_offset()); |
| |
| // Get the addend. |
| typename elfcpp::Elf_types<size>::Elf_Swxword addend = |
| reloc.get_r_addend(); |
| |
| const Sized_symbol<size>* sym; |
| Symbol_value<size> symval; |
| const Symbol_value<size> *psymval; |
| bool is_defined_in_discarded_section; |
| unsigned int shndx; |
| const Symbol* gsym = NULL; |
| if (r_sym < local_count) |
| { |
| sym = NULL; |
| psymval = object->local_symbol(r_sym); |
| |
| // If the local symbol belongs to a section we are discarding, |
| // and that section is a debug section, try to find the |
| // corresponding kept section and map this symbol to its |
| // counterpart in the kept section. The symbol must not |
| // correspond to a section we are folding. |
| bool is_ordinary; |
| shndx = psymval->input_shndx(&is_ordinary); |
| is_defined_in_discarded_section = |
| (is_ordinary |
| && shndx != elfcpp::SHN_UNDEF |
| && !object->is_section_included(shndx) |
| && !relinfo->symtab->is_section_folded(object, shndx)); |
| |
| // We need to compute the would-be final value of this local |
| // symbol. |
| if (!is_defined_in_discarded_section) |
| { |
| typedef Sized_relobj_file<size, big_endian> ObjType; |
| if (psymval->is_section_symbol()) |
| symval.set_is_section_symbol(); |
| typename ObjType::Compute_final_local_value_status status = |
| object->compute_final_local_value(r_sym, psymval, &symval, |
| relinfo->symtab); |
| if (status == ObjType::CFLV_OK) |
| { |
| // Currently we cannot handle a branch to a target in |
| // a merged section. If this is the case, issue an error |
| // and also free the merge symbol value. |
| if (!symval.has_output_value()) |
| { |
| const std::string& section_name = |
| object->section_name(shndx); |
| object->error(_("cannot handle branch to local %u " |
| "in a merged section %s"), |
| r_sym, section_name.c_str()); |
| } |
| psymval = &symval; |
| } |
| else |
| { |
| // We cannot determine the final value. |
| continue; |
| } |
| } |
| } |
| else |
| { |
| gsym = object->global_symbol(r_sym); |
| gold_assert(gsym != NULL); |
| if (gsym->is_forwarder()) |
| gsym = relinfo->symtab->resolve_forwards(gsym); |
| |
| sym = static_cast<const Sized_symbol<size>*>(gsym); |
| if (sym->has_symtab_index() && sym->symtab_index() != -1U) |
| symval.set_output_symtab_index(sym->symtab_index()); |
| else |
| symval.set_no_output_symtab_entry(); |
| |
| // We need to compute the would-be final value of this global |
| // symbol. |
| const Symbol_table* symtab = relinfo->symtab; |
| const Sized_symbol<size>* sized_symbol = |
| symtab->get_sized_symbol<size>(gsym); |
| Symbol_table::Compute_final_value_status status; |
| typename elfcpp::Elf_types<size>::Elf_Addr value = |
| symtab->compute_final_value<size>(sized_symbol, &status); |
| |
| // Skip this if the symbol has not output section. |
| if (status == Symbol_table::CFVS_NO_OUTPUT_SECTION) |
| continue; |
| symval.set_output_value(value); |
| |
| if (gsym->type() == elfcpp::STT_TLS) |
| symval.set_is_tls_symbol(); |
| else if (gsym->type() == elfcpp::STT_GNU_IFUNC) |
| symval.set_is_ifunc_symbol(); |
| psymval = &symval; |
| |
| is_defined_in_discarded_section = |
| (gsym->is_defined_in_discarded_section() |
| && gsym->is_undefined()); |
| shndx = 0; |
| } |
| |
| Symbol_value<size> symval2; |
| if (is_defined_in_discarded_section) |
| { |
| std::string name = object->section_name(relinfo->data_shndx); |
| |
| if (comdat_behavior == CB_UNDETERMINED) |
| comdat_behavior = default_comdat_behavior.get(name.c_str()); |
| |
| if (comdat_behavior == CB_PRETEND) |
| { |
| bool found; |
| typename elfcpp::Elf_types<size>::Elf_Addr value = |
| object->map_to_kept_section(shndx, name, &found); |
| if (found) |
| symval2.set_output_value(value + psymval->input_value()); |
| else |
| symval2.set_output_value(0); |
| } |
| else |
| { |
| if (comdat_behavior == CB_ERROR) |
| issue_discarded_error(relinfo, i, offset, r_sym, gsym); |
| symval2.set_output_value(0); |
| } |
| symval2.set_no_output_symtab_entry(); |
| psymval = &symval2; |
| } |
| |
| this->scan_reloc_for_stub(relinfo, r_type, sym, r_sym, psymval, |
| addend, view_address + offset); |
| } // End of iterating relocs in a section |
| } // End of Target_aarch64::scan_reloc_section_for_stubs |
| |
| |
| // Scan an input section for stub generation. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::scan_section_for_stubs( |
| const Relocate_info<size, big_endian>* relinfo, |
| unsigned int sh_type, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| bool needs_special_offset_handling, |
| const unsigned char* view, |
| Address view_address, |
| section_size_type view_size) |
| { |
| gold_assert(sh_type == elfcpp::SHT_RELA); |
| this->scan_reloc_section_for_stubs<elfcpp::SHT_RELA>( |
| relinfo, |
| prelocs, |
| reloc_count, |
| output_section, |
| needs_special_offset_handling, |
| view, |
| view_address, |
| view_size); |
| } |
| |
| |
| // Relocate a single reloc stub. |
| |
| template<int size, bool big_endian> |
| void Target_aarch64<size, big_endian>:: |
| relocate_reloc_stub(The_reloc_stub* stub, |
| const The_relocate_info*, |
| Output_section*, |
| unsigned char* view, |
| Address address, |
| section_size_type) |
| { |
| typedef AArch64_relocate_functions<size, big_endian> The_reloc_functions; |
| typedef typename The_reloc_functions::Status The_reloc_functions_status; |
| typedef typename elfcpp::Swap<32,big_endian>::Valtype Insntype; |
| |
| Insntype* ip = reinterpret_cast<Insntype*>(view); |
| int insn_number = stub->insn_num(); |
| const uint32_t* insns = stub->insns(); |
| // Check the insns are really those stub insns. |
| for (int i = 0; i < insn_number; ++i) |
| { |
| Insntype insn = elfcpp::Swap<32,big_endian>::readval(ip + i); |
| gold_assert(((uint32_t)insn == insns[i])); |
| } |
| |
| Address dest = stub->destination_address(); |
| |
| switch(stub->type()) |
| { |
| case ST_ADRP_BRANCH: |
| { |
| // 1st reloc is ADR_PREL_PG_HI21 |
| The_reloc_functions_status status = |
| The_reloc_functions::adrp(view, dest, address); |
| // An error should never arise in the above step. If so, please |
| // check 'aarch64_valid_for_adrp_p'. |
| gold_assert(status == The_reloc_functions::STATUS_OKAY); |
| |
| // 2nd reloc is ADD_ABS_LO12_NC |
| const AArch64_reloc_property* arp = |
| aarch64_reloc_property_table->get_reloc_property( |
| elfcpp::R_AARCH64_ADD_ABS_LO12_NC); |
| gold_assert(arp != NULL); |
| status = The_reloc_functions::template |
| rela_general<32>(view + 4, dest, 0, arp); |
| // An error should never arise, it is an "_NC" relocation. |
| gold_assert(status == The_reloc_functions::STATUS_OKAY); |
| } |
| break; |
| |
| case ST_LONG_BRANCH_ABS: |
| // 1st reloc is R_AARCH64_PREL64, at offset 8 |
| elfcpp::Swap<64,big_endian>::writeval(view + 8, dest); |
| break; |
| |
| case ST_LONG_BRANCH_PCREL: |
| { |
| // "PC" calculation is the 2nd insn in the stub. |
| uint64_t offset = dest - (address + 4); |
| // Offset is placed at offset 4 and 5. |
| elfcpp::Swap<64,big_endian>::writeval(view + 16, offset); |
| } |
| break; |
| |
| default: |
| gold_unreachable(); |
| } |
| } |
| |
| |
| // A class to handle the PLT data. |
| // This is an abstract base class that handles most of the linker details |
| // but does not know the actual contents of PLT entries. The derived |
| // classes below fill in those details. |
| |
| template<int size, bool big_endian> |
| class Output_data_plt_aarch64 : public Output_section_data |
| { |
| public: |
| typedef Output_data_reloc<elfcpp::SHT_RELA, true, size, big_endian> |
| Reloc_section; |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr Address; |
| |
| Output_data_plt_aarch64(Layout* layout, |
| uint64_t addralign, |
| Output_data_got_aarch64<size, big_endian>* got, |
| Output_data_space* got_plt, |
| Output_data_space* got_irelative) |
| : Output_section_data(addralign), tlsdesc_rel_(NULL), irelative_rel_(NULL), |
| got_(got), got_plt_(got_plt), got_irelative_(got_irelative), |
| count_(0), irelative_count_(0), tlsdesc_got_offset_(-1U) |
| { this->init(layout); } |
| |
| // Initialize the PLT section. |
| void |
| init(Layout* layout); |
| |
| // Add an entry to the PLT. |
| void |
| add_entry(Symbol_table*, Layout*, Symbol* gsym); |
| |
| // Add an entry to the PLT for a local STT_GNU_IFUNC symbol. |
| unsigned int |
| add_local_ifunc_entry(Symbol_table* symtab, Layout*, |
| Sized_relobj_file<size, big_endian>* relobj, |
| unsigned int local_sym_index); |
| |
| // Add the relocation for a PLT entry. |
| void |
| add_relocation(Symbol_table*, Layout*, Symbol* gsym, |
| unsigned int got_offset); |
| |
| // Add the reserved TLSDESC_PLT entry to the PLT. |
| void |
| reserve_tlsdesc_entry(unsigned int got_offset) |
| { this->tlsdesc_got_offset_ = got_offset; } |
| |
| // Return true if a TLSDESC_PLT entry has been reserved. |
| bool |
| has_tlsdesc_entry() const |
| { return this->tlsdesc_got_offset_ != -1U; } |
| |
| // Return the GOT offset for the reserved TLSDESC_PLT entry. |
| unsigned int |
| get_tlsdesc_got_offset() const |
| { return this->tlsdesc_got_offset_; } |
| |
| // Return the PLT offset of the reserved TLSDESC_PLT entry. |
| unsigned int |
| get_tlsdesc_plt_offset() const |
| { |
| return (this->first_plt_entry_offset() + |
| (this->count_ + this->irelative_count_) |
| * this->get_plt_entry_size()); |
| } |
| |
| // Return the .rela.plt section data. |
| Reloc_section* |
| rela_plt() |
| { return this->rel_; } |
| |
| // Return where the TLSDESC relocations should go. |
| Reloc_section* |
| rela_tlsdesc(Layout*); |
| |
| // Return where the IRELATIVE relocations should go in the PLT |
| // relocations. |
| Reloc_section* |
| rela_irelative(Symbol_table*, Layout*); |
| |
| // Return whether we created a section for IRELATIVE relocations. |
| bool |
| has_irelative_section() const |
| { return this->irelative_rel_ != NULL; } |
| |
| // Return the number of PLT entries. |
| unsigned int |
| entry_count() const |
| { return this->count_ + this->irelative_count_; } |
| |
| // Return the offset of the first non-reserved PLT entry. |
| unsigned int |
| first_plt_entry_offset() const |
| { return this->do_first_plt_entry_offset(); } |
| |
| // Return the size of a PLT entry. |
| unsigned int |
| get_plt_entry_size() const |
| { return this->do_get_plt_entry_size(); } |
| |
| // Return the reserved tlsdesc entry size. |
| unsigned int |
| get_plt_tlsdesc_entry_size() const |
| { return this->do_get_plt_tlsdesc_entry_size(); } |
| |
| // Return the PLT address to use for a global symbol. |
| uint64_t |
| address_for_global(const Symbol*); |
| |
| // Return the PLT address to use for a local symbol. |
| uint64_t |
| address_for_local(const Relobj*, unsigned int symndx); |
| |
| protected: |
| // Fill in the first PLT entry. |
| void |
| fill_first_plt_entry(unsigned char* pov, |
| Address got_address, |
| Address plt_address) |
| { this->do_fill_first_plt_entry(pov, got_address, plt_address); } |
| |
| // Fill in a normal PLT entry. |
| void |
| fill_plt_entry(unsigned char* pov, |
| Address got_address, |
| Address plt_address, |
| unsigned int got_offset, |
| unsigned int plt_offset) |
| { |
| this->do_fill_plt_entry(pov, got_address, plt_address, |
| got_offset, plt_offset); |
| } |
| |
| // Fill in the reserved TLSDESC PLT entry. |
| void |
| fill_tlsdesc_entry(unsigned char* pov, |
| Address gotplt_address, |
| Address plt_address, |
| Address got_base, |
| unsigned int tlsdesc_got_offset, |
| unsigned int plt_offset) |
| { |
| this->do_fill_tlsdesc_entry(pov, gotplt_address, plt_address, got_base, |
| tlsdesc_got_offset, plt_offset); |
| } |
| |
| virtual unsigned int |
| do_first_plt_entry_offset() const = 0; |
| |
| virtual unsigned int |
| do_get_plt_entry_size() const = 0; |
| |
| virtual unsigned int |
| do_get_plt_tlsdesc_entry_size() const = 0; |
| |
| virtual void |
| do_fill_first_plt_entry(unsigned char* pov, |
| Address got_addr, |
| Address plt_addr) = 0; |
| |
| virtual void |
| do_fill_plt_entry(unsigned char* pov, |
| Address got_address, |
| Address plt_address, |
| unsigned int got_offset, |
| unsigned int plt_offset) = 0; |
| |
| virtual void |
| do_fill_tlsdesc_entry(unsigned char* pov, |
| Address gotplt_address, |
| Address plt_address, |
| Address got_base, |
| unsigned int tlsdesc_got_offset, |
| unsigned int plt_offset) = 0; |
| |
| void |
| do_adjust_output_section(Output_section* os); |
| |
| // Write to a map file. |
| void |
| do_print_to_mapfile(Mapfile* mapfile) const |
| { mapfile->print_output_data(this, _("** PLT")); } |
| |
| private: |
| // Set the final size. |
| void |
| set_final_data_size(); |
| |
| // Write out the PLT data. |
| void |
| do_write(Output_file*); |
| |
| // The reloc section. |
| Reloc_section* rel_; |
| |
| // The TLSDESC relocs, if necessary. These must follow the regular |
| // PLT relocs. |
| Reloc_section* tlsdesc_rel_; |
| |
| // The IRELATIVE relocs, if necessary. These must follow the |
| // regular PLT relocations. |
| Reloc_section* irelative_rel_; |
| |
| // The .got section. |
| Output_data_got_aarch64<size, big_endian>* got_; |
| |
| // The .got.plt section. |
| Output_data_space* got_plt_; |
| |
| // The part of the .got.plt section used for IRELATIVE relocs. |
| Output_data_space* got_irelative_; |
| |
| // The number of PLT entries. |
| unsigned int count_; |
| |
| // Number of PLT entries with R_AARCH64_IRELATIVE relocs. These |
| // follow the regular PLT entries. |
| unsigned int irelative_count_; |
| |
| // GOT offset of the reserved TLSDESC_GOT entry for the lazy trampoline. |
| // Communicated to the loader via DT_TLSDESC_GOT. The magic value -1 |
| // indicates an offset is not allocated. |
| unsigned int tlsdesc_got_offset_; |
| }; |
| |
| // Initialize the PLT section. |
| |
| template<int size, bool big_endian> |
| void |
| Output_data_plt_aarch64<size, big_endian>::init(Layout* layout) |
| { |
| this->rel_ = new Reloc_section(false); |
| layout->add_output_section_data(".rela.plt", elfcpp::SHT_RELA, |
| elfcpp::SHF_ALLOC, this->rel_, |
| ORDER_DYNAMIC_PLT_RELOCS, false); |
| } |
| |
| template<int size, bool big_endian> |
| void |
| Output_data_plt_aarch64<size, big_endian>::do_adjust_output_section( |
| Output_section* os) |
| { |
| os->set_entsize(this->get_plt_entry_size()); |
| } |
| |
| // Add an entry to the PLT. |
| |
| template<int size, bool big_endian> |
| void |
| Output_data_plt_aarch64<size, big_endian>::add_entry(Symbol_table* symtab, |
| Layout* layout, Symbol* gsym) |
| { |
| gold_assert(!gsym->has_plt_offset()); |
| |
| unsigned int* pcount; |
| unsigned int plt_reserved; |
| Output_section_data_build* got; |
| |
| if (gsym->type() == elfcpp::STT_GNU_IFUNC |
| && gsym->can_use_relative_reloc(false)) |
| { |
| pcount = &this->irelative_count_; |
| plt_reserved = 0; |
| got = this->got_irelative_; |
| } |
| else |
| { |
| pcount = &this->count_; |
| plt_reserved = this->first_plt_entry_offset(); |
| got = this->got_plt_; |
| } |
| |
| gsym->set_plt_offset((*pcount) * this->get_plt_entry_size() |
| + plt_reserved); |
| |
| ++*pcount; |
| |
| section_offset_type got_offset = got->current_data_size(); |
| |
| // Every PLT entry needs a GOT entry which points back to the PLT |
| // entry (this will be changed by the dynamic linker, normally |
| // lazily when the function is called). |
| got->set_current_data_size(got_offset + size / 8); |
| |
| // Every PLT entry needs a reloc. |
| this->add_relocation(symtab, layout, gsym, got_offset); |
| |
| // Note that we don't need to save the symbol. The contents of the |
| // PLT are independent of which symbols are used. The symbols only |
| // appear in the relocations. |
| } |
| |
| // Add an entry to the PLT for a local STT_GNU_IFUNC symbol. Return |
| // the PLT offset. |
| |
| template<int size, bool big_endian> |
| unsigned int |
| Output_data_plt_aarch64<size, big_endian>::add_local_ifunc_entry( |
| Symbol_table* symtab, |
| Layout* layout, |
| Sized_relobj_file<size, big_endian>* relobj, |
| unsigned int local_sym_index) |
| { |
| unsigned int plt_offset = this->irelative_count_ * this->get_plt_entry_size(); |
| ++this->irelative_count_; |
| |
| section_offset_type got_offset = this->got_irelative_->current_data_size(); |
| |
| // Every PLT entry needs a GOT entry which points back to the PLT |
| // entry. |
| this->got_irelative_->set_current_data_size(got_offset + size / 8); |
| |
| // Every PLT entry needs a reloc. |
| Reloc_section* rela = this->rela_irelative(symtab, layout); |
| rela->add_symbolless_local_addend(relobj, local_sym_index, |
| elfcpp::R_AARCH64_IRELATIVE, |
| this->got_irelative_, got_offset, 0); |
| |
| return plt_offset; |
| } |
| |
| // Add the relocation for a PLT entry. |
| |
| template<int size, bool big_endian> |
| void |
| Output_data_plt_aarch64<size, big_endian>::add_relocation( |
| Symbol_table* symtab, Layout* layout, Symbol* gsym, unsigned int got_offset) |
| { |
| if (gsym->type() == elfcpp::STT_GNU_IFUNC |
| && gsym->can_use_relative_reloc(false)) |
| { |
| Reloc_section* rela = this->rela_irelative(symtab, layout); |
| rela->add_symbolless_global_addend(gsym, elfcpp::R_AARCH64_IRELATIVE, |
| this->got_irelative_, got_offset, 0); |
| } |
| else |
| { |
| gsym->set_needs_dynsym_entry(); |
| this->rel_->add_global(gsym, elfcpp::R_AARCH64_JUMP_SLOT, this->got_plt_, |
| got_offset, 0); |
| } |
| } |
| |
| // Return where the TLSDESC relocations should go, creating it if |
| // necessary. These follow the JUMP_SLOT relocations. |
| |
| template<int size, bool big_endian> |
| typename Output_data_plt_aarch64<size, big_endian>::Reloc_section* |
| Output_data_plt_aarch64<size, big_endian>::rela_tlsdesc(Layout* layout) |
| { |
| if (this->tlsdesc_rel_ == NULL) |
| { |
| this->tlsdesc_rel_ = new Reloc_section(false); |
| layout->add_output_section_data(".rela.plt", elfcpp::SHT_RELA, |
| elfcpp::SHF_ALLOC, this->tlsdesc_rel_, |
| ORDER_DYNAMIC_PLT_RELOCS, false); |
| gold_assert(this->tlsdesc_rel_->output_section() |
| == this->rel_->output_section()); |
| } |
| return this->tlsdesc_rel_; |
| } |
| |
| // Return where the IRELATIVE relocations should go in the PLT. These |
| // follow the JUMP_SLOT and the TLSDESC relocations. |
| |
| template<int size, bool big_endian> |
| typename Output_data_plt_aarch64<size, big_endian>::Reloc_section* |
| Output_data_plt_aarch64<size, big_endian>::rela_irelative(Symbol_table* symtab, |
| Layout* layout) |
| { |
| if (this->irelative_rel_ == NULL) |
| { |
| // Make sure we have a place for the TLSDESC relocations, in |
| // case we see any later on. |
| this->rela_tlsdesc(layout); |
| this->irelative_rel_ = new Reloc_section(false); |
| layout->add_output_section_data(".rela.plt", elfcpp::SHT_RELA, |
| elfcpp::SHF_ALLOC, this->irelative_rel_, |
| ORDER_DYNAMIC_PLT_RELOCS, false); |
| gold_assert(this->irelative_rel_->output_section() |
| == this->rel_->output_section()); |
| |
| if (parameters->doing_static_link()) |
| { |
| // A statically linked executable will only have a .rela.plt |
| // section to hold R_AARCH64_IRELATIVE relocs for |
| // STT_GNU_IFUNC symbols. The library will use these |
| // symbols to locate the IRELATIVE relocs at program startup |
| // time. |
| symtab->define_in_output_data("__rela_iplt_start", NULL, |
| Symbol_table::PREDEFINED, |
| this->irelative_rel_, 0, 0, |
| elfcpp::STT_NOTYPE, elfcpp::STB_GLOBAL, |
| elfcpp::STV_HIDDEN, 0, false, true); |
| symtab->define_in_output_data("__rela_iplt_end", NULL, |
| Symbol_table::PREDEFINED, |
| this->irelative_rel_, 0, 0, |
| elfcpp::STT_NOTYPE, elfcpp::STB_GLOBAL, |
| elfcpp::STV_HIDDEN, 0, true, true); |
| } |
| } |
| return this->irelative_rel_; |
| } |
| |
| // Return the PLT address to use for a global symbol. |
| |
| template<int size, bool big_endian> |
| uint64_t |
| Output_data_plt_aarch64<size, big_endian>::address_for_global( |
| const Symbol* gsym) |
| { |
| uint64_t offset = 0; |
| if (gsym->type() == elfcpp::STT_GNU_IFUNC |
| && gsym->can_use_relative_reloc(false)) |
| offset = (this->first_plt_entry_offset() + |
| this->count_ * this->get_plt_entry_size()); |
| return this->address() + offset + gsym->plt_offset(); |
| } |
| |
| // Return the PLT address to use for a local symbol. These are always |
| // IRELATIVE relocs. |
| |
| template<int size, bool big_endian> |
| uint64_t |
| Output_data_plt_aarch64<size, big_endian>::address_for_local( |
| const Relobj* object, |
| unsigned int r_sym) |
| { |
| return (this->address() |
| + this->first_plt_entry_offset() |
| + this->count_ * this->get_plt_entry_size() |
| + object->local_plt_offset(r_sym)); |
| } |
| |
| // Set the final size. |
| |
| template<int size, bool big_endian> |
| void |
| Output_data_plt_aarch64<size, big_endian>::set_final_data_size() |
| { |
| unsigned int count = this->count_ + this->irelative_count_; |
| unsigned int extra_size = 0; |
| if (this->has_tlsdesc_entry()) |
| extra_size += this->get_plt_tlsdesc_entry_size(); |
| this->set_data_size(this->first_plt_entry_offset() |
| + count * this->get_plt_entry_size() |
| + extra_size); |
| } |
| |
| template<int size, bool big_endian> |
| class Output_data_plt_aarch64_standard : |
| public Output_data_plt_aarch64<size, big_endian> |
| { |
| public: |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr Address; |
| Output_data_plt_aarch64_standard( |
| Layout* layout, |
| Output_data_got_aarch64<size, big_endian>* got, |
| Output_data_space* got_plt, |
| Output_data_space* got_irelative) |
| : Output_data_plt_aarch64<size, big_endian>(layout, |
| size == 32 ? 4 : 8, |
| got, got_plt, |
| got_irelative) |
| { } |
| |
| protected: |
| // Return the offset of the first non-reserved PLT entry. |
| virtual unsigned int |
| do_first_plt_entry_offset() const |
| { return this->first_plt_entry_size; } |
| |
| // Return the size of a PLT entry |
| virtual unsigned int |
| do_get_plt_entry_size() const |
| { return this->plt_entry_size; } |
| |
| // Return the size of a tlsdesc entry |
| virtual unsigned int |
| do_get_plt_tlsdesc_entry_size() const |
| { return this->plt_tlsdesc_entry_size; } |
| |
| virtual void |
| do_fill_first_plt_entry(unsigned char* pov, |
| Address got_address, |
| Address plt_address); |
| |
| virtual void |
| do_fill_plt_entry(unsigned char* pov, |
| Address got_address, |
| Address plt_address, |
| unsigned int got_offset, |
| unsigned int plt_offset); |
| |
| virtual void |
| do_fill_tlsdesc_entry(unsigned char* pov, |
| Address gotplt_address, |
| Address plt_address, |
| Address got_base, |
| unsigned int tlsdesc_got_offset, |
| unsigned int plt_offset); |
| |
| private: |
| // The size of the first plt entry size. |
| static const int first_plt_entry_size = 32; |
| // The size of the plt entry size. |
| static const int plt_entry_size = 16; |
| // The size of the plt tlsdesc entry size. |
| static const int plt_tlsdesc_entry_size = 32; |
| // Template for the first PLT entry. |
| static const uint32_t first_plt_entry[first_plt_entry_size / 4]; |
| // Template for subsequent PLT entries. |
| static const uint32_t plt_entry[plt_entry_size / 4]; |
| // The reserved TLSDESC entry in the PLT for an executable. |
| static const uint32_t tlsdesc_plt_entry[plt_tlsdesc_entry_size / 4]; |
| }; |
| |
| // The first entry in the PLT for an executable. |
| |
| template<> |
| const uint32_t |
| Output_data_plt_aarch64_standard<32, false>:: |
| first_plt_entry[first_plt_entry_size / 4] = |
| { |
| 0xa9bf7bf0, /* stp x16, x30, [sp, #-16]! */ |
| 0x90000010, /* adrp x16, PLT_GOT+0x8 */ |
| 0xb9400A11, /* ldr w17, [x16, #PLT_GOT+0x8] */ |
| 0x11002210, /* add w16, w16,#PLT_GOT+0x8 */ |
| 0xd61f0220, /* br x17 */ |
| 0xd503201f, /* nop */ |
| 0xd503201f, /* nop */ |
| 0xd503201f, /* nop */ |
| }; |
| |
| |
| template<> |
| const uint32_t |
| Output_data_plt_aarch64_standard<32, true>:: |
| first_plt_entry[first_plt_entry_size / 4] = |
| { |
| 0xa9bf7bf0, /* stp x16, x30, [sp, #-16]! */ |
| 0x90000010, /* adrp x16, PLT_GOT+0x8 */ |
| 0xb9400A11, /* ldr w17, [x16, #PLT_GOT+0x8] */ |
| 0x11002210, /* add w16, w16,#PLT_GOT+0x8 */ |
| 0xd61f0220, /* br x17 */ |
| 0xd503201f, /* nop */ |
| 0xd503201f, /* nop */ |
| 0xd503201f, /* nop */ |
| }; |
| |
| |
| template<> |
| const uint32_t |
| Output_data_plt_aarch64_standard<64, false>:: |
| first_plt_entry[first_plt_entry_size / 4] = |
| { |
| 0xa9bf7bf0, /* stp x16, x30, [sp, #-16]! */ |
| 0x90000010, /* adrp x16, PLT_GOT+16 */ |
| 0xf9400A11, /* ldr x17, [x16, #PLT_GOT+0x10] */ |
| 0x91004210, /* add x16, x16,#PLT_GOT+0x10 */ |
| 0xd61f0220, /* br x17 */ |
| 0xd503201f, /* nop */ |
| 0xd503201f, /* nop */ |
| 0xd503201f, /* nop */ |
| }; |
| |
| |
| template<> |
| const uint32_t |
| Output_data_plt_aarch64_standard<64, true>:: |
| first_plt_entry[first_plt_entry_size / 4] = |
| { |
| 0xa9bf7bf0, /* stp x16, x30, [sp, #-16]! */ |
| 0x90000010, /* adrp x16, PLT_GOT+16 */ |
| 0xf9400A11, /* ldr x17, [x16, #PLT_GOT+0x10] */ |
| 0x91004210, /* add x16, x16,#PLT_GOT+0x10 */ |
| 0xd61f0220, /* br x17 */ |
| 0xd503201f, /* nop */ |
| 0xd503201f, /* nop */ |
| 0xd503201f, /* nop */ |
| }; |
| |
| |
| template<> |
| const uint32_t |
| Output_data_plt_aarch64_standard<32, false>:: |
| plt_entry[plt_entry_size / 4] = |
| { |
| 0x90000010, /* adrp x16, PLTGOT + n * 4 */ |
| 0xb9400211, /* ldr w17, [w16, PLTGOT + n * 4] */ |
| 0x11000210, /* add w16, w16, :lo12:PLTGOT + n * 4 */ |
| 0xd61f0220, /* br x17. */ |
| }; |
| |
| |
| template<> |
| const uint32_t |
| Output_data_plt_aarch64_standard<32, true>:: |
| plt_entry[plt_entry_size / 4] = |
| { |
| 0x90000010, /* adrp x16, PLTGOT + n * 4 */ |
| 0xb9400211, /* ldr w17, [w16, PLTGOT + n * 4] */ |
| 0x11000210, /* add w16, w16, :lo12:PLTGOT + n * 4 */ |
| 0xd61f0220, /* br x17. */ |
| }; |
| |
| |
| template<> |
| const uint32_t |
| Output_data_plt_aarch64_standard<64, false>:: |
| plt_entry[plt_entry_size / 4] = |
| { |
| 0x90000010, /* adrp x16, PLTGOT + n * 8 */ |
| 0xf9400211, /* ldr x17, [x16, PLTGOT + n * 8] */ |
| 0x91000210, /* add x16, x16, :lo12:PLTGOT + n * 8 */ |
| 0xd61f0220, /* br x17. */ |
| }; |
| |
| |
| template<> |
| const uint32_t |
| Output_data_plt_aarch64_standard<64, true>:: |
| plt_entry[plt_entry_size / 4] = |
| { |
| 0x90000010, /* adrp x16, PLTGOT + n * 8 */ |
| 0xf9400211, /* ldr x17, [x16, PLTGOT + n * 8] */ |
| 0x91000210, /* add x16, x16, :lo12:PLTGOT + n * 8 */ |
| 0xd61f0220, /* br x17. */ |
| }; |
| |
| |
| template<int size, bool big_endian> |
| void |
| Output_data_plt_aarch64_standard<size, big_endian>::do_fill_first_plt_entry( |
| unsigned char* pov, |
| Address got_address, |
| Address plt_address) |
| { |
| // PLT0 of the small PLT looks like this in ELF64 - |
| // stp x16, x30, [sp, #-16]! Save the reloc and lr on stack. |
| // adrp x16, PLT_GOT + 16 Get the page base of the GOTPLT |
| // ldr x17, [x16, #:lo12:PLT_GOT+16] Load the address of the |
| // symbol resolver |
| // add x16, x16, #:lo12:PLT_GOT+16 Load the lo12 bits of the |
| // GOTPLT entry for this. |
| // br x17 |
| // PLT0 will be slightly different in ELF32 due to different got entry |
| // size. |
| memcpy(pov, this->first_plt_entry, this->first_plt_entry_size); |
| Address gotplt_2nd_ent = got_address + (size / 8) * 2; |
| |
| // Fill in the top 21 bits for this: ADRP x16, PLT_GOT + 8 * 2. |
| // ADRP: (PG(S+A)-PG(P)) >> 12) & 0x1fffff. |
| // FIXME: This only works for 64bit |
| AArch64_relocate_functions<size, big_endian>::adrp(pov + 4, |
| gotplt_2nd_ent, plt_address + 4); |
| |
| // Fill in R_AARCH64_LDST8_LO12 |
| elfcpp::Swap<32, big_endian>::writeval( |
| pov + 8, |
| ((this->first_plt_entry[2] & 0xffc003ff) |
| | ((gotplt_2nd_ent & 0xff8) << 7))); |
| |
| // Fill in R_AARCH64_ADD_ABS_LO12 |
| elfcpp::Swap<32, big_endian>::writeval( |
| pov + 12, |
| ((this->first_plt_entry[3] & 0xffc003ff) |
| | ((gotplt_2nd_ent & 0xfff) << 10))); |
| } |
| |
| |
| // Subsequent entries in the PLT for an executable. |
| // FIXME: This only works for 64bit |
| |
| template<int size, bool big_endian> |
| void |
| Output_data_plt_aarch64_standard<size, big_endian>::do_fill_plt_entry( |
| unsigned char* pov, |
| Address got_address, |
| Address plt_address, |
| unsigned int got_offset, |
| unsigned int plt_offset) |
| { |
| memcpy(pov, this->plt_entry, this->plt_entry_size); |
| |
| Address gotplt_entry_address = got_address + got_offset; |
| Address plt_entry_address = plt_address + plt_offset; |
| |
| // Fill in R_AARCH64_PCREL_ADR_HI21 |
| AArch64_relocate_functions<size, big_endian>::adrp( |
| pov, |
| gotplt_entry_address, |
| plt_entry_address); |
| |
| // Fill in R_AARCH64_LDST64_ABS_LO12 |
| elfcpp::Swap<32, big_endian>::writeval( |
| pov + 4, |
| ((this->plt_entry[1] & 0xffc003ff) |
| | ((gotplt_entry_address & 0xff8) << 7))); |
| |
| // Fill in R_AARCH64_ADD_ABS_LO12 |
| elfcpp::Swap<32, big_endian>::writeval( |
| pov + 8, |
| ((this->plt_entry[2] & 0xffc003ff) |
| | ((gotplt_entry_address & 0xfff) <<10))); |
| |
| } |
| |
| |
| template<> |
| const uint32_t |
| Output_data_plt_aarch64_standard<32, false>:: |
| tlsdesc_plt_entry[plt_tlsdesc_entry_size / 4] = |
| { |
| 0xa9bf0fe2, /* stp x2, x3, [sp, #-16]! */ |
| 0x90000002, /* adrp x2, 0 */ |
| 0x90000003, /* adrp x3, 0 */ |
| 0xb9400042, /* ldr w2, [w2, #0] */ |
| 0x11000063, /* add w3, w3, 0 */ |
| 0xd61f0040, /* br x2 */ |
| 0xd503201f, /* nop */ |
| 0xd503201f, /* nop */ |
| }; |
| |
| template<> |
| const uint32_t |
| Output_data_plt_aarch64_standard<32, true>:: |
| tlsdesc_plt_entry[plt_tlsdesc_entry_size / 4] = |
| { |
| 0xa9bf0fe2, /* stp x2, x3, [sp, #-16]! */ |
| 0x90000002, /* adrp x2, 0 */ |
| 0x90000003, /* adrp x3, 0 */ |
| 0xb9400042, /* ldr w2, [w2, #0] */ |
| 0x11000063, /* add w3, w3, 0 */ |
| 0xd61f0040, /* br x2 */ |
| 0xd503201f, /* nop */ |
| 0xd503201f, /* nop */ |
| }; |
| |
| template<> |
| const uint32_t |
| Output_data_plt_aarch64_standard<64, false>:: |
| tlsdesc_plt_entry[plt_tlsdesc_entry_size / 4] = |
| { |
| 0xa9bf0fe2, /* stp x2, x3, [sp, #-16]! */ |
| 0x90000002, /* adrp x2, 0 */ |
| 0x90000003, /* adrp x3, 0 */ |
| 0xf9400042, /* ldr x2, [x2, #0] */ |
| 0x91000063, /* add x3, x3, 0 */ |
| 0xd61f0040, /* br x2 */ |
| 0xd503201f, /* nop */ |
| 0xd503201f, /* nop */ |
| }; |
| |
| template<> |
| const uint32_t |
| Output_data_plt_aarch64_standard<64, true>:: |
| tlsdesc_plt_entry[plt_tlsdesc_entry_size / 4] = |
| { |
| 0xa9bf0fe2, /* stp x2, x3, [sp, #-16]! */ |
| 0x90000002, /* adrp x2, 0 */ |
| 0x90000003, /* adrp x3, 0 */ |
| 0xf9400042, /* ldr x2, [x2, #0] */ |
| 0x91000063, /* add x3, x3, 0 */ |
| 0xd61f0040, /* br x2 */ |
| 0xd503201f, /* nop */ |
| 0xd503201f, /* nop */ |
| }; |
| |
| template<int size, bool big_endian> |
| void |
| Output_data_plt_aarch64_standard<size, big_endian>::do_fill_tlsdesc_entry( |
| unsigned char* pov, |
| Address gotplt_address, |
| Address plt_address, |
| Address got_base, |
| unsigned int tlsdesc_got_offset, |
| unsigned int plt_offset) |
| { |
| memcpy(pov, tlsdesc_plt_entry, plt_tlsdesc_entry_size); |
| |
| // move DT_TLSDESC_GOT address into x2 |
| // move .got.plt address into x3 |
| Address tlsdesc_got_entry = got_base + tlsdesc_got_offset; |
| Address plt_entry_address = plt_address + plt_offset; |
| |
| // R_AARCH64_ADR_PREL_PG_HI21 |
| AArch64_relocate_functions<size, big_endian>::adrp( |
| pov + 4, |
| tlsdesc_got_entry, |
| plt_entry_address + 4); |
| |
| // R_AARCH64_ADR_PREL_PG_HI21 |
| AArch64_relocate_functions<size, big_endian>::adrp( |
| pov + 8, |
| gotplt_address, |
| plt_entry_address + 8); |
| |
| // R_AARCH64_LDST64_ABS_LO12 |
| elfcpp::Swap<32, big_endian>::writeval( |
| pov + 12, |
| ((this->tlsdesc_plt_entry[3] & 0xffc003ff) |
| | ((tlsdesc_got_entry & 0xff8) << 7))); |
| |
| // R_AARCH64_ADD_ABS_LO12 |
| elfcpp::Swap<32, big_endian>::writeval( |
| pov + 16, |
| ((this->tlsdesc_plt_entry[4] & 0xffc003ff) |
| | ((gotplt_address & 0xfff) << 10))); |
| } |
| |
| // Write out the PLT. This uses the hand-coded instructions above, |
| // and adjusts them as needed. This is specified by the AMD64 ABI. |
| |
| template<int size, bool big_endian> |
| void |
| Output_data_plt_aarch64<size, big_endian>::do_write(Output_file* of) |
| { |
| const off_t offset = this->offset(); |
| const section_size_type oview_size = |
| convert_to_section_size_type(this->data_size()); |
| unsigned char* const oview = of->get_output_view(offset, oview_size); |
| |
| const off_t got_file_offset = this->got_plt_->offset(); |
| gold_assert(got_file_offset + this->got_plt_->data_size() |
| == this->got_irelative_->offset()); |
| |
| const section_size_type got_size = |
| convert_to_section_size_type(this->got_plt_->data_size() |
| + this->got_irelative_->data_size()); |
| unsigned char* const got_view = of->get_output_view(got_file_offset, |
| got_size); |
| |
| unsigned char* pov = oview; |
| |
| // The base address of the .plt section. |
| typename elfcpp::Elf_types<size>::Elf_Addr plt_address = this->address(); |
| // The base address of the PLT portion of the .got section. |
| typename elfcpp::Elf_types<size>::Elf_Addr gotplt_address |
| = this->got_plt_->address(); |
| |
| this->fill_first_plt_entry(pov, gotplt_address, plt_address); |
| pov += this->first_plt_entry_offset(); |
| |
| // The first three entries in .got.plt are reserved. |
| unsigned char* got_pov = got_view; |
| memset(got_pov, 0, size / 8 * AARCH64_GOTPLT_RESERVE_COUNT); |
| got_pov += (size / 8) * AARCH64_GOTPLT_RESERVE_COUNT; |
| |
| unsigned int plt_offset = this->first_plt_entry_offset(); |
| unsigned int got_offset = (size / 8) * AARCH64_GOTPLT_RESERVE_COUNT; |
| const unsigned int count = this->count_ + this->irelative_count_; |
| for (unsigned int plt_index = 0; |
| plt_index < count; |
| ++plt_index, |
| pov += this->get_plt_entry_size(), |
| got_pov += size / 8, |
| plt_offset += this->get_plt_entry_size(), |
| got_offset += size / 8) |
| { |
| // Set and adjust the PLT entry itself. |
| this->fill_plt_entry(pov, gotplt_address, plt_address, |
| got_offset, plt_offset); |
| |
| // Set the entry in the GOT, which points to plt0. |
| elfcpp::Swap<size, big_endian>::writeval(got_pov, plt_address); |
| } |
| |
| if (this->has_tlsdesc_entry()) |
| { |
| // Set and adjust the reserved TLSDESC PLT entry. |
| unsigned int tlsdesc_got_offset = this->get_tlsdesc_got_offset(); |
| // The base address of the .base section. |
| typename elfcpp::Elf_types<size>::Elf_Addr got_base = |
| this->got_->address(); |
| this->fill_tlsdesc_entry(pov, gotplt_address, plt_address, got_base, |
| tlsdesc_got_offset, plt_offset); |
| pov += this->get_plt_tlsdesc_entry_size(); |
| } |
| |
| gold_assert(static_cast<section_size_type>(pov - oview) == oview_size); |
| gold_assert(static_cast<section_size_type>(got_pov - got_view) == got_size); |
| |
| of->write_output_view(offset, oview_size, oview); |
| of->write_output_view(got_file_offset, got_size, got_view); |
| } |
| |
| // Telling how to update the immediate field of an instruction. |
| struct AArch64_howto |
| { |
| // The immediate field mask. |
| elfcpp::Elf_Xword dst_mask; |
| |
| // The offset to apply relocation immediate |
| int doffset; |
| |
| // The second part offset, if the immediate field has two parts. |
| // -1 if the immediate field has only one part. |
| int doffset2; |
| }; |
| |
| static const AArch64_howto aarch64_howto[AArch64_reloc_property::INST_NUM] = |
| { |
| {0, -1, -1}, // DATA |
| {0x1fffe0, 5, -1}, // MOVW [20:5]-imm16 |
| {0xffffe0, 5, -1}, // LD [23:5]-imm19 |
| {0x60ffffe0, 29, 5}, // ADR [30:29]-immlo [23:5]-immhi |
| {0x60ffffe0, 29, 5}, // ADRP [30:29]-immlo [23:5]-immhi |
| {0x3ffc00, 10, -1}, // ADD [21:10]-imm12 |
| {0x3ffc00, 10, -1}, // LDST [21:10]-imm12 |
| {0x7ffe0, 5, -1}, // TBZNZ [18:5]-imm14 |
| {0xffffe0, 5, -1}, // CONDB [23:5]-imm19 |
| {0x3ffffff, 0, -1}, // B [25:0]-imm26 |
| {0x3ffffff, 0, -1}, // CALL [25:0]-imm26 |
| }; |
| |
| // AArch64 relocate function class |
| |
| template<int size, bool big_endian> |
| class AArch64_relocate_functions |
| { |
| public: |
| typedef enum |
| { |
| STATUS_OKAY, // No error during relocation. |
| STATUS_OVERFLOW, // Relocation overflow. |
| STATUS_BAD_RELOC, // Relocation cannot be applied. |
| } Status; |
| |
| typedef AArch64_relocate_functions<size, big_endian> This; |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr Address; |
| typedef Relocate_info<size, big_endian> The_relocate_info; |
| typedef AArch64_relobj<size, big_endian> The_aarch64_relobj; |
| typedef Reloc_stub<size, big_endian> The_reloc_stub; |
| typedef Stub_table<size, big_endian> The_stub_table; |
| typedef elfcpp::Rela<size, big_endian> The_rela; |
| typedef typename elfcpp::Swap<size, big_endian>::Valtype AArch64_valtype; |
| |
| // Return the page address of the address. |
| // Page(address) = address & ~0xFFF |
| |
| static inline AArch64_valtype |
| Page(Address address) |
| { |
| return (address & (~static_cast<Address>(0xFFF))); |
| } |
| |
| private: |
| // Update instruction (pointed by view) with selected bits (immed). |
| // val = (val & ~dst_mask) | (immed << doffset) |
| |
| template<int valsize> |
| static inline void |
| update_view(unsigned char* view, |
| AArch64_valtype immed, |
| elfcpp::Elf_Xword doffset, |
| elfcpp::Elf_Xword dst_mask) |
| { |
| typedef typename elfcpp::Swap<valsize, big_endian>::Valtype Valtype; |
| Valtype* wv = reinterpret_cast<Valtype*>(view); |
| Valtype val = elfcpp::Swap<valsize, big_endian>::readval(wv); |
| |
| // Clear immediate fields. |
| val &= ~dst_mask; |
| elfcpp::Swap<valsize, big_endian>::writeval(wv, |
| static_cast<Valtype>(val | (immed << doffset))); |
| } |
| |
| // Update two parts of an instruction (pointed by view) with selected |
| // bits (immed1 and immed2). |
| // val = (val & ~dst_mask) | (immed1 << doffset1) | (immed2 << doffset2) |
| |
| template<int valsize> |
| static inline void |
| update_view_two_parts( |
| unsigned char* view, |
| AArch64_valtype immed1, |
| AArch64_valtype immed2, |
| elfcpp::Elf_Xword doffset1, |
| elfcpp::Elf_Xword doffset2, |
| elfcpp::Elf_Xword dst_mask) |
| { |
| typedef typename elfcpp::Swap<valsize, big_endian>::Valtype Valtype; |
| Valtype* wv = reinterpret_cast<Valtype*>(view); |
| Valtype val = elfcpp::Swap<valsize, big_endian>::readval(wv); |
| val &= ~dst_mask; |
| elfcpp::Swap<valsize, big_endian>::writeval(wv, |
| static_cast<Valtype>(val | (immed1 << doffset1) | |
| (immed2 << doffset2))); |
| } |
| |
| // Update adr or adrp instruction with immed. |
| // In adr and adrp: [30:29] immlo [23:5] immhi |
| |
| static inline void |
| update_adr(unsigned char* view, AArch64_valtype immed) |
| { |
| elfcpp::Elf_Xword dst_mask = (0x3 << 29) | (0x7ffff << 5); |
| This::template update_view_two_parts<32>( |
| view, |
| immed & 0x3, |
| (immed & 0x1ffffc) >> 2, |
| 29, |
| 5, |
| dst_mask); |
| } |
| |
| // Update movz/movn instruction with bits immed. |
| // Set instruction to movz if is_movz is true, otherwise set instruction |
| // to movn. |
| |
| static inline void |
| update_movnz(unsigned char* view, |
| AArch64_valtype immed, |
| bool is_movz) |
| { |
| typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype; |
| Valtype* wv = reinterpret_cast<Valtype*>(view); |
| Valtype val = elfcpp::Swap<32, big_endian>::readval(wv); |
| |
| const elfcpp::Elf_Xword doffset = |
| aarch64_howto[AArch64_reloc_property::INST_MOVW].doffset; |
| const elfcpp::Elf_Xword dst_mask = |
| aarch64_howto[AArch64_reloc_property::INST_MOVW].dst_mask; |
| |
| // Clear immediate fields and opc code. |
| val &= ~(dst_mask | (0x3 << 29)); |
| |
| // Set instruction to movz or movn. |
| // movz: [30:29] is 10 movn: [30:29] is 00 |
| if (is_movz) |
| val |= (0x2 << 29); |
| |
| elfcpp::Swap<32, big_endian>::writeval(wv, |
| static_cast<Valtype>(val | (immed << doffset))); |
| } |
| |
| public: |
| |
| // Update selected bits in text. |
| |
| template<int valsize> |
| static inline typename This::Status |
| reloc_common(unsigned char* view, Address x, |
| const AArch64_reloc_property* reloc_property) |
| { |
| // Select bits from X. |
| Address immed = reloc_property->select_x_value(x); |
| |
| // Update view. |
| const AArch64_reloc_property::Reloc_inst inst = |
| reloc_property->reloc_inst(); |
| // If it is a data relocation or instruction has 2 parts of immediate |
| // fields, you should not call pcrela_general. |
| gold_assert(aarch64_howto[inst].doffset2 == -1 && |
| aarch64_howto[inst].doffset != -1); |
| This::template update_view<valsize>(view, immed, |
| aarch64_howto[inst].doffset, |
| aarch64_howto[inst].dst_mask); |
| |
| // Do check overflow or alignment if needed. |
| return (reloc_property->checkup_x_value(x) |
| ? This::STATUS_OKAY |
| : This::STATUS_OVERFLOW); |
| } |
| |
| // Construct a B insn. Note, although we group it here with other relocation |
| // operation, there is actually no 'relocation' involved here. |
| static inline void |
| construct_b(unsigned char* view, unsigned int branch_offset) |
| { |
| update_view_two_parts<32>(view, 0x05, (branch_offset >> 2), |
| 26, 0, 0xffffffff); |
| } |
| |
| // Do a simple rela relocation at unaligned addresses. |
| |
| template<int valsize> |
| static inline typename This::Status |
| rela_ua(unsigned char* view, |
| const Sized_relobj_file<size, big_endian>* object, |
| const Symbol_value<size>* psymval, |
| AArch64_valtype addend, |
| const AArch64_reloc_property* reloc_property) |
| { |
| typedef typename elfcpp::Swap_unaligned<valsize, big_endian>::Valtype |
| Valtype; |
| typename elfcpp::Elf_types<size>::Elf_Addr x = |
| psymval->value(object, addend); |
| elfcpp::Swap_unaligned<valsize, big_endian>::writeval(view, |
| static_cast<Valtype>(x)); |
| return (reloc_property->checkup_x_value(x) |
| ? This::STATUS_OKAY |
| : This::STATUS_OVERFLOW); |
| } |
| |
| // Do a simple pc-relative relocation at unaligned addresses. |
| |
| template<int valsize> |
| static inline typename This::Status |
| pcrela_ua(unsigned char* view, |
| const Sized_relobj_file<size, big_endian>* object, |
| const Symbol_value<size>* psymval, |
| AArch64_valtype addend, |
| Address address, |
| const AArch64_reloc_property* reloc_property) |
| { |
| typedef typename elfcpp::Swap_unaligned<valsize, big_endian>::Valtype |
| Valtype; |
| Address x = psymval->value(object, addend) - address; |
| elfcpp::Swap_unaligned<valsize, big_endian>::writeval(view, |
| static_cast<Valtype>(x)); |
| return (reloc_property->checkup_x_value(x) |
| ? This::STATUS_OKAY |
| : This::STATUS_OVERFLOW); |
| } |
| |
| // Do a simple rela relocation at aligned addresses. |
| |
| template<int valsize> |
| static inline typename This::Status |
| rela( |
| unsigned char* view, |
| const Sized_relobj_file<size, big_endian>* object, |
| const Symbol_value<size>* psymval, |
| AArch64_valtype addend, |
| const AArch64_reloc_property* reloc_property) |
| { |
| typedef typename elfcpp::Swap<valsize, big_endian>::Valtype Valtype; |
| Valtype* wv = reinterpret_cast<Valtype*>(view); |
| Address x = psymval->value(object, addend); |
| elfcpp::Swap<valsize, big_endian>::writeval(wv,static_cast<Valtype>(x)); |
| return (reloc_property->checkup_x_value(x) |
| ? This::STATUS_OKAY |
| : This::STATUS_OVERFLOW); |
| } |
| |
| // Do relocate. Update selected bits in text. |
| // new_val = (val & ~dst_mask) | (immed << doffset) |
| |
| template<int valsize> |
| static inline typename This::Status |
| rela_general(unsigned char* view, |
| const Sized_relobj_file<size, big_endian>* object, |
| const Symbol_value<size>* psymval, |
| AArch64_valtype addend, |
| const AArch64_reloc_property* reloc_property) |
| { |
| // Calculate relocation. |
| Address x = psymval->value(object, addend); |
| return This::template reloc_common<valsize>(view, x, reloc_property); |
| } |
| |
| // Do relocate. Update selected bits in text. |
| // new val = (val & ~dst_mask) | (immed << doffset) |
| |
| template<int valsize> |
| static inline typename This::Status |
| rela_general( |
| unsigned char* view, |
| AArch64_valtype s, |
| AArch64_valtype addend, |
| const AArch64_reloc_property* reloc_property) |
| { |
| // Calculate relocation. |
| Address x = s + addend; |
| return This::template reloc_common<valsize>(view, x, reloc_property); |
| } |
| |
| // Do address relative relocate. Update selected bits in text. |
| // new val = (val & ~dst_mask) | (immed << doffset) |
| |
| template<int valsize> |
| static inline typename This::Status |
| pcrela_general( |
| unsigned char* view, |
| const Sized_relobj_file<size, big_endian>* object, |
| const Symbol_value<size>* psymval, |
| AArch64_valtype addend, |
| Address address, |
| const AArch64_reloc_property* reloc_property) |
| { |
| // Calculate relocation. |
| Address x = psymval->value(object, addend) - address; |
| return This::template reloc_common<valsize>(view, x, reloc_property); |
| } |
| |
| |
| // Calculate (S + A) - address, update adr instruction. |
| |
| static inline typename This::Status |
| adr(unsigned char* view, |
| const Sized_relobj_file<size, big_endian>* object, |
| const Symbol_value<size>* psymval, |
| Address addend, |
| Address address, |
| const AArch64_reloc_property* /* reloc_property */) |
| { |
| AArch64_valtype x = psymval->value(object, addend) - address; |
| // Pick bits [20:0] of X. |
| AArch64_valtype immed = x & 0x1fffff; |
| update_adr(view, immed); |
| // Check -2^20 <= X < 2^20 |
| return (size == 64 && Bits<21>::has_overflow((x)) |
| ? This::STATUS_OVERFLOW |
| : This::STATUS_OKAY); |
| } |
| |
| // Calculate PG(S+A) - PG(address), update adrp instruction. |
| // R_AARCH64_ADR_PREL_PG_HI21 |
| |
| static inline typename This::Status |
| adrp( |
| unsigned char* view, |
| Address sa, |
| Address address) |
| { |
| AArch64_valtype x = This::Page(sa) - This::Page(address); |
| // Pick [32:12] of X. |
| AArch64_valtype immed = (x >> 12) & 0x1fffff; |
| update_adr(view, immed); |
| // Check -2^32 <= X < 2^32 |
| return (size == 64 && Bits<33>::has_overflow((x)) |
| ? This::STATUS_OVERFLOW |
| : This::STATUS_OKAY); |
| } |
| |
| // Calculate PG(S+A) - PG(address), update adrp instruction. |
| // R_AARCH64_ADR_PREL_PG_HI21 |
| |
| static inline typename This::Status |
| adrp(unsigned char* view, |
| const Sized_relobj_file<size, big_endian>* object, |
| const Symbol_value<size>* psymval, |
| Address addend, |
| Address address, |
| const AArch64_reloc_property* reloc_property) |
| { |
| Address sa = psymval->value(object, addend); |
| AArch64_valtype x = This::Page(sa) - This::Page(address); |
| // Pick [32:12] of X. |
| AArch64_valtype immed = (x >> 12) & 0x1fffff; |
| update_adr(view, immed); |
| return (reloc_property->checkup_x_value(x) |
| ? This::STATUS_OKAY |
| : This::STATUS_OVERFLOW); |
| } |
| |
| // Update mov[n/z] instruction. Check overflow if needed. |
| // If X >=0, set the instruction to movz and its immediate value to the |
| // selected bits S. |
| // If X < 0, set the instruction to movn and its immediate value to |
| // NOT (selected bits of). |
| |
| static inline typename This::Status |
| movnz(unsigned char* view, |
| AArch64_valtype x, |
| const AArch64_reloc_property* reloc_property) |
| { |
| // Select bits from X. |
| Address immed; |
| bool is_movz; |
| typedef typename elfcpp::Elf_types<size>::Elf_Swxword SignedW; |
| if (static_cast<SignedW>(x) >= 0) |
| { |
| immed = reloc_property->select_x_value(x); |
| is_movz = true; |
| } |
| else |
| { |
| immed = reloc_property->select_x_value(~x);; |
| is_movz = false; |
| } |
| |
| // Update movnz instruction. |
| update_movnz(view, immed, is_movz); |
| |
| // Do check overflow or alignment if needed. |
| return (reloc_property->checkup_x_value(x) |
| ? This::STATUS_OKAY |
| : This::STATUS_OVERFLOW); |
| } |
| |
| static inline bool |
| maybe_apply_stub(unsigned int, |
| const The_relocate_info*, |
| const The_rela&, |
| unsigned char*, |
| Address, |
| const Sized_symbol<size>*, |
| const Symbol_value<size>*, |
| const Sized_relobj_file<size, big_endian>*, |
| section_size_type); |
| |
| }; // End of AArch64_relocate_functions |
| |
| |
| // For a certain relocation type (usually jump/branch), test to see if the |
| // destination needs a stub to fulfil. If so, re-route the destination of the |
| // original instruction to the stub, note, at this time, the stub has already |
| // been generated. |
| |
| template<int size, bool big_endian> |
| bool |
| AArch64_relocate_functions<size, big_endian>:: |
| maybe_apply_stub(unsigned int r_type, |
| const The_relocate_info* relinfo, |
| const The_rela& rela, |
| unsigned char* view, |
| Address address, |
| const Sized_symbol<size>* gsym, |
| const Symbol_value<size>* psymval, |
| const Sized_relobj_file<size, big_endian>* object, |
| section_size_type current_group_size) |
| { |
| if (parameters->options().relocatable()) |
| return false; |
| |
| typename elfcpp::Elf_types<size>::Elf_Swxword addend = rela.get_r_addend(); |
| Address branch_target = psymval->value(object, 0) + addend; |
| int stub_type = |
| The_reloc_stub::stub_type_for_reloc(r_type, address, branch_target); |
| if (stub_type == ST_NONE) |
| return false; |
| |
| const The_aarch64_relobj* aarch64_relobj = |
| static_cast<const The_aarch64_relobj*>(object); |
| const AArch64_reloc_property* arp = |
| aarch64_reloc_property_table->get_reloc_property(r_type); |
| gold_assert(arp != NULL); |
| |
| // We don't create stubs for undefined symbols, but do for weak. |
| if (gsym |
| && !gsym->use_plt_offset(arp->reference_flags()) |
| && gsym->is_undefined()) |
| { |
| gold_debug(DEBUG_TARGET, |
| "stub: looking for a stub for undefined symbol %s in file %s", |
| gsym->name(), aarch64_relobj->name().c_str()); |
| return false; |
| } |
| |
| The_stub_table* stub_table = aarch64_relobj->stub_table(relinfo->data_shndx); |
| gold_assert(stub_table != NULL); |
| |
| unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info()); |
| typename The_reloc_stub::Key stub_key(stub_type, gsym, object, r_sym, addend); |
| The_reloc_stub* stub = stub_table->find_reloc_stub(stub_key); |
| gold_assert(stub != NULL); |
| |
| Address new_branch_target = stub_table->address() + stub->offset(); |
| typename elfcpp::Swap<size, big_endian>::Valtype branch_offset = |
| new_branch_target - address; |
| typename This::Status status = This::template |
| rela_general<32>(view, branch_offset, 0, arp); |
| if (status != This::STATUS_OKAY) |
| gold_error(_("Stub is too far away, try a smaller value " |
| "for '--stub-group-size'. The current value is 0x%lx."), |
| static_cast<unsigned long>(current_group_size)); |
| return true; |
| } |
| |
| |
| // Group input sections for stub generation. |
| // |
| // We group input sections in an output section so that the total size, |
| // including any padding space due to alignment is smaller than GROUP_SIZE |
| // unless the only input section in group is bigger than GROUP_SIZE already. |
| // Then an ARM stub table is created to follow the last input section |
| // in group. For each group an ARM stub table is created an is placed |
| // after the last group. If STUB_ALWAYS_AFTER_BRANCH is false, we further |
| // extend the group after the stub table. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::group_sections( |
| Layout* layout, |
| section_size_type group_size, |
| bool stubs_always_after_branch, |
| const Task* task) |
| { |
| // Group input sections and insert stub table |
| Layout::Section_list section_list; |
| layout->get_executable_sections(§ion_list); |
| for (Layout::Section_list::const_iterator p = section_list.begin(); |
| p != section_list.end(); |
| ++p) |
| { |
| AArch64_output_section<size, big_endian>* output_section = |
| static_cast<AArch64_output_section<size, big_endian>*>(*p); |
| output_section->group_sections(group_size, stubs_always_after_branch, |
| this, task); |
| } |
| } |
| |
| |
| // Find the AArch64_input_section object corresponding to the SHNDX-th input |
| // section of RELOBJ. |
| |
| template<int size, bool big_endian> |
| AArch64_input_section<size, big_endian>* |
| Target_aarch64<size, big_endian>::find_aarch64_input_section( |
| Relobj* relobj, unsigned int shndx) const |
| { |
| Section_id sid(relobj, shndx); |
| typename AArch64_input_section_map::const_iterator p = |
| this->aarch64_input_section_map_.find(sid); |
| return (p != this->aarch64_input_section_map_.end()) ? p->second : NULL; |
| } |
| |
| |
| // Make a new AArch64_input_section object. |
| |
| template<int size, bool big_endian> |
| AArch64_input_section<size, big_endian>* |
| Target_aarch64<size, big_endian>::new_aarch64_input_section( |
| Relobj* relobj, unsigned int shndx) |
| { |
| Section_id sid(relobj, shndx); |
| |
| AArch64_input_section<size, big_endian>* input_section = |
| new AArch64_input_section<size, big_endian>(relobj, shndx); |
| input_section->init(); |
| |
| // Register new AArch64_input_section in map for look-up. |
| std::pair<typename AArch64_input_section_map::iterator,bool> ins = |
| this->aarch64_input_section_map_.insert( |
| std::make_pair(sid, input_section)); |
| |
| // Make sure that it we have not created another AArch64_input_section |
| // for this input section already. |
| gold_assert(ins.second); |
| |
| return input_section; |
| } |
| |
| |
| // Relaxation hook. This is where we do stub generation. |
| |
| template<int size, bool big_endian> |
| bool |
| Target_aarch64<size, big_endian>::do_relax( |
| int pass, |
| const Input_objects* input_objects, |
| Symbol_table* symtab, |
| Layout* layout , |
| const Task* task) |
| { |
| gold_assert(!parameters->options().relocatable()); |
| if (pass == 1) |
| { |
| // We don't handle negative stub_group_size right now. |
| this->stub_group_size_ = abs(parameters->options().stub_group_size()); |
| if (this->stub_group_size_ == 1) |
| { |
| // Leave room for 4096 4-byte stub entries. If we exceed that, then we |
| // will fail to link. The user will have to relink with an explicit |
| // group size option. |
| this->stub_group_size_ = The_reloc_stub::MAX_BRANCH_OFFSET - |
| 4096 * 4; |
| } |
| group_sections(layout, this->stub_group_size_, true, task); |
| } |
| else |
| { |
| // If this is not the first pass, addresses and file offsets have |
| // been reset at this point, set them here. |
| for (Stub_table_iterator sp = this->stub_tables_.begin(); |
| sp != this->stub_tables_.end(); ++sp) |
| { |
| The_stub_table* stt = *sp; |
| The_aarch64_input_section* owner = stt->owner(); |
| off_t off = align_address(owner->original_size(), |
| stt->addralign()); |
| stt->set_address_and_file_offset(owner->address() + off, |
| owner->offset() + off); |
| } |
| } |
| |
| // Scan relocs for relocation stubs |
| for (Input_objects::Relobj_iterator op = input_objects->relobj_begin(); |
| op != input_objects->relobj_end(); |
| ++op) |
| { |
| The_aarch64_relobj* aarch64_relobj = |
| static_cast<The_aarch64_relobj*>(*op); |
| // Lock the object so we can read from it. This is only called |
| // single-threaded from Layout::finalize, so it is OK to lock. |
| Task_lock_obj<Object> tl(task, aarch64_relobj); |
| aarch64_relobj->scan_sections_for_stubs(this, symtab, layout); |
| } |
| |
| bool any_stub_table_changed = false; |
| for (Stub_table_iterator siter = this->stub_tables_.begin(); |
| siter != this->stub_tables_.end() && !any_stub_table_changed; ++siter) |
| { |
| The_stub_table* stub_table = *siter; |
| if (stub_table->update_data_size_changed_p()) |
| { |
| The_aarch64_input_section* owner = stub_table->owner(); |
| uint64_t address = owner->address(); |
| off_t offset = owner->offset(); |
| owner->reset_address_and_file_offset(); |
| owner->set_address_and_file_offset(address, offset); |
| |
| any_stub_table_changed = true; |
| } |
| } |
| |
| // Do not continue relaxation. |
| bool continue_relaxation = any_stub_table_changed; |
| if (!continue_relaxation) |
| for (Stub_table_iterator sp = this->stub_tables_.begin(); |
| (sp != this->stub_tables_.end()); |
| ++sp) |
| (*sp)->finalize_stubs(); |
| |
| return continue_relaxation; |
| } |
| |
| |
| // Make a new Stub_table. |
| |
| template<int size, bool big_endian> |
| Stub_table<size, big_endian>* |
| Target_aarch64<size, big_endian>::new_stub_table( |
| AArch64_input_section<size, big_endian>* owner) |
| { |
| Stub_table<size, big_endian>* stub_table = |
| new Stub_table<size, big_endian>(owner); |
| stub_table->set_address(align_address( |
| owner->address() + owner->data_size(), 8)); |
| stub_table->set_file_offset(owner->offset() + owner->data_size()); |
| stub_table->finalize_data_size(); |
| |
| this->stub_tables_.push_back(stub_table); |
| |
| return stub_table; |
| } |
| |
| |
| template<int size, bool big_endian> |
| uint64_t |
| Target_aarch64<size, big_endian>::do_reloc_addend( |
| void* arg, unsigned int r_type, uint64_t) const |
| { |
| gold_assert(r_type == elfcpp::R_AARCH64_TLSDESC); |
| uintptr_t intarg = reinterpret_cast<uintptr_t>(arg); |
| gold_assert(intarg < this->tlsdesc_reloc_info_.size()); |
| const Tlsdesc_info& ti(this->tlsdesc_reloc_info_[intarg]); |
| const Symbol_value<size>* psymval = ti.object->local_symbol(ti.r_sym); |
| gold_assert(psymval->is_tls_symbol()); |
| // The value of a TLS symbol is the offset in the TLS segment. |
| return psymval->value(ti.object, 0); |
| } |
| |
| // Return the number of entries in the PLT. |
| |
| template<int size, bool big_endian> |
| unsigned int |
| Target_aarch64<size, big_endian>::plt_entry_count() const |
| { |
| if (this->plt_ == NULL) |
| return 0; |
| return this->plt_->entry_count(); |
| } |
| |
| // Return the offset of the first non-reserved PLT entry. |
| |
| template<int size, bool big_endian> |
| unsigned int |
| Target_aarch64<size, big_endian>::first_plt_entry_offset() const |
| { |
| return this->plt_->first_plt_entry_offset(); |
| } |
| |
| // Return the size of each PLT entry. |
| |
| template<int size, bool big_endian> |
| unsigned int |
| Target_aarch64<size, big_endian>::plt_entry_size() const |
| { |
| return this->plt_->get_plt_entry_size(); |
| } |
| |
| // Define the _TLS_MODULE_BASE_ symbol in the TLS segment. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::define_tls_base_symbol( |
| Symbol_table* symtab, Layout* layout) |
| { |
| if (this->tls_base_symbol_defined_) |
| return; |
| |
| Output_segment* tls_segment = layout->tls_segment(); |
| if (tls_segment != NULL) |
| { |
| // _TLS_MODULE_BASE_ always points to the beginning of tls segment. |
| symtab->define_in_output_segment("_TLS_MODULE_BASE_", NULL, |
| Symbol_table::PREDEFINED, |
| tls_segment, 0, 0, |
| elfcpp::STT_TLS, |
| elfcpp::STB_LOCAL, |
| elfcpp::STV_HIDDEN, 0, |
| Symbol::SEGMENT_START, |
| true); |
| } |
| this->tls_base_symbol_defined_ = true; |
| } |
| |
| // Create the reserved PLT and GOT entries for the TLS descriptor resolver. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::reserve_tlsdesc_entries( |
| Symbol_table* symtab, Layout* layout) |
| { |
| if (this->plt_ == NULL) |
| this->make_plt_section(symtab, layout); |
| |
| if (!this->plt_->has_tlsdesc_entry()) |
| { |
| // Allocate the TLSDESC_GOT entry. |
| Output_data_got_aarch64<size, big_endian>* got = |
| this->got_section(symtab, layout); |
| unsigned int got_offset = got->add_constant(0); |
| |
| // Allocate the TLSDESC_PLT entry. |
| this->plt_->reserve_tlsdesc_entry(got_offset); |
| } |
| } |
| |
| // Create a GOT entry for the TLS module index. |
| |
| template<int size, bool big_endian> |
| unsigned int |
| Target_aarch64<size, big_endian>::got_mod_index_entry( |
| Symbol_table* symtab, Layout* layout, |
| Sized_relobj_file<size, big_endian>* object) |
| { |
| if (this->got_mod_index_offset_ == -1U) |
| { |
| gold_assert(symtab != NULL && layout != NULL && object != NULL); |
| Reloc_section* rela_dyn = this->rela_dyn_section(layout); |
| Output_data_got_aarch64<size, big_endian>* got = |
| this->got_section(symtab, layout); |
| unsigned int got_offset = got->add_constant(0); |
| rela_dyn->add_local(object, 0, elfcpp::R_AARCH64_TLS_DTPMOD64, got, |
| got_offset, 0); |
| got->add_constant(0); |
| this->got_mod_index_offset_ = got_offset; |
| } |
| return this->got_mod_index_offset_; |
| } |
| |
| // Optimize the TLS relocation type based on what we know about the |
| // symbol. IS_FINAL is true if the final address of this symbol is |
| // known at link time. |
| |
| template<int size, bool big_endian> |
| tls::Tls_optimization |
| Target_aarch64<size, big_endian>::optimize_tls_reloc(bool is_final, |
| int r_type) |
| { |
| // If we are generating a shared library, then we can't do anything |
| // in the linker |
| if (parameters->options().shared()) |
| return tls::TLSOPT_NONE; |
| |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_TLSGD_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC: |
| case elfcpp::R_AARCH64_TLSDESC_LD_PREL19: |
| case elfcpp::R_AARCH64_TLSDESC_ADR_PREL21: |
| case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSDESC_LD64_LO12: |
| case elfcpp::R_AARCH64_TLSDESC_ADD_LO12: |
| case elfcpp::R_AARCH64_TLSDESC_OFF_G1: |
| case elfcpp::R_AARCH64_TLSDESC_OFF_G0_NC: |
| case elfcpp::R_AARCH64_TLSDESC_LDR: |
| case elfcpp::R_AARCH64_TLSDESC_ADD: |
| case elfcpp::R_AARCH64_TLSDESC_CALL: |
| // These are General-Dynamic which permits fully general TLS |
| // access. Since we know that we are generating an executable, |
| // we can convert this to Initial-Exec. If we also know that |
| // this is a local symbol, we can further switch to Local-Exec. |
| if (is_final) |
| return tls::TLSOPT_TO_LE; |
| return tls::TLSOPT_TO_IE; |
| |
| case elfcpp::R_AARCH64_TLSLD_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G1: |
| case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G0_NC: |
| case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_HI12: |
| case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_LO12_NC: |
| // These are Local-Dynamic, which refer to local symbols in the |
| // dynamic TLS block. Since we know that we generating an |
| // executable, we can switch to Local-Exec. |
| return tls::TLSOPT_TO_LE; |
| |
| case elfcpp::R_AARCH64_TLSIE_MOVW_GOTTPREL_G1: |
| case elfcpp::R_AARCH64_TLSIE_MOVW_GOTTPREL_G0_NC: |
| case elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21: |
| case elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSIE_LD_GOTTPREL_PREL19: |
| // These are Initial-Exec relocs which get the thread offset |
| // from the GOT. If we know that we are linking against the |
| // local symbol, we can switch to Local-Exec, which links the |
| // thread offset into the instruction. |
| if (is_final) |
| return tls::TLSOPT_TO_LE; |
| return tls::TLSOPT_NONE; |
| |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G2: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1_NC: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0_NC: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12_NC: |
| // When we already have Local-Exec, there is nothing further we |
| // can do. |
| return tls::TLSOPT_NONE; |
| |
| default: |
| gold_unreachable(); |
| } |
| } |
| |
| // Returns true if this relocation type could be that of a function pointer. |
| |
| template<int size, bool big_endian> |
| inline bool |
| Target_aarch64<size, big_endian>::Scan::possible_function_pointer_reloc( |
| unsigned int r_type) |
| { |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_ADR_PREL_PG_HI21: |
| case elfcpp::R_AARCH64_ADR_PREL_PG_HI21_NC: |
| case elfcpp::R_AARCH64_ADD_ABS_LO12_NC: |
| case elfcpp::R_AARCH64_ADR_GOT_PAGE: |
| case elfcpp::R_AARCH64_LD64_GOT_LO12_NC: |
| { |
| return true; |
| } |
| } |
| return false; |
| } |
| |
| // For safe ICF, scan a relocation for a local symbol to check if it |
| // corresponds to a function pointer being taken. In that case mark |
| // the function whose pointer was taken as not foldable. |
| |
| template<int size, bool big_endian> |
| inline bool |
| Target_aarch64<size, big_endian>::Scan::local_reloc_may_be_function_pointer( |
| Symbol_table* , |
| Layout* , |
| Target_aarch64<size, big_endian>* , |
| Sized_relobj_file<size, big_endian>* , |
| unsigned int , |
| Output_section* , |
| const elfcpp::Rela<size, big_endian>& , |
| unsigned int r_type, |
| const elfcpp::Sym<size, big_endian>&) |
| { |
| // When building a shared library, do not fold any local symbols. |
| return (parameters->options().shared() |
| || possible_function_pointer_reloc(r_type)); |
| } |
| |
| // For safe ICF, scan a relocation for a global symbol to check if it |
| // corresponds to a function pointer being taken. In that case mark |
| // the function whose pointer was taken as not foldable. |
| |
| template<int size, bool big_endian> |
| inline bool |
| Target_aarch64<size, big_endian>::Scan::global_reloc_may_be_function_pointer( |
| Symbol_table* , |
| Layout* , |
| Target_aarch64<size, big_endian>* , |
| Sized_relobj_file<size, big_endian>* , |
| unsigned int , |
| Output_section* , |
| const elfcpp::Rela<size, big_endian>& , |
| unsigned int r_type, |
| Symbol* gsym) |
| { |
| // When building a shared library, do not fold symbols whose visibility |
| // is hidden, internal or protected. |
| return ((parameters->options().shared() |
| && (gsym->visibility() == elfcpp::STV_INTERNAL |
| || gsym->visibility() == elfcpp::STV_PROTECTED |
| || gsym->visibility() == elfcpp::STV_HIDDEN)) |
| || possible_function_pointer_reloc(r_type)); |
| } |
| |
| // Report an unsupported relocation against a local symbol. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::Scan::unsupported_reloc_local( |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int r_type) |
| { |
| gold_error(_("%s: unsupported reloc %u against local symbol"), |
| object->name().c_str(), r_type); |
| } |
| |
| // We are about to emit a dynamic relocation of type R_TYPE. If the |
| // dynamic linker does not support it, issue an error. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::Scan::check_non_pic(Relobj* object, |
| unsigned int r_type) |
| { |
| gold_assert(r_type != elfcpp::R_AARCH64_NONE); |
| |
| switch (r_type) |
| { |
| // These are the relocation types supported by glibc for AARCH64. |
| case elfcpp::R_AARCH64_NONE: |
| case elfcpp::R_AARCH64_COPY: |
| case elfcpp::R_AARCH64_GLOB_DAT: |
| case elfcpp::R_AARCH64_JUMP_SLOT: |
| case elfcpp::R_AARCH64_RELATIVE: |
| case elfcpp::R_AARCH64_TLS_DTPREL64: |
| case elfcpp::R_AARCH64_TLS_DTPMOD64: |
| case elfcpp::R_AARCH64_TLS_TPREL64: |
| case elfcpp::R_AARCH64_TLSDESC: |
| case elfcpp::R_AARCH64_IRELATIVE: |
| case elfcpp::R_AARCH64_ABS32: |
| case elfcpp::R_AARCH64_ABS64: |
| return; |
| |
| default: |
| break; |
| } |
| |
| // This prevents us from issuing more than one error per reloc |
| // section. But we can still wind up issuing more than one |
| // error per object file. |
| if (this->issued_non_pic_error_) |
| return; |
| gold_assert(parameters->options().output_is_position_independent()); |
| object->error(_("requires unsupported dynamic reloc; " |
| "recompile with -fPIC")); |
| this->issued_non_pic_error_ = true; |
| return; |
| } |
| |
| // Return whether we need to make a PLT entry for a relocation of the |
| // given type against a STT_GNU_IFUNC symbol. |
| |
| template<int size, bool big_endian> |
| bool |
| Target_aarch64<size, big_endian>::Scan::reloc_needs_plt_for_ifunc( |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int r_type) |
| { |
| const AArch64_reloc_property* arp = |
| aarch64_reloc_property_table->get_reloc_property(r_type); |
| gold_assert(arp != NULL); |
| |
| int flags = arp->reference_flags(); |
| if (flags & Symbol::TLS_REF) |
| { |
| gold_error(_("%s: unsupported TLS reloc %s for IFUNC symbol"), |
| object->name().c_str(), arp->name().c_str()); |
| return false; |
| } |
| return flags != 0; |
| } |
| |
| // Scan a relocation for a local symbol. |
| |
| template<int size, bool big_endian> |
| inline void |
| Target_aarch64<size, big_endian>::Scan::local( |
| Symbol_table* symtab, |
| Layout* layout, |
| Target_aarch64<size, big_endian>* target, |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int data_shndx, |
| Output_section* output_section, |
| const elfcpp::Rela<size, big_endian>& rela, |
| unsigned int r_type, |
| const elfcpp::Sym<size, big_endian>& lsym, |
| bool is_discarded) |
| { |
| if (is_discarded) |
| return; |
| |
| typedef Output_data_reloc<elfcpp::SHT_RELA, true, size, big_endian> |
| Reloc_section; |
| unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info()); |
| |
| // A local STT_GNU_IFUNC symbol may require a PLT entry. |
| bool is_ifunc = lsym.get_st_type() == elfcpp::STT_GNU_IFUNC; |
| if (is_ifunc && this->reloc_needs_plt_for_ifunc(object, r_type)) |
| target->make_local_ifunc_plt_entry(symtab, layout, object, r_sym); |
| |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_NONE: |
| break; |
| |
| case elfcpp::R_AARCH64_ABS32: |
| case elfcpp::R_AARCH64_ABS16: |
| if (parameters->options().output_is_position_independent()) |
| { |
| gold_error(_("%s: unsupported reloc %u in pos independent link."), |
| object->name().c_str(), r_type); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_ABS64: |
| // If building a shared library or pie, we need to mark this as a dynmic |
| // reloction, so that the dynamic loader can relocate it. |
| if (parameters->options().output_is_position_independent()) |
| { |
| Reloc_section* rela_dyn = target->rela_dyn_section(layout); |
| rela_dyn->add_local_relative(object, r_sym, |
| elfcpp::R_AARCH64_RELATIVE, |
| output_section, |
| data_shndx, |
| rela.get_r_offset(), |
| rela.get_r_addend(), |
| is_ifunc); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_PREL64: |
| case elfcpp::R_AARCH64_PREL32: |
| case elfcpp::R_AARCH64_PREL16: |
| break; |
| |
| case elfcpp::R_AARCH64_ADR_GOT_PAGE: |
| case elfcpp::R_AARCH64_LD64_GOT_LO12_NC: |
| case elfcpp::R_AARCH64_LD64_GOTPAGE_LO15: |
| // The above relocations are used to access GOT entries. |
| { |
| Output_data_got_aarch64<size, big_endian>* got = |
| target->got_section(symtab, layout); |
| bool is_new = false; |
| // This symbol requires a GOT entry. |
| if (is_ifunc) |
| is_new = got->add_local_plt(object, r_sym, GOT_TYPE_STANDARD); |
| else |
| is_new = got->add_local(object, r_sym, GOT_TYPE_STANDARD); |
| if (is_new && parameters->options().output_is_position_independent()) |
| target->rela_dyn_section(layout)-> |
| add_local_relative(object, |
| r_sym, |
| elfcpp::R_AARCH64_RELATIVE, |
| got, |
| object->local_got_offset(r_sym, |
| GOT_TYPE_STANDARD), |
| 0, |
| false); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_MOVW_UABS_G0: // 263 |
| case elfcpp::R_AARCH64_MOVW_UABS_G0_NC: // 264 |
| case elfcpp::R_AARCH64_MOVW_UABS_G1: // 265 |
| case elfcpp::R_AARCH64_MOVW_UABS_G1_NC: // 266 |
| case elfcpp::R_AARCH64_MOVW_UABS_G2: // 267 |
| case elfcpp::R_AARCH64_MOVW_UABS_G2_NC: // 268 |
| case elfcpp::R_AARCH64_MOVW_UABS_G3: // 269 |
| case elfcpp::R_AARCH64_MOVW_SABS_G0: // 270 |
| case elfcpp::R_AARCH64_MOVW_SABS_G1: // 271 |
| case elfcpp::R_AARCH64_MOVW_SABS_G2: // 272 |
| if (parameters->options().output_is_position_independent()) |
| { |
| gold_error(_("%s: unsupported reloc %u in pos independent link."), |
| object->name().c_str(), r_type); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_LD_PREL_LO19: // 273 |
| case elfcpp::R_AARCH64_ADR_PREL_LO21: // 274 |
| case elfcpp::R_AARCH64_ADR_PREL_PG_HI21: // 275 |
| case elfcpp::R_AARCH64_ADR_PREL_PG_HI21_NC: // 276 |
| case elfcpp::R_AARCH64_ADD_ABS_LO12_NC: // 277 |
| case elfcpp::R_AARCH64_LDST8_ABS_LO12_NC: // 278 |
| case elfcpp::R_AARCH64_LDST16_ABS_LO12_NC: // 284 |
| case elfcpp::R_AARCH64_LDST32_ABS_LO12_NC: // 285 |
| case elfcpp::R_AARCH64_LDST64_ABS_LO12_NC: // 286 |
| case elfcpp::R_AARCH64_LDST128_ABS_LO12_NC: // 299 |
| break; |
| |
| // Control flow, pc-relative. We don't need to do anything for a relative |
| // addressing relocation against a local symbol if it does not reference |
| // the GOT. |
| case elfcpp::R_AARCH64_TSTBR14: |
| case elfcpp::R_AARCH64_CONDBR19: |
| case elfcpp::R_AARCH64_JUMP26: |
| case elfcpp::R_AARCH64_CALL26: |
| break; |
| |
| case elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21: |
| case elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC: |
| { |
| tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>:: |
| optimize_tls_reloc(!parameters->options().shared(), r_type); |
| if (tlsopt == tls::TLSOPT_TO_LE) |
| break; |
| |
| layout->set_has_static_tls(); |
| // Create a GOT entry for the tp-relative offset. |
| if (!parameters->doing_static_link()) |
| { |
| Output_data_got_aarch64<size, big_endian>* got = |
| target->got_section(symtab, layout); |
| got->add_local_with_rel(object, r_sym, GOT_TYPE_TLS_OFFSET, |
| target->rela_dyn_section(layout), |
| elfcpp::R_AARCH64_TLS_TPREL64); |
| } |
| else if (!object->local_has_got_offset(r_sym, |
| GOT_TYPE_TLS_OFFSET)) |
| { |
| Output_data_got_aarch64<size, big_endian>* got = |
| target->got_section(symtab, layout); |
| got->add_local(object, r_sym, GOT_TYPE_TLS_OFFSET); |
| unsigned int got_offset = |
| object->local_got_offset(r_sym, GOT_TYPE_TLS_OFFSET); |
| const elfcpp::Elf_Xword addend = rela.get_r_addend(); |
| gold_assert(addend == 0); |
| got->add_static_reloc(got_offset, elfcpp::R_AARCH64_TLS_TPREL64, |
| object, r_sym); |
| } |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSGD_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC: |
| { |
| tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>:: |
| optimize_tls_reloc(!parameters->options().shared(), r_type); |
| if (tlsopt == tls::TLSOPT_TO_LE) |
| { |
| layout->set_has_static_tls(); |
| break; |
| } |
| gold_assert(tlsopt == tls::TLSOPT_NONE); |
| |
| Output_data_got_aarch64<size, big_endian>* got = |
| target->got_section(symtab, layout); |
| got->add_local_pair_with_rel(object,r_sym, data_shndx, |
| GOT_TYPE_TLS_PAIR, |
| target->rela_dyn_section(layout), |
| elfcpp::R_AARCH64_TLS_DTPMOD64); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G2: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1_NC: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0_NC: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12_NC: |
| { |
| layout->set_has_static_tls(); |
| bool output_is_shared = parameters->options().shared(); |
| if (output_is_shared) |
| gold_error(_("%s: unsupported TLSLE reloc %u in shared code."), |
| object->name().c_str(), r_type); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSLD_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC: |
| { |
| tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>:: |
| optimize_tls_reloc(!parameters->options().shared(), r_type); |
| if (tlsopt == tls::TLSOPT_NONE) |
| { |
| // Create a GOT entry for the module index. |
| target->got_mod_index_entry(symtab, layout, object); |
| } |
| else if (tlsopt != tls::TLSOPT_TO_LE) |
| unsupported_reloc_local(object, r_type); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G1: |
| case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G0_NC: |
| case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_HI12: |
| case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_LO12_NC: |
| break; |
| |
| case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSDESC_LD64_LO12: |
| case elfcpp::R_AARCH64_TLSDESC_ADD_LO12: |
| { |
| tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>:: |
| optimize_tls_reloc(!parameters->options().shared(), r_type); |
| target->define_tls_base_symbol(symtab, layout); |
| if (tlsopt == tls::TLSOPT_NONE) |
| { |
| // Create reserved PLT and GOT entries for the resolver. |
| target->reserve_tlsdesc_entries(symtab, layout); |
| |
| // Generate a double GOT entry with an R_AARCH64_TLSDESC reloc. |
| // The R_AARCH64_TLSDESC reloc is resolved lazily, so the GOT |
| // entry needs to be in an area in .got.plt, not .got. Call |
| // got_section to make sure the section has been created. |
| target->got_section(symtab, layout); |
| Output_data_got<size, big_endian>* got = |
| target->got_tlsdesc_section(); |
| unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info()); |
| if (!object->local_has_got_offset(r_sym, GOT_TYPE_TLS_DESC)) |
| { |
| unsigned int got_offset = got->add_constant(0); |
| got->add_constant(0); |
| object->set_local_got_offset(r_sym, GOT_TYPE_TLS_DESC, |
| got_offset); |
| Reloc_section* rt = target->rela_tlsdesc_section(layout); |
| // We store the arguments we need in a vector, and use |
| // the index into the vector as the parameter to pass |
| // to the target specific routines. |
| uintptr_t intarg = target->add_tlsdesc_info(object, r_sym); |
| void* arg = reinterpret_cast<void*>(intarg); |
| rt->add_target_specific(elfcpp::R_AARCH64_TLSDESC, arg, |
| got, got_offset, 0); |
| } |
| } |
| else if (tlsopt != tls::TLSOPT_TO_LE) |
| unsupported_reloc_local(object, r_type); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSDESC_CALL: |
| break; |
| |
| default: |
| unsupported_reloc_local(object, r_type); |
| } |
| } |
| |
| |
| // Report an unsupported relocation against a global symbol. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::Scan::unsupported_reloc_global( |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int r_type, |
| Symbol* gsym) |
| { |
| gold_error(_("%s: unsupported reloc %u against global symbol %s"), |
| object->name().c_str(), r_type, gsym->demangled_name().c_str()); |
| } |
| |
| template<int size, bool big_endian> |
| inline void |
| Target_aarch64<size, big_endian>::Scan::global( |
| Symbol_table* symtab, |
| Layout* layout, |
| Target_aarch64<size, big_endian>* target, |
| Sized_relobj_file<size, big_endian> * object, |
| unsigned int data_shndx, |
| Output_section* output_section, |
| const elfcpp::Rela<size, big_endian>& rela, |
| unsigned int r_type, |
| Symbol* gsym) |
| { |
| // A STT_GNU_IFUNC symbol may require a PLT entry. |
| if (gsym->type() == elfcpp::STT_GNU_IFUNC |
| && this->reloc_needs_plt_for_ifunc(object, r_type)) |
| target->make_plt_entry(symtab, layout, gsym); |
| |
| typedef Output_data_reloc<elfcpp::SHT_RELA, true, size, big_endian> |
| Reloc_section; |
| const AArch64_reloc_property* arp = |
| aarch64_reloc_property_table->get_reloc_property(r_type); |
| gold_assert(arp != NULL); |
| |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_NONE: |
| break; |
| |
| case elfcpp::R_AARCH64_ABS16: |
| case elfcpp::R_AARCH64_ABS32: |
| case elfcpp::R_AARCH64_ABS64: |
| { |
| // Make a PLT entry if necessary. |
| if (gsym->needs_plt_entry()) |
| { |
| target->make_plt_entry(symtab, layout, gsym); |
| // Since this is not a PC-relative relocation, we may be |
| // taking the address of a function. In that case we need to |
| // set the entry in the dynamic symbol table to the address of |
| // the PLT entry. |
| if (gsym->is_from_dynobj() && !parameters->options().shared()) |
| gsym->set_needs_dynsym_value(); |
| } |
| // Make a dynamic relocation if necessary. |
| if (gsym->needs_dynamic_reloc(arp->reference_flags())) |
| { |
| if (!parameters->options().output_is_position_independent() |
| && gsym->may_need_copy_reloc()) |
| { |
| target->copy_reloc(symtab, layout, object, |
| data_shndx, output_section, gsym, rela); |
| } |
| else if (r_type == elfcpp::R_AARCH64_ABS64 |
| && gsym->type() == elfcpp::STT_GNU_IFUNC |
| && gsym->can_use_relative_reloc(false) |
| && !gsym->is_from_dynobj() |
| && !gsym->is_undefined() |
| && !gsym->is_preemptible()) |
| { |
| // Use an IRELATIVE reloc for a locally defined STT_GNU_IFUNC |
| // symbol. This makes a function address in a PIE executable |
| // match the address in a shared library that it links against. |
| Reloc_section* rela_dyn = |
| target->rela_irelative_section(layout); |
| unsigned int r_type = elfcpp::R_AARCH64_IRELATIVE; |
| rela_dyn->add_symbolless_global_addend(gsym, r_type, |
| output_section, object, |
| data_shndx, |
| rela.get_r_offset(), |
| rela.get_r_addend()); |
| } |
| else if (r_type == elfcpp::R_AARCH64_ABS64 |
| && gsym->can_use_relative_reloc(false)) |
| { |
| Reloc_section* rela_dyn = target->rela_dyn_section(layout); |
| rela_dyn->add_global_relative(gsym, |
| elfcpp::R_AARCH64_RELATIVE, |
| output_section, |
| object, |
| data_shndx, |
| rela.get_r_offset(), |
| rela.get_r_addend(), |
| false); |
| } |
| else |
| { |
| check_non_pic(object, r_type); |
| Output_data_reloc<elfcpp::SHT_RELA, true, size, big_endian>* |
| rela_dyn = target->rela_dyn_section(layout); |
| rela_dyn->add_global( |
| gsym, r_type, output_section, object, |
| data_shndx, rela.get_r_offset(),rela.get_r_addend()); |
| } |
| } |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_PREL16: |
| case elfcpp::R_AARCH64_PREL32: |
| case elfcpp::R_AARCH64_PREL64: |
| // This is used to fill the GOT absolute address. |
| if (gsym->needs_plt_entry()) |
| { |
| target->make_plt_entry(symtab, layout, gsym); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_MOVW_UABS_G0: // 263 |
| case elfcpp::R_AARCH64_MOVW_UABS_G0_NC: // 264 |
| case elfcpp::R_AARCH64_MOVW_UABS_G1: // 265 |
| case elfcpp::R_AARCH64_MOVW_UABS_G1_NC: // 266 |
| case elfcpp::R_AARCH64_MOVW_UABS_G2: // 267 |
| case elfcpp::R_AARCH64_MOVW_UABS_G2_NC: // 268 |
| case elfcpp::R_AARCH64_MOVW_UABS_G3: // 269 |
| case elfcpp::R_AARCH64_MOVW_SABS_G0: // 270 |
| case elfcpp::R_AARCH64_MOVW_SABS_G1: // 271 |
| case elfcpp::R_AARCH64_MOVW_SABS_G2: // 272 |
| if (parameters->options().output_is_position_independent()) |
| { |
| gold_error(_("%s: unsupported reloc %u in pos independent link."), |
| object->name().c_str(), r_type); |
| } |
| // Make a PLT entry if necessary. |
| if (gsym->needs_plt_entry()) |
| { |
| target->make_plt_entry(symtab, layout, gsym); |
| // Since this is not a PC-relative relocation, we may be |
| // taking the address of a function. In that case we need to |
| // set the entry in the dynamic symbol table to the address of |
| // the PLT entry. |
| if (gsym->is_from_dynobj() && !parameters->options().shared()) |
| gsym->set_needs_dynsym_value(); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_LD_PREL_LO19: // 273 |
| case elfcpp::R_AARCH64_ADR_PREL_LO21: // 274 |
| case elfcpp::R_AARCH64_ADR_PREL_PG_HI21: // 275 |
| case elfcpp::R_AARCH64_ADR_PREL_PG_HI21_NC: // 276 |
| case elfcpp::R_AARCH64_ADD_ABS_LO12_NC: // 277 |
| case elfcpp::R_AARCH64_LDST8_ABS_LO12_NC: // 278 |
| case elfcpp::R_AARCH64_LDST16_ABS_LO12_NC: // 284 |
| case elfcpp::R_AARCH64_LDST32_ABS_LO12_NC: // 285 |
| case elfcpp::R_AARCH64_LDST64_ABS_LO12_NC: // 286 |
| case elfcpp::R_AARCH64_LDST128_ABS_LO12_NC: // 299 |
| { |
| if (gsym->needs_plt_entry()) |
| target->make_plt_entry(symtab, layout, gsym); |
| // Make a dynamic relocation if necessary. |
| if (gsym->needs_dynamic_reloc(arp->reference_flags())) |
| { |
| if (parameters->options().output_is_executable() |
| && gsym->may_need_copy_reloc()) |
| { |
| target->copy_reloc(symtab, layout, object, |
| data_shndx, output_section, gsym, rela); |
| } |
| } |
| break; |
| } |
| |
| case elfcpp::R_AARCH64_ADR_GOT_PAGE: |
| case elfcpp::R_AARCH64_LD64_GOT_LO12_NC: |
| case elfcpp::R_AARCH64_LD64_GOTPAGE_LO15: |
| { |
| // The above relocations are used to access GOT entries. |
| // Note a GOT entry is an *address* to a symbol. |
| // The symbol requires a GOT entry |
| Output_data_got_aarch64<size, big_endian>* got = |
| target->got_section(symtab, layout); |
| if (gsym->final_value_is_known()) |
| { |
| // For a STT_GNU_IFUNC symbol we want the PLT address. |
| if (gsym->type() == elfcpp::STT_GNU_IFUNC) |
| got->add_global_plt(gsym, GOT_TYPE_STANDARD); |
| else |
| got->add_global(gsym, GOT_TYPE_STANDARD); |
| } |
| else |
| { |
| // If this symbol is not fully resolved, we need to add a dynamic |
| // relocation for it. |
| Reloc_section* rela_dyn = target->rela_dyn_section(layout); |
| |
| // Use a GLOB_DAT rather than a RELATIVE reloc if: |
| // |
| // 1) The symbol may be defined in some other module. |
| // 2) We are building a shared library and this is a protected |
| // symbol; using GLOB_DAT means that the dynamic linker can use |
| // the address of the PLT in the main executable when appropriate |
| // so that function address comparisons work. |
| // 3) This is a STT_GNU_IFUNC symbol in position dependent code, |
| // again so that function address comparisons work. |
| if (gsym->is_from_dynobj() |
| || gsym->is_undefined() |
| || gsym->is_preemptible() |
| || (gsym->visibility() == elfcpp::STV_PROTECTED |
| && parameters->options().shared()) |
| || (gsym->type() == elfcpp::STT_GNU_IFUNC |
| && parameters->options().output_is_position_independent())) |
| got->add_global_with_rel(gsym, GOT_TYPE_STANDARD, |
| rela_dyn, elfcpp::R_AARCH64_GLOB_DAT); |
| else |
| { |
| // For a STT_GNU_IFUNC symbol we want to write the PLT |
| // offset into the GOT, so that function pointer |
| // comparisons work correctly. |
| bool is_new; |
| if (gsym->type() != elfcpp::STT_GNU_IFUNC) |
| is_new = got->add_global(gsym, GOT_TYPE_STANDARD); |
| else |
| { |
| is_new = got->add_global_plt(gsym, GOT_TYPE_STANDARD); |
| // Tell the dynamic linker to use the PLT address |
| // when resolving relocations. |
| if (gsym->is_from_dynobj() |
| && !parameters->options().shared()) |
| gsym->set_needs_dynsym_value(); |
| } |
| if (is_new) |
| { |
| rela_dyn->add_global_relative( |
| gsym, elfcpp::R_AARCH64_RELATIVE, |
| got, |
| gsym->got_offset(GOT_TYPE_STANDARD), |
| 0, |
| false); |
| } |
| } |
| } |
| break; |
| } |
| |
| case elfcpp::R_AARCH64_TSTBR14: |
| case elfcpp::R_AARCH64_CONDBR19: |
| case elfcpp::R_AARCH64_JUMP26: |
| case elfcpp::R_AARCH64_CALL26: |
| { |
| if (gsym->final_value_is_known()) |
| break; |
| |
| if (gsym->is_defined() && |
| !gsym->is_from_dynobj() && |
| !gsym->is_preemptible()) |
| break; |
| |
| // Make plt entry for function call. |
| target->make_plt_entry(symtab, layout, gsym); |
| break; |
| } |
| |
| case elfcpp::R_AARCH64_TLSGD_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC: // General dynamic |
| { |
| tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>:: |
| optimize_tls_reloc(gsym->final_value_is_known(), r_type); |
| if (tlsopt == tls::TLSOPT_TO_LE) |
| { |
| layout->set_has_static_tls(); |
| break; |
| } |
| gold_assert(tlsopt == tls::TLSOPT_NONE); |
| |
| // General dynamic. |
| Output_data_got_aarch64<size, big_endian>* got = |
| target->got_section(symtab, layout); |
| // Create 2 consecutive entries for module index and offset. |
| got->add_global_pair_with_rel(gsym, GOT_TYPE_TLS_PAIR, |
| target->rela_dyn_section(layout), |
| elfcpp::R_AARCH64_TLS_DTPMOD64, |
| elfcpp::R_AARCH64_TLS_DTPREL64); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSLD_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC: // Local dynamic |
| { |
| tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>:: |
| optimize_tls_reloc(!parameters->options().shared(), r_type); |
| if (tlsopt == tls::TLSOPT_NONE) |
| { |
| // Create a GOT entry for the module index. |
| target->got_mod_index_entry(symtab, layout, object); |
| } |
| else if (tlsopt != tls::TLSOPT_TO_LE) |
| unsupported_reloc_local(object, r_type); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G1: |
| case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G0_NC: |
| case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_HI12: |
| case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_LO12_NC: // Other local dynamic |
| break; |
| |
| case elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21: |
| case elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC: // Initial executable |
| { |
| tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>:: |
| optimize_tls_reloc(gsym->final_value_is_known(), r_type); |
| if (tlsopt == tls::TLSOPT_TO_LE) |
| break; |
| |
| layout->set_has_static_tls(); |
| // Create a GOT entry for the tp-relative offset. |
| Output_data_got_aarch64<size, big_endian>* got |
| = target->got_section(symtab, layout); |
| if (!parameters->doing_static_link()) |
| { |
| got->add_global_with_rel( |
| gsym, GOT_TYPE_TLS_OFFSET, |
| target->rela_dyn_section(layout), |
| elfcpp::R_AARCH64_TLS_TPREL64); |
| } |
| if (!gsym->has_got_offset(GOT_TYPE_TLS_OFFSET)) |
| { |
| got->add_global(gsym, GOT_TYPE_TLS_OFFSET); |
| unsigned int got_offset = |
| gsym->got_offset(GOT_TYPE_TLS_OFFSET); |
| const elfcpp::Elf_Xword addend = rela.get_r_addend(); |
| gold_assert(addend == 0); |
| got->add_static_reloc(got_offset, |
| elfcpp::R_AARCH64_TLS_TPREL64, gsym); |
| } |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G2: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1_NC: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0_NC: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12_NC: // Local executable |
| layout->set_has_static_tls(); |
| if (parameters->options().shared()) |
| gold_error(_("%s: unsupported TLSLE reloc type %u in shared objects."), |
| object->name().c_str(), r_type); |
| break; |
| |
| case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSDESC_LD64_LO12: |
| case elfcpp::R_AARCH64_TLSDESC_ADD_LO12: // TLS descriptor |
| { |
| target->define_tls_base_symbol(symtab, layout); |
| tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>:: |
| optimize_tls_reloc(gsym->final_value_is_known(), r_type); |
| if (tlsopt == tls::TLSOPT_NONE) |
| { |
| // Create reserved PLT and GOT entries for the resolver. |
| target->reserve_tlsdesc_entries(symtab, layout); |
| |
| // Create a double GOT entry with an R_AARCH64_TLSDESC |
| // relocation. The R_AARCH64_TLSDESC is resolved lazily, so the GOT |
| // entry needs to be in an area in .got.plt, not .got. Call |
| // got_section to make sure the section has been created. |
| target->got_section(symtab, layout); |
| Output_data_got<size, big_endian>* got = |
| target->got_tlsdesc_section(); |
| Reloc_section* rt = target->rela_tlsdesc_section(layout); |
| got->add_global_pair_with_rel(gsym, GOT_TYPE_TLS_DESC, rt, |
| elfcpp::R_AARCH64_TLSDESC, 0); |
| } |
| else if (tlsopt == tls::TLSOPT_TO_IE) |
| { |
| // Create a GOT entry for the tp-relative offset. |
| Output_data_got<size, big_endian>* got |
| = target->got_section(symtab, layout); |
| got->add_global_with_rel(gsym, GOT_TYPE_TLS_OFFSET, |
| target->rela_dyn_section(layout), |
| elfcpp::R_AARCH64_TLS_TPREL64); |
| } |
| else if (tlsopt != tls::TLSOPT_TO_LE) |
| unsupported_reloc_global(object, r_type, gsym); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSDESC_CALL: |
| break; |
| |
| default: |
| gold_error(_("%s: unsupported reloc type in global scan"), |
| aarch64_reloc_property_table-> |
| reloc_name_in_error_message(r_type).c_str()); |
| } |
| return; |
| } // End of Scan::global |
| |
| |
| // Create the PLT section. |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::make_plt_section( |
| Symbol_table* symtab, Layout* layout) |
| { |
| if (this->plt_ == NULL) |
| { |
| // Create the GOT section first. |
| this->got_section(symtab, layout); |
| |
| this->plt_ = this->make_data_plt(layout, this->got_, this->got_plt_, |
| this->got_irelative_); |
| |
| layout->add_output_section_data(".plt", elfcpp::SHT_PROGBITS, |
| (elfcpp::SHF_ALLOC |
| | elfcpp::SHF_EXECINSTR), |
| this->plt_, ORDER_PLT, false); |
| |
| // Make the sh_info field of .rela.plt point to .plt. |
| Output_section* rela_plt_os = this->plt_->rela_plt()->output_section(); |
| rela_plt_os->set_info_section(this->plt_->output_section()); |
| } |
| } |
| |
| // Return the section for TLSDESC relocations. |
| |
| template<int size, bool big_endian> |
| typename Target_aarch64<size, big_endian>::Reloc_section* |
| Target_aarch64<size, big_endian>::rela_tlsdesc_section(Layout* layout) const |
| { |
| return this->plt_section()->rela_tlsdesc(layout); |
| } |
| |
| // Create a PLT entry for a global symbol. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::make_plt_entry( |
| Symbol_table* symtab, |
| Layout* layout, |
| Symbol* gsym) |
| { |
| if (gsym->has_plt_offset()) |
| return; |
| |
| if (this->plt_ == NULL) |
| this->make_plt_section(symtab, layout); |
| |
| this->plt_->add_entry(symtab, layout, gsym); |
| } |
| |
| // Make a PLT entry for a local STT_GNU_IFUNC symbol. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::make_local_ifunc_plt_entry( |
| Symbol_table* symtab, Layout* layout, |
| Sized_relobj_file<size, big_endian>* relobj, |
| unsigned int local_sym_index) |
| { |
| if (relobj->local_has_plt_offset(local_sym_index)) |
| return; |
| if (this->plt_ == NULL) |
| this->make_plt_section(symtab, layout); |
| unsigned int plt_offset = this->plt_->add_local_ifunc_entry(symtab, layout, |
| relobj, |
| local_sym_index); |
| relobj->set_local_plt_offset(local_sym_index, plt_offset); |
| } |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::gc_process_relocs( |
| Symbol_table* symtab, |
| Layout* layout, |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int data_shndx, |
| unsigned int sh_type, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| bool needs_special_offset_handling, |
| size_t local_symbol_count, |
| const unsigned char* plocal_symbols) |
| { |
| typedef Target_aarch64<size, big_endian> Aarch64; |
| typedef gold::Default_classify_reloc<elfcpp::SHT_RELA, size, big_endian> |
| Classify_reloc; |
| |
| if (sh_type == elfcpp::SHT_REL) |
| { |
| return; |
| } |
| |
| gold::gc_process_relocs<size, big_endian, Aarch64, Scan, Classify_reloc>( |
| symtab, |
| layout, |
| this, |
| object, |
| data_shndx, |
| prelocs, |
| reloc_count, |
| output_section, |
| needs_special_offset_handling, |
| local_symbol_count, |
| plocal_symbols); |
| } |
| |
| // Scan relocations for a section. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::scan_relocs( |
| Symbol_table* symtab, |
| Layout* layout, |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int data_shndx, |
| unsigned int sh_type, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| bool needs_special_offset_handling, |
| size_t local_symbol_count, |
| const unsigned char* plocal_symbols) |
| { |
| typedef Target_aarch64<size, big_endian> Aarch64; |
| typedef gold::Default_classify_reloc<elfcpp::SHT_RELA, size, big_endian> |
| Classify_reloc; |
| |
| if (sh_type == elfcpp::SHT_REL) |
| { |
| gold_error(_("%s: unsupported REL reloc section"), |
| object->name().c_str()); |
| return; |
| } |
| |
| gold::scan_relocs<size, big_endian, Aarch64, Scan, Classify_reloc>( |
| symtab, |
| layout, |
| this, |
| object, |
| data_shndx, |
| prelocs, |
| reloc_count, |
| output_section, |
| needs_special_offset_handling, |
| local_symbol_count, |
| plocal_symbols); |
| } |
| |
| // Return the value to use for a dynamic which requires special |
| // treatment. This is how we support equality comparisons of function |
| // pointers across shared library boundaries, as described in the |
| // processor specific ABI supplement. |
| |
| template<int size, bool big_endian> |
| uint64_t |
| Target_aarch64<size, big_endian>::do_dynsym_value(const Symbol* gsym) const |
| { |
| gold_assert(gsym->is_from_dynobj() && gsym->has_plt_offset()); |
| return this->plt_address_for_global(gsym); |
| } |
| |
| |
| // Finalize the sections. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::do_finalize_sections( |
| Layout* layout, |
| const Input_objects*, |
| Symbol_table* symtab) |
| { |
| const Reloc_section* rel_plt = (this->plt_ == NULL |
| ? NULL |
| : this->plt_->rela_plt()); |
| layout->add_target_dynamic_tags(false, this->got_plt_, rel_plt, |
| this->rela_dyn_, true, false); |
| |
| // Emit any relocs we saved in an attempt to avoid generating COPY |
| // relocs. |
| if (this->copy_relocs_.any_saved_relocs()) |
| this->copy_relocs_.emit(this->rela_dyn_section(layout)); |
| |
| // Fill in some more dynamic tags. |
| Output_data_dynamic* const odyn = layout->dynamic_data(); |
| if (odyn != NULL) |
| { |
| if (this->plt_ != NULL |
| && this->plt_->output_section() != NULL |
| && this->plt_ ->has_tlsdesc_entry()) |
| { |
| unsigned int plt_offset = this->plt_->get_tlsdesc_plt_offset(); |
| unsigned int got_offset = this->plt_->get_tlsdesc_got_offset(); |
| this->got_->finalize_data_size(); |
| odyn->add_section_plus_offset(elfcpp::DT_TLSDESC_PLT, |
| this->plt_, plt_offset); |
| odyn->add_section_plus_offset(elfcpp::DT_TLSDESC_GOT, |
| this->got_, got_offset); |
| } |
| } |
| |
| // Set the size of the _GLOBAL_OFFSET_TABLE_ symbol to the size of |
| // the .got section. |
| Symbol* sym = this->global_offset_table_; |
| if (sym != NULL) |
| { |
| uint64_t data_size = this->got_->current_data_size(); |
| symtab->get_sized_symbol<size>(sym)->set_symsize(data_size); |
| |
| // If the .got section is more than 0x8000 bytes, we add |
| // 0x8000 to the value of _GLOBAL_OFFSET_TABLE_, so that 16 |
| // bit relocations have a greater chance of working. |
| if (data_size >= 0x8000) |
| symtab->get_sized_symbol<size>(sym)->set_value( |
| symtab->get_sized_symbol<size>(sym)->value() + 0x8000); |
| } |
| |
| if (parameters->doing_static_link() |
| && (this->plt_ == NULL || !this->plt_->has_irelative_section())) |
| { |
| // If linking statically, make sure that the __rela_iplt symbols |
| // were defined if necessary, even if we didn't create a PLT. |
| static const Define_symbol_in_segment syms[] = |
| { |
| { |
| "__rela_iplt_start", // name |
| elfcpp::PT_LOAD, // segment_type |
| elfcpp::PF_W, // segment_flags_set |
| elfcpp::PF(0), // segment_flags_clear |
| 0, // value |
| 0, // size |
| elfcpp::STT_NOTYPE, // type |
| elfcpp::STB_GLOBAL, // binding |
| elfcpp::STV_HIDDEN, // visibility |
| 0, // nonvis |
| Symbol::SEGMENT_START, // offset_from_base |
| true // only_if_ref |
| }, |
| { |
| "__rela_iplt_end", // name |
| elfcpp::PT_LOAD, // segment_type |
| elfcpp::PF_W, // segment_flags_set |
| elfcpp::PF(0), // segment_flags_clear |
| 0, // value |
| 0, // size |
| elfcpp::STT_NOTYPE, // type |
| elfcpp::STB_GLOBAL, // binding |
| elfcpp::STV_HIDDEN, // visibility |
| 0, // nonvis |
| Symbol::SEGMENT_START, // offset_from_base |
| true // only_if_ref |
| } |
| }; |
| |
| symtab->define_symbols(layout, 2, syms, |
| layout->script_options()->saw_sections_clause()); |
| } |
| |
| return; |
| } |
| |
| // Perform a relocation. |
| |
| template<int size, bool big_endian> |
| inline bool |
| Target_aarch64<size, big_endian>::Relocate::relocate( |
| const Relocate_info<size, big_endian>* relinfo, |
| unsigned int, |
| Target_aarch64<size, big_endian>* target, |
| Output_section* , |
| size_t relnum, |
| const unsigned char* preloc, |
| const Sized_symbol<size>* gsym, |
| const Symbol_value<size>* psymval, |
| unsigned char* view, |
| typename elfcpp::Elf_types<size>::Elf_Addr address, |
| section_size_type /* view_size */) |
| { |
| if (view == NULL) |
| return true; |
| |
| typedef AArch64_relocate_functions<size, big_endian> Reloc; |
| |
| const elfcpp::Rela<size, big_endian> rela(preloc); |
| unsigned int r_type = elfcpp::elf_r_type<size>(rela.get_r_info()); |
| const AArch64_reloc_property* reloc_property = |
| aarch64_reloc_property_table->get_reloc_property(r_type); |
| |
| if (reloc_property == NULL) |
| { |
| std::string reloc_name = |
| aarch64_reloc_property_table->reloc_name_in_error_message(r_type); |
| gold_error_at_location(relinfo, relnum, rela.get_r_offset(), |
| _("cannot relocate %s in object file"), |
| reloc_name.c_str()); |
| return true; |
| } |
| |
| const Sized_relobj_file<size, big_endian>* object = relinfo->object; |
| |
| // Pick the value to use for symbols defined in the PLT. |
| Symbol_value<size> symval; |
| if (gsym != NULL |
| && gsym->use_plt_offset(reloc_property->reference_flags())) |
| { |
| symval.set_output_value(target->plt_address_for_global(gsym)); |
| psymval = &symval; |
| } |
| else if (gsym == NULL && psymval->is_ifunc_symbol()) |
| { |
| unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info()); |
| if (object->local_has_plt_offset(r_sym)) |
| { |
| symval.set_output_value(target->plt_address_for_local(object, r_sym)); |
| psymval = &symval; |
| } |
| } |
| |
| const elfcpp::Elf_Xword addend = rela.get_r_addend(); |
| |
| // Get the GOT offset if needed. |
| // For aarch64, the GOT pointer points to the start of the GOT section. |
| bool have_got_offset = false; |
| int got_offset = 0; |
| int got_base = (target->got_ != NULL |
| ? (target->got_->current_data_size() >= 0x8000 |
| ? 0x8000 : 0) |
| : 0); |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_MOVW_GOTOFF_G0: |
| case elfcpp::R_AARCH64_MOVW_GOTOFF_G0_NC: |
| case elfcpp::R_AARCH64_MOVW_GOTOFF_G1: |
| case elfcpp::R_AARCH64_MOVW_GOTOFF_G1_NC: |
| case elfcpp::R_AARCH64_MOVW_GOTOFF_G2: |
| case elfcpp::R_AARCH64_MOVW_GOTOFF_G2_NC: |
| case elfcpp::R_AARCH64_MOVW_GOTOFF_G3: |
| case elfcpp::R_AARCH64_GOTREL64: |
| case elfcpp::R_AARCH64_GOTREL32: |
| case elfcpp::R_AARCH64_GOT_LD_PREL19: |
| case elfcpp::R_AARCH64_LD64_GOTOFF_LO15: |
| case elfcpp::R_AARCH64_ADR_GOT_PAGE: |
| case elfcpp::R_AARCH64_LD64_GOT_LO12_NC: |
| case elfcpp::R_AARCH64_LD64_GOTPAGE_LO15: |
| if (gsym != NULL) |
| { |
| gold_assert(gsym->has_got_offset(GOT_TYPE_STANDARD)); |
| got_offset = gsym->got_offset(GOT_TYPE_STANDARD) - got_base; |
| } |
| else |
| { |
| unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info()); |
| gold_assert(object->local_has_got_offset(r_sym, GOT_TYPE_STANDARD)); |
| got_offset = (object->local_got_offset(r_sym, GOT_TYPE_STANDARD) |
| - got_base); |
| } |
| have_got_offset = true; |
| break; |
| |
| default: |
| break; |
| } |
| |
| typename Reloc::Status reloc_status = Reloc::STATUS_OKAY; |
| typename elfcpp::Elf_types<size>::Elf_Addr value; |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_NONE: |
| break; |
| |
| case elfcpp::R_AARCH64_ABS64: |
| if (!parameters->options().apply_dynamic_relocs() |
| && parameters->options().output_is_position_independent() |
| && gsym != NULL |
| && gsym->needs_dynamic_reloc(reloc_property->reference_flags()) |
| && !gsym->can_use_relative_reloc(false)) |
| // We have generated an absolute dynamic relocation, so do not |
| // apply the relocation statically. (Works around bugs in older |
| // Android dynamic linkers.) |
| break; |
| reloc_status = Reloc::template rela_ua<64>( |
| view, object, psymval, addend, reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_ABS32: |
| if (!parameters->options().apply_dynamic_relocs() |
| && parameters->options().output_is_position_independent() |
| && gsym != NULL |
| && gsym->needs_dynamic_reloc(reloc_property->reference_flags())) |
| // We have generated an absolute dynamic relocation, so do not |
| // apply the relocation statically. (Works around bugs in older |
| // Android dynamic linkers.) |
| break; |
| reloc_status = Reloc::template rela_ua<32>( |
| view, object, psymval, addend, reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_ABS16: |
| if (!parameters->options().apply_dynamic_relocs() |
| && parameters->options().output_is_position_independent() |
| && gsym != NULL |
| && gsym->needs_dynamic_reloc(reloc_property->reference_flags())) |
| // We have generated an absolute dynamic relocation, so do not |
| // apply the relocation statically. (Works around bugs in older |
| // Android dynamic linkers.) |
| break; |
| reloc_status = Reloc::template rela_ua<16>( |
| view, object, psymval, addend, reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_PREL64: |
| reloc_status = Reloc::template pcrela_ua<64>( |
| view, object, psymval, addend, address, reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_PREL32: |
| reloc_status = Reloc::template pcrela_ua<32>( |
| view, object, psymval, addend, address, reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_PREL16: |
| reloc_status = Reloc::template pcrela_ua<16>( |
| view, object, psymval, addend, address, reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_MOVW_UABS_G0: |
| case elfcpp::R_AARCH64_MOVW_UABS_G0_NC: |
| case elfcpp::R_AARCH64_MOVW_UABS_G1: |
| case elfcpp::R_AARCH64_MOVW_UABS_G1_NC: |
| case elfcpp::R_AARCH64_MOVW_UABS_G2: |
| case elfcpp::R_AARCH64_MOVW_UABS_G2_NC: |
| case elfcpp::R_AARCH64_MOVW_UABS_G3: |
| reloc_status = Reloc::template rela_general<32>( |
| view, object, psymval, addend, reloc_property); |
| break; |
| case elfcpp::R_AARCH64_MOVW_SABS_G0: |
| case elfcpp::R_AARCH64_MOVW_SABS_G1: |
| case elfcpp::R_AARCH64_MOVW_SABS_G2: |
| reloc_status = Reloc::movnz(view, psymval->value(object, addend), |
| reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_LD_PREL_LO19: |
| reloc_status = Reloc::template pcrela_general<32>( |
| view, object, psymval, addend, address, reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_ADR_PREL_LO21: |
| reloc_status = Reloc::adr(view, object, psymval, addend, |
| address, reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_ADR_PREL_PG_HI21_NC: |
| case elfcpp::R_AARCH64_ADR_PREL_PG_HI21: |
| reloc_status = Reloc::adrp(view, object, psymval, addend, address, |
| reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_LDST8_ABS_LO12_NC: |
| case elfcpp::R_AARCH64_LDST16_ABS_LO12_NC: |
| case elfcpp::R_AARCH64_LDST32_ABS_LO12_NC: |
| case elfcpp::R_AARCH64_LDST64_ABS_LO12_NC: |
| case elfcpp::R_AARCH64_LDST128_ABS_LO12_NC: |
| case elfcpp::R_AARCH64_ADD_ABS_LO12_NC: |
| reloc_status = Reloc::template rela_general<32>( |
| view, object, psymval, addend, reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_CALL26: |
| if (this->skip_call_tls_get_addr_) |
| { |
| // Double check that the TLSGD insn has been optimized away. |
| typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype; |
| Insntype insn = elfcpp::Swap<32, big_endian>::readval( |
| reinterpret_cast<Insntype*>(view)); |
| gold_assert((insn & 0xff000000) == 0x91000000); |
| |
| reloc_status = Reloc::STATUS_OKAY; |
| this->skip_call_tls_get_addr_ = false; |
| // Return false to stop further processing this reloc. |
| return false; |
| } |
| // Fall through. |
| case elfcpp::R_AARCH64_JUMP26: |
| if (Reloc::maybe_apply_stub(r_type, relinfo, rela, view, address, |
| gsym, psymval, object, |
| target->stub_group_size_)) |
| break; |
| // Fall through. |
| case elfcpp::R_AARCH64_TSTBR14: |
| case elfcpp::R_AARCH64_CONDBR19: |
| reloc_status = Reloc::template pcrela_general<32>( |
| view, object, psymval, addend, address, reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_ADR_GOT_PAGE: |
| gold_assert(have_got_offset); |
| value = target->got_->address() + got_base + got_offset; |
| reloc_status = Reloc::adrp(view, value + addend, address); |
| break; |
| |
| case elfcpp::R_AARCH64_LD64_GOT_LO12_NC: |
| gold_assert(have_got_offset); |
| value = target->got_->address() + got_base + got_offset; |
| reloc_status = Reloc::template rela_general<32>( |
| view, value, addend, reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_LD64_GOTPAGE_LO15: |
| { |
| gold_assert(have_got_offset); |
| value = target->got_->address() + got_base + got_offset + addend - |
| Reloc::Page(target->got_->address() + got_base); |
| if ((value & 7) != 0) |
| reloc_status = Reloc::STATUS_OVERFLOW; |
| else |
| reloc_status = Reloc::template reloc_common<32>( |
| view, value, reloc_property); |
| break; |
| } |
| |
| case elfcpp::R_AARCH64_TLSGD_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLD_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G1: |
| case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G0_NC: |
| case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_HI12: |
| case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21: |
| case elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G2: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1_NC: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0_NC: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSDESC_LD64_LO12: |
| case elfcpp::R_AARCH64_TLSDESC_ADD_LO12: |
| case elfcpp::R_AARCH64_TLSDESC_CALL: |
| reloc_status = relocate_tls(relinfo, target, relnum, rela, r_type, |
| gsym, psymval, view, address); |
| break; |
| |
| // These are dynamic relocations, which are unexpected when linking. |
| case elfcpp::R_AARCH64_COPY: |
| case elfcpp::R_AARCH64_GLOB_DAT: |
| case elfcpp::R_AARCH64_JUMP_SLOT: |
| case elfcpp::R_AARCH64_RELATIVE: |
| case elfcpp::R_AARCH64_IRELATIVE: |
| case elfcpp::R_AARCH64_TLS_DTPREL64: |
| case elfcpp::R_AARCH64_TLS_DTPMOD64: |
| case elfcpp::R_AARCH64_TLS_TPREL64: |
| case elfcpp::R_AARCH64_TLSDESC: |
| gold_error_at_location(relinfo, relnum, rela.get_r_offset(), |
| _("unexpected reloc %u in object file"), |
| r_type); |
| break; |
| |
| default: |
| gold_error_at_location(relinfo, relnum, rela.get_r_offset(), |
| _("unsupported reloc %s"), |
| reloc_property->name().c_str()); |
| break; |
| } |
| |
| // Report any errors. |
| switch (reloc_status) |
| { |
| case Reloc::STATUS_OKAY: |
| break; |
| case Reloc::STATUS_OVERFLOW: |
| gold_error_at_location(relinfo, relnum, rela.get_r_offset(), |
| _("relocation overflow in %s"), |
| reloc_property->name().c_str()); |
| break; |
| case Reloc::STATUS_BAD_RELOC: |
| gold_error_at_location( |
| relinfo, |
| relnum, |
| rela.get_r_offset(), |
| _("unexpected opcode while processing relocation %s"), |
| reloc_property->name().c_str()); |
| break; |
| default: |
| gold_unreachable(); |
| } |
| |
| return true; |
| } |
| |
| |
| template<int size, bool big_endian> |
| inline |
| typename AArch64_relocate_functions<size, big_endian>::Status |
| Target_aarch64<size, big_endian>::Relocate::relocate_tls( |
| const Relocate_info<size, big_endian>* relinfo, |
| Target_aarch64<size, big_endian>* target, |
| size_t relnum, |
| const elfcpp::Rela<size, big_endian>& rela, |
| unsigned int r_type, const Sized_symbol<size>* gsym, |
| const Symbol_value<size>* psymval, |
| unsigned char* view, |
| typename elfcpp::Elf_types<size>::Elf_Addr address) |
| { |
| typedef AArch64_relocate_functions<size, big_endian> aarch64_reloc_funcs; |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address; |
| |
| Output_segment* tls_segment = relinfo->layout->tls_segment(); |
| const elfcpp::Elf_Xword addend = rela.get_r_addend(); |
| const AArch64_reloc_property* reloc_property = |
| aarch64_reloc_property_table->get_reloc_property(r_type); |
| gold_assert(reloc_property != NULL); |
| |
| const bool is_final = (gsym == NULL |
| ? !parameters->options().shared() |
| : gsym->final_value_is_known()); |
| tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>:: |
| optimize_tls_reloc(is_final, r_type); |
| |
| Sized_relobj_file<size, big_endian>* object = relinfo->object; |
| int tls_got_offset_type; |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_TLSGD_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC: // Global-dynamic |
| { |
| if (tlsopt == tls::TLSOPT_TO_LE) |
| { |
| if (tls_segment == NULL) |
| { |
| gold_assert(parameters->errors()->error_count() > 0 |
| || issue_undefined_symbol_error(gsym)); |
| return aarch64_reloc_funcs::STATUS_BAD_RELOC; |
| } |
| return tls_gd_to_le(relinfo, target, rela, r_type, view, |
| psymval); |
| } |
| else if (tlsopt == tls::TLSOPT_NONE) |
| { |
| tls_got_offset_type = GOT_TYPE_TLS_PAIR; |
| // Firstly get the address for the got entry. |
| typename elfcpp::Elf_types<size>::Elf_Addr got_entry_address; |
| if (gsym != NULL) |
| { |
| gold_assert(gsym->has_got_offset(tls_got_offset_type)); |
| got_entry_address = target->got_->address() + |
| gsym->got_offset(tls_got_offset_type); |
| } |
| else |
| { |
| unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info()); |
| gold_assert( |
| object->local_has_got_offset(r_sym, tls_got_offset_type)); |
| got_entry_address = target->got_->address() + |
| object->local_got_offset(r_sym, tls_got_offset_type); |
| } |
| |
| // Relocate the address into adrp/ld, adrp/add pair. |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_TLSGD_ADR_PAGE21: |
| return aarch64_reloc_funcs::adrp( |
| view, got_entry_address + addend, address); |
| |
| break; |
| |
| case elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC: |
| return aarch64_reloc_funcs::template rela_general<32>( |
| view, got_entry_address, addend, reloc_property); |
| break; |
| |
| default: |
| gold_unreachable(); |
| } |
| } |
| gold_error_at_location(relinfo, relnum, rela.get_r_offset(), |
| _("unsupported gd_to_ie relaxation on %u"), |
| r_type); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSLD_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC: // Local-dynamic |
| { |
| if (tlsopt == tls::TLSOPT_TO_LE) |
| { |
| if (tls_segment == NULL) |
| { |
| gold_assert(parameters->errors()->error_count() > 0 |
| || issue_undefined_symbol_error(gsym)); |
| return aarch64_reloc_funcs::STATUS_BAD_RELOC; |
| } |
| return this->tls_ld_to_le(relinfo, target, rela, r_type, view, |
| psymval); |
| } |
| |
| gold_assert(tlsopt == tls::TLSOPT_NONE); |
| // Relocate the field with the offset of the GOT entry for |
| // the module index. |
| typename elfcpp::Elf_types<size>::Elf_Addr got_entry_address; |
| got_entry_address = (target->got_mod_index_entry(NULL, NULL, NULL) + |
| target->got_->address()); |
| |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_TLSLD_ADR_PAGE21: |
| return aarch64_reloc_funcs::adrp( |
| view, got_entry_address + addend, address); |
| break; |
| |
| case elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC: |
| return aarch64_reloc_funcs::template rela_general<32>( |
| view, got_entry_address, addend, reloc_property); |
| break; |
| |
| default: |
| gold_unreachable(); |
| } |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G1: |
| case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G0_NC: |
| case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_HI12: |
| case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_LO12_NC: // Other local-dynamic |
| { |
| AArch64_address value = psymval->value(object, 0); |
| if (tlsopt == tls::TLSOPT_TO_LE) |
| { |
| if (tls_segment == NULL) |
| { |
| gold_assert(parameters->errors()->error_count() > 0 |
| || issue_undefined_symbol_error(gsym)); |
| return aarch64_reloc_funcs::STATUS_BAD_RELOC; |
| } |
| } |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G1: |
| return aarch64_reloc_funcs::movnz(view, value + addend, |
| reloc_property); |
| break; |
| |
| case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G0_NC: |
| case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_HI12: |
| case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_LO12_NC: |
| return aarch64_reloc_funcs::template rela_general<32>( |
| view, value, addend, reloc_property); |
| break; |
| |
| default: |
| gold_unreachable(); |
| } |
| // We should never reach here. |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21: |
| case elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC: // Initial-exec |
| { |
| if (tlsopt == tls::TLSOPT_TO_LE) |
| { |
| if (tls_segment == NULL) |
| { |
| gold_assert(parameters->errors()->error_count() > 0 |
| || issue_undefined_symbol_error(gsym)); |
| return aarch64_reloc_funcs::STATUS_BAD_RELOC; |
| } |
| return tls_ie_to_le(relinfo, target, rela, r_type, view, |
| psymval); |
| } |
| tls_got_offset_type = GOT_TYPE_TLS_OFFSET; |
| |
| // Firstly get the address for the got entry. |
| typename elfcpp::Elf_types<size>::Elf_Addr got_entry_address; |
| if (gsym != NULL) |
| { |
| gold_assert(gsym->has_got_offset(tls_got_offset_type)); |
| got_entry_address = target->got_->address() + |
| gsym->got_offset(tls_got_offset_type); |
| } |
| else |
| { |
| unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info()); |
| gold_assert( |
| object->local_has_got_offset(r_sym, tls_got_offset_type)); |
| got_entry_address = target->got_->address() + |
| object->local_got_offset(r_sym, tls_got_offset_type); |
| } |
| // Relocate the address into adrp/ld, adrp/add pair. |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21: |
| return aarch64_reloc_funcs::adrp(view, got_entry_address + addend, |
| address); |
| break; |
| case elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC: |
| return aarch64_reloc_funcs::template rela_general<32>( |
| view, got_entry_address, addend, reloc_property); |
| default: |
| gold_unreachable(); |
| } |
| } |
| // We shall never reach here. |
| break; |
| |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G2: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1_NC: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0_NC: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12_NC: |
| case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12: |
| case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12_NC: |
| { |
| gold_assert(tls_segment != NULL); |
| AArch64_address value = psymval->value(object, 0); |
| |
| if (!parameters->options().shared()) |
| { |
| AArch64_address aligned_tcb_size = |
| align_address(target->tcb_size(), |
| tls_segment->maximum_alignment()); |
| value += aligned_tcb_size; |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G2: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1: |
| case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0: |
| return aarch64_reloc_funcs::movnz(view, value + addend, |
| reloc_property); |
| default: |
| return aarch64_reloc_funcs::template |
| rela_general<32>(view, |
| value, |
| addend, |
| reloc_property); |
| } |
| } |
| else |
| gold_error(_("%s: unsupported reloc %u " |
| "in non-static TLSLE mode."), |
| object->name().c_str(), r_type); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21: |
| case elfcpp::R_AARCH64_TLSDESC_LD64_LO12: |
| case elfcpp::R_AARCH64_TLSDESC_ADD_LO12: |
| case elfcpp::R_AARCH64_TLSDESC_CALL: |
| { |
| if (tlsopt == tls::TLSOPT_TO_LE) |
| { |
| if (tls_segment == NULL) |
| { |
| gold_assert(parameters->errors()->error_count() > 0 |
| || issue_undefined_symbol_error(gsym)); |
| return aarch64_reloc_funcs::STATUS_BAD_RELOC; |
| } |
| return tls_desc_gd_to_le(relinfo, target, rela, r_type, |
| view, psymval); |
| } |
| else |
| { |
| tls_got_offset_type = (tlsopt == tls::TLSOPT_TO_IE |
| ? GOT_TYPE_TLS_OFFSET |
| : GOT_TYPE_TLS_DESC); |
| int got_tlsdesc_offset = 0; |
| if (r_type != elfcpp::R_AARCH64_TLSDESC_CALL |
| && tlsopt == tls::TLSOPT_NONE) |
| { |
| // We created GOT entries in the .got.tlsdesc portion of the |
| // .got.plt section, but the offset stored in the symbol is the |
| // offset within .got.tlsdesc. |
| got_tlsdesc_offset = (target->got_tlsdesc_->address() |
| - target->got_->address()); |
| } |
| typename elfcpp::Elf_types<size>::Elf_Addr got_entry_address; |
| if (gsym != NULL) |
| { |
| gold_assert(gsym->has_got_offset(tls_got_offset_type)); |
| got_entry_address = target->got_->address() |
| + got_tlsdesc_offset |
| + gsym->got_offset(tls_got_offset_type); |
| } |
| else |
| { |
| unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info()); |
| gold_assert( |
| object->local_has_got_offset(r_sym, tls_got_offset_type)); |
| got_entry_address = target->got_->address() + |
| got_tlsdesc_offset + |
| object->local_got_offset(r_sym, tls_got_offset_type); |
| } |
| if (tlsopt == tls::TLSOPT_TO_IE) |
| { |
| return tls_desc_gd_to_ie(relinfo, target, rela, r_type, |
| view, psymval, got_entry_address, |
| address); |
| } |
| |
| // Now do tlsdesc relocation. |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21: |
| return aarch64_reloc_funcs::adrp(view, |
| got_entry_address + addend, |
| address); |
| break; |
| case elfcpp::R_AARCH64_TLSDESC_LD64_LO12: |
| case elfcpp::R_AARCH64_TLSDESC_ADD_LO12: |
| return aarch64_reloc_funcs::template rela_general<32>( |
| view, got_entry_address, addend, reloc_property); |
| break; |
| case elfcpp::R_AARCH64_TLSDESC_CALL: |
| return aarch64_reloc_funcs::STATUS_OKAY; |
| break; |
| default: |
| gold_unreachable(); |
| } |
| } |
| } |
| break; |
| |
| default: |
| gold_error(_("%s: unsupported TLS reloc %u."), |
| object->name().c_str(), r_type); |
| } |
| return aarch64_reloc_funcs::STATUS_BAD_RELOC; |
| } // End of relocate_tls. |
| |
| |
| template<int size, bool big_endian> |
| inline |
| typename AArch64_relocate_functions<size, big_endian>::Status |
| Target_aarch64<size, big_endian>::Relocate::tls_gd_to_le( |
| const Relocate_info<size, big_endian>* relinfo, |
| Target_aarch64<size, big_endian>* target, |
| const elfcpp::Rela<size, big_endian>& rela, |
| unsigned int r_type, |
| unsigned char* view, |
| const Symbol_value<size>* psymval) |
| { |
| typedef AArch64_relocate_functions<size, big_endian> aarch64_reloc_funcs; |
| typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype; |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address; |
| |
| Insntype* ip = reinterpret_cast<Insntype*>(view); |
| Insntype insn1 = elfcpp::Swap<32, big_endian>::readval(ip); |
| Insntype insn2 = elfcpp::Swap<32, big_endian>::readval(ip + 1); |
| Insntype insn3 = elfcpp::Swap<32, big_endian>::readval(ip + 2); |
| |
| if (r_type == elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC) |
| { |
| // This is the 2nd relocs, optimization should already have been |
| // done. |
| gold_assert((insn1 & 0xfff00000) == 0x91400000); |
| return aarch64_reloc_funcs::STATUS_OKAY; |
| } |
| |
| // The original sequence is - |
| // 90000000 adrp x0, 0 <main> |
| // 91000000 add x0, x0, #0x0 |
| // 94000000 bl 0 <__tls_get_addr> |
| // optimized to sequence - |
| // d53bd040 mrs x0, tpidr_el0 |
| // 91400000 add x0, x0, #0x0, lsl #12 |
| // 91000000 add x0, x0, #0x0 |
| |
| // Unlike tls_ie_to_le, we change the 3 insns in one function call when we |
| // encounter the first relocation "R_AARCH64_TLSGD_ADR_PAGE21". Because we |
| // have to change "bl tls_get_addr", which does not have a corresponding tls |
| // relocation type. So before proceeding, we need to make sure compiler |
| // does not change the sequence. |
| if(!(insn1 == 0x90000000 // adrp x0,0 |
| && insn2 == 0x91000000 // add x0, x0, #0x0 |
| && insn3 == 0x94000000)) // bl 0 |
| { |
| // Ideally we should give up gd_to_le relaxation and do gd access. |
| // However the gd_to_le relaxation decision has been made early |
| // in the scan stage, where we did not allocate any GOT entry for |
| // this symbol. Therefore we have to exit and report error now. |
| gold_error(_("unexpected reloc insn sequence while relaxing " |
| "tls gd to le for reloc %u."), r_type); |
| return aarch64_reloc_funcs::STATUS_BAD_RELOC; |
| } |
| |
| // Write new insns. |
| insn1 = 0xd53bd040; // mrs x0, tpidr_el0 |
| insn2 = 0x91400000; // add x0, x0, #0x0, lsl #12 |
| insn3 = 0x91000000; // add x0, x0, #0x0 |
| elfcpp::Swap<32, big_endian>::writeval(ip, insn1); |
| elfcpp::Swap<32, big_endian>::writeval(ip + 1, insn2); |
| elfcpp::Swap<32, big_endian>::writeval(ip + 2, insn3); |
| |
| // Calculate tprel value. |
| Output_segment* tls_segment = relinfo->layout->tls_segment(); |
| gold_assert(tls_segment != NULL); |
| AArch64_address value = psymval->value(relinfo->object, 0); |
| const elfcpp::Elf_Xword addend = rela.get_r_addend(); |
| AArch64_address aligned_tcb_size = |
| align_address(target->tcb_size(), tls_segment->maximum_alignment()); |
| AArch64_address x = value + aligned_tcb_size; |
| |
| // After new insns are written, apply TLSLE relocs. |
| const AArch64_reloc_property* rp1 = |
| aarch64_reloc_property_table->get_reloc_property( |
| elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12); |
| const AArch64_reloc_property* rp2 = |
| aarch64_reloc_property_table->get_reloc_property( |
| elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12); |
| gold_assert(rp1 != NULL && rp2 != NULL); |
| |
| typename aarch64_reloc_funcs::Status s1 = |
| aarch64_reloc_funcs::template rela_general<32>(view + 4, |
| x, |
| addend, |
| rp1); |
| if (s1 != aarch64_reloc_funcs::STATUS_OKAY) |
| return s1; |
| |
| typename aarch64_reloc_funcs::Status s2 = |
| aarch64_reloc_funcs::template rela_general<32>(view + 8, |
| x, |
| addend, |
| rp2); |
| |
| this->skip_call_tls_get_addr_ = true; |
| return s2; |
| } // End of tls_gd_to_le |
| |
| |
| template<int size, bool big_endian> |
| inline |
| typename AArch64_relocate_functions<size, big_endian>::Status |
| Target_aarch64<size, big_endian>::Relocate::tls_ld_to_le( |
| const Relocate_info<size, big_endian>* relinfo, |
| Target_aarch64<size, big_endian>* target, |
| const elfcpp::Rela<size, big_endian>& rela, |
| unsigned int r_type, |
| unsigned char* view, |
| const Symbol_value<size>* psymval) |
| { |
| typedef AArch64_relocate_functions<size, big_endian> aarch64_reloc_funcs; |
| typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype; |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address; |
| |
| Insntype* ip = reinterpret_cast<Insntype*>(view); |
| Insntype insn1 = elfcpp::Swap<32, big_endian>::readval(ip); |
| Insntype insn2 = elfcpp::Swap<32, big_endian>::readval(ip + 1); |
| Insntype insn3 = elfcpp::Swap<32, big_endian>::readval(ip + 2); |
| |
| if (r_type == elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC) |
| { |
| // This is the 2nd relocs, optimization should already have been |
| // done. |
| gold_assert((insn1 & 0xfff00000) == 0x91400000); |
| return aarch64_reloc_funcs::STATUS_OKAY; |
| } |
| |
| // The original sequence is - |
| // 90000000 adrp x0, 0 <main> |
| // 91000000 add x0, x0, #0x0 |
| // 94000000 bl 0 <__tls_get_addr> |
| // optimized to sequence - |
| // d53bd040 mrs x0, tpidr_el0 |
| // 91400000 add x0, x0, #0x0, lsl #12 |
| // 91000000 add x0, x0, #0x0 |
| |
| // Unlike tls_ie_to_le, we change the 3 insns in one function call when we |
| // encounter the first relocation "R_AARCH64_TLSLD_ADR_PAGE21". Because we |
| // have to change "bl tls_get_addr", which does not have a corresponding tls |
| // relocation type. So before proceeding, we need to make sure compiler |
| // does not change the sequence. |
| if(!(insn1 == 0x90000000 // adrp x0,0 |
| && insn2 == 0x91000000 // add x0, x0, #0x0 |
| && insn3 == 0x94000000)) // bl 0 |
| { |
| // Ideally we should give up gd_to_le relaxation and do gd access. |
| // However the gd_to_le relaxation decision has been made early |
| // in the scan stage, where we did not allocate a GOT entry for |
| // this symbol. Therefore we have to exit and report an error now. |
| gold_error(_("unexpected reloc insn sequence while relaxing " |
| "tls gd to le for reloc %u."), r_type); |
| return aarch64_reloc_funcs::STATUS_BAD_RELOC; |
| } |
| |
| // Write new insns. |
| insn1 = 0xd53bd040; // mrs x0, tpidr_el0 |
| insn2 = 0x91400000; // add x0, x0, #0x0, lsl #12 |
| insn3 = 0x91000000; // add x0, x0, #0x0 |
| elfcpp::Swap<32, big_endian>::writeval(ip, insn1); |
| elfcpp::Swap<32, big_endian>::writeval(ip + 1, insn2); |
| elfcpp::Swap<32, big_endian>::writeval(ip + 2, insn3); |
| |
| // Calculate tprel value. |
| Output_segment* tls_segment = relinfo->layout->tls_segment(); |
| gold_assert(tls_segment != NULL); |
| AArch64_address value = psymval->value(relinfo->object, 0); |
| const elfcpp::Elf_Xword addend = rela.get_r_addend(); |
| AArch64_address aligned_tcb_size = |
| align_address(target->tcb_size(), tls_segment->maximum_alignment()); |
| AArch64_address x = value + aligned_tcb_size; |
| |
| // After new insns are written, apply TLSLE relocs. |
| const AArch64_reloc_property* rp1 = |
| aarch64_reloc_property_table->get_reloc_property( |
| elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12); |
| const AArch64_reloc_property* rp2 = |
| aarch64_reloc_property_table->get_reloc_property( |
| elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12); |
| gold_assert(rp1 != NULL && rp2 != NULL); |
| |
| typename aarch64_reloc_funcs::Status s1 = |
| aarch64_reloc_funcs::template rela_general<32>(view + 4, |
| x, |
| addend, |
| rp1); |
| if (s1 != aarch64_reloc_funcs::STATUS_OKAY) |
| return s1; |
| |
| typename aarch64_reloc_funcs::Status s2 = |
| aarch64_reloc_funcs::template rela_general<32>(view + 8, |
| x, |
| addend, |
| rp2); |
| |
| this->skip_call_tls_get_addr_ = true; |
| return s2; |
| |
| } // End of tls_ld_to_le |
| |
| template<int size, bool big_endian> |
| inline |
| typename AArch64_relocate_functions<size, big_endian>::Status |
| Target_aarch64<size, big_endian>::Relocate::tls_ie_to_le( |
| const Relocate_info<size, big_endian>* relinfo, |
| Target_aarch64<size, big_endian>* target, |
| const elfcpp::Rela<size, big_endian>& rela, |
| unsigned int r_type, |
| unsigned char* view, |
| const Symbol_value<size>* psymval) |
| { |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address; |
| typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype; |
| typedef AArch64_relocate_functions<size, big_endian> aarch64_reloc_funcs; |
| |
| AArch64_address value = psymval->value(relinfo->object, 0); |
| Output_segment* tls_segment = relinfo->layout->tls_segment(); |
| AArch64_address aligned_tcb_address = |
| align_address(target->tcb_size(), tls_segment->maximum_alignment()); |
| const elfcpp::Elf_Xword addend = rela.get_r_addend(); |
| AArch64_address x = value + addend + aligned_tcb_address; |
| // "x" is the offset to tp, we can only do this if x is within |
| // range [0, 2^32-1] |
| if (!(size == 32 || (size == 64 && (static_cast<uint64_t>(x) >> 32) == 0))) |
| { |
| gold_error(_("TLS variable referred by reloc %u is too far from TP."), |
| r_type); |
| return aarch64_reloc_funcs::STATUS_BAD_RELOC; |
| } |
| |
| Insntype* ip = reinterpret_cast<Insntype*>(view); |
| Insntype insn = elfcpp::Swap<32, big_endian>::readval(ip); |
| unsigned int regno; |
| Insntype newinsn; |
| if (r_type == elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21) |
| { |
| // Generate movz. |
| regno = (insn & 0x1f); |
| newinsn = (0xd2a00000 | regno) | (((x >> 16) & 0xffff) << 5); |
| } |
| else if (r_type == elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC) |
| { |
| // Generate movk. |
| regno = (insn & 0x1f); |
| gold_assert(regno == ((insn >> 5) & 0x1f)); |
| newinsn = (0xf2800000 | regno) | ((x & 0xffff) << 5); |
| } |
| else |
| gold_unreachable(); |
| |
| elfcpp::Swap<32, big_endian>::writeval(ip, newinsn); |
| return aarch64_reloc_funcs::STATUS_OKAY; |
| } // End of tls_ie_to_le |
| |
| |
| template<int size, bool big_endian> |
| inline |
| typename AArch64_relocate_functions<size, big_endian>::Status |
| Target_aarch64<size, big_endian>::Relocate::tls_desc_gd_to_le( |
| const Relocate_info<size, big_endian>* relinfo, |
| Target_aarch64<size, big_endian>* target, |
| const elfcpp::Rela<size, big_endian>& rela, |
| unsigned int r_type, |
| unsigned char* view, |
| const Symbol_value<size>* psymval) |
| { |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address; |
| typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype; |
| typedef AArch64_relocate_functions<size, big_endian> aarch64_reloc_funcs; |
| |
| // TLSDESC-GD sequence is like: |
| // adrp x0, :tlsdesc:v1 |
| // ldr x1, [x0, #:tlsdesc_lo12:v1] |
| // add x0, x0, :tlsdesc_lo12:v1 |
| // .tlsdesccall v1 |
| // blr x1 |
| // After desc_gd_to_le optimization, the sequence will be like: |
| // movz x0, #0x0, lsl #16 |
| // movk x0, #0x10 |
| // nop |
| // nop |
| |
| // Calculate tprel value. |
| Output_segment* tls_segment = relinfo->layout->tls_segment(); |
| gold_assert(tls_segment != NULL); |
| Insntype* ip = reinterpret_cast<Insntype*>(view); |
| const elfcpp::Elf_Xword addend = rela.get_r_addend(); |
| AArch64_address value = psymval->value(relinfo->object, addend); |
| AArch64_address aligned_tcb_size = |
| align_address(target->tcb_size(), tls_segment->maximum_alignment()); |
| AArch64_address x = value + aligned_tcb_size; |
| // x is the offset to tp, we can only do this if x is within range |
| // [0, 2^32-1]. If x is out of range, fail and exit. |
| if (size == 64 && (static_cast<uint64_t>(x) >> 32) != 0) |
| { |
| gold_error(_("TLS variable referred by reloc %u is too far from TP. " |
| "We Can't do gd_to_le relaxation.\n"), r_type); |
| return aarch64_reloc_funcs::STATUS_BAD_RELOC; |
| } |
| Insntype newinsn; |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_TLSDESC_ADD_LO12: |
| case elfcpp::R_AARCH64_TLSDESC_CALL: |
| // Change to nop |
| newinsn = 0xd503201f; |
| break; |
| |
| case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21: |
| // Change to movz. |
| newinsn = 0xd2a00000 | (((x >> 16) & 0xffff) << 5); |
| break; |
| |
| case elfcpp::R_AARCH64_TLSDESC_LD64_LO12: |
| // Change to movk. |
| newinsn = 0xf2800000 | ((x & 0xffff) << 5); |
| break; |
| |
| default: |
| gold_error(_("unsupported tlsdesc gd_to_le optimization on reloc %u"), |
| r_type); |
| gold_unreachable(); |
| } |
| elfcpp::Swap<32, big_endian>::writeval(ip, newinsn); |
| return aarch64_reloc_funcs::STATUS_OKAY; |
| } // End of tls_desc_gd_to_le |
| |
| |
| template<int size, bool big_endian> |
| inline |
| typename AArch64_relocate_functions<size, big_endian>::Status |
| Target_aarch64<size, big_endian>::Relocate::tls_desc_gd_to_ie( |
| const Relocate_info<size, big_endian>* /* relinfo */, |
| Target_aarch64<size, big_endian>* /* target */, |
| const elfcpp::Rela<size, big_endian>& rela, |
| unsigned int r_type, |
| unsigned char* view, |
| const Symbol_value<size>* /* psymval */, |
| typename elfcpp::Elf_types<size>::Elf_Addr got_entry_address, |
| typename elfcpp::Elf_types<size>::Elf_Addr address) |
| { |
| typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype; |
| typedef AArch64_relocate_functions<size, big_endian> aarch64_reloc_funcs; |
| |
| // TLSDESC-GD sequence is like: |
| // adrp x0, :tlsdesc:v1 |
| // ldr x1, [x0, #:tlsdesc_lo12:v1] |
| // add x0, x0, :tlsdesc_lo12:v1 |
| // .tlsdesccall v1 |
| // blr x1 |
| // After desc_gd_to_ie optimization, the sequence will be like: |
| // adrp x0, :tlsie:v1 |
| // ldr x0, [x0, :tlsie_lo12:v1] |
| // nop |
| // nop |
| |
| Insntype* ip = reinterpret_cast<Insntype*>(view); |
| const elfcpp::Elf_Xword addend = rela.get_r_addend(); |
| Insntype newinsn; |
| switch (r_type) |
| { |
| case elfcpp::R_AARCH64_TLSDESC_ADD_LO12: |
| case elfcpp::R_AARCH64_TLSDESC_CALL: |
| // Change to nop |
| newinsn = 0xd503201f; |
| elfcpp::Swap<32, big_endian>::writeval(ip, newinsn); |
| break; |
| |
| case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21: |
| { |
| return aarch64_reloc_funcs::adrp(view, got_entry_address + addend, |
| address); |
| } |
| break; |
| |
| case elfcpp::R_AARCH64_TLSDESC_LD64_LO12: |
| { |
| // Set ldr target register to be x0. |
| Insntype insn = elfcpp::Swap<32, big_endian>::readval(ip); |
| insn &= 0xffffffe0; |
| elfcpp::Swap<32, big_endian>::writeval(ip, insn); |
| // Do relocation. |
| const AArch64_reloc_property* reloc_property = |
| aarch64_reloc_property_table->get_reloc_property( |
| elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC); |
| return aarch64_reloc_funcs::template rela_general<32>( |
| view, got_entry_address, addend, reloc_property); |
| } |
| break; |
| |
| default: |
| gold_error(_("Don't support tlsdesc gd_to_ie optimization on reloc %u"), |
| r_type); |
| gold_unreachable(); |
| } |
| return aarch64_reloc_funcs::STATUS_OKAY; |
| } // End of tls_desc_gd_to_ie |
| |
| // Relocate section data. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::relocate_section( |
| const Relocate_info<size, big_endian>* relinfo, |
| unsigned int sh_type, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| bool needs_special_offset_handling, |
| unsigned char* view, |
| typename elfcpp::Elf_types<size>::Elf_Addr address, |
| section_size_type view_size, |
| const Reloc_symbol_changes* reloc_symbol_changes) |
| { |
| typedef typename elfcpp::Elf_types<size>::Elf_Addr Address; |
| typedef Target_aarch64<size, big_endian> Aarch64; |
| typedef typename Target_aarch64<size, big_endian>::Relocate AArch64_relocate; |
| typedef gold::Default_classify_reloc<elfcpp::SHT_RELA, size, big_endian> |
| Classify_reloc; |
| |
| gold_assert(sh_type == elfcpp::SHT_RELA); |
| |
| // See if we are relocating a relaxed input section. If so, the view |
| // covers the whole output section and we need to adjust accordingly. |
| if (needs_special_offset_handling) |
| { |
| const Output_relaxed_input_section* poris = |
| output_section->find_relaxed_input_section(relinfo->object, |
| relinfo->data_shndx); |
| if (poris != NULL) |
| { |
| Address section_address = poris->address(); |
| section_size_type section_size = poris->data_size(); |
| |
| gold_assert((section_address >= address) |
| && ((section_address + section_size) |
| <= (address + view_size))); |
| |
| off_t offset = section_address - address; |
| view += offset; |
| address += offset; |
| view_size = section_size; |
| } |
| } |
| |
| gold::relocate_section<size, big_endian, Aarch64, AArch64_relocate, |
| gold::Default_comdat_behavior, Classify_reloc>( |
| relinfo, |
| this, |
| prelocs, |
| reloc_count, |
| output_section, |
| needs_special_offset_handling, |
| view, |
| address, |
| view_size, |
| reloc_symbol_changes); |
| } |
| |
| // Scan the relocs during a relocatable link. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::scan_relocatable_relocs( |
| Symbol_table* symtab, |
| Layout* layout, |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int data_shndx, |
| unsigned int sh_type, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| bool needs_special_offset_handling, |
| size_t local_symbol_count, |
| const unsigned char* plocal_symbols, |
| Relocatable_relocs* rr) |
| { |
| typedef gold::Default_classify_reloc<elfcpp::SHT_RELA, size, big_endian> |
| Classify_reloc; |
| typedef gold::Default_scan_relocatable_relocs<Classify_reloc> |
| Scan_relocatable_relocs; |
| |
| gold_assert(sh_type == elfcpp::SHT_RELA); |
| |
| gold::scan_relocatable_relocs<size, big_endian, Scan_relocatable_relocs>( |
| symtab, |
| layout, |
| object, |
| data_shndx, |
| prelocs, |
| reloc_count, |
| output_section, |
| needs_special_offset_handling, |
| local_symbol_count, |
| plocal_symbols, |
| rr); |
| } |
| |
| // Scan the relocs for --emit-relocs. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::emit_relocs_scan( |
| Symbol_table* symtab, |
| Layout* layout, |
| Sized_relobj_file<size, big_endian>* object, |
| unsigned int data_shndx, |
| unsigned int sh_type, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| bool needs_special_offset_handling, |
| size_t local_symbol_count, |
| const unsigned char* plocal_syms, |
| Relocatable_relocs* rr) |
| { |
| typedef gold::Default_classify_reloc<elfcpp::SHT_RELA, size, big_endian> |
| Classify_reloc; |
| typedef gold::Default_emit_relocs_strategy<Classify_reloc> |
| Emit_relocs_strategy; |
| |
| gold_assert(sh_type == elfcpp::SHT_RELA); |
| |
| gold::scan_relocatable_relocs<size, big_endian, Emit_relocs_strategy>( |
| symtab, |
| layout, |
| object, |
| data_shndx, |
| prelocs, |
| reloc_count, |
| output_section, |
| needs_special_offset_handling, |
| local_symbol_count, |
| plocal_syms, |
| rr); |
| } |
| |
| // Relocate a section during a relocatable link. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::relocate_relocs( |
| const Relocate_info<size, big_endian>* relinfo, |
| unsigned int sh_type, |
| const unsigned char* prelocs, |
| size_t reloc_count, |
| Output_section* output_section, |
| typename elfcpp::Elf_types<size>::Elf_Off offset_in_output_section, |
| unsigned char* view, |
| typename elfcpp::Elf_types<size>::Elf_Addr view_address, |
| section_size_type view_size, |
| unsigned char* reloc_view, |
| section_size_type reloc_view_size) |
| { |
| typedef gold::Default_classify_reloc<elfcpp::SHT_RELA, size, big_endian> |
| Classify_reloc; |
| |
| gold_assert(sh_type == elfcpp::SHT_RELA); |
| |
| gold::relocate_relocs<size, big_endian, Classify_reloc>( |
| relinfo, |
| prelocs, |
| reloc_count, |
| output_section, |
| offset_in_output_section, |
| view, |
| view_address, |
| view_size, |
| reloc_view, |
| reloc_view_size); |
| } |
| |
| |
| // Return whether this is a 3-insn erratum sequence. |
| |
| template<int size, bool big_endian> |
| bool |
| Target_aarch64<size, big_endian>::is_erratum_843419_sequence( |
| typename elfcpp::Swap<32,big_endian>::Valtype insn1, |
| typename elfcpp::Swap<32,big_endian>::Valtype insn2, |
| typename elfcpp::Swap<32,big_endian>::Valtype insn3) |
| { |
| unsigned rt1, rt2; |
| bool load, pair; |
| |
| // The 2nd insn is a single register load or store; or register pair |
| // store. |
| if (Insn_utilities::aarch64_mem_op_p(insn2, &rt1, &rt2, &pair, &load) |
| && (!pair || (pair && !load))) |
| { |
| // The 3rd insn is a load or store instruction from the "Load/store |
| // register (unsigned immediate)" encoding class, using Rn as the |
| // base address register. |
| if (Insn_utilities::aarch64_ldst_uimm(insn3) |
| && (Insn_utilities::aarch64_rn(insn3) |
| == Insn_utilities::aarch64_rd(insn1))) |
| return true; |
| } |
| return false; |
| } |
| |
| |
| // Return whether this is a 835769 sequence. |
| // (Similarly implemented as in elfnn-aarch64.c.) |
| |
| template<int size, bool big_endian> |
| bool |
| Target_aarch64<size, big_endian>::is_erratum_835769_sequence( |
| typename elfcpp::Swap<32,big_endian>::Valtype insn1, |
| typename elfcpp::Swap<32,big_endian>::Valtype insn2) |
| { |
| uint32_t rt; |
| uint32_t rt2 = 0; |
| uint32_t rn; |
| uint32_t rm; |
| uint32_t ra; |
| bool pair; |
| bool load; |
| |
| if (Insn_utilities::aarch64_mlxl(insn2) |
| && Insn_utilities::aarch64_mem_op_p (insn1, &rt, &rt2, &pair, &load)) |
| { |
| /* Any SIMD memory op is independent of the subsequent MLA |
| by definition of the erratum. */ |
| if (Insn_utilities::aarch64_bit(insn1, 26)) |
| return true; |
| |
| /* If not SIMD, check for integer memory ops and MLA relationship. */ |
| rn = Insn_utilities::aarch64_rn(insn2); |
| ra = Insn_utilities::aarch64_ra(insn2); |
| rm = Insn_utilities::aarch64_rm(insn2); |
| |
| /* If this is a load and there's a true(RAW) dependency, we are safe |
| and this is not an erratum sequence. */ |
| if (load && |
| (rt == rn || rt == rm || rt == ra |
| || (pair && (rt2 == rn || rt2 == rm || rt2 == ra)))) |
| return false; |
| |
| /* We conservatively put out stubs for all other cases (including |
| writebacks). */ |
| return true; |
| } |
| |
| return false; |
| } |
| |
| |
| // Helper method to create erratum stub for ST_E_843419 and ST_E_835769. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::create_erratum_stub( |
| AArch64_relobj<size, big_endian>* relobj, |
| unsigned int shndx, |
| section_size_type erratum_insn_offset, |
| Address erratum_address, |
| typename Insn_utilities::Insntype erratum_insn, |
| int erratum_type, |
| unsigned int e843419_adrp_offset) |
| { |
| gold_assert(erratum_type == ST_E_843419 || erratum_type == ST_E_835769); |
| The_stub_table* stub_table = relobj->stub_table(shndx); |
| gold_assert(stub_table != NULL); |
| if (stub_table->find_erratum_stub(relobj, |
| shndx, |
| erratum_insn_offset) == NULL) |
| { |
| const int BPI = AArch64_insn_utilities<big_endian>::BYTES_PER_INSN; |
| The_erratum_stub* stub; |
| if (erratum_type == ST_E_835769) |
| stub = new The_erratum_stub(relobj, erratum_type, shndx, |
| erratum_insn_offset); |
| else if (erratum_type == ST_E_843419) |
| stub = new E843419_stub<size, big_endian>( |
| relobj, shndx, erratum_insn_offset, e843419_adrp_offset); |
| else |
| gold_unreachable(); |
| stub->set_erratum_insn(erratum_insn); |
| stub->set_erratum_address(erratum_address); |
| // For erratum ST_E_843419 and ST_E_835769, the destination address is |
| // always the next insn after erratum insn. |
| stub->set_destination_address(erratum_address + BPI); |
| stub_table->add_erratum_stub(stub); |
| } |
| } |
| |
| |
| // Scan erratum for section SHNDX range [output_address + span_start, |
| // output_address + span_end). Note here we do not share the code with |
| // scan_erratum_843419_span function, because for 843419 we optimize by only |
| // scanning the last few insns of a page, whereas for 835769, we need to scan |
| // every insn. |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::scan_erratum_835769_span( |
| AArch64_relobj<size, big_endian>* relobj, |
| unsigned int shndx, |
| const section_size_type span_start, |
| const section_size_type span_end, |
| unsigned char* input_view, |
| Address output_address) |
| { |
| typedef typename Insn_utilities::Insntype Insntype; |
| |
| const int BPI = AArch64_insn_utilities<big_endian>::BYTES_PER_INSN; |
| |
| // Adjust output_address and view to the start of span. |
| output_address += span_start; |
| input_view += span_start; |
| |
| section_size_type span_length = span_end - span_start; |
| section_size_type offset = 0; |
| for (offset = 0; offset + BPI < span_length; offset += BPI) |
| { |
| Insntype* ip = reinterpret_cast<Insntype*>(input_view + offset); |
| Insntype insn1 = ip[0]; |
| Insntype insn2 = ip[1]; |
| if (is_erratum_835769_sequence(insn1, insn2)) |
| { |
| Insntype erratum_insn = insn2; |
| // "span_start + offset" is the offset for insn1. So for insn2, it is |
| // "span_start + offset + BPI". |
| section_size_type erratum_insn_offset = span_start + offset + BPI; |
| Address erratum_address = output_address + offset + BPI; |
| gold_info(_("Erratum 835769 found and fixed at \"%s\", " |
| "section %d, offset 0x%08x."), |
| relobj->name().c_str(), shndx, |
| (unsigned int)(span_start + offset)); |
| |
| this->create_erratum_stub(relobj, shndx, |
| erratum_insn_offset, erratum_address, |
| erratum_insn, ST_E_835769); |
| offset += BPI; // Skip mac insn. |
| } |
| } |
| } // End of "Target_aarch64::scan_erratum_835769_span". |
| |
| |
| // Scan erratum for section SHNDX range |
| // [output_address + span_start, output_address + span_end). |
| |
| template<int size, bool big_endian> |
| void |
| Target_aarch64<size, big_endian>::scan_erratum_843419_span( |
| AArch64_relobj<size, big_endian>* relobj, |
| unsigned int shndx, |
| const section_size_type span_start, |
| const section_size_type span_end, |
| unsigned char* input_view, |
| Address output_address) |
| { |
| typedef typename Insn_utilities::Insntype Insntype; |
| |
| // Adjust output_address and view to the start of span. |
| output_address += span_start; |
| input_view += span_start; |
| |
| if ((output_address & 0x03) != 0) |
| return; |
| |
| section_size_type offset = 0; |
| section_size_type span_length = span_end - span_start; |
| // The first instruction must be ending at 0xFF8 or 0xFFC. |
| unsigned int page_offset = output_address & 0xFFF; |
| // Make sure starting position, that is "output_address+offset", |
| // starts at page position 0xff8 or 0xffc. |
| if (page_offset < 0xff8) |
| offset = 0xff8 - page_offset; |
| while (offset + 3 * Insn_utilities::BYTES_PER_INSN <= span_length) |
| { |
| Insntype* ip = reinterpret_cast<Insntype*>(input_view + offset); |
| Insntype insn1 = ip[0]; |
| if (Insn_utilities::is_adrp(insn1)) |
| { |
| Insntype insn2 = ip[1]; |
| Insntype insn3 = ip[2]; |
| Insntype erratum_insn; |
| unsigned insn_offset; |
| bool do_report = false; |
| if (is_erratum_843419_sequence(insn1, insn2, insn3)) |
| { |
| do_report = true; |
| erratum_insn = insn3; |
| insn_offset = 2 * Insn_utilities::BYTES_PER_INSN; |
| } |
| else if (offset + 4 * Insn_utilities::BYTES_PER_INSN <= span_length) |
| { |
| // Optionally we can have an insn between ins2 and ins3 |
| Insntype insn_opt = ip[2]; |
| // And insn_opt must not be a branch. |
| if (!Insn_utilities::aarch64_b(insn_opt) |
| && !Insn_utilities::aarch64_bl(insn_opt) |
| && !Insn_utilities::aarch64_blr(insn_opt) |
| && !Insn_utilities::aarch64_br(insn_opt)) |
| { |
| // And insn_opt must not write to dest reg in insn1. However |
| // we do a conservative scan, which means we may fix/report |
| // more than necessary, but it doesn't hurt. |
| |
| Insntype insn4 = ip[3]; |
| if (is_erratum_843419_sequence(insn1, insn2, insn4)) |
| { |
| do_report = true; |
| erratum_insn = insn4; |
| insn_offset = 3 * Insn_utilities::BYTES_PER_INSN; |
| } |
| } |
| } |
| if (do_report) |
| { |
| unsigned int erratum_insn_offset = |
| span_start + offset + insn_offset; |
| Address erratum_address = |
| output_address + offset + insn_offset; |
| create_erratum_stub(relobj, shndx, |
| erratum_insn_offset, erratum_address, |
| erratum_insn, ST_E_843419, |
| span_start + offset); |
| } |
| } |
| |
| // Advance to next candidate instruction. We only consider instruction |
| // sequences starting at a page offset of 0xff8 or 0xffc. |
| page_offset = (output_address + offset) & 0xfff; |
| if (page_offset == 0xff8) |
| offset += 4; |
| else // (page_offset == 0xffc), we move to next page's 0xff8. |
| offset += 0xffc; |
| } |
| } // End of "Target_aarch64::scan_erratum_843419_span". |
| |
| |
| // The selector for aarch64 object files. |
| |
| template<int size, bool big_endian> |
| class Target_selector_aarch64 : public Target_selector |
| { |
| public: |
| Target_selector_aarch64(); |
| |
| virtual Target* |
| do_instantiate_target() |
| { return new Target_aarch64<size, big_endian>(); } |
| }; |
| |
| template<> |
| Target_selector_aarch64<32, true>::Target_selector_aarch64() |
| : Target_selector(elfcpp::EM_AARCH64, 32, true, |
| "elf32-bigaarch64", "aarch64_elf32_be_vec") |
| { } |
| |
| template<> |
| Target_selector_aarch64<32, false>::Target_selector_aarch64() |
| : Target_selector(elfcpp::EM_AARCH64, 32, false, |
| "elf32-littleaarch64", "aarch64_elf32_le_vec") |
| { } |
| |
| template<> |
| Target_selector_aarch64<64, true>::Target_selector_aarch64() |
| : Target_selector(elfcpp::EM_AARCH64, 64, true, |
| "elf64-bigaarch64", "aarch64_elf64_be_vec") |
| { } |
| |
| template<> |
| Target_selector_aarch64<64, false>::Target_selector_aarch64() |
| : Target_selector(elfcpp::EM_AARCH64, 64, false, |
| "elf64-littleaarch64", "aarch64_elf64_le_vec") |
| { } |
| |
| Target_selector_aarch64<32, true> target_selector_aarch64elf32b; |
| Target_selector_aarch64<32, false> target_selector_aarch64elf32; |
| Target_selector_aarch64<64, true> target_selector_aarch64elfb; |
| Target_selector_aarch64<64, false> target_selector_aarch64elf; |
| |
| } // End anonymous namespace. |