mirror of
https://sourceware.org/git/binutils-gdb.git
synced 2024-11-21 01:12:32 +08:00
fd67aa1129
Adds two new external authors to etc/update-copyright.py to cover bfd/ax_tls.m4, and adds gprofng to dirs handled automatically, then updates copyright messages as follows: 1) Update cgen/utils.scm emitted copyrights. 2) Run "etc/update-copyright.py --this-year" with an extra external author I haven't committed, 'Kalray SA.', to cover gas testsuite files (which should have their copyright message removed). 3) Build with --enable-maintainer-mode --enable-cgen-maint=yes. 4) Check out */po/*.pot which we don't update frequently.
8629 lines
270 KiB
C++
8629 lines
270 KiB
C++
// aarch64.cc -- aarch64 target support for gold.
|
|
|
|
// Copyright (C) 2014-2024 Free Software Foundation, Inc.
|
|
// Written by Jing Yu <jingyu@google.com> and Han Shen <shenhan@google.com>.
|
|
|
|
// This file is part of gold.
|
|
|
|
// This program is free software; you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation; either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with this program; if not, write to the Free Software
|
|
// Foundation, Inc., 51 Franklin Street - Fifth Floor, Boston,
|
|
// MA 02110-1301, USA.
|
|
|
|
#include "gold.h"
|
|
|
|
#include <cstring>
|
|
#include <map>
|
|
#include <set>
|
|
|
|
#include "elfcpp.h"
|
|
#include "dwarf.h"
|
|
#include "parameters.h"
|
|
#include "reloc.h"
|
|
#include "aarch64.h"
|
|
#include "object.h"
|
|
#include "symtab.h"
|
|
#include "layout.h"
|
|
#include "output.h"
|
|
#include "copy-relocs.h"
|
|
#include "target.h"
|
|
#include "target-reloc.h"
|
|
#include "target-select.h"
|
|
#include "tls.h"
|
|
#include "freebsd.h"
|
|
#include "nacl.h"
|
|
#include "gc.h"
|
|
#include "icf.h"
|
|
#include "aarch64-reloc-property.h"
|
|
|
|
// The first three .got.plt entries are reserved.
|
|
const int32_t AARCH64_GOTPLT_RESERVE_COUNT = 3;
|
|
|
|
|
|
namespace
|
|
{
|
|
|
|
using namespace gold;
|
|
|
|
template<int size, bool big_endian>
|
|
class Output_data_plt_aarch64;
|
|
|
|
template<int size, bool big_endian>
|
|
class Output_data_plt_aarch64_standard;
|
|
|
|
template<int size, bool big_endian>
|
|
class Target_aarch64;
|
|
|
|
template<int size, bool big_endian>
|
|
class AArch64_relocate_functions;
|
|
|
|
// Utility class dealing with insns. This is ported from macros in
|
|
// bfd/elfnn-aarch64.cc, but wrapped inside a class as static members. This
|
|
// class is used in erratum sequence scanning.
|
|
|
|
template<bool big_endian>
|
|
class AArch64_insn_utilities
|
|
{
|
|
public:
|
|
typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype;
|
|
|
|
static const int BYTES_PER_INSN;
|
|
|
|
// Zero register encoding - 31.
|
|
static const unsigned int AARCH64_ZR;
|
|
|
|
static unsigned int
|
|
aarch64_bit(Insntype insn, int pos)
|
|
{ return ((1 << pos) & insn) >> pos; }
|
|
|
|
static unsigned int
|
|
aarch64_bits(Insntype insn, int pos, int l)
|
|
{ return (insn >> pos) & ((1 << l) - 1); }
|
|
|
|
// Get the encoding field "op31" of 3-source data processing insns. "op31" is
|
|
// the name defined in armv8 insn manual C3.5.9.
|
|
static unsigned int
|
|
aarch64_op31(Insntype insn)
|
|
{ return aarch64_bits(insn, 21, 3); }
|
|
|
|
// Get the encoding field "ra" of 3-source data processing insns. "ra" is the
|
|
// third source register. See armv8 insn manual C3.5.9.
|
|
static unsigned int
|
|
aarch64_ra(Insntype insn)
|
|
{ return aarch64_bits(insn, 10, 5); }
|
|
|
|
static bool
|
|
is_adr(const Insntype insn)
|
|
{ return (insn & 0x9F000000) == 0x10000000; }
|
|
|
|
static bool
|
|
is_adrp(const Insntype insn)
|
|
{ return (insn & 0x9F000000) == 0x90000000; }
|
|
|
|
static bool
|
|
is_mrs_tpidr_el0(const Insntype insn)
|
|
{ return (insn & 0xFFFFFFE0) == 0xd53bd040; }
|
|
|
|
static unsigned int
|
|
aarch64_rm(const Insntype insn)
|
|
{ return aarch64_bits(insn, 16, 5); }
|
|
|
|
static unsigned int
|
|
aarch64_rn(const Insntype insn)
|
|
{ return aarch64_bits(insn, 5, 5); }
|
|
|
|
static unsigned int
|
|
aarch64_rd(const Insntype insn)
|
|
{ return aarch64_bits(insn, 0, 5); }
|
|
|
|
static unsigned int
|
|
aarch64_rt(const Insntype insn)
|
|
{ return aarch64_bits(insn, 0, 5); }
|
|
|
|
static unsigned int
|
|
aarch64_rt2(const Insntype insn)
|
|
{ return aarch64_bits(insn, 10, 5); }
|
|
|
|
// Encode imm21 into adr. Signed imm21 is in the range of [-1M, 1M).
|
|
static Insntype
|
|
aarch64_adr_encode_imm(Insntype adr, int imm21)
|
|
{
|
|
gold_assert(is_adr(adr));
|
|
gold_assert(-(1 << 20) <= imm21 && imm21 < (1 << 20));
|
|
const int mask19 = (1 << 19) - 1;
|
|
const int mask2 = 3;
|
|
adr &= ~((mask19 << 5) | (mask2 << 29));
|
|
adr |= ((imm21 & mask2) << 29) | (((imm21 >> 2) & mask19) << 5);
|
|
return adr;
|
|
}
|
|
|
|
// Retrieve encoded adrp 33-bit signed imm value. This value is obtained by
|
|
// 21-bit signed imm encoded in the insn multiplied by 4k (page size) and
|
|
// 64-bit sign-extended, resulting in [-4G, 4G) with 12-lsb being 0.
|
|
static int64_t
|
|
aarch64_adrp_decode_imm(const Insntype adrp)
|
|
{
|
|
const int mask19 = (1 << 19) - 1;
|
|
const int mask2 = 3;
|
|
gold_assert(is_adrp(adrp));
|
|
// 21-bit imm encoded in adrp.
|
|
uint64_t imm = ((adrp >> 29) & mask2) | (((adrp >> 5) & mask19) << 2);
|
|
// Retrieve msb of 21-bit-signed imm for sign extension.
|
|
uint64_t msbt = (imm >> 20) & 1;
|
|
// Real value is imm multiplied by 4k. Value now has 33-bit information.
|
|
int64_t value = imm << 12;
|
|
// Sign extend to 64-bit by repeating msbt 31 (64-33) times and merge it
|
|
// with value.
|
|
return ((((uint64_t)(1) << 32) - msbt) << 33) | value;
|
|
}
|
|
|
|
static bool
|
|
aarch64_b(const Insntype insn)
|
|
{ return (insn & 0xFC000000) == 0x14000000; }
|
|
|
|
static bool
|
|
aarch64_bl(const Insntype insn)
|
|
{ return (insn & 0xFC000000) == 0x94000000; }
|
|
|
|
static bool
|
|
aarch64_blr(const Insntype insn)
|
|
{ return (insn & 0xFFFFFC1F) == 0xD63F0000; }
|
|
|
|
static bool
|
|
aarch64_br(const Insntype insn)
|
|
{ return (insn & 0xFFFFFC1F) == 0xD61F0000; }
|
|
|
|
// All ld/st ops. See C4-182 of the ARM ARM. The encoding space for
|
|
// LD_PCREL, LDST_RO, LDST_UI and LDST_UIMM cover prefetch ops.
|
|
static bool
|
|
aarch64_ld(Insntype insn) { return aarch64_bit(insn, 22) == 1; }
|
|
|
|
static bool
|
|
aarch64_ldst(Insntype insn)
|
|
{ return (insn & 0x0a000000) == 0x08000000; }
|
|
|
|
static bool
|
|
aarch64_ldst_ex(Insntype insn)
|
|
{ return (insn & 0x3f000000) == 0x08000000; }
|
|
|
|
static bool
|
|
aarch64_ldst_pcrel(Insntype insn)
|
|
{ return (insn & 0x3b000000) == 0x18000000; }
|
|
|
|
static bool
|
|
aarch64_ldst_nap(Insntype insn)
|
|
{ return (insn & 0x3b800000) == 0x28000000; }
|
|
|
|
static bool
|
|
aarch64_ldstp_pi(Insntype insn)
|
|
{ return (insn & 0x3b800000) == 0x28800000; }
|
|
|
|
static bool
|
|
aarch64_ldstp_o(Insntype insn)
|
|
{ return (insn & 0x3b800000) == 0x29000000; }
|
|
|
|
static bool
|
|
aarch64_ldstp_pre(Insntype insn)
|
|
{ return (insn & 0x3b800000) == 0x29800000; }
|
|
|
|
static bool
|
|
aarch64_ldst_ui(Insntype insn)
|
|
{ return (insn & 0x3b200c00) == 0x38000000; }
|
|
|
|
static bool
|
|
aarch64_ldst_piimm(Insntype insn)
|
|
{ return (insn & 0x3b200c00) == 0x38000400; }
|
|
|
|
static bool
|
|
aarch64_ldst_u(Insntype insn)
|
|
{ return (insn & 0x3b200c00) == 0x38000800; }
|
|
|
|
static bool
|
|
aarch64_ldst_preimm(Insntype insn)
|
|
{ return (insn & 0x3b200c00) == 0x38000c00; }
|
|
|
|
static bool
|
|
aarch64_ldst_ro(Insntype insn)
|
|
{ return (insn & 0x3b200c00) == 0x38200800; }
|
|
|
|
static bool
|
|
aarch64_ldst_uimm(Insntype insn)
|
|
{ return (insn & 0x3b000000) == 0x39000000; }
|
|
|
|
static bool
|
|
aarch64_ldst_simd_m(Insntype insn)
|
|
{ return (insn & 0xbfbf0000) == 0x0c000000; }
|
|
|
|
static bool
|
|
aarch64_ldst_simd_m_pi(Insntype insn)
|
|
{ return (insn & 0xbfa00000) == 0x0c800000; }
|
|
|
|
static bool
|
|
aarch64_ldst_simd_s(Insntype insn)
|
|
{ return (insn & 0xbf9f0000) == 0x0d000000; }
|
|
|
|
static bool
|
|
aarch64_ldst_simd_s_pi(Insntype insn)
|
|
{ return (insn & 0xbf800000) == 0x0d800000; }
|
|
|
|
// Classify an INSN if it is indeed a load/store. Return true if INSN is a
|
|
// LD/ST instruction otherwise return false. For scalar LD/ST instructions
|
|
// PAIR is FALSE, RT is returned and RT2 is set equal to RT. For LD/ST pair
|
|
// instructions PAIR is TRUE, RT and RT2 are returned.
|
|
static bool
|
|
aarch64_mem_op_p(Insntype insn, unsigned int *rt, unsigned int *rt2,
|
|
bool *pair, bool *load)
|
|
{
|
|
uint32_t opcode;
|
|
unsigned int r;
|
|
uint32_t opc = 0;
|
|
uint32_t v = 0;
|
|
uint32_t opc_v = 0;
|
|
|
|
/* Bail out quickly if INSN doesn't fall into the load-store
|
|
encoding space. */
|
|
if (!aarch64_ldst (insn))
|
|
return false;
|
|
|
|
*pair = false;
|
|
*load = false;
|
|
if (aarch64_ldst_ex (insn))
|
|
{
|
|
*rt = aarch64_rt (insn);
|
|
*rt2 = *rt;
|
|
if (aarch64_bit (insn, 21) == 1)
|
|
{
|
|
*pair = true;
|
|
*rt2 = aarch64_rt2 (insn);
|
|
}
|
|
*load = aarch64_ld (insn);
|
|
return true;
|
|
}
|
|
else if (aarch64_ldst_nap (insn)
|
|
|| aarch64_ldstp_pi (insn)
|
|
|| aarch64_ldstp_o (insn)
|
|
|| aarch64_ldstp_pre (insn))
|
|
{
|
|
*pair = true;
|
|
*rt = aarch64_rt (insn);
|
|
*rt2 = aarch64_rt2 (insn);
|
|
*load = aarch64_ld (insn);
|
|
return true;
|
|
}
|
|
else if (aarch64_ldst_pcrel (insn)
|
|
|| aarch64_ldst_ui (insn)
|
|
|| aarch64_ldst_piimm (insn)
|
|
|| aarch64_ldst_u (insn)
|
|
|| aarch64_ldst_preimm (insn)
|
|
|| aarch64_ldst_ro (insn)
|
|
|| aarch64_ldst_uimm (insn))
|
|
{
|
|
*rt = aarch64_rt (insn);
|
|
*rt2 = *rt;
|
|
if (aarch64_ldst_pcrel (insn))
|
|
*load = true;
|
|
opc = aarch64_bits (insn, 22, 2);
|
|
v = aarch64_bit (insn, 26);
|
|
opc_v = opc | (v << 2);
|
|
*load = (opc_v == 1 || opc_v == 2 || opc_v == 3
|
|
|| opc_v == 5 || opc_v == 7);
|
|
return true;
|
|
}
|
|
else if (aarch64_ldst_simd_m (insn)
|
|
|| aarch64_ldst_simd_m_pi (insn))
|
|
{
|
|
*rt = aarch64_rt (insn);
|
|
*load = aarch64_bit (insn, 22);
|
|
opcode = (insn >> 12) & 0xf;
|
|
switch (opcode)
|
|
{
|
|
case 0:
|
|
case 2:
|
|
*rt2 = *rt + 3;
|
|
break;
|
|
|
|
case 4:
|
|
case 6:
|
|
*rt2 = *rt + 2;
|
|
break;
|
|
|
|
case 7:
|
|
*rt2 = *rt;
|
|
break;
|
|
|
|
case 8:
|
|
case 10:
|
|
*rt2 = *rt + 1;
|
|
break;
|
|
|
|
default:
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
else if (aarch64_ldst_simd_s (insn)
|
|
|| aarch64_ldst_simd_s_pi (insn))
|
|
{
|
|
*rt = aarch64_rt (insn);
|
|
r = (insn >> 21) & 1;
|
|
*load = aarch64_bit (insn, 22);
|
|
opcode = (insn >> 13) & 0x7;
|
|
switch (opcode)
|
|
{
|
|
case 0:
|
|
case 2:
|
|
case 4:
|
|
*rt2 = *rt + r;
|
|
break;
|
|
|
|
case 1:
|
|
case 3:
|
|
case 5:
|
|
*rt2 = *rt + (r == 0 ? 2 : 3);
|
|
break;
|
|
|
|
case 6:
|
|
*rt2 = *rt + r;
|
|
break;
|
|
|
|
case 7:
|
|
*rt2 = *rt + (r == 0 ? 2 : 3);
|
|
break;
|
|
|
|
default:
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
return false;
|
|
} // End of "aarch64_mem_op_p".
|
|
|
|
// Return true if INSN is mac insn.
|
|
static bool
|
|
aarch64_mac(Insntype insn)
|
|
{ return (insn & 0xff000000) == 0x9b000000; }
|
|
|
|
// Return true if INSN is multiply-accumulate.
|
|
// (This is similar to implementaton in elfnn-aarch64.c.)
|
|
static bool
|
|
aarch64_mlxl(Insntype insn)
|
|
{
|
|
uint32_t op31 = aarch64_op31(insn);
|
|
if (aarch64_mac(insn)
|
|
&& (op31 == 0 || op31 == 1 || op31 == 5)
|
|
/* Exclude MUL instructions which are encoded as a multiple-accumulate
|
|
with RA = XZR. */
|
|
&& aarch64_ra(insn) != AARCH64_ZR)
|
|
{
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
}; // End of "AArch64_insn_utilities".
|
|
|
|
|
|
// Insn length in byte.
|
|
|
|
template<bool big_endian>
|
|
const int AArch64_insn_utilities<big_endian>::BYTES_PER_INSN = 4;
|
|
|
|
|
|
// Zero register encoding - 31.
|
|
|
|
template<bool big_endian>
|
|
const unsigned int AArch64_insn_utilities<big_endian>::AARCH64_ZR = 0x1f;
|
|
|
|
|
|
// Output_data_got_aarch64 class.
|
|
|
|
template<int size, bool big_endian>
|
|
class Output_data_got_aarch64 : public Output_data_got<size, big_endian>
|
|
{
|
|
public:
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr Valtype;
|
|
Output_data_got_aarch64(Symbol_table* symtab, Layout* layout)
|
|
: Output_data_got<size, big_endian>(),
|
|
symbol_table_(symtab), layout_(layout)
|
|
{ }
|
|
|
|
// Add a static entry for the GOT entry at OFFSET. GSYM is a global
|
|
// symbol and R_TYPE is the code of a dynamic relocation that needs to be
|
|
// applied in a static link.
|
|
void
|
|
add_static_reloc(unsigned int got_offset, unsigned int r_type, Symbol* gsym)
|
|
{ this->static_relocs_.push_back(Static_reloc(got_offset, r_type, gsym)); }
|
|
|
|
|
|
// Add a static reloc for the GOT entry at OFFSET. RELOBJ is an object
|
|
// defining a local symbol with INDEX. R_TYPE is the code of a dynamic
|
|
// relocation that needs to be applied in a static link.
|
|
void
|
|
add_static_reloc(unsigned int got_offset, unsigned int r_type,
|
|
Sized_relobj_file<size, big_endian>* relobj,
|
|
unsigned int index)
|
|
{
|
|
this->static_relocs_.push_back(Static_reloc(got_offset, r_type, relobj,
|
|
index));
|
|
}
|
|
|
|
|
|
protected:
|
|
// Write out the GOT table.
|
|
void
|
|
do_write(Output_file* of) {
|
|
// The first entry in the GOT is the address of the .dynamic section.
|
|
gold_assert(this->data_size() >= size / 8);
|
|
Output_section* dynamic = this->layout_->dynamic_section();
|
|
Valtype dynamic_addr = dynamic == NULL ? 0 : dynamic->address();
|
|
this->replace_constant(0, dynamic_addr);
|
|
Output_data_got<size, big_endian>::do_write(of);
|
|
|
|
// Handling static relocs
|
|
if (this->static_relocs_.empty())
|
|
return;
|
|
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address;
|
|
|
|
gold_assert(parameters->doing_static_link());
|
|
const off_t offset = this->offset();
|
|
const section_size_type oview_size =
|
|
convert_to_section_size_type(this->data_size());
|
|
unsigned char* const oview = of->get_output_view(offset, oview_size);
|
|
|
|
Output_segment* tls_segment = this->layout_->tls_segment();
|
|
gold_assert(tls_segment != NULL);
|
|
|
|
AArch64_address aligned_tcb_address =
|
|
align_address(Target_aarch64<size, big_endian>::TCB_SIZE,
|
|
tls_segment->maximum_alignment());
|
|
|
|
for (size_t i = 0; i < this->static_relocs_.size(); ++i)
|
|
{
|
|
Static_reloc& reloc(this->static_relocs_[i]);
|
|
AArch64_address value;
|
|
|
|
if (!reloc.symbol_is_global())
|
|
{
|
|
Sized_relobj_file<size, big_endian>* object = reloc.relobj();
|
|
const Symbol_value<size>* psymval =
|
|
reloc.relobj()->local_symbol(reloc.index());
|
|
|
|
// We are doing static linking. Issue an error and skip this
|
|
// relocation if the symbol is undefined or in a discarded_section.
|
|
bool is_ordinary;
|
|
unsigned int shndx = psymval->input_shndx(&is_ordinary);
|
|
if ((shndx == elfcpp::SHN_UNDEF)
|
|
|| (is_ordinary
|
|
&& shndx != elfcpp::SHN_UNDEF
|
|
&& !object->is_section_included(shndx)
|
|
&& !this->symbol_table_->is_section_folded(object, shndx)))
|
|
{
|
|
gold_error(_("undefined or discarded local symbol %u from "
|
|
" object %s in GOT"),
|
|
reloc.index(), reloc.relobj()->name().c_str());
|
|
continue;
|
|
}
|
|
value = psymval->value(object, 0);
|
|
}
|
|
else
|
|
{
|
|
const Symbol* gsym = reloc.symbol();
|
|
gold_assert(gsym != NULL);
|
|
if (gsym->is_forwarder())
|
|
gsym = this->symbol_table_->resolve_forwards(gsym);
|
|
|
|
// We are doing static linking. Issue an error and skip this
|
|
// relocation if the symbol is undefined or in a discarded_section
|
|
// unless it is a weakly_undefined symbol.
|
|
if ((gsym->is_defined_in_discarded_section()
|
|
|| gsym->is_undefined())
|
|
&& !gsym->is_weak_undefined())
|
|
{
|
|
gold_error(_("undefined or discarded symbol %s in GOT"),
|
|
gsym->name());
|
|
continue;
|
|
}
|
|
|
|
if (!gsym->is_weak_undefined())
|
|
{
|
|
const Sized_symbol<size>* sym =
|
|
static_cast<const Sized_symbol<size>*>(gsym);
|
|
value = sym->value();
|
|
}
|
|
else
|
|
value = 0;
|
|
}
|
|
|
|
unsigned got_offset = reloc.got_offset();
|
|
gold_assert(got_offset < oview_size);
|
|
|
|
typedef typename elfcpp::Swap<size, big_endian>::Valtype Valtype;
|
|
Valtype* wv = reinterpret_cast<Valtype*>(oview + got_offset);
|
|
Valtype x;
|
|
switch (reloc.r_type())
|
|
{
|
|
case elfcpp::R_AARCH64_TLS_DTPREL64:
|
|
x = value;
|
|
break;
|
|
case elfcpp::R_AARCH64_TLS_TPREL64:
|
|
x = value + aligned_tcb_address;
|
|
break;
|
|
default:
|
|
gold_unreachable();
|
|
}
|
|
elfcpp::Swap<size, big_endian>::writeval(wv, x);
|
|
}
|
|
|
|
of->write_output_view(offset, oview_size, oview);
|
|
}
|
|
|
|
private:
|
|
// Symbol table of the output object.
|
|
Symbol_table* symbol_table_;
|
|
// A pointer to the Layout class, so that we can find the .dynamic
|
|
// section when we write out the GOT section.
|
|
Layout* layout_;
|
|
|
|
// This class represent dynamic relocations that need to be applied by
|
|
// gold because we are using TLS relocations in a static link.
|
|
class Static_reloc
|
|
{
|
|
public:
|
|
Static_reloc(unsigned int got_offset, unsigned int r_type, Symbol* gsym)
|
|
: got_offset_(got_offset), r_type_(r_type), symbol_is_global_(true)
|
|
{ this->u_.global.symbol = gsym; }
|
|
|
|
Static_reloc(unsigned int got_offset, unsigned int r_type,
|
|
Sized_relobj_file<size, big_endian>* relobj, unsigned int index)
|
|
: got_offset_(got_offset), r_type_(r_type), symbol_is_global_(false)
|
|
{
|
|
this->u_.local.relobj = relobj;
|
|
this->u_.local.index = index;
|
|
}
|
|
|
|
// Return the GOT offset.
|
|
unsigned int
|
|
got_offset() const
|
|
{ return this->got_offset_; }
|
|
|
|
// Relocation type.
|
|
unsigned int
|
|
r_type() const
|
|
{ return this->r_type_; }
|
|
|
|
// Whether the symbol is global or not.
|
|
bool
|
|
symbol_is_global() const
|
|
{ return this->symbol_is_global_; }
|
|
|
|
// For a relocation against a global symbol, the global symbol.
|
|
Symbol*
|
|
symbol() const
|
|
{
|
|
gold_assert(this->symbol_is_global_);
|
|
return this->u_.global.symbol;
|
|
}
|
|
|
|
// For a relocation against a local symbol, the defining object.
|
|
Sized_relobj_file<size, big_endian>*
|
|
relobj() const
|
|
{
|
|
gold_assert(!this->symbol_is_global_);
|
|
return this->u_.local.relobj;
|
|
}
|
|
|
|
// For a relocation against a local symbol, the local symbol index.
|
|
unsigned int
|
|
index() const
|
|
{
|
|
gold_assert(!this->symbol_is_global_);
|
|
return this->u_.local.index;
|
|
}
|
|
|
|
private:
|
|
// GOT offset of the entry to which this relocation is applied.
|
|
unsigned int got_offset_;
|
|
// Type of relocation.
|
|
unsigned int r_type_;
|
|
// Whether this relocation is against a global symbol.
|
|
bool symbol_is_global_;
|
|
// A global or local symbol.
|
|
union
|
|
{
|
|
struct
|
|
{
|
|
// For a global symbol, the symbol itself.
|
|
Symbol* symbol;
|
|
} global;
|
|
struct
|
|
{
|
|
// For a local symbol, the object defining the symbol.
|
|
Sized_relobj_file<size, big_endian>* relobj;
|
|
// For a local symbol, the symbol index.
|
|
unsigned int index;
|
|
} local;
|
|
} u_;
|
|
}; // End of inner class Static_reloc
|
|
|
|
std::vector<Static_reloc> static_relocs_;
|
|
}; // End of Output_data_got_aarch64
|
|
|
|
|
|
template<int size, bool big_endian>
|
|
class AArch64_input_section;
|
|
|
|
|
|
template<int size, bool big_endian>
|
|
class AArch64_output_section;
|
|
|
|
|
|
template<int size, bool big_endian>
|
|
class AArch64_relobj;
|
|
|
|
|
|
// Stub type enum constants.
|
|
|
|
enum
|
|
{
|
|
ST_NONE = 0,
|
|
|
|
// Using adrp/add pair, 4 insns (including alignment) without mem access,
|
|
// the fastest stub. This has a limited jump distance, which is tested by
|
|
// aarch64_valid_for_adrp_p.
|
|
ST_ADRP_BRANCH = 1,
|
|
|
|
// Using ldr-absolute-address/br-register, 4 insns with 1 mem access,
|
|
// unlimited in jump distance.
|
|
ST_LONG_BRANCH_ABS = 2,
|
|
|
|
// Using ldr/calculate-pcrel/jump, 8 insns (including alignment) with 1
|
|
// mem access, slowest one. Only used in position independent executables.
|
|
ST_LONG_BRANCH_PCREL = 3,
|
|
|
|
// Stub for erratum 843419 handling.
|
|
ST_E_843419 = 4,
|
|
|
|
// Stub for erratum 835769 handling.
|
|
ST_E_835769 = 5,
|
|
|
|
// Number of total stub types.
|
|
ST_NUMBER = 6
|
|
};
|
|
|
|
|
|
// Struct that wraps insns for a particular stub. All stub templates are
|
|
// created/initialized as constants by Stub_template_repertoire.
|
|
|
|
template<bool big_endian>
|
|
struct Stub_template
|
|
{
|
|
const typename AArch64_insn_utilities<big_endian>::Insntype* insns;
|
|
const int insn_num;
|
|
};
|
|
|
|
|
|
// Simple singleton class that creates/initializes/stores all types of stub
|
|
// templates.
|
|
|
|
template<bool big_endian>
|
|
class Stub_template_repertoire
|
|
{
|
|
public:
|
|
typedef typename AArch64_insn_utilities<big_endian>::Insntype Insntype;
|
|
|
|
// Single static method to get stub template for a given stub type.
|
|
static const Stub_template<big_endian>*
|
|
get_stub_template(int type)
|
|
{
|
|
static Stub_template_repertoire<big_endian> singleton;
|
|
return singleton.stub_templates_[type];
|
|
}
|
|
|
|
private:
|
|
// Constructor - creates/initializes all stub templates.
|
|
Stub_template_repertoire();
|
|
~Stub_template_repertoire()
|
|
{ }
|
|
|
|
// Disallowing copy ctor and copy assignment operator.
|
|
Stub_template_repertoire(Stub_template_repertoire&);
|
|
Stub_template_repertoire& operator=(Stub_template_repertoire&);
|
|
|
|
// Data that stores all insn templates.
|
|
const Stub_template<big_endian>* stub_templates_[ST_NUMBER];
|
|
}; // End of "class Stub_template_repertoire".
|
|
|
|
|
|
// Constructor - creates/initilizes all stub templates.
|
|
|
|
template<bool big_endian>
|
|
Stub_template_repertoire<big_endian>::Stub_template_repertoire()
|
|
{
|
|
// Insn array definitions.
|
|
const static Insntype ST_NONE_INSNS[] = {};
|
|
|
|
const static Insntype ST_ADRP_BRANCH_INSNS[] =
|
|
{
|
|
0x90000010, /* adrp ip0, X */
|
|
/* ADR_PREL_PG_HI21(X) */
|
|
0x91000210, /* add ip0, ip0, :lo12:X */
|
|
/* ADD_ABS_LO12_NC(X) */
|
|
0xd61f0200, /* br ip0 */
|
|
0x00000000, /* alignment padding */
|
|
};
|
|
|
|
const static Insntype ST_LONG_BRANCH_ABS_INSNS[] =
|
|
{
|
|
0x58000050, /* ldr ip0, 0x8 */
|
|
0xd61f0200, /* br ip0 */
|
|
0x00000000, /* address field */
|
|
0x00000000, /* address fields */
|
|
};
|
|
|
|
const static Insntype ST_LONG_BRANCH_PCREL_INSNS[] =
|
|
{
|
|
0x58000090, /* ldr ip0, 0x10 */
|
|
0x10000011, /* adr ip1, #0 */
|
|
0x8b110210, /* add ip0, ip0, ip1 */
|
|
0xd61f0200, /* br ip0 */
|
|
0x00000000, /* address field */
|
|
0x00000000, /* address field */
|
|
0x00000000, /* alignment padding */
|
|
0x00000000, /* alignment padding */
|
|
};
|
|
|
|
const static Insntype ST_E_843419_INSNS[] =
|
|
{
|
|
0x00000000, /* Placeholder for erratum insn. */
|
|
0x14000000, /* b <label> */
|
|
};
|
|
|
|
// ST_E_835769 has the same stub template as ST_E_843419
|
|
// but we reproduce the array here so that the sizeof
|
|
// expressions in install_insn_template will work.
|
|
const static Insntype ST_E_835769_INSNS[] =
|
|
{
|
|
0x00000000, /* Placeholder for erratum insn. */
|
|
0x14000000, /* b <label> */
|
|
};
|
|
|
|
#define install_insn_template(T) \
|
|
const static Stub_template<big_endian> template_##T = { \
|
|
T##_INSNS, sizeof(T##_INSNS) / sizeof(T##_INSNS[0]) }; \
|
|
this->stub_templates_[T] = &template_##T
|
|
|
|
install_insn_template(ST_NONE);
|
|
install_insn_template(ST_ADRP_BRANCH);
|
|
install_insn_template(ST_LONG_BRANCH_ABS);
|
|
install_insn_template(ST_LONG_BRANCH_PCREL);
|
|
install_insn_template(ST_E_843419);
|
|
install_insn_template(ST_E_835769);
|
|
|
|
#undef install_insn_template
|
|
}
|
|
|
|
|
|
// Base class for stubs.
|
|
|
|
template<int size, bool big_endian>
|
|
class Stub_base
|
|
{
|
|
public:
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address;
|
|
typedef typename AArch64_insn_utilities<big_endian>::Insntype Insntype;
|
|
|
|
static const AArch64_address invalid_address =
|
|
static_cast<AArch64_address>(-1);
|
|
|
|
static const section_offset_type invalid_offset =
|
|
static_cast<section_offset_type>(-1);
|
|
|
|
Stub_base(int type)
|
|
: destination_address_(invalid_address),
|
|
offset_(invalid_offset),
|
|
type_(type)
|
|
{}
|
|
|
|
~Stub_base()
|
|
{}
|
|
|
|
// Get stub type.
|
|
int
|
|
type() const
|
|
{ return this->type_; }
|
|
|
|
// Get stub template that provides stub insn information.
|
|
const Stub_template<big_endian>*
|
|
stub_template() const
|
|
{
|
|
return Stub_template_repertoire<big_endian>::
|
|
get_stub_template(this->type());
|
|
}
|
|
|
|
// Get destination address.
|
|
AArch64_address
|
|
destination_address() const
|
|
{
|
|
gold_assert(this->destination_address_ != this->invalid_address);
|
|
return this->destination_address_;
|
|
}
|
|
|
|
// Set destination address.
|
|
void
|
|
set_destination_address(AArch64_address address)
|
|
{
|
|
gold_assert(address != this->invalid_address);
|
|
this->destination_address_ = address;
|
|
}
|
|
|
|
// Reset the destination address.
|
|
void
|
|
reset_destination_address()
|
|
{ this->destination_address_ = this->invalid_address; }
|
|
|
|
// Get offset of code stub. For Reloc_stub, it is the offset from the
|
|
// beginning of its containing stub table; for Erratum_stub, it is the offset
|
|
// from the end of reloc_stubs.
|
|
section_offset_type
|
|
offset() const
|
|
{
|
|
gold_assert(this->offset_ != this->invalid_offset);
|
|
return this->offset_;
|
|
}
|
|
|
|
// Set stub offset.
|
|
void
|
|
set_offset(section_offset_type offset)
|
|
{ this->offset_ = offset; }
|
|
|
|
// Return the stub insn.
|
|
const Insntype*
|
|
insns() const
|
|
{ return this->stub_template()->insns; }
|
|
|
|
// Return num of stub insns.
|
|
unsigned int
|
|
insn_num() const
|
|
{ return this->stub_template()->insn_num; }
|
|
|
|
// Get size of the stub.
|
|
int
|
|
stub_size() const
|
|
{
|
|
return this->insn_num() *
|
|
AArch64_insn_utilities<big_endian>::BYTES_PER_INSN;
|
|
}
|
|
|
|
// Write stub to output file.
|
|
void
|
|
write(unsigned char* view, section_size_type view_size)
|
|
{ this->do_write(view, view_size); }
|
|
|
|
protected:
|
|
// Abstract method to be implemented by sub-classes.
|
|
virtual void
|
|
do_write(unsigned char*, section_size_type) = 0;
|
|
|
|
private:
|
|
// The last insn of a stub is a jump to destination insn. This field records
|
|
// the destination address.
|
|
AArch64_address destination_address_;
|
|
// The stub offset. Note this has difference interpretations between an
|
|
// Reloc_stub and an Erratum_stub. For Reloc_stub this is the offset from the
|
|
// beginning of the containing stub_table, whereas for Erratum_stub, this is
|
|
// the offset from the end of reloc_stubs.
|
|
section_offset_type offset_;
|
|
// Stub type.
|
|
const int type_;
|
|
}; // End of "Stub_base".
|
|
|
|
|
|
// Erratum stub class. An erratum stub differs from a reloc stub in that for
|
|
// each erratum occurrence, we generate an erratum stub. We never share erratum
|
|
// stubs, whereas for reloc stubs, different branch insns share a single reloc
|
|
// stub as long as the branch targets are the same. (More to the point, reloc
|
|
// stubs can be shared because they're used to reach a specific target, whereas
|
|
// erratum stubs branch back to the original control flow.)
|
|
|
|
template<int size, bool big_endian>
|
|
class Erratum_stub : public Stub_base<size, big_endian>
|
|
{
|
|
public:
|
|
typedef AArch64_relobj<size, big_endian> The_aarch64_relobj;
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address;
|
|
typedef AArch64_insn_utilities<big_endian> Insn_utilities;
|
|
typedef typename AArch64_insn_utilities<big_endian>::Insntype Insntype;
|
|
|
|
static const int STUB_ADDR_ALIGN;
|
|
|
|
static const Insntype invalid_insn = static_cast<Insntype>(-1);
|
|
|
|
Erratum_stub(The_aarch64_relobj* relobj, int type,
|
|
unsigned shndx, unsigned int sh_offset)
|
|
: Stub_base<size, big_endian>(type), relobj_(relobj),
|
|
shndx_(shndx), sh_offset_(sh_offset),
|
|
erratum_insn_(invalid_insn),
|
|
erratum_address_(this->invalid_address)
|
|
{}
|
|
|
|
~Erratum_stub() {}
|
|
|
|
// Return the object that contains the erratum.
|
|
The_aarch64_relobj*
|
|
relobj()
|
|
{ return this->relobj_; }
|
|
|
|
// Get section index of the erratum.
|
|
unsigned int
|
|
shndx() const
|
|
{ return this->shndx_; }
|
|
|
|
// Get section offset of the erratum.
|
|
unsigned int
|
|
sh_offset() const
|
|
{ return this->sh_offset_; }
|
|
|
|
// Get the erratum insn. This is the insn located at erratum_insn_address.
|
|
Insntype
|
|
erratum_insn() const
|
|
{
|
|
gold_assert(this->erratum_insn_ != this->invalid_insn);
|
|
return this->erratum_insn_;
|
|
}
|
|
|
|
// Set the insn that the erratum happens to.
|
|
void
|
|
set_erratum_insn(Insntype insn)
|
|
{ this->erratum_insn_ = insn; }
|
|
|
|
// For 843419, the erratum insn is ld/st xt, [xn, #uimm], which may be a
|
|
// relocation spot, in this case, the erratum_insn_ recorded at scanning phase
|
|
// is no longer the one we want to write out to the stub, update erratum_insn_
|
|
// with relocated version. Also note that in this case xn must not be "PC", so
|
|
// it is safe to move the erratum insn from the origin place to the stub. For
|
|
// 835769, the erratum insn is multiply-accumulate insn, which could not be a
|
|
// relocation spot (assertion added though).
|
|
void
|
|
update_erratum_insn(Insntype insn)
|
|
{
|
|
gold_assert(this->erratum_insn_ != this->invalid_insn);
|
|
switch (this->type())
|
|
{
|
|
case ST_E_843419:
|
|
gold_assert(Insn_utilities::aarch64_ldst_uimm(insn));
|
|
gold_assert(Insn_utilities::aarch64_ldst_uimm(this->erratum_insn()));
|
|
gold_assert(Insn_utilities::aarch64_rd(insn) ==
|
|
Insn_utilities::aarch64_rd(this->erratum_insn()));
|
|
gold_assert(Insn_utilities::aarch64_rn(insn) ==
|
|
Insn_utilities::aarch64_rn(this->erratum_insn()));
|
|
// Update plain ld/st insn with relocated insn.
|
|
this->erratum_insn_ = insn;
|
|
break;
|
|
case ST_E_835769:
|
|
gold_assert(insn == this->erratum_insn());
|
|
break;
|
|
default:
|
|
gold_unreachable();
|
|
}
|
|
}
|
|
|
|
|
|
// Return the address where an erratum must be done.
|
|
AArch64_address
|
|
erratum_address() const
|
|
{
|
|
gold_assert(this->erratum_address_ != this->invalid_address);
|
|
return this->erratum_address_;
|
|
}
|
|
|
|
// Set the address where an erratum must be done.
|
|
void
|
|
set_erratum_address(AArch64_address addr)
|
|
{ this->erratum_address_ = addr; }
|
|
|
|
// Later relaxation passes of may alter the recorded erratum and destination
|
|
// address. Given an up to date output section address of shidx_ in
|
|
// relobj_ we can derive the erratum_address and destination address.
|
|
void
|
|
update_erratum_address(AArch64_address output_section_addr)
|
|
{
|
|
const int BPI = AArch64_insn_utilities<big_endian>::BYTES_PER_INSN;
|
|
AArch64_address updated_addr = output_section_addr + this->sh_offset_;
|
|
this->set_erratum_address(updated_addr);
|
|
this->set_destination_address(updated_addr + BPI);
|
|
}
|
|
|
|
// Comparator used to group Erratum_stubs in a set by (obj, shndx,
|
|
// sh_offset). We do not include 'type' in the calculation, because there is
|
|
// at most one stub type at (obj, shndx, sh_offset).
|
|
bool
|
|
operator<(const Erratum_stub<size, big_endian>& k) const
|
|
{
|
|
if (this == &k)
|
|
return false;
|
|
// We group stubs by relobj.
|
|
if (this->relobj_ != k.relobj_)
|
|
return this->relobj_ < k.relobj_;
|
|
// Then by section index.
|
|
if (this->shndx_ != k.shndx_)
|
|
return this->shndx_ < k.shndx_;
|
|
// Lastly by section offset.
|
|
return this->sh_offset_ < k.sh_offset_;
|
|
}
|
|
|
|
void
|
|
invalidate_erratum_stub()
|
|
{
|
|
gold_assert(this->erratum_insn_ != invalid_insn);
|
|
this->erratum_insn_ = invalid_insn;
|
|
}
|
|
|
|
bool
|
|
is_invalidated_erratum_stub()
|
|
{ return this->erratum_insn_ == invalid_insn; }
|
|
|
|
protected:
|
|
virtual void
|
|
do_write(unsigned char*, section_size_type);
|
|
|
|
private:
|
|
// The object that needs to be fixed.
|
|
The_aarch64_relobj* relobj_;
|
|
// The shndx in the object that needs to be fixed.
|
|
const unsigned int shndx_;
|
|
// The section offset in the obejct that needs to be fixed.
|
|
const unsigned int sh_offset_;
|
|
// The insn to be fixed.
|
|
Insntype erratum_insn_;
|
|
// The address of the above insn.
|
|
AArch64_address erratum_address_;
|
|
}; // End of "Erratum_stub".
|
|
|
|
|
|
// Erratum sub class to wrap additional info needed by 843419. In fixing this
|
|
// erratum, we may choose to replace 'adrp' with 'adr', in this case, we need
|
|
// adrp's code position (two or three insns before erratum insn itself).
|
|
|
|
template<int size, bool big_endian>
|
|
class E843419_stub : public Erratum_stub<size, big_endian>
|
|
{
|
|
public:
|
|
typedef typename AArch64_insn_utilities<big_endian>::Insntype Insntype;
|
|
|
|
E843419_stub(AArch64_relobj<size, big_endian>* relobj,
|
|
unsigned int shndx, unsigned int sh_offset,
|
|
unsigned int adrp_sh_offset)
|
|
: Erratum_stub<size, big_endian>(relobj, ST_E_843419, shndx, sh_offset),
|
|
adrp_sh_offset_(adrp_sh_offset)
|
|
{}
|
|
|
|
unsigned int
|
|
adrp_sh_offset() const
|
|
{ return this->adrp_sh_offset_; }
|
|
|
|
private:
|
|
// Section offset of "adrp". (We do not need a "adrp_shndx_" field, because we
|
|
// can obtain it from its parent.)
|
|
const unsigned int adrp_sh_offset_;
|
|
};
|
|
|
|
|
|
template<int size, bool big_endian>
|
|
const int Erratum_stub<size, big_endian>::STUB_ADDR_ALIGN = 4;
|
|
|
|
// Comparator used in set definition.
|
|
template<int size, bool big_endian>
|
|
struct Erratum_stub_less
|
|
{
|
|
bool
|
|
operator()(const Erratum_stub<size, big_endian>* s1,
|
|
const Erratum_stub<size, big_endian>* s2) const
|
|
{ return *s1 < *s2; }
|
|
};
|
|
|
|
// Erratum_stub implementation for writing stub to output file.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Erratum_stub<size, big_endian>::do_write(unsigned char* view, section_size_type)
|
|
{
|
|
typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype;
|
|
const Insntype* insns = this->insns();
|
|
uint32_t num_insns = this->insn_num();
|
|
Insntype* ip = reinterpret_cast<Insntype*>(view);
|
|
// For current implemented erratum 843419 and 835769, the first insn in the
|
|
// stub is always a copy of the problematic insn (in 843419, the mem access
|
|
// insn, in 835769, the mac insn), followed by a jump-back.
|
|
elfcpp::Swap<32, big_endian>::writeval(ip, this->erratum_insn());
|
|
for (uint32_t i = 1; i < num_insns; ++i)
|
|
elfcpp::Swap<32, big_endian>::writeval(ip + i, insns[i]);
|
|
}
|
|
|
|
|
|
// Reloc stub class.
|
|
|
|
template<int size, bool big_endian>
|
|
class Reloc_stub : public Stub_base<size, big_endian>
|
|
{
|
|
public:
|
|
typedef Reloc_stub<size, big_endian> This;
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address;
|
|
|
|
// Branch range. This is used to calculate the section group size, as well as
|
|
// determine whether a stub is needed.
|
|
static const int MAX_BRANCH_OFFSET = ((1 << 25) - 1) << 2;
|
|
static const int MIN_BRANCH_OFFSET = -((1 << 25) << 2);
|
|
|
|
// Constant used to determine if an offset fits in the adrp instruction
|
|
// encoding.
|
|
static const int MAX_ADRP_IMM = (1 << 20) - 1;
|
|
static const int MIN_ADRP_IMM = -(1 << 20);
|
|
|
|
static const int BYTES_PER_INSN = 4;
|
|
static const int STUB_ADDR_ALIGN;
|
|
|
|
// Determine whether the offset fits in the jump/branch instruction.
|
|
static bool
|
|
aarch64_valid_branch_offset_p(int64_t offset)
|
|
{ return offset >= MIN_BRANCH_OFFSET && offset <= MAX_BRANCH_OFFSET; }
|
|
|
|
// Determine whether the offset fits in the adrp immediate field.
|
|
static bool
|
|
aarch64_valid_for_adrp_p(AArch64_address location, AArch64_address dest)
|
|
{
|
|
typedef AArch64_relocate_functions<size, big_endian> Reloc;
|
|
int64_t adrp_imm = Reloc::Page (dest) - Reloc::Page (location);
|
|
adrp_imm = adrp_imm < 0 ? ~(~adrp_imm >> 12) : adrp_imm >> 12;
|
|
return adrp_imm >= MIN_ADRP_IMM && adrp_imm <= MAX_ADRP_IMM;
|
|
}
|
|
|
|
// Determine the stub type for a certain relocation or ST_NONE, if no stub is
|
|
// needed.
|
|
static int
|
|
stub_type_for_reloc(unsigned int r_type, AArch64_address address,
|
|
AArch64_address target);
|
|
|
|
Reloc_stub(int type)
|
|
: Stub_base<size, big_endian>(type)
|
|
{ }
|
|
|
|
~Reloc_stub()
|
|
{ }
|
|
|
|
// The key class used to index the stub instance in the stub table's stub map.
|
|
class Key
|
|
{
|
|
public:
|
|
Key(int type, const Symbol* symbol, const Relobj* relobj,
|
|
unsigned int r_sym, int32_t addend)
|
|
: type_(type), addend_(addend)
|
|
{
|
|
if (symbol != NULL)
|
|
{
|
|
this->r_sym_ = Reloc_stub::invalid_index;
|
|
this->u_.symbol = symbol;
|
|
}
|
|
else
|
|
{
|
|
gold_assert(relobj != NULL && r_sym != invalid_index);
|
|
this->r_sym_ = r_sym;
|
|
this->u_.relobj = relobj;
|
|
}
|
|
}
|
|
|
|
~Key()
|
|
{ }
|
|
|
|
// Return stub type.
|
|
int
|
|
type() const
|
|
{ return this->type_; }
|
|
|
|
// Return the local symbol index or invalid_index.
|
|
unsigned int
|
|
r_sym() const
|
|
{ return this->r_sym_; }
|
|
|
|
// Return the symbol if there is one.
|
|
const Symbol*
|
|
symbol() const
|
|
{ return this->r_sym_ == invalid_index ? this->u_.symbol : NULL; }
|
|
|
|
// Return the relobj if there is one.
|
|
const Relobj*
|
|
relobj() const
|
|
{ return this->r_sym_ != invalid_index ? this->u_.relobj : NULL; }
|
|
|
|
// Whether this equals to another key k.
|
|
bool
|
|
eq(const Key& k) const
|
|
{
|
|
return ((this->type_ == k.type_)
|
|
&& (this->r_sym_ == k.r_sym_)
|
|
&& ((this->r_sym_ != Reloc_stub::invalid_index)
|
|
? (this->u_.relobj == k.u_.relobj)
|
|
: (this->u_.symbol == k.u_.symbol))
|
|
&& (this->addend_ == k.addend_));
|
|
}
|
|
|
|
// Return a hash value.
|
|
size_t
|
|
hash_value() const
|
|
{
|
|
size_t name_hash_value = gold::string_hash<char>(
|
|
(this->r_sym_ != Reloc_stub::invalid_index)
|
|
? this->u_.relobj->name().c_str()
|
|
: this->u_.symbol->name());
|
|
// We only have 4 stub types.
|
|
size_t stub_type_hash_value = 0x03 & this->type_;
|
|
return (name_hash_value
|
|
^ stub_type_hash_value
|
|
^ ((this->r_sym_ & 0x3fff) << 2)
|
|
^ ((this->addend_ & 0xffff) << 16));
|
|
}
|
|
|
|
// Functors for STL associative containers.
|
|
struct hash
|
|
{
|
|
size_t
|
|
operator()(const Key& k) const
|
|
{ return k.hash_value(); }
|
|
};
|
|
|
|
struct equal_to
|
|
{
|
|
bool
|
|
operator()(const Key& k1, const Key& k2) const
|
|
{ return k1.eq(k2); }
|
|
};
|
|
|
|
private:
|
|
// Stub type.
|
|
const int type_;
|
|
// If this is a local symbol, this is the index in the defining object.
|
|
// Otherwise, it is invalid_index for a global symbol.
|
|
unsigned int r_sym_;
|
|
// If r_sym_ is an invalid index, this points to a global symbol.
|
|
// Otherwise, it points to a relobj. We used the unsized and target
|
|
// independent Symbol and Relobj classes instead of Sized_symbol<32> and
|
|
// Arm_relobj, in order to avoid making the stub class a template
|
|
// as most of the stub machinery is endianness-neutral. However, it
|
|
// may require a bit of casting done by users of this class.
|
|
union
|
|
{
|
|
const Symbol* symbol;
|
|
const Relobj* relobj;
|
|
} u_;
|
|
// Addend associated with a reloc.
|
|
int32_t addend_;
|
|
}; // End of inner class Reloc_stub::Key
|
|
|
|
protected:
|
|
// This may be overridden in the child class.
|
|
virtual void
|
|
do_write(unsigned char*, section_size_type);
|
|
|
|
private:
|
|
static const unsigned int invalid_index = static_cast<unsigned int>(-1);
|
|
}; // End of Reloc_stub
|
|
|
|
template<int size, bool big_endian>
|
|
const int Reloc_stub<size, big_endian>::STUB_ADDR_ALIGN = 4;
|
|
|
|
// Write data to output file.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Reloc_stub<size, big_endian>::
|
|
do_write(unsigned char* view, section_size_type)
|
|
{
|
|
typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype;
|
|
const uint32_t* insns = this->insns();
|
|
uint32_t num_insns = this->insn_num();
|
|
Insntype* ip = reinterpret_cast<Insntype*>(view);
|
|
for (uint32_t i = 0; i < num_insns; ++i)
|
|
elfcpp::Swap<32, big_endian>::writeval(ip + i, insns[i]);
|
|
}
|
|
|
|
|
|
// Determine the stub type for a certain relocation or ST_NONE, if no stub is
|
|
// needed.
|
|
|
|
template<int size, bool big_endian>
|
|
inline int
|
|
Reloc_stub<size, big_endian>::stub_type_for_reloc(
|
|
unsigned int r_type, AArch64_address location, AArch64_address dest)
|
|
{
|
|
int64_t branch_offset = 0;
|
|
switch(r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_CALL26:
|
|
case elfcpp::R_AARCH64_JUMP26:
|
|
branch_offset = dest - location;
|
|
break;
|
|
default:
|
|
gold_unreachable();
|
|
}
|
|
|
|
if (aarch64_valid_branch_offset_p(branch_offset))
|
|
return ST_NONE;
|
|
|
|
if (aarch64_valid_for_adrp_p(location, dest))
|
|
return ST_ADRP_BRANCH;
|
|
|
|
// Always use PC-relative addressing in case of -shared or -pie.
|
|
if (parameters->options().output_is_position_independent())
|
|
return ST_LONG_BRANCH_PCREL;
|
|
|
|
// This saves 2 insns per stub, compared to ST_LONG_BRANCH_PCREL.
|
|
// But is only applicable to non-shared or non-pie.
|
|
return ST_LONG_BRANCH_ABS;
|
|
}
|
|
|
|
// A class to hold stubs for the ARM target. This contains 2 different types of
|
|
// stubs - reloc stubs and erratum stubs.
|
|
|
|
template<int size, bool big_endian>
|
|
class Stub_table : public Output_data
|
|
{
|
|
public:
|
|
typedef Target_aarch64<size, big_endian> The_target_aarch64;
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address;
|
|
typedef AArch64_relobj<size, big_endian> The_aarch64_relobj;
|
|
typedef AArch64_input_section<size, big_endian> The_aarch64_input_section;
|
|
typedef Reloc_stub<size, big_endian> The_reloc_stub;
|
|
typedef typename The_reloc_stub::Key The_reloc_stub_key;
|
|
typedef Erratum_stub<size, big_endian> The_erratum_stub;
|
|
typedef Erratum_stub_less<size, big_endian> The_erratum_stub_less;
|
|
typedef typename The_reloc_stub_key::hash The_reloc_stub_key_hash;
|
|
typedef typename The_reloc_stub_key::equal_to The_reloc_stub_key_equal_to;
|
|
typedef Stub_table<size, big_endian> The_stub_table;
|
|
typedef Unordered_map<The_reloc_stub_key, The_reloc_stub*,
|
|
The_reloc_stub_key_hash, The_reloc_stub_key_equal_to>
|
|
Reloc_stub_map;
|
|
typedef typename Reloc_stub_map::const_iterator Reloc_stub_map_const_iter;
|
|
typedef Relocate_info<size, big_endian> The_relocate_info;
|
|
|
|
typedef std::set<The_erratum_stub*, The_erratum_stub_less> Erratum_stub_set;
|
|
typedef typename Erratum_stub_set::iterator Erratum_stub_set_iter;
|
|
|
|
Stub_table(The_aarch64_input_section* owner)
|
|
: Output_data(), owner_(owner), reloc_stubs_size_(0),
|
|
erratum_stubs_size_(0), prev_data_size_(0)
|
|
{ }
|
|
|
|
~Stub_table()
|
|
{ }
|
|
|
|
The_aarch64_input_section*
|
|
owner() const
|
|
{ return owner_; }
|
|
|
|
// Whether this stub table is empty.
|
|
bool
|
|
empty() const
|
|
{ return reloc_stubs_.empty() && erratum_stubs_.empty(); }
|
|
|
|
// Return the current data size.
|
|
off_t
|
|
current_data_size() const
|
|
{ return this->current_data_size_for_child(); }
|
|
|
|
// Add a STUB using KEY. The caller is responsible for avoiding addition
|
|
// if a STUB with the same key has already been added.
|
|
void
|
|
add_reloc_stub(The_reloc_stub* stub, const The_reloc_stub_key& key);
|
|
|
|
// Add an erratum stub into the erratum stub set. The set is ordered by
|
|
// (relobj, shndx, sh_offset).
|
|
void
|
|
add_erratum_stub(The_erratum_stub* stub);
|
|
|
|
// Find if such erratum exists for any given (obj, shndx, sh_offset).
|
|
The_erratum_stub*
|
|
find_erratum_stub(The_aarch64_relobj* a64relobj,
|
|
unsigned int shndx, unsigned int sh_offset);
|
|
|
|
// Find all the erratums for a given input section. The return value is a pair
|
|
// of iterators [begin, end).
|
|
std::pair<Erratum_stub_set_iter, Erratum_stub_set_iter>
|
|
find_erratum_stubs_for_input_section(The_aarch64_relobj* a64relobj,
|
|
unsigned int shndx);
|
|
|
|
// Compute the erratum stub address.
|
|
AArch64_address
|
|
erratum_stub_address(The_erratum_stub* stub) const
|
|
{
|
|
AArch64_address r = align_address(this->address() + this->reloc_stubs_size_,
|
|
The_erratum_stub::STUB_ADDR_ALIGN);
|
|
r += stub->offset();
|
|
return r;
|
|
}
|
|
|
|
// Finalize stubs. No-op here, just for completeness.
|
|
void
|
|
finalize_stubs()
|
|
{ }
|
|
|
|
// Look up a relocation stub using KEY. Return NULL if there is none.
|
|
The_reloc_stub*
|
|
find_reloc_stub(The_reloc_stub_key& key)
|
|
{
|
|
Reloc_stub_map_const_iter p = this->reloc_stubs_.find(key);
|
|
return (p != this->reloc_stubs_.end()) ? p->second : NULL;
|
|
}
|
|
|
|
// Relocate reloc stubs in this stub table. This does not relocate erratum stubs.
|
|
void
|
|
relocate_reloc_stubs(const The_relocate_info*,
|
|
The_target_aarch64*,
|
|
Output_section*,
|
|
unsigned char*,
|
|
AArch64_address,
|
|
section_size_type);
|
|
|
|
// Relocate an erratum stub.
|
|
void
|
|
relocate_erratum_stub(The_erratum_stub*, unsigned char*);
|
|
|
|
// Update data size at the end of a relaxation pass. Return true if data size
|
|
// is different from that of the previous relaxation pass.
|
|
bool
|
|
update_data_size_changed_p()
|
|
{
|
|
// No addralign changed here.
|
|
off_t s = align_address(this->reloc_stubs_size_,
|
|
The_erratum_stub::STUB_ADDR_ALIGN)
|
|
+ this->erratum_stubs_size_;
|
|
bool changed = (s != this->prev_data_size_);
|
|
this->prev_data_size_ = s;
|
|
return changed;
|
|
}
|
|
|
|
protected:
|
|
// Write out section contents.
|
|
void
|
|
do_write(Output_file*);
|
|
|
|
// Return the required alignment.
|
|
uint64_t
|
|
do_addralign() const
|
|
{
|
|
return std::max(The_reloc_stub::STUB_ADDR_ALIGN,
|
|
The_erratum_stub::STUB_ADDR_ALIGN);
|
|
}
|
|
|
|
// Reset address and file offset.
|
|
void
|
|
do_reset_address_and_file_offset()
|
|
{ this->set_current_data_size_for_child(this->prev_data_size_); }
|
|
|
|
// Set final data size.
|
|
void
|
|
set_final_data_size()
|
|
{ this->set_data_size(this->current_data_size()); }
|
|
|
|
private:
|
|
// Relocate one reloc stub.
|
|
void
|
|
relocate_reloc_stub(The_reloc_stub*,
|
|
const The_relocate_info*,
|
|
The_target_aarch64*,
|
|
Output_section*,
|
|
unsigned char*,
|
|
AArch64_address,
|
|
section_size_type);
|
|
|
|
private:
|
|
// Owner of this stub table.
|
|
The_aarch64_input_section* owner_;
|
|
// The relocation stubs.
|
|
Reloc_stub_map reloc_stubs_;
|
|
// The erratum stubs.
|
|
Erratum_stub_set erratum_stubs_;
|
|
// Size of reloc stubs.
|
|
off_t reloc_stubs_size_;
|
|
// Size of erratum stubs.
|
|
off_t erratum_stubs_size_;
|
|
// data size of this in the previous pass.
|
|
off_t prev_data_size_;
|
|
}; // End of Stub_table
|
|
|
|
|
|
// Add an erratum stub into the erratum stub set. The set is ordered by
|
|
// (relobj, shndx, sh_offset).
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Stub_table<size, big_endian>::add_erratum_stub(The_erratum_stub* stub)
|
|
{
|
|
std::pair<Erratum_stub_set_iter, bool> ret =
|
|
this->erratum_stubs_.insert(stub);
|
|
gold_assert(ret.second);
|
|
this->erratum_stubs_size_ = align_address(
|
|
this->erratum_stubs_size_, The_erratum_stub::STUB_ADDR_ALIGN);
|
|
stub->set_offset(this->erratum_stubs_size_);
|
|
this->erratum_stubs_size_ += stub->stub_size();
|
|
}
|
|
|
|
|
|
// Find if such erratum exists for given (obj, shndx, sh_offset).
|
|
|
|
template<int size, bool big_endian>
|
|
Erratum_stub<size, big_endian>*
|
|
Stub_table<size, big_endian>::find_erratum_stub(
|
|
The_aarch64_relobj* a64relobj, unsigned int shndx, unsigned int sh_offset)
|
|
{
|
|
// A dummy object used as key to search in the set.
|
|
The_erratum_stub key(a64relobj, ST_NONE,
|
|
shndx, sh_offset);
|
|
Erratum_stub_set_iter i = this->erratum_stubs_.find(&key);
|
|
if (i != this->erratum_stubs_.end())
|
|
{
|
|
The_erratum_stub* stub(*i);
|
|
gold_assert(stub->erratum_insn() != 0);
|
|
return stub;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
|
|
// Find all the errata for a given input section. The return value is a pair of
|
|
// iterators [begin, end).
|
|
|
|
template<int size, bool big_endian>
|
|
std::pair<typename Stub_table<size, big_endian>::Erratum_stub_set_iter,
|
|
typename Stub_table<size, big_endian>::Erratum_stub_set_iter>
|
|
Stub_table<size, big_endian>::find_erratum_stubs_for_input_section(
|
|
The_aarch64_relobj* a64relobj, unsigned int shndx)
|
|
{
|
|
typedef std::pair<Erratum_stub_set_iter, Erratum_stub_set_iter> Result_pair;
|
|
Erratum_stub_set_iter start, end;
|
|
The_erratum_stub low_key(a64relobj, ST_NONE, shndx, 0);
|
|
start = this->erratum_stubs_.lower_bound(&low_key);
|
|
if (start == this->erratum_stubs_.end())
|
|
return Result_pair(this->erratum_stubs_.end(),
|
|
this->erratum_stubs_.end());
|
|
end = start;
|
|
while (end != this->erratum_stubs_.end() &&
|
|
(*end)->relobj() == a64relobj && (*end)->shndx() == shndx)
|
|
++end;
|
|
return Result_pair(start, end);
|
|
}
|
|
|
|
|
|
// Add a STUB using KEY. The caller is responsible for avoiding addition
|
|
// if a STUB with the same key has already been added.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Stub_table<size, big_endian>::add_reloc_stub(
|
|
The_reloc_stub* stub, const The_reloc_stub_key& key)
|
|
{
|
|
gold_assert(stub->type() == key.type());
|
|
this->reloc_stubs_[key] = stub;
|
|
|
|
// Assign stub offset early. We can do this because we never remove
|
|
// reloc stubs and they are in the beginning of the stub table.
|
|
this->reloc_stubs_size_ = align_address(this->reloc_stubs_size_,
|
|
The_reloc_stub::STUB_ADDR_ALIGN);
|
|
stub->set_offset(this->reloc_stubs_size_);
|
|
this->reloc_stubs_size_ += stub->stub_size();
|
|
}
|
|
|
|
|
|
// Relocate an erratum stub.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Stub_table<size, big_endian>::
|
|
relocate_erratum_stub(The_erratum_stub* estub,
|
|
unsigned char* view)
|
|
{
|
|
// Just for convenience.
|
|
const int BPI = AArch64_insn_utilities<big_endian>::BYTES_PER_INSN;
|
|
|
|
gold_assert(!estub->is_invalidated_erratum_stub());
|
|
AArch64_address stub_address = this->erratum_stub_address(estub);
|
|
// The address of "b" in the stub that is to be "relocated".
|
|
AArch64_address stub_b_insn_address;
|
|
// Branch offset that is to be filled in "b" insn.
|
|
int b_offset = 0;
|
|
switch (estub->type())
|
|
{
|
|
case ST_E_843419:
|
|
case ST_E_835769:
|
|
// The 1st insn of the erratum could be a relocation spot,
|
|
// in this case we need to fix it with
|
|
// "(*i)->erratum_insn()".
|
|
elfcpp::Swap<32, big_endian>::writeval(
|
|
view + (stub_address - this->address()),
|
|
estub->erratum_insn());
|
|
// For the erratum, the 2nd insn is a b-insn to be patched
|
|
// (relocated).
|
|
stub_b_insn_address = stub_address + 1 * BPI;
|
|
b_offset = estub->destination_address() - stub_b_insn_address;
|
|
AArch64_relocate_functions<size, big_endian>::construct_b(
|
|
view + (stub_b_insn_address - this->address()),
|
|
((unsigned int)(b_offset)) & 0xfffffff);
|
|
break;
|
|
default:
|
|
gold_unreachable();
|
|
break;
|
|
}
|
|
estub->invalidate_erratum_stub();
|
|
}
|
|
|
|
|
|
// Relocate only reloc stubs in this stub table. This does not relocate erratum
|
|
// stubs.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Stub_table<size, big_endian>::
|
|
relocate_reloc_stubs(const The_relocate_info* relinfo,
|
|
The_target_aarch64* target_aarch64,
|
|
Output_section* output_section,
|
|
unsigned char* view,
|
|
AArch64_address address,
|
|
section_size_type view_size)
|
|
{
|
|
// "view_size" is the total size of the stub_table.
|
|
gold_assert(address == this->address() &&
|
|
view_size == static_cast<section_size_type>(this->data_size()));
|
|
for(Reloc_stub_map_const_iter p = this->reloc_stubs_.begin();
|
|
p != this->reloc_stubs_.end(); ++p)
|
|
relocate_reloc_stub(p->second, relinfo, target_aarch64, output_section,
|
|
view, address, view_size);
|
|
}
|
|
|
|
|
|
// Relocate one reloc stub. This is a helper for
|
|
// Stub_table::relocate_reloc_stubs().
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Stub_table<size, big_endian>::
|
|
relocate_reloc_stub(The_reloc_stub* stub,
|
|
const The_relocate_info* relinfo,
|
|
The_target_aarch64* target_aarch64,
|
|
Output_section* output_section,
|
|
unsigned char* view,
|
|
AArch64_address address,
|
|
section_size_type view_size)
|
|
{
|
|
// "offset" is the offset from the beginning of the stub_table.
|
|
section_size_type offset = stub->offset();
|
|
section_size_type stub_size = stub->stub_size();
|
|
// "view_size" is the total size of the stub_table.
|
|
gold_assert(offset + stub_size <= view_size);
|
|
|
|
target_aarch64->relocate_reloc_stub(stub, relinfo, output_section,
|
|
view + offset, address + offset, view_size);
|
|
}
|
|
|
|
|
|
// Write out the stubs to file.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Stub_table<size, big_endian>::do_write(Output_file* of)
|
|
{
|
|
off_t offset = this->offset();
|
|
const section_size_type oview_size =
|
|
convert_to_section_size_type(this->data_size());
|
|
unsigned char* const oview = of->get_output_view(offset, oview_size);
|
|
|
|
// Write relocation stubs.
|
|
for (typename Reloc_stub_map::const_iterator p = this->reloc_stubs_.begin();
|
|
p != this->reloc_stubs_.end(); ++p)
|
|
{
|
|
The_reloc_stub* stub = p->second;
|
|
AArch64_address address = this->address() + stub->offset();
|
|
gold_assert(address ==
|
|
align_address(address, The_reloc_stub::STUB_ADDR_ALIGN));
|
|
stub->write(oview + stub->offset(), stub->stub_size());
|
|
}
|
|
|
|
// Write erratum stubs.
|
|
unsigned int erratum_stub_start_offset =
|
|
align_address(this->reloc_stubs_size_, The_erratum_stub::STUB_ADDR_ALIGN);
|
|
for (typename Erratum_stub_set::iterator p = this->erratum_stubs_.begin();
|
|
p != this->erratum_stubs_.end(); ++p)
|
|
{
|
|
The_erratum_stub* stub(*p);
|
|
stub->write(oview + erratum_stub_start_offset + stub->offset(),
|
|
stub->stub_size());
|
|
}
|
|
|
|
of->write_output_view(this->offset(), oview_size, oview);
|
|
}
|
|
|
|
|
|
// AArch64_relobj class.
|
|
|
|
template<int size, bool big_endian>
|
|
class AArch64_relobj : public Sized_relobj_file<size, big_endian>
|
|
{
|
|
public:
|
|
typedef AArch64_relobj<size, big_endian> This;
|
|
typedef Target_aarch64<size, big_endian> The_target_aarch64;
|
|
typedef AArch64_input_section<size, big_endian> The_aarch64_input_section;
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address;
|
|
typedef Stub_table<size, big_endian> The_stub_table;
|
|
typedef Erratum_stub<size, big_endian> The_erratum_stub;
|
|
typedef typename The_stub_table::Erratum_stub_set_iter Erratum_stub_set_iter;
|
|
typedef std::vector<The_stub_table*> Stub_table_list;
|
|
static const AArch64_address invalid_address =
|
|
static_cast<AArch64_address>(-1);
|
|
|
|
AArch64_relobj(const std::string& name, Input_file* input_file, off_t offset,
|
|
const typename elfcpp::Ehdr<size, big_endian>& ehdr)
|
|
: Sized_relobj_file<size, big_endian>(name, input_file, offset, ehdr),
|
|
stub_tables_()
|
|
{ }
|
|
|
|
~AArch64_relobj()
|
|
{ }
|
|
|
|
// Return the stub table of the SHNDX-th section if there is one.
|
|
The_stub_table*
|
|
stub_table(unsigned int shndx) const
|
|
{
|
|
gold_assert(shndx < this->stub_tables_.size());
|
|
return this->stub_tables_[shndx];
|
|
}
|
|
|
|
// Set STUB_TABLE to be the stub_table of the SHNDX-th section.
|
|
void
|
|
set_stub_table(unsigned int shndx, The_stub_table* stub_table)
|
|
{
|
|
gold_assert(shndx < this->stub_tables_.size());
|
|
this->stub_tables_[shndx] = stub_table;
|
|
}
|
|
|
|
// Entrance to errata scanning.
|
|
void
|
|
scan_errata(unsigned int shndx,
|
|
const elfcpp::Shdr<size, big_endian>&,
|
|
Output_section*, const Symbol_table*,
|
|
The_target_aarch64*);
|
|
|
|
// Scan all relocation sections for stub generation.
|
|
void
|
|
scan_sections_for_stubs(The_target_aarch64*, const Symbol_table*,
|
|
const Layout*);
|
|
|
|
// Whether a section is a scannable text section.
|
|
bool
|
|
text_section_is_scannable(const elfcpp::Shdr<size, big_endian>&, unsigned int,
|
|
const Output_section*, const Symbol_table*);
|
|
|
|
// Convert regular input section with index SHNDX to a relaxed section.
|
|
void
|
|
convert_input_section_to_relaxed_section(unsigned shndx)
|
|
{
|
|
// The stubs have relocations and we need to process them after writing
|
|
// out the stubs. So relocation now must follow section write.
|
|
this->set_section_offset(shndx, -1ULL);
|
|
this->set_relocs_must_follow_section_writes();
|
|
}
|
|
|
|
// Structure for mapping symbol position.
|
|
struct Mapping_symbol_position
|
|
{
|
|
Mapping_symbol_position(unsigned int shndx, AArch64_address offset):
|
|
shndx_(shndx), offset_(offset)
|
|
{}
|
|
|
|
// "<" comparator used in ordered_map container.
|
|
bool
|
|
operator<(const Mapping_symbol_position& p) const
|
|
{
|
|
return (this->shndx_ < p.shndx_
|
|
|| (this->shndx_ == p.shndx_ && this->offset_ < p.offset_));
|
|
}
|
|
|
|
// Section index.
|
|
unsigned int shndx_;
|
|
|
|
// Section offset.
|
|
AArch64_address offset_;
|
|
};
|
|
|
|
typedef std::map<Mapping_symbol_position, char> Mapping_symbol_info;
|
|
|
|
protected:
|
|
// Post constructor setup.
|
|
void
|
|
do_setup()
|
|
{
|
|
// Call parent's setup method.
|
|
Sized_relobj_file<size, big_endian>::do_setup();
|
|
|
|
// Initialize look-up tables.
|
|
this->stub_tables_.resize(this->shnum());
|
|
}
|
|
|
|
virtual void
|
|
do_relocate_sections(
|
|
const Symbol_table* symtab, const Layout* layout,
|
|
const unsigned char* pshdrs, Output_file* of,
|
|
typename Sized_relobj_file<size, big_endian>::Views* pviews);
|
|
|
|
// Count local symbols and (optionally) record mapping info.
|
|
virtual void
|
|
do_count_local_symbols(Stringpool_template<char>*,
|
|
Stringpool_template<char>*);
|
|
|
|
private:
|
|
// Fix all errata in the object, and for each erratum, relocate corresponding
|
|
// erratum stub.
|
|
void
|
|
fix_errata_and_relocate_erratum_stubs(
|
|
typename Sized_relobj_file<size, big_endian>::Views* pviews);
|
|
|
|
// Try to fix erratum 843419 in an optimized way. Return true if patch is
|
|
// applied.
|
|
bool
|
|
try_fix_erratum_843419_optimized(
|
|
The_erratum_stub*, AArch64_address,
|
|
typename Sized_relobj_file<size, big_endian>::View_size&);
|
|
|
|
// Whether a section needs to be scanned for relocation stubs.
|
|
bool
|
|
section_needs_reloc_stub_scanning(const elfcpp::Shdr<size, big_endian>&,
|
|
const Relobj::Output_sections&,
|
|
const Symbol_table*, const unsigned char*);
|
|
|
|
// List of stub tables.
|
|
Stub_table_list stub_tables_;
|
|
|
|
// Mapping symbol information sorted by (section index, section_offset).
|
|
Mapping_symbol_info mapping_symbol_info_;
|
|
}; // End of AArch64_relobj
|
|
|
|
|
|
// Override to record mapping symbol information.
|
|
template<int size, bool big_endian>
|
|
void
|
|
AArch64_relobj<size, big_endian>::do_count_local_symbols(
|
|
Stringpool_template<char>* pool, Stringpool_template<char>* dynpool)
|
|
{
|
|
Sized_relobj_file<size, big_endian>::do_count_local_symbols(pool, dynpool);
|
|
|
|
// Only erratum-fixing work needs mapping symbols, so skip this time consuming
|
|
// processing if not fixing erratum.
|
|
if (!parameters->options().fix_cortex_a53_843419()
|
|
&& !parameters->options().fix_cortex_a53_835769())
|
|
return;
|
|
|
|
const unsigned int loccount = this->local_symbol_count();
|
|
if (loccount == 0)
|
|
return;
|
|
|
|
// Read the symbol table section header.
|
|
const unsigned int symtab_shndx = this->symtab_shndx();
|
|
elfcpp::Shdr<size, big_endian>
|
|
symtabshdr(this, this->elf_file()->section_header(symtab_shndx));
|
|
gold_assert(symtabshdr.get_sh_type() == elfcpp::SHT_SYMTAB);
|
|
|
|
// Read the local symbols.
|
|
const int sym_size =elfcpp::Elf_sizes<size>::sym_size;
|
|
gold_assert(loccount == symtabshdr.get_sh_info());
|
|
off_t locsize = loccount * sym_size;
|
|
const unsigned char* psyms = this->get_view(symtabshdr.get_sh_offset(),
|
|
locsize, true, true);
|
|
|
|
// For mapping symbol processing, we need to read the symbol names.
|
|
unsigned int strtab_shndx = this->adjust_shndx(symtabshdr.get_sh_link());
|
|
if (strtab_shndx >= this->shnum())
|
|
{
|
|
this->error(_("invalid symbol table name index: %u"), strtab_shndx);
|
|
return;
|
|
}
|
|
|
|
elfcpp::Shdr<size, big_endian>
|
|
strtabshdr(this, this->elf_file()->section_header(strtab_shndx));
|
|
if (strtabshdr.get_sh_type() != elfcpp::SHT_STRTAB)
|
|
{
|
|
this->error(_("symbol table name section has wrong type: %u"),
|
|
static_cast<unsigned int>(strtabshdr.get_sh_type()));
|
|
return;
|
|
}
|
|
|
|
const char* pnames =
|
|
reinterpret_cast<const char*>(this->get_view(strtabshdr.get_sh_offset(),
|
|
strtabshdr.get_sh_size(),
|
|
false, false));
|
|
|
|
// Skip the first dummy symbol.
|
|
psyms += sym_size;
|
|
typename Sized_relobj_file<size, big_endian>::Local_values*
|
|
plocal_values = this->local_values();
|
|
for (unsigned int i = 1; i < loccount; ++i, psyms += sym_size)
|
|
{
|
|
elfcpp::Sym<size, big_endian> sym(psyms);
|
|
Symbol_value<size>& lv((*plocal_values)[i]);
|
|
AArch64_address input_value = lv.input_value();
|
|
|
|
// Check to see if this is a mapping symbol. AArch64 mapping symbols are
|
|
// defined in "ELF for the ARM 64-bit Architecture", Table 4-4, Mapping
|
|
// symbols.
|
|
// Mapping symbols could be one of the following 4 forms -
|
|
// a) $x
|
|
// b) $x.<any...>
|
|
// c) $d
|
|
// d) $d.<any...>
|
|
const char* sym_name = pnames + sym.get_st_name();
|
|
if (sym_name[0] == '$' && (sym_name[1] == 'x' || sym_name[1] == 'd')
|
|
&& (sym_name[2] == '\0' || sym_name[2] == '.'))
|
|
{
|
|
bool is_ordinary;
|
|
unsigned int input_shndx =
|
|
this->adjust_sym_shndx(i, sym.get_st_shndx(), &is_ordinary);
|
|
gold_assert(is_ordinary);
|
|
|
|
Mapping_symbol_position msp(input_shndx, input_value);
|
|
// Insert mapping_symbol_info into map whose ordering is defined by
|
|
// (shndx, offset_within_section).
|
|
this->mapping_symbol_info_[msp] = sym_name[1];
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
// Fix all errata in the object and for each erratum, we relocate the
|
|
// corresponding erratum stub (by calling Stub_table::relocate_erratum_stub).
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
AArch64_relobj<size, big_endian>::fix_errata_and_relocate_erratum_stubs(
|
|
typename Sized_relobj_file<size, big_endian>::Views* pviews)
|
|
{
|
|
typedef typename elfcpp::Swap<32,big_endian>::Valtype Insntype;
|
|
unsigned int shnum = this->shnum();
|
|
const Relobj::Output_sections& out_sections(this->output_sections());
|
|
for (unsigned int i = 1; i < shnum; ++i)
|
|
{
|
|
The_stub_table* stub_table = this->stub_table(i);
|
|
if (!stub_table)
|
|
continue;
|
|
std::pair<Erratum_stub_set_iter, Erratum_stub_set_iter>
|
|
ipair(stub_table->find_erratum_stubs_for_input_section(this, i));
|
|
Erratum_stub_set_iter p = ipair.first, end = ipair.second;
|
|
typename Sized_relobj_file<size, big_endian>::View_size&
|
|
pview((*pviews)[i]);
|
|
AArch64_address view_offset = 0;
|
|
if (pview.is_input_output_view)
|
|
{
|
|
// In this case, write_sections has not added the output offset to
|
|
// the view's address, so we must do so. Currently this only happens
|
|
// for a relaxed section.
|
|
unsigned int index = this->adjust_shndx(i);
|
|
const Output_relaxed_input_section* poris =
|
|
out_sections[index]->find_relaxed_input_section(this, index);
|
|
gold_assert(poris != NULL);
|
|
view_offset = poris->address() - pview.address;
|
|
}
|
|
|
|
while (p != end)
|
|
{
|
|
The_erratum_stub* stub = *p;
|
|
|
|
// Double check data before fix.
|
|
gold_assert(pview.address + view_offset + stub->sh_offset()
|
|
== stub->erratum_address());
|
|
|
|
// Update previously recorded erratum insn with relocated
|
|
// version.
|
|
Insntype* ip =
|
|
reinterpret_cast<Insntype*>(
|
|
pview.view + view_offset + stub->sh_offset());
|
|
Insntype insn_to_fix = ip[0];
|
|
stub->update_erratum_insn(insn_to_fix);
|
|
|
|
// First try to see if erratum is 843419 and if it can be fixed
|
|
// without using branch-to-stub.
|
|
if (!try_fix_erratum_843419_optimized(stub, view_offset, pview))
|
|
{
|
|
// Replace the erratum insn with a branch-to-stub.
|
|
AArch64_address stub_address =
|
|
stub_table->erratum_stub_address(stub);
|
|
unsigned int b_offset = stub_address - stub->erratum_address();
|
|
AArch64_relocate_functions<size, big_endian>::construct_b(
|
|
pview.view + view_offset + stub->sh_offset(),
|
|
b_offset & 0xfffffff);
|
|
}
|
|
|
|
// Erratum fix is done (or skipped), continue to relocate erratum
|
|
// stub. Note, when erratum fix is skipped (either because we
|
|
// proactively change the code sequence or the code sequence is
|
|
// changed by relaxation, etc), we can still safely relocate the
|
|
// erratum stub, ignoring the fact the erratum could never be
|
|
// executed.
|
|
stub_table->relocate_erratum_stub(
|
|
stub,
|
|
pview.view + (stub_table->address() - pview.address));
|
|
|
|
// Next erratum stub.
|
|
++p;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
// This is an optimization for 843419. This erratum requires the sequence begin
|
|
// with 'adrp', when final value calculated by adrp fits in adr, we can just
|
|
// replace 'adrp' with 'adr', so we save 2 jumps per occurrence. (Note, however,
|
|
// in this case, we do not delete the erratum stub (too late to do so), it is
|
|
// merely generated without ever being called.)
|
|
|
|
template<int size, bool big_endian>
|
|
bool
|
|
AArch64_relobj<size, big_endian>::try_fix_erratum_843419_optimized(
|
|
The_erratum_stub* stub, AArch64_address view_offset,
|
|
typename Sized_relobj_file<size, big_endian>::View_size& pview)
|
|
{
|
|
if (stub->type() != ST_E_843419)
|
|
return false;
|
|
|
|
typedef AArch64_insn_utilities<big_endian> Insn_utilities;
|
|
typedef typename elfcpp::Swap<32,big_endian>::Valtype Insntype;
|
|
E843419_stub<size, big_endian>* e843419_stub =
|
|
reinterpret_cast<E843419_stub<size, big_endian>*>(stub);
|
|
AArch64_address pc =
|
|
pview.address + view_offset + e843419_stub->adrp_sh_offset();
|
|
unsigned int adrp_offset = e843419_stub->adrp_sh_offset ();
|
|
Insntype* adrp_view =
|
|
reinterpret_cast<Insntype*>(pview.view + view_offset + adrp_offset);
|
|
Insntype adrp_insn = adrp_view[0];
|
|
|
|
// If the instruction at adrp_sh_offset is "mrs R, tpidr_el0", it may come
|
|
// from IE -> LE relaxation etc. This is a side-effect of TLS relaxation that
|
|
// ADRP has been turned into MRS, there is no erratum risk anymore.
|
|
// Therefore, we return true to avoid doing unnecessary branch-to-stub.
|
|
if (Insn_utilities::is_mrs_tpidr_el0(adrp_insn))
|
|
return true;
|
|
|
|
// If the instruction at adrp_sh_offset is not ADRP and the instruction before
|
|
// it is "mrs R, tpidr_el0", it may come from LD -> LE relaxation etc.
|
|
// Like the above case, there is no erratum risk any more, we can safely
|
|
// return true.
|
|
if (!Insn_utilities::is_adrp(adrp_insn) && adrp_offset)
|
|
{
|
|
Insntype* prev_view =
|
|
reinterpret_cast<Insntype*>(
|
|
pview.view + view_offset + adrp_offset - 4);
|
|
Insntype prev_insn = prev_view[0];
|
|
|
|
if (Insn_utilities::is_mrs_tpidr_el0(prev_insn))
|
|
return true;
|
|
}
|
|
|
|
/* If we reach here, the first instruction must be ADRP. */
|
|
gold_assert(Insn_utilities::is_adrp(adrp_insn));
|
|
// Get adrp 33-bit signed imm value.
|
|
int64_t adrp_imm = Insn_utilities::
|
|
aarch64_adrp_decode_imm(adrp_insn);
|
|
// adrp - final value transferred to target register is calculated as:
|
|
// PC[11:0] = Zeros(12)
|
|
// adrp_dest_value = PC + adrp_imm;
|
|
int64_t adrp_dest_value = (pc & ~((1 << 12) - 1)) + adrp_imm;
|
|
// adr -final value transferred to target register is calucalted as:
|
|
// PC + adr_imm
|
|
// So we have:
|
|
// PC + adr_imm = adrp_dest_value
|
|
// ==>
|
|
// adr_imm = adrp_dest_value - PC
|
|
int64_t adr_imm = adrp_dest_value - pc;
|
|
// Check if imm fits in adr (21-bit signed).
|
|
if (-(1 << 20) <= adr_imm && adr_imm < (1 << 20))
|
|
{
|
|
// Convert 'adrp' into 'adr'.
|
|
Insntype adr_insn = adrp_insn & ((1u << 31) - 1);
|
|
adr_insn = Insn_utilities::
|
|
aarch64_adr_encode_imm(adr_insn, adr_imm);
|
|
elfcpp::Swap<32, big_endian>::writeval(adrp_view, adr_insn);
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
|
|
// Relocate sections.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
AArch64_relobj<size, big_endian>::do_relocate_sections(
|
|
const Symbol_table* symtab, const Layout* layout,
|
|
const unsigned char* pshdrs, Output_file* of,
|
|
typename Sized_relobj_file<size, big_endian>::Views* pviews)
|
|
{
|
|
// Relocate the section data.
|
|
this->relocate_section_range(symtab, layout, pshdrs, of, pviews,
|
|
1, this->shnum() - 1);
|
|
|
|
// We do not generate stubs if doing a relocatable link.
|
|
if (parameters->options().relocatable())
|
|
return;
|
|
|
|
// This part only relocates erratum stubs that belong to input sections of this
|
|
// object file.
|
|
if (parameters->options().fix_cortex_a53_843419()
|
|
|| parameters->options().fix_cortex_a53_835769())
|
|
this->fix_errata_and_relocate_erratum_stubs(pviews);
|
|
|
|
Relocate_info<size, big_endian> relinfo;
|
|
relinfo.symtab = symtab;
|
|
relinfo.layout = layout;
|
|
relinfo.object = this;
|
|
|
|
// This part relocates all reloc stubs that are contained in stub_tables of
|
|
// this object file.
|
|
unsigned int shnum = this->shnum();
|
|
The_target_aarch64* target = The_target_aarch64::current_target();
|
|
|
|
for (unsigned int i = 1; i < shnum; ++i)
|
|
{
|
|
The_aarch64_input_section* aarch64_input_section =
|
|
target->find_aarch64_input_section(this, i);
|
|
if (aarch64_input_section != NULL
|
|
&& aarch64_input_section->is_stub_table_owner()
|
|
&& !aarch64_input_section->stub_table()->empty())
|
|
{
|
|
Output_section* os = this->output_section(i);
|
|
gold_assert(os != NULL);
|
|
|
|
relinfo.reloc_shndx = elfcpp::SHN_UNDEF;
|
|
relinfo.reloc_shdr = NULL;
|
|
relinfo.data_shndx = i;
|
|
relinfo.data_shdr = pshdrs + i * elfcpp::Elf_sizes<size>::shdr_size;
|
|
|
|
typename Sized_relobj_file<size, big_endian>::View_size&
|
|
view_struct = (*pviews)[i];
|
|
gold_assert(view_struct.view != NULL);
|
|
|
|
The_stub_table* stub_table = aarch64_input_section->stub_table();
|
|
off_t offset = stub_table->address() - view_struct.address;
|
|
unsigned char* view = view_struct.view + offset;
|
|
AArch64_address address = stub_table->address();
|
|
section_size_type view_size = stub_table->data_size();
|
|
stub_table->relocate_reloc_stubs(&relinfo, target, os, view, address,
|
|
view_size);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
// Determine if an input section is scannable for stub processing. SHDR is
|
|
// the header of the section and SHNDX is the section index. OS is the output
|
|
// section for the input section and SYMTAB is the global symbol table used to
|
|
// look up ICF information.
|
|
|
|
template<int size, bool big_endian>
|
|
bool
|
|
AArch64_relobj<size, big_endian>::text_section_is_scannable(
|
|
const elfcpp::Shdr<size, big_endian>& text_shdr,
|
|
unsigned int text_shndx,
|
|
const Output_section* os,
|
|
const Symbol_table* symtab)
|
|
{
|
|
// Skip any empty sections, unallocated sections or sections whose
|
|
// type are not SHT_PROGBITS.
|
|
if (text_shdr.get_sh_size() == 0
|
|
|| (text_shdr.get_sh_flags() & elfcpp::SHF_ALLOC) == 0
|
|
|| text_shdr.get_sh_type() != elfcpp::SHT_PROGBITS)
|
|
return false;
|
|
|
|
// Skip any discarded or ICF'ed sections.
|
|
if (os == NULL || symtab->is_section_folded(this, text_shndx))
|
|
return false;
|
|
|
|
// Skip exception frame.
|
|
if (strcmp(os->name(), ".eh_frame") == 0)
|
|
return false ;
|
|
|
|
gold_assert(!this->is_output_section_offset_invalid(text_shndx) ||
|
|
os->find_relaxed_input_section(this, text_shndx) != NULL);
|
|
|
|
return true;
|
|
}
|
|
|
|
|
|
// Determine if we want to scan the SHNDX-th section for relocation stubs.
|
|
// This is a helper for AArch64_relobj::scan_sections_for_stubs().
|
|
|
|
template<int size, bool big_endian>
|
|
bool
|
|
AArch64_relobj<size, big_endian>::section_needs_reloc_stub_scanning(
|
|
const elfcpp::Shdr<size, big_endian>& shdr,
|
|
const Relobj::Output_sections& out_sections,
|
|
const Symbol_table* symtab,
|
|
const unsigned char* pshdrs)
|
|
{
|
|
unsigned int sh_type = shdr.get_sh_type();
|
|
if (sh_type != elfcpp::SHT_RELA)
|
|
return false;
|
|
|
|
// Ignore empty section.
|
|
off_t sh_size = shdr.get_sh_size();
|
|
if (sh_size == 0)
|
|
return false;
|
|
|
|
// Ignore reloc section with unexpected symbol table. The
|
|
// error will be reported in the final link.
|
|
if (this->adjust_shndx(shdr.get_sh_link()) != this->symtab_shndx())
|
|
return false;
|
|
|
|
gold_assert(sh_type == elfcpp::SHT_RELA);
|
|
unsigned int reloc_size = elfcpp::Elf_sizes<size>::rela_size;
|
|
|
|
// Ignore reloc section with unexpected entsize or uneven size.
|
|
// The error will be reported in the final link.
|
|
if (reloc_size != shdr.get_sh_entsize() || sh_size % reloc_size != 0)
|
|
return false;
|
|
|
|
// Ignore reloc section with bad info. This error will be
|
|
// reported in the final link.
|
|
unsigned int text_shndx = this->adjust_shndx(shdr.get_sh_info());
|
|
if (text_shndx >= this->shnum())
|
|
return false;
|
|
|
|
const unsigned int shdr_size = elfcpp::Elf_sizes<size>::shdr_size;
|
|
const elfcpp::Shdr<size, big_endian> text_shdr(pshdrs +
|
|
text_shndx * shdr_size);
|
|
return this->text_section_is_scannable(text_shdr, text_shndx,
|
|
out_sections[text_shndx], symtab);
|
|
}
|
|
|
|
|
|
// Scan section SHNDX for erratum 843419 and 835769.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
AArch64_relobj<size, big_endian>::scan_errata(
|
|
unsigned int shndx, const elfcpp::Shdr<size, big_endian>& shdr,
|
|
Output_section* os, const Symbol_table* symtab,
|
|
The_target_aarch64* target)
|
|
{
|
|
if (shdr.get_sh_size() == 0
|
|
|| (shdr.get_sh_flags() &
|
|
(elfcpp::SHF_ALLOC | elfcpp::SHF_EXECINSTR)) == 0
|
|
|| shdr.get_sh_type() != elfcpp::SHT_PROGBITS)
|
|
return;
|
|
|
|
if (!os || symtab->is_section_folded(this, shndx)) return;
|
|
|
|
AArch64_address output_offset = this->get_output_section_offset(shndx);
|
|
AArch64_address output_address;
|
|
if (output_offset != invalid_address)
|
|
output_address = os->address() + output_offset;
|
|
else
|
|
{
|
|
const Output_relaxed_input_section* poris =
|
|
os->find_relaxed_input_section(this, shndx);
|
|
if (!poris) return;
|
|
output_address = poris->address();
|
|
}
|
|
|
|
// Update the addresses in previously generated erratum stubs. Unlike when
|
|
// we scan relocations for stubs, if section addresses have changed due to
|
|
// other relaxations we are unlikely to scan the same erratum instances
|
|
// again.
|
|
The_stub_table* stub_table = this->stub_table(shndx);
|
|
if (stub_table)
|
|
{
|
|
std::pair<Erratum_stub_set_iter, Erratum_stub_set_iter>
|
|
ipair(stub_table->find_erratum_stubs_for_input_section(this, shndx));
|
|
for (Erratum_stub_set_iter p = ipair.first; p != ipair.second; ++p)
|
|
(*p)->update_erratum_address(output_address);
|
|
}
|
|
|
|
section_size_type input_view_size = 0;
|
|
const unsigned char* input_view =
|
|
this->section_contents(shndx, &input_view_size, false);
|
|
|
|
Mapping_symbol_position section_start(shndx, 0);
|
|
// Find the first mapping symbol record within section shndx.
|
|
typename Mapping_symbol_info::const_iterator p =
|
|
this->mapping_symbol_info_.lower_bound(section_start);
|
|
while (p != this->mapping_symbol_info_.end() &&
|
|
p->first.shndx_ == shndx)
|
|
{
|
|
typename Mapping_symbol_info::const_iterator prev = p;
|
|
++p;
|
|
if (prev->second == 'x')
|
|
{
|
|
section_size_type span_start =
|
|
convert_to_section_size_type(prev->first.offset_);
|
|
section_size_type span_end;
|
|
if (p != this->mapping_symbol_info_.end()
|
|
&& p->first.shndx_ == shndx)
|
|
span_end = convert_to_section_size_type(p->first.offset_);
|
|
else
|
|
span_end = convert_to_section_size_type(shdr.get_sh_size());
|
|
|
|
// Here we do not share the scanning code of both errata. For 843419,
|
|
// only the last few insns of each page are examined, which is fast,
|
|
// whereas, for 835769, every insn pair needs to be checked.
|
|
|
|
if (parameters->options().fix_cortex_a53_843419())
|
|
target->scan_erratum_843419_span(
|
|
this, shndx, span_start, span_end,
|
|
const_cast<unsigned char*>(input_view), output_address);
|
|
|
|
if (parameters->options().fix_cortex_a53_835769())
|
|
target->scan_erratum_835769_span(
|
|
this, shndx, span_start, span_end,
|
|
const_cast<unsigned char*>(input_view), output_address);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
// Scan relocations for stub generation.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
AArch64_relobj<size, big_endian>::scan_sections_for_stubs(
|
|
The_target_aarch64* target,
|
|
const Symbol_table* symtab,
|
|
const Layout* layout)
|
|
{
|
|
unsigned int shnum = this->shnum();
|
|
const unsigned int shdr_size = elfcpp::Elf_sizes<size>::shdr_size;
|
|
|
|
// Read the section headers.
|
|
const unsigned char* pshdrs = this->get_view(this->elf_file()->shoff(),
|
|
shnum * shdr_size,
|
|
true, true);
|
|
|
|
// To speed up processing, we set up hash tables for fast lookup of
|
|
// input offsets to output addresses.
|
|
this->initialize_input_to_output_maps();
|
|
|
|
const Relobj::Output_sections& out_sections(this->output_sections());
|
|
|
|
Relocate_info<size, big_endian> relinfo;
|
|
relinfo.symtab = symtab;
|
|
relinfo.layout = layout;
|
|
relinfo.object = this;
|
|
|
|
// Do relocation stubs scanning.
|
|
const unsigned char* p = pshdrs + shdr_size;
|
|
for (unsigned int i = 1; i < shnum; ++i, p += shdr_size)
|
|
{
|
|
const elfcpp::Shdr<size, big_endian> shdr(p);
|
|
if (parameters->options().fix_cortex_a53_843419()
|
|
|| parameters->options().fix_cortex_a53_835769())
|
|
scan_errata(i, shdr, out_sections[i], symtab, target);
|
|
if (this->section_needs_reloc_stub_scanning(shdr, out_sections, symtab,
|
|
pshdrs))
|
|
{
|
|
unsigned int index = this->adjust_shndx(shdr.get_sh_info());
|
|
AArch64_address output_offset =
|
|
this->get_output_section_offset(index);
|
|
AArch64_address output_address;
|
|
if (output_offset != invalid_address)
|
|
{
|
|
output_address = out_sections[index]->address() + output_offset;
|
|
}
|
|
else
|
|
{
|
|
// Currently this only happens for a relaxed section.
|
|
const Output_relaxed_input_section* poris =
|
|
out_sections[index]->find_relaxed_input_section(this, index);
|
|
gold_assert(poris != NULL);
|
|
output_address = poris->address();
|
|
}
|
|
|
|
// Get the relocations.
|
|
const unsigned char* prelocs = this->get_view(shdr.get_sh_offset(),
|
|
shdr.get_sh_size(),
|
|
true, false);
|
|
|
|
// Get the section contents.
|
|
section_size_type input_view_size = 0;
|
|
const unsigned char* input_view =
|
|
this->section_contents(index, &input_view_size, false);
|
|
|
|
relinfo.reloc_shndx = i;
|
|
relinfo.data_shndx = index;
|
|
unsigned int sh_type = shdr.get_sh_type();
|
|
unsigned int reloc_size;
|
|
gold_assert (sh_type == elfcpp::SHT_RELA);
|
|
reloc_size = elfcpp::Elf_sizes<size>::rela_size;
|
|
|
|
Output_section* os = out_sections[index];
|
|
target->scan_section_for_stubs(&relinfo, sh_type, prelocs,
|
|
shdr.get_sh_size() / reloc_size,
|
|
os,
|
|
output_offset == invalid_address,
|
|
input_view, output_address,
|
|
input_view_size);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
// A class to wrap an ordinary input section containing executable code.
|
|
|
|
template<int size, bool big_endian>
|
|
class AArch64_input_section : public Output_relaxed_input_section
|
|
{
|
|
public:
|
|
typedef Stub_table<size, big_endian> The_stub_table;
|
|
|
|
AArch64_input_section(Relobj* relobj, unsigned int shndx)
|
|
: Output_relaxed_input_section(relobj, shndx, 1),
|
|
stub_table_(NULL),
|
|
original_contents_(NULL), original_size_(0),
|
|
original_addralign_(1)
|
|
{ }
|
|
|
|
~AArch64_input_section()
|
|
{ delete[] this->original_contents_; }
|
|
|
|
// Initialize.
|
|
void
|
|
init();
|
|
|
|
// Set the stub_table.
|
|
void
|
|
set_stub_table(The_stub_table* st)
|
|
{ this->stub_table_ = st; }
|
|
|
|
// Whether this is a stub table owner.
|
|
bool
|
|
is_stub_table_owner() const
|
|
{ return this->stub_table_ != NULL && this->stub_table_->owner() == this; }
|
|
|
|
// Return the original size of the section.
|
|
uint32_t
|
|
original_size() const
|
|
{ return this->original_size_; }
|
|
|
|
// Return the stub table.
|
|
The_stub_table*
|
|
stub_table()
|
|
{ return stub_table_; }
|
|
|
|
protected:
|
|
// Write out this input section.
|
|
void
|
|
do_write(Output_file*);
|
|
|
|
// Return required alignment of this.
|
|
uint64_t
|
|
do_addralign() const
|
|
{
|
|
if (this->is_stub_table_owner())
|
|
return std::max(this->stub_table_->addralign(),
|
|
static_cast<uint64_t>(this->original_addralign_));
|
|
else
|
|
return this->original_addralign_;
|
|
}
|
|
|
|
// Finalize data size.
|
|
void
|
|
set_final_data_size();
|
|
|
|
// Reset address and file offset.
|
|
void
|
|
do_reset_address_and_file_offset();
|
|
|
|
// Output offset.
|
|
bool
|
|
do_output_offset(const Relobj* object, unsigned int shndx,
|
|
section_offset_type offset,
|
|
section_offset_type* poutput) const
|
|
{
|
|
if ((object == this->relobj())
|
|
&& (shndx == this->shndx())
|
|
&& (offset >= 0)
|
|
&& (offset <=
|
|
convert_types<section_offset_type, uint32_t>(this->original_size_)))
|
|
{
|
|
*poutput = offset;
|
|
return true;
|
|
}
|
|
else
|
|
return false;
|
|
}
|
|
|
|
private:
|
|
// Copying is not allowed.
|
|
AArch64_input_section(const AArch64_input_section&);
|
|
AArch64_input_section& operator=(const AArch64_input_section&);
|
|
|
|
// The relocation stubs.
|
|
The_stub_table* stub_table_;
|
|
// Original section contents. We have to make a copy here since the file
|
|
// containing the original section may not be locked when we need to access
|
|
// the contents.
|
|
unsigned char* original_contents_;
|
|
// Section size of the original input section.
|
|
uint32_t original_size_;
|
|
// Address alignment of the original input section.
|
|
uint32_t original_addralign_;
|
|
}; // End of AArch64_input_section
|
|
|
|
|
|
// Finalize data size.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
AArch64_input_section<size, big_endian>::set_final_data_size()
|
|
{
|
|
off_t off = convert_types<off_t, uint64_t>(this->original_size_);
|
|
|
|
if (this->is_stub_table_owner())
|
|
{
|
|
this->stub_table_->finalize_data_size();
|
|
off = align_address(off, this->stub_table_->addralign());
|
|
off += this->stub_table_->data_size();
|
|
}
|
|
this->set_data_size(off);
|
|
}
|
|
|
|
|
|
// Reset address and file offset.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
AArch64_input_section<size, big_endian>::do_reset_address_and_file_offset()
|
|
{
|
|
// Size of the original input section contents.
|
|
off_t off = convert_types<off_t, uint64_t>(this->original_size_);
|
|
|
|
// If this is a stub table owner, account for the stub table size.
|
|
if (this->is_stub_table_owner())
|
|
{
|
|
The_stub_table* stub_table = this->stub_table_;
|
|
|
|
// Reset the stub table's address and file offset. The
|
|
// current data size for child will be updated after that.
|
|
stub_table_->reset_address_and_file_offset();
|
|
off = align_address(off, stub_table_->addralign());
|
|
off += stub_table->current_data_size();
|
|
}
|
|
|
|
this->set_current_data_size(off);
|
|
}
|
|
|
|
|
|
// Initialize an Arm_input_section.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
AArch64_input_section<size, big_endian>::init()
|
|
{
|
|
Relobj* relobj = this->relobj();
|
|
unsigned int shndx = this->shndx();
|
|
|
|
// We have to cache original size, alignment and contents to avoid locking
|
|
// the original file.
|
|
this->original_addralign_ =
|
|
convert_types<uint32_t, uint64_t>(relobj->section_addralign(shndx));
|
|
|
|
// This is not efficient but we expect only a small number of relaxed
|
|
// input sections for stubs.
|
|
section_size_type section_size;
|
|
const unsigned char* section_contents =
|
|
relobj->section_contents(shndx, §ion_size, false);
|
|
this->original_size_ =
|
|
convert_types<uint32_t, uint64_t>(relobj->section_size(shndx));
|
|
|
|
gold_assert(this->original_contents_ == NULL);
|
|
this->original_contents_ = new unsigned char[section_size];
|
|
memcpy(this->original_contents_, section_contents, section_size);
|
|
|
|
// We want to make this look like the original input section after
|
|
// output sections are finalized.
|
|
Output_section* os = relobj->output_section(shndx);
|
|
off_t offset = relobj->output_section_offset(shndx);
|
|
gold_assert(os != NULL && !relobj->is_output_section_offset_invalid(shndx));
|
|
this->set_address(os->address() + offset);
|
|
this->set_file_offset(os->offset() + offset);
|
|
this->set_current_data_size(this->original_size_);
|
|
this->finalize_data_size();
|
|
}
|
|
|
|
|
|
// Write data to output file.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
AArch64_input_section<size, big_endian>::do_write(Output_file* of)
|
|
{
|
|
// We have to write out the original section content.
|
|
gold_assert(this->original_contents_ != NULL);
|
|
of->write(this->offset(), this->original_contents_,
|
|
this->original_size_);
|
|
|
|
// If this owns a stub table and it is not empty, write it.
|
|
if (this->is_stub_table_owner() && !this->stub_table_->empty())
|
|
this->stub_table_->write(of);
|
|
}
|
|
|
|
|
|
// Arm output section class. This is defined mainly to add a number of stub
|
|
// generation methods.
|
|
|
|
template<int size, bool big_endian>
|
|
class AArch64_output_section : public Output_section
|
|
{
|
|
public:
|
|
typedef Target_aarch64<size, big_endian> The_target_aarch64;
|
|
typedef AArch64_relobj<size, big_endian> The_aarch64_relobj;
|
|
typedef Stub_table<size, big_endian> The_stub_table;
|
|
typedef AArch64_input_section<size, big_endian> The_aarch64_input_section;
|
|
|
|
public:
|
|
AArch64_output_section(const char* name, elfcpp::Elf_Word type,
|
|
elfcpp::Elf_Xword flags)
|
|
: Output_section(name, type, flags)
|
|
{ }
|
|
|
|
~AArch64_output_section() {}
|
|
|
|
// Group input sections for stub generation.
|
|
void
|
|
group_sections(section_size_type, bool, Target_aarch64<size, big_endian>*,
|
|
const Task*);
|
|
|
|
private:
|
|
typedef Output_section::Input_section Input_section;
|
|
typedef Output_section::Input_section_list Input_section_list;
|
|
|
|
// Create a stub group.
|
|
void
|
|
create_stub_group(Input_section_list::const_iterator,
|
|
Input_section_list::const_iterator,
|
|
Input_section_list::const_iterator,
|
|
The_target_aarch64*,
|
|
std::vector<Output_relaxed_input_section*>&,
|
|
const Task*);
|
|
}; // End of AArch64_output_section
|
|
|
|
|
|
// Create a stub group for input sections from FIRST to LAST. OWNER points to
|
|
// the input section that will be the owner of the stub table.
|
|
|
|
template<int size, bool big_endian> void
|
|
AArch64_output_section<size, big_endian>::create_stub_group(
|
|
Input_section_list::const_iterator first,
|
|
Input_section_list::const_iterator last,
|
|
Input_section_list::const_iterator owner,
|
|
The_target_aarch64* target,
|
|
std::vector<Output_relaxed_input_section*>& new_relaxed_sections,
|
|
const Task* task)
|
|
{
|
|
// Currently we convert ordinary input sections into relaxed sections only
|
|
// at this point.
|
|
The_aarch64_input_section* input_section;
|
|
if (owner->is_relaxed_input_section())
|
|
gold_unreachable();
|
|
else
|
|
{
|
|
gold_assert(owner->is_input_section());
|
|
// Create a new relaxed input section. We need to lock the original
|
|
// file.
|
|
Task_lock_obj<Object> tl(task, owner->relobj());
|
|
input_section =
|
|
target->new_aarch64_input_section(owner->relobj(), owner->shndx());
|
|
new_relaxed_sections.push_back(input_section);
|
|
}
|
|
|
|
// Create a stub table.
|
|
The_stub_table* stub_table =
|
|
target->new_stub_table(input_section);
|
|
|
|
input_section->set_stub_table(stub_table);
|
|
|
|
Input_section_list::const_iterator p = first;
|
|
// Look for input sections or relaxed input sections in [first ... last].
|
|
do
|
|
{
|
|
if (p->is_input_section() || p->is_relaxed_input_section())
|
|
{
|
|
// The stub table information for input sections live
|
|
// in their objects.
|
|
The_aarch64_relobj* aarch64_relobj =
|
|
static_cast<The_aarch64_relobj*>(p->relobj());
|
|
aarch64_relobj->set_stub_table(p->shndx(), stub_table);
|
|
}
|
|
}
|
|
while (p++ != last);
|
|
}
|
|
|
|
|
|
// Group input sections for stub generation. GROUP_SIZE is roughly the limit of
|
|
// stub groups. We grow a stub group by adding input section until the size is
|
|
// just below GROUP_SIZE. The last input section will be converted into a stub
|
|
// table owner. If STUB_ALWAYS_AFTER_BRANCH is false, we also add input sectiond
|
|
// after the stub table, effectively doubling the group size.
|
|
//
|
|
// This is similar to the group_sections() function in elf32-arm.c but is
|
|
// implemented differently.
|
|
|
|
template<int size, bool big_endian>
|
|
void AArch64_output_section<size, big_endian>::group_sections(
|
|
section_size_type group_size,
|
|
bool stubs_always_after_branch,
|
|
Target_aarch64<size, big_endian>* target,
|
|
const Task* task)
|
|
{
|
|
typedef enum
|
|
{
|
|
NO_GROUP,
|
|
FINDING_STUB_SECTION,
|
|
HAS_STUB_SECTION
|
|
} State;
|
|
|
|
std::vector<Output_relaxed_input_section*> new_relaxed_sections;
|
|
|
|
State state = NO_GROUP;
|
|
section_size_type off = 0;
|
|
section_size_type group_begin_offset = 0;
|
|
section_size_type group_end_offset = 0;
|
|
section_size_type stub_table_end_offset = 0;
|
|
Input_section_list::const_iterator group_begin =
|
|
this->input_sections().end();
|
|
Input_section_list::const_iterator stub_table =
|
|
this->input_sections().end();
|
|
Input_section_list::const_iterator group_end = this->input_sections().end();
|
|
for (Input_section_list::const_iterator p = this->input_sections().begin();
|
|
p != this->input_sections().end();
|
|
++p)
|
|
{
|
|
section_size_type section_begin_offset =
|
|
align_address(off, p->addralign());
|
|
section_size_type section_end_offset =
|
|
section_begin_offset + p->data_size();
|
|
|
|
// Check to see if we should group the previously seen sections.
|
|
switch (state)
|
|
{
|
|
case NO_GROUP:
|
|
break;
|
|
|
|
case FINDING_STUB_SECTION:
|
|
// Adding this section makes the group larger than GROUP_SIZE.
|
|
if (section_end_offset - group_begin_offset >= group_size)
|
|
{
|
|
if (stubs_always_after_branch)
|
|
{
|
|
gold_assert(group_end != this->input_sections().end());
|
|
this->create_stub_group(group_begin, group_end, group_end,
|
|
target, new_relaxed_sections,
|
|
task);
|
|
state = NO_GROUP;
|
|
}
|
|
else
|
|
{
|
|
// Input sections up to stub_group_size bytes after the stub
|
|
// table can be handled by it too.
|
|
state = HAS_STUB_SECTION;
|
|
stub_table = group_end;
|
|
stub_table_end_offset = group_end_offset;
|
|
}
|
|
}
|
|
break;
|
|
|
|
case HAS_STUB_SECTION:
|
|
// Adding this section makes the post stub-section group larger
|
|
// than GROUP_SIZE.
|
|
gold_unreachable();
|
|
// NOT SUPPORTED YET. For completeness only.
|
|
if (section_end_offset - stub_table_end_offset >= group_size)
|
|
{
|
|
gold_assert(group_end != this->input_sections().end());
|
|
this->create_stub_group(group_begin, group_end, stub_table,
|
|
target, new_relaxed_sections, task);
|
|
state = NO_GROUP;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
gold_unreachable();
|
|
}
|
|
|
|
// If we see an input section and currently there is no group, start
|
|
// a new one. Skip any empty sections. We look at the data size
|
|
// instead of calling p->relobj()->section_size() to avoid locking.
|
|
if ((p->is_input_section() || p->is_relaxed_input_section())
|
|
&& (p->data_size() != 0))
|
|
{
|
|
if (state == NO_GROUP)
|
|
{
|
|
state = FINDING_STUB_SECTION;
|
|
group_begin = p;
|
|
group_begin_offset = section_begin_offset;
|
|
}
|
|
|
|
// Keep track of the last input section seen.
|
|
group_end = p;
|
|
group_end_offset = section_end_offset;
|
|
}
|
|
|
|
off = section_end_offset;
|
|
}
|
|
|
|
// Create a stub group for any ungrouped sections.
|
|
if (state == FINDING_STUB_SECTION || state == HAS_STUB_SECTION)
|
|
{
|
|
gold_assert(group_end != this->input_sections().end());
|
|
this->create_stub_group(group_begin, group_end,
|
|
(state == FINDING_STUB_SECTION
|
|
? group_end
|
|
: stub_table),
|
|
target, new_relaxed_sections, task);
|
|
}
|
|
|
|
if (!new_relaxed_sections.empty())
|
|
this->convert_input_sections_to_relaxed_sections(new_relaxed_sections);
|
|
|
|
// Update the section offsets
|
|
for (size_t i = 0; i < new_relaxed_sections.size(); ++i)
|
|
{
|
|
The_aarch64_relobj* relobj = static_cast<The_aarch64_relobj*>(
|
|
new_relaxed_sections[i]->relobj());
|
|
unsigned int shndx = new_relaxed_sections[i]->shndx();
|
|
// Tell AArch64_relobj that this input section is converted.
|
|
relobj->convert_input_section_to_relaxed_section(shndx);
|
|
}
|
|
} // End of AArch64_output_section::group_sections
|
|
|
|
|
|
AArch64_reloc_property_table* aarch64_reloc_property_table = NULL;
|
|
|
|
|
|
// The aarch64 target class.
|
|
// See the ABI at
|
|
// http://infocenter.arm.com/help/topic/com.arm.doc.ihi0056b/IHI0056B_aaelf64.pdf
|
|
template<int size, bool big_endian>
|
|
class Target_aarch64 : public Sized_target<size, big_endian>
|
|
{
|
|
public:
|
|
typedef Target_aarch64<size, big_endian> This;
|
|
typedef Output_data_reloc<elfcpp::SHT_RELA, true, size, big_endian>
|
|
Reloc_section;
|
|
typedef Relocate_info<size, big_endian> The_relocate_info;
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr Address;
|
|
typedef AArch64_relobj<size, big_endian> The_aarch64_relobj;
|
|
typedef Reloc_stub<size, big_endian> The_reloc_stub;
|
|
typedef Erratum_stub<size, big_endian> The_erratum_stub;
|
|
typedef typename Reloc_stub<size, big_endian>::Key The_reloc_stub_key;
|
|
typedef Stub_table<size, big_endian> The_stub_table;
|
|
typedef std::vector<The_stub_table*> Stub_table_list;
|
|
typedef typename Stub_table_list::iterator Stub_table_iterator;
|
|
typedef AArch64_input_section<size, big_endian> The_aarch64_input_section;
|
|
typedef AArch64_output_section<size, big_endian> The_aarch64_output_section;
|
|
typedef Unordered_map<Section_id,
|
|
AArch64_input_section<size, big_endian>*,
|
|
Section_id_hash> AArch64_input_section_map;
|
|
typedef AArch64_insn_utilities<big_endian> Insn_utilities;
|
|
const static int TCB_SIZE = size / 8 * 2;
|
|
static const Address invalid_address = static_cast<Address>(-1);
|
|
|
|
Target_aarch64(const Target::Target_info* info = &aarch64_info)
|
|
: Sized_target<size, big_endian>(info),
|
|
got_(NULL), plt_(NULL), got_plt_(NULL), got_irelative_(NULL),
|
|
got_tlsdesc_(NULL), global_offset_table_(NULL), rela_dyn_(NULL),
|
|
rela_irelative_(NULL), copy_relocs_(elfcpp::R_AARCH64_COPY),
|
|
got_mod_index_offset_(-1U),
|
|
tlsdesc_reloc_info_(), tls_base_symbol_defined_(false),
|
|
stub_tables_(), stub_group_size_(0), aarch64_input_section_map_()
|
|
{ }
|
|
|
|
// Scan the relocations to determine unreferenced sections for
|
|
// garbage collection.
|
|
void
|
|
gc_process_relocs(Symbol_table* symtab,
|
|
Layout* layout,
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int data_shndx,
|
|
unsigned int sh_type,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
bool needs_special_offset_handling,
|
|
size_t local_symbol_count,
|
|
const unsigned char* plocal_symbols);
|
|
|
|
// Scan the relocations to look for symbol adjustments.
|
|
void
|
|
scan_relocs(Symbol_table* symtab,
|
|
Layout* layout,
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int data_shndx,
|
|
unsigned int sh_type,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
bool needs_special_offset_handling,
|
|
size_t local_symbol_count,
|
|
const unsigned char* plocal_symbols);
|
|
|
|
// Finalize the sections.
|
|
void
|
|
do_finalize_sections(Layout*, const Input_objects*, Symbol_table*);
|
|
|
|
// Return the value to use for a dynamic which requires special
|
|
// treatment.
|
|
uint64_t
|
|
do_dynsym_value(const Symbol*) const;
|
|
|
|
// Relocate a section.
|
|
void
|
|
relocate_section(const Relocate_info<size, big_endian>*,
|
|
unsigned int sh_type,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
bool needs_special_offset_handling,
|
|
unsigned char* view,
|
|
typename elfcpp::Elf_types<size>::Elf_Addr view_address,
|
|
section_size_type view_size,
|
|
const Reloc_symbol_changes*);
|
|
|
|
// Scan the relocs during a relocatable link.
|
|
void
|
|
scan_relocatable_relocs(Symbol_table* symtab,
|
|
Layout* layout,
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int data_shndx,
|
|
unsigned int sh_type,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
bool needs_special_offset_handling,
|
|
size_t local_symbol_count,
|
|
const unsigned char* plocal_symbols,
|
|
Relocatable_relocs*);
|
|
|
|
// Scan the relocs for --emit-relocs.
|
|
void
|
|
emit_relocs_scan(Symbol_table* symtab,
|
|
Layout* layout,
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int data_shndx,
|
|
unsigned int sh_type,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
bool needs_special_offset_handling,
|
|
size_t local_symbol_count,
|
|
const unsigned char* plocal_syms,
|
|
Relocatable_relocs* rr);
|
|
|
|
// Relocate a section during a relocatable link.
|
|
void
|
|
relocate_relocs(
|
|
const Relocate_info<size, big_endian>*,
|
|
unsigned int sh_type,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
typename elfcpp::Elf_types<size>::Elf_Off offset_in_output_section,
|
|
unsigned char* view,
|
|
typename elfcpp::Elf_types<size>::Elf_Addr view_address,
|
|
section_size_type view_size,
|
|
unsigned char* reloc_view,
|
|
section_size_type reloc_view_size);
|
|
|
|
// Return the symbol index to use for a target specific relocation.
|
|
// The only target specific relocation is R_AARCH64_TLSDESC for a
|
|
// local symbol, which is an absolute reloc.
|
|
unsigned int
|
|
do_reloc_symbol_index(void*, unsigned int r_type) const
|
|
{
|
|
gold_assert(r_type == elfcpp::R_AARCH64_TLSDESC);
|
|
return 0;
|
|
}
|
|
|
|
// Return the addend to use for a target specific relocation.
|
|
uint64_t
|
|
do_reloc_addend(void* arg, unsigned int r_type, uint64_t addend) const;
|
|
|
|
// Return the PLT section.
|
|
uint64_t
|
|
do_plt_address_for_global(const Symbol* gsym) const
|
|
{ return this->plt_section()->address_for_global(gsym); }
|
|
|
|
uint64_t
|
|
do_plt_address_for_local(const Relobj* relobj, unsigned int symndx) const
|
|
{ return this->plt_section()->address_for_local(relobj, symndx); }
|
|
|
|
// This function should be defined in targets that can use relocation
|
|
// types to determine (implemented in local_reloc_may_be_function_pointer
|
|
// and global_reloc_may_be_function_pointer)
|
|
// if a function's pointer is taken. ICF uses this in safe mode to only
|
|
// fold those functions whose pointer is defintely not taken.
|
|
bool
|
|
do_can_check_for_function_pointers() const
|
|
{ return true; }
|
|
|
|
// Return the number of entries in the PLT.
|
|
unsigned int
|
|
plt_entry_count() const;
|
|
|
|
//Return the offset of the first non-reserved PLT entry.
|
|
unsigned int
|
|
first_plt_entry_offset() const;
|
|
|
|
// Return the size of each PLT entry.
|
|
unsigned int
|
|
plt_entry_size() const;
|
|
|
|
// Create a stub table.
|
|
The_stub_table*
|
|
new_stub_table(The_aarch64_input_section*);
|
|
|
|
// Create an aarch64 input section.
|
|
The_aarch64_input_section*
|
|
new_aarch64_input_section(Relobj*, unsigned int);
|
|
|
|
// Find an aarch64 input section instance for a given OBJ and SHNDX.
|
|
The_aarch64_input_section*
|
|
find_aarch64_input_section(Relobj*, unsigned int) const;
|
|
|
|
// Return the thread control block size.
|
|
unsigned int
|
|
tcb_size() const { return This::TCB_SIZE; }
|
|
|
|
// Scan a section for stub generation.
|
|
void
|
|
scan_section_for_stubs(const Relocate_info<size, big_endian>*, unsigned int,
|
|
const unsigned char*, size_t, Output_section*,
|
|
bool, const unsigned char*,
|
|
Address,
|
|
section_size_type);
|
|
|
|
// Scan a relocation section for stub.
|
|
template<int sh_type>
|
|
void
|
|
scan_reloc_section_for_stubs(
|
|
const The_relocate_info* relinfo,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
bool needs_special_offset_handling,
|
|
const unsigned char* view,
|
|
Address view_address,
|
|
section_size_type);
|
|
|
|
// Relocate a single reloc stub.
|
|
void
|
|
relocate_reloc_stub(The_reloc_stub*, const Relocate_info<size, big_endian>*,
|
|
Output_section*, unsigned char*, Address,
|
|
section_size_type);
|
|
|
|
// Get the default AArch64 target.
|
|
static This*
|
|
current_target()
|
|
{
|
|
gold_assert(parameters->target().machine_code() == elfcpp::EM_AARCH64
|
|
&& parameters->target().get_size() == size
|
|
&& parameters->target().is_big_endian() == big_endian);
|
|
return static_cast<This*>(parameters->sized_target<size, big_endian>());
|
|
}
|
|
|
|
|
|
// Scan erratum 843419 for a part of a section.
|
|
void
|
|
scan_erratum_843419_span(
|
|
AArch64_relobj<size, big_endian>*,
|
|
unsigned int,
|
|
const section_size_type,
|
|
const section_size_type,
|
|
unsigned char*,
|
|
Address);
|
|
|
|
// Scan erratum 835769 for a part of a section.
|
|
void
|
|
scan_erratum_835769_span(
|
|
AArch64_relobj<size, big_endian>*,
|
|
unsigned int,
|
|
const section_size_type,
|
|
const section_size_type,
|
|
unsigned char*,
|
|
Address);
|
|
|
|
protected:
|
|
void
|
|
do_select_as_default_target()
|
|
{
|
|
gold_assert(aarch64_reloc_property_table == NULL);
|
|
aarch64_reloc_property_table = new AArch64_reloc_property_table();
|
|
}
|
|
|
|
// Add a new reloc argument, returning the index in the vector.
|
|
size_t
|
|
add_tlsdesc_info(Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int r_sym)
|
|
{
|
|
this->tlsdesc_reloc_info_.push_back(Tlsdesc_info(object, r_sym));
|
|
return this->tlsdesc_reloc_info_.size() - 1;
|
|
}
|
|
|
|
virtual Output_data_plt_aarch64<size, big_endian>*
|
|
do_make_data_plt(Layout* layout,
|
|
Output_data_got_aarch64<size, big_endian>* got,
|
|
Output_data_space* got_plt,
|
|
Output_data_space* got_irelative)
|
|
{
|
|
return new Output_data_plt_aarch64_standard<size, big_endian>(
|
|
layout, got, got_plt, got_irelative);
|
|
}
|
|
|
|
|
|
// do_make_elf_object to override the same function in the base class.
|
|
Object*
|
|
do_make_elf_object(const std::string&, Input_file*, off_t,
|
|
const elfcpp::Ehdr<size, big_endian>&);
|
|
|
|
Output_data_plt_aarch64<size, big_endian>*
|
|
make_data_plt(Layout* layout,
|
|
Output_data_got_aarch64<size, big_endian>* got,
|
|
Output_data_space* got_plt,
|
|
Output_data_space* got_irelative)
|
|
{
|
|
return this->do_make_data_plt(layout, got, got_plt, got_irelative);
|
|
}
|
|
|
|
// We only need to generate stubs, and hence perform relaxation if we are
|
|
// not doing relocatable linking.
|
|
virtual bool
|
|
do_may_relax() const
|
|
{ return !parameters->options().relocatable(); }
|
|
|
|
// Relaxation hook. This is where we do stub generation.
|
|
virtual bool
|
|
do_relax(int, const Input_objects*, Symbol_table*, Layout*, const Task*);
|
|
|
|
void
|
|
group_sections(Layout* layout,
|
|
section_size_type group_size,
|
|
bool stubs_always_after_branch,
|
|
const Task* task);
|
|
|
|
void
|
|
scan_reloc_for_stub(const The_relocate_info*, unsigned int,
|
|
const Sized_symbol<size>*, unsigned int,
|
|
const Symbol_value<size>*,
|
|
typename elfcpp::Elf_types<size>::Elf_Swxword,
|
|
Address Elf_Addr);
|
|
|
|
// Make an output section.
|
|
Output_section*
|
|
do_make_output_section(const char* name, elfcpp::Elf_Word type,
|
|
elfcpp::Elf_Xword flags)
|
|
{ return new The_aarch64_output_section(name, type, flags); }
|
|
|
|
private:
|
|
// The class which scans relocations.
|
|
class Scan
|
|
{
|
|
public:
|
|
Scan()
|
|
: issued_non_pic_error_(false)
|
|
{ }
|
|
|
|
inline void
|
|
local(Symbol_table* symtab, Layout* layout, Target_aarch64* target,
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int data_shndx,
|
|
Output_section* output_section,
|
|
const elfcpp::Rela<size, big_endian>& reloc, unsigned int r_type,
|
|
const elfcpp::Sym<size, big_endian>& lsym,
|
|
bool is_discarded);
|
|
|
|
inline void
|
|
global(Symbol_table* symtab, Layout* layout, Target_aarch64* target,
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int data_shndx,
|
|
Output_section* output_section,
|
|
const elfcpp::Rela<size, big_endian>& reloc, unsigned int r_type,
|
|
Symbol* gsym);
|
|
|
|
inline bool
|
|
local_reloc_may_be_function_pointer(Symbol_table* , Layout* ,
|
|
Target_aarch64<size, big_endian>* ,
|
|
Sized_relobj_file<size, big_endian>* ,
|
|
unsigned int ,
|
|
Output_section* ,
|
|
const elfcpp::Rela<size, big_endian>& ,
|
|
unsigned int r_type,
|
|
const elfcpp::Sym<size, big_endian>&);
|
|
|
|
inline bool
|
|
global_reloc_may_be_function_pointer(Symbol_table* , Layout* ,
|
|
Target_aarch64<size, big_endian>* ,
|
|
Sized_relobj_file<size, big_endian>* ,
|
|
unsigned int ,
|
|
Output_section* ,
|
|
const elfcpp::Rela<size, big_endian>& ,
|
|
unsigned int r_type,
|
|
Symbol* gsym);
|
|
|
|
private:
|
|
static void
|
|
unsupported_reloc_local(Sized_relobj_file<size, big_endian>*,
|
|
unsigned int r_type);
|
|
|
|
static void
|
|
unsupported_reloc_global(Sized_relobj_file<size, big_endian>*,
|
|
unsigned int r_type, Symbol*);
|
|
|
|
inline bool
|
|
possible_function_pointer_reloc(unsigned int r_type);
|
|
|
|
void
|
|
check_non_pic(Relobj*, unsigned int r_type);
|
|
|
|
bool
|
|
reloc_needs_plt_for_ifunc(Sized_relobj_file<size, big_endian>*,
|
|
unsigned int r_type);
|
|
|
|
// Whether we have issued an error about a non-PIC compilation.
|
|
bool issued_non_pic_error_;
|
|
};
|
|
|
|
// The class which implements relocation.
|
|
class Relocate
|
|
{
|
|
public:
|
|
Relocate()
|
|
: skip_call_tls_get_addr_(false)
|
|
{ }
|
|
|
|
~Relocate()
|
|
{ }
|
|
|
|
// Do a relocation. Return false if the caller should not issue
|
|
// any warnings about this relocation.
|
|
inline bool
|
|
relocate(const Relocate_info<size, big_endian>*, unsigned int,
|
|
Target_aarch64*, Output_section*, size_t, const unsigned char*,
|
|
const Sized_symbol<size>*, const Symbol_value<size>*,
|
|
unsigned char*, typename elfcpp::Elf_types<size>::Elf_Addr,
|
|
section_size_type);
|
|
|
|
private:
|
|
inline typename AArch64_relocate_functions<size, big_endian>::Status
|
|
relocate_tls(const Relocate_info<size, big_endian>*,
|
|
Target_aarch64<size, big_endian>*,
|
|
size_t,
|
|
const elfcpp::Rela<size, big_endian>&,
|
|
unsigned int r_type, const Sized_symbol<size>*,
|
|
const Symbol_value<size>*,
|
|
unsigned char*,
|
|
typename elfcpp::Elf_types<size>::Elf_Addr);
|
|
|
|
inline typename AArch64_relocate_functions<size, big_endian>::Status
|
|
tls_gd_to_le(
|
|
const Relocate_info<size, big_endian>*,
|
|
Target_aarch64<size, big_endian>*,
|
|
const elfcpp::Rela<size, big_endian>&,
|
|
unsigned int,
|
|
unsigned char*,
|
|
const Symbol_value<size>*);
|
|
|
|
inline typename AArch64_relocate_functions<size, big_endian>::Status
|
|
tls_ld_to_le(
|
|
const Relocate_info<size, big_endian>*,
|
|
Target_aarch64<size, big_endian>*,
|
|
const elfcpp::Rela<size, big_endian>&,
|
|
unsigned int,
|
|
unsigned char*,
|
|
const Symbol_value<size>*);
|
|
|
|
inline typename AArch64_relocate_functions<size, big_endian>::Status
|
|
tls_ie_to_le(
|
|
const Relocate_info<size, big_endian>*,
|
|
Target_aarch64<size, big_endian>*,
|
|
const elfcpp::Rela<size, big_endian>&,
|
|
unsigned int,
|
|
unsigned char*,
|
|
const Symbol_value<size>*);
|
|
|
|
inline typename AArch64_relocate_functions<size, big_endian>::Status
|
|
tls_desc_gd_to_le(
|
|
const Relocate_info<size, big_endian>*,
|
|
Target_aarch64<size, big_endian>*,
|
|
const elfcpp::Rela<size, big_endian>&,
|
|
unsigned int,
|
|
unsigned char*,
|
|
const Symbol_value<size>*);
|
|
|
|
inline typename AArch64_relocate_functions<size, big_endian>::Status
|
|
tls_desc_gd_to_ie(
|
|
const Relocate_info<size, big_endian>*,
|
|
Target_aarch64<size, big_endian>*,
|
|
const elfcpp::Rela<size, big_endian>&,
|
|
unsigned int,
|
|
unsigned char*,
|
|
const Symbol_value<size>*,
|
|
typename elfcpp::Elf_types<size>::Elf_Addr,
|
|
typename elfcpp::Elf_types<size>::Elf_Addr);
|
|
|
|
bool skip_call_tls_get_addr_;
|
|
|
|
}; // End of class Relocate
|
|
|
|
// Adjust TLS relocation type based on the options and whether this
|
|
// is a local symbol.
|
|
static tls::Tls_optimization
|
|
optimize_tls_reloc(bool is_final, int r_type);
|
|
|
|
// Get the GOT section, creating it if necessary.
|
|
Output_data_got_aarch64<size, big_endian>*
|
|
got_section(Symbol_table*, Layout*);
|
|
|
|
// Get the GOT PLT section.
|
|
Output_data_space*
|
|
got_plt_section() const
|
|
{
|
|
gold_assert(this->got_plt_ != NULL);
|
|
return this->got_plt_;
|
|
}
|
|
|
|
// Get the GOT section for TLSDESC entries.
|
|
Output_data_got<size, big_endian>*
|
|
got_tlsdesc_section() const
|
|
{
|
|
gold_assert(this->got_tlsdesc_ != NULL);
|
|
return this->got_tlsdesc_;
|
|
}
|
|
|
|
// Create the PLT section.
|
|
void
|
|
make_plt_section(Symbol_table* symtab, Layout* layout);
|
|
|
|
// Create a PLT entry for a global symbol.
|
|
void
|
|
make_plt_entry(Symbol_table*, Layout*, Symbol*);
|
|
|
|
// Create a PLT entry for a local STT_GNU_IFUNC symbol.
|
|
void
|
|
make_local_ifunc_plt_entry(Symbol_table*, Layout*,
|
|
Sized_relobj_file<size, big_endian>* relobj,
|
|
unsigned int local_sym_index);
|
|
|
|
// Define the _TLS_MODULE_BASE_ symbol in the TLS segment.
|
|
void
|
|
define_tls_base_symbol(Symbol_table*, Layout*);
|
|
|
|
// Create the reserved PLT and GOT entries for the TLS descriptor resolver.
|
|
void
|
|
reserve_tlsdesc_entries(Symbol_table* symtab, Layout* layout);
|
|
|
|
// Create a GOT entry for the TLS module index.
|
|
unsigned int
|
|
got_mod_index_entry(Symbol_table* symtab, Layout* layout,
|
|
Sized_relobj_file<size, big_endian>* object);
|
|
|
|
// Get the PLT section.
|
|
Output_data_plt_aarch64<size, big_endian>*
|
|
plt_section() const
|
|
{
|
|
gold_assert(this->plt_ != NULL);
|
|
return this->plt_;
|
|
}
|
|
|
|
// Helper method to create erratum stubs for ST_E_843419 and ST_E_835769. For
|
|
// ST_E_843419, we need an additional field for adrp offset.
|
|
void create_erratum_stub(
|
|
AArch64_relobj<size, big_endian>* relobj,
|
|
unsigned int shndx,
|
|
section_size_type erratum_insn_offset,
|
|
Address erratum_address,
|
|
typename Insn_utilities::Insntype erratum_insn,
|
|
int erratum_type,
|
|
unsigned int e843419_adrp_offset=0);
|
|
|
|
// Return whether this is a 3-insn erratum sequence.
|
|
bool is_erratum_843419_sequence(
|
|
typename elfcpp::Swap<32,big_endian>::Valtype insn1,
|
|
typename elfcpp::Swap<32,big_endian>::Valtype insn2,
|
|
typename elfcpp::Swap<32,big_endian>::Valtype insn3);
|
|
|
|
// Return whether this is a 835769 sequence.
|
|
// (Similarly implemented as in elfnn-aarch64.c.)
|
|
bool is_erratum_835769_sequence(
|
|
typename elfcpp::Swap<32,big_endian>::Valtype,
|
|
typename elfcpp::Swap<32,big_endian>::Valtype);
|
|
|
|
// Get the dynamic reloc section, creating it if necessary.
|
|
Reloc_section*
|
|
rela_dyn_section(Layout*);
|
|
|
|
// Get the section to use for TLSDESC relocations.
|
|
Reloc_section*
|
|
rela_tlsdesc_section(Layout*) const;
|
|
|
|
// Get the section to use for IRELATIVE relocations.
|
|
Reloc_section*
|
|
rela_irelative_section(Layout*);
|
|
|
|
// Add a potential copy relocation.
|
|
void
|
|
copy_reloc(Symbol_table* symtab, Layout* layout,
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int shndx, Output_section* output_section,
|
|
Symbol* sym, const elfcpp::Rela<size, big_endian>& reloc)
|
|
{
|
|
unsigned int r_type = elfcpp::elf_r_type<size>(reloc.get_r_info());
|
|
this->copy_relocs_.copy_reloc(symtab, layout,
|
|
symtab->get_sized_symbol<size>(sym),
|
|
object, shndx, output_section,
|
|
r_type, reloc.get_r_offset(),
|
|
reloc.get_r_addend(),
|
|
this->rela_dyn_section(layout));
|
|
}
|
|
|
|
// Information about this specific target which we pass to the
|
|
// general Target structure.
|
|
static const Target::Target_info aarch64_info;
|
|
|
|
// The types of GOT entries needed for this platform.
|
|
// These values are exposed to the ABI in an incremental link.
|
|
// Do not renumber existing values without changing the version
|
|
// number of the .gnu_incremental_inputs section.
|
|
enum Got_type
|
|
{
|
|
GOT_TYPE_STANDARD = 0, // GOT entry for a regular symbol
|
|
GOT_TYPE_TLS_OFFSET = 1, // GOT entry for TLS offset
|
|
GOT_TYPE_TLS_PAIR = 2, // GOT entry for TLS module/offset pair
|
|
GOT_TYPE_TLS_DESC = 3 // GOT entry for TLS_DESC pair
|
|
};
|
|
|
|
// This type is used as the argument to the target specific
|
|
// relocation routines. The only target specific reloc is
|
|
// R_AARCh64_TLSDESC against a local symbol.
|
|
struct Tlsdesc_info
|
|
{
|
|
Tlsdesc_info(Sized_relobj_file<size, big_endian>* a_object,
|
|
unsigned int a_r_sym)
|
|
: object(a_object), r_sym(a_r_sym)
|
|
{ }
|
|
|
|
// The object in which the local symbol is defined.
|
|
Sized_relobj_file<size, big_endian>* object;
|
|
// The local symbol index in the object.
|
|
unsigned int r_sym;
|
|
};
|
|
|
|
// The GOT section.
|
|
Output_data_got_aarch64<size, big_endian>* got_;
|
|
// The PLT section.
|
|
Output_data_plt_aarch64<size, big_endian>* plt_;
|
|
// The GOT PLT section.
|
|
Output_data_space* got_plt_;
|
|
// The GOT section for IRELATIVE relocations.
|
|
Output_data_space* got_irelative_;
|
|
// The GOT section for TLSDESC relocations.
|
|
Output_data_got<size, big_endian>* got_tlsdesc_;
|
|
// The _GLOBAL_OFFSET_TABLE_ symbol.
|
|
Symbol* global_offset_table_;
|
|
// The dynamic reloc section.
|
|
Reloc_section* rela_dyn_;
|
|
// The section to use for IRELATIVE relocs.
|
|
Reloc_section* rela_irelative_;
|
|
// Relocs saved to avoid a COPY reloc.
|
|
Copy_relocs<elfcpp::SHT_RELA, size, big_endian> copy_relocs_;
|
|
// Offset of the GOT entry for the TLS module index.
|
|
unsigned int got_mod_index_offset_;
|
|
// We handle R_AARCH64_TLSDESC against a local symbol as a target
|
|
// specific relocation. Here we store the object and local symbol
|
|
// index for the relocation.
|
|
std::vector<Tlsdesc_info> tlsdesc_reloc_info_;
|
|
// True if the _TLS_MODULE_BASE_ symbol has been defined.
|
|
bool tls_base_symbol_defined_;
|
|
// List of stub_tables
|
|
Stub_table_list stub_tables_;
|
|
// Actual stub group size
|
|
section_size_type stub_group_size_;
|
|
AArch64_input_section_map aarch64_input_section_map_;
|
|
}; // End of Target_aarch64
|
|
|
|
|
|
template<>
|
|
const Target::Target_info Target_aarch64<64, false>::aarch64_info =
|
|
{
|
|
64, // size
|
|
false, // is_big_endian
|
|
elfcpp::EM_AARCH64, // machine_code
|
|
false, // has_make_symbol
|
|
false, // has_resolve
|
|
false, // has_code_fill
|
|
false, // is_default_stack_executable
|
|
true, // can_icf_inline_merge_sections
|
|
'\0', // wrap_char
|
|
"/lib/ld.so.1", // program interpreter
|
|
0x400000, // default_text_segment_address
|
|
0x10000, // abi_pagesize (overridable by -z max-page-size)
|
|
0x1000, // common_pagesize (overridable by -z common-page-size)
|
|
false, // isolate_execinstr
|
|
0, // rosegment_gap
|
|
elfcpp::SHN_UNDEF, // small_common_shndx
|
|
elfcpp::SHN_UNDEF, // large_common_shndx
|
|
0, // small_common_section_flags
|
|
0, // large_common_section_flags
|
|
NULL, // attributes_section
|
|
NULL, // attributes_vendor
|
|
"_start", // entry_symbol_name
|
|
32, // hash_entry_size
|
|
elfcpp::SHT_PROGBITS, // unwind_section_type
|
|
};
|
|
|
|
template<>
|
|
const Target::Target_info Target_aarch64<32, false>::aarch64_info =
|
|
{
|
|
32, // size
|
|
false, // is_big_endian
|
|
elfcpp::EM_AARCH64, // machine_code
|
|
false, // has_make_symbol
|
|
false, // has_resolve
|
|
false, // has_code_fill
|
|
false, // is_default_stack_executable
|
|
false, // can_icf_inline_merge_sections
|
|
'\0', // wrap_char
|
|
"/lib/ld.so.1", // program interpreter
|
|
0x400000, // default_text_segment_address
|
|
0x10000, // abi_pagesize (overridable by -z max-page-size)
|
|
0x1000, // common_pagesize (overridable by -z common-page-size)
|
|
false, // isolate_execinstr
|
|
0, // rosegment_gap
|
|
elfcpp::SHN_UNDEF, // small_common_shndx
|
|
elfcpp::SHN_UNDEF, // large_common_shndx
|
|
0, // small_common_section_flags
|
|
0, // large_common_section_flags
|
|
NULL, // attributes_section
|
|
NULL, // attributes_vendor
|
|
"_start", // entry_symbol_name
|
|
32, // hash_entry_size
|
|
elfcpp::SHT_PROGBITS, // unwind_section_type
|
|
};
|
|
|
|
template<>
|
|
const Target::Target_info Target_aarch64<64, true>::aarch64_info =
|
|
{
|
|
64, // size
|
|
true, // is_big_endian
|
|
elfcpp::EM_AARCH64, // machine_code
|
|
false, // has_make_symbol
|
|
false, // has_resolve
|
|
false, // has_code_fill
|
|
false, // is_default_stack_executable
|
|
true, // can_icf_inline_merge_sections
|
|
'\0', // wrap_char
|
|
"/lib/ld.so.1", // program interpreter
|
|
0x400000, // default_text_segment_address
|
|
0x10000, // abi_pagesize (overridable by -z max-page-size)
|
|
0x1000, // common_pagesize (overridable by -z common-page-size)
|
|
false, // isolate_execinstr
|
|
0, // rosegment_gap
|
|
elfcpp::SHN_UNDEF, // small_common_shndx
|
|
elfcpp::SHN_UNDEF, // large_common_shndx
|
|
0, // small_common_section_flags
|
|
0, // large_common_section_flags
|
|
NULL, // attributes_section
|
|
NULL, // attributes_vendor
|
|
"_start", // entry_symbol_name
|
|
32, // hash_entry_size
|
|
elfcpp::SHT_PROGBITS, // unwind_section_type
|
|
};
|
|
|
|
template<>
|
|
const Target::Target_info Target_aarch64<32, true>::aarch64_info =
|
|
{
|
|
32, // size
|
|
true, // is_big_endian
|
|
elfcpp::EM_AARCH64, // machine_code
|
|
false, // has_make_symbol
|
|
false, // has_resolve
|
|
false, // has_code_fill
|
|
false, // is_default_stack_executable
|
|
false, // can_icf_inline_merge_sections
|
|
'\0', // wrap_char
|
|
"/lib/ld.so.1", // program interpreter
|
|
0x400000, // default_text_segment_address
|
|
0x10000, // abi_pagesize (overridable by -z max-page-size)
|
|
0x1000, // common_pagesize (overridable by -z common-page-size)
|
|
false, // isolate_execinstr
|
|
0, // rosegment_gap
|
|
elfcpp::SHN_UNDEF, // small_common_shndx
|
|
elfcpp::SHN_UNDEF, // large_common_shndx
|
|
0, // small_common_section_flags
|
|
0, // large_common_section_flags
|
|
NULL, // attributes_section
|
|
NULL, // attributes_vendor
|
|
"_start", // entry_symbol_name
|
|
32, // hash_entry_size
|
|
elfcpp::SHT_PROGBITS, // unwind_section_type
|
|
};
|
|
|
|
// Get the GOT section, creating it if necessary.
|
|
|
|
template<int size, bool big_endian>
|
|
Output_data_got_aarch64<size, big_endian>*
|
|
Target_aarch64<size, big_endian>::got_section(Symbol_table* symtab,
|
|
Layout* layout)
|
|
{
|
|
if (this->got_ == NULL)
|
|
{
|
|
gold_assert(symtab != NULL && layout != NULL);
|
|
|
|
// When using -z now, we can treat .got.plt as a relro section.
|
|
// Without -z now, it is modified after program startup by lazy
|
|
// PLT relocations.
|
|
bool is_got_plt_relro = parameters->options().now();
|
|
Output_section_order got_order = (is_got_plt_relro
|
|
? ORDER_RELRO
|
|
: ORDER_RELRO_LAST);
|
|
Output_section_order got_plt_order = (is_got_plt_relro
|
|
? ORDER_RELRO
|
|
: ORDER_NON_RELRO_FIRST);
|
|
|
|
// Layout of .got and .got.plt sections.
|
|
// .got[0] &_DYNAMIC <-_GLOBAL_OFFSET_TABLE_
|
|
// ...
|
|
// .gotplt[0] reserved for ld.so (&linkmap) <--DT_PLTGOT
|
|
// .gotplt[1] reserved for ld.so (resolver)
|
|
// .gotplt[2] reserved
|
|
|
|
// Generate .got section.
|
|
this->got_ = new Output_data_got_aarch64<size, big_endian>(symtab,
|
|
layout);
|
|
layout->add_output_section_data(".got", elfcpp::SHT_PROGBITS,
|
|
(elfcpp::SHF_ALLOC | elfcpp::SHF_WRITE),
|
|
this->got_, got_order, true);
|
|
// The first word of GOT is reserved for the address of .dynamic.
|
|
// We put 0 here now. The value will be replaced later in
|
|
// Output_data_got_aarch64::do_write.
|
|
this->got_->add_constant(0);
|
|
|
|
// Define _GLOBAL_OFFSET_TABLE_ at the start of the PLT.
|
|
// _GLOBAL_OFFSET_TABLE_ value points to the start of the .got section,
|
|
// even if there is a .got.plt section.
|
|
this->global_offset_table_ =
|
|
symtab->define_in_output_data("_GLOBAL_OFFSET_TABLE_", NULL,
|
|
Symbol_table::PREDEFINED,
|
|
this->got_,
|
|
0, 0, elfcpp::STT_OBJECT,
|
|
elfcpp::STB_LOCAL,
|
|
elfcpp::STV_HIDDEN, 0,
|
|
false, false);
|
|
|
|
// Generate .got.plt section.
|
|
this->got_plt_ = new Output_data_space(size / 8, "** GOT PLT");
|
|
layout->add_output_section_data(".got.plt", elfcpp::SHT_PROGBITS,
|
|
(elfcpp::SHF_ALLOC
|
|
| elfcpp::SHF_WRITE),
|
|
this->got_plt_, got_plt_order,
|
|
is_got_plt_relro);
|
|
|
|
// The first three entries are reserved.
|
|
this->got_plt_->set_current_data_size(
|
|
AARCH64_GOTPLT_RESERVE_COUNT * (size / 8));
|
|
|
|
// If there are any IRELATIVE relocations, they get GOT entries
|
|
// in .got.plt after the jump slot entries.
|
|
this->got_irelative_ = new Output_data_space(size / 8,
|
|
"** GOT IRELATIVE PLT");
|
|
layout->add_output_section_data(".got.plt", elfcpp::SHT_PROGBITS,
|
|
(elfcpp::SHF_ALLOC
|
|
| elfcpp::SHF_WRITE),
|
|
this->got_irelative_,
|
|
got_plt_order,
|
|
is_got_plt_relro);
|
|
|
|
// If there are any TLSDESC relocations, they get GOT entries in
|
|
// .got.plt after the jump slot and IRELATIVE entries.
|
|
this->got_tlsdesc_ = new Output_data_got<size, big_endian>();
|
|
layout->add_output_section_data(".got.plt", elfcpp::SHT_PROGBITS,
|
|
(elfcpp::SHF_ALLOC
|
|
| elfcpp::SHF_WRITE),
|
|
this->got_tlsdesc_,
|
|
got_plt_order,
|
|
is_got_plt_relro);
|
|
|
|
if (!is_got_plt_relro)
|
|
{
|
|
// Those bytes can go into the relro segment.
|
|
layout->increase_relro(
|
|
AARCH64_GOTPLT_RESERVE_COUNT * (size / 8));
|
|
}
|
|
|
|
}
|
|
return this->got_;
|
|
}
|
|
|
|
// Get the dynamic reloc section, creating it if necessary.
|
|
|
|
template<int size, bool big_endian>
|
|
typename Target_aarch64<size, big_endian>::Reloc_section*
|
|
Target_aarch64<size, big_endian>::rela_dyn_section(Layout* layout)
|
|
{
|
|
if (this->rela_dyn_ == NULL)
|
|
{
|
|
gold_assert(layout != NULL);
|
|
this->rela_dyn_ = new Reloc_section(parameters->options().combreloc());
|
|
layout->add_output_section_data(".rela.dyn", elfcpp::SHT_RELA,
|
|
elfcpp::SHF_ALLOC, this->rela_dyn_,
|
|
ORDER_DYNAMIC_RELOCS, false);
|
|
}
|
|
return this->rela_dyn_;
|
|
}
|
|
|
|
// Get the section to use for IRELATIVE relocs, creating it if
|
|
// necessary. These go in .rela.dyn, but only after all other dynamic
|
|
// relocations. They need to follow the other dynamic relocations so
|
|
// that they can refer to global variables initialized by those
|
|
// relocs.
|
|
|
|
template<int size, bool big_endian>
|
|
typename Target_aarch64<size, big_endian>::Reloc_section*
|
|
Target_aarch64<size, big_endian>::rela_irelative_section(Layout* layout)
|
|
{
|
|
if (this->rela_irelative_ == NULL)
|
|
{
|
|
// Make sure we have already created the dynamic reloc section.
|
|
this->rela_dyn_section(layout);
|
|
this->rela_irelative_ = new Reloc_section(false);
|
|
layout->add_output_section_data(".rela.dyn", elfcpp::SHT_RELA,
|
|
elfcpp::SHF_ALLOC, this->rela_irelative_,
|
|
ORDER_DYNAMIC_RELOCS, false);
|
|
gold_assert(this->rela_dyn_->output_section()
|
|
== this->rela_irelative_->output_section());
|
|
}
|
|
return this->rela_irelative_;
|
|
}
|
|
|
|
|
|
// do_make_elf_object to override the same function in the base class. We need
|
|
// to use a target-specific sub-class of Sized_relobj_file<size, big_endian> to
|
|
// store backend specific information. Hence we need to have our own ELF object
|
|
// creation.
|
|
|
|
template<int size, bool big_endian>
|
|
Object*
|
|
Target_aarch64<size, big_endian>::do_make_elf_object(
|
|
const std::string& name,
|
|
Input_file* input_file,
|
|
off_t offset, const elfcpp::Ehdr<size, big_endian>& ehdr)
|
|
{
|
|
int et = ehdr.get_e_type();
|
|
// ET_EXEC files are valid input for --just-symbols/-R,
|
|
// and we treat them as relocatable objects.
|
|
if (et == elfcpp::ET_EXEC && input_file->just_symbols())
|
|
return Sized_target<size, big_endian>::do_make_elf_object(
|
|
name, input_file, offset, ehdr);
|
|
else if (et == elfcpp::ET_REL)
|
|
{
|
|
AArch64_relobj<size, big_endian>* obj =
|
|
new AArch64_relobj<size, big_endian>(name, input_file, offset, ehdr);
|
|
obj->setup();
|
|
return obj;
|
|
}
|
|
else if (et == elfcpp::ET_DYN)
|
|
{
|
|
// Keep base implementation.
|
|
Sized_dynobj<size, big_endian>* obj =
|
|
new Sized_dynobj<size, big_endian>(name, input_file, offset, ehdr);
|
|
obj->setup();
|
|
return obj;
|
|
}
|
|
else
|
|
{
|
|
gold_error(_("%s: unsupported ELF file type %d"),
|
|
name.c_str(), et);
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
|
|
// Scan a relocation for stub generation.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::scan_reloc_for_stub(
|
|
const Relocate_info<size, big_endian>* relinfo,
|
|
unsigned int r_type,
|
|
const Sized_symbol<size>* gsym,
|
|
unsigned int r_sym,
|
|
const Symbol_value<size>* psymval,
|
|
typename elfcpp::Elf_types<size>::Elf_Swxword addend,
|
|
Address address)
|
|
{
|
|
const AArch64_relobj<size, big_endian>* aarch64_relobj =
|
|
static_cast<AArch64_relobj<size, big_endian>*>(relinfo->object);
|
|
|
|
Symbol_value<size> symval;
|
|
if (gsym != NULL)
|
|
{
|
|
const AArch64_reloc_property* arp = aarch64_reloc_property_table->
|
|
get_reloc_property(r_type);
|
|
if (gsym->use_plt_offset(arp->reference_flags()))
|
|
{
|
|
// This uses a PLT, change the symbol value.
|
|
symval.set_output_value(this->plt_address_for_global(gsym));
|
|
psymval = &symval;
|
|
}
|
|
else if (gsym->is_undefined())
|
|
{
|
|
// There is no need to generate a stub symbol if the original symbol
|
|
// is undefined.
|
|
gold_debug(DEBUG_TARGET,
|
|
"stub: not creating a stub for undefined symbol %s in file %s",
|
|
gsym->name(), aarch64_relobj->name().c_str());
|
|
return;
|
|
}
|
|
}
|
|
|
|
// Get the symbol value.
|
|
typename Symbol_value<size>::Value value = psymval->value(aarch64_relobj, 0);
|
|
|
|
// Owing to pipelining, the PC relative branches below actually skip
|
|
// two instructions when the branch offset is 0.
|
|
Address destination = static_cast<Address>(-1);
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_CALL26:
|
|
case elfcpp::R_AARCH64_JUMP26:
|
|
destination = value + addend;
|
|
break;
|
|
default:
|
|
gold_unreachable();
|
|
}
|
|
|
|
int stub_type = The_reloc_stub::
|
|
stub_type_for_reloc(r_type, address, destination);
|
|
if (stub_type == ST_NONE)
|
|
return;
|
|
|
|
The_stub_table* stub_table = aarch64_relobj->stub_table(relinfo->data_shndx);
|
|
gold_assert(stub_table != NULL);
|
|
|
|
The_reloc_stub_key key(stub_type, gsym, aarch64_relobj, r_sym, addend);
|
|
The_reloc_stub* stub = stub_table->find_reloc_stub(key);
|
|
if (stub == NULL)
|
|
{
|
|
stub = new The_reloc_stub(stub_type);
|
|
stub_table->add_reloc_stub(stub, key);
|
|
}
|
|
stub->set_destination_address(destination);
|
|
} // End of Target_aarch64::scan_reloc_for_stub
|
|
|
|
|
|
// This function scans a relocation section for stub generation.
|
|
// The template parameter Relocate must be a class type which provides
|
|
// a single function, relocate(), which implements the machine
|
|
// specific part of a relocation.
|
|
|
|
// BIG_ENDIAN is the endianness of the data. SH_TYPE is the section type:
|
|
// SHT_REL or SHT_RELA.
|
|
|
|
// PRELOCS points to the relocation data. RELOC_COUNT is the number
|
|
// of relocs. OUTPUT_SECTION is the output section.
|
|
// NEEDS_SPECIAL_OFFSET_HANDLING is true if input offsets need to be
|
|
// mapped to output offsets.
|
|
|
|
// VIEW is the section data, VIEW_ADDRESS is its memory address, and
|
|
// VIEW_SIZE is the size. These refer to the input section, unless
|
|
// NEEDS_SPECIAL_OFFSET_HANDLING is true, in which case they refer to
|
|
// the output section.
|
|
|
|
template<int size, bool big_endian>
|
|
template<int sh_type>
|
|
void inline
|
|
Target_aarch64<size, big_endian>::scan_reloc_section_for_stubs(
|
|
const Relocate_info<size, big_endian>* relinfo,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* /*output_section*/,
|
|
bool /*needs_special_offset_handling*/,
|
|
const unsigned char* /*view*/,
|
|
Address view_address,
|
|
section_size_type)
|
|
{
|
|
typedef typename Reloc_types<sh_type,size,big_endian>::Reloc Reltype;
|
|
|
|
const int reloc_size =
|
|
Reloc_types<sh_type,size,big_endian>::reloc_size;
|
|
AArch64_relobj<size, big_endian>* object =
|
|
static_cast<AArch64_relobj<size, big_endian>*>(relinfo->object);
|
|
unsigned int local_count = object->local_symbol_count();
|
|
|
|
gold::Default_comdat_behavior default_comdat_behavior;
|
|
Comdat_behavior comdat_behavior = CB_UNDETERMINED;
|
|
|
|
for (size_t i = 0; i < reloc_count; ++i, prelocs += reloc_size)
|
|
{
|
|
Reltype reloc(prelocs);
|
|
typename elfcpp::Elf_types<size>::Elf_WXword r_info = reloc.get_r_info();
|
|
unsigned int r_sym = elfcpp::elf_r_sym<size>(r_info);
|
|
unsigned int r_type = elfcpp::elf_r_type<size>(r_info);
|
|
if (r_type != elfcpp::R_AARCH64_CALL26
|
|
&& r_type != elfcpp::R_AARCH64_JUMP26)
|
|
continue;
|
|
|
|
section_offset_type offset =
|
|
convert_to_section_size_type(reloc.get_r_offset());
|
|
|
|
// Get the addend.
|
|
typename elfcpp::Elf_types<size>::Elf_Swxword addend =
|
|
reloc.get_r_addend();
|
|
|
|
const Sized_symbol<size>* sym;
|
|
Symbol_value<size> symval;
|
|
const Symbol_value<size> *psymval;
|
|
bool is_defined_in_discarded_section;
|
|
unsigned int shndx;
|
|
const Symbol* gsym = NULL;
|
|
if (r_sym < local_count)
|
|
{
|
|
sym = NULL;
|
|
psymval = object->local_symbol(r_sym);
|
|
|
|
// If the local symbol belongs to a section we are discarding,
|
|
// and that section is a debug section, try to find the
|
|
// corresponding kept section and map this symbol to its
|
|
// counterpart in the kept section. The symbol must not
|
|
// correspond to a section we are folding.
|
|
bool is_ordinary;
|
|
shndx = psymval->input_shndx(&is_ordinary);
|
|
is_defined_in_discarded_section =
|
|
(is_ordinary
|
|
&& shndx != elfcpp::SHN_UNDEF
|
|
&& !object->is_section_included(shndx)
|
|
&& !relinfo->symtab->is_section_folded(object, shndx));
|
|
|
|
// We need to compute the would-be final value of this local
|
|
// symbol.
|
|
if (!is_defined_in_discarded_section)
|
|
{
|
|
typedef Sized_relobj_file<size, big_endian> ObjType;
|
|
if (psymval->is_section_symbol())
|
|
symval.set_is_section_symbol();
|
|
typename ObjType::Compute_final_local_value_status status =
|
|
object->compute_final_local_value(r_sym, psymval, &symval,
|
|
relinfo->symtab);
|
|
if (status == ObjType::CFLV_OK)
|
|
{
|
|
// Currently we cannot handle a branch to a target in
|
|
// a merged section. If this is the case, issue an error
|
|
// and also free the merge symbol value.
|
|
if (!symval.has_output_value())
|
|
{
|
|
const std::string& section_name =
|
|
object->section_name(shndx);
|
|
object->error(_("cannot handle branch to local %u "
|
|
"in a merged section %s"),
|
|
r_sym, section_name.c_str());
|
|
}
|
|
psymval = &symval;
|
|
}
|
|
else
|
|
{
|
|
// We cannot determine the final value.
|
|
continue;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
gsym = object->global_symbol(r_sym);
|
|
gold_assert(gsym != NULL);
|
|
if (gsym->is_forwarder())
|
|
gsym = relinfo->symtab->resolve_forwards(gsym);
|
|
|
|
sym = static_cast<const Sized_symbol<size>*>(gsym);
|
|
if (sym->has_symtab_index() && sym->symtab_index() != -1U)
|
|
symval.set_output_symtab_index(sym->symtab_index());
|
|
else
|
|
symval.set_no_output_symtab_entry();
|
|
|
|
// We need to compute the would-be final value of this global
|
|
// symbol.
|
|
const Symbol_table* symtab = relinfo->symtab;
|
|
const Sized_symbol<size>* sized_symbol =
|
|
symtab->get_sized_symbol<size>(gsym);
|
|
Symbol_table::Compute_final_value_status status;
|
|
typename elfcpp::Elf_types<size>::Elf_Addr value =
|
|
symtab->compute_final_value<size>(sized_symbol, &status);
|
|
|
|
// Skip this if the symbol has not output section.
|
|
if (status == Symbol_table::CFVS_NO_OUTPUT_SECTION)
|
|
continue;
|
|
symval.set_output_value(value);
|
|
|
|
if (gsym->type() == elfcpp::STT_TLS)
|
|
symval.set_is_tls_symbol();
|
|
else if (gsym->type() == elfcpp::STT_GNU_IFUNC)
|
|
symval.set_is_ifunc_symbol();
|
|
psymval = &symval;
|
|
|
|
is_defined_in_discarded_section =
|
|
(gsym->is_defined_in_discarded_section()
|
|
&& gsym->is_undefined());
|
|
shndx = 0;
|
|
}
|
|
|
|
Symbol_value<size> symval2;
|
|
if (is_defined_in_discarded_section)
|
|
{
|
|
std::string name = object->section_name(relinfo->data_shndx);
|
|
|
|
if (comdat_behavior == CB_UNDETERMINED)
|
|
comdat_behavior = default_comdat_behavior.get(name.c_str());
|
|
|
|
if (comdat_behavior == CB_PRETEND)
|
|
{
|
|
bool found;
|
|
typename elfcpp::Elf_types<size>::Elf_Addr value =
|
|
object->map_to_kept_section(shndx, name, &found);
|
|
if (found)
|
|
symval2.set_output_value(value + psymval->input_value());
|
|
else
|
|
symval2.set_output_value(0);
|
|
}
|
|
else
|
|
{
|
|
if (comdat_behavior == CB_ERROR)
|
|
issue_discarded_error(relinfo, i, offset, r_sym, gsym);
|
|
symval2.set_output_value(0);
|
|
}
|
|
symval2.set_no_output_symtab_entry();
|
|
psymval = &symval2;
|
|
}
|
|
|
|
this->scan_reloc_for_stub(relinfo, r_type, sym, r_sym, psymval,
|
|
addend, view_address + offset);
|
|
} // End of iterating relocs in a section
|
|
} // End of Target_aarch64::scan_reloc_section_for_stubs
|
|
|
|
|
|
// Scan an input section for stub generation.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::scan_section_for_stubs(
|
|
const Relocate_info<size, big_endian>* relinfo,
|
|
unsigned int sh_type,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
bool needs_special_offset_handling,
|
|
const unsigned char* view,
|
|
Address view_address,
|
|
section_size_type view_size)
|
|
{
|
|
gold_assert(sh_type == elfcpp::SHT_RELA);
|
|
this->scan_reloc_section_for_stubs<elfcpp::SHT_RELA>(
|
|
relinfo,
|
|
prelocs,
|
|
reloc_count,
|
|
output_section,
|
|
needs_special_offset_handling,
|
|
view,
|
|
view_address,
|
|
view_size);
|
|
}
|
|
|
|
|
|
// Relocate a single reloc stub.
|
|
|
|
template<int size, bool big_endian>
|
|
void Target_aarch64<size, big_endian>::
|
|
relocate_reloc_stub(The_reloc_stub* stub,
|
|
const The_relocate_info*,
|
|
Output_section*,
|
|
unsigned char* view,
|
|
Address address,
|
|
section_size_type)
|
|
{
|
|
typedef AArch64_relocate_functions<size, big_endian> The_reloc_functions;
|
|
typedef typename The_reloc_functions::Status The_reloc_functions_status;
|
|
typedef typename elfcpp::Swap<32,big_endian>::Valtype Insntype;
|
|
|
|
Insntype* ip = reinterpret_cast<Insntype*>(view);
|
|
int insn_number = stub->insn_num();
|
|
const uint32_t* insns = stub->insns();
|
|
// Check the insns are really those stub insns.
|
|
for (int i = 0; i < insn_number; ++i)
|
|
{
|
|
Insntype insn = elfcpp::Swap<32,big_endian>::readval(ip + i);
|
|
gold_assert(((uint32_t)insn == insns[i]));
|
|
}
|
|
|
|
Address dest = stub->destination_address();
|
|
|
|
switch(stub->type())
|
|
{
|
|
case ST_ADRP_BRANCH:
|
|
{
|
|
// 1st reloc is ADR_PREL_PG_HI21
|
|
The_reloc_functions_status status =
|
|
The_reloc_functions::adrp(view, dest, address);
|
|
// An error should never arise in the above step. If so, please
|
|
// check 'aarch64_valid_for_adrp_p'.
|
|
gold_assert(status == The_reloc_functions::STATUS_OKAY);
|
|
|
|
// 2nd reloc is ADD_ABS_LO12_NC
|
|
const AArch64_reloc_property* arp =
|
|
aarch64_reloc_property_table->get_reloc_property(
|
|
elfcpp::R_AARCH64_ADD_ABS_LO12_NC);
|
|
gold_assert(arp != NULL);
|
|
status = The_reloc_functions::template
|
|
rela_general<32>(view + 4, dest, 0, arp);
|
|
// An error should never arise, it is an "_NC" relocation.
|
|
gold_assert(status == The_reloc_functions::STATUS_OKAY);
|
|
}
|
|
break;
|
|
|
|
case ST_LONG_BRANCH_ABS:
|
|
// 1st reloc is R_AARCH64_PREL64, at offset 8
|
|
elfcpp::Swap<64,big_endian>::writeval(view + 8, dest);
|
|
break;
|
|
|
|
case ST_LONG_BRANCH_PCREL:
|
|
{
|
|
// "PC" calculation is the 2nd insn in the stub.
|
|
uint64_t offset = dest - (address + 4);
|
|
// Offset is placed at offset 4 and 5.
|
|
elfcpp::Swap<64,big_endian>::writeval(view + 16, offset);
|
|
}
|
|
break;
|
|
|
|
default:
|
|
gold_unreachable();
|
|
}
|
|
}
|
|
|
|
|
|
// A class to handle the PLT data.
|
|
// This is an abstract base class that handles most of the linker details
|
|
// but does not know the actual contents of PLT entries. The derived
|
|
// classes below fill in those details.
|
|
|
|
template<int size, bool big_endian>
|
|
class Output_data_plt_aarch64 : public Output_section_data
|
|
{
|
|
public:
|
|
typedef Output_data_reloc<elfcpp::SHT_RELA, true, size, big_endian>
|
|
Reloc_section;
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr Address;
|
|
|
|
Output_data_plt_aarch64(Layout* layout,
|
|
uint64_t addralign,
|
|
Output_data_got_aarch64<size, big_endian>* got,
|
|
Output_data_space* got_plt,
|
|
Output_data_space* got_irelative)
|
|
: Output_section_data(addralign), tlsdesc_rel_(NULL), irelative_rel_(NULL),
|
|
got_(got), got_plt_(got_plt), got_irelative_(got_irelative),
|
|
count_(0), irelative_count_(0), tlsdesc_got_offset_(-1U)
|
|
{ this->init(layout); }
|
|
|
|
// Initialize the PLT section.
|
|
void
|
|
init(Layout* layout);
|
|
|
|
// Add an entry to the PLT.
|
|
void
|
|
add_entry(Symbol_table*, Layout*, Symbol* gsym);
|
|
|
|
// Add an entry to the PLT for a local STT_GNU_IFUNC symbol.
|
|
unsigned int
|
|
add_local_ifunc_entry(Symbol_table* symtab, Layout*,
|
|
Sized_relobj_file<size, big_endian>* relobj,
|
|
unsigned int local_sym_index);
|
|
|
|
// Add the relocation for a PLT entry.
|
|
void
|
|
add_relocation(Symbol_table*, Layout*, Symbol* gsym,
|
|
unsigned int got_offset);
|
|
|
|
// Add the reserved TLSDESC_PLT entry to the PLT.
|
|
void
|
|
reserve_tlsdesc_entry(unsigned int got_offset)
|
|
{ this->tlsdesc_got_offset_ = got_offset; }
|
|
|
|
// Return true if a TLSDESC_PLT entry has been reserved.
|
|
bool
|
|
has_tlsdesc_entry() const
|
|
{ return this->tlsdesc_got_offset_ != -1U; }
|
|
|
|
// Return the GOT offset for the reserved TLSDESC_PLT entry.
|
|
unsigned int
|
|
get_tlsdesc_got_offset() const
|
|
{ return this->tlsdesc_got_offset_; }
|
|
|
|
// Return the PLT offset of the reserved TLSDESC_PLT entry.
|
|
unsigned int
|
|
get_tlsdesc_plt_offset() const
|
|
{
|
|
return (this->first_plt_entry_offset() +
|
|
(this->count_ + this->irelative_count_)
|
|
* this->get_plt_entry_size());
|
|
}
|
|
|
|
// Return the .rela.plt section data.
|
|
Reloc_section*
|
|
rela_plt()
|
|
{ return this->rel_; }
|
|
|
|
// Return where the TLSDESC relocations should go.
|
|
Reloc_section*
|
|
rela_tlsdesc(Layout*);
|
|
|
|
// Return where the IRELATIVE relocations should go in the PLT
|
|
// relocations.
|
|
Reloc_section*
|
|
rela_irelative(Symbol_table*, Layout*);
|
|
|
|
// Return whether we created a section for IRELATIVE relocations.
|
|
bool
|
|
has_irelative_section() const
|
|
{ return this->irelative_rel_ != NULL; }
|
|
|
|
// Return the number of PLT entries.
|
|
unsigned int
|
|
entry_count() const
|
|
{ return this->count_ + this->irelative_count_; }
|
|
|
|
// Return the offset of the first non-reserved PLT entry.
|
|
unsigned int
|
|
first_plt_entry_offset() const
|
|
{ return this->do_first_plt_entry_offset(); }
|
|
|
|
// Return the size of a PLT entry.
|
|
unsigned int
|
|
get_plt_entry_size() const
|
|
{ return this->do_get_plt_entry_size(); }
|
|
|
|
// Return the reserved tlsdesc entry size.
|
|
unsigned int
|
|
get_plt_tlsdesc_entry_size() const
|
|
{ return this->do_get_plt_tlsdesc_entry_size(); }
|
|
|
|
// Return the PLT address to use for a global symbol.
|
|
uint64_t
|
|
address_for_global(const Symbol*);
|
|
|
|
// Return the PLT address to use for a local symbol.
|
|
uint64_t
|
|
address_for_local(const Relobj*, unsigned int symndx);
|
|
|
|
protected:
|
|
// Fill in the first PLT entry.
|
|
void
|
|
fill_first_plt_entry(unsigned char* pov,
|
|
Address got_address,
|
|
Address plt_address)
|
|
{ this->do_fill_first_plt_entry(pov, got_address, plt_address); }
|
|
|
|
// Fill in a normal PLT entry.
|
|
void
|
|
fill_plt_entry(unsigned char* pov,
|
|
Address got_address,
|
|
Address plt_address,
|
|
unsigned int got_offset,
|
|
unsigned int plt_offset)
|
|
{
|
|
this->do_fill_plt_entry(pov, got_address, plt_address,
|
|
got_offset, plt_offset);
|
|
}
|
|
|
|
// Fill in the reserved TLSDESC PLT entry.
|
|
void
|
|
fill_tlsdesc_entry(unsigned char* pov,
|
|
Address gotplt_address,
|
|
Address plt_address,
|
|
Address got_base,
|
|
unsigned int tlsdesc_got_offset,
|
|
unsigned int plt_offset)
|
|
{
|
|
this->do_fill_tlsdesc_entry(pov, gotplt_address, plt_address, got_base,
|
|
tlsdesc_got_offset, plt_offset);
|
|
}
|
|
|
|
virtual unsigned int
|
|
do_first_plt_entry_offset() const = 0;
|
|
|
|
virtual unsigned int
|
|
do_get_plt_entry_size() const = 0;
|
|
|
|
virtual unsigned int
|
|
do_get_plt_tlsdesc_entry_size() const = 0;
|
|
|
|
virtual void
|
|
do_fill_first_plt_entry(unsigned char* pov,
|
|
Address got_addr,
|
|
Address plt_addr) = 0;
|
|
|
|
virtual void
|
|
do_fill_plt_entry(unsigned char* pov,
|
|
Address got_address,
|
|
Address plt_address,
|
|
unsigned int got_offset,
|
|
unsigned int plt_offset) = 0;
|
|
|
|
virtual void
|
|
do_fill_tlsdesc_entry(unsigned char* pov,
|
|
Address gotplt_address,
|
|
Address plt_address,
|
|
Address got_base,
|
|
unsigned int tlsdesc_got_offset,
|
|
unsigned int plt_offset) = 0;
|
|
|
|
void
|
|
do_adjust_output_section(Output_section* os);
|
|
|
|
// Write to a map file.
|
|
void
|
|
do_print_to_mapfile(Mapfile* mapfile) const
|
|
{ mapfile->print_output_data(this, _("** PLT")); }
|
|
|
|
private:
|
|
// Set the final size.
|
|
void
|
|
set_final_data_size();
|
|
|
|
// Write out the PLT data.
|
|
void
|
|
do_write(Output_file*);
|
|
|
|
// The reloc section.
|
|
Reloc_section* rel_;
|
|
|
|
// The TLSDESC relocs, if necessary. These must follow the regular
|
|
// PLT relocs.
|
|
Reloc_section* tlsdesc_rel_;
|
|
|
|
// The IRELATIVE relocs, if necessary. These must follow the
|
|
// regular PLT relocations.
|
|
Reloc_section* irelative_rel_;
|
|
|
|
// The .got section.
|
|
Output_data_got_aarch64<size, big_endian>* got_;
|
|
|
|
// The .got.plt section.
|
|
Output_data_space* got_plt_;
|
|
|
|
// The part of the .got.plt section used for IRELATIVE relocs.
|
|
Output_data_space* got_irelative_;
|
|
|
|
// The number of PLT entries.
|
|
unsigned int count_;
|
|
|
|
// Number of PLT entries with R_AARCH64_IRELATIVE relocs. These
|
|
// follow the regular PLT entries.
|
|
unsigned int irelative_count_;
|
|
|
|
// GOT offset of the reserved TLSDESC_GOT entry for the lazy trampoline.
|
|
// Communicated to the loader via DT_TLSDESC_GOT. The magic value -1
|
|
// indicates an offset is not allocated.
|
|
unsigned int tlsdesc_got_offset_;
|
|
};
|
|
|
|
// Initialize the PLT section.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Output_data_plt_aarch64<size, big_endian>::init(Layout* layout)
|
|
{
|
|
this->rel_ = new Reloc_section(false);
|
|
layout->add_output_section_data(".rela.plt", elfcpp::SHT_RELA,
|
|
elfcpp::SHF_ALLOC, this->rel_,
|
|
ORDER_DYNAMIC_PLT_RELOCS, false);
|
|
}
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Output_data_plt_aarch64<size, big_endian>::do_adjust_output_section(
|
|
Output_section* os)
|
|
{
|
|
os->set_entsize(this->get_plt_entry_size());
|
|
}
|
|
|
|
// Add an entry to the PLT.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Output_data_plt_aarch64<size, big_endian>::add_entry(Symbol_table* symtab,
|
|
Layout* layout, Symbol* gsym)
|
|
{
|
|
gold_assert(!gsym->has_plt_offset());
|
|
|
|
unsigned int* pcount;
|
|
unsigned int plt_reserved;
|
|
Output_section_data_build* got;
|
|
|
|
if (gsym->type() == elfcpp::STT_GNU_IFUNC
|
|
&& gsym->can_use_relative_reloc(false))
|
|
{
|
|
pcount = &this->irelative_count_;
|
|
plt_reserved = 0;
|
|
got = this->got_irelative_;
|
|
}
|
|
else
|
|
{
|
|
pcount = &this->count_;
|
|
plt_reserved = this->first_plt_entry_offset();
|
|
got = this->got_plt_;
|
|
}
|
|
|
|
gsym->set_plt_offset((*pcount) * this->get_plt_entry_size()
|
|
+ plt_reserved);
|
|
|
|
++*pcount;
|
|
|
|
section_offset_type got_offset = got->current_data_size();
|
|
|
|
// Every PLT entry needs a GOT entry which points back to the PLT
|
|
// entry (this will be changed by the dynamic linker, normally
|
|
// lazily when the function is called).
|
|
got->set_current_data_size(got_offset + size / 8);
|
|
|
|
// Every PLT entry needs a reloc.
|
|
this->add_relocation(symtab, layout, gsym, got_offset);
|
|
|
|
// Note that we don't need to save the symbol. The contents of the
|
|
// PLT are independent of which symbols are used. The symbols only
|
|
// appear in the relocations.
|
|
}
|
|
|
|
// Add an entry to the PLT for a local STT_GNU_IFUNC symbol. Return
|
|
// the PLT offset.
|
|
|
|
template<int size, bool big_endian>
|
|
unsigned int
|
|
Output_data_plt_aarch64<size, big_endian>::add_local_ifunc_entry(
|
|
Symbol_table* symtab,
|
|
Layout* layout,
|
|
Sized_relobj_file<size, big_endian>* relobj,
|
|
unsigned int local_sym_index)
|
|
{
|
|
unsigned int plt_offset = this->irelative_count_ * this->get_plt_entry_size();
|
|
++this->irelative_count_;
|
|
|
|
section_offset_type got_offset = this->got_irelative_->current_data_size();
|
|
|
|
// Every PLT entry needs a GOT entry which points back to the PLT
|
|
// entry.
|
|
this->got_irelative_->set_current_data_size(got_offset + size / 8);
|
|
|
|
// Every PLT entry needs a reloc.
|
|
Reloc_section* rela = this->rela_irelative(symtab, layout);
|
|
rela->add_symbolless_local_addend(relobj, local_sym_index,
|
|
elfcpp::R_AARCH64_IRELATIVE,
|
|
this->got_irelative_, got_offset, 0);
|
|
|
|
return plt_offset;
|
|
}
|
|
|
|
// Add the relocation for a PLT entry.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Output_data_plt_aarch64<size, big_endian>::add_relocation(
|
|
Symbol_table* symtab, Layout* layout, Symbol* gsym, unsigned int got_offset)
|
|
{
|
|
if (gsym->type() == elfcpp::STT_GNU_IFUNC
|
|
&& gsym->can_use_relative_reloc(false))
|
|
{
|
|
Reloc_section* rela = this->rela_irelative(symtab, layout);
|
|
rela->add_symbolless_global_addend(gsym, elfcpp::R_AARCH64_IRELATIVE,
|
|
this->got_irelative_, got_offset, 0);
|
|
}
|
|
else
|
|
{
|
|
gsym->set_needs_dynsym_entry();
|
|
this->rel_->add_global(gsym, elfcpp::R_AARCH64_JUMP_SLOT, this->got_plt_,
|
|
got_offset, 0);
|
|
}
|
|
}
|
|
|
|
// Return where the TLSDESC relocations should go, creating it if
|
|
// necessary. These follow the JUMP_SLOT relocations.
|
|
|
|
template<int size, bool big_endian>
|
|
typename Output_data_plt_aarch64<size, big_endian>::Reloc_section*
|
|
Output_data_plt_aarch64<size, big_endian>::rela_tlsdesc(Layout* layout)
|
|
{
|
|
if (this->tlsdesc_rel_ == NULL)
|
|
{
|
|
this->tlsdesc_rel_ = new Reloc_section(false);
|
|
layout->add_output_section_data(".rela.plt", elfcpp::SHT_RELA,
|
|
elfcpp::SHF_ALLOC, this->tlsdesc_rel_,
|
|
ORDER_DYNAMIC_PLT_RELOCS, false);
|
|
gold_assert(this->tlsdesc_rel_->output_section()
|
|
== this->rel_->output_section());
|
|
}
|
|
return this->tlsdesc_rel_;
|
|
}
|
|
|
|
// Return where the IRELATIVE relocations should go in the PLT. These
|
|
// follow the JUMP_SLOT and the TLSDESC relocations.
|
|
|
|
template<int size, bool big_endian>
|
|
typename Output_data_plt_aarch64<size, big_endian>::Reloc_section*
|
|
Output_data_plt_aarch64<size, big_endian>::rela_irelative(Symbol_table* symtab,
|
|
Layout* layout)
|
|
{
|
|
if (this->irelative_rel_ == NULL)
|
|
{
|
|
// Make sure we have a place for the TLSDESC relocations, in
|
|
// case we see any later on.
|
|
this->rela_tlsdesc(layout);
|
|
this->irelative_rel_ = new Reloc_section(false);
|
|
layout->add_output_section_data(".rela.plt", elfcpp::SHT_RELA,
|
|
elfcpp::SHF_ALLOC, this->irelative_rel_,
|
|
ORDER_DYNAMIC_PLT_RELOCS, false);
|
|
gold_assert(this->irelative_rel_->output_section()
|
|
== this->rel_->output_section());
|
|
|
|
if (parameters->doing_static_link())
|
|
{
|
|
// A statically linked executable will only have a .rela.plt
|
|
// section to hold R_AARCH64_IRELATIVE relocs for
|
|
// STT_GNU_IFUNC symbols. The library will use these
|
|
// symbols to locate the IRELATIVE relocs at program startup
|
|
// time.
|
|
symtab->define_in_output_data("__rela_iplt_start", NULL,
|
|
Symbol_table::PREDEFINED,
|
|
this->irelative_rel_, 0, 0,
|
|
elfcpp::STT_NOTYPE, elfcpp::STB_GLOBAL,
|
|
elfcpp::STV_HIDDEN, 0, false, true);
|
|
symtab->define_in_output_data("__rela_iplt_end", NULL,
|
|
Symbol_table::PREDEFINED,
|
|
this->irelative_rel_, 0, 0,
|
|
elfcpp::STT_NOTYPE, elfcpp::STB_GLOBAL,
|
|
elfcpp::STV_HIDDEN, 0, true, true);
|
|
}
|
|
}
|
|
return this->irelative_rel_;
|
|
}
|
|
|
|
// Return the PLT address to use for a global symbol.
|
|
|
|
template<int size, bool big_endian>
|
|
uint64_t
|
|
Output_data_plt_aarch64<size, big_endian>::address_for_global(
|
|
const Symbol* gsym)
|
|
{
|
|
uint64_t offset = 0;
|
|
if (gsym->type() == elfcpp::STT_GNU_IFUNC
|
|
&& gsym->can_use_relative_reloc(false))
|
|
offset = (this->first_plt_entry_offset() +
|
|
this->count_ * this->get_plt_entry_size());
|
|
return this->address() + offset + gsym->plt_offset();
|
|
}
|
|
|
|
// Return the PLT address to use for a local symbol. These are always
|
|
// IRELATIVE relocs.
|
|
|
|
template<int size, bool big_endian>
|
|
uint64_t
|
|
Output_data_plt_aarch64<size, big_endian>::address_for_local(
|
|
const Relobj* object,
|
|
unsigned int r_sym)
|
|
{
|
|
return (this->address()
|
|
+ this->first_plt_entry_offset()
|
|
+ this->count_ * this->get_plt_entry_size()
|
|
+ object->local_plt_offset(r_sym));
|
|
}
|
|
|
|
// Set the final size.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Output_data_plt_aarch64<size, big_endian>::set_final_data_size()
|
|
{
|
|
unsigned int count = this->count_ + this->irelative_count_;
|
|
unsigned int extra_size = 0;
|
|
if (this->has_tlsdesc_entry())
|
|
extra_size += this->get_plt_tlsdesc_entry_size();
|
|
this->set_data_size(this->first_plt_entry_offset()
|
|
+ count * this->get_plt_entry_size()
|
|
+ extra_size);
|
|
}
|
|
|
|
template<int size, bool big_endian>
|
|
class Output_data_plt_aarch64_standard :
|
|
public Output_data_plt_aarch64<size, big_endian>
|
|
{
|
|
public:
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr Address;
|
|
Output_data_plt_aarch64_standard(
|
|
Layout* layout,
|
|
Output_data_got_aarch64<size, big_endian>* got,
|
|
Output_data_space* got_plt,
|
|
Output_data_space* got_irelative)
|
|
: Output_data_plt_aarch64<size, big_endian>(layout,
|
|
size == 32 ? 4 : 8,
|
|
got, got_plt,
|
|
got_irelative)
|
|
{ }
|
|
|
|
protected:
|
|
// Return the offset of the first non-reserved PLT entry.
|
|
virtual unsigned int
|
|
do_first_plt_entry_offset() const
|
|
{ return this->first_plt_entry_size; }
|
|
|
|
// Return the size of a PLT entry
|
|
virtual unsigned int
|
|
do_get_plt_entry_size() const
|
|
{ return this->plt_entry_size; }
|
|
|
|
// Return the size of a tlsdesc entry
|
|
virtual unsigned int
|
|
do_get_plt_tlsdesc_entry_size() const
|
|
{ return this->plt_tlsdesc_entry_size; }
|
|
|
|
virtual void
|
|
do_fill_first_plt_entry(unsigned char* pov,
|
|
Address got_address,
|
|
Address plt_address);
|
|
|
|
virtual void
|
|
do_fill_plt_entry(unsigned char* pov,
|
|
Address got_address,
|
|
Address plt_address,
|
|
unsigned int got_offset,
|
|
unsigned int plt_offset);
|
|
|
|
virtual void
|
|
do_fill_tlsdesc_entry(unsigned char* pov,
|
|
Address gotplt_address,
|
|
Address plt_address,
|
|
Address got_base,
|
|
unsigned int tlsdesc_got_offset,
|
|
unsigned int plt_offset);
|
|
|
|
private:
|
|
// The size of the first plt entry size.
|
|
static const int first_plt_entry_size = 32;
|
|
// The size of the plt entry size.
|
|
static const int plt_entry_size = 16;
|
|
// The size of the plt tlsdesc entry size.
|
|
static const int plt_tlsdesc_entry_size = 32;
|
|
// Template for the first PLT entry.
|
|
static const uint32_t first_plt_entry[first_plt_entry_size / 4];
|
|
// Template for subsequent PLT entries.
|
|
static const uint32_t plt_entry[plt_entry_size / 4];
|
|
// The reserved TLSDESC entry in the PLT for an executable.
|
|
static const uint32_t tlsdesc_plt_entry[plt_tlsdesc_entry_size / 4];
|
|
};
|
|
|
|
// The first entry in the PLT for an executable.
|
|
|
|
template<>
|
|
const uint32_t
|
|
Output_data_plt_aarch64_standard<32, false>::
|
|
first_plt_entry[first_plt_entry_size / 4] =
|
|
{
|
|
0xa9bf7bf0, /* stp x16, x30, [sp, #-16]! */
|
|
0x90000010, /* adrp x16, PLT_GOT+0x8 */
|
|
0xb9400A11, /* ldr w17, [x16, #PLT_GOT+0x8] */
|
|
0x11002210, /* add w16, w16,#PLT_GOT+0x8 */
|
|
0xd61f0220, /* br x17 */
|
|
0xd503201f, /* nop */
|
|
0xd503201f, /* nop */
|
|
0xd503201f, /* nop */
|
|
};
|
|
|
|
|
|
template<>
|
|
const uint32_t
|
|
Output_data_plt_aarch64_standard<32, true>::
|
|
first_plt_entry[first_plt_entry_size / 4] =
|
|
{
|
|
0xa9bf7bf0, /* stp x16, x30, [sp, #-16]! */
|
|
0x90000010, /* adrp x16, PLT_GOT+0x8 */
|
|
0xb9400A11, /* ldr w17, [x16, #PLT_GOT+0x8] */
|
|
0x11002210, /* add w16, w16,#PLT_GOT+0x8 */
|
|
0xd61f0220, /* br x17 */
|
|
0xd503201f, /* nop */
|
|
0xd503201f, /* nop */
|
|
0xd503201f, /* nop */
|
|
};
|
|
|
|
|
|
template<>
|
|
const uint32_t
|
|
Output_data_plt_aarch64_standard<64, false>::
|
|
first_plt_entry[first_plt_entry_size / 4] =
|
|
{
|
|
0xa9bf7bf0, /* stp x16, x30, [sp, #-16]! */
|
|
0x90000010, /* adrp x16, PLT_GOT+16 */
|
|
0xf9400A11, /* ldr x17, [x16, #PLT_GOT+0x10] */
|
|
0x91004210, /* add x16, x16,#PLT_GOT+0x10 */
|
|
0xd61f0220, /* br x17 */
|
|
0xd503201f, /* nop */
|
|
0xd503201f, /* nop */
|
|
0xd503201f, /* nop */
|
|
};
|
|
|
|
|
|
template<>
|
|
const uint32_t
|
|
Output_data_plt_aarch64_standard<64, true>::
|
|
first_plt_entry[first_plt_entry_size / 4] =
|
|
{
|
|
0xa9bf7bf0, /* stp x16, x30, [sp, #-16]! */
|
|
0x90000010, /* adrp x16, PLT_GOT+16 */
|
|
0xf9400A11, /* ldr x17, [x16, #PLT_GOT+0x10] */
|
|
0x91004210, /* add x16, x16,#PLT_GOT+0x10 */
|
|
0xd61f0220, /* br x17 */
|
|
0xd503201f, /* nop */
|
|
0xd503201f, /* nop */
|
|
0xd503201f, /* nop */
|
|
};
|
|
|
|
|
|
template<>
|
|
const uint32_t
|
|
Output_data_plt_aarch64_standard<32, false>::
|
|
plt_entry[plt_entry_size / 4] =
|
|
{
|
|
0x90000010, /* adrp x16, PLTGOT + n * 4 */
|
|
0xb9400211, /* ldr w17, [w16, PLTGOT + n * 4] */
|
|
0x11000210, /* add w16, w16, :lo12:PLTGOT + n * 4 */
|
|
0xd61f0220, /* br x17. */
|
|
};
|
|
|
|
|
|
template<>
|
|
const uint32_t
|
|
Output_data_plt_aarch64_standard<32, true>::
|
|
plt_entry[plt_entry_size / 4] =
|
|
{
|
|
0x90000010, /* adrp x16, PLTGOT + n * 4 */
|
|
0xb9400211, /* ldr w17, [w16, PLTGOT + n * 4] */
|
|
0x11000210, /* add w16, w16, :lo12:PLTGOT + n * 4 */
|
|
0xd61f0220, /* br x17. */
|
|
};
|
|
|
|
|
|
template<>
|
|
const uint32_t
|
|
Output_data_plt_aarch64_standard<64, false>::
|
|
plt_entry[plt_entry_size / 4] =
|
|
{
|
|
0x90000010, /* adrp x16, PLTGOT + n * 8 */
|
|
0xf9400211, /* ldr x17, [x16, PLTGOT + n * 8] */
|
|
0x91000210, /* add x16, x16, :lo12:PLTGOT + n * 8 */
|
|
0xd61f0220, /* br x17. */
|
|
};
|
|
|
|
|
|
template<>
|
|
const uint32_t
|
|
Output_data_plt_aarch64_standard<64, true>::
|
|
plt_entry[plt_entry_size / 4] =
|
|
{
|
|
0x90000010, /* adrp x16, PLTGOT + n * 8 */
|
|
0xf9400211, /* ldr x17, [x16, PLTGOT + n * 8] */
|
|
0x91000210, /* add x16, x16, :lo12:PLTGOT + n * 8 */
|
|
0xd61f0220, /* br x17. */
|
|
};
|
|
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Output_data_plt_aarch64_standard<size, big_endian>::do_fill_first_plt_entry(
|
|
unsigned char* pov,
|
|
Address got_address,
|
|
Address plt_address)
|
|
{
|
|
// PLT0 of the small PLT looks like this in ELF64 -
|
|
// stp x16, x30, [sp, #-16]! Save the reloc and lr on stack.
|
|
// adrp x16, PLT_GOT + 16 Get the page base of the GOTPLT
|
|
// ldr x17, [x16, #:lo12:PLT_GOT+16] Load the address of the
|
|
// symbol resolver
|
|
// add x16, x16, #:lo12:PLT_GOT+16 Load the lo12 bits of the
|
|
// GOTPLT entry for this.
|
|
// br x17
|
|
// PLT0 will be slightly different in ELF32 due to different got entry
|
|
// size.
|
|
memcpy(pov, this->first_plt_entry, this->first_plt_entry_size);
|
|
Address gotplt_2nd_ent = got_address + (size / 8) * 2;
|
|
|
|
// Fill in the top 21 bits for this: ADRP x16, PLT_GOT + 8 * 2.
|
|
// ADRP: (PG(S+A)-PG(P)) >> 12) & 0x1fffff.
|
|
// FIXME: This only works for 64bit
|
|
AArch64_relocate_functions<size, big_endian>::adrp(pov + 4,
|
|
gotplt_2nd_ent, plt_address + 4);
|
|
|
|
// Fill in R_AARCH64_LDST8_LO12
|
|
elfcpp::Swap<32, big_endian>::writeval(
|
|
pov + 8,
|
|
((this->first_plt_entry[2] & 0xffc003ff)
|
|
| ((gotplt_2nd_ent & 0xff8) << 7)));
|
|
|
|
// Fill in R_AARCH64_ADD_ABS_LO12
|
|
elfcpp::Swap<32, big_endian>::writeval(
|
|
pov + 12,
|
|
((this->first_plt_entry[3] & 0xffc003ff)
|
|
| ((gotplt_2nd_ent & 0xfff) << 10)));
|
|
}
|
|
|
|
|
|
// Subsequent entries in the PLT for an executable.
|
|
// FIXME: This only works for 64bit
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Output_data_plt_aarch64_standard<size, big_endian>::do_fill_plt_entry(
|
|
unsigned char* pov,
|
|
Address got_address,
|
|
Address plt_address,
|
|
unsigned int got_offset,
|
|
unsigned int plt_offset)
|
|
{
|
|
memcpy(pov, this->plt_entry, this->plt_entry_size);
|
|
|
|
Address gotplt_entry_address = got_address + got_offset;
|
|
Address plt_entry_address = plt_address + plt_offset;
|
|
|
|
// Fill in R_AARCH64_PCREL_ADR_HI21
|
|
AArch64_relocate_functions<size, big_endian>::adrp(
|
|
pov,
|
|
gotplt_entry_address,
|
|
plt_entry_address);
|
|
|
|
// Fill in R_AARCH64_LDST64_ABS_LO12
|
|
elfcpp::Swap<32, big_endian>::writeval(
|
|
pov + 4,
|
|
((this->plt_entry[1] & 0xffc003ff)
|
|
| ((gotplt_entry_address & 0xff8) << 7)));
|
|
|
|
// Fill in R_AARCH64_ADD_ABS_LO12
|
|
elfcpp::Swap<32, big_endian>::writeval(
|
|
pov + 8,
|
|
((this->plt_entry[2] & 0xffc003ff)
|
|
| ((gotplt_entry_address & 0xfff) <<10)));
|
|
|
|
}
|
|
|
|
|
|
template<>
|
|
const uint32_t
|
|
Output_data_plt_aarch64_standard<32, false>::
|
|
tlsdesc_plt_entry[plt_tlsdesc_entry_size / 4] =
|
|
{
|
|
0xa9bf0fe2, /* stp x2, x3, [sp, #-16]! */
|
|
0x90000002, /* adrp x2, 0 */
|
|
0x90000003, /* adrp x3, 0 */
|
|
0xb9400042, /* ldr w2, [w2, #0] */
|
|
0x11000063, /* add w3, w3, 0 */
|
|
0xd61f0040, /* br x2 */
|
|
0xd503201f, /* nop */
|
|
0xd503201f, /* nop */
|
|
};
|
|
|
|
template<>
|
|
const uint32_t
|
|
Output_data_plt_aarch64_standard<32, true>::
|
|
tlsdesc_plt_entry[plt_tlsdesc_entry_size / 4] =
|
|
{
|
|
0xa9bf0fe2, /* stp x2, x3, [sp, #-16]! */
|
|
0x90000002, /* adrp x2, 0 */
|
|
0x90000003, /* adrp x3, 0 */
|
|
0xb9400042, /* ldr w2, [w2, #0] */
|
|
0x11000063, /* add w3, w3, 0 */
|
|
0xd61f0040, /* br x2 */
|
|
0xd503201f, /* nop */
|
|
0xd503201f, /* nop */
|
|
};
|
|
|
|
template<>
|
|
const uint32_t
|
|
Output_data_plt_aarch64_standard<64, false>::
|
|
tlsdesc_plt_entry[plt_tlsdesc_entry_size / 4] =
|
|
{
|
|
0xa9bf0fe2, /* stp x2, x3, [sp, #-16]! */
|
|
0x90000002, /* adrp x2, 0 */
|
|
0x90000003, /* adrp x3, 0 */
|
|
0xf9400042, /* ldr x2, [x2, #0] */
|
|
0x91000063, /* add x3, x3, 0 */
|
|
0xd61f0040, /* br x2 */
|
|
0xd503201f, /* nop */
|
|
0xd503201f, /* nop */
|
|
};
|
|
|
|
template<>
|
|
const uint32_t
|
|
Output_data_plt_aarch64_standard<64, true>::
|
|
tlsdesc_plt_entry[plt_tlsdesc_entry_size / 4] =
|
|
{
|
|
0xa9bf0fe2, /* stp x2, x3, [sp, #-16]! */
|
|
0x90000002, /* adrp x2, 0 */
|
|
0x90000003, /* adrp x3, 0 */
|
|
0xf9400042, /* ldr x2, [x2, #0] */
|
|
0x91000063, /* add x3, x3, 0 */
|
|
0xd61f0040, /* br x2 */
|
|
0xd503201f, /* nop */
|
|
0xd503201f, /* nop */
|
|
};
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Output_data_plt_aarch64_standard<size, big_endian>::do_fill_tlsdesc_entry(
|
|
unsigned char* pov,
|
|
Address gotplt_address,
|
|
Address plt_address,
|
|
Address got_base,
|
|
unsigned int tlsdesc_got_offset,
|
|
unsigned int plt_offset)
|
|
{
|
|
memcpy(pov, tlsdesc_plt_entry, plt_tlsdesc_entry_size);
|
|
|
|
// move DT_TLSDESC_GOT address into x2
|
|
// move .got.plt address into x3
|
|
Address tlsdesc_got_entry = got_base + tlsdesc_got_offset;
|
|
Address plt_entry_address = plt_address + plt_offset;
|
|
|
|
// R_AARCH64_ADR_PREL_PG_HI21
|
|
AArch64_relocate_functions<size, big_endian>::adrp(
|
|
pov + 4,
|
|
tlsdesc_got_entry,
|
|
plt_entry_address + 4);
|
|
|
|
// R_AARCH64_ADR_PREL_PG_HI21
|
|
AArch64_relocate_functions<size, big_endian>::adrp(
|
|
pov + 8,
|
|
gotplt_address,
|
|
plt_entry_address + 8);
|
|
|
|
// R_AARCH64_LDST64_ABS_LO12
|
|
elfcpp::Swap<32, big_endian>::writeval(
|
|
pov + 12,
|
|
((this->tlsdesc_plt_entry[3] & 0xffc003ff)
|
|
| ((tlsdesc_got_entry & 0xff8) << 7)));
|
|
|
|
// R_AARCH64_ADD_ABS_LO12
|
|
elfcpp::Swap<32, big_endian>::writeval(
|
|
pov + 16,
|
|
((this->tlsdesc_plt_entry[4] & 0xffc003ff)
|
|
| ((gotplt_address & 0xfff) << 10)));
|
|
}
|
|
|
|
// Write out the PLT. This uses the hand-coded instructions above,
|
|
// and adjusts them as needed. This is specified by the AMD64 ABI.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Output_data_plt_aarch64<size, big_endian>::do_write(Output_file* of)
|
|
{
|
|
const off_t offset = this->offset();
|
|
const section_size_type oview_size =
|
|
convert_to_section_size_type(this->data_size());
|
|
unsigned char* const oview = of->get_output_view(offset, oview_size);
|
|
|
|
const off_t got_file_offset = this->got_plt_->offset();
|
|
gold_assert(got_file_offset + this->got_plt_->data_size()
|
|
== this->got_irelative_->offset());
|
|
|
|
const section_size_type got_size =
|
|
convert_to_section_size_type(this->got_plt_->data_size()
|
|
+ this->got_irelative_->data_size());
|
|
unsigned char* const got_view = of->get_output_view(got_file_offset,
|
|
got_size);
|
|
|
|
unsigned char* pov = oview;
|
|
|
|
// The base address of the .plt section.
|
|
typename elfcpp::Elf_types<size>::Elf_Addr plt_address = this->address();
|
|
// The base address of the PLT portion of the .got section.
|
|
typename elfcpp::Elf_types<size>::Elf_Addr gotplt_address
|
|
= this->got_plt_->address();
|
|
|
|
this->fill_first_plt_entry(pov, gotplt_address, plt_address);
|
|
pov += this->first_plt_entry_offset();
|
|
|
|
// The first three entries in .got.plt are reserved.
|
|
unsigned char* got_pov = got_view;
|
|
memset(got_pov, 0, size / 8 * AARCH64_GOTPLT_RESERVE_COUNT);
|
|
got_pov += (size / 8) * AARCH64_GOTPLT_RESERVE_COUNT;
|
|
|
|
unsigned int plt_offset = this->first_plt_entry_offset();
|
|
unsigned int got_offset = (size / 8) * AARCH64_GOTPLT_RESERVE_COUNT;
|
|
const unsigned int count = this->count_ + this->irelative_count_;
|
|
for (unsigned int plt_index = 0;
|
|
plt_index < count;
|
|
++plt_index,
|
|
pov += this->get_plt_entry_size(),
|
|
got_pov += size / 8,
|
|
plt_offset += this->get_plt_entry_size(),
|
|
got_offset += size / 8)
|
|
{
|
|
// Set and adjust the PLT entry itself.
|
|
this->fill_plt_entry(pov, gotplt_address, plt_address,
|
|
got_offset, plt_offset);
|
|
|
|
// Set the entry in the GOT, which points to plt0.
|
|
elfcpp::Swap<size, big_endian>::writeval(got_pov, plt_address);
|
|
}
|
|
|
|
if (this->has_tlsdesc_entry())
|
|
{
|
|
// Set and adjust the reserved TLSDESC PLT entry.
|
|
unsigned int tlsdesc_got_offset = this->get_tlsdesc_got_offset();
|
|
// The base address of the .base section.
|
|
typename elfcpp::Elf_types<size>::Elf_Addr got_base =
|
|
this->got_->address();
|
|
this->fill_tlsdesc_entry(pov, gotplt_address, plt_address, got_base,
|
|
tlsdesc_got_offset, plt_offset);
|
|
pov += this->get_plt_tlsdesc_entry_size();
|
|
}
|
|
|
|
gold_assert(static_cast<section_size_type>(pov - oview) == oview_size);
|
|
gold_assert(static_cast<section_size_type>(got_pov - got_view) == got_size);
|
|
|
|
of->write_output_view(offset, oview_size, oview);
|
|
of->write_output_view(got_file_offset, got_size, got_view);
|
|
}
|
|
|
|
// Telling how to update the immediate field of an instruction.
|
|
struct AArch64_howto
|
|
{
|
|
// The immediate field mask.
|
|
elfcpp::Elf_Xword dst_mask;
|
|
|
|
// The offset to apply relocation immediate
|
|
int doffset;
|
|
|
|
// The second part offset, if the immediate field has two parts.
|
|
// -1 if the immediate field has only one part.
|
|
int doffset2;
|
|
};
|
|
|
|
static const AArch64_howto aarch64_howto[AArch64_reloc_property::INST_NUM] =
|
|
{
|
|
{0, -1, -1}, // DATA
|
|
{0x1fffe0, 5, -1}, // MOVW [20:5]-imm16
|
|
{0xffffe0, 5, -1}, // LD [23:5]-imm19
|
|
{0x60ffffe0, 29, 5}, // ADR [30:29]-immlo [23:5]-immhi
|
|
{0x60ffffe0, 29, 5}, // ADRP [30:29]-immlo [23:5]-immhi
|
|
{0x3ffc00, 10, -1}, // ADD [21:10]-imm12
|
|
{0x3ffc00, 10, -1}, // LDST [21:10]-imm12
|
|
{0x7ffe0, 5, -1}, // TBZNZ [18:5]-imm14
|
|
{0xffffe0, 5, -1}, // CONDB [23:5]-imm19
|
|
{0x3ffffff, 0, -1}, // B [25:0]-imm26
|
|
{0x3ffffff, 0, -1}, // CALL [25:0]-imm26
|
|
};
|
|
|
|
// AArch64 relocate function class
|
|
|
|
template<int size, bool big_endian>
|
|
class AArch64_relocate_functions
|
|
{
|
|
public:
|
|
typedef enum
|
|
{
|
|
STATUS_OKAY, // No error during relocation.
|
|
STATUS_OVERFLOW, // Relocation overflow.
|
|
STATUS_BAD_RELOC, // Relocation cannot be applied.
|
|
} Status;
|
|
|
|
typedef AArch64_relocate_functions<size, big_endian> This;
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr Address;
|
|
typedef Relocate_info<size, big_endian> The_relocate_info;
|
|
typedef AArch64_relobj<size, big_endian> The_aarch64_relobj;
|
|
typedef Reloc_stub<size, big_endian> The_reloc_stub;
|
|
typedef Stub_table<size, big_endian> The_stub_table;
|
|
typedef elfcpp::Rela<size, big_endian> The_rela;
|
|
typedef typename elfcpp::Swap<size, big_endian>::Valtype AArch64_valtype;
|
|
|
|
// Return the page address of the address.
|
|
// Page(address) = address & ~0xFFF
|
|
|
|
static inline AArch64_valtype
|
|
Page(Address address)
|
|
{
|
|
return (address & (~static_cast<Address>(0xFFF)));
|
|
}
|
|
|
|
private:
|
|
// Update instruction (pointed by view) with selected bits (immed).
|
|
// val = (val & ~dst_mask) | (immed << doffset)
|
|
|
|
template<int valsize>
|
|
static inline void
|
|
update_view(unsigned char* view,
|
|
AArch64_valtype immed,
|
|
elfcpp::Elf_Xword doffset,
|
|
elfcpp::Elf_Xword dst_mask)
|
|
{
|
|
typedef typename elfcpp::Swap<valsize, big_endian>::Valtype Valtype;
|
|
Valtype* wv = reinterpret_cast<Valtype*>(view);
|
|
Valtype val = elfcpp::Swap<valsize, big_endian>::readval(wv);
|
|
|
|
// Clear immediate fields.
|
|
val &= ~dst_mask;
|
|
elfcpp::Swap<valsize, big_endian>::writeval(wv,
|
|
static_cast<Valtype>(val | (immed << doffset)));
|
|
}
|
|
|
|
// Update two parts of an instruction (pointed by view) with selected
|
|
// bits (immed1 and immed2).
|
|
// val = (val & ~dst_mask) | (immed1 << doffset1) | (immed2 << doffset2)
|
|
|
|
template<int valsize>
|
|
static inline void
|
|
update_view_two_parts(
|
|
unsigned char* view,
|
|
AArch64_valtype immed1,
|
|
AArch64_valtype immed2,
|
|
elfcpp::Elf_Xword doffset1,
|
|
elfcpp::Elf_Xword doffset2,
|
|
elfcpp::Elf_Xword dst_mask)
|
|
{
|
|
typedef typename elfcpp::Swap<valsize, big_endian>::Valtype Valtype;
|
|
Valtype* wv = reinterpret_cast<Valtype*>(view);
|
|
Valtype val = elfcpp::Swap<valsize, big_endian>::readval(wv);
|
|
val &= ~dst_mask;
|
|
elfcpp::Swap<valsize, big_endian>::writeval(wv,
|
|
static_cast<Valtype>(val | (immed1 << doffset1) |
|
|
(immed2 << doffset2)));
|
|
}
|
|
|
|
// Update adr or adrp instruction with immed.
|
|
// In adr and adrp: [30:29] immlo [23:5] immhi
|
|
|
|
static inline void
|
|
update_adr(unsigned char* view, AArch64_valtype immed)
|
|
{
|
|
elfcpp::Elf_Xword dst_mask = (0x3 << 29) | (0x7ffff << 5);
|
|
This::template update_view_two_parts<32>(
|
|
view,
|
|
immed & 0x3,
|
|
(immed & 0x1ffffc) >> 2,
|
|
29,
|
|
5,
|
|
dst_mask);
|
|
}
|
|
|
|
// Update movz/movn instruction with bits immed.
|
|
// Set instruction to movz if is_movz is true, otherwise set instruction
|
|
// to movn.
|
|
|
|
static inline void
|
|
update_movnz(unsigned char* view,
|
|
AArch64_valtype immed,
|
|
bool is_movz)
|
|
{
|
|
typedef typename elfcpp::Swap<32, big_endian>::Valtype Valtype;
|
|
Valtype* wv = reinterpret_cast<Valtype*>(view);
|
|
Valtype val = elfcpp::Swap<32, big_endian>::readval(wv);
|
|
|
|
const elfcpp::Elf_Xword doffset =
|
|
aarch64_howto[AArch64_reloc_property::INST_MOVW].doffset;
|
|
const elfcpp::Elf_Xword dst_mask =
|
|
aarch64_howto[AArch64_reloc_property::INST_MOVW].dst_mask;
|
|
|
|
// Clear immediate fields and opc code.
|
|
val &= ~(dst_mask | (0x3 << 29));
|
|
|
|
// Set instruction to movz or movn.
|
|
// movz: [30:29] is 10 movn: [30:29] is 00
|
|
if (is_movz)
|
|
val |= (0x2 << 29);
|
|
|
|
elfcpp::Swap<32, big_endian>::writeval(wv,
|
|
static_cast<Valtype>(val | (immed << doffset)));
|
|
}
|
|
|
|
public:
|
|
|
|
// Update selected bits in text.
|
|
|
|
template<int valsize>
|
|
static inline typename This::Status
|
|
reloc_common(unsigned char* view, Address x,
|
|
const AArch64_reloc_property* reloc_property)
|
|
{
|
|
// Select bits from X.
|
|
Address immed = reloc_property->select_x_value(x);
|
|
|
|
// Update view.
|
|
const AArch64_reloc_property::Reloc_inst inst =
|
|
reloc_property->reloc_inst();
|
|
// If it is a data relocation or instruction has 2 parts of immediate
|
|
// fields, you should not call pcrela_general.
|
|
gold_assert(aarch64_howto[inst].doffset2 == -1 &&
|
|
aarch64_howto[inst].doffset != -1);
|
|
This::template update_view<valsize>(view, immed,
|
|
aarch64_howto[inst].doffset,
|
|
aarch64_howto[inst].dst_mask);
|
|
|
|
// Do check overflow or alignment if needed.
|
|
return (reloc_property->checkup_x_value(x)
|
|
? This::STATUS_OKAY
|
|
: This::STATUS_OVERFLOW);
|
|
}
|
|
|
|
// Construct a B insn. Note, although we group it here with other relocation
|
|
// operation, there is actually no 'relocation' involved here.
|
|
static inline void
|
|
construct_b(unsigned char* view, unsigned int branch_offset)
|
|
{
|
|
update_view_two_parts<32>(view, 0x05, (branch_offset >> 2),
|
|
26, 0, 0xffffffff);
|
|
}
|
|
|
|
// Do a simple rela relocation at unaligned addresses.
|
|
|
|
template<int valsize>
|
|
static inline typename This::Status
|
|
rela_ua(unsigned char* view,
|
|
const Sized_relobj_file<size, big_endian>* object,
|
|
const Symbol_value<size>* psymval,
|
|
AArch64_valtype addend,
|
|
const AArch64_reloc_property* reloc_property)
|
|
{
|
|
typedef typename elfcpp::Swap_unaligned<valsize, big_endian>::Valtype
|
|
Valtype;
|
|
typename elfcpp::Elf_types<size>::Elf_Addr x =
|
|
psymval->value(object, addend);
|
|
elfcpp::Swap_unaligned<valsize, big_endian>::writeval(view,
|
|
static_cast<Valtype>(x));
|
|
return (reloc_property->checkup_x_value(x)
|
|
? This::STATUS_OKAY
|
|
: This::STATUS_OVERFLOW);
|
|
}
|
|
|
|
// Do a simple pc-relative relocation at unaligned addresses.
|
|
|
|
template<int valsize>
|
|
static inline typename This::Status
|
|
pcrela_ua(unsigned char* view,
|
|
const Sized_relobj_file<size, big_endian>* object,
|
|
const Symbol_value<size>* psymval,
|
|
AArch64_valtype addend,
|
|
Address address,
|
|
const AArch64_reloc_property* reloc_property)
|
|
{
|
|
typedef typename elfcpp::Swap_unaligned<valsize, big_endian>::Valtype
|
|
Valtype;
|
|
Address x = psymval->value(object, addend) - address;
|
|
elfcpp::Swap_unaligned<valsize, big_endian>::writeval(view,
|
|
static_cast<Valtype>(x));
|
|
return (reloc_property->checkup_x_value(x)
|
|
? This::STATUS_OKAY
|
|
: This::STATUS_OVERFLOW);
|
|
}
|
|
|
|
// Do a simple rela relocation at aligned addresses.
|
|
|
|
template<int valsize>
|
|
static inline typename This::Status
|
|
rela(
|
|
unsigned char* view,
|
|
const Sized_relobj_file<size, big_endian>* object,
|
|
const Symbol_value<size>* psymval,
|
|
AArch64_valtype addend,
|
|
const AArch64_reloc_property* reloc_property)
|
|
{
|
|
typedef typename elfcpp::Swap<valsize, big_endian>::Valtype Valtype;
|
|
Valtype* wv = reinterpret_cast<Valtype*>(view);
|
|
Address x = psymval->value(object, addend);
|
|
elfcpp::Swap<valsize, big_endian>::writeval(wv,static_cast<Valtype>(x));
|
|
return (reloc_property->checkup_x_value(x)
|
|
? This::STATUS_OKAY
|
|
: This::STATUS_OVERFLOW);
|
|
}
|
|
|
|
// Do relocate. Update selected bits in text.
|
|
// new_val = (val & ~dst_mask) | (immed << doffset)
|
|
|
|
template<int valsize>
|
|
static inline typename This::Status
|
|
rela_general(unsigned char* view,
|
|
const Sized_relobj_file<size, big_endian>* object,
|
|
const Symbol_value<size>* psymval,
|
|
AArch64_valtype addend,
|
|
const AArch64_reloc_property* reloc_property)
|
|
{
|
|
// Calculate relocation.
|
|
Address x = psymval->value(object, addend);
|
|
return This::template reloc_common<valsize>(view, x, reloc_property);
|
|
}
|
|
|
|
// Do relocate. Update selected bits in text.
|
|
// new val = (val & ~dst_mask) | (immed << doffset)
|
|
|
|
template<int valsize>
|
|
static inline typename This::Status
|
|
rela_general(
|
|
unsigned char* view,
|
|
AArch64_valtype s,
|
|
AArch64_valtype addend,
|
|
const AArch64_reloc_property* reloc_property)
|
|
{
|
|
// Calculate relocation.
|
|
Address x = s + addend;
|
|
return This::template reloc_common<valsize>(view, x, reloc_property);
|
|
}
|
|
|
|
// Do address relative relocate. Update selected bits in text.
|
|
// new val = (val & ~dst_mask) | (immed << doffset)
|
|
|
|
template<int valsize>
|
|
static inline typename This::Status
|
|
pcrela_general(
|
|
unsigned char* view,
|
|
const Sized_relobj_file<size, big_endian>* object,
|
|
const Symbol_value<size>* psymval,
|
|
AArch64_valtype addend,
|
|
Address address,
|
|
const AArch64_reloc_property* reloc_property)
|
|
{
|
|
// Calculate relocation.
|
|
Address x = psymval->value(object, addend) - address;
|
|
return This::template reloc_common<valsize>(view, x, reloc_property);
|
|
}
|
|
|
|
|
|
// Calculate (S + A) - address, update adr instruction.
|
|
|
|
static inline typename This::Status
|
|
adr(unsigned char* view,
|
|
const Sized_relobj_file<size, big_endian>* object,
|
|
const Symbol_value<size>* psymval,
|
|
Address addend,
|
|
Address address,
|
|
const AArch64_reloc_property* /* reloc_property */)
|
|
{
|
|
AArch64_valtype x = psymval->value(object, addend) - address;
|
|
// Pick bits [20:0] of X.
|
|
AArch64_valtype immed = x & 0x1fffff;
|
|
update_adr(view, immed);
|
|
// Check -2^20 <= X < 2^20
|
|
return (size == 64 && Bits<21>::has_overflow((x))
|
|
? This::STATUS_OVERFLOW
|
|
: This::STATUS_OKAY);
|
|
}
|
|
|
|
// Calculate PG(S+A) - PG(address), update adrp instruction.
|
|
// R_AARCH64_ADR_PREL_PG_HI21
|
|
|
|
static inline typename This::Status
|
|
adrp(
|
|
unsigned char* view,
|
|
Address sa,
|
|
Address address)
|
|
{
|
|
AArch64_valtype x = This::Page(sa) - This::Page(address);
|
|
// Pick [32:12] of X.
|
|
AArch64_valtype immed = (x >> 12) & 0x1fffff;
|
|
update_adr(view, immed);
|
|
// Check -2^32 <= X < 2^32
|
|
return (size == 64 && Bits<33>::has_overflow((x))
|
|
? This::STATUS_OVERFLOW
|
|
: This::STATUS_OKAY);
|
|
}
|
|
|
|
// Calculate PG(S+A) - PG(address), update adrp instruction.
|
|
// R_AARCH64_ADR_PREL_PG_HI21
|
|
|
|
static inline typename This::Status
|
|
adrp(unsigned char* view,
|
|
const Sized_relobj_file<size, big_endian>* object,
|
|
const Symbol_value<size>* psymval,
|
|
Address addend,
|
|
Address address,
|
|
const AArch64_reloc_property* reloc_property)
|
|
{
|
|
Address sa = psymval->value(object, addend);
|
|
AArch64_valtype x = This::Page(sa) - This::Page(address);
|
|
// Pick [32:12] of X.
|
|
AArch64_valtype immed = (x >> 12) & 0x1fffff;
|
|
update_adr(view, immed);
|
|
return (reloc_property->checkup_x_value(x)
|
|
? This::STATUS_OKAY
|
|
: This::STATUS_OVERFLOW);
|
|
}
|
|
|
|
// Update mov[n/z] instruction. Check overflow if needed.
|
|
// If X >=0, set the instruction to movz and its immediate value to the
|
|
// selected bits S.
|
|
// If X < 0, set the instruction to movn and its immediate value to
|
|
// NOT (selected bits of).
|
|
|
|
static inline typename This::Status
|
|
movnz(unsigned char* view,
|
|
AArch64_valtype x,
|
|
const AArch64_reloc_property* reloc_property)
|
|
{
|
|
// Select bits from X.
|
|
Address immed;
|
|
bool is_movz;
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Swxword SignedW;
|
|
if (static_cast<SignedW>(x) >= 0)
|
|
{
|
|
immed = reloc_property->select_x_value(x);
|
|
is_movz = true;
|
|
}
|
|
else
|
|
{
|
|
immed = reloc_property->select_x_value(~x);;
|
|
is_movz = false;
|
|
}
|
|
|
|
// Update movnz instruction.
|
|
update_movnz(view, immed, is_movz);
|
|
|
|
// Do check overflow or alignment if needed.
|
|
return (reloc_property->checkup_x_value(x)
|
|
? This::STATUS_OKAY
|
|
: This::STATUS_OVERFLOW);
|
|
}
|
|
|
|
static inline bool
|
|
maybe_apply_stub(unsigned int,
|
|
const The_relocate_info*,
|
|
const The_rela&,
|
|
unsigned char*,
|
|
Address,
|
|
const Sized_symbol<size>*,
|
|
const Symbol_value<size>*,
|
|
const Sized_relobj_file<size, big_endian>*,
|
|
section_size_type);
|
|
|
|
}; // End of AArch64_relocate_functions
|
|
|
|
|
|
// For a certain relocation type (usually jump/branch), test to see if the
|
|
// destination needs a stub to fulfil. If so, re-route the destination of the
|
|
// original instruction to the stub, note, at this time, the stub has already
|
|
// been generated.
|
|
|
|
template<int size, bool big_endian>
|
|
bool
|
|
AArch64_relocate_functions<size, big_endian>::
|
|
maybe_apply_stub(unsigned int r_type,
|
|
const The_relocate_info* relinfo,
|
|
const The_rela& rela,
|
|
unsigned char* view,
|
|
Address address,
|
|
const Sized_symbol<size>* gsym,
|
|
const Symbol_value<size>* psymval,
|
|
const Sized_relobj_file<size, big_endian>* object,
|
|
section_size_type current_group_size)
|
|
{
|
|
if (parameters->options().relocatable())
|
|
return false;
|
|
|
|
typename elfcpp::Elf_types<size>::Elf_Swxword addend = rela.get_r_addend();
|
|
Address branch_target = psymval->value(object, 0) + addend;
|
|
int stub_type =
|
|
The_reloc_stub::stub_type_for_reloc(r_type, address, branch_target);
|
|
if (stub_type == ST_NONE)
|
|
return false;
|
|
|
|
const The_aarch64_relobj* aarch64_relobj =
|
|
static_cast<const The_aarch64_relobj*>(object);
|
|
const AArch64_reloc_property* arp =
|
|
aarch64_reloc_property_table->get_reloc_property(r_type);
|
|
gold_assert(arp != NULL);
|
|
|
|
// We don't create stubs for undefined symbols, but do for weak.
|
|
if (gsym
|
|
&& !gsym->use_plt_offset(arp->reference_flags())
|
|
&& gsym->is_undefined())
|
|
{
|
|
gold_debug(DEBUG_TARGET,
|
|
"stub: looking for a stub for undefined symbol %s in file %s",
|
|
gsym->name(), aarch64_relobj->name().c_str());
|
|
return false;
|
|
}
|
|
|
|
The_stub_table* stub_table = aarch64_relobj->stub_table(relinfo->data_shndx);
|
|
gold_assert(stub_table != NULL);
|
|
|
|
unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info());
|
|
typename The_reloc_stub::Key stub_key(stub_type, gsym, object, r_sym, addend);
|
|
The_reloc_stub* stub = stub_table->find_reloc_stub(stub_key);
|
|
gold_assert(stub != NULL);
|
|
|
|
Address new_branch_target = stub_table->address() + stub->offset();
|
|
typename elfcpp::Swap<size, big_endian>::Valtype branch_offset =
|
|
new_branch_target - address;
|
|
typename This::Status status = This::template
|
|
rela_general<32>(view, branch_offset, 0, arp);
|
|
if (status != This::STATUS_OKAY)
|
|
gold_error(_("Stub is too far away, try a smaller value "
|
|
"for '--stub-group-size'. The current value is 0x%lx."),
|
|
static_cast<unsigned long>(current_group_size));
|
|
return true;
|
|
}
|
|
|
|
|
|
// Group input sections for stub generation.
|
|
//
|
|
// We group input sections in an output section so that the total size,
|
|
// including any padding space due to alignment is smaller than GROUP_SIZE
|
|
// unless the only input section in group is bigger than GROUP_SIZE already.
|
|
// Then an ARM stub table is created to follow the last input section
|
|
// in group. For each group an ARM stub table is created an is placed
|
|
// after the last group. If STUB_ALWAYS_AFTER_BRANCH is false, we further
|
|
// extend the group after the stub table.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::group_sections(
|
|
Layout* layout,
|
|
section_size_type group_size,
|
|
bool stubs_always_after_branch,
|
|
const Task* task)
|
|
{
|
|
// Group input sections and insert stub table
|
|
Layout::Section_list section_list;
|
|
layout->get_executable_sections(§ion_list);
|
|
for (Layout::Section_list::const_iterator p = section_list.begin();
|
|
p != section_list.end();
|
|
++p)
|
|
{
|
|
AArch64_output_section<size, big_endian>* output_section =
|
|
static_cast<AArch64_output_section<size, big_endian>*>(*p);
|
|
output_section->group_sections(group_size, stubs_always_after_branch,
|
|
this, task);
|
|
}
|
|
}
|
|
|
|
|
|
// Find the AArch64_input_section object corresponding to the SHNDX-th input
|
|
// section of RELOBJ.
|
|
|
|
template<int size, bool big_endian>
|
|
AArch64_input_section<size, big_endian>*
|
|
Target_aarch64<size, big_endian>::find_aarch64_input_section(
|
|
Relobj* relobj, unsigned int shndx) const
|
|
{
|
|
Section_id sid(relobj, shndx);
|
|
typename AArch64_input_section_map::const_iterator p =
|
|
this->aarch64_input_section_map_.find(sid);
|
|
return (p != this->aarch64_input_section_map_.end()) ? p->second : NULL;
|
|
}
|
|
|
|
|
|
// Make a new AArch64_input_section object.
|
|
|
|
template<int size, bool big_endian>
|
|
AArch64_input_section<size, big_endian>*
|
|
Target_aarch64<size, big_endian>::new_aarch64_input_section(
|
|
Relobj* relobj, unsigned int shndx)
|
|
{
|
|
Section_id sid(relobj, shndx);
|
|
|
|
AArch64_input_section<size, big_endian>* input_section =
|
|
new AArch64_input_section<size, big_endian>(relobj, shndx);
|
|
input_section->init();
|
|
|
|
// Register new AArch64_input_section in map for look-up.
|
|
std::pair<typename AArch64_input_section_map::iterator,bool> ins =
|
|
this->aarch64_input_section_map_.insert(
|
|
std::make_pair(sid, input_section));
|
|
|
|
// Make sure that it we have not created another AArch64_input_section
|
|
// for this input section already.
|
|
gold_assert(ins.second);
|
|
|
|
return input_section;
|
|
}
|
|
|
|
|
|
// Relaxation hook. This is where we do stub generation.
|
|
|
|
template<int size, bool big_endian>
|
|
bool
|
|
Target_aarch64<size, big_endian>::do_relax(
|
|
int pass,
|
|
const Input_objects* input_objects,
|
|
Symbol_table* symtab,
|
|
Layout* layout ,
|
|
const Task* task)
|
|
{
|
|
gold_assert(!parameters->options().relocatable());
|
|
if (pass == 1)
|
|
{
|
|
// We don't handle negative stub_group_size right now.
|
|
this->stub_group_size_ = abs(parameters->options().stub_group_size());
|
|
if (this->stub_group_size_ == 1)
|
|
{
|
|
// Leave room for 4096 4-byte stub entries. If we exceed that, then we
|
|
// will fail to link. The user will have to relink with an explicit
|
|
// group size option.
|
|
this->stub_group_size_ = The_reloc_stub::MAX_BRANCH_OFFSET -
|
|
4096 * 4;
|
|
}
|
|
group_sections(layout, this->stub_group_size_, true, task);
|
|
}
|
|
else
|
|
{
|
|
// If this is not the first pass, addresses and file offsets have
|
|
// been reset at this point, set them here.
|
|
for (Stub_table_iterator sp = this->stub_tables_.begin();
|
|
sp != this->stub_tables_.end(); ++sp)
|
|
{
|
|
The_stub_table* stt = *sp;
|
|
The_aarch64_input_section* owner = stt->owner();
|
|
off_t off = align_address(owner->original_size(),
|
|
stt->addralign());
|
|
stt->set_address_and_file_offset(owner->address() + off,
|
|
owner->offset() + off);
|
|
}
|
|
}
|
|
|
|
// Scan relocs for relocation stubs
|
|
for (Input_objects::Relobj_iterator op = input_objects->relobj_begin();
|
|
op != input_objects->relobj_end();
|
|
++op)
|
|
{
|
|
The_aarch64_relobj* aarch64_relobj =
|
|
static_cast<The_aarch64_relobj*>(*op);
|
|
// Lock the object so we can read from it. This is only called
|
|
// single-threaded from Layout::finalize, so it is OK to lock.
|
|
Task_lock_obj<Object> tl(task, aarch64_relobj);
|
|
aarch64_relobj->scan_sections_for_stubs(this, symtab, layout);
|
|
}
|
|
|
|
bool any_stub_table_changed = false;
|
|
for (Stub_table_iterator siter = this->stub_tables_.begin();
|
|
siter != this->stub_tables_.end() && !any_stub_table_changed; ++siter)
|
|
{
|
|
The_stub_table* stub_table = *siter;
|
|
if (stub_table->update_data_size_changed_p())
|
|
{
|
|
The_aarch64_input_section* owner = stub_table->owner();
|
|
uint64_t address = owner->address();
|
|
off_t offset = owner->offset();
|
|
owner->reset_address_and_file_offset();
|
|
owner->set_address_and_file_offset(address, offset);
|
|
|
|
any_stub_table_changed = true;
|
|
}
|
|
}
|
|
|
|
// Do not continue relaxation.
|
|
bool continue_relaxation = any_stub_table_changed;
|
|
if (!continue_relaxation)
|
|
for (Stub_table_iterator sp = this->stub_tables_.begin();
|
|
(sp != this->stub_tables_.end());
|
|
++sp)
|
|
(*sp)->finalize_stubs();
|
|
|
|
return continue_relaxation;
|
|
}
|
|
|
|
|
|
// Make a new Stub_table.
|
|
|
|
template<int size, bool big_endian>
|
|
Stub_table<size, big_endian>*
|
|
Target_aarch64<size, big_endian>::new_stub_table(
|
|
AArch64_input_section<size, big_endian>* owner)
|
|
{
|
|
Stub_table<size, big_endian>* stub_table =
|
|
new Stub_table<size, big_endian>(owner);
|
|
stub_table->set_address(align_address(
|
|
owner->address() + owner->data_size(), 8));
|
|
stub_table->set_file_offset(owner->offset() + owner->data_size());
|
|
stub_table->finalize_data_size();
|
|
|
|
this->stub_tables_.push_back(stub_table);
|
|
|
|
return stub_table;
|
|
}
|
|
|
|
|
|
template<int size, bool big_endian>
|
|
uint64_t
|
|
Target_aarch64<size, big_endian>::do_reloc_addend(
|
|
void* arg, unsigned int r_type, uint64_t) const
|
|
{
|
|
gold_assert(r_type == elfcpp::R_AARCH64_TLSDESC);
|
|
uintptr_t intarg = reinterpret_cast<uintptr_t>(arg);
|
|
gold_assert(intarg < this->tlsdesc_reloc_info_.size());
|
|
const Tlsdesc_info& ti(this->tlsdesc_reloc_info_[intarg]);
|
|
const Symbol_value<size>* psymval = ti.object->local_symbol(ti.r_sym);
|
|
gold_assert(psymval->is_tls_symbol());
|
|
// The value of a TLS symbol is the offset in the TLS segment.
|
|
return psymval->value(ti.object, 0);
|
|
}
|
|
|
|
// Return the number of entries in the PLT.
|
|
|
|
template<int size, bool big_endian>
|
|
unsigned int
|
|
Target_aarch64<size, big_endian>::plt_entry_count() const
|
|
{
|
|
if (this->plt_ == NULL)
|
|
return 0;
|
|
return this->plt_->entry_count();
|
|
}
|
|
|
|
// Return the offset of the first non-reserved PLT entry.
|
|
|
|
template<int size, bool big_endian>
|
|
unsigned int
|
|
Target_aarch64<size, big_endian>::first_plt_entry_offset() const
|
|
{
|
|
return this->plt_->first_plt_entry_offset();
|
|
}
|
|
|
|
// Return the size of each PLT entry.
|
|
|
|
template<int size, bool big_endian>
|
|
unsigned int
|
|
Target_aarch64<size, big_endian>::plt_entry_size() const
|
|
{
|
|
return this->plt_->get_plt_entry_size();
|
|
}
|
|
|
|
// Define the _TLS_MODULE_BASE_ symbol in the TLS segment.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::define_tls_base_symbol(
|
|
Symbol_table* symtab, Layout* layout)
|
|
{
|
|
if (this->tls_base_symbol_defined_)
|
|
return;
|
|
|
|
Output_segment* tls_segment = layout->tls_segment();
|
|
if (tls_segment != NULL)
|
|
{
|
|
// _TLS_MODULE_BASE_ always points to the beginning of tls segment.
|
|
symtab->define_in_output_segment("_TLS_MODULE_BASE_", NULL,
|
|
Symbol_table::PREDEFINED,
|
|
tls_segment, 0, 0,
|
|
elfcpp::STT_TLS,
|
|
elfcpp::STB_LOCAL,
|
|
elfcpp::STV_HIDDEN, 0,
|
|
Symbol::SEGMENT_START,
|
|
true);
|
|
}
|
|
this->tls_base_symbol_defined_ = true;
|
|
}
|
|
|
|
// Create the reserved PLT and GOT entries for the TLS descriptor resolver.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::reserve_tlsdesc_entries(
|
|
Symbol_table* symtab, Layout* layout)
|
|
{
|
|
if (this->plt_ == NULL)
|
|
this->make_plt_section(symtab, layout);
|
|
|
|
if (!this->plt_->has_tlsdesc_entry())
|
|
{
|
|
// Allocate the TLSDESC_GOT entry.
|
|
Output_data_got_aarch64<size, big_endian>* got =
|
|
this->got_section(symtab, layout);
|
|
unsigned int got_offset = got->add_constant(0);
|
|
|
|
// Allocate the TLSDESC_PLT entry.
|
|
this->plt_->reserve_tlsdesc_entry(got_offset);
|
|
}
|
|
}
|
|
|
|
// Create a GOT entry for the TLS module index.
|
|
|
|
template<int size, bool big_endian>
|
|
unsigned int
|
|
Target_aarch64<size, big_endian>::got_mod_index_entry(
|
|
Symbol_table* symtab, Layout* layout,
|
|
Sized_relobj_file<size, big_endian>* object)
|
|
{
|
|
if (this->got_mod_index_offset_ == -1U)
|
|
{
|
|
gold_assert(symtab != NULL && layout != NULL && object != NULL);
|
|
Reloc_section* rela_dyn = this->rela_dyn_section(layout);
|
|
Output_data_got_aarch64<size, big_endian>* got =
|
|
this->got_section(symtab, layout);
|
|
unsigned int got_offset = got->add_constant(0);
|
|
rela_dyn->add_local(object, 0, elfcpp::R_AARCH64_TLS_DTPMOD64, got,
|
|
got_offset, 0);
|
|
got->add_constant(0);
|
|
this->got_mod_index_offset_ = got_offset;
|
|
}
|
|
return this->got_mod_index_offset_;
|
|
}
|
|
|
|
// Optimize the TLS relocation type based on what we know about the
|
|
// symbol. IS_FINAL is true if the final address of this symbol is
|
|
// known at link time.
|
|
|
|
template<int size, bool big_endian>
|
|
tls::Tls_optimization
|
|
Target_aarch64<size, big_endian>::optimize_tls_reloc(bool is_final,
|
|
int r_type)
|
|
{
|
|
// If we are generating a shared library, then we can't do anything
|
|
// in the linker
|
|
if (parameters->options().shared())
|
|
return tls::TLSOPT_NONE;
|
|
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_TLSGD_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSDESC_LD_PREL19:
|
|
case elfcpp::R_AARCH64_TLSDESC_ADR_PREL21:
|
|
case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSDESC_LD64_LO12:
|
|
case elfcpp::R_AARCH64_TLSDESC_ADD_LO12:
|
|
case elfcpp::R_AARCH64_TLSDESC_OFF_G1:
|
|
case elfcpp::R_AARCH64_TLSDESC_OFF_G0_NC:
|
|
case elfcpp::R_AARCH64_TLSDESC_LDR:
|
|
case elfcpp::R_AARCH64_TLSDESC_ADD:
|
|
case elfcpp::R_AARCH64_TLSDESC_CALL:
|
|
// These are General-Dynamic which permits fully general TLS
|
|
// access. Since we know that we are generating an executable,
|
|
// we can convert this to Initial-Exec. If we also know that
|
|
// this is a local symbol, we can further switch to Local-Exec.
|
|
if (is_final)
|
|
return tls::TLSOPT_TO_LE;
|
|
return tls::TLSOPT_TO_IE;
|
|
|
|
case elfcpp::R_AARCH64_TLSLD_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G1:
|
|
case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G0_NC:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_HI12:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_LO12_NC:
|
|
// These are Local-Dynamic, which refer to local symbols in the
|
|
// dynamic TLS block. Since we know that we generating an
|
|
// executable, we can switch to Local-Exec.
|
|
return tls::TLSOPT_TO_LE;
|
|
|
|
case elfcpp::R_AARCH64_TLSIE_MOVW_GOTTPREL_G1:
|
|
case elfcpp::R_AARCH64_TLSIE_MOVW_GOTTPREL_G0_NC:
|
|
case elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSIE_LD_GOTTPREL_PREL19:
|
|
// These are Initial-Exec relocs which get the thread offset
|
|
// from the GOT. If we know that we are linking against the
|
|
// local symbol, we can switch to Local-Exec, which links the
|
|
// thread offset into the instruction.
|
|
if (is_final)
|
|
return tls::TLSOPT_TO_LE;
|
|
return tls::TLSOPT_NONE;
|
|
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G2:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12_NC:
|
|
// When we already have Local-Exec, there is nothing further we
|
|
// can do.
|
|
return tls::TLSOPT_NONE;
|
|
|
|
default:
|
|
gold_unreachable();
|
|
}
|
|
}
|
|
|
|
// Returns true if this relocation type could be that of a function pointer.
|
|
|
|
template<int size, bool big_endian>
|
|
inline bool
|
|
Target_aarch64<size, big_endian>::Scan::possible_function_pointer_reloc(
|
|
unsigned int r_type)
|
|
{
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_ADR_PREL_PG_HI21:
|
|
case elfcpp::R_AARCH64_ADR_PREL_PG_HI21_NC:
|
|
case elfcpp::R_AARCH64_ADD_ABS_LO12_NC:
|
|
case elfcpp::R_AARCH64_ADR_GOT_PAGE:
|
|
case elfcpp::R_AARCH64_LD64_GOT_LO12_NC:
|
|
{
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
// For safe ICF, scan a relocation for a local symbol to check if it
|
|
// corresponds to a function pointer being taken. In that case mark
|
|
// the function whose pointer was taken as not foldable.
|
|
|
|
template<int size, bool big_endian>
|
|
inline bool
|
|
Target_aarch64<size, big_endian>::Scan::local_reloc_may_be_function_pointer(
|
|
Symbol_table* ,
|
|
Layout* ,
|
|
Target_aarch64<size, big_endian>* ,
|
|
Sized_relobj_file<size, big_endian>* ,
|
|
unsigned int ,
|
|
Output_section* ,
|
|
const elfcpp::Rela<size, big_endian>& ,
|
|
unsigned int r_type,
|
|
const elfcpp::Sym<size, big_endian>&)
|
|
{
|
|
// When building a shared library, do not fold any local symbols.
|
|
return (parameters->options().shared()
|
|
|| possible_function_pointer_reloc(r_type));
|
|
}
|
|
|
|
// For safe ICF, scan a relocation for a global symbol to check if it
|
|
// corresponds to a function pointer being taken. In that case mark
|
|
// the function whose pointer was taken as not foldable.
|
|
|
|
template<int size, bool big_endian>
|
|
inline bool
|
|
Target_aarch64<size, big_endian>::Scan::global_reloc_may_be_function_pointer(
|
|
Symbol_table* ,
|
|
Layout* ,
|
|
Target_aarch64<size, big_endian>* ,
|
|
Sized_relobj_file<size, big_endian>* ,
|
|
unsigned int ,
|
|
Output_section* ,
|
|
const elfcpp::Rela<size, big_endian>& ,
|
|
unsigned int r_type,
|
|
Symbol* gsym)
|
|
{
|
|
// When building a shared library, do not fold symbols whose visibility
|
|
// is hidden, internal or protected.
|
|
return ((parameters->options().shared()
|
|
&& (gsym->visibility() == elfcpp::STV_INTERNAL
|
|
|| gsym->visibility() == elfcpp::STV_PROTECTED
|
|
|| gsym->visibility() == elfcpp::STV_HIDDEN))
|
|
|| possible_function_pointer_reloc(r_type));
|
|
}
|
|
|
|
// Report an unsupported relocation against a local symbol.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::Scan::unsupported_reloc_local(
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int r_type)
|
|
{
|
|
gold_error(_("%s: unsupported reloc %u against local symbol"),
|
|
object->name().c_str(), r_type);
|
|
}
|
|
|
|
// We are about to emit a dynamic relocation of type R_TYPE. If the
|
|
// dynamic linker does not support it, issue an error.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::Scan::check_non_pic(Relobj* object,
|
|
unsigned int r_type)
|
|
{
|
|
gold_assert(r_type != elfcpp::R_AARCH64_NONE);
|
|
|
|
switch (r_type)
|
|
{
|
|
// These are the relocation types supported by glibc for AARCH64.
|
|
case elfcpp::R_AARCH64_NONE:
|
|
case elfcpp::R_AARCH64_COPY:
|
|
case elfcpp::R_AARCH64_GLOB_DAT:
|
|
case elfcpp::R_AARCH64_JUMP_SLOT:
|
|
case elfcpp::R_AARCH64_RELATIVE:
|
|
case elfcpp::R_AARCH64_TLS_DTPREL64:
|
|
case elfcpp::R_AARCH64_TLS_DTPMOD64:
|
|
case elfcpp::R_AARCH64_TLS_TPREL64:
|
|
case elfcpp::R_AARCH64_TLSDESC:
|
|
case elfcpp::R_AARCH64_IRELATIVE:
|
|
case elfcpp::R_AARCH64_ABS32:
|
|
case elfcpp::R_AARCH64_ABS64:
|
|
return;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
// This prevents us from issuing more than one error per reloc
|
|
// section. But we can still wind up issuing more than one
|
|
// error per object file.
|
|
if (this->issued_non_pic_error_)
|
|
return;
|
|
gold_assert(parameters->options().output_is_position_independent());
|
|
object->error(_("requires unsupported dynamic reloc; "
|
|
"recompile with -fPIC"));
|
|
this->issued_non_pic_error_ = true;
|
|
return;
|
|
}
|
|
|
|
// Return whether we need to make a PLT entry for a relocation of the
|
|
// given type against a STT_GNU_IFUNC symbol.
|
|
|
|
template<int size, bool big_endian>
|
|
bool
|
|
Target_aarch64<size, big_endian>::Scan::reloc_needs_plt_for_ifunc(
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int r_type)
|
|
{
|
|
const AArch64_reloc_property* arp =
|
|
aarch64_reloc_property_table->get_reloc_property(r_type);
|
|
gold_assert(arp != NULL);
|
|
|
|
int flags = arp->reference_flags();
|
|
if (flags & Symbol::TLS_REF)
|
|
{
|
|
gold_error(_("%s: unsupported TLS reloc %s for IFUNC symbol"),
|
|
object->name().c_str(), arp->name().c_str());
|
|
return false;
|
|
}
|
|
return flags != 0;
|
|
}
|
|
|
|
// Scan a relocation for a local symbol.
|
|
|
|
template<int size, bool big_endian>
|
|
inline void
|
|
Target_aarch64<size, big_endian>::Scan::local(
|
|
Symbol_table* symtab,
|
|
Layout* layout,
|
|
Target_aarch64<size, big_endian>* target,
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int data_shndx,
|
|
Output_section* output_section,
|
|
const elfcpp::Rela<size, big_endian>& rela,
|
|
unsigned int r_type,
|
|
const elfcpp::Sym<size, big_endian>& lsym,
|
|
bool is_discarded)
|
|
{
|
|
if (is_discarded)
|
|
return;
|
|
|
|
typedef Output_data_reloc<elfcpp::SHT_RELA, true, size, big_endian>
|
|
Reloc_section;
|
|
unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info());
|
|
|
|
// A local STT_GNU_IFUNC symbol may require a PLT entry.
|
|
bool is_ifunc = lsym.get_st_type() == elfcpp::STT_GNU_IFUNC;
|
|
if (is_ifunc && this->reloc_needs_plt_for_ifunc(object, r_type))
|
|
target->make_local_ifunc_plt_entry(symtab, layout, object, r_sym);
|
|
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_NONE:
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_ABS32:
|
|
case elfcpp::R_AARCH64_ABS16:
|
|
if (parameters->options().output_is_position_independent())
|
|
{
|
|
gold_error(_("%s: unsupported reloc %u in pos independent link."),
|
|
object->name().c_str(), r_type);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_ABS64:
|
|
// If building a shared library or pie, we need to mark this as a dynmic
|
|
// reloction, so that the dynamic loader can relocate it.
|
|
if (parameters->options().output_is_position_independent())
|
|
{
|
|
Reloc_section* rela_dyn = target->rela_dyn_section(layout);
|
|
rela_dyn->add_local_relative(object, r_sym,
|
|
elfcpp::R_AARCH64_RELATIVE,
|
|
output_section,
|
|
data_shndx,
|
|
rela.get_r_offset(),
|
|
rela.get_r_addend(),
|
|
is_ifunc);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_PREL64:
|
|
case elfcpp::R_AARCH64_PREL32:
|
|
case elfcpp::R_AARCH64_PREL16:
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_ADR_GOT_PAGE:
|
|
case elfcpp::R_AARCH64_LD64_GOT_LO12_NC:
|
|
case elfcpp::R_AARCH64_LD64_GOTPAGE_LO15:
|
|
// The above relocations are used to access GOT entries.
|
|
{
|
|
Output_data_got_aarch64<size, big_endian>* got =
|
|
target->got_section(symtab, layout);
|
|
bool is_new = false;
|
|
// This symbol requires a GOT entry.
|
|
if (is_ifunc)
|
|
is_new = got->add_local_plt(object, r_sym, GOT_TYPE_STANDARD);
|
|
else
|
|
is_new = got->add_local(object, r_sym, GOT_TYPE_STANDARD);
|
|
if (is_new && parameters->options().output_is_position_independent())
|
|
target->rela_dyn_section(layout)->
|
|
add_local_relative(object,
|
|
r_sym,
|
|
elfcpp::R_AARCH64_RELATIVE,
|
|
got,
|
|
object->local_got_offset(r_sym,
|
|
GOT_TYPE_STANDARD),
|
|
0,
|
|
false);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G0: // 263
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G0_NC: // 264
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G1: // 265
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G1_NC: // 266
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G2: // 267
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G2_NC: // 268
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G3: // 269
|
|
case elfcpp::R_AARCH64_MOVW_SABS_G0: // 270
|
|
case elfcpp::R_AARCH64_MOVW_SABS_G1: // 271
|
|
case elfcpp::R_AARCH64_MOVW_SABS_G2: // 272
|
|
if (parameters->options().output_is_position_independent())
|
|
{
|
|
gold_error(_("%s: unsupported reloc %u in pos independent link."),
|
|
object->name().c_str(), r_type);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_LD_PREL_LO19: // 273
|
|
case elfcpp::R_AARCH64_ADR_PREL_LO21: // 274
|
|
case elfcpp::R_AARCH64_ADR_PREL_PG_HI21: // 275
|
|
case elfcpp::R_AARCH64_ADR_PREL_PG_HI21_NC: // 276
|
|
case elfcpp::R_AARCH64_ADD_ABS_LO12_NC: // 277
|
|
case elfcpp::R_AARCH64_LDST8_ABS_LO12_NC: // 278
|
|
case elfcpp::R_AARCH64_LDST16_ABS_LO12_NC: // 284
|
|
case elfcpp::R_AARCH64_LDST32_ABS_LO12_NC: // 285
|
|
case elfcpp::R_AARCH64_LDST64_ABS_LO12_NC: // 286
|
|
case elfcpp::R_AARCH64_LDST128_ABS_LO12_NC: // 299
|
|
break;
|
|
|
|
// Control flow, pc-relative. We don't need to do anything for a relative
|
|
// addressing relocation against a local symbol if it does not reference
|
|
// the GOT.
|
|
case elfcpp::R_AARCH64_TSTBR14:
|
|
case elfcpp::R_AARCH64_CONDBR19:
|
|
case elfcpp::R_AARCH64_JUMP26:
|
|
case elfcpp::R_AARCH64_CALL26:
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC:
|
|
{
|
|
tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>::
|
|
optimize_tls_reloc(!parameters->options().shared(), r_type);
|
|
if (tlsopt == tls::TLSOPT_TO_LE)
|
|
break;
|
|
|
|
layout->set_has_static_tls();
|
|
// Create a GOT entry for the tp-relative offset.
|
|
if (!parameters->doing_static_link())
|
|
{
|
|
Output_data_got_aarch64<size, big_endian>* got =
|
|
target->got_section(symtab, layout);
|
|
got->add_local_with_rel(object, r_sym, GOT_TYPE_TLS_OFFSET,
|
|
target->rela_dyn_section(layout),
|
|
elfcpp::R_AARCH64_TLS_TPREL64);
|
|
}
|
|
else if (!object->local_has_got_offset(r_sym,
|
|
GOT_TYPE_TLS_OFFSET))
|
|
{
|
|
Output_data_got_aarch64<size, big_endian>* got =
|
|
target->got_section(symtab, layout);
|
|
got->add_local(object, r_sym, GOT_TYPE_TLS_OFFSET);
|
|
unsigned int got_offset =
|
|
object->local_got_offset(r_sym, GOT_TYPE_TLS_OFFSET);
|
|
const elfcpp::Elf_Xword addend = rela.get_r_addend();
|
|
gold_assert(addend == 0);
|
|
got->add_static_reloc(got_offset, elfcpp::R_AARCH64_TLS_TPREL64,
|
|
object, r_sym);
|
|
}
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSGD_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC:
|
|
{
|
|
tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>::
|
|
optimize_tls_reloc(!parameters->options().shared(), r_type);
|
|
if (tlsopt == tls::TLSOPT_TO_LE)
|
|
{
|
|
layout->set_has_static_tls();
|
|
break;
|
|
}
|
|
gold_assert(tlsopt == tls::TLSOPT_NONE);
|
|
|
|
Output_data_got_aarch64<size, big_endian>* got =
|
|
target->got_section(symtab, layout);
|
|
got->add_local_pair_with_rel(object,r_sym, data_shndx,
|
|
GOT_TYPE_TLS_PAIR,
|
|
target->rela_dyn_section(layout),
|
|
elfcpp::R_AARCH64_TLS_DTPMOD64);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G2:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12_NC:
|
|
{
|
|
layout->set_has_static_tls();
|
|
bool output_is_shared = parameters->options().shared();
|
|
if (output_is_shared)
|
|
gold_error(_("%s: unsupported TLSLE reloc %u in shared code."),
|
|
object->name().c_str(), r_type);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSLD_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC:
|
|
{
|
|
tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>::
|
|
optimize_tls_reloc(!parameters->options().shared(), r_type);
|
|
if (tlsopt == tls::TLSOPT_NONE)
|
|
{
|
|
// Create a GOT entry for the module index.
|
|
target->got_mod_index_entry(symtab, layout, object);
|
|
}
|
|
else if (tlsopt != tls::TLSOPT_TO_LE)
|
|
unsupported_reloc_local(object, r_type);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G1:
|
|
case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G0_NC:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_HI12:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_LO12_NC:
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSDESC_LD64_LO12:
|
|
case elfcpp::R_AARCH64_TLSDESC_ADD_LO12:
|
|
{
|
|
tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>::
|
|
optimize_tls_reloc(!parameters->options().shared(), r_type);
|
|
target->define_tls_base_symbol(symtab, layout);
|
|
if (tlsopt == tls::TLSOPT_NONE)
|
|
{
|
|
// Create reserved PLT and GOT entries for the resolver.
|
|
target->reserve_tlsdesc_entries(symtab, layout);
|
|
|
|
// Generate a double GOT entry with an R_AARCH64_TLSDESC reloc.
|
|
// The R_AARCH64_TLSDESC reloc is resolved lazily, so the GOT
|
|
// entry needs to be in an area in .got.plt, not .got. Call
|
|
// got_section to make sure the section has been created.
|
|
target->got_section(symtab, layout);
|
|
Output_data_got<size, big_endian>* got =
|
|
target->got_tlsdesc_section();
|
|
unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info());
|
|
if (!object->local_has_got_offset(r_sym, GOT_TYPE_TLS_DESC))
|
|
{
|
|
unsigned int got_offset = got->add_constant(0);
|
|
got->add_constant(0);
|
|
object->set_local_got_offset(r_sym, GOT_TYPE_TLS_DESC,
|
|
got_offset);
|
|
Reloc_section* rt = target->rela_tlsdesc_section(layout);
|
|
// We store the arguments we need in a vector, and use
|
|
// the index into the vector as the parameter to pass
|
|
// to the target specific routines.
|
|
uintptr_t intarg = target->add_tlsdesc_info(object, r_sym);
|
|
void* arg = reinterpret_cast<void*>(intarg);
|
|
rt->add_target_specific(elfcpp::R_AARCH64_TLSDESC, arg,
|
|
got, got_offset, 0);
|
|
}
|
|
}
|
|
else if (tlsopt != tls::TLSOPT_TO_LE)
|
|
unsupported_reloc_local(object, r_type);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSDESC_CALL:
|
|
break;
|
|
|
|
default:
|
|
unsupported_reloc_local(object, r_type);
|
|
}
|
|
}
|
|
|
|
|
|
// Report an unsupported relocation against a global symbol.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::Scan::unsupported_reloc_global(
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int r_type,
|
|
Symbol* gsym)
|
|
{
|
|
gold_error(_("%s: unsupported reloc %u against global symbol %s"),
|
|
object->name().c_str(), r_type, gsym->demangled_name().c_str());
|
|
}
|
|
|
|
template<int size, bool big_endian>
|
|
inline void
|
|
Target_aarch64<size, big_endian>::Scan::global(
|
|
Symbol_table* symtab,
|
|
Layout* layout,
|
|
Target_aarch64<size, big_endian>* target,
|
|
Sized_relobj_file<size, big_endian> * object,
|
|
unsigned int data_shndx,
|
|
Output_section* output_section,
|
|
const elfcpp::Rela<size, big_endian>& rela,
|
|
unsigned int r_type,
|
|
Symbol* gsym)
|
|
{
|
|
// A STT_GNU_IFUNC symbol may require a PLT entry.
|
|
if (gsym->type() == elfcpp::STT_GNU_IFUNC
|
|
&& this->reloc_needs_plt_for_ifunc(object, r_type))
|
|
target->make_plt_entry(symtab, layout, gsym);
|
|
|
|
typedef Output_data_reloc<elfcpp::SHT_RELA, true, size, big_endian>
|
|
Reloc_section;
|
|
const AArch64_reloc_property* arp =
|
|
aarch64_reloc_property_table->get_reloc_property(r_type);
|
|
gold_assert(arp != NULL);
|
|
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_NONE:
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_ABS16:
|
|
case elfcpp::R_AARCH64_ABS32:
|
|
case elfcpp::R_AARCH64_ABS64:
|
|
{
|
|
// Make a PLT entry if necessary.
|
|
if (gsym->needs_plt_entry())
|
|
{
|
|
target->make_plt_entry(symtab, layout, gsym);
|
|
// Since this is not a PC-relative relocation, we may be
|
|
// taking the address of a function. In that case we need to
|
|
// set the entry in the dynamic symbol table to the address of
|
|
// the PLT entry.
|
|
if (gsym->is_from_dynobj() && !parameters->options().shared())
|
|
gsym->set_needs_dynsym_value();
|
|
}
|
|
// Make a dynamic relocation if necessary.
|
|
if (gsym->needs_dynamic_reloc(arp->reference_flags()))
|
|
{
|
|
if (!parameters->options().output_is_position_independent()
|
|
&& gsym->may_need_copy_reloc())
|
|
{
|
|
target->copy_reloc(symtab, layout, object,
|
|
data_shndx, output_section, gsym, rela);
|
|
}
|
|
else if (r_type == elfcpp::R_AARCH64_ABS64
|
|
&& gsym->type() == elfcpp::STT_GNU_IFUNC
|
|
&& gsym->can_use_relative_reloc(false)
|
|
&& !gsym->is_from_dynobj()
|
|
&& !gsym->is_undefined()
|
|
&& !gsym->is_preemptible())
|
|
{
|
|
// Use an IRELATIVE reloc for a locally defined STT_GNU_IFUNC
|
|
// symbol. This makes a function address in a PIE executable
|
|
// match the address in a shared library that it links against.
|
|
Reloc_section* rela_dyn =
|
|
target->rela_irelative_section(layout);
|
|
unsigned int r_type = elfcpp::R_AARCH64_IRELATIVE;
|
|
rela_dyn->add_symbolless_global_addend(gsym, r_type,
|
|
output_section, object,
|
|
data_shndx,
|
|
rela.get_r_offset(),
|
|
rela.get_r_addend());
|
|
}
|
|
else if (r_type == elfcpp::R_AARCH64_ABS64
|
|
&& gsym->can_use_relative_reloc(false))
|
|
{
|
|
Reloc_section* rela_dyn = target->rela_dyn_section(layout);
|
|
rela_dyn->add_global_relative(gsym,
|
|
elfcpp::R_AARCH64_RELATIVE,
|
|
output_section,
|
|
object,
|
|
data_shndx,
|
|
rela.get_r_offset(),
|
|
rela.get_r_addend(),
|
|
false);
|
|
}
|
|
else
|
|
{
|
|
check_non_pic(object, r_type);
|
|
Output_data_reloc<elfcpp::SHT_RELA, true, size, big_endian>*
|
|
rela_dyn = target->rela_dyn_section(layout);
|
|
rela_dyn->add_global(
|
|
gsym, r_type, output_section, object,
|
|
data_shndx, rela.get_r_offset(),rela.get_r_addend());
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_PREL16:
|
|
case elfcpp::R_AARCH64_PREL32:
|
|
case elfcpp::R_AARCH64_PREL64:
|
|
// This is used to fill the GOT absolute address.
|
|
if (gsym->needs_plt_entry())
|
|
{
|
|
target->make_plt_entry(symtab, layout, gsym);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G0: // 263
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G0_NC: // 264
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G1: // 265
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G1_NC: // 266
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G2: // 267
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G2_NC: // 268
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G3: // 269
|
|
case elfcpp::R_AARCH64_MOVW_SABS_G0: // 270
|
|
case elfcpp::R_AARCH64_MOVW_SABS_G1: // 271
|
|
case elfcpp::R_AARCH64_MOVW_SABS_G2: // 272
|
|
if (parameters->options().output_is_position_independent())
|
|
{
|
|
gold_error(_("%s: unsupported reloc %u in pos independent link."),
|
|
object->name().c_str(), r_type);
|
|
}
|
|
// Make a PLT entry if necessary.
|
|
if (gsym->needs_plt_entry())
|
|
{
|
|
target->make_plt_entry(symtab, layout, gsym);
|
|
// Since this is not a PC-relative relocation, we may be
|
|
// taking the address of a function. In that case we need to
|
|
// set the entry in the dynamic symbol table to the address of
|
|
// the PLT entry.
|
|
if (gsym->is_from_dynobj() && !parameters->options().shared())
|
|
gsym->set_needs_dynsym_value();
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_LD_PREL_LO19: // 273
|
|
case elfcpp::R_AARCH64_ADR_PREL_LO21: // 274
|
|
case elfcpp::R_AARCH64_ADR_PREL_PG_HI21: // 275
|
|
case elfcpp::R_AARCH64_ADR_PREL_PG_HI21_NC: // 276
|
|
case elfcpp::R_AARCH64_ADD_ABS_LO12_NC: // 277
|
|
case elfcpp::R_AARCH64_LDST8_ABS_LO12_NC: // 278
|
|
case elfcpp::R_AARCH64_LDST16_ABS_LO12_NC: // 284
|
|
case elfcpp::R_AARCH64_LDST32_ABS_LO12_NC: // 285
|
|
case elfcpp::R_AARCH64_LDST64_ABS_LO12_NC: // 286
|
|
case elfcpp::R_AARCH64_LDST128_ABS_LO12_NC: // 299
|
|
{
|
|
if (gsym->needs_plt_entry())
|
|
target->make_plt_entry(symtab, layout, gsym);
|
|
// Make a dynamic relocation if necessary.
|
|
if (gsym->needs_dynamic_reloc(arp->reference_flags()))
|
|
{
|
|
if (parameters->options().output_is_executable()
|
|
&& gsym->may_need_copy_reloc())
|
|
{
|
|
target->copy_reloc(symtab, layout, object,
|
|
data_shndx, output_section, gsym, rela);
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
|
|
case elfcpp::R_AARCH64_ADR_GOT_PAGE:
|
|
case elfcpp::R_AARCH64_LD64_GOT_LO12_NC:
|
|
case elfcpp::R_AARCH64_LD64_GOTPAGE_LO15:
|
|
{
|
|
// The above relocations are used to access GOT entries.
|
|
// Note a GOT entry is an *address* to a symbol.
|
|
// The symbol requires a GOT entry
|
|
Output_data_got_aarch64<size, big_endian>* got =
|
|
target->got_section(symtab, layout);
|
|
if (gsym->final_value_is_known())
|
|
{
|
|
// For a STT_GNU_IFUNC symbol we want the PLT address.
|
|
if (gsym->type() == elfcpp::STT_GNU_IFUNC)
|
|
got->add_global_plt(gsym, GOT_TYPE_STANDARD);
|
|
else
|
|
got->add_global(gsym, GOT_TYPE_STANDARD);
|
|
}
|
|
else
|
|
{
|
|
// If this symbol is not fully resolved, we need to add a dynamic
|
|
// relocation for it.
|
|
Reloc_section* rela_dyn = target->rela_dyn_section(layout);
|
|
|
|
// Use a GLOB_DAT rather than a RELATIVE reloc if:
|
|
//
|
|
// 1) The symbol may be defined in some other module.
|
|
// 2) We are building a shared library and this is a protected
|
|
// symbol; using GLOB_DAT means that the dynamic linker can use
|
|
// the address of the PLT in the main executable when appropriate
|
|
// so that function address comparisons work.
|
|
// 3) This is a STT_GNU_IFUNC symbol in position dependent code,
|
|
// again so that function address comparisons work.
|
|
if (gsym->is_from_dynobj()
|
|
|| gsym->is_undefined()
|
|
|| gsym->is_preemptible()
|
|
|| (gsym->visibility() == elfcpp::STV_PROTECTED
|
|
&& parameters->options().shared())
|
|
|| (gsym->type() == elfcpp::STT_GNU_IFUNC
|
|
&& parameters->options().output_is_position_independent()))
|
|
got->add_global_with_rel(gsym, GOT_TYPE_STANDARD,
|
|
rela_dyn, elfcpp::R_AARCH64_GLOB_DAT);
|
|
else
|
|
{
|
|
// For a STT_GNU_IFUNC symbol we want to write the PLT
|
|
// offset into the GOT, so that function pointer
|
|
// comparisons work correctly.
|
|
bool is_new;
|
|
if (gsym->type() != elfcpp::STT_GNU_IFUNC)
|
|
is_new = got->add_global(gsym, GOT_TYPE_STANDARD);
|
|
else
|
|
{
|
|
is_new = got->add_global_plt(gsym, GOT_TYPE_STANDARD);
|
|
// Tell the dynamic linker to use the PLT address
|
|
// when resolving relocations.
|
|
if (gsym->is_from_dynobj()
|
|
&& !parameters->options().shared())
|
|
gsym->set_needs_dynsym_value();
|
|
}
|
|
if (is_new)
|
|
{
|
|
rela_dyn->add_global_relative(
|
|
gsym, elfcpp::R_AARCH64_RELATIVE,
|
|
got,
|
|
gsym->got_offset(GOT_TYPE_STANDARD),
|
|
0,
|
|
false);
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
|
|
case elfcpp::R_AARCH64_TSTBR14:
|
|
case elfcpp::R_AARCH64_CONDBR19:
|
|
case elfcpp::R_AARCH64_JUMP26:
|
|
case elfcpp::R_AARCH64_CALL26:
|
|
{
|
|
if (gsym->final_value_is_known())
|
|
break;
|
|
|
|
if (gsym->is_defined() &&
|
|
!gsym->is_from_dynobj() &&
|
|
!gsym->is_preemptible())
|
|
break;
|
|
|
|
// Make plt entry for function call.
|
|
target->make_plt_entry(symtab, layout, gsym);
|
|
break;
|
|
}
|
|
|
|
case elfcpp::R_AARCH64_TLSGD_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC: // General dynamic
|
|
{
|
|
tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>::
|
|
optimize_tls_reloc(gsym->final_value_is_known(), r_type);
|
|
if (tlsopt == tls::TLSOPT_TO_LE)
|
|
{
|
|
layout->set_has_static_tls();
|
|
break;
|
|
}
|
|
gold_assert(tlsopt == tls::TLSOPT_NONE);
|
|
|
|
// General dynamic.
|
|
Output_data_got_aarch64<size, big_endian>* got =
|
|
target->got_section(symtab, layout);
|
|
// Create 2 consecutive entries for module index and offset.
|
|
got->add_global_pair_with_rel(gsym, GOT_TYPE_TLS_PAIR,
|
|
target->rela_dyn_section(layout),
|
|
elfcpp::R_AARCH64_TLS_DTPMOD64,
|
|
elfcpp::R_AARCH64_TLS_DTPREL64);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSLD_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC: // Local dynamic
|
|
{
|
|
tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>::
|
|
optimize_tls_reloc(!parameters->options().shared(), r_type);
|
|
if (tlsopt == tls::TLSOPT_NONE)
|
|
{
|
|
// Create a GOT entry for the module index.
|
|
target->got_mod_index_entry(symtab, layout, object);
|
|
}
|
|
else if (tlsopt != tls::TLSOPT_TO_LE)
|
|
unsupported_reloc_local(object, r_type);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G1:
|
|
case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G0_NC:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_HI12:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_LO12_NC: // Other local dynamic
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC: // Initial executable
|
|
{
|
|
tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>::
|
|
optimize_tls_reloc(gsym->final_value_is_known(), r_type);
|
|
if (tlsopt == tls::TLSOPT_TO_LE)
|
|
break;
|
|
|
|
layout->set_has_static_tls();
|
|
// Create a GOT entry for the tp-relative offset.
|
|
Output_data_got_aarch64<size, big_endian>* got
|
|
= target->got_section(symtab, layout);
|
|
if (!parameters->doing_static_link())
|
|
{
|
|
got->add_global_with_rel(
|
|
gsym, GOT_TYPE_TLS_OFFSET,
|
|
target->rela_dyn_section(layout),
|
|
elfcpp::R_AARCH64_TLS_TPREL64);
|
|
}
|
|
if (!gsym->has_got_offset(GOT_TYPE_TLS_OFFSET))
|
|
{
|
|
got->add_global(gsym, GOT_TYPE_TLS_OFFSET);
|
|
unsigned int got_offset =
|
|
gsym->got_offset(GOT_TYPE_TLS_OFFSET);
|
|
const elfcpp::Elf_Xword addend = rela.get_r_addend();
|
|
gold_assert(addend == 0);
|
|
got->add_static_reloc(got_offset,
|
|
elfcpp::R_AARCH64_TLS_TPREL64, gsym);
|
|
}
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G2:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12_NC: // Local executable
|
|
layout->set_has_static_tls();
|
|
if (parameters->options().shared())
|
|
gold_error(_("%s: unsupported TLSLE reloc type %u in shared objects."),
|
|
object->name().c_str(), r_type);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSDESC_LD64_LO12:
|
|
case elfcpp::R_AARCH64_TLSDESC_ADD_LO12: // TLS descriptor
|
|
{
|
|
target->define_tls_base_symbol(symtab, layout);
|
|
tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>::
|
|
optimize_tls_reloc(gsym->final_value_is_known(), r_type);
|
|
if (tlsopt == tls::TLSOPT_NONE)
|
|
{
|
|
// Create reserved PLT and GOT entries for the resolver.
|
|
target->reserve_tlsdesc_entries(symtab, layout);
|
|
|
|
// Create a double GOT entry with an R_AARCH64_TLSDESC
|
|
// relocation. The R_AARCH64_TLSDESC is resolved lazily, so the GOT
|
|
// entry needs to be in an area in .got.plt, not .got. Call
|
|
// got_section to make sure the section has been created.
|
|
target->got_section(symtab, layout);
|
|
Output_data_got<size, big_endian>* got =
|
|
target->got_tlsdesc_section();
|
|
Reloc_section* rt = target->rela_tlsdesc_section(layout);
|
|
got->add_global_pair_with_rel(gsym, GOT_TYPE_TLS_DESC, rt,
|
|
elfcpp::R_AARCH64_TLSDESC, 0);
|
|
}
|
|
else if (tlsopt == tls::TLSOPT_TO_IE)
|
|
{
|
|
// Create a GOT entry for the tp-relative offset.
|
|
Output_data_got<size, big_endian>* got
|
|
= target->got_section(symtab, layout);
|
|
got->add_global_with_rel(gsym, GOT_TYPE_TLS_OFFSET,
|
|
target->rela_dyn_section(layout),
|
|
elfcpp::R_AARCH64_TLS_TPREL64);
|
|
}
|
|
else if (tlsopt != tls::TLSOPT_TO_LE)
|
|
unsupported_reloc_global(object, r_type, gsym);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSDESC_CALL:
|
|
break;
|
|
|
|
default:
|
|
gold_error(_("%s: unsupported reloc type in global scan"),
|
|
aarch64_reloc_property_table->
|
|
reloc_name_in_error_message(r_type).c_str());
|
|
}
|
|
return;
|
|
} // End of Scan::global
|
|
|
|
|
|
// Create the PLT section.
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::make_plt_section(
|
|
Symbol_table* symtab, Layout* layout)
|
|
{
|
|
if (this->plt_ == NULL)
|
|
{
|
|
// Create the GOT section first.
|
|
this->got_section(symtab, layout);
|
|
|
|
this->plt_ = this->make_data_plt(layout, this->got_, this->got_plt_,
|
|
this->got_irelative_);
|
|
|
|
layout->add_output_section_data(".plt", elfcpp::SHT_PROGBITS,
|
|
(elfcpp::SHF_ALLOC
|
|
| elfcpp::SHF_EXECINSTR),
|
|
this->plt_, ORDER_PLT, false);
|
|
|
|
// Make the sh_info field of .rela.plt point to .plt.
|
|
Output_section* rela_plt_os = this->plt_->rela_plt()->output_section();
|
|
rela_plt_os->set_info_section(this->plt_->output_section());
|
|
}
|
|
}
|
|
|
|
// Return the section for TLSDESC relocations.
|
|
|
|
template<int size, bool big_endian>
|
|
typename Target_aarch64<size, big_endian>::Reloc_section*
|
|
Target_aarch64<size, big_endian>::rela_tlsdesc_section(Layout* layout) const
|
|
{
|
|
return this->plt_section()->rela_tlsdesc(layout);
|
|
}
|
|
|
|
// Create a PLT entry for a global symbol.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::make_plt_entry(
|
|
Symbol_table* symtab,
|
|
Layout* layout,
|
|
Symbol* gsym)
|
|
{
|
|
if (gsym->has_plt_offset())
|
|
return;
|
|
|
|
if (this->plt_ == NULL)
|
|
this->make_plt_section(symtab, layout);
|
|
|
|
this->plt_->add_entry(symtab, layout, gsym);
|
|
}
|
|
|
|
// Make a PLT entry for a local STT_GNU_IFUNC symbol.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::make_local_ifunc_plt_entry(
|
|
Symbol_table* symtab, Layout* layout,
|
|
Sized_relobj_file<size, big_endian>* relobj,
|
|
unsigned int local_sym_index)
|
|
{
|
|
if (relobj->local_has_plt_offset(local_sym_index))
|
|
return;
|
|
if (this->plt_ == NULL)
|
|
this->make_plt_section(symtab, layout);
|
|
unsigned int plt_offset = this->plt_->add_local_ifunc_entry(symtab, layout,
|
|
relobj,
|
|
local_sym_index);
|
|
relobj->set_local_plt_offset(local_sym_index, plt_offset);
|
|
}
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::gc_process_relocs(
|
|
Symbol_table* symtab,
|
|
Layout* layout,
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int data_shndx,
|
|
unsigned int sh_type,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
bool needs_special_offset_handling,
|
|
size_t local_symbol_count,
|
|
const unsigned char* plocal_symbols)
|
|
{
|
|
typedef Target_aarch64<size, big_endian> Aarch64;
|
|
typedef gold::Default_classify_reloc<elfcpp::SHT_RELA, size, big_endian>
|
|
Classify_reloc;
|
|
|
|
if (sh_type == elfcpp::SHT_REL)
|
|
{
|
|
return;
|
|
}
|
|
|
|
gold::gc_process_relocs<size, big_endian, Aarch64, Scan, Classify_reloc>(
|
|
symtab,
|
|
layout,
|
|
this,
|
|
object,
|
|
data_shndx,
|
|
prelocs,
|
|
reloc_count,
|
|
output_section,
|
|
needs_special_offset_handling,
|
|
local_symbol_count,
|
|
plocal_symbols);
|
|
}
|
|
|
|
// Scan relocations for a section.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::scan_relocs(
|
|
Symbol_table* symtab,
|
|
Layout* layout,
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int data_shndx,
|
|
unsigned int sh_type,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
bool needs_special_offset_handling,
|
|
size_t local_symbol_count,
|
|
const unsigned char* plocal_symbols)
|
|
{
|
|
typedef Target_aarch64<size, big_endian> Aarch64;
|
|
typedef gold::Default_classify_reloc<elfcpp::SHT_RELA, size, big_endian>
|
|
Classify_reloc;
|
|
|
|
if (sh_type == elfcpp::SHT_REL)
|
|
{
|
|
gold_error(_("%s: unsupported REL reloc section"),
|
|
object->name().c_str());
|
|
return;
|
|
}
|
|
|
|
gold::scan_relocs<size, big_endian, Aarch64, Scan, Classify_reloc>(
|
|
symtab,
|
|
layout,
|
|
this,
|
|
object,
|
|
data_shndx,
|
|
prelocs,
|
|
reloc_count,
|
|
output_section,
|
|
needs_special_offset_handling,
|
|
local_symbol_count,
|
|
plocal_symbols);
|
|
}
|
|
|
|
// Return the value to use for a dynamic which requires special
|
|
// treatment. This is how we support equality comparisons of function
|
|
// pointers across shared library boundaries, as described in the
|
|
// processor specific ABI supplement.
|
|
|
|
template<int size, bool big_endian>
|
|
uint64_t
|
|
Target_aarch64<size, big_endian>::do_dynsym_value(const Symbol* gsym) const
|
|
{
|
|
gold_assert(gsym->is_from_dynobj() && gsym->has_plt_offset());
|
|
return this->plt_address_for_global(gsym);
|
|
}
|
|
|
|
|
|
// Finalize the sections.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::do_finalize_sections(
|
|
Layout* layout,
|
|
const Input_objects*,
|
|
Symbol_table* symtab)
|
|
{
|
|
const Reloc_section* rel_plt = (this->plt_ == NULL
|
|
? NULL
|
|
: this->plt_->rela_plt());
|
|
layout->add_target_dynamic_tags(false, this->got_plt_, rel_plt,
|
|
this->rela_dyn_, true, false, false);
|
|
|
|
// Emit any relocs we saved in an attempt to avoid generating COPY
|
|
// relocs.
|
|
if (this->copy_relocs_.any_saved_relocs())
|
|
this->copy_relocs_.emit(this->rela_dyn_section(layout));
|
|
|
|
// Fill in some more dynamic tags.
|
|
Output_data_dynamic* const odyn = layout->dynamic_data();
|
|
if (odyn != NULL)
|
|
{
|
|
if (this->plt_ != NULL
|
|
&& this->plt_->output_section() != NULL
|
|
&& this->plt_ ->has_tlsdesc_entry())
|
|
{
|
|
unsigned int plt_offset = this->plt_->get_tlsdesc_plt_offset();
|
|
unsigned int got_offset = this->plt_->get_tlsdesc_got_offset();
|
|
this->got_->finalize_data_size();
|
|
odyn->add_section_plus_offset(elfcpp::DT_TLSDESC_PLT,
|
|
this->plt_, plt_offset);
|
|
odyn->add_section_plus_offset(elfcpp::DT_TLSDESC_GOT,
|
|
this->got_, got_offset);
|
|
}
|
|
}
|
|
|
|
// Set the size of the _GLOBAL_OFFSET_TABLE_ symbol to the size of
|
|
// the .got section.
|
|
Symbol* sym = this->global_offset_table_;
|
|
if (sym != NULL)
|
|
{
|
|
uint64_t data_size = this->got_->current_data_size();
|
|
symtab->get_sized_symbol<size>(sym)->set_symsize(data_size);
|
|
|
|
// If the .got section is more than 0x8000 bytes, we add
|
|
// 0x8000 to the value of _GLOBAL_OFFSET_TABLE_, so that 16
|
|
// bit relocations have a greater chance of working.
|
|
if (data_size >= 0x8000)
|
|
symtab->get_sized_symbol<size>(sym)->set_value(
|
|
symtab->get_sized_symbol<size>(sym)->value() + 0x8000);
|
|
}
|
|
|
|
if (parameters->doing_static_link()
|
|
&& (this->plt_ == NULL || !this->plt_->has_irelative_section()))
|
|
{
|
|
// If linking statically, make sure that the __rela_iplt symbols
|
|
// were defined if necessary, even if we didn't create a PLT.
|
|
static const Define_symbol_in_segment syms[] =
|
|
{
|
|
{
|
|
"__rela_iplt_start", // name
|
|
elfcpp::PT_LOAD, // segment_type
|
|
elfcpp::PF_W, // segment_flags_set
|
|
elfcpp::PF(0), // segment_flags_clear
|
|
0, // value
|
|
0, // size
|
|
elfcpp::STT_NOTYPE, // type
|
|
elfcpp::STB_GLOBAL, // binding
|
|
elfcpp::STV_HIDDEN, // visibility
|
|
0, // nonvis
|
|
Symbol::SEGMENT_START, // offset_from_base
|
|
true // only_if_ref
|
|
},
|
|
{
|
|
"__rela_iplt_end", // name
|
|
elfcpp::PT_LOAD, // segment_type
|
|
elfcpp::PF_W, // segment_flags_set
|
|
elfcpp::PF(0), // segment_flags_clear
|
|
0, // value
|
|
0, // size
|
|
elfcpp::STT_NOTYPE, // type
|
|
elfcpp::STB_GLOBAL, // binding
|
|
elfcpp::STV_HIDDEN, // visibility
|
|
0, // nonvis
|
|
Symbol::SEGMENT_START, // offset_from_base
|
|
true // only_if_ref
|
|
}
|
|
};
|
|
|
|
symtab->define_symbols(layout, 2, syms,
|
|
layout->script_options()->saw_sections_clause());
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
// Perform a relocation.
|
|
|
|
template<int size, bool big_endian>
|
|
inline bool
|
|
Target_aarch64<size, big_endian>::Relocate::relocate(
|
|
const Relocate_info<size, big_endian>* relinfo,
|
|
unsigned int,
|
|
Target_aarch64<size, big_endian>* target,
|
|
Output_section* ,
|
|
size_t relnum,
|
|
const unsigned char* preloc,
|
|
const Sized_symbol<size>* gsym,
|
|
const Symbol_value<size>* psymval,
|
|
unsigned char* view,
|
|
typename elfcpp::Elf_types<size>::Elf_Addr address,
|
|
section_size_type /* view_size */)
|
|
{
|
|
if (view == NULL)
|
|
return true;
|
|
|
|
typedef AArch64_relocate_functions<size, big_endian> Reloc;
|
|
|
|
const elfcpp::Rela<size, big_endian> rela(preloc);
|
|
unsigned int r_type = elfcpp::elf_r_type<size>(rela.get_r_info());
|
|
const AArch64_reloc_property* reloc_property =
|
|
aarch64_reloc_property_table->get_reloc_property(r_type);
|
|
|
|
if (reloc_property == NULL)
|
|
{
|
|
std::string reloc_name =
|
|
aarch64_reloc_property_table->reloc_name_in_error_message(r_type);
|
|
gold_error_at_location(relinfo, relnum, rela.get_r_offset(),
|
|
_("cannot relocate %s in object file"),
|
|
reloc_name.c_str());
|
|
return true;
|
|
}
|
|
|
|
const Sized_relobj_file<size, big_endian>* object = relinfo->object;
|
|
|
|
// Pick the value to use for symbols defined in the PLT.
|
|
Symbol_value<size> symval;
|
|
if (gsym != NULL
|
|
&& gsym->use_plt_offset(reloc_property->reference_flags()))
|
|
{
|
|
symval.set_output_value(target->plt_address_for_global(gsym));
|
|
psymval = &symval;
|
|
}
|
|
else if (gsym == NULL && psymval->is_ifunc_symbol())
|
|
{
|
|
unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info());
|
|
if (object->local_has_plt_offset(r_sym))
|
|
{
|
|
symval.set_output_value(target->plt_address_for_local(object, r_sym));
|
|
psymval = &symval;
|
|
}
|
|
}
|
|
|
|
const elfcpp::Elf_Xword addend = rela.get_r_addend();
|
|
|
|
// Get the GOT offset if needed.
|
|
// For aarch64, the GOT pointer points to the start of the GOT section.
|
|
bool have_got_offset = false;
|
|
int got_offset = 0;
|
|
int got_base = (target->got_ != NULL
|
|
? (target->got_->current_data_size() >= 0x8000
|
|
? 0x8000 : 0)
|
|
: 0);
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_MOVW_GOTOFF_G0:
|
|
case elfcpp::R_AARCH64_MOVW_GOTOFF_G0_NC:
|
|
case elfcpp::R_AARCH64_MOVW_GOTOFF_G1:
|
|
case elfcpp::R_AARCH64_MOVW_GOTOFF_G1_NC:
|
|
case elfcpp::R_AARCH64_MOVW_GOTOFF_G2:
|
|
case elfcpp::R_AARCH64_MOVW_GOTOFF_G2_NC:
|
|
case elfcpp::R_AARCH64_MOVW_GOTOFF_G3:
|
|
case elfcpp::R_AARCH64_GOTREL64:
|
|
case elfcpp::R_AARCH64_GOTREL32:
|
|
case elfcpp::R_AARCH64_GOT_LD_PREL19:
|
|
case elfcpp::R_AARCH64_LD64_GOTOFF_LO15:
|
|
case elfcpp::R_AARCH64_ADR_GOT_PAGE:
|
|
case elfcpp::R_AARCH64_LD64_GOT_LO12_NC:
|
|
case elfcpp::R_AARCH64_LD64_GOTPAGE_LO15:
|
|
if (gsym != NULL)
|
|
{
|
|
gold_assert(gsym->has_got_offset(GOT_TYPE_STANDARD));
|
|
got_offset = gsym->got_offset(GOT_TYPE_STANDARD) - got_base;
|
|
}
|
|
else
|
|
{
|
|
unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info());
|
|
gold_assert(object->local_has_got_offset(r_sym, GOT_TYPE_STANDARD));
|
|
got_offset = (object->local_got_offset(r_sym, GOT_TYPE_STANDARD)
|
|
- got_base);
|
|
}
|
|
have_got_offset = true;
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
typename Reloc::Status reloc_status = Reloc::STATUS_OKAY;
|
|
typename elfcpp::Elf_types<size>::Elf_Addr value;
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_NONE:
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_ABS64:
|
|
if (!parameters->options().apply_dynamic_relocs()
|
|
&& parameters->options().output_is_position_independent()
|
|
&& gsym != NULL
|
|
&& gsym->needs_dynamic_reloc(reloc_property->reference_flags())
|
|
&& !gsym->can_use_relative_reloc(false))
|
|
// We have generated an absolute dynamic relocation, so do not
|
|
// apply the relocation statically. (Works around bugs in older
|
|
// Android dynamic linkers.)
|
|
break;
|
|
reloc_status = Reloc::template rela_ua<64>(
|
|
view, object, psymval, addend, reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_ABS32:
|
|
if (!parameters->options().apply_dynamic_relocs()
|
|
&& parameters->options().output_is_position_independent()
|
|
&& gsym != NULL
|
|
&& gsym->needs_dynamic_reloc(reloc_property->reference_flags()))
|
|
// We have generated an absolute dynamic relocation, so do not
|
|
// apply the relocation statically. (Works around bugs in older
|
|
// Android dynamic linkers.)
|
|
break;
|
|
reloc_status = Reloc::template rela_ua<32>(
|
|
view, object, psymval, addend, reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_ABS16:
|
|
if (!parameters->options().apply_dynamic_relocs()
|
|
&& parameters->options().output_is_position_independent()
|
|
&& gsym != NULL
|
|
&& gsym->needs_dynamic_reloc(reloc_property->reference_flags()))
|
|
// We have generated an absolute dynamic relocation, so do not
|
|
// apply the relocation statically. (Works around bugs in older
|
|
// Android dynamic linkers.)
|
|
break;
|
|
reloc_status = Reloc::template rela_ua<16>(
|
|
view, object, psymval, addend, reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_PREL64:
|
|
reloc_status = Reloc::template pcrela_ua<64>(
|
|
view, object, psymval, addend, address, reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_PREL32:
|
|
reloc_status = Reloc::template pcrela_ua<32>(
|
|
view, object, psymval, addend, address, reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_PREL16:
|
|
reloc_status = Reloc::template pcrela_ua<16>(
|
|
view, object, psymval, addend, address, reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G0:
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G0_NC:
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G1:
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G1_NC:
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G2:
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G2_NC:
|
|
case elfcpp::R_AARCH64_MOVW_UABS_G3:
|
|
reloc_status = Reloc::template rela_general<32>(
|
|
view, object, psymval, addend, reloc_property);
|
|
break;
|
|
case elfcpp::R_AARCH64_MOVW_SABS_G0:
|
|
case elfcpp::R_AARCH64_MOVW_SABS_G1:
|
|
case elfcpp::R_AARCH64_MOVW_SABS_G2:
|
|
reloc_status = Reloc::movnz(view, psymval->value(object, addend),
|
|
reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_LD_PREL_LO19:
|
|
reloc_status = Reloc::template pcrela_general<32>(
|
|
view, object, psymval, addend, address, reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_ADR_PREL_LO21:
|
|
reloc_status = Reloc::adr(view, object, psymval, addend,
|
|
address, reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_ADR_PREL_PG_HI21_NC:
|
|
case elfcpp::R_AARCH64_ADR_PREL_PG_HI21:
|
|
reloc_status = Reloc::adrp(view, object, psymval, addend, address,
|
|
reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_LDST8_ABS_LO12_NC:
|
|
case elfcpp::R_AARCH64_LDST16_ABS_LO12_NC:
|
|
case elfcpp::R_AARCH64_LDST32_ABS_LO12_NC:
|
|
case elfcpp::R_AARCH64_LDST64_ABS_LO12_NC:
|
|
case elfcpp::R_AARCH64_LDST128_ABS_LO12_NC:
|
|
case elfcpp::R_AARCH64_ADD_ABS_LO12_NC:
|
|
reloc_status = Reloc::template rela_general<32>(
|
|
view, object, psymval, addend, reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_CALL26:
|
|
if (this->skip_call_tls_get_addr_)
|
|
{
|
|
// Double check that the TLSGD insn has been optimized away.
|
|
typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype;
|
|
Insntype insn = elfcpp::Swap<32, big_endian>::readval(
|
|
reinterpret_cast<Insntype*>(view));
|
|
gold_assert((insn & 0xff000000) == 0x91000000);
|
|
|
|
reloc_status = Reloc::STATUS_OKAY;
|
|
this->skip_call_tls_get_addr_ = false;
|
|
// Return false to stop further processing this reloc.
|
|
return false;
|
|
}
|
|
// Fall through.
|
|
case elfcpp::R_AARCH64_JUMP26:
|
|
if (Reloc::maybe_apply_stub(r_type, relinfo, rela, view, address,
|
|
gsym, psymval, object,
|
|
target->stub_group_size_))
|
|
break;
|
|
// Fall through.
|
|
case elfcpp::R_AARCH64_TSTBR14:
|
|
case elfcpp::R_AARCH64_CONDBR19:
|
|
reloc_status = Reloc::template pcrela_general<32>(
|
|
view, object, psymval, addend, address, reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_ADR_GOT_PAGE:
|
|
gold_assert(have_got_offset);
|
|
value = target->got_->address() + got_base + got_offset;
|
|
reloc_status = Reloc::adrp(view, value + addend, address);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_LD64_GOT_LO12_NC:
|
|
gold_assert(have_got_offset);
|
|
value = target->got_->address() + got_base + got_offset;
|
|
reloc_status = Reloc::template rela_general<32>(
|
|
view, value, addend, reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_LD64_GOTPAGE_LO15:
|
|
{
|
|
gold_assert(have_got_offset);
|
|
value = target->got_->address() + got_base + got_offset + addend -
|
|
Reloc::Page(target->got_->address() + got_base);
|
|
if ((value & 7) != 0)
|
|
reloc_status = Reloc::STATUS_OVERFLOW;
|
|
else
|
|
reloc_status = Reloc::template reloc_common<32>(
|
|
view, value, reloc_property);
|
|
break;
|
|
}
|
|
|
|
case elfcpp::R_AARCH64_TLSGD_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLD_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G1:
|
|
case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G0_NC:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_HI12:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G2:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSDESC_LD64_LO12:
|
|
case elfcpp::R_AARCH64_TLSDESC_ADD_LO12:
|
|
case elfcpp::R_AARCH64_TLSDESC_CALL:
|
|
reloc_status = relocate_tls(relinfo, target, relnum, rela, r_type,
|
|
gsym, psymval, view, address);
|
|
break;
|
|
|
|
// These are dynamic relocations, which are unexpected when linking.
|
|
case elfcpp::R_AARCH64_COPY:
|
|
case elfcpp::R_AARCH64_GLOB_DAT:
|
|
case elfcpp::R_AARCH64_JUMP_SLOT:
|
|
case elfcpp::R_AARCH64_RELATIVE:
|
|
case elfcpp::R_AARCH64_IRELATIVE:
|
|
case elfcpp::R_AARCH64_TLS_DTPREL64:
|
|
case elfcpp::R_AARCH64_TLS_DTPMOD64:
|
|
case elfcpp::R_AARCH64_TLS_TPREL64:
|
|
case elfcpp::R_AARCH64_TLSDESC:
|
|
gold_error_at_location(relinfo, relnum, rela.get_r_offset(),
|
|
_("unexpected reloc %u in object file"),
|
|
r_type);
|
|
break;
|
|
|
|
default:
|
|
gold_error_at_location(relinfo, relnum, rela.get_r_offset(),
|
|
_("unsupported reloc %s"),
|
|
reloc_property->name().c_str());
|
|
break;
|
|
}
|
|
|
|
// Report any errors.
|
|
switch (reloc_status)
|
|
{
|
|
case Reloc::STATUS_OKAY:
|
|
break;
|
|
case Reloc::STATUS_OVERFLOW:
|
|
gold_error_at_location(relinfo, relnum, rela.get_r_offset(),
|
|
_("relocation overflow in %s"),
|
|
reloc_property->name().c_str());
|
|
break;
|
|
case Reloc::STATUS_BAD_RELOC:
|
|
gold_error_at_location(
|
|
relinfo,
|
|
relnum,
|
|
rela.get_r_offset(),
|
|
_("unexpected opcode while processing relocation %s"),
|
|
reloc_property->name().c_str());
|
|
break;
|
|
default:
|
|
gold_unreachable();
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
|
|
template<int size, bool big_endian>
|
|
inline
|
|
typename AArch64_relocate_functions<size, big_endian>::Status
|
|
Target_aarch64<size, big_endian>::Relocate::relocate_tls(
|
|
const Relocate_info<size, big_endian>* relinfo,
|
|
Target_aarch64<size, big_endian>* target,
|
|
size_t relnum,
|
|
const elfcpp::Rela<size, big_endian>& rela,
|
|
unsigned int r_type, const Sized_symbol<size>* gsym,
|
|
const Symbol_value<size>* psymval,
|
|
unsigned char* view,
|
|
typename elfcpp::Elf_types<size>::Elf_Addr address)
|
|
{
|
|
typedef AArch64_relocate_functions<size, big_endian> aarch64_reloc_funcs;
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address;
|
|
|
|
Output_segment* tls_segment = relinfo->layout->tls_segment();
|
|
const elfcpp::Elf_Xword addend = rela.get_r_addend();
|
|
const AArch64_reloc_property* reloc_property =
|
|
aarch64_reloc_property_table->get_reloc_property(r_type);
|
|
gold_assert(reloc_property != NULL);
|
|
|
|
const bool is_final = (gsym == NULL
|
|
? !parameters->options().shared()
|
|
: gsym->final_value_is_known());
|
|
tls::Tls_optimization tlsopt = Target_aarch64<size, big_endian>::
|
|
optimize_tls_reloc(is_final, r_type);
|
|
|
|
Sized_relobj_file<size, big_endian>* object = relinfo->object;
|
|
int tls_got_offset_type;
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_TLSGD_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC: // Global-dynamic
|
|
{
|
|
if (tlsopt == tls::TLSOPT_TO_LE)
|
|
{
|
|
if (tls_segment == NULL)
|
|
{
|
|
gold_assert(parameters->errors()->error_count() > 0
|
|
|| issue_undefined_symbol_error(gsym));
|
|
return aarch64_reloc_funcs::STATUS_BAD_RELOC;
|
|
}
|
|
return tls_gd_to_le(relinfo, target, rela, r_type, view,
|
|
psymval);
|
|
}
|
|
else if (tlsopt == tls::TLSOPT_NONE)
|
|
{
|
|
tls_got_offset_type = GOT_TYPE_TLS_PAIR;
|
|
// Firstly get the address for the got entry.
|
|
typename elfcpp::Elf_types<size>::Elf_Addr got_entry_address;
|
|
if (gsym != NULL)
|
|
{
|
|
gold_assert(gsym->has_got_offset(tls_got_offset_type));
|
|
got_entry_address = target->got_->address() +
|
|
gsym->got_offset(tls_got_offset_type);
|
|
}
|
|
else
|
|
{
|
|
unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info());
|
|
gold_assert(
|
|
object->local_has_got_offset(r_sym, tls_got_offset_type));
|
|
got_entry_address = target->got_->address() +
|
|
object->local_got_offset(r_sym, tls_got_offset_type);
|
|
}
|
|
|
|
// Relocate the address into adrp/ld, adrp/add pair.
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_TLSGD_ADR_PAGE21:
|
|
return aarch64_reloc_funcs::adrp(
|
|
view, got_entry_address + addend, address);
|
|
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC:
|
|
return aarch64_reloc_funcs::template rela_general<32>(
|
|
view, got_entry_address, addend, reloc_property);
|
|
break;
|
|
|
|
default:
|
|
gold_unreachable();
|
|
}
|
|
}
|
|
gold_error_at_location(relinfo, relnum, rela.get_r_offset(),
|
|
_("unsupported gd_to_ie relaxation on %u"),
|
|
r_type);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSLD_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC: // Local-dynamic
|
|
{
|
|
if (tlsopt == tls::TLSOPT_TO_LE)
|
|
{
|
|
if (tls_segment == NULL)
|
|
{
|
|
gold_assert(parameters->errors()->error_count() > 0
|
|
|| issue_undefined_symbol_error(gsym));
|
|
return aarch64_reloc_funcs::STATUS_BAD_RELOC;
|
|
}
|
|
return this->tls_ld_to_le(relinfo, target, rela, r_type, view,
|
|
psymval);
|
|
}
|
|
|
|
gold_assert(tlsopt == tls::TLSOPT_NONE);
|
|
// Relocate the field with the offset of the GOT entry for
|
|
// the module index.
|
|
typename elfcpp::Elf_types<size>::Elf_Addr got_entry_address;
|
|
got_entry_address = (target->got_mod_index_entry(NULL, NULL, NULL) +
|
|
target->got_->address());
|
|
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_TLSLD_ADR_PAGE21:
|
|
return aarch64_reloc_funcs::adrp(
|
|
view, got_entry_address + addend, address);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC:
|
|
return aarch64_reloc_funcs::template rela_general<32>(
|
|
view, got_entry_address, addend, reloc_property);
|
|
break;
|
|
|
|
default:
|
|
gold_unreachable();
|
|
}
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G1:
|
|
case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G0_NC:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_HI12:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_LO12_NC: // Other local-dynamic
|
|
{
|
|
AArch64_address value = psymval->value(object, 0);
|
|
if (tlsopt == tls::TLSOPT_TO_LE)
|
|
{
|
|
if (tls_segment == NULL)
|
|
{
|
|
gold_assert(parameters->errors()->error_count() > 0
|
|
|| issue_undefined_symbol_error(gsym));
|
|
return aarch64_reloc_funcs::STATUS_BAD_RELOC;
|
|
}
|
|
}
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G1:
|
|
return aarch64_reloc_funcs::movnz(view, value + addend,
|
|
reloc_property);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSLD_MOVW_DTPREL_G0_NC:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_HI12:
|
|
case elfcpp::R_AARCH64_TLSLD_ADD_DTPREL_LO12_NC:
|
|
return aarch64_reloc_funcs::template rela_general<32>(
|
|
view, value, addend, reloc_property);
|
|
break;
|
|
|
|
default:
|
|
gold_unreachable();
|
|
}
|
|
// We should never reach here.
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC: // Initial-exec
|
|
{
|
|
if (tlsopt == tls::TLSOPT_TO_LE)
|
|
{
|
|
if (tls_segment == NULL)
|
|
{
|
|
gold_assert(parameters->errors()->error_count() > 0
|
|
|| issue_undefined_symbol_error(gsym));
|
|
return aarch64_reloc_funcs::STATUS_BAD_RELOC;
|
|
}
|
|
return tls_ie_to_le(relinfo, target, rela, r_type, view,
|
|
psymval);
|
|
}
|
|
tls_got_offset_type = GOT_TYPE_TLS_OFFSET;
|
|
|
|
// Firstly get the address for the got entry.
|
|
typename elfcpp::Elf_types<size>::Elf_Addr got_entry_address;
|
|
if (gsym != NULL)
|
|
{
|
|
gold_assert(gsym->has_got_offset(tls_got_offset_type));
|
|
got_entry_address = target->got_->address() +
|
|
gsym->got_offset(tls_got_offset_type);
|
|
}
|
|
else
|
|
{
|
|
unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info());
|
|
gold_assert(
|
|
object->local_has_got_offset(r_sym, tls_got_offset_type));
|
|
got_entry_address = target->got_->address() +
|
|
object->local_got_offset(r_sym, tls_got_offset_type);
|
|
}
|
|
// Relocate the address into adrp/ld, adrp/add pair.
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21:
|
|
return aarch64_reloc_funcs::adrp(view, got_entry_address + addend,
|
|
address);
|
|
break;
|
|
case elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC:
|
|
return aarch64_reloc_funcs::template rela_general<32>(
|
|
view, got_entry_address, addend, reloc_property);
|
|
default:
|
|
gold_unreachable();
|
|
}
|
|
}
|
|
// We shall never reach here.
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G2:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST8_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST16_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST32_TPREL_LO12_NC:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12:
|
|
case elfcpp::R_AARCH64_TLSLE_LDST64_TPREL_LO12_NC:
|
|
{
|
|
gold_assert(tls_segment != NULL);
|
|
AArch64_address value = psymval->value(object, 0);
|
|
|
|
if (!parameters->options().shared())
|
|
{
|
|
AArch64_address aligned_tcb_size =
|
|
align_address(target->tcb_size(),
|
|
tls_segment->maximum_alignment());
|
|
value += aligned_tcb_size;
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G2:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G1:
|
|
case elfcpp::R_AARCH64_TLSLE_MOVW_TPREL_G0:
|
|
return aarch64_reloc_funcs::movnz(view, value + addend,
|
|
reloc_property);
|
|
default:
|
|
return aarch64_reloc_funcs::template
|
|
rela_general<32>(view,
|
|
value,
|
|
addend,
|
|
reloc_property);
|
|
}
|
|
}
|
|
else
|
|
gold_error(_("%s: unsupported reloc %u "
|
|
"in non-static TLSLE mode."),
|
|
object->name().c_str(), r_type);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21:
|
|
case elfcpp::R_AARCH64_TLSDESC_LD64_LO12:
|
|
case elfcpp::R_AARCH64_TLSDESC_ADD_LO12:
|
|
case elfcpp::R_AARCH64_TLSDESC_CALL:
|
|
{
|
|
if (tlsopt == tls::TLSOPT_TO_LE)
|
|
{
|
|
if (tls_segment == NULL)
|
|
{
|
|
gold_assert(parameters->errors()->error_count() > 0
|
|
|| issue_undefined_symbol_error(gsym));
|
|
return aarch64_reloc_funcs::STATUS_BAD_RELOC;
|
|
}
|
|
return tls_desc_gd_to_le(relinfo, target, rela, r_type,
|
|
view, psymval);
|
|
}
|
|
else
|
|
{
|
|
tls_got_offset_type = (tlsopt == tls::TLSOPT_TO_IE
|
|
? GOT_TYPE_TLS_OFFSET
|
|
: GOT_TYPE_TLS_DESC);
|
|
int got_tlsdesc_offset = 0;
|
|
if (r_type != elfcpp::R_AARCH64_TLSDESC_CALL
|
|
&& tlsopt == tls::TLSOPT_NONE)
|
|
{
|
|
// We created GOT entries in the .got.tlsdesc portion of the
|
|
// .got.plt section, but the offset stored in the symbol is the
|
|
// offset within .got.tlsdesc.
|
|
got_tlsdesc_offset = (target->got_tlsdesc_->address()
|
|
- target->got_->address());
|
|
}
|
|
typename elfcpp::Elf_types<size>::Elf_Addr got_entry_address;
|
|
if (gsym != NULL)
|
|
{
|
|
gold_assert(gsym->has_got_offset(tls_got_offset_type));
|
|
got_entry_address = target->got_->address()
|
|
+ got_tlsdesc_offset
|
|
+ gsym->got_offset(tls_got_offset_type);
|
|
}
|
|
else
|
|
{
|
|
unsigned int r_sym = elfcpp::elf_r_sym<size>(rela.get_r_info());
|
|
gold_assert(
|
|
object->local_has_got_offset(r_sym, tls_got_offset_type));
|
|
got_entry_address = target->got_->address() +
|
|
got_tlsdesc_offset +
|
|
object->local_got_offset(r_sym, tls_got_offset_type);
|
|
}
|
|
if (tlsopt == tls::TLSOPT_TO_IE)
|
|
{
|
|
return tls_desc_gd_to_ie(relinfo, target, rela, r_type,
|
|
view, psymval, got_entry_address,
|
|
address);
|
|
}
|
|
|
|
// Now do tlsdesc relocation.
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21:
|
|
return aarch64_reloc_funcs::adrp(view,
|
|
got_entry_address + addend,
|
|
address);
|
|
break;
|
|
case elfcpp::R_AARCH64_TLSDESC_LD64_LO12:
|
|
case elfcpp::R_AARCH64_TLSDESC_ADD_LO12:
|
|
return aarch64_reloc_funcs::template rela_general<32>(
|
|
view, got_entry_address, addend, reloc_property);
|
|
break;
|
|
case elfcpp::R_AARCH64_TLSDESC_CALL:
|
|
return aarch64_reloc_funcs::STATUS_OKAY;
|
|
break;
|
|
default:
|
|
gold_unreachable();
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
|
|
default:
|
|
gold_error(_("%s: unsupported TLS reloc %u."),
|
|
object->name().c_str(), r_type);
|
|
}
|
|
return aarch64_reloc_funcs::STATUS_BAD_RELOC;
|
|
} // End of relocate_tls.
|
|
|
|
|
|
template<int size, bool big_endian>
|
|
inline
|
|
typename AArch64_relocate_functions<size, big_endian>::Status
|
|
Target_aarch64<size, big_endian>::Relocate::tls_gd_to_le(
|
|
const Relocate_info<size, big_endian>* relinfo,
|
|
Target_aarch64<size, big_endian>* target,
|
|
const elfcpp::Rela<size, big_endian>& rela,
|
|
unsigned int r_type,
|
|
unsigned char* view,
|
|
const Symbol_value<size>* psymval)
|
|
{
|
|
typedef AArch64_relocate_functions<size, big_endian> aarch64_reloc_funcs;
|
|
typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype;
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address;
|
|
|
|
Insntype* ip = reinterpret_cast<Insntype*>(view);
|
|
Insntype insn1 = elfcpp::Swap<32, big_endian>::readval(ip);
|
|
Insntype insn2 = elfcpp::Swap<32, big_endian>::readval(ip + 1);
|
|
Insntype insn3 = elfcpp::Swap<32, big_endian>::readval(ip + 2);
|
|
|
|
if (r_type == elfcpp::R_AARCH64_TLSGD_ADD_LO12_NC)
|
|
{
|
|
// This is the 2nd relocs, optimization should already have been
|
|
// done.
|
|
gold_assert((insn1 & 0xfff00000) == 0x91400000);
|
|
return aarch64_reloc_funcs::STATUS_OKAY;
|
|
}
|
|
|
|
// The original sequence is -
|
|
// 90000000 adrp x0, 0 <main>
|
|
// 91000000 add x0, x0, #0x0
|
|
// 94000000 bl 0 <__tls_get_addr>
|
|
// optimized to sequence -
|
|
// d53bd040 mrs x0, tpidr_el0
|
|
// 91400000 add x0, x0, #0x0, lsl #12
|
|
// 91000000 add x0, x0, #0x0
|
|
|
|
// Unlike tls_ie_to_le, we change the 3 insns in one function call when we
|
|
// encounter the first relocation "R_AARCH64_TLSGD_ADR_PAGE21". Because we
|
|
// have to change "bl tls_get_addr", which does not have a corresponding tls
|
|
// relocation type. So before proceeding, we need to make sure compiler
|
|
// does not change the sequence.
|
|
if(!(insn1 == 0x90000000 // adrp x0,0
|
|
&& insn2 == 0x91000000 // add x0, x0, #0x0
|
|
&& insn3 == 0x94000000)) // bl 0
|
|
{
|
|
// Ideally we should give up gd_to_le relaxation and do gd access.
|
|
// However the gd_to_le relaxation decision has been made early
|
|
// in the scan stage, where we did not allocate any GOT entry for
|
|
// this symbol. Therefore we have to exit and report error now.
|
|
gold_error(_("unexpected reloc insn sequence while relaxing "
|
|
"tls gd to le for reloc %u."), r_type);
|
|
return aarch64_reloc_funcs::STATUS_BAD_RELOC;
|
|
}
|
|
|
|
// Write new insns.
|
|
insn1 = 0xd53bd040; // mrs x0, tpidr_el0
|
|
insn2 = 0x91400000; // add x0, x0, #0x0, lsl #12
|
|
insn3 = 0x91000000; // add x0, x0, #0x0
|
|
elfcpp::Swap<32, big_endian>::writeval(ip, insn1);
|
|
elfcpp::Swap<32, big_endian>::writeval(ip + 1, insn2);
|
|
elfcpp::Swap<32, big_endian>::writeval(ip + 2, insn3);
|
|
|
|
// Calculate tprel value.
|
|
Output_segment* tls_segment = relinfo->layout->tls_segment();
|
|
gold_assert(tls_segment != NULL);
|
|
AArch64_address value = psymval->value(relinfo->object, 0);
|
|
const elfcpp::Elf_Xword addend = rela.get_r_addend();
|
|
AArch64_address aligned_tcb_size =
|
|
align_address(target->tcb_size(), tls_segment->maximum_alignment());
|
|
AArch64_address x = value + aligned_tcb_size;
|
|
|
|
// After new insns are written, apply TLSLE relocs.
|
|
const AArch64_reloc_property* rp1 =
|
|
aarch64_reloc_property_table->get_reloc_property(
|
|
elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12);
|
|
const AArch64_reloc_property* rp2 =
|
|
aarch64_reloc_property_table->get_reloc_property(
|
|
elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12);
|
|
gold_assert(rp1 != NULL && rp2 != NULL);
|
|
|
|
typename aarch64_reloc_funcs::Status s1 =
|
|
aarch64_reloc_funcs::template rela_general<32>(view + 4,
|
|
x,
|
|
addend,
|
|
rp1);
|
|
if (s1 != aarch64_reloc_funcs::STATUS_OKAY)
|
|
return s1;
|
|
|
|
typename aarch64_reloc_funcs::Status s2 =
|
|
aarch64_reloc_funcs::template rela_general<32>(view + 8,
|
|
x,
|
|
addend,
|
|
rp2);
|
|
|
|
this->skip_call_tls_get_addr_ = true;
|
|
return s2;
|
|
} // End of tls_gd_to_le
|
|
|
|
|
|
template<int size, bool big_endian>
|
|
inline
|
|
typename AArch64_relocate_functions<size, big_endian>::Status
|
|
Target_aarch64<size, big_endian>::Relocate::tls_ld_to_le(
|
|
const Relocate_info<size, big_endian>* relinfo,
|
|
Target_aarch64<size, big_endian>* target,
|
|
const elfcpp::Rela<size, big_endian>& rela,
|
|
unsigned int r_type,
|
|
unsigned char* view,
|
|
const Symbol_value<size>* psymval)
|
|
{
|
|
typedef AArch64_relocate_functions<size, big_endian> aarch64_reloc_funcs;
|
|
typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype;
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address;
|
|
|
|
Insntype* ip = reinterpret_cast<Insntype*>(view);
|
|
Insntype insn1 = elfcpp::Swap<32, big_endian>::readval(ip);
|
|
Insntype insn2 = elfcpp::Swap<32, big_endian>::readval(ip + 1);
|
|
Insntype insn3 = elfcpp::Swap<32, big_endian>::readval(ip + 2);
|
|
|
|
if (r_type == elfcpp::R_AARCH64_TLSLD_ADD_LO12_NC)
|
|
{
|
|
// This is the 2nd relocs, optimization should already have been
|
|
// done.
|
|
gold_assert((insn1 & 0xfff00000) == 0x91400000);
|
|
return aarch64_reloc_funcs::STATUS_OKAY;
|
|
}
|
|
|
|
// The original sequence is -
|
|
// 90000000 adrp x0, 0 <main>
|
|
// 91000000 add x0, x0, #0x0
|
|
// 94000000 bl 0 <__tls_get_addr>
|
|
// optimized to sequence -
|
|
// d53bd040 mrs x0, tpidr_el0
|
|
// 91400000 add x0, x0, #0x0, lsl #12
|
|
// 91000000 add x0, x0, #0x0
|
|
|
|
// Unlike tls_ie_to_le, we change the 3 insns in one function call when we
|
|
// encounter the first relocation "R_AARCH64_TLSLD_ADR_PAGE21". Because we
|
|
// have to change "bl tls_get_addr", which does not have a corresponding tls
|
|
// relocation type. So before proceeding, we need to make sure compiler
|
|
// does not change the sequence.
|
|
if(!(insn1 == 0x90000000 // adrp x0,0
|
|
&& insn2 == 0x91000000 // add x0, x0, #0x0
|
|
&& insn3 == 0x94000000)) // bl 0
|
|
{
|
|
// Ideally we should give up gd_to_le relaxation and do gd access.
|
|
// However the gd_to_le relaxation decision has been made early
|
|
// in the scan stage, where we did not allocate a GOT entry for
|
|
// this symbol. Therefore we have to exit and report an error now.
|
|
gold_error(_("unexpected reloc insn sequence while relaxing "
|
|
"tls gd to le for reloc %u."), r_type);
|
|
return aarch64_reloc_funcs::STATUS_BAD_RELOC;
|
|
}
|
|
|
|
// Write new insns.
|
|
insn1 = 0xd53bd040; // mrs x0, tpidr_el0
|
|
insn2 = 0x91400000; // add x0, x0, #0x0, lsl #12
|
|
insn3 = 0x91000000; // add x0, x0, #0x0
|
|
elfcpp::Swap<32, big_endian>::writeval(ip, insn1);
|
|
elfcpp::Swap<32, big_endian>::writeval(ip + 1, insn2);
|
|
elfcpp::Swap<32, big_endian>::writeval(ip + 2, insn3);
|
|
|
|
// Calculate tprel value.
|
|
Output_segment* tls_segment = relinfo->layout->tls_segment();
|
|
gold_assert(tls_segment != NULL);
|
|
AArch64_address value = psymval->value(relinfo->object, 0);
|
|
const elfcpp::Elf_Xword addend = rela.get_r_addend();
|
|
AArch64_address aligned_tcb_size =
|
|
align_address(target->tcb_size(), tls_segment->maximum_alignment());
|
|
AArch64_address x = value + aligned_tcb_size;
|
|
|
|
// After new insns are written, apply TLSLE relocs.
|
|
const AArch64_reloc_property* rp1 =
|
|
aarch64_reloc_property_table->get_reloc_property(
|
|
elfcpp::R_AARCH64_TLSLE_ADD_TPREL_HI12);
|
|
const AArch64_reloc_property* rp2 =
|
|
aarch64_reloc_property_table->get_reloc_property(
|
|
elfcpp::R_AARCH64_TLSLE_ADD_TPREL_LO12);
|
|
gold_assert(rp1 != NULL && rp2 != NULL);
|
|
|
|
typename aarch64_reloc_funcs::Status s1 =
|
|
aarch64_reloc_funcs::template rela_general<32>(view + 4,
|
|
x,
|
|
addend,
|
|
rp1);
|
|
if (s1 != aarch64_reloc_funcs::STATUS_OKAY)
|
|
return s1;
|
|
|
|
typename aarch64_reloc_funcs::Status s2 =
|
|
aarch64_reloc_funcs::template rela_general<32>(view + 8,
|
|
x,
|
|
addend,
|
|
rp2);
|
|
|
|
this->skip_call_tls_get_addr_ = true;
|
|
return s2;
|
|
|
|
} // End of tls_ld_to_le
|
|
|
|
template<int size, bool big_endian>
|
|
inline
|
|
typename AArch64_relocate_functions<size, big_endian>::Status
|
|
Target_aarch64<size, big_endian>::Relocate::tls_ie_to_le(
|
|
const Relocate_info<size, big_endian>* relinfo,
|
|
Target_aarch64<size, big_endian>* target,
|
|
const elfcpp::Rela<size, big_endian>& rela,
|
|
unsigned int r_type,
|
|
unsigned char* view,
|
|
const Symbol_value<size>* psymval)
|
|
{
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address;
|
|
typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype;
|
|
typedef AArch64_relocate_functions<size, big_endian> aarch64_reloc_funcs;
|
|
|
|
AArch64_address value = psymval->value(relinfo->object, 0);
|
|
Output_segment* tls_segment = relinfo->layout->tls_segment();
|
|
AArch64_address aligned_tcb_address =
|
|
align_address(target->tcb_size(), tls_segment->maximum_alignment());
|
|
const elfcpp::Elf_Xword addend = rela.get_r_addend();
|
|
AArch64_address x = value + addend + aligned_tcb_address;
|
|
// "x" is the offset to tp, we can only do this if x is within
|
|
// range [0, 2^32-1]
|
|
if (!(size == 32 || (size == 64 && (static_cast<uint64_t>(x) >> 32) == 0)))
|
|
{
|
|
gold_error(_("TLS variable referred by reloc %u is too far from TP."),
|
|
r_type);
|
|
return aarch64_reloc_funcs::STATUS_BAD_RELOC;
|
|
}
|
|
|
|
Insntype* ip = reinterpret_cast<Insntype*>(view);
|
|
Insntype insn = elfcpp::Swap<32, big_endian>::readval(ip);
|
|
unsigned int regno;
|
|
Insntype newinsn;
|
|
if (r_type == elfcpp::R_AARCH64_TLSIE_ADR_GOTTPREL_PAGE21)
|
|
{
|
|
// Generate movz.
|
|
regno = (insn & 0x1f);
|
|
newinsn = (0xd2a00000 | regno) | (((x >> 16) & 0xffff) << 5);
|
|
}
|
|
else if (r_type == elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC)
|
|
{
|
|
// Generate movk.
|
|
regno = (insn & 0x1f);
|
|
gold_assert(regno == ((insn >> 5) & 0x1f));
|
|
newinsn = (0xf2800000 | regno) | ((x & 0xffff) << 5);
|
|
}
|
|
else
|
|
gold_unreachable();
|
|
|
|
elfcpp::Swap<32, big_endian>::writeval(ip, newinsn);
|
|
return aarch64_reloc_funcs::STATUS_OKAY;
|
|
} // End of tls_ie_to_le
|
|
|
|
|
|
template<int size, bool big_endian>
|
|
inline
|
|
typename AArch64_relocate_functions<size, big_endian>::Status
|
|
Target_aarch64<size, big_endian>::Relocate::tls_desc_gd_to_le(
|
|
const Relocate_info<size, big_endian>* relinfo,
|
|
Target_aarch64<size, big_endian>* target,
|
|
const elfcpp::Rela<size, big_endian>& rela,
|
|
unsigned int r_type,
|
|
unsigned char* view,
|
|
const Symbol_value<size>* psymval)
|
|
{
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr AArch64_address;
|
|
typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype;
|
|
typedef AArch64_relocate_functions<size, big_endian> aarch64_reloc_funcs;
|
|
|
|
// TLSDESC-GD sequence is like:
|
|
// adrp x0, :tlsdesc:v1
|
|
// ldr x1, [x0, #:tlsdesc_lo12:v1]
|
|
// add x0, x0, :tlsdesc_lo12:v1
|
|
// .tlsdesccall v1
|
|
// blr x1
|
|
// After desc_gd_to_le optimization, the sequence will be like:
|
|
// movz x0, #0x0, lsl #16
|
|
// movk x0, #0x10
|
|
// nop
|
|
// nop
|
|
|
|
// Calculate tprel value.
|
|
Output_segment* tls_segment = relinfo->layout->tls_segment();
|
|
gold_assert(tls_segment != NULL);
|
|
Insntype* ip = reinterpret_cast<Insntype*>(view);
|
|
const elfcpp::Elf_Xword addend = rela.get_r_addend();
|
|
AArch64_address value = psymval->value(relinfo->object, addend);
|
|
AArch64_address aligned_tcb_size =
|
|
align_address(target->tcb_size(), tls_segment->maximum_alignment());
|
|
AArch64_address x = value + aligned_tcb_size;
|
|
// x is the offset to tp, we can only do this if x is within range
|
|
// [0, 2^32-1]. If x is out of range, fail and exit.
|
|
if (size == 64 && (static_cast<uint64_t>(x) >> 32) != 0)
|
|
{
|
|
gold_error(_("TLS variable referred by reloc %u is too far from TP. "
|
|
"We Can't do gd_to_le relaxation.\n"), r_type);
|
|
return aarch64_reloc_funcs::STATUS_BAD_RELOC;
|
|
}
|
|
Insntype newinsn;
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_TLSDESC_ADD_LO12:
|
|
case elfcpp::R_AARCH64_TLSDESC_CALL:
|
|
// Change to nop
|
|
newinsn = 0xd503201f;
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21:
|
|
// Change to movz.
|
|
newinsn = 0xd2a00000 | (((x >> 16) & 0xffff) << 5);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSDESC_LD64_LO12:
|
|
// Change to movk.
|
|
newinsn = 0xf2800000 | ((x & 0xffff) << 5);
|
|
break;
|
|
|
|
default:
|
|
gold_error(_("unsupported tlsdesc gd_to_le optimization on reloc %u"),
|
|
r_type);
|
|
gold_unreachable();
|
|
}
|
|
elfcpp::Swap<32, big_endian>::writeval(ip, newinsn);
|
|
return aarch64_reloc_funcs::STATUS_OKAY;
|
|
} // End of tls_desc_gd_to_le
|
|
|
|
|
|
template<int size, bool big_endian>
|
|
inline
|
|
typename AArch64_relocate_functions<size, big_endian>::Status
|
|
Target_aarch64<size, big_endian>::Relocate::tls_desc_gd_to_ie(
|
|
const Relocate_info<size, big_endian>* /* relinfo */,
|
|
Target_aarch64<size, big_endian>* /* target */,
|
|
const elfcpp::Rela<size, big_endian>& rela,
|
|
unsigned int r_type,
|
|
unsigned char* view,
|
|
const Symbol_value<size>* /* psymval */,
|
|
typename elfcpp::Elf_types<size>::Elf_Addr got_entry_address,
|
|
typename elfcpp::Elf_types<size>::Elf_Addr address)
|
|
{
|
|
typedef typename elfcpp::Swap<32, big_endian>::Valtype Insntype;
|
|
typedef AArch64_relocate_functions<size, big_endian> aarch64_reloc_funcs;
|
|
|
|
// TLSDESC-GD sequence is like:
|
|
// adrp x0, :tlsdesc:v1
|
|
// ldr x1, [x0, #:tlsdesc_lo12:v1]
|
|
// add x0, x0, :tlsdesc_lo12:v1
|
|
// .tlsdesccall v1
|
|
// blr x1
|
|
// After desc_gd_to_ie optimization, the sequence will be like:
|
|
// adrp x0, :tlsie:v1
|
|
// ldr x0, [x0, :tlsie_lo12:v1]
|
|
// nop
|
|
// nop
|
|
|
|
Insntype* ip = reinterpret_cast<Insntype*>(view);
|
|
const elfcpp::Elf_Xword addend = rela.get_r_addend();
|
|
Insntype newinsn;
|
|
switch (r_type)
|
|
{
|
|
case elfcpp::R_AARCH64_TLSDESC_ADD_LO12:
|
|
case elfcpp::R_AARCH64_TLSDESC_CALL:
|
|
// Change to nop
|
|
newinsn = 0xd503201f;
|
|
elfcpp::Swap<32, big_endian>::writeval(ip, newinsn);
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSDESC_ADR_PAGE21:
|
|
{
|
|
return aarch64_reloc_funcs::adrp(view, got_entry_address + addend,
|
|
address);
|
|
}
|
|
break;
|
|
|
|
case elfcpp::R_AARCH64_TLSDESC_LD64_LO12:
|
|
{
|
|
// Set ldr target register to be x0.
|
|
Insntype insn = elfcpp::Swap<32, big_endian>::readval(ip);
|
|
insn &= 0xffffffe0;
|
|
elfcpp::Swap<32, big_endian>::writeval(ip, insn);
|
|
// Do relocation.
|
|
const AArch64_reloc_property* reloc_property =
|
|
aarch64_reloc_property_table->get_reloc_property(
|
|
elfcpp::R_AARCH64_TLSIE_LD64_GOTTPREL_LO12_NC);
|
|
return aarch64_reloc_funcs::template rela_general<32>(
|
|
view, got_entry_address, addend, reloc_property);
|
|
}
|
|
break;
|
|
|
|
default:
|
|
gold_error(_("Don't support tlsdesc gd_to_ie optimization on reloc %u"),
|
|
r_type);
|
|
gold_unreachable();
|
|
}
|
|
return aarch64_reloc_funcs::STATUS_OKAY;
|
|
} // End of tls_desc_gd_to_ie
|
|
|
|
// Relocate section data.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::relocate_section(
|
|
const Relocate_info<size, big_endian>* relinfo,
|
|
unsigned int sh_type,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
bool needs_special_offset_handling,
|
|
unsigned char* view,
|
|
typename elfcpp::Elf_types<size>::Elf_Addr address,
|
|
section_size_type view_size,
|
|
const Reloc_symbol_changes* reloc_symbol_changes)
|
|
{
|
|
typedef typename elfcpp::Elf_types<size>::Elf_Addr Address;
|
|
typedef Target_aarch64<size, big_endian> Aarch64;
|
|
typedef typename Target_aarch64<size, big_endian>::Relocate AArch64_relocate;
|
|
typedef gold::Default_classify_reloc<elfcpp::SHT_RELA, size, big_endian>
|
|
Classify_reloc;
|
|
|
|
gold_assert(sh_type == elfcpp::SHT_RELA);
|
|
|
|
// See if we are relocating a relaxed input section. If so, the view
|
|
// covers the whole output section and we need to adjust accordingly.
|
|
if (needs_special_offset_handling)
|
|
{
|
|
const Output_relaxed_input_section* poris =
|
|
output_section->find_relaxed_input_section(relinfo->object,
|
|
relinfo->data_shndx);
|
|
if (poris != NULL)
|
|
{
|
|
Address section_address = poris->address();
|
|
section_size_type section_size = poris->data_size();
|
|
|
|
gold_assert((section_address >= address)
|
|
&& ((section_address + section_size)
|
|
<= (address + view_size)));
|
|
|
|
off_t offset = section_address - address;
|
|
view += offset;
|
|
address += offset;
|
|
view_size = section_size;
|
|
}
|
|
}
|
|
|
|
gold::relocate_section<size, big_endian, Aarch64, AArch64_relocate,
|
|
gold::Default_comdat_behavior, Classify_reloc>(
|
|
relinfo,
|
|
this,
|
|
prelocs,
|
|
reloc_count,
|
|
output_section,
|
|
needs_special_offset_handling,
|
|
view,
|
|
address,
|
|
view_size,
|
|
reloc_symbol_changes);
|
|
}
|
|
|
|
// Scan the relocs during a relocatable link.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::scan_relocatable_relocs(
|
|
Symbol_table* symtab,
|
|
Layout* layout,
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int data_shndx,
|
|
unsigned int sh_type,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
bool needs_special_offset_handling,
|
|
size_t local_symbol_count,
|
|
const unsigned char* plocal_symbols,
|
|
Relocatable_relocs* rr)
|
|
{
|
|
typedef gold::Default_classify_reloc<elfcpp::SHT_RELA, size, big_endian>
|
|
Classify_reloc;
|
|
typedef gold::Default_scan_relocatable_relocs<Classify_reloc>
|
|
Scan_relocatable_relocs;
|
|
|
|
gold_assert(sh_type == elfcpp::SHT_RELA);
|
|
|
|
gold::scan_relocatable_relocs<size, big_endian, Scan_relocatable_relocs>(
|
|
symtab,
|
|
layout,
|
|
object,
|
|
data_shndx,
|
|
prelocs,
|
|
reloc_count,
|
|
output_section,
|
|
needs_special_offset_handling,
|
|
local_symbol_count,
|
|
plocal_symbols,
|
|
rr);
|
|
}
|
|
|
|
// Scan the relocs for --emit-relocs.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::emit_relocs_scan(
|
|
Symbol_table* symtab,
|
|
Layout* layout,
|
|
Sized_relobj_file<size, big_endian>* object,
|
|
unsigned int data_shndx,
|
|
unsigned int sh_type,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
bool needs_special_offset_handling,
|
|
size_t local_symbol_count,
|
|
const unsigned char* plocal_syms,
|
|
Relocatable_relocs* rr)
|
|
{
|
|
typedef gold::Default_classify_reloc<elfcpp::SHT_RELA, size, big_endian>
|
|
Classify_reloc;
|
|
typedef gold::Default_emit_relocs_strategy<Classify_reloc>
|
|
Emit_relocs_strategy;
|
|
|
|
gold_assert(sh_type == elfcpp::SHT_RELA);
|
|
|
|
gold::scan_relocatable_relocs<size, big_endian, Emit_relocs_strategy>(
|
|
symtab,
|
|
layout,
|
|
object,
|
|
data_shndx,
|
|
prelocs,
|
|
reloc_count,
|
|
output_section,
|
|
needs_special_offset_handling,
|
|
local_symbol_count,
|
|
plocal_syms,
|
|
rr);
|
|
}
|
|
|
|
// Relocate a section during a relocatable link.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::relocate_relocs(
|
|
const Relocate_info<size, big_endian>* relinfo,
|
|
unsigned int sh_type,
|
|
const unsigned char* prelocs,
|
|
size_t reloc_count,
|
|
Output_section* output_section,
|
|
typename elfcpp::Elf_types<size>::Elf_Off offset_in_output_section,
|
|
unsigned char* view,
|
|
typename elfcpp::Elf_types<size>::Elf_Addr view_address,
|
|
section_size_type view_size,
|
|
unsigned char* reloc_view,
|
|
section_size_type reloc_view_size)
|
|
{
|
|
typedef gold::Default_classify_reloc<elfcpp::SHT_RELA, size, big_endian>
|
|
Classify_reloc;
|
|
|
|
gold_assert(sh_type == elfcpp::SHT_RELA);
|
|
|
|
if (offset_in_output_section == this->invalid_address)
|
|
{
|
|
const Output_relaxed_input_section *poris
|
|
= output_section->find_relaxed_input_section(relinfo->object,
|
|
relinfo->data_shndx);
|
|
if (poris != NULL)
|
|
{
|
|
Address section_address = poris->address();
|
|
section_size_type section_size = poris->data_size();
|
|
|
|
gold_assert(section_address >= view_address
|
|
&& (section_address + section_size
|
|
<= view_address + view_size));
|
|
|
|
off_t offset = section_address - view_address;
|
|
view += offset;
|
|
view_address += offset;
|
|
view_size = section_size;
|
|
}
|
|
}
|
|
|
|
gold::relocate_relocs<size, big_endian, Classify_reloc>(
|
|
relinfo,
|
|
prelocs,
|
|
reloc_count,
|
|
output_section,
|
|
offset_in_output_section,
|
|
view,
|
|
view_address,
|
|
view_size,
|
|
reloc_view,
|
|
reloc_view_size);
|
|
}
|
|
|
|
|
|
// Return whether this is a 3-insn erratum sequence.
|
|
|
|
template<int size, bool big_endian>
|
|
bool
|
|
Target_aarch64<size, big_endian>::is_erratum_843419_sequence(
|
|
typename elfcpp::Swap<32,big_endian>::Valtype insn1,
|
|
typename elfcpp::Swap<32,big_endian>::Valtype insn2,
|
|
typename elfcpp::Swap<32,big_endian>::Valtype insn3)
|
|
{
|
|
unsigned rt1, rt2;
|
|
bool load, pair;
|
|
|
|
// The 2nd insn is a single register load or store; or register pair
|
|
// store.
|
|
if (Insn_utilities::aarch64_mem_op_p(insn2, &rt1, &rt2, &pair, &load)
|
|
&& (!pair || (pair && !load)))
|
|
{
|
|
// The 3rd insn is a load or store instruction from the "Load/store
|
|
// register (unsigned immediate)" encoding class, using Rn as the
|
|
// base address register.
|
|
if (Insn_utilities::aarch64_ldst_uimm(insn3)
|
|
&& (Insn_utilities::aarch64_rn(insn3)
|
|
== Insn_utilities::aarch64_rd(insn1)))
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
|
|
// Return whether this is a 835769 sequence.
|
|
// (Similarly implemented as in elfnn-aarch64.c.)
|
|
|
|
template<int size, bool big_endian>
|
|
bool
|
|
Target_aarch64<size, big_endian>::is_erratum_835769_sequence(
|
|
typename elfcpp::Swap<32,big_endian>::Valtype insn1,
|
|
typename elfcpp::Swap<32,big_endian>::Valtype insn2)
|
|
{
|
|
uint32_t rt;
|
|
uint32_t rt2 = 0;
|
|
uint32_t rn;
|
|
uint32_t rm;
|
|
uint32_t ra;
|
|
bool pair;
|
|
bool load;
|
|
|
|
if (Insn_utilities::aarch64_mlxl(insn2)
|
|
&& Insn_utilities::aarch64_mem_op_p (insn1, &rt, &rt2, &pair, &load))
|
|
{
|
|
/* Any SIMD memory op is independent of the subsequent MLA
|
|
by definition of the erratum. */
|
|
if (Insn_utilities::aarch64_bit(insn1, 26))
|
|
return true;
|
|
|
|
/* If not SIMD, check for integer memory ops and MLA relationship. */
|
|
rn = Insn_utilities::aarch64_rn(insn2);
|
|
ra = Insn_utilities::aarch64_ra(insn2);
|
|
rm = Insn_utilities::aarch64_rm(insn2);
|
|
|
|
/* If this is a load and there's a true(RAW) dependency, we are safe
|
|
and this is not an erratum sequence. */
|
|
if (load &&
|
|
(rt == rn || rt == rm || rt == ra
|
|
|| (pair && (rt2 == rn || rt2 == rm || rt2 == ra))))
|
|
return false;
|
|
|
|
/* We conservatively put out stubs for all other cases (including
|
|
writebacks). */
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
|
|
// Helper method to create erratum stub for ST_E_843419 and ST_E_835769.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::create_erratum_stub(
|
|
AArch64_relobj<size, big_endian>* relobj,
|
|
unsigned int shndx,
|
|
section_size_type erratum_insn_offset,
|
|
Address erratum_address,
|
|
typename Insn_utilities::Insntype erratum_insn,
|
|
int erratum_type,
|
|
unsigned int e843419_adrp_offset)
|
|
{
|
|
gold_assert(erratum_type == ST_E_843419 || erratum_type == ST_E_835769);
|
|
The_stub_table* stub_table = relobj->stub_table(shndx);
|
|
gold_assert(stub_table != NULL);
|
|
if (stub_table->find_erratum_stub(relobj,
|
|
shndx,
|
|
erratum_insn_offset) == NULL)
|
|
{
|
|
const int BPI = AArch64_insn_utilities<big_endian>::BYTES_PER_INSN;
|
|
The_erratum_stub* stub;
|
|
if (erratum_type == ST_E_835769)
|
|
stub = new The_erratum_stub(relobj, erratum_type, shndx,
|
|
erratum_insn_offset);
|
|
else if (erratum_type == ST_E_843419)
|
|
stub = new E843419_stub<size, big_endian>(
|
|
relobj, shndx, erratum_insn_offset, e843419_adrp_offset);
|
|
else
|
|
gold_unreachable();
|
|
stub->set_erratum_insn(erratum_insn);
|
|
stub->set_erratum_address(erratum_address);
|
|
// For erratum ST_E_843419 and ST_E_835769, the destination address is
|
|
// always the next insn after erratum insn.
|
|
stub->set_destination_address(erratum_address + BPI);
|
|
stub_table->add_erratum_stub(stub);
|
|
}
|
|
}
|
|
|
|
|
|
// Scan erratum for section SHNDX range [output_address + span_start,
|
|
// output_address + span_end). Note here we do not share the code with
|
|
// scan_erratum_843419_span function, because for 843419 we optimize by only
|
|
// scanning the last few insns of a page, whereas for 835769, we need to scan
|
|
// every insn.
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::scan_erratum_835769_span(
|
|
AArch64_relobj<size, big_endian>* relobj,
|
|
unsigned int shndx,
|
|
const section_size_type span_start,
|
|
const section_size_type span_end,
|
|
unsigned char* input_view,
|
|
Address output_address)
|
|
{
|
|
typedef typename Insn_utilities::Insntype Insntype;
|
|
|
|
const int BPI = AArch64_insn_utilities<big_endian>::BYTES_PER_INSN;
|
|
|
|
// Adjust output_address and view to the start of span.
|
|
output_address += span_start;
|
|
input_view += span_start;
|
|
|
|
section_size_type span_length = span_end - span_start;
|
|
section_size_type offset = 0;
|
|
for (offset = 0; offset + BPI < span_length; offset += BPI)
|
|
{
|
|
Insntype* ip = reinterpret_cast<Insntype*>(input_view + offset);
|
|
Insntype insn1 = ip[0];
|
|
Insntype insn2 = ip[1];
|
|
if (is_erratum_835769_sequence(insn1, insn2))
|
|
{
|
|
Insntype erratum_insn = insn2;
|
|
// "span_start + offset" is the offset for insn1. So for insn2, it is
|
|
// "span_start + offset + BPI".
|
|
section_size_type erratum_insn_offset = span_start + offset + BPI;
|
|
Address erratum_address = output_address + offset + BPI;
|
|
gold_info(_("Erratum 835769 found and fixed at \"%s\", "
|
|
"section %d, offset 0x%08x."),
|
|
relobj->name().c_str(), shndx,
|
|
(unsigned int)(span_start + offset));
|
|
|
|
this->create_erratum_stub(relobj, shndx,
|
|
erratum_insn_offset, erratum_address,
|
|
erratum_insn, ST_E_835769);
|
|
offset += BPI; // Skip mac insn.
|
|
}
|
|
}
|
|
} // End of "Target_aarch64::scan_erratum_835769_span".
|
|
|
|
|
|
// Scan erratum for section SHNDX range
|
|
// [output_address + span_start, output_address + span_end).
|
|
|
|
template<int size, bool big_endian>
|
|
void
|
|
Target_aarch64<size, big_endian>::scan_erratum_843419_span(
|
|
AArch64_relobj<size, big_endian>* relobj,
|
|
unsigned int shndx,
|
|
const section_size_type span_start,
|
|
const section_size_type span_end,
|
|
unsigned char* input_view,
|
|
Address output_address)
|
|
{
|
|
typedef typename Insn_utilities::Insntype Insntype;
|
|
|
|
// Adjust output_address and view to the start of span.
|
|
output_address += span_start;
|
|
input_view += span_start;
|
|
|
|
if ((output_address & 0x03) != 0)
|
|
return;
|
|
|
|
section_size_type offset = 0;
|
|
section_size_type span_length = span_end - span_start;
|
|
// The first instruction must be ending at 0xFF8 or 0xFFC.
|
|
unsigned int page_offset = output_address & 0xFFF;
|
|
// Make sure starting position, that is "output_address+offset",
|
|
// starts at page position 0xff8 or 0xffc.
|
|
if (page_offset < 0xff8)
|
|
offset = 0xff8 - page_offset;
|
|
while (offset + 3 * Insn_utilities::BYTES_PER_INSN <= span_length)
|
|
{
|
|
Insntype* ip = reinterpret_cast<Insntype*>(input_view + offset);
|
|
Insntype insn1 = ip[0];
|
|
if (Insn_utilities::is_adrp(insn1))
|
|
{
|
|
Insntype insn2 = ip[1];
|
|
Insntype insn3 = ip[2];
|
|
Insntype erratum_insn;
|
|
unsigned insn_offset;
|
|
bool do_report = false;
|
|
if (is_erratum_843419_sequence(insn1, insn2, insn3))
|
|
{
|
|
do_report = true;
|
|
erratum_insn = insn3;
|
|
insn_offset = 2 * Insn_utilities::BYTES_PER_INSN;
|
|
}
|
|
else if (offset + 4 * Insn_utilities::BYTES_PER_INSN <= span_length)
|
|
{
|
|
// Optionally we can have an insn between ins2 and ins3
|
|
Insntype insn_opt = ip[2];
|
|
// And insn_opt must not be a branch.
|
|
if (!Insn_utilities::aarch64_b(insn_opt)
|
|
&& !Insn_utilities::aarch64_bl(insn_opt)
|
|
&& !Insn_utilities::aarch64_blr(insn_opt)
|
|
&& !Insn_utilities::aarch64_br(insn_opt))
|
|
{
|
|
// And insn_opt must not write to dest reg in insn1. However
|
|
// we do a conservative scan, which means we may fix/report
|
|
// more than necessary, but it doesn't hurt.
|
|
|
|
Insntype insn4 = ip[3];
|
|
if (is_erratum_843419_sequence(insn1, insn2, insn4))
|
|
{
|
|
do_report = true;
|
|
erratum_insn = insn4;
|
|
insn_offset = 3 * Insn_utilities::BYTES_PER_INSN;
|
|
}
|
|
}
|
|
}
|
|
if (do_report)
|
|
{
|
|
unsigned int erratum_insn_offset =
|
|
span_start + offset + insn_offset;
|
|
Address erratum_address =
|
|
output_address + offset + insn_offset;
|
|
create_erratum_stub(relobj, shndx,
|
|
erratum_insn_offset, erratum_address,
|
|
erratum_insn, ST_E_843419,
|
|
span_start + offset);
|
|
}
|
|
}
|
|
|
|
// Advance to next candidate instruction. We only consider instruction
|
|
// sequences starting at a page offset of 0xff8 or 0xffc.
|
|
page_offset = (output_address + offset) & 0xfff;
|
|
if (page_offset == 0xff8)
|
|
offset += 4;
|
|
else // (page_offset == 0xffc), we move to next page's 0xff8.
|
|
offset += 0xffc;
|
|
}
|
|
} // End of "Target_aarch64::scan_erratum_843419_span".
|
|
|
|
|
|
// The selector for aarch64 object files.
|
|
|
|
template<int size, bool big_endian>
|
|
class Target_selector_aarch64 : public Target_selector
|
|
{
|
|
public:
|
|
Target_selector_aarch64();
|
|
|
|
virtual Target*
|
|
do_instantiate_target()
|
|
{ return new Target_aarch64<size, big_endian>(); }
|
|
};
|
|
|
|
template<>
|
|
Target_selector_aarch64<32, true>::Target_selector_aarch64()
|
|
: Target_selector(elfcpp::EM_AARCH64, 32, true,
|
|
"elf32-bigaarch64", "aarch64_elf32_be_vec")
|
|
{ }
|
|
|
|
template<>
|
|
Target_selector_aarch64<32, false>::Target_selector_aarch64()
|
|
: Target_selector(elfcpp::EM_AARCH64, 32, false,
|
|
"elf32-littleaarch64", "aarch64_elf32_le_vec")
|
|
{ }
|
|
|
|
template<>
|
|
Target_selector_aarch64<64, true>::Target_selector_aarch64()
|
|
: Target_selector(elfcpp::EM_AARCH64, 64, true,
|
|
"elf64-bigaarch64", "aarch64_elf64_be_vec")
|
|
{ }
|
|
|
|
template<>
|
|
Target_selector_aarch64<64, false>::Target_selector_aarch64()
|
|
: Target_selector(elfcpp::EM_AARCH64, 64, false,
|
|
"elf64-littleaarch64", "aarch64_elf64_le_vec")
|
|
{ }
|
|
|
|
Target_selector_aarch64<32, true> target_selector_aarch64elf32b;
|
|
Target_selector_aarch64<32, false> target_selector_aarch64elf32;
|
|
Target_selector_aarch64<64, true> target_selector_aarch64elfb;
|
|
Target_selector_aarch64<64, false> target_selector_aarch64elf;
|
|
|
|
} // End anonymous namespace.
|