summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--libexec/rtld-elf/mips/reloc.c287
-rw-r--r--libexec/rtld-elf/mips/rtld_start.S181
2 files changed, 315 insertions, 153 deletions
diff --git a/libexec/rtld-elf/mips/reloc.c b/libexec/rtld-elf/mips/reloc.c
index dd64644..b58e3fe 100644
--- a/libexec/rtld-elf/mips/reloc.c
+++ b/libexec/rtld-elf/mips/reloc.c
@@ -1,5 +1,4 @@
-/* $NetBSD: mdreloc.c,v 1.23 2003/07/26 15:04:38 mrg Exp $ */
-/* $NetBSD: mips_reloc.c,v 1.53 2008/07/24 04:39:25 matt Exp $ */
+/* $NetBSD: mips_reloc.c,v 1.58 2010/01/14 11:57:06 skrll Exp $ */
/*
* Copyright 1997 Michael L. Hitch <mhitch@montana.edu>
@@ -31,66 +30,102 @@
#include <sys/cdefs.h>
__FBSDID("$FreeBSD$");
-#include <sys/param.h>
-#include <sys/mman.h>
-#include <errno.h>
-#include <stdio.h>
+#include <sys/types.h>
+#include <sys/stat.h>
+#include <sys/endian.h>
+
#include <stdlib.h>
#include <string.h>
-#include <unistd.h>
+
#include "debug.h"
#include "rtld.h"
void
init_pltgot(Obj_Entry *obj)
-{
+{
if (obj->pltgot != NULL) {
obj->pltgot[0] = (Elf_Addr) &_rtld_bind_start;
+ /* XXX only if obj->pltgot[1] & 0x80000000 ?? */
obj->pltgot[1] |= (Elf_Addr) obj;
}
}
-int
+int
do_copy_relocations(Obj_Entry *dstobj)
{
/* Do nothing */
- return 0;
+ return 0;
}
-void _rtld_bind_start(void);
void _rtld_relocate_nonplt_self(Elf_Dyn *, Elf_Addr);
-int open();
-int _open();
-
/*
* It is possible for the compiler to emit relocations for unaligned data.
* We handle this situation with these inlines.
*/
-#define RELOC_ALIGNED_P(x) \
- (((uintptr_t)(x) & (sizeof(void *) - 1)) == 0)
-
-static __inline Elf_Addr
-load_ptr(void *where)
+#if ELFSIZE == 64
+/*
+ * ELF64 MIPS encodes the relocs uniquely. The first 32-bits of info contain
+ * the symbol index. The top 32-bits contain three relocation types encoded
+ * in big-endian integer with first relocation in LSB. This means for little
+ * endian we have to byte swap that interger (r_type).
+ */
+#define Elf_Sxword Elf64_Sxword
+#define ELF_R_NXTTYPE_64_P(r_type) ((((r_type) >> 8) & 0xff) == R_TYPE(64))
+#if BYTE_ORDER == LITTLE_ENDIAN
+#undef ELF_R_SYM
+#undef ELF_R_TYPE
+#define ELF_R_SYM(r_info) ((r_info) & 0xffffffff)
+#define ELF_R_TYPE(r_info) bswap32((r_info) >> 32)
+#endif
+#else
+#define ELF_R_NXTTYPE_64_P(r_type) (0)
+#define Elf_Sxword Elf32_Sword
+#endif
+
+static __inline Elf_Sxword
+load_ptr(void *where, size_t len)
{
- if (__predict_true(RELOC_ALIGNED_P(where)))
- return *(Elf_Addr *)where;
- else {
- Elf_Addr res;
-
- (void)memcpy(&res, where, sizeof(res));
- return res;
+ Elf_Sxword val;
+
+ if (__predict_true(((uintptr_t)where & (len - 1)) == 0)) {
+#if ELFSIZE == 64
+ if (len == sizeof(Elf_Sxword))
+ return *(Elf_Sxword *)where;
+#endif
+ return *(Elf_Sword *)where;
}
+
+ val = 0;
+#if BYTE_ORDER == LITTLE_ENDIAN
+ (void)memcpy(&val, where, len);
+#endif
+#if BYTE_ORDER == BIG_ENDIAN
+ (void)memcpy((uint8_t *)((&val)+1) - len, where, len);
+#endif
+ return (len == sizeof(Elf_Sxword)) ? val : (Elf_Sword)val;
}
static __inline void
-store_ptr(void *where, Elf_Addr val)
+store_ptr(void *where, Elf_Sxword val, size_t len)
{
- if (__predict_true(RELOC_ALIGNED_P(where)))
- *(Elf_Addr *)where = val;
- else
- (void)memcpy(where, &val, sizeof(val));
+ if (__predict_true(((uintptr_t)where & (len - 1)) == 0)) {
+#if ELFSIZE == 64
+ if (len == sizeof(Elf_Sxword)) {
+ *(Elf_Sxword *)where = val;
+ return;
+ }
+#endif
+ *(Elf_Sword *)where = val;
+ return;
+ }
+#if BYTE_ORDER == LITTLE_ENDIAN
+ (void)memcpy(where, &val, len);
+#endif
+#if BYTE_ORDER == BIG_ENDIAN
+ (void)memcpy(where, (const uint8_t *)((&val)+1) - len, len);
+#endif
}
void
@@ -102,7 +137,7 @@ _rtld_relocate_nonplt_self(Elf_Dyn *dynp, Elf_Addr relocbase)
Elf_Addr *where;
Elf_Addr *got = NULL;
Elf_Word local_gotno = 0, symtabno = 0, gotsym = 0;
- int i;
+ size_t i;
for (; dynp->d_tag != DT_NULL; dynp++) {
switch (dynp->d_tag) {
@@ -134,7 +169,7 @@ _rtld_relocate_nonplt_self(Elf_Dyn *dynp, Elf_Addr relocbase)
/* Relocate the local GOT entries */
got += i;
for (; i < local_gotno; i++) {
- *got++ += relocbase;
+ *got++ += relocbase;
}
sym = symtab + gotsym;
@@ -147,19 +182,41 @@ _rtld_relocate_nonplt_self(Elf_Dyn *dynp, Elf_Addr relocbase)
rellim = (const Elf_Rel *)((caddr_t)rel + relsz);
for (; rel < rellim; rel++) {
+ Elf_Word r_symndx, r_type;
+
where = (void *)(relocbase + rel->r_offset);
- switch (ELF_R_TYPE(rel->r_info)) {
- case R_TYPE(NONE):
+ r_symndx = ELF_R_SYM(rel->r_info);
+ r_type = ELF_R_TYPE(rel->r_info);
+
+ switch (r_type & 0xff) {
+ case R_TYPE(REL32): {
+ const size_t rlen =
+ ELF_R_NXTTYPE_64_P(r_type)
+ ? sizeof(Elf_Sxword)
+ : sizeof(Elf_Sword);
+ Elf_Sxword old = load_ptr(where, rlen);
+ Elf_Sxword val = old;
+#if ELFSIZE == 64
+ assert(r_type == R_TYPE(REL32)
+ || r_type == (R_TYPE(REL32)|(R_TYPE(64) << 8)));
+#endif
+ assert(r_symndx < gotsym);
+ sym = symtab + r_symndx;
+ assert(ELF_ST_BIND(sym->st_info) == STB_LOCAL);
+ val += relocbase;
+ store_ptr(where, val, sizeof(Elf_Sword));
+ dbg("REL32/L(%p) %p -> %p in <self>",
+ where, (void *)old, (void *)val);
+ store_ptr(where, val, rlen);
break;
+ }
- case R_TYPE(REL32):
- assert(ELF_R_SYM(rel->r_info) < gotsym);
- sym = symtab + ELF_R_SYM(rel->r_info);
- assert(ELF_ST_BIND(sym->st_info) == STB_LOCAL);
- store_ptr(where, load_ptr(where) + relocbase);
+ case R_TYPE(GPREL32):
+ case R_TYPE(NONE):
break;
+
default:
abort();
break;
@@ -189,9 +246,6 @@ _mips_rtld_bind(Obj_Entry *obj, Elf_Size reloff)
return (Elf_Addr)target;
}
-/*
- * Process non-PLT relocations
- */
int
reloc_non_plt(Obj_Entry *obj, Obj_Entry *obj_rtld)
{
@@ -200,12 +254,24 @@ reloc_non_plt(Obj_Entry *obj, Obj_Entry *obj_rtld)
Elf_Addr *got = obj->pltgot;
const Elf_Sym *sym, *def;
const Obj_Entry *defobj;
- int i;
+ Elf_Word i;
+#ifdef SUPPORT_OLD_BROKEN_LD
+ int broken;
+#endif
/* The relocation for the dynamic loader has already been done. */
if (obj == obj_rtld)
return (0);
+#ifdef SUPPORT_OLD_BROKEN_LD
+ broken = 0;
+ sym = obj->symtab;
+ for (i = 1; i < 12; i++)
+ if (sym[i].st_info == ELF_ST_INFO(STB_LOCAL, STT_NOTYPE))
+ broken = 1;
+ dbg("%s: broken=%d", obj->path, broken);
+#endif
+
i = (got[1] & 0x80000000) ? 2 : 1;
/* Relocate the local GOT entries */
@@ -213,16 +279,39 @@ reloc_non_plt(Obj_Entry *obj, Obj_Entry *obj_rtld)
dbg("got:%p for %d entries adding %x",
got, obj->local_gotno, (uint32_t)obj->relocbase);
for (; i < obj->local_gotno; i++) {
- *got += (Elf_Addr)obj->relocbase;
- got++;
+ *got += (Elf_Addr)obj->relocbase;
+ got++;
}
sym = obj->symtab + obj->gotsym;
-
dbg("got:%p for %d entries",
got, obj->symtabno);
/* Now do the global GOT entries */
for (i = obj->gotsym; i < obj->symtabno; i++) {
+ dbg(" doing got %d sym %p (%s, %lx)", i - obj->gotsym, sym,
+ sym->st_name + obj->strtab, (u_long) *got);
+
+#ifdef SUPPORT_OLD_BROKEN_LD
+ if (ELF_ST_TYPE(sym->st_info) == STT_FUNC &&
+ broken && sym->st_shndx == SHN_UNDEF) {
+ /*
+ * XXX DANGER WILL ROBINSON!
+ * You might think this is stupid, as it intentionally
+ * defeats lazy binding -- and you'd be right.
+ * Unfortunately, for lazy binding to work right, we
+ * need to a way to force the GOT slots used for
+ * function pointers to be resolved immediately. This
+ * is supposed to be done automatically by the linker,
+ * by not outputting a PLT slot and setting st_value
+ * to 0 if there are non-PLT references, but older
+ * versions of GNU ld do not do this.
+ */
+ def = find_symdef(i, obj, &defobj, false, NULL);
+ if (def == NULL)
+ return -1;
+ *got = def->st_value + (Elf_Addr)defobj->relocbase;
+ } else
+#endif
if (ELF_ST_TYPE(sym->st_info) == STT_FUNC &&
sym->st_value != 0 && sym->st_shndx == SHN_UNDEF) {
/*
@@ -242,81 +331,118 @@ reloc_non_plt(Obj_Entry *obj, Obj_Entry *obj_rtld)
*/
*got = sym->st_value + (Elf_Addr)obj->relocbase;
if ((Elf_Addr)(*got) == (Elf_Addr)obj->relocbase) {
- dbg("Warning2, i:%d maps to relocbase address:%x",
- i, (uint32_t)obj->relocbase);
+ dbg("Warning2, i:%d maps to relocbase address:%x",
+ i, (uint32_t)obj->relocbase);
}
} else if (sym->st_info == ELF_ST_INFO(STB_GLOBAL, STT_SECTION)) {
/* Symbols with index SHN_ABS are not relocated. */
- if (sym->st_shndx != SHN_ABS) {
+ if (sym->st_shndx != SHN_ABS) {
*got = sym->st_value +
(Elf_Addr)obj->relocbase;
if ((Elf_Addr)(*got) == (Elf_Addr)obj->relocbase) {
- dbg("Warning3, i:%d maps to relocbase address:%x",
- i, (uint32_t)obj->relocbase);
+ dbg("Warning3, i:%d maps to relocbase address:%x",
+ i, (uint32_t)obj->relocbase);
}
}
} else {
/* TODO: add cache here */
def = find_symdef(i, obj, &defobj, false, NULL);
if (def == NULL) {
- dbg("Warning4, cant find symbole %d", i);
+ dbg("Warning4, cant find symbole %d", i);
return -1;
}
*got = def->st_value + (Elf_Addr)defobj->relocbase;
if ((Elf_Addr)(*got) == (Elf_Addr)obj->relocbase) {
- dbg("Warning4, i:%d maps to relocbase address:%x",
- i, (uint32_t)obj->relocbase);
- dbg("via first obj symbol %s",
- obj->strtab + obj->symtab[i].st_name);
- dbg("found in obj %p:%s",
- defobj, defobj->path);
- }
+ dbg("Warning4, i:%d maps to relocbase address:%x",
+ i, (uint32_t)obj->relocbase);
+ dbg("via first obj symbol %s",
+ obj->strtab + obj->symtab[i].st_name);
+ dbg("found in obj %p:%s",
+ defobj, defobj->path);
+ }
}
+
+ dbg(" --> now %lx", (u_long) *got);
++sym;
++got;
}
+
got = obj->pltgot;
rellim = (const Elf_Rel *)((caddr_t)obj->rel + obj->relsize);
for (rel = obj->rel; rel < rellim; rel++) {
+ Elf_Word r_symndx, r_type;
void *where;
- Elf_Addr tmp;
- unsigned long symnum;
where = obj->relocbase + rel->r_offset;
- symnum = ELF_R_SYM(rel->r_info);
- switch (ELF_R_TYPE(rel->r_info)) {
+ r_symndx = ELF_R_SYM(rel->r_info);
+ r_type = ELF_R_TYPE(rel->r_info);
+
+ switch (r_type & 0xff) {
case R_TYPE(NONE):
break;
- case R_TYPE(REL32):
+ case R_TYPE(REL32): {
/* 32-bit PC-relative reference */
- def = obj->symtab + symnum;
- if (symnum >= obj->gotsym) {
- tmp = load_ptr(where);
- tmp += got[obj->local_gotno + symnum - obj->gotsym];
- store_ptr(where, tmp);
- break;
+ const size_t rlen =
+ ELF_R_NXTTYPE_64_P(r_type)
+ ? sizeof(Elf_Sxword)
+ : sizeof(Elf_Sword);
+ Elf_Sxword old = load_ptr(where, rlen);
+ Elf_Sxword val = old;
+
+ def = obj->symtab + r_symndx;
+
+ if (r_symndx >= obj->gotsym) {
+ val += got[obj->local_gotno + r_symndx - obj->gotsym];
+ dbg("REL32/G(%p) %p --> %p (%s) in %s",
+ where, (void *)old, (void *)val,
+ obj->strtab + def->st_name,
+ obj->path);
} else {
- tmp = load_ptr(where);
+ /*
+ * XXX: ABI DIFFERENCE!
+ *
+ * Old NetBSD binutils would generate shared
+ * libs with section-relative relocations being
+ * already adjusted for the start address of
+ * the section.
+ *
+ * New binutils, OTOH, generate shared libs
+ * with the same relocations being based at
+ * zero, so we need to add in the start address
+ * of the section.
+ *
+ * --rkb, Oct 6, 2001
+ */
if (def->st_info ==
ELF_ST_INFO(STB_LOCAL, STT_SECTION)
+#ifdef SUPPORT_OLD_BROKEN_LD
+ && !broken
+#endif
)
- tmp += (Elf_Addr)def->st_value;
+ val += (Elf_Addr)def->st_value;
+
+ val += (Elf_Addr)obj->relocbase;
- tmp += (Elf_Addr)obj->relocbase;
- store_ptr(where, tmp);
+ dbg("REL32/L(%p) %p -> %p (%s) in %s",
+ where, (void *)old, (void *)val,
+ obj->strtab + def->st_name, obj->path);
}
+ store_ptr(where, val, rlen);
break;
+ }
+
default:
dbg("sym = %lu, type = %lu, offset = %p, "
"contents = %p, symbol = %s",
- symnum, (u_long)ELF_R_TYPE(rel->r_info),
- (void *)rel->r_offset, (void *)load_ptr(where),
- obj->strtab + obj->symtab[symnum].st_name);
+ (u_long)r_symndx, (u_long)ELF_R_TYPE(rel->r_info),
+ (void *)rel->r_offset,
+ (void *)load_ptr(where, sizeof(Elf_Sword)),
+ obj->strtab + obj->symtab[r_symndx].st_name);
_rtld_error("%s: Unsupported relocation type %ld "
- "in non-PLT relocations\n",
+ "in non-PLT relocations",
obj->path, (u_long) ELF_R_TYPE(rel->r_info));
return -1;
}
@@ -331,6 +457,7 @@ reloc_non_plt(Obj_Entry *obj, Obj_Entry *obj_rtld)
int
reloc_plt(Obj_Entry *obj)
{
+#if 0
const Elf_Rel *rellim;
const Elf_Rel *rel;
@@ -345,6 +472,8 @@ reloc_plt(Obj_Entry *obj)
*where += (Elf_Addr )obj->relocbase;
}
+#endif
+ /* PLT fixups were done above in the GOT relocation. */
return (0);
}
diff --git a/libexec/rtld-elf/mips/rtld_start.S b/libexec/rtld-elf/mips/rtld_start.S
index 138729d..b4b7079 100644
--- a/libexec/rtld-elf/mips/rtld_start.S
+++ b/libexec/rtld-elf/mips/rtld_start.S
@@ -1,5 +1,4 @@
-/* $NetBSD: rtld_start.S,v 1.9 2002/10/05 11:59:05 mycroft Exp $ */
-/* $FreeBSD$ */
+/* $NetBSD: rtld_start.S,v 1.10 2009/12/14 00:41:19 matt Exp $ */
/*
* Copyright 1997 Michael L. Hitch <mhitch@montana.edu>
@@ -27,6 +26,8 @@
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+ *
+ * $FreeBSD$
*/
#include <machine/asm.h>
@@ -34,94 +35,126 @@
.globl _C_LABEL(_rtld_relocate_nonplt_self)
.globl _C_LABEL(_rtld)
+#define PTR_SIZE (1<<PTR_SCALESHIFT)
+
+/*
+ * a0 stack pointer
+ * a1 rtld cleanup (filled in by dynamic loader)
+ * a2 rtld object (filled in by dynamic loader)
+ * a3 ps_strings
+ */
LEAF(rtld_start)
- .abicalls
+ .frame sp, 4*PTR_SIZE, ra
+ .mask 0x10090000,-PTR_SIZE
.set noreorder
+ SETUP_GP
+ PTR_SUBU sp, 4*PTR_SIZE /* adjust stack pointer */
+ SETUP_GP64(s4, rtld_start)
+ SAVE_GP(0)
+ /* -> 1*PTR_SIZE(sp) for atexit */
+ /* -> 2*PTR_SIZE(sp) for obj_main */
+ move s0, a0 /* save stack pointer from a0 */
+ move s3, a3 /* save ps_strings pointer */
- .cpload t9
- addu sp, sp, -16 /* adjust stack pointer */
- /* keep it aligned */
- .cprestore 0 /* -> 0(sp) for gp */
- /* -> 4(sp) for atexit */
- /* -> 8(sp) for obj_main */
- move s0,a0 /* save stack pointer from a0 */
- move s1,a3 /* save ps_strings pointer */
-
- la a1, 1f
+ PTR_LA a1, 1f
bal 1f
- nop
-1: subu a1, ra, a1 /* relocbase */
- la t9,_C_LABEL(_rtld_relocate_nonplt_self)
- move s2,a1
- la a0,_DYNAMIC
- addu t9, a1, t9
- jalr t9
- addu a0, a1, a0 /* &_DYNAMIC */
-
+ PTR_LA t0, _C_LABEL(_rtld_relocate_nonplt_self)
+1: PTR_SUBU a1, ra, a1 /* relocbase */
+ PTR_LA a0, _DYNAMIC
+ PTR_ADDU t9, a1, t0
+ jalr t9 /* _rtld_relocate_nonplt_self(dynp, relocabase) */
+ PTR_ADDU a0, a1, a0 /* &_DYNAMIC */
- move a0, s0 /* stack pointer */
- addu a1, sp, 4 /* &exit_proc */
- addu a2, sp, 8 /* &objp */
- addu sp, sp, -16 /* arguments slot */
- jal _C_LABEL(_rtld) /* v0 = _rtld(sp, exit_proc, objp) */
- nop
- addu sp, sp, 16
+ move a0, s0 /* sp */
+ PTR_ADDU a1, sp, 2*PTR_SIZE /* &our atexit function */
+ PTR_ADDU a2, sp, 3*PTR_SIZE /* obj_main entry */
+ jal _C_LABEL(_rtld) /* v0 = _rtld(sp, cleanup, objp) */
+ nop
- move a0, s0 /* arguments pointer */
- move a3, s1 /* arguments pointer */
- lw a1, 4(sp) /* our atexit function */
- lw a2, 8(sp) /* obj_main entry */
- addu sp, sp, 16 /* readjust stack */
- move t9,v0
- move a2,s1 /* restore ps_strings */
- jr t9 /* _start(ap, cleanup, obj, ps_strings); */
- nop
+ PTR_L a1, 2*PTR_SIZE(sp) /* our atexit function */
+ PTR_L a2, 3*PTR_SIZE(sp) /* obj_main entry */
+ PTR_ADDU sp, 4*PTR_SIZE /* readjust stack */
+ move a0, s0 /* stack pointer */
+ move t9, v0
+ jr t9 /* _start(sp, cleanup, obj); */
+ move a3, s3 /* restore ps_strings */
END(rtld_start)
+#define XCALLFRAME_SIZ (12*SZREG)
+#define XCALLFRAME_RA (10*SZREG)
+#define XCALLFRAME_GP (9*SZREG)
+#define XCALLFRAME_S0 (8*SZREG)
+#define XCALLFRAME_A3 (7*SZREG)
+#define XCALLFRAME_A2 (6*SZREG)
+#define XCALLFRAME_A1 (5*SZREG)
+#define XCALLFRAME_A0 (4*SZREG)
+#if defined(__mips_n32) || defined(__mips_n64)
+#define XCALLFRAME_A7 (3*SZREG)
+#define XCALLFRAME_A6 (2*SZREG)
+#define XCALLFRAME_A5 (1*SZREG)
+#define XCALLFRAME_A4 (0*SZREG)
+#endif
+
.globl _rtld_bind_start
.ent _rtld_bind_start
_rtld_bind_start:
- /* ABI conventions for stubs:
- * t8 contains symbol index
- * t7 contains return address
- */
- .frame sp, 0, ra /* satisfy compiler */
+ .frame sp, XCALLFRAME_SIZ, $15
+ move v1, gp /* save old GP */
+#if defined(__mips_o32) || defined(__mips_o64)
+ PTR_ADDU t9, 8 /* modify T9 to point at .cpload */
+#endif
+ SETUP_GP
+ PTR_SUBU sp, XCALLFRAME_SIZ /* save arguments and sp value in stack */
+ SETUP_GP64(XCALLFRAME_GP, _rtld_bind_start)
+ SAVE_GP(XCALLFRAME_GP)
+#if defined(__mips_n32) || defined(__mips_n64)
+ REG_S a4, XCALLFRAME_A4(sp)
+ REG_S a5, XCALLFRAME_A5(sp)
+ REG_S a6, XCALLFRAME_A6(sp)
+ REG_S a7, XCALLFRAME_A7(sp)
+#endif
+ REG_S a0, XCALLFRAME_A0(sp)
+ REG_S a1, XCALLFRAME_A1(sp)
+ REG_S a2, XCALLFRAME_A2(sp)
+ REG_S a3, XCALLFRAME_A3(sp)
+ REG_S $15, XCALLFRAME_RA(sp) /* ra is in t7/t3 */
+ REG_S s0, XCALLFRAME_S0(sp)
+ move s0, sp
- move v1,gp /* save old GP */
- add t9,8 /* modify T9 to point at .cpload */
- .cpload t9
- subu sp,48 /* save arguments and sp value */
- .cprestore 36
- sw a0,16(sp)
- sw a1,20(sp)
- sw a2,24(sp)
- sw a3,28(sp)
- sw s0,32(sp)
- sw t7,40(sp)
- move s0,sp
- move a0,v1 /* old GP */
- subu a0,a0,0x7ff0 /* The offset of $gp from the */
- /* beginning of the .got section: */
+ move a0, v1 /* old GP */
+ subu a0, a0, 0x7ff0 /* The offset of $gp from the */
+ /* beginning of the .got section: */
/* $gp = .got + 0x7ff0, so */
/* .got = $gp - 0x7ff0 */
/* Simple math as you can see. */
+#if defined(__mips_n64)
+ ld a0, 8(a0) /* object = pltgot[1] & 0x7fffffff */
+#else
+ lw a0, 4(a0) /* object = pltgot[1] & 0x7fffffff */
+#endif
+ and a0, a0, 0x7fffffff
+ move a1, t8 /* symbol index */
- lw a0,4(a0) /* object = pltgot[1] & 0x7fffffff */
- and a0,a0,0x7fffffff
- move a1,t8 /* symbol index */
-
jal _C_LABEL(_mips_rtld_bind)
- nop
- move sp,s0
- lw ra,40(sp)
- lw a0,16(sp)
- lw a1,20(sp)
- lw a2,24(sp)
- lw a3,28(sp)
- lw s0,32(sp)
- addu sp,48
- move t9,v0
+ nop
+
+ move sp, s0
+ REG_L ra, XCALLFRAME_RA(sp)
+ REG_L s0, XCALLFRAME_S0(sp)
+ REG_L a0, XCALLFRAME_A0(sp)
+ REG_L a1, XCALLFRAME_A1(sp)
+ REG_L a2, XCALLFRAME_A2(sp)
+ REG_L a3, XCALLFRAME_A3(sp)
+#if defined(__mips_n32) || defined(__mips_n64)
+ REG_L a4, XCALLFRAME_A4(sp)
+ REG_L a5, XCALLFRAME_A5(sp)
+ REG_L a6, XCALLFRAME_A6(sp)
+ REG_L a7, XCALLFRAME_A7(sp)
+#endif
+ RESTORE_GP64
+ PTR_ADDU sp, XCALLFRAME_SIZ
+ move t9, v0
jr t9
- nop
- .end _rtld_bind_start
+ nop
+END(_rtld_bind_start)
OpenPOWER on IntegriCloud