aboutsummaryrefslogtreecommitdiff
path: root/sysdeps/sparc/sparc64/dl-machine.h
diff options
context:
space:
mode:
Diffstat (limited to 'sysdeps/sparc/sparc64/dl-machine.h')
-rw-r--r--sysdeps/sparc/sparc64/dl-machine.h330
1 files changed, 270 insertions, 60 deletions
diff --git a/sysdeps/sparc/sparc64/dl-machine.h b/sysdeps/sparc/sparc64/dl-machine.h
index 788e239779..33def6b455 100644
--- a/sysdeps/sparc/sparc64/dl-machine.h
+++ b/sysdeps/sparc/sparc64/dl-machine.h
@@ -21,19 +21,15 @@
#include <assert.h>
#include <string.h>
-#include <link.h>
#include <sys/param.h>
+#include <elf/ldsodefs.h>
#include <sysdep.h>
-
-/* Translate a processor-specific dynamic tag to the index into l_info. */
-#define DT_SPARC(x) (DT_SPARC_##x - DT_LOPROC + DT_NUM)
-
/* Return nonzero iff E_MACHINE is compatible with the running host. */
static inline int
elf_machine_matches_host (Elf64_Half e_machine)
{
- return e_machine == EM_SPARC64;
+ return e_machine == EM_SPARCV9;
}
/* Return the link-time address of _DYNAMIC. Conveniently, this is the
@@ -42,12 +38,11 @@ elf_machine_matches_host (Elf64_Half e_machine)
static inline Elf64_Addr
elf_machine_dynamic (void)
{
- register Elf64_Addr elf_pic_register __asm__("%l7");
+ register Elf64_Addr *elf_pic_register __asm__("%l7");
- return *(Elf64_Addr *)elf_pic_register;
+ return *elf_pic_register;
}
-
/* Return the run-time load address of the shared object. */
static inline Elf64_Addr
elf_machine_load_address (void)
@@ -67,21 +62,92 @@ elf_machine_load_address (void)
return pc - *(Elf64_Addr *)(elf_pic_register + la);
}
+/* We have 3 cases to handle. And we code different code sequences
+ for each one. I love V9 code models... */
static inline void
elf_machine_fixup_plt(struct link_map *map, const Elf64_Rela *reloc,
Elf64_Addr *reloc_addr, Elf64_Addr value)
{
- Elf64_Dyn *pltfmt = map->l_info[DT_SPARC(PLTFMT)];
- switch (pltfmt ? pltfmt->d_un.d_val : 0)
+ unsigned int *insns = (unsigned int *) reloc_addr;
+ Elf64_Addr plt_vaddr = (Elf64_Addr) reloc_addr;
+
+ /* Now move plt_vaddr up to the call instruction. */
+ plt_vaddr += (2 * 4);
+
+ /* 32-bit Sparc style, the target is in the lower 32-bits of
+ address space. */
+ if ((value >> 32) == 0)
{
- case 1: /* .got.plt with absolute addresses */
- *reloc_addr = value;
- break;
- case 2: /* .got.plt with got-relative addresses */
- *reloc_addr = value - (map->l_info[DT_PLTGOT]->d_un.d_ptr + map->l_addr);
- break;
- default:
- assert (! "unexpected .plt format type");
+ /* sethi %hi(target), %g1
+ jmpl %g1 + %lo(target), %g0 */
+
+ insns[2] = 0x81c06000 | (value & 0x3ff);
+ __asm __volatile ("flush %0 + 8" : : "r" (insns));
+
+ insns[1] = 0x03000000 | ((unsigned int)(value >> 10));
+ __asm __volatile ("flush %0 + 4" : : "r" (insns));
+ }
+ /* We can also get somewhat simple sequences if the distance between
+ the target and the PLT entry is within +/- 2GB. */
+ else if ((plt_vaddr > value
+ && ((plt_vaddr - value) >> 32) == 0)
+ || (value > plt_vaddr
+ && ((value - plt_vaddr) >> 32) == 0))
+ {
+ unsigned int displacement;
+
+ if (plt_vaddr > value)
+ displacement = (0 - (plt_vaddr - value));
+ else
+ displacement = value - plt_vaddr;
+
+ /* mov %o7, %g1
+ call displacement
+ mov %g1, %o7 */
+
+ insns[3] = 0x9e100001;
+ __asm __volatile ("flush %0 + 12" : : "r" (insns));
+
+ insns[2] = 0x40000000 | (displacement >> 2);
+ __asm __volatile ("flush %0 + 8" : : "r" (insns));
+
+ insns[1] = 0x8210000f;
+ __asm __volatile ("flush %0 + 4" : : "r" (insns));
+ }
+ /* Worst case, ho hum... */
+ else
+ {
+ unsigned int high32 = (value >> 32);
+ unsigned int low32 = (unsigned int) value;
+
+ /* ??? Some tricks can be stolen from the sparc64 egcs backend
+ constant formation code I wrote. -DaveM */
+
+ /* sethi %hh(value), %g1
+ sethi %lm(value), %g2
+ or %g1, %hl(value), %g1
+ or %g2, %lo(value), %g2
+ sllx %g1, 32, %g1
+ jmpl %g1 + %g2, %g0
+ nop */
+
+ insns[6] = 0x81c04002;
+ __asm __volatile ("flush %0 + 24" : : "r" (insns));
+
+ insns[5] = 0x83287020;
+ __asm __volatile ("flush %0 + 20" : : "r" (insns));
+
+ insns[4] = 0x8410a000 | (low32 & 0x3ff);
+ __asm __volatile ("flush %0 + 16" : : "r" (insns));
+
+ insns[3] = 0x82106000 | (high32 & 0x3ff);
+ __asm __volatile ("flush %0 + 12" : : "r" (insns));
+
+ insns[2] = 0x05000000 | (low32 >> 10);
+ __asm __volatile ("flush %0 + 8" : : "r" (insns));
+
+ insns[1] = 0x03000000 | (high32 >> 10);
+ __asm __volatile ("flush %0 + 4" : : "r" (insns));
}
}
@@ -119,7 +185,7 @@ elf_machine_rela (struct link_map *map, const Elf64_Rela *reloc,
#endif
*reloc_addr = map->l_addr + reloc->r_addend;
}
- else
+ else if (ELF64_R_TYPE (reloc->r_info) != R_SPARC_NONE) /* Who is Wilbur? */
{
const Elf64_Sym *const refsym = sym;
Elf64_Addr value;
@@ -137,6 +203,10 @@ elf_machine_rela (struct link_map *map, const Elf64_Rela *reloc,
switch (ELF64_R_TYPE (reloc->r_info))
{
case R_SPARC_COPY:
+ if (sym == NULL)
+ /* This can happen in trace mode if an object could not be
+ found. */
+ break;
if (sym->st_size > refsym->st_size
|| (_dl_verbose && sym->st_size < refsym->st_size))
{
@@ -164,6 +234,9 @@ elf_machine_rela (struct link_map *map, const Elf64_Rela *reloc,
case R_SPARC_16:
*(short *) reloc_addr = value;
break;
+ case R_SPARC_32:
+ *(unsigned int *) reloc_addr = value;
+ break;
case R_SPARC_DISP8:
*(char *) reloc_addr = (value - (Elf64_Addr) reloc_addr);
break;
@@ -171,27 +244,64 @@ elf_machine_rela (struct link_map *map, const Elf64_Rela *reloc,
*(short *) reloc_addr = (value - (Elf64_Addr) reloc_addr);
break;
case R_SPARC_DISP32:
- *(unsigned int *)reloc_addr = (value - (Elf64_Addr) reloc_addr);
- break;
- case R_SPARC_LO10:
- *(unsigned *)reloc_addr = (*(unsigned *)reloc_addr & ~0x3ff)
- | (value & 0x3ff);
+ *(unsigned int *) reloc_addr = (value - (Elf64_Addr) reloc_addr);
break;
case R_SPARC_WDISP30:
- *(unsigned *)reloc_addr = ((*(unsigned *)reloc_addr & 0xc0000000)
- | ((value - (Elf64_Addr) reloc_addr) >> 2));
+ *(unsigned int *) reloc_addr =
+ ((*(unsigned int *)reloc_addr & 0xc0000000) |
+ ((value - (Elf64_Addr) reloc_addr) >> 2));
+ break;
+
+ /* MEDLOW code model relocs */
+ case R_SPARC_LO10:
+ *(unsigned int *) reloc_addr =
+ ((*(unsigned int *)reloc_addr & ~0x3ff) |
+ (value & 0x3ff));
break;
case R_SPARC_HI22:
- *(unsigned *)reloc_addr = (*(unsigned *)reloc_addr & 0xffc00000)
- | (value >> 10);
+ *(unsigned int *) reloc_addr =
+ ((*(unsigned int *)reloc_addr & 0xffc00000) |
+ (value >> 10));
+ break;
+
+ /* MEDMID code model relocs */
+ case R_SPARC_H44:
+ *(unsigned int *) reloc_addr =
+ ((*(unsigned int *)reloc_addr & 0xffc00000) |
+ (value >> 22));
+ break;
+ case R_SPARC_M44:
+ *(unsigned int *) reloc_addr =
+ ((*(unsigned int *)reloc_addr & ~0x3ff) |
+ ((value >> 12) & 0x3ff));
+ break;
+ case R_SPARC_L44:
+ *(unsigned int *) reloc_addr =
+ ((*(unsigned int *)reloc_addr & ~0xfff) |
+ (value & 0xfff));
+ break;
+
+ /* MEDANY code model relocs */
+ case R_SPARC_HH22:
+ *(unsigned int *) reloc_addr =
+ ((*(unsigned int *)reloc_addr & 0xffc00000) |
+ (value >> 42));
+ break;
+ case R_SPARC_HM10:
+ *(unsigned int *) reloc_addr =
+ ((*(unsigned int *)reloc_addr & ~0x3ff) |
+ ((value >> 32) & 0x3ff));
+ break;
+ case R_SPARC_LM22:
+ *(unsigned int *) reloc_addr =
+ ((*(unsigned int *)reloc_addr & 0xffc00000) |
+ ((value >> 10) & 0x003fffff));
break;
case R_SPARC_JMP_SLOT:
elf_machine_fixup_plt(map, reloc, reloc_addr, value);
break;
- case R_SPARC_NONE: /* Alright, Wilbur. */
- break;
default:
assert (! "unexpected dynamic reloc type");
break;
@@ -239,16 +349,63 @@ elf_machine_lazy_rel (Elf64_Addr l_addr, const Elf64_Rela *reloc)
static inline int
elf_machine_runtime_setup (struct link_map *l, int lazy, int profile)
{
- Elf64_Addr *got;
- extern void _dl_runtime_resolve (void);
-
if (l->l_info[DT_JMPREL] && lazy)
{
- got = (Elf64_Addr *) (l->l_addr + l->l_info[DT_PLTGOT]->d_un.d_ptr);
- /* This function will get called to fix up the GOT entry indicated by
- the offset on the stack, and then jump to the resolved address. */
- got[1] = (Elf64_Addr) &_dl_runtime_resolve;
- got[2] = (Elf64_Addr) l; /* Identify this shared object. */
+ extern void _dl_runtime_resolve_0 (void);
+ extern void _dl_runtime_resolve_1 (void);
+ extern void _dl_runtime_profile_0 (void);
+ extern void _dl_runtime_profile_1 (void);
+ Elf64_Addr res0_addr, res1_addr;
+ unsigned int *plt = (unsigned int *)
+ (l->l_addr + l->l_info[DT_PLTGOT]->d_un.d_ptr);
+
+ if (! profile)
+ {
+ res0_addr = (Elf64_Addr) &_dl_runtime_resolve_0;
+ res1_addr = (Elf64_Addr) &_dl_runtime_resolve_1;
+ }
+ else
+ {
+ res0_addr = (Elf64_Addr) &_dl_runtime_profile_0;
+ res1_addr = (Elf64_Addr) &_dl_runtime_profile_1;
+ if (_dl_name_match_p (_dl_profile, l))
+ _dl_profile_map = l;
+ }
+
+ /* PLT0 looks like:
+
+ save %sp, -192, %sp
+ sethi %hh(_dl_runtime_{resolve,profile}_0), %g3
+ sethi %lm(_dl_runtime_{resolve,profile}_0), %g4
+ or %g3, %hm(_dl_runtime_{resolve,profile}_0), %g3
+ or %g4, %lo(_dl_runtime_{resolve,profile}_0), %g4
+ sllx %g3, 32, %g3
+ jmpl %g3 + %g4, %o0
+ nop
+
+ PLT1 is similar except we jump to _dl_runtime_{resolve,profile}_1. */
+
+ plt[0] = 0x9de3bf40;
+ plt[1] = 0x07000000 | (res0_addr >> (64 - 22));
+ plt[2] = 0x09000000 | ((res0_addr >> 10) & 0x003fffff);
+ plt[3] = 0x8610e000 | ((res0_addr >> 32) & 0x3ff);
+ plt[4] = 0x88112000 | (res0_addr & 0x3ff);
+ plt[5] = 0x8728f020;
+ plt[6] = 0x91c0c004;
+ plt[7] = 0x01000000;
+
+ plt[8 + 0] = 0x9de3bf40;
+ plt[8 + 1] = 0x07000000 | (res1_addr >> (64 - 22));
+ plt[8 + 2] = 0x09000000 | ((res1_addr >> 10) & 0x003fffff);
+ plt[8 + 3] = 0x8610e000 | ((res1_addr >> 32) & 0x3ff);
+ plt[8 + 4] = 0x88112000 | (res1_addr & 0x3ff);
+ plt[8 + 5] = 0x8728f020;
+ plt[8 + 6] = 0x91c0c004;
+ plt[8 + 7] = 0x01000000;
+
+ /* Now put the magic cookie at the beginning of .PLT3
+ Entry .PLT4 is unused by this implementation. */
+ *((struct link_map **)(&plt[16 + 0])) = l;
}
return lazy;
@@ -256,23 +413,68 @@ elf_machine_runtime_setup (struct link_map *l, int lazy, int profile)
/* This code is used in dl-runtime.c to call the `fixup' function
and then redirect to the address it returns. */
-#define ELF_MACHINE_RUNTIME_TRAMPOLINE asm ("\
- .globl _dl_runtime_resolve
- .type _dl_runtime_resolve, @function
-_dl_runtime_resolve:
- save %sp, -160, %sp
- mov %g1, %o0
- call fixup
- mov %g2, %o1
- jmp %o0
+#define TRAMPOLINE_TEMPLATE(tramp_name, fixup_name) \
+ asm ("\
+ .text
+ .globl " #tramp_name "_0
+ .type " #tramp_name "_0, @function
+ .align 32
+" #tramp_name "_0:
+ ldx [%o0 + 32 + 8], %l0
+ sethi %hi(1048576), %g2
+ sub %g1, %o0, %o0
+ xor %g2, -20, %g2
+ sethi %hi(5120), %g3
+ add %o0, %g2, %o0
+ sethi %hi(32768), %o2
+ udivx %o0, %g3, %g3
+ sllx %g3, 2, %g1
+ add %g1, %g3, %g1
+ sllx %g1, 10, %g2
+ sllx %g1, 5, %g1
+ sub %o0, %g2, %o0
+ udivx %o0, 24, %o0
+ add %o0, %o2, %o0
+ add %g1, %o0, %g1
+ sllx %g1, 1, %o1
+ mov %l0, %o0
+ add %o1, %g1, %o1
+ mov %i7, %o2
+ call " #fixup_name "
+ sllx %o1, 3, %o1
+ jmp %o0
+ restore
+ .size " #tramp_name "_0, . - " #tramp_name "_0
+
+ .globl " #tramp_name "_1
+ .type " #tramp_name "_1, @function
+ .align 32
+" #tramp_name "_1:
+ srlx %g1, 15, %o1
+ ldx [%o0 + 8], %o0
+ sllx %o1, 1, %o3
+ add %o1, %o3, %o1
+ mov %i7, %o2
+ call " #fixup_name "
+ sllx %o1, 3, %o1
+ jmp %o0
restore
- .size _dl_runtime_resolve, .-_dl_runtime_resolve
-");
+ .size " #tramp_name "_1, . - " #tramp_name "_1
+ .previous");
+
+#ifndef PROF
+#define ELF_MACHINE_RUNTIME_TRAMPOLINE \
+ TRAMPOLINE_TEMPLATE (_dl_runtime_resolve, fixup); \
+ TRAMPOLINE_TEMPLATE (_dl_runtime_profile, profile_fixup);
+#else
+#define ELF_MACHINE_RUNTIME_TRAMPOLINE \
+ TRAMPOLINE_TEMPLATE (_dl_runtime_resolve, fixup); \
+ TRAMPOLINE_TEMPLATE (_dl_runtime_profile, fixup);
+#endif
/* The PLT uses Elf64_Rela relocs. */
#define elf_machine_relplt elf_machine_rela
-
/* Initial entry point code for the dynamic linker.
The C function `_dl_start' is the real entry point;
its return value is the user program's entry point. */
@@ -281,8 +483,10 @@ _dl_runtime_resolve:
#define __S(x) __S1(x)
#define RTLD_START __asm__ ( "\
- .global _start
- .type _start, @function
+ .text
+ .global _start
+ .type _start, @function
+ .align 32
_start:
/* Make room for functions to drop their arguments on the stack. */
sub %sp, 6*8, %sp
@@ -292,8 +496,8 @@ _start:
/* FALLTHRU */
.size _start, .-_start
- .global _dl_start_user
- .type _dl_start_user, @function
+ .global _dl_start_user
+ .type _dl_start_user, @function
_dl_start_user:
/* Load the GOT register. */
1: call 11f
@@ -302,6 +506,12 @@ _dl_start_user:
add %l7,%o7,%l7
/* Save the user entry point address in %l0. */
mov %o0,%l0
+ /* Store the highest stack address. */
+ sethi %hi(__libc_stack_end), %g2
+ or %g2, %lo(__libc_stack_end), %g2
+ ldx [%l7 + %g2], %l1
+ add %sp, 6*8, %l2
+ stx %l2, [%l1]
/* See if we were run as a command with the executable file name as an
extra leading argument. If so, we must shift things around since we
must keep the stack doubleword aligned. */
@@ -338,7 +548,7 @@ _dl_start_user:
stx %i4, [%i1+8]
brnz,pt %i3, 13b
add %i1, 16, %i1
- /* Load _dl_main_searchlist to pass to _dl_init_next. */
+ /* Load searchlist of the main object to pass to _dl_init_next. */
2: sethi %hi(_dl_main_searchlist), %g2
or %g2, %lo(_dl_main_searchlist), %g2
ldx [%l7+%g2], %g2
@@ -349,8 +559,7 @@ _dl_start_user:
brz,pn %o0, 4f
nop
jmpl %o0, %o7
- nop
- ba,a 3b
+ sub %o7, 24, %o7
/* Clear the startup flag. */
4: sethi %hi(_dl_starting_up), %g2
or %g2, %lo(_dl_starting_up), %g2
@@ -360,7 +569,8 @@ _dl_start_user:
sethi %hi(_dl_fini), %g1
or %g1, %lo(_dl_fini), %g1
ldx [%l7+%g1], %g1
- /* Jump to the user's entry point & undo the allocation of the xtra regs. */
+ /* Jump to the user's entry point and deallocate the extra stack we got. */
jmp %l0
add %sp, 6*8, %sp
- .size _dl_start_user, .-_dl_start_user");
+ .size _dl_start_user, . - _dl_start_user
+ .previous");