| /* ----------------------------------------------------------------------- |
| ffi_linux64.c - Copyright (C) 2013 IBM |
| Copyright (C) 2011 Anthony Green |
| Copyright (C) 2011 Kyle Moffett |
| Copyright (C) 2008 Red Hat, Inc |
| Copyright (C) 2007, 2008 Free Software Foundation, Inc |
| Copyright (c) 1998 Geoffrey Keating |
| |
| PowerPC Foreign Function Interface |
| |
| Permission is hereby granted, free of charge, to any person obtaining |
| a copy of this software and associated documentation files (the |
| ``Software''), to deal in the Software without restriction, including |
| without limitation the rights to use, copy, modify, merge, publish, |
| distribute, sublicense, and/or sell copies of the Software, and to |
| permit persons to whom the Software is furnished to do so, subject to |
| the following conditions: |
| |
| The above copyright notice and this permission notice shall be included |
| in all copies or substantial portions of the Software. |
| |
| THE SOFTWARE IS PROVIDED ``AS IS'', WITHOUT WARRANTY OF ANY KIND, EXPRESS |
| OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
| MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. |
| IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY CLAIM, DAMAGES OR |
| OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, |
| ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR |
| OTHER DEALINGS IN THE SOFTWARE. |
| ----------------------------------------------------------------------- */ |
| |
| #include "ffi.h" |
| |
| #ifdef POWERPC64 |
| #include "ffi_common.h" |
| #include "ffi_powerpc.h" |
| |
| |
| /* About the LINUX64 ABI. */ |
| enum { |
| NUM_GPR_ARG_REGISTERS64 = 8, |
| NUM_FPR_ARG_REGISTERS64 = 13, |
| NUM_VEC_ARG_REGISTERS64 = 12, |
| }; |
| enum { ASM_NEEDS_REGISTERS64 = 4 }; |
| |
| |
| #if HAVE_LONG_DOUBLE_VARIANT && FFI_TYPE_LONGDOUBLE != FFI_TYPE_DOUBLE |
| /* Adjust size of ffi_type_longdouble. */ |
| void FFI_HIDDEN |
| ffi_prep_types_linux64 (ffi_abi abi) |
| { |
| if ((abi & (FFI_LINUX | FFI_LINUX_LONG_DOUBLE_128)) == FFI_LINUX) |
| { |
| ffi_type_longdouble.size = 8; |
| ffi_type_longdouble.alignment = 8; |
| } |
| else |
| { |
| ffi_type_longdouble.size = 16; |
| ffi_type_longdouble.alignment = 16; |
| } |
| } |
| #endif |
| |
| |
| static unsigned int |
| discover_homogeneous_aggregate (ffi_abi abi, |
| const ffi_type *t, |
| unsigned int *elnum) |
| { |
| switch (t->type) |
| { |
| #if FFI_TYPE_LONGDOUBLE != FFI_TYPE_DOUBLE |
| case FFI_TYPE_LONGDOUBLE: |
| /* 64-bit long doubles are equivalent to doubles. */ |
| if ((abi & FFI_LINUX_LONG_DOUBLE_128) == 0) |
| { |
| *elnum = 1; |
| return FFI_TYPE_DOUBLE; |
| } |
| /* IBM extended precision values use unaligned pairs |
| of FPRs, but according to the ABI must be considered |
| distinct from doubles. They are also limited to a |
| maximum of four members in a homogeneous aggregate. */ |
| else if ((abi & FFI_LINUX_LONG_DOUBLE_IEEE128) == 0) |
| { |
| *elnum = 2; |
| return FFI_TYPE_LONGDOUBLE; |
| } |
| /* Fall through. */ |
| #endif |
| case FFI_TYPE_FLOAT: |
| case FFI_TYPE_DOUBLE: |
| *elnum = 1; |
| return (int) t->type; |
| |
| case FFI_TYPE_STRUCT:; |
| { |
| unsigned int base_elt = 0, total_elnum = 0; |
| ffi_type **el = t->elements; |
| while (*el) |
| { |
| unsigned int el_elt, el_elnum = 0; |
| el_elt = discover_homogeneous_aggregate (abi, *el, &el_elnum); |
| if (el_elt == 0 |
| || (base_elt && base_elt != el_elt)) |
| return 0; |
| base_elt = el_elt; |
| total_elnum += el_elnum; |
| #if _CALL_ELF == 2 |
| if (total_elnum > 8) |
| return 0; |
| #else |
| if (total_elnum > 1) |
| return 0; |
| #endif |
| el++; |
| } |
| *elnum = total_elnum; |
| return base_elt; |
| } |
| |
| default: |
| return 0; |
| } |
| } |
| |
| |
| /* Perform machine dependent cif processing */ |
| static ffi_status |
| ffi_prep_cif_linux64_core (ffi_cif *cif) |
| { |
| ffi_type **ptr; |
| unsigned bytes; |
| unsigned i, fparg_count = 0, intarg_count = 0, vecarg_count = 0; |
| unsigned flags = cif->flags; |
| unsigned elt, elnum, rtype; |
| |
| #if FFI_TYPE_LONGDOUBLE == FFI_TYPE_DOUBLE |
| /* If compiled without long double support... */ |
| if ((cif->abi & FFI_LINUX_LONG_DOUBLE_128) != 0 || |
| (cif->abi & FFI_LINUX_LONG_DOUBLE_IEEE128) != 0) |
| return FFI_BAD_ABI; |
| #elif !defined(__VEC__) |
| /* If compiled without vector register support (used by assembly)... */ |
| if ((cif->abi & FFI_LINUX_LONG_DOUBLE_IEEE128) != 0) |
| return FFI_BAD_ABI; |
| #else |
| /* If the IEEE128 flag is set, but long double is only 64 bits wide... */ |
| if ((cif->abi & FFI_LINUX_LONG_DOUBLE_128) == 0 && |
| (cif->abi & FFI_LINUX_LONG_DOUBLE_IEEE128) != 0) |
| return FFI_BAD_ABI; |
| #endif |
| |
| /* The machine-independent calculation of cif->bytes doesn't work |
| for us. Redo the calculation. */ |
| #if _CALL_ELF == 2 |
| /* Space for backchain, CR, LR, TOC and the asm's temp regs. */ |
| bytes = (4 + ASM_NEEDS_REGISTERS64) * sizeof (long); |
| |
| /* Space for the general registers. */ |
| bytes += NUM_GPR_ARG_REGISTERS64 * sizeof (long); |
| #else |
| /* Space for backchain, CR, LR, cc/ld doubleword, TOC and the asm's temp |
| regs. */ |
| bytes = (6 + ASM_NEEDS_REGISTERS64) * sizeof (long); |
| |
| /* Space for the mandatory parm save area and general registers. */ |
| bytes += 2 * NUM_GPR_ARG_REGISTERS64 * sizeof (long); |
| #endif |
| |
| /* Return value handling. */ |
| rtype = cif->rtype->type; |
| #if _CALL_ELF == 2 |
| homogeneous: |
| #endif |
| switch (rtype) |
| { |
| #if FFI_TYPE_LONGDOUBLE != FFI_TYPE_DOUBLE |
| case FFI_TYPE_LONGDOUBLE: |
| if ((cif->abi & FFI_LINUX_LONG_DOUBLE_IEEE128) != 0) |
| { |
| flags |= FLAG_RETURNS_VEC; |
| break; |
| } |
| if ((cif->abi & FFI_LINUX_LONG_DOUBLE_128) != 0) |
| flags |= FLAG_RETURNS_128BITS; |
| /* Fall through. */ |
| #endif |
| case FFI_TYPE_DOUBLE: |
| flags |= FLAG_RETURNS_64BITS; |
| /* Fall through. */ |
| case FFI_TYPE_FLOAT: |
| flags |= FLAG_RETURNS_FP; |
| break; |
| |
| case FFI_TYPE_UINT128: |
| flags |= FLAG_RETURNS_128BITS; |
| /* Fall through. */ |
| case FFI_TYPE_UINT64: |
| case FFI_TYPE_SINT64: |
| case FFI_TYPE_POINTER: |
| flags |= FLAG_RETURNS_64BITS; |
| break; |
| |
| case FFI_TYPE_STRUCT: |
| #if _CALL_ELF == 2 |
| elt = discover_homogeneous_aggregate (cif->abi, cif->rtype, &elnum); |
| if (elt) |
| { |
| flags |= FLAG_RETURNS_SMST; |
| rtype = elt; |
| goto homogeneous; |
| } |
| if (cif->rtype->size <= 16) |
| { |
| flags |= FLAG_RETURNS_SMST; |
| break; |
| } |
| #endif |
| intarg_count++; |
| flags |= FLAG_RETVAL_REFERENCE; |
| /* Fall through. */ |
| case FFI_TYPE_VOID: |
| flags |= FLAG_RETURNS_NOTHING; |
| break; |
| |
| default: |
| /* Returns 32-bit integer, or similar. Nothing to do here. */ |
| break; |
| } |
| |
| for (ptr = cif->arg_types, i = cif->nargs; i > 0; i--, ptr++) |
| { |
| unsigned int align; |
| |
| switch ((*ptr)->type) |
| { |
| #if FFI_TYPE_LONGDOUBLE != FFI_TYPE_DOUBLE |
| case FFI_TYPE_LONGDOUBLE: |
| if ((cif->abi & FFI_LINUX_LONG_DOUBLE_IEEE128) != 0) |
| { |
| vecarg_count++; |
| /* Align to 16 bytes, plus the 16-byte argument. */ |
| intarg_count = (intarg_count + 3) & ~0x1; |
| if (vecarg_count > NUM_VEC_ARG_REGISTERS64) |
| flags |= FLAG_ARG_NEEDS_PSAVE; |
| break; |
| } |
| if ((cif->abi & FFI_LINUX_LONG_DOUBLE_128) != 0) |
| { |
| fparg_count++; |
| intarg_count++; |
| } |
| /* Fall through. */ |
| #endif |
| case FFI_TYPE_DOUBLE: |
| case FFI_TYPE_FLOAT: |
| fparg_count++; |
| intarg_count++; |
| if (fparg_count > NUM_FPR_ARG_REGISTERS64) |
| flags |= FLAG_ARG_NEEDS_PSAVE; |
| break; |
| |
| case FFI_TYPE_STRUCT: |
| if ((cif->abi & FFI_LINUX_STRUCT_ALIGN) != 0) |
| { |
| align = (*ptr)->alignment; |
| if (align > 16) |
| align = 16; |
| align = align / 8; |
| if (align > 1) |
| intarg_count = FFI_ALIGN (intarg_count, align); |
| } |
| intarg_count += ((*ptr)->size + 7) / 8; |
| elt = discover_homogeneous_aggregate (cif->abi, *ptr, &elnum); |
| #if FFI_TYPE_LONGDOUBLE != FFI_TYPE_DOUBLE |
| if (elt == FFI_TYPE_LONGDOUBLE && |
| (cif->abi & FFI_LINUX_LONG_DOUBLE_IEEE128) != 0) |
| { |
| vecarg_count += elnum; |
| if (vecarg_count > NUM_VEC_ARG_REGISTERS64) |
| flags |= FLAG_ARG_NEEDS_PSAVE; |
| break; |
| } |
| else |
| #endif |
| if (elt) |
| { |
| fparg_count += elnum; |
| if (fparg_count > NUM_FPR_ARG_REGISTERS64) |
| flags |= FLAG_ARG_NEEDS_PSAVE; |
| } |
| else |
| { |
| if (intarg_count > NUM_GPR_ARG_REGISTERS64) |
| flags |= FLAG_ARG_NEEDS_PSAVE; |
| } |
| break; |
| |
| case FFI_TYPE_POINTER: |
| case FFI_TYPE_UINT64: |
| case FFI_TYPE_SINT64: |
| case FFI_TYPE_INT: |
| case FFI_TYPE_UINT32: |
| case FFI_TYPE_SINT32: |
| case FFI_TYPE_UINT16: |
| case FFI_TYPE_SINT16: |
| case FFI_TYPE_UINT8: |
| case FFI_TYPE_SINT8: |
| /* Everything else is passed as a 8-byte word in a GPR, either |
| the object itself or a pointer to it. */ |
| intarg_count++; |
| if (intarg_count > NUM_GPR_ARG_REGISTERS64) |
| flags |= FLAG_ARG_NEEDS_PSAVE; |
| break; |
| default: |
| FFI_ASSERT (0); |
| } |
| } |
| |
| if (fparg_count != 0) |
| flags |= FLAG_FP_ARGUMENTS; |
| if (intarg_count > 4) |
| flags |= FLAG_4_GPR_ARGUMENTS; |
| if (vecarg_count != 0) |
| flags |= FLAG_VEC_ARGUMENTS; |
| |
| /* Space for the FPR registers, if needed. */ |
| if (fparg_count != 0) |
| bytes += NUM_FPR_ARG_REGISTERS64 * sizeof (double); |
| /* Space for the vector registers, if needed, aligned to 16 bytes. */ |
| if (vecarg_count != 0) { |
| bytes = (bytes + 15) & ~0xF; |
| bytes += NUM_VEC_ARG_REGISTERS64 * sizeof (float128); |
| } |
| |
| /* Stack space. */ |
| #if _CALL_ELF == 2 |
| if ((flags & FLAG_ARG_NEEDS_PSAVE) != 0) |
| bytes += intarg_count * sizeof (long); |
| #else |
| if (intarg_count > NUM_GPR_ARG_REGISTERS64) |
| bytes += (intarg_count - NUM_GPR_ARG_REGISTERS64) * sizeof (long); |
| #endif |
| |
| /* The stack space allocated needs to be a multiple of 16 bytes. */ |
| bytes = (bytes + 15) & ~0xF; |
| |
| cif->flags = flags; |
| cif->bytes = bytes; |
| |
| return FFI_OK; |
| } |
| |
| ffi_status FFI_HIDDEN |
| ffi_prep_cif_linux64 (ffi_cif *cif) |
| { |
| if ((cif->abi & FFI_LINUX) != 0) |
| cif->nfixedargs = cif->nargs; |
| #if _CALL_ELF != 2 |
| else if (cif->abi == FFI_COMPAT_LINUX64) |
| { |
| /* This call is from old code. Don't touch cif->nfixedargs |
| since old code will be using a smaller cif. */ |
| cif->flags |= FLAG_COMPAT; |
| /* Translate to new abi value. */ |
| cif->abi = FFI_LINUX | FFI_LINUX_LONG_DOUBLE_128; |
| } |
| #endif |
| else |
| return FFI_BAD_ABI; |
| return ffi_prep_cif_linux64_core (cif); |
| } |
| |
| ffi_status FFI_HIDDEN |
| ffi_prep_cif_linux64_var (ffi_cif *cif, |
| unsigned int nfixedargs, |
| unsigned int ntotalargs MAYBE_UNUSED) |
| { |
| if ((cif->abi & FFI_LINUX) != 0) |
| cif->nfixedargs = nfixedargs; |
| #if _CALL_ELF != 2 |
| else if (cif->abi == FFI_COMPAT_LINUX64) |
| { |
| /* This call is from old code. Don't touch cif->nfixedargs |
| since old code will be using a smaller cif. */ |
| cif->flags |= FLAG_COMPAT; |
| /* Translate to new abi value. */ |
| cif->abi = FFI_LINUX | FFI_LINUX_LONG_DOUBLE_128; |
| } |
| #endif |
| else |
| return FFI_BAD_ABI; |
| #if _CALL_ELF == 2 |
| cif->flags |= FLAG_ARG_NEEDS_PSAVE; |
| #endif |
| return ffi_prep_cif_linux64_core (cif); |
| } |
| |
| |
| /* ffi_prep_args64 is called by the assembly routine once stack space |
| has been allocated for the function's arguments. |
| |
| The stack layout we want looks like this: |
| |
| | Ret addr from ffi_call_LINUX64 8bytes | higher addresses |
| |--------------------------------------------| |
| | CR save area 8bytes | |
| |--------------------------------------------| |
| | Previous backchain pointer 8 | stack pointer here |
| |--------------------------------------------|<+ <<< on entry to |
| | Saved r28-r31 4*8 | | ffi_call_LINUX64 |
| |--------------------------------------------| | |
| | GPR registers r3-r10 8*8 | | |
| |--------------------------------------------| | |
| | FPR registers f1-f13 (optional) 13*8 | | |
| |--------------------------------------------| | |
| | VEC registers v2-v13 (optional) 12*16 | | |
| |--------------------------------------------| | |
| | Parameter save area | | |
| |--------------------------------------------| | |
| | TOC save area 8 | | |
| |--------------------------------------------| | stack | |
| | Linker doubleword 8 | | grows | |
| |--------------------------------------------| | down V |
| | Compiler doubleword 8 | | |
| |--------------------------------------------| | lower addresses |
| | Space for callee's LR 8 | | |
| |--------------------------------------------| | |
| | CR save area 8 | | |
| |--------------------------------------------| | stack pointer here |
| | Current backchain pointer 8 |-/ during |
| |--------------------------------------------| <<< ffi_call_LINUX64 |
| |
| */ |
| |
| void FFI_HIDDEN |
| ffi_prep_args64 (extended_cif *ecif, unsigned long *const stack) |
| { |
| const unsigned long bytes = ecif->cif->bytes; |
| const unsigned long flags = ecif->cif->flags; |
| |
| typedef union |
| { |
| char *c; |
| unsigned long *ul; |
| float *f; |
| double *d; |
| float128 *f128; |
| size_t p; |
| } valp; |
| |
| /* 'stacktop' points at the previous backchain pointer. */ |
| valp stacktop; |
| |
| /* 'next_arg' points at the space for gpr3, and grows upwards as |
| we use GPR registers, then continues at rest. */ |
| valp gpr_base; |
| valp gpr_end; |
| valp rest; |
| valp next_arg; |
| |
| /* 'fpr_base' points at the space for f1, and grows upwards as |
| we use FPR registers. */ |
| valp fpr_base; |
| unsigned int fparg_count; |
| |
| /* 'vec_base' points at the space for v2, and grows upwards as |
| we use vector registers. */ |
| valp vec_base; |
| unsigned int vecarg_count; |
| |
| unsigned int i, words, nargs, nfixedargs; |
| ffi_type **ptr; |
| double double_tmp; |
| union |
| { |
| void **v; |
| char **c; |
| signed char **sc; |
| unsigned char **uc; |
| signed short **ss; |
| unsigned short **us; |
| signed int **si; |
| unsigned int **ui; |
| unsigned long **ul; |
| float **f; |
| double **d; |
| float128 **f128; |
| } p_argv; |
| unsigned long gprvalue; |
| unsigned long align; |
| |
| stacktop.c = (char *) stack + bytes; |
| gpr_base.ul = stacktop.ul - ASM_NEEDS_REGISTERS64 - NUM_GPR_ARG_REGISTERS64; |
| gpr_end.ul = gpr_base.ul + NUM_GPR_ARG_REGISTERS64; |
| #if _CALL_ELF == 2 |
| rest.ul = stack + 4 + NUM_GPR_ARG_REGISTERS64; |
| #else |
| rest.ul = stack + 6 + NUM_GPR_ARG_REGISTERS64; |
| #endif |
| fpr_base.d = gpr_base.d - NUM_FPR_ARG_REGISTERS64; |
| fparg_count = 0; |
| /* Place the vector args below the FPRs, if used, else the GPRs. */ |
| if (ecif->cif->flags & FLAG_FP_ARGUMENTS) |
| vec_base.p = fpr_base.p & ~0xF; |
| else |
| vec_base.p = gpr_base.p; |
| vec_base.f128 -= NUM_VEC_ARG_REGISTERS64; |
| vecarg_count = 0; |
| next_arg.ul = gpr_base.ul; |
| |
| /* Check that everything starts aligned properly. */ |
| FFI_ASSERT (((unsigned long) (char *) stack & 0xF) == 0); |
| FFI_ASSERT (((unsigned long) stacktop.c & 0xF) == 0); |
| FFI_ASSERT (((unsigned long) gpr_base.c & 0xF) == 0); |
| FFI_ASSERT (((unsigned long) gpr_end.c & 0xF) == 0); |
| FFI_ASSERT (((unsigned long) vec_base.c & 0xF) == 0); |
| FFI_ASSERT ((bytes & 0xF) == 0); |
| |
| /* Deal with return values that are actually pass-by-reference. */ |
| if (flags & FLAG_RETVAL_REFERENCE) |
| *next_arg.ul++ = (unsigned long) (char *) ecif->rvalue; |
| |
| /* Now for the arguments. */ |
| p_argv.v = ecif->avalue; |
| nargs = ecif->cif->nargs; |
| #if _CALL_ELF != 2 |
| nfixedargs = (unsigned) -1; |
| if ((flags & FLAG_COMPAT) == 0) |
| #endif |
| nfixedargs = ecif->cif->nfixedargs; |
| for (ptr = ecif->cif->arg_types, i = 0; |
| i < nargs; |
| i++, ptr++, p_argv.v++) |
| { |
| unsigned int elt, elnum; |
| |
| switch ((*ptr)->type) |
| { |
| #if FFI_TYPE_LONGDOUBLE != FFI_TYPE_DOUBLE |
| case FFI_TYPE_LONGDOUBLE: |
| if ((ecif->cif->abi & FFI_LINUX_LONG_DOUBLE_IEEE128) != 0) |
| { |
| next_arg.p = FFI_ALIGN (next_arg.p, 16); |
| if (next_arg.ul == gpr_end.ul) |
| next_arg.ul = rest.ul; |
| if (vecarg_count < NUM_VEC_ARG_REGISTERS64 && i < nfixedargs) |
| memcpy (vec_base.f128++, *p_argv.f128, sizeof (float128)); |
| else |
| memcpy (next_arg.f128, *p_argv.f128, sizeof (float128)); |
| if (++next_arg.f128 == gpr_end.f128) |
| next_arg.f128 = rest.f128; |
| vecarg_count++; |
| FFI_ASSERT (__LDBL_MANT_DIG__ == 113); |
| FFI_ASSERT (flags & FLAG_VEC_ARGUMENTS); |
| break; |
| } |
| if ((ecif->cif->abi & FFI_LINUX_LONG_DOUBLE_128) != 0) |
| { |
| double_tmp = (*p_argv.d)[0]; |
| if (fparg_count < NUM_FPR_ARG_REGISTERS64 && i < nfixedargs) |
| { |
| *fpr_base.d++ = double_tmp; |
| # if _CALL_ELF != 2 |
| if ((flags & FLAG_COMPAT) != 0) |
| *next_arg.d = double_tmp; |
| # endif |
| } |
| else |
| *next_arg.d = double_tmp; |
| if (++next_arg.ul == gpr_end.ul) |
| next_arg.ul = rest.ul; |
| fparg_count++; |
| double_tmp = (*p_argv.d)[1]; |
| if (fparg_count < NUM_FPR_ARG_REGISTERS64 && i < nfixedargs) |
| { |
| *fpr_base.d++ = double_tmp; |
| # if _CALL_ELF != 2 |
| if ((flags & FLAG_COMPAT) != 0) |
| *next_arg.d = double_tmp; |
| # endif |
| } |
| else |
| *next_arg.d = double_tmp; |
| if (++next_arg.ul == gpr_end.ul) |
| next_arg.ul = rest.ul; |
| fparg_count++; |
| FFI_ASSERT (__LDBL_MANT_DIG__ == 106); |
| FFI_ASSERT (flags & FLAG_FP_ARGUMENTS); |
| break; |
| } |
| /* Fall through. */ |
| #endif |
| case FFI_TYPE_DOUBLE: |
| #if _CALL_ELF != 2 |
| do_double: |
| #endif |
| double_tmp = **p_argv.d; |
| if (fparg_count < NUM_FPR_ARG_REGISTERS64 && i < nfixedargs) |
| { |
| *fpr_base.d++ = double_tmp; |
| #if _CALL_ELF != 2 |
| if ((flags & FLAG_COMPAT) != 0) |
| *next_arg.d = double_tmp; |
| #endif |
| } |
| else |
| *next_arg.d = double_tmp; |
| if (++next_arg.ul == gpr_end.ul) |
| next_arg.ul = rest.ul; |
| fparg_count++; |
| FFI_ASSERT (flags & FLAG_FP_ARGUMENTS); |
| break; |
| |
| case FFI_TYPE_FLOAT: |
| #if _CALL_ELF != 2 |
| do_float: |
| #endif |
| double_tmp = **p_argv.f; |
| if (fparg_count < NUM_FPR_ARG_REGISTERS64 && i < nfixedargs) |
| { |
| *fpr_base.d++ = double_tmp; |
| #if _CALL_ELF != 2 |
| if ((flags & FLAG_COMPAT) != 0) |
| { |
| # ifndef __LITTLE_ENDIAN__ |
| next_arg.f[1] = (float) double_tmp; |
| # else |
| next_arg.f[0] = (float) double_tmp; |
| # endif |
| } |
| #endif |
| } |
| else |
| { |
| # ifndef __LITTLE_ENDIAN__ |
| next_arg.f[1] = (float) double_tmp; |
| # else |
| next_arg.f[0] = (float) double_tmp; |
| # endif |
| } |
| if (++next_arg.ul == gpr_end.ul) |
| next_arg.ul = rest.ul; |
| fparg_count++; |
| FFI_ASSERT (flags & FLAG_FP_ARGUMENTS); |
| break; |
| |
| case FFI_TYPE_STRUCT: |
| if ((ecif->cif->abi & FFI_LINUX_STRUCT_ALIGN) != 0) |
| { |
| align = (*ptr)->alignment; |
| if (align > 16) |
| align = 16; |
| if (align > 1) |
| { |
| next_arg.p = FFI_ALIGN (next_arg.p, align); |
| if (next_arg.ul == gpr_end.ul) |
| next_arg.ul = rest.ul; |
| } |
| } |
| elt = discover_homogeneous_aggregate (ecif->cif->abi, *ptr, &elnum); |
| if (elt) |
| { |
| #if _CALL_ELF == 2 |
| union { |
| void *v; |
| float *f; |
| double *d; |
| float128 *f128; |
| } arg; |
| |
| arg.v = *p_argv.v; |
| #if FFI_TYPE_LONGDOUBLE != FFI_TYPE_DOUBLE |
| if (elt == FFI_TYPE_LONGDOUBLE && |
| (ecif->cif->abi & FFI_LINUX_LONG_DOUBLE_IEEE128) != 0) |
| { |
| do |
| { |
| if (vecarg_count < NUM_VEC_ARG_REGISTERS64 |
| && i < nfixedargs) |
| memcpy (vec_base.f128++, arg.f128, sizeof (float128)); |
| else |
| memcpy (next_arg.f128, arg.f128++, sizeof (float128)); |
| if (++next_arg.f128 == gpr_end.f128) |
| next_arg.f128 = rest.f128; |
| vecarg_count++; |
| } |
| while (--elnum != 0); |
| } |
| else |
| #endif |
| if (elt == FFI_TYPE_FLOAT) |
| { |
| do |
| { |
| double_tmp = *arg.f++; |
| if (fparg_count < NUM_FPR_ARG_REGISTERS64 |
| && i < nfixedargs) |
| *fpr_base.d++ = double_tmp; |
| else |
| *next_arg.f = (float) double_tmp; |
| if (++next_arg.f == gpr_end.f) |
| next_arg.f = rest.f; |
| fparg_count++; |
| } |
| while (--elnum != 0); |
| if ((next_arg.p & 7) != 0) |
| if (++next_arg.f == gpr_end.f) |
| next_arg.f = rest.f; |
| } |
| else |
| do |
| { |
| double_tmp = *arg.d++; |
| if (fparg_count < NUM_FPR_ARG_REGISTERS64 && i < nfixedargs) |
| *fpr_base.d++ = double_tmp; |
| else |
| *next_arg.d = double_tmp; |
| if (++next_arg.d == gpr_end.d) |
| next_arg.d = rest.d; |
| fparg_count++; |
| } |
| while (--elnum != 0); |
| #else |
| if (elt == FFI_TYPE_FLOAT) |
| goto do_float; |
| else |
| goto do_double; |
| #endif |
| } |
| else |
| { |
| words = ((*ptr)->size + 7) / 8; |
| if (next_arg.ul >= gpr_base.ul && next_arg.ul + words > gpr_end.ul) |
| { |
| size_t first = gpr_end.c - next_arg.c; |
| memcpy (next_arg.c, *p_argv.c, first); |
| memcpy (rest.c, *p_argv.c + first, (*ptr)->size - first); |
| next_arg.c = rest.c + words * 8 - first; |
| } |
| else |
| { |
| char *where = next_arg.c; |
| |
| #ifndef __LITTLE_ENDIAN__ |
| /* Structures with size less than eight bytes are passed |
| left-padded. */ |
| if ((*ptr)->size < 8) |
| where += 8 - (*ptr)->size; |
| #endif |
| memcpy (where, *p_argv.c, (*ptr)->size); |
| next_arg.ul += words; |
| if (next_arg.ul == gpr_end.ul) |
| next_arg.ul = rest.ul; |
| } |
| } |
| break; |
| |
| case FFI_TYPE_UINT8: |
| gprvalue = **p_argv.uc; |
| goto putgpr; |
| case FFI_TYPE_SINT8: |
| gprvalue = **p_argv.sc; |
| goto putgpr; |
| case FFI_TYPE_UINT16: |
| gprvalue = **p_argv.us; |
| goto putgpr; |
| case FFI_TYPE_SINT16: |
| gprvalue = **p_argv.ss; |
| goto putgpr; |
| case FFI_TYPE_UINT32: |
| gprvalue = **p_argv.ui; |
| goto putgpr; |
| case FFI_TYPE_INT: |
| case FFI_TYPE_SINT32: |
| gprvalue = **p_argv.si; |
| goto putgpr; |
| |
| case FFI_TYPE_UINT64: |
| case FFI_TYPE_SINT64: |
| case FFI_TYPE_POINTER: |
| gprvalue = **p_argv.ul; |
| putgpr: |
| *next_arg.ul++ = gprvalue; |
| if (next_arg.ul == gpr_end.ul) |
| next_arg.ul = rest.ul; |
| break; |
| } |
| } |
| |
| FFI_ASSERT (flags & FLAG_4_GPR_ARGUMENTS |
| || (next_arg.ul >= gpr_base.ul |
| && next_arg.ul <= gpr_base.ul + 4)); |
| } |
| |
| |
| #if _CALL_ELF == 2 |
| #define MIN_CACHE_LINE_SIZE 8 |
| |
| static void |
| flush_icache (char *wraddr, char *xaddr, int size) |
| { |
| int i; |
| for (i = 0; i < size; i += MIN_CACHE_LINE_SIZE) |
| __asm__ volatile ("icbi 0,%0;" "dcbf 0,%1;" |
| : : "r" (xaddr + i), "r" (wraddr + i) : "memory"); |
| __asm__ volatile ("icbi 0,%0;" "dcbf 0,%1;" "sync;" "isync;" |
| : : "r"(xaddr + size - 1), "r"(wraddr + size - 1) |
| : "memory"); |
| } |
| #endif |
| |
| |
| ffi_status FFI_HIDDEN |
| ffi_prep_closure_loc_linux64 (ffi_closure *closure, |
| ffi_cif *cif, |
| void (*fun) (ffi_cif *, void *, void **, void *), |
| void *user_data, |
| void *codeloc) |
| { |
| #if _CALL_ELF == 2 |
| unsigned int *tramp = (unsigned int *) &closure->tramp[0]; |
| |
| if (cif->abi < FFI_LINUX || cif->abi >= FFI_LAST_ABI) |
| return FFI_BAD_ABI; |
| |
| tramp[0] = 0xe96c0018; /* 0: ld 11,2f-0b(12) */ |
| tramp[1] = 0xe98c0010; /* ld 12,1f-0b(12) */ |
| tramp[2] = 0x7d8903a6; /* mtctr 12 */ |
| tramp[3] = 0x4e800420; /* bctr */ |
| /* 1: .quad function_addr */ |
| /* 2: .quad context */ |
| *(void **) &tramp[4] = (void *) ffi_closure_LINUX64; |
| *(void **) &tramp[6] = codeloc; |
| flush_icache ((char *) tramp, (char *) codeloc, 4 * 4); |
| #else |
| void **tramp = (void **) &closure->tramp[0]; |
| |
| if (cif->abi < FFI_LINUX || cif->abi >= FFI_LAST_ABI) |
| return FFI_BAD_ABI; |
| |
| /* Copy function address and TOC from ffi_closure_LINUX64 OPD. */ |
| memcpy (&tramp[0], (void **) ffi_closure_LINUX64, sizeof (void *)); |
| tramp[1] = codeloc; |
| memcpy (&tramp[2], (void **) ffi_closure_LINUX64 + 1, sizeof (void *)); |
| #endif |
| |
| closure->cif = cif; |
| closure->fun = fun; |
| closure->user_data = user_data; |
| |
| return FFI_OK; |
| } |
| |
| |
| int FFI_HIDDEN |
| ffi_closure_helper_LINUX64 (ffi_cif *cif, |
| void (*fun) (ffi_cif *, void *, void **, void *), |
| void *user_data, |
| void *rvalue, |
| unsigned long *pst, |
| ffi_dblfl *pfr, |
| float128 *pvec) |
| { |
| /* rvalue is the pointer to space for return value in closure assembly */ |
| /* pst is the pointer to parameter save area |
| (r3-r10 are stored into its first 8 slots by ffi_closure_LINUX64) */ |
| /* pfr is the pointer to where f1-f13 are stored in ffi_closure_LINUX64 */ |
| /* pvec is the pointer to where v2-v13 are stored in ffi_closure_LINUX64 */ |
| |
| void **avalue; |
| ffi_type **arg_types; |
| unsigned long i, avn, nfixedargs; |
| ffi_dblfl *end_pfr = pfr + NUM_FPR_ARG_REGISTERS64; |
| float128 *end_pvec = pvec + NUM_VEC_ARG_REGISTERS64; |
| unsigned long align; |
| |
| avalue = alloca (cif->nargs * sizeof (void *)); |
| |
| /* Copy the caller's structure return value address so that the |
| closure returns the data directly to the caller. */ |
| if (cif->rtype->type == FFI_TYPE_STRUCT |
| && (cif->flags & FLAG_RETURNS_SMST) == 0) |
| { |
| rvalue = (void *) *pst; |
| pst++; |
| } |
| |
| i = 0; |
| avn = cif->nargs; |
| #if _CALL_ELF != 2 |
| nfixedargs = (unsigned) -1; |
| if ((cif->flags & FLAG_COMPAT) == 0) |
| #endif |
| nfixedargs = cif->nfixedargs; |
| arg_types = cif->arg_types; |
| |
| /* Grab the addresses of the arguments from the stack frame. */ |
| while (i < avn) |
| { |
| unsigned int elt, elnum; |
| |
| switch (arg_types[i]->type) |
| { |
| case FFI_TYPE_SINT8: |
| case FFI_TYPE_UINT8: |
| #ifndef __LITTLE_ENDIAN__ |
| avalue[i] = (char *) pst + 7; |
| pst++; |
| break; |
| #endif |
| |
| case FFI_TYPE_SINT16: |
| case FFI_TYPE_UINT16: |
| #ifndef __LITTLE_ENDIAN__ |
| avalue[i] = (char *) pst + 6; |
| pst++; |
| break; |
| #endif |
| |
| case FFI_TYPE_SINT32: |
| case FFI_TYPE_UINT32: |
| #ifndef __LITTLE_ENDIAN__ |
| avalue[i] = (char *) pst + 4; |
| pst++; |
| break; |
| #endif |
| |
| case FFI_TYPE_SINT64: |
| case FFI_TYPE_UINT64: |
| case FFI_TYPE_POINTER: |
| avalue[i] = pst; |
| pst++; |
| break; |
| |
| case FFI_TYPE_STRUCT: |
| if ((cif->abi & FFI_LINUX_STRUCT_ALIGN) != 0) |
| { |
| align = arg_types[i]->alignment; |
| if (align > 16) |
| align = 16; |
| if (align > 1) |
| pst = (unsigned long *) FFI_ALIGN ((size_t) pst, align); |
| } |
| elt = discover_homogeneous_aggregate (cif->abi, arg_types[i], &elnum); |
| if (elt) |
| { |
| #if _CALL_ELF == 2 |
| union { |
| void *v; |
| unsigned long *ul; |
| float *f; |
| double *d; |
| float128 *f128; |
| size_t p; |
| } to, from; |
| |
| /* Repackage the aggregate from its parts. The |
| aggregate size is not greater than the space taken by |
| the registers so store back to the register/parameter |
| save arrays. */ |
| #if FFI_TYPE_LONGDOUBLE != FFI_TYPE_DOUBLE |
| if (elt == FFI_TYPE_LONGDOUBLE && |
| (cif->abi & FFI_LINUX_LONG_DOUBLE_IEEE128) != 0) |
| { |
| if (pvec + elnum <= end_pvec) |
| to.v = pvec; |
| else |
| to.v = pst; |
| } |
| else |
| #endif |
| if (pfr + elnum <= end_pfr) |
| to.v = pfr; |
| else |
| to.v = pst; |
| |
| avalue[i] = to.v; |
| from.ul = pst; |
| #if FFI_TYPE_LONGDOUBLE != FFI_TYPE_DOUBLE |
| if (elt == FFI_TYPE_LONGDOUBLE && |
| (cif->abi & FFI_LINUX_LONG_DOUBLE_IEEE128) != 0) |
| { |
| do |
| { |
| if (pvec < end_pvec && i < nfixedargs) |
| memcpy (to.f128, pvec++, sizeof (float128)); |
| else |
| memcpy (to.f128, from.f128, sizeof (float128)); |
| to.f128++; |
| from.f128++; |
| } |
| while (--elnum != 0); |
| } |
| else |
| #endif |
| if (elt == FFI_TYPE_FLOAT) |
| { |
| do |
| { |
| if (pfr < end_pfr && i < nfixedargs) |
| { |
| *to.f = (float) pfr->d; |
| pfr++; |
| } |
| else |
| *to.f = *from.f; |
| to.f++; |
| from.f++; |
| } |
| while (--elnum != 0); |
| } |
| else |
| { |
| do |
| { |
| if (pfr < end_pfr && i < nfixedargs) |
| { |
| *to.d = pfr->d; |
| pfr++; |
| } |
| else |
| *to.d = *from.d; |
| to.d++; |
| from.d++; |
| } |
| while (--elnum != 0); |
| } |
| #else |
| if (elt == FFI_TYPE_FLOAT) |
| goto do_float; |
| else |
| goto do_double; |
| #endif |
| } |
| else |
| { |
| #ifndef __LITTLE_ENDIAN__ |
| /* Structures with size less than eight bytes are passed |
| left-padded. */ |
| if (arg_types[i]->size < 8) |
| avalue[i] = (char *) pst + 8 - arg_types[i]->size; |
| else |
| #endif |
| avalue[i] = pst; |
| } |
| pst += (arg_types[i]->size + 7) / 8; |
| break; |
| |
| #if FFI_TYPE_LONGDOUBLE != FFI_TYPE_DOUBLE |
| case FFI_TYPE_LONGDOUBLE: |
| if ((cif->abi & FFI_LINUX_LONG_DOUBLE_IEEE128) != 0) |
| { |
| if (((unsigned long) pst & 0xF) != 0) |
| ++pst; |
| if (pvec < end_pvec && i < nfixedargs) |
| avalue[i] = pvec++; |
| else |
| avalue[i] = pst; |
| pst += 2; |
| break; |
| } |
| else if ((cif->abi & FFI_LINUX_LONG_DOUBLE_128) != 0) |
| { |
| if (pfr + 1 < end_pfr && i + 1 < nfixedargs) |
| { |
| avalue[i] = pfr; |
| pfr += 2; |
| } |
| else |
| { |
| if (pfr < end_pfr && i < nfixedargs) |
| { |
| /* Passed partly in f13 and partly on the stack. |
| Move it all to the stack. */ |
| *pst = *(unsigned long *) pfr; |
| pfr++; |
| } |
| avalue[i] = pst; |
| } |
| pst += 2; |
| break; |
| } |
| /* Fall through. */ |
| #endif |
| case FFI_TYPE_DOUBLE: |
| #if _CALL_ELF != 2 |
| do_double: |
| #endif |
| /* On the outgoing stack all values are aligned to 8 */ |
| /* there are 13 64bit floating point registers */ |
| |
| if (pfr < end_pfr && i < nfixedargs) |
| { |
| avalue[i] = pfr; |
| pfr++; |
| } |
| else |
| avalue[i] = pst; |
| pst++; |
| break; |
| |
| case FFI_TYPE_FLOAT: |
| #if _CALL_ELF != 2 |
| do_float: |
| #endif |
| if (pfr < end_pfr && i < nfixedargs) |
| { |
| /* Float values are stored as doubles in the |
| ffi_closure_LINUX64 code. Fix them here. */ |
| pfr->f = (float) pfr->d; |
| avalue[i] = pfr; |
| pfr++; |
| } |
| else |
| { |
| #ifndef __LITTLE_ENDIAN__ |
| avalue[i] = (char *) pst + 4; |
| #else |
| avalue[i] = pst; |
| #endif |
| } |
| pst++; |
| break; |
| |
| default: |
| FFI_ASSERT (0); |
| } |
| |
| i++; |
| } |
| |
| (*fun) (cif, rvalue, avalue, user_data); |
| |
| /* Tell ffi_closure_LINUX64 how to perform return type promotions. */ |
| if ((cif->flags & FLAG_RETURNS_SMST) != 0) |
| { |
| if ((cif->flags & (FLAG_RETURNS_FP | FLAG_RETURNS_VEC)) == 0) |
| return FFI_V2_TYPE_SMALL_STRUCT + cif->rtype->size - 1; |
| else if ((cif->flags & FLAG_RETURNS_VEC) != 0) |
| return FFI_V2_TYPE_VECTOR_HOMOG; |
| else if ((cif->flags & FLAG_RETURNS_64BITS) != 0) |
| return FFI_V2_TYPE_DOUBLE_HOMOG; |
| else |
| return FFI_V2_TYPE_FLOAT_HOMOG; |
| } |
| if ((cif->flags & FLAG_RETURNS_VEC) != 0) |
| return FFI_V2_TYPE_VECTOR; |
| return cif->rtype->type; |
| } |
| #endif |