| /* Subroutines for insn-output.c for Sun SPARC. |
| Copyright (C) 1987, 88, 89, 92-96, 1997 Free Software Foundation, Inc. |
| Contributed by Michael Tiemann (tiemann@cygnus.com) |
| 64 bit SPARC V9 support by Michael Tiemann, Jim Wilson, and Doug Evans, |
| at Cygnus Support. |
| |
| This file is part of GNU CC. |
| |
| GNU CC is free software; you can redistribute it and/or modify |
| it under the terms of the GNU General Public License as published by |
| the Free Software Foundation; either version 2, or (at your option) |
| any later version. |
| |
| GNU CC is distributed in the hope that it will be useful, |
| but WITHOUT ANY WARRANTY; without even the implied warranty of |
| MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| GNU General Public License for more details. |
| |
| You should have received a copy of the GNU General Public License |
| along with GNU CC; see the file COPYING. If not, write to |
| the Free Software Foundation, 59 Temple Place - Suite 330, |
| Boston, MA 02111-1307, USA. */ |
| |
| #include <stdio.h> |
| #include "config.h" |
| #include "tree.h" |
| #include "rtl.h" |
| #include "regs.h" |
| #include "hard-reg-set.h" |
| #include "real.h" |
| #include "insn-config.h" |
| #include "conditions.h" |
| #include "insn-flags.h" |
| #include "output.h" |
| #include "insn-attr.h" |
| #include "flags.h" |
| #include "expr.h" |
| #include "recog.h" |
| |
| /* 1 if the caller has placed an "unimp" insn immediately after the call. |
| This is used in v8 code when calling a function that returns a structure. |
| v9 doesn't have this. Be careful to have this test be the same as that |
| used on the call. */ |
| |
| #define SKIP_CALLERS_UNIMP_P \ |
| (!TARGET_ARCH64 && current_function_returns_struct \ |
| && ! integer_zerop (DECL_SIZE (DECL_RESULT (current_function_decl))) \ |
| && (TREE_CODE (DECL_SIZE (DECL_RESULT (current_function_decl))) \ |
| == INTEGER_CST)) |
| |
| /* Global variables for machine-dependent things. */ |
| |
| /* Size of frame. Need to know this to emit return insns from leaf procedures. |
| ACTUAL_FSIZE is set by compute_frame_size() which is called during the |
| reload pass. This is important as the value is later used in insn |
| scheduling (to see what can go in a delay slot). |
| APPARENT_FSIZE is the size of the stack less the register save area and less |
| the outgoing argument area. It is used when saving call preserved regs. */ |
| static int apparent_fsize; |
| static int actual_fsize; |
| |
| /* Save the operands last given to a compare for use when we |
| generate a scc or bcc insn. */ |
| |
| rtx sparc_compare_op0, sparc_compare_op1; |
| |
| /* Count of named arguments (v9 only). |
| ??? INIT_CUMULATIVE_ARGS initializes these, and FUNCTION_ARG_ADVANCE |
| increments SPARC_ARG_COUNT. They are then used by |
| FUNCTION_ARG_CALLEE_COPIES to determine if the argument is really a named |
| argument or not. This hack is necessary because the NAMED argument to the |
| FUNCTION_ARG_XXX macros is not what it says it is: it does not include the |
| last named argument. */ |
| |
| int sparc_arg_count; |
| int sparc_n_named_args; |
| |
| /* We may need an epilogue if we spill too many registers. |
| If this is non-zero, then we branch here for the epilogue. */ |
| static rtx leaf_label; |
| |
| #ifdef LEAF_REGISTERS |
| |
| /* Vector to say how input registers are mapped to output |
| registers. FRAME_POINTER_REGNUM cannot be remapped by |
| this function to eliminate it. You must use -fomit-frame-pointer |
| to get that. */ |
| char leaf_reg_remap[] = |
| { 0, 1, 2, 3, 4, 5, 6, 7, |
| -1, -1, -1, -1, -1, -1, 14, -1, |
| -1, -1, -1, -1, -1, -1, -1, -1, |
| 8, 9, 10, 11, 12, 13, -1, 15, |
| |
| 32, 33, 34, 35, 36, 37, 38, 39, |
| 40, 41, 42, 43, 44, 45, 46, 47, |
| 48, 49, 50, 51, 52, 53, 54, 55, |
| 56, 57, 58, 59, 60, 61, 62, 63, |
| 64, 65, 66, 67, 68, 69, 70, 71, |
| 72, 73, 74, 75, 76, 77, 78, 79, |
| 80, 81, 82, 83, 84, 85, 86, 87, |
| 88, 89, 90, 91, 92, 93, 94, 95, |
| 96, 97, 98, 99, 100}; |
| |
| #endif |
| |
| /* Name of where we pretend to think the frame pointer points. |
| Normally, this is "%fp", but if we are in a leaf procedure, |
| this is "%sp+something". We record "something" separately as it may be |
| too big for reg+constant addressing. */ |
| |
| static char *frame_base_name; |
| static int frame_base_offset; |
| |
| static rtx find_addr_reg (); |
| static void sparc_init_modes (); |
| |
| /* Option handling. */ |
| |
| struct sparc_cpu_select sparc_select[] = |
| { |
| /* switch name, tune arch */ |
| { (char *)0, "default", 1, 1 }, |
| { (char *)0, "-mcpu=", 1, 1 }, |
| { (char *)0, "-mtune=", 1, 0 }, |
| { 0, 0 } |
| }; |
| |
| /* CPU type. This is set from TARGET_CPU_DEFAULT and -m{cpu,tune}=xxx. */ |
| enum processor_type sparc_cpu; |
| |
| /* Validate and override various options, and do some machine dependent |
| initialization. */ |
| |
| void |
| sparc_override_options () |
| { |
| /* Map TARGET_CPU_DEFAULT to value for -m{arch,tune}=. */ |
| static struct cpu_default { |
| int cpu; |
| char *name; |
| } cpu_default[] = { |
| { TARGET_CPU_sparc, "cypress" }, |
| { TARGET_CPU_v8, "v8" }, |
| { TARGET_CPU_supersparc, "supersparc" }, |
| { TARGET_CPU_sparclet, "tsc701" }, |
| { TARGET_CPU_sparclite, "f930" }, |
| { TARGET_CPU_ultrasparc, "ultrasparc" }, |
| { 0 } |
| }; |
| struct cpu_default *def; |
| /* Table of values for -m{cpu,tune}=. */ |
| static struct cpu_table { |
| char *name; |
| enum processor_type processor; |
| int disable; |
| int enable; |
| } cpu_table[] = { |
| { "v7", PROCESSOR_V7, MASK_ISA, 0 }, |
| { "cypress", PROCESSOR_CYPRESS, MASK_ISA, 0 }, |
| { "v8", PROCESSOR_V8, MASK_ISA, MASK_V8 }, |
| /* TI TMS390Z55 supersparc */ |
| { "supersparc", PROCESSOR_SUPERSPARC, MASK_ISA, MASK_V8 }, |
| { "sparclite", PROCESSOR_SPARCLITE, MASK_ISA, MASK_SPARCLITE }, |
| /* The Fujitsu MB86930 is the original sparclite chip, with no fpu. |
| The Fujitsu MB86934 is the recent sparclite chip, with an fpu. */ |
| { "f930", PROCESSOR_F930, MASK_ISA|MASK_FPU, MASK_SPARCLITE }, |
| { "f934", PROCESSOR_F934, MASK_ISA, MASK_SPARCLITE|MASK_FPU }, |
| { "sparclet", PROCESSOR_SPARCLET, MASK_ISA, MASK_SPARCLET }, |
| /* TEMIC sparclet */ |
| { "tsc701", PROCESSOR_TSC701, MASK_ISA, MASK_SPARCLET }, |
| /* "v9" is used to specify a true 64 bit architecture. |
| "v8plus" is what Sun calls Solaris2 running on UltraSPARC's. */ |
| { "v8plus", PROCESSOR_V8PLUS, MASK_ISA, MASK_V9 }, |
| #if SPARC_ARCH64 |
| { "v9", PROCESSOR_V9, MASK_ISA, MASK_V9 }, |
| #endif |
| /* TI ultrasparc */ |
| { "ultrasparc", PROCESSOR_ULTRASPARC, MASK_ISA, MASK_V9 }, |
| { 0 } |
| }; |
| struct cpu_table *cpu; |
| struct sparc_cpu_select *sel; |
| |
| int fpu = TARGET_FPU; /* save current -mfpu status */ |
| |
| /* Set the default. */ |
| for (def = &cpu_default[0]; def->name; ++def) |
| if (def->cpu == TARGET_CPU_DEFAULT) |
| break; |
| if (! def->name) |
| abort (); |
| sparc_select[0].string = def->name; |
| |
| for (sel = &sparc_select[0]; sel->name; ++sel) |
| { |
| if (sel->string) |
| { |
| for (cpu = &cpu_table[0]; cpu->name; ++cpu) |
| if (! strcmp (sel->string, cpu->name)) |
| { |
| if (sel->set_tune_p) |
| sparc_cpu = cpu->processor; |
| |
| if (sel->set_arch_p) |
| { |
| target_flags &= ~cpu->disable; |
| target_flags |= cpu->enable; |
| } |
| break; |
| } |
| |
| if (! cpu->name) |
| error ("bad value (%s) for %s switch", sel->string, sel->name); |
| } |
| } |
| |
| /* If -mfpu or -mno-fpu was explicitly used, don't override with |
| the processor default. */ |
| if (TARGET_FPU_SET) |
| target_flags = (target_flags & ~MASK_FPU) | fpu; |
| |
| /* Use the deprecated v8 insns for sparc64 in 32 bit mode. */ |
| if (TARGET_V9 && TARGET_ARCH32) |
| target_flags |= MASK_DEPRECATED_V8_INSNS; |
| |
| /* Do various machine dependent initializations. */ |
| sparc_init_modes (); |
| } |
| |
| /* Float conversions (v9 only). |
| |
| The floating point registers cannot hold DImode values because SUBREG's |
| on them get the wrong register. "(subreg:SI (reg:DI M int-reg) 0)" is the |
| same as "(subreg:SI (reg:DI N float-reg) 1)", but gcc doesn't know how to |
| turn the "0" to a "1". Therefore, we must explicitly do the conversions |
| to/from int/fp regs. `sparc64_fpconv_stack_slot' is the address of an |
| 8 byte stack slot used during the transfer. |
| ??? I could have used [%fp-16] but I didn't want to add yet another |
| dependence on this. */ |
| /* ??? Can we use assign_stack_temp here? */ |
| |
| static rtx fpconv_stack_temp; |
| |
| /* Called once for each function. */ |
| |
| void |
| sparc64_init_expanders () |
| { |
| fpconv_stack_temp = NULL_RTX; |
| } |
| |
| /* Assign a stack temp for fp/int DImode conversions. */ |
| |
| rtx |
| sparc64_fpconv_stack_temp () |
| { |
| if (fpconv_stack_temp == NULL_RTX) |
| fpconv_stack_temp = |
| assign_stack_local (DImode, GET_MODE_SIZE (DImode), 0); |
| |
| return fpconv_stack_temp; |
| } |
| |
| /* Miscellaneous utilities. */ |
| |
| /* Nonzero if CODE, a comparison, is suitable for use in v9 conditional move |
| or branch on register contents instructions. */ |
| |
| int |
| v9_regcmp_p (code) |
| enum rtx_code code; |
| { |
| return (code == EQ || code == NE || code == GE || code == LT |
| || code == LE || code == GT); |
| } |
| |
| /* Operand constraints. */ |
| |
| /* Return non-zero only if OP is a register of mode MODE, |
| or const0_rtx. Don't allow const0_rtx if TARGET_LIVE_G0 because |
| %g0 may contain anything. */ |
| |
| int |
| reg_or_0_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| if (register_operand (op, mode)) |
| return 1; |
| if (TARGET_LIVE_G0) |
| return 0; |
| if (op == const0_rtx) |
| return 1; |
| if (GET_MODE (op) == VOIDmode && GET_CODE (op) == CONST_DOUBLE |
| && CONST_DOUBLE_HIGH (op) == 0 |
| && CONST_DOUBLE_LOW (op) == 0) |
| return 1; |
| if (GET_MODE_CLASS (GET_MODE (op)) == MODE_FLOAT |
| && GET_CODE (op) == CONST_DOUBLE |
| && fp_zero_operand (op)) |
| return 1; |
| return 0; |
| } |
| |
| /* Nonzero if OP is a floating point value with value 0.0. */ |
| |
| int |
| fp_zero_operand (op) |
| rtx op; |
| { |
| REAL_VALUE_TYPE r; |
| |
| REAL_VALUE_FROM_CONST_DOUBLE (r, op); |
| return (REAL_VALUES_EQUAL (r, dconst0) && ! REAL_VALUE_MINUS_ZERO (r)); |
| } |
| |
| /* Nonzero if OP is an integer register. */ |
| |
| int |
| intreg_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return (register_operand (op, SImode) |
| || (TARGET_ARCH64 && register_operand (op, DImode))); |
| } |
| |
| /* Nonzero if OP is a floating point condition code register. */ |
| |
| int |
| fcc_reg_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| /* This can happen when recog is called from combine. Op may be a MEM. |
| Fail instead of calling abort in this case. */ |
| if (GET_CODE (op) != REG) |
| return 0; |
| |
| if (mode != VOIDmode && mode != GET_MODE (op)) |
| return 0; |
| if (mode == VOIDmode |
| && (GET_MODE (op) != CCFPmode && GET_MODE (op) != CCFPEmode)) |
| return 0; |
| |
| #if 0 /* ??? ==> 1 when %fcc0-3 are pseudos first. See gen_compare_reg(). */ |
| if (reg_renumber == 0) |
| return REGNO (op) >= FIRST_PSEUDO_REGISTER; |
| return REGNO_OK_FOR_CCFP_P (REGNO (op)); |
| #else |
| return (unsigned) REGNO (op) - SPARC_FIRST_V9_FCC_REG < 4; |
| #endif |
| } |
| |
| /* Nonzero if OP is an integer or floating point condition code register. */ |
| |
| int |
| icc_or_fcc_reg_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| if (GET_CODE (op) == REG && REGNO (op) == SPARC_ICC_REG) |
| { |
| if (mode != VOIDmode && mode != GET_MODE (op)) |
| return 0; |
| if (mode == VOIDmode |
| && GET_MODE (op) != CCmode && GET_MODE (op) != CCXmode) |
| return 0; |
| return 1; |
| } |
| |
| return fcc_reg_operand (op, mode); |
| } |
| |
| /* Nonzero if OP can appear as the dest of a RESTORE insn. */ |
| int |
| restore_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return (GET_CODE (op) == REG && GET_MODE (op) == mode |
| && (REGNO (op) < 8 || (REGNO (op) >= 24 && REGNO (op) < 32))); |
| } |
| |
| /* Call insn on SPARC can take a PC-relative constant address, or any regular |
| memory address. */ |
| |
| int |
| call_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| if (GET_CODE (op) != MEM) |
| abort (); |
| op = XEXP (op, 0); |
| return (symbolic_operand (op, mode) || memory_address_p (Pmode, op)); |
| } |
| |
| int |
| call_operand_address (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return (symbolic_operand (op, mode) || memory_address_p (Pmode, op)); |
| } |
| |
| /* Returns 1 if OP is either a symbol reference or a sum of a symbol |
| reference and a constant. */ |
| |
| int |
| symbolic_operand (op, mode) |
| register rtx op; |
| enum machine_mode mode; |
| { |
| switch (GET_CODE (op)) |
| { |
| case SYMBOL_REF: |
| case LABEL_REF: |
| return 1; |
| |
| case CONST: |
| op = XEXP (op, 0); |
| return ((GET_CODE (XEXP (op, 0)) == SYMBOL_REF |
| || GET_CODE (XEXP (op, 0)) == LABEL_REF) |
| && GET_CODE (XEXP (op, 1)) == CONST_INT); |
| |
| /* ??? This clause seems to be irrelevant. */ |
| case CONST_DOUBLE: |
| return GET_MODE (op) == mode; |
| |
| default: |
| return 0; |
| } |
| } |
| |
| /* Return truth value of statement that OP is a symbolic memory |
| operand of mode MODE. */ |
| |
| int |
| symbolic_memory_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| if (GET_CODE (op) == SUBREG) |
| op = SUBREG_REG (op); |
| if (GET_CODE (op) != MEM) |
| return 0; |
| op = XEXP (op, 0); |
| return (GET_CODE (op) == SYMBOL_REF || GET_CODE (op) == CONST |
| || GET_CODE (op) == HIGH || GET_CODE (op) == LABEL_REF); |
| } |
| |
| /* Return truth value of statement that OP is a LABEL_REF of mode MODE. */ |
| |
| int |
| label_ref_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| if (GET_CODE (op) != LABEL_REF) |
| return 0; |
| if (GET_MODE (op) != mode) |
| return 0; |
| return 1; |
| } |
| |
| /* Return 1 if the operand is an argument used in generating pic references |
| in either the medium/low or medium/anywhere code models of sparc64. */ |
| |
| int |
| sp64_medium_pic_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| /* Check for (const (minus (symbol_ref:GOT) |
| (const (minus (label) (pc))))). */ |
| if (GET_CODE (op) != CONST) |
| return 0; |
| op = XEXP (op, 0); |
| if (GET_CODE (op) != MINUS) |
| return 0; |
| if (GET_CODE (XEXP (op, 0)) != SYMBOL_REF) |
| return 0; |
| /* ??? Ensure symbol is GOT. */ |
| if (GET_CODE (XEXP (op, 1)) != CONST) |
| return 0; |
| if (GET_CODE (XEXP (XEXP (op, 1), 0)) != MINUS) |
| return 0; |
| return 1; |
| } |
| |
| /* Return 1 if the operand is a data segment reference. This includes |
| the readonly data segment, or in other words anything but the text segment. |
| This is needed in the medium/anywhere code model on v9. These values |
| are accessed with MEDANY_BASE_REG. */ |
| |
| int |
| data_segment_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| switch (GET_CODE (op)) |
| { |
| case SYMBOL_REF : |
| return ! SYMBOL_REF_FLAG (op); |
| case PLUS : |
| /* Assume canonical format of symbol + constant. |
| Fall through. */ |
| case CONST : |
| return data_segment_operand (XEXP (op, 0)); |
| default : |
| return 0; |
| } |
| } |
| |
| /* Return 1 if the operand is a text segment reference. |
| This is needed in the medium/anywhere code model on v9. */ |
| |
| int |
| text_segment_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| switch (GET_CODE (op)) |
| { |
| case LABEL_REF : |
| return 1; |
| case SYMBOL_REF : |
| return SYMBOL_REF_FLAG (op); |
| case PLUS : |
| /* Assume canonical format of symbol + constant. |
| Fall through. */ |
| case CONST : |
| return text_segment_operand (XEXP (op, 0)); |
| default : |
| return 0; |
| } |
| } |
| |
| /* Return 1 if the operand is either a register or a memory operand that is |
| not symbolic. */ |
| |
| int |
| reg_or_nonsymb_mem_operand (op, mode) |
| register rtx op; |
| enum machine_mode mode; |
| { |
| if (register_operand (op, mode)) |
| return 1; |
| |
| if (memory_operand (op, mode) && ! symbolic_memory_operand (op, mode)) |
| return 1; |
| |
| return 0; |
| } |
| |
| int |
| sparc_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| if (register_operand (op, mode)) |
| return 1; |
| if (GET_CODE (op) == CONST_INT) |
| return SMALL_INT (op); |
| if (GET_MODE (op) != mode) |
| return 0; |
| if (GET_CODE (op) == SUBREG) |
| op = SUBREG_REG (op); |
| if (GET_CODE (op) != MEM) |
| return 0; |
| |
| op = XEXP (op, 0); |
| if (GET_CODE (op) == LO_SUM) |
| return (GET_CODE (XEXP (op, 0)) == REG |
| && symbolic_operand (XEXP (op, 1), Pmode)); |
| return memory_address_p (mode, op); |
| } |
| |
| int |
| move_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| if (mode == DImode && arith_double_operand (op, mode)) |
| return 1; |
| if (register_operand (op, mode)) |
| return 1; |
| if (GET_CODE (op) == CONST_INT) |
| return (SMALL_INT (op) || (INTVAL (op) & 0x3ff) == 0); |
| |
| if (GET_MODE (op) != mode) |
| return 0; |
| if (GET_CODE (op) == SUBREG) |
| op = SUBREG_REG (op); |
| if (GET_CODE (op) != MEM) |
| return 0; |
| op = XEXP (op, 0); |
| if (GET_CODE (op) == LO_SUM) |
| return (register_operand (XEXP (op, 0), Pmode) |
| && CONSTANT_P (XEXP (op, 1))); |
| return memory_address_p (mode, op); |
| } |
| |
| int |
| splittable_symbolic_memory_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| if (GET_CODE (op) != MEM) |
| return 0; |
| if (! symbolic_operand (XEXP (op, 0), Pmode)) |
| return 0; |
| return 1; |
| } |
| |
| int |
| splittable_immediate_memory_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| if (GET_CODE (op) != MEM) |
| return 0; |
| if (! immediate_operand (XEXP (op, 0), Pmode)) |
| return 0; |
| return 1; |
| } |
| |
| /* Return truth value of whether OP is EQ or NE. */ |
| |
| int |
| eq_or_neq (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return (GET_CODE (op) == EQ || GET_CODE (op) == NE); |
| } |
| |
| /* Return 1 if this is a comparison operator, but not an EQ, NE, GEU, |
| or LTU for non-floating-point. We handle those specially. */ |
| |
| int |
| normal_comp_operator (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| enum rtx_code code = GET_CODE (op); |
| |
| if (GET_RTX_CLASS (code) != '<') |
| return 0; |
| |
| if (GET_MODE (XEXP (op, 0)) == CCFPmode |
| || GET_MODE (XEXP (op, 0)) == CCFPEmode) |
| return 1; |
| |
| return (code != NE && code != EQ && code != GEU && code != LTU); |
| } |
| |
| /* Return 1 if this is a comparison operator. This allows the use of |
| MATCH_OPERATOR to recognize all the branch insns. */ |
| |
| int |
| noov_compare_op (op, mode) |
| register rtx op; |
| enum machine_mode mode; |
| { |
| enum rtx_code code = GET_CODE (op); |
| |
| if (GET_RTX_CLASS (code) != '<') |
| return 0; |
| |
| if (GET_MODE (XEXP (op, 0)) == CC_NOOVmode) |
| /* These are the only branches which work with CC_NOOVmode. */ |
| return (code == EQ || code == NE || code == GE || code == LT); |
| return 1; |
| } |
| |
| /* Nonzero if OP is a comparison operator suitable for use in v9 |
| conditional move or branch on register contents instructions. */ |
| |
| int |
| v9_regcmp_op (op, mode) |
| register rtx op; |
| enum machine_mode mode; |
| { |
| enum rtx_code code = GET_CODE (op); |
| |
| if (GET_RTX_CLASS (code) != '<') |
| return 0; |
| |
| return v9_regcmp_p (code); |
| } |
| |
| /* Return 1 if this is a SIGN_EXTEND or ZERO_EXTEND operation. */ |
| |
| int |
| extend_op (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return GET_CODE (op) == SIGN_EXTEND || GET_CODE (op) == ZERO_EXTEND; |
| } |
| |
| /* Return nonzero if OP is an operator of mode MODE which can set |
| the condition codes explicitly. We do not include PLUS and MINUS |
| because these require CC_NOOVmode, which we handle explicitly. */ |
| |
| int |
| cc_arithop (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| if (GET_CODE (op) == AND |
| || GET_CODE (op) == IOR |
| || GET_CODE (op) == XOR) |
| return 1; |
| |
| return 0; |
| } |
| |
| /* Return nonzero if OP is an operator of mode MODE which can bitwise |
| complement its second operand and set the condition codes explicitly. */ |
| |
| int |
| cc_arithopn (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| /* XOR is not here because combine canonicalizes (xor (not ...) ...) |
| and (xor ... (not ...)) to (not (xor ...)). */ |
| return (GET_CODE (op) == AND |
| || GET_CODE (op) == IOR); |
| } |
| |
| /* Return true if OP is a register, or is a CONST_INT that can fit in a |
| signed 13 bit immediate field. This is an acceptable SImode operand for |
| most 3 address instructions. */ |
| |
| int |
| arith_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return (register_operand (op, mode) |
| || (GET_CODE (op) == CONST_INT && SMALL_INT (op))); |
| } |
| |
| /* Return true if OP is a register, or is a CONST_INT that can fit in a |
| signed 11 bit immediate field. This is an acceptable SImode operand for |
| the movcc instructions. */ |
| |
| int |
| arith11_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return (register_operand (op, mode) |
| || (GET_CODE (op) == CONST_INT && SPARC_SIMM11_P (INTVAL (op)))); |
| } |
| |
| /* Return true if OP is a register, or is a CONST_INT that can fit in a |
| signed 10 bit immediate field. This is an acceptable SImode operand for |
| the movrcc instructions. */ |
| |
| int |
| arith10_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return (register_operand (op, mode) |
| || (GET_CODE (op) == CONST_INT && SPARC_SIMM10_P (INTVAL (op)))); |
| } |
| |
| /* Return true if OP is a register, is a CONST_INT that fits in a 13 bit |
| immediate field, or is a CONST_DOUBLE whose both parts fit in a 13 bit |
| immediate field. |
| v9: Return true if OP is a register, or is a CONST_INT or CONST_DOUBLE that |
| can fit in a 13 bit immediate field. This is an acceptable DImode operand |
| for most 3 address instructions. */ |
| |
| int |
| arith_double_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return (register_operand (op, mode) |
| || (GET_CODE (op) == CONST_INT && SMALL_INT (op)) |
| || (! TARGET_ARCH64 |
| && GET_CODE (op) == CONST_DOUBLE |
| && (unsigned) (CONST_DOUBLE_LOW (op) + 0x1000) < 0x2000 |
| && (unsigned) (CONST_DOUBLE_HIGH (op) + 0x1000) < 0x2000) |
| || (TARGET_ARCH64 |
| && GET_CODE (op) == CONST_DOUBLE |
| && (unsigned) (CONST_DOUBLE_LOW (op) + 0x1000) < 0x2000 |
| && ((CONST_DOUBLE_HIGH (op) == -1 |
| && (CONST_DOUBLE_LOW (op) & 0x1000) == 0x1000) |
| || (CONST_DOUBLE_HIGH (op) == 0 |
| && (CONST_DOUBLE_LOW (op) & 0x1000) == 0)))); |
| } |
| |
| /* Return true if OP is a register, or is a CONST_INT or CONST_DOUBLE that |
| can fit in an 11 bit immediate field. This is an acceptable DImode |
| operand for the movcc instructions. */ |
| /* ??? Replace with arith11_operand? */ |
| |
| int |
| arith11_double_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return (register_operand (op, mode) |
| || (GET_CODE (op) == CONST_DOUBLE |
| && (GET_MODE (op) == mode || GET_MODE (op) == VOIDmode) |
| && (unsigned) (CONST_DOUBLE_LOW (op) + 0x400) < 0x800 |
| && ((CONST_DOUBLE_HIGH (op) == -1 |
| && (CONST_DOUBLE_LOW (op) & 0x400) == 0x400) |
| || (CONST_DOUBLE_HIGH (op) == 0 |
| && (CONST_DOUBLE_LOW (op) & 0x400) == 0))) |
| || (GET_CODE (op) == CONST_INT |
| && (GET_MODE (op) == mode || GET_MODE (op) == VOIDmode) |
| && (unsigned) (INTVAL (op) + 0x400) < 0x800)); |
| } |
| |
| /* Return true if OP is a register, or is a CONST_INT or CONST_DOUBLE that |
| can fit in an 10 bit immediate field. This is an acceptable DImode |
| operand for the movrcc instructions. */ |
| /* ??? Replace with arith10_operand? */ |
| |
| int |
| arith10_double_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return (register_operand (op, mode) |
| || (GET_CODE (op) == CONST_DOUBLE |
| && (GET_MODE (op) == mode || GET_MODE (op) == VOIDmode) |
| && (unsigned) (CONST_DOUBLE_LOW (op) + 0x200) < 0x400 |
| && ((CONST_DOUBLE_HIGH (op) == -1 |
| && (CONST_DOUBLE_LOW (op) & 0x200) == 0x200) |
| || (CONST_DOUBLE_HIGH (op) == 0 |
| && (CONST_DOUBLE_LOW (op) & 0x200) == 0))) |
| || (GET_CODE (op) == CONST_INT |
| && (GET_MODE (op) == mode || GET_MODE (op) == VOIDmode) |
| && (unsigned) (INTVAL (op) + 0x200) < 0x400)); |
| } |
| |
| /* Return truth value of whether OP is a integer which fits the |
| range constraining immediate operands in most three-address insns, |
| which have a 13 bit immediate field. */ |
| |
| int |
| small_int (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return (GET_CODE (op) == CONST_INT && SMALL_INT (op)); |
| } |
| |
| /* Recognize operand values for the umul instruction. That instruction sign |
| extends immediate values just like all other sparc instructions, but |
| interprets the extended result as an unsigned number. */ |
| |
| int |
| uns_small_int (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| #if HOST_BITS_PER_WIDE_INT > 32 |
| /* All allowed constants will fit a CONST_INT. */ |
| return (GET_CODE (op) == CONST_INT |
| && ((INTVAL (op) >= 0 && INTVAL (op) < 0x1000) |
| || (INTVAL (op) >= 0xFFFFF000 && INTVAL (op) < 0x100000000L))); |
| #else |
| return ((GET_CODE (op) == CONST_INT && (unsigned) INTVAL (op) < 0x1000) |
| || (GET_CODE (op) == CONST_DOUBLE |
| && CONST_DOUBLE_HIGH (op) == 0 |
| && (unsigned) CONST_DOUBLE_LOW (op) - 0xFFFFF000 < 0x1000)); |
| #endif |
| } |
| |
| int |
| uns_arith_operand (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return register_operand (op, mode) || uns_small_int (op, mode); |
| } |
| |
| /* Return truth value of statement that OP is a call-clobbered register. */ |
| int |
| clobbered_register (op, mode) |
| rtx op; |
| enum machine_mode mode; |
| { |
| return (GET_CODE (op) == REG && call_used_regs[REGNO (op)]); |
| } |
| |
| /* X and Y are two things to compare using CODE. Emit the compare insn and |
| return the rtx for the cc reg in the proper mode. */ |
| |
| rtx |
| gen_compare_reg (code, x, y) |
| enum rtx_code code; |
| rtx x, y; |
| { |
| enum machine_mode mode = SELECT_CC_MODE (code, x, y); |
| rtx cc_reg; |
| |
| /* ??? We don't have movcc patterns so we cannot generate pseudo regs for the |
| fcc regs (cse can't tell they're really call clobbered regs and will |
| remove a duplicate comparison even if there is an intervening function |
| call - it will then try to reload the cc reg via an int reg which is why |
| we need the movcc patterns). It is possible to provide the movcc |
| patterns by using the ldxfsr/stxfsr v9 insns. I tried it: you need two |
| registers (say %g1,%g5) and it takes about 6 insns. A better fix would be |
| to tell cse that CCFPE mode registers (even pseudos) are call |
| clobbered. */ |
| |
| /* ??? This is an experiment. Rather than making changes to cse which may |
| or may not be easy/clean, we do our own cse. This is possible because |
| we will generate hard registers. Cse knows they're call clobbered (it |
| doesn't know the same thing about pseudos). If we guess wrong, no big |
| deal, but if we win, great! */ |
| |
| if (TARGET_V9 && GET_MODE_CLASS (GET_MODE (x)) == MODE_FLOAT) |
| #if 1 /* experiment */ |
| { |
| int reg; |
| /* We cycle through the registers to ensure they're all exercised. */ |
| static int next_fcc_reg = 0; |
| /* Previous x,y for each fcc reg. */ |
| static rtx prev_args[4][2]; |
| |
| /* Scan prev_args for x,y. */ |
| for (reg = 0; reg < 4; reg++) |
| if (prev_args[reg][0] == x && prev_args[reg][1] == y) |
| break; |
| if (reg == 4) |
| { |
| reg = next_fcc_reg; |
| prev_args[reg][0] = x; |
| prev_args[reg][1] = y; |
| next_fcc_reg = (next_fcc_reg + 1) & 3; |
| } |
| cc_reg = gen_rtx (REG, mode, reg + SPARC_FIRST_V9_FCC_REG); |
| } |
| #else |
| cc_reg = gen_reg_rtx (mode); |
| #endif /* ! experiment */ |
| else if (GET_MODE_CLASS (GET_MODE (x)) == MODE_FLOAT) |
| cc_reg = gen_rtx (REG, mode, SPARC_FCC_REG); |
| else |
| cc_reg = gen_rtx (REG, mode, SPARC_ICC_REG); |
| |
| emit_insn (gen_rtx (SET, VOIDmode, cc_reg, |
| gen_rtx (COMPARE, mode, x, y))); |
| |
| return cc_reg; |
| } |
| |
| /* This function is used for v9 only. |
| CODE is the code for an Scc's comparison. |
| OPERANDS[0] is the target of the Scc insn. |
| OPERANDS[1] is the value we compare against const0_rtx (which hasn't |
| been generated yet). |
| |
| This function is needed to turn |
| |
| (set (reg:SI 110) |
| (gt (reg:CCX 100 %icc) |
| (const_int 0))) |
| into |
| (set (reg:SI 110) |
| (gt:DI (reg:CCX 100 %icc) |
| (const_int 0))) |
| |
| IE: The instruction recognizer needs to see the mode of the comparison to |
| find the right instruction. We could use "gt:DI" right in the |
| define_expand, but leaving it out allows us to handle DI, SI, etc. |
| |
| We refer to the global sparc compare operands sparc_compare_op0 and |
| sparc_compare_op1. */ |
| |
| int |
| gen_v9_scc (compare_code, operands) |
| enum rtx_code compare_code; |
| register rtx *operands; |
| { |
| rtx temp, op0, op1; |
| |
| if (! TARGET_ARCH64 |
| && (GET_MODE (sparc_compare_op0) == DImode |
| || GET_MODE (operands[0]) == DImode)) |
| return 0; |
| |
| /* Handle the case where operands[0] == sparc_compare_op0. |
| We "early clobber" the result. */ |
| if (REGNO (operands[0]) == REGNO (sparc_compare_op0)) |
| { |
| op0 = gen_reg_rtx (GET_MODE (sparc_compare_op0)); |
| emit_move_insn (op0, sparc_compare_op0); |
| } |
| else |
| op0 = sparc_compare_op0; |
| /* For consistency in the following. */ |
| op1 = sparc_compare_op1; |
| |
| /* Try to use the movrCC insns. */ |
| if (TARGET_ARCH64 |
| && GET_MODE_CLASS (GET_MODE (op0)) == MODE_INT |
| && op1 == const0_rtx |
| && v9_regcmp_p (compare_code)) |
| { |
| /* Special case for op0 != 0. This can be done with one instruction if |
| operands[0] == sparc_compare_op0. We don't assume they are equal |
| now though. */ |
| |
| if (compare_code == NE |
| && GET_MODE (operands[0]) == DImode |
| && GET_MODE (op0) == DImode) |
| { |
| emit_insn (gen_rtx (SET, VOIDmode, operands[0], op0)); |
| emit_insn (gen_rtx (SET, VOIDmode, operands[0], |
| gen_rtx (IF_THEN_ELSE, DImode, |
| gen_rtx (compare_code, DImode, |
| op0, const0_rtx), |
| const1_rtx, |
| operands[0]))); |
| return 1; |
| } |
| |
| emit_insn (gen_rtx (SET, VOIDmode, operands[0], const0_rtx)); |
| if (GET_MODE (op0) != DImode) |
| { |
| temp = gen_reg_rtx (DImode); |
| convert_move (temp, op0, 0); |
| } |
| else |
| temp = op0; |
| emit_insn (gen_rtx (SET, VOIDmode, operands[0], |
| gen_rtx (IF_THEN_ELSE, GET_MODE (operands[0]), |
| gen_rtx (compare_code, DImode, |
| temp, const0_rtx), |
| const1_rtx, |
| operands[0]))); |
| return 1; |
| } |
| else |
| { |
| operands[1] = gen_compare_reg (compare_code, op0, op1); |
| |
| switch (GET_MODE (operands[1])) |
| { |
| case CCmode : |
| case CCXmode : |
| case CCFPEmode : |
| case CCFPmode : |
| break; |
| default : |
| abort (); |
| } |
| emit_insn (gen_rtx (SET, VOIDmode, operands[0], const0_rtx)); |
| emit_insn (gen_rtx (SET, VOIDmode, operands[0], |
| gen_rtx (IF_THEN_ELSE, GET_MODE (operands[0]), |
| gen_rtx (compare_code, |
| GET_MODE (operands[1]), |
| operands[1], const0_rtx), |
| const1_rtx, operands[0]))); |
| return 1; |
| } |
| } |
| |
| /* Emit a conditional jump insn for the v9 architecture using comparison code |
| CODE and jump target LABEL. |
| This function exists to take advantage of the v9 brxx insns. */ |
| |
| void |
| emit_v9_brxx_insn (code, op0, label) |
| enum rtx_code code; |
| rtx op0, label; |
| { |
| emit_jump_insn (gen_rtx (SET, VOIDmode, |
| pc_rtx, |
| gen_rtx (IF_THEN_ELSE, VOIDmode, |
| gen_rtx (code, GET_MODE (op0), |
| op0, const0_rtx), |
| gen_rtx (LABEL_REF, VOIDmode, label), |
| pc_rtx))); |
| } |
| |
| /* Return nonzero if a return peephole merging return with |
| setting of output register is ok. */ |
| int |
| leaf_return_peephole_ok () |
| { |
| return (actual_fsize == 0); |
| } |
| |
| /* Return nonzero if TRIAL can go into the function epilogue's |
| delay slot. SLOT is the slot we are trying to fill. */ |
| |
| int |
| eligible_for_epilogue_delay (trial, slot) |
| rtx trial; |
| int slot; |
| { |
| rtx pat, src; |
| |
| if (slot >= 1) |
| return 0; |
| |
| if (GET_CODE (trial) != INSN || GET_CODE (PATTERN (trial)) != SET) |
| return 0; |
| |
| if (get_attr_length (trial) != 1) |
| return 0; |
| |
| /* If %g0 is live, there are lots of things we can't handle. |
| Rather than trying to find them all now, let's punt and only |
| optimize things as necessary. */ |
| if (TARGET_LIVE_G0) |
| return 0; |
| |
| /* In the case of a true leaf function, anything can go into the delay slot. |
| A delay slot only exists however if the frame size is zero, otherwise |
| we will put an insn to adjust the stack after the return. */ |
| if (leaf_function) |
| { |
| if (leaf_return_peephole_ok ()) |
| return ((get_attr_in_uncond_branch_delay (trial) |
| == IN_BRANCH_DELAY_TRUE)); |
| return 0; |
| } |
| |
| /* If only trivial `restore' insns work, nothing can go in the |
| delay slot. */ |
| else if (TARGET_BROKEN_SAVERESTORE) |
| return 0; |
| |
| pat = PATTERN (trial); |
| |
| /* Otherwise, only operations which can be done in tandem with |
| a `restore' insn can go into the delay slot. */ |
| if (GET_CODE (SET_DEST (pat)) != REG |
| || REGNO (SET_DEST (pat)) >= 32 |
| || REGNO (SET_DEST (pat)) < 24) |
| return 0; |
| |
| /* The set of insns matched here must agree precisely with the set of |
| patterns paired with a RETURN in sparc.md. */ |
| |
| src = SET_SRC (pat); |
| |
| /* This matches "*return_[qhs]". */ |
| if (arith_operand (src, GET_MODE (src))) |
| return GET_MODE_SIZE (GET_MODE (src)) <= GET_MODE_SIZE (SImode); |
| |
| /* This matches "*return_di". */ |
| else if (arith_double_operand (src, GET_MODE (src))) |
| return GET_MODE_SIZE (GET_MODE (src)) <= GET_MODE_SIZE (DImode); |
| |
| /* This matches "*return_sf_no_fpu". */ |
| else if (! TARGET_FPU && restore_operand (SET_DEST (pat), SFmode) |
| && register_operand (src, SFmode)) |
| return 1; |
| |
| /* This matches "*return_addsi". */ |
| else if (GET_CODE (src) == PLUS |
| && arith_operand (XEXP (src, 0), SImode) |
| && arith_operand (XEXP (src, 1), SImode) |
| && (register_operand (XEXP (src, 0), SImode) |
| || register_operand (XEXP (src, 1), SImode))) |
| return 1; |
| |
| /* This matches "*return_adddi". */ |
| else if (GET_CODE (src) == PLUS |
| && arith_double_operand (XEXP (src, 0), DImode) |
| && arith_double_operand (XEXP (src, 1), DImode) |
| && (register_operand (XEXP (src, 0), DImode) |
| || register_operand (XEXP (src, 1), DImode))) |
| return 1; |
| |
| /* This matches "*return_subsi". */ |
| else if (GET_CODE (src) == MINUS |
| && register_operand (XEXP (src, 0), SImode) |
| && small_int (XEXP (src, 1), VOIDmode) |
| && INTVAL (XEXP (src, 1)) != -4096) |
| return 1; |
| |
| return 0; |
| } |
| |
| int |
| short_branch (uid1, uid2) |
| int uid1, uid2; |
| { |
| unsigned int delta = insn_addresses[uid1] - insn_addresses[uid2]; |
| if (delta + 1024 < 2048) |
| return 1; |
| /* warning ("long branch, distance %d", delta); */ |
| return 0; |
| } |
| |
| /* Return non-zero if REG is not used after INSN. |
| We assume REG is a reload reg, and therefore does |
| not live past labels or calls or jumps. */ |
| int |
| reg_unused_after (reg, insn) |
| rtx reg; |
| rtx insn; |
| { |
| enum rtx_code code, prev_code = UNKNOWN; |
| |
| while (insn = NEXT_INSN (insn)) |
| { |
| if (prev_code == CALL_INSN && call_used_regs[REGNO (reg)]) |
| return 1; |
| |
| code = GET_CODE (insn); |
| if (GET_CODE (insn) == CODE_LABEL) |
| return 1; |
| |
| if (GET_RTX_CLASS (code) == 'i') |
| { |
| rtx set = single_set (insn); |
| int in_src = set && reg_overlap_mentioned_p (reg, SET_SRC (set)); |
| if (set && in_src) |
| return 0; |
| if (set && reg_overlap_mentioned_p (reg, SET_DEST (set))) |
| return 1; |
| if (set == 0 && reg_overlap_mentioned_p (reg, PATTERN (insn))) |
| return 0; |
| } |
| prev_code = code; |
| } |
| return 1; |
| } |
| |
| /* The rtx for the global offset table which is a special form |
| that *is* a position independent symbolic constant. */ |
| static rtx pic_pc_rtx; |
| |
| /* Ensure that we are not using patterns that are not OK with PIC. */ |
| |
| int |
| check_pic (i) |
| int i; |
| { |
| switch (flag_pic) |
| { |
| case 1: |
| if (GET_CODE (recog_operand[i]) == SYMBOL_REF |
| || (GET_CODE (recog_operand[i]) == CONST |
| && ! rtx_equal_p (pic_pc_rtx, recog_operand[i]))) |
| abort (); |
| case 2: |
| default: |
| return 1; |
| } |
| } |
| |
| /* Return true if X is an address which needs a temporary register when |
| reloaded while generating PIC code. */ |
| |
| int |
| pic_address_needs_scratch (x) |
| rtx x; |
| { |
| /* An address which is a symbolic plus a non SMALL_INT needs a temp reg. */ |
| if (GET_CODE (x) == CONST && GET_CODE (XEXP (x, 0)) == PLUS |
| && GET_CODE (XEXP (XEXP (x, 0), 0)) == SYMBOL_REF |
| && GET_CODE (XEXP (XEXP (x, 0), 1)) == CONST_INT |
| && ! SMALL_INT (XEXP (XEXP (x, 0), 1))) |
| return 1; |
| |
| return 0; |
| } |
| |
| /* Legitimize PIC addresses. If the address is already position-independent, |
| we return ORIG. Newly generated position-independent addresses go into a |
| reg. This is REG if non zero, otherwise we allocate register(s) as |
| necessary. */ |
| |
| rtx |
| legitimize_pic_address (orig, mode, reg) |
| rtx orig; |
| enum machine_mode mode; |
| rtx reg; |
| { |
| if (GET_CODE (orig) == SYMBOL_REF) |
| { |
| rtx pic_ref, address; |
| rtx insn; |
| |
| if (reg == 0) |
| { |
| if (reload_in_progress || reload_completed) |
| abort (); |
| else |
| reg = gen_reg_rtx (Pmode); |
| } |
| |
| if (flag_pic == 2) |
| { |
| /* If not during reload, allocate another temp reg here for loading |
| in the address, so that these instructions can be optimized |
| properly. */ |
| rtx temp_reg = ((reload_in_progress || reload_completed) |
| ? reg : gen_reg_rtx (Pmode)); |
| |
| /* Must put the SYMBOL_REF inside an UNSPEC here so that cse |
| won't get confused into thinking that these two instructions |
| are loading in the true address of the symbol. If in the |
| future a PIC rtx exists, that should be used instead. */ |
| emit_insn (gen_pic_sethi_si (temp_reg, orig)); |
| emit_insn (gen_pic_lo_sum_si (temp_reg, temp_reg, orig)); |
| |
| address = temp_reg; |
| } |
| else |
| address = orig; |
| |
| pic_ref = gen_rtx (MEM, Pmode, |
| gen_rtx (PLUS, Pmode, |
| pic_offset_table_rtx, address)); |
| current_function_uses_pic_offset_table = 1; |
| RTX_UNCHANGING_P (pic_ref) = 1; |
| insn = emit_move_insn (reg, pic_ref); |
| /* Put a REG_EQUAL note on this insn, so that it can be optimized |
| by loop. */ |
| REG_NOTES (insn) = gen_rtx (EXPR_LIST, REG_EQUAL, orig, |
| REG_NOTES (insn)); |
| return reg; |
| } |
| else if (GET_CODE (orig) == CONST) |
| { |
| rtx base, offset; |
| |
| if (GET_CODE (XEXP (orig, 0)) == PLUS |
| && XEXP (XEXP (orig, 0), 0) == pic_offset_table_rtx) |
| return orig; |
| |
| if (reg == 0) |
| { |
| if (reload_in_progress || reload_completed) |
| abort (); |
| else |
| reg = gen_reg_rtx (Pmode); |
| } |
| |
| if (GET_CODE (XEXP (orig, 0)) == PLUS) |
| { |
| base = legitimize_pic_address (XEXP (XEXP (orig, 0), 0), Pmode, reg); |
| offset = legitimize_pic_address (XEXP (XEXP (orig, 0), 1), Pmode, |
| base == reg ? 0 : reg); |
| } |
| else |
| abort (); |
| |
| if (GET_CODE (offset) == CONST_INT) |
| { |
| if (SMALL_INT (offset)) |
| return plus_constant_for_output (base, INTVAL (offset)); |
| else if (! reload_in_progress && ! reload_completed) |
| offset = force_reg (Pmode, offset); |
| else |
| /* If we reach here, then something is seriously wrong. */ |
| abort (); |
| } |
| return gen_rtx (PLUS, Pmode, base, offset); |
| } |
| else if (GET_CODE (orig) == LABEL_REF) |
| /* ??? Why do we do this? */ |
| current_function_uses_pic_offset_table = 1; |
| |
| return orig; |
| } |
| |
| /* Set up PIC-specific rtl. This should not cause any insns |
| to be emitted. */ |
| |
| void |
| initialize_pic () |
| { |
| } |
| |
| /* Emit special PIC prologues and epilogues. */ |
| |
| void |
| finalize_pic () |
| { |
| /* The table we use to reference PIC data. */ |
| rtx global_offset_table; |
| /* Labels to get the PC in the prologue of this function. */ |
| rtx l1, l2; |
| rtx seq; |
| int orig_flag_pic = flag_pic; |
| |
| if (current_function_uses_pic_offset_table == 0) |
| return; |
| |
| if (! flag_pic) |
| abort (); |
| |
| flag_pic = 0; |
| |
| /* ??? sparc64 pic currently under construction. */ |
| |
| start_sequence (); |
| |
| l1 = gen_label_rtx (); |
| |
| /* Initialize every time through, since we can't easily |
| know this to be permanent. */ |
| global_offset_table = gen_rtx (SYMBOL_REF, Pmode, "_GLOBAL_OFFSET_TABLE_"); |
| pic_pc_rtx = gen_rtx (CONST, Pmode, |
| gen_rtx (MINUS, Pmode, |
| global_offset_table, |
| gen_rtx (CONST, Pmode, |
| gen_rtx (MINUS, Pmode, |
| gen_rtx (LABEL_REF, VOIDmode, l1), |
| pc_rtx)))); |
| |
| if (! TARGET_ARCH64) |
| { |
| l2 = gen_label_rtx (); |
| emit_label (l1); |
| /* Note that we pun calls and jumps here! */ |
| emit_jump_insn (gen_get_pc_sp32 (l2)); |
| emit_label (l2); |
| |
| emit_insn (gen_rtx (SET, VOIDmode, pic_offset_table_rtx, |
| gen_rtx (HIGH, Pmode, pic_pc_rtx))); |
| |
| emit_insn (gen_rtx (SET, VOIDmode, |
| pic_offset_table_rtx, |
| gen_rtx (LO_SUM, Pmode, |
| pic_offset_table_rtx, pic_pc_rtx))); |
| emit_insn (gen_rtx (SET, VOIDmode, |
| pic_offset_table_rtx, |
| gen_rtx (PLUS, Pmode, |
| pic_offset_table_rtx, |
| gen_rtx (REG, Pmode, 15)))); |
| |
| /* emit_insn (gen_rtx (ASM_INPUT, VOIDmode, "!#PROLOGUE# 1")); */ |
| LABEL_PRESERVE_P (l1) = 1; |
| LABEL_PRESERVE_P (l2) = 1; |
| } |
| else |
| { |
| /* ??? This definately isn't right for -mfullany. */ |
| /* ??? And it doesn't quite seem right for the others either. */ |
| emit_label (l1); |
| emit_insn (gen_get_pc_sp64 (gen_rtx (REG, Pmode, 1))); |
| |
| /* Don't let the scheduler separate the previous insn from `l1'. */ |
| emit_insn (gen_blockage ()); |
| |
| emit_insn (gen_rtx (SET, VOIDmode, pic_offset_table_rtx, |
| gen_rtx (HIGH, Pmode, pic_pc_rtx))); |
| |
| emit_insn (gen_rtx (SET, VOIDmode, |
| pic_offset_table_rtx, |
| gen_rtx (LO_SUM, Pmode, |
| pic_offset_table_rtx, pic_pc_rtx))); |
| emit_insn (gen_rtx (SET, VOIDmode, |
| pic_offset_table_rtx, |
| gen_rtx (PLUS, Pmode, |
| pic_offset_table_rtx, gen_rtx (REG, Pmode, 1)))); |
| |
| /* emit_insn (gen_rtx (ASM_INPUT, VOIDmode, "!#PROLOGUE# 1")); */ |
| LABEL_PRESERVE_P (l1) = 1; |
| } |
| |
| flag_pic = orig_flag_pic; |
| |
| seq = gen_sequence (); |
| end_sequence (); |
| emit_insn_after (seq, get_insns ()); |
| |
| /* Need to emit this whether or not we obey regdecls, |
| since setjmp/longjmp can cause life info to screw up. */ |
| emit_insn (gen_rtx (USE, VOIDmode, pic_offset_table_rtx)); |
| } |
| |
| /* Emit insns to move operands[1] into operands[0]. |
| |
| Return 1 if we have written out everything that needs to be done to |
| do the move. Otherwise, return 0 and the caller will emit the move |
| normally. */ |
| |
| int |
| emit_move_sequence (operands, mode) |
| rtx *operands; |
| enum machine_mode mode; |
| { |
| register rtx operand0 = operands[0]; |
| register rtx operand1 = operands[1]; |
| |
| if (CONSTANT_P (operand1) && flag_pic |
| && pic_address_needs_scratch (operand1)) |
| operands[1] = operand1 = legitimize_pic_address (operand1, mode, 0); |
| |
| /* Handle most common case first: storing into a register. */ |
| if (register_operand (operand0, mode)) |
| { |
| if (register_operand (operand1, mode) |
| || (GET_CODE (operand1) == CONST_INT && SMALL_INT (operand1)) |
| || (GET_CODE (operand1) == CONST_DOUBLE |
| && arith_double_operand (operand1, DImode)) |
| || (GET_CODE (operand1) == HIGH && GET_MODE (operand1) != DImode) |
| /* Only `general_operands' can come here, so MEM is ok. */ |
| || GET_CODE (operand1) == MEM) |
| { |
| /* Run this case quickly. */ |
| emit_insn (gen_rtx (SET, VOIDmode, operand0, operand1)); |
| return 1; |
| } |
| } |
| else if (GET_CODE (operand0) == MEM) |
| { |
| if (register_operand (operand1, mode) |
| || (operand1 == const0_rtx && ! TARGET_LIVE_G0)) |
| { |
| /* Run this case quickly. */ |
| emit_insn (gen_rtx (SET, VOIDmode, operand0, operand1)); |
| return 1; |
| } |
| if (! reload_in_progress) |
| { |
| operands[0] = validize_mem (operand0); |
| operands[1] = operand1 = force_reg (mode, operand1); |
| } |
| } |
| |
| if (GET_CODE (operand1) == LABEL_REF |
| && mode == SImode && flag_pic) |
| { |
| if (TARGET_ARCH64) |
| abort (); |
| emit_insn (gen_move_pic_label_si (operand0, operand1)); |
| return 1; |
| } |
| /* Non-pic LABEL_REF's in sparc64 are expensive to do the normal way, |
| so always use special code. */ |
| else if (GET_CODE (operand1) == LABEL_REF |
| && mode == DImode) |
| { |
| if (! TARGET_ARCH64) |
| abort (); |
| emit_insn (gen_move_label_di (operand0, operand1)); |
| return 1; |
| } |
| /* DImode HIGH values in sparc64 need a clobber added. */ |
| else if (TARGET_ARCH64 |
| && GET_CODE (operand1) == HIGH && GET_MODE (operand1) == DImode) |
| { |
| emit_insn (gen_sethi_di_sp64 (operand0, XEXP (operand1, 0))); |
| return 1; |
| } |
| /* Simplify the source if we need to. */ |
| else if (GET_CODE (operand1) != HIGH && immediate_operand (operand1, mode)) |
| { |
| if (flag_pic && symbolic_operand (operand1, mode)) |
| { |
| rtx temp_reg = reload_in_progress ? operand0 : 0; |
| |
| operands[1] = legitimize_pic_address (operand1, mode, temp_reg); |
| } |
| else if (GET_CODE (operand1) == CONST_INT |
| ? (! SMALL_INT (operand1) |
| && (INTVAL (operand1) & 0x3ff) != 0) |
| : (GET_CODE (operand1) == CONST_DOUBLE |
| ? ! arith_double_operand (operand1, DImode) |
| : 1)) |
| { |
| /* For DImode values, temp must be operand0 because of the way |
| HI and LO_SUM work. The LO_SUM operator only copies half of |
| the LSW from the dest of the HI operator. If the LO_SUM dest is |
| not the same as the HI dest, then the MSW of the LO_SUM dest will |
| never be set. |
| |
| ??? The real problem here is that the ...(HI:DImode pattern emits |
| multiple instructions, and the ...(LO_SUM:DImode pattern emits |
| one instruction. This fails, because the compiler assumes that |
| LO_SUM copies all bits of the first operand to its dest. Better |
| would be to have the HI pattern emit one instruction and the |
| LO_SUM pattern multiple instructions. Even better would be |
| to use four rtl insns. */ |
| rtx temp = ((reload_in_progress || mode == DImode) |
| ? operand0 : gen_reg_rtx (mode)); |
| |
| if (TARGET_ARCH64 && mode == DImode) |
| emit_insn (gen_sethi_di_sp64 (temp, operand1)); |
| else |
| emit_insn (gen_rtx (SET, VOIDmode, temp, |
| gen_rtx (HIGH, mode, operand1))); |
| |
| operands[1] = gen_rtx (LO_SUM, mode, temp, operand1); |
| } |
| } |
| |
| /* Now have insn-emit do whatever it normally does. */ |
| return 0; |
| } |
| |
| /* Return the best assembler insn template |
| for moving operands[1] into operands[0] as a fullword. */ |
| |
| char * |
| singlemove_string (operands) |
| rtx *operands; |
| { |
| if (GET_CODE (operands[0]) == MEM) |
| { |
| if (GET_CODE (operands[1]) != MEM) |
| return "st %r1,%0"; |
| else |
| abort (); |
| } |
| else if (GET_CODE (operands[1]) == MEM) |
| return "ld %1,%0"; |
| else if (GET_CODE (operands[1]) == CONST_DOUBLE) |
| { |
| REAL_VALUE_TYPE r; |
| long i; |
| |
| /* Must be SFmode, otherwise this doesn't make sense. */ |
| if (GET_MODE (operands[1]) != SFmode) |
| abort (); |
| |
| REAL_VALUE_FROM_CONST_DOUBLE (r, operands[1]); |
| REAL_VALUE_TO_TARGET_SINGLE (r, i); |
| operands[1] = gen_rtx (CONST_INT, VOIDmode, i); |
| |
| if (CONST_OK_FOR_LETTER_P (i, 'I')) |
| return "mov %1,%0"; |
| else if ((i & 0x000003FF) != 0) |
| return "sethi %%hi(%a1),%0\n\tor %0,%%lo(%a1),%0"; |
| else |
| return "sethi %%hi(%a1),%0"; |
| } |
| else if (GET_CODE (operands[1]) == CONST_INT |
| && ! CONST_OK_FOR_LETTER_P (INTVAL (operands[1]), 'I')) |
| { |
| int i = INTVAL (operands[1]); |
| |
| /* If all low order 10 bits are clear, then we only need a single |
| sethi insn to load the constant. */ |
| if ((i & 0x000003FF) != 0) |
| return "sethi %%hi(%a1),%0\n\tor %0,%%lo(%a1),%0"; |
| else |
| return "sethi %%hi(%a1),%0"; |
| } |
| /* Operand 1 must be a register, or a 'I' type CONST_INT. */ |
| return "mov %1,%0"; |
| } |
| |
| /* Return non-zero if it is OK to assume that the given memory operand is |
| aligned at least to a 8-byte boundary. This should only be called |
| for memory accesses whose size is 8 bytes or larger. */ |
| |
| int |
| mem_aligned_8 (mem) |
| register rtx mem; |
| { |
| register rtx addr; |
| register rtx base; |
| register rtx offset; |
| |
| if (GET_CODE (mem) != MEM) |
| return 0; /* It's gotta be a MEM! */ |
| |
| addr = XEXP (mem, 0); |
| |
| /* Now that all misaligned double parms are copied on function entry, |
| we can assume any 64-bit object is 64-bit aligned except those which |
| are at unaligned offsets from the stack or frame pointer. If the |
| TARGET_UNALIGNED_DOUBLES switch is given, we do not make this |
| assumption. */ |
| |
| /* See what register we use in the address. */ |
| base = 0; |
| if (GET_CODE (addr) == PLUS) |
| { |
| if (GET_CODE (XEXP (addr, 0)) == REG |
| && GET_CODE (XEXP (addr, 1)) == CONST_INT) |
| { |
| base = XEXP (addr, 0); |
| offset = XEXP (addr, 1); |
| } |
| } |
| else if (GET_CODE (addr) == REG) |
| { |
| base = addr; |
| offset = const0_rtx; |
| } |
| |
| /* If it's the stack or frame pointer, check offset alignment. |
| We can have improper alignment in the function entry code. */ |
| if (base |
| && (REGNO (base) == FRAME_POINTER_REGNUM |
| || REGNO (base) == STACK_POINTER_REGNUM)) |
| { |
| if (((INTVAL (offset) - SPARC_STACK_BIAS) & 0x7) == 0) |
| return 1; |
| } |
| /* Anything else we know is properly aligned unless TARGET_UNALIGNED_DOUBLES |
| is true, in which case we can only assume that an access is aligned if |
| it is to a constant address, or the address involves a LO_SUM. |
| |
| We used to assume an address was aligned if MEM_IN_STRUCT_P was true. |
| That assumption was deleted so that gcc generated code can be used with |
| memory allocators that only guarantee 4 byte alignment. */ |
| else if (! TARGET_UNALIGNED_DOUBLES || CONSTANT_P (addr) |
| || GET_CODE (addr) == LO_SUM) |
| return 1; |
| |
| /* An obviously unaligned address. */ |
| return 0; |
| } |
| |
| enum optype { REGOP, OFFSOP, MEMOP, PUSHOP, POPOP, CNSTOP, RNDOP }; |
| |
| /* Output assembler code to perform a doubleword move insn |
| with operands OPERANDS. This is very similar to the following |
| output_move_quad function. */ |
| |
| char * |
| output_move_double (operands) |
| rtx *operands; |
| { |
| register rtx op0 = operands[0]; |
| register rtx op1 = operands[1]; |
| register enum optype optype0; |
| register enum optype optype1; |
| rtx latehalf[2]; |
| rtx addreg0 = 0; |
| rtx addreg1 = 0; |
| int highest_first = 0; |
| int no_addreg1_decrement = 0; |
| |
| /* First classify both operands. */ |
| |
| if (REG_P (op0)) |
| optype0 = REGOP; |
| else if (offsettable_memref_p (op0)) |
| optype0 = OFFSOP; |
| else if (GET_CODE (op0) == MEM) |
| optype0 = MEMOP; |
| else |
| optype0 = RNDOP; |
| |
| if (REG_P (op1)) |
| optype1 = REGOP; |
| else if (CONSTANT_P (op1)) |
| optype1 = CNSTOP; |
| else if (offsettable_memref_p (op1)) |
| optype1 = OFFSOP; |
| else if (GET_CODE (op1) == MEM) |
| optype1 = MEMOP; |
| else |
| optype1 = RNDOP; |
| |
| /* Check for the cases that the operand constraints are not |
| supposed to allow to happen. Abort if we get one, |
| because generating code for these cases is painful. */ |
| |
| if (optype0 == RNDOP || optype1 == RNDOP |
| || (optype0 == MEM && optype1 == MEM)) |
| abort (); |
| |
| /* If an operand is an unoffsettable memory ref, find a register |
| we can increment temporarily to make it refer to the second word. */ |
| |
| if (optype0 == MEMOP) |
| addreg0 = find_addr_reg (XEXP (op0, 0)); |
| |
| if (optype1 == MEMOP) |
| addreg1 = find_addr_reg (XEXP (op1, 0)); |
| |
| /* Ok, we can do one word at a time. |
| Set up in LATEHALF the operands to use for the |
| high-numbered (least significant) word and in some cases alter the |
| operands in OPERANDS to be suitable for the low-numbered word. */ |
| |
| if (optype0 == REGOP) |
| latehalf[0] = gen_rtx (REG, SImode, REGNO (op0) + 1); |
| else if (optype0 == OFFSOP) |
| latehalf[0] = adj_offsettable_operand (op0, 4); |
| else |
| latehalf[0] = op0; |
| |
| if (optype1 == REGOP) |
| latehalf[1] = gen_rtx (REG, SImode, REGNO (op1) + 1); |
| else if (optype1 == OFFSOP) |
| latehalf[1] = adj_offsettable_operand (op1, 4); |
| else if (optype1 == CNSTOP) |
| { |
| if (TARGET_ARCH64) |
| { |
| if (arith_double_operand (op1, DImode)) |
| { |
| operands[1] = gen_rtx (CONST_INT, VOIDmode, |
| CONST_DOUBLE_LOW (op1)); |
| return "mov %1,%0"; |
| } |
| else |
| { |
| /* The only way to handle CONST_DOUBLEs or other 64 bit |
| constants here is to use a temporary, such as is done |
| for the V9 DImode sethi insn pattern. This is not |
| a practical solution, so abort if we reach here. |
| The md file should always force such constants to |
| memory. */ |
| abort (); |
| } |
| } |
| else |
| split_double (op1, &operands[1], &latehalf[1]); |
| } |
| else |
| latehalf[1] = op1; |
| |
| /* Easy case: try moving both words at once. Check for moving between |
| an even/odd register pair and a memory location. */ |
| if ((optype0 == REGOP && optype1 != REGOP && optype1 != CNSTOP |
| && (TARGET_ARCH64 || (REGNO (op0) & 1) == 0)) |
| || (optype0 != REGOP && optype0 != CNSTOP && optype1 == REGOP |
| && (TARGET_ARCH64 || (REGNO (op1) & 1) == 0))) |
| { |
| register rtx mem,reg; |
| |
| if (optype0 == REGOP) |
| mem = op1, reg = op0; |
| else |
| mem = op0, reg = op1; |
| |
| /* In v9, ldd can be used for word aligned addresses, so technically |
| some of this logic is unneeded. We still avoid ldd if the address |
| is obviously unaligned though. */ |
| |
| if (mem_aligned_8 (mem) |
| /* If this is a floating point register higher than %f31, |
| then we *must* use an aligned load, since `ld' will not accept |
| the register number. */ |
| || (TARGET_V9 && REGNO (reg) >= 64)) |
| { |
| if (FP_REG_P (reg) || ! TARGET_ARCH64) |
| return (mem == op1 ? "ldd %1,%0" : "std %1,%0"); |
| else |
| return (mem == op1 ? "ldx %1,%0" : "stx %1,%0"); |
| } |
| } |
| |
| if (TARGET_ARCH64) |
| { |
| if (optype0 == REGOP && optype1 == REGOP) |
| { |
| if (FP_REG_P (op0)) |
| return "fmovd %1,%0"; |
| else |
| return "mov %1,%0"; |
| } |
| } |
| |
| /* If the first move would clobber the source of the second one, |
| do them in the other order. */ |
| |
| /* Overlapping registers. */ |
| if (optype0 == REGOP && optype1 == REGOP |
| && REGNO (op0) == REGNO (latehalf[1])) |
| { |
| /* Do that word. */ |
| output_asm_insn (singlemove_string (latehalf), latehalf); |
| /* Do low-numbered word. */ |
| return singlemove_string (operands); |
| } |
| /* Loading into a register which overlaps a register used in the address. */ |
| else if (optype0 == REGOP && optype1 != REGOP |
| && reg_overlap_mentioned_p (op0, op1)) |
| { |
| /* If both halves of dest are used in the src memory address, |
| add the two regs and put them in the low reg (op0). |
| Then it works to load latehalf first. */ |
| if (reg_mentioned_p (op0, XEXP (op1, 0)) |
| && reg_mentioned_p (latehalf[0], XEXP (op1, 0))) |
| { |
| rtx xops[2]; |
| xops[0] = latehalf[0]; |
| xops[1] = op0; |
| output_asm_insn ("add %1,%0,%1", xops); |
| operands[1] = gen_rtx (MEM, DImode, op0); |
| latehalf[1] = adj_offsettable_operand (operands[1], 4); |
| addreg1 = 0; |
| highest_first = 1; |
| } |
| /* Only one register in the dest is used in the src memory address, |
| and this is the first register of the dest, so we want to do |
| the late half first here also. */ |
| else if (! reg_mentioned_p (latehalf[0], XEXP (op1, 0))) |
| highest_first = 1; |
| /* Only one register in the dest is used in the src memory address, |
| and this is the second register of the dest, so we want to do |
| the late half last. If addreg1 is set, and addreg1 is the same |
| register as latehalf, then we must suppress the trailing decrement, |
| because it would clobber the value just loaded. */ |
| else if (addreg1 && reg_mentioned_p (addreg1, latehalf[0])) |
| no_addreg1_decrement = 1; |
| } |
| |
| /* Normal case: do the two words, low-numbered first. |
| Overlap case (highest_first set): do high-numbered word first. */ |
| |
| if (! highest_first) |
| output_asm_insn (singlemove_string (operands), operands); |
| |
| /* Make any unoffsettable addresses point at high-numbered word. */ |
| if (addreg0) |
| output_asm_insn ("add %0,0x4,%0", &addreg0); |
| if (addreg1) |
| output_asm_insn ("add %0,0x4,%0", &addreg1); |
| |
| /* Do that word. */ |
| output_asm_insn (singlemove_string (latehalf), latehalf); |
| |
| /* Undo the adds we just did. */ |
| if (addreg0) |
| output_asm_insn ("add %0,-0x4,%0", &addreg0); |
| if (addreg1 && ! no_addreg1_decrement) |
| output_asm_insn ("add %0,-0x4,%0", &addreg1); |
| |
| if (highest_first) |
| output_asm_insn (singlemove_string (operands), operands); |
| |
| return ""; |
| } |
| |
| /* Output assembler code to perform a quadword move insn |
| with operands OPERANDS. This is very similar to the preceding |
| output_move_double function. */ |
| |
| char * |
| output_move_quad (operands) |
| rtx *operands; |
| { |
| register rtx op0 = operands[0]; |
| register rtx op1 = operands[1]; |
| register enum optype optype0; |
| register enum optype optype1; |
| rtx wordpart[4][2]; |
| rtx addreg0 = 0; |
| rtx addreg1 = 0; |
| |
| /* First classify both operands. */ |
| |
| if (REG_P (op0)) |
| optype0 = REGOP; |
| else if (offsettable_memref_p (op0)) |
| optype0 = OFFSOP; |
| else if (GET_CODE (op0) == MEM) |
| optype0 = MEMOP; |
| else |
| optype0 = RNDOP; |
| |
| if (REG_P (op1)) |
| optype1 = REGOP; |
| else if (CONSTANT_P (op1)) |
| optype1 = CNSTOP; |
| else if (offsettable_memref_p (op1)) |
| optype1 = OFFSOP; |
| else if (GET_CODE (op1) == MEM) |
| optype1 = MEMOP; |
| else |
| optype1 = RNDOP; |
| |
| /* Check for the cases that the operand constraints are not |
| supposed to allow to happen. Abort if we get one, |
| because generating code for these cases is painful. */ |
| |
| if (optype0 == RNDOP || optype1 == RNDOP |
| || (optype0 == MEM && optype1 == MEM)) |
| abort (); |
| |
| /* If an operand is an unoffsettable memory ref, find a register |
| we can increment temporarily to make it refer to the later words. */ |
| |
| if (optype0 == MEMOP) |
| addreg0 = find_addr_reg (XEXP (op0, 0)); |
| |
| if (optype1 == MEMOP) |
| addreg1 = find_addr_reg (XEXP (op1, 0)); |
| |
| /* Ok, we can do one word at a time. |
| Set up in wordpart the operands to use for each word of the arguments. */ |
| |
| if (optype0 == REGOP) |
| { |
| wordpart[0][0] = gen_rtx (REG, SImode, REGNO (op0) + 0); |
| wordpart[1][0] = gen_rtx (REG, SImode, REGNO (op0) + 1); |
| wordpart[2][0] = gen_rtx (REG, SImode, REGNO (op0) + 2); |
| wordpart[3][0] = gen_rtx (REG, SImode, REGNO (op0) + 3); |
| } |
| else if (optype0 == OFFSOP) |
| { |
| wordpart[0][0] = adj_offsettable_operand (op0, 0); |
| wordpart[1][0] = adj_offsettable_operand (op0, 4); |
| wordpart[2][0] = adj_offsettable_operand (op0, 8); |
| wordpart[3][0] = adj_offsettable_operand (op0, 12); |
| } |
| else |
| { |
| wordpart[0][0] = op0; |
| wordpart[1][0] = op0; |
| wordpart[2][0] = op0; |
| wordpart[3][0] = op0; |
| } |
| |
| if (optype1 == REGOP) |
| { |
| wordpart[0][1] = gen_rtx (REG, SImode, REGNO (op1) + 0); |
| wordpart[1][1] = gen_rtx (REG, SImode, REGNO (op1) + 1); |
| wordpart[2][1] = gen_rtx (REG, SImode, REGNO (op1) + 2); |
| wordpart[3][1] = gen_rtx (REG, SImode, REGNO (op1) + 3); |
| } |
| else if (optype1 == OFFSOP) |
| { |
| wordpart[0][1] = adj_offsettable_operand (op1, 0); |
| wordpart[1][1] = adj_offsettable_operand (op1, 4); |
| wordpart[2][1] = adj_offsettable_operand (op1, 8); |
| wordpart[3][1] = adj_offsettable_operand (op1, 12); |
| } |
| else if (optype1 == CNSTOP) |
| { |
| REAL_VALUE_TYPE r; |
| long l[4]; |
| |
| /* This only works for TFmode floating point constants. */ |
| if (GET_CODE (op1) != CONST_DOUBLE || GET_MODE (op1) != TFmode) |
| abort (); |
| |
| REAL_VALUE_FROM_CONST_DOUBLE (r, op1); |
| REAL_VALUE_TO_TARGET_LONG_DOUBLE (r, l); |
| |
| wordpart[0][1] = GEN_INT (l[0]); |
| wordpart[1][1] = GEN_INT (l[1]); |
| wordpart[2][1] = GEN_INT (l[2]); |
| wordpart[3][1] = GEN_INT (l[3]); |
| } |
| else |
| { |
| wordpart[0][1] = op1; |
| wordpart[1][1] = op1; |
| wordpart[2][1] = op1; |
| wordpart[3][1] = op1; |
| } |
| |
| /* Easy case: try moving the quad as two pairs. Check for moving between |
| an even/odd register pair and a memory location. |
| Also handle new v9 fp regs here. */ |
| /* ??? Should also handle the case of non-offsettable addresses here. |
| We can at least do the first pair as a ldd/std, and then do the third |
| and fourth words individually. */ |
| if ((optype0 == REGOP && optype1 == OFFSOP && (REGNO (op0) & 1) == 0) |
| || (optype0 == OFFSOP && optype1 == REGOP && (REGNO (op1) & 1) == 0)) |
| { |
| rtx mem, reg; |
| |
| if (optype0 == REGOP) |
| mem = op1, reg = op0; |
| else |
| mem = op0, reg = op1; |
| |
| if (mem_aligned_8 (mem) |
| /* If this is a floating point register higher than %f31, |
| then we *must* use an aligned load, since `ld' will not accept |
| the register number. */ |
| || (TARGET_V9 && REGNO (reg) >= 64)) |
| { |
| if (TARGET_V9 && FP_REG_P (reg) && TARGET_HARD_QUAD) |
| { |
| if ((REGNO (reg) & 3) != 0) |
| abort (); |
| return (mem == op1 ? "ldq %1,%0" : "stq %1,%0"); |
| } |
| operands[2] = adj_offsettable_operand (mem, 8); |
| /* ??? In arch64 case, shouldn't we use ldd/std for fp regs. */ |
| if (mem == op1) |
| return TARGET_ARCH64 ? "ldx %1,%0;ldx %2,%R0" : "ldd %1,%0;ldd %2,%S0"; |
| else |
| return TARGET_ARCH64 ? "stx %1,%0;stx %R1,%2" : "std %1,%0;std %S1,%2"; |
| } |
| } |
| |
| /* If the first move would clobber the source of the second one, |
| do them in the other order. */ |
| |
| /* Overlapping registers. */ |
| if (optype0 == REGOP && optype1 == REGOP |
| && (REGNO (op0) == REGNO (wordpart[1][3]) |
| || REGNO (op0) == REGNO (wordpart[1][2]) |
| || REGNO (op0) == REGNO (wordpart[1][1]))) |
| { |
| /* Do fourth word. */ |
| output_asm_insn (singlemove_string (wordpart[3]), wordpart[3]); |
| /* Do the third word. */ |
| output_asm_insn (singlemove_string (wordpart[2]), wordpart[2]); |
| /* Do the second word. */ |
| output_asm_insn (singlemove_string (wordpart[1]), wordpart[1]); |
| /* Do lowest-numbered word. */ |
| return singlemove_string (wordpart[0]); |
| } |
| /* Loading into a register which overlaps a register used in the address. */ |
| if (optype0 == REGOP && optype1 != REGOP |
| && reg_overlap_mentioned_p (op0, op1)) |
| { |
| /* ??? Not implemented yet. This is a bit complicated, because we |
| must load which ever part overlaps the address last. If the address |
| is a double-reg address, then there are two parts which need to |
| be done last, which is impossible. We would need a scratch register |
| in that case. */ |
| abort (); |
| } |
| |
| /* Normal case: move the four words in lowest to highest address order. */ |
| |
| output_asm_insn (singlemove_string (wordpart[0]), wordpart[0]); |
| |
| /* Make any unoffsettable addresses point at the second word. */ |
| if (addreg0) |
| output_asm_insn ("add %0,0x4,%0", &addreg0); |
| if (addreg1) |
| output_asm_insn ("add %0,0x4,%0", &addreg1); |
| |
| /* Do the second word. */ |
| output_asm_insn (singlemove_string (wordpart[1]), wordpart[1]); |
| |
| /* Make any unoffsettable addresses point at the third word. */ |
| if (addreg0) |
| output_asm_insn ("add %0,0x4,%0", &addreg0); |
| if (addreg1) |
| output_asm_insn ("add %0,0x4,%0", &addreg1); |
| |
| /* Do the third word. */ |
| output_asm_insn (singlemove_string (wordpart[2]), wordpart[2]); |
| |
| /* Make any unoffsettable addresses point at the fourth word. */ |
| if (addreg0) |
| output_asm_insn ("add %0,0x4,%0", &addreg0); |
| if (addreg1) |
| output_asm_insn ("add %0,0x4,%0", &addreg1); |
| |
| /* Do the fourth word. */ |
| output_asm_insn (singlemove_string (wordpart[3]), wordpart[3]); |
| |
| /* Undo the adds we just did. */ |
| if (addreg0) |
| output_asm_insn ("add %0,-0xc,%0", &addreg0); |
| if (addreg1) |
| output_asm_insn ("add %0,-0xc,%0", &addreg1); |
| |
| return ""; |
| } |
| |
| /* Output assembler code to perform a doubleword move insn with operands |
| OPERANDS, one of which must be a floating point register. */ |
| |
| char * |
| output_fp_move_double (operands) |
| rtx *operands; |
| { |
| if (FP_REG_P (operands[0])) |
| { |
| if (FP_REG_P (operands[1])) |
| { |
| if (TARGET_V9) |
| return "fmovd %1,%0"; |
| else |
| return "fmovs %1,%0\n\tfmovs %R1,%R0"; |
| } |
| else if (GET_CODE (operands[1]) == REG) |
| abort (); |
| else |
| return output_move_double (operands); |
| } |
| else if (FP_REG_P (operands[1])) |
| { |
| if (GET_CODE (operands[0]) == REG) |
| abort (); |
| else |
| return output_move_double (operands); |
| } |
| else abort (); |
| } |
| |
| /* Output assembler code to perform a quadword move insn with operands |
| OPERANDS, one of which must be a floating point register. */ |
| |
| char * |
| output_fp_move_quad (operands) |
| rtx *operands; |
| { |
| register rtx op0 = operands[0]; |
| register rtx op1 = operands[1]; |
| |
| if (FP_REG_P (op0)) |
| { |
| if (FP_REG_P (op1)) |
| { |
| if (TARGET_V9) |
| return "fmovq %1,%0"; |
| else |
| return "fmovs %1,%0\n\tfmovs %R1,%R0\n\tfmovs %S1,%S0\n\tfmovs %T1,%T0"; |
| } |
| else if (GET_CODE (op1) == REG) |
| abort (); |
| else |
| return output_move_quad (operands); |
| } |
| else if (FP_REG_P (op1)) |
| { |
| if (GET_CODE (op0) == REG) |
| abort (); |
| else |
| return output_move_quad (operands); |
| } |
| else |
| abort (); |
| } |
| |
| /* Return a REG that occurs in ADDR with coefficient 1. |
| ADDR can be effectively incremented by incrementing REG. */ |
| |
| static rtx |
| find_addr_reg (addr) |
| rtx addr; |
| { |
| while (GET_CODE (addr) == PLUS) |
| { |
| /* We absolutely can not fudge the frame pointer here, because the |
| frame pointer must always be 8 byte aligned. It also confuses |
| debuggers. */ |
| if (GET_CODE (XEXP (addr, 0)) == REG |
| && REGNO (XEXP (addr, 0)) != FRAME_POINTER_REGNUM) |
| addr = XEXP (addr, 0); |
| else if (GET_CODE (XEXP (addr, 1)) == REG |
| && REGNO (XEXP (addr, 1)) != FRAME_POINTER_REGNUM) |
| addr = XEXP (addr, 1); |
| else if (CONSTANT_P (XEXP (addr, 0))) |
| addr = XEXP (addr, 1); |
| else if (CONSTANT_P (XEXP (addr, 1))) |
| addr = XEXP (addr, 0); |
| else |
| abort (); |
| } |
| if (GET_CODE (addr) == REG) |
| return addr; |
| abort (); |
| } |
| |
| #if 0 /* not currently used */ |
| |
| void |
| output_sized_memop (opname, mode, signedp) |
| char *opname; |
| enum machine_mode mode; |
| int signedp; |
| { |
| static char *ld_size_suffix_u[] = { "ub", "uh", "", "?", "d" }; |
| static char *ld_size_suffix_s[] = { "sb", "sh", "", "?", "d" }; |
| static char *st_size_suffix[] = { "b", "h", "", "?", "d" }; |
| char **opnametab, *modename; |
| |
| if (opname[0] == 'l') |
| if (signedp) |
| opnametab = ld_size_suffix_s; |
| else |
| opnametab = ld_size_suffix_u; |
| else |
| opnametab = st_size_suffix; |
| modename = opnametab[GET_MODE_SIZE (mode) >> 1]; |
| |
| fprintf (asm_out_file, "\t%s%s", opname, modename); |
| } |
| |
| void |
| output_move_with_extension (operands) |
| rtx *operands; |
| { |
| if (GET_MODE (operands[2]) == HImode) |
| output_asm_insn ("sll %2,0x10,%0", operands); |
| else if (GET_MODE (operands[2]) == QImode) |
| output_asm_insn ("sll %2,0x18,%0", operands); |
| else |
| abort (); |
| } |
| #endif /* not currently used */ |
| |
| #if 0 |
| /* ??? These are only used by the movstrsi pattern, but we get better code |
| in general without that, because emit_block_move can do just as good a |
| job as this function does when alignment and size are known. When they |
| aren't known, a call to strcpy may be faster anyways, because it is |
| likely to be carefully crafted assembly language code, and below we just |
| do a byte-wise copy. |
| |
| Also, emit_block_move expands into multiple read/write RTL insns, which |
| can then be optimized, whereas our movstrsi pattern can not be optimized |
| at all. */ |
| |
| /* Load the address specified by OPERANDS[3] into the register |
| specified by OPERANDS[0]. |
| |
| OPERANDS[3] may be the result of a sum, hence it could either be: |
| |
| (1) CONST |
| (2) REG |
| (2) REG + CONST_INT |
| (3) REG + REG + CONST_INT |
| (4) REG + REG (special case of 3). |
| |
| Note that (3) is not a legitimate address. |
| All cases are handled here. */ |
| |
| void |
| output_load_address (operands) |
| rtx *operands; |
| { |
| rtx base, offset; |
| |
| if (CONSTANT_P (operands[3])) |
| { |
| output_asm_insn ("set %3,%0", operands); |
| return; |
| } |
| |
| if (REG_P (operands[3])) |
| { |
| if (REGNO (operands[0]) != REGNO (operands[3])) |
| output_asm_insn ("mov %3,%0", operands); |
| return; |
| } |
| |
| if (GET_CODE (operands[3]) != PLUS) |
| abort (); |
| |
| base = XEXP (operands[3], 0); |
| offset = XEXP (operands[3], 1); |
| |
| if (GET_CODE (base) == CONST_INT) |
| { |
| rtx tmp = base; |
| base = offset; |
| offset = tmp; |
| } |
| |
| if (GET_CODE (offset) != CONST_INT) |
| { |
| /* Operand is (PLUS (REG) (REG)). */ |
| base = operands[3]; |
| offset = const0_rtx; |
| } |
| |
| if (REG_P (base)) |
| { |
| operands[6] = base; |
| operands[7] = offset; |
| if (SMALL_INT (offset)) |
| output_asm_insn ("add %6,%7,%0", operands); |
| else |
| output_asm_insn ("set %7,%0\n\tadd %0,%6,%0", operands); |
| } |
| else if (GET_CODE (base) == PLUS) |
| { |
| operands[6] = XEXP (base, 0); |
| operands[7] = XEXP (base, 1); |
| operands[8] = offset; |
| |
| if (SMALL_INT (offset)) |
| output_asm_insn ("add %6,%7,%0\n\tadd %0,%8,%0", operands); |
| else |
| output_asm_insn ("set %8,%0\n\tadd %0,%6,%0\n\tadd %0,%7,%0", operands); |
| } |
| else |
| abort (); |
| } |
| |
| /* Output code to place a size count SIZE in register REG. |
| ALIGN is the size of the unit of transfer. |
| |
| Because block moves are pipelined, we don't include the |
| first element in the transfer of SIZE to REG. */ |
| |
| static void |
| output_size_for_block_move (size, reg, align) |
| rtx size, reg; |
| rtx align; |
| { |
| rtx xoperands[3]; |
| |
| xoperands[0] = reg; |
| xoperands[1] = size; |
| xoperands[2] = align; |
| if (GET_CODE (size) == REG) |
| output_asm_insn ("sub %1,%2,%0", xoperands); |
| else |
| { |
| xoperands[1] |
| = gen_rtx (CONST_INT, VOIDmode, INTVAL (size) - INTVAL (align)); |
| output_asm_insn ("set %1,%0", xoperands); |
| } |
| } |
| |
| /* Emit code to perform a block move. |
| |
| OPERANDS[0] is the destination. |
| OPERANDS[1] is the source. |
| OPERANDS[2] is the size. |
| OPERANDS[3] is the alignment safe to use. |
| OPERANDS[4] is a register we can safely clobber as a temp. */ |
| |
| char * |
| output_block_move (operands) |
| rtx *operands; |
| { |
| /* A vector for our computed operands. Note that load_output_address |
| makes use of (and can clobber) up to the 8th element of this vector. */ |
| rtx xoperands[10]; |
| rtx zoperands[10]; |
| static int movstrsi_label = 0; |
| int i; |
| rtx temp1 = operands[4]; |
| rtx sizertx = operands[2]; |
| rtx alignrtx = operands[3]; |
| int align = INTVAL (alignrtx); |
| char label3[30], label5[30]; |
| |
| xoperands[0] = operands[0]; |
| xoperands[1] = operands[1]; |
| xoperands[2] = temp1; |
| |
| /* We can't move more than this many bytes at a time because we have only |
| one register, %g1, to move them through. */ |
| if (align > UNITS_PER_WORD) |
| { |
| align = UNITS_PER_WORD; |
| alignrtx = gen_rtx (CONST_INT, VOIDmode, UNITS_PER_WORD); |
| } |
| |
| /* We consider 8 ld/st pairs, for a total of 16 inline insns to be |
| reasonable here. (Actually will emit a maximum of 18 inline insns for |
| the case of size == 31 and align == 4). */ |
| |
| if (GET_CODE (sizertx) == CONST_INT && (INTVAL (sizertx) / align) <= 8 |
| && memory_address_p (QImode, plus_constant_for_output (xoperands[0], |
| INTVAL (sizertx))) |
| && memory_address_p (QImode, plus_constant_for_output (xoperands[1], |
| INTVAL (sizertx)))) |
| { |
| int size = INTVAL (sizertx); |
| int offset = 0; |
| |
| /* We will store different integers into this particular RTX. */ |
| xoperands[2] = rtx_alloc (CONST_INT); |
| PUT_MODE (xoperands[2], VOIDmode); |
| |
| /* This case is currently not handled. Abort instead of generating |
| bad code. */ |
| if (align > UNITS_PER_WORD) |
| abort (); |
| |
| if (TARGET_ARCH64 && align >= 8) |
| { |
| for (i = (size >> 3) - 1; i >= 0; i--) |
| { |
| INTVAL (xoperands[2]) = (i << 3) + offset; |
| output_asm_insn ("ldx [%a1+%2],%%g1\n\tstx %%g1,[%a0+%2]", |
| xoperands); |
| } |
| offset += (size & ~0x7); |
| size = size & 0x7; |
| if (size == 0) |
| return ""; |
| } |
| |
| if (align >= 4) |
| { |
| for (i = (size >> 2) - 1; i >= 0; i--) |
| { |
| INTVAL (xoperands[2]) = (i << 2) + offset; |
| output_asm_insn ("ld [%a1+%2],%%g1\n\tst %%g1,[%a0+%2]", |
| xoperands); |
| } |
| offset += (size & ~0x3); |
| size = size & 0x3; |
| if (size == 0) |
| return ""; |
| } |
| |
| if (align >= 2) |
| { |
| for (i = (size >> 1) - 1; i >= 0; i--) |
| { |
| INTVAL (xoperands[2]) = (i << 1) + offset; |
| output_asm_insn ("lduh [%a1+%2],%%g1\n\tsth %%g1,[%a0+%2]", |
| xoperands); |
| } |
| offset += (size & ~0x1); |
| size = size & 0x1; |
| if (size == 0) |
| return ""; |
| } |
| |
| if (align >= 1) |
| { |
| for (i = size - 1; i >= 0; i--) |
| { |
| INTVAL (xoperands[2]) = i + offset; |
| output_asm_insn ("ldub [%a1+%2],%%g1\n\tstb %%g1,[%a0+%2]", |
| xoperands); |
| } |
| return ""; |
| } |
| |
| /* We should never reach here. */ |
| abort (); |
| } |
| |
| /* If the size isn't known to be a multiple of the alignment, |
| we have to do it in smaller pieces. If we could determine that |
| the size was a multiple of 2 (or whatever), we could be smarter |
| about this. */ |
| if (GET_CODE (sizertx) != CONST_INT) |
| align = 1; |
| else |
| { |
| int size = INTVAL (sizertx); |
| while (size % align) |
| align >>= 1; |
| } |
| |
| if (align != INTVAL (alignrtx)) |
| alignrtx = gen_rtx (CONST_INT, VOIDmode, align); |
| |
| xoperands[3] = gen_rtx (CONST_INT, VOIDmode, movstrsi_label++); |
| xoperands[4] = gen_rtx (CONST_INT, VOIDmode, align); |
| xoperands[5] = gen_rtx (CONST_INT, VOIDmode, movstrsi_label++); |
| |
| ASM_GENERATE_INTERNAL_LABEL (label3, "Lm", INTVAL (xoperands[3])); |
| ASM_GENERATE_INTERNAL_LABEL (label5, "Lm", INTVAL (xoperands[5])); |
| |
| /* This is the size of the transfer. Emit code to decrement the size |
| value by ALIGN, and store the result in the temp1 register. */ |
| output_size_for_block_move (sizertx, temp1, alignrtx); |
| |
| /* Must handle the case when the size is zero or negative, so the first thing |
| we do is compare the size against zero, and only copy bytes if it is |
| zero or greater. Note that we have already subtracted off the alignment |
| once, so we must copy 1 alignment worth of bytes if the size is zero |
| here. |
| |
| The SUN assembler complains about labels in branch delay slots, so we |
| do this before outputting the load address, so that there will always |
| be a harmless insn between the branch here and the next label emitted |
| below. */ |
| |
| { |
| char pattern[100]; |
| |
| sprintf (pattern, "cmp %%2,0\n\tbl %s", &label5[1]); |
| output_asm_insn (pattern, xoperands); |
| } |
| |
| zoperands[0] = operands[0]; |
| zoperands[3] = plus_constant_for_output (operands[0], align); |
| output_load_address (zoperands); |
| |
| /* ??? This might be much faster if the loops below were preconditioned |
| and unrolled. |
| |
| That is, at run time, copy enough bytes one at a time to ensure that the |
| target and source addresses are aligned to the the largest possible |
| alignment. Then use a preconditioned unrolled loop to copy say 16 |
| bytes at a time. Then copy bytes one at a time until finish the rest. */ |
| |
| /* Output the first label separately, so that it is spaced properly. */ |
| |
| ASM_OUTPUT_INTERNAL_LABEL (asm_out_file, "Lm", INTVAL (xoperands[3])); |
| |
| { |
| char pattern[200]; |
| register char *ld_suffix = ((align == 1) ? "ub" : (align == 2) ? "uh" |
| : (align == 8 && TARGET_ARCH64) ? "x" : ""); |
| register char *st_suffix = ((align == 1) ? "b" : (align == 2) ? "h" |
| : (align == 8 && TARGET_ARCH64) ? "x" : ""); |
| |
| sprintf (pattern, "ld%s [%%1+%%2],%%%%g1\n\tsubcc %%2,%%4,%%2\n\tbge %s\n\tst%s %%%%g1,[%%0+%%2]\n%s:", ld_suffix, &label3[1], st_suffix, &label5[1]); |
| output_asm_insn (pattern, xoperands); |
| } |
| |
| return ""; |
| } |
| #endif |
| |
| /* Output reasonable peephole for set-on-condition-code insns. |
| Note that these insns assume a particular way of defining |
| labels. Therefore, *both* sparc.h and this function must |
| be changed if a new syntax is needed. */ |
| |
| char * |
| output_scc_insn (operands, insn) |
| rtx operands[]; |
| rtx insn; |
| { |
| static char string[100]; |
| rtx label = 0, next = insn; |
| int need_label = 0; |
| |
| /* Try doing a jump optimization which jump.c can't do for us |
| because we did not expose that setcc works by using branches. |
| |
| If this scc insn is followed by an unconditional branch, then have |
| the jump insn emitted here jump to that location, instead of to |
| the end of the scc sequence as usual. */ |
| |
| do |
| { |
| if (GET_CODE (next) == CODE_LABEL) |
| label = next; |
| next = NEXT_INSN (next); |
| if (next == 0) |
| break; |
| } |
| while (GET_CODE (next) == NOTE || GET_CODE (next) == CODE_LABEL); |
| |
| /* If we are in a sequence, and the following insn is a sequence also, |
| then just following the current insn's next field will take us to the |
| first insn of the next sequence, which is the wrong place. We don't |
| want to optimize with a branch that has had its delay slot filled. |
| Avoid this by verifying that NEXT_INSN (PREV_INSN (next)) == next |
| which fails only if NEXT is such a branch. */ |
| |
| if (next && GET_CODE (next) == JUMP_INSN && simplejump_p (next) |
| && (! final_sequence || NEXT_INSN (PREV_INSN (next)) == next)) |
| label = JUMP_LABEL (next); |
| /* If not optimizing, jump label fields are not set. To be safe, always |
| check here to whether label is still zero. */ |
| if (label == 0) |
| { |
| label = gen_label_rtx (); |
| need_label = 1; |
| } |
| |
| LABEL_NUSES (label) += 1; |
| |
| /* operands[3] is an unused slot. */ |
| operands[3] = label; |
| |
| /* If we are in a delay slot, assume it is the delay slot of an fpcc |
| insn since our type isn't allowed anywhere else. */ |
| |
| /* ??? Fpcc instructions no longer have delay slots, so this code is |
| probably obsolete. */ |
| |
| /* The fastest way to emit code for this is an annulled branch followed |
| by two move insns. This will take two cycles if the branch is taken, |
| and three cycles if the branch is not taken. |
| |
| However, if we are in the delay slot of another branch, this won't work, |
| because we can't put a branch in the delay slot of another branch. |
| The above sequence would effectively take 3 or 4 cycles respectively |
| since a no op would have be inserted between the two branches. |
| In this case, we want to emit a move, annulled branch, and then the |
| second move. This sequence always takes 3 cycles, and hence is faster |
| when we are in a branch delay slot. */ |
| |
| if (final_sequence) |
| { |
| strcpy (string, "mov 0,%0\n\t"); |
| strcat (string, output_cbranch (operands[2], 3, 0, 1, 0)); |
| strcat (string, "\n\tmov 1,%0"); |
| } |
| else |
| { |
| strcpy (string, output_cbranch (operands[2], 3, 0, 1, 0)); |
| strcat (string, "\n\tmov 1,%0\n\tmov 0,%0"); |
| } |
| |
| if (need_label) |
| strcat (string, "\n%l3:"); |
| |
| return string; |
| } |
| |
| /* Vectors to keep interesting information about registers where it can easily |
| be got. We use to use the actual mode value as the bit number, but there |
| are more than 32 modes now. Instead we use two tables: one indexed by |
| hard register number, and one indexed by mode. */ |
| |
| /* The purpose of sparc_mode_class is to shrink the range of modes so that |
| they all fit (as bit numbers) in a 32 bit word (again). Each real mode is |
| mapped into one sparc_mode_class mode. */ |
| |
| enum sparc_mode_class { |
| S_MODE, D_MODE, T_MODE, O_MODE, |
| SF_MODE, DF_MODE, TF_MODE, OF_MODE, |
| CC_MODE, CCFP_MODE |
| }; |
| |
| /* Modes for single-word and smaller quantities. */ |
| #define S_MODES ((1 << (int) S_MODE) | (1 << (int) SF_MODE)) |
| |
| /* Modes for double-word and smaller quantities. */ |
| #define D_MODES (S_MODES | (1 << (int) D_MODE) | (1 << DF_MODE)) |
| |
| /* Modes for quad-word and smaller quantities. */ |
| #define T_MODES (D_MODES | (1 << (int) T_MODE) | (1 << (int) TF_MODE)) |
| |
| /* Modes for single-float quantities. We must allow any single word or |
| smaller quantity. This is because the fix/float conversion instructions |
| take integer inputs/outputs from the float registers. */ |
| #define SF_MODES (S_MODES) |
| |
| /* Modes for double-float and smaller quantities. */ |
| #define DF_MODES (S_MODES | D_MODES) |
| |
| /* ??? Sparc64 fp regs cannot hold DImode values. */ |
| #define DF_MODES64 (SF_MODES | DF_MODE /* | D_MODE*/) |
| |
| /* Modes for double-float only quantities. */ |
| /* ??? Sparc64 fp regs cannot hold DImode values. |
| See fix_truncsfdi2. */ |
| #define DF_ONLY_MODES ((1 << (int) DF_MODE) /*| (1 << (int) D_MODE)*/) |
| |
| /* Modes for double-float and larger quantities. */ |
| #define DF_UP_MODES (DF_ONLY_MODES | TF_ONLY_MODES) |
| |
| /* Modes for quad-float only quantities. */ |
| #define TF_ONLY_MODES (1 << (int) TF_MODE) |
| |
| /* Modes for quad-float and smaller quantities. */ |
| #define TF_MODES (DF_MODES | TF_ONLY_MODES) |
| |
| /* ??? Sparc64 fp regs cannot hold DImode values. |
| See fix_truncsfdi2. */ |
| #define TF_MODES64 (DF_MODES64 | TF_ONLY_MODES) |
| |
| /* Modes for condition codes. */ |
| #define CC_MODES (1 << (int) CC_MODE) |
| #define CCFP_MODES (1 << (int) CCFP_MODE) |
| |
| /* Value is 1 if register/mode pair is acceptable on sparc. |
| The funny mixture of D and T modes is because integer operations |
| do not specially operate on tetra quantities, so non-quad-aligned |
| registers can hold quadword quantities (except %o4 and %i4 because |
| they cross fixed registers). */ |
| |
| /* This points to either the 32 bit or the 64 bit version. */ |
| int *hard_regno_mode_classes; |
| |
| static int hard_32bit_mode_classes[] = { |
| S_MODES, S_MODES, T_MODES, S_MODES, T_MODES, S_MODES, D_MODES, S_MODES, |
| T_MODES, S_MODES, T_MODES, S_MODES, D_MODES, S_MODES, D_MODES, S_MODES, |
| T_MODES, S_MODES, T_MODES, S_MODES, T_MODES, S_MODES, D_MODES, S_MODES, |
| T_MODES, S_MODES, T_MODES, S_MODES, D_MODES, S_MODES, D_MODES, S_MODES, |
| |
| TF_MODES, SF_MODES, DF_MODES, SF_MODES, TF_MODES, SF_MODES, DF_MODES, SF_MODES, |
| TF_MODES, SF_MODES, DF_MODES, SF_MODES, TF_MODES, SF_MODES, DF_MODES, SF_MODES, |
| TF_MODES, SF_MODES, DF_MODES, SF_MODES, TF_MODES, SF_MODES, DF_MODES, SF_MODES, |
| TF_MODES, SF_MODES, DF_MODES, SF_MODES, TF_MODES, SF_MODES, DF_MODES, SF_MODES, |
| |
| /* FP regs f32 to f63. Only the even numbered registers actually exist, |
| and none can hold SFmode/SImode values. */ |
| DF_UP_MODES, 0, DF_ONLY_MODES, 0, DF_UP_MODES, 0, DF_ONLY_MODES, 0, |
| DF_UP_MODES, 0, DF_ONLY_MODES, 0, DF_UP_MODES, 0, DF_ONLY_MODES, 0, |
| DF_UP_MODES, 0, DF_ONLY_MODES, 0, DF_UP_MODES, 0, DF_ONLY_MODES, 0, |
| DF_UP_MODES, 0, DF_ONLY_MODES, 0, DF_UP_MODES, 0, DF_ONLY_MODES, 0, |
| |
| /* %fcc[0123] */ |
| CCFP_MODES, CCFP_MODES, CCFP_MODES, CCFP_MODES, |
| |
| /* %icc */ |
| CC_MODES |
| }; |
| |
| static int hard_64bit_mode_classes[] = { |
| D_MODES, D_MODES, T_MODES, D_MODES, T_MODES, D_MODES, T_MODES, D_MODES, |
| T_MODES, D_MODES, T_MODES, D_MODES, T_MODES, D_MODES, T_MODES, D_MODES, |
| T_MODES, D_MODES, T_MODES, D_MODES, T_MODES, D_MODES, T_MODES, D_MODES, |
| T_MODES, D_MODES, T_MODES, D_MODES, T_MODES, D_MODES, T_MODES, D_MODES, |
| |
| TF_MODES64, SF_MODES, DF_MODES64, SF_MODES, TF_MODES64, SF_MODES, DF_MODES64, SF_MODES, |
| TF_MODES64, SF_MODES, DF_MODES64, SF_MODES, TF_MODES64, SF_MODES, DF_MODES64, SF_MODES, |
| TF_MODES64, SF_MODES, DF_MODES64, SF_MODES, TF_MODES64, SF_MODES, DF_MODES64, SF_MODES, |
| TF_MODES64, SF_MODES, DF_MODES64, SF_MODES, TF_MODES64, SF_MODES, DF_MODES64, SF_MODES, |
| |
| /* FP regs f32 to f63. Only the even numbered registers actually exist, |
| and none can hold SFmode/SImode values. */ |
| DF_UP_MODES, 0, DF_ONLY_MODES, 0, DF_UP_MODES, 0, DF_ONLY_MODES, 0, |
| DF_UP_MODES, 0, DF_ONLY_MODES, 0, DF_UP_MODES, 0, DF_ONLY_MODES, 0, |
| DF_UP_MODES, 0, DF_ONLY_MODES, 0, DF_UP_MODES, 0, DF_ONLY_MODES, 0, |
| DF_UP_MODES, 0, DF_ONLY_MODES, 0, DF_UP_MODES, 0, DF_ONLY_MODES, 0, |
| |
| /* %fcc[0123] */ |
| CCFP_MODES, CCFP_MODES, CCFP_MODES, CCFP_MODES, |
| |
| /* %icc */ |
| CC_MODES |
| }; |
| |
| int sparc_mode_class [NUM_MACHINE_MODES]; |
| |
| enum reg_class sparc_regno_reg_class[FIRST_PSEUDO_REGISTER]; |
| |
| static void |
| sparc_init_modes () |
| { |
| int i; |
| |
| for (i = 0; i < NUM_MACHINE_MODES; i++) |
| { |
| switch (GET_MODE_CLASS (i)) |
| { |
| case MODE_INT: |
| case MODE_PARTIAL_INT: |
| case MODE_COMPLEX_INT: |
| if (GET_MODE_SIZE (i) <= 4) |
| sparc_mode_class[i] = 1 << (int) S_MODE; |
| else if (GET_MODE_SIZE (i) == 8) |
| sparc_mode_class[i] = 1 << (int) D_MODE; |
| else if (GET_MODE_SIZE (i) == 16) |
| sparc_mode_class[i] = 1 << (int) T_MODE; |
| else if (GET_MODE_SIZE (i) == 32) |
| sparc_mode_class[i] = 1 << (int) O_MODE; |
| else |
| sparc_mode_class[i] = 0; |
| break; |
| case MODE_FLOAT: |
| case MODE_COMPLEX_FLOAT: |
| if (GET_MODE_SIZE (i) <= 4) |
| sparc_mode_class[i] = 1 << (int) SF_MODE; |
| else if (GET_MODE_SIZE (i) == 8) |
| sparc_mode_class[i] = 1 << (int) DF_MODE; |
| else if (GET_MODE_SIZE (i) == 16) |
| sparc_mode_class[i] = 1 << (int) TF_MODE; |
| else if (GET_MODE_SIZE (i) == 32) |
| sparc_mode_class[i] = 1 << (int) OF_MODE; |
| else |
| sparc_mode_class[i] = 0; |
| break; |
| case MODE_CC: |
| default: |
| /* mode_class hasn't been initialized yet for EXTRA_CC_MODES, so |
| we must explicitly check for them here. */ |
| if (i == (int) CCFPmode || i == (int) CCFPEmode) |
| sparc_mode_class[i] = 1 << (int) CCFP_MODE; |
| else if (i == (int) CCmode || i == (int) CC_NOOVmode |
| || i == (int) CCXmode || i == (int) CCX_NOOVmode) |
| sparc_mode_class[i] = 1 << (int) CC_MODE; |
| else |
| sparc_mode_class[i] = 0; |
| break; |
| } |
| } |
| |
| if (TARGET_ARCH64) |
| hard_regno_mode_classes = hard_64bit_mode_classes; |
| else |
| hard_regno_mode_classes = hard_32bit_mode_classes; |
| |
| /* Initialize the array used by REGNO_REG_CLASS. */ |
| for (i = 0; i < FIRST_PSEUDO_REGISTER; i++) |
| { |
| if (i < 32) |
| sparc_regno_reg_class[i] = GENERAL_REGS; |
| else if (i < 64) |
| sparc_regno_reg_class[i] = FP_REGS; |
| else if (i < 96) |
| sparc_regno_reg_class[i] = EXTRA_FP_REGS; |
| else if (i < 100) |
| sparc_regno_reg_class[i] = FPCC_REGS; |
| else |
| sparc_regno_reg_class[i] = NO_REGS; |
| } |
| } |
| |
| /* Save non call used registers from LOW to HIGH at BASE+OFFSET. |
| N_REGS is the number of 4-byte regs saved thus far. This applies even to |
| v9 int regs as it simplifies the code. */ |
| |
| #ifdef __GNUC__ |
| __inline__ |
| #endif |
| static int |
| save_regs (file, low, high, base, offset, n_regs, real_offset) |
| FILE *file; |
| int low, high; |
| char *base; |
| int offset; |
| int n_regs; |
| int real_offset; |
| { |
| int i; |
| |
| if (TARGET_ARCH64 && high <= 32) |
| { |
| for (i = low; i < high; i++) |
| { |
| if (regs_ever_live[i] && ! call_used_regs[i]) |
| { |
| fprintf (file, "\tstx %s,[%s+%d]\n", |
| reg_names[i], base, offset + 4 * n_regs); |
| #ifdef DWARF2_DEBUGGING_INFO |
| if (write_symbols == DWARF2_DEBUG) |
| dwarf2out_reg_save ("", i, real_offset + 4 * n_regs); |
| #endif |
| n_regs += 2; |
| } |
| } |
| } |
| else |
| { |
| for (i = low; i < high; i += 2) |
| { |
| if (regs_ever_live[i] && ! call_used_regs[i]) |
| if (regs_ever_live[i+1] && ! call_used_regs[i+1]) |
| { |
| fprintf (file, "\tstd %s,[%s+%d]\n", |
| reg_names[i], base, offset + 4 * n_regs); |
| #ifdef DWARF2_DEBUGGING_INFO |
| if (write_symbols == DWARF2_DEBUG) |
| { |
| char *l = (char *) dwarf2out_cfi_label (); |
| dwarf2out_reg_save (l, i, real_offset + 4 * n_regs); |
| dwarf2out_reg_save (l, i+1, real_offset + 4 * n_regs + 4); |
| } |
| #endif |
| n_regs += 2; |
| } |
| else |
| { |
| fprintf (file, "\tst %s,[%s+%d]\n", |
| reg_names[i], base, offset + 4 * n_regs); |
| #ifdef DWARF2_DEBUGGING_INFO |
| if (write_symbols == DWARF2_DEBUG) |
| dwarf2out_reg_save ("", i, real_offset + 4 * n_regs); |
| #endif |
| n_regs += 2; |
| } |
| else if (regs_ever_live[i+1] && ! call_used_regs[i+1]) |
| { |
| fprintf (file, "\tst %s,[%s+%d]\n", |
| reg_names[i+1], base, offset + 4 * n_regs + 4); |
| #ifdef DWARF2_DEBUGGING_INFO |
| if (write_symbols == DWARF2_DEBUG) |
| dwarf2out_reg_save ("", i + 1, real_offset + 4 * n_regs + 4); |
| #endif |
| n_regs += 2; |
| } |
| } |
| } |
| return n_regs; |
| } |
| |
| /* Restore non call used registers from LOW to HIGH at BASE+OFFSET. |
| |
| N_REGS is the number of 4-byte regs saved thus far. This applies even to |
| v9 int regs as it simplifies the code. */ |
| |
| #ifdef __GNUC__ |
| __inline__ |
| #endif |
| static int |
| restore_regs (file, low, high, base, offset, n_regs) |
| FILE *file; |
| int low, high; |
| char *base; |
| int offset; |
| int n_regs; |
| { |
| int i; |
| |
| if (TARGET_ARCH64 && high <= 32) |
| { |
| for (i = low; i < high; i++) |
| { |
| if (regs_ever_live[i] && ! call_used_regs[i]) |
| fprintf (file, "\tldx [%s+%d], %s\n", |
| base, offset + 4 * n_regs, reg_names[i]), |
| n_regs += 2; |
| } |
| } |
| else |
| { |
| for (i = low; i < high; i += 2) |
| { |
| if (regs_ever_live[i] && ! call_used_regs[i]) |
| if (regs_ever_live[i+1] && ! call_used_regs[i+1]) |
| fprintf (file, "\tldd [%s+%d], %s\n", |
| base, offset + 4 * n_regs, reg_names[i]), |
| n_regs += 2; |
| else |
| fprintf (file, "\tld [%s+%d],%s\n", |
| base, offset + 4 * n_regs, reg_names[i]), |
| n_regs += 2; |
| else if (regs_ever_live[i+1] && ! call_used_regs[i+1]) |
| fprintf (file, "\tld [%s+%d],%s\n", |
| base, offset + 4 * n_regs + 4, reg_names[i+1]), |
| n_regs += 2; |
| } |
| } |
| return n_regs; |
| } |
| |
| /* Static variables we want to share between prologue and epilogue. */ |
| |
| /* Number of live general or floating point registers needed to be saved |
| (as 4-byte quantities). This is only done if TARGET_EPILOGUE. */ |
| static int num_gfregs; |
| |
| /* Compute the frame size required by the function. This function is called |
| during the reload pass and also by output_function_prologue(). */ |
| |
| int |
| compute_frame_size (size, leaf_function) |
| int size; |
| int leaf_function; |
| { |
| int n_regs = 0, i; |
| int outgoing_args_size = (current_function_outgoing_args_size |
| #if ! SPARC_ARCH64 |
| + REG_PARM_STACK_SPACE (current_function_decl) |
| #endif |
| ); |
| |
| if (TARGET_EPILOGUE) |
| { |
| /* N_REGS is the number of 4-byte regs saved thus far. This applies |
| even to v9 int regs to be consistent with save_regs/restore_regs. */ |
| |
| if (TARGET_ARCH64) |
| { |
| for (i = 0; i < 8; i++) |
| if (regs_ever_live[i] && ! call_used_regs[i]) |
| n_regs += 2; |
| } |
| else |
| { |
| for (i = |