| /* Definitions of target machine for GCC for IA-32. |
| Copyright (C) 1988-2015 Free Software Foundation, Inc. |
| |
| This file is part of GCC. |
| |
| GCC is free software; you can redistribute it and/or modify |
| it under the terms of the GNU General Public License as published by |
| the Free Software Foundation; either version 3, or (at your option) |
| any later version. |
| |
| GCC is distributed in the hope that it will be useful, |
| but WITHOUT ANY WARRANTY; without even the implied warranty of |
| MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| GNU General Public License for more details. |
| |
| Under Section 7 of GPL version 3, you are granted additional |
| permissions described in the GCC Runtime Library Exception, version |
| 3.1, as published by the Free Software Foundation. |
| |
| You should have received a copy of the GNU General Public License and |
| a copy of the GCC Runtime Library Exception along with this program; |
| see the files COPYING3 and COPYING.RUNTIME respectively. If not, see |
| <http://www.gnu.org/licenses/>. */ |
| |
| /* The purpose of this file is to define the characteristics of the i386, |
| independent of assembler syntax or operating system. |
| |
| Three other files build on this one to describe a specific assembler syntax: |
| bsd386.h, att386.h, and sun386.h. |
| |
| The actual tm.h file for a particular system should include |
| this file, and then the file for the appropriate assembler syntax. |
| |
| Many macros that specify assembler syntax are omitted entirely from |
| this file because they really belong in the files for particular |
| assemblers. These include RP, IP, LPREFIX, PUT_OP_SIZE, USE_STAR, |
| ADDR_BEG, ADDR_END, PRINT_IREG, PRINT_SCALE, PRINT_B_I_S, and many |
| that start with ASM_ or end in ASM_OP. */ |
| |
| /* Redefines for option macros. */ |
| |
| #define TARGET_64BIT TARGET_ISA_64BIT |
| #define TARGET_64BIT_P(x) TARGET_ISA_64BIT_P(x) |
| #define TARGET_MMX TARGET_ISA_MMX |
| #define TARGET_MMX_P(x) TARGET_ISA_MMX_P(x) |
| #define TARGET_3DNOW TARGET_ISA_3DNOW |
| #define TARGET_3DNOW_P(x) TARGET_ISA_3DNOW_P(x) |
| #define TARGET_3DNOW_A TARGET_ISA_3DNOW_A |
| #define TARGET_3DNOW_A_P(x) TARGET_ISA_3DNOW_A_P(x) |
| #define TARGET_SSE TARGET_ISA_SSE |
| #define TARGET_SSE_P(x) TARGET_ISA_SSE_P(x) |
| #define TARGET_SSE2 TARGET_ISA_SSE2 |
| #define TARGET_SSE2_P(x) TARGET_ISA_SSE2_P(x) |
| #define TARGET_SSE3 TARGET_ISA_SSE3 |
| #define TARGET_SSE3_P(x) TARGET_ISA_SSE3_P(x) |
| #define TARGET_SSSE3 TARGET_ISA_SSSE3 |
| #define TARGET_SSSE3_P(x) TARGET_ISA_SSSE3_P(x) |
| #define TARGET_SSE4_1 TARGET_ISA_SSE4_1 |
| #define TARGET_SSE4_1_P(x) TARGET_ISA_SSE4_1_P(x) |
| #define TARGET_SSE4_2 TARGET_ISA_SSE4_2 |
| #define TARGET_SSE4_2_P(x) TARGET_ISA_SSE4_2_P(x) |
| #define TARGET_AVX TARGET_ISA_AVX |
| #define TARGET_AVX_P(x) TARGET_ISA_AVX_P(x) |
| #define TARGET_AVX2 TARGET_ISA_AVX2 |
| #define TARGET_AVX2_P(x) TARGET_ISA_AVX2_P(x) |
| #define TARGET_AVX512F TARGET_ISA_AVX512F |
| #define TARGET_AVX512F_P(x) TARGET_ISA_AVX512F_P(x) |
| #define TARGET_AVX512PF TARGET_ISA_AVX512PF |
| #define TARGET_AVX512PF_P(x) TARGET_ISA_AVX512PF_P(x) |
| #define TARGET_AVX512ER TARGET_ISA_AVX512ER |
| #define TARGET_AVX512ER_P(x) TARGET_ISA_AVX512ER_P(x) |
| #define TARGET_AVX512CD TARGET_ISA_AVX512CD |
| #define TARGET_AVX512CD_P(x) TARGET_ISA_AVX512CD_P(x) |
| #define TARGET_AVX512DQ TARGET_ISA_AVX512DQ |
| #define TARGET_AVX512DQ_P(x) TARGET_ISA_AVX512DQ_P(x) |
| #define TARGET_AVX512BW TARGET_ISA_AVX512BW |
| #define TARGET_AVX512BW_P(x) TARGET_ISA_AVX512BW_P(x) |
| #define TARGET_AVX512VL TARGET_ISA_AVX512VL |
| #define TARGET_AVX512VL_P(x) TARGET_ISA_AVX512VL_P(x) |
| #define TARGET_AVX512VBMI TARGET_ISA_AVX512VBMI |
| #define TARGET_AVX512VBMI_P(x) TARGET_ISA_AVX512VBMI_P(x) |
| #define TARGET_AVX512IFMA TARGET_ISA_AVX512IFMA |
| #define TARGET_AVX512IFMA_P(x) TARGET_ISA_AVX512IFMA_P(x) |
| #define TARGET_FMA TARGET_ISA_FMA |
| #define TARGET_FMA_P(x) TARGET_ISA_FMA_P(x) |
| #define TARGET_SSE4A TARGET_ISA_SSE4A |
| #define TARGET_SSE4A_P(x) TARGET_ISA_SSE4A_P(x) |
| #define TARGET_FMA4 TARGET_ISA_FMA4 |
| #define TARGET_FMA4_P(x) TARGET_ISA_FMA4_P(x) |
| #define TARGET_XOP TARGET_ISA_XOP |
| #define TARGET_XOP_P(x) TARGET_ISA_XOP_P(x) |
| #define TARGET_LWP TARGET_ISA_LWP |
| #define TARGET_LWP_P(x) TARGET_ISA_LWP_P(x) |
| #define TARGET_ROUND TARGET_ISA_ROUND |
| #define TARGET_ABM TARGET_ISA_ABM |
| #define TARGET_ABM_P(x) TARGET_ISA_ABM_P(x) |
| #define TARGET_BMI TARGET_ISA_BMI |
| #define TARGET_BMI_P(x) TARGET_ISA_BMI_P(x) |
| #define TARGET_BMI2 TARGET_ISA_BMI2 |
| #define TARGET_BMI2_P(x) TARGET_ISA_BMI2_P(x) |
| #define TARGET_LZCNT TARGET_ISA_LZCNT |
| #define TARGET_LZCNT_P(x) TARGET_ISA_LZCNT_P(x) |
| #define TARGET_TBM TARGET_ISA_TBM |
| #define TARGET_TBM_P(x) TARGET_ISA_TBM_P(x) |
| #define TARGET_POPCNT TARGET_ISA_POPCNT |
| #define TARGET_POPCNT_P(x) TARGET_ISA_POPCNT_P(x) |
| #define TARGET_SAHF TARGET_ISA_SAHF |
| #define TARGET_SAHF_P(x) TARGET_ISA_SAHF_P(x) |
| #define TARGET_MOVBE TARGET_ISA_MOVBE |
| #define TARGET_MOVBE_P(x) TARGET_ISA_MOVBE_P(x) |
| #define TARGET_CRC32 TARGET_ISA_CRC32 |
| #define TARGET_CRC32_P(x) TARGET_ISA_CRC32_P(x) |
| #define TARGET_AES TARGET_ISA_AES |
| #define TARGET_AES_P(x) TARGET_ISA_AES_P(x) |
| #define TARGET_SHA TARGET_ISA_SHA |
| #define TARGET_SHA_P(x) TARGET_ISA_SHA_P(x) |
| #define TARGET_CLFLUSHOPT TARGET_ISA_CLFLUSHOPT |
| #define TARGET_CLFLUSHOPT_P(x) TARGET_ISA_CLFLUSHOPT_P(x) |
| #define TARGET_XSAVEC TARGET_ISA_XSAVEC |
| #define TARGET_XSAVEC_P(x) TARGET_ISA_XSAVEC_P(x) |
| #define TARGET_XSAVES TARGET_ISA_XSAVES |
| #define TARGET_XSAVES_P(x) TARGET_ISA_XSAVES_P(x) |
| #define TARGET_PCLMUL TARGET_ISA_PCLMUL |
| #define TARGET_PCLMUL_P(x) TARGET_ISA_PCLMUL_P(x) |
| #define TARGET_CMPXCHG16B TARGET_ISA_CX16 |
| #define TARGET_CMPXCHG16B_P(x) TARGET_ISA_CX16_P(x) |
| #define TARGET_FSGSBASE TARGET_ISA_FSGSBASE |
| #define TARGET_FSGSBASE_P(x) TARGET_ISA_FSGSBASE_P(x) |
| #define TARGET_RDRND TARGET_ISA_RDRND |
| #define TARGET_RDRND_P(x) TARGET_ISA_RDRND_P(x) |
| #define TARGET_F16C TARGET_ISA_F16C |
| #define TARGET_F16C_P(x) TARGET_ISA_F16C_P(x) |
| #define TARGET_RTM TARGET_ISA_RTM |
| #define TARGET_RTM_P(x) TARGET_ISA_RTM_P(x) |
| #define TARGET_HLE TARGET_ISA_HLE |
| #define TARGET_HLE_P(x) TARGET_ISA_HLE_P(x) |
| #define TARGET_RDSEED TARGET_ISA_RDSEED |
| #define TARGET_RDSEED_P(x) TARGET_ISA_RDSEED_P(x) |
| #define TARGET_PRFCHW TARGET_ISA_PRFCHW |
| #define TARGET_PRFCHW_P(x) TARGET_ISA_PRFCHW_P(x) |
| #define TARGET_ADX TARGET_ISA_ADX |
| #define TARGET_ADX_P(x) TARGET_ISA_ADX_P(x) |
| #define TARGET_FXSR TARGET_ISA_FXSR |
| #define TARGET_FXSR_P(x) TARGET_ISA_FXSR_P(x) |
| #define TARGET_XSAVE TARGET_ISA_XSAVE |
| #define TARGET_XSAVE_P(x) TARGET_ISA_XSAVE_P(x) |
| #define TARGET_XSAVEOPT TARGET_ISA_XSAVEOPT |
| #define TARGET_XSAVEOPT_P(x) TARGET_ISA_XSAVEOPT_P(x) |
| #define TARGET_PREFETCHWT1 TARGET_ISA_PREFETCHWT1 |
| #define TARGET_PREFETCHWT1_P(x) TARGET_ISA_PREFETCHWT1_P(x) |
| #define TARGET_MPX TARGET_ISA_MPX |
| #define TARGET_MPX_P(x) TARGET_ISA_MPX_P(x) |
| #define TARGET_CLWB TARGET_ISA_CLWB |
| #define TARGET_CLWB_P(x) TARGET_ISA_CLWB_P(x) |
| #define TARGET_MWAITX TARGET_ISA_MWAITX |
| #define TARGET_MWAITX_P(x) TARGET_ISA_MWAITX_P(x) |
| |
| #define TARGET_LP64 TARGET_ABI_64 |
| #define TARGET_LP64_P(x) TARGET_ABI_64_P(x) |
| #define TARGET_X32 TARGET_ABI_X32 |
| #define TARGET_X32_P(x) TARGET_ABI_X32_P(x) |
| #define TARGET_16BIT TARGET_CODE16 |
| #define TARGET_16BIT_P(x) TARGET_CODE16_P(x) |
| |
| /* SSE4.1 defines round instructions */ |
| #define OPTION_MASK_ISA_ROUND OPTION_MASK_ISA_SSE4_1 |
| #define TARGET_ISA_ROUND ((ix86_isa_flags & OPTION_MASK_ISA_ROUND) != 0) |
| |
| #include "config/vxworks-dummy.h" |
| |
| #include "config/i386/i386-opts.h" |
| |
| #define MAX_STRINGOP_ALGS 4 |
| |
| /* Specify what algorithm to use for stringops on known size. |
| When size is unknown, the UNKNOWN_SIZE alg is used. When size is |
| known at compile time or estimated via feedback, the SIZE array |
| is walked in order until MAX is greater then the estimate (or -1 |
| means infinity). Corresponding ALG is used then. |
| When NOALIGN is true the code guaranting the alignment of the memory |
| block is skipped. |
| |
| For example initializer: |
| {{256, loop}, {-1, rep_prefix_4_byte}} |
| will use loop for blocks smaller or equal to 256 bytes, rep prefix will |
| be used otherwise. */ |
| struct stringop_algs |
| { |
| const enum stringop_alg unknown_size; |
| const struct stringop_strategy { |
| const int max; |
| const enum stringop_alg alg; |
| int noalign; |
| } size [MAX_STRINGOP_ALGS]; |
| }; |
| |
| /* Define the specific costs for a given cpu */ |
| |
| struct processor_costs { |
| const int add; /* cost of an add instruction */ |
| const int lea; /* cost of a lea instruction */ |
| const int shift_var; /* variable shift costs */ |
| const int shift_const; /* constant shift costs */ |
| const int mult_init[5]; /* cost of starting a multiply |
| in QImode, HImode, SImode, DImode, TImode*/ |
| const int mult_bit; /* cost of multiply per each bit set */ |
| const int divide[5]; /* cost of a divide/mod |
| in QImode, HImode, SImode, DImode, TImode*/ |
| int movsx; /* The cost of movsx operation. */ |
| int movzx; /* The cost of movzx operation. */ |
| const int large_insn; /* insns larger than this cost more */ |
| const int move_ratio; /* The threshold of number of scalar |
| memory-to-memory move insns. */ |
| const int movzbl_load; /* cost of loading using movzbl */ |
| const int int_load[3]; /* cost of loading integer registers |
| in QImode, HImode and SImode relative |
| to reg-reg move (2). */ |
| const int int_store[3]; /* cost of storing integer register |
| in QImode, HImode and SImode */ |
| const int fp_move; /* cost of reg,reg fld/fst */ |
| const int fp_load[3]; /* cost of loading FP register |
| in SFmode, DFmode and XFmode */ |
| const int fp_store[3]; /* cost of storing FP register |
| in SFmode, DFmode and XFmode */ |
| const int mmx_move; /* cost of moving MMX register. */ |
| const int mmx_load[2]; /* cost of loading MMX register |
| in SImode and DImode */ |
| const int mmx_store[2]; /* cost of storing MMX register |
| in SImode and DImode */ |
| const int sse_move; /* cost of moving SSE register. */ |
| const int sse_load[3]; /* cost of loading SSE register |
| in SImode, DImode and TImode*/ |
| const int sse_store[3]; /* cost of storing SSE register |
| in SImode, DImode and TImode*/ |
| const int mmxsse_to_integer; /* cost of moving mmxsse register to |
| integer and vice versa. */ |
| const int l1_cache_size; /* size of l1 cache, in kilobytes. */ |
| const int l2_cache_size; /* size of l2 cache, in kilobytes. */ |
| const int prefetch_block; /* bytes moved to cache for prefetch. */ |
| const int simultaneous_prefetches; /* number of parallel prefetch |
| operations. */ |
| const int branch_cost; /* Default value for BRANCH_COST. */ |
| const int fadd; /* cost of FADD and FSUB instructions. */ |
| const int fmul; /* cost of FMUL instruction. */ |
| const int fdiv; /* cost of FDIV instruction. */ |
| const int fabs; /* cost of FABS instruction. */ |
| const int fchs; /* cost of FCHS instruction. */ |
| const int fsqrt; /* cost of FSQRT instruction. */ |
| /* Specify what algorithm |
| to use for stringops on unknown size. */ |
| struct stringop_algs *memcpy, *memset; |
| const int scalar_stmt_cost; /* Cost of any scalar operation, excluding |
| load and store. */ |
| const int scalar_load_cost; /* Cost of scalar load. */ |
| const int scalar_store_cost; /* Cost of scalar store. */ |
| const int vec_stmt_cost; /* Cost of any vector operation, excluding |
| load, store, vector-to-scalar and |
| scalar-to-vector operation. */ |
| const int vec_to_scalar_cost; /* Cost of vect-to-scalar operation. */ |
| const int scalar_to_vec_cost; /* Cost of scalar-to-vector operation. */ |
| const int vec_align_load_cost; /* Cost of aligned vector load. */ |
| const int vec_unalign_load_cost; /* Cost of unaligned vector load. */ |
| const int vec_store_cost; /* Cost of vector store. */ |
| const int cond_taken_branch_cost; /* Cost of taken branch for vectorizer |
| cost model. */ |
| const int cond_not_taken_branch_cost;/* Cost of not taken branch for |
| vectorizer cost model. */ |
| }; |
| |
| extern const struct processor_costs *ix86_cost; |
| extern const struct processor_costs ix86_size_cost; |
| |
| #define ix86_cur_cost() \ |
| (optimize_insn_for_size_p () ? &ix86_size_cost: ix86_cost) |
| |
| /* Macros used in the machine description to test the flags. */ |
| |
| /* configure can arrange to change it. */ |
| |
| #ifndef TARGET_CPU_DEFAULT |
| #define TARGET_CPU_DEFAULT PROCESSOR_GENERIC |
| #endif |
| |
| #ifndef TARGET_FPMATH_DEFAULT |
| #define TARGET_FPMATH_DEFAULT \ |
| (TARGET_64BIT && TARGET_SSE ? FPMATH_SSE : FPMATH_387) |
| #endif |
| |
| #ifndef TARGET_FPMATH_DEFAULT_P |
| #define TARGET_FPMATH_DEFAULT_P(x) \ |
| (TARGET_64BIT_P(x) && TARGET_SSE_P(x) ? FPMATH_SSE : FPMATH_387) |
| #endif |
| |
| #define TARGET_FLOAT_RETURNS_IN_80387 TARGET_FLOAT_RETURNS |
| #define TARGET_FLOAT_RETURNS_IN_80387_P(x) TARGET_FLOAT_RETURNS_P(x) |
| |
| /* 64bit Sledgehammer mode. For libgcc2 we make sure this is a |
| compile-time constant. */ |
| #ifdef IN_LIBGCC2 |
| #undef TARGET_64BIT |
| #ifdef __x86_64__ |
| #define TARGET_64BIT 1 |
| #else |
| #define TARGET_64BIT 0 |
| #endif |
| #else |
| #ifndef TARGET_BI_ARCH |
| #undef TARGET_64BIT |
| #undef TARGET_64BIT_P |
| #if TARGET_64BIT_DEFAULT |
| #define TARGET_64BIT 1 |
| #define TARGET_64BIT_P(x) 1 |
| #else |
| #define TARGET_64BIT 0 |
| #define TARGET_64BIT_P(x) 0 |
| #endif |
| #endif |
| #endif |
| |
| #define HAS_LONG_COND_BRANCH 1 |
| #define HAS_LONG_UNCOND_BRANCH 1 |
| |
| #define TARGET_386 (ix86_tune == PROCESSOR_I386) |
| #define TARGET_486 (ix86_tune == PROCESSOR_I486) |
| #define TARGET_PENTIUM (ix86_tune == PROCESSOR_PENTIUM) |
| #define TARGET_PENTIUMPRO (ix86_tune == PROCESSOR_PENTIUMPRO) |
| #define TARGET_GEODE (ix86_tune == PROCESSOR_GEODE) |
| #define TARGET_K6 (ix86_tune == PROCESSOR_K6) |
| #define TARGET_ATHLON (ix86_tune == PROCESSOR_ATHLON) |
| #define TARGET_PENTIUM4 (ix86_tune == PROCESSOR_PENTIUM4) |
| #define TARGET_K8 (ix86_tune == PROCESSOR_K8) |
| #define TARGET_ATHLON_K8 (TARGET_K8 || TARGET_ATHLON) |
| #define TARGET_NOCONA (ix86_tune == PROCESSOR_NOCONA) |
| #define TARGET_CORE2 (ix86_tune == PROCESSOR_CORE2) |
| #define TARGET_NEHALEM (ix86_tune == PROCESSOR_NEHALEM) |
| #define TARGET_SANDYBRIDGE (ix86_tune == PROCESSOR_SANDYBRIDGE) |
| #define TARGET_HASWELL (ix86_tune == PROCESSOR_HASWELL) |
| #define TARGET_BONNELL (ix86_tune == PROCESSOR_BONNELL) |
| #define TARGET_SILVERMONT (ix86_tune == PROCESSOR_SILVERMONT) |
| #define TARGET_KNL (ix86_tune == PROCESSOR_KNL) |
| #define TARGET_INTEL (ix86_tune == PROCESSOR_INTEL) |
| #define TARGET_GENERIC (ix86_tune == PROCESSOR_GENERIC) |
| #define TARGET_AMDFAM10 (ix86_tune == PROCESSOR_AMDFAM10) |
| #define TARGET_BDVER1 (ix86_tune == PROCESSOR_BDVER1) |
| #define TARGET_BDVER2 (ix86_tune == PROCESSOR_BDVER2) |
| #define TARGET_BDVER3 (ix86_tune == PROCESSOR_BDVER3) |
| #define TARGET_BDVER4 (ix86_tune == PROCESSOR_BDVER4) |
| #define TARGET_BTVER1 (ix86_tune == PROCESSOR_BTVER1) |
| #define TARGET_BTVER2 (ix86_tune == PROCESSOR_BTVER2) |
| |
| /* Feature tests against the various tunings. */ |
| enum ix86_tune_indices { |
| #undef DEF_TUNE |
| #define DEF_TUNE(tune, name, selector) tune, |
| #include "x86-tune.def" |
| #undef DEF_TUNE |
| X86_TUNE_LAST |
| }; |
| |
| extern unsigned char ix86_tune_features[X86_TUNE_LAST]; |
| |
| #define TARGET_USE_LEAVE ix86_tune_features[X86_TUNE_USE_LEAVE] |
| #define TARGET_PUSH_MEMORY ix86_tune_features[X86_TUNE_PUSH_MEMORY] |
| #define TARGET_ZERO_EXTEND_WITH_AND \ |
| ix86_tune_features[X86_TUNE_ZERO_EXTEND_WITH_AND] |
| #define TARGET_UNROLL_STRLEN ix86_tune_features[X86_TUNE_UNROLL_STRLEN] |
| #define TARGET_BRANCH_PREDICTION_HINTS \ |
| ix86_tune_features[X86_TUNE_BRANCH_PREDICTION_HINTS] |
| #define TARGET_DOUBLE_WITH_ADD ix86_tune_features[X86_TUNE_DOUBLE_WITH_ADD] |
| #define TARGET_USE_SAHF ix86_tune_features[X86_TUNE_USE_SAHF] |
| #define TARGET_MOVX ix86_tune_features[X86_TUNE_MOVX] |
| #define TARGET_PARTIAL_REG_STALL ix86_tune_features[X86_TUNE_PARTIAL_REG_STALL] |
| #define TARGET_PARTIAL_FLAG_REG_STALL \ |
| ix86_tune_features[X86_TUNE_PARTIAL_FLAG_REG_STALL] |
| #define TARGET_LCP_STALL \ |
| ix86_tune_features[X86_TUNE_LCP_STALL] |
| #define TARGET_USE_HIMODE_FIOP ix86_tune_features[X86_TUNE_USE_HIMODE_FIOP] |
| #define TARGET_USE_SIMODE_FIOP ix86_tune_features[X86_TUNE_USE_SIMODE_FIOP] |
| #define TARGET_USE_MOV0 ix86_tune_features[X86_TUNE_USE_MOV0] |
| #define TARGET_USE_CLTD ix86_tune_features[X86_TUNE_USE_CLTD] |
| #define TARGET_USE_XCHGB ix86_tune_features[X86_TUNE_USE_XCHGB] |
| #define TARGET_SPLIT_LONG_MOVES ix86_tune_features[X86_TUNE_SPLIT_LONG_MOVES] |
| #define TARGET_READ_MODIFY_WRITE ix86_tune_features[X86_TUNE_READ_MODIFY_WRITE] |
| #define TARGET_READ_MODIFY ix86_tune_features[X86_TUNE_READ_MODIFY] |
| #define TARGET_PROMOTE_QImode ix86_tune_features[X86_TUNE_PROMOTE_QIMODE] |
| #define TARGET_FAST_PREFIX ix86_tune_features[X86_TUNE_FAST_PREFIX] |
| #define TARGET_SINGLE_STRINGOP ix86_tune_features[X86_TUNE_SINGLE_STRINGOP] |
| #define TARGET_MISALIGNED_MOVE_STRING_PRO_EPILOGUES \ |
| ix86_tune_features[X86_TUNE_MISALIGNED_MOVE_STRING_PRO_EPILOGUES] |
| #define TARGET_QIMODE_MATH ix86_tune_features[X86_TUNE_QIMODE_MATH] |
| #define TARGET_HIMODE_MATH ix86_tune_features[X86_TUNE_HIMODE_MATH] |
| #define TARGET_PROMOTE_QI_REGS ix86_tune_features[X86_TUNE_PROMOTE_QI_REGS] |
| #define TARGET_PROMOTE_HI_REGS ix86_tune_features[X86_TUNE_PROMOTE_HI_REGS] |
| #define TARGET_SINGLE_POP ix86_tune_features[X86_TUNE_SINGLE_POP] |
| #define TARGET_DOUBLE_POP ix86_tune_features[X86_TUNE_DOUBLE_POP] |
| #define TARGET_SINGLE_PUSH ix86_tune_features[X86_TUNE_SINGLE_PUSH] |
| #define TARGET_DOUBLE_PUSH ix86_tune_features[X86_TUNE_DOUBLE_PUSH] |
| #define TARGET_INTEGER_DFMODE_MOVES \ |
| ix86_tune_features[X86_TUNE_INTEGER_DFMODE_MOVES] |
| #define TARGET_PARTIAL_REG_DEPENDENCY \ |
| ix86_tune_features[X86_TUNE_PARTIAL_REG_DEPENDENCY] |
| #define TARGET_SSE_PARTIAL_REG_DEPENDENCY \ |
| ix86_tune_features[X86_TUNE_SSE_PARTIAL_REG_DEPENDENCY] |
| #define TARGET_SSE_UNALIGNED_LOAD_OPTIMAL \ |
| ix86_tune_features[X86_TUNE_SSE_UNALIGNED_LOAD_OPTIMAL] |
| #define TARGET_SSE_UNALIGNED_STORE_OPTIMAL \ |
| ix86_tune_features[X86_TUNE_SSE_UNALIGNED_STORE_OPTIMAL] |
| #define TARGET_SSE_PACKED_SINGLE_INSN_OPTIMAL \ |
| ix86_tune_features[X86_TUNE_SSE_PACKED_SINGLE_INSN_OPTIMAL] |
| #define TARGET_SSE_SPLIT_REGS ix86_tune_features[X86_TUNE_SSE_SPLIT_REGS] |
| #define TARGET_SSE_TYPELESS_STORES \ |
| ix86_tune_features[X86_TUNE_SSE_TYPELESS_STORES] |
| #define TARGET_SSE_LOAD0_BY_PXOR ix86_tune_features[X86_TUNE_SSE_LOAD0_BY_PXOR] |
| #define TARGET_MEMORY_MISMATCH_STALL \ |
| ix86_tune_features[X86_TUNE_MEMORY_MISMATCH_STALL] |
| #define TARGET_PROLOGUE_USING_MOVE \ |
| ix86_tune_features[X86_TUNE_PROLOGUE_USING_MOVE] |
| #define TARGET_EPILOGUE_USING_MOVE \ |
| ix86_tune_features[X86_TUNE_EPILOGUE_USING_MOVE] |
| #define TARGET_SHIFT1 ix86_tune_features[X86_TUNE_SHIFT1] |
| #define TARGET_USE_FFREEP ix86_tune_features[X86_TUNE_USE_FFREEP] |
| #define TARGET_INTER_UNIT_MOVES_TO_VEC \ |
| ix86_tune_features[X86_TUNE_INTER_UNIT_MOVES_TO_VEC] |
| #define TARGET_INTER_UNIT_MOVES_FROM_VEC \ |
| ix86_tune_features[X86_TUNE_INTER_UNIT_MOVES_FROM_VEC] |
| #define TARGET_INTER_UNIT_CONVERSIONS \ |
| ix86_tune_features[X86_TUNE_INTER_UNIT_CONVERSIONS] |
| #define TARGET_FOUR_JUMP_LIMIT ix86_tune_features[X86_TUNE_FOUR_JUMP_LIMIT] |
| #define TARGET_SCHEDULE ix86_tune_features[X86_TUNE_SCHEDULE] |
| #define TARGET_USE_BT ix86_tune_features[X86_TUNE_USE_BT] |
| #define TARGET_USE_INCDEC ix86_tune_features[X86_TUNE_USE_INCDEC] |
| #define TARGET_PAD_RETURNS ix86_tune_features[X86_TUNE_PAD_RETURNS] |
| #define TARGET_PAD_SHORT_FUNCTION \ |
| ix86_tune_features[X86_TUNE_PAD_SHORT_FUNCTION] |
| #define TARGET_EXT_80387_CONSTANTS \ |
| ix86_tune_features[X86_TUNE_EXT_80387_CONSTANTS] |
| #define TARGET_AVOID_VECTOR_DECODE \ |
| ix86_tune_features[X86_TUNE_AVOID_VECTOR_DECODE] |
| #define TARGET_TUNE_PROMOTE_HIMODE_IMUL \ |
| ix86_tune_features[X86_TUNE_PROMOTE_HIMODE_IMUL] |
| #define TARGET_SLOW_IMUL_IMM32_MEM \ |
| ix86_tune_features[X86_TUNE_SLOW_IMUL_IMM32_MEM] |
| #define TARGET_SLOW_IMUL_IMM8 ix86_tune_features[X86_TUNE_SLOW_IMUL_IMM8] |
| #define TARGET_MOVE_M1_VIA_OR ix86_tune_features[X86_TUNE_MOVE_M1_VIA_OR] |
| #define TARGET_NOT_UNPAIRABLE ix86_tune_features[X86_TUNE_NOT_UNPAIRABLE] |
| #define TARGET_NOT_VECTORMODE ix86_tune_features[X86_TUNE_NOT_VECTORMODE] |
| #define TARGET_USE_VECTOR_FP_CONVERTS \ |
| ix86_tune_features[X86_TUNE_USE_VECTOR_FP_CONVERTS] |
| #define TARGET_USE_VECTOR_CONVERTS \ |
| ix86_tune_features[X86_TUNE_USE_VECTOR_CONVERTS] |
| #define TARGET_SLOW_PSHUFB \ |
| ix86_tune_features[X86_TUNE_SLOW_PSHUFB] |
| #define TARGET_VECTOR_PARALLEL_EXECUTION \ |
| ix86_tune_features[X86_TUNE_VECTOR_PARALLEL_EXECUTION] |
| #define TARGET_AVOID_4BYTE_PREFIXES \ |
| ix86_tune_features[X86_TUNE_AVOID_4BYTE_PREFIXES] |
| #define TARGET_FUSE_CMP_AND_BRANCH_32 \ |
| ix86_tune_features[X86_TUNE_FUSE_CMP_AND_BRANCH_32] |
| #define TARGET_FUSE_CMP_AND_BRANCH_64 \ |
| ix86_tune_features[X86_TUNE_FUSE_CMP_AND_BRANCH_64] |
| #define TARGET_FUSE_CMP_AND_BRANCH \ |
| (TARGET_64BIT ? TARGET_FUSE_CMP_AND_BRANCH_64 \ |
| : TARGET_FUSE_CMP_AND_BRANCH_32) |
| #define TARGET_FUSE_CMP_AND_BRANCH_SOFLAGS \ |
| ix86_tune_features[X86_TUNE_FUSE_CMP_AND_BRANCH_SOFLAGS] |
| #define TARGET_FUSE_ALU_AND_BRANCH \ |
| ix86_tune_features[X86_TUNE_FUSE_ALU_AND_BRANCH] |
| #define TARGET_OPT_AGU ix86_tune_features[X86_TUNE_OPT_AGU] |
| #define TARGET_AVOID_LEA_FOR_ADDR \ |
| ix86_tune_features[X86_TUNE_AVOID_LEA_FOR_ADDR] |
| #define TARGET_VECTORIZE_DOUBLE \ |
| ix86_tune_features[X86_TUNE_VECTORIZE_DOUBLE] |
| #define TARGET_SOFTWARE_PREFETCHING_BENEFICIAL \ |
| ix86_tune_features[X86_TUNE_SOFTWARE_PREFETCHING_BENEFICIAL] |
| #define TARGET_AVX128_OPTIMAL \ |
| ix86_tune_features[X86_TUNE_AVX128_OPTIMAL] |
| #define TARGET_REASSOC_INT_TO_PARALLEL \ |
| ix86_tune_features[X86_TUNE_REASSOC_INT_TO_PARALLEL] |
| #define TARGET_REASSOC_FP_TO_PARALLEL \ |
| ix86_tune_features[X86_TUNE_REASSOC_FP_TO_PARALLEL] |
| #define TARGET_GENERAL_REGS_SSE_SPILL \ |
| ix86_tune_features[X86_TUNE_GENERAL_REGS_SSE_SPILL] |
| #define TARGET_AVOID_MEM_OPND_FOR_CMOVE \ |
| ix86_tune_features[X86_TUNE_AVOID_MEM_OPND_FOR_CMOVE] |
| #define TARGET_SPLIT_MEM_OPND_FOR_FP_CONVERTS \ |
| ix86_tune_features[X86_TUNE_SPLIT_MEM_OPND_FOR_FP_CONVERTS] |
| #define TARGET_ADJUST_UNROLL \ |
| ix86_tune_features[X86_TUNE_ADJUST_UNROLL] |
| #define TARGET_AVOID_FALSE_DEP_FOR_BMI \ |
| ix86_tune_features[X86_TUNE_AVOID_FALSE_DEP_FOR_BMI] |
| |
| /* Feature tests against the various architecture variations. */ |
| enum ix86_arch_indices { |
| X86_ARCH_CMOV, |
| X86_ARCH_CMPXCHG, |
| X86_ARCH_CMPXCHG8B, |
| X86_ARCH_XADD, |
| X86_ARCH_BSWAP, |
| |
| X86_ARCH_LAST |
| }; |
| |
| extern unsigned char ix86_arch_features[X86_ARCH_LAST]; |
| |
| #define TARGET_CMOV ix86_arch_features[X86_ARCH_CMOV] |
| #define TARGET_CMPXCHG ix86_arch_features[X86_ARCH_CMPXCHG] |
| #define TARGET_CMPXCHG8B ix86_arch_features[X86_ARCH_CMPXCHG8B] |
| #define TARGET_XADD ix86_arch_features[X86_ARCH_XADD] |
| #define TARGET_BSWAP ix86_arch_features[X86_ARCH_BSWAP] |
| |
| /* For sane SSE instruction set generation we need fcomi instruction. |
| It is safe to enable all CMOVE instructions. Also, RDRAND intrinsic |
| expands to a sequence that includes conditional move. */ |
| #define TARGET_CMOVE (TARGET_CMOV || TARGET_SSE || TARGET_RDRND) |
| |
| #define TARGET_FISTTP (TARGET_SSE3 && TARGET_80387) |
| |
| extern unsigned char x86_prefetch_sse; |
| #define TARGET_PREFETCH_SSE x86_prefetch_sse |
| |
| #define ASSEMBLER_DIALECT (ix86_asm_dialect) |
| |
| #define TARGET_SSE_MATH ((ix86_fpmath & FPMATH_SSE) != 0) |
| #define TARGET_MIX_SSE_I387 \ |
| ((ix86_fpmath & (FPMATH_SSE | FPMATH_387)) == (FPMATH_SSE | FPMATH_387)) |
| |
| #define TARGET_GNU_TLS (ix86_tls_dialect == TLS_DIALECT_GNU) |
| #define TARGET_GNU2_TLS (ix86_tls_dialect == TLS_DIALECT_GNU2) |
| #define TARGET_ANY_GNU_TLS (TARGET_GNU_TLS || TARGET_GNU2_TLS) |
| #define TARGET_SUN_TLS 0 |
| |
| #ifndef TARGET_64BIT_DEFAULT |
| #define TARGET_64BIT_DEFAULT 0 |
| #endif |
| #ifndef TARGET_TLS_DIRECT_SEG_REFS_DEFAULT |
| #define TARGET_TLS_DIRECT_SEG_REFS_DEFAULT 0 |
| #endif |
| |
| #define TARGET_SSP_GLOBAL_GUARD (ix86_stack_protector_guard == SSP_GLOBAL) |
| #define TARGET_SSP_TLS_GUARD (ix86_stack_protector_guard == SSP_TLS) |
| |
| /* Fence to use after loop using storent. */ |
| |
| extern tree x86_mfence; |
| #define FENCE_FOLLOWING_MOVNT x86_mfence |
| |
| /* Once GDB has been enhanced to deal with functions without frame |
| pointers, we can change this to allow for elimination of |
| the frame pointer in leaf functions. */ |
| #define TARGET_DEFAULT 0 |
| |
| /* Extra bits to force. */ |
| #define TARGET_SUBTARGET_DEFAULT 0 |
| #define TARGET_SUBTARGET_ISA_DEFAULT 0 |
| |
| /* Extra bits to force on w/ 32-bit mode. */ |
| #define TARGET_SUBTARGET32_DEFAULT 0 |
| #define TARGET_SUBTARGET32_ISA_DEFAULT 0 |
| |
| /* Extra bits to force on w/ 64-bit mode. */ |
| #define TARGET_SUBTARGET64_DEFAULT 0 |
| #define TARGET_SUBTARGET64_ISA_DEFAULT 0 |
| |
| /* Replace MACH-O, ifdefs by in-line tests, where possible. |
| (a) Macros defined in config/i386/darwin.h */ |
| #define TARGET_MACHO 0 |
| #define TARGET_MACHO_BRANCH_ISLANDS 0 |
| #define MACHOPIC_ATT_STUB 0 |
| /* (b) Macros defined in config/darwin.h */ |
| #define MACHO_DYNAMIC_NO_PIC_P 0 |
| #define MACHOPIC_INDIRECT 0 |
| #define MACHOPIC_PURE 0 |
| |
| /* For the RDOS */ |
| #define TARGET_RDOS 0 |
| |
| /* For the Windows 64-bit ABI. */ |
| #define TARGET_64BIT_MS_ABI (TARGET_64BIT && ix86_cfun_abi () == MS_ABI) |
| |
| /* For the Windows 32-bit ABI. */ |
| #define TARGET_32BIT_MS_ABI (!TARGET_64BIT && ix86_cfun_abi () == MS_ABI) |
| |
| /* This is re-defined by cygming.h. */ |
| #define TARGET_SEH 0 |
| |
| /* This is re-defined by cygming.h. */ |
| #define TARGET_PECOFF 0 |
| |
| /* The default abi used by target. */ |
| #define DEFAULT_ABI SYSV_ABI |
| |
| /* The default TLS segment register used by target. */ |
| #define DEFAULT_TLS_SEG_REG (TARGET_64BIT ? SEG_FS : SEG_GS) |
| |
| /* Subtargets may reset this to 1 in order to enable 96-bit long double |
| with the rounding mode forced to 53 bits. */ |
| #define TARGET_96_ROUND_53_LONG_DOUBLE 0 |
| |
| /* -march=native handling only makes sense with compiler running on |
| an x86 or x86_64 chip. If changing this condition, also change |
| the condition in driver-i386.c. */ |
| #if defined(__i386__) || defined(__x86_64__) |
| /* In driver-i386.c. */ |
| extern const char *host_detect_local_cpu (int argc, const char **argv); |
| #define EXTRA_SPEC_FUNCTIONS \ |
| { "local_cpu_detect", host_detect_local_cpu }, |
| #define HAVE_LOCAL_CPU_DETECT |
| #endif |
| |
| #if TARGET_64BIT_DEFAULT |
| #define OPT_ARCH64 "!m32" |
| #define OPT_ARCH32 "m32" |
| #else |
| #define OPT_ARCH64 "m64|mx32" |
| #define OPT_ARCH32 "m64|mx32:;" |
| #endif |
| |
| /* Support for configure-time defaults of some command line options. |
| The order here is important so that -march doesn't squash the |
| tune or cpu values. */ |
| #define OPTION_DEFAULT_SPECS \ |
| {"tune", "%{!mtune=*:%{!mcpu=*:%{!march=*:-mtune=%(VALUE)}}}" }, \ |
| {"tune_32", "%{" OPT_ARCH32 ":%{!mtune=*:%{!mcpu=*:%{!march=*:-mtune=%(VALUE)}}}}" }, \ |
| {"tune_64", "%{" OPT_ARCH64 ":%{!mtune=*:%{!mcpu=*:%{!march=*:-mtune=%(VALUE)}}}}" }, \ |
| {"cpu", "%{!mtune=*:%{!mcpu=*:%{!march=*:-mtune=%(VALUE)}}}" }, \ |
| {"cpu_32", "%{" OPT_ARCH32 ":%{!mtune=*:%{!mcpu=*:%{!march=*:-mtune=%(VALUE)}}}}" }, \ |
| {"cpu_64", "%{" OPT_ARCH64 ":%{!mtune=*:%{!mcpu=*:%{!march=*:-mtune=%(VALUE)}}}}" }, \ |
| {"arch", "%{!march=*:-march=%(VALUE)}"}, \ |
| {"arch_32", "%{" OPT_ARCH32 ":%{!march=*:-march=%(VALUE)}}"}, \ |
| {"arch_64", "%{" OPT_ARCH64 ":%{!march=*:-march=%(VALUE)}}"}, |
| |
| /* Specs for the compiler proper */ |
| |
| #ifndef CC1_CPU_SPEC |
| #define CC1_CPU_SPEC_1 "" |
| |
| #ifndef HAVE_LOCAL_CPU_DETECT |
| #define CC1_CPU_SPEC CC1_CPU_SPEC_1 |
| #else |
| #define CC1_CPU_SPEC CC1_CPU_SPEC_1 \ |
| "%{march=native:%>march=native %:local_cpu_detect(arch) \ |
| %{!mtune=*:%>mtune=native %:local_cpu_detect(tune)}} \ |
| %{mtune=native:%>mtune=native %:local_cpu_detect(tune)}" |
| #endif |
| #endif |
| |
| /* Target CPU builtins. */ |
| #define TARGET_CPU_CPP_BUILTINS() ix86_target_macros () |
| |
| /* Target Pragmas. */ |
| #define REGISTER_TARGET_PRAGMAS() ix86_register_pragmas () |
| |
| #ifndef CC1_SPEC |
| #define CC1_SPEC "%(cc1_cpu) " |
| #endif |
| |
| /* This macro defines names of additional specifications to put in the |
| specs that can be used in various specifications like CC1_SPEC. Its |
| definition is an initializer with a subgrouping for each command option. |
| |
| Each subgrouping contains a string constant, that defines the |
| specification name, and a string constant that used by the GCC driver |
| program. |
| |
| Do not define this macro if it does not need to do anything. */ |
| |
| #ifndef SUBTARGET_EXTRA_SPECS |
| #define SUBTARGET_EXTRA_SPECS |
| #endif |
| |
| #define EXTRA_SPECS \ |
| { "cc1_cpu", CC1_CPU_SPEC }, \ |
| SUBTARGET_EXTRA_SPECS |
| |
| |
| /* Set the value of FLT_EVAL_METHOD in float.h. When using only the |
| FPU, assume that the fpcw is set to extended precision; when using |
| only SSE, rounding is correct; when using both SSE and the FPU, |
| the rounding precision is indeterminate, since either may be chosen |
| apparently at random. */ |
| #define TARGET_FLT_EVAL_METHOD \ |
| (TARGET_MIX_SSE_I387 ? -1 : TARGET_SSE_MATH ? 0 : 2) |
| |
| /* Whether to allow x87 floating-point arithmetic on MODE (one of |
| SFmode, DFmode and XFmode) in the current excess precision |
| configuration. */ |
| #define X87_ENABLE_ARITH(MODE) \ |
| (flag_excess_precision == EXCESS_PRECISION_FAST || (MODE) == XFmode) |
| |
| /* Likewise, whether to allow direct conversions from integer mode |
| IMODE (HImode, SImode or DImode) to MODE. */ |
| #define X87_ENABLE_FLOAT(MODE, IMODE) \ |
| (flag_excess_precision == EXCESS_PRECISION_FAST \ |
| || (MODE) == XFmode \ |
| || ((MODE) == DFmode && (IMODE) == SImode) \ |
| || (IMODE) == HImode) |
| |
| /* target machine storage layout */ |
| |
| #define SHORT_TYPE_SIZE 16 |
| #define INT_TYPE_SIZE 32 |
| #define LONG_TYPE_SIZE (TARGET_X32 ? 32 : BITS_PER_WORD) |
| #define POINTER_SIZE (TARGET_X32 ? 32 : BITS_PER_WORD) |
| #define LONG_LONG_TYPE_SIZE 64 |
| #define FLOAT_TYPE_SIZE 32 |
| #define DOUBLE_TYPE_SIZE 64 |
| #define LONG_DOUBLE_TYPE_SIZE \ |
| (TARGET_LONG_DOUBLE_64 ? 64 : (TARGET_LONG_DOUBLE_128 ? 128 : 80)) |
| |
| #define WIDEST_HARDWARE_FP_SIZE 80 |
| |
| #if defined (TARGET_BI_ARCH) || TARGET_64BIT_DEFAULT |
| #define MAX_BITS_PER_WORD 64 |
| #else |
| #define MAX_BITS_PER_WORD 32 |
| #endif |
| |
| /* Define this if most significant byte of a word is the lowest numbered. */ |
| /* That is true on the 80386. */ |
| |
| #define BITS_BIG_ENDIAN 0 |
| |
| /* Define this if most significant byte of a word is the lowest numbered. */ |
| /* That is not true on the 80386. */ |
| #define BYTES_BIG_ENDIAN 0 |
| |
| /* Define this if most significant word of a multiword number is the lowest |
| numbered. */ |
| /* Not true for 80386 */ |
| #define WORDS_BIG_ENDIAN 0 |
| |
| /* Width of a word, in units (bytes). */ |
| #define UNITS_PER_WORD (TARGET_64BIT ? 8 : 4) |
| |
| #ifndef IN_LIBGCC2 |
| #define MIN_UNITS_PER_WORD 4 |
| #endif |
| |
| /* Allocation boundary (in *bits*) for storing arguments in argument list. */ |
| #define PARM_BOUNDARY BITS_PER_WORD |
| |
| /* Boundary (in *bits*) on which stack pointer should be aligned. */ |
| #define STACK_BOUNDARY \ |
| (TARGET_64BIT && ix86_abi == MS_ABI ? 128 : BITS_PER_WORD) |
| |
| /* Stack boundary of the main function guaranteed by OS. */ |
| #define MAIN_STACK_BOUNDARY (TARGET_64BIT ? 128 : 32) |
| |
| /* Minimum stack boundary. */ |
| #define MIN_STACK_BOUNDARY BITS_PER_WORD |
| |
| /* Boundary (in *bits*) on which the stack pointer prefers to be |
| aligned; the compiler cannot rely on having this alignment. */ |
| #define PREFERRED_STACK_BOUNDARY ix86_preferred_stack_boundary |
| |
| /* It should be MIN_STACK_BOUNDARY. But we set it to 128 bits for |
| both 32bit and 64bit, to support codes that need 128 bit stack |
| alignment for SSE instructions, but can't realign the stack. */ |
| #define PREFERRED_STACK_BOUNDARY_DEFAULT 128 |
| |
| /* 1 if -mstackrealign should be turned on by default. It will |
| generate an alternate prologue and epilogue that realigns the |
| runtime stack if nessary. This supports mixing codes that keep a |
| 4-byte aligned stack, as specified by i386 psABI, with codes that |
| need a 16-byte aligned stack, as required by SSE instructions. */ |
| #define STACK_REALIGN_DEFAULT 0 |
| |
| /* Boundary (in *bits*) on which the incoming stack is aligned. */ |
| #define INCOMING_STACK_BOUNDARY ix86_incoming_stack_boundary |
| |
| /* According to Windows x64 software convention, the maximum stack allocatable |
| in the prologue is 4G - 8 bytes. Furthermore, there is a limited set of |
| instructions allowed to adjust the stack pointer in the epilog, forcing the |
| use of frame pointer for frames larger than 2 GB. This theorical limit |
| is reduced by 256, an over-estimated upper bound for the stack use by the |
| prologue. |
| We define only one threshold for both the prolog and the epilog. When the |
| frame size is larger than this threshold, we allocate the area to save SSE |
| regs, then save them, and then allocate the remaining. There is no SEH |
| unwind info for this later allocation. */ |
| #define SEH_MAX_FRAME_SIZE ((2U << 30) - 256) |
| |
| /* Target OS keeps a vector-aligned (128-bit, 16-byte) stack. This is |
| mandatory for the 64-bit ABI, and may or may not be true for other |
| operating systems. */ |
| #define TARGET_KEEPS_VECTOR_ALIGNED_STACK TARGET_64BIT |
| |
| /* Minimum allocation boundary for the code of a function. */ |
| #define FUNCTION_BOUNDARY 8 |
| |
| /* C++ stores the virtual bit in the lowest bit of function pointers. */ |
| #define TARGET_PTRMEMFUNC_VBIT_LOCATION ptrmemfunc_vbit_in_pfn |
| |
| /* Minimum size in bits of the largest boundary to which any |
| and all fundamental data types supported by the hardware |
| might need to be aligned. No data type wants to be aligned |
| rounder than this. |
| |
| Pentium+ prefers DFmode values to be aligned to 64 bit boundary |
| and Pentium Pro XFmode values at 128 bit boundaries. |
| |
| When increasing the maximum, also update |
| TARGET_ABSOLUTE_BIGGEST_ALIGNMENT. */ |
| |
| #define BIGGEST_ALIGNMENT \ |
| (TARGET_AVX512F ? 512 : (TARGET_AVX ? 256 : 128)) |
| |
| /* Maximum stack alignment. */ |
| #define MAX_STACK_ALIGNMENT MAX_OFILE_ALIGNMENT |
| |
| /* Alignment value for attribute ((aligned)). It is a constant since |
| it is the part of the ABI. We shouldn't change it with -mavx. */ |
| #define ATTRIBUTE_ALIGNED_VALUE 128 |
| |
| /* Decide whether a variable of mode MODE should be 128 bit aligned. */ |
| #define ALIGN_MODE_128(MODE) \ |
| ((MODE) == XFmode || SSE_REG_MODE_P (MODE)) |
| |
| /* The published ABIs say that doubles should be aligned on word |
| boundaries, so lower the alignment for structure fields unless |
| -malign-double is set. */ |
| |
| /* ??? Blah -- this macro is used directly by libobjc. Since it |
| supports no vector modes, cut out the complexity and fall back |
| on BIGGEST_FIELD_ALIGNMENT. */ |
| #ifdef IN_TARGET_LIBS |
| #ifdef __x86_64__ |
| #define BIGGEST_FIELD_ALIGNMENT 128 |
| #else |
| #define BIGGEST_FIELD_ALIGNMENT 32 |
| #endif |
| #else |
| #define ADJUST_FIELD_ALIGN(FIELD, COMPUTED) \ |
| x86_field_alignment (FIELD, COMPUTED) |
| #endif |
| |
| /* If defined, a C expression to compute the alignment given to a |
| constant that is being placed in memory. EXP is the constant |
| and ALIGN is the alignment that the object would ordinarily have. |
| The value of this macro is used instead of that alignment to align |
| the object. |
| |
| If this macro is not defined, then ALIGN is used. |
| |
| The typical use of this macro is to increase alignment for string |
| constants to be word aligned so that `strcpy' calls that copy |
| constants can be done inline. */ |
| |
| #define CONSTANT_ALIGNMENT(EXP, ALIGN) ix86_constant_alignment ((EXP), (ALIGN)) |
| |
| /* If defined, a C expression to compute the alignment for a static |
| variable. TYPE is the data type, and ALIGN is the alignment that |
| the object would ordinarily have. The value of this macro is used |
| instead of that alignment to align the object. |
| |
| If this macro is not defined, then ALIGN is used. |
| |
| One use of this macro is to increase alignment of medium-size |
| data to make it all fit in fewer cache lines. Another is to |
| cause character arrays to be word-aligned so that `strcpy' calls |
| that copy constants to character arrays can be done inline. */ |
| |
| #define DATA_ALIGNMENT(TYPE, ALIGN) \ |
| ix86_data_alignment ((TYPE), (ALIGN), true) |
| |
| /* Similar to DATA_ALIGNMENT, but for the cases where the ABI mandates |
| some alignment increase, instead of optimization only purposes. E.g. |
| AMD x86-64 psABI says that variables with array type larger than 15 bytes |
| must be aligned to 16 byte boundaries. |
| |
| If this macro is not defined, then ALIGN is used. */ |
| |
| #define DATA_ABI_ALIGNMENT(TYPE, ALIGN) \ |
| ix86_data_alignment ((TYPE), (ALIGN), false) |
| |
| /* If defined, a C expression to compute the alignment for a local |
| variable. TYPE is the data type, and ALIGN is the alignment that |
| the object would ordinarily have. The value of this macro is used |
| instead of that alignment to align the object. |
| |
| If this macro is not defined, then ALIGN is used. |
| |
| One use of this macro is to increase alignment of medium-size |
| data to make it all fit in fewer cache lines. */ |
| |
| #define LOCAL_ALIGNMENT(TYPE, ALIGN) \ |
| ix86_local_alignment ((TYPE), VOIDmode, (ALIGN)) |
| |
| /* If defined, a C expression to compute the alignment for stack slot. |
| TYPE is the data type, MODE is the widest mode available, and ALIGN |
| is the alignment that the slot would ordinarily have. The value of |
| this macro is used instead of that alignment to align the slot. |
| |
| If this macro is not defined, then ALIGN is used when TYPE is NULL, |
| Otherwise, LOCAL_ALIGNMENT will be used. |
| |
| One use of this macro is to set alignment of stack slot to the |
| maximum alignment of all possible modes which the slot may have. */ |
| |
| #define STACK_SLOT_ALIGNMENT(TYPE, MODE, ALIGN) \ |
| ix86_local_alignment ((TYPE), (MODE), (ALIGN)) |
| |
| /* If defined, a C expression to compute the alignment for a local |
| variable DECL. |
| |
| If this macro is not defined, then |
| LOCAL_ALIGNMENT (TREE_TYPE (DECL), DECL_ALIGN (DECL)) will be used. |
| |
| One use of this macro is to increase alignment of medium-size |
| data to make it all fit in fewer cache lines. */ |
| |
| #define LOCAL_DECL_ALIGNMENT(DECL) \ |
| ix86_local_alignment ((DECL), VOIDmode, DECL_ALIGN (DECL)) |
| |
| /* If defined, a C expression to compute the minimum required alignment |
| for dynamic stack realignment purposes for EXP (a TYPE or DECL), |
| MODE, assuming normal alignment ALIGN. |
| |
| If this macro is not defined, then (ALIGN) will be used. */ |
| |
| #define MINIMUM_ALIGNMENT(EXP, MODE, ALIGN) \ |
| ix86_minimum_alignment (EXP, MODE, ALIGN) |
| |
| |
| /* Set this nonzero if move instructions will actually fail to work |
| when given unaligned data. */ |
| #define STRICT_ALIGNMENT 0 |
| |
| /* If bit field type is int, don't let it cross an int, |
| and give entire struct the alignment of an int. */ |
| /* Required on the 386 since it doesn't have bit-field insns. */ |
| #define PCC_BITFIELD_TYPE_MATTERS 1 |
| |
| /* Standard register usage. */ |
| |
| /* This processor has special stack-like registers. See reg-stack.c |
| for details. */ |
| |
| #define STACK_REGS |
| |
| #define IS_STACK_MODE(MODE) \ |
| (((MODE) == SFmode && !(TARGET_SSE && TARGET_SSE_MATH)) \ |
| || ((MODE) == DFmode && !(TARGET_SSE2 && TARGET_SSE_MATH)) \ |
| || (MODE) == XFmode) |
| |
| /* Number of actual hardware registers. |
| The hardware registers are assigned numbers for the compiler |
| from 0 to just below FIRST_PSEUDO_REGISTER. |
| All registers that the compiler knows about must be given numbers, |
| even those that are not normally considered general registers. |
| |
| In the 80386 we give the 8 general purpose registers the numbers 0-7. |
| We number the floating point registers 8-15. |
| Note that registers 0-7 can be accessed as a short or int, |
| while only 0-3 may be used with byte `mov' instructions. |
| |
| Reg 16 does not correspond to any hardware register, but instead |
| appears in the RTL as an argument pointer prior to reload, and is |
| eliminated during reloading in favor of either the stack or frame |
| pointer. */ |
| |
| #define FIRST_PSEUDO_REGISTER 81 |
| |
| /* Number of hardware registers that go into the DWARF-2 unwind info. |
| If not defined, equals FIRST_PSEUDO_REGISTER. */ |
| |
| #define DWARF_FRAME_REGISTERS 17 |
| |
| /* 1 for registers that have pervasive standard uses |
| and are not available for the register allocator. |
| On the 80386, the stack pointer is such, as is the arg pointer. |
| |
| REX registers are disabled for 32bit targets in |
| TARGET_CONDITIONAL_REGISTER_USAGE. */ |
| |
| #define FIXED_REGISTERS \ |
| /*ax,dx,cx,bx,si,di,bp,sp,st,st1,st2,st3,st4,st5,st6,st7*/ \ |
| { 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, \ |
| /*arg,flags,fpsr,fpcr,frame*/ \ |
| 1, 1, 1, 1, 1, \ |
| /*xmm0,xmm1,xmm2,xmm3,xmm4,xmm5,xmm6,xmm7*/ \ |
| 0, 0, 0, 0, 0, 0, 0, 0, \ |
| /* mm0, mm1, mm2, mm3, mm4, mm5, mm6, mm7*/ \ |
| 0, 0, 0, 0, 0, 0, 0, 0, \ |
| /* r8, r9, r10, r11, r12, r13, r14, r15*/ \ |
| 0, 0, 0, 0, 0, 0, 0, 0, \ |
| /*xmm8,xmm9,xmm10,xmm11,xmm12,xmm13,xmm14,xmm15*/ \ |
| 0, 0, 0, 0, 0, 0, 0, 0, \ |
| /*xmm16,xmm17,xmm18,xmm19,xmm20,xmm21,xmm22,xmm23*/ \ |
| 0, 0, 0, 0, 0, 0, 0, 0, \ |
| /*xmm24,xmm25,xmm26,xmm27,xmm28,xmm29,xmm30,xmm31*/ \ |
| 0, 0, 0, 0, 0, 0, 0, 0, \ |
| /* k0, k1, k2, k3, k4, k5, k6, k7*/ \ |
| 0, 0, 0, 0, 0, 0, 0, 0, \ |
| /* b0, b1, b2, b3*/ \ |
| 0, 0, 0, 0 } |
| |
| /* 1 for registers not available across function calls. |
| These must include the FIXED_REGISTERS and also any |
| registers that can be used without being saved. |
| The latter must include the registers where values are returned |
| and the register where structure-value addresses are passed. |
| Aside from that, you can include as many other registers as you like. |
| |
| Value is set to 1 if the register is call used unconditionally. |
| Bit one is set if the register is call used on TARGET_32BIT ABI. |
| Bit two is set if the register is call used on TARGET_64BIT ABI. |
| Bit three is set if the register is call used on TARGET_64BIT_MS_ABI. |
| |
| Proper values are computed in TARGET_CONDITIONAL_REGISTER_USAGE. */ |
| |
| #define CALL_USED_REGISTERS \ |
| /*ax,dx,cx,bx,si,di,bp,sp,st,st1,st2,st3,st4,st5,st6,st7*/ \ |
| { 1, 1, 1, 0, 4, 4, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, \ |
| /*arg,flags,fpsr,fpcr,frame*/ \ |
| 1, 1, 1, 1, 1, \ |
| /*xmm0,xmm1,xmm2,xmm3,xmm4,xmm5,xmm6,xmm7*/ \ |
| 1, 1, 1, 1, 1, 1, 6, 6, \ |
| /* mm0, mm1, mm2, mm3, mm4, mm5, mm6, mm7*/ \ |
| 1, 1, 1, 1, 1, 1, 1, 1, \ |
| /* r8, r9, r10, r11, r12, r13, r14, r15*/ \ |
| 1, 1, 1, 1, 2, 2, 2, 2, \ |
| /*xmm8,xmm9,xmm10,xmm11,xmm12,xmm13,xmm14,xmm15*/ \ |
| 6, 6, 6, 6, 6, 6, 6, 6, \ |
| /*xmm16,xmm17,xmm18,xmm19,xmm20,xmm21,xmm22,xmm23*/ \ |
| 6, 6, 6, 6, 6, 6, 6, 6, \ |
| /*xmm24,xmm25,xmm26,xmm27,xmm28,xmm29,xmm30,xmm31*/ \ |
| 6, 6, 6, 6, 6, 6, 6, 6, \ |
| /* k0, k1, k2, k3, k4, k5, k6, k7*/ \ |
| 1, 1, 1, 1, 1, 1, 1, 1, \ |
| /* b0, b1, b2, b3*/ \ |
| 1, 1, 1, 1 } |
| |
| /* Order in which to allocate registers. Each register must be |
| listed once, even those in FIXED_REGISTERS. List frame pointer |
| late and fixed registers last. Note that, in general, we prefer |
| registers listed in CALL_USED_REGISTERS, keeping the others |
| available for storage of persistent values. |
| |
| The ADJUST_REG_ALLOC_ORDER actually overwrite the order, |
| so this is just empty initializer for array. */ |
| |
| #define REG_ALLOC_ORDER \ |
| { 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17,\ |
| 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, \ |
| 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, \ |
| 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, \ |
| 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, \ |
| 78, 79, 80 } |
| |
| /* ADJUST_REG_ALLOC_ORDER is a macro which permits reg_alloc_order |
| to be rearranged based on a particular function. When using sse math, |
| we want to allocate SSE before x87 registers and vice versa. */ |
| |
| #define ADJUST_REG_ALLOC_ORDER x86_order_regs_for_local_alloc () |
| |
| |
| #define OVERRIDE_ABI_FORMAT(FNDECL) ix86_call_abi_override (FNDECL) |
| |
| /* Return number of consecutive hard regs needed starting at reg REGNO |
| to hold something of mode MODE. |
| This is ordinarily the length in words of a value of mode MODE |
| but can be less for certain modes in special long registers. |
| |
| Actually there are no two word move instructions for consecutive |
| registers. And only registers 0-3 may have mov byte instructions |
| applied to them. */ |
| |
| #define HARD_REGNO_NREGS(REGNO, MODE) \ |
| (STACK_REGNO_P (REGNO) || SSE_REGNO_P (REGNO) || MMX_REGNO_P (REGNO) \ |
| || MASK_REGNO_P (REGNO) || BND_REGNO_P (REGNO) \ |
| ? (COMPLEX_MODE_P (MODE) ? 2 : 1) \ |
| : ((MODE) == XFmode \ |
| ? (TARGET_64BIT ? 2 : 3) \ |
| : (MODE) == XCmode \ |
| ? (TARGET_64BIT ? 4 : 6) \ |
| : ((GET_MODE_SIZE (MODE) + UNITS_PER_WORD - 1) / UNITS_PER_WORD))) |
| |
| #define HARD_REGNO_NREGS_HAS_PADDING(REGNO, MODE) \ |
| ((TARGET_128BIT_LONG_DOUBLE && !TARGET_64BIT) \ |
| ? (STACK_REGNO_P (REGNO) || SSE_REGNO_P (REGNO) || MMX_REGNO_P (REGNO) \ |
| ? 0 \ |
| : ((MODE) == XFmode || (MODE) == XCmode)) \ |
| : 0) |
| |
| #define HARD_REGNO_NREGS_WITH_PADDING(REGNO, MODE) ((MODE) == XFmode ? 4 : 8) |
| |
| #define VALID_AVX256_REG_MODE(MODE) \ |
| ((MODE) == V32QImode || (MODE) == V16HImode || (MODE) == V8SImode \ |
| || (MODE) == V4DImode || (MODE) == V2TImode || (MODE) == V8SFmode \ |
| || (MODE) == V4DFmode) |
| |
| #define VALID_AVX256_REG_OR_OI_MODE(MODE) \ |
| (VALID_AVX256_REG_MODE (MODE) || (MODE) == OImode) |
| |
| #define VALID_AVX512F_SCALAR_MODE(MODE) \ |
| ((MODE) == DImode || (MODE) == DFmode || (MODE) == SImode \ |
| || (MODE) == SFmode) |
| |
| #define VALID_AVX512F_REG_MODE(MODE) \ |
| ((MODE) == V8DImode || (MODE) == V8DFmode || (MODE) == V64QImode \ |
| || (MODE) == V16SImode || (MODE) == V16SFmode || (MODE) == V32HImode \ |
| || (MODE) == V4TImode) |
| |
| #define VALID_AVX512VL_128_REG_MODE(MODE) \ |
| ((MODE) == V2DImode || (MODE) == V2DFmode || (MODE) == V16QImode \ |
| || (MODE) == V4SImode || (MODE) == V4SFmode || (MODE) == V8HImode) |
| |
| #define VALID_SSE2_REG_MODE(MODE) \ |
| ((MODE) == V16QImode || (MODE) == V8HImode || (MODE) == V2DFmode \ |
| || (MODE) == V2DImode || (MODE) == DFmode) |
| |
| #define VALID_SSE_REG_MODE(MODE) \ |
| ((MODE) == V1TImode || (MODE) == TImode \ |
| || (MODE) == V4SFmode || (MODE) == V4SImode \ |
| || (MODE) == SFmode || (MODE) == TFmode) |
| |
| #define VALID_MMX_REG_MODE_3DNOW(MODE) \ |
| ((MODE) == V2SFmode || (MODE) == SFmode) |
| |
| #define VALID_MMX_REG_MODE(MODE) \ |
| ((MODE == V1DImode) || (MODE) == DImode \ |
| || (MODE) == V2SImode || (MODE) == SImode \ |
| || (MODE) == V4HImode || (MODE) == V8QImode) |
| |
| #define VALID_BND_REG_MODE(MODE) \ |
| (TARGET_64BIT ? (MODE) == BND64mode : (MODE) == BND32mode) |
| |
| #define VALID_DFP_MODE_P(MODE) \ |
| ((MODE) == SDmode || (MODE) == DDmode || (MODE) == TDmode) |
| |
| #define VALID_FP_MODE_P(MODE) \ |
| ((MODE) == SFmode || (MODE) == DFmode || (MODE) == XFmode \ |
| || (MODE) == SCmode || (MODE) == DCmode || (MODE) == XCmode) \ |
| |
| #define VALID_INT_MODE_P(MODE) \ |
| ((MODE) == QImode || (MODE) == HImode || (MODE) == SImode \ |
| || (MODE) == DImode \ |
| || (MODE) == CQImode || (MODE) == CHImode || (MODE) == CSImode \ |
| || (MODE) == CDImode \ |
| || (TARGET_64BIT && ((MODE) == TImode || (MODE) == CTImode \ |
| || (MODE) == TFmode || (MODE) == TCmode))) |
| |
| /* Return true for modes passed in SSE registers. */ |
| #define SSE_REG_MODE_P(MODE) \ |
| ((MODE) == V1TImode || (MODE) == TImode || (MODE) == V16QImode \ |
| || (MODE) == TFmode || (MODE) == V8HImode || (MODE) == V2DFmode \ |
| || (MODE) == V2DImode || (MODE) == V4SFmode || (MODE) == V4SImode \ |
| || (MODE) == V32QImode || (MODE) == V16HImode || (MODE) == V8SImode \ |
| || (MODE) == V4DImode || (MODE) == V8SFmode || (MODE) == V4DFmode \ |
| || (MODE) == V2TImode || (MODE) == V8DImode || (MODE) == V64QImode \ |
| || (MODE) == V16SImode || (MODE) == V32HImode || (MODE) == V8DFmode \ |
| || (MODE) == V16SFmode) |
| |
| #define VALID_MASK_REG_MODE(MODE) ((MODE) == HImode || (MODE) == QImode) |
| |
| #define VALID_MASK_AVX512BW_MODE(MODE) ((MODE) == SImode || (MODE) == DImode) |
| |
| /* Value is 1 if hard register REGNO can hold a value of machine-mode MODE. */ |
| |
| #define HARD_REGNO_MODE_OK(REGNO, MODE) \ |
| ix86_hard_regno_mode_ok ((REGNO), (MODE)) |
| |
| /* Value is 1 if it is a good idea to tie two pseudo registers |
| when one has mode MODE1 and one has mode MODE2. |
| If HARD_REGNO_MODE_OK could produce different values for MODE1 and MODE2, |
| for any hard reg, then this must be 0 for correct output. */ |
| |
| #define MODES_TIEABLE_P(MODE1, MODE2) ix86_modes_tieable_p (MODE1, MODE2) |
| |
| /* It is possible to write patterns to move flags; but until someone |
| does it, */ |
| #define AVOID_CCMODE_COPIES |
| |
| /* Specify the modes required to caller save a given hard regno. |
| We do this on i386 to prevent flags from being saved at all. |
| |
| Kill any attempts to combine saving of modes. */ |
| |
| #define HARD_REGNO_CALLER_SAVE_MODE(REGNO, NREGS, MODE) \ |
| (CC_REGNO_P (REGNO) ? VOIDmode \ |
| : (MODE) == VOIDmode && (NREGS) != 1 ? VOIDmode \ |
| : (MODE) == VOIDmode ? choose_hard_reg_mode ((REGNO), (NREGS), false) \ |
| : (MODE) == HImode && !(TARGET_PARTIAL_REG_STALL \ |
| || MASK_REGNO_P (REGNO)) ? SImode \ |
| : (MODE) == QImode && !(TARGET_64BIT || QI_REGNO_P (REGNO) \ |
| || MASK_REGNO_P (REGNO)) ? SImode \ |
| : (MODE)) |
| |
| /* The only ABI that saves SSE registers across calls is Win64 (thus no |
| need to check the current ABI here), and with AVX enabled Win64 only |
| guarantees that the low 16 bytes are saved. */ |
| #define HARD_REGNO_CALL_PART_CLOBBERED(REGNO, MODE) \ |
| (SSE_REGNO_P (REGNO) && GET_MODE_SIZE (MODE) > 16) |
| |
| /* Specify the registers used for certain standard purposes. |
| The values of these macros are register numbers. */ |
| |
| /* on the 386 the pc register is %eip, and is not usable as a general |
| register. The ordinary mov instructions won't work */ |
| /* #define PC_REGNUM */ |
| |
| /* Register to use for pushing function arguments. */ |
| #define STACK_POINTER_REGNUM 7 |
| |
| /* Base register for access to local variables of the function. */ |
| #define HARD_FRAME_POINTER_REGNUM 6 |
| |
| /* Base register for access to local variables of the function. */ |
| #define FRAME_POINTER_REGNUM 20 |
| |
| /* First floating point reg */ |
| #define FIRST_FLOAT_REG 8 |
| |
| /* First & last stack-like regs */ |
| #define FIRST_STACK_REG FIRST_FLOAT_REG |
| #define LAST_STACK_REG (FIRST_FLOAT_REG + 7) |
| |
| #define FIRST_SSE_REG (FRAME_POINTER_REGNUM + 1) |
| #define LAST_SSE_REG (FIRST_SSE_REG + 7) |
| |
| #define FIRST_MMX_REG (LAST_SSE_REG + 1) /*29*/ |
| #define LAST_MMX_REG (FIRST_MMX_REG + 7) |
| |
| #define FIRST_REX_INT_REG (LAST_MMX_REG + 1) /*37*/ |
| #define LAST_REX_INT_REG (FIRST_REX_INT_REG + 7) |
| |
| #define FIRST_REX_SSE_REG (LAST_REX_INT_REG + 1) /*45*/ |
| #define LAST_REX_SSE_REG (FIRST_REX_SSE_REG + 7) |
| |
| #define FIRST_EXT_REX_SSE_REG (LAST_REX_SSE_REG + 1) /*53*/ |
| #define LAST_EXT_REX_SSE_REG (FIRST_EXT_REX_SSE_REG + 15) /*68*/ |
| |
| #define FIRST_MASK_REG (LAST_EXT_REX_SSE_REG + 1) /*69*/ |
| #define LAST_MASK_REG (FIRST_MASK_REG + 7) /*76*/ |
| |
| #define FIRST_BND_REG (LAST_MASK_REG + 1) /*77*/ |
| #define LAST_BND_REG (FIRST_BND_REG + 3) /*80*/ |
| |
| /* Override this in other tm.h files to cope with various OS lossage |
| requiring a frame pointer. */ |
| #ifndef SUBTARGET_FRAME_POINTER_REQUIRED |
| #define SUBTARGET_FRAME_POINTER_REQUIRED 0 |
| #endif |
| |
| /* Make sure we can access arbitrary call frames. */ |
| #define SETUP_FRAME_ADDRESSES() ix86_setup_frame_addresses () |
| |
| /* Base register for access to arguments of the function. */ |
| #define ARG_POINTER_REGNUM 16 |
| |
| /* Register to hold the addressing base for position independent |
| code access to data items. We don't use PIC pointer for 64bit |
| mode. Define the regnum to dummy value to prevent gcc from |
| pessimizing code dealing with EBX. |
| |
| To avoid clobbering a call-saved register unnecessarily, we renumber |
| the pic register when possible. The change is visible after the |
| prologue has been emitted. */ |
| |
| #define REAL_PIC_OFFSET_TABLE_REGNUM (TARGET_64BIT ? R15_REG : BX_REG) |
| |
| #define PIC_OFFSET_TABLE_REGNUM \ |
| (ix86_use_pseudo_pic_reg () \ |
| ? (pic_offset_table_rtx \ |
| ? INVALID_REGNUM \ |
| : REAL_PIC_OFFSET_TABLE_REGNUM) \ |
| : INVALID_REGNUM) |
| |
| #define GOT_SYMBOL_NAME "_GLOBAL_OFFSET_TABLE_" |
| |
| /* This is overridden by <cygwin.h>. */ |
| #define MS_AGGREGATE_RETURN 0 |
| |
| #define KEEP_AGGREGATE_RETURN_POINTER 0 |
| |
| /* Define the classes of registers for register constraints in the |
| machine description. Also define ranges of constants. |
| |
| One of the classes must always be named ALL_REGS and include all hard regs. |
| If there is more than one class, another class must be named NO_REGS |
| and contain no registers. |
| |
| The name GENERAL_REGS must be the name of a class (or an alias for |
| another name such as ALL_REGS). This is the class of registers |
| that is allowed by "g" or "r" in a register constraint. |
| Also, registers outside this class are allocated only when |
| instructions express preferences for them. |
| |
| The classes must be numbered in nondecreasing order; that is, |
| a larger-numbered class must never be contained completely |
| in a smaller-numbered class. |
| |
| For any two classes, it is very desirable that there be another |
| class that represents their union. |
| |
| It might seem that class BREG is unnecessary, since no useful 386 |
| opcode needs reg %ebx. But some systems pass args to the OS in ebx, |
| and the "b" register constraint is useful in asms for syscalls. |
| |
| The flags, fpsr and fpcr registers are in no class. */ |
| |
| enum reg_class |
| { |
| NO_REGS, |
| AREG, DREG, CREG, BREG, SIREG, DIREG, |
| AD_REGS, /* %eax/%edx for DImode */ |
| Q_REGS, /* %eax %ebx %ecx %edx */ |
| NON_Q_REGS, /* %esi %edi %ebp %esp */ |
| INDEX_REGS, /* %eax %ebx %ecx %edx %esi %edi %ebp */ |
| LEGACY_REGS, /* %eax %ebx %ecx %edx %esi %edi %ebp %esp */ |
| CLOBBERED_REGS, /* call-clobbered integer registers */ |
| GENERAL_REGS, /* %eax %ebx %ecx %edx %esi %edi %ebp %esp |
| %r8 %r9 %r10 %r11 %r12 %r13 %r14 %r15 */ |
| FP_TOP_REG, FP_SECOND_REG, /* %st(0) %st(1) */ |
| FLOAT_REGS, |
| SSE_FIRST_REG, |
| NO_REX_SSE_REGS, |
| SSE_REGS, |
| EVEX_SSE_REGS, |
| BND_REGS, |
| ALL_SSE_REGS, |
| MMX_REGS, |
| FP_TOP_SSE_REGS, |
| FP_SECOND_SSE_REGS, |
| FLOAT_SSE_REGS, |
| FLOAT_INT_REGS, |
| INT_SSE_REGS, |
| FLOAT_INT_SSE_REGS, |
| MASK_EVEX_REGS, |
| MASK_REGS, |
| ALL_REGS, LIM_REG_CLASSES |
| }; |
| |
| #define N_REG_CLASSES ((int) LIM_REG_CLASSES) |
| |
| #define INTEGER_CLASS_P(CLASS) \ |
| reg_class_subset_p ((CLASS), GENERAL_REGS) |
| #define FLOAT_CLASS_P(CLASS) \ |
| reg_class_subset_p ((CLASS), FLOAT_REGS) |
| #define SSE_CLASS_P(CLASS) \ |
| reg_class_subset_p ((CLASS), ALL_SSE_REGS) |
| #define MMX_CLASS_P(CLASS) \ |
| ((CLASS) == MMX_REGS) |
| #define MAYBE_INTEGER_CLASS_P(CLASS) \ |
| reg_classes_intersect_p ((CLASS), GENERAL_REGS) |
| #define MAYBE_FLOAT_CLASS_P(CLASS) \ |
| reg_classes_intersect_p ((CLASS), FLOAT_REGS) |
| #define MAYBE_SSE_CLASS_P(CLASS) \ |
| reg_classes_intersect_p ((CLASS), ALL_SSE_REGS) |
| #define MAYBE_MMX_CLASS_P(CLASS) \ |
| reg_classes_intersect_p ((CLASS), MMX_REGS) |
| #define MAYBE_MASK_CLASS_P(CLASS) \ |
| reg_classes_intersect_p ((CLASS), MASK_REGS) |
| |
| #define Q_CLASS_P(CLASS) \ |
| reg_class_subset_p ((CLASS), Q_REGS) |
| |
| #define MAYBE_NON_Q_CLASS_P(CLASS) \ |
| reg_classes_intersect_p ((CLASS), NON_Q_REGS) |
| |
| /* Give names of register classes as strings for dump file. */ |
| |
| #define REG_CLASS_NAMES \ |
| { "NO_REGS", \ |
| "AREG", "DREG", "CREG", "BREG", \ |
| "SIREG", "DIREG", \ |
| "AD_REGS", \ |
| "Q_REGS", "NON_Q_REGS", \ |
| "INDEX_REGS", \ |
| "LEGACY_REGS", \ |
| "CLOBBERED_REGS", \ |
| "GENERAL_REGS", \ |
| "FP_TOP_REG", "FP_SECOND_REG", \ |
| "FLOAT_REGS", \ |
| "SSE_FIRST_REG", \ |
| "NO_REX_SSE_REGS", \ |
| "SSE_REGS", \ |
| "EVEX_SSE_REGS", \ |
| "BND_REGS", \ |
| "ALL_SSE_REGS", \ |
| "MMX_REGS", \ |
| "FP_TOP_SSE_REGS", \ |
| "FP_SECOND_SSE_REGS", \ |
| "FLOAT_SSE_REGS", \ |
| "FLOAT_INT_REGS", \ |
| "INT_SSE_REGS", \ |
| "FLOAT_INT_SSE_REGS", \ |
| "MASK_EVEX_REGS", \ |
| "MASK_REGS", \ |
| "ALL_REGS" } |
| |
| /* Define which registers fit in which classes. This is an initializer |
| for a vector of HARD_REG_SET of length N_REG_CLASSES. |
| |
| Note that CLOBBERED_REGS are calculated by |
| TARGET_CONDITIONAL_REGISTER_USAGE. */ |
| |
| #define REG_CLASS_CONTENTS \ |
| { { 0x00, 0x0, 0x0 }, \ |
| { 0x01, 0x0, 0x0 }, /* AREG */ \ |
| { 0x02, 0x0, 0x0 }, /* DREG */ \ |
| { 0x04, 0x0, 0x0 }, /* CREG */ \ |
| { 0x08, 0x0, 0x0 }, /* BREG */ \ |
| { 0x10, 0x0, 0x0 }, /* SIREG */ \ |
| { 0x20, 0x0, 0x0 }, /* DIREG */ \ |
| { 0x03, 0x0, 0x0 }, /* AD_REGS */ \ |
| { 0x0f, 0x0, 0x0 }, /* Q_REGS */ \ |
| { 0x1100f0, 0x1fe0, 0x0 }, /* NON_Q_REGS */ \ |
| { 0x7f, 0x1fe0, 0x0 }, /* INDEX_REGS */ \ |
| { 0x1100ff, 0x0, 0x0 }, /* LEGACY_REGS */ \ |
| { 0x07, 0x0, 0x0 }, /* CLOBBERED_REGS */ \ |
| { 0x1100ff, 0x1fe0, 0x0 }, /* GENERAL_REGS */ \ |
| { 0x100, 0x0, 0x0 }, /* FP_TOP_REG */ \ |
| { 0x0200, 0x0, 0x0 }, /* FP_SECOND_REG */ \ |
| { 0xff00, 0x0, 0x0 }, /* FLOAT_REGS */ \ |
| { 0x200000, 0x0, 0x0 }, /* SSE_FIRST_REG */ \ |
| { 0x1fe00000, 0x000000, 0x0 }, /* NO_REX_SSE_REGS */ \ |
| { 0x1fe00000, 0x1fe000, 0x0 }, /* SSE_REGS */ \ |
| { 0x0,0xffe00000, 0x1f }, /* EVEX_SSE_REGS */ \ |
| { 0x0, 0x0,0x1e000 }, /* BND_REGS */ \ |
| { 0x1fe00000,0xffffe000, 0x1f }, /* ALL_SSE_REGS */ \ |
| { 0xe0000000, 0x1f, 0x0 }, /* MMX_REGS */ \ |
| { 0x1fe00100,0xffffe000, 0x1f }, /* FP_TOP_SSE_REG */ \ |
| { 0x1fe00200,0xffffe000, 0x1f }, /* FP_SECOND_SSE_REG */ \ |
| { 0x1fe0ff00,0xffffe000, 0x1f }, /* FLOAT_SSE_REGS */ \ |
| { 0x11ffff, 0x1fe0, 0x0 }, /* FLOAT_INT_REGS */ \ |
| { 0x1ff100ff,0xffffffe0, 0x1f }, /* INT_SSE_REGS */ \ |
| { 0x1ff1ffff,0xffffffe0, 0x1f }, /* FLOAT_INT_SSE_REGS */ \ |
| { 0x0, 0x0, 0x1fc0 }, /* MASK_EVEX_REGS */ \ |
| { 0x0, 0x0, 0x1fe0 }, /* MASK_REGS */ \ |
| { 0xffffffff,0xffffffff,0x1ffff } \ |
| } |
| |
| /* The same information, inverted: |
| Return the class number of the smallest class containing |
| reg number REGNO. This could be a conditional expression |
| or could index an array. */ |
| |
| #define REGNO_REG_CLASS(REGNO) (regclass_map[REGNO]) |
| |
| /* When this hook returns true for MODE, the compiler allows |
| registers explicitly used in the rtl to be used as spill registers |
| but prevents the compiler from extending the lifetime of these |
| registers. */ |
| #define TARGET_SMALL_REGISTER_CLASSES_FOR_MODE_P hook_bool_mode_true |
| |
| #define QI_REG_P(X) (REG_P (X) && QI_REGNO_P (REGNO (X))) |
| #define QI_REGNO_P(N) IN_RANGE ((N), AX_REG, BX_REG) |
| |
| #define GENERAL_REG_P(X) \ |
| (REG_P (X) && GENERAL_REGNO_P (REGNO (X))) |
| #define GENERAL_REGNO_P(N) \ |
| (IN_RANGE ((N), AX_REG, SP_REG) || REX_INT_REGNO_P (N)) |
| |
| #define ANY_QI_REG_P(X) (REG_P (X) && ANY_QI_REGNO_P (REGNO (X))) |
| #define ANY_QI_REGNO_P(N) \ |
| (TARGET_64BIT ? GENERAL_REGNO_P (N) : QI_REGNO_P (N)) |
| |
| #define REX_INT_REG_P(X) (REG_P (X) && REX_INT_REGNO_P (REGNO (X))) |
| #define REX_INT_REGNO_P(N) \ |
| IN_RANGE ((N), FIRST_REX_INT_REG, LAST_REX_INT_REG) |
| |
| #define STACK_REG_P(X) (REG_P (X) && STACK_REGNO_P (REGNO (X))) |
| #define STACK_REGNO_P(N) IN_RANGE ((N), FIRST_STACK_REG, LAST_STACK_REG) |
| |
| #define ANY_FP_REG_P(X) (REG_P (X) && ANY_FP_REGNO_P (REGNO (X))) |
| #define ANY_FP_REGNO_P(N) (STACK_REGNO_P (N) || SSE_REGNO_P (N)) |
| |
| #define X87_FLOAT_MODE_P(MODE) \ |
| (TARGET_80387 && ((MODE) == SFmode || (MODE) == DFmode || (MODE) == XFmode)) |
| |
| #define SSE_REG_P(X) (REG_P (X) && SSE_REGNO_P (REGNO (X))) |
| #define SSE_REGNO_P(N) \ |
| (IN_RANGE ((N), FIRST_SSE_REG, LAST_SSE_REG) \ |
| || REX_SSE_REGNO_P (N) \ |
| || EXT_REX_SSE_REGNO_P (N)) |
| |
| #define REX_SSE_REGNO_P(N) \ |
| IN_RANGE ((N), FIRST_REX_SSE_REG, LAST_REX_SSE_REG) |
| |
| #define EXT_REX_SSE_REGNO_P(N) \ |
| IN_RANGE ((N), FIRST_EXT_REX_SSE_REG, LAST_EXT_REX_SSE_REG) |
| |
| #define SSE_REGNO(N) \ |
| ((N) < 8 ? FIRST_SSE_REG + (N) \ |
| : (N) <= LAST_REX_SSE_REG ? (FIRST_REX_SSE_REG + (N) - 8) \ |
| : (FIRST_EXT_REX_SSE_REG + (N) - 16)) |
| |
| #define MASK_REG_P(X) (REG_P (X) && MASK_REGNO_P (REGNO (X))) |
| #define MASK_REGNO_P(N) IN_RANGE ((N), FIRST_MASK_REG, LAST_MASK_REG) |
| #define ANY_MASK_REG_P(X) (REG_P (X) && MASK_REGNO_P (REGNO (X))) |
| |
| #define SSE_FLOAT_MODE_P(MODE) \ |
| ((TARGET_SSE && (MODE) == SFmode) || (TARGET_SSE2 && (MODE) == DFmode)) |
| |
| #define FMA4_VEC_FLOAT_MODE_P(MODE) \ |
| (TARGET_FMA4 && ((MODE) == V4SFmode || (MODE) == V2DFmode \ |
| || (MODE) == V8SFmode || (MODE) == V4DFmode)) |
| |
| #define MMX_REG_P(X) (REG_P (X) && MMX_REGNO_P (REGNO (X))) |
| #define MMX_REGNO_P(N) IN_RANGE ((N), FIRST_MMX_REG, LAST_MMX_REG) |
| |
| #define STACK_TOP_P(X) (REG_P (X) && REGNO (X) == FIRST_STACK_REG) |
| |
| #define CC_REG_P(X) (REG_P (X) && CC_REGNO_P (REGNO (X))) |
| #define CC_REGNO_P(X) ((X) == FLAGS_REG || (X) == FPSR_REG) |
| |
| #define BND_REGNO_P(N) IN_RANGE ((N), FIRST_BND_REG, LAST_BND_REG) |
| #define ANY_BND_REG_P(X) (REG_P (X) && BND_REGNO_P (REGNO (X))) |
| |
| /* The class value for index registers, and the one for base regs. */ |
| |
| #define INDEX_REG_CLASS INDEX_REGS |
| #define BASE_REG_CLASS GENERAL_REGS |
| |
| /* Place additional restrictions on the register class to use when it |
| is necessary to be able to hold a value of mode MODE in a reload |
| register for which class CLASS would ordinarily be used. |
| |
| We avoid classes containing registers from multiple units due to |
| the limitation in ix86_secondary_memory_needed. We limit these |
| classes to their "natural mode" single unit register class, depending |
| on the unit availability. |
| |
| Please note that reg_class_subset_p is not commutative, so these |
| conditions mean "... if (CLASS) includes ALL registers from the |
| register set." */ |
| |
| #define LIMIT_RELOAD_CLASS(MODE, CLASS) \ |
| (((MODE) == QImode && !TARGET_64BIT \ |
| && reg_class_subset_p (Q_REGS, (CLASS))) ? Q_REGS \ |
| : (((MODE) == SImode || (MODE) == DImode) \ |
| && reg_class_subset_p (GENERAL_REGS, (CLASS))) ? GENERAL_REGS \ |
| : (SSE_FLOAT_MODE_P (MODE) && TARGET_SSE_MATH \ |
| && reg_class_subset_p (SSE_REGS, (CLASS))) ? SSE_REGS \ |
| : (X87_FLOAT_MODE_P (MODE) \ |
| && reg_class_subset_p (FLOAT_REGS, (CLASS))) ? FLOAT_REGS \ |
| : (CLASS)) |
| |
| /* If we are copying between general and FP registers, we need a memory |
| location. The same is true for SSE and MMX registers. */ |
| #define SECONDARY_MEMORY_NEEDED(CLASS1, CLASS2, MODE) \ |
| ix86_secondary_memory_needed ((CLASS1), (CLASS2), (MODE), 1) |
| |
| /* Get_secondary_mem widens integral modes to BITS_PER_WORD. |
| There is no need to emit full 64 bit move on 64 bit targets |
| for integral modes that can be moved using 32 bit move. */ |
| #define SECONDARY_MEMORY_NEEDED_MODE(MODE) \ |
| (GET_MODE_BITSIZE (MODE) < 32 && INTEGRAL_MODE_P (MODE) \ |
| ? mode_for_size (32, GET_MODE_CLASS (MODE), 0) \ |
| : MODE) |
| |
| /* Return a class of registers that cannot change FROM mode to TO mode. */ |
| |
| #define CANNOT_CHANGE_MODE_CLASS(FROM, TO, CLASS) \ |
| ix86_cannot_change_mode_class (FROM, TO, CLASS) |
| |
| /* Stack layout; function entry, exit and calling. */ |
| |
| /* Define this if pushing a word on the stack |
| makes the stack pointer a smaller address. */ |
| #define STACK_GROWS_DOWNWARD |
| |
| /* Define this to nonzero if the nominal address of the stack frame |
| is at the high-address end of the local variables; |
| that is, each additional local variable allocated |
| goes at a more negative offset in the frame. */ |
| #define FRAME_GROWS_DOWNWARD 1 |
| |
| /* Offset within stack frame to start allocating local variables at. |
| If FRAME_GROWS_DOWNWARD, this is the offset to the END of the |
| first local allocated. Otherwise, it is the offset to the BEGINNING |
| of the first local allocated. */ |
| #define STARTING_FRAME_OFFSET 0 |
| |
| /* If we generate an insn to push BYTES bytes, this says how many the stack |
| pointer really advances by. On 386, we have pushw instruction that |
| decrements by exactly 2 no matter what the position was, there is no pushb. |
| |
| But as CIE data alignment factor on this arch is -4 for 32bit targets |
| and -8 for 64bit targets, we need to make sure all stack pointer adjustments |
| are in multiple of 4 for 32bit targets and 8 for 64bit targets. */ |
| |
| #define PUSH_ROUNDING(BYTES) \ |
| (((BYTES) + UNITS_PER_WORD - 1) & -UNITS_PER_WORD) |
| |
| /* If defined, the maximum amount of space required for outgoing arguments |
| will be computed and placed into the variable `crtl->outgoing_args_size'. |
| No space will be pushed onto the stack for each call; instead, the |
| function prologue should increase the stack frame size by this amount. |
| |
| In 32bit mode enabling argument accumulation results in about 5% code size |
| growth becuase move instructions are less compact than push. In 64bit |
| mode the difference is less drastic but visible. |
| |
| FIXME: Unlike earlier implementations, the size of unwind info seems to |
| actually grow with accumulation. Is that because accumulated args |
| unwind info became unnecesarily bloated? |
| |
| With the 64-bit MS ABI, we can generate correct code with or without |
| accumulated args, but because of OUTGOING_REG_PARM_STACK_SPACE the code |
| generated without accumulated args is terrible. |
| |
| If stack probes are required, the space used for large function |
| arguments on the stack must also be probed, so enable |
| -maccumulate-outgoing-args so this happens in the prologue. */ |
| |
| #define ACCUMULATE_OUTGOING_ARGS \ |
| ((TARGET_ACCUMULATE_OUTGOING_ARGS && optimize_function_for_speed_p (cfun)) \ |
| || TARGET_STACK_PROBE || TARGET_64BIT_MS_ABI) |
| |
| /* If defined, a C expression whose value is nonzero when we want to use PUSH |
| instructions to pass outgoing arguments. */ |
| |
| #define PUSH_ARGS (TARGET_PUSH_ARGS && !ACCUMULATE_OUTGOING_ARGS) |
| |
| /* We want the stack and args grow in opposite directions, even if |
| PUSH_ARGS is 0. */ |
| #define PUSH_ARGS_REVERSED 1 |
| |
| /* Offset of first parameter from the argument pointer register value. */ |
| #define FIRST_PARM_OFFSET(FNDECL) 0 |
| |
| /* Define this macro if functions should assume that stack space has been |
| allocated for arguments even when their values are passed in registers. |
| |
| The value of this macro is the size, in bytes, of the area reserved for |
| arguments passed in registers for the function represented by FNDECL. |
| |
| This space can be allocated by the caller, or be a part of the |
| machine-dependent stack frame: `OUTGOING_REG_PARM_STACK_SPACE' says |
| which. */ |
| #define REG_PARM_STACK_SPACE(FNDECL) ix86_reg_parm_stack_space (FNDECL) |
| |
| #define OUTGOING_REG_PARM_STACK_SPACE(FNTYPE) \ |
| (TARGET_64BIT && ix86_function_type_abi (FNTYPE) == MS_ABI) |
| |
| /* Define how to find the value returned by a library function |
| assuming the value has mode MODE. */ |
| |
| #define LIBCALL_VALUE(MODE) ix86_libcall_value (MODE) |
| |
| /* Define the size of the result block used for communication between |
| untyped_call and untyped_return. The block contains a DImode value |
| followed by the block used by fnsave and frstor. */ |
| |
| #define APPLY_RESULT_SIZE (8+108) |
| |
| /* 1 if N is a possible register number for function argument passing. */ |
| #define FUNCTION_ARG_REGNO_P(N) ix86_function_arg_regno_p (N) |
| |
| /* Define a data type for recording info about an argument list |
| during the scan of that argument list. This data type should |
| hold all necessary information about the function itself |
| and about the args processed so far, enough to enable macros |
| such as FUNCTION_ARG to determine where the next arg should go. */ |
| |
| typedef struct ix86_args { |
| int words; /* # words passed so far */ |
| int nregs; /* # registers available for passing */ |
| int regno; /* next available register number */ |
| int fastcall; /* fastcall or thiscall calling convention |
| is used */ |
| int sse_words; /* # sse words passed so far */ |
| int sse_nregs; /* # sse registers available for passing */ |
| int warn_avx512f; /* True when we want to warn |
| about AVX512F ABI. */ |
| int warn_avx; /* True when we want to warn about AVX ABI. */ |
| int warn_sse; /* True when we want to warn about SSE ABI. */ |
| int warn_mmx; /* True when we want to warn about MMX ABI. */ |
| int sse_regno; /* next available sse register number */ |
| int mmx_words; /* # mmx words passed so far */ |
| int mmx_nregs; /* # mmx registers available for passing */ |
| int mmx_regno; /* next available mmx register number */ |
| int maybe_vaarg; /* true for calls to possibly vardic fncts. */ |
| int caller; /* true if it is caller. */ |
| int float_in_sse; /* Set to 1 or 2 for 32bit targets if |
| SFmode/DFmode arguments should be passed |
| in SSE registers. Otherwise 0. */ |
| int bnd_regno; /* next available bnd register number */ |
| int bnds_in_bt; /* number of bounds expected in BT. */ |
| int force_bnd_pass; /* number of bounds expected for stdarg arg. */ |
| int stdarg; /* Set to 1 if function is stdarg. */ |
| enum calling_abi call_abi; /* Set to SYSV_ABI for sysv abi. Otherwise |
| MS_ABI for ms abi. */ |
| tree decl; /* Callee decl. */ |
| } CUMULATIVE_ARGS; |
| |
| /* Initialize a variable CUM of type CUMULATIVE_ARGS |
| for a call to a function whose data type is FNTYPE. |
| For a library call, FNTYPE is 0. */ |
| |
| #define INIT_CUMULATIVE_ARGS(CUM, FNTYPE, LIBNAME, FNDECL, N_NAMED_ARGS) \ |
| init_cumulative_args (&(CUM), (FNTYPE), (LIBNAME), (FNDECL), \ |
| (N_NAMED_ARGS) != -1) |
| |
| /* Output assembler code to FILE to increment profiler label # LABELNO |
| for profiling a function entry. */ |
| |
| #define FUNCTION_PROFILER(FILE, LABELNO) x86_function_profiler (FILE, LABELNO) |
| |
| #define MCOUNT_NAME "_mcount" |
| |
| #define MCOUNT_NAME_BEFORE_PROLOGUE "__fentry__" |
| |
| #define PROFILE_COUNT_REGISTER "edx" |
| |
| /* EXIT_IGNORE_STACK should be nonzero if, when returning from a function, |
| the stack pointer does not matter. The value is tested only in |
| functions that have frame pointers. |
| No definition is equivalent to always zero. */ |
| /* Note on the 386 it might be more efficient not to define this since |
| we have to restore it ourselves from the frame pointer, in order to |
| use pop */ |
| |
| #define EXIT_IGNORE_STACK 1 |
| |
| /* Output assembler code for a block containing the constant parts |
| of a trampoline, leaving space for the variable parts. */ |
| |
| /* On the 386, the trampoline contains two instructions: |
| mov #STATIC,ecx |
| jmp FUNCTION |
| The trampoline is generated entirely at runtime. The operand of JMP |
| is the address of FUNCTION relative to the instruction following the |
| JMP (which is 5 bytes long). */ |
| |
| /* Length in units of the trampoline for entering a nested function. */ |
| |
| #define TRAMPOLINE_SIZE (TARGET_64BIT ? 24 : 10) |
| |
| /* Definitions for register eliminations. |
| |
| This is an array of structures. Each structure initializes one pair |
| of eliminable registers. The "from" register number is given first, |
| followed by "to". Eliminations of the same "from" register are listed |
| in order of preference. |
| |
| There are two registers that can always be eliminated on the i386. |
| The frame pointer and the arg pointer can be replaced by either the |
| hard frame pointer or to the stack pointer, depending upon the |
| circumstances. The hard frame pointer is not used before reload and |
| so it is not eligible for elimination. */ |
| |
| #define ELIMINABLE_REGS \ |
| {{ ARG_POINTER_REGNUM, STACK_POINTER_REGNUM}, \ |
| { ARG_POINTER_REGNUM, HARD_FRAME_POINTER_REGNUM}, \ |
| { FRAME_POINTER_REGNUM, STACK_POINTER_REGNUM}, \ |
| { FRAME_POINTER_REGNUM, HARD_FRAME_POINTER_REGNUM}} \ |
| |
| /* Define the offset between two registers, one to be eliminated, and the other |
| its replacement, at the start of a routine. */ |
| |
| #define INITIAL_ELIMINATION_OFFSET(FROM, TO, OFFSET) \ |
| ((OFFSET) = ix86_initial_elimination_offset ((FROM), (TO))) |
| |
| /* Addressing modes, and classification of registers for them. */ |
| |
| /* Macros to check register numbers against specific register classes. */ |
| |
| /* These assume that REGNO is a hard or pseudo reg number. |
| They give nonzero only if REGNO is a hard reg of the suitable class |
| or a pseudo reg currently allocated to a suitable hard reg. |
| Since they use reg_renumber, they are safe only once reg_renumber |
| has been allocated, which happens in reginfo.c during register |
| allocation. */ |
| |
| #define REGNO_OK_FOR_INDEX_P(REGNO) \ |
| ((REGNO) < STACK_POINTER_REGNUM \ |
| || REX_INT_REGNO_P (REGNO) \ |
| || (unsigned) reg_renumber[(REGNO)] < STACK_POINTER_REGNUM \ |
| || REX_INT_REGNO_P ((unsigned) reg_renumber[(REGNO)])) |
| |
| #define REGNO_OK_FOR_BASE_P(REGNO) \ |
| (GENERAL_REGNO_P (REGNO) \ |
| || (REGNO) == ARG_POINTER_REGNUM \ |
| || (REGNO) == FRAME_POINTER_REGNUM \ |
| || GENERAL_REGNO_P ((unsigned) reg_renumber[(REGNO)])) |
| |
| /* The macros REG_OK_FOR..._P assume that the arg is a REG rtx |
| and check its validity for a certain class. |
| We have two alternate definitions for each of them. |
| The usual definition accepts all pseudo regs; the other rejects |
| them unless they have been allocated suitable hard regs. |
| The symbol REG_OK_STRICT causes the latter definition to be used. |
| |
| Most source files want to accept pseudo regs in the hope that |
| they will get allocated to the class that the insn wants them to be in. |
| Source files for reload pass need to be strict. |
| After reload, it makes no difference, since pseudo regs have |
| been eliminated by then. */ |
| |
| |
| /* Non strict versions, pseudos are ok. */ |
| #define REG_OK_FOR_INDEX_NONSTRICT_P(X) \ |
| (REGNO (X) < STACK_POINTER_REGNUM \ |
| || REX_INT_REGNO_P (REGNO (X)) \ |
| || REGNO (X) >= FIRST_PSEUDO_REGISTER) |
| |
| #define REG_OK_FOR_BASE_NONSTRICT_P(X) \ |
| (GENERAL_REGNO_P (REGNO (X)) \ |
| || REGNO (X) == ARG_POINTER_REGNUM \ |
| || REGNO (X) == FRAME_POINTER_REGNUM \ |
| || REGNO (X) >= FIRST_PSEUDO_REGISTER) |
| |
| /* Strict versions, hard registers only */ |
| #define REG_OK_FOR_INDEX_STRICT_P(X) REGNO_OK_FOR_INDEX_P (REGNO (X)) |
| #define REG_OK_FOR_BASE_STRICT_P(X) REGNO_OK_FOR_BASE_P (REGNO (X)) |
| |
| #ifndef REG_OK_STRICT |
| #define REG_OK_FOR_INDEX_P(X) REG_OK_FOR_INDEX_NONSTRICT_P (X) |
| #define REG_OK_FOR_BASE_P(X) REG_OK_FOR_BASE_NONSTRICT_P (X) |
| |
| #else |
| #define REG_OK_FOR_INDEX_P(X) REG_OK_FOR_INDEX_STRICT_P (X) |
| #define REG_OK_FOR_BASE_P(X) REG_OK_FOR_BASE_STRICT_P (X) |
| #endif |
| |
| /* TARGET_LEGITIMATE_ADDRESS_P recognizes an RTL expression |
| that is a valid memory address for an instruction. |
| The MODE argument is the machine mode for the MEM expression |
| that wants to use this address. |
| |
| The other macros defined here are used only in TARGET_LEGITIMATE_ADDRESS_P, |
| except for CONSTANT_ADDRESS_P which is usually machine-independent. |
| |
| See legitimize_pic_address in i386.c for details as to what |
| constitutes a legitimate address when -fpic is used. */ |
| |
| #define MAX_REGS_PER_ADDRESS 2 |
| |
| #define CONSTANT_ADDRESS_P(X) constant_address_p (X) |
| |
| /* Try a machine-dependent way of reloading an illegitimate address |
| operand. If we find one, push the reload and jump to WIN. This |
| macro is used in only one place: `find_reloads_address' in reload.c. */ |
| |
| #define LEGITIMIZE_RELOAD_ADDRESS(X, MODE, OPNUM, TYPE, INDL, WIN) \ |
| do { \ |
| if (ix86_legitimize_reload_address ((X), (MODE), (OPNUM), \ |
| (int)(TYPE), (INDL))) \ |
| goto WIN; \ |
| } while (0) |
| |
| /* If defined, a C expression to determine the base term of address X. |
| This macro is used in only one place: `find_base_term' in alias.c. |
| |
| It is always safe for this macro to not be defined. It exists so |
| that alias analysis can understand machine-dependent addresses. |
| |
| The typical use of this macro is to handle addresses containing |
| a label_ref or symbol_ref within an UNSPEC. */ |
| |
| #define FIND_BASE_TERM(X) ix86_find_base_term (X) |
| |
| /* Nonzero if the constant value X is a legitimate general operand |
| when generating PIC code. It is given that flag_pic is on and |
| that X satisfies CONSTANT_P or is a CONST_DOUBLE. */ |
| |
| #define LEGITIMATE_PIC_OPERAND_P(X) legitimate_pic_operand_p (X) |
| |
| #define SYMBOLIC_CONST(X) \ |
| (GET_CODE (X) == SYMBOL_REF \ |
| || GET_CODE (X) == LABEL_REF \ |
| || (GET_CODE (X) == CONST && symbolic_reference_mentioned_p (X))) |
| |
| /* Max number of args passed in registers. If this is more than 3, we will |
| have problems with ebx (register #4), since it is a caller save register and |
| is also used as the pic register in ELF. So for now, don't allow more than |
| 3 registers to be passed in registers. */ |
| |
| /* Abi specific values for REGPARM_MAX and SSE_REGPARM_MAX */ |
| #define X86_64_REGPARM_MAX 6 |
| #define X86_64_MS_REGPARM_MAX 4 |
| |
| #define X86_32_REGPARM_MAX 3 |
| |
| #define REGPARM_MAX \ |
| (TARGET_64BIT \ |
| ? (TARGET_64BIT_MS_ABI \ |
| ? X86_64_MS_REGPARM_MAX \ |
| : X86_64_REGPARM_MAX) \ |
| : X86_32_REGPARM_MAX) |
| |
| #define X86_64_SSE_REGPARM_MAX 8 |
| #define X86_64_MS_SSE_REGPARM_MAX 4 |
| |
| #define X86_32_SSE_REGPARM_MAX (TARGET_SSE ? (TARGET_MACHO ? 4 : 3) : 0) |
| |
| #define SSE_REGPARM_MAX \ |
| (TARGET_64BIT \ |
| ? (TARGET_64BIT_MS_ABI \ |
| ? X86_64_MS_SSE_REGPARM_MAX \ |
| : X86_64_SSE_REGPARM_MAX) \ |
| : X86_32_SSE_REGPARM_MAX) |
| |
| #define MMX_REGPARM_MAX (TARGET_64BIT ? 0 : (TARGET_MMX ? 3 : 0)) |
| |
| /* Specify the machine mode that this machine uses |
| for the index in the tablejump instruction. */ |
| #define CASE_VECTOR_MODE \ |
| (!TARGET_LP64 || (flag_pic && ix86_cmodel != CM_LARGE_PIC) ? SImode : DImode) |
| |
| /* Define this as 1 if `char' should by default be signed; else as 0. */ |
| #define DEFAULT_SIGNED_CHAR 1 |
| |
| /* Max number of bytes we can move from memory to memory |
| in one reasonably fast instruction. */ |
| #define MOVE_MAX 16 |
| |
| /* MOVE_MAX_PIECES is the number of bytes at a time which we can |
| move efficiently, as opposed to MOVE_MAX which is the maximum |
| number of bytes we can move with a single instruction. */ |
| #define MOVE_MAX_PIECES UNITS_PER_WORD |
| |
| /* If a memory-to-memory move would take MOVE_RATIO or more simple |
| move-instruction pairs, we will do a movmem or libcall instead. |
| Increasing the value will always make code faster, but eventually |
| incurs high cost in increased code size. |
| |
| If you don't define this, a reasonable default is used. */ |
| |
| #define MOVE_RATIO(speed) ((speed) ? ix86_cost->move_ratio : 3) |
| |
| /* If a clear memory operation would take CLEAR_RATIO or more simple |
| move-instruction sequences, we will do a clrmem or libcall instead. */ |
| |
| #define CLEAR_RATIO(speed) ((speed) ? MIN (6, ix86_cost->move_ratio) : 2) |
| |
| /* Define if shifts truncate the shift count which implies one can |
| omit a sign-extension or zero-extension of a shift count. |
| |
| On i386, shifts do truncate the count. But bit test instructions |
| take the modulo of the bit offset operand. */ |
| |
| /* #define SHIFT_COUNT_TRUNCATED */ |
| |
| /* Value is 1 if truncating an integer of INPREC bits to OUTPREC bits |
| is done just by pretending it is already truncated. */ |
| #define TRULY_NOOP_TRUNCATION(OUTPREC, INPREC) 1 |
| |
| /* A macro to update M and UNSIGNEDP when an object whose type is |
| TYPE and which has the specified mode and signedness is to be |
| stored in a register. This macro is only called when TYPE is a |
| scalar type. |
| |
| On i386 it is sometimes useful to promote HImode and QImode |
| quantities to SImode. The choice depends on target type. */ |
| |
| #define PROMOTE_MODE(MODE, UNSIGNEDP, TYPE) \ |
| do { \ |
| if (((MODE) == HImode && TARGET_PROMOTE_HI_REGS) \ |
| || ((MODE) == QImode && TARGET_PROMOTE_QI_REGS)) \ |
| (MODE) = SImode; \ |
| } while (0) |
| |
| /* Specify the machine mode that pointers have. |
| After generation of rtl, the compiler makes no further distinction |
| between pointers and any other objects of this machine mode. */ |
| #define Pmode (ix86_pmode == PMODE_DI ? DImode : SImode) |
| |
| /* Specify the machine mode that bounds have. */ |
| #define BNDmode (ix86_pmode == PMODE_DI ? BND64mode : BND32mode) |
| |
| /* A C expression whose value is zero if pointers that need to be extended |
| from being `POINTER_SIZE' bits wide to `Pmode' are sign-extended and |
| greater then zero if they are zero-extended and less then zero if the |
| ptr_extend instruction should be used. */ |
| |
| #define POINTERS_EXTEND_UNSIGNED 1 |
| |
| /* A function address in a call instruction |
| is a byte address (for indexing purposes) |
| so give the MEM rtx a byte's mode. */ |
| #define FUNCTION_MODE QImode |
| |
| |
| /* A C expression for the cost of a branch instruction. A value of 1 |
| is the default; other values are interpreted relative to that. */ |
| |
| #define BRANCH_COST(speed_p, predictable_p) \ |
| (!(speed_p) ? 2 : (predictable_p) ? 0 : ix86_branch_cost) |
| |
| /* An integer expression for the size in bits of the largest integer machine |
| mode that should actually be used. We allow pairs of registers. */ |
| #define MAX_FIXED_MODE_SIZE GET_MODE_BITSIZE (TARGET_64BIT ? TImode : DImode) |
| |
| /* Define this macro as a C expression which is nonzero if accessing |
| less than a word of memory (i.e. a `char' or a `short') is no |
| faster than accessing a word of memory, i.e., if such access |
| require more than one instruction or if there is no difference in |
| cost between byte and (aligned) word loads. |
| |
| When this macro is not defined, the compiler will access a field by |
| finding the smallest containing object; when it is defined, a |
| fullword load will be used if alignment permits. Unless bytes |
| accesses are faster than word accesses, using word accesses is |
| preferable since it may eliminate subsequent memory access if |
| subsequent accesses occur to other fields in the same word of the |
| structure, but to different bytes. */ |
| |
| #define SLOW_BYTE_ACCESS 0 |
| |
| /* Nonzero if access to memory by shorts is slow and undesirable. */ |
| #define SLOW_SHORT_ACCESS 0 |
| |
| /* Define this macro to be the value 1 if unaligned accesses have a |
| cost many times greater than aligned accesses, for example if they |
| are emulated in a trap handler. |
| |
| When this macro is nonzero, the compiler will act as if |
| `STRICT_ALIGNMENT' were nonzero when generating code for block |
| moves. This can cause significantly more instructions to be |
| produced. Therefore, do not set this macro nonzero if unaligned |
| accesses only add a cycle or two to the time for a memory access. |
| |
| If the value of this macro is always zero, it need not be defined. */ |
| |
| /* #define SLOW_UNALIGNED_ACCESS(MODE, ALIGN) 0 */ |
| |
| /* Define this macro if it is as good or better to call a constant |
| function address than to call an address kept in a register. |
| |
| Desirable on the 386 because a CALL with a constant address is |
| faster than one with a register address. */ |
| |
| #define NO_FUNCTION_CSE |
| |
| /* Given a comparison code (EQ, NE, etc.) and the first operand of a COMPARE, |
| return the mode to be used for the comparison. |
| |
| For floating-point equality comparisons, CCFPEQmode should be used. |
| VOIDmode should be used in all other cases. |
| |
| For integer comparisons against zero, reduce to CCNOmode or CCZmode if |
| possible, to allow for more combinations. */ |
| |
| #define SELECT_CC_MODE(OP, X, Y) ix86_cc_mode ((OP), (X), (Y)) |
| |
| /* Return nonzero if MODE implies a floating point inequality can be |
| reversed. */ |
| |
| #define REVERSIBLE_CC_MODE(MODE) 1 |
| |
| /* A C expression whose value is reversed condition code of the CODE for |
| comparison done in CC_MODE mode. */ |
| #define REVERSE_CONDITION(CODE, MODE) ix86_reverse_condition ((CODE), (MODE)) |
| |
| |
| /* Control the assembler format that we output, to the extent |
| this does not vary between assemblers. */ |
| |
| /* How to refer to registers in assembler output. |
| This sequence is indexed by compiler's hard-register-number (see above). */ |
| |
| /* In order to refer to the first 8 regs as 32-bit regs, prefix an "e". |
| For non floating point regs, the following are the HImode names. |
| |
| For float regs, the stack top is sometimes referred to as "%st(0)" |
| instead of just "%st". TARGET_PRINT_OPERAND handles this with the |
| "y" code. */ |
| |
| #define HI_REGISTER_NAMES \ |
| {"ax","dx","cx","bx","si","di","bp","sp", \ |
| "st","st(1)","st(2)","st(3)","st(4)","st(5)","st(6)","st(7)", \ |
| "argp", "flags", "fpsr", "fpcr", "frame", \ |
| "xmm0","xmm1","xmm2","xmm3","xmm4","xmm5","xmm6","xmm7", \ |
| "mm0", "mm1", "mm2", "mm3", "mm4", "mm5", "mm6", "mm7", \ |
| "r8", "r9", "r10", "r11", "r12", "r13", "r14", "r15", \ |
| "xmm8", "xmm9", "xmm10", "xmm11", "xmm12", "xmm13", "xmm14", "xmm15", \ |
| "xmm16", "xmm17", "xmm18", "xmm19", \ |
| "xmm20", "xmm21", "xmm22", "xmm23", \ |
| "xmm24", "xmm25", "xmm26", "xmm27", \ |
| "xmm28", "xmm29", "xmm30", "xmm31", \ |
| "k0", "k1", "k2", "k3", "k4", "k5", "k6", "k7", \ |
| "bnd0", "bnd1", "bnd2", "bnd3" } |
| |
| #define REGISTER_NAMES HI_REGISTER_NAMES |
| |
| /* Table of additional register names to use in user input. */ |
| |
| #define ADDITIONAL_REGISTER_NAMES \ |
| { { "eax", 0 }, { "edx", 1 }, { "ecx", 2 }, { "ebx", 3 }, \ |
| { "esi", 4 }, { "edi", 5 }, { "ebp", 6 }, { "esp", 7 }, \ |
| { "rax", 0 }, { "rdx", 1 }, { "rcx", 2 }, { "rbx", 3 }, \ |
| { "rsi", 4 }, { "rdi", 5 }, { "rbp", 6 }, { "rsp", 7 }, \ |
| { "al", 0 }, { "dl", 1 }, { "cl", 2 }, { "bl", 3 }, \ |
| { "ah", 0 }, { "dh", 1 }, { "ch", 2 }, { "bh", 3 }, \ |
| { "ymm0", 21}, { "ymm1", 22}, { "ymm2", 23}, { "ymm3", 24}, \ |
| { "ymm4", 25}, { "ymm5", 26}, { "ymm6", 27}, { "ymm7", 28}, \ |
| { "ymm8", 45}, { "ymm9", 46}, { "ymm10", 47}, { "ymm11", 48}, \ |
| { "ymm12", 49}, { "ymm13", 50}, { "ymm14", 51}, { "ymm15", 52}, \ |
| { "ymm16", 53}, { "ymm17", 54}, { "ymm18", 55}, { "ymm19", 56}, \ |
| { "ymm20", 57}, { "ymm21", 58}, { "ymm22", 59}, { "ymm23", 60}, \ |
| { "ymm24", 61}, { "ymm25", 62}, { "ymm26", 63}, { "ymm27", 64}, \ |
| { "ymm28", 65}, { "ymm29", 66}, { "ymm30", 67}, { "ymm31", 68}, \ |
| { "zmm0", 21}, { "zmm1", 22}, { "zmm2", 23}, { "zmm3", 24}, \ |
| { "zmm4", 25}, { "zmm5", 26}, { "zmm6", 27}, { "zmm7", 28}, \ |
| { "zmm8", 45}, { "zmm9", 46}, { "zmm10", 47}, { "zmm11", 48}, \ |
| { "zmm12", 49}, { "zmm13", 50}, { "zmm14", 51}, { "zmm15", 52}, \ |
| { "zmm16", 53}, { "zmm17", 54}, { "zmm18", 55}, { "zmm19", 56}, \ |
| { "zmm20", 57}, { "zmm21", 58}, { "zmm22", 59}, { "zmm23", 60}, \ |
| { "zmm24", 61}, { "zmm25", 62}, { "zmm26", 63}, { "zmm27", 64}, \ |
| { "zmm28", 65}, { "zmm29", 66}, { "zmm30", 67}, { "zmm31", 68} } |
| |
| /* Note we are omitting these since currently I don't know how |
| to get gcc to use these, since they want the same but different |
| number as al, and ax. |
| */ |
| |
| #define QI_REGISTER_NAMES \ |
| {"al", "dl", "cl", "bl", "sil", "dil", "bpl", "spl",} |
| |
| /* These parallel the array above, and can be used to access bits 8:15 |
| of regs 0 through 3. */ |
| |
| #define QI_HIGH_REGISTER_NAMES \ |
| {"ah", "dh", "ch", "bh", } |
| |
| /* How to renumber registers for dbx and gdb. */ |
| |
| #define DBX_REGISTER_NUMBER(N) \ |
| (TARGET_64BIT ? dbx64_register_map[(N)] : dbx_register_map[(N)]) |
| |
| extern int const dbx_register_map[FIRST_PSEUDO_REGISTER]; |
| extern int const dbx64_register_map[FIRST_PSEUDO_REGISTER]; |
| extern int const svr4_dbx_register_map[FIRST_PSEUDO_REGISTER]; |
| |
| extern int const x86_64_ms_sysv_extra_clobbered_registers[12]; |
| |
| /* Before the prologue, RA is at 0(%esp). */ |
| #define INCOMING_RETURN_ADDR_RTX \ |
| gen_rtx_MEM (VOIDmode, gen_rtx_REG (VOIDmode, STACK_POINTER_REGNUM)) |
| |
| /* After the prologue, RA is at -4(AP) in the current frame. */ |
| #define RETURN_ADDR_RTX(COUNT, FRAME) \ |
| ((COUNT) == 0 \ |
| ? gen_rtx_MEM (Pmode, plus_constant (Pmode, arg_pointer_rtx, \ |
| -UNITS_PER_WORD)) \ |
| : gen_rtx_MEM (Pmode, plus_constant (Pmode, FRAME, UNITS_PER_WORD))) |
| |
| /* PC is dbx register 8; let's use that column for RA. */ |
| #define DWARF_FRAME_RETURN_COLUMN (TARGET_64BIT ? 16 : 8) |
| |
| /* Before the prologue, the top of the frame is at 4(%esp). */ |
| #define INCOMING_FRAME_SP_OFFSET UNITS_PER_WORD |
| |
| /* Describe how we implement __builtin_eh_return. */ |
| #define EH_RETURN_DATA_REGNO(N) ((N) <= DX_REG ? (N) : INVALID_REGNUM) |
| #define EH_RETURN_STACKADJ_RTX gen_rtx_REG (Pmode, CX_REG) |
| |
| |
| /* Select a format to encode pointers in exception handling data. CODE |
| is 0 for data, 1 for code labels, 2 for function pointers. GLOBAL is |
| true if the symbol may be affected by dynamic relocations. |
| |
| ??? All x86 object file formats are capable of representing this. |
| After all, the relocation needed is the same as for the call insn. |
| Whether or not a particular assembler allows us to enter such, I |
| guess we'll have to see. */ |
| #define ASM_PREFERRED_EH_DATA_FORMAT(CODE, GLOBAL) \ |
| asm_preferred_eh_data_format ((CODE), (GLOBAL)) |
| |
| /* This is how to output an insn to push a register on the stack. |
| It need not be very fast code. */ |
| |
| #define ASM_OUTPUT_REG_PUSH(FILE, REGNO) \ |
| do { \ |
| if (TARGET_64BIT) \ |
| asm_fprintf ((FILE), "\tpush{q}\t%%r%s\n", \ |
| reg_names[(REGNO)] + (REX_INT_REGNO_P (REGNO) != 0)); \ |
| else \ |
| asm_fprintf ((FILE), "\tpush{l}\t%%e%s\n", reg_names[(REGNO)]); \ |
| } while (0) |
| |
| /* This is how to output an insn to pop a register from the stack. |
| It need not be very fast code. */ |
| |
| #define ASM_OUTPUT_REG_POP(FILE, REGNO) \ |
| do { \ |
| if (TARGET_64BIT) \ |
| asm_fprintf ((FILE), "\tpop{q}\t%%r%s\n", \ |
| reg_names[(REGNO)] + (REX_INT_REGNO_P (REGNO) != 0)); \ |
| else \ |
| asm_fprintf ((FILE), "\tpop{l}\t%%e%s\n", reg_names[(REGNO)]); \ |
| } while (0) |
| |
| /* This is how to output an element of a case-vector that is absolute. */ |
| |
| #define ASM_OUTPUT_ADDR_VEC_ELT(FILE, VALUE) \ |
| ix86_output_addr_vec_elt ((FILE), (VALUE)) |
| |
| /* This is how to output an element of a case-vector that is relative. */ |
| |
| #define ASM_OUTPUT_ADDR_DIFF_ELT(FILE, BODY, VALUE, REL) \ |
| ix86_output_addr_diff_elt ((FILE), (VALUE), (REL)) |
| |
| /* When we see %v, we will print the 'v' prefix if TARGET_AVX is true. */ |
| |
| #define ASM_OUTPUT_AVX_PREFIX(STREAM, PTR) \ |
| { \ |
| if ((PTR)[0] == '%' && (PTR)[1] == 'v') \ |
| (PTR) += TARGET_AVX ? 1 : 2; \ |
| } |
| |
| /* A C statement or statements which output an assembler instruction |
| opcode to the stdio stream STREAM. The macro-operand PTR is a |
| variable of type `char *' which points to the opcode name in |
| its "internal" form--the form that is written in the machine |
| description. */ |
| |
| #define ASM_OUTPUT_OPCODE(STREAM, PTR) \ |
| ASM_OUTPUT_AVX_PREFIX ((STREAM), (PTR)) |
| |
| /* A C statement to output to the stdio stream FILE an assembler |
| command to pad the location counter to a multiple of 1<<LOG |
| bytes if it is within MAX_SKIP bytes. */ |
| |
| #ifdef HAVE_GAS_MAX_SKIP_P2ALIGN |
| #undef ASM_OUTPUT_MAX_SKIP_PAD |
| #define ASM_OUTPUT_MAX_SKIP_PAD(FILE, LOG, MAX_SKIP) \ |
| if ((LOG) != 0) \ |
| { \ |
| if ((MAX_SKIP) == 0) \ |
| fprintf ((FILE), "\t.p2align %d\n", (LOG)); \ |
| else \ |
| fprintf ((FILE), "\t.p2align %d,,%d\n", (LOG), (MAX_SKIP)); \ |
| } |
| #endif |
| |
| /* Write the extra assembler code needed to declare a function |
| properly. */ |
| |
| #undef ASM_OUTPUT_FUNCTION_LABEL |
| #define ASM_OUTPUT_FUNCTION_LABEL(FILE, NAME, DECL) \ |
| ix86_asm_output_function_label (FILE, NAME, DECL) |
| |
| /* Under some conditions we need jump tables in the text section, |
| because the assembler cannot handle label differences between |
| sections. This is the case for x86_64 on Mach-O for example. */ |
| |
| #define JUMP_TABLES_IN_TEXT_SECTION \ |
| (flag_pic && ((TARGET_MACHO && TARGET_64BIT) \ |
| || (!TARGET_64BIT && !HAVE_AS_GOTOFF_IN_DATA))) |
| |
| /* Switch to init or fini section via SECTION_OP, emit a call to FUNC, |
| and switch back. For x86 we do this only to save a few bytes that |
| would otherwise be unused in the text section. */ |
| #define CRT_MKSTR2(VAL) #VAL |
| #define CRT_MKSTR(x) CRT_MKSTR2(x) |
| |
| #define CRT_CALL_STATIC_FUNCTION(SECTION_OP, FUNC) \ |
| asm (SECTION_OP "\n\t" \ |
| "call " CRT_MKSTR(__USER_LABEL_PREFIX__) #FUNC "\n" \ |
| TEXT_SECTION_ASM_OP); |
| |
| /* Default threshold for putting data in large sections |
| with x86-64 medium memory model */ |
| #define DEFAULT_LARGE_SECTION_THRESHOLD 65536 |
| |
| /* Which processor to tune code generation for. These must be in sync |
| with processor_target_table in i386.c. */ |
| |
| enum processor_type |
| { |
| PROCESSOR_GENERIC = 0, |
| PROCESSOR_I386, /* 80386 */ |
| PROCESSOR_I486, /* 80486DX, 80486SX, 80486DX[24] */ |
| PROCESSOR_PENTIUM, |
| PROCESSOR_PENTIUMPRO, |
| PROCESSOR_PENTIUM4, |
| PROCESSOR_NOCONA, |
| PROCESSOR_CORE2, |
| PROCESSOR_NEHALEM, |
| PROCESSOR_SANDYBRIDGE, |
| PROCESSOR_HASWELL, |
| PROCESSOR_BONNELL, |
| PROCESSOR_SILVERMONT, |
| PROCESSOR_KNL, |
| PROCESSOR_INTEL, |
| PROCESSOR_GEODE, |
| PROCESSOR_K6, |
| PROCESSOR_ATHLON, |
| PROCESSOR_K8, |
| PROCESSOR_AMDFAM10, |
| PROCESSOR_BDVER1, |
| PROCESSOR_BDVER2, |
| PROCESSOR_BDVER3, |
| PROCESSOR_BDVER4, |
| PROCESSOR_BTVER1, |
| PROCESSOR_BTVER2, |
| PROCESSOR_max |
| }; |
| |
| extern enum processor_type ix86_tune; |
| extern enum processor_type ix86_arch; |
| |
| /* Size of the RED_ZONE area. */ |
| #define RED_ZONE_SIZE 128 |
| /* Reserved area of the red zone for temporaries. */ |
| #define RED_ZONE_RESERVE 8 |
| |
| extern unsigned int ix86_preferred_stack_boundary; |
| extern unsigned int ix86_incoming_stack_boundary; |
| |
| /* Smallest class containing REGNO. */ |
| extern enum reg_class const regclass_map[FIRST_PSEUDO_REGISTER]; |
| |
| enum ix86_fpcmp_strategy { |
| IX86_FPCMP_SAHF, |
| IX86_FPCMP_COMI, |
| IX86_FPCMP_ARITH |
| }; |
| |
| /* To properly truncate FP values into integers, we need to set i387 control |
| word. We can't emit proper mode switching code before reload, as spills |
| generated by reload may truncate values incorrectly, but we still can avoid |
| redundant computation of new control word by the mode switching pass. |
| The fldcw instructions are still emitted redundantly, but this is probably |
| not going to be noticeable problem, as most CPUs do have fast path for |
| the sequence. |
| |
| The machinery is to emit simple truncation instructions and split them |
| before reload to instructions having USEs of two memory locations that |
| are filled by this code to old and new control word. |
| |
| Post-reload pass may be later used to eliminate the redundant fildcw if |
| needed. */ |
| |
| enum ix86_entity |
| { |
| AVX_U128 = 0, |
| I387_TRUNC, |
| I387_FLOOR, |
| I387_CEIL, |
| I387_MASK_PM, |
| MAX_386_ENTITIES |
| }; |
| |
| enum ix86_stack_slot |
| { |
| SLOT_TEMP = 0, |
| SLOT_CW_STORED, |
| SLOT_CW_TRUNC, |
| SLOT_CW_FLOOR, |
| SLOT_CW_CEIL, |
| SLOT_CW_MASK_PM, |
| MAX_386_STACK_LOCALS |
| }; |
| |
| enum avx_u128_state |
| { |
| AVX_U128_CLEAN, |
| AVX_U128_DIRTY, |
| AVX_U128_ANY |
| }; |
| |
| /* Define this macro if the port needs extra instructions inserted |
| for mode switching in an optimizing compilation. */ |
| |
| #define OPTIMIZE_MODE_SWITCHING(ENTITY) \ |
| ix86_optimize_mode_switching[(ENTITY)] |
| |
| /* If you define `OPTIMIZE_MODE_SWITCHING', you have to define this as |
| initializer for an array of integers. Each initializer element N |
| refers to an entity that needs mode switching, and specifies the |
| number of different modes that might need to be set for this |
| entity. The position of the initializer in the initializer - |
| starting counting at zero - determines the integer that is used to |
| refer to the mode-switched entity in question. */ |
| |
| #define NUM_MODES_FOR_MODE_SWITCHING \ |
| { AVX_U128_ANY, I387_CW_ANY, I387_CW_ANY, I387_CW_ANY, I387_CW_ANY } |
| |
| |
| /* Avoid renaming of stack registers, as doing so in combination with |
| scheduling just increases amount of live registers at time and in |
| the turn amount of fxch instructions needed. |
| |
| ??? Maybe Pentium chips benefits from renaming, someone can try.... |
| |
| Don't rename evex to non-evex sse registers. */ |
| |
| #define HARD_REGNO_RENAME_OK(SRC, TARGET) (!STACK_REGNO_P (SRC) && \ |
| (EXT_REX_SSE_REGNO_P (SRC) == \ |
| EXT_REX_SSE_REGNO_P (TARGET))) |
| |
| |
| #define FASTCALL_PREFIX '@' |
| |
| /* Machine specific frame tracking during prologue/epilogue generation. */ |
| |
| #ifndef USED_FOR_TARGET |
| struct GTY(()) machine_frame_state |
| { |
| /* This pair tracks the currently active CFA as reg+offset. When reg |
| is drap_reg, we don't bother trying to record here the real CFA when |
| it might really be a DW_CFA_def_cfa_expression. */ |
| rtx cfa_reg; |
| HOST_WIDE_INT cfa_offset; |
| |
| /* The current offset (canonically from the CFA) of ESP and EBP. |
| When stack frame re-alignment is active, these may not be relative |
| to the CFA. However, in all cases they are relative to the offsets |
| of the saved registers stored in ix86_frame. */ |
| HOST_WIDE_INT sp_offset; |
| HOST_WIDE_INT fp_offset; |
| |
| /* The size of the red-zone that may be assumed for the purposes of |
| eliding register restore notes in the epilogue. This may be zero |
| if no red-zone is in effect, or may be reduced from the real |
| red-zone value by a maximum runtime stack re-alignment value. */ |
| int red_zone_offset; |
| |
| /* Indicate whether each of ESP, EBP or DRAP currently holds a valid |
| value within the frame. If false then the offset above should be |
| ignored. Note that DRAP, if valid, *always* points to the CFA and |
| thus has an offset of zero. */ |
| BOOL_BITFIELD sp_valid : 1; |
| BOOL_BITFIELD fp_valid : 1; |
| BOOL_BITFIELD drap_valid : 1; |
| |
| /* Indicate whether the local stack frame has been re-aligned. When |
| set, the SP/FP offsets above are relative to the aligned frame |
| and not the CFA. */ |
| BOOL_BITFIELD realigned : 1; |
| }; |
| |
| /* Private to winnt.c. */ |
| struct seh_frame_state; |
| |
| struct GTY(()) machine_function { |
| struct stack_local_entry *stack_locals; |
| const char *some_ld_name; |
| int varargs_gpr_size; |
| int varargs_fpr_size; |
| int optimize_mode_switching[MAX_386_ENTITIES]; |
| |
| /* Number of saved registers USE_FAST_PROLOGUE_EPILOGUE |
| has been computed for. */ |
| int use_fast_prologue_epilogue_nregs; |
| |
| /* For -fsplit-stack support: A stack local which holds a pointer to |
| the stack arguments for a function with a variable number of |
| arguments. This is set at the start of the function and is used |
| to initialize the overflow_arg_area field of the va_list |
| structure. */ |
| rtx split_stack_varargs_pointer; |
| |
| /* This value is used for amd64 targets and specifies the current abi |
| to be used. MS_ABI means ms abi. Otherwise SYSV_ABI means sysv abi. */ |
| ENUM_BITFIELD(calling_abi) call_abi : 8; |
| |
| /* Nonzero if the function accesses a previous frame. */ |
| BOOL_BITFIELD accesses_prev_frame : 1; |
| |
| /* Nonzero if the function requires a CLD in the prologue. */ |
| BOOL_BITFIELD needs_cld : 1; |
| |
| /* Set by ix86_compute_frame_layout and used by prologue/epilogue |
| expander to determine the style used. */ |
| BOOL_BITFIELD use_fast_prologue_epilogue : 1; |
| |
| /* If true, the current function needs the default PIC register, not |
| an alternate register (on x86) and must not use the red zone (on |
| x86_64), even if it's a leaf function. We don't want the |
| function to be regarded as non-leaf because TLS calls need not |
| affect register allocation. This flag is set when a TLS call |
| instruction is expanded within a function, and never reset, even |
| if all such instructions are optimized away. Use the |
| ix86_current_function_calls_tls_descriptor macro for a better |
| approximation. */ |
| BOOL_BITFIELD tls_descriptor_call_expanded_p : 1; |
| |
| /* If true, the current function has a STATIC_CHAIN is placed on the |
| stack below the return address. */ |
| BOOL_BITFIELD static_chain_on_stack : 1; |
| |
| /* If true, it is safe to not save/restore DRAP register. */ |
| BOOL_BITFIELD no_drap_save_restore : 1; |
| |
| /* During prologue/epilogue generation, the current frame state. |
| Otherwise, the frame state at the end of the prologue. */ |
| struct machine_frame_state fs; |
| |
| /* During SEH output, this is non-null. */ |
| struct seh_frame_state * GTY((skip(""))) seh; |
| }; |
| #endif |
| |
| #define ix86_stack_locals (cfun->machine->stack_locals) |
| #define ix86_varargs_gpr_size (cfun->machine->varargs_gpr_size) |
| #define ix86_varargs_fpr_size (cfun->machine->varargs_fpr_size) |
| #define ix86_optimize_mode_switching (cfun->machine->optimize_mode_switching) |
| #define ix86_current_function_needs_cld (cfun->machine->needs_cld) |
| #define ix86_tls_descriptor_calls_expanded_in_cfun \ |
| (cfun->machine->tls_descriptor_call_expanded_p) |
| /* Since tls_descriptor_call_expanded is not cleared, even if all TLS |
| calls are optimized away, we try to detect cases in which it was |
| optimized away. Since such instructions (use (reg REG_SP)), we can |
| verify whether there's any such instruction live by testing that |
| REG_SP is live. */ |
| #define ix86_current_function_calls_tls_descriptor \ |
| (ix86_tls_descriptor_calls_expanded_in_cfun && df_regs_ever_live_p (SP_REG)) |
| #define ix86_static_chain_on_stack (cfun->machine->static_chain_on_stack) |
| |
| /* Control behavior of x86_file_start. */ |
| #define X86_FILE_START_VERSION_DIRECTIVE false |
| #define X86_FILE_START_FLTUSED false |
| |
| /* Flag to mark data that is in the large address area. */ |
| #define SYMBOL_FLAG_FAR_ADDR (SYMBOL_FLAG_MACH_DEP << 0) |
| #define SYMBOL_REF_FAR_ADDR_P(X) \ |
| ((SYMBOL_REF_FLAGS (X) & SYMBOL_FLAG_FAR_ADDR) != 0) |
| |
| /* Flags to mark dllimport/dllexport. Used by PE ports, but handy to |
| have defined always, to avoid ifdefing. */ |
| #define SYMBOL_FLAG_DLLIMPORT (SYMBOL_FLAG_MACH_DEP << 1) |
| #define SYMBOL_REF_DLLIMPORT_P(X) \ |
| ((SYMBOL_REF_FLAGS (X) & SYMBOL_FLAG_DLLIMPORT) != 0) |
| |
| #define SYMBOL_FLAG_DLLEXPORT (SYMBOL_FLAG_MACH_DEP << 2) |
| #define SYMBOL_REF_DLLEXPORT_P(X) \ |
| ((SYMBOL_REF_FLAGS (X) & SYMBOL_FLAG_DLLEXPORT) != 0) |
| |
| #define SYMBOL_FLAG_STUBVAR (SYMBOL_FLAG_MACH_DEP << 4) |
| #define SYMBOL_REF_STUBVAR_P(X) \ |
| ((SYMBOL_REF_FLAGS (X) & SYMBOL_FLAG_STUBVAR) != 0) |
| |
| extern void debug_ready_dispatch (void); |
| extern void debug_dispatch_window (int); |
| |
| /* The value at zero is only defined for the BMI instructions |
| LZCNT and TZCNT, not the BSR/BSF insns in the original isa. */ |
| #define CTZ_DEFINED_VALUE_AT_ZERO(MODE, VALUE) \ |
| ((VALUE) = GET_MODE_BITSIZE (MODE), TARGET_BMI ? 1 : 0) |
| #define CLZ_DEFINED_VALUE_AT_ZERO(MODE, VALUE) \ |
| ((VALUE) = GET_MODE_BITSIZE (MODE), TARGET_LZCNT ? 1 : 0) |
| |
| |
| /* Flags returned by ix86_get_callcvt (). */ |
| #define IX86_CALLCVT_CDECL 0x1 |
| #define IX86_CALLCVT_STDCALL 0x2 |
| #define IX86_CALLCVT_FASTCALL 0x4 |
| #define IX86_CALLCVT_THISCALL 0x8 |
| #define IX86_CALLCVT_REGPARM 0x10 |
| #define IX86_CALLCVT_SSEREGPARM 0x20 |
| |
| #define IX86_BASE_CALLCVT(FLAGS) \ |
| ((FLAGS) & (IX86_CALLCVT_CDECL | IX86_CALLCVT_STDCALL \ |
| | IX86_CALLCVT_FASTCALL | IX86_CALLCVT_THISCALL)) |
| |
| #define RECIP_MASK_NONE 0x00 |
| #define RECIP_MASK_DIV 0x01 |
| #define RECIP_MASK_SQRT 0x02 |
| #define RECIP_MASK_VEC_DIV 0x04 |
| #define RECIP_MASK_VEC_SQRT 0x08 |
| #define RECIP_MASK_ALL (RECIP_MASK_DIV | RECIP_MASK_SQRT \ |
| | RECIP_MASK_VEC_DIV | RECIP_MASK_VEC_SQRT) |
| #define RECIP_MASK_DEFAULT (RECIP_MASK_VEC_DIV | RECIP_MASK_VEC_SQRT) |
| |
| #define TARGET_RECIP_DIV ((recip_mask & RECIP_MASK_DIV) != 0) |
| #define TARGET_RECIP_SQRT ((recip_mask & RECIP_MASK_SQRT) != 0) |
| #define TARGET_RECIP_VEC_DIV ((recip_mask & RECIP_MASK_VEC_DIV) != 0) |
| #define TARGET_RECIP_VEC_SQRT ((recip_mask & RECIP_MASK_VEC_SQRT) != 0) |
| |
| #define IX86_HLE_ACQUIRE (1 << 16) |
| #define IX86_HLE_RELEASE (1 << 17) |
| |
| /* For switching between functions with different target attributes. */ |
| #define SWITCHABLE_TARGET 1 |
| |
| /* |
| Local variables: |
| version-control: t |
| End: |
| */ |