| /* Subroutines shared by all languages that are variants of C. |
| Copyright (C) 1992-2017 Free Software Foundation, Inc. |
| |
| This file is part of GCC. |
| |
| GCC is free software; you can redistribute it and/or modify it under |
| the terms of the GNU General Public License as published by the Free |
| Software Foundation; either version 3, or (at your option) any later |
| version. |
| |
| GCC is distributed in the hope that it will be useful, but WITHOUT ANY |
| WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
| for more details. |
| |
| You should have received a copy of the GNU General Public License |
| along with GCC; see the file COPYING3. If not see |
| <http://www.gnu.org/licenses/>. */ |
| |
| #define GCC_C_COMMON_C |
| |
| #include "config.h" |
| #include "system.h" |
| #include "coretypes.h" |
| #include "target.h" |
| #include "function.h" |
| #include "tree.h" |
| #include "memmodel.h" |
| #include "c-common.h" |
| #include "gimple-expr.h" |
| #include "tm_p.h" |
| #include "stringpool.h" |
| #include "cgraph.h" |
| #include "diagnostic.h" |
| #include "intl.h" |
| #include "stor-layout.h" |
| #include "calls.h" |
| #include "attribs.h" |
| #include "varasm.h" |
| #include "trans-mem.h" |
| #include "c-objc.h" |
| #include "common/common-target.h" |
| #include "langhooks.h" |
| #include "tree-inline.h" |
| #include "toplev.h" |
| #include "tree-iterator.h" |
| #include "opts.h" |
| #include "gimplify.h" |
| #include "substring-locations.h" |
| #include "spellcheck.h" |
| |
| cpp_reader *parse_in; /* Declared in c-pragma.h. */ |
| |
| /* Mode used to build pointers (VOIDmode means ptr_mode). */ |
| |
| machine_mode c_default_pointer_mode = VOIDmode; |
| |
| /* The following symbols are subsumed in the c_global_trees array, and |
| listed here individually for documentation purposes. |
| |
| INTEGER_TYPE and REAL_TYPE nodes for the standard data types. |
| |
| tree short_integer_type_node; |
| tree long_integer_type_node; |
| tree long_long_integer_type_node; |
| |
| tree short_unsigned_type_node; |
| tree long_unsigned_type_node; |
| tree long_long_unsigned_type_node; |
| |
| tree truthvalue_type_node; |
| tree truthvalue_false_node; |
| tree truthvalue_true_node; |
| |
| tree ptrdiff_type_node; |
| |
| tree unsigned_char_type_node; |
| tree signed_char_type_node; |
| tree wchar_type_node; |
| |
| tree char16_type_node; |
| tree char32_type_node; |
| |
| tree float_type_node; |
| tree double_type_node; |
| tree long_double_type_node; |
| |
| tree complex_integer_type_node; |
| tree complex_float_type_node; |
| tree complex_double_type_node; |
| tree complex_long_double_type_node; |
| |
| tree dfloat32_type_node; |
| tree dfloat64_type_node; |
| tree_dfloat128_type_node; |
| |
| tree intQI_type_node; |
| tree intHI_type_node; |
| tree intSI_type_node; |
| tree intDI_type_node; |
| tree intTI_type_node; |
| |
| tree unsigned_intQI_type_node; |
| tree unsigned_intHI_type_node; |
| tree unsigned_intSI_type_node; |
| tree unsigned_intDI_type_node; |
| tree unsigned_intTI_type_node; |
| |
| tree widest_integer_literal_type_node; |
| tree widest_unsigned_literal_type_node; |
| |
| Nodes for types `void *' and `const void *'. |
| |
| tree ptr_type_node, const_ptr_type_node; |
| |
| Nodes for types `char *' and `const char *'. |
| |
| tree string_type_node, const_string_type_node; |
| |
| Type `char[SOMENUMBER]'. |
| Used when an array of char is needed and the size is irrelevant. |
| |
| tree char_array_type_node; |
| |
| Type `wchar_t[SOMENUMBER]' or something like it. |
| Used when a wide string literal is created. |
| |
| tree wchar_array_type_node; |
| |
| Type `char16_t[SOMENUMBER]' or something like it. |
| Used when a UTF-16 string literal is created. |
| |
| tree char16_array_type_node; |
| |
| Type `char32_t[SOMENUMBER]' or something like it. |
| Used when a UTF-32 string literal is created. |
| |
| tree char32_array_type_node; |
| |
| Type `int ()' -- used for implicit declaration of functions. |
| |
| tree default_function_type; |
| |
| A VOID_TYPE node, packaged in a TREE_LIST. |
| |
| tree void_list_node; |
| |
| The lazily created VAR_DECLs for __FUNCTION__, __PRETTY_FUNCTION__, |
| and __func__. (C doesn't generate __FUNCTION__ and__PRETTY_FUNCTION__ |
| VAR_DECLS, but C++ does.) |
| |
| tree function_name_decl_node; |
| tree pretty_function_name_decl_node; |
| tree c99_function_name_decl_node; |
| |
| Stack of nested function name VAR_DECLs. |
| |
| tree saved_function_name_decls; |
| |
| */ |
| |
| tree c_global_trees[CTI_MAX]; |
| |
| /* Switches common to the C front ends. */ |
| |
| /* Nonzero means don't output line number information. */ |
| |
| char flag_no_line_commands; |
| |
| /* Nonzero causes -E output not to be done, but directives such as |
| #define that have side effects are still obeyed. */ |
| |
| char flag_no_output; |
| |
| /* Nonzero means dump macros in some fashion. */ |
| |
| char flag_dump_macros; |
| |
| /* Nonzero means pass #include lines through to the output. */ |
| |
| char flag_dump_includes; |
| |
| /* Nonzero means process PCH files while preprocessing. */ |
| |
| bool flag_pch_preprocess; |
| |
| /* The file name to which we should write a precompiled header, or |
| NULL if no header will be written in this compile. */ |
| |
| const char *pch_file; |
| |
| /* Nonzero if an ISO standard was selected. It rejects macros in the |
| user's namespace. */ |
| int flag_iso; |
| |
| /* C/ObjC language option variables. */ |
| |
| |
| /* Nonzero means allow type mismatches in conditional expressions; |
| just make their values `void'. */ |
| |
| int flag_cond_mismatch; |
| |
| /* Nonzero means enable C89 Amendment 1 features. */ |
| |
| int flag_isoc94; |
| |
| /* Nonzero means use the ISO C99 (or C11) dialect of C. */ |
| |
| int flag_isoc99; |
| |
| /* Nonzero means use the ISO C11 dialect of C. */ |
| |
| int flag_isoc11; |
| |
| /* Nonzero means that we have builtin functions, and main is an int. */ |
| |
| int flag_hosted = 1; |
| |
| |
| /* ObjC language option variables. */ |
| |
| |
| /* Tells the compiler that this is a special run. Do not perform any |
| compiling, instead we are to test some platform dependent features |
| and output a C header file with appropriate definitions. */ |
| |
| int print_struct_values; |
| |
| /* Tells the compiler what is the constant string class for ObjC. */ |
| |
| const char *constant_string_class_name; |
| |
| |
| /* C++ language option variables. */ |
| |
| /* The reference version of the ABI for -Wabi. */ |
| |
| int warn_abi_version = -1; |
| |
| /* Nonzero means generate separate instantiation control files and |
| juggle them at link time. */ |
| |
| int flag_use_repository; |
| |
| /* The C++ dialect being used. Default set in c_common_post_options. */ |
| |
| enum cxx_dialect cxx_dialect = cxx_unset; |
| |
| /* Maximum template instantiation depth. This limit exists to limit the |
| time it takes to notice excessively recursive template instantiations. |
| |
| The default is lower than the 1024 recommended by the C++0x standard |
| because G++ runs out of stack before 1024 with highly recursive template |
| argument deduction substitution (g++.dg/cpp0x/enum11.C). */ |
| |
| int max_tinst_depth = 900; |
| |
| /* The elements of `ridpointers' are identifier nodes for the reserved |
| type names and storage classes. It is indexed by a RID_... value. */ |
| tree *ridpointers; |
| |
| tree (*make_fname_decl) (location_t, tree, int); |
| |
| /* Nonzero means don't warn about problems that occur when the code is |
| executed. */ |
| int c_inhibit_evaluation_warnings; |
| |
| /* Whether we are building a boolean conversion inside |
| convert_for_assignment, or some other late binary operation. If |
| build_binary_op is called for C (from code shared by C and C++) in |
| this case, then the operands have already been folded and the |
| result will not be folded again, so C_MAYBE_CONST_EXPR should not |
| be generated. */ |
| bool in_late_binary_op; |
| |
| /* Whether lexing has been completed, so subsequent preprocessor |
| errors should use the compiler's input_location. */ |
| bool done_lexing = false; |
| |
| /* Information about how a function name is generated. */ |
| struct fname_var_t |
| { |
| tree *const decl; /* pointer to the VAR_DECL. */ |
| const unsigned rid; /* RID number for the identifier. */ |
| const int pretty; /* How pretty is it? */ |
| }; |
| |
| /* The three ways of getting then name of the current function. */ |
| |
| const struct fname_var_t fname_vars[] = |
| { |
| /* C99 compliant __func__, must be first. */ |
| {&c99_function_name_decl_node, RID_C99_FUNCTION_NAME, 0}, |
| /* GCC __FUNCTION__ compliant. */ |
| {&function_name_decl_node, RID_FUNCTION_NAME, 0}, |
| /* GCC __PRETTY_FUNCTION__ compliant. */ |
| {&pretty_function_name_decl_node, RID_PRETTY_FUNCTION_NAME, 1}, |
| {NULL, 0, 0}, |
| }; |
| |
| /* Global visibility options. */ |
| struct visibility_flags visibility_options; |
| |
| static tree check_case_value (location_t, tree); |
| static bool check_case_bounds (location_t, tree, tree, tree *, tree *, |
| bool *); |
| |
| |
| static void check_nonnull_arg (void *, tree, unsigned HOST_WIDE_INT); |
| static bool nonnull_check_p (tree, unsigned HOST_WIDE_INT); |
| static int resort_field_decl_cmp (const void *, const void *); |
| |
| /* Reserved words. The third field is a mask: keywords are disabled |
| if they match the mask. |
| |
| Masks for languages: |
| C --std=c89: D_C99 | D_CXXONLY | D_OBJC | D_CXX_OBJC |
| C --std=c99: D_CXXONLY | D_OBJC |
| ObjC is like C except that D_OBJC and D_CXX_OBJC are not set |
| C++ --std=c++98: D_CONLY | D_CXX11 | D_OBJC |
| C++ --std=c++11: D_CONLY | D_OBJC |
| ObjC++ is like C++ except that D_OBJC is not set |
| |
| If -fno-asm is used, D_ASM is added to the mask. If |
| -fno-gnu-keywords is used, D_EXT is added. If -fno-asm and C in |
| C89 mode, D_EXT89 is added for both -fno-asm and -fno-gnu-keywords. |
| In C with -Wc++-compat, we warn if D_CXXWARN is set. |
| |
| Note the complication of the D_CXX_OBJC keywords. These are |
| reserved words such as 'class'. In C++, 'class' is a reserved |
| word. In Objective-C++ it is too. In Objective-C, it is a |
| reserved word too, but only if it follows an '@' sign. |
| */ |
| const struct c_common_resword c_common_reswords[] = |
| { |
| { "_Alignas", RID_ALIGNAS, D_CONLY }, |
| { "_Alignof", RID_ALIGNOF, D_CONLY }, |
| { "_Atomic", RID_ATOMIC, D_CONLY }, |
| { "_Bool", RID_BOOL, D_CONLY }, |
| { "_Complex", RID_COMPLEX, 0 }, |
| { "_Cilk_spawn", RID_CILK_SPAWN, 0 }, |
| { "_Cilk_sync", RID_CILK_SYNC, 0 }, |
| { "_Cilk_for", RID_CILK_FOR, 0 }, |
| { "_Imaginary", RID_IMAGINARY, D_CONLY }, |
| { "_Float16", RID_FLOAT16, D_CONLY }, |
| { "_Float32", RID_FLOAT32, D_CONLY }, |
| { "_Float64", RID_FLOAT64, D_CONLY }, |
| { "_Float128", RID_FLOAT128, D_CONLY }, |
| { "_Float32x", RID_FLOAT32X, D_CONLY }, |
| { "_Float64x", RID_FLOAT64X, D_CONLY }, |
| { "_Float128x", RID_FLOAT128X, D_CONLY }, |
| { "_Decimal32", RID_DFLOAT32, D_CONLY | D_EXT }, |
| { "_Decimal64", RID_DFLOAT64, D_CONLY | D_EXT }, |
| { "_Decimal128", RID_DFLOAT128, D_CONLY | D_EXT }, |
| { "_Fract", RID_FRACT, D_CONLY | D_EXT }, |
| { "_Accum", RID_ACCUM, D_CONLY | D_EXT }, |
| { "_Sat", RID_SAT, D_CONLY | D_EXT }, |
| { "_Static_assert", RID_STATIC_ASSERT, D_CONLY }, |
| { "_Noreturn", RID_NORETURN, D_CONLY }, |
| { "_Generic", RID_GENERIC, D_CONLY }, |
| { "_Thread_local", RID_THREAD, D_CONLY }, |
| { "__FUNCTION__", RID_FUNCTION_NAME, 0 }, |
| { "__PRETTY_FUNCTION__", RID_PRETTY_FUNCTION_NAME, 0 }, |
| { "__alignof", RID_ALIGNOF, 0 }, |
| { "__alignof__", RID_ALIGNOF, 0 }, |
| { "__asm", RID_ASM, 0 }, |
| { "__asm__", RID_ASM, 0 }, |
| { "__attribute", RID_ATTRIBUTE, 0 }, |
| { "__attribute__", RID_ATTRIBUTE, 0 }, |
| { "__auto_type", RID_AUTO_TYPE, D_CONLY }, |
| { "__bases", RID_BASES, D_CXXONLY }, |
| { "__builtin_addressof", RID_ADDRESSOF, D_CXXONLY }, |
| { "__builtin_call_with_static_chain", |
| RID_BUILTIN_CALL_WITH_STATIC_CHAIN, D_CONLY }, |
| { "__builtin_choose_expr", RID_CHOOSE_EXPR, D_CONLY }, |
| { "__builtin_complex", RID_BUILTIN_COMPLEX, D_CONLY }, |
| { "__builtin_launder", RID_BUILTIN_LAUNDER, D_CXXONLY }, |
| { "__builtin_shuffle", RID_BUILTIN_SHUFFLE, 0 }, |
| { "__builtin_offsetof", RID_OFFSETOF, 0 }, |
| { "__builtin_types_compatible_p", RID_TYPES_COMPATIBLE_P, D_CONLY }, |
| { "__builtin_va_arg", RID_VA_ARG, 0 }, |
| { "__complex", RID_COMPLEX, 0 }, |
| { "__complex__", RID_COMPLEX, 0 }, |
| { "__const", RID_CONST, 0 }, |
| { "__const__", RID_CONST, 0 }, |
| { "__decltype", RID_DECLTYPE, D_CXXONLY }, |
| { "__direct_bases", RID_DIRECT_BASES, D_CXXONLY }, |
| { "__extension__", RID_EXTENSION, 0 }, |
| { "__func__", RID_C99_FUNCTION_NAME, 0 }, |
| { "__has_nothrow_assign", RID_HAS_NOTHROW_ASSIGN, D_CXXONLY }, |
| { "__has_nothrow_constructor", RID_HAS_NOTHROW_CONSTRUCTOR, D_CXXONLY }, |
| { "__has_nothrow_copy", RID_HAS_NOTHROW_COPY, D_CXXONLY }, |
| { "__has_trivial_assign", RID_HAS_TRIVIAL_ASSIGN, D_CXXONLY }, |
| { "__has_trivial_constructor", RID_HAS_TRIVIAL_CONSTRUCTOR, D_CXXONLY }, |
| { "__has_trivial_copy", RID_HAS_TRIVIAL_COPY, D_CXXONLY }, |
| { "__has_trivial_destructor", RID_HAS_TRIVIAL_DESTRUCTOR, D_CXXONLY }, |
| { "__has_unique_object_representations", RID_HAS_UNIQUE_OBJ_REPRESENTATIONS, |
| D_CXXONLY }, |
| { "__has_virtual_destructor", RID_HAS_VIRTUAL_DESTRUCTOR, D_CXXONLY }, |
| { "__imag", RID_IMAGPART, 0 }, |
| { "__imag__", RID_IMAGPART, 0 }, |
| { "__inline", RID_INLINE, 0 }, |
| { "__inline__", RID_INLINE, 0 }, |
| { "__is_abstract", RID_IS_ABSTRACT, D_CXXONLY }, |
| { "__is_aggregate", RID_IS_AGGREGATE, D_CXXONLY }, |
| { "__is_base_of", RID_IS_BASE_OF, D_CXXONLY }, |
| { "__is_class", RID_IS_CLASS, D_CXXONLY }, |
| { "__is_empty", RID_IS_EMPTY, D_CXXONLY }, |
| { "__is_enum", RID_IS_ENUM, D_CXXONLY }, |
| { "__is_final", RID_IS_FINAL, D_CXXONLY }, |
| { "__is_literal_type", RID_IS_LITERAL_TYPE, D_CXXONLY }, |
| { "__is_pod", RID_IS_POD, D_CXXONLY }, |
| { "__is_polymorphic", RID_IS_POLYMORPHIC, D_CXXONLY }, |
| { "__is_same_as", RID_IS_SAME_AS, D_CXXONLY }, |
| { "__is_standard_layout", RID_IS_STD_LAYOUT, D_CXXONLY }, |
| { "__is_trivial", RID_IS_TRIVIAL, D_CXXONLY }, |
| { "__is_trivially_assignable", RID_IS_TRIVIALLY_ASSIGNABLE, D_CXXONLY }, |
| { "__is_trivially_constructible", RID_IS_TRIVIALLY_CONSTRUCTIBLE, D_CXXONLY }, |
| { "__is_trivially_copyable", RID_IS_TRIVIALLY_COPYABLE, D_CXXONLY }, |
| { "__is_union", RID_IS_UNION, D_CXXONLY }, |
| { "__label__", RID_LABEL, 0 }, |
| { "__null", RID_NULL, 0 }, |
| { "__real", RID_REALPART, 0 }, |
| { "__real__", RID_REALPART, 0 }, |
| { "__restrict", RID_RESTRICT, 0 }, |
| { "__restrict__", RID_RESTRICT, 0 }, |
| { "__signed", RID_SIGNED, 0 }, |
| { "__signed__", RID_SIGNED, 0 }, |
| { "__thread", RID_THREAD, 0 }, |
| { "__transaction_atomic", RID_TRANSACTION_ATOMIC, 0 }, |
| { "__transaction_relaxed", RID_TRANSACTION_RELAXED, 0 }, |
| { "__transaction_cancel", RID_TRANSACTION_CANCEL, 0 }, |
| { "__typeof", RID_TYPEOF, 0 }, |
| { "__typeof__", RID_TYPEOF, 0 }, |
| { "__underlying_type", RID_UNDERLYING_TYPE, D_CXXONLY }, |
| { "__volatile", RID_VOLATILE, 0 }, |
| { "__volatile__", RID_VOLATILE, 0 }, |
| { "__GIMPLE", RID_GIMPLE, D_CONLY }, |
| { "__PHI", RID_PHI, D_CONLY }, |
| { "__RTL", RID_RTL, D_CONLY }, |
| { "alignas", RID_ALIGNAS, D_CXXONLY | D_CXX11 | D_CXXWARN }, |
| { "alignof", RID_ALIGNOF, D_CXXONLY | D_CXX11 | D_CXXWARN }, |
| { "asm", RID_ASM, D_ASM }, |
| { "auto", RID_AUTO, 0 }, |
| { "bool", RID_BOOL, D_CXXONLY | D_CXXWARN }, |
| { "break", RID_BREAK, 0 }, |
| { "case", RID_CASE, 0 }, |
| { "catch", RID_CATCH, D_CXX_OBJC | D_CXXWARN }, |
| { "char", RID_CHAR, 0 }, |
| { "char16_t", RID_CHAR16, D_CXXONLY | D_CXX11 | D_CXXWARN }, |
| { "char32_t", RID_CHAR32, D_CXXONLY | D_CXX11 | D_CXXWARN }, |
| { "class", RID_CLASS, D_CXX_OBJC | D_CXXWARN }, |
| { "const", RID_CONST, 0 }, |
| { "constexpr", RID_CONSTEXPR, D_CXXONLY | D_CXX11 | D_CXXWARN }, |
| { "const_cast", RID_CONSTCAST, D_CXXONLY | D_CXXWARN }, |
| { "continue", RID_CONTINUE, 0 }, |
| { "decltype", RID_DECLTYPE, D_CXXONLY | D_CXX11 | D_CXXWARN }, |
| { "default", RID_DEFAULT, 0 }, |
| { "delete", RID_DELETE, D_CXXONLY | D_CXXWARN }, |
| { "do", RID_DO, 0 }, |
| { "double", RID_DOUBLE, 0 }, |
| { "dynamic_cast", RID_DYNCAST, D_CXXONLY | D_CXXWARN }, |
| { "else", RID_ELSE, 0 }, |
| { "enum", RID_ENUM, 0 }, |
| { "explicit", RID_EXPLICIT, D_CXXONLY | D_CXXWARN }, |
| { "export", RID_EXPORT, D_CXXONLY | D_CXXWARN }, |
| { "extern", RID_EXTERN, 0 }, |
| { "false", RID_FALSE, D_CXXONLY | D_CXXWARN }, |
| { "float", RID_FLOAT, 0 }, |
| { "for", RID_FOR, 0 }, |
| { "friend", RID_FRIEND, D_CXXONLY | D_CXXWARN }, |
| { "goto", RID_GOTO, 0 }, |
| { "if", RID_IF, 0 }, |
| { "inline", RID_INLINE, D_EXT89 }, |
| { "int", RID_INT, 0 }, |
| { "long", RID_LONG, 0 }, |
| { "mutable", RID_MUTABLE, D_CXXONLY | D_CXXWARN }, |
| { "namespace", RID_NAMESPACE, D_CXXONLY | D_CXXWARN }, |
| { "new", RID_NEW, D_CXXONLY | D_CXXWARN }, |
| { "noexcept", RID_NOEXCEPT, D_CXXONLY | D_CXX11 | D_CXXWARN }, |
| { "nullptr", RID_NULLPTR, D_CXXONLY | D_CXX11 | D_CXXWARN }, |
| { "operator", RID_OPERATOR, D_CXXONLY | D_CXXWARN }, |
| { "private", RID_PRIVATE, D_CXX_OBJC | D_CXXWARN }, |
| { "protected", RID_PROTECTED, D_CXX_OBJC | D_CXXWARN }, |
| { "public", RID_PUBLIC, D_CXX_OBJC | D_CXXWARN }, |
| { "register", RID_REGISTER, 0 }, |
| { "reinterpret_cast", RID_REINTCAST, D_CXXONLY | D_CXXWARN }, |
| { "restrict", RID_RESTRICT, D_CONLY | D_C99 }, |
| { "return", RID_RETURN, 0 }, |
| { "short", RID_SHORT, 0 }, |
| { "signed", RID_SIGNED, 0 }, |
| { "sizeof", RID_SIZEOF, 0 }, |
| { "static", RID_STATIC, 0 }, |
| { "static_assert", RID_STATIC_ASSERT, D_CXXONLY | D_CXX11 | D_CXXWARN }, |
| { "static_cast", RID_STATCAST, D_CXXONLY | D_CXXWARN }, |
| { "struct", RID_STRUCT, 0 }, |
| { "switch", RID_SWITCH, 0 }, |
| { "template", RID_TEMPLATE, D_CXXONLY | D_CXXWARN }, |
| { "this", RID_THIS, D_CXXONLY | D_CXXWARN }, |
| { "thread_local", RID_THREAD, D_CXXONLY | D_CXX11 | D_CXXWARN }, |
| { "throw", RID_THROW, D_CXX_OBJC | D_CXXWARN }, |
| { "true", RID_TRUE, D_CXXONLY | D_CXXWARN }, |
| { "try", RID_TRY, D_CXX_OBJC | D_CXXWARN }, |
| { "typedef", RID_TYPEDEF, 0 }, |
| { "typename", RID_TYPENAME, D_CXXONLY | D_CXXWARN }, |
| { "typeid", RID_TYPEID, D_CXXONLY | D_CXXWARN }, |
| { "typeof", RID_TYPEOF, D_ASM | D_EXT }, |
| { "union", RID_UNION, 0 }, |
| { "unsigned", RID_UNSIGNED, 0 }, |
| { "using", RID_USING, D_CXXONLY | D_CXXWARN }, |
| { "virtual", RID_VIRTUAL, D_CXXONLY | D_CXXWARN }, |
| { "void", RID_VOID, 0 }, |
| { "volatile", RID_VOLATILE, 0 }, |
| { "wchar_t", RID_WCHAR, D_CXXONLY }, |
| { "while", RID_WHILE, 0 }, |
| |
| /* C++ transactional memory. */ |
| { "synchronized", RID_SYNCHRONIZED, D_CXX_OBJC | D_TRANSMEM }, |
| { "atomic_noexcept", RID_ATOMIC_NOEXCEPT, D_CXXONLY | D_TRANSMEM }, |
| { "atomic_cancel", RID_ATOMIC_CANCEL, D_CXXONLY | D_TRANSMEM }, |
| { "atomic_commit", RID_TRANSACTION_ATOMIC, D_CXXONLY | D_TRANSMEM }, |
| |
| /* Concepts-related keywords */ |
| { "concept", RID_CONCEPT, D_CXX_CONCEPTS_FLAGS | D_CXXWARN }, |
| { "requires", RID_REQUIRES, D_CXX_CONCEPTS_FLAGS | D_CXXWARN }, |
| |
| /* These Objective-C keywords are recognized only immediately after |
| an '@'. */ |
| { "compatibility_alias", RID_AT_ALIAS, D_OBJC }, |
| { "defs", RID_AT_DEFS, D_OBJC }, |
| { "encode", RID_AT_ENCODE, D_OBJC }, |
| { "end", RID_AT_END, D_OBJC }, |
| { "implementation", RID_AT_IMPLEMENTATION, D_OBJC }, |
| { "interface", RID_AT_INTERFACE, D_OBJC }, |
| { "protocol", RID_AT_PROTOCOL, D_OBJC }, |
| { "selector", RID_AT_SELECTOR, D_OBJC }, |
| { "finally", RID_AT_FINALLY, D_OBJC }, |
| { "optional", RID_AT_OPTIONAL, D_OBJC }, |
| { "required", RID_AT_REQUIRED, D_OBJC }, |
| { "property", RID_AT_PROPERTY, D_OBJC }, |
| { "package", RID_AT_PACKAGE, D_OBJC }, |
| { "synthesize", RID_AT_SYNTHESIZE, D_OBJC }, |
| { "dynamic", RID_AT_DYNAMIC, D_OBJC }, |
| /* These are recognized only in protocol-qualifier context |
| (see above) */ |
| { "bycopy", RID_BYCOPY, D_OBJC }, |
| { "byref", RID_BYREF, D_OBJC }, |
| { "in", RID_IN, D_OBJC }, |
| { "inout", RID_INOUT, D_OBJC }, |
| { "oneway", RID_ONEWAY, D_OBJC }, |
| { "out", RID_OUT, D_OBJC }, |
| /* These are recognized inside a property attribute list */ |
| { "assign", RID_ASSIGN, D_OBJC }, |
| { "copy", RID_COPY, D_OBJC }, |
| { "getter", RID_GETTER, D_OBJC }, |
| { "nonatomic", RID_NONATOMIC, D_OBJC }, |
| { "readonly", RID_READONLY, D_OBJC }, |
| { "readwrite", RID_READWRITE, D_OBJC }, |
| { "retain", RID_RETAIN, D_OBJC }, |
| { "setter", RID_SETTER, D_OBJC }, |
| }; |
| |
| const unsigned int num_c_common_reswords = |
| sizeof c_common_reswords / sizeof (struct c_common_resword); |
| |
| /* Return identifier for address space AS. */ |
| |
| const char * |
| c_addr_space_name (addr_space_t as) |
| { |
| int rid = RID_FIRST_ADDR_SPACE + as; |
| gcc_assert (ridpointers [rid]); |
| return IDENTIFIER_POINTER (ridpointers [rid]); |
| } |
| |
| /* Push current bindings for the function name VAR_DECLS. */ |
| |
| void |
| start_fname_decls (void) |
| { |
| unsigned ix; |
| tree saved = NULL_TREE; |
| |
| for (ix = 0; fname_vars[ix].decl; ix++) |
| { |
| tree decl = *fname_vars[ix].decl; |
| |
| if (decl) |
| { |
| saved = tree_cons (decl, build_int_cst (integer_type_node, ix), |
| saved); |
| *fname_vars[ix].decl = NULL_TREE; |
| } |
| } |
| if (saved || saved_function_name_decls) |
| /* Normally they'll have been NULL, so only push if we've got a |
| stack, or they are non-NULL. */ |
| saved_function_name_decls = tree_cons (saved, NULL_TREE, |
| saved_function_name_decls); |
| } |
| |
| /* Finish up the current bindings, adding them into the current function's |
| statement tree. This must be done _before_ finish_stmt_tree is called. |
| If there is no current function, we must be at file scope and no statements |
| are involved. Pop the previous bindings. */ |
| |
| void |
| finish_fname_decls (void) |
| { |
| unsigned ix; |
| tree stmts = NULL_TREE; |
| tree stack = saved_function_name_decls; |
| |
| for (; stack && TREE_VALUE (stack); stack = TREE_CHAIN (stack)) |
| append_to_statement_list (TREE_VALUE (stack), &stmts); |
| |
| if (stmts) |
| { |
| tree *bodyp = &DECL_SAVED_TREE (current_function_decl); |
| |
| if (TREE_CODE (*bodyp) == BIND_EXPR) |
| bodyp = &BIND_EXPR_BODY (*bodyp); |
| |
| append_to_statement_list_force (*bodyp, &stmts); |
| *bodyp = stmts; |
| } |
| |
| for (ix = 0; fname_vars[ix].decl; ix++) |
| *fname_vars[ix].decl = NULL_TREE; |
| |
| if (stack) |
| { |
| /* We had saved values, restore them. */ |
| tree saved; |
| |
| for (saved = TREE_PURPOSE (stack); saved; saved = TREE_CHAIN (saved)) |
| { |
| tree decl = TREE_PURPOSE (saved); |
| unsigned ix = TREE_INT_CST_LOW (TREE_VALUE (saved)); |
| |
| *fname_vars[ix].decl = decl; |
| } |
| stack = TREE_CHAIN (stack); |
| } |
| saved_function_name_decls = stack; |
| } |
| |
| /* Return the text name of the current function, suitably prettified |
| by PRETTY_P. Return string must be freed by caller. */ |
| |
| const char * |
| fname_as_string (int pretty_p) |
| { |
| const char *name = "top level"; |
| char *namep; |
| int vrb = 2, len; |
| cpp_string cstr = { 0, 0 }, strname; |
| |
| if (!pretty_p) |
| { |
| name = ""; |
| vrb = 0; |
| } |
| |
| if (current_function_decl) |
| name = lang_hooks.decl_printable_name (current_function_decl, vrb); |
| |
| len = strlen (name) + 3; /* Two for '"'s. One for NULL. */ |
| |
| namep = XNEWVEC (char, len); |
| snprintf (namep, len, "\"%s\"", name); |
| strname.text = (unsigned char *) namep; |
| strname.len = len - 1; |
| |
| if (cpp_interpret_string (parse_in, &strname, 1, &cstr, CPP_STRING)) |
| { |
| XDELETEVEC (namep); |
| return (const char *) cstr.text; |
| } |
| |
| return namep; |
| } |
| |
| /* Return the VAR_DECL for a const char array naming the current |
| function. If the VAR_DECL has not yet been created, create it |
| now. RID indicates how it should be formatted and IDENTIFIER_NODE |
| ID is its name (unfortunately C and C++ hold the RID values of |
| keywords in different places, so we can't derive RID from ID in |
| this language independent code. LOC is the location of the |
| function. */ |
| |
| tree |
| fname_decl (location_t loc, unsigned int rid, tree id) |
| { |
| unsigned ix; |
| tree decl = NULL_TREE; |
| |
| for (ix = 0; fname_vars[ix].decl; ix++) |
| if (fname_vars[ix].rid == rid) |
| break; |
| |
| decl = *fname_vars[ix].decl; |
| if (!decl) |
| { |
| /* If a tree is built here, it would normally have the lineno of |
| the current statement. Later this tree will be moved to the |
| beginning of the function and this line number will be wrong. |
| To avoid this problem set the lineno to 0 here; that prevents |
| it from appearing in the RTL. */ |
| tree stmts; |
| location_t saved_location = input_location; |
| input_location = UNKNOWN_LOCATION; |
| |
| stmts = push_stmt_list (); |
| decl = (*make_fname_decl) (loc, id, fname_vars[ix].pretty); |
| stmts = pop_stmt_list (stmts); |
| if (!IS_EMPTY_STMT (stmts)) |
| saved_function_name_decls |
| = tree_cons (decl, stmts, saved_function_name_decls); |
| *fname_vars[ix].decl = decl; |
| input_location = saved_location; |
| } |
| if (!ix && !current_function_decl) |
| pedwarn (loc, 0, "%qD is not defined outside of function scope", decl); |
| |
| return decl; |
| } |
| |
| /* Given a STRING_CST, give it a suitable array-of-chars data type. */ |
| |
| tree |
| fix_string_type (tree value) |
| { |
| int length = TREE_STRING_LENGTH (value); |
| int nchars; |
| tree e_type, i_type, a_type; |
| |
| /* Compute the number of elements, for the array type. */ |
| if (TREE_TYPE (value) == char_array_type_node || !TREE_TYPE (value)) |
| { |
| nchars = length; |
| e_type = char_type_node; |
| } |
| else if (TREE_TYPE (value) == char16_array_type_node) |
| { |
| nchars = length / (TYPE_PRECISION (char16_type_node) / BITS_PER_UNIT); |
| e_type = char16_type_node; |
| } |
| else if (TREE_TYPE (value) == char32_array_type_node) |
| { |
| nchars = length / (TYPE_PRECISION (char32_type_node) / BITS_PER_UNIT); |
| e_type = char32_type_node; |
| } |
| else |
| { |
| nchars = length / (TYPE_PRECISION (wchar_type_node) / BITS_PER_UNIT); |
| e_type = wchar_type_node; |
| } |
| |
| /* C89 2.2.4.1, C99 5.2.4.1 (Translation limits). The analogous |
| limit in C++98 Annex B is very large (65536) and is not normative, |
| so we do not diagnose it (warn_overlength_strings is forced off |
| in c_common_post_options). */ |
| if (warn_overlength_strings) |
| { |
| const int nchars_max = flag_isoc99 ? 4095 : 509; |
| const int relevant_std = flag_isoc99 ? 99 : 90; |
| if (nchars - 1 > nchars_max) |
| /* Translators: The %d after 'ISO C' will be 90 or 99. Do not |
| separate the %d from the 'C'. 'ISO' should not be |
| translated, but it may be moved after 'C%d' in languages |
| where modifiers follow nouns. */ |
| pedwarn (input_location, OPT_Woverlength_strings, |
| "string length %qd is greater than the length %qd " |
| "ISO C%d compilers are required to support", |
| nchars - 1, nchars_max, relevant_std); |
| } |
| |
| /* Create the array type for the string constant. The ISO C++ |
| standard says that a string literal has type `const char[N]' or |
| `const wchar_t[N]'. We use the same logic when invoked as a C |
| front-end with -Wwrite-strings. |
| ??? We should change the type of an expression depending on the |
| state of a warning flag. We should just be warning -- see how |
| this is handled in the C++ front-end for the deprecated implicit |
| conversion from string literals to `char*' or `wchar_t*'. |
| |
| The C++ front end relies on TYPE_MAIN_VARIANT of a cv-qualified |
| array type being the unqualified version of that type. |
| Therefore, if we are constructing an array of const char, we must |
| construct the matching unqualified array type first. The C front |
| end does not require this, but it does no harm, so we do it |
| unconditionally. */ |
| i_type = build_index_type (size_int (nchars - 1)); |
| a_type = build_array_type (e_type, i_type); |
| if (c_dialect_cxx() || warn_write_strings) |
| a_type = c_build_qualified_type (a_type, TYPE_QUAL_CONST); |
| |
| TREE_TYPE (value) = a_type; |
| TREE_CONSTANT (value) = 1; |
| TREE_READONLY (value) = 1; |
| TREE_STATIC (value) = 1; |
| return value; |
| } |
| |
| /* Given a string of type STRING_TYPE, determine what kind of string |
| token would give an equivalent execution encoding: CPP_STRING, |
| CPP_STRING16, or CPP_STRING32. Return CPP_OTHER in case of error. |
| This may not be exactly the string token type that initially created |
| the string, since CPP_WSTRING is indistinguishable from the 16/32 bit |
| string type at this point. |
| |
| This effectively reverses part of the logic in lex_string and |
| fix_string_type. */ |
| |
| static enum cpp_ttype |
| get_cpp_ttype_from_string_type (tree string_type) |
| { |
| gcc_assert (string_type); |
| if (TREE_CODE (string_type) == POINTER_TYPE) |
| string_type = TREE_TYPE (string_type); |
| |
| if (TREE_CODE (string_type) != ARRAY_TYPE) |
| return CPP_OTHER; |
| |
| tree element_type = TREE_TYPE (string_type); |
| if (TREE_CODE (element_type) != INTEGER_TYPE) |
| return CPP_OTHER; |
| |
| int bits_per_character = TYPE_PRECISION (element_type); |
| switch (bits_per_character) |
| { |
| case 8: |
| return CPP_STRING; /* It could have also been CPP_UTF8STRING. */ |
| case 16: |
| return CPP_STRING16; |
| case 32: |
| return CPP_STRING32; |
| } |
| |
| return CPP_OTHER; |
| } |
| |
| /* The global record of string concatentations, for use in |
| extracting locations within string literals. */ |
| |
| GTY(()) string_concat_db *g_string_concat_db; |
| |
| /* Implementation of LANG_HOOKS_GET_SUBSTRING_LOCATION. */ |
| |
| const char * |
| c_get_substring_location (const substring_loc &substr_loc, |
| location_t *out_loc) |
| { |
| enum cpp_ttype tok_type |
| = get_cpp_ttype_from_string_type (substr_loc.get_string_type ()); |
| if (tok_type == CPP_OTHER) |
| return "unrecognized string type"; |
| |
| return get_source_location_for_substring (parse_in, g_string_concat_db, |
| substr_loc.get_fmt_string_loc (), |
| tok_type, |
| substr_loc.get_caret_idx (), |
| substr_loc.get_start_idx (), |
| substr_loc.get_end_idx (), |
| out_loc); |
| } |
| |
| |
| /* Fold X for consideration by one of the warning functions when checking |
| whether an expression has a constant value. */ |
| |
| tree |
| fold_for_warn (tree x) |
| { |
| if (c_dialect_cxx ()) |
| return c_fully_fold (x, /*for_init*/false, /*maybe_constp*/NULL); |
| else |
| /* The C front-end has already folded X appropriately. */ |
| return x; |
| } |
| |
| /* Return true iff T is a boolean promoted to int. */ |
| |
| bool |
| bool_promoted_to_int_p (tree t) |
| { |
| return (CONVERT_EXPR_P (t) |
| && TREE_TYPE (t) == integer_type_node |
| && TREE_CODE (TREE_TYPE (TREE_OPERAND (t, 0))) == BOOLEAN_TYPE); |
| } |
| |
| /* vector_targets_convertible_p is used for vector pointer types. The |
| callers perform various checks that the qualifiers are satisfactory, |
| while OTOH vector_targets_convertible_p ignores the number of elements |
| in the vectors. That's fine with vector pointers as we can consider, |
| say, a vector of 8 elements as two consecutive vectors of 4 elements, |
| and that does not require and conversion of the pointer values. |
| In contrast, vector_types_convertible_p and |
| vector_types_compatible_elements_p are used for vector value types. */ |
| /* True if pointers to distinct types T1 and T2 can be converted to |
| each other without an explicit cast. Only returns true for opaque |
| vector types. */ |
| bool |
| vector_targets_convertible_p (const_tree t1, const_tree t2) |
| { |
| if (VECTOR_TYPE_P (t1) && VECTOR_TYPE_P (t2) |
| && (TYPE_VECTOR_OPAQUE (t1) || TYPE_VECTOR_OPAQUE (t2)) |
| && tree_int_cst_equal (TYPE_SIZE (t1), TYPE_SIZE (t2))) |
| return true; |
| |
| return false; |
| } |
| |
| /* vector_types_convertible_p is used for vector value types. |
| It could in principle call vector_targets_convertible_p as a subroutine, |
| but then the check for vector type would be duplicated with its callers, |
| and also the purpose of vector_targets_convertible_p would become |
| muddled. |
| Where vector_types_convertible_p returns true, a conversion might still be |
| needed to make the types match. |
| In contrast, vector_targets_convertible_p is used for vector pointer |
| values, and vector_types_compatible_elements_p is used specifically |
| in the context for binary operators, as a check if use is possible without |
| conversion. */ |
| /* True if vector types T1 and T2 can be converted to each other |
| without an explicit cast. If EMIT_LAX_NOTE is true, and T1 and T2 |
| can only be converted with -flax-vector-conversions yet that is not |
| in effect, emit a note telling the user about that option if such |
| a note has not previously been emitted. */ |
| bool |
| vector_types_convertible_p (const_tree t1, const_tree t2, bool emit_lax_note) |
| { |
| static bool emitted_lax_note = false; |
| bool convertible_lax; |
| |
| if ((TYPE_VECTOR_OPAQUE (t1) || TYPE_VECTOR_OPAQUE (t2)) |
| && tree_int_cst_equal (TYPE_SIZE (t1), TYPE_SIZE (t2))) |
| return true; |
| |
| convertible_lax = |
| (tree_int_cst_equal (TYPE_SIZE (t1), TYPE_SIZE (t2)) |
| && (TREE_CODE (TREE_TYPE (t1)) != REAL_TYPE || |
| TYPE_VECTOR_SUBPARTS (t1) == TYPE_VECTOR_SUBPARTS (t2)) |
| && (INTEGRAL_TYPE_P (TREE_TYPE (t1)) |
| == INTEGRAL_TYPE_P (TREE_TYPE (t2)))); |
| |
| if (!convertible_lax || flag_lax_vector_conversions) |
| return convertible_lax; |
| |
| if (TYPE_VECTOR_SUBPARTS (t1) == TYPE_VECTOR_SUBPARTS (t2) |
| && lang_hooks.types_compatible_p (TREE_TYPE (t1), TREE_TYPE (t2))) |
| return true; |
| |
| if (emit_lax_note && !emitted_lax_note) |
| { |
| emitted_lax_note = true; |
| inform (input_location, "use -flax-vector-conversions to permit " |
| "conversions between vectors with differing " |
| "element types or numbers of subparts"); |
| } |
| |
| return false; |
| } |
| |
| /* Build a VEC_PERM_EXPR if V0, V1 and MASK are not error_mark_nodes |
| and have vector types, V0 has the same type as V1, and the number of |
| elements of V0, V1, MASK is the same. |
| |
| In case V1 is a NULL_TREE it is assumed that __builtin_shuffle was |
| called with two arguments. In this case implementation passes the |
| first argument twice in order to share the same tree code. This fact |
| could enable the mask-values being twice the vector length. This is |
| an implementation accident and this semantics is not guaranteed to |
| the user. */ |
| tree |
| c_build_vec_perm_expr (location_t loc, tree v0, tree v1, tree mask, |
| bool complain) |
| { |
| tree ret; |
| bool wrap = true; |
| bool maybe_const = false; |
| bool two_arguments = false; |
| |
| if (v1 == NULL_TREE) |
| { |
| two_arguments = true; |
| v1 = v0; |
| } |
| |
| if (v0 == error_mark_node || v1 == error_mark_node |
| || mask == error_mark_node) |
| return error_mark_node; |
| |
| if (!VECTOR_INTEGER_TYPE_P (TREE_TYPE (mask))) |
| { |
| if (complain) |
| error_at (loc, "__builtin_shuffle last argument must " |
| "be an integer vector"); |
| return error_mark_node; |
| } |
| |
| if (!VECTOR_TYPE_P (TREE_TYPE (v0)) |
| || !VECTOR_TYPE_P (TREE_TYPE (v1))) |
| { |
| if (complain) |
| error_at (loc, "__builtin_shuffle arguments must be vectors"); |
| return error_mark_node; |
| } |
| |
| if (TYPE_MAIN_VARIANT (TREE_TYPE (v0)) != TYPE_MAIN_VARIANT (TREE_TYPE (v1))) |
| { |
| if (complain) |
| error_at (loc, "__builtin_shuffle argument vectors must be of " |
| "the same type"); |
| return error_mark_node; |
| } |
| |
| if (TYPE_VECTOR_SUBPARTS (TREE_TYPE (v0)) |
| != TYPE_VECTOR_SUBPARTS (TREE_TYPE (mask)) |
| && TYPE_VECTOR_SUBPARTS (TREE_TYPE (v1)) |
| != TYPE_VECTOR_SUBPARTS (TREE_TYPE (mask))) |
| { |
| if (complain) |
| error_at (loc, "__builtin_shuffle number of elements of the " |
| "argument vector(s) and the mask vector should " |
| "be the same"); |
| return error_mark_node; |
| } |
| |
| if (GET_MODE_BITSIZE (TYPE_MODE (TREE_TYPE (TREE_TYPE (v0)))) |
| != GET_MODE_BITSIZE (TYPE_MODE (TREE_TYPE (TREE_TYPE (mask))))) |
| { |
| if (complain) |
| error_at (loc, "__builtin_shuffle argument vector(s) inner type " |
| "must have the same size as inner type of the mask"); |
| return error_mark_node; |
| } |
| |
| if (!c_dialect_cxx ()) |
| { |
| /* Avoid C_MAYBE_CONST_EXPRs inside VEC_PERM_EXPR. */ |
| v0 = c_fully_fold (v0, false, &maybe_const); |
| wrap &= maybe_const; |
| |
| if (two_arguments) |
| v1 = v0 = save_expr (v0); |
| else |
| { |
| v1 = c_fully_fold (v1, false, &maybe_const); |
| wrap &= maybe_const; |
| } |
| |
| mask = c_fully_fold (mask, false, &maybe_const); |
| wrap &= maybe_const; |
| } |
| else if (two_arguments) |
| v1 = v0 = save_expr (v0); |
| |
| ret = build3_loc (loc, VEC_PERM_EXPR, TREE_TYPE (v0), v0, v1, mask); |
| |
| if (!c_dialect_cxx () && !wrap) |
| ret = c_wrap_maybe_const (ret, true); |
| |
| return ret; |
| } |
| |
| /* Like tree.c:get_narrower, but retain conversion from C++0x scoped enum |
| to integral type. */ |
| |
| tree |
| c_common_get_narrower (tree op, int *unsignedp_ptr) |
| { |
| op = get_narrower (op, unsignedp_ptr); |
| |
| if (TREE_CODE (TREE_TYPE (op)) == ENUMERAL_TYPE |
| && ENUM_IS_SCOPED (TREE_TYPE (op))) |
| { |
| /* C++0x scoped enumerations don't implicitly convert to integral |
| type; if we stripped an explicit conversion to a larger type we |
| need to replace it so common_type will still work. */ |
| tree type = c_common_type_for_size (TYPE_PRECISION (TREE_TYPE (op)), |
| TYPE_UNSIGNED (TREE_TYPE (op))); |
| op = fold_convert (type, op); |
| } |
| return op; |
| } |
| |
| /* This is a helper function of build_binary_op. |
| |
| For certain operations if both args were extended from the same |
| smaller type, do the arithmetic in that type and then extend. |
| |
| BITWISE indicates a bitwise operation. |
| For them, this optimization is safe only if |
| both args are zero-extended or both are sign-extended. |
| Otherwise, we might change the result. |
| Eg, (short)-1 | (unsigned short)-1 is (int)-1 |
| but calculated in (unsigned short) it would be (unsigned short)-1. |
| */ |
| tree |
| shorten_binary_op (tree result_type, tree op0, tree op1, bool bitwise) |
| { |
| int unsigned0, unsigned1; |
| tree arg0, arg1; |
| int uns; |
| tree type; |
| |
| /* Cast OP0 and OP1 to RESULT_TYPE. Doing so prevents |
| excessive narrowing when we call get_narrower below. For |
| example, suppose that OP0 is of unsigned int extended |
| from signed char and that RESULT_TYPE is long long int. |
| If we explicitly cast OP0 to RESULT_TYPE, OP0 would look |
| like |
| |
| (long long int) (unsigned int) signed_char |
| |
| which get_narrower would narrow down to |
| |
| (unsigned int) signed char |
| |
| If we do not cast OP0 first, get_narrower would return |
| signed_char, which is inconsistent with the case of the |
| explicit cast. */ |
| op0 = convert (result_type, op0); |
| op1 = convert (result_type, op1); |
| |
| arg0 = c_common_get_narrower (op0, &unsigned0); |
| arg1 = c_common_get_narrower (op1, &unsigned1); |
| |
| /* UNS is 1 if the operation to be done is an unsigned one. */ |
| uns = TYPE_UNSIGNED (result_type); |
| |
| /* Handle the case that OP0 (or OP1) does not *contain* a conversion |
| but it *requires* conversion to FINAL_TYPE. */ |
| |
| if ((TYPE_PRECISION (TREE_TYPE (op0)) |
| == TYPE_PRECISION (TREE_TYPE (arg0))) |
| && TREE_TYPE (op0) != result_type) |
| unsigned0 = TYPE_UNSIGNED (TREE_TYPE (op0)); |
| if ((TYPE_PRECISION (TREE_TYPE (op1)) |
| == TYPE_PRECISION (TREE_TYPE (arg1))) |
| && TREE_TYPE (op1) != result_type) |
| unsigned1 = TYPE_UNSIGNED (TREE_TYPE (op1)); |
| |
| /* Now UNSIGNED0 is 1 if ARG0 zero-extends to FINAL_TYPE. */ |
| |
| /* For bitwise operations, signedness of nominal type |
| does not matter. Consider only how operands were extended. */ |
| if (bitwise) |
| uns = unsigned0; |
| |
| /* Note that in all three cases below we refrain from optimizing |
| an unsigned operation on sign-extended args. |
| That would not be valid. */ |
| |
| /* Both args variable: if both extended in same way |
| from same width, do it in that width. |
| Do it unsigned if args were zero-extended. */ |
| if ((TYPE_PRECISION (TREE_TYPE (arg0)) |
| < TYPE_PRECISION (result_type)) |
| && (TYPE_PRECISION (TREE_TYPE (arg1)) |
| == TYPE_PRECISION (TREE_TYPE (arg0))) |
| && unsigned0 == unsigned1 |
| && (unsigned0 || !uns)) |
| return c_common_signed_or_unsigned_type |
| (unsigned0, common_type (TREE_TYPE (arg0), TREE_TYPE (arg1))); |
| |
| else if (TREE_CODE (arg0) == INTEGER_CST |
| && (unsigned1 || !uns) |
| && (TYPE_PRECISION (TREE_TYPE (arg1)) |
| < TYPE_PRECISION (result_type)) |
| && (type |
| = c_common_signed_or_unsigned_type (unsigned1, |
| TREE_TYPE (arg1))) |
| && !POINTER_TYPE_P (type) |
| && int_fits_type_p (arg0, type)) |
| return type; |
| |
| else if (TREE_CODE (arg1) == INTEGER_CST |
| && (unsigned0 || !uns) |
| && (TYPE_PRECISION (TREE_TYPE (arg0)) |
| < TYPE_PRECISION (result_type)) |
| && (type |
| = c_common_signed_or_unsigned_type (unsigned0, |
| TREE_TYPE (arg0))) |
| && !POINTER_TYPE_P (type) |
| && int_fits_type_p (arg1, type)) |
| return type; |
| |
| return result_type; |
| } |
| |
| /* Returns true iff any integer value of type FROM_TYPE can be represented as |
| real of type TO_TYPE. This is a helper function for unsafe_conversion_p. */ |
| |
| static bool |
| int_safely_convertible_to_real_p (const_tree from_type, const_tree to_type) |
| { |
| tree type_low_bound = TYPE_MIN_VALUE (from_type); |
| tree type_high_bound = TYPE_MAX_VALUE (from_type); |
| REAL_VALUE_TYPE real_low_bound = |
| real_value_from_int_cst (0, type_low_bound); |
| REAL_VALUE_TYPE real_high_bound = |
| real_value_from_int_cst (0, type_high_bound); |
| |
| return exact_real_truncate (TYPE_MODE (to_type), &real_low_bound) |
| && exact_real_truncate (TYPE_MODE (to_type), &real_high_bound); |
| } |
| |
| /* Checks if expression EXPR of complex/real/integer type cannot be converted |
| to the complex/real/integer type TYPE. Function returns non-zero when: |
| * EXPR is a constant which cannot be exactly converted to TYPE. |
| * EXPR is not a constant and size of EXPR's type > than size of TYPE, |
| for EXPR type and TYPE being both integers or both real, or both |
| complex. |
| * EXPR is not a constant of complex type and TYPE is a real or |
| an integer. |
| * EXPR is not a constant of real type and TYPE is an integer. |
| * EXPR is not a constant of integer type which cannot be |
| exactly converted to real type. |
| |
| Function allows conversions between types of different signedness and |
| can return SAFE_CONVERSION (zero) in that case. Function can produce |
| signedness warnings if PRODUCE_WARNS is true. |
| |
| Function allows conversions from complex constants to non-complex types, |
| provided that imaginary part is zero and real part can be safely converted |
| to TYPE. */ |
| |
| enum conversion_safety |
| unsafe_conversion_p (location_t loc, tree type, tree expr, bool produce_warns) |
| { |
| enum conversion_safety give_warning = SAFE_CONVERSION; /* is 0 or false */ |
| tree expr_type = TREE_TYPE (expr); |
| loc = expansion_point_location_if_in_system_header (loc); |
| |
| if (TREE_CODE (expr) == REAL_CST || TREE_CODE (expr) == INTEGER_CST) |
| { |
| /* If type is complex, we are interested in compatibility with |
| underlying type. */ |
| if (TREE_CODE (type) == COMPLEX_TYPE) |
| type = TREE_TYPE (type); |
| |
| /* Warn for real constant that is not an exact integer converted |
| to integer type. */ |
| if (TREE_CODE (expr_type) == REAL_TYPE |
| && TREE_CODE (type) == INTEGER_TYPE) |
| { |
| if (!real_isinteger (TREE_REAL_CST_PTR (expr), TYPE_MODE (expr_type))) |
| give_warning = UNSAFE_REAL; |
| } |
| /* Warn for an integer constant that does not fit into integer type. */ |
| else if (TREE_CODE (expr_type) == INTEGER_TYPE |
| && TREE_CODE (type) == INTEGER_TYPE |
| && !int_fits_type_p (expr, type)) |
| { |
| if (TYPE_UNSIGNED (type) && !TYPE_UNSIGNED (expr_type) |
| && tree_int_cst_sgn (expr) < 0) |
| { |
| if (produce_warns) |
| warning_at (loc, OPT_Wsign_conversion, "negative integer" |
| " implicitly converted to unsigned type"); |
| } |
| else if (!TYPE_UNSIGNED (type) && TYPE_UNSIGNED (expr_type)) |
| { |
| if (produce_warns) |
| warning_at (loc, OPT_Wsign_conversion, "conversion of unsigned" |
| " constant value to negative integer"); |
| } |
| else |
| give_warning = UNSAFE_OTHER; |
| } |
| else if (TREE_CODE (type) == REAL_TYPE) |
| { |
| /* Warn for an integer constant that does not fit into real type. */ |
| if (TREE_CODE (expr_type) == INTEGER_TYPE) |
| { |
| REAL_VALUE_TYPE a = real_value_from_int_cst (0, expr); |
| if (!exact_real_truncate (TYPE_MODE (type), &a)) |
| give_warning = UNSAFE_REAL; |
| } |
| /* Warn for a real constant that does not fit into a smaller |
| real type. */ |
| else if (TREE_CODE (expr_type) == REAL_TYPE |
| && TYPE_PRECISION (type) < TYPE_PRECISION (expr_type)) |
| { |
| REAL_VALUE_TYPE a = TREE_REAL_CST (expr); |
| if (!exact_real_truncate (TYPE_MODE (type), &a)) |
| give_warning = UNSAFE_REAL; |
| } |
| } |
| } |
| |
| else if (TREE_CODE (expr) == COMPLEX_CST) |
| { |
| tree imag_part = TREE_IMAGPART (expr); |
| /* Conversion from complex constant with zero imaginary part, |
| perform check for conversion of real part. */ |
| if ((TREE_CODE (imag_part) == REAL_CST |
| && real_zerop (imag_part)) |
| || (TREE_CODE (imag_part) == INTEGER_CST |
| && integer_zerop (imag_part))) |
| /* Note: in this branch we use recursive call to unsafe_conversion_p |
| with different type of EXPR, but it is still safe, because when EXPR |
| is a constant, it's type is not used in text of generated warnings |
| (otherwise they could sound misleading). */ |
| return unsafe_conversion_p (loc, type, TREE_REALPART (expr), |
| produce_warns); |
| /* Conversion from complex constant with non-zero imaginary part. */ |
| else |
| { |
| /* Conversion to complex type. |
| Perform checks for both real and imaginary parts. */ |
| if (TREE_CODE (type) == COMPLEX_TYPE) |
| { |
| /* Unfortunately, produce_warns must be false in two subsequent |
| calls of unsafe_conversion_p, because otherwise we could |
| produce strange "double" warnings, if both real and imaginary |
| parts have conversion problems related to signedness. |
| |
| For example: |
| int32_t _Complex a = 0x80000000 + 0x80000000i; |
| |
| Possible solution: add a separate function for checking |
| constants and combine result of two calls appropriately. */ |
| enum conversion_safety re_safety = |
| unsafe_conversion_p (loc, type, TREE_REALPART (expr), false); |
| enum conversion_safety im_safety = |
| unsafe_conversion_p (loc, type, imag_part, false); |
| |
| /* Merge the results into appropriate single warning. */ |
| |
| /* Note: this case includes SAFE_CONVERSION, i.e. success. */ |
| if (re_safety == im_safety) |
| give_warning = re_safety; |
| else if (!re_safety && im_safety) |
| give_warning = im_safety; |
| else if (re_safety && !im_safety) |
| give_warning = re_safety; |
| else |
| give_warning = UNSAFE_OTHER; |
| } |
| /* Warn about conversion from complex to real or integer type. */ |
| else |
| give_warning = UNSAFE_IMAGINARY; |
| } |
| } |
| |
| /* Checks for remaining case: EXPR is not constant. */ |
| else |
| { |
| /* Warn for real types converted to integer types. */ |
| if (TREE_CODE (expr_type) == REAL_TYPE |
| && TREE_CODE (type) == INTEGER_TYPE) |
| give_warning = UNSAFE_REAL; |
| |
| else if (TREE_CODE (expr_type) == INTEGER_TYPE |
| && TREE_CODE (type) == INTEGER_TYPE) |
| { |
| /* Don't warn about unsigned char y = 0xff, x = (int) y; */ |
| expr = get_unwidened (expr, 0); |
| expr_type = TREE_TYPE (expr); |
| |
| /* Don't warn for short y; short x = ((int)y & 0xff); */ |
| if (TREE_CODE (expr) == BIT_AND_EXPR |
| || TREE_CODE (expr) == BIT_IOR_EXPR |
| || TREE_CODE (expr) == BIT_XOR_EXPR) |
| { |
| /* If both args were extended from a shortest type, |
| use that type if that is safe. */ |
| expr_type = shorten_binary_op (expr_type, |
| TREE_OPERAND (expr, 0), |
| TREE_OPERAND (expr, 1), |
| /* bitwise */1); |
| |
| if (TREE_CODE (expr) == BIT_AND_EXPR) |
| { |
| tree op0 = TREE_OPERAND (expr, 0); |
| tree op1 = TREE_OPERAND (expr, 1); |
| bool unsigned0 = TYPE_UNSIGNED (TREE_TYPE (op0)); |
| bool unsigned1 = TYPE_UNSIGNED (TREE_TYPE (op1)); |
| |
| /* If one of the operands is a non-negative constant |
| that fits in the target type, then the type of the |
| other operand does not matter. */ |
| if ((TREE_CODE (op0) == INTEGER_CST |
| && int_fits_type_p (op0, c_common_signed_type (type)) |
| && int_fits_type_p (op0, c_common_unsigned_type (type))) |
| || (TREE_CODE (op1) == INTEGER_CST |
| && int_fits_type_p (op1, c_common_signed_type (type)) |
| && int_fits_type_p (op1, |
| c_common_unsigned_type (type)))) |
| return SAFE_CONVERSION; |
| /* If constant is unsigned and fits in the target |
| type, then the result will also fit. */ |
| else if ((TREE_CODE (op0) == INTEGER_CST |
| && unsigned0 |
| && int_fits_type_p (op0, type)) |
| || (TREE_CODE (op1) == INTEGER_CST |
| && unsigned1 |
| && int_fits_type_p (op1, type))) |
| return SAFE_CONVERSION; |
| } |
| } |
| /* Warn for integer types converted to smaller integer types. */ |
| if (TYPE_PRECISION (type) < TYPE_PRECISION (expr_type)) |
| give_warning = UNSAFE_OTHER; |
| |
| /* When they are the same width but different signedness, |
| then the value may change. */ |
| else if (((TYPE_PRECISION (type) == TYPE_PRECISION (expr_type) |
| && TYPE_UNSIGNED (expr_type) != TYPE_UNSIGNED (type)) |
| /* Even when converted to a bigger type, if the type is |
| unsigned but expr is signed, then negative values |
| will be changed. */ |
| || (TYPE_UNSIGNED (type) && !TYPE_UNSIGNED (expr_type))) |
| && produce_warns) |
| warning_at (loc, OPT_Wsign_conversion, "conversion to %qT from %qT " |
| "may change the sign of the result", |
| type, expr_type); |
| } |
| |
| /* Warn for integer types converted to real types if and only if |
| all the range of values of the integer type cannot be |
| represented by the real type. */ |
| else if (TREE_CODE (expr_type) == INTEGER_TYPE |
| && TREE_CODE (type) == REAL_TYPE) |
| { |
| /* Don't warn about char y = 0xff; float x = (int) y; */ |
| expr = get_unwidened (expr, 0); |
| expr_type = TREE_TYPE (expr); |
| |
| if (!int_safely_convertible_to_real_p (expr_type, type)) |
| give_warning = UNSAFE_OTHER; |
| } |
| |
| /* Warn for real types converted to smaller real types. */ |
| else if (TREE_CODE (expr_type) == REAL_TYPE |
| && TREE_CODE (type) == REAL_TYPE |
| && TYPE_PRECISION (type) < TYPE_PRECISION (expr_type)) |
| give_warning = UNSAFE_REAL; |
| |
| /* Check conversion between two complex types. */ |
| else if (TREE_CODE (expr_type) == COMPLEX_TYPE |
| && TREE_CODE (type) == COMPLEX_TYPE) |
| { |
| /* Extract underlying types (i.e., type of real and imaginary |
| parts) of expr_type and type. */ |
| tree from_type = TREE_TYPE (expr_type); |
| tree to_type = TREE_TYPE (type); |
| |
| /* Warn for real types converted to integer types. */ |
| if (TREE_CODE (from_type) == REAL_TYPE |
| && TREE_CODE (to_type) == INTEGER_TYPE) |
| give_warning = UNSAFE_REAL; |
| |
| /* Warn for real types converted to smaller real types. */ |
| else if (TREE_CODE (from_type) == REAL_TYPE |
| && TREE_CODE (to_type) == REAL_TYPE |
| && TYPE_PRECISION (to_type) < TYPE_PRECISION (from_type)) |
| give_warning = UNSAFE_REAL; |
| |
| /* Check conversion for complex integer types. Here implementation |
| is simpler than for real-domain integers because it does not |
| involve sophisticated cases, such as bitmasks, casts, etc. */ |
| else if (TREE_CODE (from_type) == INTEGER_TYPE |
| && TREE_CODE (to_type) == INTEGER_TYPE) |
| { |
| /* Warn for integer types converted to smaller integer types. */ |
| if (TYPE_PRECISION (to_type) < TYPE_PRECISION (from_type)) |
| give_warning = UNSAFE_OTHER; |
| |
| /* Check for different signedness, see case for real-domain |
| integers (above) for a more detailed comment. */ |
| else if (((TYPE_PRECISION (to_type) == TYPE_PRECISION (from_type) |
| && TYPE_UNSIGNED (to_type) != TYPE_UNSIGNED (from_type)) |
| || (TYPE_UNSIGNED (to_type) && !TYPE_UNSIGNED (from_type))) |
| && produce_warns) |
| warning_at (loc, OPT_Wsign_conversion, |
| "conversion to %qT from %qT " |
| "may change the sign of the result", |
| type, expr_type); |
| } |
| else if (TREE_CODE (from_type) == INTEGER_TYPE |
| && TREE_CODE (to_type) == REAL_TYPE |
| && !int_safely_convertible_to_real_p (from_type, to_type)) |
| give_warning = UNSAFE_OTHER; |
| } |
| |
| /* Warn for complex types converted to real or integer types. */ |
| else if (TREE_CODE (expr_type) == COMPLEX_TYPE |
| && TREE_CODE (type) != COMPLEX_TYPE) |
| give_warning = UNSAFE_IMAGINARY; |
| } |
| |
| return give_warning; |
| } |
| |
| |
| /* Convert EXPR to TYPE, warning about conversion problems with constants. |
| Invoke this function on every expression that is converted implicitly, |
| i.e. because of language rules and not because of an explicit cast. */ |
| |
| tree |
| convert_and_check (location_t loc, tree type, tree expr) |
| { |
| tree result; |
| tree expr_for_warning; |
| |
| /* Convert from a value with possible excess precision rather than |
| via the semantic type, but do not warn about values not fitting |
| exactly in the semantic type. */ |
| if (TREE_CODE (expr) == EXCESS_PRECISION_EXPR) |
| { |
| tree orig_type = TREE_TYPE (expr); |
| expr = TREE_OPERAND (expr, 0); |
| expr_for_warning = convert (orig_type, expr); |
| if (orig_type == type) |
| return expr_for_warning; |
| } |
| else |
| expr_for_warning = expr; |
| |
| if (TREE_TYPE (expr) == type) |
| return expr; |
| |
| result = convert (type, expr); |
| |
| if (c_inhibit_evaluation_warnings == 0 |
| && !TREE_OVERFLOW_P (expr) |
| && result != error_mark_node) |
| warnings_for_convert_and_check (loc, type, expr_for_warning, result); |
| |
| return result; |
| } |
| |
| /* A node in a list that describes references to variables (EXPR), which are |
| either read accesses if WRITER is zero, or write accesses, in which case |
| WRITER is the parent of EXPR. */ |
| struct tlist |
| { |
| struct tlist *next; |
| tree expr, writer; |
| }; |
| |
| /* Used to implement a cache the results of a call to verify_tree. We only |
| use this for SAVE_EXPRs. */ |
| struct tlist_cache |
| { |
| struct tlist_cache *next; |
| struct tlist *cache_before_sp; |
| struct tlist *cache_after_sp; |
| tree expr; |
| }; |
| |
| /* Obstack to use when allocating tlist structures, and corresponding |
| firstobj. */ |
| static struct obstack tlist_obstack; |
| static char *tlist_firstobj = 0; |
| |
| /* Keep track of the identifiers we've warned about, so we can avoid duplicate |
| warnings. */ |
| static struct tlist *warned_ids; |
| /* SAVE_EXPRs need special treatment. We process them only once and then |
| cache the results. */ |
| static struct tlist_cache *save_expr_cache; |
| |
| static void add_tlist (struct tlist **, struct tlist *, tree, int); |
| static void merge_tlist (struct tlist **, struct tlist *, int); |
| static void verify_tree (tree, struct tlist **, struct tlist **, tree); |
| static bool warning_candidate_p (tree); |
| static bool candidate_equal_p (const_tree, const_tree); |
| static void warn_for_collisions (struct tlist *); |
| static void warn_for_collisions_1 (tree, tree, struct tlist *, int); |
| static struct tlist *new_tlist (struct tlist *, tree, tree); |
| |
| /* Create a new struct tlist and fill in its fields. */ |
| static struct tlist * |
| new_tlist (struct tlist *next, tree t, tree writer) |
| { |
| struct tlist *l; |
| l = XOBNEW (&tlist_obstack, struct tlist); |
| l->next = next; |
| l->expr = t; |
| l->writer = writer; |
| return l; |
| } |
| |
| /* Add duplicates of the nodes found in ADD to the list *TO. If EXCLUDE_WRITER |
| is nonnull, we ignore any node we find which has a writer equal to it. */ |
| |
| static void |
| add_tlist (struct tlist **to, struct tlist *add, tree exclude_writer, int copy) |
| { |
| while (add) |
| { |
| struct tlist *next = add->next; |
| if (!copy) |
| add->next = *to; |
| if (!exclude_writer || !candidate_equal_p (add->writer, exclude_writer)) |
| *to = copy ? new_tlist (*to, add->expr, add->writer) : add; |
| add = next; |
| } |
| } |
| |
| /* Merge the nodes of ADD into TO. This merging process is done so that for |
| each variable that already exists in TO, no new node is added; however if |
| there is a write access recorded in ADD, and an occurrence on TO is only |
| a read access, then the occurrence in TO will be modified to record the |
| write. */ |
| |
| static void |
| merge_tlist (struct tlist **to, struct tlist *add, int copy) |
| { |
| struct tlist **end = to; |
| |
| while (*end) |
| end = &(*end)->next; |
| |
| while (add) |
| { |
| int found = 0; |
| struct tlist *tmp2; |
| struct tlist *next = add->next; |
| |
| for (tmp2 = *to; tmp2; tmp2 = tmp2->next) |
| if (candidate_equal_p (tmp2->expr, add->expr)) |
| { |
| found = 1; |
| if (!tmp2->writer) |
| tmp2->writer = add->writer; |
| } |
| if (!found) |
| { |
| *end = copy ? new_tlist (NULL, add->expr, add->writer) : add; |
| end = &(*end)->next; |
| *end = 0; |
| } |
| add = next; |
| } |
| } |
| |
| /* WRITTEN is a variable, WRITER is its parent. Warn if any of the variable |
| references in list LIST conflict with it, excluding reads if ONLY writers |
| is nonzero. */ |
| |
| static void |
| warn_for_collisions_1 (tree written, tree writer, struct tlist *list, |
| int only_writes) |
| { |
| struct tlist *tmp; |
| |
| /* Avoid duplicate warnings. */ |
| for (tmp = warned_ids; tmp; tmp = tmp->next) |
| if (candidate_equal_p (tmp->expr, written)) |
| return; |
| |
| while (list) |
| { |
| if (candidate_equal_p (list->expr, written) |
| && !candidate_equal_p (list->writer, writer) |
| && (!only_writes || list->writer)) |
| { |
| warned_ids = new_tlist (warned_ids, written, NULL_TREE); |
| warning_at (EXPR_LOC_OR_LOC (writer, input_location), |
| OPT_Wsequence_point, "operation on %qE may be undefined", |
| list->expr); |
| } |
| list = list->next; |
| } |
| } |
| |
| /* Given a list LIST of references to variables, find whether any of these |
| can cause conflicts due to missing sequence points. */ |
| |
| static void |
| warn_for_collisions (struct tlist *list) |
| { |
| struct tlist *tmp; |
| |
| for (tmp = list; tmp; tmp = tmp->next) |
| { |
| if (tmp->writer) |
| warn_for_collisions_1 (tmp->expr, tmp->writer, list, 0); |
| } |
| } |
| |
| /* Return nonzero if X is a tree that can be verified by the sequence point |
| warnings. */ |
| |
| static bool |
| warning_candidate_p (tree x) |
| { |
| if (DECL_P (x) && DECL_ARTIFICIAL (x)) |
| return false; |
| |
| if (TREE_CODE (x) == BLOCK) |
| return false; |
| |
| /* VOID_TYPE_P (TREE_TYPE (x)) is workaround for cp/tree.c |
| (lvalue_p) crash on TRY/CATCH. */ |
| if (TREE_TYPE (x) == NULL_TREE || VOID_TYPE_P (TREE_TYPE (x))) |
| return false; |
| |
| if (!lvalue_p (x)) |
| return false; |
| |
| /* No point to track non-const calls, they will never satisfy |
| operand_equal_p. */ |
| if (TREE_CODE (x) == CALL_EXPR && (call_expr_flags (x) & ECF_CONST) == 0) |
| return false; |
| |
| if (TREE_CODE (x) == STRING_CST) |
| return false; |
| |
| return true; |
| } |
| |
| /* Return nonzero if X and Y appear to be the same candidate (or NULL) */ |
| static bool |
| candidate_equal_p (const_tree x, const_tree y) |
| { |
| return (x == y) || (x && y && operand_equal_p (x, y, 0)); |
| } |
| |
| /* Walk the tree X, and record accesses to variables. If X is written by the |
| parent tree, WRITER is the parent. |
| We store accesses in one of the two lists: PBEFORE_SP, and PNO_SP. If this |
| expression or its only operand forces a sequence point, then everything up |
| to the sequence point is stored in PBEFORE_SP. Everything else gets stored |
| in PNO_SP. |
| Once we return, we will have emitted warnings if any subexpression before |
| such a sequence point could be undefined. On a higher level, however, the |
| sequence point may not be relevant, and we'll merge the two lists. |
| |
| Example: (b++, a) + b; |
| The call that processes the COMPOUND_EXPR will store the increment of B |
| in PBEFORE_SP, and the use of A in PNO_SP. The higher-level call that |
| processes the PLUS_EXPR will need to merge the two lists so that |
| eventually, all accesses end up on the same list (and we'll warn about the |
| unordered subexpressions b++ and b. |
| |
| A note on merging. If we modify the former example so that our expression |
| becomes |
| (b++, b) + a |
| care must be taken not simply to add all three expressions into the final |
| PNO_SP list. The function merge_tlist takes care of that by merging the |
| before-SP list of the COMPOUND_EXPR into its after-SP list in a special |
| way, so that no more than one access to B is recorded. */ |
| |
| static void |
| verify_tree (tree x, struct tlist **pbefore_sp, struct tlist **pno_sp, |
| tree writer) |
| { |
| struct tlist *tmp_before, *tmp_nosp, *tmp_list2, *tmp_list3; |
| enum tree_code code; |
| enum tree_code_class cl; |
| |
| /* X may be NULL if it is the operand of an empty statement expression |
| ({ }). */ |
| if (x == NULL) |
| return; |
| |
| restart: |
| code = TREE_CODE (x); |
| cl = TREE_CODE_CLASS (code); |
| |
| if (warning_candidate_p (x)) |
| *pno_sp = new_tlist (*pno_sp, x, writer); |
| |
| switch (code) |
| { |
| case CONSTRUCTOR: |
| case SIZEOF_EXPR: |
| return; |
| |
| case COMPOUND_EXPR: |
| case TRUTH_ANDIF_EXPR: |
| case TRUTH_ORIF_EXPR: |
| tmp_before = tmp_nosp = tmp_list2 = tmp_list3 = 0; |
| verify_tree (TREE_OPERAND (x, 0), &tmp_before, &tmp_nosp, NULL_TREE); |
| warn_for_collisions (tmp_nosp); |
| merge_tlist (pbefore_sp, tmp_before, 0); |
| merge_tlist (pbefore_sp, tmp_nosp, 0); |
| verify_tree (TREE_OPERAND (x, 1), &tmp_list3, &tmp_list2, NULL_TREE); |
| warn_for_collisions (tmp_list2); |
| merge_tlist (pbefore_sp, tmp_list3, 0); |
| merge_tlist (pno_sp, tmp_list2, 0); |
| return; |
| |
| case COND_EXPR: |
| tmp_before = tmp_list2 = 0; |
| verify_tree (TREE_OPERAND (x, 0), &tmp_before, &tmp_list2, NULL_TREE); |
| warn_for_collisions (tmp_list2); |
| merge_tlist (pbefore_sp, tmp_before, 0); |
| merge_tlist (pbefore_sp, tmp_list2, 0); |
| |
| tmp_list3 = tmp_nosp = 0; |
| verify_tree (TREE_OPERAND (x, 1), &tmp_list3, &tmp_nosp, NULL_TREE); |
| warn_for_collisions (tmp_nosp); |
| merge_tlist (pbefore_sp, tmp_list3, 0); |
| |
| tmp_list3 = tmp_list2 = 0; |
| verify_tree (TREE_OPERAND (x, 2), &tmp_list3, &tmp_list2, NULL_TREE); |
| warn_for_collisions (tmp_list2); |
| merge_tlist (pbefore_sp, tmp_list3, 0); |
| /* Rather than add both tmp_nosp and tmp_list2, we have to merge the |
| two first, to avoid warning for (a ? b++ : b++). */ |
| merge_tlist (&tmp_nosp, tmp_list2, 0); |
| add_tlist (pno_sp, tmp_nosp, NULL_TREE, 0); |
| return; |
| |
| case PREDECREMENT_EXPR: |
| case PREINCREMENT_EXPR: |
| case POSTDECREMENT_EXPR: |
| case POSTINCREMENT_EXPR: |
| verify_tree (TREE_OPERAND (x, 0), pno_sp, pno_sp, x); |
| return; |
| |
| case MODIFY_EXPR: |
| tmp_before = tmp_nosp = tmp_list3 = 0; |
| verify_tree (TREE_OPERAND (x, 1), &tmp_before, &tmp_nosp, NULL_TREE); |
| verify_tree (TREE_OPERAND (x, 0), &tmp_list3, &tmp_list3, x); |
| /* Expressions inside the LHS are not ordered wrt. the sequence points |
| in the RHS. Example: |
| *a = (a++, 2) |
| Despite the fact that the modification of "a" is in the before_sp |
| list (tmp_before), it conflicts with the use of "a" in the LHS. |
| We can handle this by adding the contents of tmp_list3 |
| to those of tmp_before, and redoing the collision warnings for that |
| list. */ |
| add_tlist (&tmp_before, tmp_list3, x, 1); |
| warn_for_collisions (tmp_before); |
| /* Exclude the LHS itself here; we first have to merge it into the |
| tmp_nosp list. This is done to avoid warning for "a = a"; if we |
| didn't exclude the LHS, we'd get it twice, once as a read and once |
| as a write. */ |
| add_tlist (pno_sp, tmp_list3, x, 0); |
| warn_for_collisions_1 (TREE_OPERAND (x, 0), x, tmp_nosp, 1); |
| |
| merge_tlist (pbefore_sp, tmp_before, 0); |
| if (warning_candidate_p (TREE_OPERAND (x, 0))) |
| merge_tlist (&tmp_nosp, new_tlist (NULL, TREE_OPERAND (x, 0), x), 0); |
| add_tlist (pno_sp, tmp_nosp, NULL_TREE, 1); |
| return; |
| |
| case CALL_EXPR: |
| /* We need to warn about conflicts among arguments and conflicts between |
| args and the function address. Side effects of the function address, |
| however, are not ordered by the sequence point of the call. */ |
| { |
| call_expr_arg_iterator iter; |
| tree arg; |
| tmp_before = tmp_nosp = 0; |
| verify_tree (CALL_EXPR_FN (x), &tmp_before, &tmp_nosp, NULL_TREE); |
| FOR_EACH_CALL_EXPR_ARG (arg, iter, x) |
| { |
| tmp_list2 = tmp_list3 = 0; |
| verify_tree (arg, &tmp_list2, &tmp_list3, NULL_TREE); |
| merge_tlist (&tmp_list3, tmp_list2, 0); |
| add_tlist (&tmp_before, tmp_list3, NULL_TREE, 0); |
| } |
| add_tlist (&tmp_before, tmp_nosp, NULL_TREE, 0); |
| warn_for_collisions (tmp_before); |
| add_tlist (pbefore_sp, tmp_before, NULL_TREE, 0); |
| return; |
| } |
| |
| case TREE_LIST: |
| /* Scan all the list, e.g. indices of multi dimensional array. */ |
| while (x) |
| { |
| tmp_before = tmp_nosp = 0; |
| verify_tree (TREE_VALUE (x), &tmp_before, &tmp_nosp, NULL_TREE); |
| merge_tlist (&tmp_nosp, tmp_before, 0); |
| add_tlist (pno_sp, tmp_nosp, NULL_TREE, 0); |
| x = TREE_CHAIN (x); |
| } |
| return; |
| |
| case SAVE_EXPR: |
| { |
| struct tlist_cache *t; |
| for (t = save_expr_cache; t; t = t->next) |
| if (candidate_equal_p (t->expr, x)) |
| break; |
| |
| if (!t) |
| { |
| t = XOBNEW (&tlist_obstack, struct tlist_cache); |
| t->next = save_expr_cache; |
| t->expr = x; |
| save_expr_cache = t; |
| |
| tmp_before = tmp_nosp = 0; |
| verify_tree (TREE_OPERAND (x, 0), &tmp_before, &tmp_nosp, NULL_TREE); |
| warn_for_collisions (tmp_nosp); |
| |
| tmp_list3 = 0; |
| merge_tlist (&tmp_list3, tmp_nosp, 0); |
| t->cache_before_sp = tmp_before; |
| t->cache_after_sp = tmp_list3; |
| } |
| merge_tlist (pbefore_sp, t->cache_before_sp, 1); |
| add_tlist (pno_sp, t->cache_after_sp, NULL_TREE, 1); |
| return; |
| } |
| |
| case ADDR_EXPR: |
| x = TREE_OPERAND (x, 0); |
| if (DECL_P (x)) |
| return; |
| writer = 0; |
| goto restart; |
| |
| default: |
| /* For other expressions, simply recurse on their operands. |
| Manual tail recursion for unary expressions. |
| Other non-expressions need not be processed. */ |
| if (cl == tcc_unary) |
| { |
| x = TREE_OPERAND (x, 0); |
| writer = 0; |
| goto restart; |
| } |
| else if (IS_EXPR_CODE_CLASS (cl)) |
| { |
| int lp; |
| int max = TREE_OPERAND_LENGTH (x); |
| for (lp = 0; lp < max; lp++) |
| { |
| tmp_before = tmp_nosp = 0; |
| verify_tree (TREE_OPERAND (x, lp), &tmp_before, &tmp_nosp, 0); |
| merge_tlist (&tmp_nosp, tmp_before, 0); |
| add_tlist (pno_sp, tmp_nosp, NULL_TREE, 0); |
| } |
| } |
| return; |
| } |
| } |
| |
| /* Try to warn for undefined behavior in EXPR due to missing sequence |
| points. */ |
| |
| DEBUG_FUNCTION void |
| verify_sequence_points (tree expr) |
| { |
| struct tlist *before_sp = 0, *after_sp = 0; |
| |
| warned_ids = 0; |
| save_expr_cache = 0; |
| if (tlist_firstobj == 0) |
| { |
| gcc_obstack_init (&tlist_obstack); |
| tlist_firstobj = (char *) obstack_alloc (&tlist_obstack, 0); |
| } |
| |
| verify_tree (expr, &before_sp, &after_sp, 0); |
| warn_for_collisions (after_sp); |
| obstack_free (&tlist_obstack, tlist_firstobj); |
| } |
| |
| /* Validate the expression after `case' and apply default promotions. */ |
| |
| static tree |
| check_case_value (location_t loc, tree value) |
| { |
| if (value == NULL_TREE) |
| return value; |
| |
| if (TREE_CODE (value) == INTEGER_CST) |
| /* Promote char or short to int. */ |
| value = perform_integral_promotions (value); |
| else if (value != error_mark_node) |
| { |
| error_at (loc, "case label does not reduce to an integer constant"); |
| value = error_mark_node; |
| } |
| |
| constant_expression_warning (value); |
| |
| return value; |
| } |
| |
| /* See if the case values LOW and HIGH are in the range of the original |
| type (i.e. before the default conversion to int) of the switch testing |
| expression. |
| TYPE is the promoted type of the testing expression, and ORIG_TYPE is |
| the type before promoting it. CASE_LOW_P is a pointer to the lower |
| bound of the case label, and CASE_HIGH_P is the upper bound or NULL |
| if the case is not a case range. |
| The caller has to make sure that we are not called with NULL for |
| CASE_LOW_P (i.e. the default case). OUTSIDE_RANGE_P says whether there |
| was a case value that doesn't fit into the range of the ORIG_TYPE. |
| Returns true if the case label is in range of ORIG_TYPE (saturated or |
| untouched) or false if the label is out of range. */ |
| |
| static bool |
| check_case_bounds (location_t loc, tree type, tree orig_type, |
| tree *case_low_p, tree *case_high_p, |
| bool *outside_range_p) |
| { |
| tree min_value, max_value; |
| tree case_low = *case_low_p; |
| tree case_high = case_high_p ? *case_high_p : case_low; |
| |
| /* If there was a problem with the original type, do nothing. */ |
| if (orig_type == error_mark_node) |
| return true; |
| |
| min_value = TYPE_MIN_VALUE (orig_type); |
| max_value = TYPE_MAX_VALUE (orig_type); |
| |
| /* We'll really need integer constants here. */ |
| case_low = fold (case_low); |
| case_high = fold (case_high); |
| |
| /* Case label is less than minimum for type. */ |
| if (tree_int_cst_compare (case_low, min_value) < 0 |
| && tree_int_cst_compare (case_high, min_value) < 0) |
| { |
| warning_at (loc, 0, "case label value is less than minimum value " |
| "for type"); |
| *outside_range_p = true; |
| return false; |
| } |
| |
| /* Case value is greater than maximum for type. */ |
| if (tree_int_cst_compare (case_low, max_value) > 0 |
| && tree_int_cst_compare (case_high, max_value) > 0) |
| { |
| warning_at (loc, 0, "case label value exceeds maximum value for type"); |
| *outside_range_p = true; |
| return false; |
| } |
| |
| /* Saturate lower case label value to minimum. */ |
| if (tree_int_cst_compare (case_high, min_value) >= 0 |
| && tree_int_cst_compare (case_low, min_value) < 0) |
| { |
| warning_at (loc, 0, "lower value in case label range" |
| " less than minimum value for type"); |
| *outside_range_p = true; |
| case_low = min_value; |
| } |
| |
| /* Saturate upper case label value to maximum. */ |
| if (tree_int_cst_compare (case_low, max_value) <= 0 |
| && tree_int_cst_compare (case_high, max_value) > 0) |
| { |
| warning_at (loc, 0, "upper value in case label range" |
| " exceeds maximum value for type"); |
| *outside_range_p = true; |
| case_high = max_value; |
| } |
| |
| if (*case_low_p != case_low) |
| *case_low_p = convert (type, case_low); |
| if (case_high_p && *case_high_p != case_high) |
| *case_high_p = convert (type, case_high); |
| |
| return true; |
| } |
| |
| /* Return an integer type with BITS bits of precision, |
| that is unsigned if UNSIGNEDP is nonzero, otherwise signed. */ |
| |
| tree |
| c_common_type_for_size (unsigned int bits, int unsignedp) |
| { |
| int i; |
| |
| if (bits == TYPE_PRECISION (integer_type_node)) |
| return unsignedp ? unsigned_type_node : integer_type_node; |
| |
| if (bits == TYPE_PRECISION (signed_char_type_node)) |
| return unsignedp ? unsigned_char_type_node : signed_char_type_node; |
| |
| if (bits == TYPE_PRECISION (short_integer_type_node)) |
| return unsignedp ? short_unsigned_type_node : short_integer_type_node; |
| |
| if (bits == TYPE_PRECISION (long_integer_type_node)) |
| return unsignedp ? long_unsigned_type_node : long_integer_type_node; |
| |
| if (bits == TYPE_PRECISION (long_long_integer_type_node)) |
| return (unsignedp ? long_long_unsigned_type_node |
| : long_long_integer_type_node); |
| |
| for (i = 0; i < NUM_INT_N_ENTS; i ++) |
| if (int_n_enabled_p[i] |
| && bits == int_n_data[i].bitsize) |
| return (unsignedp ? int_n_trees[i].unsigned_type |
| : int_n_trees[i].signed_type); |
| |
| if (bits == TYPE_PRECISION (widest_integer_literal_type_node)) |
| return (unsignedp ? widest_unsigned_literal_type_node |
| : widest_integer_literal_type_node); |
| |
| if (bits <= TYPE_PRECISION (intQI_type_node)) |
| return unsignedp ? unsigned_intQI_type_node : intQI_type_node; |
| |
| if (bits <= TYPE_PRECISION (intHI_type_node)) |
| return unsignedp ? unsigned_intHI_type_node : intHI_type_node; |
| |
| if (bits <= TYPE_PRECISION (intSI_type_node)) |
| return unsignedp ? unsigned_intSI_type_node : intSI_type_node; |
| |
| if (bits <= TYPE_PRECISION (intDI_type_node)) |
| return unsignedp ? unsigned_intDI_type_node : intDI_type_node; |
| |
| return 0; |
| } |
| |
| /* Return a fixed-point type that has at least IBIT ibits and FBIT fbits |
| that is unsigned if UNSIGNEDP is nonzero, otherwise signed; |
| and saturating if SATP is nonzero, otherwise not saturating. */ |
| |
| tree |
| c_common_fixed_point_type_for_size (unsigned int ibit, unsigned int fbit, |
| int unsignedp, int satp) |
| { |
| machine_mode mode; |
| if (ibit == 0) |
| mode = unsignedp ? UQQmode : QQmode; |
| else |
| mode = unsignedp ? UHAmode : HAmode; |
| |
| for (; mode != VOIDmode; mode = GET_MODE_WIDER_MODE (mode)) |
| if (GET_MODE_IBIT (mode) >= ibit && GET_MODE_FBIT (mode) >= fbit) |
| break; |
| |
| if (mode == VOIDmode || !targetm.scalar_mode_supported_p (mode)) |
| { |
| sorry ("GCC cannot support operators with integer types and " |
| "fixed-point types that have too many integral and " |
| "fractional bits together"); |
| return 0; |
| } |
| |
| return c_common_type_for_mode (mode, satp); |
| } |
| |
| /* Used for communication between c_common_type_for_mode and |
| c_register_builtin_type. */ |
| tree registered_builtin_types; |
| |
| /* Return a data type that has machine mode MODE. |
| If the mode is an integer, |
| then UNSIGNEDP selects between signed and unsigned types. |
| If the mode is a fixed-point mode, |
| then UNSIGNEDP selects between saturating and nonsaturating types. */ |
| |
| tree |
| c_common_type_for_mode (machine_mode mode, int unsignedp) |
| { |
| tree t; |
| int i; |
| |
| if (mode == TYPE_MODE (integer_type_node)) |
| return unsignedp ? unsigned_type_node : integer_type_node; |
| |
| if (mode == TYPE_MODE (signed_char_type_node)) |
| return unsignedp ? unsigned_char_type_node : signed_char_type_node; |
| |
| if (mode == TYPE_MODE (short_integer_type_node)) |
| return unsignedp ? short_unsigned_type_node : short_integer_type_node; |
| |
| if (mode == TYPE_MODE (long_integer_type_node)) |
| return unsignedp ? long_unsigned_type_node : long_integer_type_node; |
| |
| if (mode == TYPE_MODE (long_long_integer_type_node)) |
| return unsignedp ? long_long_unsigned_type_node : long_long_integer_type_node; |
| |
| for (i = 0; i < NUM_INT_N_ENTS; i ++) |
| if (int_n_enabled_p[i] |
| && mode == int_n_data[i].m) |
| return (unsignedp ? int_n_trees[i].unsigned_type |
| : int_n_trees[i].signed_type); |
| |
| if (mode == QImode) |
| return unsignedp ? unsigned_intQI_type_node : intQI_type_node; |
| |
| if (mode == HImode) |
| return unsignedp ? unsigned_intHI_type_node : intHI_type_node; |
| |
| if (mode == SImode) |
| return unsignedp ? unsigned_intSI_type_node : intSI_type_node; |
| |
| if (mode == DImode) |
| return unsignedp ? unsigned_intDI_type_node : intDI_type_node; |
| |
| #if HOST_BITS_PER_WIDE_INT >= 64 |
| if (mode == TYPE_MODE (intTI_type_node)) |
| return unsignedp ? unsigned_intTI_type_node : intTI_type_node; |
| #endif |
| |
| if (mode == TYPE_MODE (float_type_node)) |
| return float_type_node; |
| |
| if (mode == TYPE_MODE (double_type_node)) |
| return double_type_node; |
| |
| if (mode == TYPE_MODE (long_double_type_node)) |
| return long_double_type_node; |
| |
| for (i = 0; i < NUM_FLOATN_NX_TYPES; i++) |
| if (FLOATN_NX_TYPE_NODE (i) != NULL_TREE |
| && mode == TYPE_MODE (FLOATN_NX_TYPE_NODE (i))) |
| return FLOATN_NX_TYPE_NODE (i); |
| |
| if (mode == TYPE_MODE (void_type_node)) |
| return void_type_node; |
| |
| if (mode == TYPE_MODE (build_pointer_type (char_type_node))) |
| return (unsignedp |
| ? make_unsigned_type (GET_MODE_PRECISION (mode)) |
| : make_signed_type (GET_MODE_PRECISION (mode))); |
| |
| if (mode == TYPE_MODE (build_pointer_type (integer_type_node))) |
| return (unsignedp |
| ? make_unsigned_type (GET_MODE_PRECISION (mode)) |
| : make_signed_type (GET_MODE_PRECISION (mode))); |
| |
| if (COMPLEX_MODE_P (mode)) |
| { |
| machine_mode inner_mode; |
| tree inner_type; |
| |
| if (mode == TYPE_MODE (complex_float_type_node)) |
| return complex_float_type_node; |
| if (mode == TYPE_MODE (complex_double_type_node)) |
| return complex_double_type_node; |
| if (mode == TYPE_MODE (complex_long_double_type_node)) |
| return complex_long_double_type_node; |
| |
| for (i = 0; i < NUM_FLOATN_NX_TYPES; i++) |
| if (COMPLEX_FLOATN_NX_TYPE_NODE (i) != NULL_TREE |
| && mode == TYPE_MODE (COMPLEX_FLOATN_NX_TYPE_NODE (i))) |
| return COMPLEX_FLOATN_NX_TYPE_NODE (i); |
| |
| if (mode == TYPE_MODE (complex_integer_type_node) && !unsignedp) |
| return complex_integer_type_node; |
| |
| inner_mode = GET_MODE_INNER (mode); |
| inner_type = c_common_type_for_mode (inner_mode, unsignedp); |
| if (inner_type != NULL_TREE) |
| return build_complex_type (inner_type); |
| } |
| else if (VECTOR_MODE_P (mode)) |
| { |
| machine_mode inner_mode = GET_MODE_INNER (mode); |
| tree inner_type = c_common_type_for_mode (inner_mode, unsignedp); |
| if (inner_type != NULL_TREE) |
| return build_vector_type_for_mode (inner_type, mode); |
| } |
| |
| if (mode == TYPE_MODE (dfloat32_type_node)) |
| return dfloat32_type_node; |
| if (mode == TYPE_MODE (dfloat64_type_node)) |
| return dfloat64_type_node; |
| if (mode == TYPE_MODE (dfloat128_type_node)) |
| return dfloat128_type_node; |
| |
| if (ALL_SCALAR_FIXED_POINT_MODE_P (mode)) |
| { |
| if (mode == TYPE_MODE (short_fract_type_node)) |
| return unsignedp ? sat_short_fract_type_node : short_fract_type_node; |
| if (mode == TYPE_MODE (fract_type_node)) |
| return unsignedp ? sat_fract_type_node : fract_type_node; |
| if (mode == TYPE_MODE (long_fract_type_node)) |
| return unsignedp ? sat_long_fract_type_node : long_fract_type_node; |
| if (mode == TYPE_MODE (long_long_fract_type_node)) |
| return unsignedp ? sat_long_long_fract_type_node |
| : long_long_fract_type_node; |
| |
| if (mode == TYPE_MODE (unsigned_short_fract_type_node)) |
| return unsignedp ? sat_unsigned_short_fract_type_node |
| : unsigned_short_fract_type_node; |
| if (mode == TYPE_MODE (unsigned_fract_type_node)) |
| return unsignedp ? sat_unsigned_fract_type_node |
| : unsigned_fract_type_node; |
| if (mode == TYPE_MODE (unsigned_long_fract_type_node)) |
| return unsignedp ? sat_unsigned_long_fract_type_node |
| : unsigned_long_fract_type_node; |
| if (mode == TYPE_MODE (unsigned_long_long_fract_type_node)) |
| return unsignedp ? sat_unsigned_long_long_fract_type_node |
| : unsigned_long_long_fract_type_node; |
| |
| if (mode == TYPE_MODE (short_accum_type_node)) |
| return unsignedp ? sat_short_accum_type_node : short_accum_type_node; |
| if (mode == TYPE_MODE (accum_type_node)) |
| return unsignedp ? sat_accum_type_node : accum_type_node; |
| if (mode == TYPE_MODE (long_accum_type_node)) |
| return unsignedp ? sat_long_accum_type_node : long_accum_type_node; |
| if (mode == TYPE_MODE (long_long_accum_type_node)) |
| return unsignedp ? sat_long_long_accum_type_node |
| : long_long_accum_type_node; |
| |
| if (mode == TYPE_MODE (unsigned_short_accum_type_node)) |
| return unsignedp ? sat_unsigned_short_accum_type_node |
| : unsigned_short_accum_type_node; |
| if (mode == TYPE_MODE (unsigned_accum_type_node)) |
| return unsignedp ? sat_unsigned_accum_type_node |
| : unsigned_accum_type_node; |
| if (mode == TYPE_MODE (unsigned_long_accum_type_node)) |
| return unsignedp ? sat_unsigned_long_accum_type_node |
| : unsigned_long_accum_type_node; |
| if (mode == TYPE_MODE (unsigned_long_long_accum_type_node)) |
| return unsignedp ? sat_unsigned_long_long_accum_type_node |
| : unsigned_long_long_accum_type_node; |
| |
| if (mode == QQmode) |
| return unsignedp ? sat_qq_type_node : qq_type_node; |
| if (mode == HQmode) |
| return unsignedp ? sat_hq_type_node : hq_type_node; |
| if (mode == SQmode) |
| return unsignedp ? sat_sq_type_node : sq_type_node; |
| if (mode == DQmode) |
| return unsignedp ? sat_dq_type_node : dq_type_node; |
| if (mode == TQmode) |
| return unsignedp ? sat_tq_type_node : tq_type_node; |
| |
| if (mode == UQQmode) |
| return unsignedp ? sat_uqq_type_node : uqq_type_node; |
| if (mode == UHQmode) |
| return unsignedp ? sat_uhq_type_node : uhq_type_node; |
| if (mode == USQmode) |
| return unsignedp ? sat_usq_type_node : usq_type_node; |
| if (mode == UDQmode) |
| return unsignedp ? sat_udq_type_node : udq_type_node; |
| if (mode == UTQmode) |
| return unsignedp ? sat_utq_type_node : utq_type_node; |
| |
| if (mode == HAmode) |
| return unsignedp ? sat_ha_type_node : ha_type_node; |
| if (mode == SAmode) |
| return unsignedp ? sat_sa_type_node : sa_type_node; |
| if (mode == DAmode) |
| return unsignedp ? sat_da_type_node : da_type_node; |
| if (mode == TAmode) |
| return unsignedp ? sat_ta_type_node : ta_type_node; |
| |
| if (mode == UHAmode) |
| return unsignedp ? sat_uha_type_node : uha_type_node; |
| if (mode == USAmode) |
| return unsignedp ? sat_usa_type_node : usa_type_node; |
| if (mode == UDAmode) |
| return unsignedp ? sat_uda_type_node : uda_type_node; |
| if (mode == UTAmode) |
| return unsignedp ? sat_uta_type_node : uta_type_node; |
| } |
| |
| for (t = registered_builtin_types; t; t = TREE_CHAIN (t)) |
| if (TYPE_MODE (TREE_VALUE (t)) == mode |
| && !!unsignedp == !!TYPE_UNSIGNED (TREE_VALUE (t))) |
| return TREE_VALUE (t); |
| |
| return 0; |
| } |
| |
| tree |
| c_common_unsigned_type (tree type) |
| { |
| return c_common_signed_or_unsigned_type (1, type); |
| } |
| |
| /* Return a signed type the same as TYPE in other respects. */ |
| |
| tree |
| c_common_signed_type (tree type) |
| { |
| return c_common_signed_or_unsigned_type (0, type); |
| } |
| |
| /* Return a type the same as TYPE except unsigned or |
| signed according to UNSIGNEDP. */ |
| |
| tree |
| c_common_signed_or_unsigned_type (int unsignedp, tree type) |
| { |
| tree type1; |
| int i; |
| |
| /* This block of code emulates the behavior of the old |
| c_common_unsigned_type. In particular, it returns |
| long_unsigned_type_node if passed a long, even when a int would |
| have the same size. This is necessary for warnings to work |
| correctly in archs where sizeof(int) == sizeof(long) */ |
| |
| type1 = TYPE_MAIN_VARIANT (type); |
| if (type1 == signed_char_type_node || type1 == char_type_node || type1 == unsigned_char_type_node) |
| return unsignedp ? unsigned_char_type_node : signed_char_type_node; |
| if (type1 == integer_type_node || type1 == unsigned_type_node) |
| return unsignedp ? unsigned_type_node : integer_type_node; |
| if (type1 == short_integer_type_node || type1 == short_unsigned_type_node) |
| return unsignedp ? short_unsigned_type_node : short_integer_type_node; |
| if (type1 == long_integer_type_node || type1 == long_unsigned_type_node) |
| return unsignedp ? long_unsigned_type_node : long_integer_type_node; |
| if (type1 == long_long_integer_type_node || type1 == long_long_unsigned_type_node) |
| return unsignedp ? long_long_unsigned_type_node : long_long_integer_type_node; |
| |
| for (i = 0; i < NUM_INT_N_ENTS; i ++) |
| if (int_n_enabled_p[i] |
| && (type1 == int_n_trees[i].unsigned_type |
| || type1 == int_n_trees[i].signed_type)) |
| return (unsignedp ? int_n_trees[i].unsigned_type |
| : int_n_trees[i].signed_type); |
| |
| #if HOST_BITS_PER_WIDE_INT >= 64 |
| if (type1 == intTI_type_node || type1 == unsigned_intTI_type_node) |
| return unsignedp ? unsigned_intTI_type_node : intTI_type_node; |
| #endif |
| if (type1 == intDI_type_node || type1 == unsigned_intDI_type_node) |
| return unsignedp ? unsigned_intDI_type_node : intDI_type_node; |
| if (type1 == intSI_type_node || type1 == unsigned_intSI_type_node) |
| return unsignedp ? unsigned_intSI_type_node : intSI_type_node; |
| if (type1 == intHI_type_node || type1 == unsigned_intHI_type_node) |
| return unsignedp ? unsigned_intHI_type_node : intHI_type_node; |
| if (type1 == intQI_type_node || type1 == unsigned_intQI_type_node) |
| return unsignedp ? unsigned_intQI_type_node : intQI_type_node; |
| |
| #define C_COMMON_FIXED_TYPES(NAME) \ |
| if (type1 == short_ ## NAME ## _type_node \ |
| || type1 == unsigned_short_ ## NAME ## _type_node) \ |
| return unsignedp ? unsigned_short_ ## NAME ## _type_node \ |
| : short_ ## NAME ## _type_node; \ |
| if (type1 == NAME ## _type_node \ |
| || type1 == unsigned_ ## NAME ## _type_node) \ |
| return unsignedp ? unsigned_ ## NAME ## _type_node \ |
| : NAME ## _type_node; \ |
| if (type1 == long_ ## NAME ## _type_node \ |
| || type1 == unsigned_long_ ## NAME ## _type_node) \ |
| return unsignedp ? unsigned_long_ ## NAME ## _type_node \ |
| : long_ ## NAME ## _type_node; \ |
| if (type1 == long_long_ ## NAME ## _type_node \ |
| || type1 == unsigned_long_long_ ## NAME ## _type_node) \ |
| return unsignedp ? unsigned_long_long_ ## NAME ## _type_node \ |
| : long_long_ ## NAME ## _type_node; |
| |
| #define C_COMMON_FIXED_MODE_TYPES(NAME) \ |
| if (type1 == NAME ## _type_node \ |
| || type1 == u ## NAME ## _type_node) \ |
| return unsignedp ? u ## NAME ## _type_node \ |
| : NAME ## _type_node; |
| |
| #define C_COMMON_FIXED_TYPES_SAT(NAME) \ |
| if (type1 == sat_ ## short_ ## NAME ## _type_node \ |
| || type1 == sat_ ## unsigned_short_ ## NAME ## _type_node) \ |
| return unsignedp ? sat_ ## unsigned_short_ ## NAME ## _type_node \ |
| : sat_ ## short_ ## NAME ## _type_node; \ |
| if (type1 == sat_ ## NAME ## _type_node \ |
| || type1 == sat_ ## unsigned_ ## NAME ## _type_node) \ |
| return unsignedp ? sat_ ## unsigned_ ## NAME ## _type_node \ |
| : sat_ ## NAME ## _type_node; \ |
| if (type1 == sat_ ## long_ ## NAME ## _type_node \ |
| || type1 == sat_ ## unsigned_long_ ## NAME ## _type_node) \ |
| return unsignedp ? sat_ ## unsigned_long_ ## NAME ## _type_node \ |
| : sat_ ## long_ ## NAME ## _type_node; \ |
| if (type1 == sat_ ## long_long_ ## NAME ## _type_node \ |
| || type1 == sat_ ## unsigned_long_long_ ## NAME ## _type_node) \ |
| return unsignedp ? sat_ ## unsigned_long_long_ ## NAME ## _type_node \ |
| : sat_ ## long_long_ ## NAME ## _type_node; |
| |
| #define C_COMMON_FIXED_MODE_TYPES_SAT(NAME) \ |
| if (type1 == sat_ ## NAME ## _type_node \ |
| || type1 == sat_ ## u ## NAME ## _type_node) \ |
| return unsignedp ? sat_ ## u ## NAME ## _type_node \ |
| : sat_ ## NAME ## _type_node; |
| |
| C_COMMON_FIXED_TYPES (fract); |
| C_COMMON_FIXED_TYPES_SAT (fract); |
| C_COMMON_FIXED_TYPES (accum); |
| C_COMMON_FIXED_TYPES_SAT (accum); |
| |
| C_COMMON_FIXED_MODE_TYPES (qq); |
| C_COMMON_FIXED_MODE_TYPES (hq); |
| C_COMMON_FIXED_MODE_TYPES (sq); |
| C_COMMON_FIXED_MODE_TYPES (dq); |
| C_COMMON_FIXED_MODE_TYPES (tq); |
| C_COMMON_FIXED_MODE_TYPES_SAT (qq); |
| C_COMMON_FIXED_MODE_TYPES_SAT (hq); |
| C_COMMON_FIXED_MODE_TYPES_SAT (sq); |
| C_COMMON_FIXED_MODE_TYPES_SAT (dq); |
| C_COMMON_FIXED_MODE_TYPES_SAT (tq); |
| C_COMMON_FIXED_MODE_TYPES (ha); |
| C_COMMON_FIXED_MODE_TYPES (sa); |
| C_COMMON_FIXED_MODE_TYPES (da); |
| C_COMMON_FIXED_MODE_TYPES (ta); |
| C_COMMON_FIXED_MODE_TYPES_SAT (ha); |
| C_COMMON_FIXED_MODE_TYPES_SAT (sa); |
| C_COMMON_FIXED_MODE_TYPES_SAT (da); |
| C_COMMON_FIXED_MODE_TYPES_SAT (ta); |
| |
| /* For ENUMERAL_TYPEs in C++, must check the mode of the types, not |
| the precision; they have precision set to match their range, but |
| may use a wider mode to match an ABI. If we change modes, we may |
| wind up with bad conversions. For INTEGER_TYPEs in C, must check |
| the precision as well, so as to yield correct results for |
| bit-field types. C++ does not have these separate bit-field |
| types, and producing a signed or unsigned variant of an |
| ENUMERAL_TYPE may cause other problems as well. */ |
| |
| if (!INTEGRAL_TYPE_P (type) |
| || TYPE_UNSIGNED (type) == unsignedp) |
| return type; |
| |
| #define TYPE_OK(node) \ |
| (TYPE_MODE (type) == TYPE_MODE (node) \ |
| && TYPE_PRECISION (type) == TYPE_PRECISION (node)) |
| if (TYPE_OK (signed_char_type_node)) |
| return unsignedp ? unsigned_char_type_node : signed_char_type_node; |
| if (TYPE_OK (integer_type_node)) |
| return unsignedp ? unsigned_type_node : integer_type_node; |
| if (TYPE_OK (short_integer_type_node)) |
| return unsignedp ? short_unsigned_type_node : short_integer_type_node; |
| if (TYPE_OK (long_integer_type_node)) |
| return unsignedp ? long_unsigned_type_node : long_integer_type_node; |
| if (TYPE_OK (long_long_integer_type_node)) |
| return (unsignedp ? long_long_unsigned_type_node |
| : long_long_integer_type_node); |
| |
| for (i = 0; i < NUM_INT_N_ENTS; i ++) |
| if (int_n_enabled_p[i] |
| && TYPE_MODE (type) == int_n_data[i].m |
| && TYPE_PRECISION (type) == int_n_data[i].bitsize) |
| return (unsignedp ? int_n_trees[i].unsigned_type |
| : int_n_trees[i].signed_type); |
| |
| #if HOST_BITS_PER_WIDE_INT >= 64 |
| if (TYPE_OK (intTI_type_node)) |
| return unsignedp ? unsigned_intTI_type_node : intTI_type_node; |
| #endif |
| if (TYPE_OK (intDI_type_node)) |
| return unsignedp ? unsigned_intDI_type_node : intDI_type_node; |
| if (TYPE_OK (intSI_type_node)) |
| return unsignedp ? unsigned_intSI_type_node : intSI_type_node; |
| if (TYPE_OK (intHI_type_node)) |
| return unsignedp ? unsigned_intHI_type_node : intHI_type_node; |
| if (TYPE_OK (intQI_type_node)) |
| return unsignedp ? unsigned_intQI_type_node : intQI_type_node; |
| #undef TYPE_OK |
| |
| return build_nonstandard_integer_type (TYPE_PRECISION (type), unsignedp); |
| } |
| |
| /* Build a bit-field integer type for the given WIDTH and UNSIGNEDP. */ |
| |
| tree |
| c_build_bitfield_integer_type (unsigned HOST_WIDE_INT width, int unsignedp) |
| { |
| int i; |
| |
| /* Extended integer types of the same width as a standard type have |
| lesser rank, so those of the same width as int promote to int or |
| unsigned int and are valid for printf formats expecting int or |
| unsigned int. To avoid such special cases, avoid creating |
| extended integer types for bit-fields if a standard integer type |
| is available. */ |
| if (width == TYPE_PRECISION (integer_type_node)) |
| return unsignedp ? unsigned_type_node : integer_type_node; |
| if (width == TYPE_PRECISION (signed_char_type_node)) |
| return unsignedp ? unsigned_char_type_node : signed_char_type_node; |
| if (width == TYPE_PRECISION (short_integer_type_node)) |
| return unsignedp ? short_unsigned_type_node : short_integer_type_node; |
| if (width == TYPE_PRECISION (long_integer_type_node)) |
| return unsignedp ? long_unsigned_type_node : long_integer_type_node; |
| if (width == TYPE_PRECISION (long_long_integer_type_node)) |
| return (unsignedp ? long_long_unsigned_type_node |
| : long_long_integer_type_node); |
| for (i = 0; i < NUM_INT_N_ENTS; i ++) |
| if (int_n_enabled_p[i] |
| && width == int_n_data[i].bitsize) |
| return (unsignedp ? int_n_trees[i].unsigned_type |
| : int_n_trees[i].signed_type); |
| return build_nonstandard_integer_type (width, unsignedp); |
| } |
| |
| /* The C version of the register_builtin_type langhook. */ |
| |
| void |
| c_register_builtin_type (tree type, const char* name) |
| { |
| tree decl; |
| |
| decl = build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, get_identifier (name), type); |
| DECL_ARTIFICIAL (decl) = 1; |
| if (!TYPE_NAME (type)) |
| TYPE_NAME (type) = decl; |
| pushdecl (decl); |
| |
| registered_builtin_types = tree_cons (0, type, registered_builtin_types); |
| } |
| |
| /* Print an error message for invalid operands to arith operation |
| CODE with TYPE0 for operand 0, and TYPE1 for operand 1. |
| RICHLOC is a rich location for the message, containing either |
| three separate locations for each of the operator and operands |
| |
| lhs op rhs |
| ~~~ ^~ ~~~ |
| |
| (C FE), or one location ranging over all over them |
| |
| lhs op rhs |
| ~~~~^~~~~~ |
| |
| (C++ FE). */ |
| |
| void |
| binary_op_error (rich_location *richloc, enum tree_code code, |
| tree type0, tree type1) |
| { |
| const char *opname; |
| |
| switch (code) |
| { |
| case PLUS_EXPR: |
| opname = "+"; break; |
| case MINUS_EXPR: |
| opname = "-"; break; |
| case MULT_EXPR: |
| opname = "*"; break; |
| case MAX_EXPR: |
| opname = "max"; break; |
| case MIN_EXPR: |
| opname = "min"; break; |
| case EQ_EXPR: |
| opname = "=="; break; |
| case NE_EXPR: |
| opname = "!="; break; |
| case LE_EXPR: |
| opname = "<="; break; |
| case GE_EXPR: |
| opname = ">="; break; |
| case LT_EXPR: |
| opname = "<"; break; |
| case GT_EXPR: |
| opname = ">"; break; |
| case LSHIFT_EXPR: |
| opname = "<<"; break; |
| case RSHIFT_EXPR: |
| opname = ">>"; break; |
| case TRUNC_MOD_EXPR: |
| case FLOOR_MOD_EXPR: |
| opname = "%"; break; |
| case TRUNC_DIV_EXPR: |
| case FLOOR_DIV_EXPR: |
| opname = "/"; break; |
| case BIT_AND_EXPR: |
| opname = "&"; break; |
| case BIT_IOR_EXPR: |
| opname = "|"; break; |
| case TRUTH_ANDIF_EXPR: |
| opname = "&&"; break; |
| case TRUTH_ORIF_EXPR: |
| opname = "||"; break; |
| case BIT_XOR_EXPR: |
| opname = "^"; break; |
| default: |
| gcc_unreachable (); |
| } |
| error_at_rich_loc (richloc, |
| "invalid operands to binary %s (have %qT and %qT)", |
| opname, type0, type1); |
| } |
| |
| /* Given an expression as a tree, return its original type. Do this |
| by stripping any conversion that preserves the sign and precision. */ |
| static tree |
| expr_original_type (tree expr) |
| { |
| STRIP_SIGN_NOPS (expr); |
| return TREE_TYPE (expr); |
| } |
| |
| /* Subroutine of build_binary_op, used for comparison operations. |
| See if the operands have both been converted from subword integer types |
| and, if so, perhaps change them both back to their original type. |
| This function is also responsible for converting the two operands |
| to the proper common type for comparison. |
| |
| The arguments of this function are all pointers to local variables |
| of build_binary_op: OP0_PTR is &OP0, OP1_PTR is &OP1, |
| RESTYPE_PTR is &RESULT_TYPE and RESCODE_PTR is &RESULTCODE. |
| |
| LOC is the location of the comparison. |
| |
| If this function returns nonzero, it means that the comparison has |
| a constant value. What this function returns is an expression for |
| that value. */ |
| |
| tree |
| shorten_compare (location_t loc, tree *op0_ptr, tree *op1_ptr, |
| tree *restype_ptr, enum tree_code *rescode_ptr) |
| { |
| tree type; |
| tree op0 = *op0_ptr; |
| tree op1 = *op1_ptr; |
| int unsignedp0, unsignedp1; |
| int real1, real2; |
| tree primop0, primop1; |
| enum tree_code code = *rescode_ptr; |
| |
| /* Throw away any conversions to wider types |
| already present in the operands. */ |
| |
| primop0 = c_common_get_narrower (op0, &unsignedp0); |
| primop1 = c_common_get_narrower (op1, &unsignedp1); |
| |
| /* If primopN is first sign-extended from primopN's precision to opN's |
| precision, then zero-extended from opN's precision to |
| *restype_ptr precision, shortenings might be invalid. */ |
| if (TYPE_PRECISION (TREE_TYPE (primop0)) < TYPE_PRECISION (TREE_TYPE (op0)) |
| && TYPE_PRECISION (TREE_TYPE (op0)) < TYPE_PRECISION (*restype_ptr) |
| && !unsignedp0 |
| && TYPE_UNSIGNED (TREE_TYPE (op0))) |
| primop0 = op0; |
| if (TYPE_PRECISION (TREE_TYPE (primop1)) < TYPE_PRECISION (TREE_TYPE (op1)) |
| && TYPE_PRECISION (TREE_TYPE (op1)) < TYPE_PRECISION (*restype_ptr) |
| && !unsignedp1 |
| && TYPE_UNSIGNED (TREE_TYPE (op1))) |
| primop1 = op1; |
| |
| /* Handle the case that OP0 does not *contain* a conversion |
| but it *requires* conversion to FINAL_TYPE. */ |
| |
| if (op0 == primop0 && TREE_TYPE (op0) != *restype_ptr) |
| unsignedp0 = TYPE_UNSIGNED (TREE_TYPE (op0)); |
| if (op1 == primop1 && TREE_TYPE (op1) != *restype_ptr) |
| unsignedp1 = TYPE_UNSIGNED (TREE_TYPE (op1)); |
| |
| /* If one of the operands must be floated, we cannot optimize. */ |
| real1 = TREE_CODE (TREE_TYPE (primop0)) == REAL_TYPE; |
| real2 = TREE_CODE (TREE_TYPE (primop1)) == REAL_TYPE; |
| |
| /* If first arg is constant, swap the args (changing operation |
| so value is preserved), for canonicalization. Don't do this if |
| the second arg is 0. */ |
| |
| if (TREE_CONSTANT (primop0) |
| && !integer_zerop (primop1) && !real_zerop (primop1) |
| && !fixed_zerop (primop1)) |
| { |
| std::swap (primop0, primop1); |
| std::swap (op0, op1); |
| *op0_ptr = op0; |
| *op1_ptr = op1; |
| std::swap (unsignedp0, unsignedp1); |
| std::swap (real1, real2); |
| |
| switch (code) |
| { |
| case LT_EXPR: |
| code = GT_EXPR; |
| break; |
| case GT_EXPR: |
| code = LT_EXPR; |
| break; |
| case LE_EXPR: |
| code = GE_EXPR; |
| break; |
| case GE_EXPR: |
| code = LE_EXPR; |
| break; |
| default: |
| break; |
| } |
| *rescode_ptr = code; |
| } |
| |
| /* If comparing an integer against a constant more bits wide, |
| maybe we can deduce a value of 1 or 0 independent of the data. |
| Or else truncate the constant now |
| rather than extend the variable at run time. |
| |
| This is only interesting if the constant is the wider arg. |
| Also, it is not safe if the constant is unsigned and the |
| variable arg is signed, since in this case the variable |
| would be sign-extended and then regarded as unsigned. |
| Our technique fails in this case because the lowest/highest |
| possible unsigned results don't follow naturally from the |
| lowest/highest possible values of the variable operand. |
| For just EQ_EXPR and NE_EXPR there is another technique that |
| could be used: see if the constant can be faithfully represented |
| in the other operand's type, by truncating it and reextending it |
| and see if that preserves the constant's value. */ |
| |
| if (!real1 && !real2 |
| && TREE_CODE (TREE_TYPE (primop0)) != FIXED_POINT_TYPE |
| && TREE_CODE (primop1) == INTEGER_CST |
| && TYPE_PRECISION (TREE_TYPE (primop0)) < TYPE_PRECISION (*restype_ptr)) |
| { |
| int min_gt, max_gt, min_lt, max_lt; |
| tree maxval, minval; |
| /* 1 if comparison is nominally unsigned. */ |
| int unsignedp = TYPE_UNSIGNED (*restype_ptr); |
| tree val; |
| |
| type = c_common_signed_or_unsigned_type (unsignedp0, |
| TREE_TYPE (primop0)); |
| |
| maxval = TYPE_MAX_VALUE (type); |
| minval = TYPE_MIN_VALUE (type); |
| |
| if (unsignedp && !unsignedp0) |
| *restype_ptr = c_common_signed_type (*restype_ptr); |
| |
| if (TREE_TYPE (primop1) != *restype_ptr) |
| { |
| /* Convert primop1 to target type, but do not introduce |
| additional overflow. We know primop1 is an int_cst. */ |
| primop1 = force_fit_type (*restype_ptr, |
| wi::to_wide |
| (primop1, |
| TYPE_PRECISION (*restype_ptr)), |
| 0, TREE_OVERFLOW (primop1)); |
| } |
| if (type != *restype_ptr) |
| { |
| minval = convert (*restype_ptr, minval); |
| maxval = convert (*restype_ptr, maxval); |
| } |
| |
| min_gt = tree_int_cst_lt (primop1, minval); |
| max_gt = tree_int_cst_lt (primop1, maxval); |
| min_lt = tree_int_cst_lt (minval, primop1); |
| max_lt = tree_int_cst_lt (maxval, primop1); |
| |
| val = 0; |
| /* This used to be a switch, but Genix compiler can't handle that. */ |
| if (code == NE_EXPR) |
| { |
| if (max_lt || min_gt) |
| val = truthvalue_true_node; |
| } |
| else if (code == EQ_EXPR) |
| { |
| if (max_lt || min_gt) |
| val = truthvalue_false_node; |
| } |
| else if (code == LT_EXPR) |
| { |
| if (max_lt) |
| val = truthvalue_true_node; |
| if (!min_lt) |
| val = truthvalue_false_node; |
| } |
| else if (code == GT_EXPR) |
| { |
| if (min_gt) |
| val = truthvalue_true_node; |
| if (!max_gt) |
| val = truthvalue_false_node; |
| } |
| else if (code == LE_EXPR) |
| { |
| if (!max_gt) |
| val = truthvalue_true_node; |
| if (min_gt) |
| val = truthvalue_false_node; |
| } |
| else if (code == GE_EXPR) |
| { |
| if (!min_lt) |
| val = truthvalue_true_node; |
| if (max_lt) |
| val = truthvalue_false_node; |
| } |
| |
| /* If primop0 was sign-extended and unsigned comparison specd, |
| we did a signed comparison above using the signed type bounds. |
| But the comparison we output must be unsigned. |
| |
| Also, for inequalities, VAL is no good; but if the signed |
| comparison had *any* fixed result, it follows that the |
| unsigned comparison just tests the sign in reverse |
| (positive values are LE, negative ones GE). |
| So we can generate an unsigned comparison |
| against an extreme value of the signed type. */ |
| |
| if (unsignedp && !unsignedp0) |
| { |
| if (val != 0) |
| switch (code) |
| { |
| case LT_EXPR: |
| case GE_EXPR: |
| primop1 = TYPE_MIN_VALUE (type); |
| val = 0; |
| break; |
| |
| case LE_EXPR: |
| case GT_EXPR: |
| primop1 = TYPE_MAX_VALUE (type); |
| val = 0; |
| break; |
| |
| default: |
| break; |
| } |
| type = c_common_unsigned_type (type); |
| } |
| |
| if (TREE_CODE (primop0) != INTEGER_CST |
| /* Don't warn if it's from a (non-system) macro. */ |
| && !(from_macro_expansion_at |
| (expansion_point_location_if_in_system_header |
| (EXPR_LOCATION (primop0))))) |
| { |
| if (val == truthvalue_false_node) |
| warning_at (loc, OPT_Wtype_limits, |
| "comparison is always false due to limited range of data type"); |
| if (val == truthvalue_true_node) |
| warning_at (loc, OPT_Wtype_limits, |
| "comparison is always true due to limited range of data type"); |
| } |
| |
| if (val != 0) |
| { |
| /* Don't forget to evaluate PRIMOP0 if it has side effects. */ |
| if (TREE_SIDE_EFFECTS (primop0)) |
| return build2 (COMPOUND_EXPR, TREE_TYPE (val), primop0, val); |
| return val; |
| } |
| |
| /* Value is not predetermined, but do the comparison |
| in the type of the operand that is not constant. |
| TYPE is already properly set. */ |
| } |
| |
| /* If either arg is decimal float and the other is float, find the |
| proper common type to use for comparison. */ |
| else if (real1 && real2 |
| && DECIMAL_FLOAT_MODE_P (TYPE_MODE (TREE_TYPE (primop0))) |
| && DECIMAL_FLOAT_MODE_P (TYPE_MODE (TREE_TYPE (primop1)))) |
| type = common_type (TREE_TYPE (primop0), TREE_TYPE (primop1)); |
| |
| /* If either arg is decimal float and the other is float, fail. */ |
| else if (real1 && real2 |
| && (DECIMAL_FLOAT_MODE_P (TYPE_MODE (TREE_TYPE (primop0))) |
| || DECIMAL_FLOAT_MODE_P (TYPE_MODE (TREE_TYPE (primop1))))) |
| return 0; |
| |
| else if (real1 && real2 |
| && (TYPE_PRECISION (TREE_TYPE (primop0)) |
| == TYPE_PRECISION (TREE_TYPE (primop1)))) |
| type = TREE_TYPE (primop0); |
| |
| /* If args' natural types are both narrower than nominal type |
| and both extend in the same manner, compare them |
| in the type of the wider arg. |
| Otherwise must actually extend both to the nominal |
| common type lest different ways of extending |
| alter the result. |
| (eg, (short)-1 == (unsigned short)-1 should be 0.) */ |
| |
| else if (unsignedp0 == unsignedp1 && real1 == real2 |
| && TYPE_PRECISION (TREE_TYPE (primop0)) < TYPE_PRECISION (*restype_ptr) |
| && TYPE_PRECISION (TREE_TYPE (primop1)) < TYPE_PRECISION (*restype_ptr)) |
| { |
| type = common_type (TREE_TYPE (primop0), TREE_TYPE (primop1)); |
| type = c_common_signed_or_unsigned_type (unsignedp0 |
| || TYPE_UNSIGNED (*restype_ptr), |
| type); |
| /* Make sure shorter operand is extended the right way |
| to match the longer operand. */ |
| primop0 |
| = convert (c_common_signed_or_unsigned_type (unsignedp0, |
| TREE_TYPE (primop0)), |
| primop0); |
| primop1 |
| = convert (c_common_signed_or_unsigned_type (unsignedp1, |
| TREE_TYPE (primop1)), |
| primop1); |
| } |
| else |
| { |
| /* Here we must do the comparison on the nominal type |
| using the args exactly as we received them. */ |
| type = *restype_ptr; |
| primop0 = op0; |
| primop1 = op1; |
| |
| if (!real1 && !real2 && integer_zerop (primop1) |
| && TYPE_UNSIGNED (*restype_ptr)) |
| { |
| tree value = 0; |
| /* All unsigned values are >= 0, so we warn. However, |
| if OP0 is a constant that is >= 0, the signedness of |
| the comparison isn't an issue, so suppress the |
| warning. */ |
| bool warn = |
| warn_type_limits && !in_system_header_at (loc) |
| && !(TREE_CODE (primop0) == INTEGER_CST |
| && !TREE_OVERFLOW (convert (c_common_signed_type (type), |
| primop0))) |
| /* Do not warn for enumeration types. */ |
| && (TREE_CODE (expr_original_type (primop0)) != ENUMERAL_TYPE); |
| |
| switch (code) |
| { |
| case GE_EXPR: |
| if (warn) |
| warning_at (loc, OPT_Wtype_limits, |
| "comparison of unsigned expression >= 0 is always true"); |
| value = truthvalue_true_node; |
| break; |
| |
| case LT_EXPR: |
| if (warn) |
| warning_at (loc, OPT_Wtype_limits, |
| "comparison of unsigned expression < 0 is always false"); |
| value = truthvalue_false_node; |
| break; |
| |
| default: |
| break; |
| } |
| |
| if (value != 0) |
| { |
| /* Don't forget to evaluate PRIMOP0 if it has side effects. */ |
| if (TREE_SIDE_EFFECTS (primop0)) |
| return build2 (COMPOUND_EXPR, TREE_TYPE (value), |
| primop0, value); |
| return value; |
| } |
| } |
| } |
| |
| *op0_ptr = convert (type, primop0); |
| *op1_ptr = convert (type, primop1); |
| |
| *restype_ptr = truthvalue_type_node; |
| |
| return 0; |
| } |
| |
| /* Return a tree for the sum or difference (RESULTCODE says which) |
| of pointer PTROP and integer INTOP. */ |
| |
| tree |
| pointer_int_sum (location_t loc, enum tree_code resultcode, |
| tree ptrop, tree intop, bool complain) |
| { |
| tree size_exp, ret; |
| |
| /* The result is a pointer of the same type that is being added. */ |
| tree result_type = TREE_TYPE (ptrop); |
| |
| if (TREE_CODE (TREE_TYPE (result_type)) == VOID_TYPE) |
| { |
| if (complain && warn_pointer_arith) |
| pedwarn (loc, OPT_Wpointer_arith, |
| "pointer of type %<void *%> used in arithmetic"); |
| else if (!complain) |
| return error_mark_node; |
| size_exp = integer_one_node; |
| } |
| else if (TREE_CODE (TREE_TYPE (result_type)) == FUNCTION_TYPE) |
| { |
| if (complain && warn_pointer_arith) |
| pedwarn (loc, OPT_Wpointer_arith, |
| "pointer to a function used in arithmetic"); |
| else if (!complain) |
| return error_mark_node; |
| size_exp = integer_one_node; |
| } |
| else |
| size_exp = size_in_bytes_loc (loc, TREE_TYPE (result_type)); |
| |
| /* We are manipulating pointer values, so we don't need to warn |
| about relying on undefined signed overflow. We disable the |
| warning here because we use integer types so fold won't know that |
| they are really pointers. */ |
| fold_defer_overflow_warnings (); |
| |
| /* If what we are about to multiply by the size of the elements |
| contains a constant term, apply distributive law |
| and multiply that constant term separately. |
| This helps produce common subexpressions. */ |
| if ((TREE_CODE (intop) == PLUS_EXPR || TREE_CODE (intop) == MINUS_EXPR) |
| && !TREE_CONSTANT (intop) |
| && TREE_CONSTANT (TREE_OPERAND (intop, 1)) |
| && TREE_CONSTANT (size_exp) |
| /* If the constant comes from pointer subtraction, |
| skip this optimization--it would cause an error. */ |
| && TREE_CODE (TREE_TYPE (TREE_OPERAND (intop, 0))) == INTEGER_TYPE |
| /* If the constant is unsigned, and smaller than the pointer size, |
| then we must skip this optimization. This is because it could cause |
| an overflow error if the constant is negative but INTOP is not. */ |
| && (!TYPE_UNSIGNED (TREE_TYPE (intop)) |
| || (TYPE_PRECISION (TREE_TYPE (intop)) |
| == TYPE_PRECISION (TREE_TYPE (ptrop))))) |
| { |
| enum tree_code subcode = resultcode; |
| tree int_type = TREE_TYPE (intop); |
| if (TREE_CODE (intop) == MINUS_EXPR) |
| subcode = (subcode == PLUS_EXPR ? MINUS_EXPR : PLUS_EXPR); |
| /* Convert both subexpression types to the type of intop, |
| because weird cases involving pointer arithmetic |
| can result in a sum or difference with different type args. */ |
| ptrop = build_binary_op (EXPR_LOCATION (TREE_OPERAND (intop, 1)), |
| subcode, ptrop, |
| convert (int_type, TREE_OPERAND (intop, 1)), 1); |
| intop = convert (int_type, TREE_OPERAND (intop, 0)); |
| } |
| |
| /* Convert the integer argument to a type the same size as sizetype |
| so the multiply won't overflow spuriously. */ |
| if (TYPE_PRECISION (TREE_TYPE (intop)) != TYPE_PRECISION (sizetype) |
| || TYPE_UNSIGNED (TREE_TYPE (intop)) != TYPE_UNSIGNED (sizetype)) |
| intop = convert (c_common_type_for_size (TYPE_PRECISION (sizetype), |
| TYPE_UNSIGNED (sizetype)), intop); |
| |
| /* Replace the integer argument with a suitable product by the object size. |
| Do this multiplication as signed, then convert to the appropriate type |
| for the pointer operation and disregard an overflow that occurred only |
| because of the sign-extension change in the latter conversion. */ |
| { |
| tree t = fold_build2_loc (loc, MULT_EXPR, TREE_TYPE (intop), intop, |
| convert (TREE_TYPE (intop), size_exp)); |
| intop = convert (sizetype, t); |
| if (TREE_OVERFLOW_P (intop) && !TREE_OVERFLOW (t)) |
| intop = wide_int_to_tree (TREE_TYPE (intop), intop); |
| } |
| |
| /* Create the sum or difference. */ |
| if (resultcode == MINUS_EXPR) |
| intop = fold_build1_loc (loc, NEGATE_EXPR, sizetype, intop); |
| |
| ret = fold_build_pointer_plus_loc (loc, ptrop, intop); |
| |
| fold_undefer_and_ignore_overflow_warnings (); |
| |
| return ret; |
| } |
| |
| /* Wrap a C_MAYBE_CONST_EXPR around an expression that is fully folded |
| and if NON_CONST is known not to be permitted in an evaluated part |
| of a constant expression. */ |
| |
| tree |
| c_wrap_maybe_const (tree expr, bool non_const) |
| { |
| bool nowarning = TREE_NO_WARNING (expr); |
| location_t loc = EXPR_LOCATION (expr); |
| |
| /* This should never be called for C++. */ |
| if (c_dialect_cxx ()) |
| gcc_unreachable (); |
| |
| /* The result of folding may have a NOP_EXPR to set TREE_NO_WARNING. */ |
| STRIP_TYPE_NOPS (expr); |
| expr = build2 (C_MAYBE_CONST_EXPR, TREE_TYPE (expr), NULL, expr); |
| C_MAYBE_CONST_EXPR_NON_CONST (expr) = non_const; |
| if (nowarning) |
| TREE_NO_WARNING (expr) = 1; |
| protected_set_expr_location (expr, loc); |
| |
| return expr; |
| } |
| |
| /* Wrap a SAVE_EXPR around EXPR, if appropriate. Like save_expr, but |
| for C folds the inside expression and wraps a C_MAYBE_CONST_EXPR |
| around the SAVE_EXPR if needed so that c_fully_fold does not need |
| to look inside SAVE_EXPRs. */ |
| |
| tree |
| c_save_expr (tree expr) |
| { |
| bool maybe_const = true; |
| if (c_dialect_cxx ()) |
| return save_expr (expr); |
| expr = c_fully_fold (expr, false, &maybe_const); |
| expr = save_expr (expr); |
| if (!maybe_const) |
| expr = c_wrap_maybe_const (expr, true); |
| return expr; |
| } |
| |
| /* Return whether EXPR is a declaration whose address can never be |
| NULL. */ |
| |
| bool |
| decl_with_nonnull_addr_p (const_tree expr) |
| { |
| return (DECL_P (expr) |
| && (TREE_CODE (expr) == PARM_DECL |
| || TREE_CODE (expr) == LABEL_DECL |
| || !DECL_WEAK (expr))); |
| } |
| |
| /* Prepare expr to be an argument of a TRUTH_NOT_EXPR, |
| or for an `if' or `while' statement or ?..: exp. It should already |
| have been validated to be of suitable type; otherwise, a bad |
| diagnostic may result. |
| |
| The EXPR is located at LOCATION. |
| |
| This preparation consists of taking the ordinary |
| representation of an expression expr and producing a valid tree |
| boolean expression describing whether expr is nonzero. We could |
| simply always do build_binary_op (NE_EXPR, expr, truthvalue_false_node, 1), |
| but we optimize comparisons, &&, ||, and !. |
| |
| The resulting type should always be `truthvalue_type_node'. */ |
| |
| tree |
| c_common_truthvalue_conversion (location_t location, tree expr) |
| { |
| switch (TREE_CODE (expr)) |
| { |
| case EQ_EXPR: case NE_EXPR: case UNEQ_EXPR: case LTGT_EXPR: |
| case LE_EXPR: case GE_EXPR: case LT_EXPR: case GT_EXPR: |
| case UNLE_EXPR: case UNGE_EXPR: case UNLT_EXPR: case UNGT_EXPR: |
| case ORDERED_EXPR: case UNORDERED_EXPR: |
| if (TREE_TYPE (expr) == truthvalue_type_node) |
| return expr; |
| expr = build2 (TREE_CODE (expr), truthvalue_type_node, |
| TREE_OPERAND (expr, 0), TREE_OPERAND (expr, 1)); |
| goto ret; |
| |
| case TRUTH_ANDIF_EXPR: |
| case TRUTH_ORIF_EXPR: |
| case TRUTH_AND_EXPR: |
| case TRUTH_OR_EXPR: |
| case TRUTH_XOR_EXPR: |
| if (TREE_TYPE (expr) == truthvalue_type_node) |
| return expr; |
| expr = build2 (TREE_CODE (expr), truthvalue_type_node, |
| c_common_truthvalue_conversion (location, |
| TREE_OPERAND (expr, 0)), |
| c_common_truthvalue_conversion (location, |
| TREE_OPERAND (expr, 1))); |
| goto ret; |
| |
| case TRUTH_NOT_EXPR: |
| if (TREE_TYPE (expr) == truthvalue_type_node) |
| return expr; |
| expr = build1 (TREE_CODE (expr), truthvalue_type_node, |
| c_common_truthvalue_conversion (location, |
| TREE_OPERAND (expr, 0))); |
| goto ret; |
| |
| case ERROR_MARK: |
| return expr; |
| |
| case INTEGER_CST: |
| if (TREE_CODE (TREE_TYPE (expr)) == ENUMERAL_TYPE |
| && !integer_zerop (expr) |
| && !integer_onep (expr)) |
| warning_at (location, OPT_Wint_in_bool_context, |
| "enum constant in boolean context"); |
| return integer_zerop (expr) ? truthvalue_false_node |
| : truthvalue_true_node; |
| |
| case REAL_CST: |
| return real_compare (NE_EXPR, &TREE_REAL_CST (expr), &dconst0) |
| ? truthvalue_true_node |
| : truthvalue_false_node; |
| |
| case FIXED_CST: |
| return fixed_compare (NE_EXPR, &TREE_FIXED_CST (expr), |
| &FCONST0 (TYPE_MODE (TREE_TYPE (expr)))) |
| ? truthvalue_true_node |
| : truthvalue_false_node; |
| |
| case FUNCTION_DECL: |
| expr = build_unary_op (location, ADDR_EXPR, expr, false); |
| /* Fall through. */ |
| |
| case ADDR_EXPR: |
| { |
| tree inner = TREE_OPERAND (expr, 0); |
| if (decl_with_nonnull_addr_p (inner)) |
| { |
| /* Common Ada/Pascal programmer's mistake. */ |
| warning_at (location, |
| OPT_Waddress, |
| "the address of %qD will always evaluate as %<true%>", |
| inner); |
| return truthvalue_true_node; |
| } |
| break; |
| } |
| |
| case COMPLEX_EXPR: |
| expr = build_binary_op (EXPR_LOCATION (expr), |
| (TREE_SIDE_EFFECTS (TREE_OPERAND (expr, 1)) |
| ? TRUTH_OR_EXPR : TRUTH_ORIF_EXPR), |
| c_common_truthvalue_conversion (location, |
| TREE_OPERAND (expr, 0)), |
| c_common_truthvalue_conversion (location, |
| TREE_OPERAND (expr, 1)), |
| 0); |
| goto ret; |
| |
| case NEGATE_EXPR: |
| case ABS_EXPR: |
| case FLOAT_EXPR: |
| case EXCESS_PRECISION_EXPR: |
| /* These don't change whether an object is nonzero or zero. */ |
| return c_common_truthvalue_conversion (location, TREE_OPERAND (expr, 0)); |
| |
| case LROTATE_EXPR: |
| case RROTATE_EXPR: |
| /* These don't change whether an object is zero or nonzero, but |
| we can't ignore them if their second arg has side-effects. */ |
| if (TREE_SIDE_EFFECTS (TREE_OPERAND (expr, 1))) |
| { |
| expr = build2 (COMPOUND_EXPR, truthvalue_type_node, |
| TREE_OPERAND (expr, 1), |
| c_common_truthvalue_conversion |
| (location, TREE_OPERAND (expr, 0))); |
| goto ret; |
| } |
| else |
| return c_common_truthvalue_conversion (location, |
| TREE_OPERAND (expr, 0)); |
| |
| case MULT_EXPR: |
| warning_at (EXPR_LOCATION (expr), OPT_Wint_in_bool_context, |
| "%<*%> in boolean context, suggest %<&&%> instead"); |
| break; |
| |
| case LSHIFT_EXPR: |
| /* We will only warn on signed shifts here, because the majority of |
| false positive warnings happen in code where unsigned arithmetic |
| was used in anticipation of a possible overflow. |
| Furthermore, if we see an unsigned type here we know that the |
| result of the shift is not subject to integer promotion rules. */ |
| if (TREE_CODE (TREE_TYPE (expr)) == INTEGER_TYPE |
| && !TYPE_UNSIGNED (TREE_TYPE (expr))) |
| warning_at (EXPR_LOCATION (expr), OPT_Wint_in_bool_context, |
| "%<<<%> in boolean context, did you mean %<<%> ?"); |
| break; |
| |
| case COND_EXPR: |
| if (warn_int_in_bool_context |
| && !from_macro_definition_at (EXPR_LOCATION (expr))) |
| { |
| tree val1 = fold_for_warn (TREE_OPERAND (expr, 1)); |
| tree val2 = fold_for_warn (TREE_OPERAND (expr, 2)); |
| if (TREE_CODE (val1) == INTEGER_CST |
| && TREE_CODE (val2) == INTEGER_CST |
| && !integer_zerop (val1) |
| && !integer_zerop (val2) |
| && (!integer_onep (val1) |
| || !integer_onep (val2))) |
| warning_at (EXPR_LOCATION (expr), OPT_Wint_in_bool_context, |
| "?: using integer constants in boolean context, " |
| "the expression will always evaluate to %<true%>"); |
| else if ((TREE_CODE (val1) == INTEGER_CST |
| && !integer_zerop (val1) |
| && !integer_onep (val1)) |
| || (TREE_CODE (val2) == INTEGER_CST |
| && !integer_zerop (val2) |
| && !integer_onep (val2))) |
| warning_at (EXPR_LOCATION (expr), OPT_Wint_in_bool_context, |
| "?: using integer constants in boolean context"); |
| } |
| /* Distribute the conversion into the arms of a COND_EXPR. */ |
| if (c_dialect_cxx ()) |
| /* Avoid premature folding. */ |
| break; |
| else |
| { |
| int w = warn_int_in_bool_context; |
| warn_int_in_bool_context = 0; |
| /* Folding will happen later for C. */ |
| expr = build3 (COND_EXPR, truthvalue_type_node, |
| TREE_OPERAND (expr, 0), |
| c_common_truthvalue_conversion (location, |
| TREE_OPERAND (expr, 1)), |
| c_common_truthvalue_conversion (location, |
| TREE_OPERAND (expr, 2))); |
| warn_int_in_bool_context = w; |
| goto ret; |
| } |
| |
| CASE_CONVERT: |
| { |
| tree totype = TREE_TYPE (expr); |
| tree fromtype = TREE_TYPE (TREE_OPERAND (expr, 0)); |
| |
| if (POINTER_TYPE_P (totype) |
| && !c_inhibit_evaluation_warnings |
| && TREE_CODE (fromtype) == REFERENCE_TYPE) |
| { |
| tree inner = expr; |
| STRIP_NOPS (inner); |
| |
| if (DECL_P (inner)) |
| warning_at (location, |
| OPT_Waddress, |
| "the compiler can assume that the address of " |
| "%qD will always evaluate to %<true%>", |
| inner); |
| } |
| |
| /* Don't cancel the effect of a CONVERT_EXPR from a REFERENCE_TYPE, |
| since that affects how `default_conversion' will behave. */ |
| if (TREE_CODE (totype) == REFERENCE_TYPE |
| || TREE_CODE (fromtype) == REFERENCE_TYPE) |
| break; |
| /* Don't strip a conversion from C++0x scoped enum, since they |
| don't implicitly convert to other types. */ |
| if (TREE_CODE (fromtype) == ENUMERAL_TYPE |
| && ENUM_IS_SCOPED (fromtype)) |
| break; |
| /* If this isn't narrowing the argument, we can ignore it. */ |
| if (TYPE_PRECISION (totype) >= TYPE_PRECISION (fromtype)) |
| return c_common_truthvalue_conversion (location, |
| TREE_OPERAND (expr, 0)); |
| } |
| break; |
| |
| case MODIFY_EXPR: |
| if (!TREE_NO_WARNING (expr) |
| && warn_parentheses) |
| { |
| warning_at (location, OPT_Wparentheses, |
| "suggest parentheses around assignment used as " |
| "truth value"); |
| TREE_NO_WARNING (expr) = 1; |
| } |
| break; |
| |
| default: |
| break; |
| } |
| |
| if (TREE_CODE (TREE_TYPE (expr)) == COMPLEX_TYPE) |
| { |
| tree t = (in_late_binary_op ? save_expr (expr) : c_save_expr (expr)); |
| expr = (build_binary_op |
| (EXPR_LOCATION (expr), |
| (TREE_SIDE_EFFECTS (expr) |
| ? TRUTH_OR_EXPR : TRUTH_ORIF_EXPR), |
| c_common_truthvalue_conversion |
| (location, |
| build_unary_op (location, REALPART_EXPR, t, false)), |
| c_common_truthvalue_conversion |
| (location, |
| build_unary_op (location, IMAGPART_EXPR, t, false)), |
| 0)); |
| goto ret; |
| } |
| |
| if (TREE_CODE (TREE_TYPE (expr)) == FIXED_POINT_TYPE) |
| { |
| tree fixed_zero_node = build_fixed (TREE_TYPE (expr), |
| FCONST0 (TYPE_MODE |
| (TREE_TYPE (expr)))); |
| return build_binary_op (location, NE_EXPR, expr, fixed_zero_node, 1); |
| } |
| else |
| return build_binary_op (location, NE_EXPR, expr, integer_zero_node, 1); |
| |
| ret: |
| protected_set_expr_location (expr, location); |
| return expr; |
| } |
| |
| static void def_builtin_1 (enum built_in_function fncode, |
| const char *name, |
| enum built_in_class fnclass, |
| tree fntype, tree libtype, |
| bool both_p, bool fallback_p, bool nonansi_p, |
| tree fnattrs, bool implicit_p); |
| |
| |
| /* Apply the TYPE_QUALS to the new DECL. */ |
| |
| void |
| c_apply_type_quals_to_decl (int type_quals, tree decl) |
| { |
| tree type = TREE_TYPE (decl); |
| |
| if (type == error_mark_node) |
| return; |
| |
| if ((type_quals & TYPE_QUAL_CONST) |
| || (type && TREE_CODE (type) == REFERENCE_TYPE)) |
| /* We used to check TYPE_NEEDS_CONSTRUCTING here, but now a constexpr |
| constructor can produce constant init, so rely on cp_finish_decl to |
| clear TREE_READONLY if the variable has non-constant init. */ |
| TREE_READONLY (decl) = 1; |
| if (type_quals & TYPE_QUAL_VOLATILE) |
| { |
| TREE_SIDE_EFFECTS (decl) = 1; |
| TREE_THIS_VOLATILE (decl) = 1; |
| } |
| if (type_quals & TYPE_QUAL_RESTRICT) |
| { |
| while (type && TREE_CODE (type) == ARRAY_TYPE) |
| /* Allow 'restrict' on arrays of pointers. |
| FIXME currently we just ignore it. */ |
| type = TREE_TYPE (type); |
| if (!type |
| || !POINTER_TYPE_P (type) |
| || !C_TYPE_OBJECT_OR_INCOMPLETE_P (TREE_TYPE (type))) |
| error ("invalid use of %<restrict%>"); |
| } |
| } |
| |
| struct c_type_hasher : ggc_ptr_hash<tree_node> |
| { |
| static hashval_t hash (tree); |
| static bool equal (tree, tree); |
| }; |
| |
| /* Hash function for the problem of multiple type definitions in |
| different files. This must hash all types that will compare |
| equal via comptypes to the same value. In practice it hashes |
| on some of the simple stuff and leaves the details to comptypes. */ |
| |
| hashval_t |
| c_type_hasher::hash (tree t) |
| { |
| int n_elements; |
| int shift, size; |
| tree t2; |
| switch (TREE_CODE (t)) |
| { |
| /* For pointers, hash on pointee type plus some swizzling. */ |
| case POINTER_TYPE: |
| return hash (TREE_TYPE (t)) ^ 0x3003003; |
| /* Hash on number of elements and total size. */ |
| case ENUMERAL_TYPE: |
| shift = 3; |
| t2 = TYPE_VALUES (t); |
| break; |
| case RECORD_TYPE: |
| shift = 0; |
| t2 = TYPE_FIELDS (t); |
| break; |
| case QUAL_UNION_TYPE: |
| shift = 1; |
| t2 = TYPE_FIELDS (t); |
| break; |
| case UNION_TYPE: |
| shift = 2; |
| t2 = TYPE_FIELDS (t); |
| break; |
| default: |
| gcc_unreachable (); |
| } |
| /* FIXME: We want to use a DECL_CHAIN iteration method here, but |
| TYPE_VALUES of ENUMERAL_TYPEs is stored as a TREE_LIST. */ |
| n_elements = list_length (t2); |
| /* We might have a VLA here. */ |
| if (TREE_CODE (TYPE_SIZE (t)) != INTEGER_CST) |
| size = 0; |
| else |
| size = TREE_INT_CST_LOW (TYPE_SIZE (t)); |
| return ((size << 24) | (n_elements << shift)); |
| } |
| |
| bool |
| c_type_hasher::equal (tree t1, tree t2) |
| { |
| return lang_hooks.types_compatible_p (t1, t2); |
| } |
| |
| static GTY(()) hash_table<c_type_hasher> *type_hash_table; |
| |
| /* Return the typed-based alias set for T, which may be an expression |
| or a type. Return -1 if we don't do anything special. */ |
| |
| alias_set_type |
| c_common_get_alias_set (tree t) |
| { |
| /* For VLAs, use the alias set of the element type rather than the |
| default of alias set 0 for types compared structurally. */ |
| if (TYPE_P (t) && TYPE_STRUCTURAL_EQUALITY_P (t)) |
| { |
| if (TREE_CODE (t) == ARRAY_TYPE) |
| return get_alias_set (TREE_TYPE (t)); |
| return -1; |
| } |
| |
| /* That's all the expressions we handle specially. */ |
| if (!TYPE_P (t)) |
| return -1; |
| |
| /* The C standard guarantees that any object may be accessed via an |
| lvalue that has character type. */ |
| if (t == char_type_node |
| || t == signed_char_type_node |
| || t == unsigned_char_type_node) |
| return 0; |
| |
| /* The C standard specifically allows aliasing between signed and |
| unsigned variants of the same type. We treat the signed |
| variant as canonical. */ |
| if (TREE_CODE (t) == INTEGER_TYPE && TYPE_UNSIGNED (t)) |
| { |
| tree t1 = c_common_signed_type (t); |
| |
| /* t1 == t can happen for boolean nodes which are always unsigned. */ |
| if (t1 != t) |
| return get_alias_set (t1); |
| } |
| |
| /* Handle the case of multiple type nodes referring to "the same" type, |
| which occurs with IMA. These share an alias set. FIXME: Currently only |
| C90 is handled. (In C99 type compatibility is not transitive, which |
| complicates things mightily. The alias set splay trees can theoretically |
| represent this, but insertion is tricky when you consider all the |
| different orders things might arrive in.) */ |
| |
| if (c_language != clk_c || flag_isoc99) |
| return -1; |
| |
| /* Save time if there's only one input file. */ |
| if (num_in_fnames == 1) |
| return -1; |
| |
| /* Pointers need special handling if they point to any type that |
| needs special handling (below). */ |
| if (TREE_CODE (t) == POINTER_TYPE) |
| { |
| tree t2; |
| /* Find bottom type under any nested POINTERs. */ |
| for (t2 = TREE_TYPE (t); |
| TREE_CODE (t2) == POINTER_TYPE; |
| t2 = TREE_TYPE (t2)) |
| ; |
| if (!RECORD_OR_UNION_TYPE_P (t2) |
| && TREE_CODE (t2) != ENUMERAL_TYPE) |
| return -1; |
| if (TYPE_SIZE (t2) == 0) |
| return -1; |
| } |
| /* These are the only cases that need special handling. */ |
| if (!RECORD_OR_UNION_TYPE_P (t) |
| && TREE_CODE (t) != ENUMERAL_TYPE |
| && TREE_CODE (t) != POINTER_TYPE) |
| return -1; |
| /* Undefined? */ |
| if (TYPE_SIZE (t) == 0) |
| return -1; |
| |
| /* Look up t in hash table. Only one of the compatible types within each |
| alias set is recorded in the table. */ |
| if (!type_hash_table) |
| type_hash_table = hash_table<c_type_hasher>::create_ggc (1021); |
| tree *slot = type_hash_table->find_slot (t, INSERT); |
| if (*slot != NULL) |
| { |
| TYPE_ALIAS_SET (t) = TYPE_ALIAS_SET ((tree)*slot); |
| return TYPE_ALIAS_SET ((tree)*slot); |
| } |
| else |
| /* Our caller will assign and record (in t) a new alias set; all we need |
| to do is remember t in the hash table. */ |
| *slot = t; |
| |
| return -1; |
| } |
| |
| /* Compute the value of 'sizeof (TYPE)' or '__alignof__ (TYPE)', where |
| the IS_SIZEOF parameter indicates which operator is being applied. |
| The COMPLAIN flag controls whether we should diagnose possibly |
| ill-formed constructs or not. LOC is the location of the SIZEOF or |
| TYPEOF operator. If MIN_ALIGNOF, the least alignment required for |
| a type in any context should be returned, rather than the normal |
| alignment for that type. */ |
| |
| tree |
| c_sizeof_or_alignof_type (location_t loc, |
| tree type, bool is_sizeof, bool min_alignof, |
| int complain) |
| { |
| const char *op_name; |
| tree value = NULL; |
| enum tree_code type_code = TREE_CODE (type); |
| |
| op_name = is_sizeof ? "sizeof" : "__alignof__"; |
| |
| if (type_code == FUNCTION_TYPE) |
| { |
| if (is_sizeof) |
| { |
| if (complain && warn_pointer_arith) |
| pedwarn (loc, OPT_Wpointer_arith, |
| "invalid application of %<sizeof%> to a function type"); |
| else if (!complain) |
| return error_mark_node; |
| value = size_one_node; |
| } |
| else |
| { |
| if (complain) |
| { |
| if (c_dialect_cxx ()) |
| pedwarn (loc, OPT_Wpedantic, "ISO C++ does not permit " |
| "%<alignof%> applied to a function type"); |
| else |
| pedwarn (loc, OPT_Wpedantic, "ISO C does not permit " |
| "%<_Alignof%> applied to a function type"); |
| } |
| value = size_int (FUNCTION_BOUNDARY / BITS_PER_UNIT); |
| } |
| } |
| else if (type_code == VOID_TYPE || type_code == ERROR_MARK) |
| { |
| if (type_code == VOID_TYPE |
| && complain && warn_pointer_arith) |
| pedwarn (loc, OPT_Wpointer_arith, |
| "invalid application of %qs to a void type", op_name); |
| else if (!complain) |
| return error_mark_node; |
| value = size_one_node; |
| } |
| else if (!COMPLETE_TYPE_P (type) |
| && (!c_dialect_cxx () || is_sizeof || type_code != ARRAY_TYPE)) |
| { |
| if (complain) |
| error_at (loc, "invalid application of %qs to incomplete type %qT", |
| op_name, type); |
| return error_mark_node; |
| } |
| else if (c_dialect_cxx () && type_code == ARRAY_TYPE |
| && !COMPLETE_TYPE_P (TREE_TYPE (type))) |
| { |
| if (complain) |
| error_at (loc, "invalid application of %qs to array type %qT of " |
| "incomplete element type", op_name, type); |
| return error_mark_node; |
| } |
| else |
| { |
| if (is_sizeof) |
| /* Convert in case a char is more than one unit. */ |
| value = size_binop_loc (loc, CEIL_DIV_EXPR, TYPE_SIZE_UNIT (type), |
| size_int (TYPE_PRECISION (char_type_node) |
| / BITS_PER_UNIT)); |
| else if (min_alignof) |
| value = size_int (min_align_of_type (type)); |
| else |
| value = size_int (TYPE_ALIGN_UNIT (type)); |
| } |
| |
| /* VALUE will have the middle-end integer type sizetype. |
| However, we should really return a value of type `size_t', |
| which is just a typedef for an ordinary integer type. */ |
| value = fold_convert_loc (loc, size_type_node, value); |
| |
| return value; |
| } |
| |
| /* Implement the __alignof keyword: Return the minimum required |
| alignment of EXPR, measured in bytes. For VAR_DECLs, |
| FUNCTION_DECLs and FIELD_DECLs return DECL_ALIGN (which can be set |
| from an "aligned" __attribute__ specification). LOC is the |
| location of the ALIGNOF operator. */ |
| |
| tree |
| c_alignof_expr (location_t loc, tree expr) |
| { |
| tree t; |
| |
| if (VAR_OR_FUNCTION_DECL_P (expr)) |
| t = size_int (DECL_ALIGN_UNIT (expr)); |
| |
| else if (TREE_CODE (expr) == COMPONENT_REF |
| && DECL_C_BIT_FIELD (TREE_OPERAND (expr, 1))) |
| { |
| error_at (loc, "%<__alignof%> applied to a bit-field"); |
| t = size_one_node; |
| } |
| else if (TREE_CODE (expr) == COMPONENT_REF |
| && TREE_CODE (TREE_OPERAND (expr, 1)) == FIELD_DECL) |
| t = size_int (DECL_ALIGN_UNIT (TREE_OPERAND (expr, 1))); |
| |
| else if (INDIRECT_REF_P (expr)) |
| { |
| tree t = TREE_OPERAND (expr, 0); |
| tree best = t; |
| int bestalign = TYPE_ALIGN (TREE_TYPE (TREE_TYPE (t))); |
| |
| while (CONVERT_EXPR_P (t) |
| && TREE_CODE (TREE_TYPE (TREE_OPERAND (t, 0))) == POINTER_TYPE) |
| { |
| int thisalign; |
| |
| t = TREE_OPERAND (t, 0); |
| thisalign = TYPE_ALIGN (TREE_TYPE (TREE_TYPE (t))); |
| if (thisalign > bestalign) |
| best = t, bestalign = thisalign; |
| } |
| return c_alignof (loc, TREE_TYPE (TREE_TYPE (best))); |
| } |
| else |
| return c_alignof (loc, TREE_TYPE (expr)); |
| |
| return fold_convert_loc (loc, size_type_node, t); |
| } |
| |
| /* Handle C and C++ default attributes. */ |
| |
| enum built_in_attribute |
| { |
| #define DEF_ATTR_NULL_TREE(ENUM) ENUM, |
| #define DEF_ATTR_INT(ENUM, VALUE) ENUM, |
| #define DEF_ATTR_STRING(ENUM, VALUE) ENUM, |
| #define DEF_ATTR_IDENT(ENUM, STRING) ENUM, |
| #define DEF_ATTR_TREE_LIST(ENUM, PURPOSE, VALUE, CHAIN) ENUM, |
| #include "builtin-attrs.def" |
| #undef DEF_ATTR_NULL_TREE |
| #undef DEF_ATTR_INT |
| #undef DEF_ATTR_STRING |
| #undef DEF_ATTR_IDENT |
| #undef DEF_ATTR_TREE_LIST |
| ATTR_LAST |
| }; |
| |
| static GTY(()) tree built_in_attributes[(int) ATTR_LAST]; |
| |
| static void c_init_attributes (void); |
| |
| enum c_builtin_type |
| { |
| #define DEF_PRIMITIVE_TYPE(NAME, VALUE) NAME, |
| #define DEF_FUNCTION_TYPE_0(NAME, RETURN) NAME, |
| #define DEF_FUNCTION_TYPE_1(NAME, RETURN, ARG1) NAME, |
| #define DEF_FUNCTION_TYPE_2(NAME, RETURN, ARG1, ARG2) NAME, |
| #define DEF_FUNCTION_TYPE_3(NAME, RETURN, ARG1, ARG2, ARG3) NAME, |
| #define DEF_FUNCTION_TYPE_4(NAME, RETURN, ARG1, ARG2, ARG3, ARG4) NAME, |
| #define DEF_FUNCTION_TYPE_5(NAME, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5) NAME, |
| #define DEF_FUNCTION_TYPE_6(NAME, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6) NAME, |
| #define DEF_FUNCTION_TYPE_7(NAME, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6, ARG7) NAME, |
| #define DEF_FUNCTION_TYPE_8(NAME, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6, ARG7, ARG8) NAME, |
| #define DEF_FUNCTION_TYPE_9(NAME, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6, ARG7, ARG8, ARG9) NAME, |
| #define DEF_FUNCTION_TYPE_10(NAME, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6, ARG7, ARG8, ARG9, ARG10) NAME, |
| #define DEF_FUNCTION_TYPE_11(NAME, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6, ARG7, ARG8, ARG9, ARG10, ARG11) NAME, |
| #define DEF_FUNCTION_TYPE_VAR_0(NAME, RETURN) NAME, |
| #define DEF_FUNCTION_TYPE_VAR_1(NAME, RETURN, ARG1) NAME, |
| #define DEF_FUNCTION_TYPE_VAR_2(NAME, RETURN, ARG1, ARG2) NAME, |
| #define DEF_FUNCTION_TYPE_VAR_3(NAME, RETURN, ARG1, ARG2, ARG3) NAME, |
| #define DEF_FUNCTION_TYPE_VAR_4(NAME, RETURN, ARG1, ARG2, ARG3, ARG4) NAME, |
| #define DEF_FUNCTION_TYPE_VAR_5(NAME, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5) \ |
| NAME, |
| #define DEF_FUNCTION_TYPE_VAR_6(NAME, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6) NAME, |
| #define DEF_FUNCTION_TYPE_VAR_7(NAME, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6, ARG7) NAME, |
| #define DEF_POINTER_TYPE(NAME, TYPE) NAME, |
| #include "builtin-types.def" |
| #undef DEF_PRIMITIVE_TYPE |
| #undef DEF_FUNCTION_TYPE_0 |
| #undef DEF_FUNCTION_TYPE_1 |
| #undef DEF_FUNCTION_TYPE_2 |
| #undef DEF_FUNCTION_TYPE_3 |
| #undef DEF_FUNCTION_TYPE_4 |
| #undef DEF_FUNCTION_TYPE_5 |
| #undef DEF_FUNCTION_TYPE_6 |
| #undef DEF_FUNCTION_TYPE_7 |
| #undef DEF_FUNCTION_TYPE_8 |
| #undef DEF_FUNCTION_TYPE_9 |
| #undef DEF_FUNCTION_TYPE_10 |
| #undef DEF_FUNCTION_TYPE_11 |
| #undef DEF_FUNCTION_TYPE_VAR_0 |
| #undef DEF_FUNCTION_TYPE_VAR_1 |
| #undef DEF_FUNCTION_TYPE_VAR_2 |
| #undef DEF_FUNCTION_TYPE_VAR_3 |
| #undef DEF_FUNCTION_TYPE_VAR_4 |
| #undef DEF_FUNCTION_TYPE_VAR_5 |
| #undef DEF_FUNCTION_TYPE_VAR_6 |
| #undef DEF_FUNCTION_TYPE_VAR_7 |
| #undef DEF_POINTER_TYPE |
| BT_LAST |
| }; |
| |
| typedef enum c_builtin_type builtin_type; |
| |
| /* A temporary array for c_common_nodes_and_builtins. Used in |
| communication with def_fn_type. */ |
| static tree builtin_types[(int) BT_LAST + 1]; |
| |
| /* A helper function for c_common_nodes_and_builtins. Build function type |
| for DEF with return type RET and N arguments. If VAR is true, then the |
| function should be variadic after those N arguments. |
| |
| Takes special care not to ICE if any of the types involved are |
| error_mark_node, which indicates that said type is not in fact available |
| (see builtin_type_for_size). In which case the function type as a whole |
| should be error_mark_node. */ |
| |
| static void |
| def_fn_type (builtin_type def, builtin_type ret, bool var, int n, ...) |
| { |
| tree t; |
| tree *args = XALLOCAVEC (tree, n); |
| va_list list; |
| int i; |
| |
| va_start (list, n); |
| for (i = 0; i < n; ++i) |
| { |
| builtin_type a = (builtin_type) va_arg (list, int); |
| t = builtin_types[a]; |
| if (t == error_mark_node) |
| goto egress; |
| args[i] = t; |
| } |
| |
| t = builtin_types[ret]; |
| if (t == error_mark_node) |
| goto egress; |
| if (var) |
| t = build_varargs_function_type_array (t, n, args); |
| else |
| t = build_function_type_array (t, n, args); |
| |
| egress: |
| builtin_types[def] = t; |
| va_end (list); |
| } |
| |
| /* Build builtin functions common to both C and C++ language |
| frontends. */ |
| |
| static void |
| c_define_builtins (tree va_list_ref_type_node, tree va_list_arg_type_node) |
| { |
| #define DEF_PRIMITIVE_TYPE(ENUM, VALUE) \ |
| builtin_types[ENUM] = VALUE; |
| #define DEF_FUNCTION_TYPE_0(ENUM, RETURN) \ |
| def_fn_type (ENUM, RETURN, 0, 0); |
| #define DEF_FUNCTION_TYPE_1(ENUM, RETURN, ARG1) \ |
| def_fn_type (ENUM, RETURN, 0, 1, ARG1); |
| #define DEF_FUNCTION_TYPE_2(ENUM, RETURN, ARG1, ARG2) \ |
| def_fn_type (ENUM, RETURN, 0, 2, ARG1, ARG2); |
| #define DEF_FUNCTION_TYPE_3(ENUM, RETURN, ARG1, ARG2, ARG3) \ |
| def_fn_type (ENUM, RETURN, 0, 3, ARG1, ARG2, ARG3); |
| #define DEF_FUNCTION_TYPE_4(ENUM, RETURN, ARG1, ARG2, ARG3, ARG4) \ |
| def_fn_type (ENUM, RETURN, 0, 4, ARG1, ARG2, ARG3, ARG4); |
| #define DEF_FUNCTION_TYPE_5(ENUM, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5) \ |
| def_fn_type (ENUM, RETURN, 0, 5, ARG1, ARG2, ARG3, ARG4, ARG5); |
| #define DEF_FUNCTION_TYPE_6(ENUM, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6) \ |
| def_fn_type (ENUM, RETURN, 0, 6, ARG1, ARG2, ARG3, ARG4, ARG5, ARG6); |
| #define DEF_FUNCTION_TYPE_7(ENUM, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6, ARG7) \ |
| def_fn_type (ENUM, RETURN, 0, 7, ARG1, ARG2, ARG3, ARG4, ARG5, ARG6, ARG7); |
| #define DEF_FUNCTION_TYPE_8(ENUM, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6, ARG7, ARG8) \ |
| def_fn_type (ENUM, RETURN, 0, 8, ARG1, ARG2, ARG3, ARG4, ARG5, ARG6, \ |
| ARG7, ARG8); |
| #define DEF_FUNCTION_TYPE_9(ENUM, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6, ARG7, ARG8, ARG9) \ |
| def_fn_type (ENUM, RETURN, 0, 9, ARG1, ARG2, ARG3, ARG4, ARG5, ARG6, \ |
| ARG7, ARG8, ARG9); |
| #define DEF_FUNCTION_TYPE_10(ENUM, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6, ARG7, ARG8, ARG9, ARG10) \ |
| def_fn_type (ENUM, RETURN, 0, 10, ARG1, ARG2, ARG3, ARG4, ARG5, ARG6, \ |
| ARG7, ARG8, ARG9, ARG10); |
| #define DEF_FUNCTION_TYPE_11(ENUM, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6, ARG7, ARG8, ARG9, ARG10, ARG11) \ |
| def_fn_type (ENUM, RETURN, 0, 11, ARG1, ARG2, ARG3, ARG4, ARG5, ARG6, \ |
| ARG7, ARG8, ARG9, ARG10, ARG11); |
| #define DEF_FUNCTION_TYPE_VAR_0(ENUM, RETURN) \ |
| def_fn_type (ENUM, RETURN, 1, 0); |
| #define DEF_FUNCTION_TYPE_VAR_1(ENUM, RETURN, ARG1) \ |
| def_fn_type (ENUM, RETURN, 1, 1, ARG1); |
| #define DEF_FUNCTION_TYPE_VAR_2(ENUM, RETURN, ARG1, ARG2) \ |
| def_fn_type (ENUM, RETURN, 1, 2, ARG1, ARG2); |
| #define DEF_FUNCTION_TYPE_VAR_3(ENUM, RETURN, ARG1, ARG2, ARG3) \ |
| def_fn_type (ENUM, RETURN, 1, 3, ARG1, ARG2, ARG3); |
| #define DEF_FUNCTION_TYPE_VAR_4(ENUM, RETURN, ARG1, ARG2, ARG3, ARG4) \ |
| def_fn_type (ENUM, RETURN, 1, 4, ARG1, ARG2, ARG3, ARG4); |
| #define DEF_FUNCTION_TYPE_VAR_5(ENUM, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5) \ |
| def_fn_type (ENUM, RETURN, 1, 5, ARG1, ARG2, ARG3, ARG4, ARG5); |
| #define DEF_FUNCTION_TYPE_VAR_6(ENUM, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6) \ |
| def_fn_type (ENUM, RETURN, 1, 6, ARG1, ARG2, ARG3, ARG4, ARG5, ARG6); |
| #define DEF_FUNCTION_TYPE_VAR_7(ENUM, RETURN, ARG1, ARG2, ARG3, ARG4, ARG5, \ |
| ARG6, ARG7) \ |
| def_fn_type (ENUM, RETURN, 1, 7, ARG1, ARG2, ARG3, ARG4, ARG5, ARG6, ARG7); |
| #define DEF_POINTER_TYPE(ENUM, TYPE) \ |
| builtin_types[(int) ENUM] = build_pointer_type (builtin_types[(int) TYPE]); |
| |
| #include "builtin-types.def" |
| |
| #undef DEF_PRIMITIVE_TYPE |
| #undef DEF_FUNCTION_TYPE_0 |
| #undef DEF_FUNCTION_TYPE_1 |
| #undef DEF_FUNCTION_TYPE_2 |
| #undef DEF_FUNCTION_TYPE_3 |
| #undef DEF_FUNCTION_TYPE_4 |
| #undef DEF_FUNCTION_TYPE_5 |
| #undef DEF_FUNCTION_TYPE_6 |
| #undef DEF_FUNCTION_TYPE_7 |
| #undef DEF_FUNCTION_TYPE_8 |
| #undef DEF_FUNCTION_TYPE_9 |
| #undef DEF_FUNCTION_TYPE_10 |
| #undef DEF_FUNCTION_TYPE_11 |
| #undef DEF_FUNCTION_TYPE_VAR_0 |
| #undef DEF_FUNCTION_TYPE_VAR_1 |
| #undef DEF_FUNCTION_TYPE_VAR_2 |
| #undef DEF_FUNCTION_TYPE_VAR_3 |
| #undef DEF_FUNCTION_TYPE_VAR_4 |
| #undef DEF_FUNCTION_TYPE_VAR_5 |
| #undef DEF_FUNCTION_TYPE_VAR_6 |
| #undef DEF_FUNCTION_TYPE_VAR_7 |
| #undef DEF_POINTER_TYPE |
| builtin_types[(int) BT_LAST] = NULL_TREE; |
| |
| c_init_attributes (); |
| |
| #define DEF_BUILTIN(ENUM, NAME, CLASS, TYPE, LIBTYPE, BOTH_P, FALLBACK_P, \ |
| NONANSI_P, ATTRS, IMPLICIT, COND) \ |
| if (NAME && COND) \ |
| def_builtin_1 (ENUM, NAME, CLASS, \ |
| builtin_types[(int) TYPE], \ |
| builtin_types[(int) LIBTYPE], \ |
| BOTH_P, FALLBACK_P, NONANSI_P, \ |
| built_in_attributes[(int) ATTRS], IMPLICIT); |
| #include "builtins.def" |
| |
| targetm.init_builtins (); |
| |
| build_common_builtin_nodes (); |
| |
| if (flag_cilkplus) |
| cilk_init_builtins (); |
| } |
| |
| /* Like get_identifier, but avoid warnings about null arguments when |
| the argument may be NULL for targets where GCC lacks stdint.h type |
| information. */ |
| |
| static inline tree |
| c_get_ident (const char *id) |
| { |
| return get_identifier (id); |
| } |
| |
| /* Build tree nodes and builtin functions common to both C and C++ language |
| frontends. */ |
| |
| void |
| c_common_nodes_and_builtins (void) |
| { |
| int char16_type_size; |
| int char32_type_size; |
| int wchar_type_size; |
| tree array_domain_type; |
| tree va_list_ref_type_node; |
| tree va_list_arg_type_node; |
| int i; |
| |
| build_common_tree_nodes (flag_signed_char); |
| |
| /* Define `int' and `char' first so that dbx will output them first. */ |
| record_builtin_type (RID_INT, NULL, integer_type_node); |
| record_builtin_type (RID_CHAR, "char", char_type_node); |
| |
| /* `signed' is the same as `int'. FIXME: the declarations of "signed", |
| "unsigned long", "long long unsigned" and "unsigned short" were in C++ |
| but not C. Are the conditionals here needed? */ |
| if (c_dialect_cxx ()) |
| record_builtin_type (RID_SIGNED, NULL, integer_type_node); |
| record_builtin_type (RID_LONG, "long int", long_integer_type_node); |
| record_builtin_type (RID_UNSIGNED, "unsigned int", unsigned_type_node); |
| record_builtin_type (RID_MAX, "long unsigned int", |
| long_unsigned_type_node); |
| |
| for (i = 0; i < NUM_INT_N_ENTS; i ++) |
| { |
| char name[25]; |
| |
| sprintf (name, "__int%d", int_n_data[i].bitsize); |
| record_builtin_type ((enum rid)(RID_FIRST_INT_N + i), name, |
| int_n_trees[i].signed_type); |
| sprintf (name, "__int%d unsigned", int_n_data[i].bitsize); |
| record_builtin_type (RID_MAX, name, int_n_trees[i].unsigned_type); |
| } |
| |
| if (c_dialect_cxx ()) |
| record_builtin_type (RID_MAX, "unsigned long", long_unsigned_type_node); |
| record_builtin_type (RID_MAX, "long long int", |
| long_long_integer_type_node); |
| record_builtin_type (RID_MAX, "long long unsigned int", |
| long_long_unsigned_type_node); |
| if (c_dialect_cxx ()) |
| record_builtin_type (RID_MAX, "long long unsigned", |
| long_long_unsigned_type_node); |
| record_builtin_type (RID_SHORT, "short int", short_integer_type_node); |
| record_builtin_type (RID_MAX, "short unsigned int", |
| short_unsigned_type_node); |
| if (c_dialect_cxx ()) |
| record_builtin_type (RID_MAX, "unsigned short", |
| short_unsigned_type_node); |
| |
| /* Define both `signed char' and `unsigned char'. */ |
| record_builtin_type (RID_MAX, "signed char", signed_char_type_node); |
| record_builtin_type (RID_MAX, "unsigned char", unsigned_char_type_node); |
| |
| /* These are types that c_common_type_for_size and |
| c_common_type_for_mode use. */ |
| lang_hooks.decls.pushdecl (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, NULL_TREE, |
| intQI_type_node)); |
| lang_hooks.decls.pushdecl (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, NULL_TREE, |
| intHI_type_node)); |
| lang_hooks.decls.pushdecl (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, NULL_TREE, |
| intSI_type_node)); |
| lang_hooks.decls.pushdecl (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, NULL_TREE, |
| intDI_type_node)); |
| #if HOST_BITS_PER_WIDE_INT >= 64 |
| /* Note that this is different than the __int128 type that's part of |
| the generic __intN support. */ |
| if (targetm.scalar_mode_supported_p (TImode)) |
| lang_hooks.decls.pushdecl (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, |
| get_identifier ("__int128_t"), |
| intTI_type_node)); |
| #endif |
| lang_hooks.decls.pushdecl (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, NULL_TREE, |
| unsigned_intQI_type_node)); |
| lang_hooks.decls.pushdecl (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, NULL_TREE, |
| unsigned_intHI_type_node)); |
| lang_hooks.decls.pushdecl (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, NULL_TREE, |
| unsigned_intSI_type_node)); |
| lang_hooks.decls.pushdecl (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, NULL_TREE, |
| unsigned_intDI_type_node)); |
| #if HOST_BITS_PER_WIDE_INT >= 64 |
| if (targetm.scalar_mode_supported_p (TImode)) |
| lang_hooks.decls.pushdecl (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, |
| get_identifier ("__uint128_t"), |
| unsigned_intTI_type_node)); |
| #endif |
| |
| /* Create the widest literal types. */ |
| if (targetm.scalar_mode_supported_p (TImode)) |
| { |
| widest_integer_literal_type_node = intTI_type_node; |
| widest_unsigned_literal_type_node = unsigned_intTI_type_node; |
| } |
| else |
| { |
| widest_integer_literal_type_node = intDI_type_node; |
| widest_unsigned_literal_type_node = unsigned_intDI_type_node; |
| } |
| |
| signed_size_type_node = c_common_signed_type (size_type_node); |
| |
| pid_type_node = |
| TREE_TYPE (identifier_global_value (get_identifier (PID_TYPE))); |
| |
| record_builtin_type (RID_FLOAT, NULL, float_type_node); |
| record_builtin_type (RID_DOUBLE, NULL, double_type_node); |
| record_builtin_type (RID_MAX, "long double", long_double_type_node); |
| |
| if (!c_dialect_cxx ()) |
| for (i = 0; i < NUM_FLOATN_NX_TYPES; i++) |
| if (FLOATN_NX_TYPE_NODE (i) != NULL_TREE) |
| record_builtin_type ((enum rid) (RID_FLOATN_NX_FIRST + i), NULL, |
| FLOATN_NX_TYPE_NODE (i)); |
| |
| /* Only supported decimal floating point extension if the target |
| actually supports underlying modes. */ |
| if (targetm.scalar_mode_supported_p (SDmode) |
| && targetm.scalar_mode_supported_p (DDmode) |
| && targetm.scalar_mode_supported_p (TDmode)) |
| { |
| record_builtin_type (RID_DFLOAT32, NULL, dfloat32_type_node); |
| record_builtin_type (RID_DFLOAT64, NULL, dfloat64_type_node); |
| record_builtin_type (RID_DFLOAT128, NULL, dfloat128_type_node); |
| } |
| |
| if (targetm.fixed_point_supported_p ()) |
| { |
| record_builtin_type (RID_MAX, "short _Fract", short_fract_type_node); |
| record_builtin_type (RID_FRACT, NULL, fract_type_node); |
| record_builtin_type (RID_MAX, "long _Fract", long_fract_type_node); |
| record_builtin_type (RID_MAX, "long long _Fract", |
| long_long_fract_type_node); |
| record_builtin_type (RID_MAX, "unsigned short _Fract", |
| unsigned_short_fract_type_node); |
| record_builtin_type (RID_MAX, "unsigned _Fract", |
| unsigned_fract_type_node); |
| record_builtin_type (RID_MAX, "unsigned long _Fract", |
| unsigned_long_fract_type_node); |
| record_builtin_type (RID_MAX, "unsigned long long _Fract", |
| unsigned_long_long_fract_type_node); |
| record_builtin_type (RID_MAX, "_Sat short _Fract", |
| sat_short_fract_type_node); |
| record_builtin_type (RID_MAX, "_Sat _Fract", sat_fract_type_node); |
| record_builtin_type (RID_MAX, "_Sat long _Fract", |
| sat_long_fract_type_node); |
| record_builtin_type (RID_MAX, "_Sat long long _Fract", |
| sat_long_long_fract_type_node); |
| record_builtin_type (RID_MAX, "_Sat unsigned short _Fract", |
| sat_unsigned_short_fract_type_node); |
| record_builtin_type (RID_MAX, "_Sat unsigned _Fract", |
| sat_unsigned_fract_type_node); |
| record_builtin_type (RID_MAX, "_Sat unsigned long _Fract", |
| sat_unsigned_long_fract_type_node); |
| record_builtin_type (RID_MAX, "_Sat unsigned long long _Fract", |
| sat_unsigned_long_long_fract_type_node); |
| record_builtin_type (RID_MAX, "short _Accum", short_accum_type_node); |
| record_builtin_type (RID_ACCUM, NULL, accum_type_node); |
| record_builtin_type (RID_MAX, "long _Accum", long_accum_type_node); |
| record_builtin_type (RID_MAX, "long long _Accum", |
| long_long_accum_type_node); |
| record_builtin_type (RID_MAX, "unsigned short _Accum", |
| unsigned_short_accum_type_node); |
| record_builtin_type (RID_MAX, "unsigned _Accum", |
| unsigned_accum_type_node); |
| record_builtin_type (RID_MAX, "unsigned long _Accum", |
| unsigned_long_accum_type_node); |
| record_builtin_type (RID_MAX, "unsigned long long _Accum", |
| unsigned_long_long_accum_type_node); |
| record_builtin_type (RID_MAX, "_Sat short _Accum", |
| sat_short_accum_type_node); |
| record_builtin_type (RID_MAX, "_Sat _Accum", sat_accum_type_node); |
| record_builtin_type (RID_MAX, "_Sat long _Accum", |
| sat_long_accum_type_node); |
| record_builtin_type (RID_MAX, "_Sat long long _Accum", |
| sat_long_long_accum_type_node); |
| record_builtin_type (RID_MAX, "_Sat unsigned short _Accum", |
| sat_unsigned_short_accum_type_node); |
| record_builtin_type (RID_MAX, "_Sat unsigned _Accum", |
| sat_unsigned_accum_type_node); |
| record_builtin_type (RID_MAX, "_Sat unsigned long _Accum", |
| sat_unsigned_long_accum_type_node); |
| record_builtin_type (RID_MAX, "_Sat unsigned long long _Accum", |
| sat_unsigned_long_long_accum_type_node); |
| |
| } |
| |
| lang_hooks.decls.pushdecl (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, |
| get_identifier ("complex int"), |
| complex_integer_type_node)); |
| lang_hooks.decls.pushdecl (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, |
| get_identifier ("complex float"), |
| complex_float_type_node)); |
| lang_hooks.decls.pushdecl (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, |
| get_identifier ("complex double"), |
| complex_double_type_node)); |
| lang_hooks.decls.pushdecl |
| (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, get_identifier ("complex long double"), |
| complex_long_double_type_node)); |
| |
| if (!c_dialect_cxx ()) |
| for (i = 0; i < NUM_FLOATN_NX_TYPES; i++) |
| if (COMPLEX_FLOATN_NX_TYPE_NODE (i) != NULL_TREE) |
| { |
| char buf[30]; |
| sprintf (buf, "complex _Float%d%s", floatn_nx_types[i].n, |
| floatn_nx_types[i].extended ? "x" : ""); |
| lang_hooks.decls.pushdecl |
| (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, |
| get_identifier (buf), |
| COMPLEX_FLOATN_NX_TYPE_NODE (i))); |
| } |
| |
| if (c_dialect_cxx ()) |
| { |
| /* For C++, make fileptr_type_node a distinct void * type until |
| FILE type is defined. */ |
| fileptr_type_node = build_variant_type_copy (ptr_type_node); |
| /* Likewise for const struct tm*. */ |
| const_tm_ptr_type_node = build_variant_type_copy (const_ptr_type_node); |
| } |
| |
| record_builtin_type (RID_VOID, NULL, void_type_node); |
| |
| /* Set the TYPE_NAME for any variants that were built before |
| record_builtin_type gave names to the built-in types. */ |
| { |
| tree void_name = TYPE_NAME (void_type_node); |
| TYPE_NAME (void_type_node) = NULL_TREE; |
| TYPE_NAME (build_qualified_type (void_type_node, TYPE_QUAL_CONST)) |
| = void_name; |
| TYPE_NAME (void_type_node) = void_name; |
| } |
| |
| void_list_node = build_void_list_node (); |
| |
| /* Make a type to be the domain of a few array types |
| whose domains don't really matter. |
| 200 is small enough that it always fits in size_t |
| and large enough that it can hold most function names for the |
| initializations of __FUNCTION__ and __PRETTY_FUNCTION__. */ |
| array_domain_type = build_index_type (size_int (200)); |
| |
| /* Make a type for arrays of characters. |
| With luck nothing will ever really depend on the length of this |
| array type. */ |
| char_array_type_node |
| = build_array_type (char_type_node, array_domain_type); |
| |
| string_type_node = build_pointer_type (char_type_node); |
| const_string_type_node |
| = build_pointer_type (build_qualified_type |
| (char_type_node, TYPE_QUAL_CONST)); |
| |
| /* This is special for C++ so functions can be overloaded. */ |
| wchar_type_node = get_identifier (MODIFIED_WCHAR_TYPE); |
| wchar_type_node = TREE_TYPE (identifier_global_value (wchar_type_node)); |
| wchar_type_size = TYPE_PRECISION (wchar_type_node); |
| underlying_wchar_type_node = wchar_type_node; |
| if (c_dialect_cxx ()) |
| { |
| if (TYPE_UNSIGNED (wchar_type_node)) |
| wchar_type_node = make_unsigned_type (wchar_type_size); |
| else |
| wchar_type_node = make_signed_type (wchar_type_size); |
| record_builtin_type (RID_WCHAR, "wchar_t", wchar_type_node); |
| } |
| |
| /* This is for wide string constants. */ |
| wchar_array_type_node |
| = build_array_type (wchar_type_node, array_domain_type); |
| |
| /* Define 'char16_t'. */ |
| char16_type_node = get_identifier (CHAR16_TYPE); |
| char16_type_node = TREE_TYPE (identifier_global_value (char16_type_node)); |
| char16_type_size = TYPE_PRECISION (char16_type_node); |
| if (c_dialect_cxx ()) |
| { |
| char16_type_node = make_unsigned_type (char16_type_size); |
| |
| if (cxx_dialect >= cxx11) |
| record_builtin_type (RID_CHAR16, "char16_t", char16_type_node); |
| } |
| |
| /* This is for UTF-16 string constants. */ |
| char16_array_type_node |
| = build_array_type (char16_type_node, array_domain_type); |
| |
| /* Define 'char32_t'. */ |
| char32_type_node = get_identifier (CHAR32_TYPE); |
| char32_type_node = TREE_TYPE (identifier_global_value (char32_type_node)); |
| char32_type_size = TYPE_PRECISION (char32_type_node); |
| if (c_dialect_cxx ()) |
| { |
| char32_type_node = make_unsigned_type (char32_type_size); |
| |
| if (cxx_dialect >= cxx11) |
| record_builtin_type (RID_CHAR32, "char32_t", char32_type_node); |
| } |
| |
| /* This is for UTF-32 string constants. */ |
| char32_array_type_node |
| = build_array_type (char32_type_node, array_domain_type); |
| |
| wint_type_node = |
| TREE_TYPE (identifier_global_value (get_identifier (WINT_TYPE))); |
| |
| intmax_type_node = |
| TREE_TYPE (identifier_global_value (get_identifier (INTMAX_TYPE))); |
| uintmax_type_node = |
| TREE_TYPE (identifier_global_value (get_identifier (UINTMAX_TYPE))); |
| |
| if (SIG_ATOMIC_TYPE) |
| sig_atomic_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (SIG_ATOMIC_TYPE))); |
| if (INT8_TYPE) |
| int8_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (INT8_TYPE))); |
| if (INT16_TYPE) |
| int16_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (INT16_TYPE))); |
| if (INT32_TYPE) |
| int32_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (INT32_TYPE))); |
| if (INT64_TYPE) |
| int64_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (INT64_TYPE))); |
| if (UINT8_TYPE) |
| uint8_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (UINT8_TYPE))); |
| if (UINT16_TYPE) |
| c_uint16_type_node = uint16_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (UINT16_TYPE))); |
| if (UINT32_TYPE) |
| c_uint32_type_node = uint32_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (UINT32_TYPE))); |
| if (UINT64_TYPE) |
| c_uint64_type_node = uint64_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (UINT64_TYPE))); |
| if (INT_LEAST8_TYPE) |
| int_least8_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (INT_LEAST8_TYPE))); |
| if (INT_LEAST16_TYPE) |
| int_least16_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (INT_LEAST16_TYPE))); |
| if (INT_LEAST32_TYPE) |
| int_least32_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (INT_LEAST32_TYPE))); |
| if (INT_LEAST64_TYPE) |
| int_least64_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (INT_LEAST64_TYPE))); |
| if (UINT_LEAST8_TYPE) |
| uint_least8_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (UINT_LEAST8_TYPE))); |
| if (UINT_LEAST16_TYPE) |
| uint_least16_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (UINT_LEAST16_TYPE))); |
| if (UINT_LEAST32_TYPE) |
| uint_least32_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (UINT_LEAST32_TYPE))); |
| if (UINT_LEAST64_TYPE) |
| uint_least64_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (UINT_LEAST64_TYPE))); |
| if (INT_FAST8_TYPE) |
| int_fast8_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (INT_FAST8_TYPE))); |
| if (INT_FAST16_TYPE) |
| int_fast16_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (INT_FAST16_TYPE))); |
| if (INT_FAST32_TYPE) |
| int_fast32_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (INT_FAST32_TYPE))); |
| if (INT_FAST64_TYPE) |
| int_fast64_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (INT_FAST64_TYPE))); |
| if (UINT_FAST8_TYPE) |
| uint_fast8_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (UINT_FAST8_TYPE))); |
| if (UINT_FAST16_TYPE) |
| uint_fast16_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (UINT_FAST16_TYPE))); |
| if (UINT_FAST32_TYPE) |
| uint_fast32_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (UINT_FAST32_TYPE))); |
| if (UINT_FAST64_TYPE) |
| uint_fast64_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (UINT_FAST64_TYPE))); |
| if (INTPTR_TYPE) |
| intptr_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (INTPTR_TYPE))); |
| if (UINTPTR_TYPE) |
| uintptr_type_node = |
| TREE_TYPE (identifier_global_value (c_get_ident (UINTPTR_TYPE))); |
| |
| default_function_type |
| = build_varargs_function_type_list (integer_type_node, NULL_TREE); |
| unsigned_ptrdiff_type_node = c_common_unsigned_type (ptrdiff_type_node); |
| |
| lang_hooks.decls.pushdecl |
| (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, get_identifier ("__builtin_va_list"), |
| va_list_type_node)); |
| if (targetm.enum_va_list_p) |
| { |
| int l; |
| const char *pname; |
| tree ptype; |
| |
| for (l = 0; targetm.enum_va_list_p (l, &pname, &ptype); ++l) |
| { |
| lang_hooks.decls.pushdecl |
| (build_decl (UNKNOWN_LOCATION, |
| TYPE_DECL, get_identifier (pname), |
| ptype)); |
| |
| } |
| } |
| |
| if (TREE_CODE (va_list_type_node) == ARRAY_TYPE) |
| { |
| va_list_arg_type_node = va_list_ref_type_node = |
| build_pointer_type (TREE_TYPE (va_list_type_node)); |
| } |
| else |
| { |
| va_list_arg_type_node = va_list_type_node; |
| va_list_ref_type_node = build_reference_type (va_list_type_node); |
| } |
| |
| if (!flag_preprocess_only) |
| c_define_builtins (va_list_ref_type_node, va_list_arg_type_node); |
| |
| main_identifier_node = get_identifier ("main"); |
| |
| /* Create the built-in __null node. It is important that this is |
| not shared. */ |
| null_node = make_int_cst (1, 1); |
| TREE_TYPE (null_node) = c_common_type_for_size (POINTER_SIZE, 0); |
| |
| /* Since builtin_types isn't gc'ed, don't export these nodes. */ |
| memset (builtin_types, 0, sizeof (builtin_types)); |
| } |
| |
| /* The number of named compound-literals generated thus far. */ |
| static GTY(()) int compound_literal_number; |
| |
| /* Set DECL_NAME for DECL, a VAR_DECL for a compound-literal. */ |
| |
| void |
| set_compound_literal_name (tree decl) |
| { |
| char *name; |
| ASM_FORMAT_PRIVATE_NAME (name, "__compound_literal", |
| compound_literal_number); |
| compound_literal_number++; |
| DECL_NAME (decl) = get_identifier (name); |
| } |
| |
| /* build_va_arg helper function. Return a VA_ARG_EXPR with location LOC, type |
| TYPE and operand OP. */ |
| |
| static tree |
| build_va_arg_1 (location_t loc, tree type, tree op) |
| { |
| tree expr = build1 (VA_ARG_EXPR, type, op); |
| SET_EXPR_LOCATION (expr, loc); |
| return expr; |
| } |
| |
| /* Return a VA_ARG_EXPR corresponding to a source-level expression |
| va_arg (EXPR, TYPE) at source location LOC. */ |
| |
| tree |
| build_va_arg (location_t loc, tree expr, tree type) |
| { |
| tree va_type = TREE_TYPE (expr); |
| tree canon_va_type = (va_type == error_mark_node |
| ? error_mark_node |
| : targetm.canonical_va_list_type (va_type)); |
| |
| if (va_type == error_mark_node |
| || canon_va_type == NULL_TREE) |
| { |
| if (canon_va_type == NULL_TREE) |
| error_at (loc, "first argument to %<va_arg%> not of type %<va_list%>"); |
| |
| /* Let's handle things neutrallly, if expr: |
| - has undeclared type, or |
| - is not an va_list type. */ |
| return build_va_arg_1 (loc, type, error_mark_node); |
| } |
| |
| if (TREE_CODE (canon_va_type) != ARRAY_TYPE) |
| { |
| /* Case 1: Not an array type. */ |
| |
| /* Take the address, to get '&ap'. Note that &ap is not a va_list |
| type. */ |
| mark_addressable (expr); |
| expr = build1 (ADDR_EXPR, build_pointer_type (TREE_TYPE (expr)), expr); |
| |
| return build_va_arg_1 (loc, type, expr); |
| } |
| |
| /* Case 2: Array type. |
| |
| Background: |
| |
| For contrast, let's start with the simple case (case 1). If |
| canon_va_type is not an array type, but say a char *, then when |
| passing-by-value a va_list, the type of the va_list param decl is |
| the same as for another va_list decl (all ap's are char *): |
| |
| f2_1 (char * ap) |
| D.1815 = VA_ARG (&ap, 0B, 1); |
| return D.1815; |
| |
| f2 (int i) |
| char * ap.0; |
| char * ap; |
| __builtin_va_start (&ap, 0); |
| ap.0 = ap; |
| res = f2_1 (ap.0); |
| __builtin_va_end (&ap); |
| D.1812 = res; |
| return D.1812; |
| |
| However, if canon_va_type is ARRAY_TYPE, then when passing-by-value a |
| va_list the type of the va_list param decl (case 2b, struct * ap) is not |
| the same as for another va_list decl (case 2a, struct ap[1]). |
| |
| f2_1 (struct * ap) |
| D.1844 = VA_ARG (ap, 0B, 0); |
| return D.1844; |
| |
| f2 (int i) |
| struct ap[1]; |
| __builtin_va_start (&ap, 0); |
| res = f2_1 (&ap); |
| __builtin_va_end (&ap); |
| D.1841 = res; |
| return D.1841; |
| |
| Case 2b is different because: |
| - on the callee side, the parm decl has declared type va_list, but |
| grokdeclarator changes the type of the parm decl to a pointer to the |
| array elem type. |
| - on the caller side, the pass-by-value uses &ap. |
| |
| We unify these two cases (case 2a: va_list is array type, |
| case 2b: va_list is pointer to array elem type), by adding '&' for the |
| array type case, such that we have a pointer to array elem in both |
| cases. */ |
| |
| if (TREE_CODE (va_type) == ARRAY_TYPE) |
| { |
| /* Case 2a: va_list is array type. */ |
| |
| /* Take the address, to get '&ap'. Make sure it's a pointer to array |
| elem type. */ |
| mark_addressable (expr); |
| expr = build1 (ADDR_EXPR, build_pointer_type (TREE_TYPE (canon_va_type)), |
| expr); |
| |
| /* Verify that &ap is still recognized as having va_list type. */ |
| tree canon_expr_type |
| = targetm.canonical_va_list_type (TREE_TYPE (expr)); |
| gcc_assert (canon_expr_type != NULL_TREE); |
| } |
| else |
| { |
| /* Case 2b: va_list is pointer to array elem type. */ |
| gcc_assert (POINTER_TYPE_P (va_type)); |
| |
| /* Comparison as in std_canonical_va_list_type. */ |
| gcc_assert (TYPE_MAIN_VARIANT (TREE_TYPE (va_type)) |
| == TYPE_MAIN_VARIANT (TREE_TYPE (canon_va_type))); |
| |
| /* Don't take the address. We've already got '&ap'. */ |
| ; |
| } |
| |
| return build_va_arg_1 (loc, type, expr); |
| } |
| |
| |
| /* Linked list of disabled built-in functions. */ |
| |
| struct disabled_builtin |
| { |
| const char *name; |
| struct disabled_builtin *next; |
| }; |
| static disabled_builtin *disabled_builtins = NULL; |
| |
| static bool builtin_function_disabled_p (const char *); |
| |
| /* Disable a built-in function specified by -fno-builtin-NAME. If NAME |
| begins with "__builtin_", give an error. */ |
| |
| void |
| disable_builtin_function (const char *name) |
| { |
| if (strncmp (name, "__builtin_", strlen ("__builtin_")) == 0) |
| error ("cannot disable built-in function %qs", name); |
| else |
| { |
| disabled_builtin *new_disabled_builtin = XNEW (disabled_builtin); |
| new_disabled_builtin->name = name; |
| new_disabled_builtin->next = disabled_builtins; |
| disabled_builtins = new_disabled_builtin; |
| } |
| } |
| |
| |
| /* Return true if the built-in function NAME has been disabled, false |
| otherwise. */ |
| |
| static bool |
| builtin_function_disabled_p (const char *name) |
| { |
| disabled_builtin *p; |
| for (p = disabled_builtins; p != NULL; p = p->next) |
| { |
| if (strcmp (name, p->name) == 0) |
| return true; |
| } |
| return false; |
| } |
| |
| |
| /* Worker for DEF_BUILTIN. |
| Possibly define a builtin function with one or two names. |
| Does not declare a non-__builtin_ function if flag_no_builtin, or if |
| nonansi_p and flag_no_nonansi_builtin. */ |
| |
| static void |
| def_builtin_1 (enum built_in_function fncode, |
| const char *name, |
| enum built_in_class fnclass, |
| tree fntype, tree libtype, |
| bool both_p, bool fallback_p, bool nonansi_p, |
| tree fnattrs, bool implicit_p) |
| { |
| tree decl; |
| const char *libname; |
| |
| if (fntype == error_mark_node) |
| return; |
| |
| gcc_assert ((!both_p && !fallback_p) |
| || !strncmp (name, "__builtin_", |
| strlen ("__builtin_"))); |
| |
| libname = name + strlen ("__builtin_"); |
| decl = add_builtin_function (name, fntype, fncode, fnclass, |
| (fallback_p ? libname : NULL), |
| fnattrs); |
| |
| set_builtin_decl (fncode, decl, implicit_p); |
| |
| if (both_p |
| && !flag_no_builtin && !builtin_function_disabled_p (libname) |
| && !(nonansi_p && flag_no_nonansi_builtin)) |
| add_builtin_function (libname, libtype, fncode, fnclass, |
| NULL, fnattrs); |
| } |
| |
| /* Nonzero if the type T promotes to int. This is (nearly) the |
| integral promotions defined in ISO C99 6.3.1.1/2. */ |
| |
| bool |
| c_promoting_integer_type_p (const_tree t) |
| { |
| switch (TREE_CODE (t)) |
| { |
| case INTEGER_TYPE: |
| return (TYPE_MAIN_VARIANT (t) == char_type_node |
| || TYPE_MAIN_VARIANT (t) == signed_char_type_node |
| || TYPE_MAIN_VARIANT (t) == unsigned_char_type_node |
| || TYPE_MAIN_VARIANT (t) == short_integer_type_node |
| || TYPE_MAIN_VARIANT (t) == short_unsigned_type_node |
| || TYPE_PRECISION (t) < TYPE_PRECISION (integer_type_node)); |
| |
| case ENUMERAL_TYPE: |
| /* ??? Technically all enumerations not larger than an int |
| promote to an int. But this is used along code paths |
| that only want to notice a size change. */ |
| return TYPE_PRECISION (t) < TYPE_PRECISION (integer_type_node); |
| |
| case BOOLEAN_TYPE: |
| return 1; |
| |
| default: |
| return 0; |
| } |
| } |
| |
| /* Return 1 if PARMS specifies a fixed number of parameters |
| and none of their types is affected by default promotions. */ |
| |
| int |
| self_promoting_args_p (const_tree parms) |
| { |
| const_tree t; |
| for (t = parms; t; t = TREE_CHAIN (t)) |
| { |
| tree type = TREE_VALUE (t); |
| |
| if (type == error_mark_node) |
| continue; |
| |
| if (TREE_CHAIN (t) == 0 && type != void_type_node) |
| return 0; |
| |
| if (type == 0) |
| return 0; |
| |
| if (TYPE_MAIN_VARIANT (type) == float_type_node) |
| return 0; |
| |
| if (c_promoting_integer_type_p (type)) |
| return 0; |
| } |
| return 1; |
| } |
| |
| /* Recursively remove any '*' or '&' operator from TYPE. */ |
| tree |
| strip_pointer_operator (tree t) |
| { |
| while (POINTER_TYPE_P (t)) |
| t = TREE_TYPE (t); |
| return t; |
| } |
| |
| /* Recursively remove pointer or array type from TYPE. */ |
| tree |
| strip_pointer_or_array_types (tree t) |
| { |
| while (TREE_CODE (t) == ARRAY_TYPE || POINTER_TYPE_P (t)) |
| t = TREE_TYPE (t); |
| return t; |
| } |
| |
| /* Used to compare case labels. K1 and K2 are actually tree nodes |
| representing case labels, or NULL_TREE for a `default' label. |
| Returns -1 if K1 is ordered before K2, -1 if K1 is ordered after |
| K2, and 0 if K1 and K2 are equal. */ |
| |
| int |
| case_compare (splay_tree_key k1, splay_tree_key k2) |
| { |
| /* Consider a NULL key (such as arises with a `default' label) to be |
| smaller than anything else. */ |
| if (!k1) |
| return k2 ? -1 : 0; |
| else if (!k2) |
| return k1 ? 1 : 0; |
| |
| return tree_int_cst_compare ((tree) k1, (tree) k2); |
| } |
| |
| /* Process a case label, located at LOC, for the range LOW_VALUE |
| ... HIGH_VALUE. If LOW_VALUE and HIGH_VALUE are both NULL_TREE |
| then this case label is actually a `default' label. If only |
| HIGH_VALUE is NULL_TREE, then case label was declared using the |
| usual C/C++ syntax, rather than the GNU case range extension. |
| CASES is a tree containing all the case ranges processed so far; |
| COND is the condition for the switch-statement itself. |
| OUTSIDE_RANGE_P says whether there was a case value that doesn't |
| fit into the range of the ORIG_TYPE. Returns the CASE_LABEL_EXPR |
| created, or ERROR_MARK_NODE if no CASE_LABEL_EXPR is created. */ |
| |
| tree |
| c_add_case_label (location_t loc, splay_tree cases, tree cond, tree orig_type, |
| tree low_value, tree high_value, bool *outside_range_p) |
| { |
| tree type; |
| tree label; |
| tree case_label; |
| splay_tree_node node; |
| |
| /* Create the LABEL_DECL itself. */ |
| label = create_artificial_label (loc); |
| |
| /* If there was an error processing the switch condition, bail now |
| before we get more confused. */ |
| if (!cond || cond == error_mark_node) |
| goto error_out; |
| |
| if ((low_value && TREE_TYPE (low_value) |
| && POINTER_TYPE_P (TREE_TYPE (low_value))) |
| || (high_value && TREE_TYPE (high_value) |
| && POINTER_TYPE_P (TREE_TYPE (high_value)))) |
| { |
| error_at (loc, "pointers are not permitted as case values"); |
| goto error_out; |
| } |
| |
| /* Case ranges are a GNU extension. */ |
| if (high_value) |
| pedwarn (loc, OPT_Wpedantic, |
| "range expressions in switch statements are non-standard"); |
| |
| type = TREE_TYPE (cond); |
| if (low_value) |
| { |
| low_value = check_case_value (loc, low_value); |
| low_value = convert_and_check (loc, type, low_value); |
| if (low_value == error_mark_node) |
| goto error_out; |
| } |
| if (high_value) |
| { |
| high_value = check_case_value (loc, high_value); |
| high_value = convert_and_check (loc, type, high_value); |
| if (high_value == error_mark_node) |
| goto error_out; |
| } |
| |
| if (low_value && high_value) |
| { |
| /* If the LOW_VALUE and HIGH_VALUE are the same, then this isn't |
| really a case range, even though it was written that way. |
| Remove the HIGH_VALUE to simplify later processing. */ |
| if (tree_int_cst_equal (low_value, high_value)) |
| high_value = NULL_TREE; |
| else if (!tree_int_cst_lt (low_value, high_value)) |
| warning_at (loc, 0, "empty range specified"); |
| } |
| |
| /* See if the case is in range of the type of the original testing |
| expression. If both low_value and high_value are out of range, |
| don't insert the case label and return NULL_TREE. */ |
| if (low_value |
| && !check_case_bounds (loc, type, orig_type, |
| &low_value, high_value ? &high_value : NULL, |
| outside_range_p)) |
| return NULL_TREE; |
| |
| /* Look up the LOW_VALUE in the table of case labels we already |
| have. */ |
| node = splay_tree_lookup (cases, (splay_tree_key) low_value); |
| /* If there was not an exact match, check for overlapping ranges. |
| There's no need to do this if there's no LOW_VALUE or HIGH_VALUE; |
| that's a `default' label and the only overlap is an exact match. */ |
| if (!node && (low_value || high_value)) |
| { |
| splay_tree_node low_bound; |
| splay_tree_node high_bound; |
| |
| /* Even though there wasn't an exact match, there might be an |
| overlap between this case range and another case range. |
| Since we've (inductively) not allowed any overlapping case |
| ranges, we simply need to find the greatest low case label |
| that is smaller that LOW_VALUE, and the smallest low case |
| label that is greater than LOW_VALUE. If there is an overlap |
| it will occur in one of these two ranges. */ |
| low_bound = splay_tree_predecessor (cases, |
| (splay_tree_key) low_value); |
| high_bound = splay_tree_successor (cases, |
| (splay_tree_key) low_value); |
| |
| /* Check to see if the LOW_BOUND overlaps. It is smaller than |
| the LOW_VALUE, so there is no need to check unless the |
| LOW_BOUND is in fact itself a case range. */ |
| if (low_bound |
| && CASE_HIGH ((tree) low_bound->value) |
| && tree_int_cst_compare (CASE_HIGH ((tree) low_bound->value), |
| low_value) >= 0) |
| node = low_bound; |
| /* Check to see if the HIGH_BOUND overlaps. The low end of that |
| range is bigger than the low end of the current range, so we |
| are only interested if the current range is a real range, and |
| not an ordinary case label. */ |
| else if (high_bound |
| && high_value |
| && (tree_int_cst_compare ((tree) high_bound->key, |
| high_value) |
| <= 0)) |
| node = high_bound; |
| } |
| /* If there was an overlap, issue an error. */ |
| if (node) |
| { |
| tree duplicate = CASE_LABEL ((tree) node->value); |
| |
| if (high_value) |
| { |
| error_at (loc, "duplicate (or overlapping) case value"); |
| inform (DECL_SOURCE_LOCATION (duplicate), |
| "this is the first entry overlapping that value"); |
| } |
| else if (low_value) |
| { |
| error_at (loc, "duplicate case value") ; |
| inform (DECL_SOURCE_LOCATION (duplicate), "previously used here"); |
| } |
| else |
| { |
| error_at (loc, "multiple default labels in one switch"); |
| inform (DECL_SOURCE_LOCATION (duplicate), |
| "this is the first default label"); |
| } |
| goto error_out; |
| } |
| |
| /* Add a CASE_LABEL to the statement-tree. */ |
| case_label = add_stmt (build_case_label (low_value, high_value, label)); |
| /* Register this case label in the splay tree. */ |
| splay_tree_insert (cases, |
| (splay_tree_key) low_value, |
| (splay_tree_value) case_label); |
| |
| return case_label; |
| |
| error_out: |
| /* Add a label so that the back-end doesn't think that the beginning of |
| the switch is unreachable. Note that we do not add a case label, as |
| that just leads to duplicates and thence to failure later on. */ |
| if (!cases->root) |
| { |
| tree t = create_artificial_label (loc); |
| add_stmt (build_stmt (loc, LABEL_EXPR, t)); |
| } |
| return error_mark_node; |
| } |
| |
| /* Finish an expression taking the address of LABEL (an |
| IDENTIFIER_NODE). Returns an expression for the address. |
| |
| LOC is the location for the expression returned. */ |
| |
| tree |
| finish_label_address_expr (tree label, location_t loc) |
| { |
| tree result; |
| |
| pedwarn (input_location, OPT_Wpedantic, "taking the address of a label is non-standard"); |
| |
| if (label == error_mark_node) |
| return error_mark_node; |
| |
| label = lookup_label (label); |
| if (label == NULL_TREE) |
| result = null_pointer_node; |
| else |
| { |
| TREE_USED (label) = 1; |
| result = build1 (ADDR_EXPR, ptr_type_node, label); |
| /* The current function is not necessarily uninlinable. |
| Computed gotos are incompatible with inlining, but the value |
| here could be used only in a diagnostic, for example. */ |
| protected_set_expr_location (result, loc); |
| } |
| |
| return result; |
| } |
| |
| |
| /* Given a boolean expression ARG, return a tree representing an increment |
| or decrement (as indicated by CODE) of ARG. The front end must check for |
| invalid cases (e.g., decrement in C++). */ |
| tree |
| boolean_increment (enum tree_code code, tree arg) |
| { |
| tree val; |
| tree true_res = build_int_cst (TREE_TYPE (arg), 1); |
| |
| arg = stabilize_reference (arg); |
| switch (code) |
| { |
| case PREINCREMENT_EXPR: |
| val = build2 (MODIFY_EXPR, TREE_TYPE (arg), arg, true_res); |
| break; |
| case POSTINCREMENT_EXPR: |
| val = build2 (MODIFY_EXPR, TREE_TYPE (arg), arg, true_res); |
| arg = save_expr (arg); |
| val = build2 (COMPOUND_EXPR, TREE_TYPE (arg), val, arg); |
| val = build2 (COMPOUND_EXPR, TREE_TYPE (arg), arg, val); |
| break; |
| case PREDECREMENT_EXPR: |
| val = build2 (MODIFY_EXPR, TREE_TYPE (arg), arg, |
| invert_truthvalue_loc (input_location, arg)); |
| break; |
| case POSTDECREMENT_EXPR: |
| val = build2 (MODIFY_EXPR, TREE_TYPE (arg), arg, |
| invert_truthvalue_loc (input_location, arg)); |
| arg = save_expr (arg); |
| val = build2 (COMPOUND_EXPR, TREE_TYPE (arg), val, arg); |
| val = build2 (COMPOUND_EXPR, TREE_TYPE (arg), arg, val); |
| break; |
| default: |
| gcc_unreachable (); |
| } |
| TREE_SIDE_EFFECTS (val) = 1; |
| return val; |
| } |
| |
| /* Built-in macros for stddef.h and stdint.h, that require macros |
| defined in this file. */ |
| void |
| c_stddef_cpp_builtins(void) |
| { |
| builtin_define_with_value ("__SIZE_TYPE__", SIZE_TYPE, 0); |
| builtin_define_with_value ("__PTRDIFF_TYPE__", PTRDIFF_TYPE, 0); |
| builtin_define_with_value ("__WCHAR_TYPE__", MODIFIED_WCHAR_TYPE, 0); |
| builtin_define_with_value ("__WINT_TYPE__", WINT_TYPE, 0); |
| builtin_define_with_value ("__INTMAX_TYPE__", INTMAX_TYPE, 0); |
| builtin_define_with_value ("__UINTMAX_TYPE__", UINTMAX_TYPE, 0); |
| builtin_define_with_value ("__CHAR16_TYPE__", CHAR16_TYPE, 0); |
| builtin_define_with_value ("__CHAR32_TYPE__", CHAR32_TYPE, 0); |
| if (SIG_ATOMIC_TYPE) |
| builtin_define_with_value ("__SIG_ATOMIC_TYPE__", SIG_ATOMIC_TYPE, 0); |
| if (INT8_TYPE) |
| builtin_define_with_value ("__INT8_TYPE__", INT8_TYPE, 0); |
| if (INT16_TYPE) |
| builtin_define_with_value ("__INT16_TYPE__", INT16_TYPE, 0); |
| if (INT32_TYPE) |
| builtin_define_with_value ("__INT32_TYPE__", INT32_TYPE, 0); |
| if (INT64_TYPE) |
| builtin_define_with_value ("__INT64_TYPE__", INT64_TYPE, 0); |
| if (UINT8_TYPE) |
| builtin_define_with_value ("__UINT8_TYPE__", UINT8_TYPE, 0); |
| if (UINT16_TYPE) |
| builtin_define_with_value ("__UINT16_TYPE__", UINT16_TYPE, 0); |
| if (UINT32_TYPE) |
| builtin_define_with_value ("__UINT32_TYPE__", UINT32_TYPE, 0); |
| if (UINT64_TYPE) |
| builtin_define_with_value ("__UINT64_TYPE__", UINT64_TYPE, 0); |
| if (INT_LEAST8_TYPE) |
| builtin_define_with_value ("__INT_LEAST8_TYPE__", INT_LEAST8_TYPE, 0); |
| if (INT_LEAST16_TYPE) |
| builtin_define_with_value ("__INT_LEAST16_TYPE__", INT_LEAST16_TYPE, 0); |
| if (INT_LEAST32_TYPE) |
| builtin_define_with_value ("__INT_LEAST32_TYPE__", INT_LEAST32_TYPE, 0); |
| if (INT_LEAST64_TYPE) |
| builtin_define_with_value ("__INT_LEAST64_TYPE__", INT_LEAST64_TYPE, 0); |
| if (UINT_LEAST8_TYPE) |
| builtin_define_with_value ("__UINT_LEAST8_TYPE__", UINT_LEAST8_TYPE, 0); |
| if (UINT_LEAST16_TYPE) |
| builtin_define_with_value ("__UINT_LEAST16_TYPE__", UINT_LEAST16_TYPE, 0); |
| if (UINT_LEAST32_TYPE) |
| builtin_define_with_value ("__UINT_LEAST32_TYPE__", UINT_LEAST32_TYPE, 0); |
| if (UINT_LEAST64_TYPE) |
| builtin_define_with_value ("__UINT_LEAST64_TYPE__", UINT_LEAST64_TYPE, 0); |
| if (INT_FAST8_TYPE) |
| builtin_define_with_value ("__INT_FAST8_TYPE__", INT_FAST8_TYPE, 0); |
| if (INT_FAST16_TYPE) |
| builtin_define_with_value ("__INT_FAST16_TYPE__", INT_FAST16_TYPE, 0); |
| if (INT_FAST32_TYPE) |
| builtin_define_with_value ("__INT_FAST32_TYPE__", INT_FAST32_TYPE, 0); |
| if (INT_FAST64_TYPE) |
| builtin_define_with_value ("__INT_FAST64_TYPE__", INT_FAST64_TYPE, 0); |
| if (UINT_FAST8_TYPE) |
| builtin_define_with_value ("__UINT_FAST8_TYPE__", UINT_FAST8_TYPE, 0); |
| if (UINT_FAST16_TYPE) |
| builtin_define_with_value ("__UINT_FAST16_TYPE__", UINT_FAST16_TYPE, 0); |
| if (UINT_FAST32_TYPE) |
| builtin_define_with_value ("__UINT_FAST32_TYPE__", UINT_FAST32_TYPE, 0); |
| if (UINT_FAST64_TYPE) |
| builtin_define_with_value ("__UINT_FAST64_TYPE__", UINT_FAST64_TYPE, 0); |
| if (INTPTR_TYPE) |
| builtin_define_with_value ("__INTPTR_TYPE__", INTPTR_TYPE, 0); |
| if (UINTPTR_TYPE) |
| builtin_define_with_value ("__UINTPTR_TYPE__", UINTPTR_TYPE, 0); |
| } |
| |
| static void |
| c_init_attributes (void) |
| { |
| /* Fill in the built_in_attributes array. */ |
| #define DEF_ATTR_NULL_TREE(ENUM) \ |
| built_in_attributes[(int) ENUM] = NULL_TREE; |
| #define DEF_ATTR_INT(ENUM, VALUE) \ |
| built_in_attributes[(int) ENUM] = build_int_cst (integer_type_node, VALUE); |
| #define DEF_ATTR_STRING(ENUM, VALUE) \ |
| built_in_attributes[(int) ENUM] = build_string (strlen (VALUE), VALUE); |
| #define DEF_ATTR_IDENT(ENUM, STRING) \ |
| built_in_attributes[(int) ENUM] = get_identifier (STRING); |
| #define DEF_ATTR_TREE_LIST(ENUM, PURPOSE, VALUE, CHAIN) \ |
| built_in_attributes[(int) ENUM] \ |
| = tree_cons (built_in_attributes[(int) PURPOSE], \ |
| built_in_attributes[(int) VALUE], \ |
| built_in_attributes[(int) CHAIN]); |
| #include "builtin-attrs.def" |
| #undef DEF_ATTR_NULL_TREE |
| #undef DEF_ATTR_INT |
| #undef DEF_ATTR_IDENT |
| #undef DEF_ATTR_TREE_LIST |
| } |
| |
| /* Check whether ALIGN is a valid user-specified alignment. If so, |
| return its base-2 log; if not, output an error and return -1. If |
| ALLOW_ZERO then 0 is valid and should result in a return of -1 with |
| no error. */ |
| int |
| check_user_alignment (const_tree align, bool allow_zero) |
| { |
| int i; |
| |
| if (error_operand_p (align)) |
| return -1; |
| if (TREE_CODE (align) != INTEGER_CST |
| || !INTEGRAL_TYPE_P (TREE_TYPE (align))) |
| { |
| error ("requested alignment is not an integer constant"); |
| return -1; |
| } |
| else if (allow_zero && integer_zerop (align)) |
| return -1; |
| else if (tree_int_cst_sgn (align) == -1 |
| || (i = tree_log2 (align)) == -1) |
| { |
| error ("requested alignment is not a positive power of 2"); |
| return -1; |
| } |
| else if (i >= HOST_BITS_PER_INT - LOG2_BITS_PER_UNIT) |
| { |
| error ("requested alignment is too large"); |
| return -1; |
| } |
| return i; |
| } |
| |
| /* Determine the ELF symbol visibility for DECL, which is either a |
| variable or a function. It is an error to use this function if a |
| definition of DECL is not available in this translation unit. |
| Returns true if the final visibility has been determined by this |
| function; false if the caller is free to make additional |
| modifications. */ |
| |
| bool |
| c_determine_visibility (tree decl) |
| { |
| gcc_assert (VAR_OR_FUNCTION_DECL_P (decl)); |
| |
| /* If the user explicitly specified the visibility with an |
| attribute, honor that. DECL_VISIBILITY will have been set during |
| the processing of the attribute. We check for an explicit |
| attribute, rather than just checking DECL_VISIBILITY_SPECIFIED, |
| to distinguish the use of an attribute from the use of a "#pragma |
| GCC visibility push(...)"; in the latter case we still want other |
| considerations to be able to overrule the #pragma. */ |
| if (lookup_attribute ("visibility", DECL_ATTRIBUTES (decl)) |
| || (TARGET_DLLIMPORT_DECL_ATTRIBUTES |
| && (lookup_attribute ("dllimport", DECL_ATTRIBUTES (decl)) |
| || lookup_attribute ("dllexport", DECL_ATTRIBUTES (decl))))) |
| return true; |
| |
| /* Set default visibility to whatever the user supplied with |
| visibility_specified depending on #pragma GCC visibility. */ |
| if (!DECL_VISIBILITY_SPECIFIED (decl)) |
| { |
| if (visibility_options.inpragma |
| || DECL_VISIBILITY (decl) != default_visibility) |
| { |
| DECL_VISIBILITY (decl) = default_visibility; |
| DECL_VISIBILITY_SPECIFIED (decl) = visibility_options.inpragma; |
| /* If visibility changed and DECL already has DECL_RTL, ensure |
| symbol flags are updated. */ |
| if (((VAR_P (decl) && TREE_STATIC (decl)) |
| || TREE_CODE (decl) == FUNCTION_DECL) |
| && DECL_RTL_SET_P (decl)) |
| make_decl_rtl (decl); |
| } |
| } |
| return false; |
| } |
| |
| /* Data to communicate through check_function_arguments_recurse between |
| check_function_nonnull and check_nonnull_arg. */ |
| |
| struct nonnull_arg_ctx |
| { |
| location_t loc; |
| bool warned_p; |
| }; |
| |
| /* Check the argument list of a function call for null in argument slots |
| that are marked as requiring a non-null pointer argument. The NARGS |
| arguments are passed in the array ARGARRAY. Return true if we have |
| warned. */ |
| |
| static bool |
| check_function_nonnull (location_t loc, tree attrs, int nargs, tree *argarray) |
| { |
| tree a; |
| int i; |
| |
| attrs = lookup_attribute ("nonnull", attrs); |
| if (attrs == NULL_TREE) |
| return false; |
| |
| a = attrs; |
| /* See if any of the nonnull attributes has no arguments. If so, |
| then every pointer argument is checked (in which case the check |
| for pointer type is done in check_nonnull_arg). */ |
| if (TREE_VALUE (a) != NULL_TREE) |
| do |
| a = lookup_attribute ("nonnull", TREE_CHAIN (a)); |
| while (a != NULL_TREE && TREE_VALUE (a) != NULL_TREE); |
| |
| struct nonnull_arg_ctx ctx = { loc, false }; |
| if (a != NULL_TREE) |
| for (i = 0; i < nargs; i++) |
| check_function_arguments_recurse (check_nonnull_arg, &ctx, argarray[i], |
| i + 1); |
| else |
| { |
| /* Walk the argument list. If we encounter an argument number we |
| should check for non-null, do it. */ |
| for (i = 0; i < nargs; i++) |
| { |
| for (a = attrs; ; a = TREE_CHAIN (a)) |
| { |
| a = lookup_attribute ("nonnull", a); |
| if (a == NULL_TREE || nonnull_check_p (TREE_VALUE (a), i + 1)) |
| break; |
| } |
| |
| if (a != NULL_TREE) |
| check_function_arguments_recurse (check_nonnull_arg, &ctx, |
| argarray[i], i + 1); |
| } |
| } |
| return ctx.warned_p; |
| } |
| |
| /* Check that the Nth argument of a function call (counting backwards |
| from the end) is a (pointer)0. The NARGS arguments are passed in the |
| array ARGARRAY. */ |
| |
| static void |
| check_function_sentinel (const_tree fntype, int nargs, tree *argarray) |
| { |
| tree attr = lookup_attribute ("sentinel", TYPE_ATTRIBUTES (fntype)); |
| |
| if (attr) |
| { |
| int len = 0; |
| int pos = 0; |
| tree sentinel; |
| function_args_iterator iter; |
| tree t; |
| |
| /* Skip over the named arguments. */ |
| FOREACH_FUNCTION_ARGS (fntype, t, iter) |
| { |
| if (len == nargs) |
| break; |
| len++; |
| } |
| |
| if (TREE_VALUE (attr)) |
| { |
| tree p = TREE_VALUE (TREE_VALUE (attr)); |
| pos = TREE_INT_CST_LOW (p); |
| } |
| |
| /* The sentinel must be one of the varargs, i.e. |
| in position >= the number of fixed arguments. */ |
| if ((nargs - 1 - pos) < len) |
| { |
| warning (OPT_Wformat_, |
| "not enough variable arguments to fit a sentinel"); |
| return; |
| } |
| |
| /* Validate the sentinel. */ |
| sentinel = argarray[nargs - 1 - pos]; |
| if ((!POINTER_TYPE_P (TREE_TYPE (sentinel)) |
| || !integer_zerop (sentinel)) |
| /* Although __null (in C++) is only an integer we allow it |
| nevertheless, as we are guaranteed that it's exactly |
| as wide as a pointer, and we don't want to force |
| users to cast the NULL they have written there. |
| We warn with -Wstrict-null-sentinel, though. */ |
| && (warn_strict_null_sentinel || null_node != sentinel)) |
| warning (OPT_Wformat_, "missing sentinel in function call"); |
| } |
| } |
| |
| /* Check that the same argument isn't passed to restrict arguments |
| and other arguments. */ |
| |
| static void |
| check_function_restrict (const_tree fndecl, const_tree fntype, |
| int nargs, tree *argarray) |
| { |
| int i; |
| tree parms; |
| |
| if (fndecl |
| && TREE_CODE (fndecl) == FUNCTION_DECL |
| && DECL_ARGUMENTS (fndecl)) |
| parms = DECL_ARGUMENTS (fndecl); |
| else |
| parms = TYPE_ARG_TYPES (fntype); |
| |
| for (i = 0; i < nargs; i++) |
| TREE_VISITED (argarray[i]) = 0; |
| |
| for (i = 0; i < nargs && parms && parms != void_list_node; i++) |
| { |
| tree type; |
| if (TREE_CODE (parms) == PARM_DECL) |
| { |
| type = TREE_TYPE (parms); |
| parms = DECL_CHAIN (parms); |
| } |
| else |
| { |
| type = TREE_VALUE (parms); |
| parms = TREE_CHAIN (parms); |
| } |
| if (POINTER_TYPE_P (type) |
| && TYPE_RESTRICT (type) |
| && !TYPE_READONLY (TREE_TYPE (type))) |
| warn_for_restrict (i, argarray, nargs); |
| } |
| |
| for (i = 0; i < nargs; i++) |
| TREE_VISITED (argarray[i]) = 0; |
| } |
| |
| /* Helper for check_function_nonnull; given a list of operands which |
| must be non-null in ARGS, determine if operand PARAM_NUM should be |
| checked. */ |
| |
| static bool |
| nonnull_check_p (tree args, unsigned HOST_WIDE_INT param_num) |
| { |
| unsigned HOST_WIDE_INT arg_num = 0; |
| |
| for (; args; args = TREE_CHAIN (args)) |
| { |
| bool found = get_nonnull_operand (TREE_VALUE (args), &arg_num); |
| |
| gcc_assert (found); |
| |
| if (arg_num == param_num) |
| return true; |
| } |
| return false; |
| } |
| |
| /* Check that the function argument PARAM (which is operand number |
| PARAM_NUM) is non-null. This is called by check_function_nonnull |
| via check_function_arguments_recurse. */ |
| |
| static void |
| check_nonnull_arg (void *ctx, tree param, unsigned HOST_WIDE_INT param_num) |
| { |
| struct nonnull_arg_ctx *pctx = (struct nonnull_arg_ctx *) ctx; |
| |
| /* Just skip checking the argument if it's not a pointer. This can |
| happen if the "nonnull" attribute was given without an operand |
| list (which means to check every pointer argument). */ |
| |
| if (TREE_CODE (TREE_TYPE (param)) != POINTER_TYPE) |
| return; |
| |
| /* Diagnose the simple cases of null arguments. */ |
| if (integer_zerop (fold_for_warn (param))) |
| { |
| warning_at (pctx->loc, OPT_Wnonnull, "null argument where non-null " |
| "required (argument %lu)", (unsigned long) param_num); |
| pctx->warned_p = true; |
| } |
| } |
| |
| /* Helper for nonnull attribute handling; fetch the operand number |
| from the attribute argument list. */ |
| |
| bool |
| get_nonnull_operand (tree arg_num_expr, unsigned HOST_WIDE_INT *valp) |
| { |
| /* Verify the arg number is a small constant. */ |
| if (tree_fits_uhwi_p (arg_num_expr)) |
| { |
| *valp = TREE_INT_CST_LOW (arg_num_expr); |
| return true; |
| } |
| else |
| return false; |
| } |
| |
| /* Arguments being collected for optimization. */ |
| typedef const char *const_char_p; /* For DEF_VEC_P. */ |
| static GTY(()) vec<const_char_p, va_gc> *optimize_args; |
| |
| |
| /* Inner function to convert a TREE_LIST to argv string to parse the optimize |
| options in ARGS. ATTR_P is true if this is for attribute(optimize), and |
| false for #pragma GCC optimize. */ |
| |
| bool |
| parse_optimize_options (tree args, bool attr_p) |
| { |
| bool ret = true; |
| unsigned opt_argc; |
| unsigned i; |
| const char **opt_argv; |
| struct cl_decoded_option *decoded_options; |
| unsigned int decoded_options_count; |
| tree ap; |
| |
| /* Build up argv vector. Just in case the string is stored away, use garbage |
| collected strings. */ |
| vec_safe_truncate (optimize_args, 0); |
| vec_safe_push (optimize_args, (const char *) NULL); |
| |
| for (ap = args; ap != NULL_TREE; ap = TREE_CHAIN (ap)) |
| { |
| tree value = TREE_VALUE (ap); |
| |
| if (TREE_CODE (value) == INTEGER_CST) |
| { |
| char buffer[20]; |
| sprintf (buffer, "-O%ld", (long) TREE_INT_CST_LOW (value)); |
| vec_safe_push (optimize_args, ggc_strdup (buffer)); |
| } |
| |
| else if (TREE_CODE (value) == STRING_CST) |
| { |
| /* Split string into multiple substrings. */ |
| size_t len = TREE_STRING_LENGTH (value); |
| char *p = ASTRDUP (TREE_STRING_POINTER (value)); |
| char *end = p + len; |
| char *comma; |
| char *next_p = p; |
| |
| while (next_p != NULL) |
| { |
| size_t len2; |
| char *q, *r; |
| |
| p = next_p; |
| comma = strchr (p, ','); |
| if (comma) |
| { |
| len2 = comma - p; |
| *comma = '\0'; |
| next_p = comma+1; |
| } |
| else |
| { |
| len2 = end - p; |
| next_p = NULL; |
| } |
| |
| r = q = (char *) ggc_alloc_atomic (len2 + 3); |
| |
| /* If the user supplied -Oxxx or -fxxx, only allow -Oxxx or -fxxx |
| options. */ |
| if (*p == '-' && p[1] != 'O' && p[1] != 'f') |
| { |
| ret = false; |
| if (attr_p) |
| warning (OPT_Wattributes, |
| "bad option %qs to attribute %<optimize%>", p); |
| else |
| warning (OPT_Wpragmas, |
| "bad option %qs to pragma %<optimize%>", p); |
| continue; |
| } |
| |
| if (*p != '-') |
| { |
| *r++ = '-'; |
| |
| /* Assume that Ox is -Ox, a numeric value is -Ox, a s by |
| itself is -Os, and any other switch begins with a -f. */ |
| if ((*p >= '0' && *p <= '9') |
| || (p[0] == 's' && p[1] == '\0')) |
| *r++ = 'O'; |
| else if (*p != 'O') |
| *r++ = 'f'; |
| } |
| |
| memcpy (r, p, len2); |
| r[len2] = '\0'; |
| vec_safe_push (optimize_args, (const char *) q); |
| } |
| |
| } |
| } |
| |
| opt_argc = optimize_args->length (); |
| opt_argv = (const char **) alloca (sizeof (char *) * (opt_argc + 1)); |
| |
| for (i = 1; i < opt_argc; i++) |
| opt_argv[i] = (*optimize_args)[i]; |
| |
| /* Now parse the options. */ |
| decode_cmdline_options_to_array_default_mask (opt_argc, opt_argv, |
| &decoded_options, |
| &decoded_options_count); |
| /* Drop non-Optimization options. */ |
| unsigned j = 1; |
| for (i = 1; i < decoded_options_count; ++i) |
| { |
| if (! (cl_options[decoded_options[i].opt_index].flags & CL_OPTIMIZATION)) |
| { |
| ret = false; |
| if (attr_p) |
| warning (OPT_Wattributes, |
| "bad option %qs to attribute %<optimize%>", |
| decoded_options[i].orig_option_with_args_text); |
| else |
| warning (OPT_Wpragmas, |
| "bad option %qs to pragma %<optimize%>", |
| decoded_options[i].orig_option_with_args_text); |
| continue; |
| } |
| if (i != j) |
| decoded_options[j] = decoded_options[i]; |
| j++; |
| } |
| decoded_options_count = j; |
| /* And apply them. */ |
| decode_options (&global_options, &global_options_set, |
| decoded_options, decoded_options_count, |
| input_location, global_dc, NULL); |
| |
| targetm.override_options_after_change(); |
| |
| optimize_args->truncate (0); |
| return ret; |
| } |
| |
| /* Check whether ATTR is a valid attribute fallthrough. */ |
| |
| bool |
| attribute_fallthrough_p (tree attr) |
| { |
| if (attr == error_mark_node) |
| return false; |
| tree t = lookup_attribute ("fallthrough", attr); |
| if (t == NULL_TREE) |
| return false; |
| /* This attribute shall appear at most once in each attribute-list. */ |
| if (lookup_attribute ("fallthrough", TREE_CHAIN (t))) |
| warning (OPT_Wattributes, "%<fallthrough%> attribute specified multiple " |
| "times"); |
| /* No attribute-argument-clause shall be present. */ |
| else if (TREE_VALUE (t) != NULL_TREE) |
| warning (OPT_Wattributes, "%<fallthrough%> attribute specified with " |
| "a parameter"); |
| /* Warn if other attributes are found. */ |
| for (t = attr; t != NULL_TREE; t = TREE_CHAIN (t)) |
| { |
| tree name = get_attribute_name (t); |
| if (!is_attribute_p ("fallthrough", name)) |
| warning (OPT_Wattributes, "%qE attribute ignored", name); |
| } |
| return true; |
| } |
| |
| |
| /* Check for valid arguments being passed to a function with FNTYPE. |
| There are NARGS arguments in the array ARGARRAY. LOC should be used for |
| diagnostics. Return true if -Wnonnull warning has been diagnosed. */ |
| bool |
| check_function_arguments (location_t loc, const_tree fndecl, const_tree fntype, |
| int nargs, tree *argarray) |
| { |
| bool warned_p = false; |
| |
| /* Check for null being passed in a pointer argument that must be |
| non-null. We also need to do this if format checking is enabled. */ |
| |
| if (warn_nonnull) |
| warned_p = check_function_nonnull (loc, TYPE_ATTRIBUTES (fntype), |
| nargs, argarray); |
| |
| /* Check for errors in format strings. */ |
| |
| if (warn_format || warn_suggest_attribute_format) |
| check_function_format (TYPE_ATTRIBUTES (fntype), nargs, argarray); |
| |
| if (warn_format) |
| check_function_sentinel (fntype, nargs, argarray); |
| |
| if (warn_restrict) |
| check_function_restrict (fndecl, fntype, nargs, argarray); |
| return warned_p; |
| } |
| |
| /* Generic argument checking recursion routine. PARAM is the argument to |
| be checked. PARAM_NUM is the number of the argument. CALLBACK is invoked |
| once the argument is resolved. CTX is context for the callback. */ |
| void |
| check_function_arguments_recurse (void (*callback) |
| (void *, tree, unsigned HOST_WIDE_INT), |
| void *ctx, tree param, |
| unsigned HOST_WIDE_INT param_num) |
| { |
| if (CONVERT_EXPR_P (param) |
| && (TYPE_PRECISION (TREE_TYPE (param)) |
| == TYPE_PRECISION (TREE_TYPE (TREE_OPERAND (param, 0))))) |
| { |
| /* Strip coercion. */ |
| check_function_arguments_recurse (callback, ctx, |
| TREE_OPERAND (param, 0), param_num); |
| return; |
| } |
| |
| if (TREE_CODE (param) == CALL_EXPR) |
| { |
| tree type = TREE_TYPE (TREE_TYPE (CALL_EXPR_FN (param))); |
| tree attrs; |
| bool found_format_arg = false; |
| |
| /* See if this is a call to a known internationalization function |
| that modifies a format arg. Such a function may have multiple |
| format_arg attributes (for example, ngettext). */ |
| |
| for (attrs = TYPE_ATTRIBUTES (type); |
| attrs; |
| attrs = TREE_CHAIN (attrs)) |
| if (is_attribute_p ("format_arg", TREE_PURPOSE (attrs))) |
| { |
| tree inner_arg; |
| tree format_num_expr; |
| int format_num; |
| int i; |
| call_expr_arg_iterator iter; |
| |
| /* Extract the argument number, which was previously checked |
| to be valid. */ |
| format_num_expr = TREE_VALUE (TREE_VALUE (attrs)); |
| |
| format_num = tree_to_uhwi (format_num_expr); |
| |
| for (inner_arg = first_call_expr_arg (param, &iter), i = 1; |
| inner_arg != 0; |
| inner_arg = next_call_expr_arg (&iter), i++) |
| if (i == format_num) |
| { |
| check_function_arguments_recurse (callback, ctx, |
| inner_arg, param_num); |
| found_format_arg = true; |
| break; |
| } |
| } |
| |
| /* If we found a format_arg attribute and did a recursive check, |
| we are done with checking this argument. Otherwise, we continue |
| and this will be considered a non-literal. */ |
| if (found_format_arg) |
| return; |
| } |
| |
| if (TREE_CODE (param) == COND_EXPR) |
| { |
| /* Simplify to avoid warning for an impossible case. */ |
| param = fold_for_warn (param); |
| if (TREE_CODE (param) == COND_EXPR) |
| { |
| /* Check both halves of the conditional expression. */ |
| check_function_arguments_recurse (callback, ctx, |
| TREE_OPERAND (param, 1), |
| param_num); |
| check_function_arguments_recurse (callback, ctx, |
| TREE_OPERAND (param, 2), |
| param_num); |
| return; |
| } |
| } |
| |
| (*callback) (ctx, param, param_num); |
| } |
| |
| /* Checks for a builtin function FNDECL that the number of arguments |
| NARGS against the required number REQUIRED and issues an error if |
| there is a mismatch. Returns true if the number of arguments is |
| correct, otherwise false. LOC is the location of FNDECL. */ |
| |
| static bool |
| builtin_function_validate_nargs (location_t loc, tree fndecl, int nargs, |
| int required) |
| { |
| if (nargs < required) |
| { |
| error_at (loc, "too few arguments to function %qE", fndecl); |
| return false; |
| } |
| else if (nargs > required) |
| { |
| error_at (loc, "too many arguments to function %qE", fndecl); |
| return false; |
| } |
| return true; |
| } |
| |
| /* Helper macro for check_builtin_function_arguments. */ |
| #define ARG_LOCATION(N) \ |
| (arg_loc.is_empty () \ |
| ? EXPR_LOC_OR_LOC (args[(N)], input_location) \ |
| : expansion_point_location (arg_loc[(N)])) |
| |
| /* Verifies the NARGS arguments ARGS to the builtin function FNDECL. |
| Returns false if there was an error, otherwise true. LOC is the |
| location of the function; ARG_LOC is a vector of locations of the |
| arguments. */ |
| |
| bool |
| check_builtin_function_arguments (location_t loc, vec<location_t> arg_loc, |
| tree fndecl, int nargs, tree *args) |
| { |
| if (!DECL_BUILT_IN (fndecl) |
| || DECL_BUILT_IN_CLASS (fndecl) != BUILT_IN_NORMAL) |
| return true; |
| |
| switch (DECL_FUNCTION_CODE (fndecl)) |
| { |
| case BUILT_IN_ALLOCA_WITH_ALIGN: |
| { |
| /* Get the requested alignment (in bits) if it's a constant |
| integer expression. */ |
| unsigned HOST_WIDE_INT align |
| = tree_fits_uhwi_p (args[1]) ? tree_to_uhwi (args[1]) : 0; |
| |
| /* Determine if the requested alignment is a power of 2. */ |
| if ((align & (align - 1))) |
| align = 0; |
| |
| /* The maximum alignment in bits corresponding to the same |
| maximum in bytes enforced in check_user_alignment(). */ |
| unsigned maxalign = (UINT_MAX >> 1) + 1; |
| |
| /* Reject invalid alignments. */ |
| if (align < BITS_PER_UNIT || maxalign < align) |
| { |
| error_at (ARG_LOCATION (1), |
| "second argument to function %qE must be a constant " |
| "integer power of 2 between %qi and %qu bits", |
| fndecl, BITS_PER_UNIT, maxalign); |
| return false; |
| } |
| return true; |
| } |
| |
| case BUILT_IN_CONSTANT_P: |
| return builtin_function_validate_nargs (loc, fndecl, nargs, 1); |
| |
| case BUILT_IN_ISFINITE: |
| case BUILT_IN_ISINF: |
| case BUILT_IN_ISINF_SIGN: |
| case BUILT_IN_ISNAN: |
| case BUILT_IN_ISNORMAL: |
| case BUILT_IN_SIGNBIT: |
| if (builtin_function_validate_nargs (loc, fndecl, nargs, 1)) |
| { |
| if (TREE_CODE (TREE_TYPE (args[0])) != REAL_TYPE) |
| { |
| error_at (ARG_LOCATION (0), "non-floating-point argument in " |
| "call to function %qE", fndecl); |
| return false; |
| } |
| return true; |
| } |
| return false; |
| |
| case BUILT_IN_ISGREATER: |
| case BUILT_IN_ISGREATEREQUAL: |
| case BUILT_IN_ISLESS: |
| case BUILT_IN_ISLESSEQUAL: |
| case BUILT_IN_ISLESSGREATER: |
| case BUILT_IN_ISUNORDERED: |
| if (builtin_function_validate_nargs (loc, fndecl, nargs, 2)) |
| { |
| enum tree_code code0, code1; |
| code0 = TREE_CODE (TREE_TYPE (args[0])); |
| code1 = TREE_CODE (TREE_TYPE (args[1])); |
| if (!((code0 == REAL_TYPE && code1 == REAL_TYPE) |
| || (code0 == REAL_TYPE && code1 == INTEGER_TYPE) |
| || (code0 == INTEGER_TYPE && code1 == REAL_TYPE))) |
| { |
| error_at (loc, "non-floating-point arguments in call to " |
| "function %qE", fndecl); |
| return false; |
| } |
| return true; |
| } |
| return false; |
| |
| case BUILT_IN_FPCLASSIFY: |
| if (builtin_function_validate_nargs (loc, fndecl, nargs, 6)) |
| { |
| for (unsigned int i = 0; i < 5; i++) |
| if (TREE_CODE (args[i]) != INTEGER_CST) |
| { |
| error_at (ARG_LOCATION (i), "non-const integer argument %u in " |
| "call to function %qE", i + 1, fndecl); |
| return false; |
| } |
| |
| if (TREE_CODE (TREE_TYPE (args[5])) != REAL_TYPE) |
| { |
| error_at (ARG_LOCATION (5), "non-floating-point argument in " |
| "call to function %qE", fndecl); |
| return false; |
| } |
| return true; |
| } |
| return false; |
| |
| case BUILT_IN_ASSUME_ALIGNED: |
| if (builtin_function_validate_nargs (loc, fndecl, nargs, 2 + (nargs > 2))) |
| { |
| if (nargs >= 3 && TREE_CODE (TREE_TYPE (args[2])) != INTEGER_TYPE) |
| { |
| error_at (ARG_LOCATION (2), "non-integer argument 3 in call to " |
| "function %qE", fndecl); |
| return false; |
| } |
| return true; |
| } |
| return false; |
| |
| case BUILT_IN_ADD_OVERFLOW: |
| case BUILT_IN_SUB_OVERFLOW: |
| case BUILT_IN_MUL_OVERFLOW: |
| if (builtin_function_validate_nargs (loc, fndecl, nargs, 3)) |
| { |
| unsigned i; |
| for (i = 0; i < 2; i++) |
| if (!INTEGRAL_TYPE_P (TREE_TYPE (args[i]))) |
| { |
| error_at (ARG_LOCATION (i), "argument %u in call to function " |
| "%qE does not have integral type", i + 1, fndecl); |
| return false; |
| } |
| if (TREE_CODE (TREE_TYPE (args[2])) != POINTER_TYPE |
| || !INTEGRAL_TYPE_P (TREE_TYPE (TREE_TYPE (args[2])))) |
| { |
| error_at (ARG_LOCATION (2), "argument 3 in call to function %qE " |
| "does not have pointer to integral type", fndecl); |
| return false; |
| } |
| else if (TREE_CODE (TREE_TYPE (TREE_TYPE (args[2]))) == ENUMERAL_TYPE) |
| { |
| error_at (ARG_LOCATION (2), "argument 3 in call to function %qE " |
| "has pointer to enumerated type", fndecl); |
| return false; |
| } |
| else if (TREE_CODE (TREE_TYPE (TREE_TYPE (args[2]))) == BOOLEAN_TYPE) |
| { |
| error_at (ARG_LOCATION (2), "argument 3 in call to function %qE " |
| "has pointer to boolean type", fndecl); |
| return false; |
| } |
| return true; |
| } |
| return false; |
| |
| case BUILT_IN_ADD_OVERFLOW_P: |
| case BUILT_IN_SUB_OVERFLOW_P: |
| case BUILT_IN_MUL_OVERFLOW_P: |
| if (builtin_function_validate_nargs (loc, fndecl, nargs, 3)) |
| { |
| unsigned i; |
| for (i = 0; i < 3; i++) |
| if (!INTEGRAL_TYPE_P (TREE_TYPE (args[i]))) |
| { |
| error_at (ARG_LOCATION (i), "argument %u in call to function " |
| "%qE does not have integral type", i + 1, fndecl); |
| return false; |
| } |
| if (TREE_CODE (TREE_TYPE (args[2])) == ENUMERAL_TYPE) |
| { |
| error_at (ARG_LOCATION (2), "argument 3 in call to function " |
| "%qE has enumerated type", fndecl); |
| return false; |
| } |
| else if (TREE_CODE (TREE_TYPE (args[2])) == BOOLEAN_TYPE) |
| { |
| error_at (ARG_LOCATION (2), "argument 3 in call to function " |
| "%qE has boolean type", fndecl); |
| return false; |
| } |
| return true; |
| } |
| return false; |
| |
| default: |
| return true; |
| } |
| } |
| |
| /* Function to help qsort sort FIELD_DECLs by name order. */ |
| |
| int |
| field_decl_cmp (const void *x_p, const void *y_p) |
| { |
| const tree *const x = (const tree *const) x_p; |
| const tree *const y = (const tree *const) y_p; |
| |
| if (DECL_NAME (*x) == DECL_NAME (*y)) |
| /* A nontype is "greater" than a type. */ |
| return (TREE_CODE (*y) == TYPE_DECL) - (TREE_CODE (*x) == TYPE_DECL); |
| if (DECL_NAME (*x) == NULL_TREE) |
| return -1; |
| if (DECL_NAME (*y) == NULL_TREE) |
| return 1; |
| if (DECL_NAME (*x) < DECL_NAME (*y)) |
| return -1; |
| return 1; |
| } |
| |
| static struct { |
| gt_pointer_operator new_value; |
| void *cookie; |
| } resort_data; |
| |
| /* This routine compares two fields like field_decl_cmp but using the |
| pointer operator in resort_data. */ |
| |
| static int |
| resort_field_decl_cmp (const void *x_p, const void *y_p) |
| { |
| const tree *const x = (const tree *const) x_p; |
| const tree *const y = (const tree *const) y_p; |
| |
| if (DECL_NAME (*x) == DECL_NAME (*y)) |
| /* A nontype is "greater" than a type. */ |
| return (TREE_CODE (*y) == TYPE_DECL) - (TREE_CODE (*x) == TYPE_DECL); |
| if (DECL_NAME (*x) == NULL_TREE) |
| return -1; |
| if (DECL_NAME (*y) == NULL_TREE) |
| return 1; |
| { |
| tree d1 = DECL_NAME (*x); |
| tree d2 = DECL_NAME (*y); |
| resort_data.new_value (&d1, resort_data.cookie); |
| resort_data.new_value (&d2, resort_data.cookie); |
| if (d1 < d2) |
| return -1; |
| } |
| return 1; |
| } |
| |
| /* Resort DECL_SORTED_FIELDS because pointers have been reordered. */ |
| |
| void |
| resort_sorted_fields (void *obj, |
| void * ARG_UNUSED (orig_obj), |
| gt_pointer_operator new_value, |
| void *cookie) |
| { |
| struct sorted_fields_type *sf = (struct sorted_fields_type *) obj; |
| resort_data.new_value = new_value; |
| resort_data.cookie = cookie; |
| qsort (&sf->elts[0], sf->len, sizeof (tree), |
| resort_field_decl_cmp); |
| } |
| |
| /* Subroutine of c_parse_error. |
| Return the result of concatenating LHS and RHS. RHS is really |
| a string literal, its first character is indicated by RHS_START and |
| RHS_SIZE is its length (including the terminating NUL character). |
| |
| The caller is responsible for deleting the returned pointer. */ |
| |
| static char * |
| catenate_strings (const char *lhs, const char *rhs_start, int rhs_size) |
| { |
| const int lhs_size = strlen (lhs); |
| char *result = XNEWVEC (char, lhs_size + rhs_size); |
| strncpy (result, lhs, lhs_size); |
| strncpy (result + lhs_size, rhs_start, rhs_size); |
| return result; |
| } |
| |
| /* Issue the error given by GMSGID, indicating that it occurred before |
| TOKEN, which had the associated VALUE. */ |
| |
| void |
| c_parse_error (const char *gmsgid, enum cpp_ttype token_type, |
| tree value, unsigned char token_flags) |
| { |
| #define catenate_messages(M1, M2) catenate_strings ((M1), (M2), sizeof (M2)) |
| |
| char *message = NULL; |
| |
| if (token_type == CPP_EOF) |
| message = catenate_messages (gmsgid, " at end of input"); |
| else if (token_type == CPP_CHAR |
| || token_type == CPP_WCHAR |
| || token_type == CPP_CHAR16 |
| || token_type == CPP_CHAR32 |
| || token_type == CPP_UTF8CHAR) |
| { |
| unsigned int val = TREE_INT_CST_LOW (value); |
| const char *prefix; |
| |
| switch (token_type) |
| { |
| default: |
| prefix = ""; |
| break; |
| case CPP_WCHAR: |
| prefix = "L"; |
| break; |
| case CPP_CHAR16: |
| prefix = "u"; |
| break; |
| case CPP_CHAR32: |
| prefix = "U"; |
| break; |
| case CPP_UTF8CHAR: |
| prefix = "u8"; |
| break; |
| } |
| |
| if (val <= UCHAR_MAX && ISGRAPH (val)) |
| message = catenate_messages (gmsgid, " before %s'%c'"); |
| else |
| message = catenate_messages (gmsgid, " before %s'\\x%x'"); |
| |
| error (message, prefix, val); |
| free (message); |
| message = NULL; |
| } |
| else if (token_type == CPP_CHAR_USERDEF |
| || token_type == CPP_WCHAR_USERDEF |
| || token_type == CPP_CHAR16_USERDEF |
| || token_type == CPP_CHAR32_USERDEF |
| || token_type == CPP_UTF8CHAR_USERDEF) |
| message = catenate_messages (gmsgid, |
| " before user-defined character literal"); |
| else if (token_type == CPP_STRING_USERDEF |
| || token_type == CPP_WSTRING_USERDEF |
| || token_type == CPP_STRING16_USERDEF |
| || token_type == CPP_STRING32_USERDEF |
| || token_type == CPP_UTF8STRING_USERDEF) |
| message = catenate_messages (gmsgid, " before user-defined string literal"); |
| else if (token_type == CPP_STRING |
| || token_type == CPP_WSTRING |
| || token_type == CPP_STRING16 |
| || token_type == CPP_STRING32 |
| || token_type == CPP_UTF8STRING) |
| message = catenate_messages (gmsgid, " before string constant"); |
| else if (token_type == CPP_NUMBER) |
| message = catenate_messages (gmsgid, " before numeric constant"); |
| else if (token_type == CPP_NAME) |
| { |
| message = catenate_messages (gmsgid, " before %qE"); |
| error (message, value); |
| free (message); |
| message = NULL; |
| } |
| else if (token_type == CPP_PRAGMA) |
| message = catenate_messages (gmsgid, " before %<#pragma%>"); |
| else if (token_type == CPP_PRAGMA_EOL) |
| message = catenate_messages (gmsgid, " before end of line"); |
| else if (token_type == CPP_DECLTYPE) |
| message = catenate_messages (gmsgid, " before %<decltype%>"); |
| else if (token_type < N_TTYPES) |
| { |
| message = catenate_messages (gmsgid, " before %qs token"); |
| error (message, cpp_type2name (token_type, token_flags)); |
| free (message); |
| message = NULL; |
| } |
| else |
| error (gmsgid); |
| |
| if (message) |
| { |
| error (message); |
| free (message); |
| } |
| #undef catenate_messages |
| } |
| |
| /* Return the gcc option code associated with the reason for a cpp |
| message, or 0 if none. */ |
| |
| static int |
| c_option_controlling_cpp_error (int reason) |
| { |
| const struct cpp_reason_option_codes_t *entry; |
| |
| for (entry = cpp_reason_option_codes; entry->reason != CPP_W_NONE; entry++) |
| { |
| if (entry->reason == reason) |
| return entry->option_code; |
| } |
| return 0; |
| } |
| |
| /* Callback from cpp_error for PFILE to print diagnostics from the |
| preprocessor. The diagnostic is of type LEVEL, with REASON set |
| to the reason code if LEVEL is represents a warning, at location |
| RICHLOC unless this is after lexing and the compiler's location |
| should be used instead; MSG is the translated message and AP |
| the arguments. Returns true if a diagnostic was emitted, false |
| otherwise. */ |
| |
| bool |
| c_cpp_error (cpp_reader *pfile ATTRIBUTE_UNUSED, int level, int reason, |
| rich_location *richloc, |
| const char *msg, va_list *ap) |
| { |
| diagnostic_info diagnostic; |
| diagnostic_t dlevel; |
| bool save_warn_system_headers = global_dc->dc_warn_system_headers; |
| bool ret; |
| |
| switch (level) |
| { |
| case CPP_DL_WARNING_SYSHDR: |
| if (flag_no_output) |
| return false; |
| global_dc->dc_warn_system_headers = 1; |
| /* Fall through. */ |
| case CPP_DL_WARNING: |
| if (flag_no_output) |
| return false; |
| dlevel = DK_WARNING; |
| break; |
| case CPP_DL_PEDWARN: |
| if (flag_no_output && !flag_pedantic_errors) |
| return false; |
| dlevel = DK_PEDWARN; |
| break; |
| case CPP_DL_ERROR: |
| dlevel = DK_ERROR; |
| break; |
| case CPP_DL_ICE: |
| dlevel = DK_ICE; |
| break; |
| case CPP_DL_NOTE: |
| dlevel = DK_NOTE; |
| break; |
| case CPP_DL_FATAL: |
| dlevel = DK_FATAL; |
| break; |
| default: |
| gcc_unreachable (); |
| } |
| if (done_lexing) |
| richloc->set_range (line_table, 0, input_location, true); |
| diagnostic_set_info_translated (&diagnostic, msg, ap, |
| richloc, dlevel); |
| diagnostic_override_option_index (&diagnostic, |
| c_option_controlling_cpp_error (reason)); |
| ret = report_diagnostic (&diagnostic); |
| if (level == CPP_DL_WARNING_SYSHDR) |
| global_dc->dc_warn_system_headers = save_warn_system_headers; |
| return ret; |
| } |
| |
| /* Convert a character from the host to the target execution character |
| set. cpplib handles this, mostly. */ |
| |
| HOST_WIDE_INT |
| c_common_to_target_charset (HOST_WIDE_INT c) |
| { |
| /* Character constants in GCC proper are sign-extended under -fsigned-char, |
| zero-extended under -fno-signed-char. cpplib insists that characters |
| and character constants are always unsigned. Hence we must convert |
| back and forth. */ |
| cppchar_t uc = ((cppchar_t)c) & ((((cppchar_t)1) << CHAR_BIT)-1); |
| |
| uc = cpp_host_to_exec_charset (parse_in, uc); |
| |
| if (flag_signed_char) |
| return ((HOST_WIDE_INT)uc) << (HOST_BITS_PER_WIDE_INT - CHAR_TYPE_SIZE) |
| >> (HOST_BITS_PER_WIDE_INT - CHAR_TYPE_SIZE); |
| else |
| return uc; |
| } |
| |
| /* Fold an offsetof-like expression. EXPR is a nested sequence of component |
| references with an INDIRECT_REF of a constant at the bottom; much like the |
| traditional rendering of offsetof as a macro. TYPE is the desired type of |
| the whole expression. Return the folded result. */ |
| |
| tree |
| fold_offsetof (tree expr, tree type, enum tree_code ctx) |
| { |
| tree base, off, t; |
| tree_code code = TREE_CODE (expr); |
| switch (code) |
| { |
| case ERROR_MARK: |
| return expr; |
| |
| case VAR_DECL: |
| error ("cannot apply %<offsetof%> to static data member %qD", expr); |
| return error_mark_node; |
| |
| case CALL_EXPR: |
| case TARGET_EXPR: |
| error ("cannot apply %<offsetof%> when %<operator[]%> is overloaded"); |
| return error_mark_node; |
| |
| case NOP_EXPR: |
| case INDIRECT_REF: |
| if (!TREE_CONSTANT (TREE_OPERAND (expr, 0))) |
| { |
| error ("cannot apply %<offsetof%> to a non constant address"); |
| return error_mark_node; |
| } |
| return convert (type, TREE_OPERAND (expr, 0)); |
| |
| case COMPONENT_REF: |
| base = fold_offsetof (TREE_OPERAND (expr, 0), type, code); |
| if (base == error_mark_node) |
| return base; |
| |
| t = TREE_OPERAND (expr, 1); |
| if (DECL_C_BIT_FIELD (t)) |
| { |
| error ("attempt to take address of bit-field structure " |
| "member %qD", t); |
| return error_mark_node; |
| } |
| off = size_binop_loc (input_location, PLUS_EXPR, DECL_FIELD_OFFSET (t), |
| size_int (tree_to_uhwi (DECL_FIELD_BIT_OFFSET (t)) |
| / BITS_PER_UNIT)); |
| break; |
| |
| case ARRAY_REF: |
| base = fold_offsetof (TREE_OPERAND (expr, 0), type, code); |
| if (base == error_mark_node) |
| return base; |
| |
| t = TREE_OPERAND (expr, 1); |
| |
| /* Check if the offset goes beyond the upper bound of the array. */ |
| if (TREE_CODE (t) == INTEGER_CST && tree_int_cst_sgn (t) >= 0) |
| { |
| tree upbound = array_ref_up_bound (expr); |
| if (upbound != NULL_TREE |
| && TREE_CODE (upbound) == INTEGER_CST |
| && !tree_int_cst_equal (upbound, |
| TYPE_MAX_VALUE (TREE_TYPE (upbound)))) |
| { |
| if (ctx != ARRAY_REF && ctx != COMPONENT_REF) |
| upbound = size_binop (PLUS_EXPR, upbound, |
| build_int_cst (TREE_TYPE (upbound), 1)); |
| if (tree_int_cst_lt (upbound, t)) |
| { |
| tree v; |
| |
| for (v = TREE_OPERAND (expr, 0); |
| TREE_CODE (v) == COMPONENT_REF; |
| v = TREE_OPERAND (v, 0)) |
| if (TREE_CODE (TREE_TYPE (TREE_OPERAND (v, 0))) |
| == RECORD_TYPE) |
| { |
| tree fld_chain = DECL_CHAIN (TREE_OPERAND (v, 1)); |
| for (; fld_chain; fld_chain = DECL_CHAIN (fld_chain)) |
| if (TREE_CODE (fld_chain) == FIELD_DECL) |
| break; |
| |
| if (fld_chain) |
| break; |
| } |
| /* Don't warn if the array might be considered a poor |
| man's flexible array member with a very permissive |
| definition thereof. */ |
| if (TREE_CODE (v) == ARRAY_REF |
| || TREE_CODE (v) == COMPONENT_REF) |
| warning (OPT_Warray_bounds, |
| "index %E denotes an offset " |
| "greater than size of %qT", |
| t, TREE_TYPE (TREE_OPERAND (expr, 0))); |
| } |
| } |
| } |
| |
| t = convert (sizetype, t); |
| off = size_binop (MULT_EXPR, TYPE_SIZE_UNIT (TREE_TYPE (expr)), t); |
| break; |
| |
| case COMPOUND_EXPR: |
| /* Handle static members of volatile structs. */ |
| t = TREE_OPERAND (expr, 1); |
| gcc_assert (VAR_P (t)); |
| return fold_offsetof (t, type); |
| |
| default: |
| gcc_unreachable (); |
| } |
| |
| if (!POINTER_TYPE_P (type)) |
| return size_binop (PLUS_EXPR, base, convert (type, off)); |
| return fold_build_pointer_plus (base, off); |
| } |
| |
| /* *PTYPE is an incomplete array. Complete it with a domain based on |
| INITIAL_VALUE. If INITIAL_VALUE is not present, use 1 if DO_DEFAULT |
| is true. Return 0 if successful, 1 if INITIAL_VALUE can't be deciphered, |
| 2 if INITIAL_VALUE was NULL, and 3 if INITIAL_VALUE was empty. */ |
| |
| int |
| complete_array_type (tree *ptype, tree initial_value, bool do_default) |
| { |
| tree maxindex, type, main_type, elt, unqual_elt; |
| int failure = 0, quals; |
| bool overflow_p = false; |
| |
| maxindex = size_zero_node; |
| if (initial_value) |
| { |
| if (TREE_CODE (initial_value) == STRING_CST) |
| { |
| int eltsize |
| = int_size_in_bytes (TREE_TYPE (TREE_TYPE (initial_value))); |
| maxindex = size_int (TREE_STRING_LENGTH (initial_value)/eltsize - 1); |
| } |
| else if (TREE_CODE (initial_value) == CONSTRUCTOR) |
| { |
| vec<constructor_elt, va_gc> *v = CONSTRUCTOR_ELTS (initial_value); |
| |
| if (vec_safe_is_empty (v)) |
| { |
| if (pedantic) |
| failure = 3; |
| maxindex = ssize_int (-1); |
| } |
| else |
| { |
| tree curindex; |
| unsigned HOST_WIDE_INT cnt; |
| constructor_elt *ce; |
| bool fold_p = false; |
| |
| if ((*v)[0].index) |
| maxindex = (*v)[0].index, fold_p = true; |
| |
| curindex = maxindex; |
| |
| for (cnt = 1; vec_safe_iterate (v, cnt, &ce); cnt++) |
| { |
| bool curfold_p = false; |
| if (ce->index) |
| curindex = ce->index, curfold_p = true; |
| else |
| { |
| if (fold_p) |
| { |
| /* Since we treat size types now as ordinary |
| unsigned types, we need an explicit overflow |
| check. */ |
| tree orig = curindex; |
| curindex = fold_convert (sizetype, curindex); |
| overflow_p |= tree_int_cst_lt (curindex, orig); |
| } |
| curindex = size_binop (PLUS_EXPR, curindex, |
| size_one_node); |
| } |
| if (tree_int_cst_lt (maxindex, curindex)) |
| maxindex = curindex, fold_p = curfold_p; |
| } |
| if (fold_p) |
| { |
| tree orig = maxindex; |
| maxindex = fold_convert (sizetype, maxindex); |
| overflow_p |= tree_int_cst_lt (maxindex, orig); |
| } |
| } |
| } |
| else |
| { |
| /* Make an error message unless that happened already. */ |
| if (initial_value != error_mark_node) |
| failure = 1; |
| } |
| } |
| else |
| { |
| failure = 2; |
| if (!do_default) |
| return failure; |
| } |
| |
| type = *ptype; |
| elt = TREE_TYPE (type); |
| quals = TYPE_QUALS (strip_array_types (elt)); |
| if (quals == 0) |
| unqual_elt = elt; |
| else |
| unqual_elt = c_build_qualified_type (elt, KEEP_QUAL_ADDR_SPACE (quals)); |
| |
| /* Using build_distinct_type_copy and modifying things afterward instead |
| of using build_array_type to create a new type preserves all of the |
| TYPE_LANG_FLAG_? bits that the front end may have set. */ |
| main_type = build_distinct_type_copy (TYPE_MAIN_VARIANT (type)); |
| TREE_TYPE (main_type) = unqual_elt; |
| TYPE_DOMAIN (main_type) |
| = build_range_type (TREE_TYPE (maxindex), |
| build_int_cst (TREE_TYPE (maxindex), 0), maxindex); |
| TYPE_TYPELESS_STORAGE (main_type) = TYPE_TYPELESS_STORAGE (type); |
| layout_type (main_type); |
| |
| /* Make sure we have the canonical MAIN_TYPE. */ |
| inchash::hash hstate; |
| hstate.add_object (TYPE_HASH (unqual_elt)); |
| hstate.add_object (TYPE_HASH (TYPE_DOMAIN (main_type))); |
| if (!AGGREGATE_TYPE_P (unqual_elt)) |
| hstate.add_flag (TYPE_TYPELESS_STORAGE (main_type)); |
| main_type = type_hash_canon (hstate.end (), main_type); |
| |
| /* Fix the canonical type. */ |
| if (TYPE_STRUCTURAL_EQUALITY_P (TREE_TYPE (main_type)) |
| || TYPE_STRUCTURAL_EQUALITY_P (TYPE_DOMAIN (main_type))) |
| SET_TYPE_STRUCTURAL_EQUALITY (main_type); |
| else if (TYPE_CANONICAL (TREE_TYPE (main_type)) != TREE_TYPE (main_type) |
| || (TYPE_CANONICAL (TYPE_DOMAIN (main_type)) |
| != TYPE_DOMAIN (main_type))) |
| TYPE_CANONICAL (main_type) |
| = build_array_type (TYPE_CANONICAL (TREE_TYPE (main_type)), |
| TYPE_CANONICAL (TYPE_DOMAIN (main_type)), |
| TYPE_TYPELESS_STORAGE (main_type)); |
| else |
| TYPE_CANONICAL (main_type) = main_type; |
| |
| if (quals == 0) |
| type = main_type; |
| else |
| type = c_build_qualified_type (main_type, quals); |
| |
| if (COMPLETE_TYPE_P (type) |
| && TREE_CODE (TYPE_SIZE_UNIT (type)) == INTEGER_CST |
| && (overflow_p || TREE_OVERFLOW (TYPE_SIZE_UNIT (type)))) |
| { |
| error ("size of array is too large"); |
| /* If we proceed with the array type as it is, we'll eventually |
| crash in tree_to_[su]hwi(). */ |
| type = error_mark_node; |
| } |
| |
| *ptype = type; |
| return failure; |
| } |
| |
| /* Like c_mark_addressable but don't check register qualifier. */ |
| void |
| c_common_mark_addressable_vec (tree t) |
| { |
| if (TREE_CODE (t) == C_MAYBE_CONST_EXPR) |
| t = C_MAYBE_CONST_EXPR_EXPR (t); |
| while (handled_component_p (t)) |
| t = TREE_OPERAND (t, 0); |
| if (!VAR_P (t) |
| && TREE_CODE (t) != PARM_DECL |
| && TREE_CODE (t) != COMPOUND_LITERAL_EXPR) |
| return; |
| if (!VAR_P (t) || !DECL_HARD_REGISTER (t)) |
| TREE_ADDRESSABLE (t) = 1; |
| if (TREE_CODE (t) == COMPOUND_LITERAL_EXPR) |
| TREE_ADDRESSABLE (COMPOUND_LITERAL_EXPR_DECL (t)) = 1; |
| } |
| |
| |
| |
| /* Used to help initialize the builtin-types.def table. When a type of |
| the correct size doesn't exist, use error_mark_node instead of NULL. |
| The later results in segfaults even when a decl using the type doesn't |
| get invoked. */ |
| |
| tree |
| builtin_type_for_size (int size, bool unsignedp) |
| { |
| tree type = c_common_type_for_size (size, unsignedp); |
| return type ? type : error_mark_node; |
| } |
| |
| /* A helper function for resolve_overloaded_builtin in resolving the |
| overloaded __sync_ builtins. Returns a positive power of 2 if the |
| first operand of PARAMS is a pointer to a supported data type. |
| Returns 0 if an error is encountered. |
| FETCH is true when FUNCTION is one of the _FETCH_OP_ or _OP_FETCH_ |
| built-ins. */ |
| |
| static int |
| sync_resolve_size (tree function, vec<tree, va_gc> *params, bool fetch) |
| { |
| /* Type of the argument. */ |
| tree argtype; |
| /* Type the argument points to. */ |
| tree type; |
| int size; |
| |
| if (vec_safe_is_empty (params)) |
| { |
| error ("too few arguments to function %qE", function); |
| return 0; |
| } |
| |
| argtype = type = TREE_TYPE ((*params)[0]); |
| if (TREE_CODE (type) == ARRAY_TYPE && c_dialect_cxx ()) |
| { |
| /* Force array-to-pointer decay for C++. */ |
| (*params)[0] = default_conversion ((*params)[0]); |
| type = TREE_TYPE ((*params)[0]); |
| } |
| if (TREE_CODE (type) != POINTER_TYPE) |
| goto incompatible; |
| |
| type = TREE_TYPE (type); |
| if (!INTEGRAL_TYPE_P (type) && !POINTER_TYPE_P (type)) |
| goto incompatible; |
| |
| if (fetch && TREE_CODE (type) == BOOLEAN_TYPE) |
| goto incompatible; |
| |
| size = tree_to_uhwi (TYPE_SIZE_UNIT (type)); |
| if (size == 1 || size == 2 || size == 4 || size == 8 || size == 16) |
| return size; |
| |
| incompatible: |
| /* Issue the diagnostic only if the argument is valid, otherwise |
| it would be redundant at best and could be misleading. */ |
| if (argtype != error_mark_node) |
| error ("operand type %qT is incompatible with argument %d of %qE", |
| argtype, 1, function); |
| return 0; |
| } |
| |
| /* A helper function for resolve_overloaded_builtin. Adds casts to |
| PARAMS to make arguments match up with those of FUNCTION. Drops |
| the variadic arguments at the end. Returns false if some error |
| was encountered; true on success. */ |
| |
| static bool |
| sync_resolve_params (location_t loc, tree orig_function, tree function, |
| vec<tree, va_gc> *params, bool orig_format) |
| { |
| function_args_iterator iter; |
| tree ptype; |
| unsigned int parmnum; |
| |
| function_args_iter_init (&iter, TREE_TYPE (function)); |
| /* We've declared the implementation functions to use "volatile void *" |
| as the pointer parameter, so we shouldn't get any complaints from the |
| call to check_function_arguments what ever type the user used. */ |
| function_args_iter_next (&iter); |
| ptype = TREE_TYPE (TREE_TYPE ((*params)[0])); |
| ptype = TYPE_MAIN_VARIANT (ptype); |
| |
| /* For the rest of the values, we need to cast these to FTYPE, so that we |
| don't get warnings for passing pointer types, etc. */ |
| parmnum = 0; |
| while (1) |
| { |
| tree val, arg_type; |
| |
| arg_type = function_args_iter_cond (&iter); |
| /* XXX void_type_node belies the abstraction. */ |
| if (arg_type == void_type_node) |
| break; |
| |
| ++parmnum; |
| if (params->length () <= parmnum) |
| { |
| error_at (loc, "too few arguments to function %qE", orig_function); |
| return false; |
| } |
| |
| /* Only convert parameters if arg_type is unsigned integer type with |
| new format sync routines, i.e. don't attempt to convert pointer |
| arguments (e.g. EXPECTED argument of __atomic_compare_exchange_n), |
| bool arguments (e.g. WEAK argument) or signed int arguments (memmodel |
| kinds). */ |
| if (TREE_CODE (arg_type) == INTEGER_TYPE && TYPE_UNSIGNED (arg_type)) |
| { |
| /* Ideally for the first conversion we'd use convert_for_assignment |
| so that we get warnings for anything that doesn't match the pointer |
| type. This isn't portable across the C and C++ front ends atm. */ |
| val = (*params)[parmnum]; |
| val = convert (ptype, val); |
| val = convert (arg_type, val); |
| (*params)[parmnum] = val; |
| } |
| |
| function_args_iter_next (&iter); |
| } |
| |
| /* __atomic routines are not variadic. */ |
| if (!orig_format && params->length () != parmnum + 1) |
| { |
| error_at (loc, "too many arguments to function %qE", orig_function); |
| return false; |
| } |
| |
| /* The definition of these primitives is variadic, with the remaining |
| being "an optional list of variables protected by the memory barrier". |
| No clue what that's supposed to mean, precisely, but we consider all |
| call-clobbered variables to be protected so we're safe. */ |
| params->truncate (parmnum + 1); |
| |
| return true; |
| } |
| |
| /* A helper function for resolve_overloaded_builtin. Adds a cast to |
| RESULT to make it match the type of the first pointer argument in |
| PARAMS. */ |
| |
| static tree |
| sync_resolve_return (tree first_param, tree result, bool orig_format) |
| { |
| tree ptype = TREE_TYPE (TREE_TYPE (first_param)); |
| tree rtype = TREE_TYPE (result); |
| ptype = TYPE_MAIN_VARIANT (ptype); |
| |
| /* New format doesn't require casting unless the types are the same size. */ |
| if (orig_format || tree_int_cst_equal (TYPE_SIZE (ptype), TYPE_SIZE (rtype))) |
| return convert (ptype, result); |
| else |
| return result; |
| } |
| |
| /* This function verifies the PARAMS to generic atomic FUNCTION. |
| It returns the size if all the parameters are the same size, otherwise |
| 0 is returned if the parameters are invalid. */ |
| |
| static int |
| get_atomic_generic_size (location_t loc, tree function, |
| vec<tree, va_gc> *params) |
| { |
| unsigned int n_param; |
| unsigned int n_model; |
| unsigned int x; |
| int size_0; |
| tree type_0; |
| |
| /* Determine the parameter makeup. */ |
| switch (DECL_FUNCTION_CODE (function)) |
| { |
| case BUILT_IN_ATOMIC_EXCHANGE: |
| n_param = 4; |
| n_model = 1; |
| break; |
| case BUILT_IN_ATOMIC_LOAD: |
| case BUILT_IN_ATOMIC_STORE: |
| n_param = 3; |
| n_model = 1; |
| break; |
| case BUILT_IN_ATOMIC_COMPARE_EXCHANGE: |
| n_param = 6; |
| n_model = 2; |
| break; |
| default: |
| gcc_unreachable (); |
| } |
| |
| if (vec_safe_length (params) != n_param) |
| { |
| error_at (loc, "incorrect number of arguments to function %qE", function); |
| return 0; |
| } |
| |
| /* Get type of first parameter, and determine its size. */ |
| type_0 = TREE_TYPE ((*params)[0]); |
| if (TREE_CODE (type_0) == ARRAY_TYPE && c_dialect_cxx ()) |
| { |
| /* Force array-to-pointer decay for C++. */ |
| (*params)[0] = default_conversion ((*params)[0]); |
| type_0 = TREE_TYPE ((*params)[0]); |
| } |
| if (TREE_CODE (type_0) != POINTER_TYPE || VOID_TYPE_P (TREE_TYPE (type_0))) |
| { |
| error_at (loc, "argument 1 of %qE must be a non-void pointer type", |
| function); |
| return 0; |
| } |
| |
| /* Types must be compile time constant sizes. */ |
| if (TREE_CODE ((TYPE_SIZE_UNIT (TREE_TYPE (type_0)))) != INTEGER_CST) |
| { |
| error_at (loc, |
| "argument 1 of %qE must be a pointer to a constant size type", |
| function); |
| return 0; |
| } |
| |
| size_0 = tree_to_uhwi (TYPE_SIZE_UNIT (TREE_TYPE (type_0))); |
| |
| /* Zero size objects are not allowed. */ |
| if (size_0 == 0) |
| { |
| error_at (loc, |
| "argument 1 of %qE must be a pointer to a nonzero size object", |
| function); |
| return 0; |
| } |
| |
| /* Check each other parameter is a pointer and the same size. */ |
| for (x = 0; x < n_param - n_model; x++) |
| { |
| int size; |
| tree type = TREE_TYPE ((*params)[x]); |
| /* __atomic_compare_exchange has a bool in the 4th position, skip it. */ |
| if (n_param == 6 && x == 3) |
| continue; |
| if (TREE_CODE (type) == ARRAY_TYPE && c_dialect_cxx ()) |
| { |
| /* Force array-to-pointer decay for C++. */ |
| (*params)[x] = default_conversion ((*params)[x]); |
| type = TREE_TYPE ((*params)[x]); |
| } |
| if (!POINTER_TYPE_P (type)) |
| { |
| error_at (loc, "argument %d of %qE must be a pointer type", x + 1, |
| function); |
| return 0; |
| } |
| else if (TYPE_SIZE_UNIT (TREE_TYPE (type)) |
| && TREE_CODE ((TYPE_SIZE_UNIT (TREE_TYPE (type)))) |
| != INTEGER_CST) |
| { |
| error_at (loc, "argument %d of %qE must be a pointer to a constant " |
| "size type", x + 1, function); |
| return 0; |
| } |
| else if (FUNCTION_POINTER_TYPE_P (type)) |
| { |
| error_at (loc, "argument %d of %qE must not be a pointer to a " |
| "function", x + 1, function); |
| return 0; |
| } |
| tree type_size = TYPE_SIZE_UNIT (TREE_TYPE (type)); |
| size = type_size ? tree_to_uhwi (type_size) : 0; |
| if (size != size_0) |
| { |
| error_at (loc, "size mismatch in argument %d of %qE", x + 1, |
| function); |
| return 0; |
| } |
| } |
| |
| /* Check memory model parameters for validity. */ |
| for (x = n_param - n_model ; x < n_param; x++) |
| { |
| tree p = (*params)[x]; |
| if (TREE_CODE (p) == INTEGER_CST) |
| { |
| /* memmodel_base masks the low 16 bits, thus ignore any bits above |
| it by using TREE_INT_CST_LOW instead of tree_to_*hwi. Those high |
| bits will be checked later during expansion in target specific |
| way. */ |
| if (memmodel_base (TREE_INT_CST_LOW (p)) >= MEMMODEL_LAST) |
| warning_at (loc, OPT_Winvalid_memory_model, |
| "invalid memory model argument %d of %qE", x + 1, |
| function); |
| } |
| else |
| if (!INTEGRAL_TYPE_P (TREE_TYPE (p))) |
| { |
| error_at (loc, "non-integer memory model argument %d of %qE", x + 1, |
| function); |
| return 0; |
| } |
| } |
| |
| return size_0; |
| } |
| |
| |
| /* This will take an __atomic_ generic FUNCTION call, and add a size parameter N |
| at the beginning of the parameter list PARAMS representing the size of the |
| objects. This is to match the library ABI requirement. LOC is the location |
| of the function call. |
| The new function is returned if it needed rebuilding, otherwise NULL_TREE is |
| returned to allow the external call to be constructed. */ |
| |
| static tree |
| add_atomic_size_parameter (unsigned n, location_t loc, tree function, |
| vec<tree, va_gc> *params) |
| { |
| tree size_node; |
| |
| /* Insert a SIZE_T parameter as the first param. If there isn't |
| enough space, allocate a new vector and recursively re-build with that. */ |
| if (!params->space (1)) |
| { |
| unsigned int z, len; |
| vec<tree, va_gc> *v; |
| tree f; |
| |
| len = params->length (); |
| vec_alloc (v, len + 1); |
| v->quick_push (build_int_cst (size_type_node, n)); |
| for (z = 0; z < len; z++) |
| v->quick_push ((*params)[z]); |
| f = build_function_call_vec (loc, vNULL, function, v, NULL); |
| vec_free (v); |
| return f; |
| } |
| |
| /* Add the size parameter and leave as a function call for processing. */ |
| size_node = build_int_cst (size_type_node, n); |
| params->quick_insert (0, size_node); |
| return NULL_TREE; |
| } |
| |
| |
| /* Return whether atomic operations for naturally aligned N-byte |
| arguments are supported, whether inline or through libatomic. */ |
| static bool |
| atomic_size_supported_p (int n) |
| { |
| switch (n) |
| { |
| case 1: |
| case 2: |
| case 4: |
| case 8: |
| return true; |
| |
| case 16: |
| return targetm.scalar_mode_supported_p (TImode); |
| |
| default: |
| return false; |
| } |
| } |
| |
| /* This will process an __atomic_exchange function call, determine whether it |
| needs to be mapped to the _N variation, or turned into a library call. |
| LOC is the location of the builtin call. |
| FUNCTION is the DECL that has been invoked; |
| PARAMS is the argument list for the call. The return value is non-null |
| TRUE is returned if it is translated into the proper format for a call to the |
| external library, and NEW_RETURN is set the tree for that function. |
| FALSE is returned if processing for the _N variation is required, and |
| NEW_RETURN is set to the return value the result is copied into. */ |
| static bool |
| resolve_overloaded_atomic_exchange (location_t loc, tree function, |
| vec<tree, va_gc> *params, tree *new_return) |
| { |
| tree p0, p1, p2, p3; |
| tree I_type, I_type_ptr; |
| int n = get_atomic_generic_size (loc, function, params); |
| |
| /* Size of 0 is an error condition. */ |
| if (n == 0) |
| { |
| *new_return = error_mark_node; |
| return true; |
| } |
| |
| /* If not a lock-free size, change to the library generic format. */ |
| if (!atomic_size_supported_p (n)) |
| { |
| *new_return = add_atomic_size_parameter (n, loc, function, params); |
| return true; |
| } |
| |
| /* Otherwise there is a lockfree match, transform the call from: |
| void fn(T* mem, T* desired, T* return, model) |
| into |
| *return = (T) (fn (In* mem, (In) *desired, model)) */ |
| |
| p0 = (*params)[0]; |
| p1 = (*params)[1]; |
| p2 = (*params)[2]; |
| p3 = (*params)[3]; |
| |
| /* Create pointer to appropriate size. */ |
| I_type = builtin_type_for_size (BITS_PER_UNIT * n, 1); |
| I_type_ptr = build_pointer_type (I_type); |
| |
| /* Convert object pointer to required type. */ |
| p0 = build1 (VIEW_CONVERT_EXPR, I_type_ptr, p0); |
| (*params)[0] = p0; |
| /* Convert new value to required type, and dereference it. */ |
| p1 = build_indirect_ref (loc, p1, RO_UNARY_STAR); |
| p1 = build1 (VIEW_CONVERT_EXPR, I_type, p1); |
| (*params)[1] = p1; |
| |
| /* Move memory model to the 3rd position, and end param list. */ |
| (*params)[2] = p3; |
| params->truncate (3); |
| |
| /* Convert return pointer and dereference it for later assignment. */ |
| *new_return = build_indirect_ref (loc, p2, RO_UNARY_STAR); |
| |
| return false; |
| } |
| |
| |
| /* This will process an __atomic_compare_exchange function call, determine |
| whether it needs to be mapped to the _N variation, or turned into a lib call. |
| LOC is the location of the builtin call. |
| FUNCTION is the DECL that has been invoked; |
| PARAMS is the argument list for the call. The return value is non-null |
| TRUE is returned if it is translated into the proper format for a call to the |
| external library, and NEW_RETURN is set the tree for that function. |
| FALSE is returned if processing for the _N variation is required. */ |
| |
| static bool |
| resolve_overloaded_atomic_compare_exchange (location_t loc, tree function, |
| vec<tree, va_gc> *params, |
| tree *new_return) |
| { |
| tree p0, p1, p2; |
| tree I_type, I_type_ptr; |
| int n = get_atomic_generic_size (loc, function, params); |
| |
| /* Size of 0 is an error condition. */ |
| if (n == 0) |
| { |
| *new_return = error_mark_node; |
| return true; |
| } |
| |
| /* If not a lock-free size, change to the library generic format. */ |
| if (!atomic_size_supported_p (n)) |
| { |
| /* The library generic format does not have the weak parameter, so |
| remove it from the param list. Since a parameter has been removed, |
| we can be sure that there is room for the SIZE_T parameter, meaning |
| there will not be a recursive rebuilding of the parameter list, so |
| there is no danger this will be done twice. */ |
| if (n > 0) |
| { |
| (*params)[3] = (*params)[4]; |
| (*params)[4] = (*params)[5]; |
| params->truncate (5); |
| } |
| *new_return = add_atomic_size_parameter (n, loc, function, params); |
| return true; |
| } |
| |
| /* Otherwise, there is a match, so the call needs to be transformed from: |
| bool fn(T* mem, T* desired, T* return, weak, success, failure) |
| into |
| bool fn ((In *)mem, (In *)expected, (In) *desired, weak, succ, fail) */ |
| |
| p0 = (*params)[0]; |
| p1 = (*params)[1]; |
| p2 = (*params)[2]; |
| |
| /* Create pointer to appropriate size. */ |
| I_type = builtin_type_for_size (BITS_PER_UNIT * n, 1); |
| I_type_ptr = build_pointer_type (I_type); |
| |
| /* Convert object pointer to required type. */ |
| p0 = build1 (VIEW_CONVERT_EXPR, I_type_ptr, p0); |
| (*params)[0] = p0; |
| |
| /* Convert expected pointer to required type. */ |
| p1 = build1 (VIEW_CONVERT_EXPR, I_type_ptr, p1); |
| (*params)[1] = p1; |
| |
| /* Convert desired value to required type, and dereference it. */ |
| p2 = build_indirect_ref (loc, p2, RO_UNARY_STAR); |
| p2 = build1 (VIEW_CONVERT_EXPR, I_type, p2); |
| (*params)[2] = p2; |
| |
| /* The rest of the parameters are fine. NULL means no special return value |
| processing.*/ |
| *new_return = NULL; |
| return false; |
| } |
| |
| |
| /* This will process an __atomic_load function call, determine whether it |
| needs to be mapped to the _N variation, or turned into a library call. |
| LOC is the location of the builtin call. |
| FUNCTION is the DECL that has been invoked; |
| PARAMS is the argument list for the call. The return value is non-null |
| TRUE is returned if it is translated into the proper format for a call to the |
| external library, and NEW_RETURN is set the tree for that function. |
| FALSE is returned if processing for the _N variation is required, and |
| NEW_RETURN is set to the return value the result is copied into. */ |
| |
| static bool |
| resolve_overloaded_atomic_load (location_t loc, tree function, |
| vec<tree, va_gc> *params, tree *new_return) |
| { |
| tree p0, p1, p2; |
| tree I_type, I_type_ptr; |
| int n = get_atomic_generic_size (loc, function, params); |
| |
| /* Size of 0 is an error condition. */ |
| if (n == 0) |
| { |
| *new_return = error_mark_node; |
| return true; |
| } |
| |
| /* If not a lock-free size, change to the library generic format. */ |
| if (!atomic_size_supported_p (n)) |
| { |
| *new_return = add_atomic_size_parameter (n, loc, function, params); |
| return true; |
| } |
| |
| /* Otherwise, there is a match, so the call needs to be transformed from: |
| void fn(T* mem, T* return, model) |
| into |
| *return = (T) (fn ((In *) mem, model)) */ |
| |
| p0 = (*params)[0]; |
| p1 = (*params)[1]; |
| p2 = (*params)[2]; |
| |
| /* Create pointer to appropriate size. */ |
| I_type = builtin_type_for_size (BITS_PER_UNIT * n, 1); |
| I_type_ptr = build_pointer_type (I_type); |
| |
| /* Convert object pointer to required type. */ |
| p0 = build1 (VIEW_CONVERT_EXPR, I_type_ptr, p0); |
| (*params)[0] = p0; |
| |
| /* Move memory model to the 2nd position, and end param list. */ |
| (*params)[1] = p2; |
| params->truncate (2); |
| |
| /* Convert return pointer and dereference it for later assignment. */ |
| *new_return = build_indirect_ref (loc, p1, RO_UNARY_STAR); |
| |
| return false; |
| } |
| |
| |
| /* This will process an __atomic_store function call, determine whether it |
| needs to be mapped to the _N variation, or turned into a library call. |
| LOC is the location of the builtin call. |
| FUNCTION is the DECL that has been invoked; |
| PARAMS is the argument list for the call. The return value is non-null |
| TRUE is returned if it is translated into the proper format for a call to the |
| external library, and NEW_RETURN is set the tree for that function. |
| FALSE is returned if processing for the _N variation is required, and |
| NEW_RETURN is set to the return value the result is copied into. */ |
| |
| static bool |
| resolve_overloaded_atomic_store (location_t loc, tree function, |
| vec<tree, va_gc> *params, tree *new_return) |
| { |
| tree p0, p1; |
| tree I_type, I_type_ptr; |
| int n = get_atomic_generic_size (loc, function, params); |
| |
| /* Size of 0 is an error condition. */ |
| if (n == 0) |
| { |
| *new_return = error_mark_node; |
| return true; |
| } |
| |
| /* If not a lock-free size, change to the library generic format. */ |
| if (!atomic_size_supported_p (n)) |
| { |
| *new_return = add_atomic_size_parameter (n, loc, function, params); |
| return true; |
| } |
| |
| /* Otherwise, there is a match, so the call needs to be transformed from: |
| void fn(T* mem, T* value, model) |
| into |
| fn ((In *) mem, (In) *value, model) */ |
| |
| p0 = (*params)[0]; |
| p1 = (*params)[1]; |
| |
| /* Create pointer to appropriate size. */ |
| I_type = builtin_type_for_size (BITS_PER_UNIT * n, 1); |
| I_type_ptr = build_pointer_type (I_type); |
| |
| /* Convert object pointer to required type. */ |
| p0 = build1 (VIEW_CONVERT_EXPR, I_type_ptr, p0); |
| (*params)[0] = p0; |
| |
| /* Convert new value to required type, and dereference it. */ |
| p1 = build_indirect_ref (loc, p1, RO_UNARY_STAR); |
| p1 = build1 (VIEW_CONVERT_EXPR, I_type, p1); |
| (*params)[1] = p1; |
| |
| /* The memory model is in the right spot already. Return is void. */ |
| *new_return = NULL_TREE; |
| |
| return false; |
| } |
| |
| |
| /* Some builtin functions are placeholders for other expressions. This |
| function should be called immediately after parsing the call expression |
| before surrounding code has committed to the type of the expression. |
| |
| LOC is the location of the builtin call. |
| |
| FUNCTION is the DECL that has been invoked; it is known to be a builtin. |
| PARAMS is the argument list for the call. The return value is non-null |
| when expansion is complete, and null if normal processing should |
| continue. */ |
| |
| tree |
| resolve_overloaded_builtin (location_t loc, tree function, |
| vec<tree, va_gc> *params) |
| { |
| enum built_in_function orig_code = DECL_FUNCTION_CODE (function); |
| |
| /* Is function one of the _FETCH_OP_ or _OP_FETCH_ built-ins? |
| Those are not valid to call with a pointer to _Bool (or C++ bool) |
| and so must be rejected. */ |
| bool fetch_op = true; |
| bool orig_format = true; |
| tree new_return = NULL_TREE; |
| |
| switch (DECL_BUILT_IN_CLASS (function)) |
| { |
| case BUILT_IN_NORMAL: |
| break; |
| case BUILT_IN_MD: |
| if (targetm.resolve_overloaded_builtin) |
| return targetm.resolve_overloaded_builtin (loc, function, params); |
| else |
| return NULL_TREE; |
| default: |
| return NULL_TREE; |
| } |
| |
| /* Handle BUILT_IN_NORMAL here. */ |
| switch (orig_code) |
| { |
| case BUILT_IN_ATOMIC_EXCHANGE: |
| case BUILT_IN_ATOMIC_COMPARE_EXCHANGE: |
| case BUILT_IN_ATOMIC_LOAD: |
| case BUILT_IN_ATOMIC_STORE: |
| { |
| /* Handle these 4 together so that they can fall through to the next |
| case if the call is transformed to an _N variant. */ |
| switch (orig_code) |
| { |
| case BUILT_IN_ATOMIC_EXCHANGE: |
| { |
| if (resolve_overloaded_atomic_exchange (loc, function, params, |
| &new_return)) |
| return new_return; |
| /* Change to the _N variant. */ |
| orig_code = BUILT_IN_ATOMIC_EXCHANGE_N; |
| break; |
| } |
| |
| case BUILT_IN_ATOMIC_COMPARE_EXCHANGE: |
| { |
| if (resolve_overloaded_atomic_compare_exchange (loc, function, |
| params, |
| &new_return)) |
| return new_return; |
| /* Change to the _N variant. */ |
| orig_code = BUILT_IN_ATOMIC_COMPARE_EXCHANGE_N; |
| break; |
| } |
| case BUILT_IN_ATOMIC_LOAD: |
| { |
| if (resolve_overloaded_atomic_load (loc, function, params, |
| &new_return)) |
| return new_return; |
| /* Change to the _N variant. */ |
| orig_code = BUILT_IN_ATOMIC_LOAD_N; |
| break; |
| } |
| case BUILT_IN_ATOMIC_STORE: |
| { |
| if (resolve_overloaded_atomic_store (loc, function, params, |
| &new_return)) |
| return new_return; |
| /* Change to the _N variant. */ |
| orig_code = BUILT_IN_ATOMIC_STORE_N; |
| break; |
| } |
| default: |
| gcc_unreachable (); |
| } |
| } |
| /* FALLTHRU */ |
| case BUILT_IN_ATOMIC_EXCHANGE_N: |
| case BUILT_IN_ATOMIC_COMPARE_EXCHANGE_N: |
| case BUILT_IN_ATOMIC_LOAD_N: |
| case BUILT_IN_ATOMIC_STORE_N: |
| fetch_op = false; |
| /* FALLTHRU */ |
| case BUILT_IN_ATOMIC_ADD_FETCH_N: |
| case BUILT_IN_ATOMIC_SUB_FETCH_N: |
| case BUILT_IN_ATOMIC_AND_FETCH_N: |
| case BUILT_IN_ATOMIC_NAND_FETCH_N: |
| case BUILT_IN_ATOMIC_XOR_FETCH_N: |
| case BUILT_IN_ATOMIC_OR_FETCH_N: |
| case BUILT_IN_ATOMIC_FETCH_ADD_N: |
| case BUILT_IN_ATOMIC_FETCH_SUB_N: |
| case BUILT_IN_ATOMIC_FETCH_AND_N: |
| case BUILT_IN_ATOMIC_FETCH_NAND_N: |
| case BUILT_IN_ATOMIC_FETCH_XOR_N: |
| case BUILT_IN_ATOMIC_FETCH_OR_N: |
| orig_format = false; |
| /* FALLTHRU */ |
| case BUILT_IN_SYNC_FETCH_AND_ADD_N: |
| case BUILT_IN_SYNC_FETCH_AND_SUB_N: |
| case BUILT_IN_SYNC_FETCH_AND_OR_N: |
| case BUILT_IN_SYNC_FETCH_AND_AND_N: |
| case BUILT_IN_SYNC_FETCH_AND_XOR_N: |
| case BUILT_IN_SYNC_FETCH_AND_NAND_N: |
| case BUILT_IN_SYNC_ADD_AND_FETCH_N: |
| case BUILT_IN_SYNC_SUB_AND_FETCH_N: |
| case BUILT_IN_SYNC_OR_AND_FETCH_N: |
| case BUILT_IN_SYNC_AND_AND_FETCH_N: |
| case BUILT_IN_SYNC_XOR_AND_FETCH_N: |
| case BUILT_IN_SYNC_NAND_AND_FETCH_N: |
| case BUILT_IN_SYNC_BOOL_COMPARE_AND_SWAP_N: |
| case BUILT_IN_SYNC_VAL_COMPARE_AND_SWAP_N: |
| case BUILT_IN_SYNC_LOCK_TEST_AND_SET_N: |
| case BUILT_IN_SYNC_LOCK_RELEASE_N: |
| { |
| /* The following are not _FETCH_OPs and must be accepted with |
| pointers to _Bool (or C++ bool). */ |
| if (fetch_op) |
| fetch_op = |
| (orig_code != BUILT_IN_SYNC_BOOL_COMPARE_AND_SWAP_N |
| && orig_code != BUILT_IN_SYNC_VAL_COMPARE_AND_SWAP_N |
| && orig_code != BUILT_IN_SYNC_LOCK_TEST_AND_SET_N |
| && orig_code != BUILT_IN_SYNC_LOCK_RELEASE_N); |
| |
| int n = sync_resolve_size (function, params, fetch_op); |
| tree new_function, first_param, result; |
| enum built_in_function fncode; |
| |
| if (n == 0) |
| return error_mark_node; |
| |
| fncode = (enum built_in_function)((int)orig_code + exact_log2 (n) + 1); |
| new_function = builtin_decl_explicit (fncode); |
| if (!sync_resolve_params (loc, function, new_function, params, |
| orig_format)) |
| return error_mark_node; |
| |
| first_param = (*params)[0]; |
| result = build_function_call_vec (loc, vNULL, new_function, params, |
| NULL); |
| if (result == error_mark_node) |
| return result; |
| if (orig_code != BUILT_IN_SYNC_BOOL_COMPARE_AND_SWAP_N |
| && orig_code != BUILT_IN_SYNC_LOCK_RELEASE_N |
| && orig_code != BUILT_IN_ATOMIC_STORE_N |
| && orig_code != BUILT_IN_ATOMIC_COMPARE_EXCHANGE_N) |
| result = sync_resolve_return (first_param, result, orig_format); |
| |
| if (fetch_op) |
| /* Prevent -Wunused-value warning. */ |
| TREE_USED (result) = true; |
| |
| /* If new_return is set, assign function to that expr and cast the |
| result to void since the generic interface returned void. */ |
| if (new_return) |
| { |
| /* Cast function result from I{1,2,4,8,16} to the required type. */ |
| result = build1 (VIEW_CONVERT_EXPR, TREE_TYPE (new_return), result); |
| result = build2 (MODIFY_EXPR, TREE_TYPE (new_return), new_return, |
| result); |
| TREE_SIDE_EFFECTS (result) = 1; |
| protected_set_expr_location (result, loc); |
| result = convert (void_type_node, result); |
| } |
| return result; |
| } |
| |
| default: |
| return NULL_TREE; |
| } |
| } |
| |
| /* vector_types_compatible_elements_p is used in type checks of vectors |
| values used as operands of binary operators. Where it returns true, and |
| the other checks of the caller succeed (being vector types in he first |
| place, and matching number of elements), we can just treat the types |
| as essentially the same. |
| Contrast with vector_targets_convertible_p, which is used for vector |
| pointer types, and vector_types_convertible_p, which will allow |
| language-specific matches under the control of flag_lax_vector_conversions, |
| and might still require a conversion. */ |
| /* True if vector types T1 and T2 can be inputs to the same binary |
| operator without conversion. |
| We don't check the overall vector size here because some of our callers |
| want to give different error messages when the vectors are compatible |
| except for the element count. */ |
| |
| bool |
| vector_types_compatible_elements_p (tree t1, tree t2) |
| { |
| bool opaque = TYPE_VECTOR_OPAQUE (t1) || TYPE_VECTOR_OPAQUE (t2); |
| t1 = TREE_TYPE (t1); |
| t2 = TREE_TYPE (t2); |
| |
| enum tree_code c1 = TREE_CODE (t1), c2 = TREE_CODE (t2); |
| |
| gcc_assert ((c1 == INTEGER_TYPE || c1 == REAL_TYPE || c1 == FIXED_POINT_TYPE) |
| && (c2 == INTEGER_TYPE || c2 == REAL_TYPE |
| || c2 == FIXED_POINT_TYPE)); |
| |
| t1 = c_common_signed_type (t1); |
| t2 = c_common_signed_type (t2); |
| /* Equality works here because c_common_signed_type uses |
| TYPE_MAIN_VARIANT. */ |
| if (t1 == t2) |
| return true; |
| if (opaque && c1 == c2 |
| && (c1 == INTEGER_TYPE || c1 == REAL_TYPE) |
| && TYPE_PRECISION (t1) == TYPE_PRECISION (t2)) |
| return true; |
| return false; |
| } |
| |
| /* Check for missing format attributes on function pointers. LTYPE is |
| the new type or left-hand side type. RTYPE is the old type or |
| right-hand side type. Returns TRUE if LTYPE is missing the desired |
| attribute. */ |
| |
| bool |
| check_missing_format_attribute (tree ltype, tree rtype) |
| { |
| tree const ttr = TREE_TYPE (rtype), ttl = TREE_TYPE (ltype); |
| tree ra; |
| |
| for (ra = TYPE_ATTRIBUTES (ttr); ra; ra = TREE_CHAIN (ra)) |
| if (is_attribute_p ("format", TREE_PURPOSE (ra))) |
| break; |
| if (ra) |
| { |
| tree la; |
| for (la = TYPE_ATTRIBUTES (ttl); la; la = TREE_CHAIN (la)) |
| if (is_attribute_p ("format", TREE_PURPOSE (la))) |
| break; |
| return !la; |
| } |
| else |
| return false; |
| } |
| |
| /* Setup a TYPE_DECL node as a typedef representation. |
| |
| X is a TYPE_DECL for a typedef statement. Create a brand new |
| ..._TYPE node (which will be just a variant of the existing |
| ..._TYPE node with identical properties) and then install X |
| as the TYPE_NAME of this brand new (duplicate) ..._TYPE node. |
| |
| The whole point here is to end up with a situation where each |
| and every ..._TYPE node the compiler creates will be uniquely |
| associated with AT MOST one node representing a typedef name. |
| This way, even though the compiler substitutes corresponding |
| ..._TYPE nodes for TYPE_DECL (i.e. "typedef name") nodes very |
| early on, later parts of the compiler can always do the reverse |
| translation and get back the corresponding typedef name. For |
| example, given: |
| |
| typedef struct S MY_TYPE; |
| MY_TYPE object; |
| |
| Later parts of the compiler might only know that `object' was of |
| type `struct S' if it were not for code just below. With this |
| code however, later parts of the compiler see something like: |
| |
| struct S' == struct S |
| typedef struct S' MY_TYPE; |
| struct S' object; |
| |
| And they can then deduce (from the node for type struct S') that |
| the original object declaration was: |
| |
| MY_TYPE object; |
| |
| Being able to do this is important for proper support of protoize, |
| and also for generating precise symbolic debugging information |
| which takes full account of the programmer's (typedef) vocabulary. |
| |
| Obviously, we don't want to generate a duplicate ..._TYPE node if |
| the TYPE_DECL node that we are now processing really represents a |
| standard built-in type. */ |
| |
| void |
| set_underlying_type (tree x) |
| { |
| if (x == error_mark_node) |
| return; |
| if (DECL_IS_BUILTIN (x) && TREE_CODE (TREE_TYPE (x)) != ARRAY_TYPE) |
| { |
| if (TYPE_NAME (TREE_TYPE (x)) == 0) |
| TYPE_NAME (TREE_TYPE (x)) = x; |
| } |
| else if (TREE_TYPE (x) != error_mark_node |
| && DECL_ORIGINAL_TYPE (x) == NULL_TREE) |
| { |
| tree tt = TREE_TYPE (x); |
| DECL_ORIGINAL_TYPE (x) = tt; |
| tt = build_variant_type_copy (tt); |
| TYPE_STUB_DECL (tt) = TYPE_STUB_DECL (DECL_ORIGINAL_TYPE (x)); |
| TYPE_NAME (tt) = x; |
| |
| /* Mark the type as used only when its type decl is decorated |
| with attribute unused. */ |
| if (lookup_attribute ("unused", DECL_ATTRIBUTES (x))) |
| TREE_USED (tt) = 1; |
| |
| TREE_TYPE (x) = tt; |
| } |
| } |
| |
| /* Record the types used by the current global variable declaration |
| being parsed, so that we can decide later to emit their debug info. |
| Those types are in types_used_by_cur_var_decl, and we are going to |
| store them in the types_used_by_vars_hash hash table. |
| DECL is the declaration of the global variable that has been parsed. */ |
| |
| void |
| record_types_used_by_current_var_decl (tree decl) |
| { |
| gcc_assert (decl && DECL_P (decl) && TREE_STATIC (decl)); |
| |
| while (types_used_by_cur_var_decl && !types_used_by_cur_var_decl->is_empty ()) |
| { |
| tree type = types_used_by_cur_var_decl->pop (); |
| types_used_by_var_decl_insert (type, decl); |
| } |
| } |
| |
| /* The C and C++ parsers both use vectors to hold function arguments. |
| For efficiency, we keep a cache of unused vectors. This is the |
| cache. */ |
| |
| typedef vec<tree, va_gc> *tree_gc_vec; |
| static GTY((deletable)) vec<tree_gc_vec, va_gc> *tree_vector_cache; |
| |
| /* Return a new vector from the cache. If the cache is empty, |
| allocate a new vector. These vectors are GC'ed, so it is OK if the |
| pointer is not released.. */ |
| |
| vec<tree, va_gc> * |
| make_tree_vector (void) |
| { |
| if (tree_vector_cache && !tree_vector_cache->is_empty ()) |
| return tree_vector_cache->pop (); |
| else |
| { |
| /* Passing 0 to vec::alloc returns NULL, and our callers require |
| that we always return a non-NULL value. The vector code uses |
| 4 when growing a NULL vector, so we do too. */ |
| vec<tree, va_gc> *v; |
| vec_alloc (v, 4); |
| return v; |
| } |
| } |
| |
| /* Release a vector of trees back to the cache. */ |
| |
| void |
| release_tree_vector (vec<tree, va_gc> *vec) |
| { |
| if (vec != NULL) |
| { |
| vec->truncate (0); |
| vec_safe_push (tree_vector_cache, vec); |
| } |
| } |
| |
| /* Get a new tree vector holding a single tree. */ |
| |
| vec<tree, va_gc> * |
| make_tree_vector_single (tree t) |
| { |
| vec<tree, va_gc> *ret = make_tree_vector (); |
| ret->quick_push (t); |
| return ret; |
| } |
| |
| /* Get a new tree vector of the TREE_VALUEs of a TREE_LIST chain. */ |
| |
| vec<tree, va_gc> * |
| make_tree_vector_from_list (tree list) |
| { |
| vec<tree, va_gc> *ret = make_tree_vector (); |
| for (; list; list = TREE_CHAIN (list)) |
| vec_safe_push (ret, TREE_VALUE (list)); |
| return ret; |
| } |
| |
| /* Get a new tree vector of the values of a CONSTRUCTOR. */ |
| |
| vec<tree, va_gc> * |
| make_tree_vector_from_ctor (tree ctor) |
| { |
| vec<tree,va_gc> *ret = make_tree_vector (); |
| vec_safe_reserve (ret, CONSTRUCTOR_NELTS (ctor)); |
| for (unsigned i = 0; i < CONSTRUCTOR_NELTS (ctor); ++i) |
| ret->quick_push (CONSTRUCTOR_ELT (ctor, i)->value); |
| return ret; |
| } |
| |
| /* Get a new tree vector which is a copy of an existing one. */ |
| |
| vec<tree, va_gc> * |
| make_tree_vector_copy (const vec<tree, va_gc> *orig) |
| { |
| vec<tree, va_gc> *ret; |
| unsigned int ix; |
| tree t; |
| |
| ret = make_tree_vector (); |
| vec_safe_reserve (ret, vec_safe_length (orig)); |
| FOR_EACH_VEC_SAFE_ELT (orig, ix, t) |
| ret->quick_push (t); |
| return ret; |
| } |
| |
| /* Return true if KEYWORD starts a type specifier. */ |
| |
| bool |
| keyword_begins_type_specifier (enum rid keyword) |
| { |
| switch (keyword) |
| { |
| case RID_AUTO_TYPE: |
| case RID_INT: |
| case RID_CHAR: |
| case RID_FLOAT: |
| case RID_DOUBLE: |
| case RID_VOID: |
| case RID_UNSIGNED: |
| case RID_LONG: |
| case RID_SHORT: |
| case RID_SIGNED: |
| CASE_RID_FLOATN_NX: |
| case RID_DFLOAT32: |
| case RID_DFLOAT64: |
| case RID_DFLOAT128: |
| case RID_FRACT: |
| case RID_ACCUM: |
| case RID_BOOL: |
| case RID_WCHAR: |
| case RID_CHAR16: |
| case RID_CHAR32: |
| case RID_SAT: |
| case RID_COMPLEX: |
| case RID_TYPEOF: |
| case RID_STRUCT: |
| case RID_CLASS: |
| case RID_UNION: |
| case RID_ENUM: |
| return true; |
| default: |
| if (keyword >= RID_FIRST_INT_N |
| && keyword < RID_FIRST_INT_N + NUM_INT_N_ENTS |
| && int_n_enabled_p[keyword-RID_FIRST_INT_N]) |
| return true; |
| return false; |
| } |
| } |
| |
| /* Return true if KEYWORD names a type qualifier. */ |
| |
| bool |
| keyword_is_type_qualifier (enum rid keyword) |
| { |
| switch (keyword) |
| { |
| case RID_CONST: |
| case RID_VOLATILE: |
| case RID_RESTRICT: |
| case RID_ATOMIC: |
| return true; |
| default: |
| return false; |
| } |
| } |
| |
| /* Return true if KEYWORD names a storage class specifier. |
| |
| RID_TYPEDEF is not included in this list despite `typedef' being |
| listed in C99 6.7.1.1. 6.7.1.3 indicates that `typedef' is listed as |
| such for syntactic convenience only. */ |
| |
| bool |
| keyword_is_storage_class_specifier (enum rid keyword) |
| { |
| switch (keyword) |
| { |
| case RID_STATIC: |
| case RID_EXTERN: |
| case RID_REGISTER: |
| case RID_AUTO: |
| case RID_MUTABLE: |
| case RID_THREAD: |
| return true; |
| default: |
| return false; |
| } |
| } |
| |
| /* Return true if KEYWORD names a function-specifier [dcl.fct.spec]. */ |
| |
| static bool |
| keyword_is_function_specifier (enum rid keyword) |
| { |
| switch (keyword) |
| { |
| case RID_INLINE: |
| case RID_NORETURN: |
| case RID_VIRTUAL: |
| case RID_EXPLICIT: |
| return true; |
| default: |
| return false; |
| } |
| } |
| |
| /* Return true if KEYWORD names a decl-specifier [dcl.spec] or a |
| declaration-specifier (C99 6.7). */ |
| |
| bool |
| keyword_is_decl_specifier (enum rid keyword) |
| { |
| if (keyword_is_storage_class_specifier (keyword) |
| || keyword_is_type_qualifier (keyword) |
| || keyword_is_function_specifier (keyword)) |
| return true; |
| |
| switch (keyword) |
| { |
| case RID_TYPEDEF: |
| case RID_FRIEND: |
| case RID_CONSTEXPR: |
| return true; |
| default: |
| return false; |
| } |
| } |
| |
| /* Initialize language-specific-bits of tree_contains_struct. */ |
| |
| void |
| c_common_init_ts (void) |
| { |
| MARK_TS_TYPED (C_MAYBE_CONST_EXPR); |
| MARK_TS_TYPED (EXCESS_PRECISION_EXPR); |
| MARK_TS_TYPED (ARRAY_NOTATION_REF); |
| } |
| |
| /* Build a user-defined numeric literal out of an integer constant type VALUE |
| with identifier SUFFIX. */ |
| |
| tree |
| build_userdef_literal (tree suffix_id, tree value, |
| enum overflow_type overflow, tree num_string) |
| { |
| tree literal = make_node (USERDEF_LITERAL); |
| USERDEF_LITERAL_SUFFIX_ID (literal) = suffix_id; |
| USERDEF_LITERAL_VALUE (literal) = value; |
| USERDEF_LITERAL_OVERFLOW (literal) = overflow; |
| USERDEF_LITERAL_NUM_STRING (literal) = num_string; |
| return literal; |
| } |
| |
| /* For vector[index], convert the vector to an array of the underlying type. |
| Return true if the resulting ARRAY_REF should not be an lvalue. */ |
| |
| bool |
| convert_vector_to_array_for_subscript (location_t loc, |
| tree *vecp, tree index) |
| { |
| bool ret = false; |
| if (VECTOR_TYPE_P (TREE_TYPE (*vecp))) |
| { |
| tree type = TREE_TYPE (*vecp); |
| |
| ret = !lvalue_p (*vecp); |
| |
| if (TREE_CODE (index) == INTEGER_CST) |
| if (!tree_fits_uhwi_p (index) |
| || tree_to_uhwi (index) >= TYPE_VECTOR_SUBPARTS (type)) |
| warning_at (loc, OPT_Warray_bounds, "index value is out of bound"); |
| |
| /* We are building an ARRAY_REF so mark the vector as addressable |
| to not run into the gimplifiers premature setting of DECL_GIMPLE_REG_P |
| for function parameters. */ |
| c_common_mark_addressable_vec (*vecp); |
| |
| *vecp = build1 (VIEW_CONVERT_EXPR, |
| build_array_type_nelts (TREE_TYPE (type), |
| TYPE_VECTOR_SUBPARTS (type)), |
| *vecp); |
| } |
| return ret; |
| } |
| |
| /* Determine which of the operands, if any, is a scalar that needs to be |
| converted to a vector, for the range of operations. */ |
| enum stv_conv |
| scalar_to_vector (location_t loc, enum tree_code code, tree op0, tree op1, |
| bool complain) |
| { |
| tree type0 = TREE_TYPE (op0); |
| tree type1 = TREE_TYPE (op1); |
| bool integer_only_op = false; |
| enum stv_conv ret = stv_firstarg; |
| |
| gcc_assert (VECTOR_TYPE_P (type0) || VECTOR_TYPE_P (type1)); |
| switch (code) |
| { |
| /* Most GENERIC binary expressions require homogeneous arguments. |
| LSHIFT_EXPR and RSHIFT_EXPR are exceptions and accept a first |
| argument that is a vector and a second one that is a scalar, so |
| we never return stv_secondarg for them. */ |
| case RSHIFT_EXPR: |
| case LSHIFT_EXPR: |
| if (TREE_CODE (type0) == INTEGER_TYPE |
| && TREE_CODE (TREE_TYPE (type1)) == INTEGER_TYPE) |
| { |
| if (unsafe_conversion_p (loc, TREE_TYPE (type1), op0, false)) |
| { |
| if (complain) |
| error_at (loc, "conversion of scalar %qT to vector %qT " |
| "involves truncation", type0, type1); |
| return stv_error; |
| } |
| else |
| return stv_firstarg; |
| } |
| break; |
| |
| case BIT_IOR_EXPR: |
| case BIT_XOR_EXPR: |
| case BIT_AND_EXPR: |
| integer_only_op = true; |
| /* fall through */ |
| |
| case VEC_COND_EXPR: |
| |
| case PLUS_EXPR: |
| case MINUS_EXPR: |
| case MULT_EXPR: |
| case TRUNC_DIV_EXPR: |
| case CEIL_DIV_EXPR: |
| case FLOOR_DIV_EXPR: |
| case ROUND_DIV_EXPR: |
| case EXACT_DIV_EXPR: |
| case TRUNC_MOD_EXPR: |
| case FLOOR_MOD_EXPR: |
| case RDIV_EXPR: |
| case EQ_EXPR: |
| case NE_EXPR: |
| case LE_EXPR: |
| case GE_EXPR: |
| case LT_EXPR: |
| case GT_EXPR: |
| /* What about UNLT_EXPR? */ |
| if (VECTOR_TYPE_P (type0)) |
| { |
| ret = stv_secondarg; |
| std::swap (type0, type1); |
| std::swap (op0, op1); |
| } |
| |
| if (TREE_CODE (type0) == INTEGER_TYPE |
| && TREE_CODE (TREE_TYPE (type1)) == INTEGER_TYPE) |
| { |
| if (unsafe_conversion_p (loc, TREE_TYPE (type1), op0, false)) |
| { |
| if (complain) |
| error_at (loc, "conversion of scalar %qT to vector %qT " |
| "involves truncation", type0, type1); |
| return stv_error; |
| } |
| return ret; |
| } |
| else if (!integer_only_op |
| /* Allow integer --> real conversion if safe. */ |
| && (TREE_CODE (type0) == REAL_TYPE |
| || TREE_CODE (type0) == INTEGER_TYPE) |
| && SCALAR_FLOAT_TYPE_P (TREE_TYPE (type1))) |
| { |
| if (unsafe_conversion_p (loc, TREE_TYPE (type1), op0, false)) |
| { |
| if (complain) |
| error_at (loc, "conversion of scalar %qT to vector %qT " |
| "involves truncation", type0, type1); |
| return stv_error; |
| } |
| return ret; |
| } |
| default: |
| break; |
| } |
| |
| return stv_nothing; |
| } |
| |
| /* Return the alignment of std::max_align_t. |
| |
| [support.types.layout] The type max_align_t is a POD type whose alignment |
| requirement is at least as great as that of every scalar type, and whose |
| alignment requirement is supported in every context. */ |
| |
| unsigned |
| max_align_t_align () |
| { |
| unsigned int max_align = MAX (TYPE_ALIGN (long_long_integer_type_node), |
| TYPE_ALIGN (long_double_type_node)); |
| if (float128_type_node != NULL_TREE) |
| max_align = MAX (max_align, TYPE_ALIGN (float128_type_node)); |
| return max_align; |
| } |
| |
| /* Return true iff ALIGN is an integral constant that is a fundamental |
| alignment, as defined by [basic.align] in the c++-11 |
| specifications. |
| |
| That is: |
| |
| [A fundamental alignment is represented by an alignment less than or |
| equal to the greatest alignment supported by the implementation |
| in all contexts, which is equal to alignof(max_align_t)]. */ |
| |
| bool |
| cxx_fundamental_alignment_p (unsigned align) |
| { |
| return (align <= max_align_t_align ()); |
| } |
| |
| /* Return true if T is a pointer to a zero-sized aggregate. */ |
| |
| bool |
| pointer_to_zero_sized_aggr_p (tree t) |
| { |
| if (!POINTER_TYPE_P (t)) |
| return false; |
| t = TREE_TYPE (t); |
| return (TYPE_SIZE (t) && integer_zerop (TYPE_SIZE (t))); |
| } |
| |
| /* For an EXPR of a FUNCTION_TYPE that references a GCC built-in function |
| with no library fallback or for an ADDR_EXPR whose operand is such type |
| issues an error pointing to the location LOC. |
| Returns true when the expression has been diagnosed and false |
| otherwise. */ |
| |
| bool |
| reject_gcc_builtin (const_tree expr, location_t loc /* = UNKNOWN_LOCATION */) |
| { |
| if (TREE_CODE (expr) == ADDR_EXPR) |
| expr = TREE_OPERAND (expr, 0); |
| |
| if (TREE_TYPE (expr) |
| && TREE_CODE (TREE_TYPE (expr)) == FUNCTION_TYPE |
| && TREE_CODE (expr) == FUNCTION_DECL |
| /* The intersection of DECL_BUILT_IN and DECL_IS_BUILTIN avoids |
| false positives for user-declared built-ins such as abs or |
| strlen, and for C++ operators new and delete. |
| The c_decl_implicit() test avoids false positives for implicitly |
| declared built-ins with library fallbacks (such as abs). */ |
| && DECL_BUILT_IN (expr) |
| && DECL_IS_BUILTIN (expr) |
| && !c_decl_implicit (expr) |
| && !DECL_ASSEMBLER_NAME_SET_P (expr)) |
| { |
| if (loc == UNKNOWN_LOCATION) |
| loc = EXPR_LOC_OR_LOC (expr, input_location); |
| |
| /* Reject arguments that are built-in functions with |
| no library fallback. */ |
| error_at (loc, "built-in function %qE must be directly called", expr); |
| |
| return true; |
| } |
| |
| return false; |
| } |
| |
| /* Check if array size calculations overflow or if the array covers more |
| than half of the address space. Return true if the size of the array |
| is valid, false otherwise. TYPE is the type of the array and NAME is |
| the name of the array, or NULL_TREE for unnamed arrays. */ |
| |
| bool |
| valid_array_size_p (location_t loc, tree type, tree name) |
| { |
| if (type != error_mark_node |
| && COMPLETE_TYPE_P (type) |
| && TREE_CODE (TYPE_SIZE_UNIT (type)) == INTEGER_CST |
| && !valid_constant_size_p (TYPE_SIZE_UNIT (type))) |
| { |
| if (name) |
| error_at (loc, "size of array %qE is too large", name); |
| else |
| error_at (loc, "size of unnamed array is too large"); |
| return false; |
| } |
| return true; |
| } |
| |
| /* Read SOURCE_DATE_EPOCH from environment to have a deterministic |
| timestamp to replace embedded current dates to get reproducible |
| results. Returns -1 if SOURCE_DATE_EPOCH is not defined. */ |
| |
| time_t |
| cb_get_source_date_epoch (cpp_reader *pfile ATTRIBUTE_UNUSED) |
| { |
| char *source_date_epoch; |
| int64_t epoch; |
| char *endptr; |
| |
| source_date_epoch = getenv ("SOURCE_DATE_EPOCH"); |
| if (!source_date_epoch) |
| return (time_t) -1; |
| |
| errno = 0; |
| #if defined(INT64_T_IS_LONG) |
| epoch = strtol (source_date_epoch, &endptr, 10); |
| #else |
| epoch = strtoll (source_date_epoch, &endptr, 10); |
| #endif |
| if (errno != 0 || endptr == source_date_epoch || *endptr != '\0' |
| || epoch < 0 || epoch > MAX_SOURCE_DATE_EPOCH) |
| { |
| error_at (input_location, "environment variable SOURCE_DATE_EPOCH must " |
| "expand to a non-negative integer less than or equal to %wd", |
| MAX_SOURCE_DATE_EPOCH); |
| return (time_t) -1; |
| } |
| |
| return (time_t) epoch; |
| } |
| |
| /* Callback for libcpp for offering spelling suggestions for misspelled |
| directives. GOAL is an unrecognized string; CANDIDATES is a |
| NULL-terminated array of candidate strings. Return the closest |
| match to GOAL within CANDIDATES, or NULL if none are good |
| suggestions. */ |
| |
| const char * |
| cb_get_suggestion (cpp_reader *, const char *goal, |
| const char *const *candidates) |
| { |
| best_match<const char *, const char *> bm (goal); |
| while (*candidates) |
| bm.consider (*candidates++); |
| return bm.get_best_meaningful_candidate (); |
| } |
| |
| /* Return the latice point which is the wider of the two FLT_EVAL_METHOD |
| modes X, Y. This isn't just >, as the FLT_EVAL_METHOD values added |
| by C TS 18661-3 for interchange types that are computed in their |
| native precision are larger than the C11 values for evaluating in the |
| precision of float/double/long double. If either mode is |
| FLT_EVAL_METHOD_UNPREDICTABLE, return that. */ |
| |
| enum flt_eval_method |
| excess_precision_mode_join (enum flt_eval_method x, |
| enum flt_eval_method y) |
| { |
| if (x == FLT_EVAL_METHOD_UNPREDICTABLE |
| || y == FLT_EVAL_METHOD_UNPREDICTABLE) |
| return FLT_EVAL_METHOD_UNPREDICTABLE; |
| |
| /* GCC only supports one interchange type right now, _Float16. If |
| we're evaluating _Float16 in 16-bit precision, then flt_eval_method |
| will be FLT_EVAL_METHOD_PROMOTE_TO_FLOAT16. */ |
| if (x == FLT_EVAL_METHOD_PROMOTE_TO_FLOAT16) |
| return y; |
| if (y == FLT_EVAL_METHOD_PROMOTE_TO_FLOAT16) |
| return x; |
| |
| /* Other values for flt_eval_method are directly comparable, and we want |
| the maximum. */ |
| return MAX (x, y); |
| } |
| |
| /* Return the value that should be set for FLT_EVAL_METHOD in the |
| context of ISO/IEC TS 18861-3. |
| |
| This relates to the effective excess precision seen by the user, |
| which is the join point of the precision the target requests for |
| -fexcess-precision={standard,fast} and the implicit excess precision |
| the target uses. */ |
| |
| static enum flt_eval_method |
| c_ts18661_flt_eval_method (void) |
| { |
| enum flt_eval_method implicit |
| = targetm.c.excess_precision (EXCESS_PRECISION_TYPE_IMPLICIT); |
| |
| enum excess_precision_type flag_type |
| = (flag_excess_precision_cmdline == EXCESS_PRECISION_STANDARD |
| ? EXCESS_PRECISION_TYPE_STANDARD |
| : EXCESS_PRECISION_TYPE_FAST); |
| |
| enum flt_eval_method requested |
| = targetm.c.excess_precision (flag_type); |
| |
| return excess_precision_mode_join (implicit, requested); |
| } |
| |
| /* As c_cpp_ts18661_flt_eval_method, but clamps the expected values to |
| those that were permitted by C11. That is to say, eliminates |
| FLT_EVAL_METHOD_PROMOTE_TO_FLOAT16. */ |
| |
| static enum flt_eval_method |
| c_c11_flt_eval_method (void) |
| { |
| return excess_precision_mode_join (c_ts18661_flt_eval_method (), |
| FLT_EVAL_METHOD_PROMOTE_TO_FLOAT); |
| } |
| |
| /* Return the value that should be set for FLT_EVAL_METHOD. |
| MAYBE_C11_ONLY_P is TRUE if we should check |
| FLAG_PERMITTED_EVAL_METHODS as to whether we should limit the possible |
| values we can return to those from C99/C11, and FALSE otherwise. |
| See the comments on c_ts18661_flt_eval_method for what value we choose |
| to set here. */ |
| |
| int |
| c_flt_eval_method (bool maybe_c11_only_p) |
| { |
| if (maybe_c11_only_p |
| && flag_permitted_flt_eval_methods |
| == PERMITTED_FLT_EVAL_METHODS_C11) |
| return c_c11_flt_eval_method (); |
| else |
| return c_ts18661_flt_eval_method (); |
| } |
| |
| #include "gt-c-family-c-common.h" |