blob: 4c34949a97f1d9d2e4ad48ddfc582396b27bdd25 [file] [log] [blame]
/* Subroutines used for code generation on Renesas RL78 processors.
Copyright (C) 2011-2021 Free Software Foundation, Inc.
Contributed by Red Hat.
This file is part of GCC.
GCC is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3, or (at your option)
any later version.
GCC is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with GCC; see the file COPYING3. If not see
<http://www.gnu.org/licenses/>. */
#define IN_TARGET_CODE 1
#include "config.h"
#include "system.h"
#include "coretypes.h"
#include "backend.h"
#include "target.h"
#include "rtl.h"
#include "tree.h"
#include "df.h"
#include "memmodel.h"
#include "tm_p.h"
#include "stringpool.h"
#include "attribs.h"
#include "optabs.h"
#include "emit-rtl.h"
#include "recog.h"
#include "diagnostic-core.h"
#include "varasm.h"
#include "stor-layout.h"
#include "calls.h"
#include "output.h"
#include "insn-attr.h"
#include "explow.h"
#include "expr.h"
#include "reload.h"
#include "cfgrtl.h"
#include "langhooks.h"
#include "tree-pass.h"
#include "context.h"
#include "tm-constrs.h" /* for satisfies_constraint_*(). */
#include "builtins.h"
/* This file should be included last. */
#include "target-def.h"
static inline bool is_interrupt_func (const_tree decl);
static inline bool is_brk_interrupt_func (const_tree decl);
static void rl78_reorg (void);
static const char *rl78_strip_name_encoding (const char *);
static const char *rl78_strip_nonasm_name_encoding (const char *);
static section * rl78_select_section (tree, int, unsigned HOST_WIDE_INT);
/* Debugging statements are tagged with DEBUG0 only so that they can
be easily enabled individually, by replacing the '0' with '1' as
needed. */
#define DEBUG0 0
#define DEBUG1 1
/* REGISTER_NAMES has the names for individual 8-bit registers, but
these have the names we need to use when referring to 16-bit
register pairs. */
static const char * const word_regnames[] =
{
"ax", "AX", "bc", "BC", "de", "DE", "hl", "HL",
"r8", "r9", "r10", "r11", "r12", "r13", "r14", "r15",
"r16", "r17", "r18", "r19", "r20", "r21", "r22", "r23",
"r24", "r25", "r26", "r27", "r28", "r29", "r30", "r31",
"sp", "ap", "psw", "es", "cs"
};
/* used by rl78_addsi3_internal for formatting insns output */
static char fmt_buffer[1024];
/* Structure for G13 MDUC registers. */
struct mduc_reg_type
{
unsigned int address;
enum machine_mode mode;
};
struct mduc_reg_type mduc_regs[] =
{
{0xf00e8, E_QImode},
{0xffff0, E_HImode},
{0xffff2, E_HImode},
{0xf2224, E_HImode},
{0xf00e0, E_HImode},
{0xf00e2, E_HImode}
};
struct GTY(()) machine_function
{
/* If set, the rest of the fields have been computed. */
int computed;
/* Which register pairs need to be pushed in the prologue. */
int need_to_push [FIRST_PSEUDO_REGISTER / 2];
/* These fields describe the frame layout... */
/* arg pointer */
/* 4 bytes for saved PC */
int framesize_regs;
/* frame pointer */
int framesize_locals;
int framesize_outgoing;
/* stack pointer */
int framesize;
/* If set, recog is allowed to match against the "real" patterns. */
int real_insns_ok;
/* If set, recog is allowed to match against the "virtual" patterns. */
int virt_insns_ok;
/* Set if the current function needs to clean up any trampolines. */
int trampolines_used;
/* True if the ES register is used and hence
needs to be saved inside interrupt handlers. */
bool uses_es;
};
/* This is our init_machine_status, as set in
rl78_option_override. */
static struct machine_function *
rl78_init_machine_status (void)
{
struct machine_function *m;
m = ggc_cleared_alloc<machine_function> ();
m->virt_insns_ok = 1;
return m;
}
/* This pass converts virtual instructions using virtual registers, to
real instructions using real registers. Rather than run it as
reorg, we reschedule it before vartrack to help with debugging. */
namespace
{
const pass_data pass_data_rl78_devirt =
{
RTL_PASS, /* type */
"devirt", /* name */
OPTGROUP_NONE, /* optinfo_flags */
TV_MACH_DEP, /* tv_id */
0, /* properties_required */
0, /* properties_provided */
0, /* properties_destroyed */
0, /* todo_flags_start */
0, /* todo_flags_finish */
};
class pass_rl78_devirt : public rtl_opt_pass
{
public:
pass_rl78_devirt (gcc::context *ctxt)
: rtl_opt_pass (pass_data_rl78_devirt, ctxt)
{
}
/* opt_pass methods: */
virtual unsigned int execute (function *)
{
rl78_reorg ();
return 0;
}
};
} // anon namespace
rtl_opt_pass *
make_pass_rl78_devirt (gcc::context *ctxt)
{
return new pass_rl78_devirt (ctxt);
}
/* Redundant move elimination pass. Must be run after the basic block
reordering pass for the best effect. */
static unsigned int
move_elim_pass (void)
{
rtx_insn *insn, *ninsn;
rtx prev = NULL_RTX;
for (insn = get_insns (); insn; insn = ninsn)
{
rtx set;
ninsn = next_nonnote_nondebug_insn (insn);
if ((set = single_set (insn)) == NULL_RTX)
{
prev = NULL_RTX;
continue;
}
/* If we have two SET insns in a row (without anything
between them) and the source of the second one is the
destination of the first one, and vice versa, then we
can eliminate the second SET. */
if (prev
&& rtx_equal_p (SET_DEST (prev), SET_SRC (set))
&& rtx_equal_p (SET_DEST (set), SET_SRC (prev))
/* ... and none of the operands are volatile. */
&& ! volatile_refs_p (SET_SRC (prev))
&& ! volatile_refs_p (SET_DEST (prev))
&& ! volatile_refs_p (SET_SRC (set))
&& ! volatile_refs_p (SET_DEST (set)))
{
if (dump_file)
fprintf (dump_file, " Delete insn %d because it is redundant\n",
INSN_UID (insn));
delete_insn (insn);
prev = NULL_RTX;
}
else
prev = set;
}
if (dump_file)
print_rtl_with_bb (dump_file, get_insns (), TDF_NONE);
return 0;
}
namespace
{
const pass_data pass_data_rl78_move_elim =
{
RTL_PASS, /* type */
"move_elim", /* name */
OPTGROUP_NONE, /* optinfo_flags */
TV_MACH_DEP, /* tv_id */
0, /* properties_required */
0, /* properties_provided */
0, /* properties_destroyed */
0, /* todo_flags_start */
0, /* todo_flags_finish */
};
class pass_rl78_move_elim : public rtl_opt_pass
{
public:
pass_rl78_move_elim (gcc::context *ctxt)
: rtl_opt_pass (pass_data_rl78_move_elim, ctxt)
{
}
/* opt_pass methods: */
virtual unsigned int execute (function *) { return move_elim_pass (); }
};
} // anon namespace
rtl_opt_pass *
make_pass_rl78_move_elim (gcc::context *ctxt)
{
return new pass_rl78_move_elim (ctxt);
}
#undef TARGET_ASM_FILE_START
#define TARGET_ASM_FILE_START rl78_asm_file_start
static void
rl78_asm_file_start (void)
{
int i;
if (TARGET_G10)
{
/* The memory used is 0xffec8 to 0xffedf; real registers are in
0xffee0 to 0xffee7. */
for (i = 8; i < 32; i++)
fprintf (asm_out_file, "r%d\t=\t0x%x\n", i, 0xffec0 + i);
}
else
{
for (i = 0; i < 8; i++)
{
fprintf (asm_out_file, "r%d\t=\t0x%x\n", 8 + i, 0xffef0 + i);
fprintf (asm_out_file, "r%d\t=\t0x%x\n", 16 + i, 0xffee8 + i);
fprintf (asm_out_file, "r%d\t=\t0x%x\n", 24 + i, 0xffee0 + i);
}
}
opt_pass *rl78_devirt_pass = make_pass_rl78_devirt (g);
struct register_pass_info rl78_devirt_info =
{
rl78_devirt_pass,
"pro_and_epilogue",
1,
PASS_POS_INSERT_BEFORE
};
opt_pass *rl78_move_elim_pass = make_pass_rl78_move_elim (g);
struct register_pass_info rl78_move_elim_info =
{
rl78_move_elim_pass,
"bbro",
1,
PASS_POS_INSERT_AFTER
};
register_pass (& rl78_devirt_info);
register_pass (& rl78_move_elim_info);
}
void
rl78_output_symbol_ref (FILE * file, rtx sym)
{
tree type = SYMBOL_REF_DECL (sym);
const char *str = XSTR (sym, 0);
if (str[0] == '*')
{
fputs (str + 1, file);
}
else
{
str = rl78_strip_nonasm_name_encoding (str);
if (type && TREE_CODE (type) == FUNCTION_DECL)
{
fprintf (file, "%%code(");
assemble_name (file, str);
fprintf (file, ")");
}
else
assemble_name (file, str);
}
}
#undef TARGET_OPTION_OVERRIDE
#define TARGET_OPTION_OVERRIDE rl78_option_override
#define MUST_SAVE_MDUC_REGISTERS \
(TARGET_SAVE_MDUC_REGISTERS \
&& (is_interrupt_func (NULL_TREE)) && RL78_MUL_G13)
static void
rl78_option_override (void)
{
flag_omit_frame_pointer = 1;
flag_no_function_cse = 1;
flag_split_wide_types = 0;
init_machine_status = rl78_init_machine_status;
if (TARGET_ALLREGS)
{
int i;
for (i = 24; i < 32; i++)
fixed_regs[i] = 0;
}
if (TARGET_ES0
&& strcmp (lang_hooks.name, "GNU C")
&& strcmp (lang_hooks.name, "GNU C11")
&& strcmp (lang_hooks.name, "GNU C17")
&& strcmp (lang_hooks.name, "GNU C2X")
&& strcmp (lang_hooks.name, "GNU C89")
&& strcmp (lang_hooks.name, "GNU C99")
/* Compiling with -flto results in a language of GNU GIMPLE being used... */
&& strcmp (lang_hooks.name, "GNU GIMPLE"))
/* Address spaces are currently only supported by C. */
error ("%<-mes0%> can only be used with C");
if (TARGET_SAVE_MDUC_REGISTERS && !(TARGET_G13 || RL78_MUL_G13))
warning (0, "mduc registers only saved for G13 target");
switch (rl78_cpu_type)
{
case CPU_UNINIT:
rl78_cpu_type = CPU_G14;
if (rl78_mul_type == MUL_UNINIT)
rl78_mul_type = MUL_NONE;
break;
case CPU_G10:
switch (rl78_mul_type)
{
case MUL_UNINIT: rl78_mul_type = MUL_NONE; break;
case MUL_NONE: break;
case MUL_G13: error ("%<-mmul=g13%> cannot be used with "
"%<-mcpu=g10%>"); break;
case MUL_G14: error ("%<-mmul=g14%> cannot be used with "
"%<-mcpu=g10%>"); break;
}
break;
case CPU_G13:
switch (rl78_mul_type)
{
case MUL_UNINIT: rl78_mul_type = MUL_G13; break;
case MUL_NONE: break;
case MUL_G13: break;
/* The S2 core does not have mul/div instructions. */
case MUL_G14: error ("%<-mmul=g14%> cannot be used with "
"%<-mcpu=g13%>"); break;
}
break;
case CPU_G14:
switch (rl78_mul_type)
{
case MUL_UNINIT: rl78_mul_type = MUL_G14; break;
case MUL_NONE: break;
case MUL_G14: break;
/* The G14 core does not have the hardware multiply peripheral used by the
G13 core, hence you cannot use G13 multipliy routines on G14 hardware. */
case MUL_G13: error ("%<-mmul=g13%> cannot be used with "
"%<-mcpu=g14%>"); break;
}
break;
}
}
/* Most registers are 8 bits. Some are 16 bits because, for example,
gcc doesn't like dealing with $FP as a register pair (the second
half of $fp is also 2 to keep reload happy wrt register pairs, but
no register class includes it). This table maps register numbers
to size in bytes. */
static const int register_sizes[] =
{
1, 1, 1, 1, 1, 1, 1, 1,
1, 1, 1, 1, 1, 1, 1, 1,
1, 1, 1, 1, 1, 1, 2, 2,
1, 1, 1, 1, 1, 1, 1, 1,
2, 2, 1, 1, 1
};
/* Predicates used in the MD patterns. This one is true when virtual
insns may be matched, which typically means before (or during) the
devirt pass. */
bool
rl78_virt_insns_ok (void)
{
if (cfun)
return cfun->machine->virt_insns_ok;
return true;
}
/* Predicates used in the MD patterns. This one is true when real
insns may be matched, which typically means after (or during) the
devirt pass. */
bool
rl78_real_insns_ok (void)
{
if (cfun)
return cfun->machine->real_insns_ok;
return false;
}
#undef TARGET_HARD_REGNO_NREGS
#define TARGET_HARD_REGNO_NREGS rl78_hard_regno_nregs
static unsigned int
rl78_hard_regno_nregs (unsigned int regno, machine_mode mode)
{
int rs = register_sizes[regno];
if (rs < 1)
rs = 1;
return ((GET_MODE_SIZE (mode) + rs - 1) / rs);
}
#undef TARGET_HARD_REGNO_MODE_OK
#define TARGET_HARD_REGNO_MODE_OK rl78_hard_regno_mode_ok
static bool
rl78_hard_regno_mode_ok (unsigned int regno, machine_mode mode)
{
int s = GET_MODE_SIZE (mode);
if (s < 1)
return false;
/* These are not to be used by gcc. */
if (regno == 23 || regno == ES_REG || regno == CS_REG)
return false;
/* $fp can always be accessed as a 16-bit value. */
if (regno == FP_REG && s == 2)
return true;
if (regno < SP_REG)
{
/* Since a reg-reg move is really a reg-mem move, we must
enforce alignment. */
if (s > 1 && (regno % 2))
return false;
return true;
}
if (s == CC_REGNUM)
return (mode == BImode);
/* All other registers must be accessed in their natural sizes. */
if (s == register_sizes [regno])
return true;
return false;
}
#undef TARGET_MODES_TIEABLE_P
#define TARGET_MODES_TIEABLE_P rl78_modes_tieable_p
static bool
rl78_modes_tieable_p (machine_mode mode1, machine_mode mode2)
{
return ((GET_MODE_CLASS (mode1) == MODE_FLOAT
|| GET_MODE_CLASS (mode1) == MODE_COMPLEX_FLOAT)
== (GET_MODE_CLASS (mode2) == MODE_FLOAT
|| GET_MODE_CLASS (mode2) == MODE_COMPLEX_FLOAT));
}
/* Simplify_gen_subreg() doesn't handle memory references the way we
need it to below, so we use this function for when we must get a
valid subreg in a "natural" state. */
static rtx
rl78_subreg (machine_mode mode, rtx r, machine_mode omode, int byte)
{
if (GET_CODE (r) == MEM)
return adjust_address (r, mode, byte);
else
return simplify_gen_subreg (mode, r, omode, byte);
}
/* Used by movsi. Split SImode moves into two HImode moves, using
appropriate patterns for the upper and lower halves of symbols. */
void
rl78_expand_movsi (rtx *operands)
{
rtx op00, op02, op10, op12;
op00 = rl78_subreg (HImode, operands[0], SImode, 0);
op02 = rl78_subreg (HImode, operands[0], SImode, 2);
if (GET_CODE (operands[1]) == CONST
|| GET_CODE (operands[1]) == SYMBOL_REF)
{
op10 = gen_rtx_ZERO_EXTRACT (HImode, operands[1], GEN_INT (16), GEN_INT (0));
op10 = gen_rtx_CONST (HImode, op10);
op12 = gen_rtx_ZERO_EXTRACT (HImode, operands[1], GEN_INT (16), GEN_INT (16));
op12 = gen_rtx_CONST (HImode, op12);
}
else
{
op10 = rl78_subreg (HImode, operands[1], SImode, 0);
op12 = rl78_subreg (HImode, operands[1], SImode, 2);
}
if (rtx_equal_p (operands[0], operands[1]))
;
else if (rtx_equal_p (op00, op12))
{
emit_move_insn (op02, op12);
emit_move_insn (op00, op10);
}
else
{
emit_move_insn (op00, op10);
emit_move_insn (op02, op12);
}
}
/* Generate code to move an SImode value. */
void
rl78_split_movsi (rtx *operands, machine_mode omode)
{
rtx op00, op02, op10, op12;
op00 = rl78_subreg (HImode, operands[0], omode, 0);
op02 = rl78_subreg (HImode, operands[0], omode, 2);
if (GET_CODE (operands[1]) == CONST
|| GET_CODE (operands[1]) == SYMBOL_REF)
{
op10 = gen_rtx_ZERO_EXTRACT (HImode, operands[1], GEN_INT (16), GEN_INT (0));
op10 = gen_rtx_CONST (HImode, op10);
op12 = gen_rtx_ZERO_EXTRACT (HImode, operands[1], GEN_INT (16), GEN_INT (16));
op12 = gen_rtx_CONST (HImode, op12);
}
else
{
op10 = rl78_subreg (HImode, operands[1], omode, 0);
op12 = rl78_subreg (HImode, operands[1], omode, 2);
}
if (rtx_equal_p (operands[0], operands[1]))
;
else if (rtx_equal_p (op00, op12))
{
operands[2] = op02;
operands[4] = op12;
operands[3] = op00;
operands[5] = op10;
}
else
{
operands[2] = op00;
operands[4] = op10;
operands[3] = op02;
operands[5] = op12;
}
}
void
rl78_split_movdi (rtx *operands, enum machine_mode omode)
{
rtx op00, op04, op10, op14;
op00 = rl78_subreg (SImode, operands[0], omode, 0);
op04 = rl78_subreg (SImode, operands[0], omode, 4);
op10 = rl78_subreg (SImode, operands[1], omode, 0);
op14 = rl78_subreg (SImode, operands[1], omode, 4);
emit_insn (gen_movsi (op00, op10));
emit_insn (gen_movsi (op04, op14));
}
/* Used by various two-operand expanders which cannot accept all
operands in the "far" namespace. Force some such operands into
registers so that each pattern has at most one far operand. */
int
rl78_force_nonfar_2 (rtx *operands, rtx (*gen)(rtx,rtx))
{
int did = 0;
rtx temp_reg = NULL;
/* FIXME: in the future, be smarter about only doing this if the
other operand is also far, assuming the devirtualizer can also
handle that. */
if (rl78_far_p (operands[0]))
{
temp_reg = operands[0];
operands[0] = gen_reg_rtx (GET_MODE (operands[0]));
did = 1;
}
if (!did)
return 0;
emit_insn (gen (operands[0], operands[1]));
if (temp_reg)
emit_move_insn (temp_reg, operands[0]);
return 1;
}
/* Likewise, but for three-operand expanders. */
int
rl78_force_nonfar_3 (rtx *operands, rtx (*gen)(rtx,rtx,rtx))
{
int did = 0;
rtx temp_reg = NULL;
/* FIXME: Likewise. */
if (rl78_far_p (operands[1]))
{
rtx temp_reg = gen_reg_rtx (GET_MODE (operands[1]));
emit_move_insn (temp_reg, operands[1]);
operands[1] = temp_reg;
did = 1;
}
if (rl78_far_p (operands[0]))
{
temp_reg = operands[0];
operands[0] = gen_reg_rtx (GET_MODE (operands[0]));
did = 1;
}
if (!did)
return 0;
emit_insn (gen (operands[0], operands[1], operands[2]));
if (temp_reg)
emit_move_insn (temp_reg, operands[0]);
return 1;
}
int
rl78_one_far_p (rtx *operands, int n)
{
rtx which = NULL;
int i, c = 0;
for (i = 0; i < n; i ++)
if (rl78_far_p (operands[i]))
{
if (which == NULL)
which = operands[i];
else if (rtx_equal_p (operands[i], which))
continue;
c ++;
}
return c <= 1;
}
#undef TARGET_CAN_ELIMINATE
#define TARGET_CAN_ELIMINATE rl78_can_eliminate
static bool
rl78_can_eliminate (const int from ATTRIBUTE_UNUSED, const int to ATTRIBUTE_UNUSED)
{
return true;
}
/* Returns true if the given register needs to be saved by the
current function. */
static bool
need_to_save (unsigned int regno)
{
if (is_interrupt_func (cfun->decl))
{
/* We don't know what devirt will need */
if (regno < 8)
return true;
/* We don't need to save registers that have
been reserved for interrupt handlers. */
if (regno > 23)
return false;
/* If the handler is a non-leaf function then it may call
non-interrupt aware routines which will happily clobber
any call_used registers, so we have to preserve them.
We do not have to worry about the frame pointer register
though, as that is handled below. */
if (!crtl->is_leaf && call_used_or_fixed_reg_p (regno) && regno < 22)
return true;
/* Otherwise we only have to save a register, call_used
or not, if it is used by this handler. */
return df_regs_ever_live_p (regno);
}
if (regno == FRAME_POINTER_REGNUM
&& (frame_pointer_needed || df_regs_ever_live_p (regno)))
return true;
if (fixed_regs[regno])
return false;
if (crtl->calls_eh_return)
return true;
if (df_regs_ever_live_p (regno)
&& !call_used_or_fixed_reg_p (regno))
return true;
return false;
}
/* We use this to wrap all emitted insns in the prologue. */
static rtx
F (rtx x)
{
RTX_FRAME_RELATED_P (x) = 1;
return x;
}
/* Compute all the frame-related fields in our machine_function
structure. */
static void
rl78_compute_frame_info (void)
{
int i;
cfun->machine->computed = 1;
cfun->machine->framesize_regs = 0;
cfun->machine->framesize_locals = get_frame_size ();
cfun->machine->framesize_outgoing = crtl->outgoing_args_size;
for (i = 0; i < 16; i ++)
if (need_to_save (i * 2) || need_to_save (i * 2 + 1))
{
cfun->machine->need_to_push [i] = 1;
cfun->machine->framesize_regs += 2;
}
else
cfun->machine->need_to_push [i] = 0;
if ((cfun->machine->framesize_locals + cfun->machine->framesize_outgoing) & 1)
cfun->machine->framesize_locals ++;
cfun->machine->framesize = (cfun->machine->framesize_regs
+ cfun->machine->framesize_locals
+ cfun->machine->framesize_outgoing);
}
/* Returns true if the provided function has the specified attribute. */
static inline bool
has_func_attr (const_tree decl, const char * func_attr)
{
if (decl == NULL_TREE)
decl = current_function_decl;
return lookup_attribute (func_attr, DECL_ATTRIBUTES (decl)) != NULL_TREE;
}
/* Returns true if the provided function has the "interrupt" attribute. */
static inline bool
is_interrupt_func (const_tree decl)
{
return has_func_attr (decl, "interrupt") || has_func_attr (decl, "brk_interrupt");
}
/* Returns true if the provided function has the "brk_interrupt" attribute. */
static inline bool
is_brk_interrupt_func (const_tree decl)
{
return has_func_attr (decl, "brk_interrupt");
}
/* Check "interrupt" attributes. */
static tree
rl78_handle_func_attribute (tree * node,
tree name,
tree args ATTRIBUTE_UNUSED,
int flags ATTRIBUTE_UNUSED,
bool * no_add_attrs)
{
gcc_assert (DECL_P (* node));
if (TREE_CODE (* node) != FUNCTION_DECL)
{
warning (OPT_Wattributes, "%qE attribute only applies to functions",
name);
* no_add_attrs = true;
}
/* FIXME: We ought to check that the interrupt and exception
handler attributes have been applied to void functions. */
return NULL_TREE;
}
/* Check "naked" attributes. */
static tree
rl78_handle_naked_attribute (tree * node,
tree name ATTRIBUTE_UNUSED,
tree args,
int flags ATTRIBUTE_UNUSED,
bool * no_add_attrs)
{
gcc_assert (DECL_P (* node));
gcc_assert (args == NULL_TREE);
if (TREE_CODE (* node) != FUNCTION_DECL)
{
warning (OPT_Wattributes, "naked attribute only applies to functions");
* no_add_attrs = true;
}
/* Disable warnings about this function - eg reaching the end without
seeing a return statement - because the programmer is doing things
that gcc does not know about. */
TREE_NO_WARNING (* node) = 1;
return NULL_TREE;
}
/* Check "saddr" attributes. */
static tree
rl78_handle_saddr_attribute (tree * node,
tree name,
tree args ATTRIBUTE_UNUSED,
int flags ATTRIBUTE_UNUSED,
bool * no_add_attrs)
{
gcc_assert (DECL_P (* node));
if (TREE_CODE (* node) == FUNCTION_DECL)
{
warning (OPT_Wattributes, "%qE attribute doesn%'t apply to functions",
name);
* no_add_attrs = true;
}
return NULL_TREE;
}
/* Check "vector" attribute. */
static tree
rl78_handle_vector_attribute (tree * node,
tree name,
tree args,
int flags ATTRIBUTE_UNUSED,
bool * no_add_attrs)
{
gcc_assert (DECL_P (* node));
gcc_assert (args != NULL_TREE);
if (TREE_CODE (* node) != FUNCTION_DECL)
{
warning (OPT_Wattributes, "%qE attribute only applies to functions",
name);
* no_add_attrs = true;
}
return NULL_TREE;
}
#undef TARGET_ATTRIBUTE_TABLE
#define TARGET_ATTRIBUTE_TABLE rl78_attribute_table
/* Table of RL78-specific attributes. */
const struct attribute_spec rl78_attribute_table[] =
{
/* Name, min_len, max_len, decl_req, type_req, fn_type_req,
affects_type_identity, handler, exclude. */
{ "interrupt", 0, -1, true, false, false, false,
rl78_handle_func_attribute, NULL },
{ "brk_interrupt", 0, 0, true, false, false, false,
rl78_handle_func_attribute, NULL },
{ "naked", 0, 0, true, false, false, false,
rl78_handle_naked_attribute, NULL },
{ "saddr", 0, 0, true, false, false, false,
rl78_handle_saddr_attribute, NULL },
{ "vector", 1, -1, true, false, false, false,
rl78_handle_vector_attribute, NULL },
{ NULL, 0, 0, false, false, false, false, NULL, NULL }
};
/* Break down an address RTX into its component base/index/addend
portions and return TRUE if the address is of a valid form, else
FALSE. */
static bool
characterize_address (rtx x, rtx *base, rtx *index, rtx *addend)
{
*base = NULL_RTX;
*index = NULL_RTX;
*addend = NULL_RTX;
if (GET_CODE (x) == UNSPEC
&& XINT (x, 1) == UNS_ES_ADDR)
x = XVECEXP (x, 0, 1);
if (GET_CODE (x) == REG)
{
*base = x;
return true;
}
/* We sometimes get these without the CONST wrapper */
if (GET_CODE (x) == PLUS
&& GET_CODE (XEXP (x, 0)) == SYMBOL_REF
&& GET_CODE (XEXP (x, 1)) == CONST_INT)
{
*addend = x;
return true;
}
if (GET_CODE (x) == PLUS)
{
*base = XEXP (x, 0);
x = XEXP (x, 1);
if (GET_CODE (*base) == SUBREG)
{
if (GET_MODE (*base) == HImode
&& GET_MODE (XEXP (*base, 0)) == SImode
&& GET_CODE (XEXP (*base, 0)) == REG)
{
/* This is a throw-away rtx just to tell everyone
else what effective register we're using. */
*base = gen_rtx_REG (HImode, REGNO (XEXP (*base, 0)));
}
}
if (GET_CODE (*base) != REG
&& GET_CODE (x) == REG)
{
rtx tmp = *base;
*base = x;
x = tmp;
}
if (GET_CODE (*base) != REG)
return false;
if (GET_CODE (x) == ZERO_EXTEND
&& GET_CODE (XEXP (x, 0)) == REG)
{
*index = XEXP (x, 0);
return false;
}
}
switch (GET_CODE (x))
{
case PLUS:
if (GET_CODE (XEXP (x, 0)) == SYMBOL_REF
&& GET_CODE (XEXP (x, 0)) == CONST_INT)
{
*addend = x;
return true;
}
/* fall through */
case MEM:
case REG:
return false;
case SUBREG:
switch (GET_CODE (XEXP (x, 0)))
{
case CONST:
case SYMBOL_REF:
case CONST_INT:
*addend = x;
return true;
default:
return false;
}
case CONST:
case SYMBOL_REF:
case CONST_INT:
*addend = x;
return true;
default:
return false;
}
return false;
}
/* Used by the Whb constraint. Match addresses that use HL+B or HL+C
addressing. */
bool
rl78_hl_b_c_addr_p (rtx op)
{
rtx hl, bc;
if (GET_CODE (op) != PLUS)
return false;
hl = XEXP (op, 0);
bc = XEXP (op, 1);
if (GET_CODE (hl) == ZERO_EXTEND)
{
rtx tmp = hl;
hl = bc;
bc = tmp;
}
if (GET_CODE (hl) != REG)
return false;
if (GET_CODE (bc) != ZERO_EXTEND)
return false;
bc = XEXP (bc, 0);
if (GET_CODE (bc) != REG)
return false;
if (REGNO (hl) != HL_REG)
return false;
if (REGNO (bc) != B_REG && REGNO (bc) != C_REG)
return false;
return true;
}
#define REG_IS(r, regno) (((r) == (regno)) || ((r) >= FIRST_PSEUDO_REGISTER && !(strict)))
/* Return the appropriate mode for a named address address. */
#undef TARGET_ADDR_SPACE_ADDRESS_MODE
#define TARGET_ADDR_SPACE_ADDRESS_MODE rl78_addr_space_address_mode
static scalar_int_mode
rl78_addr_space_address_mode (addr_space_t addrspace)
{
switch (addrspace)
{
case ADDR_SPACE_GENERIC:
return HImode;
case ADDR_SPACE_NEAR:
return HImode;
case ADDR_SPACE_FAR:
return SImode;
default:
gcc_unreachable ();
}
}
/* Used in various constraints and predicates to match operands in the
"far" address space. */
int
rl78_far_p (rtx x)
{
if (! MEM_P (x))
return 0;
#if DEBUG0
fprintf (stderr, "\033[35mrl78_far_p: "); debug_rtx (x);
fprintf (stderr, " = %d\033[0m\n", MEM_ADDR_SPACE (x) == ADDR_SPACE_FAR);
#endif
/* Not all far addresses are legitimate, because the devirtualizer
can't handle them. */
if (! rl78_as_legitimate_address (GET_MODE (x), XEXP (x, 0), false, ADDR_SPACE_FAR))
return 0;
return GET_MODE_BITSIZE (rl78_addr_space_address_mode (MEM_ADDR_SPACE (x))) == 32;
}
/* Return the appropriate mode for a named address pointer. */
#undef TARGET_ADDR_SPACE_POINTER_MODE
#define TARGET_ADDR_SPACE_POINTER_MODE rl78_addr_space_pointer_mode
static scalar_int_mode
rl78_addr_space_pointer_mode (addr_space_t addrspace)
{
switch (addrspace)
{
case ADDR_SPACE_GENERIC:
return HImode;
case ADDR_SPACE_NEAR:
return HImode;
case ADDR_SPACE_FAR:
return SImode;
default:
gcc_unreachable ();
}
}
/* Returns TRUE for valid addresses. */
#undef TARGET_VALID_POINTER_MODE
#define TARGET_VALID_POINTER_MODE rl78_valid_pointer_mode
static bool
rl78_valid_pointer_mode (scalar_int_mode m)
{
return (m == HImode || m == SImode);
}
#undef TARGET_LEGITIMATE_CONSTANT_P
#define TARGET_LEGITIMATE_CONSTANT_P rl78_is_legitimate_constant
static bool
rl78_is_legitimate_constant (machine_mode mode ATTRIBUTE_UNUSED, rtx x ATTRIBUTE_UNUSED)
{
return true;
}
#undef TARGET_LRA_P
#define TARGET_LRA_P hook_bool_void_false
#undef TARGET_ADDR_SPACE_LEGITIMATE_ADDRESS_P
#define TARGET_ADDR_SPACE_LEGITIMATE_ADDRESS_P rl78_as_legitimate_address
bool
rl78_as_legitimate_address (machine_mode mode ATTRIBUTE_UNUSED, rtx x,
bool strict ATTRIBUTE_UNUSED, addr_space_t as ATTRIBUTE_UNUSED)
{
rtx base, index, addend;
bool is_far_addr = false;
int as_bits;
as_bits = GET_MODE_BITSIZE (rl78_addr_space_address_mode (as));
if (GET_CODE (x) == UNSPEC
&& XINT (x, 1) == UNS_ES_ADDR)
{
x = XVECEXP (x, 0, 1);
is_far_addr = true;
}
if (as_bits == 16 && is_far_addr)
return false;
if (! characterize_address (x, &base, &index, &addend))
return false;
/* We can't extract the high/low portions of a PLUS address
involving a register during devirtualization, so make sure all
such __far addresses do not have addends. This forces GCC to do
the sum separately. */
if (addend && base && as_bits == 32 && GET_MODE (base) == SImode)
return false;
if (base && index)
{
int ir = REGNO (index);
int br = REGNO (base);
#define OK(test, debug) if (test) { /*fprintf(stderr, "%d: OK %s\n", __LINE__, debug);*/ return true; }
OK (REG_IS (br, HL_REG) && REG_IS (ir, B_REG), "[hl+b]");
OK (REG_IS (br, HL_REG) && REG_IS (ir, C_REG), "[hl+c]");
return false;
}
if (strict && base && GET_CODE (base) == REG && REGNO (base) >= FIRST_PSEUDO_REGISTER)
return false;
if (! cfun->machine->virt_insns_ok && base && GET_CODE (base) == REG
&& REGNO (base) >= 8 && REGNO (base) <= 31)
return false;
return true;
}
/* Determine if one named address space is a subset of another. */
#undef TARGET_ADDR_SPACE_SUBSET_P
#define TARGET_ADDR_SPACE_SUBSET_P rl78_addr_space_subset_p
static bool
rl78_addr_space_subset_p (addr_space_t subset, addr_space_t superset)
{
int subset_bits;
int superset_bits;
subset_bits = GET_MODE_BITSIZE (rl78_addr_space_address_mode (subset));
superset_bits = GET_MODE_BITSIZE (rl78_addr_space_address_mode (superset));
return (subset_bits <= superset_bits);
}
#undef TARGET_ADDR_SPACE_CONVERT
#define TARGET_ADDR_SPACE_CONVERT rl78_addr_space_convert
/* Convert from one address space to another. */
static rtx
rl78_addr_space_convert (rtx op, tree from_type, tree to_type)
{
addr_space_t from_as = TYPE_ADDR_SPACE (TREE_TYPE (from_type));
addr_space_t to_as = TYPE_ADDR_SPACE (TREE_TYPE (to_type));
rtx result;
int to_bits;
int from_bits;
to_bits = GET_MODE_BITSIZE (rl78_addr_space_address_mode (to_as));
from_bits = GET_MODE_BITSIZE (rl78_addr_space_address_mode (from_as));
if (to_bits < from_bits)
{
rtx tmp;
/* This is unpredictable, as we're truncating off usable address
bits. */
warning (OPT_Waddress, "converting far pointer to near pointer");
result = gen_reg_rtx (HImode);
if (GET_CODE (op) == SYMBOL_REF
|| (GET_CODE (op) == REG && REGNO (op) >= FIRST_PSEUDO_REGISTER))
tmp = gen_rtx_raw_SUBREG (HImode, op, 0);
else
tmp = simplify_subreg (HImode, op, SImode, 0);
gcc_assert (tmp != NULL_RTX);
emit_move_insn (result, tmp);
return result;
}
else if (to_bits > from_bits)
{
/* This always works. */
result = gen_reg_rtx (SImode);
emit_move_insn (rl78_subreg (HImode, result, SImode, 0), op);
if (TREE_CODE (from_type) == POINTER_TYPE
&& TREE_CODE (TREE_TYPE (from_type)) == FUNCTION_TYPE)
emit_move_insn (rl78_subreg (HImode, result, SImode, 2), const0_rtx);
else
emit_move_insn (rl78_subreg (HImode, result, SImode, 2), GEN_INT (0x0f));
return result;
}
else
return op;
gcc_unreachable ();
}
/* Implements REGNO_MODE_CODE_OK_FOR_BASE_P. */
bool
rl78_regno_mode_code_ok_for_base_p (int regno, machine_mode mode ATTRIBUTE_UNUSED,
addr_space_t address_space ATTRIBUTE_UNUSED,
int outer_code ATTRIBUTE_UNUSED, int index_code)
{
if (regno <= SP_REG && regno >= 16)
return true;
if (index_code == REG)
return (regno == HL_REG);
if (regno == C_REG || regno == B_REG || regno == E_REG || regno == L_REG)
return true;
return false;
}
/* Implements MODE_CODE_BASE_REG_CLASS. */
enum reg_class
rl78_mode_code_base_reg_class (machine_mode mode ATTRIBUTE_UNUSED,
addr_space_t address_space ATTRIBUTE_UNUSED,
int outer_code ATTRIBUTE_UNUSED,
int index_code ATTRIBUTE_UNUSED)
{
return V_REGS;
}
/* Typical stack layout should looks like this after the function's prologue:
| |
-- ^
| | \ |
| | arguments saved | Increasing
| | on the stack | addresses
PARENT arg pointer -> | | /
-------------------------- ---- -------------------
CHILD |ret | return address
--
| | \
| | call saved
| | registers
frame pointer -> | | /
--
| | \
| | local
| | variables
| | /
--
| | \
| | outgoing | Decreasing
| | arguments | addresses
current stack pointer -> | | / |
-------------------------- ---- ------------------ V
| | */
/* Implements INITIAL_ELIMINATION_OFFSET. The frame layout is
described in the machine_Function struct definition, above. */
int
rl78_initial_elimination_offset (int from, int to)
{
int rv = 0; /* as if arg to arg */
rl78_compute_frame_info ();
switch (to)
{
case STACK_POINTER_REGNUM:
rv += cfun->machine->framesize_outgoing;
rv += cfun->machine->framesize_locals;
/* Fall through. */
case FRAME_POINTER_REGNUM:
rv += cfun->machine->framesize_regs;
rv += 4;
break;
default:
gcc_unreachable ();
}
switch (from)
{
case FRAME_POINTER_REGNUM:
rv -= 4;
rv -= cfun->machine->framesize_regs;
case ARG_POINTER_REGNUM:
break;
default:
gcc_unreachable ();
}
return rv;
}
static bool
rl78_is_naked_func (void)
{
return (lookup_attribute ("naked", DECL_ATTRIBUTES (current_function_decl)) != NULL_TREE);
}
/* Check if the block uses mul/div insns for G13 target. */
static bool
check_mduc_usage (void)
{
rtx_insn * insn;
basic_block bb;
FOR_EACH_BB_FN (bb, cfun)
{
FOR_BB_INSNS (bb, insn)
{
if (INSN_P (insn)
&& (get_attr_is_g13_muldiv_insn (insn) == IS_G13_MULDIV_INSN_YES))
return true;
}
}
return false;
}
/* Expand the function prologue (from the prologue pattern). */
void
rl78_expand_prologue (void)
{
int i, fs;
rtx sp = gen_rtx_REG (HImode, STACK_POINTER_REGNUM);
rtx ax = gen_rtx_REG (HImode, AX_REG);
int rb = 0;
if (rl78_is_naked_func ())
return;
/* Always re-compute the frame info - the register usage may have changed. */
rl78_compute_frame_info ();
if (MUST_SAVE_MDUC_REGISTERS && (!crtl->is_leaf || check_mduc_usage ()))
cfun->machine->framesize += ARRAY_SIZE (mduc_regs) * 2;
if (flag_stack_usage_info)
current_function_static_stack_size = cfun->machine->framesize;
if (is_interrupt_func (cfun->decl) && !TARGET_G10)
for (i = 0; i < 4; i++)
if (cfun->machine->need_to_push [i])
{
/* Select Bank 0 if we are using any registers from Bank 0. */
emit_insn (gen_sel_rb (GEN_INT (0)));
break;
}
for (i = 0; i < 16; i++)
if (cfun->machine->need_to_push [i])
{
int reg = i * 2;
if (TARGET_G10)
{
if (reg >= 8)
{
emit_move_insn (ax, gen_rtx_REG (HImode, reg));
reg = AX_REG;
}
}
else
{
int need_bank = i/4;
if (need_bank != rb)
{
emit_insn (gen_sel_rb (GEN_INT (need_bank)));
rb = need_bank;
}
}
F (emit_insn (gen_push (gen_rtx_REG (HImode, reg))));
}
if (rb != 0)
emit_insn (gen_sel_rb (GEN_INT (0)));
/* Save ES register inside interrupt functions if it is used. */
if (is_interrupt_func (cfun->decl) && cfun->machine->uses_es)
{
emit_insn (gen_movqi_from_es (gen_rtx_REG (QImode, A_REG)));
F (emit_insn (gen_push (ax)));
}
/* Save MDUC registers inside interrupt routine. */
if (MUST_SAVE_MDUC_REGISTERS && (!crtl->is_leaf || check_mduc_usage ()))
{
for (unsigned i = 0; i < ARRAY_SIZE (mduc_regs); i++)
{
mduc_reg_type *reg = mduc_regs + i;
rtx mem_mduc = gen_rtx_MEM (reg->mode, GEN_INT (reg->address));
MEM_VOLATILE_P (mem_mduc) = 1;
if (reg->mode == QImode)
emit_insn (gen_movqi (gen_rtx_REG (QImode, A_REG), mem_mduc));
else
emit_insn (gen_movhi (gen_rtx_REG (HImode, AX_REG), mem_mduc));
emit_insn (gen_push (gen_rtx_REG (HImode, AX_REG)));
}
}
if (frame_pointer_needed)
{
F (emit_move_insn (ax, sp));
F (emit_move_insn (gen_rtx_REG (HImode, FRAME_POINTER_REGNUM), ax));
}
fs = cfun->machine->framesize_locals + cfun->machine->framesize_outgoing;
if (fs > 0)
{
/* If we need to subtract more than 254*3 then it is faster and
smaller to move SP into AX and perform the subtraction there. */
if (fs > 254 * 3)
{
rtx insn;
emit_move_insn (ax, sp);
emit_insn (gen_subhi3 (ax, ax, GEN_INT (fs)));
insn = F (emit_move_insn (sp, ax));
add_reg_note (insn, REG_FRAME_RELATED_EXPR,
gen_rtx_SET (sp, gen_rtx_PLUS (HImode, sp,
GEN_INT (-fs))));
}
else
{
while (fs > 0)
{
int fs_byte = (fs > 254) ? 254 : fs;
F (emit_insn (gen_subhi3 (sp, sp, GEN_INT (fs_byte))));
fs -= fs_byte;
}
}
}
}
/* Expand the function epilogue (from the epilogue pattern). */
void
rl78_expand_epilogue (void)
{
int i, fs;
rtx sp = gen_rtx_REG (HImode, STACK_POINTER_REGNUM);
rtx ax = gen_rtx_REG (HImode, AX_REG);
int rb = 0;
if (rl78_is_naked_func ())
return;
if (frame_pointer_needed)
{
emit_move_insn (ax, gen_rtx_REG (HImode, FRAME_POINTER_REGNUM));
emit_move_insn (sp, ax);
}
else
{
fs = cfun->machine->framesize_locals + cfun->machine->framesize_outgoing;
if (fs > 254 * 3)
{
emit_move_insn (ax, sp);
emit_insn (gen_addhi3 (ax, ax, GEN_INT (fs)));
emit_move_insn (sp, ax);
}
else
{
while (fs > 0)
{
int fs_byte = (fs > 254) ? 254 : fs;
emit_insn (gen_addhi3 (sp, sp, GEN_INT (fs_byte)));
fs -= fs_byte;
}
}
}
/* Restore MDUC registers from interrupt routine. */
if (MUST_SAVE_MDUC_REGISTERS && (!crtl->is_leaf || check_mduc_usage ()))
{
for (int i = ARRAY_SIZE (mduc_regs) - 1; i >= 0; i--)
{
mduc_reg_type *reg = mduc_regs + i;
rtx mem_mduc = gen_rtx_MEM (reg->mode, GEN_INT (reg->address));
emit_insn (gen_pop (gen_rtx_REG (HImode, AX_REG)));
MEM_VOLATILE_P (mem_mduc) = 1;
if (reg->mode == QImode)
emit_insn (gen_movqi (mem_mduc, gen_rtx_REG (QImode, A_REG)));
else
emit_insn (gen_movhi (mem_mduc, gen_rtx_REG (HImode, AX_REG)));
}
}
if (is_interrupt_func (cfun->decl) && cfun->machine->uses_es)
{
emit_insn (gen_pop (gen_rtx_REG (HImode, AX_REG)));
emit_insn (gen_movqi_to_es (gen_rtx_REG (QImode, A_REG)));
}
for (i = 15; i >= 0; i--)
if (cfun->machine->need_to_push [i])
{
rtx dest = gen_rtx_REG (HImode, i * 2);
if (TARGET_G10)
{
if (i < 8)
emit_insn (gen_pop (dest));
else
{
emit_insn (gen_pop (ax));
emit_move_insn (dest, ax);
/* Generate a USE of the pop'd register so that DCE will not eliminate the move. */
emit_insn (gen_use (dest));
}
}
else
{
int need_bank = i / 4;
if (need_bank != rb)
{
emit_insn (gen_sel_rb (GEN_INT (need_bank)));
rb = need_bank;
}
emit_insn (gen_pop (dest));
}
}
if (rb != 0)
emit_insn (gen_sel_rb (GEN_INT (0)));
if (cfun->machine->trampolines_used)
emit_insn (gen_trampoline_uninit ());
if (is_brk_interrupt_func (cfun->decl))
emit_jump_insn (gen_brk_interrupt_return ());
else if (is_interrupt_func (cfun->decl))
emit_jump_insn (gen_interrupt_return ());
else
emit_jump_insn (gen_rl78_return ());
}
/* Likewise, for exception handlers. */
void
rl78_expand_eh_epilogue (rtx x ATTRIBUTE_UNUSED)
{
/* FIXME - replace this with an indirect jump with stack adjust. */
emit_jump_insn (gen_rl78_return ());
}
#undef TARGET_ASM_FUNCTION_PROLOGUE
#define TARGET_ASM_FUNCTION_PROLOGUE rl78_start_function
static void
add_vector_labels (FILE *file, const char *aname)
{
tree vec_attr;
tree val_attr;
const char *vname = "vect";
const char *s;
int vnum;
/* This node is for the vector/interrupt tag itself */
vec_attr = lookup_attribute (aname, DECL_ATTRIBUTES (current_function_decl));
if (!vec_attr)
return;
/* Now point it at the first argument */
vec_attr = TREE_VALUE (vec_attr);
/* Iterate through the arguments. */
while (vec_attr)
{
val_attr = TREE_VALUE (vec_attr);
switch (TREE_CODE (val_attr))
{
case STRING_CST:
s = TREE_STRING_POINTER (val_attr);
goto string_id_common;
case IDENTIFIER_NODE:
s = IDENTIFIER_POINTER (val_attr);
string_id_common:
if (strcmp (s, "$default") == 0)
{
fprintf (file, "\t.global\t$tableentry$default$%s\n", vname);
fprintf (file, "$tableentry$default$%s:\n", vname);
}
else
vname = s;
break;
case INTEGER_CST:
vnum = TREE_INT_CST_LOW (val_attr);
fprintf (file, "\t.global\t$tableentry$%d$%s\n", vnum, vname);
fprintf (file, "$tableentry$%d$%s:\n", vnum, vname);
break;
default:
;
}
vec_attr = TREE_CHAIN (vec_attr);
}
}
/* We don't use this to actually emit the function prologue. We use
this to insert a comment in the asm file describing the
function. */
static void
rl78_start_function (FILE *file)
{
int i;
add_vector_labels (file, "interrupt");
add_vector_labels (file, "vector");
if (cfun->machine->framesize == 0)
return;
fprintf (file, "\t; start of function\n");
if (cfun->machine->framesize_regs)
{
fprintf (file, "\t; push %d:", cfun->machine->framesize_regs);
for (i = 0; i < 16; i ++)
if (cfun->machine->need_to_push[i])
fprintf (file, " %s", word_regnames[i*2]);
fprintf (file, "\n");
}
if (frame_pointer_needed)
fprintf (file, "\t; $fp points here (r22)\n");
if (cfun->machine->framesize_locals)
fprintf (file, "\t; locals: %d byte%s\n", cfun->machine->framesize_locals,
cfun->machine->framesize_locals == 1 ? "" : "s");
if (cfun->machine->framesize_outgoing)
fprintf (file, "\t; outgoing: %d byte%s\n", cfun->machine->framesize_outgoing,
cfun->machine->framesize_outgoing == 1 ? "" : "s");
if (cfun->machine->uses_es)
fprintf (file, "\t; uses ES register\n");
if (MUST_SAVE_MDUC_REGISTERS)
fprintf (file, "\t; preserves MDUC registers\n");
}
/* Return an RTL describing where a function return value of type RET_TYPE
is held. */
#undef TARGET_FUNCTION_VALUE
#define TARGET_FUNCTION_VALUE rl78_function_value
static rtx
rl78_function_value (const_tree ret_type,
const_tree fn_decl_or_type ATTRIBUTE_UNUSED,
bool outgoing ATTRIBUTE_UNUSED)
{
machine_mode mode = TYPE_MODE (ret_type);
return gen_rtx_REG (mode, 8);
}
#undef TARGET_PROMOTE_FUNCTION_MODE
#define TARGET_PROMOTE_FUNCTION_MODE rl78_promote_function_mode
static machine_mode
rl78_promote_function_mode (const_tree type ATTRIBUTE_UNUSED,
machine_mode mode,
int *punsignedp ATTRIBUTE_UNUSED,
const_tree funtype ATTRIBUTE_UNUSED, int for_return ATTRIBUTE_UNUSED)
{
return mode;
}
#undef TARGET_FUNCTION_ARG
#define TARGET_FUNCTION_ARG rl78_function_arg
static rtx
rl78_function_arg (cumulative_args_t, const function_arg_info &)
{
return NULL_RTX;
}
#undef TARGET_FUNCTION_ARG_ADVANCE
#define TARGET_FUNCTION_ARG_ADVANCE rl78_function_arg_advance
static void
rl78_function_arg_advance (cumulative_args_t cum_v,
const function_arg_info &arg)
{
int rounded_size;
CUMULATIVE_ARGS * cum = get_cumulative_args (cum_v);
rounded_size = arg.promoted_size_in_bytes ();
if (rounded_size & 1)
rounded_size ++;
(*cum) += rounded_size;
}
#undef TARGET_FUNCTION_ARG_BOUNDARY
#define TARGET_FUNCTION_ARG_BOUNDARY rl78_function_arg_boundary
static unsigned int
rl78_function_arg_boundary (machine_mode mode ATTRIBUTE_UNUSED,
const_tree type ATTRIBUTE_UNUSED)
{
return 16;
}
/* Supported modifier letters:
A - address of a MEM
S - SADDR form of a real register
v - real register corresponding to a virtual register
m - minus - negative of CONST_INT value.
C - inverse of a conditional (NE vs EQ for example)
C - complement of an integer
z - collapsed conditional
s - shift count mod 8
S - shift count mod 16
r - reverse shift count (8-(count mod 8))
B - bit position
h - bottom HI of an SI
H - top HI of an SI
q - bottom QI of an HI
Q - top QI of an HI
e - third QI of an SI (i.e. where the ES register gets values from)
E - fourth QI of an SI (i.e. MSB)
p - Add +0 to a zero-indexed HL based address.
*/
/* Implements the bulk of rl78_print_operand, below. We do it this
way because we need to test for a constant at the top level and
insert the '#', but not test for it anywhere else as we recurse
down into the operand. */
static void
rl78_print_operand_1 (FILE * file, rtx op, int letter)
{
int need_paren;
switch (GET_CODE (op))
{
case MEM:
if (letter == 'A')
rl78_print_operand_1 (file, XEXP (op, 0), letter);
else
{
if (rl78_far_p (op))
{
fprintf (file, "es:");
if (GET_CODE (XEXP (op, 0)) == UNSPEC)
op = gen_rtx_MEM (GET_MODE (op), XVECEXP (XEXP (op, 0), 0, 1));
}
if (letter == 'H')
{
op = adjust_address (op, HImode, 2);
letter = 0;
}
if (letter == 'h')
{
op = adjust_address (op, HImode, 0);
letter = 0;
}
if (letter == 'Q')
{
op = adjust_address (op, QImode, 1);
letter = 0;
}
if (letter == 'q')
{
op = adjust_address (op, QImode, 0);
letter = 0;
}
if (letter == 'e')
{
op = adjust_address (op, QImode, 2);
letter = 0;
}
if (letter == 'E')
{
op = adjust_address (op, QImode, 3);
letter = 0;
}
if (CONSTANT_P (XEXP (op, 0)))
{
if (!rl78_saddr_p (op))
fprintf (file, "!");
rl78_print_operand_1 (file, XEXP (op, 0), letter);
}
else if (GET_CODE (XEXP (op, 0)) == PLUS
&& GET_CODE (XEXP (XEXP (op, 0), 0)) == SYMBOL_REF)
{
if (!rl78_saddr_p (op))
fprintf (file, "!");
rl78_print_operand_1 (file, XEXP (op, 0), letter);
}
else if (GET_CODE (XEXP (op, 0)) == PLUS
&& GET_CODE (XEXP (XEXP (op, 0), 0)) == REG
&& REGNO (XEXP (XEXP (op, 0), 0)) == 2)
{
rl78_print_operand_1 (file, XEXP (XEXP (op, 0), 1), 'u');
fprintf (file, "[");
rl78_print_operand_1 (file, XEXP (XEXP (op, 0), 0), 0);
if (letter == 'p' && GET_CODE (XEXP (op, 0)) == REG)
fprintf (file, "+0");
fprintf (file, "]");
}
else
{
op = XEXP (op, 0);
fprintf (file, "[");
rl78_print_operand_1 (file, op, letter);
if (letter == 'p' && REG_P (op) && REGNO (op) == 6)
fprintf (file, "+0");
fprintf (file, "]");
}
}
break;
case REG:
if (letter == 'Q')
fprintf (file, "%s", reg_names [REGNO (op) | 1]);
else if (letter == 'H')
fprintf (file, "%s", reg_names [REGNO (op) + 2]);
else if (letter == 'q')
fprintf (file, "%s", reg_names [REGNO (op) & ~1]);
else if (letter == 'e')
fprintf (file, "%s", reg_names [REGNO (op) + 2]);
else if (letter == 'E')
fprintf (file, "%s", reg_names [REGNO (op) + 3]);
else if (letter == 'S')
fprintf (file, "0x%x", 0xffef8 + REGNO (op));
else if (GET_MODE (op) == HImode
&& ! (REGNO (op) & ~0xfe))
{
if (letter == 'v')
fprintf (file, "%s", word_regnames [REGNO (op) % 8]);
else
fprintf (file, "%s", word_regnames [REGNO (op)]);
}
else
fprintf (file, "%s", reg_names [REGNO (op)]);
break;
case CONST_INT:
if (letter == 'Q')
fprintf (file, "%ld", INTVAL (op) >> 8);
else if (letter == 'H')
fprintf (file, "%ld", INTVAL (op) >> 16);
else if (letter == 'q')
fprintf (file, "%ld", INTVAL (op) & 0xff);
else if (letter == 'h')
fprintf (file, "%ld", INTVAL (op) & 0xffff);
else if (letter == 'e')
fprintf (file, "%ld", (INTVAL (op) >> 16) & 0xff);
else if (letter == 'B')
{
int ival = INTVAL (op);
if (ival == -128)
ival = 0x80;
if (exact_log2 (ival) >= 0)
fprintf (file, "%d", exact_log2 (ival));
else
fprintf (file, "%d", exact_log2 (~ival & 0xff));
}
else if (letter == 'E')
fprintf (file, "%ld", (INTVAL (op) >> 24) & 0xff);
else if (letter == 'm')
fprintf (file, "%ld", - INTVAL (op));
else if (letter == 's')
fprintf (file, "%ld", INTVAL (op) % 8);
else if (letter == 'S')
fprintf (file, "%ld", INTVAL (op) % 16);
else if (letter == 'r')
fprintf (file, "%ld", 8 - (INTVAL (op) % 8));
else if (letter == 'C')
fprintf (file, "%ld", (INTVAL (op) ^ 0x8000) & 0xffff);
else
fprintf (file, "%ld", INTVAL (op));
break;
case CONST:
rl78_print_operand_1 (file, XEXP (op, 0), letter);
break;
case ZERO_EXTRACT:
{
int bits = INTVAL (XEXP (op, 1));
int ofs = INTVAL (XEXP (op, 2));
if (bits == 16 && ofs == 0)
fprintf (file, "%%lo16(");
else if (bits == 16 && ofs == 16)
fprintf (file, "%%hi16(");
else if (bits == 8 && ofs == 16)
fprintf (file, "%%hi8(");
else
gcc_unreachable ();
rl78_print_operand_1 (file, XEXP (op, 0), 0);
fprintf (file, ")");
}
break;
case ZERO_EXTEND:
if (GET_CODE (XEXP (op, 0)) == REG)
fprintf (file, "%s", reg_names [REGNO (XEXP (op, 0))]);
else
print_rtl (file, op);
break;
case PLUS:
need_paren = 0;
if (letter == 'H')
{
fprintf (file, "%%hi16(");
need_paren = 1;
letter = 0;
}
if (letter == 'h')
{
fprintf (file, "%%lo16(");
need_paren = 1;
letter = 0;
}
if (letter == 'e')
{
fprintf (file, "%%hi8(");
need_paren = 1;
letter = 0;
}
if (letter == 'q' || letter == 'Q')
output_operand_lossage ("q/Q modifiers invalid for symbol references");
if (GET_CODE (XEXP (op, 0)) == ZERO_EXTEND)
{
if (GET_CODE (XEXP (op, 1)) == SYMBOL_REF
&& SYMBOL_REF_DECL (XEXP (op, 1))
&& TREE_CODE (SYMBOL_REF_DECL (XEXP (op, 1))) == FUNCTION_DECL)
{
fprintf (file, "%%code(");
assemble_name (file, rl78_strip_nonasm_name_encoding (XSTR (XEXP (op, 1), 0)));
fprintf (file, "+");
rl78_print_operand_1 (file, XEXP (op, 0), letter);
fprintf (file, ")");
}
else
{
rl78_print_operand_1 (file, XEXP (op, 1), letter);
fprintf (file, "+");
rl78_print_operand_1 (file, XEXP (op, 0), letter);
}
}
else
{
if (GET_CODE (XEXP (op, 0)) == SYMBOL_REF
&& SYMBOL_REF_DECL (XEXP (op, 0))
&& TREE_CODE (SYMBOL_REF_DECL (XEXP (op, 0))) == FUNCTION_DECL)
{
fprintf (file, "%%code(");
assemble_name (file, rl78_strip_nonasm_name_encoding (XSTR (XEXP (op, 0), 0)));
fprintf (file, "+");
rl78_print_operand_1 (file, XEXP (op, 1), letter);
fprintf (file, ")");
}
else
{
rl78_print_operand_1 (file, XEXP (op, 0), letter);
fprintf (file, "+");
rl78_print_operand_1 (file, XEXP (op, 1), letter);
}
}
if (need_paren)
fprintf (file, ")");
break;
case SUBREG:
if (GET_MODE (op) == HImode
&& SUBREG_BYTE (op) == 0)
{
fprintf (file, "%%lo16(");
rl78_print_operand_1 (file, SUBREG_REG (op), 0);
fprintf (file, ")");
}
else if (GET_MODE (op) == HImode
&& SUBREG_BYTE (op) == 2)
{
fprintf (file, "%%hi16(");
rl78_print_operand_1 (file, SUBREG_REG (op), 0);
fprintf (file, ")");
}
else
{
fprintf (file, "(%s)", GET_RTX_NAME (GET_CODE (op)));
}
break;
case SYMBOL_REF:
need_paren = 0;
if (letter == 'H')
{
fprintf (file, "%%hi16(");
need_paren = 1;
letter = 0;
}
if (letter == 'h')
{
fprintf (file, "%%lo16(");
need_paren = 1;
letter = 0;
}
if (letter == 'e')
{
fprintf (file, "%%hi8(");
need_paren = 1;
letter = 0;
}
if (letter == 'q' || letter == 'Q')
output_operand_lossage ("q/Q modifiers invalid for symbol references");
if (SYMBOL_REF_DECL (op) && TREE_CODE (SYMBOL_REF_DECL (op)) == FUNCTION_DECL)
{
fprintf (file, "%%code(");
assemble_name (file, rl78_strip_nonasm_name_encoding (XSTR (op, 0)));
fprintf (file, ")");
}
else
assemble_name (file, rl78_strip_nonasm_name_encoding (XSTR (op, 0)));
if (need_paren)
fprintf (file, ")");
break;
case CODE_LABEL:
case LABEL_REF:
output_asm_label (op);
break;
case LTU:
if (letter == 'z')
fprintf (file, "#comparison eliminated");
else
fprintf (file, letter == 'C' ? "nc" : "c");
break;
case LEU:
if (letter == 'z')
fprintf (file, "br");
else
fprintf (file, letter == 'C' ? "h" : "nh");
break;
case GEU:
if (letter == 'z')
fprintf (file, "br");
else
fprintf (file, letter == 'C' ? "c" : "nc");
break;
case GTU:
if (letter == 'z')
fprintf (file, "#comparison eliminated");
else
fprintf (file, letter == 'C' ? "nh" : "h");
break;
case EQ:
if (letter == 'z')
fprintf (file, "br");
else
fprintf (file, letter == 'C' ? "nz" : "z");
break;
case NE:
if (letter == 'z')
fprintf (file, "#comparison eliminated");
else
fprintf (file, letter == 'C' ? "z" : "nz");
break;
/* Note: these assume appropriate adjustments were made so that
unsigned comparisons, which is all this chip has, will
work. */
case LT:
if (letter == 'z')
fprintf (file, "#comparison eliminated");
else
fprintf (file, letter == 'C' ? "nc" : "c");
break;
case LE:
if (letter == 'z')
fprintf (file, "br");
else
fprintf (file, letter == 'C' ? "h" : "nh");
break;
case GE:
if (letter == 'z')
fprintf (file, "br");
else
fprintf (file, letter == 'C' ? "c" : "nc");
break;
case GT:
if (letter == 'z')
fprintf (file, "#comparison eliminated");
else
fprintf (file, letter == 'C' ? "nh" : "h");
break;
default:
fprintf (file, "(%s)", GET_RTX_NAME (GET_CODE (op)));
break;
}
}
#undef TARGET_PRINT_OPERAND
#define TARGET_PRINT_OPERAND rl78_print_operand
static void
rl78_print_operand (FILE * file, rtx op, int letter)
{
if (CONSTANT_P (op) && letter != 'u' && letter != 's' && letter != 'r' && letter != 'S' && letter != 'B')
fprintf (file, "#");
rl78_print_operand_1 (file, op, letter);
}
#undef TARGET_TRAMPOLINE_INIT
#define TARGET_TRAMPOLINE_INIT rl78_trampoline_init
/* Note that the RL78's addressing makes it very difficult to do
trampolines on the stack. So, libgcc has a small pool of
trampolines from which one is allocated to this task. */
static void
rl78_trampoline_init (rtx m_tramp, tree fndecl, rtx static_chain)
{
rtx mov_addr, thunk_addr;
rtx function = XEXP (DECL_RTL (fndecl), 0);
mov_addr = adjust_address (m_tramp, HImode, 0);
thunk_addr = gen_reg_rtx (HImode);
function = force_reg (HImode, function);
static_chain = force_reg (HImode, static_chain);
emit_insn (gen_trampoline_init (thunk_addr, function, static_chain));
emit_move_insn (mov_addr, thunk_addr);
cfun->machine->trampolines_used = 1;
}
#undef TARGET_TRAMPOLINE_ADJUST_ADDRESS
#define TARGET_TRAMPOLINE_ADJUST_ADDRESS rl78_trampoline_adjust_address
static rtx
rl78_trampoline_adjust_address (rtx m_tramp)
{
rtx x = gen_rtx_MEM (HImode, m_tramp);
return x;
}
/* Expander for cbranchqi4 and cbranchhi4. RL78 is missing some of
the "normal" compares, specifically, it only has unsigned compares,
so we must synthesize the missing ones. */
void
rl78_expand_compare (rtx *operands)
{
if (GET_CODE (operands[2]) == MEM)
operands[2] = copy_to_mode_reg (GET_MODE (operands[2]), operands[2]);
}
/* Define this to 1 if you are debugging the peephole optimizers. */
#define DEBUG_PEEP 0
/* Predicate used to enable the peephole2 patterns in rl78-virt.md.
The default "word" size is a byte so we can effectively use all the
registers, but we want to do 16-bit moves whenever possible. This
function determines when such a move is an option. */
bool
rl78_peep_movhi_p (rtx *operands)
{
int i;
rtx m, a;
/* (set (op0) (op1))
(set (op2) (op3)) */
if (! rl78_virt_insns_ok ())
return false;
#if DEBUG_PEEP
fprintf (stderr, "\033[33m");
debug_rtx (operands[0]);
debug_rtx (operands[1]);
debug_rtx (operands[2]);
debug_rtx (operands[3]);
fprintf (stderr, "\033[0m");
#endif
/* You can move a constant to memory as QImode, but not HImode. */
if (GET_CODE (operands[0]) == MEM
&& GET_CODE (operands[1]) != REG)
{
#if DEBUG_PEEP
fprintf (stderr, "no peep: move constant to memory\n");
#endif
return false;
}
if (rtx_equal_p (operands[0], operands[3]))
{
#if DEBUG_PEEP
fprintf (stderr, "no peep: overlapping\n");
#endif
return false;
}
for (i = 0; i < 2; i ++)
{
if (GET_CODE (operands[i]) != GET_CODE (operands[i+2]))
{
#if DEBUG_PEEP
fprintf (stderr, "no peep: different codes\n");
#endif
return false;
}
if (GET_MODE (operands[i]) != GET_MODE (operands[i+2]))
{
#if DEBUG_PEEP
fprintf (stderr, "no peep: different modes\n");
#endif
return false;
}
switch (GET_CODE (operands[i]))
{
case REG:
/* LSB MSB */
if (REGNO (operands[i]) + 1 != REGNO (operands[i+2])
|| GET_MODE (operands[i]) != QImode)
{
#if DEBUG_PEEP
fprintf (stderr, "no peep: wrong regnos %d %d %d\n",
REGNO (operands[i]), REGNO (operands[i+2]),
i);
#endif
return false;
}
if (! rl78_hard_regno_mode_ok (REGNO (operands[i]), HImode))
{
#if DEBUG_PEEP
fprintf (stderr, "no peep: reg %d not HI\n", REGNO (operands[i]));
#endif
return false;
}
break;
case CONST_INT:
break;
case MEM:
if (GET_MODE (operands[i]) != QImode)
return false;
if (MEM_ALIGN (operands[i]) < 16)
return false;
a = XEXP (operands[i], 0);
if (GET_CODE (a) == CONST)
a = XEXP (a, 0);
if (GET_CODE (a) == PLUS)
a = XEXP (a, 1);
if (GET_CODE (a) == CONST_INT
&& INTVAL (a) & 1)
{
#if DEBUG_PEEP
fprintf (stderr, "no peep: misaligned mem %d\n", i);
debug_rtx (operands[i]);
#endif
return false;
}
m = adjust_address (operands[i], QImode, 1);
if (! rtx_equal_p (m, operands[i+2]))
{
#if DEBUG_PEEP
fprintf (stderr, "no peep: wrong mem %d\n", i);
debug_rtx (m);
debug_rtx (operands[i+2]);
#endif
return false;
}
break;
default:
#if DEBUG_PEEP
fprintf (stderr, "no peep: wrong rtx %d\n", i);
#endif
return false;
}
}
#if DEBUG_PEEP
fprintf (stderr, "\033[32mpeep!\033[0m\n");
#endif
return true;
}
/* Likewise, when a peephole is activated, this function helps compute
the new operands. */
void
rl78_setup_peep_movhi (rtx *operands)
{
int i;
for (i = 0; i < 2; i ++)
{
switch (GET_CODE (operands[i]))
{
case REG:
operands[i+4] = gen_rtx_REG (HImode, REGNO (operands[i]));
break;
case CONST_INT:
operands[i+4] = GEN_INT ((INTVAL (operands[i]) & 0xff) + ((char) INTVAL (operands[i+2])) * 256);
break;
case MEM:
operands[i+4] = adjust_address (operands[i], HImode, 0);
break;
default:
break;
}
}
}
/*
How Devirtualization works in the RL78 GCC port
Background
The RL78 is an 8-bit port with some 16-bit operations. It has 32
bytes of register space, in four banks, memory-mapped. One bank is
the "selected" bank and holds the registers used for primary
operations. Since the registers are memory mapped, often you can
still refer to the unselected banks via memory accesses.
Virtual Registers
The GCC port uses bank 0 as the "selected" registers (A, X, BC, etc)
and refers to the other banks via their memory addresses, although
they're treated as regular registers internally. These "virtual"
registers are R8 through R23 (bank3 is reserved for asm-based
interrupt handlers).
There are four machine description files:
rl78.md - common register-independent patterns and definitions
rl78-expand.md - expanders
rl78-virt.md - patterns that match BEFORE devirtualization
rl78-real.md - patterns that match AFTER devirtualization
At least through register allocation and reload, gcc is told that it
can do pretty much anything - but may only use the virtual registers.
GCC cannot properly create the varying addressing modes that the RL78
supports in an efficient way.
Sometime after reload, the RL78 backend "devirtualizes" the RTL. It
uses the "valloc" attribute in rl78-virt.md for determining the rules
by which it will replace virtual registers with real registers (or
not) and how to make up addressing modes. For example, insns tagged
with "ro1" have a single read-only parameter, which may need to be
moved from memory/constant/vreg to a suitable real register. As part
of devirtualization, a flag is toggled, disabling the rl78-virt.md
patterns and enabling the rl78-real.md patterns. The new patterns'
constraints are used to determine the real registers used. NOTE:
patterns in rl78-virt.md essentially ignore the constrains and rely on
predicates, where the rl78-real.md ones essentially ignore the
predicates and rely on the constraints.
The devirtualization pass is scheduled via the pass manager (despite
being called "rl78_reorg") so it can be scheduled prior to var-track
(the idea is to let gdb know about the new registers). Ideally, it
would be scheduled right after pro/epilogue generation, so the
post-reload optimizers could operate on the real registers, but when I
tried that there were some issues building the target libraries.
During devirtualization, a simple register move optimizer is run. It
would be better to run a full CSE/propogation pass on it though, but
that has not yet been attempted.
*/
#define DEBUG_ALLOC 0
#define OP(x) (*recog_data.operand_loc[x])
/* This array is used to hold knowledge about the contents of the
real registers (A ... H), the memory-based registers (r8 ... r31)
and the first NUM_STACK_LOCS words on the stack. We use this to
avoid generating redundant move instructions.
A value in the range 0 .. 31 indicates register A .. r31.
A value in the range 32 .. 63 indicates stack slot (value - 32).
A value of NOT_KNOWN indicates that the contents of that location
are not known. */
#define NUM_STACK_LOCS 32
#define NOT_KNOWN 127
static unsigned char content_memory [32 + NUM_STACK_LOCS];
static unsigned char saved_update_index = NOT_KNOWN;
static unsigned char saved_update_value;
static machine_mode saved_update_mode;
static inline void
clear_content_memory (void)
{
memset (content_memory, NOT_KNOWN, sizeof content_memory);
if (dump_file)
fprintf (dump_file, " clear content memory\n");
saved_update_index = NOT_KNOWN;
}
/* Convert LOC into an index into the content_memory array.
If LOC cannot be converted, return NOT_KNOWN. */
static unsigned char
get_content_index (rtx loc)
{
machine_mode mode;
if (loc == NULL_RTX)
return NOT_KNOWN;
if (REG_P (loc))
{
if (REGNO (loc) < 32)
return REGNO (loc);
return NOT_KNOWN;
}
mode = GET_MODE (loc);
if (! rl78_stack_based_mem (loc, mode))
return NOT_KNOWN;
loc = XEXP (loc, 0);
if (REG_P (loc))
/* loc = MEM (SP) */
return 32;
/* loc = MEM (PLUS (SP, INT)). */
loc = XEXP (loc, 1);
if (INTVAL (loc) < NUM_STACK_LOCS)
return 32 + INTVAL (loc);
return NOT_KNOWN;
}
/* Return a string describing content INDEX in mode MODE.
WARNING: Can return a pointer to a static buffer. */
static const char *
get_content_name (unsigned char index, machine_mode mode)
{
static char buffer [128];
if (index == NOT_KNOWN)
return "Unknown";
if (index > 31)
sprintf (buffer, "stack slot %d", index - 32);
else if (mode == HImode)
sprintf (buffer, "%s%s",
reg_names [index + 1], reg_names [index]);
else
return reg_names [index];
return buffer;
}
#if DEBUG_ALLOC
static void
display_content_memory (FILE * file)
{
unsigned int i;
fprintf (file, " Known memory contents:\n");
for (i = 0; i < sizeof content_memory; i++)
if (content_memory[i] != NOT_KNOWN)
{
fprintf (file, " %s contains a copy of ", get_content_name (i, QImode));
fprintf (file, "%s\n", get_content_name (content_memory [i], QImode));
}
}
#endif
static void
update_content (unsigned char index, unsigned char val, machine_mode mode)
{
unsigned int i;
gcc_assert (index < sizeof content_memory);
content_memory [index] = val;
if (val != NOT_KNOWN)
content_memory [val] = index;
/* Make the entry in dump_file *before* VAL is increased below. */
if (dump_file)
{
fprintf (dump_file, " %s now contains ", get_content_name (index, mode));
if (val == NOT_KNOWN)
fprintf (dump_file, "Unknown\n");
else
fprintf (dump_file, "%s and vice versa\n", get_content_name (val, mode));
}
if (mode == HImode)
{
val = val == NOT_KNOWN ? val : val + 1;
content_memory [index + 1] = val;
if (val != NOT_KNOWN)
{
content_memory [val] = index + 1;
-- val;
}
}
/* Any other places that had INDEX recorded as their contents are now invalid. */
for (i = 0; i < sizeof content_memory; i++)
{
if (i == index
|| (val != NOT_KNOWN && i == val))
{
if (mode == HImode)
++ i;
continue;
}
if (content_memory[i] == index
|| (val != NOT_KNOWN && content_memory[i] == val))
{
content_memory[i] = NOT_KNOWN;
if (dump_file)
fprintf (dump_file, " %s cleared\n", get_content_name (i, mode));
if (mode == HImode)
content_memory[++ i] = NOT_KNOWN;
}
}
}
/* Record that LOC contains VALUE.
For HImode locations record that LOC+1 contains VALUE+1.
If LOC is not a register or stack slot, do nothing.
If VALUE is not a register or stack slot, clear the recorded content. */
static void
record_content (rtx loc, rtx value)
{
machine_mode mode;
unsigned char index;
unsigned char val;
if ((index = get_content_index (loc)) == NOT_KNOWN)
return;
val = get_content_index (value);
mode = GET_MODE (loc);
if (val == index)
{
if (! optimize)
return;
/* This should not happen when optimizing. */
#if 1
fprintf (stderr, "ASSIGNMENT of location to itself detected! [%s]\n",
get_content_name (val, mode));
return;
#else
gcc_unreachable ();
#endif
}
update_content (index, val, mode);
}
/* Returns TRUE if LOC already contains a copy of VALUE. */
static bool
already_contains (rtx loc, rtx value)
{
unsigned char index;
unsigned char val;
if ((index = get_content_index (loc)) == NOT_KNOWN)
return false;
if ((val = get_content_index (value)) == NOT_KNOWN)
return false;
if (content_memory [index] != val)
return false;
if (GET_MODE (loc) == HImode)
return content_memory [index + 1] == val + 1;
return true;
}
bool
rl78_es_addr (rtx addr)
{
if (GET_CODE (addr) == MEM)
addr = XEXP (addr, 0);
if (GET_CODE (addr) != UNSPEC)
return false;
if (XINT (addr, 1) != UNS_ES_ADDR)
return false;
return true;
}
rtx
rl78_es_base (rtx addr)
{
if (GET_CODE (addr) == MEM)
addr = XEXP (addr, 0);
addr = XVECEXP (addr, 0, 1);
if (GET_CODE (addr) == CONST
&& GET_CODE (XEXP (addr, 0)) == ZERO_EXTRACT)
addr = XEXP (XEXP (addr, 0), 0);
/* Mode doesn't matter here. */
return gen_rtx_MEM (HImode, addr);
}
/* Rescans an insn to see if it's recognized again. This is done
carefully to ensure that all the constraint information is accurate
for the newly matched insn. */
static bool
insn_ok_now (rtx_insn * insn)
{
rtx pattern = PATTERN (insn);
int i;
INSN_CODE (insn) = -1;
if (recog (pattern, insn, 0) > -1)
{
extract_insn (insn);
if (constrain_operands (1, get_preferred_alternatives (insn)))
{
#if DEBUG_ALLOC
fprintf (stderr, "\033[32m");
debug_rtx (insn);
fprintf (stderr, "\033[0m");
#endif
if (SET_P (pattern))
record_content (SET_DEST (pattern), SET_SRC (pattern));
/* We need to detect far addresses that haven't been
converted to es/lo16 format. */
for (i=0; i<recog_data.n_operands; i++)
if (GET_CODE (OP (i)) == MEM
&& GET_MODE (XEXP (OP (i), 0)) == SImode
&& GET_CODE (XEXP (OP (i), 0)) != UNSPEC)
goto not_ok;
return true;
}
}
/* INSN is not OK as-is. It may not be recognized in real mode or
it might not have satisfied its constraints in real mode. Either
way it will require fixups.
It is vital we always re-recognize at this point as some insns
have fewer operands in real mode than virtual mode. If we do
not re-recognize, then the recog_data will refer to real mode
operands and we may read invalid data. Usually this isn't a
problem, but once in a while the data we read is bogus enough
to cause a segfault or other undesirable behavior. */
not_ok:
/* We need to re-recog the insn with virtual registers to get
the operands. */
INSN_CODE (insn) = -1;
cfun->machine->virt_insns_ok = 1;
if (recog (pattern, insn, 0) > -1)
{
extract_insn (insn);
/* In theory this should always be true. */
if (constrain_operands (0, get_preferred_alternatives (insn)))
{
cfun->machine->virt_insns_ok = 0;
return false;
}
}
#if DEBUG_ALLOC
fprintf (stderr, "\033[41;30m Unrecognized *virtual* insn \033[0m\n");
debug_rtx (insn);
#endif
gcc_unreachable ();
return false;
}
#if DEBUG_ALLOC
#define WORKED fprintf (stderr, "\033[48;5;22m Worked at line %d \033[0m\n", __LINE__)
#define FAILEDSOFAR fprintf (stderr, "\033[48;5;52m FAILED at line %d \033[0m\n", __LINE__)
#define FAILED fprintf (stderr, "\033[48;5;52m FAILED at line %d \033[0m\n", __LINE__), gcc_unreachable ()
#define MAYBE_OK(insn) if (insn_ok_now (insn)) { WORKED; return; } else { FAILEDSOFAR; }
#define MUST_BE_OK(insn) if (insn_ok_now (insn)) { WORKED; return; } FAILED
#else
#define FAILED gcc_unreachable ()
#define MAYBE_OK(insn) if (insn_ok_now (insn)) return;
#define MUST_BE_OK(insn) if (insn_ok_now (insn)) return; FAILED
#endif
/* Registers into which we move the contents of virtual registers. */
#define X gen_rtx_REG (QImode, X_REG)
#define A gen_rtx_REG (QImode, A_REG)
#define C gen_rtx_REG (QImode, C_REG)
#define B gen_rtx_REG (QImode, B_REG)
#define E gen_rtx_REG (QImode, E_REG)
#define D gen_rtx_REG (QImode, D_REG)
#define L gen_rtx_REG (QImode, L_REG)
#define H gen_rtx_REG (QImode, H_REG)
#define AX gen_rtx_REG (HImode, AX_REG)
#define BC gen_rtx_REG (HImode, BC_REG)
#define DE gen_rtx_REG (HImode, DE_REG)
#define HL gen_rtx_REG (HImode, HL_REG)
/* Returns TRUE if R is a virtual register. */
static inline bool
is_virtual_register (rtx r)
{
return (GET_CODE (r) == REG
&& REGNO (r) >= 8
&& REGNO (r) < 32);
}
/* In all these alloc routines, we expect the following: the insn
pattern is unshared, the insn was previously recognized and failed
due to predicates or constraints, and the operand data is in
recog_data. */
static int virt_insn_was_frame;
/* Hook for all insns we emit. Re-mark them as FRAME_RELATED if
needed. */
static rtx
EM2 (int line ATTRIBUTE_UNUSED, rtx r)
{
#if DEBUG_ALLOC
fprintf (stderr, "\033[36m%d: ", line);
debug_rtx (r);
fprintf (stderr, "\033[0m");
#endif
/*SCHED_GROUP_P (r) = 1;*/
if (virt_insn_was_frame)
RTX_FRAME_RELATED_P (r) = 1;
return r;
}
#define EM(x) EM2 (__LINE__, x)
/* Return a suitable RTX for the low half of a __far address. */
static rtx
rl78_lo16 (rtx addr)
{
rtx r;
if (GET_CODE (addr) == SYMBOL_REF
|| GET_CODE (addr) == CONST)
{
r = gen_rtx_ZERO_EXTRACT (HImode, addr, GEN_INT (16), GEN_INT (0));
r = gen_rtx_CONST (HImode, r);
}
else
r = rl78_subreg (HImode, addr, SImode, 0);
r = gen_es_addr (r);
cfun->machine->uses_es = true;
return r;
}
/* Return a suitable RTX for the high half's lower byte of a __far address. */
static rtx
rl78_hi8 (rtx addr)
{
if (GET_CODE (addr) == SYMBOL_REF
|| GET_CODE (addr) == CONST)
{
rtx r = gen_rtx_ZERO_EXTRACT (QImode, addr, GEN_INT (8), GEN_INT (16));
r = gen_rtx_CONST (QImode, r);
return r;
}
return rl78_subreg (QImode, addr, SImode, 2);
}
static void
add_postponed_content_update (rtx to, rtx value)
{
unsigned char index;
if ((index = get_content_index (to)) == NOT_KNOWN)
return;
gcc_assert (saved_update_index == NOT_KNOWN);
saved_update_index = index;
saved_update_value = get_content_index (value);
saved_update_mode = GET_MODE (to);
}
static void
process_postponed_content_update (void)
{
if (saved_update_index != NOT_KNOWN)
{
update_content (saved_update_index, saved_update_value, saved_update_mode);
saved_update_index = NOT_KNOWN;
}
}
/* Generate and emit a move of (register) FROM into TO. if WHERE is not NULL
then if BEFORE is true then emit the insn before WHERE, otherwise emit it
after WHERE. If TO already contains FROM then do nothing. Returns TO if
BEFORE is true, FROM otherwise. */
static rtx
gen_and_emit_move (rtx to, rtx from, rtx_insn *where, bool before)
{
machine_mode mode = GET_MODE (to);
if (optimize && before && already_contains (to, from))
{
#if DEBUG_ALLOC
display_content_memory (stderr);
#endif
if (dump_file)
{
fprintf (dump_file, " Omit move of %s into ",
get_content_name (get_content_index (from), mode));
fprintf (dump_file, "%s as it already contains this value\n",
get_content_name (get_content_index (to), mode));
}
}
else
{
rtx move = mode == QImode ? gen_movqi (to, from) : gen_movhi (to, from);
EM (move);
if (where == NULL_RTX)
emit_insn (move);
else if (before)
emit_insn_before (move, where);
else
{
rtx note = find_reg_note (where, REG_EH_REGION, NULL_RTX);
/* If necessary move REG_EH_REGION notes forward.
cf. compiling gcc.dg/pr44545.c. */
if (note != NULL_RTX)
{
add_reg_note (move, REG_EH_REGION, XEXP (note, 0));
remove_note (where, note);
}
emit_insn_after (move, where);
}
if (before)
record_content (to, from);
else
add_postponed_content_update (to, from);
}
return before ? to : from;
}
/* If M is MEM(REG) or MEM(PLUS(REG,INT)) and REG is virtual then
copy it into NEWBASE and return the updated MEM. Otherwise just
return M. Any needed insns are emitted before BEFORE. */
static rtx
transcode_memory_rtx (rtx m, rtx newbase, rtx_insn *before)
{
rtx base, index, addendr;
int addend = 0;
int need_es = 0;
if (! MEM_P (m))
return m;
if (GET_MODE (XEXP (m, 0)) == SImode)
{
rtx new_m;
rtx seg = rl78_hi8 (XEXP (m, 0));
if (!TARGET_ES0)
{
emit_insn_before (EM (gen_movqi (A, seg)), before);
emit_insn_before (EM (gen_movqi_to_es (A)), before);
}
record_content (A, NULL_RTX);
new_m = gen_rtx_MEM (GET_MODE (m), rl78_lo16 (XEXP (m, 0)));
MEM_COPY_ATTRIBUTES (new_m, m);
m = new_m;
need_es = 1;
}
characterize_address (XEXP (m, 0), & base, & index, & addendr);
gcc_assert (index == NULL_RTX);
if (base == NULL_RTX)
return m;
if (addendr && GET_CODE (addendr) == CONST_INT)
addend = INTVAL (addendr);
gcc_assert (REG_P (base));
gcc_assert (REG_P (newbase));
int limit = 256 - GET_MODE_SIZE (GET_MODE (m));
if (REGNO (base) == SP_REG)
{
if (addend >= 0 && addend <= limit)
return m;
}
/* BASE should be a virtual register. We copy it to NEWBASE. If
the addend is out of range for DE/HL, we use AX to compute the full
address. */
if (addend < 0
|| (addend > limit && REGNO (newbase) != BC_REG)
|| (addendr
&& (GET_CODE (addendr) != CONST_INT)
&& ((REGNO (newbase) != BC_REG))
))
{
/* mov ax, vreg
add ax, #imm
mov hl, ax */
EM (emit_insn_before (gen_movhi (AX, base), before));
EM (emit_insn_before (gen_addhi3 (AX, AX, addendr), before));
EM (emit_insn_before (gen_movhi (newbase, AX), before));
record_content (AX, NULL_RTX);
record_content (newbase, NULL_RTX);
base = newbase;
addend = 0;
addendr = 0;
}
else
{
base = gen_and_emit_move (newbase, base, before, true);
}
if (addend)
{
record_content (base, NULL_RTX);
base = gen_rtx_PLUS (HImode, base, GEN_INT (addend));
}
else if (addendr)
{
record_content (base, NULL_RTX);
base = gen_rtx_PLUS (HImode, base, addendr);
}
if (need_es)
{
m = change_address (m, GET_MODE (m), gen_es_addr (base));
cfun->machine->uses_es = true;
}
else
m = change_address (m, GET_MODE (m), base);
return m;
}
/* Copy SRC to accumulator (A or AX), placing any generated insns
before BEFORE. Returns accumulator RTX. */
static rtx
move_to_acc (int opno, rtx_insn *before)
{
rtx src = OP (opno);
machine_mode mode = GET_MODE (src);
if (REG_P (src) && REGNO (src) < 2)
return src;
if (mode == VOIDmode)
mode = recog_data.operand_mode[opno];
return gen_and_emit_move (mode == QImode ? A : AX, src, before, true);
}
static void
force_into_acc (rtx src, rtx_insn *before)
{
machine_mode mode = GET_MODE (src);
rtx move;
if (REG_P (src) && REGNO (src) < 2)
return;
move = mode == QImode ? gen_movqi (A, src) : gen_movhi (AX, src);
EM (move);
emit_insn_before (move, before);
record_content (AX, NULL_RTX);
}
/* Copy accumulator (A or AX) to DEST, placing any generated insns
after AFTER. Returns accumulator RTX. */
static rtx
move_from_acc (unsigned int opno, rtx_insn *after)
{
rtx dest = OP (opno);
machine_mode mode = GET_MODE (dest);
if (REG_P (dest) && REGNO (dest) < 2)
return dest;
return gen_and_emit_move (dest, mode == QImode ? A : AX, after, false);
}
/* Copy accumulator (A or AX) to REGNO, placing any generated insns
before BEFORE. Returns reg RTX. */
static rtx
move_acc_to_reg (rtx acc, int regno, rtx_insn *before)
{
machine_mode mode = GET_MODE (acc);
rtx reg;
reg = gen_rtx_REG (mode, regno);
return gen_and_emit_move (reg, acc, before, true);
}
/* Copy SRC to X, placing any generated insns before BEFORE.
Returns X RTX. */
static rtx
move_to_x (int opno, rtx_insn *before)
{
rtx src = OP (opno);
machine_mode mode = GET_MODE (src);
rtx reg;
if (mode == VOIDmode)
mode = recog_data.operand_mode[opno];
reg = (mode == QImode) ? X : AX;
if (mode == QImode || ! is_virtual_register (OP (opno)))
{
OP (opno) = move_to_acc (opno, before);
OP (opno) = move_acc_to_reg (OP (opno), X_REG, before);
return reg;
}
return gen_and_emit_move (reg, src, before, true);
}
/* Copy OP (opno) to H or HL, placing any generated insns before BEFORE.
Returns H/HL RTX. */
static rtx
move_to_hl (int opno, rtx_insn *before)
{
rtx src = OP (opno);
machine_mode mode = GET_MODE (src);
rtx reg;
if (mode == VOIDmode)
mode = recog_data.operand_mode[opno];
reg = (mode == QImode) ? L : HL;
if (mode == QImode || ! is_virtual_register (OP (opno)))
{
OP (opno) = move_to_acc (opno, before);
OP (opno) = move_acc_to_reg (OP (opno), L_REG, before);
return reg;
}
return gen_and_emit_move (reg, src, before, true);
}
/* Copy OP (opno) to E or DE, placing any generated insns before BEFORE.
Returns E/DE RTX. */
static rtx
move_to_de (int opno, rtx_insn *before)
{
rtx src = OP (opno);
machine_mode mode = GET_MODE (src);
rtx reg;
if (mode == VOIDmode)
mode = recog_data.operand_mode[opno];
reg = (mode == QImode) ? E : DE;
if (mode == QImode || ! is_virtual_register (OP (opno)))
{
OP (opno) = move_to_acc (opno, before);
OP (opno) = move_acc_to_reg (OP (opno), E_REG, before);
}
else
{
gen_and_emit_move (reg, src, before, true);
}
return reg;
}
/* Devirtualize an insn of the form (SET (op) (unop (op))). */
static void
rl78_alloc_physical_registers_op1 (rtx_insn * insn)
{
/* op[0] = func op[1] */
/* We first try using A as the destination, then copying it
back. */
if (rtx_equal_p (OP (0), OP (1)))
{
OP (0) =
OP (1) = transcode_memory_rtx (OP (1), DE, insn);
}
else
{
/* If necessary, load the operands into BC and HL.
Check to see if we already have OP (0) in HL
and if so, swap the order.
It is tempting to perform this optimization when OP(0) does
not hold a MEM, but this leads to bigger code in general.
The problem is that if OP(1) holds a MEM then swapping it
into BC means a BC-relative load is used and these are 3
bytes long vs 1 byte for an HL load. */
if (MEM_P (OP (0))
&& already_contains (HL, XEXP (OP (0), 0)))
{
OP (0) = transcode_memory_rtx (OP (0), HL, insn);
OP (1) = transcode_memory_rtx (OP (1), BC, insn);
}
else
{
OP (0) = transcode_memory_rtx (OP (0), BC, insn);
OP (1) = transcode_memory_rtx (OP (1), HL, insn);
}
}
MAYBE_OK (insn);
OP (0) = move_from_acc (0, insn);
MAYBE_OK (insn);
/* Try copying the src to acc first, then. This is for, for
example, ZERO_EXTEND or NOT. */
OP (1) = move_to_acc (1, insn);
MUST_BE_OK (insn);
}
/* Returns true if operand OPNUM contains a constraint of type CONSTRAINT.
Assumes that the current insn has already been recognised and hence the
constraint data has been filled in. */
static bool
has_constraint (unsigned int opnum, enum constraint_num constraint)
{
const char * p = recog_data.constraints[opnum];
/* No constraints means anything is accepted. */
if (p == NULL || *p == 0 || *p == ',')
return true;
do
{
char c;
unsigned int len;
c = *p;
len = CONSTRAINT_LEN (c, p);
gcc_assert (len > 0);
switch (c)
{
case 0:
case ',':
return false;
default:
if (lookup_constraint (p) == constraint)
return true;
}
p += len;
}
while (1);
}
/* Devirtualize an insn of the form (SET (op) (binop (op) (op))). */
static void
rl78_alloc_physical_registers_op2 (rtx_insn * insn)
{
rtx_insn *prev;
rtx_insn *first;
bool hl_used;
int tmp_id;
rtx saved_op1;
if (rtx_equal_p (OP (0), OP (1)))
{
if (MEM_P (OP (2)))
{
OP (0) =
OP (1) = transcode_memory_rtx (OP (1), DE, insn);
OP (2) = transcode_memory_rtx (OP (2), HL, insn);
}
else
{
OP (0) =
OP (1) = transcode_memory_rtx (OP (1), HL, insn);
OP (2) = transcode_memory_rtx (OP (2), DE, insn);
}
}
else if (rtx_equal_p (OP (0), OP (2)))
{
OP (1) = transcode_memory_rtx (OP (1), DE, insn);
OP (0) =
OP (2) = transcode_memory_rtx (OP (2), HL, insn);
}
else
{
OP (0) = transcode_memory_rtx (OP (0), BC, insn);
OP (1) = transcode_memory_rtx (OP (1), DE, insn);
OP (2) = transcode_memory_rtx (OP (2), HL, insn);
}
MAYBE_OK (insn);
prev = prev_nonnote_nondebug_insn (insn);
if (recog_data.constraints[1][0] == '%'
&& is_virtual_register (OP (1))
&& ! is_virtual_register (OP (2))
&& ! CONSTANT_P (OP (2)))
{
rtx tmp = OP (1);
OP (1) = OP (2);
OP (2) = tmp;
}
/* Make a note of whether (H)L is being used. It matters
because if OP (2) also needs reloading, then we must take
care not to corrupt HL. */
hl_used = reg_mentioned_p (L, OP (0)) || reg_mentioned_p (L, OP (1));
/* If HL is not currently being used and dest == op1 then there are
some possible optimizations available by reloading one of the
operands into HL, before trying to use the accumulator. */
if (optimize
&& ! hl_used
&& rtx_equal_p (OP (0), OP (1)))
{
/* If op0 is a Ws1 type memory address then switching the base
address register to HL might allow us to perform an in-memory
operation. (eg for the INCW instruction).
FIXME: Adding the move into HL is costly if this optimization is not
going to work, so for now, make sure that we know that the new insn will
match the requirements of the addhi3_real pattern. Really we ought to
generate a candidate sequence, test that, and then install it if the
results are good. */
if (satisfies_constraint_Ws1 (OP (0))
&& has_constraint (0, CONSTRAINT_Wh1)
&& (satisfies_constraint_K (OP (2)) || satisfies_constraint_L (OP (2))))
{
rtx base, index, addend, newbase;
characterize_address (XEXP (OP (0), 0), & base, & index, & addend);
gcc_assert (index == NULL_RTX);
gcc_assert (REG_P (base) && REGNO (base) == SP_REG);
/* Ws1 addressing allows an offset of 0, Wh1 addressing requires a non-zero offset. */
if (addend != NULL_RTX)
{
newbase = gen_and_emit_move (HL, base, insn, true);
record_content (newbase, NULL_RTX);
newbase = gen_rtx_PLUS (HImode, newbase, addend);
OP (0) = OP (1) = change_address (OP (0), VOIDmode, newbase);
/* We do not want to fail here as this means that
we have inserted useless insns into the stream. */
MUST_BE_OK (insn);
}
}
else if (REG_P (OP (0))
&& satisfies_constraint_Ws1 (OP (2))
&& has_constraint (2, CONSTRAINT_Wh1))
{
rtx base, index, addend, newbase;
characterize_address (XEXP (OP (2), 0), & base, & index, & addend);
gcc_assert (index == NULL_RTX);
gcc_assert (REG_P (base) && REGNO (base) == SP_REG);
/* Ws1 addressing allows an offset of 0, Wh1 addressing requires a non-zero offset. */
if (addend != NULL_RTX)
{
gen_and_emit_move (HL, base, insn, true);
if (REGNO (OP (0)) != X_REG)
{
OP (1) = move_to_acc (1, insn);
OP (0) = move_from_acc (0, insn);
}
record_content (HL, NULL_RTX);
newbase = gen_rtx_PLUS (HImode, HL, addend);
OP (2) = change_address (OP (2), VOIDmode, newbase);
/* We do not want to fail here as this means that
we have inserted useless insns into the stream. */
MUST_BE_OK (insn);
}
}
}
OP (0) = move_from_acc (0, insn);
tmp_id = get_max_insn_count ();
saved_op1 = OP (1);
if (rtx_equal_p (OP (1), OP (2)))
OP (2) = OP (1) = move_to_acc (1, insn);
else
OP (1) = move_to_acc (1, insn);
MAYBE_OK (insn);
/* If we omitted the move of OP1 into the accumulator (because
it was already there from a previous insn), then force the
generation of the move instruction now. We know that we
are about to emit a move into HL (or DE) via AX, and hence
our optimization to remove the load of OP1 is no longer valid. */
if (tmp_id == get_max_insn_count ())
force_into_acc (saved_op1, insn);
/* We have to copy op2 to HL (or DE), but that involves AX, which
already has a live value. Emit it before those insns. */
if (prev)
first = next_nonnote_nondebug_insn (prev);
else
for (first = insn; prev_nonnote_nondebug_insn (first); first = prev_nonnote_nondebug_insn (first))
;
OP (2) = hl_used ? move_to_de (2, first) : move_to_hl (2, first);
MUST_BE_OK (insn);
}
/* Devirtualize an insn of the form SET (PC) (MEM/REG). */
static void
rl78_alloc_physical_registers_ro1 (rtx_insn * insn)
{
OP (0) = transcode_memory_rtx (OP (0), BC, insn);
MAYBE_OK (insn);
OP (0) = move_to_acc (0, insn);
MUST_BE_OK (insn);
}
/* Devirtualize a compare insn. */
static void
rl78_alloc_physical_registers_cmp (rtx_insn * insn)
{
int tmp_id;
rtx saved_op1;
rtx_insn *prev = prev_nonnote_nondebug_insn (insn);
rtx_insn *first;
OP (1) = transcode_memory_rtx (OP (1), DE, insn);
OP (2) = transcode_memory_rtx (OP (2), HL, insn);
/* HI compares have to have OP (1) in AX, but QI
compares do not, so it is worth checking here. */
MAYBE_OK (insn);
/* For an HImode compare, OP (1) must always be in AX.
But if OP (1) is a REG (and not AX), then we can avoid
a reload of OP (1) if we reload OP (2) into AX and invert
the comparison. */
if (REG_P (OP (1))
&& REGNO (OP (1)) != AX_REG
&& GET_MODE (OP (1)) == HImode
&& MEM_P (OP (2)))
{
rtx cmp = XEXP (SET_SRC (PATTERN (insn)), 0);
OP (2) = move_to_acc (2, insn);
switch (GET_CODE (cmp))
{
case EQ:
case NE:
break;
case LTU: cmp = gen_rtx_GTU (HImode, OP (2), OP (1)); break;
case GTU: cmp = gen_rtx_LTU (HImode, OP (2), OP (1)); break;
case LEU: cmp = gen_rtx_GEU (HImode, OP (2), OP (1)); break;
case GEU: cmp = gen_rtx_LEU (HImode, OP (2), OP (1)); break;
case LT:
case GT:
case LE:
case GE:
#if DEBUG_ALLOC
debug_rtx (insn);
#endif
default:
gcc_unreachable ();
}
if (GET_CODE (cmp) == EQ || GET_CODE (cmp) == NE)
PATTERN (insn) = gen_cbranchhi4_real (cmp, OP (2), OP (1), OP (3));
else
PATTERN (insn) = gen_cbranchhi4_real_inverted (cmp, OP (2), OP (1), OP (3));
MUST_BE_OK (insn);
}
/* Surprisingly, gcc can generate a comparison of a register with itself, but this
should be handled by the second alternative of the cbranchhi_real pattern. */
if (rtx_equal_p (OP (1), OP (2)))
{
OP (1) = OP (2) = BC;
MUST_BE_OK (insn);
}
tmp_id = get_max_insn_count ();
saved_op1 = OP (1);
OP (1) = move_to_acc (1, insn);
MAYBE_OK (insn);
/* If we omitted the move of OP1 into the accumulator (because
it was already there from a previous insn), then force the
generation of the move instruction now. We know that we
are about to emit a move into HL via AX, and hence our
optimization to remove the load of OP1 is no longer valid. */
if (tmp_id == get_max_insn_count ())
force_into_acc (saved_op1, insn);
/* We have to copy op2 to HL, but that involves the acc, which
already has a live value. Emit it before those insns. */
if (prev)
first = next_nonnote_nondebug_insn (prev);
else
for (first = insn; prev_nonnote_nondebug_insn (first); first = prev_nonnote_nondebug_insn (first))
;
OP (2) = move_to_hl (2, first);
MUST_BE_OK (insn);
}
/* Like op2, but AX = A * X. */
static void
rl78_alloc_physical_registers_umul (rtx_insn * insn)
{
rtx_insn *prev = prev_nonnote_nondebug_insn (insn);
rtx_insn *first;
int tmp_id;
rtx saved_op1;
OP (0) = transcode_memory_rtx (OP (0), BC, insn);
OP (1) = transcode_memory_rtx (OP (1), DE, insn);
OP (2) = transcode_memory_rtx (OP (2), HL, insn);
MAYBE_OK (insn);
if (recog_data.constraints[1][0] == '%'
&& is_virtual_register (OP (1))
&& !is_virtual_register (OP (2))
&& !CONSTANT_P (OP (2)))
{
rtx tmp = OP (1);
OP (1) = OP (2);
OP (2) = tmp;
}
OP (0) = move_from_acc (0, insn);
tmp_id = get_max_insn_count ();
saved_op1 = OP (1);
if (rtx_equal_p (OP (1), OP (2)))
{
gcc_assert (GET_MODE (OP (2)) == QImode);
/* The MULU instruction does not support duplicate arguments
but we know that if we copy OP (2) to X it will do so via
A and thus OP (1) will already be loaded into A. */
OP (2) = move_to_x (2, insn);
OP (1) = A;
}
else
OP (1) = move_to_acc (1, insn);
MAYBE_OK (insn);
/* If we omitted the move of OP1 into the accumulator (because
it was already there from a previous insn), then force the
generation of the move instruction now. We know that we
are about to emit a move into HL (or DE) via AX, and hence
our optimization to remove the load of OP1 is no longer valid. */
if (tmp_id == get_max_insn_count ())
force_into_acc (saved_op1, insn);
/* We have to copy op2 to X, but that involves the acc, which
already has a live value. Emit it before those insns. */
if (prev)
first = next_nonnote_nondebug_insn (prev);
else
for (first = insn; prev_nonnote_nondebug_insn (first); first = prev_nonnote_nondebug_insn (first))
;
OP (2) = move_to_x (2, first);
MUST_BE_OK (insn);
}
static void
rl78_alloc_address_registers_macax (rtx_insn * insn)
{
int which, op;
bool replace_in_op0 = false;
bool replace_in_op1 = false;
MAYBE_OK (insn);
/* Two different MEMs are not allowed. */
which = 0;
for (op = 2; op >= 0; op --)
{
if (MEM_P (OP (op)))
{
if (op == 0 && replace_in_op0)
continue;
if (op == 1 && replace_in_op1)
continue;
switch (which)
{
case 0:
/* If we replace a MEM, make sure that we replace it for all
occurrences of the same MEM in the insn. */
replace_in_op0 = (op > 0 && rtx_equal_p (OP (op), OP (0)));
replace_in_op1 = (op > 1 && rtx_equal_p (OP (op), OP (1)));
OP (op) = transcode_memory_rtx (OP (op), HL, insn);
if (op == 2
&& MEM_P (OP (op))
&& ((GET_CODE (XEXP (OP (op), 0)) == REG
&& REGNO (XEXP (OP (op), 0)) == SP_REG)
|| (GET_CODE (XEXP (OP (op), 0)) == PLUS
&& REGNO (XEXP (XEXP (OP (op), 0), 0)) == SP_REG)))
{
emit_insn_before (gen_movhi (HL, gen_rtx_REG (HImode, SP_REG)), insn);
OP (op) = replace_rtx (OP (op), gen_rtx_REG (HImode, SP_REG), HL);
}
if (replace_in_op0)
OP (0) = OP (op);
if (replace_in_op1)
OP (1) = OP (op);
break;
case 1:
OP (op) = transcode_memory_rtx (OP (op), DE, insn);
break;
case 2:
OP (op) = transcode_memory_rtx (OP (op), BC, insn);
break;
}
which ++;
}
}
MUST_BE_OK (insn);
}
static void
rl78_alloc_address_registers_div (rtx_insn * insn)
{
MUST_BE_OK (insn);
}
/* Scan all insns and devirtualize them. */
static void
rl78_alloc_physical_registers (void)
{
/* During most of the compile, gcc is dealing with virtual
registers. At this point, we need to assign physical registers
to the vitual ones, and copy in/out as needed. */
rtx_insn *insn, *curr;
enum attr_valloc valloc_method;
for (insn = get_insns (); insn; insn = curr)
{
int i;
curr = next_nonnote_nondebug_insn (insn);
if (INSN_P (insn)
&& (GET_CODE (PATTERN (insn)) == SET
|| GET_CODE (PATTERN (insn)) == CALL)
&& INSN_CODE (insn) == -1)
{
if (GET_CODE (SET_SRC (PATTERN (insn))) == ASM_OPERANDS)
continue;
i = recog (PATTERN (insn), insn, 0);
if (i == -1)
{
debug_rtx (insn);
gcc_unreachable ();
}
INSN_CODE (insn) = i;
}
}
cfun->machine->virt_insns_ok = 0;
cfun->machine->real_insns_ok = 1;
clear_content_memory ();
for (insn = get_insns (); insn; insn = curr)
{
rtx pattern;
curr = insn ? next_nonnote_nondebug_insn (insn) : NULL;
if (!INSN_P (insn))
{
if (LABEL_P (insn))
clear_content_memory ();
continue;
}
if (dump_file)
fprintf (dump_file, "Converting insn %d\n", INSN_UID (insn));
pattern = PATTERN (insn);
if (GET_CODE (pattern) == PARALLEL)
pattern = XVECEXP (pattern, 0, 0);
if (JUMP_P (insn) || CALL_P (insn) || GET_CODE (pattern) == CALL)
clear_content_memory ();
if (GET_CODE (pattern) != SET
&& GET_CODE (pattern) != CALL)
continue;
if (GET_CODE (pattern) == SET
&& GET_CODE (SET_SRC (pattern)) == ASM_OPERANDS)
continue;
valloc_method = get_attr_valloc (insn);
PATTERN (insn) = copy_rtx_if_shared (PATTERN (insn));
if (valloc_method == VALLOC_MACAX)
{
record_content (AX, NULL_RTX);
record_content (BC, NULL_RTX);
record_content (DE, NULL_RTX);
}
else if (valloc_method == VALLOC_DIVHI)
{
record_content (AX, NULL_RTX);
record_content (BC, NULL_RTX);
}
else if (valloc_method == VALLOC_DIVSI)
{
record_content (AX, NULL_RTX);
record_content (BC, NULL_RTX);
record_content (DE, NULL_RTX);
record_content (HL, NULL_RTX);
}
if (insn_ok_now (insn))
continue;
INSN_CODE (insn) = -1;
if (RTX_FRAME_RELATED_P (insn))
virt_insn_was_frame = 1;
else
virt_insn_was_frame = 0;
switch (valloc_method)
{
case VALLOC_OP1:
rl78_alloc_physical_registers_op1 (insn);
break;
case VALLOC_OP2:
rl78_alloc_physical_registers_op2 (insn);
break;
case VALLOC_RO1:
rl78_alloc_physical_registers_ro1 (insn);
break;
case VALLOC_CMP:
rl78_alloc_physical_registers_cmp (insn);
break;
case VALLOC_UMUL:
rl78_alloc_physical_registers_umul (insn);
record_content (AX, NULL_RTX);
break;
case VALLOC_MACAX:
/* Macro that clobbers AX. */
rl78_alloc_address_registers_macax (insn);
record_content (AX, NULL_RTX);
record_content (BC, NULL_RTX);
record_content (DE, NULL_RTX);
break;
case VALLOC_DIVSI:
rl78_alloc_address_registers_div (insn);
record_content (AX, NULL_RTX);
record_content (BC, NULL_RTX);
record_content (DE, NULL_RTX);
record_content (HL, NULL_RTX);
break;
case VALLOC_DIVHI:
rl78_alloc_address_registers_div (insn);
record_content (AX, NULL_RTX);
record_content (BC, NULL_RTX);
break;
default:
gcc_unreachable ();
}
if (JUMP_P (insn) || CALL_P (insn) || GET_CODE (pattern) == CALL)
clear_content_memory ();
else
process_postponed_content_update ();
}
#if DEBUG_ALLOC
fprintf (stderr, "\033[0m");
#endif
}
/* Add REG_DEAD notes using DEAD[reg] for rtx S which is part of INSN.
This function scans for uses of registers; the last use (i.e. first
encounter when scanning backwards) triggers a REG_DEAD note if the
reg was previously in DEAD[]. */
static void
rl78_note_reg_uses (char *dead, rtx s, rtx insn)
{
const char *fmt;
int i, r;
enum rtx_code code;
if (!s)
return;
code = GET_CODE (s);
switch (code)
{
/* Compare registers by number. */
case REG:
r = REGNO (s);
if (dump_file)
{
fprintf (dump_file, "note use reg %d size %d on insn %d\n",
r, GET_MODE_SIZE (GET_MODE (s)), INSN_UID (insn));
print_rtl_single (dump_file, s);
}
if (dead [r])
add_reg_note (insn, REG_DEAD, gen_rtx_REG (GET_MODE (s), r));
for (i = 0; i < GET_MODE_SIZE (GET_MODE (s)); i ++)
dead [r + i] = 0;
return;
/* These codes have no constituent expressions
and are unique. */
case SCRATCH:
case PC:
return;
case CONST_INT:
case CONST_VECTOR:
case CONST_DOUBLE:
case CONST_FIXED:
/* These are kept unique for a given value. */
return;
default:
break;
}
fmt = GET_RTX_FORMAT (code);
for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
{
if (fmt[i] == 'E')
{
int j;
for (j = XVECLEN (s, i) - 1; j >= 0; j--)
rl78_note_reg_uses (dead, XVECEXP (s, i, j), insn);
}
else if (fmt[i] == 'e')
rl78_note_reg_uses (dead, XEXP (s, i), insn);
}
}
/* Like the previous function, but scan for SETs instead. */
static void
rl78_note_reg_set (char *dead, rtx d, rtx insn)
{
int r, i;
bool is_dead;
if (GET_CODE (d) == MEM)
rl78_note_reg_uses (dead, XEXP (d, 0), insn);
if (GET_CODE (d) != REG)
return;
/* Do not mark the reg unused unless all QImode parts of it are dead. */
r = REGNO (d);
is_dead = true;
for (i = 0; i < GET_MODE_SIZE (GET_MODE (d)); i ++)
if (!dead [r + i])
is_dead = false;
if(is_dead)
add_reg_note (insn, REG_UNUSED, gen_rtx_REG (GET_MODE (d), r));
if (dump_file)
fprintf (dump_file, "note set reg %d size %d\n", r, GET_MODE_SIZE (GET_MODE (d)));
for (i = 0; i < GET_MODE_SIZE (GET_MODE (d)); i ++)
dead [r + i] = 1;
}
/* This is a rather crude register death pass. Death status is reset
at every jump or call insn. */
static void
rl78_calculate_death_notes (void)
{
char dead[FIRST_PSEUDO_REGISTER];
rtx p, s, d;
rtx_insn *insn;
int i;
memset (dead, 0, sizeof (dead));
for (insn = get_last_insn ();
insn;
insn = prev_nonnote_nondebug_insn (insn))
{
if (dump_file)
{
fprintf (dump_file, "\n--------------------------------------------------");
fprintf (dump_file, "\nDead:");
for (i = 0; i < FIRST_PSEUDO_REGISTER; i ++)
if (dead[i])
fprintf (dump_file, " %s", reg_names[i]);
fprintf (dump_file, "\n");
print_rtl_single (dump_file, insn);
}
switch (GET_CODE (insn))
{
case INSN:
p = PATTERN (insn);
if (GET_CODE (p) == PARALLEL)
{
rtx q = XVECEXP (p, 0 ,1);
/* This happens with the DIV patterns. */
if (GET_CODE (q) == SET)
{
s = SET_SRC (q);
d = SET_DEST (q);
rl78_note_reg_set (dead, d, insn);
rl78_note_reg_uses (dead, s, insn);
}
p = XVECEXP (p, 0, 0);
}
switch (GET_CODE (p))
{
case SET:
s = SET_SRC (p);
d = SET_DEST (p);
rl78_note_reg_set (dead, d, insn);
rl78_note_reg_uses (dead, s, insn);
break;
case USE:
rl78_note_reg_uses (dead, p, insn);
break;
default:
break;
}
break;
case JUMP_INSN:
if (INSN_CODE (insn) == CODE_FOR_rl78_return)
{
memset (dead, 1, sizeof (dead));
/* We expect a USE just prior to this, which will mark
the actual return registers. The USE will have a
death note, but we aren't going to be modifying it
after this pass. */
break;
}
/* FALLTHRU */
case CALL_INSN:
memset (dead, 0, sizeof (dead));
break;
default:
break;
}
if (dump_file)
print_rtl_single (dump_file, insn);
}
}
/* Helper function to reset the origins in RP and the age in AGE for
all registers. */
static void
reset_origins (int *rp, int *age)
{
int i;
for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
{
rp[i] = i;
age[i] = 0;
}
}
static void
set_origin (rtx pat, rtx_insn * insn, int * origins, int * age)
{
rtx src = SET_SRC (pat);
rtx dest = SET_DEST (pat);
int mb = GET_MODE_SIZE (GET_MODE (dest));
int i;
if (GET_CODE (dest) == REG)
{
int dr = REGNO (dest);
if (GET_CODE (src) == REG)
{
int sr = REGNO (src);
bool same = true;
int best_age, best_reg;
/* See if the copy is not needed. */
for (i = 0; i < mb; i ++)
if (origins[dr + i] != origins[sr + i])
same = false;
if (same)
{
if (dump_file)
fprintf (dump_file, "deleting because dest already has correct value\n");
delete_insn (insn);
return;
}
if (dr < 8 || sr >= 8)
{
int ar;
best_age = -1;
best_reg = -1;
/* See if the copy can be made from another
bank 0 register instead, instead of the
virtual src register. */
for (ar = 0; ar < 8; ar += mb)
{
same = true;
for (i = 0; i < mb; i ++)
if (origins[ar + i] != origins[sr + i])
same = false;
/* The chip has some reg-reg move limitations. */
if (mb == 1 && dr > 3)
same = false;
if (same)
{
if (best_age == -1 || best_age > age[sr + i])
{
best_age = age[sr + i];
best_reg = sr;
}
}
}
if (best_reg != -1)
{
/* FIXME: copy debug info too. */
SET_SRC (pat) = gen_rtx_REG (GET_MODE (src), best_reg);
sr = best_reg;
}
}
for (i = 0; i < mb; i++)
{
origins[dr + i] = origins[sr + i];
age[dr + i] = age[sr + i] + 1;
}
}
else
{
/* The destination is computed, its origin is itself. */
if (dump_file)
fprintf (dump_file, "resetting origin of r%d for %d byte%s\n",
dr, mb, mb == 1 ? "" : "s");
for (i = 0; i < mb; i ++)
{
origins[dr + i] = dr + i;
age[dr + i] = 0;
}
}
/* Any registers marked with that reg as an origin are reset. */
for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
if (origins[i] >= dr && origins[i] < dr + mb)
{
origins[i] = i;
age[i] = 0;
}
}
/* Special case - our MUL patterns uses AX and sometimes BC. */
if (get_attr_valloc (insn) == VALLOC_MACAX)
{
if (dump_file)
fprintf (dump_file, "Resetting origin of AX/BC for MUL pattern.\n");
for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
if (i <= 3 || origins[i] <= 3)
{
origins[i] = i;
age[i] = 0;
}
}
else if (get_attr_valloc (insn) == VALLOC_DIVHI)
{
if (dump_file)
fprintf (dump_file, "Resetting origin of AX/DE for DIVHI pattern.\n");
for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
if (i == A_REG
|| i == X_REG
|| i == D_REG
|| i == E_REG
|| origins[i] == A_REG
|| origins[i] == X_REG
|| origins[i] == D_REG
|| origins[i] == E_REG)
{
origins[i] = i;
age[i] = 0;
}
}
else if (get_attr_valloc (insn) == VALLOC_DIVSI)
{
if (dump_file)
fprintf (dump_file, "Resetting origin of AX/BC/DE/HL for DIVSI pattern.\n");
for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
if (i <= 7 || origins[i] <= 7)
{
origins[i] = i;
age[i] = 0;
}
}
if (GET_CODE (src) == ASHIFT
|| GET_CODE (src) == ASHIFTRT
|| GET_CODE (src) == LSHIFTRT)
{
rtx count = XEXP (src, 1);
if (GET_CODE (count) == REG)
{
/* Special case - our pattern clobbers the count register. */
int r = REGNO (count);
if (dump_file)
fprintf (dump_file, "Resetting origin of r%d for shift.\n", r);
for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
if (i == r || origins[i] == r)
{
origins[i] = i;
age[i] = 0;
}
}
}
}
/* The idea behind this optimization is to look for cases where we
move data from A to B to C, and instead move from A to B, and A to
C. If B is a virtual register or memory, this is a big win on its
own. If B turns out to be unneeded after this, it's a bigger win.
For each register, we try to determine where it's value originally
came from, if it's propogated purely through moves (and not
computes). The ORIGINS[] array has the regno for the "origin" of
the value in the [regno] it's indexed by. */
static void
rl78_propogate_register_origins (void)
{
int origins[FIRST_PSEUDO_REGISTER];
int age[FIRST_PSEUDO_REGISTER];
int i;
rtx_insn *insn, *ninsn = NULL;
rtx pat;
reset_origins (origins, age);
for (insn = get_insns (); insn; insn = ninsn)
{
ninsn = next_nonnote_nondebug_insn (insn);
if (dump_file)
{
fprintf (dump_file, "\n");
fprintf (dump_file, "Origins:");
for (i = 0; i < FIRST_PSEUDO_REGISTER; i ++)
if (origins[i] != i)
fprintf (dump_file, " r%d=r%d", i, origins[i]);
fprintf (dump_file, "\n");
print_rtl_single (dump_file, insn);
}
switch (GET_CODE (insn))
{
case CODE_LABEL:
case BARRIER:
case CALL_INSN:
case JUMP_INSN:
reset_origins (origins, age);
break;
default:
break;
case INSN:
pat = PATTERN (insn);
if (GET_CODE (pat) == PARALLEL)
{
rtx clobber = XVECEXP (pat, 0, 1);
pat = XVECEXP (pat, 0, 0);
if (GET_CODE (clobber) == CLOBBER
&& GET_CODE (XEXP (clobber, 0)) == REG)
{
int cr = REGNO (XEXP (clobber, 0));
int mb = GET_MODE_SIZE (GET_MODE (XEXP (clobber, 0)));
if (dump_file)
fprintf (dump_file, "reset origins of %d regs at %d\n", mb, cr);
for (i = 0; i < mb; i++)
{
origins[cr + i] = cr + i;
age[cr + i] = 0;
}
}
/* This happens with the DIV patterns. */
else if (GET_CODE (clobber) == SET)
{
set_origin (clobber, insn, origins, age);
}
else
break;
}
if (GET_CODE (pat) == SET)
{
set_origin (pat, insn, origins, age);
}
else if (GET_CODE (pat) == CLOBBER
&& GET_CODE (XEXP (pat, 0)) == REG)
{
if (REG_P (XEXP (pat, 0)))
{
unsigned int reg = REGNO (XEXP (pat, 0));
origins[reg] = reg;
age[reg] = 0;
}
}
}
}
}
/* Remove any SETs where the destination is unneeded. */
static void
rl78_remove_unused_sets (void)
{
rtx_insn *insn, *ninsn = NULL;
rtx dest;
for (insn = get_insns (); insn; insn = ninsn)
{
ninsn = next_nonnote_nondebug_insn (insn);
rtx set = single_set (insn);
if (set == NULL)
continue;
dest = SET_DEST (set);
if (GET_CODE (dest) != REG || REGNO (dest) > 23)
continue;
if (find_regno_note (insn, REG_UNUSED, REGNO (dest)))
{
if (dump_file)
fprintf (dump_file, "deleting because the set register is never used.\n");
delete_insn (insn);
}
}
}
/* This is the top of the devritualization pass. */
static void
rl78_reorg (void)
{
/* split2 only happens when optimizing, but we need all movSIs to be
split now. */
if (optimize <= 0)
split_all_insns ();
rl78_alloc_physical_registers ();
if (dump_file)
{
fprintf (dump_file, "\n================DEVIRT:=AFTER=ALLOC=PHYSICAL=REGISTERS================\n");
print_rtl_with_bb (dump_file, get_insns (), TDF_NONE);
}
rl78_propogate_register_origins ();
rl78_calculate_death_notes ();
if (dump_file)
{
fprintf (dump_file, "\n================DEVIRT:=AFTER=PROPOGATION=============================\n");
print_rtl_with_bb (dump_file, get_insns (), TDF_NONE);
fprintf (dump_file, "\n======================================================================\n");
}
rl78_remove_unused_sets ();
/* The code after devirtualizing has changed so much that at this point
we might as well just rescan everything. Note that
df_rescan_all_insns is not going to help here because it does not
touch the artificial uses and defs. */
df_finish_pass (true);
if (optimize > 1)
df_live_add_problem ();
df_scan_alloc (NULL);
df_scan_blocks ();
if (optimize)
df_analyze ();
}
#undef TARGET_RETURN_IN_MEMORY
#define TARGET_RETURN_IN_MEMORY rl78_return_in_memory
static bool
rl78_return_in_memory (const_tree type, const_tree fntype ATTRIBUTE_UNUSED)
{
const HOST_WIDE_INT size = int_size_in_bytes (type);
return (size == -1 || size > 8);
}
#undef TARGET_RTX_COSTS
#define TARGET_RTX_COSTS rl78_rtx_costs
static bool
rl78_rtx_costs (rtx x,
machine_mode mode,
int outer_code ATTRIBUTE_UNUSED,
int opno ATTRIBUTE_UNUSED,
int * total,
bool speed ATTRIBUTE_UNUSED)
{
int code = GET_CODE (x);
if (code == IF_THEN_ELSE)
{
*total = COSTS_N_INSNS (10);
return true;
}
if (mode == HImode)
{
if (code == MULT && ! speed)
{
* total = COSTS_N_INSNS (8);
return true;
}
return false;
}
if (mode == SImode)
{
switch (code)
{
case MULT:
if (! speed)
/* If we are compiling for space then we do not want to use the
inline SImode multiplication patterns or shift sequences.
The cost is not set to 1 or 5 however as we have to allow for
the possibility that we might be converting a leaf function
into a non-leaf function. (There is no way to tell here).
A value of 13 seems to be a reasonable compromise for the
moment. */
* total = COSTS_N_INSNS (13);
else if (RL78_MUL_G14)
*total = COSTS_N_INSNS (14);
else if (RL78_MUL_G13)
*total = COSTS_N_INSNS (29);
else
*total = COSTS_N_INSNS (500);
return true;
case PLUS:
*total = COSTS_N_INSNS (8);
return true;
case ASHIFT:
case ASHIFTRT:
case LSHIFTRT:
if (GET_CODE (XEXP (x, 1)) == CONST_INT)
{
switch (INTVAL (XEXP (x, 1)))
{
case 0: *total = COSTS_N_INSNS (0); break;
case 1: *total = COSTS_N_INSNS (6); break;
case 2: case 3: case 4: case 5: case 6: case 7:
*total = COSTS_N_INSNS (10); break;
case 8: *total = COSTS_N_INSNS (6); break;
case 9: case 10: case 11: case 12: case 13: case 14: case 15:
*total = COSTS_N_INSNS (10); break;
case 16: *total = COSTS_N_INSNS (3); break;
case 17: case 18: case 19: case 20: case 21: case 22: case 23:
*total = COSTS_N_INSNS (4); break;
case 24: *total = COSTS_N_INSNS (4); break;
case 25: case 26: case 27: case 28: case 29: case 30: case 31:
*total = COSTS_N_INSNS (5); break;
}
}
else
*total = COSTS_N_INSNS (10+4*16);
return true;
default:
break;
}
}
return false;
}
static GTY(()) section * saddr_section;
static GTY(()) section * frodata_section;
int
rl78_saddr_p (rtx x)
{
const char * c;
if (MEM_P (x))
x = XEXP (x, 0);
if (GET_CODE (x) == PLUS)
x = XEXP (x, 0);
if (GET_CODE (x) != SYMBOL_REF)
return 0;
c = XSTR (x, 0);
if (memcmp (c, "@s.", 3) == 0)
return 1;
return 0;
}
int
rl78_sfr_p (rtx x)
{
if (MEM_P (x))
x = XEXP (x, 0);
if (GET_CODE (x) != CONST_INT)
return 0;
if ((INTVAL (x) & 0xFF00) != 0xFF00)
return 0;
return 1;
}
#undef TARGET_STRIP_NAME_ENCODING
#define TARGET_STRIP_NAME_ENCODING rl78_strip_name_encoding
static const char *
rl78_strip_name_encoding (const char * sym)
{
while (1)
{
if (*sym == '*')
sym++;
else if (*sym == '@' && sym[2] == '.')
sym += 3;
else
return sym;
}
}
/* Like rl78_strip_name_encoding, but does not strip leading asterisks. This
is important if the stripped name is going to be passed to assemble_name()
as that handles asterisk prefixed names in a special manner. */
static const char *
rl78_strip_nonasm_name_encoding (const char * sym)
{
while (1)
{
if (*sym == '@' && sym[2] == '.')
sym += 3;
else
return sym;
}
}
static int
rl78_attrlist_to_encoding (tree list, tree decl ATTRIBUTE_UNUSED)
{
while (list)
{
if (is_attribute_p ("saddr", TREE_PURPOSE (list)))
return 's';
list = TREE_CHAIN (list);
}
return 0;
}
#define RL78_ATTRIBUTES(decl) \
(TYPE_P (decl)) ? TYPE_ATTRIBUTES (decl) \
: DECL_ATTRIBUTES (decl) \
? (DECL_ATTRIBUTES (decl)) \
: TYPE_ATTRIBUTES (TREE_TYPE (decl))
#undef TARGET_ENCODE_SECTION_INFO
#define TARGET_ENCODE_SECTION_INFO rl78_encode_section_info
static void
rl78_encode_section_info (tree decl, rtx rtl, int first)
{
rtx rtlname;
const char * oldname;
char encoding;
char * newname;
tree idp;
tree type;
tree rl78_attributes;
if (!first)
return;
rtlname = XEXP (rtl, 0);
if (GET_CODE (rtlname) == SYMBOL_REF)
oldname = XSTR (rtlname, 0);
else if (GET_CODE (rtlname) == MEM
&& GET_CODE (XEXP (rtlname, 0)) == SYMBOL_REF)
oldname = XSTR (XEXP (rtlname, 0), 0);
else
gcc_unreachable ();
type = TREE_TYPE (decl);
if (type == error_mark_node)
return;
if (! DECL_P (decl))
return;
rl78_attributes = RL78_ATTRIBUTES (decl);
encoding = rl78_attrlist_to_encoding (rl78_attributes, decl);
if (encoding)
{
newname = (char *) alloca (strlen (oldname) + 4);
sprintf (newname, "@%c.%s", encoding, oldname);
idp = get_identifier (newname);
XEXP (rtl, 0) =
gen_rtx_SYMBOL_REF (Pmode, IDENTIFIER_POINTER (idp));
SYMBOL_REF_WEAK (XEXP (rtl, 0)) = DECL_WEAK (decl);
SET_SYMBOL_REF_DECL (XEXP (rtl, 0), decl);
}
}
#undef TARGET_ASM_INIT_SECTIONS
#define TARGET_ASM_INIT_SECTIONS rl78_asm_init_sections
static void
rl78_asm_init_sections (void)
{
saddr_section
= get_unnamed_section (SECTION_WRITE, output_section_asm_op,
"\t.section .saddr,\"aw\",@progbits");
frodata_section
= get_unnamed_section (SECTION_WRITE, output_section_asm_op,
"\t.section .frodata,\"aw\",@progbits");
}
#undef TARGET_ASM_SELECT_SECTION
#define TARGET_ASM_SELECT_SECTION rl78_select_section
static section *
rl78_select_section (tree decl,
int reloc,
unsigned HOST_WIDE_INT align)
{
int readonly = 1;
switch (TREE_CODE (decl))
{
case VAR_DECL:
if (!TREE_READONLY (decl)
|| TREE_SIDE_EFFECTS (decl)
|| !DECL_INITIAL (decl)
|| (DECL_INITIAL (decl) != error_mark_node
&& !TREE_CONSTANT (DECL_INITIAL (decl))))
readonly = 0;
break;
case CONSTRUCTOR:
if (! TREE_CONSTANT (decl))
readonly = 0;
break;
default:
break;
}
if (TREE_CODE (decl) == VAR_DECL)
{
const char *name = XSTR (XEXP (DECL_RTL (decl), 0), 0);
if (name[0] == '@' && name[2] == '.')
switch (name[1])
{
case 's':
return saddr_section;
}
if (TYPE_ADDR_SPACE (TREE_TYPE (decl)) == ADDR_SPACE_FAR
&& readonly)
{
return frodata_section;
}
}
if (readonly)
return TARGET_ES0 ? frodata_section : readonly_data_section;
switch (categorize_decl_for_section (decl, reloc))
{
case SECCAT_TEXT: return text_section;
case SECCAT_DATA: return data_section;
case SECCAT_BSS: return bss_section;
case SECCAT_RODATA: return TARGET_ES0 ? frodata_section : readonly_data_section;
default:
return default_select_section (decl, reloc, align);
}
}
void
rl78_output_labelref (FILE *file, const char *str)
{
const char *str2;
str2 = targetm.strip_name_encoding (str);
if (str2[0] != '.')
fputs (user_label_prefix, file);
fputs (str2, file);
}
void
rl78_output_aligned_common (FILE *stream,
tree decl ATTRIBUTE_UNUSED,
const char *name,
int size, int align, int global)
{
/* We intentionally don't use rl78_section_tag() here. */
if (name[0] == '@' && name[2] == '.')
{
const char *sec = 0;
switch (name[1])
{
case 's':
switch_to_section (saddr_section);
sec = ".saddr";
break;
}
if (sec)
{
const char *name2;
int p2align = 0;
while (align > BITS_PER_UNIT)
{
align /= 2;
p2align ++;
}
name2 = targetm.strip_name_encoding (name);
if (global)
fprintf (stream, "\t.global\t_%s\n", name2);
fprintf (stream, "\t.p2align %d\n", p2align);
fprintf (stream, "\t.type\t_%s,@object\n", name2);
fprintf (stream, "\t.size\t_%s,%d\n", name2, size);
fprintf (stream, "_%s:\n\t.zero\t%d\n", name2, size);
return;
}
}
if (!global)
{
fprintf (stream, "\t.local\t");
assemble_name (stream, name);
fprintf (stream, "\n");
}
fprintf (stream, "\t.comm\t");
assemble_name (stream, name);
fprintf (stream, ",%u,%u\n", size, align / BITS_PER_UNIT);
}
#undef TARGET_INSERT_ATTRIBUTES
#define TARGET_INSERT_ATTRIBUTES rl78_insert_attributes
static void
rl78_insert_attributes (tree decl, tree *attributes ATTRIBUTE_UNUSED)
{
if (TARGET_ES0
&& TREE_CODE (decl) == VAR_DECL
&& TREE_READONLY (decl)
&& TREE_ADDRESSABLE (decl)
&& TYPE_ADDR_SPACE (TREE_TYPE (decl)) == ADDR_SPACE_GENERIC)
{
tree type = TREE_TYPE (decl);
tree attr = TYPE_ATTRIBUTES (type);
int q = TYPE_QUALS_NO_ADDR_SPACE (type) | ENCODE_QUAL_ADDR_SPACE (ADDR_SPACE_FAR);
TREE_TYPE (decl) = build_type_attribute_qual_variant (type, attr, q);
}
}
#undef TARGET_ASM_INTEGER
#define TARGET_ASM_INTEGER rl78_asm_out_integer
static bool
rl78_asm_out_integer (rtx x, unsigned int size, int aligned_p)
{
if (default_assemble_integer (x, size, aligned_p))
return true;
if (size == 4)
{
assemble_integer_with_op (".long\t", x);
return true;
}
return false;
}
#undef TARGET_UNWIND_WORD_MODE
#define TARGET_UNWIND_WORD_MODE rl78_unwind_word_mode
static scalar_int_mode
rl78_unwind_word_mode (void)
{
return HImode;
}
#ifndef USE_COLLECT2
#undef TARGET_ASM_CONSTRUCTOR
#define TARGET_ASM_CONSTRUCTOR rl78_asm_constructor
#undef TARGET_ASM_DESTRUCTOR
#define TARGET_ASM_DESTRUCTOR rl78_asm_destructor
static void
rl78_asm_ctor_dtor (rtx symbol, int priority, bool is_ctor)
{
section *sec;
if (priority != DEFAULT_INIT_PRIORITY)
{
/* This section of the function is based upon code copied
from: gcc/varasm.c:get_cdtor_priority_section(). */
char buf[18];
sprintf (buf, "%s.%.5u", is_ctor ? ".ctors" : ".dtors",
MAX_INIT_PRIORITY - priority);
sec = get_section (buf, 0, NULL);
}
else
sec = is_ctor ? ctors_section : dtors_section;
assemble_addr_to_section (symbol, sec);
}
static void
rl78_asm_constructor (rtx symbol, int priority)
{
rl78_asm_ctor_dtor (symbol, priority, true);
}
static void
rl78_asm_destructor (rtx symbol, int priority)
{
rl78_asm_ctor_dtor (symbol, priority, false);
}
#endif /* ! USE_COLLECT2 */
/* Scan backwards through the insn chain looking to see if the flags
have been set for a comparison of OP against OPERAND. Start with
the insn *before* the current insn. */
bool
rl78_flags_already_set (rtx op, rtx operand)
{
/* We only track the Z flag. */
if (GET_CODE (op) != EQ && GET_CODE (op) != NE)
return false;
/* This should not happen, but let's be paranoid. */
if (current_output_insn == NULL_RTX)
return false;
rtx_insn *insn;
bool res = false;
for (insn = prev_nonnote_nondebug_insn (current_output_insn);
insn != NULL_RTX;
insn = prev_nonnote_nondebug_insn (insn))
{
if (LABEL_P (insn))
break;
if (! INSN_P (insn))
continue;
/* Make sure that the insn can be recognized. */
if (recog_memoized (insn) == -1)
continue;
enum attr_update_Z updated = get_attr_update_Z (insn);
rtx set = single_set (insn);
bool must_break = (set != NULL_RTX && rtx_equal_p (operand, SET_DEST (set)));
switch (updated)
{
case UPDATE_Z_NO:
break;
case UPDATE_Z_CLOBBER:
must_break = true;
break;
case UPDATE_Z_UPDATE_Z:
res = must_break;
must_break = true;
break;
default:
gcc_unreachable ();
}
if (must_break)
break;
}
/* We have to re-recognize the current insn as the call(s) to
get_attr_update_Z() above will have overwritten the recog_data cache. */
recog_memoized (current_output_insn);
cleanup_subreg_operands (current_output_insn);
constrain_operands_cached (current_output_insn, 1);
return res;
}
const char *
rl78_addsi3_internal (rtx * operands, unsigned int alternative)
{
const char *addH2 = "addw ax, %H2\n\t";
/* If we are adding in a constant symbolic address when -mes0
is active then we know that the address must be <64K and
that it is invalid to access anything above 64K relative to
this address. So we can skip adding in the high bytes. */
if (TARGET_ES0
&& GET_CODE (operands[2]) == SYMBOL_REF
&& TREE_CODE (SYMBOL_REF_DECL (operands[2])) == VAR_DECL
&& TREE_READONLY (SYMBOL_REF_DECL (operands[2]))
&& ! TREE_SIDE_EFFECTS (SYMBOL_REF_DECL (operands[2])))
return "movw ax, %h1\n\taddw ax, %h2\n\tmovw %h0, ax";
if(CONST_INT_P(operands[2]))
{
if((INTVAL(operands[2]) & 0xFFFF0000) == 0)
{
addH2 = "";
}
else if((INTVAL(operands[2]) & 0xFFFF0000) == 0x00010000)
{
addH2 = "incw ax\n\t";
}
else if((INTVAL(operands[2]) & 0xFFFF0000) == 0xFFFF0000)
{
addH2 = "decw ax\n\t";
}
}
switch (alternative)
{
case 0:
case 1:
snprintf(fmt_buffer, sizeof(fmt_buffer),
"movw ax, %%h1\n\taddw ax, %%h2\n\tmovw %%h0, ax\n\tmovw ax, %%H1\n\tsknc\n\tincw ax\n\t%smovw %%H0,ax", addH2);
break;
case 2:
snprintf(fmt_buffer, sizeof(fmt_buffer),
"movw ax, %%h1\n\taddw ax, %%h2\n\tmovw bc, ax\n\tmovw ax, %%H1\n\tsknc\n\tincw ax\n\t%smovw %%H0, ax\n\tmovw ax, bc\n\tmovw %%h0, ax", addH2);
break;
default:
gcc_unreachable ();
}
return fmt_buffer;
}
rtx
rl78_emit_libcall (const char *name, enum rtx_code code,
enum machine_mode dmode, enum machine_mode smode,
int noperands, rtx *operands)
{
rtx ret;
rtx_insn *insns;
rtx libcall;
rtx equiv;
start_sequence ();
libcall = gen_rtx_SYMBOL_REF (Pmode, name);
switch (noperands)
{
case 2:
ret = emit_library_call_value (libcall, NULL_RTX, LCT_CONST,
dmode, operands[1], smode);
equiv = gen_rtx_fmt_e (code, dmode, operands[1]);
break;
case 3:
ret = emit_library_call_value (libcall, NULL_RTX,
LCT_CONST, dmode,
operands[1], smode, operands[2],
smode);
equiv = gen_rtx_fmt_ee (code, dmode, operands[1], operands[2]);
break;
default:
gcc_unreachable ();
}
insns = get_insns ();
end_sequence ();
emit_libcall_block (insns, operands[0], ret, equiv);
return ret;
}
#undef TARGET_PREFERRED_RELOAD_CLASS
#define TARGET_PREFERRED_RELOAD_CLASS rl78_preferred_reload_class
static reg_class_t
rl78_preferred_reload_class (rtx x ATTRIBUTE_UNUSED, reg_class_t rclass)
{
if (rclass == NO_REGS)
rclass = V_REGS;
return rclass;
}
struct gcc_target targetm = TARGET_INITIALIZER;
#include "gt-rl78.h"