/* Target machine subroutines for Altera Nios II. Copyright (C) 2012-2016 Free Software Foundation, Inc. Contributed by Jonah Graham (jgraham@altera.com), Will Reece (wreece@altera.com), and Jeff DaSilva (jdasilva@altera.com). Contributed by Mentor Graphics, Inc. This file is part of GCC. GCC is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 3, or (at your option) any later version. GCC is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with GCC; see the file COPYING3. If not see . */ #include "config.h" #include "system.h" #include "coretypes.h" #include "backend.h" #include "target.h" #include "rtl.h" #include "tree.h" #include "df.h" #include "memmodel.h" #include "tm_p.h" #include "optabs.h" #include "regs.h" #include "emit-rtl.h" #include "recog.h" #include "diagnostic-core.h" #include "output.h" #include "insn-attr.h" #include "flags.h" #include "explow.h" #include "calls.h" #include "varasm.h" #include "expr.h" #include "toplev.h" #include "langhooks.h" #include "stor-layout.h" #include "builtins.h" /* This file should be included last. */ #include "target-def.h" /* Forward function declarations. */ static bool prologue_saved_reg_p (unsigned); static void nios2_load_pic_register (void); static void nios2_register_custom_code (unsigned int, enum nios2_ccs_code, int); static const char *nios2_unspec_reloc_name (int); static void nios2_register_builtin_fndecl (unsigned, tree); static rtx nios2_ldst_parallel (bool, bool, bool, rtx, int, unsigned HOST_WIDE_INT, bool); /* Threshold for data being put into the small data/bss area, instead of the normal data area (references to the small data/bss area take 1 instruction, and use the global pointer, references to the normal data area takes 2 instructions). */ unsigned HOST_WIDE_INT nios2_section_threshold = NIOS2_DEFAULT_GVALUE; struct GTY (()) machine_function { /* Current frame information, to be filled in by nios2_compute_frame_layout with register save masks, and offsets for the current function. */ /* Mask of registers to save. */ unsigned int save_mask; /* Number of bytes that the entire frame takes up. */ int total_size; /* Number of bytes that variables take up. */ int var_size; /* Number of bytes that outgoing arguments take up. */ int args_size; /* Number of bytes needed to store registers in frame. */ int save_reg_size; /* Number of bytes used to store callee-saved registers. */ int callee_save_reg_size; /* Offset from new stack pointer to store registers. */ int save_regs_offset; /* Offset from save_regs_offset to store frame pointer register. */ int fp_save_offset; /* != 0 if function has a variable argument list. */ int uses_anonymous_args; /* != 0 if frame layout already calculated. */ int initialized; }; /* State to track the assignment of custom codes to FPU/custom builtins. */ static enum nios2_ccs_code custom_code_status[256]; static int custom_code_index[256]; /* Set to true if any conflicts (re-use of a code between 0-255) are found. */ static bool custom_code_conflict = false; /* Definition of builtin function types for nios2. */ #define N2_FTYPES \ N2_FTYPE(1, (SF)) \ N2_FTYPE(1, (VOID)) \ N2_FTYPE(2, (DF, DF)) \ N2_FTYPE(3, (DF, DF, DF)) \ N2_FTYPE(2, (DF, SF)) \ N2_FTYPE(2, (DF, SI)) \ N2_FTYPE(2, (DF, UI)) \ N2_FTYPE(2, (SF, DF)) \ N2_FTYPE(2, (SF, SF)) \ N2_FTYPE(3, (SF, SF, SF)) \ N2_FTYPE(2, (SF, SI)) \ N2_FTYPE(2, (SF, UI)) \ N2_FTYPE(2, (SI, CVPTR)) \ N2_FTYPE(2, (SI, DF)) \ N2_FTYPE(3, (SI, DF, DF)) \ N2_FTYPE(2, (SI, SF)) \ N2_FTYPE(3, (SI, SF, SF)) \ N2_FTYPE(2, (SI, SI)) \ N2_FTYPE(3, (SI, SI, SI)) \ N2_FTYPE(3, (SI, VPTR, SI)) \ N2_FTYPE(2, (UI, CVPTR)) \ N2_FTYPE(2, (UI, DF)) \ N2_FTYPE(2, (UI, SF)) \ N2_FTYPE(2, (VOID, DF)) \ N2_FTYPE(2, (VOID, SF)) \ N2_FTYPE(2, (VOID, SI)) \ N2_FTYPE(3, (VOID, SI, SI)) \ N2_FTYPE(2, (VOID, VPTR)) \ N2_FTYPE(3, (VOID, VPTR, SI)) #define N2_FTYPE_OP1(R) N2_FTYPE_ ## R ## _VOID #define N2_FTYPE_OP2(R, A1) N2_FTYPE_ ## R ## _ ## A1 #define N2_FTYPE_OP3(R, A1, A2) N2_FTYPE_ ## R ## _ ## A1 ## _ ## A2 /* Expand ftcode enumeration. */ enum nios2_ftcode { #define N2_FTYPE(N,ARGS) N2_FTYPE_OP ## N ARGS, N2_FTYPES #undef N2_FTYPE N2_FTYPE_MAX }; /* Return the tree function type, based on the ftcode. */ static tree nios2_ftype (enum nios2_ftcode ftcode) { static tree types[(int) N2_FTYPE_MAX]; tree N2_TYPE_SF = float_type_node; tree N2_TYPE_DF = double_type_node; tree N2_TYPE_SI = integer_type_node; tree N2_TYPE_UI = unsigned_type_node; tree N2_TYPE_VOID = void_type_node; static const_tree N2_TYPE_CVPTR, N2_TYPE_VPTR; if (!N2_TYPE_CVPTR) { /* const volatile void *. */ N2_TYPE_CVPTR = build_pointer_type (build_qualified_type (void_type_node, (TYPE_QUAL_CONST | TYPE_QUAL_VOLATILE))); /* volatile void *. */ N2_TYPE_VPTR = build_pointer_type (build_qualified_type (void_type_node, TYPE_QUAL_VOLATILE)); } if (types[(int) ftcode] == NULL_TREE) switch (ftcode) { #define N2_FTYPE_ARGS1(R) N2_TYPE_ ## R #define N2_FTYPE_ARGS2(R,A1) N2_TYPE_ ## R, N2_TYPE_ ## A1 #define N2_FTYPE_ARGS3(R,A1,A2) N2_TYPE_ ## R, N2_TYPE_ ## A1, N2_TYPE_ ## A2 #define N2_FTYPE(N,ARGS) \ case N2_FTYPE_OP ## N ARGS: \ types[(int) ftcode] \ = build_function_type_list (N2_FTYPE_ARGS ## N ARGS, NULL_TREE); \ break; N2_FTYPES #undef N2_FTYPE default: gcc_unreachable (); } return types[(int) ftcode]; } /* Definition of FPU instruction descriptions. */ struct nios2_fpu_insn_info { const char *name; int num_operands, *optvar; int opt, no_opt; #define N2F_DF 0x1 #define N2F_DFREQ 0x2 #define N2F_UNSAFE 0x4 #define N2F_FINITE 0x8 #define N2F_NO_ERRNO 0x10 unsigned int flags; enum insn_code icode; enum nios2_ftcode ftcode; }; /* Base macro for defining FPU instructions. */ #define N2FPU_INSN_DEF_BASE(insn, nop, flags, icode, args) \ { #insn, nop, &nios2_custom_ ## insn, OPT_mcustom_##insn##_, \ OPT_mno_custom_##insn, flags, CODE_FOR_ ## icode, \ N2_FTYPE_OP ## nop args } /* Arithmetic and math functions; 2 or 3 operand FP operations. */ #define N2FPU_OP2(mode) (mode, mode) #define N2FPU_OP3(mode) (mode, mode, mode) #define N2FPU_INSN_DEF(code, icode, nop, flags, m, M) \ N2FPU_INSN_DEF_BASE (f ## code ## m, nop, flags, \ icode ## m ## f ## nop, N2FPU_OP ## nop (M ## F)) #define N2FPU_INSN_SF(code, nop, flags) \ N2FPU_INSN_DEF (code, code, nop, flags, s, S) #define N2FPU_INSN_DF(code, nop, flags) \ N2FPU_INSN_DEF (code, code, nop, flags | N2F_DF, d, D) /* Compare instructions, 3 operand FP operation with a SI result. */ #define N2FPU_CMP_DEF(code, flags, m, M) \ N2FPU_INSN_DEF_BASE (fcmp ## code ## m, 3, flags, \ nios2_s ## code ## m ## f, (SI, M ## F, M ## F)) #define N2FPU_CMP_SF(code) N2FPU_CMP_DEF (code, 0, s, S) #define N2FPU_CMP_DF(code) N2FPU_CMP_DEF (code, N2F_DF, d, D) /* The order of definition needs to be maintained consistent with enum n2fpu_code in nios2-opts.h. */ struct nios2_fpu_insn_info nios2_fpu_insn[] = { /* Single precision instructions. */ N2FPU_INSN_SF (add, 3, 0), N2FPU_INSN_SF (sub, 3, 0), N2FPU_INSN_SF (mul, 3, 0), N2FPU_INSN_SF (div, 3, 0), /* Due to textual difference between min/max and smin/smax. */ N2FPU_INSN_DEF (min, smin, 3, N2F_FINITE, s, S), N2FPU_INSN_DEF (max, smax, 3, N2F_FINITE, s, S), N2FPU_INSN_SF (neg, 2, 0), N2FPU_INSN_SF (abs, 2, 0), N2FPU_INSN_SF (sqrt, 2, 0), N2FPU_INSN_SF (sin, 2, N2F_UNSAFE), N2FPU_INSN_SF (cos, 2, N2F_UNSAFE), N2FPU_INSN_SF (tan, 2, N2F_UNSAFE), N2FPU_INSN_SF (atan, 2, N2F_UNSAFE), N2FPU_INSN_SF (exp, 2, N2F_UNSAFE), N2FPU_INSN_SF (log, 2, N2F_UNSAFE), /* Single precision compares. */ N2FPU_CMP_SF (eq), N2FPU_CMP_SF (ne), N2FPU_CMP_SF (lt), N2FPU_CMP_SF (le), N2FPU_CMP_SF (gt), N2FPU_CMP_SF (ge), /* Double precision instructions. */ N2FPU_INSN_DF (add, 3, 0), N2FPU_INSN_DF (sub, 3, 0), N2FPU_INSN_DF (mul, 3, 0), N2FPU_INSN_DF (div, 3, 0), /* Due to textual difference between min/max and smin/smax. */ N2FPU_INSN_DEF (min, smin, 3, N2F_FINITE, d, D), N2FPU_INSN_DEF (max, smax, 3, N2F_FINITE, d, D), N2FPU_INSN_DF (neg, 2, 0), N2FPU_INSN_DF (abs, 2, 0), N2FPU_INSN_DF (sqrt, 2, 0), N2FPU_INSN_DF (sin, 2, N2F_UNSAFE), N2FPU_INSN_DF (cos, 2, N2F_UNSAFE), N2FPU_INSN_DF (tan, 2, N2F_UNSAFE), N2FPU_INSN_DF (atan, 2, N2F_UNSAFE), N2FPU_INSN_DF (exp, 2, N2F_UNSAFE), N2FPU_INSN_DF (log, 2, N2F_UNSAFE), /* Double precision compares. */ N2FPU_CMP_DF (eq), N2FPU_CMP_DF (ne), N2FPU_CMP_DF (lt), N2FPU_CMP_DF (le), N2FPU_CMP_DF (gt), N2FPU_CMP_DF (ge), /* Conversion instructions. */ N2FPU_INSN_DEF_BASE (floatis, 2, 0, floatsisf2, (SF, SI)), N2FPU_INSN_DEF_BASE (floatus, 2, 0, floatunssisf2, (SF, UI)), N2FPU_INSN_DEF_BASE (floatid, 2, 0, floatsidf2, (DF, SI)), N2FPU_INSN_DEF_BASE (floatud, 2, 0, floatunssidf2, (DF, UI)), N2FPU_INSN_DEF_BASE (round, 2, N2F_NO_ERRNO, lroundsfsi2, (SI, SF)), N2FPU_INSN_DEF_BASE (fixsi, 2, 0, fix_truncsfsi2, (SI, SF)), N2FPU_INSN_DEF_BASE (fixsu, 2, 0, fixuns_truncsfsi2, (UI, SF)), N2FPU_INSN_DEF_BASE (fixdi, 2, 0, fix_truncdfsi2, (SI, DF)), N2FPU_INSN_DEF_BASE (fixdu, 2, 0, fixuns_truncdfsi2, (UI, DF)), N2FPU_INSN_DEF_BASE (fextsd, 2, 0, extendsfdf2, (DF, SF)), N2FPU_INSN_DEF_BASE (ftruncds, 2, 0, truncdfsf2, (SF, DF)), /* X, Y access instructions. */ N2FPU_INSN_DEF_BASE (fwrx, 2, N2F_DFREQ, nios2_fwrx, (VOID, DF)), N2FPU_INSN_DEF_BASE (fwry, 2, N2F_DFREQ, nios2_fwry, (VOID, SF)), N2FPU_INSN_DEF_BASE (frdxlo, 1, N2F_DFREQ, nios2_frdxlo, (SF)), N2FPU_INSN_DEF_BASE (frdxhi, 1, N2F_DFREQ, nios2_frdxhi, (SF)), N2FPU_INSN_DEF_BASE (frdy, 1, N2F_DFREQ, nios2_frdy, (SF)) }; /* Some macros for ease of access. */ #define N2FPU(code) nios2_fpu_insn[(int) code] #define N2FPU_ENABLED_P(code) (N2FPU_N(code) >= 0) #define N2FPU_N(code) (*N2FPU(code).optvar) #define N2FPU_NAME(code) (N2FPU(code).name) #define N2FPU_ICODE(code) (N2FPU(code).icode) #define N2FPU_FTCODE(code) (N2FPU(code).ftcode) #define N2FPU_FINITE_P(code) (N2FPU(code).flags & N2F_FINITE) #define N2FPU_UNSAFE_P(code) (N2FPU(code).flags & N2F_UNSAFE) #define N2FPU_NO_ERRNO_P(code) (N2FPU(code).flags & N2F_NO_ERRNO) #define N2FPU_DOUBLE_P(code) (N2FPU(code).flags & N2F_DF) #define N2FPU_DOUBLE_REQUIRED_P(code) (N2FPU(code).flags & N2F_DFREQ) /* Same as above, but for cases where using only the op part is shorter. */ #define N2FPU_OP(op) N2FPU(n2fpu_ ## op) #define N2FPU_OP_NAME(op) N2FPU_NAME(n2fpu_ ## op) #define N2FPU_OP_ENABLED_P(op) N2FPU_ENABLED_P(n2fpu_ ## op) /* Export the FPU insn enabled predicate to nios2.md. */ bool nios2_fpu_insn_enabled (enum n2fpu_code code) { return N2FPU_ENABLED_P (code); } /* Return true if COND comparison for mode MODE is enabled under current settings. */ static bool nios2_fpu_compare_enabled (enum rtx_code cond, machine_mode mode) { if (mode == SFmode) switch (cond) { case EQ: return N2FPU_OP_ENABLED_P (fcmpeqs); case NE: return N2FPU_OP_ENABLED_P (fcmpnes); case GT: return N2FPU_OP_ENABLED_P (fcmpgts); case GE: return N2FPU_OP_ENABLED_P (fcmpges); case LT: return N2FPU_OP_ENABLED_P (fcmplts); case LE: return N2FPU_OP_ENABLED_P (fcmples); default: break; } else if (mode == DFmode) switch (cond) { case EQ: return N2FPU_OP_ENABLED_P (fcmpeqd); case NE: return N2FPU_OP_ENABLED_P (fcmpned); case GT: return N2FPU_OP_ENABLED_P (fcmpgtd); case GE: return N2FPU_OP_ENABLED_P (fcmpged); case LT: return N2FPU_OP_ENABLED_P (fcmpltd); case LE: return N2FPU_OP_ENABLED_P (fcmpled); default: break; } return false; } /* Stack layout and calling conventions. */ #define NIOS2_STACK_ALIGN(LOC) \ (((LOC) + ((PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT) - 1)) \ & ~((PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT) - 1)) /* Return the bytes needed to compute the frame pointer from the current stack pointer. */ static int nios2_compute_frame_layout (void) { unsigned int regno; unsigned int save_mask = 0; int total_size; int var_size; int out_args_size; int save_reg_size; int callee_save_reg_size; if (cfun->machine->initialized) return cfun->machine->total_size; /* Calculate space needed for gp registers. */ save_reg_size = 0; for (regno = 0; regno <= LAST_GP_REG; regno++) if (prologue_saved_reg_p (regno)) { save_mask |= 1 << regno; save_reg_size += 4; } /* If we are saving any callee-save register, then assume push.n/pop.n should be used. Make sure RA is saved, and contiguous registers starting from r16-- are all saved. */ if (TARGET_HAS_CDX && save_reg_size != 0) { if ((save_mask & (1 << RA_REGNO)) == 0) { save_mask |= 1 << RA_REGNO; save_reg_size += 4; } for (regno = 23; regno >= 16; regno--) if ((save_mask & (1 << regno)) != 0) { /* Starting from highest numbered callee-saved register that is used, make sure all regs down to r16 is saved, to maintain contiguous range for push.n/pop.n. */ unsigned int i; for (i = regno - 1; i >= 16; i--) if ((save_mask & (1 << i)) == 0) { save_mask |= 1 << i; save_reg_size += 4; } break; } } callee_save_reg_size = save_reg_size; /* If we call eh_return, we need to save the EH data registers. */ if (crtl->calls_eh_return) { unsigned i; unsigned r; for (i = 0; (r = EH_RETURN_DATA_REGNO (i)) != INVALID_REGNUM; i++) if (!(save_mask & (1 << r))) { save_mask |= 1 << r; save_reg_size += 4; } } cfun->machine->fp_save_offset = 0; if (save_mask & (1 << HARD_FRAME_POINTER_REGNUM)) { int fp_save_offset = 0; for (regno = 0; regno < HARD_FRAME_POINTER_REGNUM; regno++) if (save_mask & (1 << regno)) fp_save_offset += 4; cfun->machine->fp_save_offset = fp_save_offset; } var_size = NIOS2_STACK_ALIGN (get_frame_size ()); out_args_size = NIOS2_STACK_ALIGN (crtl->outgoing_args_size); total_size = var_size + out_args_size; save_reg_size = NIOS2_STACK_ALIGN (save_reg_size); total_size += save_reg_size; total_size += NIOS2_STACK_ALIGN (crtl->args.pretend_args_size); /* Save other computed information. */ cfun->machine->save_mask = save_mask; cfun->machine->total_size = total_size; cfun->machine->var_size = var_size; cfun->machine->args_size = out_args_size; cfun->machine->save_reg_size = save_reg_size; cfun->machine->callee_save_reg_size = callee_save_reg_size; cfun->machine->initialized = reload_completed; cfun->machine->save_regs_offset = out_args_size + var_size; return total_size; } /* Generate save/restore of register REGNO at SP + OFFSET. Used by the prologue/epilogue expand routines. */ static void save_reg (int regno, unsigned offset) { rtx reg = gen_rtx_REG (SImode, regno); rtx addr = plus_constant (Pmode, stack_pointer_rtx, offset, false); rtx_insn *insn = emit_move_insn (gen_frame_mem (Pmode, addr), reg); RTX_FRAME_RELATED_P (insn) = 1; } static void restore_reg (int regno, unsigned offset) { rtx reg = gen_rtx_REG (SImode, regno); rtx addr = plus_constant (Pmode, stack_pointer_rtx, offset, false); rtx_insn *insn = emit_move_insn (reg, gen_frame_mem (Pmode, addr)); /* Tag epilogue unwind note. */ add_reg_note (insn, REG_CFA_RESTORE, reg); RTX_FRAME_RELATED_P (insn) = 1; } /* This routine tests for the base register update SET in load/store multiple RTL insns, used in pop_operation_p and ldstwm_operation_p. */ static bool base_reg_adjustment_p (rtx set, rtx *base_reg, rtx *offset) { if (GET_CODE (set) == SET && REG_P (SET_DEST (set)) && GET_CODE (SET_SRC (set)) == PLUS && REG_P (XEXP (SET_SRC (set), 0)) && rtx_equal_p (SET_DEST (set), XEXP (SET_SRC (set), 0)) && CONST_INT_P (XEXP (SET_SRC (set), 1))) { *base_reg = XEXP (SET_SRC (set), 0); *offset = XEXP (SET_SRC (set), 1); return true; } return false; } /* Does the CFA note work for push/pop prologue/epilogue instructions. */ static void nios2_create_cfa_notes (rtx_insn *insn, bool epilogue_p) { int i = 0; rtx base_reg, offset, elt, pat = PATTERN (insn); if (epilogue_p) { elt = XVECEXP (pat, 0, 0); if (GET_CODE (elt) == RETURN) i++; elt = XVECEXP (pat, 0, i); if (base_reg_adjustment_p (elt, &base_reg, &offset)) { add_reg_note (insn, REG_CFA_ADJUST_CFA, copy_rtx (elt)); i++; } for (; i < XVECLEN (pat, 0); i++) { elt = SET_DEST (XVECEXP (pat, 0, i)); gcc_assert (REG_P (elt)); add_reg_note (insn, REG_CFA_RESTORE, elt); } } else { /* Tag each of the prologue sets. */ for (i = 0; i < XVECLEN (pat, 0); i++) RTX_FRAME_RELATED_P (XVECEXP (pat, 0, i)) = 1; } } /* Temp regno used inside prologue/epilogue. */ #define TEMP_REG_NUM 8 /* Emit conditional trap for checking stack limit. SIZE is the number of additional bytes required. GDB prologue analysis depends on this generating a direct comparison to the SP register, so the adjustment to add SIZE needs to be done on the other operand to the comparison. Use TEMP_REG_NUM as a temporary, if necessary. */ static void nios2_emit_stack_limit_check (int size) { rtx sum = NULL_RTX; if (GET_CODE (stack_limit_rtx) == SYMBOL_REF) { /* This generates a %hiadj/%lo pair with the constant size add handled by the relocations. */ sum = gen_rtx_REG (Pmode, TEMP_REG_NUM); emit_move_insn (sum, plus_constant (Pmode, stack_limit_rtx, size)); } else if (!REG_P (stack_limit_rtx)) sorry ("Unknown form for stack limit expression"); else if (size == 0) sum = stack_limit_rtx; else if (SMALL_INT (size)) { sum = gen_rtx_REG (Pmode, TEMP_REG_NUM); emit_move_insn (sum, plus_constant (Pmode, stack_limit_rtx, size)); } else { sum = gen_rtx_REG (Pmode, TEMP_REG_NUM); emit_move_insn (sum, gen_int_mode (size, Pmode)); emit_insn (gen_add2_insn (sum, stack_limit_rtx)); } emit_insn (gen_ctrapsi4 (gen_rtx_LTU (VOIDmode, stack_pointer_rtx, sum), stack_pointer_rtx, sum, GEN_INT (3))); } static rtx_insn * nios2_emit_add_constant (rtx reg, HOST_WIDE_INT immed) { rtx_insn *insn; if (SMALL_INT (immed)) insn = emit_insn (gen_add2_insn (reg, gen_int_mode (immed, Pmode))); else { rtx tmp = gen_rtx_REG (Pmode, TEMP_REG_NUM); emit_move_insn (tmp, gen_int_mode (immed, Pmode)); insn = emit_insn (gen_add2_insn (reg, tmp)); } return insn; } static rtx_insn * nios2_adjust_stack (int sp_adjust, bool epilogue_p) { enum reg_note note_kind = REG_NOTE_MAX; rtx_insn *insn = NULL; if (sp_adjust) { if (SMALL_INT (sp_adjust)) insn = emit_insn (gen_add2_insn (stack_pointer_rtx, gen_int_mode (sp_adjust, Pmode))); else { rtx tmp = gen_rtx_REG (Pmode, TEMP_REG_NUM); emit_move_insn (tmp, gen_int_mode (sp_adjust, Pmode)); insn = emit_insn (gen_add2_insn (stack_pointer_rtx, tmp)); /* Attach a note indicating what happened. */ if (!epilogue_p) note_kind = REG_FRAME_RELATED_EXPR; } if (epilogue_p) note_kind = REG_CFA_ADJUST_CFA; if (note_kind != REG_NOTE_MAX) { rtx cfa_adj = gen_rtx_SET (stack_pointer_rtx, plus_constant (Pmode, stack_pointer_rtx, sp_adjust)); add_reg_note (insn, note_kind, cfa_adj); } RTX_FRAME_RELATED_P (insn) = 1; } return insn; } void nios2_expand_prologue (void) { unsigned int regno; int total_frame_size, save_offset; int sp_offset; /* offset from base_reg to final stack value. */ int save_regs_base; /* offset from base_reg to register save area. */ rtx_insn *insn; total_frame_size = nios2_compute_frame_layout (); if (flag_stack_usage_info) current_function_static_stack_size = total_frame_size; /* When R2 CDX push.n/stwm is available, arrange for stack frame to be built using them. */ if (TARGET_HAS_CDX && (cfun->machine->save_reg_size != 0 || cfun->machine->uses_anonymous_args)) { unsigned int regmask = cfun->machine->save_mask; unsigned int callee_save_regs = regmask & 0xffff0000; unsigned int caller_save_regs = regmask & 0x0000ffff; int push_immed = 0; int pretend_args_size = NIOS2_STACK_ALIGN (crtl->args.pretend_args_size); rtx stack_mem = gen_frame_mem (SImode, plus_constant (Pmode, stack_pointer_rtx, -4)); /* Check that there is room for the entire stack frame before doing any SP adjustments or pushes. */ if (crtl->limit_stack) nios2_emit_stack_limit_check (total_frame_size); if (pretend_args_size) { if (cfun->machine->uses_anonymous_args) { /* Emit a stwm to push copy of argument registers onto the stack for va_arg processing. */ unsigned int r, mask = 0, n = pretend_args_size / 4; for (r = LAST_ARG_REGNO - n + 1; r <= LAST_ARG_REGNO; r++) mask |= (1 << r); insn = emit_insn (nios2_ldst_parallel (false, false, false, stack_mem, -pretend_args_size, mask, false)); /* Tag first SP adjustment as frame-related. */ RTX_FRAME_RELATED_P (XVECEXP (PATTERN (insn), 0, 0)) = 1; RTX_FRAME_RELATED_P (insn) = 1; } else nios2_adjust_stack (-pretend_args_size, false); } if (callee_save_regs) { /* Emit a push.n to save registers and optionally allocate push_immed extra bytes on the stack. */ int sp_adjust; if (caller_save_regs) /* Can't allocate extra stack space yet. */ push_immed = 0; else if (cfun->machine->save_regs_offset <= 60) /* Stack adjustment fits entirely in the push.n. */ push_immed = cfun->machine->save_regs_offset; else if (frame_pointer_needed && cfun->machine->fp_save_offset == 0) /* Deferring the entire stack adjustment until later allows us to use a mov.n instead of a 32-bit addi instruction to set the frame pointer. */ push_immed = 0; else /* Splitting the stack adjustment between the push.n and an explicit adjustment makes it more likely that we can use spdeci.n for the explicit part. */ push_immed = 60; sp_adjust = -(cfun->machine->callee_save_reg_size + push_immed); insn = emit_insn (nios2_ldst_parallel (false, false, false, stack_mem, sp_adjust, callee_save_regs, false)); nios2_create_cfa_notes (insn, false); RTX_FRAME_RELATED_P (insn) = 1; } if (caller_save_regs) { /* Emit a stwm to save the EH data regs, r4-r7. */ int caller_save_size = (cfun->machine->save_reg_size - cfun->machine->callee_save_reg_size); gcc_assert ((caller_save_regs & ~0xf0) == 0); insn = emit_insn (nios2_ldst_parallel (false, false, false, stack_mem, -caller_save_size, caller_save_regs, false)); nios2_create_cfa_notes (insn, false); RTX_FRAME_RELATED_P (insn) = 1; } save_regs_base = push_immed; sp_offset = -(cfun->machine->save_regs_offset - push_immed); } /* The non-CDX cases decrement the stack pointer, to prepare for individual register saves to the stack. */ else if (!SMALL_INT (total_frame_size)) { /* We need an intermediary point, this will point at the spill block. */ nios2_adjust_stack (cfun->machine->save_regs_offset - total_frame_size, false); save_regs_base = 0; sp_offset = -cfun->machine->save_regs_offset; if (crtl->limit_stack) nios2_emit_stack_limit_check (cfun->machine->save_regs_offset); } else if (total_frame_size) { nios2_adjust_stack (-total_frame_size, false); save_regs_base = cfun->machine->save_regs_offset; sp_offset = 0; if (crtl->limit_stack) nios2_emit_stack_limit_check (0); } else save_regs_base = sp_offset = 0; /* Save the registers individually in the non-CDX case. */ if (!TARGET_HAS_CDX) { save_offset = save_regs_base + cfun->machine->save_reg_size; for (regno = LAST_GP_REG; regno > 0; regno--) if (cfun->machine->save_mask & (1 << regno)) { save_offset -= 4; save_reg (regno, save_offset); } } /* Set the hard frame pointer. */ if (frame_pointer_needed) { int fp_save_offset = save_regs_base + cfun->machine->fp_save_offset; insn = (fp_save_offset == 0 ? emit_move_insn (hard_frame_pointer_rtx, stack_pointer_rtx) : emit_insn (gen_add3_insn (hard_frame_pointer_rtx, stack_pointer_rtx, gen_int_mode (fp_save_offset, Pmode)))); RTX_FRAME_RELATED_P (insn) = 1; } /* Allocate sp_offset more bytes in the stack frame. */ nios2_adjust_stack (sp_offset, false); /* Load the PIC register if needed. */ if (crtl->uses_pic_offset_table) nios2_load_pic_register (); /* If we are profiling, make sure no instructions are scheduled before the call to mcount. */ if (crtl->profile) emit_insn (gen_blockage ()); } void nios2_expand_epilogue (bool sibcall_p) { rtx_insn *insn; rtx cfa_adj; int total_frame_size; int sp_adjust, save_offset; unsigned int regno; if (!sibcall_p && nios2_can_use_return_insn ()) { emit_jump_insn (gen_return ()); return; } emit_insn (gen_blockage ()); total_frame_size = nios2_compute_frame_layout (); if (frame_pointer_needed) { /* Recover the stack pointer. */ insn = (cfun->machine->fp_save_offset == 0 ? emit_move_insn (stack_pointer_rtx, hard_frame_pointer_rtx) : emit_insn (gen_add3_insn (stack_pointer_rtx, hard_frame_pointer_rtx, gen_int_mode (-cfun->machine->fp_save_offset, Pmode)))); cfa_adj = plus_constant (Pmode, stack_pointer_rtx, (total_frame_size - cfun->machine->save_regs_offset)); add_reg_note (insn, REG_CFA_DEF_CFA, cfa_adj); RTX_FRAME_RELATED_P (insn) = 1; save_offset = 0; sp_adjust = total_frame_size - cfun->machine->save_regs_offset; } else if (!SMALL_INT (total_frame_size)) { nios2_adjust_stack (cfun->machine->save_regs_offset, true); save_offset = 0; sp_adjust = total_frame_size - cfun->machine->save_regs_offset; } else { save_offset = cfun->machine->save_regs_offset; sp_adjust = total_frame_size; } if (!TARGET_HAS_CDX) { /* Generate individual register restores. */ save_offset += cfun->machine->save_reg_size; for (regno = LAST_GP_REG; regno > 0; regno--) if (cfun->machine->save_mask & (1 << regno)) { save_offset -= 4; restore_reg (regno, save_offset); } nios2_adjust_stack (sp_adjust, true); } else if (cfun->machine->save_reg_size == 0) { /* Nothing to restore, just recover the stack position. */ nios2_adjust_stack (sp_adjust, true); } else { /* Emit CDX pop.n/ldwm to restore registers and optionally return. */ unsigned int regmask = cfun->machine->save_mask; unsigned int callee_save_regs = regmask & 0xffff0000; unsigned int caller_save_regs = regmask & 0x0000ffff; int callee_save_size = cfun->machine->callee_save_reg_size; int caller_save_size = cfun->machine->save_reg_size - callee_save_size; int pretend_args_size = NIOS2_STACK_ALIGN (crtl->args.pretend_args_size); bool ret_p = (!pretend_args_size && !crtl->calls_eh_return && !sibcall_p); if (!ret_p || caller_save_size > 0) sp_adjust = save_offset; else sp_adjust = (save_offset > 60 ? save_offset - 60 : 0); save_offset -= sp_adjust; nios2_adjust_stack (sp_adjust, true); if (caller_save_regs) { /* Emit a ldwm to restore EH data regs. */ rtx stack_mem = gen_frame_mem (SImode, stack_pointer_rtx); insn = emit_insn (nios2_ldst_parallel (true, true, true, stack_mem, caller_save_size, caller_save_regs, false)); RTX_FRAME_RELATED_P (insn) = 1; nios2_create_cfa_notes (insn, true); } if (callee_save_regs) { int sp_adjust = save_offset + callee_save_size; rtx stack_mem; if (ret_p) { /* Emit a pop.n to restore regs and return. */ stack_mem = gen_frame_mem (SImode, gen_rtx_PLUS (Pmode, stack_pointer_rtx, gen_int_mode (sp_adjust - 4, Pmode))); insn = emit_jump_insn (nios2_ldst_parallel (true, false, false, stack_mem, sp_adjust, callee_save_regs, ret_p)); RTX_FRAME_RELATED_P (insn) = 1; /* No need to attach CFA notes since we cannot step over a return. */ return; } else { /* If no return, we have to use the ldwm form. */ stack_mem = gen_frame_mem (SImode, stack_pointer_rtx); insn = emit_insn (nios2_ldst_parallel (true, true, true, stack_mem, sp_adjust, callee_save_regs, ret_p)); RTX_FRAME_RELATED_P (insn) = 1; nios2_create_cfa_notes (insn, true); } } if (pretend_args_size) nios2_adjust_stack (pretend_args_size, true); } /* Add in the __builtin_eh_return stack adjustment. */ if (crtl->calls_eh_return) emit_insn (gen_add2_insn (stack_pointer_rtx, EH_RETURN_STACKADJ_RTX)); if (!sibcall_p) emit_jump_insn (gen_simple_return ()); } bool nios2_expand_return (void) { /* If CDX is available, generate a pop.n instruction to do both the stack pop and return. */ if (TARGET_HAS_CDX) { int total_frame_size = nios2_compute_frame_layout (); int sp_adjust = (cfun->machine->save_regs_offset + cfun->machine->callee_save_reg_size); gcc_assert (sp_adjust == total_frame_size); if (sp_adjust != 0) { rtx mem = gen_frame_mem (SImode, plus_constant (Pmode, stack_pointer_rtx, sp_adjust - 4, false)); rtx_insn *insn = emit_jump_insn (nios2_ldst_parallel (true, false, false, mem, sp_adjust, cfun->machine->save_mask, true)); RTX_FRAME_RELATED_P (insn) = 1; /* No need to create CFA notes since we can't step over a return. */ return true; } } return false; } /* Implement RETURN_ADDR_RTX. Note, we do not support moving back to a previous frame. */ rtx nios2_get_return_address (int count) { if (count != 0) return const0_rtx; return get_hard_reg_initial_val (Pmode, RA_REGNO); } /* Emit code to change the current function's return address to ADDRESS. SCRATCH is available as a scratch register, if needed. ADDRESS and SCRATCH are both word-mode GPRs. */ void nios2_set_return_address (rtx address, rtx scratch) { nios2_compute_frame_layout (); if (cfun->machine->save_mask & (1 << RA_REGNO)) { unsigned offset = cfun->machine->save_reg_size - 4; rtx base; if (frame_pointer_needed) base = hard_frame_pointer_rtx; else { base = stack_pointer_rtx; offset += cfun->machine->save_regs_offset; if (!SMALL_INT (offset)) { emit_move_insn (scratch, gen_int_mode (offset, Pmode)); emit_insn (gen_add2_insn (scratch, base)); base = scratch; offset = 0; } } if (offset) base = plus_constant (Pmode, base, offset); emit_move_insn (gen_rtx_MEM (Pmode, base), address); } else emit_move_insn (gen_rtx_REG (Pmode, RA_REGNO), address); } /* Implement FUNCTION_PROFILER macro. */ void nios2_function_profiler (FILE *file, int labelno ATTRIBUTE_UNUSED) { fprintf (file, "\tmov\tr8, ra\n"); if (flag_pic == 1) { fprintf (file, "\tnextpc\tr2\n"); fprintf (file, "\t1: movhi\tr3, %%hiadj(_gp_got - 1b)\n"); fprintf (file, "\taddi\tr3, r3, %%lo(_gp_got - 1b)\n"); fprintf (file, "\tadd\tr2, r2, r3\n"); fprintf (file, "\tldw\tr2, %%call(_mcount)(r2)\n"); fprintf (file, "\tcallr\tr2\n"); } else if (flag_pic == 2) { fprintf (file, "\tnextpc\tr2\n"); fprintf (file, "\t1: movhi\tr3, %%hiadj(_gp_got - 1b)\n"); fprintf (file, "\taddi\tr3, r3, %%lo(_gp_got - 1b)\n"); fprintf (file, "\tadd\tr2, r2, r3\n"); fprintf (file, "\tmovhi\tr3, %%call_hiadj(_mcount)\n"); fprintf (file, "\taddi\tr3, r3, %%call_lo(_mcount)\n"); fprintf (file, "\tadd\tr3, r2, r3\n"); fprintf (file, "\tldw\tr2, 0(r3)\n"); fprintf (file, "\tcallr\tr2\n"); } else fprintf (file, "\tcall\t_mcount\n"); fprintf (file, "\tmov\tra, r8\n"); } /* Dump stack layout. */ static void nios2_dump_frame_layout (FILE *file) { fprintf (file, "\t%s Current Frame Info\n", ASM_COMMENT_START); fprintf (file, "\t%s total_size = %d\n", ASM_COMMENT_START, cfun->machine->total_size); fprintf (file, "\t%s var_size = %d\n", ASM_COMMENT_START, cfun->machine->var_size); fprintf (file, "\t%s args_size = %d\n", ASM_COMMENT_START, cfun->machine->args_size); fprintf (file, "\t%s save_reg_size = %d\n", ASM_COMMENT_START, cfun->machine->save_reg_size); fprintf (file, "\t%s initialized = %d\n", ASM_COMMENT_START, cfun->machine->initialized); fprintf (file, "\t%s save_regs_offset = %d\n", ASM_COMMENT_START, cfun->machine->save_regs_offset); fprintf (file, "\t%s is_leaf = %d\n", ASM_COMMENT_START, crtl->is_leaf); fprintf (file, "\t%s frame_pointer_needed = %d\n", ASM_COMMENT_START, frame_pointer_needed); fprintf (file, "\t%s pretend_args_size = %d\n", ASM_COMMENT_START, crtl->args.pretend_args_size); } /* Return true if REGNO should be saved in the prologue. */ static bool prologue_saved_reg_p (unsigned regno) { gcc_assert (GP_REG_P (regno)); if (df_regs_ever_live_p (regno) && !call_used_regs[regno]) return true; if (regno == HARD_FRAME_POINTER_REGNUM && frame_pointer_needed) return true; if (regno == PIC_OFFSET_TABLE_REGNUM && crtl->uses_pic_offset_table) return true; if (regno == RA_REGNO && df_regs_ever_live_p (RA_REGNO)) return true; return false; } /* Implement TARGET_CAN_ELIMINATE. */ static bool nios2_can_eliminate (const int from ATTRIBUTE_UNUSED, const int to) { if (to == STACK_POINTER_REGNUM) return !frame_pointer_needed; return true; } /* Implement INITIAL_ELIMINATION_OFFSET macro. */ int nios2_initial_elimination_offset (int from, int to) { int offset; nios2_compute_frame_layout (); /* Set OFFSET to the offset from the stack pointer. */ switch (from) { case FRAME_POINTER_REGNUM: offset = cfun->machine->args_size; break; case ARG_POINTER_REGNUM: offset = cfun->machine->total_size; offset -= crtl->args.pretend_args_size; break; default: gcc_unreachable (); } /* If we are asked for the frame pointer offset, then adjust OFFSET by the offset from the frame pointer to the stack pointer. */ if (to == HARD_FRAME_POINTER_REGNUM) offset -= (cfun->machine->save_regs_offset + cfun->machine->fp_save_offset); return offset; } /* Return nonzero if this function is known to have a null epilogue. This allows the optimizer to omit jumps to jumps if no stack was created. */ int nios2_can_use_return_insn (void) { int total_frame_size; if (!reload_completed || crtl->profile) return 0; total_frame_size = nios2_compute_frame_layout (); /* If CDX is available, check if we can return using a single pop.n instruction. */ if (TARGET_HAS_CDX && !frame_pointer_needed && cfun->machine->save_regs_offset <= 60 && (cfun->machine->save_mask & 0x80000000) != 0 && (cfun->machine->save_mask & 0xffff) == 0 && crtl->args.pretend_args_size == 0) return true; return total_frame_size == 0; } /* Check and signal some warnings/errors on FPU insn options. */ static void nios2_custom_check_insns (void) { unsigned int i, j; bool errors = false; for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) if (N2FPU_ENABLED_P (i) && N2FPU_DOUBLE_P (i)) { for (j = 0; j < ARRAY_SIZE (nios2_fpu_insn); j++) if (N2FPU_DOUBLE_REQUIRED_P (j) && ! N2FPU_ENABLED_P (j)) { error ("switch %<-mcustom-%s%> is required for double " "precision floating point", N2FPU_NAME (j)); errors = true; } break; } /* Warn if the user has certain exotic operations that won't get used without -funsafe-math-optimizations. See expand_builtin () in builtins.c. */ if (!flag_unsafe_math_optimizations) for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) if (N2FPU_ENABLED_P (i) && N2FPU_UNSAFE_P (i)) warning (0, "switch %<-mcustom-%s%> has no effect unless " "-funsafe-math-optimizations is specified", N2FPU_NAME (i)); /* Warn if the user is trying to use -mcustom-fmins et. al, that won't get used without -ffinite-math-only. See fold_builtin_fmin_fmax () in builtins.c. */ if (!flag_finite_math_only) for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) if (N2FPU_ENABLED_P (i) && N2FPU_FINITE_P (i)) warning (0, "switch %<-mcustom-%s%> has no effect unless " "-ffinite-math-only is specified", N2FPU_NAME (i)); /* Warn if the user is trying to use a custom rounding instruction that won't get used without -fno-math-errno. See expand_builtin_int_roundingfn_2 () in builtins.c. */ if (flag_errno_math) for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) if (N2FPU_ENABLED_P (i) && N2FPU_NO_ERRNO_P (i)) warning (0, "switch %<-mcustom-%s%> has no effect unless " "-fno-math-errno is specified", N2FPU_NAME (i)); if (errors || custom_code_conflict) fatal_error (input_location, "conflicting use of -mcustom switches, target attributes, " "and/or __builtin_custom_ functions"); } static void nios2_set_fpu_custom_code (enum n2fpu_code code, int n, bool override_p) { if (override_p || N2FPU_N (code) == -1) N2FPU_N (code) = n; nios2_register_custom_code (n, CCS_FPU, (int) code); } /* Type to represent a standard FPU config. */ struct nios2_fpu_config { const char *name; bool set_sp_constants; int code[n2fpu_code_num]; }; #define NIOS2_FPU_CONFIG_NUM 3 static struct nios2_fpu_config custom_fpu_config[NIOS2_FPU_CONFIG_NUM]; static void nios2_init_fpu_configs (void) { struct nios2_fpu_config* cfg; int i = 0; #define NEXT_FPU_CONFIG \ do { \ cfg = &custom_fpu_config[i++]; \ memset (cfg, -1, sizeof (struct nios2_fpu_config));\ } while (0) NEXT_FPU_CONFIG; cfg->name = "60-1"; cfg->set_sp_constants = true; cfg->code[n2fpu_fmuls] = 252; cfg->code[n2fpu_fadds] = 253; cfg->code[n2fpu_fsubs] = 254; NEXT_FPU_CONFIG; cfg->name = "60-2"; cfg->set_sp_constants = true; cfg->code[n2fpu_fmuls] = 252; cfg->code[n2fpu_fadds] = 253; cfg->code[n2fpu_fsubs] = 254; cfg->code[n2fpu_fdivs] = 255; NEXT_FPU_CONFIG; cfg->name = "72-3"; cfg->set_sp_constants = true; cfg->code[n2fpu_floatus] = 243; cfg->code[n2fpu_fixsi] = 244; cfg->code[n2fpu_floatis] = 245; cfg->code[n2fpu_fcmpgts] = 246; cfg->code[n2fpu_fcmples] = 249; cfg->code[n2fpu_fcmpeqs] = 250; cfg->code[n2fpu_fcmpnes] = 251; cfg->code[n2fpu_fmuls] = 252; cfg->code[n2fpu_fadds] = 253; cfg->code[n2fpu_fsubs] = 254; cfg->code[n2fpu_fdivs] = 255; #undef NEXT_FPU_CONFIG gcc_assert (i == NIOS2_FPU_CONFIG_NUM); } static struct nios2_fpu_config * nios2_match_custom_fpu_cfg (const char *cfgname, const char *endp) { int i; for (i = 0; i < NIOS2_FPU_CONFIG_NUM; i++) { bool match = !(endp != NULL ? strncmp (custom_fpu_config[i].name, cfgname, endp - cfgname) : strcmp (custom_fpu_config[i].name, cfgname)); if (match) return &custom_fpu_config[i]; } return NULL; } /* Use CFGNAME to lookup FPU config, ENDP if not NULL marks end of string. OVERRIDE is true if loaded config codes should overwrite current state. */ static void nios2_handle_custom_fpu_cfg (const char *cfgname, const char *endp, bool override) { struct nios2_fpu_config *cfg = nios2_match_custom_fpu_cfg (cfgname, endp); if (cfg) { unsigned int i; for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) if (cfg->code[i] >= 0) nios2_set_fpu_custom_code ((enum n2fpu_code) i, cfg->code[i], override); if (cfg->set_sp_constants) flag_single_precision_constant = 1; } else warning (0, "ignoring unrecognized switch %<-mcustom-fpu-cfg%> " "value %<%s%>", cfgname); /* Guard against errors in the standard configurations. */ nios2_custom_check_insns (); } /* Check individual FPU insn options, and register custom code. */ static void nios2_handle_custom_fpu_insn_option (int fpu_insn_index) { int param = N2FPU_N (fpu_insn_index); if (0 <= param && param <= 255) nios2_register_custom_code (param, CCS_FPU, fpu_insn_index); /* Valid values are 0-255, but also allow -1 so that the -mno-custom- switches work. */ else if (param != -1) error ("switch %<-mcustom-%s%> value %d must be between 0 and 255", N2FPU_NAME (fpu_insn_index), param); } /* Allocate a chunk of memory for per-function machine-dependent data. */ static struct machine_function * nios2_init_machine_status (void) { return ggc_cleared_alloc (); } /* Implement TARGET_OPTION_OVERRIDE. */ static void nios2_option_override (void) { unsigned int i; #ifdef SUBTARGET_OVERRIDE_OPTIONS SUBTARGET_OVERRIDE_OPTIONS; #endif /* Check for unsupported options. */ if (flag_pic && !TARGET_LINUX_ABI) sorry ("position-independent code requires the Linux ABI"); if (flag_pic && stack_limit_rtx && GET_CODE (stack_limit_rtx) == SYMBOL_REF) sorry ("PIC support for -fstack-limit-symbol"); /* Function to allocate machine-dependent function status. */ init_machine_status = &nios2_init_machine_status; nios2_section_threshold = (global_options_set.x_g_switch_value ? g_switch_value : NIOS2_DEFAULT_GVALUE); if (nios2_gpopt_option == gpopt_unspecified) { /* Default to -mgpopt unless -fpic or -fPIC. */ if (flag_pic) nios2_gpopt_option = gpopt_none; else nios2_gpopt_option = gpopt_local; } /* If we don't have mul, we don't have mulx either! */ if (!TARGET_HAS_MUL && TARGET_HAS_MULX) target_flags &= ~MASK_HAS_MULX; /* Optional BMX and CDX instructions only make sense for R2. */ if (!TARGET_ARCH_R2) { if (TARGET_HAS_BMX) error ("BMX instructions are only supported with R2 architecture"); if (TARGET_HAS_CDX) error ("CDX instructions are only supported with R2 architecture"); } /* R2 is little-endian only. */ if (TARGET_ARCH_R2 && TARGET_BIG_ENDIAN) error ("R2 architecture is little-endian only"); /* Initialize default FPU configurations. */ nios2_init_fpu_configs (); /* Set up default handling for floating point custom instructions. Putting things in this order means that the -mcustom-fpu-cfg= switch will always be overridden by individual -mcustom-fadds= switches, regardless of the order in which they were specified on the command line. This behavior of prioritization of individual -mcustom-= options before the -mcustom-fpu-cfg= switch is maintained for compatibility. */ if (nios2_custom_fpu_cfg_string && *nios2_custom_fpu_cfg_string) nios2_handle_custom_fpu_cfg (nios2_custom_fpu_cfg_string, NULL, false); /* Handle options for individual FPU insns. */ for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) nios2_handle_custom_fpu_insn_option (i); nios2_custom_check_insns (); /* Save the initial options in case the user does function specific options. */ target_option_default_node = target_option_current_node = build_target_option_node (&global_options); } /* Return true if CST is a constant within range of movi/movui/movhi. */ static bool nios2_simple_const_p (const_rtx cst) { HOST_WIDE_INT val = INTVAL (cst); return SMALL_INT (val) || SMALL_INT_UNSIGNED (val) || UPPER16_INT (val); } /* Compute a (partial) cost for rtx X. Return true if the complete cost has been computed, and false if subexpressions should be scanned. In either case, *TOTAL contains the cost result. */ static bool nios2_rtx_costs (rtx x, machine_mode mode ATTRIBUTE_UNUSED, int outer_code ATTRIBUTE_UNUSED, int opno ATTRIBUTE_UNUSED, int *total, bool speed ATTRIBUTE_UNUSED) { int code = GET_CODE (x); switch (code) { case CONST_INT: if (INTVAL (x) == 0) { *total = COSTS_N_INSNS (0); return true; } else if (nios2_simple_const_p (x)) { *total = COSTS_N_INSNS (2); return true; } else { *total = COSTS_N_INSNS (4); return true; } case LABEL_REF: case SYMBOL_REF: case CONST: case CONST_DOUBLE: { *total = COSTS_N_INSNS (4); return true; } case AND: { /* Recognize 'nor' insn pattern. */ if (GET_CODE (XEXP (x, 0)) == NOT && GET_CODE (XEXP (x, 1)) == NOT) { *total = COSTS_N_INSNS (1); return true; } return false; } case MULT: { *total = COSTS_N_INSNS (1); return false; } case SIGN_EXTEND: { *total = COSTS_N_INSNS (3); return false; } case ZERO_EXTEND: { *total = COSTS_N_INSNS (1); return false; } case ZERO_EXTRACT: if (TARGET_HAS_BMX) { *total = COSTS_N_INSNS (1); return true; } return false; default: return false; } } /* Implement TARGET_PREFERRED_RELOAD_CLASS. */ static reg_class_t nios2_preferred_reload_class (rtx x ATTRIBUTE_UNUSED, reg_class_t regclass) { return regclass == NO_REGS ? GENERAL_REGS : regclass; } /* Emit a call to __tls_get_addr. TI is the argument to this function. RET is an RTX for the return value location. The entire insn sequence is returned. */ static GTY(()) rtx nios2_tls_symbol; static rtx nios2_call_tls_get_addr (rtx ti) { rtx arg = gen_rtx_REG (Pmode, FIRST_ARG_REGNO); rtx ret = gen_rtx_REG (Pmode, FIRST_RETVAL_REGNO); rtx fn; rtx_insn *insn; if (!nios2_tls_symbol) nios2_tls_symbol = init_one_libfunc ("__tls_get_addr"); emit_move_insn (arg, ti); fn = gen_rtx_MEM (QImode, nios2_tls_symbol); insn = emit_call_insn (gen_call_value (ret, fn, const0_rtx)); RTL_CONST_CALL_P (insn) = 1; use_reg (&CALL_INSN_FUNCTION_USAGE (insn), ret); use_reg (&CALL_INSN_FUNCTION_USAGE (insn), arg); return ret; } /* Return true for large offsets requiring hiadj/lo relocation pairs. */ static bool nios2_large_offset_p (int unspec) { gcc_assert (nios2_unspec_reloc_name (unspec) != NULL); if (flag_pic == 2 /* FIXME: TLS GOT offset relocations will eventually also get this treatment, after binutils support for those are also completed. */ && (unspec == UNSPEC_PIC_SYM || unspec == UNSPEC_PIC_CALL_SYM)) return true; /* 'gotoff' offsets are always hiadj/lo. */ if (unspec == UNSPEC_PIC_GOTOFF_SYM) return true; return false; } /* Return true for conforming unspec relocations. Also used in constraints.md and predicates.md. */ bool nios2_unspec_reloc_p (rtx op) { return (GET_CODE (op) == CONST && GET_CODE (XEXP (op, 0)) == UNSPEC && ! nios2_large_offset_p (XINT (XEXP (op, 0), 1))); } static bool nios2_large_unspec_reloc_p (rtx op) { return (GET_CODE (op) == CONST && GET_CODE (XEXP (op, 0)) == UNSPEC && nios2_large_offset_p (XINT (XEXP (op, 0), 1))); } /* Helper to generate unspec constant. */ static rtx nios2_unspec_offset (rtx loc, int unspec) { return gen_rtx_CONST (Pmode, gen_rtx_UNSPEC (Pmode, gen_rtvec (1, loc), unspec)); } /* Generate GOT pointer based address with large offset. */ static rtx nios2_large_got_address (rtx offset, rtx tmp) { if (!tmp) tmp = gen_reg_rtx (Pmode); emit_move_insn (tmp, offset); return gen_rtx_PLUS (Pmode, tmp, pic_offset_table_rtx); } /* Generate a GOT pointer based address. */ static rtx nios2_got_address (rtx loc, int unspec) { rtx offset = nios2_unspec_offset (loc, unspec); crtl->uses_pic_offset_table = 1; if (nios2_large_offset_p (unspec)) return force_reg (Pmode, nios2_large_got_address (offset, NULL_RTX)); return gen_rtx_PLUS (Pmode, pic_offset_table_rtx, offset); } /* Generate the code to access LOC, a thread local SYMBOL_REF. The return value will be a valid address and move_operand (either a REG or a LO_SUM). */ static rtx nios2_legitimize_tls_address (rtx loc) { rtx tmp, mem, tp; enum tls_model model = SYMBOL_REF_TLS_MODEL (loc); switch (model) { case TLS_MODEL_GLOBAL_DYNAMIC: tmp = gen_reg_rtx (Pmode); emit_move_insn (tmp, nios2_got_address (loc, UNSPEC_ADD_TLS_GD)); return nios2_call_tls_get_addr (tmp); case TLS_MODEL_LOCAL_DYNAMIC: tmp = gen_reg_rtx (Pmode); emit_move_insn (tmp, nios2_got_address (loc, UNSPEC_ADD_TLS_LDM)); return gen_rtx_PLUS (Pmode, nios2_call_tls_get_addr (tmp), nios2_unspec_offset (loc, UNSPEC_ADD_TLS_LDO)); case TLS_MODEL_INITIAL_EXEC: tmp = gen_reg_rtx (Pmode); mem = gen_const_mem (Pmode, nios2_got_address (loc, UNSPEC_LOAD_TLS_IE)); emit_move_insn (tmp, mem); tp = gen_rtx_REG (Pmode, TP_REGNO); return gen_rtx_PLUS (Pmode, tp, tmp); case TLS_MODEL_LOCAL_EXEC: tp = gen_rtx_REG (Pmode, TP_REGNO); return gen_rtx_PLUS (Pmode, tp, nios2_unspec_offset (loc, UNSPEC_ADD_TLS_LE)); default: gcc_unreachable (); } } /* Divide Support If -O3 is used, we want to output a table lookup for divides between small numbers (both num and den >= 0 and < 0x10). The overhead of this method in the worst case is 40 bytes in the text section (10 insns) and 256 bytes in the data section. Additional divides do not incur additional penalties in the data section. Code speed is improved for small divides by about 5x when using this method in the worse case (~9 cycles vs ~45). And in the worst case divides not within the table are penalized by about 10% (~5 cycles vs ~45). However in the typical case the penalty is not as bad because doing the long divide in only 45 cycles is quite optimistic. ??? would be nice to have some benchmarks other than Dhrystone to back this up. This bit of expansion is to create this instruction sequence as rtl. or $8, $4, $5 slli $9, $4, 4 cmpgeui $3, $8, 16 beq $3, $0, .L3 or $10, $9, $5 add $12, $11, divide_table ldbu $2, 0($12) br .L1 .L3: call slow_div .L1: # continue here with result in $2 ??? Ideally I would like the libcall block to contain all of this code, but I don't know how to do that. What it means is that if the divide can be eliminated, it may not completely disappear. ??? The __divsi3_table label should ideally be moved out of this block and into a global. If it is placed into the sdata section we can save even more cycles by doing things gp relative. */ void nios2_emit_expensive_div (rtx *operands, machine_mode mode) { rtx or_result, shift_left_result; rtx lookup_value; rtx_code_label *lab1, *lab3; rtx_insn *insns; rtx libfunc; rtx final_result; rtx_insn *tmp; rtx table; /* It may look a little generic, but only SImode is supported for now. */ gcc_assert (mode == SImode); libfunc = optab_libfunc (sdiv_optab, SImode); lab1 = gen_label_rtx (); lab3 = gen_label_rtx (); or_result = expand_simple_binop (SImode, IOR, operands[1], operands[2], 0, 0, OPTAB_LIB_WIDEN); emit_cmp_and_jump_insns (or_result, GEN_INT (15), GTU, 0, GET_MODE (or_result), 0, lab3); JUMP_LABEL (get_last_insn ()) = lab3; shift_left_result = expand_simple_binop (SImode, ASHIFT, operands[1], GEN_INT (4), 0, 0, OPTAB_LIB_WIDEN); lookup_value = expand_simple_binop (SImode, IOR, shift_left_result, operands[2], 0, 0, OPTAB_LIB_WIDEN); table = gen_rtx_PLUS (SImode, lookup_value, gen_rtx_SYMBOL_REF (SImode, "__divsi3_table")); convert_move (operands[0], gen_rtx_MEM (QImode, table), 1); tmp = emit_jump_insn (gen_jump (lab1)); JUMP_LABEL (tmp) = lab1; emit_barrier (); emit_label (lab3); LABEL_NUSES (lab3) = 1; start_sequence (); final_result = emit_library_call_value (libfunc, NULL_RTX, LCT_CONST, SImode, 2, operands[1], SImode, operands[2], SImode); insns = get_insns (); end_sequence (); emit_libcall_block (insns, operands[0], final_result, gen_rtx_DIV (SImode, operands[1], operands[2])); emit_label (lab1); LABEL_NUSES (lab1) = 1; } /* Branches and compares. */ /* Return in *ALT_CODE and *ALT_OP, an alternate equivalent constant comparison, e.g. >= 1 into > 0. */ static void nios2_alternate_compare_const (enum rtx_code code, rtx op, enum rtx_code *alt_code, rtx *alt_op, machine_mode mode) { HOST_WIDE_INT opval = INTVAL (op); enum rtx_code scode = signed_condition (code); bool dec_p = (scode == LT || scode == GE); if (code == EQ || code == NE) { *alt_code = code; *alt_op = op; return; } *alt_op = (dec_p ? gen_int_mode (opval - 1, mode) : gen_int_mode (opval + 1, mode)); /* The required conversion between [>,>=] and [<,<=] is captured by a reverse + swap of condition codes. */ *alt_code = reverse_condition (swap_condition (code)); { /* Test if the incremented/decremented value crosses the over/underflow boundary. Supposedly, such boundary cases should already be transformed into always-true/false or EQ conditions, so use an assertion here. */ unsigned HOST_WIDE_INT alt_opval = INTVAL (*alt_op); if (code == scode) alt_opval ^= (1 << (GET_MODE_BITSIZE (mode) - 1)); alt_opval &= GET_MODE_MASK (mode); gcc_assert (dec_p ? alt_opval != GET_MODE_MASK (mode) : alt_opval != 0); } } /* Return true if the constant comparison is supported by nios2. */ static bool nios2_valid_compare_const_p (enum rtx_code code, rtx op) { switch (code) { case EQ: case NE: case GE: case LT: return SMALL_INT (INTVAL (op)); case GEU: case LTU: return SMALL_INT_UNSIGNED (INTVAL (op)); default: return false; } } /* Checks if the FPU comparison in *CMP, *OP1, and *OP2 can be supported in the current configuration. Perform modifications if MODIFY_P is true. Returns true if FPU compare can be done. */ bool nios2_validate_fpu_compare (machine_mode mode, rtx *cmp, rtx *op1, rtx *op2, bool modify_p) { bool rev_p = false; enum rtx_code code = GET_CODE (*cmp); if (!nios2_fpu_compare_enabled (code, mode)) { code = swap_condition (code); if (nios2_fpu_compare_enabled (code, mode)) rev_p = true; else return false; } if (modify_p) { if (rev_p) { rtx tmp = *op1; *op1 = *op2; *op2 = tmp; } *op1 = force_reg (mode, *op1); *op2 = force_reg (mode, *op2); *cmp = gen_rtx_fmt_ee (code, mode, *op1, *op2); } return true; } /* Checks and modifies the comparison in *CMP, *OP1, and *OP2 into valid nios2 supported form. Returns true if success. */ bool nios2_validate_compare (machine_mode mode, rtx *cmp, rtx *op1, rtx *op2) { enum rtx_code code = GET_CODE (*cmp); enum rtx_code alt_code; rtx alt_op2; if (GET_MODE_CLASS (mode) == MODE_FLOAT) return nios2_validate_fpu_compare (mode, cmp, op1, op2, true); if (!reg_or_0_operand (*op2, mode)) { /* Create alternate constant compare. */ nios2_alternate_compare_const (code, *op2, &alt_code, &alt_op2, mode); /* If alterate op2 is zero(0), we can use it directly, possibly swapping the compare code. */ if (alt_op2 == const0_rtx) { code = alt_code; *op2 = alt_op2; goto check_rebuild_cmp; } /* Check if either constant compare can be used. */ if (nios2_valid_compare_const_p (code, *op2)) return true; else if (nios2_valid_compare_const_p (alt_code, alt_op2)) { code = alt_code; *op2 = alt_op2; goto rebuild_cmp; } /* We have to force op2 into a register now. Try to pick one with a lower cost. */ if (! nios2_simple_const_p (*op2) && nios2_simple_const_p (alt_op2)) { code = alt_code; *op2 = alt_op2; } *op2 = force_reg (SImode, *op2); } check_rebuild_cmp: if (code == GT || code == GTU || code == LE || code == LEU) { rtx t = *op1; *op1 = *op2; *op2 = t; code = swap_condition (code); } rebuild_cmp: *cmp = gen_rtx_fmt_ee (code, mode, *op1, *op2); return true; } /* Addressing Modes. */ /* Implement TARGET_LEGITIMATE_CONSTANT_P. */ static bool nios2_legitimate_constant_p (machine_mode mode ATTRIBUTE_UNUSED, rtx x) { rtx base, offset; split_const (x, &base, &offset); return GET_CODE (base) != SYMBOL_REF || !SYMBOL_REF_TLS_MODEL (base); } /* Implement TARGET_CANNOT_FORCE_CONST_MEM. */ static bool nios2_cannot_force_const_mem (machine_mode mode ATTRIBUTE_UNUSED, rtx x) { return nios2_legitimate_constant_p (mode, x) == false; } /* Return true if register REGNO is a valid base register. STRICT_P is true if REG_OK_STRICT is in effect. */ bool nios2_regno_ok_for_base_p (int regno, bool strict_p) { if (!HARD_REGISTER_NUM_P (regno)) { if (!strict_p) return true; if (!reg_renumber) return false; regno = reg_renumber[regno]; } /* The fake registers will be eliminated to either the stack or hard frame pointer, both of which are usually valid base registers. Reload deals with the cases where the eliminated form isn't valid. */ return (GP_REG_P (regno) || regno == FRAME_POINTER_REGNUM || regno == ARG_POINTER_REGNUM); } /* Return true if OFFSET is permitted in a load/store address expression. Normally any 16-bit value is permitted, but on R2 if we may be emitting the IO forms of these instructions we must restrict the offset to fit in a 12-bit field instead. */ static bool nios2_valid_addr_offset_p (rtx offset) { return (CONST_INT_P (offset) && ((TARGET_ARCH_R2 && (TARGET_BYPASS_CACHE || TARGET_BYPASS_CACHE_VOLATILE)) ? SMALL_INT12 (INTVAL (offset)) : SMALL_INT (INTVAL (offset)))); } /* Return true if the address expression formed by BASE + OFFSET is valid. */ static bool nios2_valid_addr_expr_p (rtx base, rtx offset, bool strict_p) { if (!strict_p && GET_CODE (base) == SUBREG) base = SUBREG_REG (base); return (REG_P (base) && nios2_regno_ok_for_base_p (REGNO (base), strict_p) && (offset == NULL_RTX || nios2_valid_addr_offset_p (offset) || nios2_unspec_reloc_p (offset))); } /* Implement TARGET_LEGITIMATE_ADDRESS_P. */ static bool nios2_legitimate_address_p (machine_mode mode ATTRIBUTE_UNUSED, rtx operand, bool strict_p) { switch (GET_CODE (operand)) { /* Direct. */ case SYMBOL_REF: if (SYMBOL_REF_TLS_MODEL (operand)) return false; /* Else, fall through. */ case CONST: if (gprel_constant_p (operand)) return true; /* Else, fall through. */ case LABEL_REF: case CONST_INT: case CONST_DOUBLE: return false; /* Register indirect. */ case REG: return nios2_regno_ok_for_base_p (REGNO (operand), strict_p); /* Register indirect with displacement. */ case PLUS: { rtx op0 = XEXP (operand, 0); rtx op1 = XEXP (operand, 1); return (nios2_valid_addr_expr_p (op0, op1, strict_p) || nios2_valid_addr_expr_p (op1, op0, strict_p)); } default: break; } return false; } /* Return true if SECTION is a small section name. */ static bool nios2_small_section_name_p (const char *section) { return (strcmp (section, ".sbss") == 0 || strncmp (section, ".sbss.", 6) == 0 || strcmp (section, ".sdata") == 0 || strncmp (section, ".sdata.", 7) == 0); } /* Return true if EXP should be placed in the small data section. */ static bool nios2_in_small_data_p (const_tree exp) { /* We want to merge strings, so we never consider them small data. */ if (TREE_CODE (exp) == STRING_CST) return false; if (TREE_CODE (exp) == VAR_DECL) { if (DECL_SECTION_NAME (exp)) { const char *section = DECL_SECTION_NAME (exp); if (nios2_small_section_name_p (section)) return true; } else { HOST_WIDE_INT size = int_size_in_bytes (TREE_TYPE (exp)); /* If this is an incomplete type with size 0, then we can't put it in sdata because it might be too big when completed. */ if (size > 0 && (unsigned HOST_WIDE_INT) size <= nios2_section_threshold) return true; } } return false; } /* Return true if symbol is in small data section. */ static bool nios2_symbol_ref_in_small_data_p (rtx sym) { tree decl; gcc_assert (GET_CODE (sym) == SYMBOL_REF); decl = SYMBOL_REF_DECL (sym); /* TLS variables are not accessed through the GP. */ if (SYMBOL_REF_TLS_MODEL (sym) != 0) return false; /* On Nios II R2, there is no GP-relative relocation that can be used with "io" instructions. So, if we are implicitly generating those instructions, we cannot emit GP-relative accesses. */ if (TARGET_ARCH_R2 && (TARGET_BYPASS_CACHE || TARGET_BYPASS_CACHE_VOLATILE)) return false; /* If the user has explicitly placed the symbol in a small data section via an attribute, generate gp-relative addressing even if the symbol is external, weak, or larger than we'd automatically put in the small data section. OTOH, if the symbol is located in some non-small-data section, we can't use gp-relative accesses on it unless the user has requested gpopt_data or gpopt_all. */ switch (nios2_gpopt_option) { case gpopt_none: /* Don't generate a gp-relative addressing mode if that's been disabled. */ return false; case gpopt_local: /* Use GP-relative addressing for small data symbols that are not external or weak or uninitialized common, plus any symbols that have explicitly been placed in a small data section. */ if (decl && DECL_SECTION_NAME (decl)) return nios2_small_section_name_p (DECL_SECTION_NAME (decl)); return (SYMBOL_REF_SMALL_P (sym) && !SYMBOL_REF_EXTERNAL_P (sym) && !(decl && DECL_WEAK (decl)) && !(decl && DECL_COMMON (decl) && (DECL_INITIAL (decl) == NULL || (!in_lto_p && DECL_INITIAL (decl) == error_mark_node)))); case gpopt_global: /* Use GP-relative addressing for small data symbols, even if they are external or weak. Note that SYMBOL_REF_SMALL_P is also true of symbols that have explicitly been placed in a small data section. */ return SYMBOL_REF_SMALL_P (sym); case gpopt_data: /* Use GP-relative addressing for all data symbols regardless of the object size, but not for code symbols. This option is equivalent to the user asserting that the entire data section is accessible from the GP. */ return !SYMBOL_REF_FUNCTION_P (sym); case gpopt_all: /* Use GP-relative addressing for everything, including code. Effectively, the user has asserted that the entire program fits within the 64K range of the GP offset. */ return true; default: /* We shouldn't get here. */ return false; } } /* Implement TARGET_SECTION_TYPE_FLAGS. */ static unsigned int nios2_section_type_flags (tree decl, const char *name, int reloc) { unsigned int flags; flags = default_section_type_flags (decl, name, reloc); if (nios2_small_section_name_p (name)) flags |= SECTION_SMALL; return flags; } /* Return true if SYMBOL_REF X binds locally. */ static bool nios2_symbol_binds_local_p (const_rtx x) { return (SYMBOL_REF_DECL (x) ? targetm.binds_local_p (SYMBOL_REF_DECL (x)) : SYMBOL_REF_LOCAL_P (x)); } /* Position independent code related. */ /* Emit code to load the PIC register. */ static void nios2_load_pic_register (void) { rtx tmp = gen_rtx_REG (Pmode, TEMP_REG_NUM); emit_insn (gen_load_got_register (pic_offset_table_rtx, tmp)); emit_insn (gen_add3_insn (pic_offset_table_rtx, pic_offset_table_rtx, tmp)); } /* Generate a PIC address as a MEM rtx. */ static rtx nios2_load_pic_address (rtx sym, int unspec, rtx tmp) { if (flag_pic == 2 && GET_CODE (sym) == SYMBOL_REF && nios2_symbol_binds_local_p (sym)) /* Under -fPIC, generate a GOTOFF address for local symbols. */ { rtx offset = nios2_unspec_offset (sym, UNSPEC_PIC_GOTOFF_SYM); crtl->uses_pic_offset_table = 1; return nios2_large_got_address (offset, tmp); } return gen_const_mem (Pmode, nios2_got_address (sym, unspec)); } /* Nonzero if the constant value X is a legitimate general operand when generating PIC code. It is given that flag_pic is on and that X satisfies CONSTANT_P or is a CONST_DOUBLE. */ bool nios2_legitimate_pic_operand_p (rtx x) { if (nios2_large_unspec_reloc_p (x)) return true; return ! (GET_CODE (x) == SYMBOL_REF || GET_CODE (x) == LABEL_REF || GET_CODE (x) == CONST); } /* Return TRUE if X is a thread-local symbol. */ static bool nios2_tls_symbol_p (rtx x) { return (targetm.have_tls && GET_CODE (x) == SYMBOL_REF && SYMBOL_REF_TLS_MODEL (x) != 0); } /* Legitimize addresses that are CONSTANT_P expressions. */ static rtx nios2_legitimize_constant_address (rtx addr) { rtx base, offset; split_const (addr, &base, &offset); if (nios2_tls_symbol_p (base)) base = nios2_legitimize_tls_address (base); else if (flag_pic) base = nios2_load_pic_address (base, UNSPEC_PIC_SYM, NULL_RTX); else return addr; if (offset != const0_rtx) { gcc_assert (can_create_pseudo_p ()); return gen_rtx_PLUS (Pmode, force_reg (Pmode, base), (CONST_INT_P (offset) ? (SMALL_INT (INTVAL (offset)) ? offset : force_reg (Pmode, offset)) : offset)); } return base; } /* Implement TARGET_LEGITIMIZE_ADDRESS. */ static rtx nios2_legitimize_address (rtx x, rtx oldx ATTRIBUTE_UNUSED, machine_mode mode ATTRIBUTE_UNUSED) { if (CONSTANT_P (x)) return nios2_legitimize_constant_address (x); /* For the TLS LE (Local Exec) model, the compiler may try to combine constant offsets with unspec relocs, creating address RTXs looking like this: (plus:SI (reg:SI 23 r23) (const:SI (plus:SI (unspec:SI [(symbol_ref:SI ("var"))] UNSPEC_ADD_TLS_LE) (const_int 48 [0x30])))) This usually happens when 'var' is a thread-local struct variable, and access of a field in var causes the addend. We typically want this combining, so transform the above into this form, which is allowed: (plus:SI (reg:SI 23 r23) (const:SI (unspec:SI [(const:SI (plus:SI (symbol_ref:SI ("var")) (const_int 48 [0x30])))] UNSPEC_ADD_TLS_LE))) Which will be output as '%tls_le(var+48)(r23)' in assembly. */ if (GET_CODE (x) == PLUS && GET_CODE (XEXP (x, 1)) == CONST) { rtx unspec, offset; split_const (XEXP (x, 1), &unspec, &offset); if (GET_CODE (unspec) == UNSPEC && !nios2_large_offset_p (XINT (unspec, 1)) && offset != const0_rtx) { rtx reg = force_reg (Pmode, XEXP (x, 0)); unspec = copy_rtx (unspec); XVECEXP (unspec, 0, 0) = plus_constant (Pmode, XVECEXP (unspec, 0, 0), INTVAL (offset)); x = gen_rtx_PLUS (Pmode, reg, gen_rtx_CONST (Pmode, unspec)); } } return x; } static rtx nios2_delegitimize_address (rtx x) { x = delegitimize_mem_from_attrs (x); if (GET_CODE (x) == CONST && GET_CODE (XEXP (x, 0)) == UNSPEC) { switch (XINT (XEXP (x, 0), 1)) { case UNSPEC_PIC_SYM: case UNSPEC_PIC_CALL_SYM: case UNSPEC_PIC_GOTOFF_SYM: case UNSPEC_ADD_TLS_GD: case UNSPEC_ADD_TLS_LDM: case UNSPEC_LOAD_TLS_IE: case UNSPEC_ADD_TLS_LE: x = XVECEXP (XEXP (x, 0), 0, 0); gcc_assert (CONSTANT_P (x)); break; } } return x; } /* Main expander function for RTL moves. */ bool nios2_emit_move_sequence (rtx *operands, machine_mode mode) { rtx to = operands[0]; rtx from = operands[1]; if (!register_operand (to, mode) && !reg_or_0_operand (from, mode)) { gcc_assert (can_create_pseudo_p ()); from = copy_to_mode_reg (mode, from); } if (CONSTANT_P (from)) { if (CONST_INT_P (from)) { if (!SMALL_INT (INTVAL (from)) && !SMALL_INT_UNSIGNED (INTVAL (from)) && !UPPER16_INT (INTVAL (from))) { HOST_WIDE_INT high = (INTVAL (from) + 0x8000) & ~0xffff; HOST_WIDE_INT low = INTVAL (from) & 0xffff; emit_move_insn (to, gen_int_mode (high, SImode)); emit_insn (gen_add2_insn (to, gen_int_mode (low, HImode))); set_unique_reg_note (get_last_insn (), REG_EQUAL, copy_rtx (from)); return true; } } else if (!gprel_constant_p (from)) { if (!nios2_large_unspec_reloc_p (from)) from = nios2_legitimize_constant_address (from); if (CONSTANT_P (from)) { emit_insn (gen_rtx_SET (to, gen_rtx_HIGH (Pmode, copy_rtx (from)))); emit_insn (gen_rtx_SET (to, gen_rtx_LO_SUM (Pmode, to, from))); set_unique_reg_note (get_last_insn (), REG_EQUAL, copy_rtx (operands[1])); return true; } } } operands[0] = to; operands[1] = from; return false; } /* The function with address *ADDR is being called. If the address needs to be loaded from the GOT, emit the instruction to do so and update *ADDR to point to the rtx for the loaded value. If REG != NULL_RTX, it is used as the target/scratch register in the GOT address calculation. */ void nios2_adjust_call_address (rtx *call_op, rtx reg) { if (MEM_P (*call_op)) call_op = &XEXP (*call_op, 0); rtx addr = *call_op; if (flag_pic && CONSTANT_P (addr)) { rtx tmp = reg ? reg : NULL_RTX; if (!reg) reg = gen_reg_rtx (Pmode); addr = nios2_load_pic_address (addr, UNSPEC_PIC_CALL_SYM, tmp); emit_insn (gen_rtx_SET (reg, addr)); *call_op = reg; } } /* Output assembly language related definitions. */ /* Implement TARGET_PRINT_OPERAND_PUNCT_VALID_P. */ static bool nios2_print_operand_punct_valid_p (unsigned char code) { return (code == '.' || code == '!'); } /* Print the operand OP to file stream FILE modified by LETTER. LETTER can be one of: i: print i/hi/ui suffixes (used for mov instruction variants), when OP is the appropriate immediate operand. u: like 'i', except without "ui" suffix case (used for cmpgeu/cmpltu) o: print "io" if OP needs volatile access (due to TARGET_BYPASS_CACHE or TARGET_BYPASS_CACHE_VOLATILE). x: print i/hi/ci/chi suffixes for the and instruction, when OP is the appropriate immediate operand. z: prints the third register immediate operand in assembly instructions. Outputs const0_rtx as the 'zero' register instead of '0'. y: same as 'z', but for specifically for logical instructions, where the processing for immediates are slightly different. H: for %hiadj L: for %lo D: for the upper 32-bits of a 64-bit double value R: prints reverse condition. A: prints (reg) operand for ld[s]ex and st[s]ex. .: print .n suffix for 16-bit instructions. !: print r.n suffix for 16-bit instructions. Used for jmpr.n. */ static void nios2_print_operand (FILE *file, rtx op, int letter) { /* First take care of the format letters that just insert a string into the output stream. */ switch (letter) { case '.': if (current_output_insn && get_attr_length (current_output_insn) == 2) fprintf (file, ".n"); return; case '!': if (current_output_insn && get_attr_length (current_output_insn) == 2) fprintf (file, "r.n"); return; case 'x': if (CONST_INT_P (op)) { HOST_WIDE_INT val = INTVAL (op); HOST_WIDE_INT low = val & 0xffff; HOST_WIDE_INT high = (val >> 16) & 0xffff; if (val != 0) { if (high != 0) { if (low != 0) { gcc_assert (TARGET_ARCH_R2); if (high == 0xffff) fprintf (file, "c"); else if (low == 0xffff) fprintf (file, "ch"); else gcc_unreachable (); } else fprintf (file, "h"); } fprintf (file, "i"); } } return; case 'u': case 'i': if (CONST_INT_P (op)) { HOST_WIDE_INT val = INTVAL (op); HOST_WIDE_INT low = val & 0xffff; HOST_WIDE_INT high = (val >> 16) & 0xffff; if (val != 0) { if (low == 0 && high != 0) fprintf (file, "h"); else if (high == 0 && (low & 0x8000) != 0 && letter != 'u') fprintf (file, "u"); } } if (CONSTANT_P (op) && op != const0_rtx) fprintf (file, "i"); return; case 'o': if (GET_CODE (op) == MEM && ((MEM_VOLATILE_P (op) && TARGET_BYPASS_CACHE_VOLATILE) || TARGET_BYPASS_CACHE)) { gcc_assert (current_output_insn && get_attr_length (current_output_insn) == 4); fprintf (file, "io"); } return; default: break; } /* Handle comparison operator names. */ if (comparison_operator (op, VOIDmode)) { enum rtx_code cond = GET_CODE (op); if (letter == 0) { fprintf (file, "%s", GET_RTX_NAME (cond)); return; } if (letter == 'R') { fprintf (file, "%s", GET_RTX_NAME (reverse_condition (cond))); return; } } /* Now handle the cases where we actually need to format an operand. */ switch (GET_CODE (op)) { case REG: if (letter == 0 || letter == 'z' || letter == 'y') { fprintf (file, "%s", reg_names[REGNO (op)]); return; } else if (letter == 'D') { fprintf (file, "%s", reg_names[REGNO (op)+1]); return; } break; case CONST_INT: { rtx int_rtx = op; HOST_WIDE_INT val = INTVAL (int_rtx); HOST_WIDE_INT low = val & 0xffff; HOST_WIDE_INT high = (val >> 16) & 0xffff; if (letter == 'y') { if (val == 0) fprintf (file, "zero"); else { if (high != 0) { if (low != 0) { gcc_assert (TARGET_ARCH_R2); if (high == 0xffff) /* andci. */ int_rtx = gen_int_mode (low, SImode); else if (low == 0xffff) /* andchi. */ int_rtx = gen_int_mode (high, SImode); else gcc_unreachable (); } else /* andhi. */ int_rtx = gen_int_mode (high, SImode); } else /* andi. */ int_rtx = gen_int_mode (low, SImode); output_addr_const (file, int_rtx); } return; } else if (letter == 'z') { if (val == 0) fprintf (file, "zero"); else { if (low == 0 && high != 0) int_rtx = gen_int_mode (high, SImode); else if (low != 0) { gcc_assert (high == 0 || high == 0xffff); int_rtx = gen_int_mode (low, high == 0 ? SImode : HImode); } else gcc_unreachable (); output_addr_const (file, int_rtx); } return; } } /* Else, fall through. */ case CONST: case LABEL_REF: case SYMBOL_REF: case CONST_DOUBLE: if (letter == 0 || letter == 'z') { output_addr_const (file, op); return; } else if (letter == 'H' || letter == 'L') { fprintf (file, "%%"); if (GET_CODE (op) == CONST && GET_CODE (XEXP (op, 0)) == UNSPEC) { rtx unspec = XEXP (op, 0); int unspec_reloc = XINT (unspec, 1); gcc_assert (nios2_large_offset_p (unspec_reloc)); fprintf (file, "%s_", nios2_unspec_reloc_name (unspec_reloc)); op = XVECEXP (unspec, 0, 0); } fprintf (file, letter == 'H' ? "hiadj(" : "lo("); output_addr_const (file, op); fprintf (file, ")"); return; } break; case SUBREG: case MEM: if (letter == 'A') { /* Address of '(reg)' form, with no index. */ fprintf (file, "(%s)", reg_names[REGNO (XEXP (op, 0))]); return; } if (letter == 0) { output_address (VOIDmode, op); return; } break; case CODE_LABEL: if (letter == 0) { output_addr_const (file, op); return; } break; default: break; } output_operand_lossage ("Unsupported operand for code '%c'", letter); gcc_unreachable (); } /* Return true if this is a GP-relative accessible reference. */ bool gprel_constant_p (rtx op) { if (GET_CODE (op) == SYMBOL_REF && nios2_symbol_ref_in_small_data_p (op)) return true; else if (GET_CODE (op) == CONST && GET_CODE (XEXP (op, 0)) == PLUS) return gprel_constant_p (XEXP (XEXP (op, 0), 0)); return false; } /* Return the name string for a supported unspec reloc offset. */ static const char * nios2_unspec_reloc_name (int unspec) { switch (unspec) { case UNSPEC_PIC_SYM: return "got"; case UNSPEC_PIC_CALL_SYM: return "call"; case UNSPEC_PIC_GOTOFF_SYM: return "gotoff"; case UNSPEC_LOAD_TLS_IE: return "tls_ie"; case UNSPEC_ADD_TLS_LE: return "tls_le"; case UNSPEC_ADD_TLS_GD: return "tls_gd"; case UNSPEC_ADD_TLS_LDM: return "tls_ldm"; case UNSPEC_ADD_TLS_LDO: return "tls_ldo"; default: return NULL; } } /* Implement TARGET_ASM_OUTPUT_ADDR_CONST_EXTRA. */ static bool nios2_output_addr_const_extra (FILE *file, rtx op) { const char *name; gcc_assert (GET_CODE (op) == UNSPEC); /* Support for printing out const unspec relocations. */ name = nios2_unspec_reloc_name (XINT (op, 1)); if (name) { fprintf (file, "%%%s(", name); output_addr_const (file, XVECEXP (op, 0, 0)); fprintf (file, ")"); return true; } return false; } /* Implement TARGET_PRINT_OPERAND_ADDRESS. */ static void nios2_print_operand_address (FILE *file, machine_mode mode, rtx op) { switch (GET_CODE (op)) { case CONST: case CONST_INT: case LABEL_REF: case CONST_DOUBLE: case SYMBOL_REF: if (gprel_constant_p (op)) { fprintf (file, "%%gprel("); output_addr_const (file, op); fprintf (file, ")(%s)", reg_names[GP_REGNO]); return; } break; case PLUS: { rtx op0 = XEXP (op, 0); rtx op1 = XEXP (op, 1); if (REG_P (op0) && CONSTANT_P (op1)) { output_addr_const (file, op1); fprintf (file, "(%s)", reg_names[REGNO (op0)]); return; } else if (REG_P (op1) && CONSTANT_P (op0)) { output_addr_const (file, op0); fprintf (file, "(%s)", reg_names[REGNO (op1)]); return; } } break; case REG: fprintf (file, "0(%s)", reg_names[REGNO (op)]); return; case MEM: { rtx base = XEXP (op, 0); nios2_print_operand_address (file, mode, base); return; } default: break; } fprintf (stderr, "Missing way to print address\n"); debug_rtx (op); gcc_unreachable (); } /* Implement TARGET_ASM_OUTPUT_DWARF_DTPREL. */ static void nios2_output_dwarf_dtprel (FILE *file, int size, rtx x) { gcc_assert (size == 4); fprintf (file, "\t.4byte\t%%tls_ldo("); output_addr_const (file, x); fprintf (file, ")"); } /* Implemet TARGET_ASM_FILE_END. */ static void nios2_asm_file_end (void) { /* The Nios II Linux stack is mapped non-executable by default, so add a .note.GNU-stack section for switching to executable stacks only when trampolines are generated. */ if (TARGET_LINUX_ABI && trampolines_created) file_end_indicate_exec_stack (); } /* Implement TARGET_ASM_FUNCTION_PROLOGUE. */ static void nios2_asm_function_prologue (FILE *file, HOST_WIDE_INT size ATTRIBUTE_UNUSED) { if (flag_verbose_asm || flag_debug_asm) { nios2_compute_frame_layout (); nios2_dump_frame_layout (file); } } /* Emit assembly of custom FPU instructions. */ const char * nios2_fpu_insn_asm (enum n2fpu_code code) { static char buf[256]; const char *op1, *op2, *op3; int ln = 256, n = 0; int N = N2FPU_N (code); int num_operands = N2FPU (code).num_operands; const char *insn_name = N2FPU_NAME (code); tree ftype = nios2_ftype (N2FPU_FTCODE (code)); machine_mode dst_mode = TYPE_MODE (TREE_TYPE (ftype)); machine_mode src_mode = TYPE_MODE (TREE_VALUE (TYPE_ARG_TYPES (ftype))); /* Prepare X register for DF input operands. */ if (GET_MODE_SIZE (src_mode) == 8 && num_operands == 3) n = snprintf (buf, ln, "custom\t%d, zero, %%1, %%D1 # fwrx %%1\n\t", N2FPU_N (n2fpu_fwrx)); if (src_mode == SFmode) { if (dst_mode == VOIDmode) { /* The fwry case. */ op1 = op3 = "zero"; op2 = "%0"; num_operands -= 1; } else { op1 = (dst_mode == DFmode ? "%D0" : "%0"); op2 = "%1"; op3 = (num_operands == 2 ? "zero" : "%2"); } } else if (src_mode == DFmode) { if (dst_mode == VOIDmode) { /* The fwrx case. */ op1 = "zero"; op2 = "%0"; op3 = "%D0"; num_operands -= 1; } else { op1 = (dst_mode == DFmode ? "%D0" : "%0"); op2 = (num_operands == 2 ? "%1" : "%2"); op3 = (num_operands == 2 ? "%D1" : "%D2"); } } else if (src_mode == VOIDmode) { /* frdxlo, frdxhi, frdy cases. */ gcc_assert (dst_mode == SFmode); op1 = "%0"; op2 = op3 = "zero"; } else if (src_mode == SImode) { /* Conversion operators. */ gcc_assert (num_operands == 2); op1 = (dst_mode == DFmode ? "%D0" : "%0"); op2 = "%1"; op3 = "zero"; } else gcc_unreachable (); /* Main instruction string. */ n += snprintf (buf + n, ln - n, "custom\t%d, %s, %s, %s # %s %%0%s%s", N, op1, op2, op3, insn_name, (num_operands >= 2 ? ", %1" : ""), (num_operands == 3 ? ", %2" : "")); /* Extraction of Y register for DF results. */ if (dst_mode == DFmode) snprintf (buf + n, ln - n, "\n\tcustom\t%d, %%0, zero, zero # frdy %%0", N2FPU_N (n2fpu_frdy)); return buf; } /* Function argument related. */ /* Define where to put the arguments to a function. Value is zero to push the argument on the stack, or a hard register in which to store the argument. MODE is the argument's machine mode. TYPE is the data type of the argument (as a tree). This is null for libcalls where that information may not be available. CUM is a variable of type CUMULATIVE_ARGS which gives info about the preceding args and about the function being called. NAMED is nonzero if this argument is a named parameter (otherwise it is an extra parameter matching an ellipsis). */ static rtx nios2_function_arg (cumulative_args_t cum_v, machine_mode mode, const_tree type ATTRIBUTE_UNUSED, bool named ATTRIBUTE_UNUSED) { CUMULATIVE_ARGS *cum = get_cumulative_args (cum_v); rtx return_rtx = NULL_RTX; if (cum->regs_used < NUM_ARG_REGS) return_rtx = gen_rtx_REG (mode, FIRST_ARG_REGNO + cum->regs_used); return return_rtx; } /* Return number of bytes, at the beginning of the argument, that must be put in registers. 0 is the argument is entirely in registers or entirely in memory. */ static int nios2_arg_partial_bytes (cumulative_args_t cum_v, machine_mode mode, tree type ATTRIBUTE_UNUSED, bool named ATTRIBUTE_UNUSED) { CUMULATIVE_ARGS *cum = get_cumulative_args (cum_v); HOST_WIDE_INT param_size; if (mode == BLKmode) { param_size = int_size_in_bytes (type); gcc_assert (param_size >= 0); } else param_size = GET_MODE_SIZE (mode); /* Convert to words (round up). */ param_size = (UNITS_PER_WORD - 1 + param_size) / UNITS_PER_WORD; if (cum->regs_used < NUM_ARG_REGS && cum->regs_used + param_size > NUM_ARG_REGS) return (NUM_ARG_REGS - cum->regs_used) * UNITS_PER_WORD; return 0; } /* Update the data in CUM to advance over an argument of mode MODE and data type TYPE; TYPE is null for libcalls where that information may not be available. */ static void nios2_function_arg_advance (cumulative_args_t cum_v, machine_mode mode, const_tree type ATTRIBUTE_UNUSED, bool named ATTRIBUTE_UNUSED) { CUMULATIVE_ARGS *cum = get_cumulative_args (cum_v); HOST_WIDE_INT param_size; if (mode == BLKmode) { param_size = int_size_in_bytes (type); gcc_assert (param_size >= 0); } else param_size = GET_MODE_SIZE (mode); /* Convert to words (round up). */ param_size = (UNITS_PER_WORD - 1 + param_size) / UNITS_PER_WORD; if (cum->regs_used + param_size > NUM_ARG_REGS) cum->regs_used = NUM_ARG_REGS; else cum->regs_used += param_size; } enum direction nios2_function_arg_padding (machine_mode mode, const_tree type) { /* On little-endian targets, the first byte of every stack argument is passed in the first byte of the stack slot. */ if (!BYTES_BIG_ENDIAN) return upward; /* Otherwise, integral types are padded downward: the last byte of a stack argument is passed in the last byte of the stack slot. */ if (type != 0 ? INTEGRAL_TYPE_P (type) || POINTER_TYPE_P (type) : GET_MODE_CLASS (mode) == MODE_INT) return downward; /* Arguments smaller than a stack slot are padded downward. */ if (mode != BLKmode) return (GET_MODE_BITSIZE (mode) >= PARM_BOUNDARY) ? upward : downward; return ((int_size_in_bytes (type) >= (PARM_BOUNDARY / BITS_PER_UNIT)) ? upward : downward); } enum direction nios2_block_reg_padding (machine_mode mode, tree type, int first ATTRIBUTE_UNUSED) { return nios2_function_arg_padding (mode, type); } /* Emit RTL insns to initialize the variable parts of a trampoline. FNADDR is an RTX for the address of the function's pure code. CXT is an RTX for the static chain value for the function. On Nios II, we handle this by a library call. */ static void nios2_trampoline_init (rtx m_tramp, tree fndecl, rtx cxt) { rtx fnaddr = XEXP (DECL_RTL (fndecl), 0); rtx ctx_reg = force_reg (Pmode, cxt); rtx addr = force_reg (Pmode, XEXP (m_tramp, 0)); emit_library_call (gen_rtx_SYMBOL_REF (Pmode, "__trampoline_setup"), LCT_NORMAL, VOIDmode, 3, addr, Pmode, fnaddr, Pmode, ctx_reg, Pmode); } /* Implement TARGET_FUNCTION_VALUE. */ static rtx nios2_function_value (const_tree ret_type, const_tree fn ATTRIBUTE_UNUSED, bool outgoing ATTRIBUTE_UNUSED) { return gen_rtx_REG (TYPE_MODE (ret_type), FIRST_RETVAL_REGNO); } /* Implement TARGET_LIBCALL_VALUE. */ static rtx nios2_libcall_value (machine_mode mode, const_rtx fun ATTRIBUTE_UNUSED) { return gen_rtx_REG (mode, FIRST_RETVAL_REGNO); } /* Implement TARGET_FUNCTION_VALUE_REGNO_P. */ static bool nios2_function_value_regno_p (const unsigned int regno) { return regno == FIRST_RETVAL_REGNO; } /* Implement TARGET_RETURN_IN_MEMORY. */ static bool nios2_return_in_memory (const_tree type, const_tree fntype ATTRIBUTE_UNUSED) { return (int_size_in_bytes (type) > (2 * UNITS_PER_WORD) || int_size_in_bytes (type) == -1); } /* TODO: It may be possible to eliminate the copyback and implement own va_arg type. */ static void nios2_setup_incoming_varargs (cumulative_args_t cum_v, machine_mode mode, tree type, int *pretend_size, int second_time) { CUMULATIVE_ARGS *cum = get_cumulative_args (cum_v); CUMULATIVE_ARGS local_cum; cumulative_args_t local_cum_v = pack_cumulative_args (&local_cum); int regs_to_push; int pret_size; cfun->machine->uses_anonymous_args = 1; local_cum = *cum; nios2_function_arg_advance (local_cum_v, mode, type, true); regs_to_push = NUM_ARG_REGS - local_cum.regs_used; /* If we can use CDX stwm to push the arguments on the stack, nios2_expand_prologue will do that instead. */ if (!TARGET_HAS_CDX && !second_time && regs_to_push > 0) { rtx ptr = virtual_incoming_args_rtx; rtx mem = gen_rtx_MEM (BLKmode, ptr); emit_insn (gen_blockage ()); move_block_from_reg (local_cum.regs_used + FIRST_ARG_REGNO, mem, regs_to_push); emit_insn (gen_blockage ()); } pret_size = regs_to_push * UNITS_PER_WORD; if (pret_size) *pretend_size = pret_size; } /* Init FPU builtins. */ static void nios2_init_fpu_builtins (int start_code) { tree fndecl; char builtin_name[64] = "__builtin_custom_"; unsigned int i, n = strlen ("__builtin_custom_"); for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) { snprintf (builtin_name + n, sizeof (builtin_name) - n, "%s", N2FPU_NAME (i)); fndecl = add_builtin_function (builtin_name, nios2_ftype (N2FPU_FTCODE (i)), start_code + i, BUILT_IN_MD, NULL, NULL_TREE); nios2_register_builtin_fndecl (start_code + i, fndecl); } } /* Helper function for expanding FPU builtins. */ static rtx nios2_expand_fpu_builtin (tree exp, unsigned int code, rtx target) { struct expand_operand ops[MAX_RECOG_OPERANDS]; enum insn_code icode = N2FPU_ICODE (code); int nargs, argno, opno = 0; int num_operands = N2FPU (code).num_operands; machine_mode dst_mode = TYPE_MODE (TREE_TYPE (exp)); bool has_target_p = (dst_mode != VOIDmode); if (N2FPU_N (code) < 0) fatal_error (input_location, "Cannot call %<__builtin_custom_%s%> without specifying switch" " %<-mcustom-%s%>", N2FPU_NAME (code), N2FPU_NAME (code)); if (has_target_p) create_output_operand (&ops[opno++], target, dst_mode); else /* Subtract away the count of the VOID return, mainly for fwrx/fwry. */ num_operands -= 1; nargs = call_expr_nargs (exp); for (argno = 0; argno < nargs; argno++) { tree arg = CALL_EXPR_ARG (exp, argno); create_input_operand (&ops[opno++], expand_normal (arg), TYPE_MODE (TREE_TYPE (arg))); } if (!maybe_expand_insn (icode, num_operands, ops)) { error ("invalid argument to built-in function"); return has_target_p ? gen_reg_rtx (ops[0].mode) : const0_rtx; } return has_target_p ? ops[0].value : const0_rtx; } /* Nios II has custom instruction built-in functions of the forms: __builtin_custom_n __builtin_custom_nX __builtin_custom_nXX __builtin_custom_Xn __builtin_custom_XnX __builtin_custom_XnXX where each X could be either 'i' (int), 'f' (float), or 'p' (void*). Therefore with 0-1 return values, and 0-2 arguments, we have a total of (3 + 1) * (1 + 3 + 9) == 52 custom builtin functions. */ #define NUM_CUSTOM_BUILTINS ((3 + 1) * (1 + 3 + 9)) static char custom_builtin_name[NUM_CUSTOM_BUILTINS][5]; static void nios2_init_custom_builtins (int start_code) { tree builtin_ftype, ret_type, fndecl; char builtin_name[32] = "__builtin_custom_"; int n = strlen ("__builtin_custom_"); int builtin_code = 0; int lhs, rhs1, rhs2; struct { tree type; const char *c; } op[4]; /* z */ op[0].c = ""; op[0].type = NULL_TREE; /* f */ op[1].c = "f"; op[1].type = float_type_node; /* i */ op[2].c = "i"; op[2].type = integer_type_node; /* p */ op[3].c = "p"; op[3].type = ptr_type_node; /* We enumerate through the possible operand types to create all the __builtin_custom_XnXX function tree types. Note that these may slightly overlap with the function types created for other fixed builtins. */ for (lhs = 0; lhs < 4; lhs++) for (rhs1 = 0; rhs1 < 4; rhs1++) for (rhs2 = 0; rhs2 < 4; rhs2++) { if (rhs1 == 0 && rhs2 != 0) continue; ret_type = (op[lhs].type ? op[lhs].type : void_type_node); builtin_ftype = build_function_type_list (ret_type, integer_type_node, op[rhs1].type, op[rhs2].type, NULL_TREE); snprintf (builtin_name + n, 32 - n, "%sn%s%s", op[lhs].c, op[rhs1].c, op[rhs2].c); /* Save copy of parameter string into custom_builtin_name[]. */ strncpy (custom_builtin_name[builtin_code], builtin_name + n, 5); fndecl = add_builtin_function (builtin_name, builtin_ftype, start_code + builtin_code, BUILT_IN_MD, NULL, NULL_TREE); nios2_register_builtin_fndecl (start_code + builtin_code, fndecl); builtin_code += 1; } } /* Helper function for expanding custom builtins. */ static rtx nios2_expand_custom_builtin (tree exp, unsigned int index, rtx target) { bool has_target_p = (TREE_TYPE (exp) != void_type_node); machine_mode tmode = VOIDmode; int nargs, argno; rtx value, insn, unspec_args[3]; tree arg; /* XnXX form. */ if (has_target_p) { tmode = TYPE_MODE (TREE_TYPE (exp)); if (!target || GET_MODE (target) != tmode || !REG_P (target)) target = gen_reg_rtx (tmode); } nargs = call_expr_nargs (exp); for (argno = 0; argno < nargs; argno++) { arg = CALL_EXPR_ARG (exp, argno); value = expand_normal (arg); unspec_args[argno] = value; if (argno == 0) { if (!custom_insn_opcode (value, VOIDmode)) error ("custom instruction opcode must be compile time " "constant in the range 0-255 for __builtin_custom_%s", custom_builtin_name[index]); } else /* For other arguments, force into a register. */ unspec_args[argno] = force_reg (TYPE_MODE (TREE_TYPE (arg)), unspec_args[argno]); } /* Fill remaining unspec operands with zero. */ for (; argno < 3; argno++) unspec_args[argno] = const0_rtx; insn = (has_target_p ? gen_rtx_SET (target, gen_rtx_UNSPEC_VOLATILE (tmode, gen_rtvec_v (3, unspec_args), UNSPECV_CUSTOM_XNXX)) : gen_rtx_UNSPEC_VOLATILE (VOIDmode, gen_rtvec_v (3, unspec_args), UNSPECV_CUSTOM_NXX)); emit_insn (insn); return has_target_p ? target : const0_rtx; } /* Main definition of built-in functions. Nios II has a small number of fixed builtins, plus a large number of FPU insn builtins, and builtins for generating custom instructions. */ struct nios2_builtin_desc { enum insn_code icode; enum nios2_arch_type arch; enum nios2_ftcode ftype; const char *name; }; #define N2_BUILTINS \ N2_BUILTIN_DEF (sync, R1, N2_FTYPE_VOID_VOID) \ N2_BUILTIN_DEF (ldbio, R1, N2_FTYPE_SI_CVPTR) \ N2_BUILTIN_DEF (ldbuio, R1, N2_FTYPE_UI_CVPTR) \ N2_BUILTIN_DEF (ldhio, R1, N2_FTYPE_SI_CVPTR) \ N2_BUILTIN_DEF (ldhuio, R1, N2_FTYPE_UI_CVPTR) \ N2_BUILTIN_DEF (ldwio, R1, N2_FTYPE_SI_CVPTR) \ N2_BUILTIN_DEF (stbio, R1, N2_FTYPE_VOID_VPTR_SI) \ N2_BUILTIN_DEF (sthio, R1, N2_FTYPE_VOID_VPTR_SI) \ N2_BUILTIN_DEF (stwio, R1, N2_FTYPE_VOID_VPTR_SI) \ N2_BUILTIN_DEF (rdctl, R1, N2_FTYPE_SI_SI) \ N2_BUILTIN_DEF (wrctl, R1, N2_FTYPE_VOID_SI_SI) \ N2_BUILTIN_DEF (rdprs, R1, N2_FTYPE_SI_SI_SI) \ N2_BUILTIN_DEF (flushd, R1, N2_FTYPE_VOID_VPTR) \ N2_BUILTIN_DEF (flushda, R1, N2_FTYPE_VOID_VPTR) \ N2_BUILTIN_DEF (wrpie, R2, N2_FTYPE_SI_SI) \ N2_BUILTIN_DEF (eni, R2, N2_FTYPE_VOID_SI) \ N2_BUILTIN_DEF (ldex, R2, N2_FTYPE_SI_CVPTR) \ N2_BUILTIN_DEF (ldsex, R2, N2_FTYPE_SI_CVPTR) \ N2_BUILTIN_DEF (stex, R2, N2_FTYPE_SI_VPTR_SI) \ N2_BUILTIN_DEF (stsex, R2, N2_FTYPE_SI_VPTR_SI) enum nios2_builtin_code { #define N2_BUILTIN_DEF(name, arch, ftype) NIOS2_BUILTIN_ ## name, N2_BUILTINS #undef N2_BUILTIN_DEF NUM_FIXED_NIOS2_BUILTINS }; static const struct nios2_builtin_desc nios2_builtins[] = { #define N2_BUILTIN_DEF(name, arch, ftype) \ { CODE_FOR_ ## name, ARCH_ ## arch, ftype, "__builtin_" #name }, N2_BUILTINS #undef N2_BUILTIN_DEF }; /* Start/ends of FPU/custom insn builtin index ranges. */ static unsigned int nios2_fpu_builtin_base; static unsigned int nios2_custom_builtin_base; static unsigned int nios2_custom_builtin_end; /* Implement TARGET_INIT_BUILTINS. */ static void nios2_init_builtins (void) { unsigned int i; /* Initialize fixed builtins. */ for (i = 0; i < ARRAY_SIZE (nios2_builtins); i++) { const struct nios2_builtin_desc *d = &nios2_builtins[i]; tree fndecl = add_builtin_function (d->name, nios2_ftype (d->ftype), i, BUILT_IN_MD, NULL, NULL); nios2_register_builtin_fndecl (i, fndecl); } /* Initialize FPU builtins. */ nios2_fpu_builtin_base = ARRAY_SIZE (nios2_builtins); nios2_init_fpu_builtins (nios2_fpu_builtin_base); /* Initialize custom insn builtins. */ nios2_custom_builtin_base = nios2_fpu_builtin_base + ARRAY_SIZE (nios2_fpu_insn); nios2_custom_builtin_end = nios2_custom_builtin_base + NUM_CUSTOM_BUILTINS; nios2_init_custom_builtins (nios2_custom_builtin_base); } /* Array of fndecls for TARGET_BUILTIN_DECL. */ #define NIOS2_NUM_BUILTINS \ (ARRAY_SIZE (nios2_builtins) + ARRAY_SIZE (nios2_fpu_insn) + NUM_CUSTOM_BUILTINS) static GTY(()) tree nios2_builtin_decls[NIOS2_NUM_BUILTINS]; static void nios2_register_builtin_fndecl (unsigned code, tree fndecl) { nios2_builtin_decls[code] = fndecl; } /* Implement TARGET_BUILTIN_DECL. */ static tree nios2_builtin_decl (unsigned code, bool initialize_p ATTRIBUTE_UNUSED) { gcc_assert (nios2_custom_builtin_end == ARRAY_SIZE (nios2_builtin_decls)); if (code >= nios2_custom_builtin_end) return error_mark_node; if (code >= nios2_fpu_builtin_base && code < nios2_custom_builtin_base && ! N2FPU_ENABLED_P (code - nios2_fpu_builtin_base)) return error_mark_node; return nios2_builtin_decls[code]; } /* Low-level built-in expand routine. */ static rtx nios2_expand_builtin_insn (const struct nios2_builtin_desc *d, int n, struct expand_operand *ops, bool has_target_p) { if (maybe_expand_insn (d->icode, n, ops)) return has_target_p ? ops[0].value : const0_rtx; else { error ("invalid argument to built-in function %s", d->name); return has_target_p ? gen_reg_rtx (ops[0].mode) : const0_rtx; } } /* Expand ldio/stio and ldex/ldsex/stex/stsex form load-store instruction builtins. */ static rtx nios2_expand_ldst_builtin (tree exp, rtx target, const struct nios2_builtin_desc *d) { bool has_target_p; rtx addr, mem, val; struct expand_operand ops[MAX_RECOG_OPERANDS]; machine_mode mode = insn_data[d->icode].operand[0].mode; addr = expand_normal (CALL_EXPR_ARG (exp, 0)); mem = gen_rtx_MEM (mode, addr); if (insn_data[d->icode].operand[0].allows_mem) { /* stxio/stex/stsex. */ val = expand_normal (CALL_EXPR_ARG (exp, 1)); if (CONST_INT_P (val)) val = force_reg (mode, gen_int_mode (INTVAL (val), mode)); val = simplify_gen_subreg (mode, val, GET_MODE (val), 0); create_output_operand (&ops[0], mem, mode); create_input_operand (&ops[1], val, mode); if (insn_data[d->icode].n_operands == 3) { /* stex/stsex status value, returned as result of function. */ create_output_operand (&ops[2], target, mode); has_target_p = true; } else has_target_p = false; } else { /* ldxio. */ create_output_operand (&ops[0], target, mode); create_input_operand (&ops[1], mem, mode); has_target_p = true; } return nios2_expand_builtin_insn (d, insn_data[d->icode].n_operands, ops, has_target_p); } /* Expand rdctl/wrctl builtins. */ static rtx nios2_expand_rdwrctl_builtin (tree exp, rtx target, const struct nios2_builtin_desc *d) { bool has_target_p = (insn_data[d->icode].operand[0].predicate == register_operand); rtx ctlcode = expand_normal (CALL_EXPR_ARG (exp, 0)); struct expand_operand ops[MAX_RECOG_OPERANDS]; if (!rdwrctl_operand (ctlcode, VOIDmode)) { error ("Control register number must be in range 0-31 for %s", d->name); return has_target_p ? gen_reg_rtx (SImode) : const0_rtx; } if (has_target_p) { create_output_operand (&ops[0], target, SImode); create_integer_operand (&ops[1], INTVAL (ctlcode)); } else { rtx val = expand_normal (CALL_EXPR_ARG (exp, 1)); create_integer_operand (&ops[0], INTVAL (ctlcode)); create_input_operand (&ops[1], val, SImode); } return nios2_expand_builtin_insn (d, 2, ops, has_target_p); } static rtx nios2_expand_rdprs_builtin (tree exp, rtx target, const struct nios2_builtin_desc *d) { rtx reg = expand_normal (CALL_EXPR_ARG (exp, 0)); rtx imm = expand_normal (CALL_EXPR_ARG (exp, 1)); struct expand_operand ops[MAX_RECOG_OPERANDS]; if (!rdwrctl_operand (reg, VOIDmode)) { error ("Register number must be in range 0-31 for %s", d->name); return gen_reg_rtx (SImode); } if (!rdprs_dcache_operand (imm, VOIDmode)) { error ("The immediate value must fit into a %d-bit integer for %s", (TARGET_ARCH_R2) ? 12 : 16, d->name); return gen_reg_rtx (SImode); } create_output_operand (&ops[0], target, SImode); create_input_operand (&ops[1], reg, SImode); create_integer_operand (&ops[2], INTVAL (imm)); return nios2_expand_builtin_insn (d, 3, ops, true); } static rtx nios2_expand_cache_builtin (tree exp, rtx target ATTRIBUTE_UNUSED, const struct nios2_builtin_desc *d) { rtx mem, addr; struct expand_operand ops[MAX_RECOG_OPERANDS]; addr = expand_normal (CALL_EXPR_ARG (exp, 0)); mem = gen_rtx_MEM (SImode, addr); create_input_operand (&ops[0], mem, SImode); return nios2_expand_builtin_insn (d, 1, ops, false); } static rtx nios2_expand_wrpie_builtin (tree exp, rtx target, const struct nios2_builtin_desc *d) { rtx val; struct expand_operand ops[MAX_RECOG_OPERANDS]; val = expand_normal (CALL_EXPR_ARG (exp, 0)); create_input_operand (&ops[1], val, SImode); create_output_operand (&ops[0], target, SImode); return nios2_expand_builtin_insn (d, 2, ops, true); } static rtx nios2_expand_eni_builtin (tree exp, rtx target ATTRIBUTE_UNUSED, const struct nios2_builtin_desc *d) { rtx imm = expand_normal (CALL_EXPR_ARG (exp, 0)); struct expand_operand ops[MAX_RECOG_OPERANDS]; if (INTVAL (imm) != 0 && INTVAL (imm) != 1) { error ("The ENI instruction operand must be either 0 or 1"); return const0_rtx; } create_integer_operand (&ops[0], INTVAL (imm)); return nios2_expand_builtin_insn (d, 1, ops, false); } /* Implement TARGET_EXPAND_BUILTIN. Expand an expression EXP that calls a built-in function, with result going to TARGET if that's convenient (and in mode MODE if that's convenient). SUBTARGET may be used as the target for computing one of EXP's operands. IGNORE is nonzero if the value is to be ignored. */ static rtx nios2_expand_builtin (tree exp, rtx target, rtx subtarget ATTRIBUTE_UNUSED, machine_mode mode ATTRIBUTE_UNUSED, int ignore ATTRIBUTE_UNUSED) { tree fndecl = TREE_OPERAND (CALL_EXPR_FN (exp), 0); unsigned int fcode = DECL_FUNCTION_CODE (fndecl); if (fcode < nios2_fpu_builtin_base) { const struct nios2_builtin_desc *d = &nios2_builtins[fcode]; if (d->arch > nios2_arch_option) { error ("Builtin function %s requires Nios II R%d", d->name, (int) d->arch); /* Given it is invalid, just generate a normal call. */ return expand_call (exp, target, ignore); } switch (fcode) { case NIOS2_BUILTIN_sync: emit_insn (gen_sync ()); return const0_rtx; case NIOS2_BUILTIN_ldbio: case NIOS2_BUILTIN_ldbuio: case NIOS2_BUILTIN_ldhio: case NIOS2_BUILTIN_ldhuio: case NIOS2_BUILTIN_ldwio: case NIOS2_BUILTIN_stbio: case NIOS2_BUILTIN_sthio: case NIOS2_BUILTIN_stwio: case NIOS2_BUILTIN_ldex: case NIOS2_BUILTIN_ldsex: case NIOS2_BUILTIN_stex: case NIOS2_BUILTIN_stsex: return nios2_expand_ldst_builtin (exp, target, d); case NIOS2_BUILTIN_rdctl: case NIOS2_BUILTIN_wrctl: return nios2_expand_rdwrctl_builtin (exp, target, d); case NIOS2_BUILTIN_rdprs: return nios2_expand_rdprs_builtin (exp, target, d); case NIOS2_BUILTIN_flushd: case NIOS2_BUILTIN_flushda: return nios2_expand_cache_builtin (exp, target, d); case NIOS2_BUILTIN_wrpie: return nios2_expand_wrpie_builtin (exp, target, d); case NIOS2_BUILTIN_eni: return nios2_expand_eni_builtin (exp, target, d); default: gcc_unreachable (); } } else if (fcode < nios2_custom_builtin_base) /* FPU builtin range. */ return nios2_expand_fpu_builtin (exp, fcode - nios2_fpu_builtin_base, target); else if (fcode < nios2_custom_builtin_end) /* Custom insn builtin range. */ return nios2_expand_custom_builtin (exp, fcode - nios2_custom_builtin_base, target); else gcc_unreachable (); } /* Implement TARGET_INIT_LIBFUNCS. */ static void ATTRIBUTE_UNUSED nios2_init_libfuncs (void) { init_sync_libfuncs (UNITS_PER_WORD); } /* Register a custom code use, and signal error if a conflict was found. */ static void nios2_register_custom_code (unsigned int N, enum nios2_ccs_code status, int index) { gcc_assert (N <= 255); if (status == CCS_FPU) { if (custom_code_status[N] == CCS_FPU && index != custom_code_index[N]) { custom_code_conflict = true; error ("switch %<-mcustom-%s%> conflicts with switch %<-mcustom-%s%>", N2FPU_NAME (custom_code_index[N]), N2FPU_NAME (index)); } else if (custom_code_status[N] == CCS_BUILTIN_CALL) { custom_code_conflict = true; error ("call to %<__builtin_custom_%s%> conflicts with switch " "%<-mcustom-%s%>", custom_builtin_name[custom_code_index[N]], N2FPU_NAME (index)); } } else if (status == CCS_BUILTIN_CALL) { if (custom_code_status[N] == CCS_FPU) { custom_code_conflict = true; error ("call to %<__builtin_custom_%s%> conflicts with switch " "%<-mcustom-%s%>", custom_builtin_name[index], N2FPU_NAME (custom_code_index[N])); } else { /* Note that code conflicts between different __builtin_custom_xnxx calls are not checked. */ } } else gcc_unreachable (); custom_code_status[N] = status; custom_code_index[N] = index; } /* Mark a custom code as not in use. */ static void nios2_deregister_custom_code (unsigned int N) { if (N <= 255) { custom_code_status[N] = CCS_UNUSED; custom_code_index[N] = 0; } } /* Target attributes can affect per-function option state, so we need to save/restore the custom code tracking info using the TARGET_OPTION_SAVE/TARGET_OPTION_RESTORE hooks. */ static void nios2_option_save (struct cl_target_option *ptr, struct gcc_options *opts ATTRIBUTE_UNUSED) { unsigned int i; for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) ptr->saved_fpu_custom_code[i] = N2FPU_N (i); memcpy (ptr->saved_custom_code_status, custom_code_status, sizeof (custom_code_status)); memcpy (ptr->saved_custom_code_index, custom_code_index, sizeof (custom_code_index)); } static void nios2_option_restore (struct gcc_options *opts ATTRIBUTE_UNUSED, struct cl_target_option *ptr) { unsigned int i; for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) N2FPU_N (i) = ptr->saved_fpu_custom_code[i]; memcpy (custom_code_status, ptr->saved_custom_code_status, sizeof (custom_code_status)); memcpy (custom_code_index, ptr->saved_custom_code_index, sizeof (custom_code_index)); } /* Inner function to process the attribute((target(...))), take an argument and set the current options from the argument. If we have a list, recursively go over the list. */ static bool nios2_valid_target_attribute_rec (tree args) { if (TREE_CODE (args) == TREE_LIST) { bool ret = true; for (; args; args = TREE_CHAIN (args)) if (TREE_VALUE (args) && !nios2_valid_target_attribute_rec (TREE_VALUE (args))) ret = false; return ret; } else if (TREE_CODE (args) == STRING_CST) { char *argstr = ASTRDUP (TREE_STRING_POINTER (args)); while (argstr && *argstr != '\0') { bool no_opt = false, end_p = false; char *eq = NULL, *p; while (ISSPACE (*argstr)) argstr++; p = argstr; while (*p != '\0' && *p != ',') { if (!eq && *p == '=') eq = p; ++p; } if (*p == '\0') end_p = true; else *p = '\0'; if (eq) *eq = '\0'; if (!strncmp (argstr, "no-", 3)) { no_opt = true; argstr += 3; } if (!strncmp (argstr, "custom-fpu-cfg", 14)) { char *end_eq = p; if (no_opt) { error ("custom-fpu-cfg option does not support %"); return false; } if (!eq) { error ("custom-fpu-cfg option requires configuration" " argument"); return false; } /* Increment and skip whitespace. */ while (ISSPACE (*(++eq))) ; /* Decrement and skip to before any trailing whitespace. */ while (ISSPACE (*(--end_eq))) ; nios2_handle_custom_fpu_cfg (eq, end_eq + 1, true); } else if (!strncmp (argstr, "custom-", 7)) { int code = -1; unsigned int i; for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) if (!strncmp (argstr + 7, N2FPU_NAME (i), strlen (N2FPU_NAME (i)))) { /* Found insn. */ code = i; break; } if (code >= 0) { if (no_opt) { if (eq) { error ("% does not accept arguments", N2FPU_NAME (code)); return false; } /* Disable option by setting to -1. */ nios2_deregister_custom_code (N2FPU_N (code)); N2FPU_N (code) = -1; } else { char *t; if (eq) while (ISSPACE (*(++eq))) ; if (!eq || eq == p) { error ("% requires argument", N2FPU_NAME (code)); return false; } for (t = eq; t != p; ++t) { if (ISSPACE (*t)) continue; if (!ISDIGIT (*t)) { error ("`custom-%s=' argument requires " "numeric digits", N2FPU_NAME (code)); return false; } } /* Set option to argument. */ N2FPU_N (code) = atoi (eq); nios2_handle_custom_fpu_insn_option (code); } } else { error ("% is not recognised as FPU instruction", argstr + 7); return false; } } else { error ("%<%s%> is unknown", argstr); return false; } if (end_p) break; else argstr = p + 1; } return true; } else gcc_unreachable (); } /* Return a TARGET_OPTION_NODE tree of the target options listed or NULL. */ static tree nios2_valid_target_attribute_tree (tree args) { if (!nios2_valid_target_attribute_rec (args)) return NULL_TREE; nios2_custom_check_insns (); return build_target_option_node (&global_options); } /* Hook to validate attribute((target("string"))). */ static bool nios2_valid_target_attribute_p (tree fndecl, tree ARG_UNUSED (name), tree args, int ARG_UNUSED (flags)) { struct cl_target_option cur_target; bool ret = true; tree old_optimize = build_optimization_node (&global_options); tree new_target, new_optimize; tree func_optimize = DECL_FUNCTION_SPECIFIC_OPTIMIZATION (fndecl); /* If the function changed the optimization levels as well as setting target options, start with the optimizations specified. */ if (func_optimize && func_optimize != old_optimize) cl_optimization_restore (&global_options, TREE_OPTIMIZATION (func_optimize)); /* The target attributes may also change some optimization flags, so update the optimization options if necessary. */ cl_target_option_save (&cur_target, &global_options); new_target = nios2_valid_target_attribute_tree (args); new_optimize = build_optimization_node (&global_options); if (!new_target) ret = false; else if (fndecl) { DECL_FUNCTION_SPECIFIC_TARGET (fndecl) = new_target; if (old_optimize != new_optimize) DECL_FUNCTION_SPECIFIC_OPTIMIZATION (fndecl) = new_optimize; } cl_target_option_restore (&global_options, &cur_target); if (old_optimize != new_optimize) cl_optimization_restore (&global_options, TREE_OPTIMIZATION (old_optimize)); return ret; } /* Remember the last target of nios2_set_current_function. */ static GTY(()) tree nios2_previous_fndecl; /* Establish appropriate back-end context for processing the function FNDECL. The argument might be NULL to indicate processing at top level, outside of any function scope. */ static void nios2_set_current_function (tree fndecl) { tree old_tree = (nios2_previous_fndecl ? DECL_FUNCTION_SPECIFIC_TARGET (nios2_previous_fndecl) : NULL_TREE); tree new_tree = (fndecl ? DECL_FUNCTION_SPECIFIC_TARGET (fndecl) : NULL_TREE); if (fndecl && fndecl != nios2_previous_fndecl) { nios2_previous_fndecl = fndecl; if (old_tree == new_tree) ; else if (new_tree) { cl_target_option_restore (&global_options, TREE_TARGET_OPTION (new_tree)); target_reinit (); } else if (old_tree) { struct cl_target_option *def = TREE_TARGET_OPTION (target_option_current_node); cl_target_option_restore (&global_options, def); target_reinit (); } } } /* Hook to validate the current #pragma GCC target and set the FPU custom code option state. If ARGS is NULL, then POP_TARGET is used to reset the options. */ static bool nios2_pragma_target_parse (tree args, tree pop_target) { tree cur_tree; if (! args) { cur_tree = ((pop_target) ? pop_target : target_option_default_node); cl_target_option_restore (&global_options, TREE_TARGET_OPTION (cur_tree)); } else { cur_tree = nios2_valid_target_attribute_tree (args); if (!cur_tree) return false; } target_option_current_node = cur_tree; return true; } /* Implement TARGET_MERGE_DECL_ATTRIBUTES. We are just using this hook to add some additional error checking to the default behavior. GCC does not provide a target hook for merging the target options, and only correctly handles merging empty vs non-empty option data; see merge_decls() in c-decl.c. So here we require either that at least one of the decls has empty target options, or that the target options/data be identical. */ static tree nios2_merge_decl_attributes (tree olddecl, tree newdecl) { tree oldopts = lookup_attribute ("target", DECL_ATTRIBUTES (olddecl)); tree newopts = lookup_attribute ("target", DECL_ATTRIBUTES (newdecl)); if (newopts && oldopts && newopts != oldopts) { tree oldtree = DECL_FUNCTION_SPECIFIC_TARGET (olddecl); tree newtree = DECL_FUNCTION_SPECIFIC_TARGET (newdecl); if (oldtree && newtree && oldtree != newtree) { struct cl_target_option *olddata = TREE_TARGET_OPTION (oldtree); struct cl_target_option *newdata = TREE_TARGET_OPTION (newtree); if (olddata != newdata && memcmp (olddata, newdata, sizeof (struct cl_target_option))) error ("%qE redeclared with conflicting %qs attributes", DECL_NAME (newdecl), "target"); } } return merge_attributes (DECL_ATTRIBUTES (olddecl), DECL_ATTRIBUTES (newdecl)); } /* Implement TARGET_ASM_OUTPUT_MI_THUNK. */ static void nios2_asm_output_mi_thunk (FILE *file, tree thunk_fndecl ATTRIBUTE_UNUSED, HOST_WIDE_INT delta, HOST_WIDE_INT vcall_offset, tree function) { rtx this_rtx, funexp; rtx_insn *insn; /* Pretend to be a post-reload pass while generating rtl. */ reload_completed = 1; if (flag_pic) nios2_load_pic_register (); /* Mark the end of the (empty) prologue. */ emit_note (NOTE_INSN_PROLOGUE_END); /* Find the "this" pointer. If the function returns a structure, the structure return pointer is in $5. */ if (aggregate_value_p (TREE_TYPE (TREE_TYPE (function)), function)) this_rtx = gen_rtx_REG (Pmode, FIRST_ARG_REGNO + 1); else this_rtx = gen_rtx_REG (Pmode, FIRST_ARG_REGNO); /* Add DELTA to THIS_RTX. */ nios2_emit_add_constant (this_rtx, delta); /* If needed, add *(*THIS_RTX + VCALL_OFFSET) to THIS_RTX. */ if (vcall_offset) { rtx tmp; tmp = gen_rtx_REG (Pmode, 2); emit_move_insn (tmp, gen_rtx_MEM (Pmode, this_rtx)); nios2_emit_add_constant (tmp, vcall_offset); emit_move_insn (tmp, gen_rtx_MEM (Pmode, tmp)); emit_insn (gen_add2_insn (this_rtx, tmp)); } /* Generate a tail call to the target function. */ if (!TREE_USED (function)) { assemble_external (function); TREE_USED (function) = 1; } funexp = XEXP (DECL_RTL (function), 0); /* Function address needs to be constructed under PIC, provide r2 to use here. */ nios2_adjust_call_address (&funexp, gen_rtx_REG (Pmode, 2)); insn = emit_call_insn (gen_sibcall_internal (funexp, const0_rtx)); SIBLING_CALL_P (insn) = 1; /* Run just enough of rest_of_compilation to get the insns emitted. There's not really enough bulk here to make other passes such as instruction scheduling worth while. Note that use_thunk calls assemble_start_function and assemble_end_function. */ insn = get_insns (); shorten_branches (insn); final_start_function (insn, file, 1); final (insn, file, 1); final_end_function (); /* Stop pretending to be a post-reload pass. */ reload_completed = 0; } /* Utility function to break a memory address into base register + constant offset. Return false if something unexpected is seen. */ static bool split_mem_address (rtx addr, rtx *base_reg, rtx *offset) { if (REG_P (addr)) { *base_reg = addr; *offset = const0_rtx; return true; } else if (GET_CODE (addr) == PLUS) { *base_reg = XEXP (addr, 0); *offset = XEXP (addr, 1); return true; } return false; } /* Splits out the operands of an ALU insn, places them in *LHS, *RHS1, *RHS2. */ static void split_alu_insn (rtx_insn *insn, rtx *lhs, rtx *rhs1, rtx *rhs2) { rtx pat = PATTERN (insn); gcc_assert (GET_CODE (pat) == SET); *lhs = SET_DEST (pat); *rhs1 = XEXP (SET_SRC (pat), 0); if (GET_RTX_CLASS (GET_CODE (SET_SRC (pat))) != RTX_UNARY) *rhs2 = XEXP (SET_SRC (pat), 1); return; } /* Returns true if OP is a REG and assigned a CDX reg. */ static bool cdxreg (rtx op) { return REG_P (op) && (!reload_completed || CDX_REG_P (REGNO (op))); } /* Returns true if OP is within range of CDX addi.n immediates. */ static bool cdx_add_immed (rtx op) { if (CONST_INT_P (op)) { HOST_WIDE_INT ival = INTVAL (op); return ival <= 128 && ival > 0 && (ival & (ival - 1)) == 0; } return false; } /* Returns true if OP is within range of CDX andi.n immediates. */ static bool cdx_and_immed (rtx op) { if (CONST_INT_P (op)) { HOST_WIDE_INT ival = INTVAL (op); return (ival == 1 || ival == 2 || ival == 3 || ival == 4 || ival == 8 || ival == 0xf || ival == 0x10 || ival == 0x1f || ival == 0x20 || ival == 0x3f || ival == 0x7f || ival == 0x80 || ival == 0xff || ival == 0x7ff || ival == 0xff00 || ival == 0xffff); } return false; } /* Returns true if OP is within range of CDX movi.n immediates. */ static bool cdx_mov_immed (rtx op) { if (CONST_INT_P (op)) { HOST_WIDE_INT ival = INTVAL (op); return ((ival >= 0 && ival <= 124) || ival == 0xff || ival == -2 || ival == -1); } return false; } /* Returns true if OP is within range of CDX slli.n/srli.n immediates. */ static bool cdx_shift_immed (rtx op) { if (CONST_INT_P (op)) { HOST_WIDE_INT ival = INTVAL (op); return (ival == 1 || ival == 2 || ival == 3 || ival == 8 || ival == 12 || ival == 16 || ival == 24 || ival == 31); } return false; } /* Classification of different kinds of add instructions. */ enum nios2_add_insn_kind { nios2_add_n_kind, nios2_addi_n_kind, nios2_subi_n_kind, nios2_spaddi_n_kind, nios2_spinci_n_kind, nios2_spdeci_n_kind, nios2_add_kind, nios2_addi_kind }; static const char *nios2_add_insn_names[] = { "add.n", "addi.n", "subi.n", "spaddi.n", "spinci.n", "spdeci.n", "add", "addi" }; static bool nios2_add_insn_narrow[] = { true, true, true, true, true, true, false, false}; /* Function to classify kinds of add instruction patterns. */ static enum nios2_add_insn_kind nios2_add_insn_classify (rtx_insn *insn ATTRIBUTE_UNUSED, rtx lhs, rtx rhs1, rtx rhs2) { if (TARGET_HAS_CDX) { if (cdxreg (lhs) && cdxreg (rhs1)) { if (cdxreg (rhs2)) return nios2_add_n_kind; if (CONST_INT_P (rhs2)) { HOST_WIDE_INT ival = INTVAL (rhs2); if (ival > 0 && cdx_add_immed (rhs2)) return nios2_addi_n_kind; if (ival < 0 && cdx_add_immed (GEN_INT (-ival))) return nios2_subi_n_kind; } } else if (rhs1 == stack_pointer_rtx && CONST_INT_P (rhs2)) { HOST_WIDE_INT imm7 = INTVAL (rhs2) >> 2; HOST_WIDE_INT rem = INTVAL (rhs2) & 3; if (rem == 0 && (imm7 & ~0x7f) == 0) { if (cdxreg (lhs)) return nios2_spaddi_n_kind; if (lhs == stack_pointer_rtx) return nios2_spinci_n_kind; } imm7 = -INTVAL(rhs2) >> 2; rem = -INTVAL (rhs2) & 3; if (lhs == stack_pointer_rtx && rem == 0 && (imm7 & ~0x7f) == 0) return nios2_spdeci_n_kind; } } return ((REG_P (rhs2) || rhs2 == const0_rtx) ? nios2_add_kind : nios2_addi_kind); } /* Emit assembly language for the different kinds of add instructions. */ const char* nios2_add_insn_asm (rtx_insn *insn, rtx *operands) { static char buf[256]; int ln = 256; enum nios2_add_insn_kind kind = nios2_add_insn_classify (insn, operands[0], operands[1], operands[2]); if (kind == nios2_subi_n_kind) snprintf (buf, ln, "subi.n\t%%0, %%1, %d", (int) -INTVAL (operands[2])); else if (kind == nios2_spaddi_n_kind) snprintf (buf, ln, "spaddi.n\t%%0, %%2"); else if (kind == nios2_spinci_n_kind) snprintf (buf, ln, "spinci.n\t%%2"); else if (kind == nios2_spdeci_n_kind) snprintf (buf, ln, "spdeci.n\t%d", (int) -INTVAL (operands[2])); else snprintf (buf, ln, "%s\t%%0, %%1, %%z2", nios2_add_insn_names[(int)kind]); return buf; } /* This routine, which the default "length" attribute computation is based on, encapsulates information about all the cases where CDX provides a narrow 2-byte instruction form. */ bool nios2_cdx_narrow_form_p (rtx_insn *insn) { rtx pat, lhs, rhs1, rhs2; enum attr_type type; if (!TARGET_HAS_CDX) return false; type = get_attr_type (insn); pat = PATTERN (insn); gcc_assert (reload_completed); switch (type) { case TYPE_CONTROL: if (GET_CODE (pat) == SIMPLE_RETURN) return true; if (GET_CODE (pat) == PARALLEL) pat = XVECEXP (pat, 0, 0); if (GET_CODE (pat) == SET) pat = SET_SRC (pat); if (GET_CODE (pat) == IF_THEN_ELSE) { /* Conditional branch patterns; for these we only check the comparison to find beqz.n/bnez.n cases. For the 'nios2_cbranch' pattern, we cannot also check the branch range here. That will be done at the md pattern "length" attribute computation. */ rtx cmp = XEXP (pat, 0); return ((GET_CODE (cmp) == EQ || GET_CODE (cmp) == NE) && cdxreg (XEXP (cmp, 0)) && XEXP (cmp, 1) == const0_rtx); } if (GET_CODE (pat) == TRAP_IF) /* trap.n is always usable. */ return true; if (GET_CODE (pat) == CALL) pat = XEXP (XEXP (pat, 0), 0); if (REG_P (pat)) /* Control instructions taking a register operand are indirect jumps and calls. The CDX instructions have a 5-bit register field so any reg is valid. */ return true; else { gcc_assert (!insn_variable_length_p (insn)); return false; } case TYPE_ADD: { enum nios2_add_insn_kind kind; split_alu_insn (insn, &lhs, &rhs1, &rhs2); kind = nios2_add_insn_classify (insn, lhs, rhs1, rhs2); return nios2_add_insn_narrow[(int)kind]; } case TYPE_LD: { bool ret; HOST_WIDE_INT offset, rem = 0; rtx addr, reg = SET_DEST (pat), mem = SET_SRC (pat); if (GET_CODE (mem) == SIGN_EXTEND) /* No CDX form for sign-extended load. */ return false; if (GET_CODE (mem) == ZERO_EXTEND) /* The load alternatives in the zero_extend* patterns. */ mem = XEXP (mem, 0); if (MEM_P (mem)) { /* ldxio. */ if ((MEM_VOLATILE_P (mem) && TARGET_BYPASS_CACHE_VOLATILE) || TARGET_BYPASS_CACHE) return false; addr = XEXP (mem, 0); /* GP-based references are never narrow. */ if (gprel_constant_p (addr)) return false; ret = split_mem_address (addr, &rhs1, &rhs2); gcc_assert (ret); } else return false; offset = INTVAL (rhs2); if (GET_MODE (mem) == SImode) { rem = offset & 3; offset >>= 2; /* ldwsp.n case. */ if (rtx_equal_p (rhs1, stack_pointer_rtx) && rem == 0 && (offset & ~0x1f) == 0) return true; } else if (GET_MODE (mem) == HImode) { rem = offset & 1; offset >>= 1; } /* ldbu.n, ldhu.n, ldw.n cases. */ return (cdxreg (reg) && cdxreg (rhs1) && rem == 0 && (offset & ~0xf) == 0); } case TYPE_ST: if (GET_CODE (pat) == PARALLEL) /* stex, stsex. */ return false; else { bool ret; HOST_WIDE_INT offset, rem = 0; rtx addr, reg = SET_SRC (pat), mem = SET_DEST (pat); if (!MEM_P (mem)) return false; /* stxio. */ if ((MEM_VOLATILE_P (mem) && TARGET_BYPASS_CACHE_VOLATILE) || TARGET_BYPASS_CACHE) return false; addr = XEXP (mem, 0); /* GP-based references are never narrow. */ if (gprel_constant_p (addr)) return false; ret = split_mem_address (addr, &rhs1, &rhs2); gcc_assert (ret); offset = INTVAL (rhs2); if (GET_MODE (mem) == SImode) { rem = offset & 3; offset >>= 2; /* stwsp.n case. */ if (rtx_equal_p (rhs1, stack_pointer_rtx) && rem == 0 && (offset & ~0x1f) == 0) return true; /* stwz.n case. */ else if (reg == const0_rtx && cdxreg (rhs1) && rem == 0 && (offset & ~0x3f) == 0) return true; } else if (GET_MODE (mem) == HImode) { rem = offset & 1; offset >>= 1; } else { gcc_assert (GET_MODE (mem) == QImode); /* stbz.n case. */ if (reg == const0_rtx && cdxreg (rhs1) && (offset & ~0x3f) == 0) return true; } /* stbu.n, sthu.n, stw.n cases. */ return (cdxreg (reg) && cdxreg (rhs1) && rem == 0 && (offset & ~0xf) == 0); } case TYPE_MOV: lhs = SET_DEST (pat); rhs1 = SET_SRC (pat); if (CONST_INT_P (rhs1)) return (cdxreg (lhs) && cdx_mov_immed (rhs1)); gcc_assert (REG_P (lhs) && REG_P (rhs1)); return true; case TYPE_AND: /* Some zero_extend* alternatives are and insns. */ if (GET_CODE (SET_SRC (pat)) == ZERO_EXTEND) return (cdxreg (SET_DEST (pat)) && cdxreg (XEXP (SET_SRC (pat), 0))); split_alu_insn (insn, &lhs, &rhs1, &rhs2); if (CONST_INT_P (rhs2)) return (cdxreg (lhs) && cdxreg (rhs1) && cdx_and_immed (rhs2)); return (cdxreg (lhs) && cdxreg (rhs2) && (!reload_completed || rtx_equal_p (lhs, rhs1))); case TYPE_OR: case TYPE_XOR: /* Note the two-address limitation for CDX form. */ split_alu_insn (insn, &lhs, &rhs1, &rhs2); return (cdxreg (lhs) && cdxreg (rhs2) && (!reload_completed || rtx_equal_p (lhs, rhs1))); case TYPE_SUB: split_alu_insn (insn, &lhs, &rhs1, &rhs2); return (cdxreg (lhs) && cdxreg (rhs1) && cdxreg (rhs2)); case TYPE_NEG: case TYPE_NOT: split_alu_insn (insn, &lhs, &rhs1, NULL); return (cdxreg (lhs) && cdxreg (rhs1)); case TYPE_SLL: case TYPE_SRL: split_alu_insn (insn, &lhs, &rhs1, &rhs2); return (cdxreg (lhs) && ((cdxreg (rhs1) && cdx_shift_immed (rhs2)) || (cdxreg (rhs2) && (!reload_completed || rtx_equal_p (lhs, rhs1))))); case TYPE_NOP: case TYPE_PUSH: case TYPE_POP: return true; default: break; } return false; } /* Main function to implement the pop_operation predicate that check pop.n insn pattern integrity. The CDX pop.n patterns mostly hardcode the restored registers, so the main checking is for the SP offsets. */ bool pop_operation_p (rtx op) { int i; HOST_WIDE_INT last_offset = -1, len = XVECLEN (op, 0); rtx base_reg, offset; if (len < 3 /* At least has a return, SP-update, and RA restore. */ || GET_CODE (XVECEXP (op, 0, 0)) != RETURN || !base_reg_adjustment_p (XVECEXP (op, 0, 1), &base_reg, &offset) || !rtx_equal_p (base_reg, stack_pointer_rtx) || !CONST_INT_P (offset) || (INTVAL (offset) & 3) != 0) return false; for (i = len - 1; i > 1; i--) { rtx set = XVECEXP (op, 0, i); rtx curr_base_reg, curr_offset; if (GET_CODE (set) != SET || !MEM_P (SET_SRC (set)) || !split_mem_address (XEXP (SET_SRC (set), 0), &curr_base_reg, &curr_offset) || !rtx_equal_p (base_reg, curr_base_reg) || !CONST_INT_P (curr_offset)) return false; if (i == len - 1) { last_offset = INTVAL (curr_offset); if ((last_offset & 3) != 0 || last_offset > 60) return false; } else { last_offset += 4; if (INTVAL (curr_offset) != last_offset) return false; } } if (last_offset < 0 || last_offset + 4 != INTVAL (offset)) return false; return true; } /* Masks of registers that are valid for CDX ldwm/stwm instructions. The instruction can encode subsets drawn from either R2-R13 or R14-R23 + FP + RA. */ #define CDX_LDSTWM_VALID_REGS_0 0x00003ffc #define CDX_LDSTWM_VALID_REGS_1 0x90ffc000 static bool nios2_ldstwm_regset_p (unsigned int regno, unsigned int *regset) { if (*regset == 0) { if (CDX_LDSTWM_VALID_REGS_0 & (1 << regno)) *regset = CDX_LDSTWM_VALID_REGS_0; else if (CDX_LDSTWM_VALID_REGS_1 & (1 << regno)) *regset = CDX_LDSTWM_VALID_REGS_1; else return false; return true; } else return (*regset & (1 << regno)) != 0; } /* Main function to implement ldwm_operation/stwm_operation predicates that check ldwm/stwm insn pattern integrity. */ bool ldstwm_operation_p (rtx op, bool load_p) { int start, i, end = XVECLEN (op, 0) - 1, last_regno = -1; unsigned int regset = 0; rtx base_reg, offset; rtx first_elt = XVECEXP (op, 0, 0); bool inc_p = true; bool wb_p = base_reg_adjustment_p (first_elt, &base_reg, &offset); if (GET_CODE (XVECEXP (op, 0, end)) == RETURN) end--; start = wb_p ? 1 : 0; for (i = start; i <= end; i++) { int regno; rtx reg, mem, elt = XVECEXP (op, 0, i); /* Return early if not a SET at all. */ if (GET_CODE (elt) != SET) return false; reg = load_p ? SET_DEST (elt) : SET_SRC (elt); mem = load_p ? SET_SRC (elt) : SET_DEST (elt); if (!REG_P (reg) || !MEM_P (mem)) return false; regno = REGNO (reg); if (!nios2_ldstwm_regset_p (regno, ®set)) return false; /* If no writeback to determine direction, use offset of first MEM. */ if (wb_p) inc_p = INTVAL (offset) > 0; else if (i == start) { rtx first_base, first_offset; if (!split_mem_address (XEXP (mem, 0), &first_base, &first_offset)) return false; base_reg = first_base; inc_p = INTVAL (first_offset) >= 0; } /* Ensure that the base register is not loaded into. */ if (load_p && regno == (int) REGNO (base_reg)) return false; /* Check for register order inc/dec integrity. */ if (last_regno >= 0) { if (inc_p && last_regno >= regno) return false; if (!inc_p && last_regno <= regno) return false; } last_regno = regno; } return true; } /* Helper for nios2_ldst_parallel, for generating a parallel vector SET element. */ static rtx gen_ldst (bool load_p, int regno, rtx base_mem, int offset) { rtx reg = gen_rtx_REG (SImode, regno); rtx mem = adjust_address_nv (base_mem, SImode, offset); return gen_rtx_SET (load_p ? reg : mem, load_p ? mem : reg); } /* A general routine for creating the body RTL pattern of ldwm/stwm/push.n/pop.n insns. LOAD_P: true/false for load/store direction. REG_INC_P: whether registers are incrementing/decrementing in the *RTL vector* (not necessarily the order defined in the ISA specification). OFFSET_INC_P: Same as REG_INC_P, but for the memory offset order. BASE_MEM: starting MEM. BASE_UPDATE: amount to update base register; zero means no writeback. REGMASK: register mask to load/store. RET_P: true if to tag a (return) element at the end. Note that this routine does not do any checking. It's the job of the caller to do the right thing, and the insn patterns to do the safe-guarding. */ static rtx nios2_ldst_parallel (bool load_p, bool reg_inc_p, bool offset_inc_p, rtx base_mem, int base_update, unsigned HOST_WIDE_INT regmask, bool ret_p) { rtvec p; int regno, b = 0, i = 0, n = 0, len = popcount_hwi (regmask); if (ret_p) len++, i++, b++; if (base_update != 0) len++, i++; p = rtvec_alloc (len); for (regno = (reg_inc_p ? 0 : 31); regno != (reg_inc_p ? 32 : -1); regno += (reg_inc_p ? 1 : -1)) if ((regmask & (1 << regno)) != 0) { int offset = (offset_inc_p ? 4 : -4) * n++; RTVEC_ELT (p, i++) = gen_ldst (load_p, regno, base_mem, offset); } if (ret_p) RTVEC_ELT (p, 0) = ret_rtx; if (base_update != 0) { rtx reg, offset; if (!split_mem_address (XEXP (base_mem, 0), ®, &offset)) gcc_unreachable (); RTVEC_ELT (p, b) = gen_rtx_SET (reg, plus_constant (Pmode, reg, base_update)); } return gen_rtx_PARALLEL (VOIDmode, p); } /* CDX ldwm/stwm peephole optimization pattern related routines. */ /* Data structure and sorting function for ldwm/stwm peephole optimizers. */ struct ldstwm_operand { int offset; /* Offset from base register. */ rtx reg; /* Register to store at this offset. */ rtx mem; /* Original mem. */ bool bad; /* True if this load/store can't be combined. */ bool rewrite; /* True if we should rewrite using scratch. */ }; static int compare_ldstwm_operands (const void *arg1, const void *arg2) { const struct ldstwm_operand *op1 = (const struct ldstwm_operand *) arg1; const struct ldstwm_operand *op2 = (const struct ldstwm_operand *) arg2; if (op1->bad) return op2->bad ? 0 : 1; else if (op2->bad) return -1; else return op1->offset - op2->offset; } /* Helper function: return true if a load/store using REGNO with address BASEREG and offset OFFSET meets the constraints for a 2-byte CDX ldw.n, stw.n, ldwsp.n, or stwsp.n instruction. */ static bool can_use_cdx_ldstw (int regno, int basereg, int offset) { if (CDX_REG_P (regno) && CDX_REG_P (basereg) && (offset & 0x3) == 0 && 0 <= offset && offset < 0x40) return true; else if (basereg == SP_REGNO && offset >= 0 && offset < 0x80 && (offset & 0x3) == 0) return true; return false; } /* This function is called from peephole2 optimizers to try to merge a series of individual loads and stores into a ldwm or stwm. It can also rewrite addresses inside the individual loads and stores using a common base register using a scratch register and smaller offsets if that allows them to use CDX ldw.n or stw.n instructions instead of 4-byte loads or stores. N is the number of insns we are trying to merge. SCRATCH is non-null if there is a scratch register available. The OPERANDS array contains alternating REG (even) and MEM (odd) operands. */ bool gen_ldstwm_peep (bool load_p, int n, rtx scratch, rtx *operands) { /* CDX ldwm/stwm instructions allow a maximum of 12 registers to be specified. */ #define MAX_LDSTWM_OPS 12 struct ldstwm_operand sort[MAX_LDSTWM_OPS]; int basereg = -1; int baseoffset; int i, m, lastoffset, lastreg; unsigned int regmask = 0, usemask = 0, regset; bool needscratch; int newbasereg; int nbytes; if (!TARGET_HAS_CDX) return false; if (n < 2 || n > MAX_LDSTWM_OPS) return false; /* Check all the operands for validity and initialize the sort array. The places where we return false here are all situations that aren't expected to ever happen -- invalid patterns, invalid registers, etc. */ for (i = 0; i < n; i++) { rtx base, offset; rtx reg = operands[i]; rtx mem = operands[i + n]; int r, o, regno; bool bad = false; if (!REG_P (reg) || !MEM_P (mem)) return false; regno = REGNO (reg); if (regno > 31) return false; if (load_p && (regmask & (1 << regno)) != 0) return false; regmask |= 1 << regno; if (!split_mem_address (XEXP (mem, 0), &base, &offset)) return false; r = REGNO (base); o = INTVAL (offset); if (basereg == -1) basereg = r; else if (r != basereg) bad = true; usemask |= 1 << r; sort[i].bad = bad; sort[i].rewrite = false; sort[i].offset = o; sort[i].reg = reg; sort[i].mem = mem; } /* If we are doing a series of register loads, we can't safely reorder them if any of the regs used in addr expressions are also being set. */ if (load_p && (regmask & usemask)) return false; /* Sort the array by increasing mem offset order, then check that offsets are valid and register order matches mem order. At the end of this loop, m is the number of loads/stores we will try to combine; the rest are leftovers. */ qsort (sort, n, sizeof (struct ldstwm_operand), compare_ldstwm_operands); baseoffset = sort[0].offset; needscratch = baseoffset != 0; if (needscratch && !scratch) return false; lastreg = regmask = regset = 0; lastoffset = baseoffset; for (m = 0; m < n && !sort[m].bad; m++) { int thisreg = REGNO (sort[m].reg); if (sort[m].offset != lastoffset || (m > 0 && lastreg >= thisreg) || !nios2_ldstwm_regset_p (thisreg, ®set)) break; lastoffset += 4; lastreg = thisreg; regmask |= (1 << thisreg); } /* For loads, make sure we are not overwriting the scratch reg. The peephole2 pattern isn't supposed to match unless the register is unused all the way through, so this isn't supposed to happen anyway. */ if (load_p && needscratch && ((1 << REGNO (scratch)) & regmask) != 0) return false; newbasereg = needscratch ? (int) REGNO (scratch) : basereg; /* We may be able to combine only the first m of the n total loads/stores into a single instruction. If m < 2, there's no point in emitting a ldwm/stwm at all, but we might be able to do further optimizations if we have a scratch. We will count the instruction lengths of the old and new patterns and store the savings in nbytes. */ if (m < 2) { if (!needscratch) return false; m = 0; nbytes = 0; } else nbytes = -4; /* Size of ldwm/stwm. */ if (needscratch) { int bo = baseoffset > 0 ? baseoffset : -baseoffset; if (CDX_REG_P (newbasereg) && CDX_REG_P (basereg) && bo <= 128 && bo > 0 && (bo & (bo - 1)) == 0) nbytes -= 2; /* Size of addi.n/subi.n. */ else nbytes -= 4; /* Size of non-CDX addi. */ } /* Count the size of the input load/store instructions being replaced. */ for (i = 0; i < m; i++) if (can_use_cdx_ldstw (REGNO (sort[i].reg), basereg, sort[i].offset)) nbytes += 2; else nbytes += 4; /* We may also be able to save a bit if we can rewrite non-CDX load/stores that can't be combined into the ldwm/stwm into CDX load/stores using the scratch reg. For example, this might happen if baseoffset is large, by bringing in the offsets in the load/store instructions within the range that fits in the CDX instruction. */ if (needscratch && CDX_REG_P (newbasereg)) for (i = m; i < n && !sort[i].bad; i++) if (!can_use_cdx_ldstw (REGNO (sort[i].reg), basereg, sort[i].offset) && can_use_cdx_ldstw (REGNO (sort[i].reg), newbasereg, sort[i].offset - baseoffset)) { sort[i].rewrite = true; nbytes += 2; } /* Are we good to go? */ if (nbytes <= 0) return false; /* Emit the scratch load. */ if (needscratch) emit_insn (gen_rtx_SET (scratch, XEXP (sort[0].mem, 0))); /* Emit the ldwm/stwm insn. */ if (m > 0) { rtvec p = rtvec_alloc (m); for (i = 0; i < m; i++) { int offset = sort[i].offset; rtx mem, reg = sort[i].reg; rtx base_reg = gen_rtx_REG (Pmode, newbasereg); if (needscratch) offset -= baseoffset; mem = gen_rtx_MEM (SImode, plus_constant (Pmode, base_reg, offset)); if (load_p) RTVEC_ELT (p, i) = gen_rtx_SET (reg, mem); else RTVEC_ELT (p, i) = gen_rtx_SET (mem, reg); } emit_insn (gen_rtx_PARALLEL (VOIDmode, p)); } /* Emit any leftover load/stores as individual instructions, doing the previously-noted rewrites to use the scratch reg. */ for (i = m; i < n; i++) { rtx reg = sort[i].reg; rtx mem = sort[i].mem; if (sort[i].rewrite) { int offset = sort[i].offset - baseoffset; mem = gen_rtx_MEM (SImode, plus_constant (Pmode, scratch, offset)); } if (load_p) emit_move_insn (reg, mem); else emit_move_insn (mem, reg); } return true; } /* Implement TARGET_MACHINE_DEPENDENT_REORG: We use this hook when emitting CDX code to enforce the 4-byte alignment requirement for labels that are used as the targets of jmpi instructions. CDX code can otherwise contain a mix of 16-bit and 32-bit instructions aligned on any 16-bit boundary, but functions and jmpi labels have to be 32-bit aligned because of the way the address is encoded in the instruction. */ static unsigned char *label_align; static int min_labelno, max_labelno; static void nios2_reorg (void) { bool changed = true; rtx_insn *insn; if (!TARGET_HAS_CDX) return; /* Initialize the data structures. */ if (label_align) free (label_align); max_labelno = max_label_num (); min_labelno = get_first_label_num (); label_align = XCNEWVEC (unsigned char, max_labelno - min_labelno + 1); /* Iterate on inserting alignment and adjusting branch lengths until no more changes. */ while (changed) { changed = false; shorten_branches (get_insns ()); for (insn = get_insns (); insn != 0; insn = NEXT_INSN (insn)) if (JUMP_P (insn) && insn_variable_length_p (insn)) { rtx label = JUMP_LABEL (insn); /* We use the current fact that all cases of 'jmpi' doing the actual branch in the machine description has a computed length of 6 or 8. Length 4 and below are all PC-relative 'br' branches without the jump-align problem. */ if (label && LABEL_P (label) && get_attr_length (insn) > 4) { int index = CODE_LABEL_NUMBER (label) - min_labelno; if (label_align[index] != 2) { label_align[index] = 2; changed = true; } } } } } /* Implement LABEL_ALIGN, using the information gathered in nios2_reorg. */ int nios2_label_align (rtx label) { int n = CODE_LABEL_NUMBER (label); if (label_align && n >= min_labelno && n <= max_labelno) return MAX (label_align[n - min_labelno], align_labels_log); return align_labels_log; } /* Implement ADJUST_REG_ALLOC_ORDER. We use the default ordering for R1 and non-CDX R2 code; for CDX we tweak thing to prefer the registers that can be used as operands to instructions that have 3-bit register fields. */ void nios2_adjust_reg_alloc_order (void) { const int cdx_reg_alloc_order[] = { /* Call-clobbered GPRs within CDX 3-bit encoded range. */ 2, 3, 4, 5, 6, 7, /* Call-saved GPRs within CDX 3-bit encoded range. */ 16, 17, /* Other call-clobbered GPRs. */ 8, 9, 10, 11, 12, 13, 14, 15, /* Other call-saved GPRs. RA placed first since it is always saved. */ 31, 18, 19, 20, 21, 22, 23, 28, /* Fixed GPRs, not used by the register allocator. */ 0, 1, 24, 25, 26, 27, 29, 30, 32, 33, 34, 35, 36, 37, 38, 39 }; if (TARGET_HAS_CDX) memcpy (reg_alloc_order, cdx_reg_alloc_order, sizeof (int) * FIRST_PSEUDO_REGISTER); } /* Initialize the GCC target structure. */ #undef TARGET_ASM_FUNCTION_PROLOGUE #define TARGET_ASM_FUNCTION_PROLOGUE nios2_asm_function_prologue #undef TARGET_IN_SMALL_DATA_P #define TARGET_IN_SMALL_DATA_P nios2_in_small_data_p #undef TARGET_SECTION_TYPE_FLAGS #define TARGET_SECTION_TYPE_FLAGS nios2_section_type_flags #undef TARGET_INIT_BUILTINS #define TARGET_INIT_BUILTINS nios2_init_builtins #undef TARGET_EXPAND_BUILTIN #define TARGET_EXPAND_BUILTIN nios2_expand_builtin #undef TARGET_BUILTIN_DECL #define TARGET_BUILTIN_DECL nios2_builtin_decl #undef TARGET_FUNCTION_OK_FOR_SIBCALL #define TARGET_FUNCTION_OK_FOR_SIBCALL hook_bool_tree_tree_true #undef TARGET_CAN_ELIMINATE #define TARGET_CAN_ELIMINATE nios2_can_eliminate #undef TARGET_FUNCTION_ARG #define TARGET_FUNCTION_ARG nios2_function_arg #undef TARGET_FUNCTION_ARG_ADVANCE #define TARGET_FUNCTION_ARG_ADVANCE nios2_function_arg_advance #undef TARGET_ARG_PARTIAL_BYTES #define TARGET_ARG_PARTIAL_BYTES nios2_arg_partial_bytes #undef TARGET_TRAMPOLINE_INIT #define TARGET_TRAMPOLINE_INIT nios2_trampoline_init #undef TARGET_FUNCTION_VALUE #define TARGET_FUNCTION_VALUE nios2_function_value #undef TARGET_LIBCALL_VALUE #define TARGET_LIBCALL_VALUE nios2_libcall_value #undef TARGET_FUNCTION_VALUE_REGNO_P #define TARGET_FUNCTION_VALUE_REGNO_P nios2_function_value_regno_p #undef TARGET_RETURN_IN_MEMORY #define TARGET_RETURN_IN_MEMORY nios2_return_in_memory #undef TARGET_PROMOTE_PROTOTYPES #define TARGET_PROMOTE_PROTOTYPES hook_bool_const_tree_true #undef TARGET_SETUP_INCOMING_VARARGS #define TARGET_SETUP_INCOMING_VARARGS nios2_setup_incoming_varargs #undef TARGET_MUST_PASS_IN_STACK #define TARGET_MUST_PASS_IN_STACK must_pass_in_stack_var_size #undef TARGET_LEGITIMATE_CONSTANT_P #define TARGET_LEGITIMATE_CONSTANT_P nios2_legitimate_constant_p #undef TARGET_LEGITIMIZE_ADDRESS #define TARGET_LEGITIMIZE_ADDRESS nios2_legitimize_address #undef TARGET_DELEGITIMIZE_ADDRESS #define TARGET_DELEGITIMIZE_ADDRESS nios2_delegitimize_address #undef TARGET_LEGITIMATE_ADDRESS_P #define TARGET_LEGITIMATE_ADDRESS_P nios2_legitimate_address_p #undef TARGET_LRA_P #define TARGET_LRA_P hook_bool_void_false #undef TARGET_PREFERRED_RELOAD_CLASS #define TARGET_PREFERRED_RELOAD_CLASS nios2_preferred_reload_class #undef TARGET_RTX_COSTS #define TARGET_RTX_COSTS nios2_rtx_costs #undef TARGET_HAVE_TLS #define TARGET_HAVE_TLS TARGET_LINUX_ABI #undef TARGET_CANNOT_FORCE_CONST_MEM #define TARGET_CANNOT_FORCE_CONST_MEM nios2_cannot_force_const_mem #undef TARGET_ASM_OUTPUT_DWARF_DTPREL #define TARGET_ASM_OUTPUT_DWARF_DTPREL nios2_output_dwarf_dtprel #undef TARGET_PRINT_OPERAND_PUNCT_VALID_P #define TARGET_PRINT_OPERAND_PUNCT_VALID_P nios2_print_operand_punct_valid_p #undef TARGET_PRINT_OPERAND #define TARGET_PRINT_OPERAND nios2_print_operand #undef TARGET_PRINT_OPERAND_ADDRESS #define TARGET_PRINT_OPERAND_ADDRESS nios2_print_operand_address #undef TARGET_ASM_OUTPUT_ADDR_CONST_EXTRA #define TARGET_ASM_OUTPUT_ADDR_CONST_EXTRA nios2_output_addr_const_extra #undef TARGET_ASM_FILE_END #define TARGET_ASM_FILE_END nios2_asm_file_end #undef TARGET_OPTION_OVERRIDE #define TARGET_OPTION_OVERRIDE nios2_option_override #undef TARGET_OPTION_SAVE #define TARGET_OPTION_SAVE nios2_option_save #undef TARGET_OPTION_RESTORE #define TARGET_OPTION_RESTORE nios2_option_restore #undef TARGET_SET_CURRENT_FUNCTION #define TARGET_SET_CURRENT_FUNCTION nios2_set_current_function #undef TARGET_OPTION_VALID_ATTRIBUTE_P #define TARGET_OPTION_VALID_ATTRIBUTE_P nios2_valid_target_attribute_p #undef TARGET_OPTION_PRAGMA_PARSE #define TARGET_OPTION_PRAGMA_PARSE nios2_pragma_target_parse #undef TARGET_MERGE_DECL_ATTRIBUTES #define TARGET_MERGE_DECL_ATTRIBUTES nios2_merge_decl_attributes #undef TARGET_ASM_CAN_OUTPUT_MI_THUNK #define TARGET_ASM_CAN_OUTPUT_MI_THUNK \ hook_bool_const_tree_hwi_hwi_const_tree_true #undef TARGET_ASM_OUTPUT_MI_THUNK #define TARGET_ASM_OUTPUT_MI_THUNK nios2_asm_output_mi_thunk #undef TARGET_MACHINE_DEPENDENT_REORG #define TARGET_MACHINE_DEPENDENT_REORG nios2_reorg struct gcc_target targetm = TARGET_INITIALIZER; #include "gt-nios2.h"