| /* Target machine subroutines for Altera Nios II. |
| Copyright (C) 2012-2014 Free Software Foundation, Inc. |
| Contributed by Jonah Graham (jgraham@altera.com), |
| Will Reece (wreece@altera.com), and Jeff DaSilva (jdasilva@altera.com). |
| Contributed by Mentor Graphics, Inc. |
| |
| This file is part of GCC. |
| |
| GCC is free software; you can redistribute it and/or modify it |
| under the terms of the GNU General Public License as published |
| by the Free Software Foundation; either version 3, or (at your |
| option) any later version. |
| |
| GCC is distributed in the hope that it will be useful, but WITHOUT |
| ANY WARRANTY; without even the implied warranty of MERCHANTABILITY |
| or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public |
| License for more details. |
| |
| You should have received a copy of the GNU General Public License |
| along with GCC; see the file COPYING3. If not see |
| <http://www.gnu.org/licenses/>. */ |
| |
| #include "config.h" |
| #include "system.h" |
| #include "coretypes.h" |
| #include "tm.h" |
| #include "rtl.h" |
| #include "tree.h" |
| #include "regs.h" |
| #include "hard-reg-set.h" |
| #include "insn-config.h" |
| #include "conditions.h" |
| #include "output.h" |
| #include "insn-attr.h" |
| #include "flags.h" |
| #include "recog.h" |
| #include "expr.h" |
| #include "optabs.h" |
| #include "function.h" |
| #include "ggc.h" |
| #include "basic-block.h" |
| #include "diagnostic-core.h" |
| #include "toplev.h" |
| #include "target.h" |
| #include "target-def.h" |
| #include "tm_p.h" |
| #include "langhooks.h" |
| #include "df.h" |
| #include "debug.h" |
| #include "real.h" |
| #include "reload.h" |
| #include "stor-layout.h" |
| #include "varasm.h" |
| #include "calls.h" |
| |
| /* Forward function declarations. */ |
| static bool prologue_saved_reg_p (unsigned); |
| static void nios2_load_pic_register (void); |
| static void nios2_register_custom_code (unsigned int, enum nios2_ccs_code, int); |
| static const char *nios2_unspec_reloc_name (int); |
| static void nios2_register_builtin_fndecl (unsigned, tree); |
| |
| /* Threshold for data being put into the small data/bss area, instead |
| of the normal data area (references to the small data/bss area take |
| 1 instruction, and use the global pointer, references to the normal |
| data area takes 2 instructions). */ |
| unsigned HOST_WIDE_INT nios2_section_threshold = NIOS2_DEFAULT_GVALUE; |
| |
| struct GTY (()) machine_function |
| { |
| /* Current frame information, to be filled in by nios2_compute_frame_layout |
| with register save masks, and offsets for the current function. */ |
| |
| /* Mask of registers to save. */ |
| unsigned int save_mask; |
| /* Number of bytes that the entire frame takes up. */ |
| int total_size; |
| /* Number of bytes that variables take up. */ |
| int var_size; |
| /* Number of bytes that outgoing arguments take up. */ |
| int args_size; |
| /* Number of bytes needed to store registers in frame. */ |
| int save_reg_size; |
| /* Offset from new stack pointer to store registers. */ |
| int save_regs_offset; |
| /* Offset from save_regs_offset to store frame pointer register. */ |
| int fp_save_offset; |
| /* != 0 if frame layout already calculated. */ |
| int initialized; |
| }; |
| |
| /* State to track the assignment of custom codes to FPU/custom builtins. */ |
| static enum nios2_ccs_code custom_code_status[256]; |
| static int custom_code_index[256]; |
| /* Set to true if any conflicts (re-use of a code between 0-255) are found. */ |
| static bool custom_code_conflict = false; |
| |
| |
| /* Definition of builtin function types for nios2. */ |
| |
| #define N2_FTYPES \ |
| N2_FTYPE(1, (SF)) \ |
| N2_FTYPE(1, (VOID)) \ |
| N2_FTYPE(2, (DF, DF)) \ |
| N2_FTYPE(3, (DF, DF, DF)) \ |
| N2_FTYPE(2, (DF, SF)) \ |
| N2_FTYPE(2, (DF, SI)) \ |
| N2_FTYPE(2, (DF, UI)) \ |
| N2_FTYPE(2, (SF, DF)) \ |
| N2_FTYPE(2, (SF, SF)) \ |
| N2_FTYPE(3, (SF, SF, SF)) \ |
| N2_FTYPE(2, (SF, SI)) \ |
| N2_FTYPE(2, (SF, UI)) \ |
| N2_FTYPE(2, (SI, CVPTR)) \ |
| N2_FTYPE(2, (SI, DF)) \ |
| N2_FTYPE(3, (SI, DF, DF)) \ |
| N2_FTYPE(2, (SI, SF)) \ |
| N2_FTYPE(3, (SI, SF, SF)) \ |
| N2_FTYPE(2, (SI, SI)) \ |
| N2_FTYPE(2, (UI, CVPTR)) \ |
| N2_FTYPE(2, (UI, DF)) \ |
| N2_FTYPE(2, (UI, SF)) \ |
| N2_FTYPE(2, (VOID, DF)) \ |
| N2_FTYPE(2, (VOID, SF)) \ |
| N2_FTYPE(3, (VOID, SI, SI)) \ |
| N2_FTYPE(3, (VOID, VPTR, SI)) |
| |
| #define N2_FTYPE_OP1(R) N2_FTYPE_ ## R ## _VOID |
| #define N2_FTYPE_OP2(R, A1) N2_FTYPE_ ## R ## _ ## A1 |
| #define N2_FTYPE_OP3(R, A1, A2) N2_FTYPE_ ## R ## _ ## A1 ## _ ## A2 |
| |
| /* Expand ftcode enumeration. */ |
| enum nios2_ftcode { |
| #define N2_FTYPE(N,ARGS) N2_FTYPE_OP ## N ARGS, |
| N2_FTYPES |
| #undef N2_FTYPE |
| N2_FTYPE_MAX |
| }; |
| |
| /* Return the tree function type, based on the ftcode. */ |
| static tree |
| nios2_ftype (enum nios2_ftcode ftcode) |
| { |
| static tree types[(int) N2_FTYPE_MAX]; |
| |
| tree N2_TYPE_SF = float_type_node; |
| tree N2_TYPE_DF = double_type_node; |
| tree N2_TYPE_SI = integer_type_node; |
| tree N2_TYPE_UI = unsigned_type_node; |
| tree N2_TYPE_VOID = void_type_node; |
| |
| static const_tree N2_TYPE_CVPTR, N2_TYPE_VPTR; |
| if (!N2_TYPE_CVPTR) |
| { |
| /* const volatile void *. */ |
| N2_TYPE_CVPTR |
| = build_pointer_type (build_qualified_type (void_type_node, |
| (TYPE_QUAL_CONST |
| | TYPE_QUAL_VOLATILE))); |
| /* volatile void *. */ |
| N2_TYPE_VPTR |
| = build_pointer_type (build_qualified_type (void_type_node, |
| TYPE_QUAL_VOLATILE)); |
| } |
| if (types[(int) ftcode] == NULL_TREE) |
| switch (ftcode) |
| { |
| #define N2_FTYPE_ARGS1(R) N2_TYPE_ ## R |
| #define N2_FTYPE_ARGS2(R,A1) N2_TYPE_ ## R, N2_TYPE_ ## A1 |
| #define N2_FTYPE_ARGS3(R,A1,A2) N2_TYPE_ ## R, N2_TYPE_ ## A1, N2_TYPE_ ## A2 |
| #define N2_FTYPE(N,ARGS) \ |
| case N2_FTYPE_OP ## N ARGS: \ |
| types[(int) ftcode] \ |
| = build_function_type_list (N2_FTYPE_ARGS ## N ARGS, NULL_TREE); \ |
| break; |
| N2_FTYPES |
| #undef N2_FTYPE |
| default: gcc_unreachable (); |
| } |
| return types[(int) ftcode]; |
| } |
| |
| |
| /* Definition of FPU instruction descriptions. */ |
| |
| struct nios2_fpu_insn_info |
| { |
| const char *name; |
| int num_operands, *optvar; |
| int opt, no_opt; |
| #define N2F_DF 0x1 |
| #define N2F_DFREQ 0x2 |
| #define N2F_UNSAFE 0x4 |
| #define N2F_FINITE 0x8 |
| unsigned int flags; |
| enum insn_code icode; |
| enum nios2_ftcode ftcode; |
| }; |
| |
| /* Base macro for defining FPU instructions. */ |
| #define N2FPU_INSN_DEF_BASE(insn, nop, flags, icode, args) \ |
| { #insn, nop, &nios2_custom_ ## insn, OPT_mcustom_##insn##_, \ |
| OPT_mno_custom_##insn, flags, CODE_FOR_ ## icode, \ |
| N2_FTYPE_OP ## nop args } |
| |
| /* Arithmetic and math functions; 2 or 3 operand FP operations. */ |
| #define N2FPU_OP2(mode) (mode, mode) |
| #define N2FPU_OP3(mode) (mode, mode, mode) |
| #define N2FPU_INSN_DEF(code, icode, nop, flags, m, M) \ |
| N2FPU_INSN_DEF_BASE (f ## code ## m, nop, flags, \ |
| icode ## m ## f ## nop, N2FPU_OP ## nop (M ## F)) |
| #define N2FPU_INSN_SF(code, nop, flags) \ |
| N2FPU_INSN_DEF (code, code, nop, flags, s, S) |
| #define N2FPU_INSN_DF(code, nop, flags) \ |
| N2FPU_INSN_DEF (code, code, nop, flags | N2F_DF, d, D) |
| |
| /* Compare instructions, 3 operand FP operation with a SI result. */ |
| #define N2FPU_CMP_DEF(code, flags, m, M) \ |
| N2FPU_INSN_DEF_BASE (fcmp ## code ## m, 3, flags, \ |
| nios2_s ## code ## m ## f, (SI, M ## F, M ## F)) |
| #define N2FPU_CMP_SF(code) N2FPU_CMP_DEF (code, 0, s, S) |
| #define N2FPU_CMP_DF(code) N2FPU_CMP_DEF (code, N2F_DF, d, D) |
| |
| /* The order of definition needs to be maintained consistent with |
| enum n2fpu_code in nios2-opts.h. */ |
| struct nios2_fpu_insn_info nios2_fpu_insn[] = |
| { |
| /* Single precision instructions. */ |
| N2FPU_INSN_SF (add, 3, 0), |
| N2FPU_INSN_SF (sub, 3, 0), |
| N2FPU_INSN_SF (mul, 3, 0), |
| N2FPU_INSN_SF (div, 3, 0), |
| /* Due to textual difference between min/max and smin/smax. */ |
| N2FPU_INSN_DEF (min, smin, 3, N2F_FINITE, s, S), |
| N2FPU_INSN_DEF (max, smax, 3, N2F_FINITE, s, S), |
| N2FPU_INSN_SF (neg, 2, 0), |
| N2FPU_INSN_SF (abs, 2, 0), |
| N2FPU_INSN_SF (sqrt, 2, 0), |
| N2FPU_INSN_SF (sin, 2, N2F_UNSAFE), |
| N2FPU_INSN_SF (cos, 2, N2F_UNSAFE), |
| N2FPU_INSN_SF (tan, 2, N2F_UNSAFE), |
| N2FPU_INSN_SF (atan, 2, N2F_UNSAFE), |
| N2FPU_INSN_SF (exp, 2, N2F_UNSAFE), |
| N2FPU_INSN_SF (log, 2, N2F_UNSAFE), |
| /* Single precision compares. */ |
| N2FPU_CMP_SF (eq), N2FPU_CMP_SF (ne), |
| N2FPU_CMP_SF (lt), N2FPU_CMP_SF (le), |
| N2FPU_CMP_SF (gt), N2FPU_CMP_SF (ge), |
| |
| /* Double precision instructions. */ |
| N2FPU_INSN_DF (add, 3, 0), |
| N2FPU_INSN_DF (sub, 3, 0), |
| N2FPU_INSN_DF (mul, 3, 0), |
| N2FPU_INSN_DF (div, 3, 0), |
| /* Due to textual difference between min/max and smin/smax. */ |
| N2FPU_INSN_DEF (min, smin, 3, N2F_FINITE, d, D), |
| N2FPU_INSN_DEF (max, smax, 3, N2F_FINITE, d, D), |
| N2FPU_INSN_DF (neg, 2, 0), |
| N2FPU_INSN_DF (abs, 2, 0), |
| N2FPU_INSN_DF (sqrt, 2, 0), |
| N2FPU_INSN_DF (sin, 2, N2F_UNSAFE), |
| N2FPU_INSN_DF (cos, 2, N2F_UNSAFE), |
| N2FPU_INSN_DF (tan, 2, N2F_UNSAFE), |
| N2FPU_INSN_DF (atan, 2, N2F_UNSAFE), |
| N2FPU_INSN_DF (exp, 2, N2F_UNSAFE), |
| N2FPU_INSN_DF (log, 2, N2F_UNSAFE), |
| /* Double precision compares. */ |
| N2FPU_CMP_DF (eq), N2FPU_CMP_DF (ne), |
| N2FPU_CMP_DF (lt), N2FPU_CMP_DF (le), |
| N2FPU_CMP_DF (gt), N2FPU_CMP_DF (ge), |
| |
| /* Conversion instructions. */ |
| N2FPU_INSN_DEF_BASE (floatis, 2, 0, floatsisf2, (SF, SI)), |
| N2FPU_INSN_DEF_BASE (floatus, 2, 0, floatunssisf2, (SF, UI)), |
| N2FPU_INSN_DEF_BASE (floatid, 2, 0, floatsidf2, (DF, SI)), |
| N2FPU_INSN_DEF_BASE (floatud, 2, 0, floatunssidf2, (DF, UI)), |
| N2FPU_INSN_DEF_BASE (fixsi, 2, 0, fix_truncsfsi2, (SI, SF)), |
| N2FPU_INSN_DEF_BASE (fixsu, 2, 0, fixuns_truncsfsi2, (UI, SF)), |
| N2FPU_INSN_DEF_BASE (fixdi, 2, 0, fix_truncdfsi2, (SI, DF)), |
| N2FPU_INSN_DEF_BASE (fixdu, 2, 0, fixuns_truncdfsi2, (UI, DF)), |
| N2FPU_INSN_DEF_BASE (fextsd, 2, 0, extendsfdf2, (DF, SF)), |
| N2FPU_INSN_DEF_BASE (ftruncds, 2, 0, truncdfsf2, (SF, DF)), |
| |
| /* X, Y access instructions. */ |
| N2FPU_INSN_DEF_BASE (fwrx, 2, N2F_DFREQ, nios2_fwrx, (VOID, DF)), |
| N2FPU_INSN_DEF_BASE (fwry, 2, N2F_DFREQ, nios2_fwry, (VOID, SF)), |
| N2FPU_INSN_DEF_BASE (frdxlo, 1, N2F_DFREQ, nios2_frdxlo, (SF)), |
| N2FPU_INSN_DEF_BASE (frdxhi, 1, N2F_DFREQ, nios2_frdxhi, (SF)), |
| N2FPU_INSN_DEF_BASE (frdy, 1, N2F_DFREQ, nios2_frdy, (SF)) |
| }; |
| |
| /* Some macros for ease of access. */ |
| #define N2FPU(code) nios2_fpu_insn[(int) code] |
| #define N2FPU_ENABLED_P(code) (N2FPU_N(code) >= 0) |
| #define N2FPU_N(code) (*N2FPU(code).optvar) |
| #define N2FPU_NAME(code) (N2FPU(code).name) |
| #define N2FPU_ICODE(code) (N2FPU(code).icode) |
| #define N2FPU_FTCODE(code) (N2FPU(code).ftcode) |
| #define N2FPU_FINITE_P(code) (N2FPU(code).flags & N2F_FINITE) |
| #define N2FPU_UNSAFE_P(code) (N2FPU(code).flags & N2F_UNSAFE) |
| #define N2FPU_DOUBLE_P(code) (N2FPU(code).flags & N2F_DF) |
| #define N2FPU_DOUBLE_REQUIRED_P(code) (N2FPU(code).flags & N2F_DFREQ) |
| |
| /* Same as above, but for cases where using only the op part is shorter. */ |
| #define N2FPU_OP(op) N2FPU(n2fpu_ ## op) |
| #define N2FPU_OP_NAME(op) N2FPU_NAME(n2fpu_ ## op) |
| #define N2FPU_OP_ENABLED_P(op) N2FPU_ENABLED_P(n2fpu_ ## op) |
| |
| /* Export the FPU insn enabled predicate to nios2.md. */ |
| bool |
| nios2_fpu_insn_enabled (enum n2fpu_code code) |
| { |
| return N2FPU_ENABLED_P (code); |
| } |
| |
| /* Return true if COND comparison for mode MODE is enabled under current |
| settings. */ |
| |
| static bool |
| nios2_fpu_compare_enabled (enum rtx_code cond, enum machine_mode mode) |
| { |
| if (mode == SFmode) |
| switch (cond) |
| { |
| case EQ: return N2FPU_OP_ENABLED_P (fcmpeqs); |
| case NE: return N2FPU_OP_ENABLED_P (fcmpnes); |
| case GT: return N2FPU_OP_ENABLED_P (fcmpgts); |
| case GE: return N2FPU_OP_ENABLED_P (fcmpges); |
| case LT: return N2FPU_OP_ENABLED_P (fcmplts); |
| case LE: return N2FPU_OP_ENABLED_P (fcmples); |
| default: break; |
| } |
| else if (mode == DFmode) |
| switch (cond) |
| { |
| case EQ: return N2FPU_OP_ENABLED_P (fcmpeqd); |
| case NE: return N2FPU_OP_ENABLED_P (fcmpned); |
| case GT: return N2FPU_OP_ENABLED_P (fcmpgtd); |
| case GE: return N2FPU_OP_ENABLED_P (fcmpged); |
| case LT: return N2FPU_OP_ENABLED_P (fcmpltd); |
| case LE: return N2FPU_OP_ENABLED_P (fcmpled); |
| default: break; |
| } |
| return false; |
| } |
| |
| /* Stack layout and calling conventions. */ |
| |
| #define NIOS2_STACK_ALIGN(LOC) \ |
| (((LOC) + ((PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT) - 1)) \ |
| & ~((PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT) - 1)) |
| |
| /* Return the bytes needed to compute the frame pointer from the current |
| stack pointer. */ |
| static int |
| nios2_compute_frame_layout (void) |
| { |
| unsigned int regno; |
| unsigned int save_mask = 0; |
| int total_size; |
| int var_size; |
| int out_args_size; |
| int save_reg_size; |
| |
| if (cfun->machine->initialized) |
| return cfun->machine->total_size; |
| |
| var_size = NIOS2_STACK_ALIGN (get_frame_size ()); |
| out_args_size = NIOS2_STACK_ALIGN (crtl->outgoing_args_size); |
| total_size = var_size + out_args_size; |
| |
| /* Calculate space needed for gp registers. */ |
| save_reg_size = 0; |
| for (regno = 0; regno <= LAST_GP_REG; regno++) |
| if (prologue_saved_reg_p (regno)) |
| { |
| save_mask |= 1 << regno; |
| save_reg_size += 4; |
| } |
| |
| /* If we call eh_return, we need to save the EH data registers. */ |
| if (crtl->calls_eh_return) |
| { |
| unsigned i; |
| unsigned r; |
| |
| for (i = 0; (r = EH_RETURN_DATA_REGNO (i)) != INVALID_REGNUM; i++) |
| if (!(save_mask & (1 << r))) |
| { |
| save_mask |= 1 << r; |
| save_reg_size += 4; |
| } |
| } |
| |
| cfun->machine->fp_save_offset = 0; |
| if (save_mask & (1 << HARD_FRAME_POINTER_REGNUM)) |
| { |
| int fp_save_offset = 0; |
| for (regno = 0; regno < HARD_FRAME_POINTER_REGNUM; regno++) |
| if (save_mask & (1 << regno)) |
| fp_save_offset += 4; |
| |
| cfun->machine->fp_save_offset = fp_save_offset; |
| } |
| |
| save_reg_size = NIOS2_STACK_ALIGN (save_reg_size); |
| total_size += save_reg_size; |
| total_size += NIOS2_STACK_ALIGN (crtl->args.pretend_args_size); |
| |
| /* Save other computed information. */ |
| cfun->machine->save_mask = save_mask; |
| cfun->machine->total_size = total_size; |
| cfun->machine->var_size = var_size; |
| cfun->machine->args_size = out_args_size; |
| cfun->machine->save_reg_size = save_reg_size; |
| cfun->machine->initialized = reload_completed; |
| cfun->machine->save_regs_offset = out_args_size + var_size; |
| |
| return total_size; |
| } |
| |
| /* Generate save/restore of register REGNO at SP + OFFSET. Used by the |
| prologue/epilogue expand routines. */ |
| static void |
| save_reg (int regno, unsigned offset) |
| { |
| rtx reg = gen_rtx_REG (SImode, regno); |
| rtx addr = gen_rtx_PLUS (Pmode, stack_pointer_rtx, |
| gen_int_mode (offset, Pmode)); |
| rtx insn = emit_move_insn (gen_frame_mem (Pmode, addr), reg); |
| RTX_FRAME_RELATED_P (insn) = 1; |
| } |
| |
| static void |
| restore_reg (int regno, unsigned offset) |
| { |
| rtx reg = gen_rtx_REG (SImode, regno); |
| rtx addr = gen_rtx_PLUS (Pmode, stack_pointer_rtx, |
| gen_int_mode (offset, Pmode)); |
| rtx insn = emit_move_insn (reg, gen_frame_mem (Pmode, addr)); |
| /* Tag epilogue unwind note. */ |
| add_reg_note (insn, REG_CFA_RESTORE, reg); |
| RTX_FRAME_RELATED_P (insn) = 1; |
| } |
| |
| /* Emit conditional trap for checking stack limit. */ |
| static void |
| nios2_emit_stack_limit_check (void) |
| { |
| if (REG_P (stack_limit_rtx)) |
| emit_insn (gen_ctrapsi4 (gen_rtx_LTU (VOIDmode, stack_pointer_rtx, |
| stack_limit_rtx), |
| stack_pointer_rtx, stack_limit_rtx, GEN_INT (3))); |
| else |
| sorry ("only register based stack limit is supported"); |
| } |
| |
| /* Temp regno used inside prologue/epilogue. */ |
| #define TEMP_REG_NUM 8 |
| |
| void |
| nios2_expand_prologue (void) |
| { |
| unsigned int regno; |
| int total_frame_size, save_offset; |
| int sp_offset; /* offset from base_reg to final stack value. */ |
| int save_regs_base; /* offset from base_reg to register save area. */ |
| rtx insn; |
| |
| total_frame_size = nios2_compute_frame_layout (); |
| |
| if (flag_stack_usage_info) |
| current_function_static_stack_size = total_frame_size; |
| |
| /* Decrement the stack pointer. */ |
| if (!SMALL_INT (total_frame_size)) |
| { |
| /* We need an intermediary point, this will point at the spill block. */ |
| insn = emit_insn |
| (gen_add2_insn (stack_pointer_rtx, |
| gen_int_mode (cfun->machine->save_regs_offset |
| - total_frame_size, Pmode))); |
| RTX_FRAME_RELATED_P (insn) = 1; |
| save_regs_base = 0; |
| sp_offset = -cfun->machine->save_regs_offset; |
| } |
| else if (total_frame_size) |
| { |
| insn = emit_insn (gen_add2_insn (stack_pointer_rtx, |
| gen_int_mode (-total_frame_size, |
| Pmode))); |
| RTX_FRAME_RELATED_P (insn) = 1; |
| save_regs_base = cfun->machine->save_regs_offset; |
| sp_offset = 0; |
| } |
| else |
| save_regs_base = sp_offset = 0; |
| |
| if (crtl->limit_stack) |
| nios2_emit_stack_limit_check (); |
| |
| save_offset = save_regs_base + cfun->machine->save_reg_size; |
| |
| for (regno = LAST_GP_REG; regno > 0; regno--) |
| if (cfun->machine->save_mask & (1 << regno)) |
| { |
| save_offset -= 4; |
| save_reg (regno, save_offset); |
| } |
| |
| if (frame_pointer_needed) |
| { |
| int fp_save_offset = save_regs_base + cfun->machine->fp_save_offset; |
| insn = emit_insn (gen_add3_insn (hard_frame_pointer_rtx, |
| stack_pointer_rtx, |
| gen_int_mode (fp_save_offset, Pmode))); |
| RTX_FRAME_RELATED_P (insn) = 1; |
| } |
| |
| if (sp_offset) |
| { |
| rtx sp_adjust |
| = gen_rtx_SET (VOIDmode, stack_pointer_rtx, |
| plus_constant (Pmode, stack_pointer_rtx, sp_offset)); |
| if (SMALL_INT (sp_offset)) |
| insn = emit_insn (sp_adjust); |
| else |
| { |
| rtx tmp = gen_rtx_REG (Pmode, TEMP_REG_NUM); |
| emit_move_insn (tmp, gen_int_mode (sp_offset, Pmode)); |
| insn = emit_insn (gen_add2_insn (stack_pointer_rtx, tmp)); |
| /* Attach the sp_adjust as a note indicating what happened. */ |
| add_reg_note (insn, REG_FRAME_RELATED_EXPR, sp_adjust); |
| } |
| RTX_FRAME_RELATED_P (insn) = 1; |
| |
| if (crtl->limit_stack) |
| nios2_emit_stack_limit_check (); |
| } |
| |
| /* Load the PIC register if needed. */ |
| if (crtl->uses_pic_offset_table) |
| nios2_load_pic_register (); |
| |
| /* If we are profiling, make sure no instructions are scheduled before |
| the call to mcount. */ |
| if (crtl->profile) |
| emit_insn (gen_blockage ()); |
| } |
| |
| void |
| nios2_expand_epilogue (bool sibcall_p) |
| { |
| rtx insn, cfa_adj; |
| int total_frame_size; |
| int sp_adjust, save_offset; |
| unsigned int regno; |
| |
| if (!sibcall_p && nios2_can_use_return_insn ()) |
| { |
| emit_jump_insn (gen_return ()); |
| return; |
| } |
| |
| emit_insn (gen_blockage ()); |
| |
| total_frame_size = nios2_compute_frame_layout (); |
| if (frame_pointer_needed) |
| { |
| /* Recover the stack pointer. */ |
| insn = emit_insn (gen_add3_insn |
| (stack_pointer_rtx, hard_frame_pointer_rtx, |
| gen_int_mode (-cfun->machine->fp_save_offset, Pmode))); |
| cfa_adj = plus_constant (Pmode, stack_pointer_rtx, |
| (total_frame_size |
| - cfun->machine->save_regs_offset)); |
| add_reg_note (insn, REG_CFA_DEF_CFA, cfa_adj); |
| RTX_FRAME_RELATED_P (insn) = 1; |
| |
| save_offset = 0; |
| sp_adjust = total_frame_size - cfun->machine->save_regs_offset; |
| } |
| else if (!SMALL_INT (total_frame_size)) |
| { |
| rtx tmp = gen_rtx_REG (Pmode, TEMP_REG_NUM); |
| emit_move_insn (tmp, gen_int_mode (cfun->machine->save_regs_offset, |
| Pmode)); |
| insn = emit_insn (gen_add2_insn (stack_pointer_rtx, tmp)); |
| cfa_adj = gen_rtx_SET (VOIDmode, stack_pointer_rtx, |
| plus_constant (Pmode, stack_pointer_rtx, |
| cfun->machine->save_regs_offset)); |
| add_reg_note (insn, REG_CFA_ADJUST_CFA, cfa_adj); |
| RTX_FRAME_RELATED_P (insn) = 1; |
| save_offset = 0; |
| sp_adjust = total_frame_size - cfun->machine->save_regs_offset; |
| } |
| else |
| { |
| save_offset = cfun->machine->save_regs_offset; |
| sp_adjust = total_frame_size; |
| } |
| |
| save_offset += cfun->machine->save_reg_size; |
| |
| for (regno = LAST_GP_REG; regno > 0; regno--) |
| if (cfun->machine->save_mask & (1 << regno)) |
| { |
| save_offset -= 4; |
| restore_reg (regno, save_offset); |
| } |
| |
| if (sp_adjust) |
| { |
| insn = emit_insn (gen_add2_insn (stack_pointer_rtx, |
| gen_int_mode (sp_adjust, Pmode))); |
| cfa_adj = gen_rtx_SET (VOIDmode, stack_pointer_rtx, |
| plus_constant (Pmode, stack_pointer_rtx, |
| sp_adjust)); |
| add_reg_note (insn, REG_CFA_ADJUST_CFA, cfa_adj); |
| RTX_FRAME_RELATED_P (insn) = 1; |
| } |
| |
| /* Add in the __builtin_eh_return stack adjustment. */ |
| if (crtl->calls_eh_return) |
| emit_insn (gen_add2_insn (stack_pointer_rtx, EH_RETURN_STACKADJ_RTX)); |
| |
| if (!sibcall_p) |
| emit_jump_insn (gen_simple_return ()); |
| } |
| |
| /* Implement RETURN_ADDR_RTX. Note, we do not support moving |
| back to a previous frame. */ |
| rtx |
| nios2_get_return_address (int count) |
| { |
| if (count != 0) |
| return const0_rtx; |
| |
| return get_hard_reg_initial_val (Pmode, RA_REGNO); |
| } |
| |
| /* Emit code to change the current function's return address to |
| ADDRESS. SCRATCH is available as a scratch register, if needed. |
| ADDRESS and SCRATCH are both word-mode GPRs. */ |
| void |
| nios2_set_return_address (rtx address, rtx scratch) |
| { |
| nios2_compute_frame_layout (); |
| if (cfun->machine->save_mask & (1 << RA_REGNO)) |
| { |
| unsigned offset = cfun->machine->save_reg_size - 4; |
| rtx base; |
| |
| if (frame_pointer_needed) |
| base = hard_frame_pointer_rtx; |
| else |
| { |
| base = stack_pointer_rtx; |
| offset += cfun->machine->save_regs_offset; |
| |
| if (!SMALL_INT (offset)) |
| { |
| emit_move_insn (scratch, gen_int_mode (offset, Pmode)); |
| emit_insn (gen_add2_insn (scratch, base)); |
| base = scratch; |
| offset = 0; |
| } |
| } |
| if (offset) |
| base = plus_constant (Pmode, base, offset); |
| emit_move_insn (gen_rtx_MEM (Pmode, base), address); |
| } |
| else |
| emit_move_insn (gen_rtx_REG (Pmode, RA_REGNO), address); |
| } |
| |
| /* Implement FUNCTION_PROFILER macro. */ |
| void |
| nios2_function_profiler (FILE *file, int labelno ATTRIBUTE_UNUSED) |
| { |
| fprintf (file, "\tmov\tr8, ra\n"); |
| if (flag_pic == 1) |
| { |
| fprintf (file, "\tnextpc\tr2\n"); |
| fprintf (file, "\t1: movhi\tr3, %%hiadj(_gp_got - 1b)\n"); |
| fprintf (file, "\taddi\tr3, r3, %%lo(_gp_got - 1b)\n"); |
| fprintf (file, "\tadd\tr2, r2, r3\n"); |
| fprintf (file, "\tldw\tr2, %%call(_mcount)(r2)\n"); |
| fprintf (file, "\tcallr\tr2\n"); |
| } |
| else if (flag_pic == 2) |
| { |
| fprintf (file, "\tnextpc\tr2\n"); |
| fprintf (file, "\t1: movhi\tr3, %%hiadj(_gp_got - 1b)\n"); |
| fprintf (file, "\taddi\tr3, r3, %%lo(_gp_got - 1b)\n"); |
| fprintf (file, "\tadd\tr2, r2, r3\n"); |
| fprintf (file, "\tmovhi\tr3, %%call_hiadj(_mcount)\n"); |
| fprintf (file, "\taddi\tr3, r3, %%call_lo(_mcount)\n"); |
| fprintf (file, "\tadd\tr3, r2, r3\n"); |
| fprintf (file, "\tldw\tr2, 0(r3)\n"); |
| fprintf (file, "\tcallr\tr2\n"); |
| } |
| else |
| fprintf (file, "\tcall\t_mcount\n"); |
| fprintf (file, "\tmov\tra, r8\n"); |
| } |
| |
| /* Dump stack layout. */ |
| static void |
| nios2_dump_frame_layout (FILE *file) |
| { |
| fprintf (file, "\t%s Current Frame Info\n", ASM_COMMENT_START); |
| fprintf (file, "\t%s total_size = %d\n", ASM_COMMENT_START, |
| cfun->machine->total_size); |
| fprintf (file, "\t%s var_size = %d\n", ASM_COMMENT_START, |
| cfun->machine->var_size); |
| fprintf (file, "\t%s args_size = %d\n", ASM_COMMENT_START, |
| cfun->machine->args_size); |
| fprintf (file, "\t%s save_reg_size = %d\n", ASM_COMMENT_START, |
| cfun->machine->save_reg_size); |
| fprintf (file, "\t%s initialized = %d\n", ASM_COMMENT_START, |
| cfun->machine->initialized); |
| fprintf (file, "\t%s save_regs_offset = %d\n", ASM_COMMENT_START, |
| cfun->machine->save_regs_offset); |
| fprintf (file, "\t%s is_leaf = %d\n", ASM_COMMENT_START, |
| crtl->is_leaf); |
| fprintf (file, "\t%s frame_pointer_needed = %d\n", ASM_COMMENT_START, |
| frame_pointer_needed); |
| fprintf (file, "\t%s pretend_args_size = %d\n", ASM_COMMENT_START, |
| crtl->args.pretend_args_size); |
| } |
| |
| /* Return true if REGNO should be saved in the prologue. */ |
| static bool |
| prologue_saved_reg_p (unsigned regno) |
| { |
| gcc_assert (GP_REG_P (regno)); |
| |
| if (df_regs_ever_live_p (regno) && !call_used_regs[regno]) |
| return true; |
| |
| if (regno == HARD_FRAME_POINTER_REGNUM && frame_pointer_needed) |
| return true; |
| |
| if (regno == PIC_OFFSET_TABLE_REGNUM && crtl->uses_pic_offset_table) |
| return true; |
| |
| if (regno == RA_REGNO && df_regs_ever_live_p (RA_REGNO)) |
| return true; |
| |
| return false; |
| } |
| |
| /* Implement TARGET_CAN_ELIMINATE. */ |
| static bool |
| nios2_can_eliminate (const int from ATTRIBUTE_UNUSED, const int to) |
| { |
| if (to == STACK_POINTER_REGNUM) |
| return !frame_pointer_needed; |
| return true; |
| } |
| |
| /* Implement INITIAL_ELIMINATION_OFFSET macro. */ |
| int |
| nios2_initial_elimination_offset (int from, int to) |
| { |
| int offset; |
| |
| nios2_compute_frame_layout (); |
| |
| /* Set OFFSET to the offset from the stack pointer. */ |
| switch (from) |
| { |
| case FRAME_POINTER_REGNUM: |
| offset = cfun->machine->args_size; |
| break; |
| |
| case ARG_POINTER_REGNUM: |
| offset = cfun->machine->total_size; |
| offset -= crtl->args.pretend_args_size; |
| break; |
| |
| default: |
| gcc_unreachable (); |
| } |
| |
| /* If we are asked for the frame pointer offset, then adjust OFFSET |
| by the offset from the frame pointer to the stack pointer. */ |
| if (to == HARD_FRAME_POINTER_REGNUM) |
| offset -= (cfun->machine->save_regs_offset |
| + cfun->machine->fp_save_offset); |
| |
| return offset; |
| } |
| |
| /* Return nonzero if this function is known to have a null epilogue. |
| This allows the optimizer to omit jumps to jumps if no stack |
| was created. */ |
| int |
| nios2_can_use_return_insn (void) |
| { |
| if (!reload_completed || crtl->profile) |
| return 0; |
| |
| return nios2_compute_frame_layout () == 0; |
| } |
| |
| |
| /* Check and signal some warnings/errors on FPU insn options. */ |
| static void |
| nios2_custom_check_insns (void) |
| { |
| unsigned int i, j; |
| bool errors = false; |
| |
| for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) |
| if (N2FPU_ENABLED_P (i) && N2FPU_DOUBLE_P (i)) |
| { |
| for (j = 0; j < ARRAY_SIZE (nios2_fpu_insn); j++) |
| if (N2FPU_DOUBLE_REQUIRED_P (j) && ! N2FPU_ENABLED_P (j)) |
| { |
| error ("switch %<-mcustom-%s%> is required for double " |
| "precision floating point", N2FPU_NAME (j)); |
| errors = true; |
| } |
| break; |
| } |
| |
| /* Warn if the user has certain exotic operations that won't get used |
| without -funsafe-math-optimizations. See expand_builtin () in |
| builtins.c. */ |
| if (!flag_unsafe_math_optimizations) |
| for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) |
| if (N2FPU_ENABLED_P (i) && N2FPU_UNSAFE_P (i)) |
| warning (0, "switch %<-mcustom-%s%> has no effect unless " |
| "-funsafe-math-optimizations is specified", N2FPU_NAME (i)); |
| |
| /* Warn if the user is trying to use -mcustom-fmins et. al, that won't |
| get used without -ffinite-math-only. See fold_builtin_fmin_fmax () |
| in builtins.c. */ |
| if (!flag_finite_math_only) |
| for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) |
| if (N2FPU_ENABLED_P (i) && N2FPU_FINITE_P (i)) |
| warning (0, "switch %<-mcustom-%s%> has no effect unless " |
| "-ffinite-math-only is specified", N2FPU_NAME (i)); |
| |
| if (errors || custom_code_conflict) |
| fatal_error ("conflicting use of -mcustom switches, target attributes, " |
| "and/or __builtin_custom_ functions"); |
| } |
| |
| static void |
| nios2_set_fpu_custom_code (enum n2fpu_code code, int n, bool override_p) |
| { |
| if (override_p || N2FPU_N (code) == -1) |
| N2FPU_N (code) = n; |
| nios2_register_custom_code (n, CCS_FPU, (int) code); |
| } |
| |
| /* Type to represent a standard FPU config. */ |
| struct nios2_fpu_config |
| { |
| const char *name; |
| bool set_sp_constants; |
| int code[n2fpu_code_num]; |
| }; |
| |
| #define NIOS2_FPU_CONFIG_NUM 3 |
| static struct nios2_fpu_config custom_fpu_config[NIOS2_FPU_CONFIG_NUM]; |
| |
| static void |
| nios2_init_fpu_configs (void) |
| { |
| struct nios2_fpu_config* cfg; |
| int i = 0; |
| #define NEXT_FPU_CONFIG \ |
| do { \ |
| cfg = &custom_fpu_config[i++]; \ |
| memset (cfg, -1, sizeof (struct nios2_fpu_config));\ |
| } while (0) |
| |
| NEXT_FPU_CONFIG; |
| cfg->name = "60-1"; |
| cfg->set_sp_constants = true; |
| cfg->code[n2fpu_fmuls] = 252; |
| cfg->code[n2fpu_fadds] = 253; |
| cfg->code[n2fpu_fsubs] = 254; |
| |
| NEXT_FPU_CONFIG; |
| cfg->name = "60-2"; |
| cfg->set_sp_constants = true; |
| cfg->code[n2fpu_fmuls] = 252; |
| cfg->code[n2fpu_fadds] = 253; |
| cfg->code[n2fpu_fsubs] = 254; |
| cfg->code[n2fpu_fdivs] = 255; |
| |
| NEXT_FPU_CONFIG; |
| cfg->name = "72-3"; |
| cfg->set_sp_constants = true; |
| cfg->code[n2fpu_floatus] = 243; |
| cfg->code[n2fpu_fixsi] = 244; |
| cfg->code[n2fpu_floatis] = 245; |
| cfg->code[n2fpu_fcmpgts] = 246; |
| cfg->code[n2fpu_fcmples] = 249; |
| cfg->code[n2fpu_fcmpeqs] = 250; |
| cfg->code[n2fpu_fcmpnes] = 251; |
| cfg->code[n2fpu_fmuls] = 252; |
| cfg->code[n2fpu_fadds] = 253; |
| cfg->code[n2fpu_fsubs] = 254; |
| cfg->code[n2fpu_fdivs] = 255; |
| |
| #undef NEXT_FPU_CONFIG |
| gcc_assert (i == NIOS2_FPU_CONFIG_NUM); |
| } |
| |
| static struct nios2_fpu_config * |
| nios2_match_custom_fpu_cfg (const char *cfgname, const char *endp) |
| { |
| int i; |
| for (i = 0; i < NIOS2_FPU_CONFIG_NUM; i++) |
| { |
| bool match = !(endp != NULL |
| ? strncmp (custom_fpu_config[i].name, cfgname, |
| endp - cfgname) |
| : strcmp (custom_fpu_config[i].name, cfgname)); |
| if (match) |
| return &custom_fpu_config[i]; |
| } |
| return NULL; |
| } |
| |
| /* Use CFGNAME to lookup FPU config, ENDP if not NULL marks end of string. |
| OVERRIDE is true if loaded config codes should overwrite current state. */ |
| static void |
| nios2_handle_custom_fpu_cfg (const char *cfgname, const char *endp, |
| bool override) |
| { |
| struct nios2_fpu_config *cfg = nios2_match_custom_fpu_cfg (cfgname, endp); |
| if (cfg) |
| { |
| unsigned int i; |
| for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) |
| if (cfg->code[i] >= 0) |
| nios2_set_fpu_custom_code ((enum n2fpu_code) i, cfg->code[i], |
| override); |
| if (cfg->set_sp_constants) |
| flag_single_precision_constant = 1; |
| } |
| else |
| warning (0, "ignoring unrecognized switch %<-mcustom-fpu-cfg%> " |
| "value %<%s%>", cfgname); |
| |
| /* Guard against errors in the standard configurations. */ |
| nios2_custom_check_insns (); |
| } |
| |
| /* Check individual FPU insn options, and register custom code. */ |
| static void |
| nios2_handle_custom_fpu_insn_option (int fpu_insn_index) |
| { |
| int param = N2FPU_N (fpu_insn_index); |
| |
| if (0 <= param && param <= 255) |
| nios2_register_custom_code (param, CCS_FPU, fpu_insn_index); |
| |
| /* Valid values are 0-255, but also allow -1 so that the |
| -mno-custom-<opt> switches work. */ |
| else if (param != -1) |
| error ("switch %<-mcustom-%s%> value %d must be between 0 and 255", |
| N2FPU_NAME (fpu_insn_index), param); |
| } |
| |
| /* Allocate a chunk of memory for per-function machine-dependent data. */ |
| static struct machine_function * |
| nios2_init_machine_status (void) |
| { |
| return ggc_alloc_cleared_machine_function (); |
| } |
| |
| /* Implement TARGET_OPTION_OVERRIDE. */ |
| static void |
| nios2_option_override (void) |
| { |
| unsigned int i; |
| |
| #ifdef SUBTARGET_OVERRIDE_OPTIONS |
| SUBTARGET_OVERRIDE_OPTIONS; |
| #endif |
| |
| /* Check for unsupported options. */ |
| if (flag_pic && !TARGET_LINUX_ABI) |
| sorry ("position-independent code requires the Linux ABI"); |
| |
| /* Function to allocate machine-dependent function status. */ |
| init_machine_status = &nios2_init_machine_status; |
| |
| nios2_section_threshold |
| = (global_options_set.x_g_switch_value |
| ? g_switch_value : NIOS2_DEFAULT_GVALUE); |
| |
| /* Default to -mgpopt unless -fpic or -fPIC. */ |
| if (TARGET_GPOPT == -1 && flag_pic) |
| TARGET_GPOPT = 0; |
| |
| /* If we don't have mul, we don't have mulx either! */ |
| if (!TARGET_HAS_MUL && TARGET_HAS_MULX) |
| target_flags &= ~MASK_HAS_MULX; |
| |
| /* Initialize default FPU configurations. */ |
| nios2_init_fpu_configs (); |
| |
| /* Set up default handling for floating point custom instructions. |
| |
| Putting things in this order means that the -mcustom-fpu-cfg= |
| switch will always be overridden by individual -mcustom-fadds= |
| switches, regardless of the order in which they were specified |
| on the command line. |
| |
| This behavior of prioritization of individual -mcustom-<insn>= |
| options before the -mcustom-fpu-cfg= switch is maintained for |
| compatibility. */ |
| if (nios2_custom_fpu_cfg_string && *nios2_custom_fpu_cfg_string) |
| nios2_handle_custom_fpu_cfg (nios2_custom_fpu_cfg_string, NULL, false); |
| |
| /* Handle options for individual FPU insns. */ |
| for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) |
| nios2_handle_custom_fpu_insn_option (i); |
| |
| nios2_custom_check_insns (); |
| |
| /* Save the initial options in case the user does function specific |
| options. */ |
| target_option_default_node = target_option_current_node |
| = build_target_option_node (&global_options); |
| } |
| |
| |
| /* Return true if CST is a constant within range of movi/movui/movhi. */ |
| static bool |
| nios2_simple_const_p (const_rtx cst) |
| { |
| HOST_WIDE_INT val = INTVAL (cst); |
| return SMALL_INT (val) || SMALL_INT_UNSIGNED (val) || UPPER16_INT (val); |
| } |
| |
| /* Compute a (partial) cost for rtx X. Return true if the complete |
| cost has been computed, and false if subexpressions should be |
| scanned. In either case, *TOTAL contains the cost result. */ |
| static bool |
| nios2_rtx_costs (rtx x, int code, int outer_code ATTRIBUTE_UNUSED, |
| int opno ATTRIBUTE_UNUSED, |
| int *total, bool speed ATTRIBUTE_UNUSED) |
| { |
| switch (code) |
| { |
| case CONST_INT: |
| if (INTVAL (x) == 0) |
| { |
| *total = COSTS_N_INSNS (0); |
| return true; |
| } |
| else if (nios2_simple_const_p (x)) |
| { |
| *total = COSTS_N_INSNS (2); |
| return true; |
| } |
| else |
| { |
| *total = COSTS_N_INSNS (4); |
| return true; |
| } |
| |
| case LABEL_REF: |
| case SYMBOL_REF: |
| case CONST: |
| case CONST_DOUBLE: |
| { |
| *total = COSTS_N_INSNS (4); |
| return true; |
| } |
| |
| case AND: |
| { |
| /* Recognize 'nor' insn pattern. */ |
| if (GET_CODE (XEXP (x, 0)) == NOT |
| && GET_CODE (XEXP (x, 1)) == NOT) |
| { |
| *total = COSTS_N_INSNS (1); |
| return true; |
| } |
| return false; |
| } |
| |
| case MULT: |
| { |
| *total = COSTS_N_INSNS (1); |
| return false; |
| } |
| case SIGN_EXTEND: |
| { |
| *total = COSTS_N_INSNS (3); |
| return false; |
| } |
| case ZERO_EXTEND: |
| { |
| *total = COSTS_N_INSNS (1); |
| return false; |
| } |
| |
| default: |
| return false; |
| } |
| } |
| |
| /* Implement TARGET_PREFERRED_RELOAD_CLASS. */ |
| static reg_class_t |
| nios2_preferred_reload_class (rtx x ATTRIBUTE_UNUSED, reg_class_t regclass) |
| { |
| return regclass == NO_REGS ? GENERAL_REGS : regclass; |
| } |
| |
| /* Emit a call to __tls_get_addr. TI is the argument to this function. |
| RET is an RTX for the return value location. The entire insn sequence |
| is returned. */ |
| static GTY(()) rtx nios2_tls_symbol; |
| |
| static rtx |
| nios2_call_tls_get_addr (rtx ti) |
| { |
| rtx arg = gen_rtx_REG (Pmode, FIRST_ARG_REGNO); |
| rtx ret = gen_rtx_REG (Pmode, FIRST_RETVAL_REGNO); |
| rtx fn, insn; |
| |
| if (!nios2_tls_symbol) |
| nios2_tls_symbol = init_one_libfunc ("__tls_get_addr"); |
| |
| emit_move_insn (arg, ti); |
| fn = gen_rtx_MEM (QImode, nios2_tls_symbol); |
| insn = emit_call_insn (gen_call_value (ret, fn, const0_rtx)); |
| RTL_CONST_CALL_P (insn) = 1; |
| use_reg (&CALL_INSN_FUNCTION_USAGE (insn), ret); |
| use_reg (&CALL_INSN_FUNCTION_USAGE (insn), arg); |
| |
| return ret; |
| } |
| |
| /* Return true for large offsets requiring hiadj/lo relocation pairs. */ |
| static bool |
| nios2_large_offset_p (int unspec) |
| { |
| gcc_assert (nios2_unspec_reloc_name (unspec) != NULL); |
| |
| if (flag_pic == 2 |
| /* FIXME: TLS GOT offset relocations will eventually also get this |
| treatment, after binutils support for those are also completed. */ |
| && (unspec == UNSPEC_PIC_SYM || unspec == UNSPEC_PIC_CALL_SYM)) |
| return true; |
| |
| /* 'gotoff' offsets are always hiadj/lo. */ |
| if (unspec == UNSPEC_PIC_GOTOFF_SYM) |
| return true; |
| |
| return false; |
| } |
| |
| /* Return true for conforming unspec relocations. Also used in |
| constraints.md and predicates.md. */ |
| bool |
| nios2_unspec_reloc_p (rtx op) |
| { |
| return (GET_CODE (op) == CONST |
| && GET_CODE (XEXP (op, 0)) == UNSPEC |
| && ! nios2_large_offset_p (XINT (XEXP (op, 0), 1))); |
| } |
| |
| /* Helper to generate unspec constant. */ |
| static rtx |
| nios2_unspec_offset (rtx loc, int unspec) |
| { |
| return gen_rtx_CONST (Pmode, gen_rtx_UNSPEC (Pmode, gen_rtvec (1, loc), |
| unspec)); |
| } |
| |
| /* Generate GOT pointer based address with large offset. */ |
| static rtx |
| nios2_large_got_address (rtx offset) |
| { |
| rtx addr = gen_reg_rtx (Pmode); |
| emit_insn (gen_add3_insn (addr, pic_offset_table_rtx, |
| force_reg (Pmode, offset))); |
| return addr; |
| } |
| |
| /* Generate a GOT pointer based address. */ |
| static rtx |
| nios2_got_address (rtx loc, int unspec) |
| { |
| rtx offset = nios2_unspec_offset (loc, unspec); |
| crtl->uses_pic_offset_table = 1; |
| |
| if (nios2_large_offset_p (unspec)) |
| return nios2_large_got_address (offset); |
| |
| return gen_rtx_PLUS (Pmode, pic_offset_table_rtx, offset); |
| } |
| |
| /* Generate the code to access LOC, a thread local SYMBOL_REF. The |
| return value will be a valid address and move_operand (either a REG |
| or a LO_SUM). */ |
| static rtx |
| nios2_legitimize_tls_address (rtx loc) |
| { |
| rtx tmp, mem, tp; |
| enum tls_model model = SYMBOL_REF_TLS_MODEL (loc); |
| |
| switch (model) |
| { |
| case TLS_MODEL_GLOBAL_DYNAMIC: |
| tmp = gen_reg_rtx (Pmode); |
| emit_move_insn (tmp, nios2_got_address (loc, UNSPEC_ADD_TLS_GD)); |
| return nios2_call_tls_get_addr (tmp); |
| |
| case TLS_MODEL_LOCAL_DYNAMIC: |
| tmp = gen_reg_rtx (Pmode); |
| emit_move_insn (tmp, nios2_got_address (loc, UNSPEC_ADD_TLS_LDM)); |
| return gen_rtx_PLUS (Pmode, nios2_call_tls_get_addr (tmp), |
| nios2_unspec_offset (loc, UNSPEC_ADD_TLS_LDO)); |
| |
| case TLS_MODEL_INITIAL_EXEC: |
| tmp = gen_reg_rtx (Pmode); |
| mem = gen_const_mem (Pmode, nios2_got_address (loc, UNSPEC_LOAD_TLS_IE)); |
| emit_move_insn (tmp, mem); |
| tp = gen_rtx_REG (Pmode, TP_REGNO); |
| return gen_rtx_PLUS (Pmode, tp, tmp); |
| |
| case TLS_MODEL_LOCAL_EXEC: |
| tp = gen_rtx_REG (Pmode, TP_REGNO); |
| return gen_rtx_PLUS (Pmode, tp, |
| nios2_unspec_offset (loc, UNSPEC_ADD_TLS_LE)); |
| default: |
| gcc_unreachable (); |
| } |
| } |
| |
| /* Divide Support |
| |
| If -O3 is used, we want to output a table lookup for |
| divides between small numbers (both num and den >= 0 |
| and < 0x10). The overhead of this method in the worst |
| case is 40 bytes in the text section (10 insns) and |
| 256 bytes in the data section. Additional divides do |
| not incur additional penalties in the data section. |
| |
| Code speed is improved for small divides by about 5x |
| when using this method in the worse case (~9 cycles |
| vs ~45). And in the worst case divides not within the |
| table are penalized by about 10% (~5 cycles vs ~45). |
| However in the typical case the penalty is not as bad |
| because doing the long divide in only 45 cycles is |
| quite optimistic. |
| |
| ??? would be nice to have some benchmarks other |
| than Dhrystone to back this up. |
| |
| This bit of expansion is to create this instruction |
| sequence as rtl. |
| or $8, $4, $5 |
| slli $9, $4, 4 |
| cmpgeui $3, $8, 16 |
| beq $3, $0, .L3 |
| or $10, $9, $5 |
| add $12, $11, divide_table |
| ldbu $2, 0($12) |
| br .L1 |
| .L3: |
| call slow_div |
| .L1: |
| # continue here with result in $2 |
| |
| ??? Ideally I would like the libcall block to contain all |
| of this code, but I don't know how to do that. What it |
| means is that if the divide can be eliminated, it may not |
| completely disappear. |
| |
| ??? The __divsi3_table label should ideally be moved out |
| of this block and into a global. If it is placed into the |
| sdata section we can save even more cycles by doing things |
| gp relative. */ |
| void |
| nios2_emit_expensive_div (rtx *operands, enum machine_mode mode) |
| { |
| rtx or_result, shift_left_result; |
| rtx lookup_value; |
| rtx lab1, lab3; |
| rtx insns; |
| rtx libfunc; |
| rtx final_result; |
| rtx tmp; |
| rtx table; |
| |
| /* It may look a little generic, but only SImode is supported for now. */ |
| gcc_assert (mode == SImode); |
| libfunc = optab_libfunc (sdiv_optab, SImode); |
| |
| lab1 = gen_label_rtx (); |
| lab3 = gen_label_rtx (); |
| |
| or_result = expand_simple_binop (SImode, IOR, |
| operands[1], operands[2], |
| 0, 0, OPTAB_LIB_WIDEN); |
| |
| emit_cmp_and_jump_insns (or_result, GEN_INT (15), GTU, 0, |
| GET_MODE (or_result), 0, lab3); |
| JUMP_LABEL (get_last_insn ()) = lab3; |
| |
| shift_left_result = expand_simple_binop (SImode, ASHIFT, |
| operands[1], GEN_INT (4), |
| 0, 0, OPTAB_LIB_WIDEN); |
| |
| lookup_value = expand_simple_binop (SImode, IOR, |
| shift_left_result, operands[2], |
| 0, 0, OPTAB_LIB_WIDEN); |
| table = gen_rtx_PLUS (SImode, lookup_value, |
| gen_rtx_SYMBOL_REF (SImode, "__divsi3_table")); |
| convert_move (operands[0], gen_rtx_MEM (QImode, table), 1); |
| |
| tmp = emit_jump_insn (gen_jump (lab1)); |
| JUMP_LABEL (tmp) = lab1; |
| emit_barrier (); |
| |
| emit_label (lab3); |
| LABEL_NUSES (lab3) = 1; |
| |
| start_sequence (); |
| final_result = emit_library_call_value (libfunc, NULL_RTX, |
| LCT_CONST, SImode, 2, |
| operands[1], SImode, |
| operands[2], SImode); |
| |
| insns = get_insns (); |
| end_sequence (); |
| emit_libcall_block (insns, operands[0], final_result, |
| gen_rtx_DIV (SImode, operands[1], operands[2])); |
| |
| emit_label (lab1); |
| LABEL_NUSES (lab1) = 1; |
| } |
| |
| |
| /* Branches and compares. */ |
| |
| /* Return in *ALT_CODE and *ALT_OP, an alternate equivalent constant |
| comparison, e.g. >= 1 into > 0. */ |
| static void |
| nios2_alternate_compare_const (enum rtx_code code, rtx op, |
| enum rtx_code *alt_code, rtx *alt_op, |
| enum machine_mode mode) |
| { |
| HOST_WIDE_INT opval = INTVAL (op); |
| enum rtx_code scode = signed_condition (code); |
| bool dec_p = (scode == LT || scode == GE); |
| |
| if (code == EQ || code == NE) |
| { |
| *alt_code = code; |
| *alt_op = op; |
| return; |
| } |
| |
| *alt_op = (dec_p |
| ? gen_int_mode (opval - 1, mode) |
| : gen_int_mode (opval + 1, mode)); |
| |
| /* The required conversion between [>,>=] and [<,<=] is captured |
| by a reverse + swap of condition codes. */ |
| *alt_code = reverse_condition (swap_condition (code)); |
| |
| { |
| /* Test if the incremented/decremented value crosses the over/underflow |
| boundary. Supposedly, such boundary cases should already be transformed |
| into always-true/false or EQ conditions, so use an assertion here. */ |
| unsigned HOST_WIDE_INT alt_opval = INTVAL (*alt_op); |
| if (code == scode) |
| alt_opval ^= (1 << (GET_MODE_BITSIZE (mode) - 1)); |
| alt_opval &= GET_MODE_MASK (mode); |
| gcc_assert (dec_p ? alt_opval != GET_MODE_MASK (mode) : alt_opval != 0); |
| } |
| } |
| |
| /* Return true if the constant comparison is supported by nios2. */ |
| static bool |
| nios2_valid_compare_const_p (enum rtx_code code, rtx op) |
| { |
| switch (code) |
| { |
| case EQ: case NE: case GE: case LT: |
| return SMALL_INT (INTVAL (op)); |
| case GEU: case LTU: |
| return SMALL_INT_UNSIGNED (INTVAL (op)); |
| default: |
| return false; |
| } |
| } |
| |
| /* Checks if the FPU comparison in *CMP, *OP1, and *OP2 can be supported in |
| the current configuration. Perform modifications if MODIFY_P is true. |
| Returns true if FPU compare can be done. */ |
| |
| bool |
| nios2_validate_fpu_compare (enum machine_mode mode, rtx *cmp, rtx *op1, rtx *op2, |
| bool modify_p) |
| { |
| bool rev_p = false; |
| enum rtx_code code = GET_CODE (*cmp); |
| |
| if (!nios2_fpu_compare_enabled (code, mode)) |
| { |
| code = swap_condition (code); |
| if (nios2_fpu_compare_enabled (code, mode)) |
| rev_p = true; |
| else |
| return false; |
| } |
| |
| if (modify_p) |
| { |
| if (rev_p) |
| { |
| rtx tmp = *op1; |
| *op1 = *op2; |
| *op2 = tmp; |
| } |
| *op1 = force_reg (mode, *op1); |
| *op2 = force_reg (mode, *op2); |
| *cmp = gen_rtx_fmt_ee (code, mode, *op1, *op2); |
| } |
| return true; |
| } |
| |
| /* Checks and modifies the comparison in *CMP, *OP1, and *OP2 into valid |
| nios2 supported form. Returns true if success. */ |
| bool |
| nios2_validate_compare (enum machine_mode mode, rtx *cmp, rtx *op1, rtx *op2) |
| { |
| enum rtx_code code = GET_CODE (*cmp); |
| enum rtx_code alt_code; |
| rtx alt_op2; |
| |
| if (GET_MODE_CLASS (mode) == MODE_FLOAT) |
| return nios2_validate_fpu_compare (mode, cmp, op1, op2, true); |
| |
| if (!reg_or_0_operand (*op2, mode)) |
| { |
| /* Create alternate constant compare. */ |
| nios2_alternate_compare_const (code, *op2, &alt_code, &alt_op2, mode); |
| |
| /* If alterate op2 is zero(0), we can use it directly, possibly |
| swapping the compare code. */ |
| if (alt_op2 == const0_rtx) |
| { |
| code = alt_code; |
| *op2 = alt_op2; |
| goto check_rebuild_cmp; |
| } |
| |
| /* Check if either constant compare can be used. */ |
| if (nios2_valid_compare_const_p (code, *op2)) |
| return true; |
| else if (nios2_valid_compare_const_p (alt_code, alt_op2)) |
| { |
| code = alt_code; |
| *op2 = alt_op2; |
| goto rebuild_cmp; |
| } |
| |
| /* We have to force op2 into a register now. Try to pick one |
| with a lower cost. */ |
| if (! nios2_simple_const_p (*op2) |
| && nios2_simple_const_p (alt_op2)) |
| { |
| code = alt_code; |
| *op2 = alt_op2; |
| } |
| *op2 = force_reg (SImode, *op2); |
| } |
| check_rebuild_cmp: |
| if (code == GT || code == GTU || code == LE || code == LEU) |
| { |
| rtx t = *op1; *op1 = *op2; *op2 = t; |
| code = swap_condition (code); |
| } |
| rebuild_cmp: |
| *cmp = gen_rtx_fmt_ee (code, mode, *op1, *op2); |
| return true; |
| } |
| |
| |
| /* Addressing Modes. */ |
| |
| /* Implement TARGET_LEGITIMATE_CONSTANT_P. */ |
| static bool |
| nios2_legitimate_constant_p (enum machine_mode mode ATTRIBUTE_UNUSED, rtx x) |
| { |
| rtx base, offset; |
| split_const (x, &base, &offset); |
| return GET_CODE (base) != SYMBOL_REF || !SYMBOL_REF_TLS_MODEL (base); |
| } |
| |
| /* Implement TARGET_CANNOT_FORCE_CONST_MEM. */ |
| static bool |
| nios2_cannot_force_const_mem (enum machine_mode mode ATTRIBUTE_UNUSED, rtx x) |
| { |
| return nios2_legitimate_constant_p (mode, x) == false; |
| } |
| |
| /* Return true if register REGNO is a valid base register. |
| STRICT_P is true if REG_OK_STRICT is in effect. */ |
| |
| bool |
| nios2_regno_ok_for_base_p (int regno, bool strict_p) |
| { |
| if (!HARD_REGISTER_NUM_P (regno)) |
| { |
| if (!strict_p) |
| return true; |
| |
| if (!reg_renumber) |
| return false; |
| |
| regno = reg_renumber[regno]; |
| } |
| |
| /* The fake registers will be eliminated to either the stack or |
| hard frame pointer, both of which are usually valid base registers. |
| Reload deals with the cases where the eliminated form isn't valid. */ |
| return (GP_REG_P (regno) |
| || regno == FRAME_POINTER_REGNUM |
| || regno == ARG_POINTER_REGNUM); |
| } |
| |
| /* Return true if the address expression formed by BASE + OFFSET is |
| valid. */ |
| static bool |
| nios2_valid_addr_expr_p (rtx base, rtx offset, bool strict_p) |
| { |
| if (!strict_p && GET_CODE (base) == SUBREG) |
| base = SUBREG_REG (base); |
| return (REG_P (base) |
| && nios2_regno_ok_for_base_p (REGNO (base), strict_p) |
| && (offset == NULL_RTX |
| || const_arith_operand (offset, Pmode) |
| || nios2_unspec_reloc_p (offset))); |
| } |
| |
| /* Implement TARGET_LEGITIMATE_ADDRESS_P. */ |
| static bool |
| nios2_legitimate_address_p (enum machine_mode mode ATTRIBUTE_UNUSED, |
| rtx operand, bool strict_p) |
| { |
| switch (GET_CODE (operand)) |
| { |
| /* Direct. */ |
| case SYMBOL_REF: |
| if (SYMBOL_REF_TLS_MODEL (operand)) |
| return false; |
| |
| if (nios2_symbol_ref_in_small_data_p (operand)) |
| return true; |
| |
| /* Else, fall through. */ |
| case LABEL_REF: |
| case CONST_INT: |
| case CONST: |
| case CONST_DOUBLE: |
| return false; |
| |
| /* Register indirect. */ |
| case REG: |
| return nios2_regno_ok_for_base_p (REGNO (operand), strict_p); |
| |
| /* Register indirect with displacement. */ |
| case PLUS: |
| { |
| rtx op0 = XEXP (operand, 0); |
| rtx op1 = XEXP (operand, 1); |
| |
| return (nios2_valid_addr_expr_p (op0, op1, strict_p) |
| || nios2_valid_addr_expr_p (op1, op0, strict_p)); |
| } |
| |
| default: |
| break; |
| } |
| return false; |
| } |
| |
| /* Return true if SECTION is a small section name. */ |
| static bool |
| nios2_small_section_name_p (const char *section) |
| { |
| return (strcmp (section, ".sbss") == 0 |
| || strncmp (section, ".sbss.", 6) == 0 |
| || strcmp (section, ".sdata") == 0 |
| || strncmp (section, ".sdata.", 7) == 0); |
| } |
| |
| /* Return true if EXP should be placed in the small data section. */ |
| static bool |
| nios2_in_small_data_p (const_tree exp) |
| { |
| /* We want to merge strings, so we never consider them small data. */ |
| if (TREE_CODE (exp) == STRING_CST) |
| return false; |
| |
| if (TREE_CODE (exp) == VAR_DECL) |
| { |
| if (DECL_SECTION_NAME (exp)) |
| { |
| const char *section = TREE_STRING_POINTER (DECL_SECTION_NAME (exp)); |
| if (nios2_section_threshold > 0 |
| && nios2_small_section_name_p (section)) |
| return true; |
| } |
| else |
| { |
| HOST_WIDE_INT size = int_size_in_bytes (TREE_TYPE (exp)); |
| |
| /* If this is an incomplete type with size 0, then we can't put it |
| in sdata because it might be too big when completed. */ |
| if (size > 0 |
| && (unsigned HOST_WIDE_INT) size <= nios2_section_threshold) |
| return true; |
| } |
| } |
| |
| return false; |
| } |
| |
| /* Return true if symbol is in small data section. */ |
| |
| bool |
| nios2_symbol_ref_in_small_data_p (rtx sym) |
| { |
| gcc_assert (GET_CODE (sym) == SYMBOL_REF); |
| return |
| (TARGET_GPOPT |
| /* GP-relative access cannot be used for externally defined symbols, |
| because the compilation unit that defines the symbol may place it |
| in a section that cannot be reached from GP. */ |
| && !SYMBOL_REF_EXTERNAL_P (sym) |
| /* True if a symbol is both small and not weak. */ |
| && SYMBOL_REF_SMALL_P (sym) |
| && !(SYMBOL_REF_DECL (sym) && DECL_WEAK (SYMBOL_REF_DECL (sym))) |
| /* TLS variables are not accessed through the GP. */ |
| && SYMBOL_REF_TLS_MODEL (sym) == 0); |
| |
| } |
| |
| /* Implement TARGET_SECTION_TYPE_FLAGS. */ |
| |
| static unsigned int |
| nios2_section_type_flags (tree decl, const char *name, int reloc) |
| { |
| unsigned int flags; |
| |
| flags = default_section_type_flags (decl, name, reloc); |
| |
| if (nios2_small_section_name_p (name)) |
| flags |= SECTION_SMALL; |
| |
| return flags; |
| } |
| |
| /* Return true if SYMBOL_REF X binds locally. */ |
| |
| static bool |
| nios2_symbol_binds_local_p (const_rtx x) |
| { |
| return (SYMBOL_REF_DECL (x) |
| ? targetm.binds_local_p (SYMBOL_REF_DECL (x)) |
| : SYMBOL_REF_LOCAL_P (x)); |
| } |
| |
| /* Position independent code related. */ |
| |
| /* Emit code to load the PIC register. */ |
| static void |
| nios2_load_pic_register (void) |
| { |
| rtx tmp = gen_rtx_REG (Pmode, TEMP_REG_NUM); |
| |
| emit_insn (gen_load_got_register (pic_offset_table_rtx, tmp)); |
| emit_insn (gen_add3_insn (pic_offset_table_rtx, pic_offset_table_rtx, tmp)); |
| } |
| |
| /* Generate a PIC address as a MEM rtx. */ |
| static rtx |
| nios2_load_pic_address (rtx sym, int unspec) |
| { |
| if (flag_pic == 2 |
| && GET_CODE (sym) == SYMBOL_REF |
| && nios2_symbol_binds_local_p (sym)) |
| /* Under -fPIC, generate a GOTOFF address for local symbols. */ |
| return nios2_got_address (sym, UNSPEC_PIC_GOTOFF_SYM); |
| |
| return gen_const_mem (Pmode, nios2_got_address (sym, unspec)); |
| } |
| |
| /* Nonzero if the constant value X is a legitimate general operand |
| when generating PIC code. It is given that flag_pic is on and |
| that X satisfies CONSTANT_P or is a CONST_DOUBLE. */ |
| bool |
| nios2_legitimate_pic_operand_p (rtx x) |
| { |
| if (GET_CODE (x) == CONST |
| && GET_CODE (XEXP (x, 0)) == UNSPEC |
| && nios2_large_offset_p (XINT (XEXP (x, 0), 1))) |
| return true; |
| |
| return ! (GET_CODE (x) == SYMBOL_REF |
| || GET_CODE (x) == LABEL_REF || GET_CODE (x) == CONST); |
| } |
| |
| /* Return TRUE if X is a thread-local symbol. */ |
| static bool |
| nios2_tls_symbol_p (rtx x) |
| { |
| return (targetm.have_tls && GET_CODE (x) == SYMBOL_REF |
| && SYMBOL_REF_TLS_MODEL (x) != 0); |
| } |
| |
| /* Legitimize addresses that are CONSTANT_P expressions. */ |
| static rtx |
| nios2_legitimize_constant_address (rtx addr) |
| { |
| rtx base, offset; |
| split_const (addr, &base, &offset); |
| |
| if (nios2_tls_symbol_p (base)) |
| base = nios2_legitimize_tls_address (base); |
| else if (flag_pic) |
| base = nios2_load_pic_address (base, UNSPEC_PIC_SYM); |
| else |
| return addr; |
| |
| if (offset != const0_rtx) |
| { |
| gcc_assert (can_create_pseudo_p ()); |
| return gen_rtx_PLUS (Pmode, force_reg (Pmode, base), |
| (CONST_INT_P (offset) |
| ? (SMALL_INT (INTVAL (offset)) |
| ? offset : force_reg (Pmode, offset)) |
| : offset)); |
| } |
| return base; |
| } |
| |
| /* Implement TARGET_LEGITIMIZE_ADDRESS. */ |
| static rtx |
| nios2_legitimize_address (rtx x, rtx oldx ATTRIBUTE_UNUSED, |
| enum machine_mode mode ATTRIBUTE_UNUSED) |
| { |
| if (CONSTANT_P (x)) |
| return nios2_legitimize_constant_address (x); |
| |
| /* For the TLS LE (Local Exec) model, the compiler may try to |
| combine constant offsets with unspec relocs, creating address RTXs |
| looking like this: |
| (plus:SI (reg:SI 23 r23) |
| (const:SI |
| (plus:SI |
| (unspec:SI [(symbol_ref:SI ("var"))] UNSPEC_ADD_TLS_LE) |
| (const_int 48 [0x30])))) |
| |
| This usually happens when 'var' is a thread-local struct variable, |
| and access of a field in var causes the addend. |
| |
| We typically want this combining, so transform the above into this |
| form, which is allowed: |
| (plus:SI (reg:SI 23 r23) |
| (const:SI |
| (unspec:SI |
| [(const:SI |
| (plus:SI (symbol_ref:SI ("var")) |
| (const_int 48 [0x30])))] UNSPEC_ADD_TLS_LE))) |
| |
| Which will be output as '%tls_le(var+48)(r23)' in assembly. */ |
| if (GET_CODE (x) == PLUS |
| && GET_CODE (XEXP (x, 0)) == REG |
| && GET_CODE (XEXP (x, 1)) == CONST) |
| { |
| rtx unspec, offset, reg = XEXP (x, 0); |
| split_const (XEXP (x, 1), &unspec, &offset); |
| if (GET_CODE (unspec) == UNSPEC |
| && !nios2_large_offset_p (XINT (unspec, 1)) |
| && offset != const0_rtx) |
| { |
| unspec = copy_rtx (unspec); |
| XVECEXP (unspec, 0, 0) |
| = plus_constant (Pmode, XVECEXP (unspec, 0, 0), INTVAL (offset)); |
| x = gen_rtx_PLUS (Pmode, reg, gen_rtx_CONST (Pmode, unspec)); |
| } |
| } |
| |
| return x; |
| } |
| |
| static rtx |
| nios2_delegitimize_address (rtx x) |
| { |
| x = delegitimize_mem_from_attrs (x); |
| |
| if (GET_CODE (x) == CONST && GET_CODE (XEXP (x, 0)) == UNSPEC) |
| { |
| switch (XINT (XEXP (x, 0), 1)) |
| { |
| case UNSPEC_PIC_SYM: |
| case UNSPEC_PIC_CALL_SYM: |
| case UNSPEC_PIC_GOTOFF_SYM: |
| case UNSPEC_ADD_TLS_GD: |
| case UNSPEC_ADD_TLS_LDM: |
| case UNSPEC_LOAD_TLS_IE: |
| case UNSPEC_ADD_TLS_LE: |
| x = XVECEXP (XEXP (x, 0), 0, 0); |
| gcc_assert (GET_CODE (x) == SYMBOL_REF); |
| break; |
| } |
| } |
| return x; |
| } |
| |
| /* Main expander function for RTL moves. */ |
| int |
| nios2_emit_move_sequence (rtx *operands, enum machine_mode mode) |
| { |
| rtx to = operands[0]; |
| rtx from = operands[1]; |
| |
| if (!register_operand (to, mode) && !reg_or_0_operand (from, mode)) |
| { |
| gcc_assert (can_create_pseudo_p ()); |
| from = copy_to_mode_reg (mode, from); |
| } |
| |
| if (GET_CODE (from) == SYMBOL_REF || GET_CODE (from) == LABEL_REF |
| || (GET_CODE (from) == CONST |
| && GET_CODE (XEXP (from, 0)) != UNSPEC)) |
| from = nios2_legitimize_constant_address (from); |
| |
| operands[0] = to; |
| operands[1] = from; |
| return 0; |
| } |
| |
| /* The function with address *ADDR is being called. If the address |
| needs to be loaded from the GOT, emit the instruction to do so and |
| update *ADDR to point to the rtx for the loaded value. */ |
| void |
| nios2_adjust_call_address (rtx *call_op) |
| { |
| rtx addr; |
| gcc_assert (MEM_P (*call_op)); |
| addr = XEXP (*call_op, 0); |
| if (flag_pic && CONSTANT_P (addr)) |
| { |
| rtx reg = gen_reg_rtx (Pmode); |
| emit_move_insn (reg, nios2_load_pic_address (addr, UNSPEC_PIC_CALL_SYM)); |
| XEXP (*call_op, 0) = reg; |
| } |
| } |
| |
| |
| /* Output assembly language related definitions. */ |
| |
| /* Print the operand OP to file stream FILE modified by LETTER. |
| LETTER can be one of: |
| |
| i: print "i" if OP is an immediate, except 0 |
| o: print "io" if OP is volatile |
| z: for const0_rtx print $0 instead of 0 |
| H: for %hiadj |
| L: for %lo |
| U: for upper half of 32 bit value |
| D: for the upper 32-bits of a 64-bit double value |
| R: prints reverse condition. |
| */ |
| static void |
| nios2_print_operand (FILE *file, rtx op, int letter) |
| { |
| |
| switch (letter) |
| { |
| case 'i': |
| if (CONSTANT_P (op) && op != const0_rtx) |
| fprintf (file, "i"); |
| return; |
| |
| case 'o': |
| if (GET_CODE (op) == MEM |
| && ((MEM_VOLATILE_P (op) && TARGET_BYPASS_CACHE_VOLATILE) |
| || TARGET_BYPASS_CACHE)) |
| fprintf (file, "io"); |
| return; |
| |
| default: |
| break; |
| } |
| |
| if (comparison_operator (op, VOIDmode)) |
| { |
| enum rtx_code cond = GET_CODE (op); |
| if (letter == 0) |
| { |
| fprintf (file, "%s", GET_RTX_NAME (cond)); |
| return; |
| } |
| if (letter == 'R') |
| { |
| fprintf (file, "%s", GET_RTX_NAME (reverse_condition (cond))); |
| return; |
| } |
| } |
| |
| switch (GET_CODE (op)) |
| { |
| case REG: |
| if (letter == 0 || letter == 'z') |
| { |
| fprintf (file, "%s", reg_names[REGNO (op)]); |
| return; |
| } |
| else if (letter == 'D') |
| { |
| fprintf (file, "%s", reg_names[REGNO (op)+1]); |
| return; |
| } |
| break; |
| |
| case CONST_INT: |
| if (INTVAL (op) == 0 && letter == 'z') |
| { |
| fprintf (file, "zero"); |
| return; |
| } |
| |
| if (letter == 'U') |
| { |
| HOST_WIDE_INT val = INTVAL (op); |
| val = (val >> 16) & 0xFFFF; |
| output_addr_const (file, gen_int_mode (val, SImode)); |
| return; |
| } |
| /* Else, fall through. */ |
| |
| case CONST: |
| case LABEL_REF: |
| case SYMBOL_REF: |
| case CONST_DOUBLE: |
| if (letter == 0 || letter == 'z') |
| { |
| output_addr_const (file, op); |
| return; |
| } |
| else if (letter == 'H' || letter == 'L') |
| { |
| fprintf (file, "%%"); |
| if (GET_CODE (op) == CONST |
| && GET_CODE (XEXP (op, 0)) == UNSPEC) |
| { |
| rtx unspec = XEXP (op, 0); |
| int unspec_reloc = XINT (unspec, 1); |
| gcc_assert (nios2_large_offset_p (unspec_reloc)); |
| fprintf (file, "%s_", nios2_unspec_reloc_name (unspec_reloc)); |
| op = XVECEXP (unspec, 0, 0); |
| } |
| fprintf (file, letter == 'H' ? "hiadj(" : "lo("); |
| output_addr_const (file, op); |
| fprintf (file, ")"); |
| return; |
| } |
| break; |
| |
| case SUBREG: |
| case MEM: |
| if (letter == 0) |
| { |
| output_address (op); |
| return; |
| } |
| break; |
| |
| case CODE_LABEL: |
| if (letter == 0) |
| { |
| output_addr_const (file, op); |
| return; |
| } |
| break; |
| |
| default: |
| break; |
| } |
| |
| output_operand_lossage ("Unsupported operand for code '%c'", letter); |
| gcc_unreachable (); |
| } |
| |
| /* Return true if this is a GP-relative accessible reference. */ |
| static bool |
| gprel_constant_p (rtx op) |
| { |
| if (GET_CODE (op) == SYMBOL_REF |
| && nios2_symbol_ref_in_small_data_p (op)) |
| return true; |
| else if (GET_CODE (op) == CONST |
| && GET_CODE (XEXP (op, 0)) == PLUS) |
| return gprel_constant_p (XEXP (XEXP (op, 0), 0)); |
| |
| return false; |
| } |
| |
| /* Return the name string for a supported unspec reloc offset. */ |
| static const char * |
| nios2_unspec_reloc_name (int unspec) |
| { |
| switch (unspec) |
| { |
| case UNSPEC_PIC_SYM: |
| return "got"; |
| case UNSPEC_PIC_CALL_SYM: |
| return "call"; |
| case UNSPEC_PIC_GOTOFF_SYM: |
| return "gotoff"; |
| case UNSPEC_LOAD_TLS_IE: |
| return "tls_ie"; |
| case UNSPEC_ADD_TLS_LE: |
| return "tls_le"; |
| case UNSPEC_ADD_TLS_GD: |
| return "tls_gd"; |
| case UNSPEC_ADD_TLS_LDM: |
| return "tls_ldm"; |
| case UNSPEC_ADD_TLS_LDO: |
| return "tls_ldo"; |
| default: |
| return NULL; |
| } |
| } |
| |
| /* Implement TARGET_ASM_OUTPUT_ADDR_CONST_EXTRA. */ |
| static bool |
| nios2_output_addr_const_extra (FILE *file, rtx op) |
| { |
| const char *name; |
| gcc_assert (GET_CODE (op) == UNSPEC); |
| |
| /* Support for printing out const unspec relocations. */ |
| name = nios2_unspec_reloc_name (XINT (op, 1)); |
| if (name) |
| { |
| fprintf (file, "%%%s(", name); |
| output_addr_const (file, XVECEXP (op, 0, 0)); |
| fprintf (file, ")"); |
| return true; |
| } |
| return false; |
| } |
| |
| /* Implement TARGET_PRINT_OPERAND_ADDRESS. */ |
| static void |
| nios2_print_operand_address (FILE *file, rtx op) |
| { |
| switch (GET_CODE (op)) |
| { |
| case CONST: |
| case CONST_INT: |
| case LABEL_REF: |
| case CONST_DOUBLE: |
| case SYMBOL_REF: |
| if (gprel_constant_p (op)) |
| { |
| fprintf (file, "%%gprel("); |
| output_addr_const (file, op); |
| fprintf (file, ")(%s)", reg_names[GP_REGNO]); |
| return; |
| } |
| |
| break; |
| |
| case PLUS: |
| { |
| rtx op0 = XEXP (op, 0); |
| rtx op1 = XEXP (op, 1); |
| |
| if (REG_P (op0) && CONSTANT_P (op1)) |
| { |
| output_addr_const (file, op1); |
| fprintf (file, "(%s)", reg_names[REGNO (op0)]); |
| return; |
| } |
| else if (REG_P (op1) && CONSTANT_P (op0)) |
| { |
| output_addr_const (file, op0); |
| fprintf (file, "(%s)", reg_names[REGNO (op1)]); |
| return; |
| } |
| } |
| break; |
| |
| case REG: |
| fprintf (file, "0(%s)", reg_names[REGNO (op)]); |
| return; |
| |
| case MEM: |
| { |
| rtx base = XEXP (op, 0); |
| nios2_print_operand_address (file, base); |
| return; |
| } |
| default: |
| break; |
| } |
| |
| fprintf (stderr, "Missing way to print address\n"); |
| debug_rtx (op); |
| gcc_unreachable (); |
| } |
| |
| /* Implement TARGET_ASM_OUTPUT_DWARF_DTPREL. */ |
| static void |
| nios2_output_dwarf_dtprel (FILE *file, int size, rtx x) |
| { |
| gcc_assert (size == 4); |
| fprintf (file, "\t.4byte\t%%tls_ldo("); |
| output_addr_const (file, x); |
| fprintf (file, ")"); |
| } |
| |
| /* Implement TARGET_ASM_FUNCTION_PROLOGUE. */ |
| static void |
| nios2_asm_function_prologue (FILE *file, HOST_WIDE_INT size ATTRIBUTE_UNUSED) |
| { |
| if (flag_verbose_asm || flag_debug_asm) |
| { |
| nios2_compute_frame_layout (); |
| nios2_dump_frame_layout (file); |
| } |
| } |
| |
| /* Emit assembly of custom FPU instructions. */ |
| const char * |
| nios2_fpu_insn_asm (enum n2fpu_code code) |
| { |
| static char buf[256]; |
| const char *op1, *op2, *op3; |
| int ln = 256, n = 0; |
| |
| int N = N2FPU_N (code); |
| int num_operands = N2FPU (code).num_operands; |
| const char *insn_name = N2FPU_NAME (code); |
| tree ftype = nios2_ftype (N2FPU_FTCODE (code)); |
| enum machine_mode dst_mode = TYPE_MODE (TREE_TYPE (ftype)); |
| enum machine_mode src_mode = TYPE_MODE (TREE_VALUE (TYPE_ARG_TYPES (ftype))); |
| |
| /* Prepare X register for DF input operands. */ |
| if (GET_MODE_SIZE (src_mode) == 8 && num_operands == 3) |
| n = snprintf (buf, ln, "custom\t%d, zero, %%1, %%D1 # fwrx %%1\n\t", |
| N2FPU_N (n2fpu_fwrx)); |
| |
| if (src_mode == SFmode) |
| { |
| if (dst_mode == VOIDmode) |
| { |
| /* The fwry case. */ |
| op1 = op3 = "zero"; |
| op2 = "%0"; |
| num_operands -= 1; |
| } |
| else |
| { |
| op1 = (dst_mode == DFmode ? "%D0" : "%0"); |
| op2 = "%1"; |
| op3 = (num_operands == 2 ? "zero" : "%2"); |
| } |
| } |
| else if (src_mode == DFmode) |
| { |
| if (dst_mode == VOIDmode) |
| { |
| /* The fwrx case. */ |
| op1 = "zero"; |
| op2 = "%0"; |
| op3 = "%D0"; |
| num_operands -= 1; |
| } |
| else |
| { |
| op1 = (dst_mode == DFmode ? "%D0" : "%0"); |
| op2 = (num_operands == 2 ? "%1" : "%2"); |
| op3 = (num_operands == 2 ? "%D1" : "%D2"); |
| } |
| } |
| else if (src_mode == VOIDmode) |
| { |
| /* frdxlo, frdxhi, frdy cases. */ |
| gcc_assert (dst_mode == SFmode); |
| op1 = "%0"; |
| op2 = op3 = "zero"; |
| } |
| else if (src_mode == SImode) |
| { |
| /* Conversion operators. */ |
| gcc_assert (num_operands == 2); |
| op1 = (dst_mode == DFmode ? "%D0" : "%0"); |
| op2 = "%1"; |
| op3 = "zero"; |
| } |
| else |
| gcc_unreachable (); |
| |
| /* Main instruction string. */ |
| n += snprintf (buf + n, ln - n, "custom\t%d, %s, %s, %s # %s %%0%s%s", |
| N, op1, op2, op3, insn_name, |
| (num_operands >= 2 ? ", %1" : ""), |
| (num_operands == 3 ? ", %2" : "")); |
| |
| /* Extraction of Y register for DF results. */ |
| if (dst_mode == DFmode) |
| snprintf (buf + n, ln - n, "\n\tcustom\t%d, %%0, zero, zero # frdy %%0", |
| N2FPU_N (n2fpu_frdy)); |
| return buf; |
| } |
| |
| |
| |
| /* Function argument related. */ |
| |
| /* Define where to put the arguments to a function. Value is zero to |
| push the argument on the stack, or a hard register in which to |
| store the argument. |
| |
| MODE is the argument's machine mode. |
| TYPE is the data type of the argument (as a tree). |
| This is null for libcalls where that information may |
| not be available. |
| CUM is a variable of type CUMULATIVE_ARGS which gives info about |
| the preceding args and about the function being called. |
| NAMED is nonzero if this argument is a named parameter |
| (otherwise it is an extra parameter matching an ellipsis). */ |
| |
| static rtx |
| nios2_function_arg (cumulative_args_t cum_v, enum machine_mode mode, |
| const_tree type ATTRIBUTE_UNUSED, |
| bool named ATTRIBUTE_UNUSED) |
| { |
| CUMULATIVE_ARGS *cum = get_cumulative_args (cum_v); |
| rtx return_rtx = NULL_RTX; |
| |
| if (cum->regs_used < NUM_ARG_REGS) |
| return_rtx = gen_rtx_REG (mode, FIRST_ARG_REGNO + cum->regs_used); |
| |
| return return_rtx; |
| } |
| |
| /* Return number of bytes, at the beginning of the argument, that must be |
| put in registers. 0 is the argument is entirely in registers or entirely |
| in memory. */ |
| |
| static int |
| nios2_arg_partial_bytes (cumulative_args_t cum_v, |
| enum machine_mode mode, tree type ATTRIBUTE_UNUSED, |
| bool named ATTRIBUTE_UNUSED) |
| { |
| CUMULATIVE_ARGS *cum = get_cumulative_args (cum_v); |
| HOST_WIDE_INT param_size; |
| |
| if (mode == BLKmode) |
| { |
| param_size = int_size_in_bytes (type); |
| gcc_assert (param_size >= 0); |
| } |
| else |
| param_size = GET_MODE_SIZE (mode); |
| |
| /* Convert to words (round up). */ |
| param_size = (UNITS_PER_WORD - 1 + param_size) / UNITS_PER_WORD; |
| |
| if (cum->regs_used < NUM_ARG_REGS |
| && cum->regs_used + param_size > NUM_ARG_REGS) |
| return (NUM_ARG_REGS - cum->regs_used) * UNITS_PER_WORD; |
| |
| return 0; |
| } |
| |
| /* Update the data in CUM to advance over an argument of mode MODE |
| and data type TYPE; TYPE is null for libcalls where that information |
| may not be available. */ |
| |
| static void |
| nios2_function_arg_advance (cumulative_args_t cum_v, enum machine_mode mode, |
| const_tree type ATTRIBUTE_UNUSED, |
| bool named ATTRIBUTE_UNUSED) |
| { |
| CUMULATIVE_ARGS *cum = get_cumulative_args (cum_v); |
| HOST_WIDE_INT param_size; |
| |
| if (mode == BLKmode) |
| { |
| param_size = int_size_in_bytes (type); |
| gcc_assert (param_size >= 0); |
| } |
| else |
| param_size = GET_MODE_SIZE (mode); |
| |
| /* Convert to words (round up). */ |
| param_size = (UNITS_PER_WORD - 1 + param_size) / UNITS_PER_WORD; |
| |
| if (cum->regs_used + param_size > NUM_ARG_REGS) |
| cum->regs_used = NUM_ARG_REGS; |
| else |
| cum->regs_used += param_size; |
| } |
| |
| enum direction |
| nios2_function_arg_padding (enum machine_mode mode, const_tree type) |
| { |
| /* On little-endian targets, the first byte of every stack argument |
| is passed in the first byte of the stack slot. */ |
| if (!BYTES_BIG_ENDIAN) |
| return upward; |
| |
| /* Otherwise, integral types are padded downward: the last byte of a |
| stack argument is passed in the last byte of the stack slot. */ |
| if (type != 0 |
| ? INTEGRAL_TYPE_P (type) || POINTER_TYPE_P (type) |
| : GET_MODE_CLASS (mode) == MODE_INT) |
| return downward; |
| |
| /* Arguments smaller than a stack slot are padded downward. */ |
| if (mode != BLKmode) |
| return (GET_MODE_BITSIZE (mode) >= PARM_BOUNDARY) ? upward : downward; |
| |
| return ((int_size_in_bytes (type) >= (PARM_BOUNDARY / BITS_PER_UNIT)) |
| ? upward : downward); |
| } |
| |
| enum direction |
| nios2_block_reg_padding (enum machine_mode mode, tree type, |
| int first ATTRIBUTE_UNUSED) |
| { |
| return nios2_function_arg_padding (mode, type); |
| } |
| |
| /* Emit RTL insns to initialize the variable parts of a trampoline. |
| FNADDR is an RTX for the address of the function's pure code. |
| CXT is an RTX for the static chain value for the function. |
| On Nios II, we handle this by a library call. */ |
| static void |
| nios2_trampoline_init (rtx m_tramp, tree fndecl, rtx cxt) |
| { |
| rtx fnaddr = XEXP (DECL_RTL (fndecl), 0); |
| rtx ctx_reg = force_reg (Pmode, cxt); |
| rtx addr = force_reg (Pmode, XEXP (m_tramp, 0)); |
| |
| emit_library_call (gen_rtx_SYMBOL_REF (Pmode, "__trampoline_setup"), |
| LCT_NORMAL, VOIDmode, 3, addr, Pmode, fnaddr, Pmode, |
| ctx_reg, Pmode); |
| } |
| |
| /* Implement TARGET_FUNCTION_VALUE. */ |
| static rtx |
| nios2_function_value (const_tree ret_type, const_tree fn ATTRIBUTE_UNUSED, |
| bool outgoing ATTRIBUTE_UNUSED) |
| { |
| return gen_rtx_REG (TYPE_MODE (ret_type), FIRST_RETVAL_REGNO); |
| } |
| |
| /* Implement TARGET_LIBCALL_VALUE. */ |
| static rtx |
| nios2_libcall_value (enum machine_mode mode, const_rtx fun ATTRIBUTE_UNUSED) |
| { |
| return gen_rtx_REG (mode, FIRST_RETVAL_REGNO); |
| } |
| |
| /* Implement TARGET_FUNCTION_VALUE_REGNO_P. */ |
| static bool |
| nios2_function_value_regno_p (const unsigned int regno) |
| { |
| return regno == FIRST_RETVAL_REGNO; |
| } |
| |
| /* Implement TARGET_RETURN_IN_MEMORY. */ |
| static bool |
| nios2_return_in_memory (const_tree type, const_tree fntype ATTRIBUTE_UNUSED) |
| { |
| return (int_size_in_bytes (type) > (2 * UNITS_PER_WORD) |
| || int_size_in_bytes (type) == -1); |
| } |
| |
| /* TODO: It may be possible to eliminate the copyback and implement |
| own va_arg type. */ |
| static void |
| nios2_setup_incoming_varargs (cumulative_args_t cum_v, |
| enum machine_mode mode, tree type, |
| int *pretend_size, int second_time) |
| { |
| CUMULATIVE_ARGS *cum = get_cumulative_args (cum_v); |
| CUMULATIVE_ARGS local_cum; |
| cumulative_args_t local_cum_v = pack_cumulative_args (&local_cum); |
| int regs_to_push; |
| int pret_size; |
| |
| local_cum = *cum; |
| nios2_function_arg_advance (local_cum_v, mode, type, 1); |
| |
| regs_to_push = NUM_ARG_REGS - local_cum.regs_used; |
| |
| if (!second_time && regs_to_push > 0) |
| { |
| rtx ptr = virtual_incoming_args_rtx; |
| rtx mem = gen_rtx_MEM (BLKmode, ptr); |
| emit_insn (gen_blockage ()); |
| move_block_from_reg (local_cum.regs_used + FIRST_ARG_REGNO, mem, |
| regs_to_push); |
| emit_insn (gen_blockage ()); |
| } |
| |
| pret_size = regs_to_push * UNITS_PER_WORD; |
| if (pret_size) |
| *pretend_size = pret_size; |
| } |
| |
| |
| |
| /* Init FPU builtins. */ |
| static void |
| nios2_init_fpu_builtins (int start_code) |
| { |
| tree fndecl; |
| char builtin_name[64] = "__builtin_custom_"; |
| unsigned int i, n = strlen ("__builtin_custom_"); |
| |
| for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) |
| { |
| snprintf (builtin_name + n, sizeof (builtin_name) - n, |
| "%s", N2FPU_NAME (i)); |
| fndecl = |
| add_builtin_function (builtin_name, nios2_ftype (N2FPU_FTCODE (i)), |
| start_code + i, BUILT_IN_MD, NULL, NULL_TREE); |
| nios2_register_builtin_fndecl (start_code + i, fndecl); |
| } |
| } |
| |
| /* Helper function for expanding FPU builtins. */ |
| static rtx |
| nios2_expand_fpu_builtin (tree exp, unsigned int code, rtx target) |
| { |
| struct expand_operand ops[MAX_RECOG_OPERANDS]; |
| enum insn_code icode = N2FPU_ICODE (code); |
| int nargs, argno, opno = 0; |
| int num_operands = N2FPU (code).num_operands; |
| enum machine_mode dst_mode = TYPE_MODE (TREE_TYPE (exp)); |
| bool has_target_p = (dst_mode != VOIDmode); |
| |
| if (N2FPU_N (code) < 0) |
| fatal_error ("Cannot call %<__builtin_custom_%s%> without specifying switch" |
| " %<-mcustom-%s%>", N2FPU_NAME (code), N2FPU_NAME (code)); |
| if (has_target_p) |
| create_output_operand (&ops[opno++], target, dst_mode); |
| else |
| /* Subtract away the count of the VOID return, mainly for fwrx/fwry. */ |
| num_operands -= 1; |
| nargs = call_expr_nargs (exp); |
| for (argno = 0; argno < nargs; argno++) |
| { |
| tree arg = CALL_EXPR_ARG (exp, argno); |
| create_input_operand (&ops[opno++], expand_normal (arg), |
| TYPE_MODE (TREE_TYPE (arg))); |
| } |
| if (!maybe_expand_insn (icode, num_operands, ops)) |
| { |
| error ("invalid argument to built-in function"); |
| return has_target_p ? gen_reg_rtx (ops[0].mode) : const0_rtx; |
| } |
| return has_target_p ? ops[0].value : const0_rtx; |
| } |
| |
| /* Nios II has custom instruction built-in functions of the forms: |
| __builtin_custom_n |
| __builtin_custom_nX |
| __builtin_custom_nXX |
| __builtin_custom_Xn |
| __builtin_custom_XnX |
| __builtin_custom_XnXX |
| |
| where each X could be either 'i' (int), 'f' (float), or 'p' (void*). |
| Therefore with 0-1 return values, and 0-2 arguments, we have a |
| total of (3 + 1) * (1 + 3 + 9) == 52 custom builtin functions. |
| */ |
| #define NUM_CUSTOM_BUILTINS ((3 + 1) * (1 + 3 + 9)) |
| static char custom_builtin_name[NUM_CUSTOM_BUILTINS][5]; |
| |
| static void |
| nios2_init_custom_builtins (int start_code) |
| { |
| tree builtin_ftype, ret_type, fndecl; |
| char builtin_name[32] = "__builtin_custom_"; |
| int n = strlen ("__builtin_custom_"); |
| int builtin_code = 0; |
| int lhs, rhs1, rhs2; |
| |
| struct { tree type; const char *c; } op[4]; |
| /* z */ op[0].c = ""; op[0].type = NULL_TREE; |
| /* f */ op[1].c = "f"; op[1].type = float_type_node; |
| /* i */ op[2].c = "i"; op[2].type = integer_type_node; |
| /* p */ op[3].c = "p"; op[3].type = ptr_type_node; |
| |
| /* We enumerate through the possible operand types to create all the |
| __builtin_custom_XnXX function tree types. Note that these may slightly |
| overlap with the function types created for other fixed builtins. */ |
| |
| for (lhs = 0; lhs < 4; lhs++) |
| for (rhs1 = 0; rhs1 < 4; rhs1++) |
| for (rhs2 = 0; rhs2 < 4; rhs2++) |
| { |
| if (rhs1 == 0 && rhs2 != 0) |
| continue; |
| ret_type = (op[lhs].type ? op[lhs].type : void_type_node); |
| builtin_ftype |
| = build_function_type_list (ret_type, integer_type_node, |
| op[rhs1].type, op[rhs2].type, |
| NULL_TREE); |
| snprintf (builtin_name + n, 32 - n, "%sn%s%s", |
| op[lhs].c, op[rhs1].c, op[rhs2].c); |
| /* Save copy of parameter string into custom_builtin_name[]. */ |
| strncpy (custom_builtin_name[builtin_code], builtin_name + n, 5); |
| fndecl = |
| add_builtin_function (builtin_name, builtin_ftype, |
| start_code + builtin_code, |
| BUILT_IN_MD, NULL, NULL_TREE); |
| nios2_register_builtin_fndecl (start_code + builtin_code, fndecl); |
| builtin_code += 1; |
| } |
| } |
| |
| /* Helper function for expanding custom builtins. */ |
| static rtx |
| nios2_expand_custom_builtin (tree exp, unsigned int index, rtx target) |
| { |
| bool has_target_p = (TREE_TYPE (exp) != void_type_node); |
| enum machine_mode tmode = VOIDmode; |
| int nargs, argno; |
| rtx value, insn, unspec_args[3]; |
| tree arg; |
| |
| /* XnXX form. */ |
| if (has_target_p) |
| { |
| tmode = TYPE_MODE (TREE_TYPE (exp)); |
| if (!target || GET_MODE (target) != tmode |
| || !REG_P (target)) |
| target = gen_reg_rtx (tmode); |
| } |
| |
| nargs = call_expr_nargs (exp); |
| for (argno = 0; argno < nargs; argno++) |
| { |
| arg = CALL_EXPR_ARG (exp, argno); |
| value = expand_normal (arg); |
| unspec_args[argno] = value; |
| if (argno == 0) |
| { |
| if (!custom_insn_opcode (value, VOIDmode)) |
| error ("custom instruction opcode must be compile time " |
| "constant in the range 0-255 for __builtin_custom_%s", |
| custom_builtin_name[index]); |
| } |
| else |
| /* For other arguments, force into a register. */ |
| unspec_args[argno] = force_reg (TYPE_MODE (TREE_TYPE (arg)), |
| unspec_args[argno]); |
| } |
| /* Fill remaining unspec operands with zero. */ |
| for (; argno < 3; argno++) |
| unspec_args[argno] = const0_rtx; |
| |
| insn = (has_target_p |
| ? gen_rtx_SET (VOIDmode, target, |
| gen_rtx_UNSPEC_VOLATILE (tmode, |
| gen_rtvec_v (3, unspec_args), |
| UNSPECV_CUSTOM_XNXX)) |
| : gen_rtx_UNSPEC_VOLATILE (VOIDmode, gen_rtvec_v (3, unspec_args), |
| UNSPECV_CUSTOM_NXX)); |
| emit_insn (insn); |
| return has_target_p ? target : const0_rtx; |
| } |
| |
| |
| |
| |
| /* Main definition of built-in functions. Nios II has a small number of fixed |
| builtins, plus a large number of FPU insn builtins, and builtins for |
| generating custom instructions. */ |
| |
| struct nios2_builtin_desc |
| { |
| enum insn_code icode; |
| enum nios2_ftcode ftype; |
| const char *name; |
| }; |
| |
| #define N2_BUILTINS \ |
| N2_BUILTIN_DEF (sync, N2_FTYPE_VOID_VOID) \ |
| N2_BUILTIN_DEF (ldbio, N2_FTYPE_SI_CVPTR) \ |
| N2_BUILTIN_DEF (ldbuio, N2_FTYPE_UI_CVPTR) \ |
| N2_BUILTIN_DEF (ldhio, N2_FTYPE_SI_CVPTR) \ |
| N2_BUILTIN_DEF (ldhuio, N2_FTYPE_UI_CVPTR) \ |
| N2_BUILTIN_DEF (ldwio, N2_FTYPE_SI_CVPTR) \ |
| N2_BUILTIN_DEF (stbio, N2_FTYPE_VOID_VPTR_SI) \ |
| N2_BUILTIN_DEF (sthio, N2_FTYPE_VOID_VPTR_SI) \ |
| N2_BUILTIN_DEF (stwio, N2_FTYPE_VOID_VPTR_SI) \ |
| N2_BUILTIN_DEF (rdctl, N2_FTYPE_SI_SI) \ |
| N2_BUILTIN_DEF (wrctl, N2_FTYPE_VOID_SI_SI) |
| |
| enum nios2_builtin_code { |
| #define N2_BUILTIN_DEF(name, ftype) NIOS2_BUILTIN_ ## name, |
| N2_BUILTINS |
| #undef N2_BUILTIN_DEF |
| NUM_FIXED_NIOS2_BUILTINS |
| }; |
| |
| static const struct nios2_builtin_desc nios2_builtins[] = { |
| #define N2_BUILTIN_DEF(name, ftype) \ |
| { CODE_FOR_ ## name, ftype, "__builtin_" #name }, |
| N2_BUILTINS |
| #undef N2_BUILTIN_DEF |
| }; |
| |
| /* Start/ends of FPU/custom insn builtin index ranges. */ |
| static unsigned int nios2_fpu_builtin_base; |
| static unsigned int nios2_custom_builtin_base; |
| static unsigned int nios2_custom_builtin_end; |
| |
| /* Implement TARGET_INIT_BUILTINS. */ |
| static void |
| nios2_init_builtins (void) |
| { |
| unsigned int i; |
| |
| /* Initialize fixed builtins. */ |
| for (i = 0; i < ARRAY_SIZE (nios2_builtins); i++) |
| { |
| const struct nios2_builtin_desc *d = &nios2_builtins[i]; |
| tree fndecl = |
| add_builtin_function (d->name, nios2_ftype (d->ftype), i, |
| BUILT_IN_MD, NULL, NULL); |
| nios2_register_builtin_fndecl (i, fndecl); |
| } |
| |
| /* Initialize FPU builtins. */ |
| nios2_fpu_builtin_base = ARRAY_SIZE (nios2_builtins); |
| nios2_init_fpu_builtins (nios2_fpu_builtin_base); |
| |
| /* Initialize custom insn builtins. */ |
| nios2_custom_builtin_base |
| = nios2_fpu_builtin_base + ARRAY_SIZE (nios2_fpu_insn); |
| nios2_custom_builtin_end |
| = nios2_custom_builtin_base + NUM_CUSTOM_BUILTINS; |
| nios2_init_custom_builtins (nios2_custom_builtin_base); |
| } |
| |
| /* Array of fndecls for TARGET_BUILTIN_DECL. */ |
| #define NIOS2_NUM_BUILTINS \ |
| (ARRAY_SIZE (nios2_builtins) + ARRAY_SIZE (nios2_fpu_insn) + NUM_CUSTOM_BUILTINS) |
| static GTY(()) tree nios2_builtin_decls[NIOS2_NUM_BUILTINS]; |
| |
| static void |
| nios2_register_builtin_fndecl (unsigned code, tree fndecl) |
| { |
| nios2_builtin_decls[code] = fndecl; |
| } |
| |
| /* Implement TARGET_BUILTIN_DECL. */ |
| static tree |
| nios2_builtin_decl (unsigned code, bool initialize_p ATTRIBUTE_UNUSED) |
| { |
| gcc_assert (nios2_custom_builtin_end == ARRAY_SIZE (nios2_builtin_decls)); |
| |
| if (code >= nios2_custom_builtin_end) |
| return error_mark_node; |
| |
| if (code >= nios2_fpu_builtin_base |
| && code < nios2_custom_builtin_base |
| && ! N2FPU_ENABLED_P (code - nios2_fpu_builtin_base)) |
| return error_mark_node; |
| |
| return nios2_builtin_decls[code]; |
| } |
| |
| |
| /* Low-level built-in expand routine. */ |
| static rtx |
| nios2_expand_builtin_insn (const struct nios2_builtin_desc *d, int n, |
| struct expand_operand *ops, bool has_target_p) |
| { |
| if (maybe_expand_insn (d->icode, n, ops)) |
| return has_target_p ? ops[0].value : const0_rtx; |
| else |
| { |
| error ("invalid argument to built-in function %s", d->name); |
| return has_target_p ? gen_reg_rtx (ops[0].mode) : const0_rtx; |
| } |
| } |
| |
| /* Expand ldio/stio form load-store instruction builtins. */ |
| static rtx |
| nios2_expand_ldstio_builtin (tree exp, rtx target, |
| const struct nios2_builtin_desc *d) |
| { |
| bool has_target_p; |
| rtx addr, mem, val; |
| struct expand_operand ops[MAX_RECOG_OPERANDS]; |
| enum machine_mode mode = insn_data[d->icode].operand[0].mode; |
| |
| addr = expand_normal (CALL_EXPR_ARG (exp, 0)); |
| mem = gen_rtx_MEM (mode, addr); |
| |
| if (insn_data[d->icode].operand[0].allows_mem) |
| { |
| /* stxio. */ |
| val = expand_normal (CALL_EXPR_ARG (exp, 1)); |
| if (CONST_INT_P (val)) |
| val = force_reg (mode, gen_int_mode (INTVAL (val), mode)); |
| val = simplify_gen_subreg (mode, val, GET_MODE (val), 0); |
| create_output_operand (&ops[0], mem, mode); |
| create_input_operand (&ops[1], val, mode); |
| has_target_p = false; |
| } |
| else |
| { |
| /* ldxio. */ |
| create_output_operand (&ops[0], target, mode); |
| create_input_operand (&ops[1], mem, mode); |
| has_target_p = true; |
| } |
| return nios2_expand_builtin_insn (d, 2, ops, has_target_p); |
| } |
| |
| /* Expand rdctl/wrctl builtins. */ |
| static rtx |
| nios2_expand_rdwrctl_builtin (tree exp, rtx target, |
| const struct nios2_builtin_desc *d) |
| { |
| bool has_target_p = (insn_data[d->icode].operand[0].predicate |
| == register_operand); |
| rtx ctlcode = expand_normal (CALL_EXPR_ARG (exp, 0)); |
| struct expand_operand ops[MAX_RECOG_OPERANDS]; |
| if (!rdwrctl_operand (ctlcode, VOIDmode)) |
| { |
| error ("Control register number must be in range 0-31 for %s", |
| d->name); |
| return has_target_p ? gen_reg_rtx (SImode) : const0_rtx; |
| } |
| if (has_target_p) |
| { |
| create_output_operand (&ops[0], target, SImode); |
| create_integer_operand (&ops[1], INTVAL (ctlcode)); |
| } |
| else |
| { |
| rtx val = expand_normal (CALL_EXPR_ARG (exp, 1)); |
| create_integer_operand (&ops[0], INTVAL (ctlcode)); |
| create_input_operand (&ops[1], val, SImode); |
| } |
| return nios2_expand_builtin_insn (d, 2, ops, has_target_p); |
| } |
| |
| /* Implement TARGET_EXPAND_BUILTIN. Expand an expression EXP that calls |
| a built-in function, with result going to TARGET if that's convenient |
| (and in mode MODE if that's convenient). |
| SUBTARGET may be used as the target for computing one of EXP's operands. |
| IGNORE is nonzero if the value is to be ignored. */ |
| |
| static rtx |
| nios2_expand_builtin (tree exp, rtx target, rtx subtarget ATTRIBUTE_UNUSED, |
| enum machine_mode mode ATTRIBUTE_UNUSED, |
| int ignore ATTRIBUTE_UNUSED) |
| { |
| tree fndecl = TREE_OPERAND (CALL_EXPR_FN (exp), 0); |
| unsigned int fcode = DECL_FUNCTION_CODE (fndecl); |
| |
| if (fcode < nios2_fpu_builtin_base) |
| { |
| const struct nios2_builtin_desc *d = &nios2_builtins[fcode]; |
| |
| switch (fcode) |
| { |
| case NIOS2_BUILTIN_sync: |
| emit_insn (gen_sync ()); |
| return const0_rtx; |
| |
| case NIOS2_BUILTIN_ldbio: |
| case NIOS2_BUILTIN_ldbuio: |
| case NIOS2_BUILTIN_ldhio: |
| case NIOS2_BUILTIN_ldhuio: |
| case NIOS2_BUILTIN_ldwio: |
| case NIOS2_BUILTIN_stbio: |
| case NIOS2_BUILTIN_sthio: |
| case NIOS2_BUILTIN_stwio: |
| return nios2_expand_ldstio_builtin (exp, target, d); |
| |
| case NIOS2_BUILTIN_rdctl: |
| case NIOS2_BUILTIN_wrctl: |
| return nios2_expand_rdwrctl_builtin (exp, target, d); |
| |
| default: |
| gcc_unreachable (); |
| } |
| } |
| else if (fcode < nios2_custom_builtin_base) |
| /* FPU builtin range. */ |
| return nios2_expand_fpu_builtin (exp, fcode - nios2_fpu_builtin_base, |
| target); |
| else if (fcode < nios2_custom_builtin_end) |
| /* Custom insn builtin range. */ |
| return nios2_expand_custom_builtin (exp, fcode - nios2_custom_builtin_base, |
| target); |
| else |
| gcc_unreachable (); |
| } |
| |
| /* Implement TARGET_INIT_LIBFUNCS. */ |
| static void |
| nios2_init_libfuncs (void) |
| { |
| /* For Linux, we have access to kernel support for atomic operations. */ |
| if (TARGET_LINUX_ABI) |
| init_sync_libfuncs (UNITS_PER_WORD); |
| } |
| |
| |
| |
| /* Register a custom code use, and signal error if a conflict was found. */ |
| static void |
| nios2_register_custom_code (unsigned int N, enum nios2_ccs_code status, |
| int index) |
| { |
| gcc_assert (N <= 255); |
| |
| if (status == CCS_FPU) |
| { |
| if (custom_code_status[N] == CCS_FPU && index != custom_code_index[N]) |
| { |
| custom_code_conflict = true; |
| error ("switch %<-mcustom-%s%> conflicts with switch %<-mcustom-%s%>", |
| N2FPU_NAME (custom_code_index[N]), N2FPU_NAME (index)); |
| } |
| else if (custom_code_status[N] == CCS_BUILTIN_CALL) |
| { |
| custom_code_conflict = true; |
| error ("call to %<__builtin_custom_%s%> conflicts with switch " |
| "%<-mcustom-%s%>", custom_builtin_name[custom_code_index[N]], |
| N2FPU_NAME (index)); |
| } |
| } |
| else if (status == CCS_BUILTIN_CALL) |
| { |
| if (custom_code_status[N] == CCS_FPU) |
| { |
| custom_code_conflict = true; |
| error ("call to %<__builtin_custom_%s%> conflicts with switch " |
| "%<-mcustom-%s%>", custom_builtin_name[index], |
| N2FPU_NAME (custom_code_index[N])); |
| } |
| else |
| { |
| /* Note that code conflicts between different __builtin_custom_xnxx |
| calls are not checked. */ |
| } |
| } |
| else |
| gcc_unreachable (); |
| |
| custom_code_status[N] = status; |
| custom_code_index[N] = index; |
| } |
| |
| /* Mark a custom code as not in use. */ |
| static void |
| nios2_deregister_custom_code (unsigned int N) |
| { |
| if (N <= 255) |
| { |
| custom_code_status[N] = CCS_UNUSED; |
| custom_code_index[N] = 0; |
| } |
| } |
| |
| /* Target attributes can affect per-function option state, so we need to |
| save/restore the custom code tracking info using the |
| TARGET_OPTION_SAVE/TARGET_OPTION_RESTORE hooks. */ |
| |
| static void |
| nios2_option_save (struct cl_target_option *ptr, |
| struct gcc_options *opts ATTRIBUTE_UNUSED) |
| { |
| unsigned int i; |
| for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) |
| ptr->saved_fpu_custom_code[i] = N2FPU_N (i); |
| memcpy (ptr->saved_custom_code_status, custom_code_status, |
| sizeof (custom_code_status)); |
| memcpy (ptr->saved_custom_code_index, custom_code_index, |
| sizeof (custom_code_index)); |
| } |
| |
| static void |
| nios2_option_restore (struct gcc_options *opts ATTRIBUTE_UNUSED, |
| struct cl_target_option *ptr) |
| { |
| unsigned int i; |
| for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) |
| N2FPU_N (i) = ptr->saved_fpu_custom_code[i]; |
| memcpy (custom_code_status, ptr->saved_custom_code_status, |
| sizeof (custom_code_status)); |
| memcpy (custom_code_index, ptr->saved_custom_code_index, |
| sizeof (custom_code_index)); |
| } |
| |
| /* Inner function to process the attribute((target(...))), take an argument and |
| set the current options from the argument. If we have a list, recursively |
| go over the list. */ |
| |
| static bool |
| nios2_valid_target_attribute_rec (tree args) |
| { |
| if (TREE_CODE (args) == TREE_LIST) |
| { |
| bool ret = true; |
| for (; args; args = TREE_CHAIN (args)) |
| if (TREE_VALUE (args) |
| && !nios2_valid_target_attribute_rec (TREE_VALUE (args))) |
| ret = false; |
| return ret; |
| } |
| else if (TREE_CODE (args) == STRING_CST) |
| { |
| char *argstr = ASTRDUP (TREE_STRING_POINTER (args)); |
| while (argstr && *argstr != '\0') |
| { |
| bool no_opt = false, end_p = false; |
| char *eq = NULL, *p; |
| while (ISSPACE (*argstr)) |
| argstr++; |
| p = argstr; |
| while (*p != '\0' && *p != ',') |
| { |
| if (!eq && *p == '=') |
| eq = p; |
| ++p; |
| } |
| if (*p == '\0') |
| end_p = true; |
| else |
| *p = '\0'; |
| if (eq) *eq = '\0'; |
| |
| if (!strncmp (argstr, "no-", 3)) |
| { |
| no_opt = true; |
| argstr += 3; |
| } |
| if (!strncmp (argstr, "custom-fpu-cfg", 14)) |
| { |
| char *end_eq = p; |
| if (no_opt) |
| { |
| error ("custom-fpu-cfg option does not support %<no-%>"); |
| return false; |
| } |
| if (!eq) |
| { |
| error ("custom-fpu-cfg option requires configuration" |
| " argument"); |
| return false; |
| } |
| /* Increment and skip whitespace. */ |
| while (ISSPACE (*(++eq))) ; |
| /* Decrement and skip to before any trailing whitespace. */ |
| while (ISSPACE (*(--end_eq))) ; |
| |
| nios2_handle_custom_fpu_cfg (eq, end_eq + 1, true); |
| } |
| else if (!strncmp (argstr, "custom-", 7)) |
| { |
| int code = -1; |
| unsigned int i; |
| for (i = 0; i < ARRAY_SIZE (nios2_fpu_insn); i++) |
| if (!strncmp (argstr + 7, N2FPU_NAME (i), |
| strlen (N2FPU_NAME (i)))) |
| { |
| /* Found insn. */ |
| code = i; |
| break; |
| } |
| if (code >= 0) |
| { |
| if (no_opt) |
| { |
| if (eq) |
| { |
| error ("%<no-custom-%s%> does not accept arguments", |
| N2FPU_NAME (code)); |
| return false; |
| } |
| /* Disable option by setting to -1. */ |
| nios2_deregister_custom_code (N2FPU_N (code)); |
| N2FPU_N (code) = -1; |
| } |
| else |
| { |
| char *t; |
| if (eq) |
| while (ISSPACE (*(++eq))) ; |
| if (!eq || eq == p) |
| { |
| error ("%<custom-%s=%> requires argument", |
| N2FPU_NAME (code)); |
| return false; |
| } |
| for (t = eq; t != p; ++t) |
| { |
| if (ISSPACE (*t)) |
| continue; |
| if (!ISDIGIT (*t)) |
| { |
| error ("`custom-%s=' argument requires " |
| "numeric digits", N2FPU_NAME (code)); |
| return false; |
| } |
| } |
| /* Set option to argument. */ |
| N2FPU_N (code) = atoi (eq); |
| nios2_handle_custom_fpu_insn_option (code); |
| } |
| } |
| else |
| { |
| error ("%<custom-%s=%> is not recognised as FPU instruction", |
| argstr + 7); |
| return false; |
| } |
| } |
| else |
| { |
| error ("%<%s%> is unknown", argstr); |
| return false; |
| } |
| |
| if (end_p) |
| break; |
| else |
| argstr = p + 1; |
| } |
| return true; |
| } |
| else |
| gcc_unreachable (); |
| } |
| |
| /* Return a TARGET_OPTION_NODE tree of the target options listed or NULL. */ |
| |
| static tree |
| nios2_valid_target_attribute_tree (tree args) |
| { |
| if (!nios2_valid_target_attribute_rec (args)) |
| return NULL_TREE; |
| nios2_custom_check_insns (); |
| return build_target_option_node (&global_options); |
| } |
| |
| /* Hook to validate attribute((target("string"))). */ |
| |
| static bool |
| nios2_valid_target_attribute_p (tree fndecl, tree ARG_UNUSED (name), |
| tree args, int ARG_UNUSED (flags)) |
| { |
| struct cl_target_option cur_target; |
| bool ret = true; |
| tree old_optimize = build_optimization_node (&global_options); |
| tree new_target, new_optimize; |
| tree func_optimize = DECL_FUNCTION_SPECIFIC_OPTIMIZATION (fndecl); |
| |
| /* If the function changed the optimization levels as well as setting target |
| options, start with the optimizations specified. */ |
| if (func_optimize && func_optimize != old_optimize) |
| cl_optimization_restore (&global_options, |
| TREE_OPTIMIZATION (func_optimize)); |
| |
| /* The target attributes may also change some optimization flags, so update |
| the optimization options if necessary. */ |
| cl_target_option_save (&cur_target, &global_options); |
| new_target = nios2_valid_target_attribute_tree (args); |
| new_optimize = build_optimization_node (&global_options); |
| |
| if (!new_target) |
| ret = false; |
| |
| else if (fndecl) |
| { |
| DECL_FUNCTION_SPECIFIC_TARGET (fndecl) = new_target; |
| |
| if (old_optimize != new_optimize) |
| DECL_FUNCTION_SPECIFIC_OPTIMIZATION (fndecl) = new_optimize; |
| } |
| |
| cl_target_option_restore (&global_options, &cur_target); |
| |
| if (old_optimize != new_optimize) |
| cl_optimization_restore (&global_options, |
| TREE_OPTIMIZATION (old_optimize)); |
| return ret; |
| } |
| |
| /* Remember the last target of nios2_set_current_function. */ |
| static GTY(()) tree nios2_previous_fndecl; |
| |
| /* Establish appropriate back-end context for processing the function |
| FNDECL. The argument might be NULL to indicate processing at top |
| level, outside of any function scope. */ |
| static void |
| nios2_set_current_function (tree fndecl) |
| { |
| tree old_tree = (nios2_previous_fndecl |
| ? DECL_FUNCTION_SPECIFIC_TARGET (nios2_previous_fndecl) |
| : NULL_TREE); |
| |
| tree new_tree = (fndecl |
| ? DECL_FUNCTION_SPECIFIC_TARGET (fndecl) |
| : NULL_TREE); |
| |
| if (fndecl && fndecl != nios2_previous_fndecl) |
| { |
| nios2_previous_fndecl = fndecl; |
| if (old_tree == new_tree) |
| ; |
| |
| else if (new_tree) |
| { |
| cl_target_option_restore (&global_options, |
| TREE_TARGET_OPTION (new_tree)); |
| target_reinit (); |
| } |
| |
| else if (old_tree) |
| { |
| struct cl_target_option *def |
| = TREE_TARGET_OPTION (target_option_current_node); |
| |
| cl_target_option_restore (&global_options, def); |
| target_reinit (); |
| } |
| } |
| } |
| |
| /* Hook to validate the current #pragma GCC target and set the FPU custom |
| code option state. If ARGS is NULL, then POP_TARGET is used to reset |
| the options. */ |
| static bool |
| nios2_pragma_target_parse (tree args, tree pop_target) |
| { |
| tree cur_tree; |
| if (! args) |
| { |
| cur_tree = ((pop_target) |
| ? pop_target |
| : target_option_default_node); |
| cl_target_option_restore (&global_options, |
| TREE_TARGET_OPTION (cur_tree)); |
| } |
| else |
| { |
| cur_tree = nios2_valid_target_attribute_tree (args); |
| if (!cur_tree) |
| return false; |
| } |
| |
| target_option_current_node = cur_tree; |
| return true; |
| } |
| |
| /* Implement TARGET_MERGE_DECL_ATTRIBUTES. |
| We are just using this hook to add some additional error checking to |
| the default behavior. GCC does not provide a target hook for merging |
| the target options, and only correctly handles merging empty vs non-empty |
| option data; see merge_decls() in c-decl.c. |
| So here we require either that at least one of the decls has empty |
| target options, or that the target options/data be identical. */ |
| static tree |
| nios2_merge_decl_attributes (tree olddecl, tree newdecl) |
| { |
| tree oldopts = lookup_attribute ("target", DECL_ATTRIBUTES (olddecl)); |
| tree newopts = lookup_attribute ("target", DECL_ATTRIBUTES (newdecl)); |
| if (newopts && oldopts && newopts != oldopts) |
| { |
| tree oldtree = DECL_FUNCTION_SPECIFIC_TARGET (olddecl); |
| tree newtree = DECL_FUNCTION_SPECIFIC_TARGET (newdecl); |
| if (oldtree && newtree && oldtree != newtree) |
| { |
| struct cl_target_option *olddata = TREE_TARGET_OPTION (oldtree); |
| struct cl_target_option *newdata = TREE_TARGET_OPTION (newtree); |
| if (olddata != newdata |
| && memcmp (olddata, newdata, sizeof (struct cl_target_option))) |
| error ("%qE redeclared with conflicting %qs attributes", |
| DECL_NAME (newdecl), "target"); |
| } |
| } |
| return merge_attributes (DECL_ATTRIBUTES (olddecl), |
| DECL_ATTRIBUTES (newdecl)); |
| } |
| |
| |
| /* Initialize the GCC target structure. */ |
| #undef TARGET_ASM_FUNCTION_PROLOGUE |
| #define TARGET_ASM_FUNCTION_PROLOGUE nios2_asm_function_prologue |
| |
| #undef TARGET_IN_SMALL_DATA_P |
| #define TARGET_IN_SMALL_DATA_P nios2_in_small_data_p |
| |
| #undef TARGET_SECTION_TYPE_FLAGS |
| #define TARGET_SECTION_TYPE_FLAGS nios2_section_type_flags |
| |
| #undef TARGET_INIT_BUILTINS |
| #define TARGET_INIT_BUILTINS nios2_init_builtins |
| #undef TARGET_EXPAND_BUILTIN |
| #define TARGET_EXPAND_BUILTIN nios2_expand_builtin |
| #undef TARGET_BUILTIN_DECL |
| #define TARGET_BUILTIN_DECL nios2_builtin_decl |
| |
| #undef TARGET_INIT_LIBFUNCS |
| #define TARGET_INIT_LIBFUNCS nios2_init_libfuncs |
| |
| #undef TARGET_FUNCTION_OK_FOR_SIBCALL |
| #define TARGET_FUNCTION_OK_FOR_SIBCALL hook_bool_tree_tree_true |
| |
| #undef TARGET_CAN_ELIMINATE |
| #define TARGET_CAN_ELIMINATE nios2_can_eliminate |
| |
| #undef TARGET_FUNCTION_ARG |
| #define TARGET_FUNCTION_ARG nios2_function_arg |
| |
| #undef TARGET_FUNCTION_ARG_ADVANCE |
| #define TARGET_FUNCTION_ARG_ADVANCE nios2_function_arg_advance |
| |
| #undef TARGET_ARG_PARTIAL_BYTES |
| #define TARGET_ARG_PARTIAL_BYTES nios2_arg_partial_bytes |
| |
| #undef TARGET_TRAMPOLINE_INIT |
| #define TARGET_TRAMPOLINE_INIT nios2_trampoline_init |
| |
| #undef TARGET_FUNCTION_VALUE |
| #define TARGET_FUNCTION_VALUE nios2_function_value |
| |
| #undef TARGET_LIBCALL_VALUE |
| #define TARGET_LIBCALL_VALUE nios2_libcall_value |
| |
| #undef TARGET_FUNCTION_VALUE_REGNO_P |
| #define TARGET_FUNCTION_VALUE_REGNO_P nios2_function_value_regno_p |
| |
| #undef TARGET_RETURN_IN_MEMORY |
| #define TARGET_RETURN_IN_MEMORY nios2_return_in_memory |
| |
| #undef TARGET_PROMOTE_PROTOTYPES |
| #define TARGET_PROMOTE_PROTOTYPES hook_bool_const_tree_true |
| |
| #undef TARGET_SETUP_INCOMING_VARARGS |
| #define TARGET_SETUP_INCOMING_VARARGS nios2_setup_incoming_varargs |
| |
| #undef TARGET_MUST_PASS_IN_STACK |
| #define TARGET_MUST_PASS_IN_STACK must_pass_in_stack_var_size |
| |
| #undef TARGET_LEGITIMATE_CONSTANT_P |
| #define TARGET_LEGITIMATE_CONSTANT_P nios2_legitimate_constant_p |
| |
| #undef TARGET_LEGITIMIZE_ADDRESS |
| #define TARGET_LEGITIMIZE_ADDRESS nios2_legitimize_address |
| |
| #undef TARGET_DELEGITIMIZE_ADDRESS |
| #define TARGET_DELEGITIMIZE_ADDRESS nios2_delegitimize_address |
| |
| #undef TARGET_LEGITIMATE_ADDRESS_P |
| #define TARGET_LEGITIMATE_ADDRESS_P nios2_legitimate_address_p |
| |
| #undef TARGET_PREFERRED_RELOAD_CLASS |
| #define TARGET_PREFERRED_RELOAD_CLASS nios2_preferred_reload_class |
| |
| #undef TARGET_RTX_COSTS |
| #define TARGET_RTX_COSTS nios2_rtx_costs |
| |
| #undef TARGET_HAVE_TLS |
| #define TARGET_HAVE_TLS TARGET_LINUX_ABI |
| |
| #undef TARGET_CANNOT_FORCE_CONST_MEM |
| #define TARGET_CANNOT_FORCE_CONST_MEM nios2_cannot_force_const_mem |
| |
| #undef TARGET_ASM_OUTPUT_DWARF_DTPREL |
| #define TARGET_ASM_OUTPUT_DWARF_DTPREL nios2_output_dwarf_dtprel |
| |
| #undef TARGET_PRINT_OPERAND |
| #define TARGET_PRINT_OPERAND nios2_print_operand |
| |
| #undef TARGET_PRINT_OPERAND_ADDRESS |
| #define TARGET_PRINT_OPERAND_ADDRESS nios2_print_operand_address |
| |
| #undef TARGET_ASM_OUTPUT_ADDR_CONST_EXTRA |
| #define TARGET_ASM_OUTPUT_ADDR_CONST_EXTRA nios2_output_addr_const_extra |
| |
| #undef TARGET_OPTION_OVERRIDE |
| #define TARGET_OPTION_OVERRIDE nios2_option_override |
| |
| #undef TARGET_OPTION_SAVE |
| #define TARGET_OPTION_SAVE nios2_option_save |
| |
| #undef TARGET_OPTION_RESTORE |
| #define TARGET_OPTION_RESTORE nios2_option_restore |
| |
| #undef TARGET_SET_CURRENT_FUNCTION |
| #define TARGET_SET_CURRENT_FUNCTION nios2_set_current_function |
| |
| #undef TARGET_OPTION_VALID_ATTRIBUTE_P |
| #define TARGET_OPTION_VALID_ATTRIBUTE_P nios2_valid_target_attribute_p |
| |
| #undef TARGET_OPTION_PRAGMA_PARSE |
| #define TARGET_OPTION_PRAGMA_PARSE nios2_pragma_target_parse |
| |
| #undef TARGET_MERGE_DECL_ATTRIBUTES |
| #define TARGET_MERGE_DECL_ATTRIBUTES nios2_merge_decl_attributes |
| |
| struct gcc_target targetm = TARGET_INITIALIZER; |
| |
| #include "gt-nios2.h" |