;; Machine description of the Synopsys DesignWare ARC cpu for GNU C compiler
;; Copyright (C) 1994-2022 Free Software Foundation, Inc.
;; Sources derived from work done by Sankhya Technologies (www.sankhya.com) on
;; behalf of Synopsys Inc.
;; Position Independent Code support added,Code cleaned up,
;; Comments and Support For ARC700 instructions added by
;; Saurabh Verma (saurabh.verma@codito.com)
;; Ramana Radhakrishnan(ramana.radhakrishnan@codito.com)
;;
;; Performance improvements by
;; Joern Rennecke (joern.rennecke@embecosm.com)
;;
;; This file is part of GCC.
;; GCC is free software; you can redistribute it and/or modify
;; it under the terms of the GNU General Public License as published by
;; the Free Software Foundation; either version 3, or (at your option)
;; any later version.
;; GCC is distributed in the hope that it will be useful,
;; but WITHOUT ANY WARRANTY; without even the implied warranty of
;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
;; GNU General Public License for more details.
;; You should have received a copy of the GNU General Public License
;; along with GCC; see the file COPYING3. If not see
;; .
;; See file "rtl.def" for documentation on define_insn, match_*, et. al.
;; dest, src Two operand instruction's syntax
;; dest, src1, src2 Three operand instruction's syntax
;; ARC and ARCompact PREDICATES:
;;
;; comparison_operator LT, GT, LE, GE, LTU, GTU, LEU, GEU, EQ, NE
;; memory_operand memory [m]
;; immediate_operand immediate constant [IKLMNOP]
;; register_operand register [rq]
;; general_operand register, memory, constant [rqmIKLMNOP]
;; Note that the predicates are only used when selecting a pattern
;; to determine if an operand is valid.
;; The constraints then select which of the possible valid operands
;; is present (and guide register selection). The actual assembly
;; instruction is then selected on the basis of the constraints.
;; ARC and ARCompact CONSTRAINTS:
;;
;; b stack pointer r28
;; f frame pointer r27
;; Rgp global pointer r26
;; g general reg, memory, constant
;; m memory
;; p memory address
;; q registers commonly used in
;; 16-bit insns r0-r3, r12-r15
;; c core registers r0-r60, ap, pcl
;; r general registers r0-r28, blink, ap, pcl
;;
;; H fp 16-bit constant
;; I signed 12-bit immediate (for ARCompact)
;; K unsigned 3-bit immediate (for ARCompact)
;; L unsigned 6-bit immediate (for ARCompact)
;; M unsinged 5-bit immediate (for ARCompact)
;; O unsinged 7-bit immediate (for ARCompact)
;; P unsinged 8-bit immediate (for ARCompact)
;; N constant '1' (for ARCompact)
;; TODO:
;; -> prefetch instruction
;; -----------------------------------------------------------------------------
;; Include DFA scheduluers
(include ("arc600.md"))
(include ("arc700.md"))
(include ("arcEM.md"))
(include ("arcHS.md"))
(include ("arcHS4x.md"))
;; Predicates
(include ("predicates.md"))
(include ("constraints.md"))
;; -----------------------------------------------------------------------------
;; UNSPEC Usage:
;; ~~~~~~~~~~~~
;; -----------------------------------------------------------------------------
;; Symbolic name Value Desc.
;; -----------------------------------------------------------------------------
;; UNSPEC_PLT 3 symbol to be referenced through the PLT
;; UNSPEC_GOT 4 symbol to be rerenced through the GOT
;; UNSPEC_GOTOFF 5 Local symbol.To be referenced relative to the
;; GOTBASE.(Referenced as @GOTOFF)
;; UNSPEC_GOTOFFPC 6 Local symbol. To be referenced pc-relative.
;; ----------------------------------------------------------------------------
(define_c_enum "unspec" [
DUMMY_0
DUMMY_1
DUMMY_2
ARC_UNSPEC_PLT
ARC_UNSPEC_GOT
ARC_UNSPEC_GOTOFF
ARC_UNSPEC_GOTOFFPC
UNSPEC_TLS_GD
UNSPEC_TLS_LD
UNSPEC_TLS_IE
UNSPEC_TLS_OFF
UNSPEC_ARC_NORM
UNSPEC_ARC_NORMW
UNSPEC_ARC_SWAP
UNSPEC_ARC_DIVAW
UNSPEC_ARC_DIRECT
UNSPEC_ARC_LP
UNSPEC_ARC_CASESI
UNSPEC_ARC_FFS
UNSPEC_ARC_FLS
UNSPEC_ARC_MEMBAR
UNSPEC_ARC_DMACH
UNSPEC_ARC_DMACHU
UNSPEC_ARC_DMACWH
UNSPEC_ARC_DMACWHU
UNSPEC_ARC_DMPYWH
UNSPEC_ARC_QMACH
UNSPEC_ARC_QMACHU
UNSPEC_ARC_QMPYH
UNSPEC_ARC_QMPYHU
UNSPEC_ARC_VMAC2H
UNSPEC_ARC_VMAC2HU
UNSPEC_ARC_VMPY2H
UNSPEC_ARC_VMPY2HU
VUNSPEC_ARC_RTIE
VUNSPEC_ARC_SYNC
VUNSPEC_ARC_BRK
VUNSPEC_ARC_FLAG
VUNSPEC_ARC_SLEEP
VUNSPEC_ARC_SWI
VUNSPEC_ARC_CORE_READ
VUNSPEC_ARC_CORE_WRITE
VUNSPEC_ARC_LR
VUNSPEC_ARC_SR
VUNSPEC_ARC_TRAP_S
VUNSPEC_ARC_UNIMP_S
VUNSPEC_ARC_KFLAG
VUNSPEC_ARC_CLRI
VUNSPEC_ARC_SETI
VUNSPEC_ARC_NOP
VUNSPEC_ARC_STACK_IRQ
VUNSPEC_ARC_DEXCL
VUNSPEC_ARC_DEXCL_NORES
VUNSPEC_ARC_LR_HIGH
VUNSPEC_ARC_EX
VUNSPEC_ARC_CAS
VUNSPEC_ARC_SC
VUNSPEC_ARC_LL
VUNSPEC_ARC_BLOCKAGE
VUNSPEC_ARC_EH_RETURN
VUNSPEC_ARC_ARC600_RTIE
VUNSPEC_ARC_ARC600_STALL
VUNSPEC_ARC_LDDI
VUNSPEC_ARC_STDI
])
(define_constants
[(R0_REG 0)
(R1_REG 1)
(R2_REG 2)
(R3_REG 3)
(R4_REG 4)
(R9_REG 9)
(R10_REG 10)
(R12_REG 12)
(R15_REG 15)
(R16_REG 16)
(R25_REG 25)
(SP_REG 28)
(ILINK1_REG 29)
(ILINK2_REG 30)
(R30_REG 30)
(RETURN_ADDR_REGNUM 31)
(R32_REG 32)
(R33_REG 33)
(R34_REG 34)
(R35_REG 35)
(R36_REG 36)
(R37_REG 37)
(R38_REG 38)
(R39_REG 39)
(R40_REG 40)
(R41_REG 41)
(R42_REG 42)
(R43_REG 43)
(R44_REG 44)
(R45_REG 45)
(R46_REG 46)
(R47_REG 47)
(R48_REG 48)
(R49_REG 49)
(R50_REG 50)
(R51_REG 51)
(R52_REG 52)
(R53_REG 53)
(R54_REG 54)
(R55_REG 55)
(R56_REG 56)
(R57_REG 57)
(R58_REG 58)
(R59_REG 59)
(MUL64_OUT_REG 58)
(MUL32x16_REG 56)
(ARCV2_ACC 58)
(LP_COUNT 60)
(CC_REG 61)
(PCL_REG 63)
]
)
;; What is the insn_cost for this insn? The target hook can still override
;; this. For optimizing for size the "length" attribute is used instead.
(define_attr "cost" "" (const_int 0))
(define_attr "is_sfunc" "no,yes" (const_string "no"))
;; Insn type. Used to default other attribute values.
; While the attribute is_sfunc is set for any call of a special function,
; the instruction type sfunc is used only for the special call sequence
; that loads the (pc-relative) function address into r12 and then calls
; via r12.
(define_attr "type"
"move,load,store,cmove,unary,binary,compare,shift,uncond_branch,jump,branch,
brcc,brcc_no_delay_slot,call,sfunc,call_no_delay_slot,rtie,
multi,umulti, two_cycle_core,lr,sr,divaw,loop_setup,loop_end,return,
misc,spfp,dpfp_mult,dpfp_addsub,mulmac_600,cc_arith,
simd_vload, simd_vload128, simd_vstore, simd_vmove, simd_vmove_else_zero,
simd_vmove_with_acc, simd_varith_1cycle, simd_varith_2cycle,
simd_varith_with_acc, simd_vlogic, simd_vlogic_with_acc,
simd_vcompare, simd_vpermute, simd_vpack, simd_vpack_with_acc,
simd_valign, simd_valign_with_acc, simd_vcontrol,
simd_vspecial_3cycle, simd_vspecial_4cycle, simd_dma, mul16_em, div_rem,
fpu, fpu_fuse, fpu_sdiv, fpu_ddiv, fpu_cvt, block"
(cond [(eq_attr "is_sfunc" "yes")
(cond [(match_test "!TARGET_LONG_CALLS_SET && (!TARGET_MEDIUM_CALLS || GET_CODE (PATTERN (insn)) != COND_EXEC)") (const_string "call")
(match_test "flag_pic") (const_string "sfunc")]
(const_string "call_no_delay_slot"))]
(const_string "binary")))
;; The following three attributes are mixed case so that they can be
;; used conveniently with the CALL_ATTR macro.
(define_attr "is_CALL" "no,yes"
(cond [(eq_attr "is_sfunc" "yes") (const_string "yes")
(eq_attr "type" "call,call_no_delay_slot") (const_string "yes")]
(const_string "no")))
(define_attr "is_SIBCALL" "no,yes" (const_string "no"))
(define_attr "is_NON_SIBCALL" "no,yes"
(cond [(eq_attr "is_SIBCALL" "yes") (const_string "no")
(eq_attr "is_CALL" "yes") (const_string "yes")]
(const_string "no")))
;; true for compact instructions (those with _s suffix)
;; "maybe" means compact unless we conditionalize the insn.
(define_attr "iscompact" "true,maybe,true_limm,maybe_limm,false"
(cond [(eq_attr "type" "sfunc")
(const_string "maybe")]
(const_string "false")))
; Is there an instruction that we are actually putting into the delay slot?
(define_attr "delay_slot_filled" "no,yes"
(cond [(match_test "NEXT_INSN (PREV_INSN (insn)) == insn")
(const_string "no")
(match_test "!TARGET_AT_DBR_CONDEXEC
&& JUMP_P (insn)
&& INSN_ANNULLED_BRANCH_P (insn)
&& !INSN_FROM_TARGET_P (NEXT_INSN (insn))")
(const_string "no")]
(const_string "yes")))
; Is a delay slot present for purposes of shorten_branches?
; We have to take the length of this insn into account for forward branches
; even if we don't put the insn actually into a delay slot.
(define_attr "delay_slot_present" "no,yes"
(cond [(match_test "NEXT_INSN (PREV_INSN (insn)) == insn")
(const_string "no")]
(const_string "yes")))
; We can't use get_attr_length (NEXT_INSN (insn)) because this gives the
; length of a different insn with the same uid.
(define_attr "delay_slot_length" ""
(cond [(match_test "NEXT_INSN (PREV_INSN (insn)) == insn")
(const_int 0)]
(symbol_ref "get_attr_length (NEXT_INSN (PREV_INSN (insn)))
- get_attr_length (insn)")))
; for ARCv2 we need to disable/enable different instruction alternatives
(define_attr "cpu_facility" "std,av1,av2,fpx,cd"
(const_string "std"))
; We should consider all the instructions enabled until otherwise
(define_attr "enabled" "no,yes"
(cond [(and (eq_attr "cpu_facility" "av1")
(match_test "TARGET_V2"))
(const_string "no")
(and (eq_attr "cpu_facility" "av2")
(not (match_test "TARGET_V2")))
(const_string "no")
(and (eq_attr "cpu_facility" "fpx")
(match_test "TARGET_FP_DP_AX"))
(const_string "no")
(and (eq_attr "cpu_facility" "cd")
(not (and (match_test "TARGET_V2")
(match_test "TARGET_CODE_DENSITY"))))
(const_string "no")
]
(const_string "yes")))
(define_attr "predicable" "no,yes" (const_string "no"))
;; if 'predicable' were not so brain-dead, we would specify:
;; (cond [(eq_attr "cond" "!canuse") (const_string "no")
;; (eq_attr "iscompact" "maybe") (const_string "no")]
;; (const_string "yes"))
;; and then for everything but calls, we could just set the cond attribute.
;; Condition codes: this one is used by final_prescan_insn to speed up
;; conditionalizing instructions. It saves having to scan the rtl to see if
;; it uses or alters the condition codes.
;; USE: This insn uses the condition codes (eg: a conditional branch).
;; CANUSE: This insn can use the condition codes (for conditional execution).
;; SET: All condition codes are set by this insn.
;; SET_ZN: the Z and N flags are set by this insn.
;; SET_ZNC: the Z, N, and C flags are set by this insn.
;; CLOB: The condition codes are set to unknown values by this insn.
;; NOCOND: This insn can't use and doesn't affect the condition codes.
(define_attr "cond" "use,canuse,canuse_limm,canuse_limm_add,set,set_zn,clob,nocond"
(cond
[(and (eq_attr "predicable" "yes")
(eq_attr "is_sfunc" "no")
(eq_attr "delay_slot_filled" "no"))
(const_string "canuse")
(eq_attr "type" "call")
(cond [(eq_attr "delay_slot_filled" "yes") (const_string "nocond")
(match_test "!flag_pic") (const_string "canuse_limm")]
(const_string "nocond"))
(eq_attr "iscompact" "maybe,false")
(cond [ (and (eq_attr "type" "move")
(match_operand 1 "immediate_operand" ""))
(if_then_else
(ior (match_operand 1 "u6_immediate_operand" "")
(match_operand 1 "long_immediate_operand" ""))
(const_string "canuse")
(const_string "canuse_limm"))
(eq_attr "type" "binary")
(cond [(ne (symbol_ref "REGNO (operands[0])")
(symbol_ref "REGNO (operands[1])"))
(const_string "nocond")
(match_operand 2 "register_operand" "")
(const_string "canuse")
(match_operand 2 "u6_immediate_operand" "")
(const_string "canuse")
(match_operand 2 "long_immediate_operand" "")
(const_string "canuse")
(match_operand 2 "const_int_operand" "")
(const_string "canuse_limm")]
(const_string "nocond"))
(eq_attr "type" "compare")
(const_string "set")
(eq_attr "type" "cmove,branch")
(const_string "use")
(eq_attr "is_sfunc" "yes")
(cond [(match_test "(TARGET_MEDIUM_CALLS
&& !TARGET_LONG_CALLS_SET
&& flag_pic)")
(const_string "canuse_limm_add")
(match_test "(TARGET_MEDIUM_CALLS
&& !TARGET_LONG_CALLS_SET)")
(const_string "canuse_limm")]
(const_string "canuse"))
]
(const_string "nocond"))]
(cond [(eq_attr "type" "compare")
(const_string "set")
(eq_attr "type" "cmove,branch")
(const_string "use")
]
(const_string "nocond"))))
/* ??? Having all these patterns gives ifcvt more freedom to generate
inefficient code. It seem to operate on the premise that
register-register copies and registers are free. I see better code
with -fno-if-convert now than without. */
(define_cond_exec
[(match_operator 0 "proper_comparison_operator"
[(reg CC_REG) (const_int 0)])]
"true"
"")
;; Length (in # of bytes, long immediate constants counted too).
;; ??? There's a nasty interaction between the conditional execution fsm
;; and insn lengths: insns with shimm values cannot be conditionally executed.
(define_attr "length" ""
(cond
[(eq_attr "iscompact" "true")
(const_int 2)
(eq_attr "iscompact" "maybe")
(cond
[(eq_attr "type" "sfunc")
(cond [(match_test "GET_CODE (PATTERN (insn)) == COND_EXEC")
(const_int 12)]
(const_int 10))
(match_test "GET_CODE (PATTERN (insn)) == COND_EXEC") (const_int 4)
(match_test "find_reg_note (insn, REG_SAVE_NOTE, GEN_INT (1))")
(const_int 4)]
(const_int 2))
(eq_attr "iscompact" "true_limm")
(const_int 6)
(eq_attr "iscompact" "maybe_limm")
(cond [(match_test "GET_CODE (PATTERN (insn)) == COND_EXEC") (const_int 8)]
(const_int 6))
(eq_attr "type" "load")
(if_then_else
(match_operand 1 "long_immediate_loadstore_operand" "")
(const_int 8) (const_int 4))
(eq_attr "type" "store")
(if_then_else
(ior (match_operand 0 "long_immediate_loadstore_operand" "")
(match_operand 1 "immediate_operand" ""))
(const_int 8) (const_int 4))
(eq_attr "type" "move,unary")
(cond
[(match_operand 1 "u6_immediate_operand" "") (const_int 4)
(match_operand 1 "register_operand" "") (const_int 4)
(match_operand 1 "long_immediate_operand" "") (const_int 8)
(match_test "GET_CODE (PATTERN (insn)) == COND_EXEC") (const_int 8)]
(const_int 4))
(and (eq_attr "type" "shift")
(match_operand 1 "immediate_operand"))
(const_int 8)
(eq_attr "type" "binary,shift")
(if_then_else
(ior (match_operand 2 "long_immediate_operand" "")
(and (ne (symbol_ref "REGNO (operands[0])")
(symbol_ref "REGNO (operands[1])"))
(eq (match_operand 2 "u6_immediate_operand" "")
(const_int 0))))
(const_int 8) (const_int 4))
(eq_attr "type" "cmove")
(if_then_else (match_operand 1 "register_operand" "")
(const_int 4) (const_int 8))
(eq_attr "type" "call_no_delay_slot") (const_int 8)
]
(const_int 4))
)
;; The length here is the length of a single asm. Unfortunately it might be
;; 4 or 8 so we must allow for 8. That's ok though. How often will users
;; lament asm's not being put in delay slots?
;;
(define_asm_attributes
[(set_attr "length" "8")
(set_attr "type" "multi")
(set_attr "cond" "clob") ])
;; Delay slots.
;; The first two cond clauses and the default are necessary for correctness;
;; the remaining cond clause is mainly an optimization, as otherwise nops
;; would be inserted; however, if we didn't do this optimization, we would
;; have to be more conservative in our length calculations.
(define_attr "in_delay_slot" "false,true"
(cond [(eq_attr "type" "uncond_branch,jump,branch,
call,sfunc,call_no_delay_slot,
brcc, brcc_no_delay_slot,loop_setup,loop_end")
(const_string "false")
(match_test "arc_write_ext_corereg (insn)")
(const_string "false")
(gt (symbol_ref "arc_hazard (prev_active_insn (insn),
next_active_insn (insn))")
(symbol_ref "(arc_hazard (prev_active_insn (insn), insn)
+ arc_hazard (insn, next_active_insn (insn)))"))
(const_string "false")
(match_test "find_reg_note (insn, REG_SAVE_NOTE, GEN_INT (2))")
(const_string "false")
(eq_attr "iscompact" "maybe") (const_string "true")
]
(if_then_else (eq_attr "length" "2,4")
(const_string "true")
(const_string "false"))))
; must not put an insn inside that refers to blink.
(define_attr "in_call_delay_slot" "false,true"
(cond [(eq_attr "in_delay_slot" "false")
(const_string "false")
(match_test "arc_regno_use_in (RETURN_ADDR_REGNUM, PATTERN (insn))")
(const_string "false")]
(const_string "true")))
(define_attr "in_sfunc_delay_slot" "false,true"
(cond [(eq_attr "in_call_delay_slot" "false")
(const_string "false")
(match_test "arc_regno_use_in (12, PATTERN (insn))")
(const_string "false")]
(const_string "true")))
;; Instructions that we can put into a delay slot and conditionalize.
(define_attr "cond_delay_insn" "no,yes"
(cond [(eq_attr "cond" "!canuse") (const_string "no")
(eq_attr "type" "call,branch,uncond_branch,jump,brcc")
(const_string "no")
(match_test "find_reg_note (insn, REG_SAVE_NOTE, GEN_INT (2))")
(const_string "no")
(eq_attr "length" "2,4") (const_string "yes")]
(const_string "no")))
(define_attr "in_ret_delay_slot" "no,yes"
(cond [(eq_attr "in_delay_slot" "false")
(const_string "no")
(match_test "regno_clobbered_p
(RETURN_ADDR_REGNUM, insn, SImode, 1)")
(const_string "no")]
(const_string "yes")))
(define_attr "cond_ret_delay_insn" "no,yes"
(cond [(eq_attr "in_ret_delay_slot" "no") (const_string "no")
(eq_attr "cond_delay_insn" "no") (const_string "no")]
(const_string "yes")))
(define_attr "annul_ret_delay_insn" "no,yes"
(cond [(eq_attr "cond_ret_delay_insn" "yes") (const_string "yes")
(match_test "TARGET_AT_DBR_CONDEXEC") (const_string "no")
(eq_attr "type" "!call,branch,uncond_branch,jump,brcc,return,sfunc")
(const_string "yes")]
(const_string "no")))
;; Delay slot definition for ARCompact ISA
;; ??? FIXME:
;; When outputting an annul-true insn elegible for cond-exec
;; in a cbranch delay slot, unless optimizing for size, we use cond-exec
;; for ARC600; we could also use this for ARC700 if the branch can't be
;; unaligned and is at least somewhat likely (add parameter for this).
(define_delay (eq_attr "type" "call")
[(eq_attr "in_call_delay_slot" "true")
(eq_attr "in_call_delay_slot" "true")
(nil)])
(define_delay (and (match_test "!TARGET_AT_DBR_CONDEXEC")
(eq_attr "type" "brcc"))
[(eq_attr "in_delay_slot" "true")
(eq_attr "in_delay_slot" "true")
(nil)])
(define_delay (and (match_test "TARGET_AT_DBR_CONDEXEC")
(eq_attr "type" "brcc"))
[(eq_attr "in_delay_slot" "true")
(nil)
(nil)])
(define_delay
(eq_attr "type" "return")
[(eq_attr "in_ret_delay_slot" "yes")
(eq_attr "annul_ret_delay_insn" "yes")
(eq_attr "cond_ret_delay_insn" "yes")])
(define_delay (eq_attr "type" "loop_end")
[(eq_attr "in_delay_slot" "true")
(eq_attr "in_delay_slot" "true")
(nil)])
;; For ARC600, unexposing the delay sloy incurs a penalty also in the
;; non-taken case, so the only meaningful way to have an annull-true
;; filled delay slot is to conditionalize the delay slot insn.
(define_delay (and (match_test "TARGET_AT_DBR_CONDEXEC")
(eq_attr "type" "branch,uncond_branch,jump")
(match_test "!optimize_size"))
[(eq_attr "in_delay_slot" "true")
(eq_attr "cond_delay_insn" "yes")
(eq_attr "cond_delay_insn" "yes")])
;; For ARC700, anything goes for annulled-true insns, since there is no
;; penalty for the unexposed delay slot when the branch is not taken,
;; however, we must avoid things that have a delay slot themselvese to
;; avoid confusing gcc.
(define_delay (and (match_test "!TARGET_AT_DBR_CONDEXEC")
(eq_attr "type" "branch,uncond_branch,jump")
(match_test "!optimize_size"))
[(eq_attr "in_delay_slot" "true")
(eq_attr "type" "!call,branch,uncond_branch,jump,brcc,return,sfunc")
(eq_attr "cond_delay_insn" "yes")])
;; -mlongcall -fpic sfuncs use r12 to load the function address
(define_delay (eq_attr "type" "sfunc")
[(eq_attr "in_sfunc_delay_slot" "true")
(eq_attr "in_sfunc_delay_slot" "true")
(nil)])
;; ??? need to use a working strategy for canuse_limm:
;; - either canuse_limm is not eligible for delay slots, and has no
;; delay slots, or arc_reorg has to treat them as nocond, or it has to
;; somehow modify them to become inelegible for delay slots if a decision
;; is made that makes conditional execution required.
(define_attr "tune" "none,arc600,arc7xx,arc700_4_2_std,arc700_4_2_xmac, \
archs4x, archs4xd"
(const
(cond [(symbol_ref "arc_tune == ARC_TUNE_ARC600")
(const_string "arc600")
(symbol_ref "arc_tune == ARC_TUNE_ARC7XX")
(const_string "arc7xx")
(symbol_ref "arc_tune == ARC_TUNE_ARC700_4_2_STD")
(const_string "arc700_4_2_std")
(symbol_ref "arc_tune == ARC_TUNE_ARC700_4_2_XMAC")
(const_string "arc700_4_2_xmac")
(ior (symbol_ref "arc_tune == ARC_TUNE_ARCHS4X")
(symbol_ref "arc_tune == ARC_TUNE_ARCHS4X_REL31A"))
(const_string "archs4x")
(ior (symbol_ref "arc_tune == ARC_TUNE_ARCHS4XD")
(symbol_ref "arc_tune == ARC_TUNE_ARCHS4XD_SLOW"))
(const_string "archs4xd")]
(const_string "none"))))
(define_attr "tune_arc700" "false,true"
(if_then_else (eq_attr "tune" "arc7xx, arc700_4_2_std, arc700_4_2_xmac")
(const_string "true")
(const_string "false")))
(define_attr "tune_dspmpy" "none, slow, fast"
(const
(cond [(ior (symbol_ref "arc_tune == ARC_TUNE_ARCHS4X")
(symbol_ref "arc_tune == ARC_TUNE_ARCHS4XD"))
(const_string "fast")
(symbol_ref "arc_tune == ARC_TUNE_ARCHS4XD_SLOW")
(const_string "slow")]
(const_string "none"))))
(define_attr "tune_store" "none, normal, rel31a"
(const
(cond [(ior (symbol_ref "arc_tune == ARC_TUNE_ARCHS4X")
(symbol_ref "arc_tune == ARC_TUNE_ARCHS4XD"))
(const_string "normal")
(symbol_ref "arc_tune == ARC_TUNE_ARCHS4X_REL31A")
(const_string "rel31a")]
(const_string "none"))))
;; Move instructions.
(define_expand "movqi"
[(set (match_operand:QI 0 "move_dest_operand" "")
(match_operand:QI 1 "general_operand" ""))]
""
"if (prepare_move_operands (operands, QImode)) DONE;")
; In order to allow the ccfsm machinery to do its work, the leading compact
; alternatives say 'canuse' - there is another alternative that will match
; when the condition codes are used.
; Likewise, the length of an alternative that might be shifted to conditional
; execution must reflect this, lest out-of-range branches are created.
; The iscompact attribute allows the epilogue expander to know for which
; insns it should lengthen the return insn.
(define_insn "*movqi_insn"
[(set (match_operand:QI 0 "move_dest_operand" "=q, q,r,q, h, w, w,???w,h, w,q,S,!*x, r,r, Ucm,m,???m, m,Usc")
(match_operand:QI 1 "move_src_operand" "rL,rP,q,P,hCm1,cL, I,?Rac,i,?i,T,q,Usd,Ucm,m,?Rac,c,?Rac,Cm3,i"))]
"register_operand (operands[0], QImode)
|| register_operand (operands[1], QImode)
|| (satisfies_constraint_Cm3 (operands[1])
&& memory_operand (operands[0], QImode))"
"@
mov%? %0,%1%&
mov%? %0,%1%&
mov%? %0,%1%&
mov%? %0,%1%&
mov%? %0,%1%&
mov%? %0,%1
mov%? %0,%1
mov%? %0,%1
mov%? %0,%1
mov%? %0,%1
ldb%? %0,%1%&
stb%? %1,%0%&
ldb%? %0,%1%&
xldb%U1 %0,%1
ldb%U1%V1 %0,%1
xstb%U0 %1,%0
stb%U0%V0 %1,%0
stb%U0%V0 %1,%0
stb%U0%V0 %1,%0
stb%U0%V0 %1,%0"
[(set_attr "type" "move,move,move,move,move,move,move,move,move,move,load,store,load,load,load,store,store,store,store,store")
(set_attr "iscompact" "maybe,maybe,maybe,true,true,false,false,false,maybe_limm,false,true,true,true,false,false,false,false,false,false,false")
(set_attr "predicable" "yes,no,yes,no,no,yes,no,yes,yes,yes,no,no,no,no,no,no,no,no,no,no")
(set_attr "cpu_facility" "av1,av1,av1,av2,av2,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*")])
(define_expand "movhi"
[(set (match_operand:HI 0 "move_dest_operand" "")
(match_operand:HI 1 "general_operand" ""))]
""
"if (prepare_move_operands (operands, HImode)) DONE;")
(define_insn "*movhi_insn"
[(set (match_operand:HI 0 "move_dest_operand" "=q, q,r,q, h, w, w,???w,q,h, w,q,S, r,r, Ucm,m,???m, m,VUsc")
(match_operand:HI 1 "move_src_operand" " rL,rP,q,P,hCm1,cL, I,?Rac,i,i,?i,T,q,Ucm,m,?Rac,c,?Rac,Cm3,i"))]
"register_operand (operands[0], HImode)
|| register_operand (operands[1], HImode)
|| (CONSTANT_P (operands[1])
/* Don't use a LIMM that we could load with a single insn - we loose
delay-slot filling opportunities. */
&& !satisfies_constraint_I (operands[1])
&& satisfies_constraint_Usc (operands[0]))
|| (satisfies_constraint_Cm3 (operands[1])
&& memory_operand (operands[0], HImode))"
"@
mov%? %0,%1%&
mov%? %0,%1%&
mov%? %0,%1%&
mov%? %0,%1%&
mov%? %0,%1%&
mov%? %0,%1
mov%? %0,%1
mov%? %0,%1
mov%? %0,%1%&
mov%? %0,%1
mov%? %0,%1
ld%_%? %0,%1%&
st%_%? %1,%0%&
xld%_%U1 %0,%1
ld%_%U1%V1 %0,%1
xst%_%U0 %1,%0
st%_%U0%V0 %1,%0
st%_%U0%V0 %1,%0
st%_%U0%V0 %1,%0
st%_%U0%V0 %1,%0"
[(set_attr "type" "move,move,move,move,move,move,move,move,move,move,move,load,store,load,load,store,store,store,store,store")
(set_attr "iscompact" "maybe,maybe,maybe,true,true,false,false,false,maybe_limm,maybe_limm,false,true,true,false,false,false,false,false,false,false")
(set_attr "predicable" "yes,no,yes,no,no,yes,no,yes,yes,yes,yes,no,no,no,no,no,no,no,no,no")
(set_attr "cpu_facility" "av1,av1,av1,av2,av2,*,*,*,*,*,*,*,*,*,*,*,*,*,av2,*")])
(define_expand "movsi"
[(set (match_operand:SI 0 "move_dest_operand" "")
(match_operand:SI 1 "general_operand" ""))]
""
"if (prepare_move_operands (operands, SImode)) DONE;")
; In order to allow the ccfsm machinery to do its work, the leading compact
; alternatives say 'canuse' - there is another alternative that will match
; when the condition codes are used.
; The length of an alternative that might be shifted to conditional
; execution must reflect this, lest out-of-range branches are created.
; the iscompact attribute allows the epilogue expander to know for which
; insns it should lengthen the return insn.
(define_insn_and_split "*movsi_insn" ; 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28
[(set (match_operand:SI 0 "move_dest_operand" "=q, q,r,q, h, rl,r, r, r, r, ?r, r, q, h, rl, q, S, Us<,qRck,!*x, r,!*Rsd,!*Rcd,r,Ucm, Usd,m, m,VUsc")
(match_operand:SI 1 "move_src_operand" "rL,rP,q,P,hCm1,rLl,I,Clo,Chi,Cbi,Cpc,Clb,Cax,Cal,Cal,Uts,q,qRck, Us>,Usd,Ucm, Usd, Ucd,m, r,!*Rzd,r,Cm3, C32"))]
"register_operand (operands[0], SImode)
|| register_operand (operands[1], SImode)
|| (CONSTANT_P (operands[1])
&& (!satisfies_constraint_I (operands[1]) || !optimize_size)
&& satisfies_constraint_Usc (operands[0]))
|| (satisfies_constraint_Cm3 (operands[1])
&& memory_operand (operands[0], SImode))"
"@
mov%?\\t%0,%1 ;0
mov%?\\t%0,%1 ;1
mov%?\\t%0,%1 ;2
mov%?\\t%0,%1 ;3
mov%?\\t%0,%1 ;4
mov%?\\t%0,%1 ;5
mov%?\\t%0,%1 ;6
movl.cl\\t%0,%1 ;7
movh.cl\\t%0,%L1>>16 ;8
* return INTVAL (operands[1]) & 0xffffff ? \"movbi.cl\\t%0,%1 >> %p1,%p1,8;9\" : \"movbi.cl\\t%0,%L1 >> 24,24,8;9\";
add\\t%0,%1 ;10
add\\t%0,pcl,%1@pcl ;11
#
mov%?\\t%0,%j1 ;13
mov%?\\t%0,%j1 ;14
ld%?\\t%0,%1 ;15
st%?\\t%1,%0 ;16
* return arc_short_long (insn, \"push%?\\t%1%&\", \"st%U0\\t%1,%0%&\");
* return arc_short_long (insn, \"pop%?\\t%0%&\", \"ld%U1\\t%0,%1%&\");
ld%?\\t%0,%1 ;19
xld%U1\\t%0,%1 ;20
ld%?\\t%0,%1 ;21
ld%?\\t%0,%1 ;22
ld%U1%V1\\t%0,%1 ;23
xst%U0\\t%1,%0 ;24
st%?\\t%1,%0%& ;25
st%U0%V0\\t%1,%0 ;26
st%U0%V0\\t%1,%0 ;37
st%U0%V0\\t%1,%0 ;28"
"reload_completed
&& GET_CODE (PATTERN (insn)) != COND_EXEC
&& register_operand (operands[0], SImode)
&& IN_RANGE (REGNO (operands[0]) ^ 4, 4, 11)
&& satisfies_constraint_Cax (operands[1])"
[(const_int 0)]
"
arc_split_mov_const (operands);
DONE;
"
; 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28
[(set_attr "type" "move, move, move,move,move, move, move,shift,shift,shift,binary,binary,multi,move, move,load,store,store,load,load, load,load,load, load,store,store,store,store,store")
(set_attr "iscompact" "maybe,maybe,maybe,true,true,false,false,false,false,false, false, false,false,true,false,true, true, true,true,true,false,true,true,false,false, true,false,false,false")
(set_attr "length" "*,*,*,*,*,4,4,4,4,4,8,8,*,6,*,*,*,*,*,*,4,*,4,*,*,*,*,*,8")
(set_attr "predicable" "yes,no,yes,no,no,yes,no,no,no,yes,no,no,no,yes,yes,no,no,no,no,no,no,no,no,no,no,no,no,no,no")
(set_attr "cpu_facility" "av1,av1,av1,av2,av2,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,*,av2,av2,*,*,av2,*,av2,*")])
;; Sometimes generated by the epilogue code. We don't want to
;; recognize these addresses in general, because the limm is costly,
;; and we can't use them for stores. */
(define_insn "*movsi_pre_mod"
[(set (match_operand:SI 0 "register_operand" "=w")
(mem:SI (pre_modify
(reg:SI SP_REG)
(plus:SI (reg:SI SP_REG)
(match_operand 1 "immediate_operand" "Cal")))))]
"reload_completed"
"ld.a %0,[sp,%1]"
[(set_attr "type" "load")
(set_attr "length" "8")])
;; Store a value to directly to memory. The location might also be cached.
;; Since the cached copy can cause a write-back at unpredictable times,
;; we first write cached, then we write uncached.
(define_insn "store_direct"
[(set (match_operand:SI 0 "move_dest_operand" "=m")
(unspec:SI [(match_operand:SI 1 "register_operand" "c")]
UNSPEC_ARC_DIRECT))]
""
"st%U0 %1,%0\;st%U0.di %1,%0"
[(set_attr "type" "store")])
;; Combiner patterns for compare with zero
(define_mode_iterator SQH [QI HI])
(define_mode_attr SQH_postfix [(QI "b") (HI "%_")])
(define_code_iterator SEZ [sign_extend zero_extend])
(define_code_attr SEZ_prefix [(sign_extend "sex") (zero_extend "ext")])
; Optab prefix for sign/zero-extending operations
(define_code_attr su_optab [(sign_extend "") (zero_extend "u")])
(define_insn "*xt_cmp0_noout"
[(set (match_operand 0 "cc_set_register" "")
(compare:CC_ZN (SEZ:SI (match_operand:SQH 1 "register_operand" "r"))
(const_int 0)))]
""
".f\\t0,%1"
[(set_attr "type" "compare")
(set_attr "cond" "set_zn")])
(define_insn "*xt_cmp0"
[(set (match_operand 0 "cc_set_register" "")
(compare:CC_ZN (SEZ:SI (match_operand:SQH 1 "register_operand" "r"))
(const_int 0)))
(set (match_operand:SI 2 "register_operand" "=r")
(SEZ:SI (match_dup 1)))]
""
".f\\t%2,%1"
[(set_attr "type" "compare")
(set_attr "cond" "set_zn")])
(define_insn "*xbfu_cmp0_noout"
[(set (match_operand 0 "cc_set_register" "")
(compare:CC_Z
(zero_extract:SI
(match_operand:SI 1 "register_operand" " r,r")
(match_operand:SI 2 "const_int_operand" "C3p,n")
(match_operand:SI 3 "const_int_operand" " n,n"))
(const_int 0)))]
"TARGET_HS && TARGET_BARREL_SHIFTER"
{
int assemble_op2 = (((INTVAL (operands[2]) - 1) & 0x1f) << 5) | (INTVAL (operands[3]) & 0x1f);
operands[2] = GEN_INT (assemble_op2);
return "xbfu%?.f\\t0,%1,%2";
}
[(set_attr "type" "shift")
(set_attr "iscompact" "false")
(set_attr "length" "4,8")
(set_attr "predicable" "no")
(set_attr "cond" "set_zn")])
(define_insn "*xbfu_cmp0"
[(set (match_operand 4 "cc_set_register" "")
(compare:CC_Z
(zero_extract:SI
(match_operand:SI 1 "register_operand" "0 ,r,0")
(match_operand:SI 2 "const_int_operand" "C3p,n,n")
(match_operand:SI 3 "const_int_operand" "n ,n,n"))
(const_int 0)))
(set (match_operand:SI 0 "register_operand" "=r,r,r")
(zero_extract:SI (match_dup 1) (match_dup 2) (match_dup 3)))]
"TARGET_HS && TARGET_BARREL_SHIFTER"
{
int assemble_op2 = (((INTVAL (operands[2]) - 1) & 0x1f) << 5) | (INTVAL (operands[3]) & 0x1f);
operands[2] = GEN_INT (assemble_op2);
return "xbfu%?.f\\t%0,%1,%2";
}
[(set_attr "type" "shift")
(set_attr "iscompact" "false")
(set_attr "length" "4,8,8")
(set_attr "predicable" "yes,no,yes")
(set_attr "cond" "set_zn")])
; splitting to 'tst' allows short insns and combination into brcc.
(define_insn_and_split "*movsi_set_cc_insn"
[(set (match_operand 2 "cc_set_register" "")
(match_operator 3 "zn_compare_operator"
[(match_operand:SI 1 "nonmemory_operand" "rL,rI,Cal")
(const_int 0)]))
(set (match_operand:SI 0 "register_operand" "=r,r,r")
(match_dup 1))]
""
"mov%?.f\\t%0,%1"
"reload_completed && operands_match_p (operands[0], operands[1])"
[(set (match_dup 2) (match_dup 3))]
""
[(set_attr "type" "compare")
(set_attr "predicable" "yes,no,yes")
(set_attr "cond" "set_zn")
(set_attr "length" "4,4,8")])
(define_insn "unary_comparison"
[(set (match_operand:CC_ZN 0 "cc_set_register" "")
(match_operator:CC_ZN 3 "zn_compare_operator"
[(match_operator:SI 2 "unary_operator"
[(match_operand:SI 1 "register_operand" "c")])
(const_int 0)]))]
""
"%O2.f 0,%1"
[(set_attr "type" "compare")
(set_attr "cond" "set_zn")])
; this pattern is needed by combiner for cases like if (c=(~b)) { ... }
(define_insn "*unary_comparison_result_used"
[(set (match_operand 2 "cc_register" "")
(match_operator 4 "zn_compare_operator"
[(match_operator:SI 3 "unary_operator"
[(match_operand:SI 1 "register_operand" "c")])
(const_int 0)]))
(set (match_operand:SI 0 "register_operand" "=w")
(match_dup 3))]
""
"%O3.f %0,%1"
[(set_attr "type" "compare")
(set_attr "cond" "set_zn")
(set_attr "length" "4")])
; reload is too stingy with reloads for Rrq/Cbf/Rrq when it sees
; a c/???Cal/X alternative, so we say it's c/???Cal/c instead,
; even if we don't need the clobber.
(define_insn_and_split "*tst_movb"
[(set
(match_operand 0 "cc_register" "")
(match_operator 4 "zn_compare_operator"
[(and:SI
(match_operand:SI 1 "register_operand" "%q, q, c, c, c, c, q, q, c")
(match_operand:SI 2 "nonmemory_operand" "q,C0p,cI,C1p,Ccp,Chs,Cbf,Cbf,???Cal"))
(const_int 0)]))
(clobber (match_scratch:SI 3 "=X,X,X,X,X,X,Rrq,1,c"))]
"TARGET_NPS_BITOPS"
"movb.f.cl %3,%1,%p2,%p2,%s2"
"TARGET_NPS_BITOPS && reload_completed
&& (extract_constrain_insn_cached (insn), (which_alternative & ~1) != 6)"
[(set (match_dup 0) (match_dup 4))])
(define_insn "*tst"
[(set
(match_operand 0 "cc_register" "")
(match_operator 3 "zn_compare_operator"
[(and:SI
(match_operand:SI 1 "register_operand"
"%q, q, c, c, c, c, c, c")
(match_operand:SI 2 "nonmemory_operand"
" q,C0p,cI,cL,C1p,Ccp,Chs,Cal"))
(const_int 0)]))]
"reload_completed
|| !satisfies_constraint_Cbf (operands[2])
|| satisfies_constraint_C0p (operands[2])
|| satisfies_constraint_I (operands[2])
|| satisfies_constraint_C1p (operands[2])
|| satisfies_constraint_Chs (operands[2])"
"*
switch (which_alternative)
{
case 0: case 2: case 3: case 7:
return \"tst%? %1,%2\";
case 1:
return \"btst%? %1,%z2\";
case 4:
return \"bmsk%?.f 0,%1,%Z2%&\";
case 5:
return \"bclr%?.f 0,%1,%M2%&\";
case 6:
return \"asr.f 0,%1,%p2\";
default:
gcc_unreachable ();
}
"
[(set_attr "iscompact" "maybe,maybe,false,false,false,false,false,false")
(set_attr "type" "compare,compare,compare,compare,compare,compare,binary,compare")
(set_attr "length" "*,*,4,4,4,4,4,8")
(set_attr "predicable" "no,yes,no,yes,no,no,no,yes")
(set_attr "cond" "set_zn")])
; ??? Sometimes, if an AND with a constant can be expressed as a zero_extract,
; combine will do that and not try the AND.
; It would take 66 constraint combinations to describe the zero_extract
; constants that are covered by the 12-bit signed constant for tst
; (excluding the ones that are better done by mov or btst).
; so we rather use an extra pattern for tst;
; since this is about constants, reload shouldn't care.
(define_insn "*tst_bitfield_tst"
[(set (match_operand:CC_ZN 0 "cc_set_register" "")
(match_operator 4 "zn_compare_operator"
[(zero_extract:SI
(match_operand:SI 1 "register_operand" "c")
(match_operand:SI 2 "const_int_operand" "n")
(match_operand:SI 3 "const_int_operand" "n"))
(const_int 0)]))]
"INTVAL (operands[2]) > 1
&& (INTVAL (operands[3]) + INTVAL (operands[2]) <= 11
|| (INTVAL (operands[3]) <= 11
&& INTVAL (operands[3]) + INTVAL (operands[2]) == 32))"
"tst %1,((1<<%2)-1)<<%3"
[(set_attr "type" "compare")
(set_attr "cond" "set_zn")
(set_attr "length" "4")])
; Likewise for asr.f.
(define_insn "*tst_bitfield_asr"
[(set (match_operand:CC_ZN 0 "cc_set_register" "")
(match_operator 4 "zn_compare_operator"
[(zero_extract:SI
(match_operand:SI 1 "register_operand" "c")
(match_operand:SI 2 "const_int_operand" "n")
(match_operand:SI 3 "const_int_operand" "n"))
(const_int 0)]))]
"INTVAL (operands[2]) > 1
&& INTVAL (operands[3]) + INTVAL (operands[2]) == 32"
"asr.f 0,%1,%3"
[(set_attr "type" "shift")
(set_attr "cond" "set_zn")
(set_attr "length" "4")])
(define_insn "*tst_bitfield"
[(set (match_operand:CC_ZN 0 "cc_set_register" "")
(match_operator 5 "zn_compare_operator"
[(zero_extract:SI
(match_operand:SI 1 "register_operand" "%q,c, c,Rrq,c")
(match_operand:SI 2 "const_int_operand" "N,N, n,Cbn,n")
(match_operand:SI 3 "const_int_operand" "n,n,C_0,Cbn,n"))
(const_int 0)]))
(clobber (match_scratch:SI 4 "=X,X,X,Rrq,X"))]
""
"@
btst%? %1,%3
btst %1,%3
bmsk.f 0,%1,%2-1
movb.f.cl %4,%1,%3,%3,%2
and.f 0,%1,((1<<%2)-1)<<%3"
[(set_attr "iscompact" "maybe,false,false,false,false")
(set_attr "type" "compare,compare,compare,shift,compare")
(set_attr "cond" "set_zn")
(set_attr "length" "*,4,4,4,8")])
;; The next two patterns are for plos, ior, xor, and, and mult.
(define_insn "*commutative_binary_cmp0_noout"
[(set (match_operand 0 "cc_set_register" "")
(match_operator 4 "zn_compare_operator"
[(match_operator:SI 3 "commutative_operator"
[(match_operand:SI 1 "register_operand" "%r,r")
(match_operand:SI 2 "nonmemory_operand" "rL,Cal")])
(const_int 0)]))]
""
"%O3.f\\t0,%1,%2"
[(set_attr "type" "compare")
(set_attr "cond" "set_zn")
(set_attr "length" "4,8")])
(define_insn "*commutative_binary_cmp0"
[(set (match_operand 3 "cc_set_register" "")
(match_operator 5 "zn_compare_operator"
[(match_operator:SI 4 "commutative_operator"
[(match_operand:SI 1 "register_operand" "%0, 0,r,r")
(match_operand:SI 2 "nonmemory_operand" "rL,rI,r,Cal")])
(const_int 0)]))
(set (match_operand:SI 0 "register_operand" "=r,r,r,r")
(match_dup 4))]
""
"%O4.f\\t%0,%1,%2"
[(set_attr "type" "compare")
(set_attr "cond" "set_zn")
(set_attr "predicable" "yes,yes,no,no")
(set_attr "length" "4,4,4,8")])
; for flag setting 'add' instructions like if (a+b) { ...}
; the combiner needs this pattern
(define_insn "*addsi_compare"
[(set (reg:CC_ZN CC_REG)
(compare:CC_ZN (match_operand:SI 0 "register_operand" "c")
(neg:SI (match_operand:SI 1 "register_operand" "c"))))]
""
"add.f 0,%0,%1"
[(set_attr "cond" "set")
(set_attr "type" "compare")
(set_attr "length" "4")])
; for flag setting 'add' instructions like if (a+b < a) { ...}
; the combiner needs this pattern
(define_insn "addsi_compare_2"
[(set (reg:CC_C CC_REG)
(compare:CC_C (plus:SI (match_operand:SI 0 "register_operand" "c,c")
(match_operand:SI 1 "nonmemory_operand" "cL,Cal"))
(match_dup 0)))]
""
"add.f 0,%0,%1"
[(set_attr "cond" "set")
(set_attr "type" "compare")
(set_attr "length" "4,8")])
(define_insn "*addsi_compare_3"
[(set (reg:CC_C CC_REG)
(compare:CC_C (plus:SI (match_operand:SI 0 "register_operand" "c")
(match_operand:SI 1 "register_operand" "c"))
(match_dup 1)))]
""
"add.f 0,%0,%1"
[(set_attr "cond" "set")
(set_attr "type" "compare")
(set_attr "length" "4")])
; this pattern is needed by combiner for cases like if (c=a+b) { ... }
(define_insn "*commutative_binary_comparison_result_used"
[(set (match_operand 3 "cc_register" "")
(match_operator 5 "zn_compare_operator"
; We can accept any commutative operator except mult because
; our 'w' class below could try to use LP_COUNT.
[(match_operator:SI 4 "commutative_operator_sans_mult"
[(match_operand:SI 1 "register_operand" "c,0,c")
(match_operand:SI 2 "nonmemory_operand" "cL,I,?Cal")])
(const_int 0)]))
(set (match_operand:SI 0 "register_operand" "=w,w,w")
(match_dup 4))]
""
"%O4.f %0,%1,%2 ; non-mult commutative"
[(set_attr "type" "compare,compare,compare")
(set_attr "cond" "set_zn,set_zn,set_zn")
(set_attr "length" "4,4,8")])
; a MULT-specific version of this pattern to avoid touching the
; LP_COUNT register
(define_insn "*commutative_binary_mult_comparison_result_used"
[(set (match_operand 3 "cc_register" "")
(match_operator 5 "zn_compare_operator"
[(match_operator:SI 4 "mult_operator"
[(match_operand:SI 1 "register_operand" "c,0,c")
(match_operand:SI 2 "nonmemory_operand" "cL,I,?Cal")])
(const_int 0)]))
; Make sure to use the W class to not touch LP_COUNT.
(set (match_operand:SI 0 "register_operand" "=W,W,W")
(match_dup 4))]
"!TARGET_ARC600_FAMILY"
"%O4.f %0,%1,%2 ; mult commutative"
[(set_attr "type" "compare,compare,compare")
(set_attr "cond" "set_zn,set_zn,set_zn")
(set_attr "length" "4,4,8")])
(define_insn "*noncommutative_binary_cmp0"
[(set (match_operand 3 "cc_set_register" "")
(match_operator 5 "zn_compare_operator"
[(match_operator:SI 4 "noncommutative_operator"
[(match_operand:SI 1 "register_operand" "0,r,0, 0,r")
(match_operand:SI 2 "nonmemory_operand" "rL,r,I,Cal,Cal")])
(const_int 0)]))
(set (match_operand:SI 0 "register_operand" "=r,r,r,r,r")
(match_dup 4))]
""
"%O4%?.f\\t%0,%1,%2"
[(set_attr "type" "compare")
(set_attr "cond" "set_zn")
(set_attr "predicable" "yes,no,no,yes,no")
(set_attr "length" "4,4,4,8,8")])
(define_insn "*noncommutative_binary_cmp0_noout"
[(set (match_operand 0 "cc_set_register" "")
(match_operator 3 "zn_compare_operator"
[(match_operator:SI 4 "noncommutative_operator"
[(match_operand:SI 1 "register_operand" "r,r")
(match_operand:SI 2 "nonmemory_operand" "rL,Cal")])
(const_int 0)]))]
""
"%O4.f\\t0,%1,%2"
[(set_attr "type" "compare")
(set_attr "cond" "set_zn")
(set_attr "length" "4,8")])
;;rsub variants
(define_insn "*rsub_cmp0"
[(set (match_operand 4 "cc_set_register" "")
(match_operator 3 "zn_compare_operator"
[(minus:SI
(match_operand:SI 1 "nonmemory_operand" "rL,Cal")
(match_operand:SI 2 "register_operand" "r,r"))
(const_int 0)]))
(set (match_operand:SI 0 "register_operand" "=r,r")
(minus:SI (match_dup 1) (match_dup 2)))]
""
"rsub.f\\t%0,%2,%1"
[(set_attr "type" "compare")
(set_attr "cond" "set_zn")
(set_attr "length" "4,8")])
(define_insn "*rsub_cmp0_noout"
[(set (match_operand 0 "cc_set_register" "")
(match_operator 3 "zn_compare_operator"
[(minus:SI
(match_operand:SI 1 "nonmemory_operand" "rL,Cal")
(match_operand:SI 2 "register_operand" "r,r"))
(const_int 0)]))]
""
"rsub.f\\t0,%2,%1"
[(set_attr "type" "compare")
(set_attr "cond" "set_zn")
(set_attr "length" "4,8")])
(define_expand "bic_f_zn"
[(parallel
[(set (reg:CC_ZN CC_REG)
(compare:CC_ZN
(and:SI (match_operand:SI 1 "register_operand" "")
(not:SI (match_operand:SI 2 "nonmemory_operand" "")))
(const_int 0)))
(set (match_operand:SI 0 "register_operand" "")
(and:SI (match_dup 1) (not:SI (match_dup 2))))])]
"")
(define_insn "*bic_f"
[(set (match_operand 3 "cc_set_register" "")
(match_operator 4 "zn_compare_operator"
[(and:SI (match_operand:SI 1 "register_operand" "c,0,c")
(not:SI
(match_operand:SI 2 "nonmemory_operand" "cL,I,?Cal")))
(const_int 0)]))
(set (match_operand:SI 0 "register_operand" "=w,w,w")
(and:SI (match_dup 1) (not:SI (match_dup 2))))]
""
"bic.f %0,%1,%2"
[(set_attr "type" "compare,compare,compare")
(set_attr "cond" "set_zn,set_zn,set_zn")
(set_attr "length" "4,4,8")])
(define_insn "*bic_cmp0_noout"
[(set (match_operand 0 "cc_set_register" "")
(compare:CC_ZN
(and:SI (not:SI (match_operand:SI 1 "nonmemory_operand" "Lr,Cal,r"))
(match_operand:SI 2 "nonmemory_operand" "r,r,Cal"))
(const_int 0)))]
"register_operand (operands[1], SImode)
|| register_operand (operands[2], SImode)"
"bic.f\\t0,%2,%1"
[(set_attr "type" "unary")
(set_attr "cond" "set_zn")
(set_attr "length" "4,8,8")])
(define_insn "*bic_cmp0"
[(set (match_operand 0 "cc_set_register" "")
(compare:CC_ZN
(and:SI (not:SI (match_operand:SI 1 "nonmemory_operand" "Lr,Cal,r"))
(match_operand:SI 2 "nonmemory_operand" "r,r,Cal"))
(const_int 0)))
(set (match_operand:SI 3 "register_operand" "=r,r,r")
(and:SI (not:SI (match_dup 1)) (match_dup 2)))]
"register_operand (operands[1], SImode)
|| register_operand (operands[2], SImode)"
"bic.f\\t%3,%2,%1"
[(set_attr "type" "unary")
(set_attr "cond" "set_zn")
(set_attr "length" "4,8,8")])
(define_expand "movdi"
[(set (match_operand:DI 0 "move_dest_operand" "")
(match_operand:DI 1 "general_operand" ""))]
""
"
if (prepare_move_operands (operands, DImode))
DONE;
")
(define_insn_and_split "*movdi_insn"
[(set (match_operand:DI 0 "move_dest_operand" "=r, r,r, m")
(match_operand:DI 1 "move_double_src_operand" "r,Hi,m,rCm3"))]
"register_operand (operands[0], DImode)
|| register_operand (operands[1], DImode)
|| (satisfies_constraint_Cm3 (operands[1])
&& memory_operand (operands[0], DImode))"
"@
vadd2\\t%0,%1,0
#
ldd%U1%V1\\t%0,%1
std%U0%V0\\t%1,%0"
"&& reload_completed && arc_split_move_p (operands)"
[(const_int 0)]
{
arc_split_move (operands);
DONE;
}
[(set_attr "type" "move,move,load,store")
(set_attr "length" "8,16,16,16")])
;; Floating point move insns.
(define_expand "movsf"
[(set (match_operand:SF 0 "move_dest_operand" "")
(match_operand:SF 1 "general_operand" ""))]
""
"if (prepare_move_operands (operands, SFmode)) DONE;")
(define_insn "*movsf_insn"
[(set (match_operand:SF 0 "move_dest_operand" "=h,h, r,r, q,S,Usc,r,m")
(match_operand:SF 1 "move_src_operand" "hCfZ,E,rCfZ,E,Uts,q, E,m,r"))]
"register_operand (operands[0], SFmode)
|| register_operand (operands[1], SFmode)"
"@
mov%?\\t%0,%1
mov%?\\t%0,%1 ; %A1
mov%?\\t%0,%1
mov%?\\t%0,%1 ; %A1
ld%?%U1\\t%0,%1
st%?\\t%1,%0
st%U0%V0\\t%1,%0
ld%U1%V1\\t%0,%1
st%U0%V0\\t%1,%0"
[(set_attr "type" "move,move,move,move,load,store,store,load,store")
(set_attr "predicable" "no,no,yes,yes,no,no,no,no,no")
(set_attr "length" "*,*,4,*,*,*,*,*,*")
(set_attr "iscompact" "true,true_limm,false,false,true,true,false,false,false")])
(define_expand "movdf"
[(set (match_operand:DF 0 "move_dest_operand" "")
(match_operand:DF 1 "general_operand" ""))]
""
"if (prepare_move_operands (operands, DFmode)) DONE;")
(define_insn_and_split "*movdf_insn"
[(set (match_operand:DF 0 "move_dest_operand" "=D,r,r,r,r,m")
(match_operand:DF 1 "move_double_src_operand" "r,D,r,E,m,r"))]
"(register_operand (operands[0], DFmode)
|| register_operand (operands[1], DFmode))"
"@
#
#
vadd2\\t%0,%1,0
#
ldd%U1%V1\\t%0,%1
std%U0%V0\\t%1,%0"
"&& reload_completed && arc_split_move_p (operands)"
[(const_int 0)]
{
arc_split_move (operands);
DONE;
}
[(set_attr "type" "move,move,move,move,load,store")
(set_attr "length" "4,16,8,16,16,16")])
(define_insn_and_split "*movdf_insn_nolrsr"
[(set (match_operand:DF 0 "register_operand" "=r")
(match_operand:DF 1 "arc_double_register_operand" "D"))
(use (match_operand:SI 2 "" "N")) ; aka const1_rtx
]
"TARGET_DPFP && TARGET_DPFP_DISABLE_LRSR"
"#"
"&& 1"
[
; mov r0, 0
(set (match_dup 0) (match_dup 3))
; daddh?? r1, r0, r0
(parallel [
(set (match_dup 1) (plus:DF (match_dup 1) (match_dup 0)))
(use (const_int 1))
(use (const_int 1))
(use (match_dup 0)) ; used to block can_combine_p
(set (match_dup 0) (plus:DF (match_dup 1) (match_dup 0))) ; r1 in op 0
])
; We have to do this twice, once to read the value into R0 and
; second time to put back the contents which the first DEXCLx
; will have overwritten
; dexcl2 r0, r1, r0
(parallel [
(set (match_dup 4) ; aka r0result
; aka DF, r1, r0
(unspec_volatile:SI [(match_dup 5) (match_dup 4)]
VUNSPEC_ARC_DEXCL))
(clobber (match_dup 1))
])
; Generate the second, which makes sure operand5 and operand4 values
; are put back in the Dx register properly.
(set (match_dup 1) (unspec_volatile:DF
[(match_dup 5) (match_dup 4)]
VUNSPEC_ARC_DEXCL_NORES))
; Note: we cannot use a (clobber (match_scratch)) here because
; the combine pass will end up replacing uses of it with 0
]
"operands[3] = CONST0_RTX (DFmode);
operands[4] = simplify_gen_subreg (SImode, operands[0], DFmode, 0);
operands[5] = simplify_gen_subreg (SImode, operands[0], DFmode, 4);"
[(set_attr "type" "move")])
;; Load/Store with update instructions.
;;
;; Some of these we can get by using pre-decrement or pre-increment, but the
;; hardware can also do cases where the increment is not the size of the
;; object.
;;
;; In all these cases, we use operands 0 and 1 for the register being
;; incremented because those are the operands that local-alloc will
;; tie and these are the pair most likely to be tieable (and the ones
;; that will benefit the most).
;;
;; We use match_operator here because we need to know whether the memory
;; object is volatile or not.
;; Note: loadqi_update has no 16-bit variant
(define_insn "*loadqi_update"
[(set (match_operand:QI 3 "dest_reg_operand" "=r,r")
(match_operator:QI 4 "any_mem_operand"
[(plus:SI (match_operand:SI 1 "register_operand" "0,0")
(match_operand:SI 2 "nonmemory_operand" "rCm2,Cal"))]))
(set (match_operand:SI 0 "dest_reg_operand" "=r,r")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"ldb.a%V4 %3,[%0,%2]"
[(set_attr "type" "load,load")
(set_attr "length" "4,8")])
(define_insn "*load_zeroextendqisi_update"
[(set (match_operand:SI 3 "dest_reg_operand" "=r,r")
(zero_extend:SI (match_operator:QI 4 "any_mem_operand"
[(plus:SI (match_operand:SI 1 "register_operand" "0,0")
(match_operand:SI 2 "nonmemory_operand" "rCm2,Cal"))])))
(set (match_operand:SI 0 "dest_reg_operand" "=r,r")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"ldb.a%V4 %3,[%0,%2]"
[(set_attr "type" "load,load")
(set_attr "length" "4,8")])
(define_insn "*load_signextendqisi_update"
[(set (match_operand:SI 3 "dest_reg_operand" "=r,r")
(sign_extend:SI (match_operator:QI 4 "any_mem_operand"
[(plus:SI (match_operand:SI 1 "register_operand" "0,0")
(match_operand:SI 2 "nonmemory_operand" "rCm2,Cal"))])))
(set (match_operand:SI 0 "dest_reg_operand" "=r,r")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"ldb.x.a%V4 %3,[%0,%2]"
[(set_attr "type" "load,load")
(set_attr "length" "4,8")])
(define_insn "*storeqi_update"
[(set (match_operator:QI 4 "any_mem_operand"
[(plus:SI (match_operand:SI 1 "register_operand" "0")
(match_operand:SI 2 "short_immediate_operand" "I"))])
(match_operand:QI 3 "register_operand" "c"))
(set (match_operand:SI 0 "dest_reg_operand" "=w")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"stb.a%V4 %3,[%0,%2]"
[(set_attr "type" "store")
(set_attr "length" "4")])
;; ??? pattern may have to be re-written
;; Note: no 16-bit variant for this pattern
(define_insn "*loadhi_update"
[(set (match_operand:HI 3 "dest_reg_operand" "=r,r")
(match_operator:HI 4 "any_mem_operand"
[(plus:SI (match_operand:SI 1 "register_operand" "0,0")
(match_operand:SI 2 "nonmemory_operand" "rCm2,Cal"))]))
(set (match_operand:SI 0 "dest_reg_operand" "=w,w")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"ld%_.a%V4 %3,[%0,%2]"
[(set_attr "type" "load,load")
(set_attr "length" "4,8")])
(define_insn "*load_zeroextendhisi_update"
[(set (match_operand:SI 3 "dest_reg_operand" "=r,r")
(zero_extend:SI (match_operator:HI 4 "any_mem_operand"
[(plus:SI (match_operand:SI 1 "register_operand" "0,0")
(match_operand:SI 2 "nonmemory_operand" "rCm2,Cal"))])))
(set (match_operand:SI 0 "dest_reg_operand" "=r,r")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"ld%_.a%V4 %3,[%0,%2]"
[(set_attr "type" "load,load")
(set_attr "length" "4,8")])
;; Note: no 16-bit variant for this instruction
(define_insn "*load_signextendhisi_update"
[(set (match_operand:SI 3 "dest_reg_operand" "=r,r")
(sign_extend:SI (match_operator:HI 4 "any_mem_operand"
[(plus:SI (match_operand:SI 1 "register_operand" "0,0")
(match_operand:SI 2 "nonmemory_operand" "rCm2,Cal"))])))
(set (match_operand:SI 0 "dest_reg_operand" "=w,w")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"ld%_.x.a%V4 %3,[%0,%2]"
[(set_attr "type" "load,load")
(set_attr "length" "4,8")])
(define_insn "*storehi_update"
[(set (match_operator:HI 4 "any_mem_operand"
[(plus:SI (match_operand:SI 1 "register_operand" "0")
(match_operand:SI 2 "short_immediate_operand" "I"))])
(match_operand:HI 3 "register_operand" "c"))
(set (match_operand:SI 0 "dest_reg_operand" "=w")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"st%_.a%V4 %3,[%0,%2]"
[(set_attr "type" "store")
(set_attr "length" "4")])
;; No 16-bit variant for this instruction pattern
(define_insn "*loadsi_update"
[(set (match_operand:SI 3 "dest_reg_operand" "=r,r")
(match_operator:SI 4 "any_mem_operand"
[(plus:SI (match_operand:SI 1 "register_operand" "0,0")
(match_operand:SI 2 "nonmemory_operand" "rCm2,Cal"))]))
(set (match_operand:SI 0 "dest_reg_operand" "=w,w")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"ld.a%V4 %3,[%0,%2]"
[(set_attr "type" "load,load")
(set_attr "length" "4,8")])
(define_insn "*storesi_update"
[(set (match_operator:SI 4 "any_mem_operand"
[(plus:SI (match_operand:SI 1 "register_operand" "0")
(match_operand:SI 2 "short_immediate_operand" "I"))])
(match_operand:SI 3 "register_operand" "c"))
(set (match_operand:SI 0 "dest_reg_operand" "=w")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"st.a%V4 %3,[%0,%2]"
[(set_attr "type" "store")
(set_attr "length" "4")])
(define_insn "*loadsf_update"
[(set (match_operand:SF 3 "dest_reg_operand" "=r,r")
(match_operator:SF 4 "any_mem_operand"
[(plus:SI (match_operand:SI 1 "register_operand" "0,0")
(match_operand:SI 2 "nonmemory_operand" "rCm2,Cal"))]))
(set (match_operand:SI 0 "dest_reg_operand" "=w,w")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"ld.a%V4 %3,[%0,%2]"
[(set_attr "type" "load,load")
(set_attr "length" "4,8")])
(define_insn "*storesf_update"
[(set (match_operator:SF 4 "any_mem_operand"
[(plus:SI (match_operand:SI 1 "register_operand" "0")
(match_operand:SI 2 "short_immediate_operand" "I"))])
(match_operand:SF 3 "register_operand" "c"))
(set (match_operand:SI 0 "dest_reg_operand" "=w")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"st.a%V4 %3,[%0,%2]"
[(set_attr "type" "store")
(set_attr "length" "4")])
;; Conditional move instructions.
(define_expand "movsicc"
[(set (match_operand:SI 0 "dest_reg_operand" "")
(if_then_else:SI (match_operand 1 "comparison_operator" "")
(match_operand:SI 2 "nonmemory_operand" "")
(match_operand:SI 3 "register_operand" "")))]
""
"
operands[1] = gen_compare_reg (operands[1], VOIDmode);
if (operands[1] == NULL_RTX)
FAIL;
")
(define_expand "movdicc"
[(set (match_operand:DI 0 "dest_reg_operand" "")
(if_then_else:DI(match_operand 1 "comparison_operator" "")
(match_operand:DI 2 "nonmemory_operand" "")
(match_operand:DI 3 "register_operand" "")))]
""
"
operands[1] = gen_compare_reg (operands[1], VOIDmode);
if (operands[1] == NULL_RTX)
FAIL;
")
(define_expand "movsfcc"
[(set (match_operand:SF 0 "dest_reg_operand" "")
(if_then_else:SF (match_operand 1 "comparison_operator" "")
(match_operand:SF 2 "nonmemory_operand" "")
(match_operand:SF 3 "register_operand" "")))]
""
"
operands[1] = gen_compare_reg (operands[1], VOIDmode);
if (operands[1] == NULL_RTX)
FAIL;
")
(define_expand "movdfcc"
[(set (match_operand:DF 0 "dest_reg_operand" "")
(if_then_else:DF (match_operand 1 "comparison_operator" "")
(match_operand:DF 2 "nonmemory_operand" "")
(match_operand:DF 3 "register_operand" "")))]
""
"
operands[1] = gen_compare_reg (operands[1], VOIDmode);
if (operands[1] == NULL_RTX)
FAIL;
")
(define_insn "*movsicc_insn"
[(set (match_operand:SI 0 "dest_reg_operand" "=w,w")
(if_then_else:SI (match_operator 3 "proper_comparison_operator"
[(match_operand 4 "cc_register" "") (const_int 0)])
(match_operand:SI 1 "nonmemory_operand" "cL,Cal")
(match_operand:SI 2 "register_operand" "0,0")))]
""
{
if (rtx_equal_p (operands[1], const0_rtx) && GET_CODE (operands[3]) == NE
&& IN_RANGE (REGNO (operands[0]) ^ 4, 4, 11))
return "sub%?.ne %0,%0,%0";
/* ??? might be good for speed on ARC600 too, *if* properly scheduled. */
if ((optimize_size && (!TARGET_ARC600_FAMILY))
&& rtx_equal_p (operands[1], constm1_rtx)
&& GET_CODE (operands[3]) == LTU)
return "sbc.cs %0,%0,%0";
return "mov.%d3 %0,%1";
}
[(set_attr "type" "cmove,cmove")
(set_attr "length" "4,8")])
;; When there's a mask of a single bit, and then a compare to 0 or 1,
;; if the single bit is the sign bit, then GCC likes to convert this
;; into a sign extend and a compare less than, or greater to zero.
;; This is usually fine, except for the NXP400 where we have access to
;; a bit test instruction, along with a special short load instruction
;; (from CMEM), that doesn't support sign-extension on load.
;;
;; This peephole optimisation attempts to restore the use of bit-test
;; in those cases where it is useful to do so.
(define_peephole2
[(set (match_operand:SI 0 "register_operand" "")
(sign_extend:SI
(match_operand:QI 1 "any_mem_operand" "")))
(set (reg:CC_ZN CC_REG)
(compare:CC_ZN (match_dup 0)
(const_int 0)))
(set (pc)
(if_then_else (match_operator 2 "ge_lt_comparison_operator"
[(reg:CC_ZN CC_REG) (const_int 0)])
(match_operand 3 "" "")
(match_operand 4 "" "")))]
"TARGET_NPS_CMEM
&& cmem_address (XEXP (operands[1], 0), SImode)
&& peep2_reg_dead_p (2, operands[0])
&& peep2_regno_dead_p (3, CC_REG)"
[(set (match_dup 0)
(zero_extend:SI
(match_dup 1)))
(set (reg:CC_ZN CC_REG)
(compare:CC_ZN (zero_extract:SI
(match_dup 0)
(const_int 1)
(const_int 7))
(const_int 0)))
(set (pc)
(if_then_else (match_dup 2)
(match_dup 3)
(match_dup 4)))]
"if (GET_CODE (operands[2]) == GE)
operands[2] = gen_rtx_EQ (VOIDmode, gen_rtx_REG (CC_ZNmode, 61), const0_rtx);
else
operands[2] = gen_rtx_NE (VOIDmode, gen_rtx_REG (CC_ZNmode, 61), const0_rtx);")
; Try to generate more short moves, and/or less limms, by substituting a
; conditional move with a conditional sub.
(define_peephole2
[(set (match_operand:SI 0 "compact_register_operand")
(match_operand:SI 1 "const_int_operand"))
(set (match_dup 0)
(if_then_else:SI (match_operator 3 "proper_comparison_operator"
[(match_operand 4 "cc_register" "") (const_int 0)])
(match_operand:SI 2 "const_int_operand" "")
(match_dup 0)))]
"!satisfies_constraint_P (operands[1])
&& satisfies_constraint_P (operands[2])
&& UNSIGNED_INT6 (INTVAL (operands[2]) - INTVAL (operands[1]))"
[(set (match_dup 0) (match_dup 2))
(cond_exec
(match_dup 3)
(set (match_dup 0)
(plus:SI (match_dup 0) (match_dup 1))))]
"operands[3] = gen_rtx_fmt_ee (REVERSE_CONDITION (GET_CODE (operands[3]),
GET_MODE (operands[4])),
VOIDmode, operands[4], const0_rtx);
operands[1] = GEN_INT (INTVAL (operands[1]) - INTVAL (operands[2]));")
(define_insn "*movdicc_insn"
[(set (match_operand:DI 0 "dest_reg_operand" "=&w,w")
(if_then_else:DI (match_operator 3 "proper_comparison_operator"
[(match_operand 4 "cc_register" "") (const_int 0)])
(match_operand:DI 1 "nonmemory_operand" "c,i")
(match_operand:DI 2 "register_operand" "0,0")))]
""
"*
{
switch (which_alternative)
{
default:
case 0 :
/* We normally copy the low-numbered register first. However, if
the first register operand 0 is the same as the second register of
operand 1, we must copy in the opposite order. */
if (REGNO (operands[0]) == REGNO (operands[1]) + 1)
return \"mov.%d3 %R0,%R1\;mov.%d3 %0,%1\";
else
return \"mov.%d3 %0,%1\;mov.%d3 %R0,%R1\";
case 1 :
return \"mov.%d3 %L0,%L1\;mov.%d3 %H0,%H1\";
}
}"
[(set_attr "type" "cmove,cmove")
(set_attr "length" "8,16")])
(define_insn "*movsfcc_insn"
[(set (match_operand:SF 0 "dest_reg_operand" "=w,w")
(if_then_else:SF (match_operator 3 "proper_comparison_operator"
[(match_operand 4 "cc_register" "") (const_int 0)])
(match_operand:SF 1 "nonmemory_operand" "c,E")
(match_operand:SF 2 "register_operand" "0,0")))]
""
"@
mov.%d3 %0,%1
mov.%d3 %0,%1 ; %A1"
[(set_attr "type" "cmove,cmove")])
(define_insn "*movdfcc_insn"
[(set (match_operand:DF 0 "dest_reg_operand" "=w,w")
(if_then_else:DF (match_operator 1 "proper_comparison_operator"
[(match_operand 4 "cc_register" "") (const_int 0)])
(match_operand:DF 2 "nonmemory_operand" "c,E")
(match_operand:DF 3 "register_operand" "0,0")))]
""
"*
{
switch (which_alternative)
{
default:
case 0 :
/* We normally copy the low-numbered register first. However, if
the first register operand 0 is the same as the second register of
operand 1, we must copy in the opposite order. */
if (REGNO (operands[0]) == REGNO (operands[2]) + 1)
return \"mov.%d1 %R0,%R2\;mov.%d1 %0,%2\";
else
return \"mov.%d1 %0,%2\;mov.%d1 %R0,%R2\";
case 1 :
return \"mov.%d1 %L0,%L2\;mov.%d1 %H0,%H2; %A2 \";
}
}"
[(set_attr "type" "cmove,cmove")
(set_attr "length" "8,16")])
;; -------------------------------------------------------------------
;; Sign/Zero extension
;; -------------------------------------------------------------------
(define_insn "*zero_extendqihi2_i"
[(set (match_operand:HI 0 "dest_reg_operand" "=q,q,r,r,r,r")
(zero_extend:HI
(match_operand:QI 1 "nonvol_nonimm_operand" "0,q,0,r,Ucm,m")))]
""
"@
extb%?\\t%0,%1
extb%?\\t%0,%1
bmsk%?\\t%0,%1,7
extb\\t%0,%1
xldb%U1\\t%0,%1
ldb%U1\\t%0,%1"
[(set_attr "type" "unary,unary,unary,unary,load,load")
(set_attr "iscompact" "maybe,true,false,false,false,false")
(set_attr "predicable" "no,no,yes,no,no,no")])
(define_expand "zero_extendqihi2"
[(set (match_operand:HI 0 "dest_reg_operand" "")
(zero_extend:HI (match_operand:QI 1 "nonvol_nonimm_operand" "")))]
""
""
)
(define_insn "*zero_extendqisi2_ac"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,q,r,r,q,!*x,r,r")
(zero_extend:SI
(match_operand:QI 1 "nonvol_nonimm_operand" "0,q,0,r,T,Usd,Ucm,m")))]
""
"@
extb%?\\t%0,%1
extb%?\\t%0,%1
bmsk%?\\t%0,%1,7
extb\\t%0,%1
ldb%?\\t%0,%1
ldb%?\\t%0,%1
xldb%U1\\t%0,%1
ldb%U1\\t%0,%1"
[(set_attr "type" "unary,unary,unary,unary,load,load,load,load")
(set_attr "iscompact" "maybe,true,false,false,true,true,false,false")
(set_attr "predicable" "no,no,yes,no,no,no,no,no")])
(define_expand "zero_extendqisi2"
[(set (match_operand:SI 0 "dest_reg_operand" "")
(zero_extend:SI (match_operand:QI 1 "nonvol_nonimm_operand" "")))]
""
""
)
(define_insn "*zero_extendhisi2_i"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,q,r,r,!x,q,r,r")
(zero_extend:SI
(match_operand:HI 1 "nonvol_nonimm_operand" "0,q,0,r,Usd,T,Ucm,m")))]
""
"@
ext%_%?\\t%0,%1
ext%_%?\\t%0,%1
bmsk%?\\t%0,%1,15
ext%_\\t%0,%1
ld%_%?\\t%0,%1
ld%_%?\\t%0,%1
xldw%U1\\t%0,%1
ld%_%U1%V1\\t%0,%1"
[(set_attr "type" "unary,unary,unary,unary,load,load,load,load")
(set_attr "iscompact" "maybe,true,false,false,true,true,false,false")
(set_attr "predicable" "no,no,yes,no,no,no,no,no")])
(define_expand "zero_extendhisi2"
[(set (match_operand:SI 0 "dest_reg_operand" "")
(zero_extend:SI (match_operand:HI 1 "nonvol_nonimm_operand" "")))]
""
""
)
;; Sign extension instructions.
(define_insn "*extendqihi2_i"
[(set (match_operand:HI 0 "dest_reg_operand" "=q,r,r,r")
(sign_extend:HI
(match_operand:QI 1 "nonvol_nonimm_operand" "q,r,Uex,m")))]
""
"@
sexb%?\\t%0,%1
sexb\\t%0,%1
ldb.x%U1\\t%0,%1
ldb.x%U1\\t%0,%1"
[(set_attr "type" "unary,unary,load,load")
(set_attr "iscompact" "true,false,false,false")
(set_attr "length" "*,*,*,8")])
(define_expand "extendqihi2"
[(set (match_operand:HI 0 "dest_reg_operand" "")
(sign_extend:HI (match_operand:QI 1 "nonvol_nonimm_operand" "")))]
""
""
)
(define_insn "*extendqisi2_ac"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,r,r,r")
(sign_extend:SI
(match_operand:QI 1 "nonvol_nonimm_operand" "q,r,Uex,m")))]
""
"@
sexb%?\\t%0,%1
sexb\\t%0,%1
ldb.x%U1\\t%0,%1
ldb.x%U1\\t%0,%1"
[(set_attr "type" "unary,unary,load,load")
(set_attr "iscompact" "true,false,false,false")
(set_attr "length" "*,*,*,8")])
(define_expand "extendqisi2"
[(set (match_operand:SI 0 "dest_reg_operand" "")
(sign_extend:SI (match_operand:QI 1 "nonvol_nonimm_operand" "")))]
""
""
)
(define_insn "*extendhisi2_i"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,r,q,r,r")
(sign_extend:SI
(match_operand:HI 1 "nonvol_nonimm_operand" "q,r,Ucd,Uex,m")))]
""
"@
sex%_%?\\t%0,%1
sex%_\\t%0,%1
ldh%?.x\\t%0,%1%&
ld%_.x%U1%V1\\t%0,%1
ld%_.x%U1%V1\\t%0,%1"
[(set_attr "type" "unary,unary,load,load,load")
(set_attr "iscompact" "true,false,true,false,false")
(set_attr "length" "*,*,*,4,8")])
(define_expand "extendhisi2"
[(set (match_operand:SI 0 "dest_reg_operand" "")
(sign_extend:SI (match_operand:HI 1 "nonvol_nonimm_operand" "")))]
""
""
)
;; Unary arithmetic insns
;; We allow constant operands to enable late constant propagation, but it is
;; not worth while to have more than one dedicated alternative to output them -
;; if we are really worried about getting these the maximum benefit of all
;; the available alternatives, we should add an extra pass to fold such
;; operations to movsi.
;; Absolute instructions
(define_insn "abssi2"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,w,w")
(abs:SI (match_operand:SI 1 "nonmemory_operand" "q,cL,Cal")))]
""
"abs%? %0,%1%&"
[(set_attr "type" "two_cycle_core")
(set_attr "length" "*,4,8")
(set_attr "iscompact" "true,false,false")])
;; Maximum and minimum insns
(define_insn "smaxsi3"
[(set (match_operand:SI 0 "dest_reg_operand" "=r, r, r")
(smax:SI (match_operand:SI 1 "register_operand" "%0, r, r")
(match_operand:SI 2 "nonmemory_operand" "rL,rL,Cal")))]
""
"max%?\\t%0,%1,%2"
[(set_attr "type" "two_cycle_core")
(set_attr "length" "4,4,8")
(set_attr "predicable" "yes,no,no")]
)
(define_insn "sminsi3"
[(set (match_operand:SI 0 "dest_reg_operand" "=r, r, r")
(smin:SI (match_operand:SI 1 "register_operand" "%0, r, r")
(match_operand:SI 2 "nonmemory_operand" "rL,rL,Cal")))]
""
"min%?\\t%0,%1,%2"
[(set_attr "type" "two_cycle_core")
(set_attr "length" "4,4,8")
(set_attr "predicable" "yes,no,no")]
)
;; Arithmetic instructions.
; We say an insn can be conditionalized if this doesn't introduce a long
; immediate. We set the type such that we still have good scheduling if the
; insn is conditionalized.
; ??? It would make sense to allow introduction of long immediates, but
; we'd need to communicate to the ccfsm machinery the extra cost.
; The alternatives in the constraints still serve three purposes:
; - estimate insn size assuming conditional execution
; - guide reload to re-order the second and third operand to get a better fit.
; - give tentative insn type to guide scheduling
; N.B. "%" for commutativity doesn't help when there is another matching
; (but longer) alternative.
; We avoid letting this pattern use LP_COUNT as a register by specifying
; register class 'W' instead of 'w'.
(define_insn_and_split "*addsi3_mixed"
;; 0 1 2 3 4 5 6 7 8 9 a b c d e f 10 11 12
[(set (match_operand:SI 0 "dest_reg_operand" "=q,q, h,!*Rsd, q,Rcb, q, q, q, r,r, r, W, W,W, W, q, r, W")
(plus:SI (match_operand:SI 1 "register_operand" "%0,c, 0, q, 0, 0,Rcb, q, 0, 0,r, 0, c, c,0, 0, 0, 0, c")
(match_operand:SI 2 "nonmemory_operand" "cL,0, Cm1, L,CL2,Csp,CM4,qK,cO,rL,0,rCca,cLCmL,Cca,I,C2a,Cal,Cal,Cal")))]
""
{
arc_output_addsi (operands, arc_ccfsm_cond_exec_p (), true);
return "";
}
"&& reload_completed && get_attr_length (insn) == 8
&& satisfies_constraint_I (operands[2])
&& GET_CODE (PATTERN (insn)) != COND_EXEC"
[(set (match_dup 0) (match_dup 3)) (set (match_dup 0) (match_dup 4))]
"split_addsi (operands);"
[(set_attr "type" "*,*,*,*,two_cycle_core,two_cycle_core,*,*,*,*,*,two_cycle_core,*,two_cycle_core,*,two_cycle_core,*,*,*")
(set (attr "iscompact")
(cond [(match_test "~arc_output_addsi (operands, false, false) & 2")
(const_string "false")
(match_operand 2 "long_immediate_operand" "")
(const_string "maybe_limm")]
(const_string "maybe")))
(set_attr "length" "*,*,*,*,*,*,*,*,*,4,4,4,4,4,4,4,*,8,8")
(set_attr "predicable" "no,no,no,no,no,no,no,no,no,yes,yes,yes,no,no,no,no,no,yes,no")
(set_attr "cond" "canuse,nocond,nocond,nocond,canuse,canuse,nocond,nocond,nocond,canuse,canuse,canuse,nocond,nocond,canuse_limm,canuse_limm,canuse,canuse,nocond")
])
;; ARCv2 MPYW and MPYUW
(define_expand "mulhisi3"
[(set (match_operand:SI 0 "register_operand" "")
(mult:SI (sign_extend:SI (match_operand:HI 1 "register_operand" ""))
(sign_extend:SI (match_operand:HI 2 "nonmemory_operand" ""))))]
"TARGET_MPYW"
"{
if (CONSTANT_P (operands[2]))
{
emit_insn (gen_mulhisi3_imm (operands[0], operands[1], operands[2]));
DONE;
}
}"
)
(define_insn "mulhisi3_imm"
[(set (match_operand:SI 0 "register_operand" "=r,r,r, r, r")
(mult:SI (sign_extend:SI (match_operand:HI 1 "register_operand" "0,r,0, 0, r"))
(match_operand:HI 2 "short_const_int_operand" "L,L,I,C16,C16")))]
"TARGET_MPYW"
"mpyw%? %0,%1,%2"
[(set_attr "length" "4,4,4,8,8")
(set_attr "iscompact" "false")
(set_attr "type" "mul16_em")
(set_attr "predicable" "yes,no,no,yes,no")
(set_attr "cond" "canuse,nocond,nocond,canuse_limm,nocond")
])
(define_insn "mulhisi3_reg"
[(set (match_operand:SI 0 "register_operand" "=q,r,r")
(mult:SI (sign_extend:SI (match_operand:HI 1 "register_operand" "0,0,r"))
(sign_extend:SI (match_operand:HI 2 "nonmemory_operand" "q,r,r"))))]
"TARGET_MPYW"
"mpyw%? %0,%1,%2"
[(set_attr "length" "*,4,4")
(set_attr "iscompact" "maybe,false,false")
(set_attr "type" "mul16_em")
(set_attr "predicable" "yes,yes,no")
(set_attr "cond" "canuse,canuse,nocond")
])
(define_expand "umulhisi3"
[(set (match_operand:SI 0 "register_operand" "")
(mult:SI (zero_extend:SI (match_operand:HI 1 "register_operand" ""))
(zero_extend:SI (match_operand:HI 2 "arc_short_operand" ""))))]
"TARGET_MPYW"
"{
if (CONSTANT_P (operands[2]))
{
emit_insn (gen_umulhisi3_imm (operands[0], operands[1], operands[2]));
DONE;
}
}"
)
(define_insn "umulhisi3_imm"
[(set (match_operand:SI 0 "register_operand" "=r, r, r, r, r")
(mult:SI (zero_extend:SI (match_operand:HI 1 "register_operand" "%0, r, 0, 0, r"))
(match_operand:HI 2 "short_unsigned_const_operand" " L, L,J12,J16,J16")))]
"TARGET_MPYW"
"mpyuw%? %0,%1,%2"
[(set_attr "length" "4,4,4,8,8")
(set_attr "iscompact" "false")
(set_attr "type" "mul16_em")
(set_attr "predicable" "yes,no,no,yes,no")
(set_attr "cond" "canuse,nocond,nocond,canuse_limm,nocond")
])
(define_insn "umulhisi3_reg"
[(set (match_operand:SI 0 "register_operand" "=q, r, r")
(mult:SI (zero_extend:SI (match_operand:HI 1 "register_operand" "%0, 0, r"))
(zero_extend:SI (match_operand:HI 2 "register_operand" "q, r, r"))))]
"TARGET_MPYW"
"mpyuw%? %0,%1,%2"
[(set_attr "length" "*,4,4")
(set_attr "iscompact" "maybe,false,false")
(set_attr "type" "mul16_em")
(set_attr "predicable" "yes,yes,no")
(set_attr "cond" "canuse,canuse,nocond")
])
;; ARC700/ARC600/V2 multiply
;; SI <- SI * SI
(define_expand "mulsi3"
[(set (match_operand:SI 0 "register_operand" "")
(mult:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))]
"TARGET_ANY_MPY"
{
if (TARGET_MUL64_SET)
{
emit_insn (gen_mulsi64 (operands[0], operands[1], operands[2]));
DONE;
}
else if (TARGET_MULMAC_32BY16_SET)
{
emit_insn (gen_mulsi32x16 (operands[0], operands[1], operands[2]));
DONE;
}
})
(define_insn_and_split "mulsi32x16"
[(set (match_operand:SI 0 "register_operand" "=w")
(mult:SI (match_operand:SI 1 "register_operand" "%c")
(match_operand:SI 2 "nonmemory_operand" "ci")))
(clobber (reg:DI MUL32x16_REG))]
"TARGET_MULMAC_32BY16_SET"
"#"
"TARGET_MULMAC_32BY16_SET && reload_completed"
[(const_int 0)]
{
if (immediate_operand (operands[2], SImode)
&& INTVAL (operands[2]) >= 0
&& INTVAL (operands[2]) <= 65535)
{
emit_insn (gen_umul_600 (operands[1], operands[2],
gen_acc2 (), gen_acc1 ()));
emit_move_insn (operands[0], gen_acc2 ());
DONE;
}
emit_insn (gen_umul_600 (operands[1], operands[2],
gen_acc2 (), gen_acc1 ()));
emit_insn (gen_mac_600 (operands[1], operands[2],
gen_acc2 (), gen_acc1 ()));
emit_move_insn (operands[0], gen_acc2 ());
DONE;
}
[(set_attr "type" "multi")
(set_attr "length" "8")])
; mululw conditional execution without a LIMM clobbers an input register;
; we'd need a different pattern to describe this.
; To make the conditional execution valid for the LIMM alternative, we
; have to emit the LIMM before the register operand.
(define_insn "umul_600"
[(set (match_operand:SI 2 "acc2_operand" "")
(mult:SI (match_operand:SI 0 "register_operand" "c,c,c")
(zero_extract:SI (match_operand:SI 1 "nonmemory_operand"
"c,L,Cal")
(const_int 16)
(const_int 0))))
(clobber (match_operand:SI 3 "acc1_operand" ""))]
"TARGET_MULMAC_32BY16_SET"
"mululw 0, %0, %1"
[(set_attr "length" "4,4,8")
(set_attr "type" "mulmac_600")
(set_attr "predicable" "no")
(set_attr "cond" "nocond")])
(define_insn "mac_600"
[(set (match_operand:SI 2 "acc2_operand" "")
(plus:SI
(mult:SI (match_operand:SI 0 "register_operand" "c,c,c")
(ashift:SI
(zero_extract:SI (match_operand:SI 1 "nonmemory_operand" "c,L,Cal")
(const_int 16)
(const_int 16))
(const_int 16)))
(match_dup 2)))
(clobber (match_operand:SI 3 "acc1_operand" ""))]
"TARGET_MULMAC_32BY16_SET"
"machlw%? 0, %0, %1"
[(set_attr "length" "4,4,8")
(set_attr "type" "mulmac_600, mulmac_600, mulmac_600")
(set_attr "predicable" "no, no, yes")
(set_attr "cond" "nocond, canuse_limm, canuse")])
; The gcc-internal representation may differ from the hardware
; register number in order to allow the generic code to correctly
; split the concatenation of mhi and mlo.
(define_insn_and_split "mulsi64"
[(set (match_operand:SI 0 "register_operand" "=w")
(mult:SI (match_operand:SI 1 "register_operand" "%c")
(match_operand:SI 2 "nonmemory_operand" "ci")))
(clobber (reg:DI MUL64_OUT_REG))]
"TARGET_MUL64_SET"
"#"
"TARGET_MUL64_SET && reload_completed"
[(const_int 0)]
{
rtx mhi = gen_rtx_REG (SImode, R59_REG);
rtx mlo = gen_rtx_REG (SImode, R58_REG);
emit_insn (gen_mulsi_600 (operands[1], operands[2], mlo, mhi));
emit_move_insn (operands[0], mlo);
DONE;
}
[(set_attr "type" "multi")
(set_attr "length" "8")])
(define_insn "mulsi_600"
[(set (match_operand:SI 2 "mlo_operand" "")
(mult:SI (match_operand:SI 0 "register_operand" "%q,c,c,c")
(match_operand:SI 1 "nonmemory_operand" "q,cL,I,Cal")))
(clobber (match_operand:SI 3 "mhi_operand" ""))]
"TARGET_MUL64_SET"
"mul64%?\\t0,%0,%1"
[(set_attr "length" "*,4,4,8")
(set_attr "iscompact" "maybe,false,false,false")
(set_attr "type" "multi,multi,multi,multi")
(set_attr "predicable" "yes,yes,no,yes")
(set_attr "cond" "canuse,canuse,canuse_limm,canuse")])
(define_insn_and_split "mulsidi_600"
[(set (match_operand:DI 0 "register_operand" "=r,r, r")
(mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "%r,r, r"))
(sign_extend:DI (match_operand:SI 2 "nonmemory_operand" "rL,L,C32"))))
(clobber (reg:DI R58_REG))]
"TARGET_MUL64_SET"
"#"
"TARGET_MUL64_SET && reload_completed"
[(const_int 0)]
{
int hi = !TARGET_BIG_ENDIAN;
int lo = !hi;
rtx lr = operand_subword (operands[0], lo, 0, DImode);
rtx hr = operand_subword (operands[0], hi, 0, DImode);
emit_insn (gen_mul64 (operands[1], operands[2]));
emit_move_insn (lr, gen_rtx_REG (SImode, R58_REG));
emit_move_insn (hr, gen_rtx_REG (SImode, R59_REG));
DONE;
}
[(set_attr "type" "multi")
(set_attr "length" "4,4,8")])
(define_insn "mul64"
[(set (reg:DI MUL64_OUT_REG)
(mult:DI
(sign_extend:DI (match_operand:SI 0 "register_operand" "%q, c,c, c"))
(sign_extend:DI (match_operand:SI 1 "nonmemory_operand" "q,cL,L,C32"))))]
"TARGET_MUL64_SET"
"mul64%? \t0, %0, %1%&"
[(set_attr "length" "*,4,4,8")
(set_attr "iscompact" "maybe,false,false,false")
(set_attr "type" "multi,multi,multi,multi")
(set_attr "predicable" "yes,yes,no,yes")
(set_attr "cond" "canuse,canuse,canuse_limm,canuse")])
(define_insn_and_split "umulsidi_600"
[(set (match_operand:DI 0 "register_operand" "=r,r, r")
(mult:DI (zero_extend:DI (match_operand:SI 1 "register_operand" "%r,r, r"))
(zero_extend:DI (match_operand:SI 2 "nonmemory_operand" "rL,L,C32"))))
(clobber (reg:DI R58_REG))]
"TARGET_MUL64_SET"
"#"
"TARGET_MUL64_SET && reload_completed"
[(const_int 0)]
{
int hi = !TARGET_BIG_ENDIAN;
int lo = !hi;
rtx lr = operand_subword (operands[0], lo, 0, DImode);
rtx hr = operand_subword (operands[0], hi, 0, DImode);
emit_insn (gen_mulu64 (operands[1], operands[2]));
emit_move_insn (lr, gen_rtx_REG (SImode, R58_REG));
emit_move_insn (hr, gen_rtx_REG (SImode, R59_REG));
DONE;
}
[(set_attr "type" "umulti")
(set_attr "length" "4,4,8")])
(define_insn "mulu64"
[(set (reg:DI MUL64_OUT_REG)
(mult:DI
(zero_extend:DI (match_operand:SI 0 "register_operand" "%c,c,c"))
(zero_extend:DI (match_operand:SI 1 "nonmemory_operand" "cL,L,C32"))))]
"TARGET_MUL64_SET"
"mulu64%? \t0, %0, %1%&"
[(set_attr "length" "4,4,8")
(set_attr "iscompact" "false")
(set_attr "type" "umulti")
(set_attr "predicable" "yes,no,yes")
(set_attr "cond" "canuse,canuse_limm,canuse")])
; ARC700 mpy* instructions: This is a multi-cycle extension, and thus 'w'
; may not be used as destination constraint.
; The result of mpy and mpyu is the same except for flag setting (if enabled),
; but mpyu is faster for the standard multiplier.
; Note: we must make sure LP_COUNT is not one of the destination
; registers, since it cannot be the destination of a multi-cycle insn
; like MPY or MPYU.
(define_insn "mulsi3_700"
[(set (match_operand:SI 0 "mpy_dest_reg_operand" "=r, r,r, r,r")
(mult:SI (match_operand:SI 1 "register_operand" "%0, r,0, 0,r")
(match_operand:SI 2 "nonmemory_operand" "rL,rL,I,Cal,Cal")))]
"TARGET_ARC700_MPY"
"mpyu%?\\t%0,%1,%2"
[(set_attr "length" "4,4,4,8,8")
(set_attr "type" "umulti")
(set_attr "predicable" "yes,no,no,yes,no")
(set_attr "cond" "canuse,nocond,canuse_limm,canuse,nocond")])
; ARCv2 has no penalties between mpy and mpyu. So, we use mpy because of its
; short variant. LP_COUNT constraints are still valid.
(define_insn "mulsi3_v2"
[(set (match_operand:SI 0 "mpy_dest_reg_operand" "=q,q, r, r,r, r, r")
(mult:SI (match_operand:SI 1 "register_operand" "%0,q, 0, r,0, 0, c")
(match_operand:SI 2 "nonmemory_operand" "q,0,rL,rL,I,Cal,Cal")))]
"TARGET_MULTI"
"@
mpy%?\\t%0,%1,%2
mpy%?\\t%0,%2,%1
mpy%?\\t%0,%1,%2
mpy%?\\t%0,%1,%2
mpy%?\\t%0,%1,%2
mpy%?\\t%0,%1,%2
mpy%?\\t%0,%1,%2"
[(set_attr "length" "*,*,4,4,4,8,8")
(set_attr "iscompact" "maybe,maybe,false,false,false,false,false")
(set_attr "type" "umulti")
(set_attr "predicable" "no,no,yes,no,no,yes,no")
(set_attr "cond" "nocond,nocond,canuse,nocond,canuse_limm,canuse,nocond")])
(define_expand "mulsidi3"
[(set (match_operand:DI 0 "register_operand" "")
(mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" ""))
(sign_extend:DI (match_operand:SI 2 "nonmemory_operand" ""))))]
"TARGET_ANY_MPY"
{
if (TARGET_PLUS_MACD)
{
if (CONST_INT_P (operands[2]))
{
emit_insn (gen_mpyd_imm_arcv2hs (operands[0], operands[1], operands[2]));
}
else
{
emit_insn (gen_mpyd_arcv2hs (operands[0], operands[1], operands[2]));
}
DONE;
}
if (TARGET_MPY)
{
operands[2] = force_reg (SImode, operands[2]);
if (!register_operand (operands[0], DImode))
{
rtx result = gen_reg_rtx (DImode);
operands[2] = force_reg (SImode, operands[2]);
emit_insn (gen_mulsidi3 (result, operands[1], operands[2]));
emit_move_insn (operands[0], result);
DONE;
}
}
else if (TARGET_MUL64_SET)
{
emit_insn (gen_mulsidi_600 (operands[0], operands[1], operands[2]));
DONE;
}
else if (TARGET_MULMAC_32BY16_SET)
{
operands[2] = force_reg (SImode, operands[2]);
emit_insn (gen_mulsidi64 (operands[0], operands[1], operands[2]));
DONE;
}
operands[2] = force_reg (SImode, operands[2]);
})
(define_insn_and_split "mulsidi64"
[(set (match_operand:DI 0 "register_operand" "=w")
(mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "%c"))
(sign_extend:DI (match_operand:SI 2 "extend_operand" "ci"))))
(clobber (reg:DI MUL32x16_REG))]
"TARGET_MULMAC_32BY16_SET"
"#"
"TARGET_MULMAC_32BY16_SET && reload_completed"
[(const_int 0)]
{
rtx result_hi = gen_highpart (SImode, operands[0]);
rtx result_low = gen_lowpart (SImode, operands[0]);
emit_insn (gen_mul64_600 (operands[1], operands[2]));
emit_insn (gen_mac64_600 (result_hi, operands[1], operands[2]));
emit_move_insn (result_low, gen_acc2 ());
DONE;
}
[(set_attr "type" "multi")
(set_attr "length" "8")])
(define_insn "mul64_600"
[(set (reg:DI MUL32x16_REG)
(mult:DI (sign_extend:DI (match_operand:SI 0 "register_operand"
"c,c,c"))
(zero_extract:DI (match_operand:SI 1 "nonmemory_operand"
"c,L,Cal")
(const_int 16)
(const_int 0))))
]
"TARGET_MULMAC_32BY16_SET"
"mullw%? 0, %0, %1"
[(set_attr "length" "4,4,8")
(set_attr "type" "mulmac_600")
(set_attr "predicable" "no,no,yes")
(set_attr "cond" "nocond, canuse_limm, canuse")])
;; ??? check if this is canonical rtl
(define_insn "mac64_600"
[(set (reg:DI MUL32x16_REG)
(plus:DI
(mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "c,c,c"))
(ashift:DI
(sign_extract:DI (match_operand:SI 2 "nonmemory_operand" "c,L,Cal")
(const_int 16) (const_int 16))
(const_int 16)))
(reg:DI MUL32x16_REG)))
(set (match_operand:SI 0 "register_operand" "=w,w,w")
(zero_extract:SI
(plus:DI
(mult:DI (sign_extend:DI (match_dup 1))
(ashift:DI
(sign_extract:DI (match_dup 2)
(const_int 16) (const_int 16))
(const_int 16)))
(reg:DI MUL32x16_REG))
(const_int 32) (const_int 32)))]
"TARGET_MULMAC_32BY16_SET"
"machlw%? %0, %1, %2"
[(set_attr "length" "4,4,8")
(set_attr "type" "mulmac_600")
(set_attr "predicable" "no,no,yes")
(set_attr "cond" "nocond, canuse_limm, canuse")])
;; DI <- DI(signed SI) * DI(signed SI)
(define_insn_and_split "mulsidi3_700"
[(set (match_operand:DI 0 "register_operand" "=&r")
(mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "%c"))
(sign_extend:DI (match_operand:SI 2 "extend_operand" "cL"))))]
"TARGET_MPY && !TARGET_PLUS_MACD"
"#"
"&& reload_completed"
[(const_int 0)]
{
int hi = TARGET_BIG_ENDIAN ? 0 : UNITS_PER_WORD;
int lo = TARGET_BIG_ENDIAN ? UNITS_PER_WORD : 0;
rtx l0 = simplify_gen_subreg (word_mode, operands[0], DImode, lo);
rtx h0 = simplify_gen_subreg (word_mode, operands[0], DImode, hi);
emit_insn (gen_mulsi3_highpart (h0, operands[1], operands[2]));
emit_insn (gen_mulsi3 (l0, operands[1], operands[2]));
DONE;
}
[(set_attr "type" "multi")
(set_attr "length" "8")])
(define_insn "mulsi3_highpart"
[(set (match_operand:SI 0 "register_operand" "=r,r,r,r")
(truncate:SI
(lshiftrt:DI
(mult:DI
(sign_extend:DI (match_operand:SI 1 "register_operand" "%0,r,0,r"))
(sign_extend:DI (match_operand:SI 2 "extend_operand" "r,r,i,i")))
(const_int 32))))]
"TARGET_MPY"
"mpy%+%?\\t%0,%1,%2"
[(set_attr "length" "4,4,8,8")
(set_attr "type" "multi")
(set_attr "predicable" "yes,no,yes,no")
(set_attr "cond" "canuse,nocond,canuse,nocond")])
; Note that mpyhu has the same latency as mpy / mpyh,
; thus we use the type multi.
(define_insn "*umulsi3_highpart_i"
[(set (match_operand:SI 0 "register_operand" "=r,r,r,r")
(truncate:SI
(lshiftrt:DI
(mult:DI
(zero_extend:DI (match_operand:SI 1 "register_operand" "%0,r,0,r"))
(zero_extend:DI (match_operand:SI 2 "extend_operand" "r,r,i,i")))
(const_int 32))))]
"TARGET_MPY"
"mpy%+u%?\\t%0,%1,%2"
[(set_attr "length" "4,4,8,8")
(set_attr "type" "multi")
(set_attr "predicable" "yes,no,yes,no")
(set_attr "cond" "canuse,nocond,canuse,nocond")])
;; (zero_extend:DI (const_int)) leads to internal errors in combine, so we
;; need a separate pattern for immediates
;; ??? This is fine for combine, but not for reload.
(define_insn "umulsi3_highpart_int"
[(set (match_operand:SI 0 "register_operand" "=r, r, r,r, r")
(truncate:SI
(lshiftrt:DI
(mult:DI
(zero_extend:DI (match_operand:SI 1 "register_operand" " 0, r, 0, 0, r"))
(match_operand:DI 2 "immediate_usidi_operand" "L, L, I,Cal,Cal"))
(const_int 32))))]
"TARGET_MPY"
"mpy%+u%?\\t%0,%1,%2"
[(set_attr "length" "4,4,4,8,8")
(set_attr "type" "multi")
(set_attr "predicable" "yes,no,no,yes,no")
(set_attr "cond" "canuse,nocond,canuse_limm,canuse,nocond")])
(define_expand "umulsi3_highpart"
[(set (match_operand:SI 0 "general_operand" "")
(truncate:SI
(lshiftrt:DI
(mult:DI
(zero_extend:DI (match_operand:SI 1 "register_operand" ""))
(zero_extend:DI (match_operand:SI 2 "nonmemory_operand" "")))
(const_int 32))))]
"TARGET_MPY"
"
{
rtx target = operands[0];
if (!register_operand (target, SImode))
target = gen_reg_rtx (SImode);
if (CONST_INT_P (operands[2]) && INTVAL (operands[2]) < 0)
operands[2] = simplify_const_unary_operation (ZERO_EXTEND, DImode,
operands[2], SImode);
else if (!immediate_operand (operands[2], SImode))
operands[2] = gen_rtx_ZERO_EXTEND (DImode, operands[2]);
emit_insn (gen_umulsi3_highpart_int (target, operands[1], operands[2]));
if (target != operands[0])
emit_move_insn (operands[0], target);
DONE;
}")
(define_expand "umulsidi3"
[(set (match_operand:DI 0 "register_operand" "")
(mult:DI (zero_extend:DI (match_operand:SI 1 "register_operand" ""))
(zero_extend:DI (match_operand:SI 2 "nonmemory_operand" ""))))]
"TARGET_ANY_MPY"
{
if (TARGET_PLUS_MACD)
{
if (CONST_INT_P (operands[2]))
{
emit_insn (gen_mpydu_imm_arcv2hs (operands[0], operands[1], operands[2]));
}
else
{
emit_insn (gen_mpydu_arcv2hs (operands[0], operands[1], operands[2]));
}
DONE;
}
if (TARGET_MPY)
{
operands[2] = force_reg (SImode, operands[2]);
if (!register_operand (operands[0], DImode))
{
rtx result = gen_reg_rtx (DImode);
emit_insn (gen_umulsidi3 (result, operands[1], operands[2]));
emit_move_insn (operands[0], result);
DONE;
}
}
else if (TARGET_MUL64_SET)
{
operands[2] = force_reg (SImode, operands[2]);
emit_insn (gen_umulsidi_600 (operands[0], operands[1], operands[2]));
DONE;
}
else if (TARGET_MULMAC_32BY16_SET)
{
operands[2] = force_reg (SImode, operands[2]);
emit_insn (gen_umulsidi64 (operands[0], operands[1], operands[2]));
DONE;
}
else
{
gcc_unreachable ();
}
})
(define_insn_and_split "umulsidi64"
[(set (match_operand:DI 0 "register_operand" "=w")
(mult:DI (zero_extend:DI (match_operand:SI 1 "register_operand" "%c"))
(zero_extend:DI (match_operand:SI 2 "extend_operand" "ci"))))
(clobber (reg:DI MUL32x16_REG))]
"TARGET_MULMAC_32BY16_SET"
"#"
"TARGET_MULMAC_32BY16_SET && reload_completed"
[(const_int 0)]
{
rtx result_hi;
rtx result_low;
result_hi = gen_highpart (SImode, operands[0]);
result_low = gen_lowpart (SImode, operands[0]);
emit_insn (gen_umul64_600 (operands[1], operands[2]));
emit_insn (gen_umac64_600 (result_hi, operands[1], operands[2]));
emit_move_insn (result_low, gen_acc2 ());
DONE;
}
[(set_attr "type" "multi")
(set_attr "length" "8")])
(define_insn "umul64_600"
[(set (reg:DI MUL32x16_REG)
(mult:DI (zero_extend:DI (match_operand:SI 0 "register_operand"
"c,c,c"))
(zero_extract:DI (match_operand:SI 1 "nonmemory_operand"
"c,L,Cal")
(const_int 16)
(const_int 0))))
]
"TARGET_MULMAC_32BY16_SET"
"mululw 0, %0, %1"
[(set_attr "length" "4,4,8")
(set_attr "type" "mulmac_600")
(set_attr "predicable" "no")
(set_attr "cond" "nocond")])
(define_insn "umac64_600"
[(set (reg:DI MUL32x16_REG)
(plus:DI
(mult:DI (zero_extend:DI (match_operand:SI 1 "register_operand" "c,c,c"))
(ashift:DI
(zero_extract:DI (match_operand:SI 2 "nonmemory_operand" "c,L,Cal")
(const_int 16) (const_int 16))
(const_int 16)))
(reg:DI MUL32x16_REG)))
(set (match_operand:SI 0 "register_operand" "=w,w,w")
(zero_extract:SI
(plus:DI
(mult:DI (zero_extend:DI (match_dup 1))
(ashift:DI
(zero_extract:DI (match_dup 2)
(const_int 16) (const_int 16))
(const_int 16)))
(reg:DI MUL32x16_REG))
(const_int 32) (const_int 32)))]
"TARGET_MULMAC_32BY16_SET"
"machulw%? %0, %1, %2"
[(set_attr "length" "4,4,8")
(set_attr "type" "mulmac_600")
(set_attr "predicable" "no,no,yes")
(set_attr "cond" "nocond, canuse_limm, canuse")])
;; DI <- DI(unsigned SI) * DI(unsigned SI)
(define_insn_and_split "umulsidi3_700"
[(set (match_operand:DI 0 "dest_reg_operand" "=&r")
(mult:DI (zero_extend:DI (match_operand:SI 1 "register_operand" "%c"))
(zero_extend:DI (match_operand:SI 2 "extend_operand" "cL"))))]
"TARGET_MPY && !TARGET_PLUS_MACD"
"#"
"TARGET_MPY && !TARGET_PLUS_MACD && reload_completed"
[(const_int 0)]
{
int hi = !TARGET_BIG_ENDIAN;
int lo = !hi;
rtx l0 = operand_subword (operands[0], lo, 0, DImode);
rtx h0 = operand_subword (operands[0], hi, 0, DImode);
emit_insn (gen_umulsi3_highpart (h0, operands[1], operands[2]));
emit_insn (gen_mulsi3 (l0, operands[1], operands[2]));
DONE;
}
[(set_attr "type" "umulti")
(set_attr "length" "8")])
(define_expand "addsi3"
[(set (match_operand:SI 0 "dest_reg_operand" "")
(plus:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))]
""
"if (flag_pic && arc_raw_symbolic_reference_mentioned_p (operands[2], false))
{
operands[2]=force_reg(SImode, operands[2]);
}
")
(define_expand "adddi3"
[(set (match_operand:DI 0 "register_operand" "")
(plus:DI (match_operand:DI 1 "register_operand" "")
(match_operand:DI 2 "nonmemory_operand" "")))
(clobber (reg:CC CC_REG))]
""
"
rtx l0 = gen_lowpart (SImode, operands[0]);
rtx h0 = gen_highpart (SImode, operands[0]);
rtx l1 = gen_lowpart (SImode, operands[1]);
rtx h1 = gen_highpart (SImode, operands[1]);
rtx l2 = simplify_gen_subreg (SImode, operands[2], DImode,
subreg_lowpart_offset (SImode, DImode));
rtx h2 = simplify_gen_subreg (SImode, operands[2], DImode,
subreg_highpart_offset (SImode, DImode));
if (l2 == const0_rtx)
{
if (!rtx_equal_p (l0, l1) && !rtx_equal_p (l0, h1))
emit_move_insn (l0, l1);
emit_insn (gen_addsi3 (h0, h1, h2));
if (!rtx_equal_p (l0, l1) && rtx_equal_p (l0, h1))
emit_move_insn (l0, l1);
DONE;
}
if (rtx_equal_p (l0, h1))
{
if (h2 != const0_rtx)
emit_insn (gen_addsi3 (h0, h1, h2));
else if (!rtx_equal_p (h0, h1))
emit_move_insn (h0, h1);
emit_insn (gen_add_f (l0, l1, l2));
emit_insn
(gen_rtx_COND_EXEC
(VOIDmode,
gen_rtx_LTU (VOIDmode, gen_rtx_REG (CC_Cmode, CC_REG), GEN_INT (0)),
gen_rtx_SET (h0, plus_constant (SImode, h0, 1))));
DONE;
}
emit_insn (gen_add_f (l0, l1, l2));
emit_insn (gen_adc (h0, h1, h2));
DONE;
")
(define_insn "add_f"
[(set (reg:CC_C CC_REG)
(compare:CC_C
(plus:SI (match_operand:SI 1 "nonmemory_operand" "%r,L,0,I,Cal,r")
(match_operand:SI 2 "nonmemory_operand" "rL,r,I,0, r,rCal"))
(match_dup 1)))
(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r,r,r,r")
(plus:SI (match_dup 1) (match_dup 2)))]
"register_operand (operands[1], SImode)
|| register_operand (operands[2], SImode)"
"@
add.f\\t%0,%1,%2
add.f\\t%0,%2,%1
add.f\\t%0,%1,%2
add.f\\t%0,%2,%1
add.f\\t%0,%2,%1
add.f\\t%0,%1,%2"
[(set_attr "cond" "set")
(set_attr "type" "compare")
(set_attr "length" "4,4,4,4,8,8")])
(define_insn "*add_f_2"
[(set (reg:CC_C CC_REG)
(compare:CC_C
(plus:SI (match_operand:SI 1 "register_operand" "r ,0,r")
(match_operand:SI 2 "nonmemory_operand" "rL,I,rCal"))
(match_dup 2)))
(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"add.f\\t%0,%1,%2"
[(set_attr "cond" "set")
(set_attr "type" "compare")
(set_attr "length" "4,4,8")])
(define_insn "adc"
[(set (match_operand:SI 0 "register_operand" "=r, r,r,r, r,r")
(plus:SI
(plus:SI
(ltu:SI (reg:CC_C CC_REG) (const_int 0))
(match_operand:SI 1 "nonmemory_operand" "%r, 0,r,0,Cal,r"))
(match_operand:SI 2 "nonmemory_operand" "r,C_0,L,I, r,Cal")))]
"register_operand (operands[1], SImode)
|| register_operand (operands[2], SImode)"
"@
adc\\t%0,%1,%2
add.cs\\t%0,%1,1
adc\\t%0,%1,%2
adc\\t%0,%1,%2
adc\\t%0,%1,%2
adc\\t%0,%1,%2"
[(set_attr "cond" "use")
(set_attr "type" "cc_arith")
(set_attr "length" "4,4,4,4,8,8")])
; combiner-splitter cmp / scc -> cmp / adc
(define_split
[(set (match_operand:SI 0 "dest_reg_operand" "")
(gtu:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "register_operand" "")))
(clobber (reg CC_REG))]
""
[(set (reg:CC_C CC_REG) (compare:CC_C (match_dup 2) (match_dup 1)))
(set (match_dup 0) (ltu:SI (reg:CC_C CC_REG) (const_int 0)))])
; combine won't work when an intermediate result is used later...
; add %0,%1,%2 ` cmp %0,%[12] -> add.f %0,%1,%2
(define_peephole2
[(set (match_operand:SI 0 "dest_reg_operand" "")
(plus:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))
(set (reg:CC_C CC_REG)
(compare:CC_C (match_dup 0)
(match_operand:SI 3 "nonmemory_operand" "")))]
"rtx_equal_p (operands[1], operands[3])
|| rtx_equal_p (operands[2], operands[3])"
[(parallel
[(set (reg:CC_C CC_REG)
(compare:CC_C (plus:SI (match_dup 1) (match_dup 2)) (match_dup 1)))
(set (match_dup 0)
(plus:SI (match_dup 1) (match_dup 2)))])])
; ??? need to delve into combine to find out why this is not useful.
; We'd like to be able to grok various C idioms for carry bit usage.
;(define_insn "*adc_0"
; [(set (match_operand:SI 0 "dest_reg_operand" "=w")
; (plus:SI (ltu:SI (reg:CC_C CC_REG) (const_int 0))
; (match_operand:SI 1 "register_operand" "c")))]
; ""
; "adc %0,%1,0"
; [(set_attr "cond" "use")
; (set_attr "type" "cc_arith")
; (set_attr "length" "4")])
;
;(define_split
; [(set (match_operand:SI 0 "dest_reg_operand" "=w")
; (plus:SI (gtu:SI (match_operand:SI 1 "register_operand" "c")
; (match_operand:SI 2 "register_operand" "c"))
; (match_operand:SI 3 "register_operand" "c")))
; (clobber (reg CC_REG))]
; ""
; [(set (reg:CC_C CC_REG) (compare:CC_C (match_dup 2) (match_dup 1)))
; (set (match_dup 0)
; (plus:SI (ltu:SI (reg:CC_C CC_REG) (const_int 0))
; (match_dup 3)))])
(define_expand "subsi3"
[(set (match_operand:SI 0 "dest_reg_operand" "")
(minus:SI (match_operand:SI 1 "nonmemory_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))]
""
"
{
int c = 1;
if (!register_operand (operands[2], SImode))
{
operands[1] = force_reg (SImode, operands[1]);
c = 2;
}
if (flag_pic && arc_raw_symbolic_reference_mentioned_p (operands[c], false))
operands[c] = force_reg (SImode, operands[c]);
}")
; the casesi expander might generate a sub of zero, so we have to recognize it.
; combine should make such an insn go away.
(define_insn_and_split "subsi3_insn"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,q,r, r,r,r,r, r, r, r")
(minus:SI (match_operand:SI 1 "nonmemory_operand" "0,q,0,rL,r,L,I,Cal,Cal, r")
(match_operand:SI 2 "nonmemory_operand" "q,q,r, 0,r,r,0, 0, r,Cal")))]
"register_operand (operands[1], SImode)
|| register_operand (operands[2], SImode)"
"@
sub%?\\t%0,%1,%2%&
sub%?\\t%0,%1,%2%&
sub%?\\t%0,%1,%2
rsub%?\\t%0,%2,%1
sub\\t%0,%1,%2
rsub\\t%0,%2,%1
rsub\\t%0,%2,%1
rsub%?\\t%0,%2,%1
rsub\\t%0,%2,%1
sub\\t%0,%1,%2"
"reload_completed && get_attr_length (insn) == 8
&& satisfies_constraint_I (operands[1])
&& GET_CODE (PATTERN (insn)) != COND_EXEC"
[(set (match_dup 0) (match_dup 3)) (set (match_dup 0) (match_dup 4))]
"split_subsi (operands);"
[(set_attr "iscompact" "maybe,maybe,false,false,false,false,false,false,false, false")
(set_attr "length" "*,*,4,4,4,4,4,8,8,8")
(set_attr "predicable" "yes,no,yes,yes,no,no,no,yes,no,no")
(set_attr "cond" "canuse,nocond,canuse,canuse,nocond,nocond,canuse_limm,canuse,nocond,nocond")
(set_attr "cpu_facility" "*,cd,*,*,*,*,*,*,*,*")
])
(define_expand "subdi3"
[(set (match_operand:DI 0 "register_operand" "")
(minus:DI (match_operand:DI 1 "register_operand" "")
(match_operand:DI 2 "nonmemory_operand" "")))
(clobber (reg:CC CC_REG))]
""
"
rtx l0 = gen_lowpart (SImode, operands[0]);
rtx h0 = gen_highpart (SImode, operands[0]);
rtx l1 = gen_lowpart (SImode, operands[1]);
rtx h1 = gen_highpart (SImode, operands[1]);
rtx l2 = simplify_gen_subreg (SImode, operands[2], DImode,
subreg_lowpart_offset (SImode, DImode));
rtx h2 = simplify_gen_subreg (SImode, operands[2], DImode,
subreg_highpart_offset (SImode, DImode));
if (rtx_equal_p (l0, h1) || rtx_equal_p (l0, h2))
{
h1 = simplify_gen_binary (MINUS, SImode, h1, h2);
if (!rtx_equal_p (h0, h1))
emit_insn (gen_rtx_SET (h0, h1));
emit_insn (gen_sub_f (l0, l1, l2));
emit_insn
(gen_rtx_COND_EXEC
(VOIDmode,
gen_rtx_LTU (VOIDmode, gen_rtx_REG (CC_Cmode, CC_REG), GEN_INT (0)),
gen_rtx_SET (h0, plus_constant (SImode, h0, -1))));
DONE;
}
emit_insn (gen_sub_f (l0, l1, l2));
emit_insn (gen_sbc (h0, h1, h2));
DONE;
")
(define_insn "*sbc_0"
[(set (match_operand:SI 0 "dest_reg_operand" "=w")
(minus:SI (match_operand:SI 1 "register_operand" "c")
(ltu:SI (match_operand:CC_C 2 "cc_use_register")
(const_int 0))))]
""
"sbc %0,%1,0"
[(set_attr "cond" "use")
(set_attr "type" "cc_arith")
(set_attr "length" "4")])
(define_insn "sbc"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r,r,r,r")
(minus:SI
(minus:SI
(match_operand:SI 1 "nonmemory_operand" "r, 0,r,0, r,Cal")
(ltu:SI (reg:CC_C CC_REG) (const_int 0)))
(match_operand:SI 2 "nonmemory_operand" "r,C_0,L,I,Cal,r")))]
"register_operand (operands[1], SImode)
|| register_operand (operands[2], SImode)"
"@
sbc\\t%0,%1,%2
sub.cs\\t%0,%1,1
sbc\\t%0,%1,%2
sbc\\t%0,%1,%2
sbc\\t%0,%1,%2
sbc\\t%0,%1,%2"
[(set_attr "cond" "use")
(set_attr "type" "cc_arith")
(set_attr "length" "4,4,4,4,8,8")])
(define_insn "sub_f"
[(set (reg:CC CC_REG)
(compare:CC (match_operand:SI 1 "nonmemory_operand" " r,L,0,I,r,Cal")
(match_operand:SI 2 "nonmemory_operand" "rL,r,I,0,Cal,r")))
(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r,r,r,r")
(minus:SI (match_dup 1) (match_dup 2)))]
"register_operand (operands[1], SImode)
|| register_operand (operands[2], SImode)"
"@
sub.f\\t%0,%1,%2
rsub.f\\t%0,%2,%1
sub.f\\t%0,%1,%2
rsub.f\\t%0,%2,%1
sub.f\\t%0,%1,%2
sub.f\\t%0,%1,%2"
[(set_attr "type" "compare")
(set_attr "length" "4,4,4,4,8,8")])
; combine won't work when an intermediate result is used later...
; add %0,%1,%2 ` cmp %0,%[12] -> add.f %0,%1,%2
(define_peephole2
[(set (reg:CC CC_REG)
(compare:CC (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))
(set (match_operand:SI 0 "dest_reg_operand" "")
(minus:SI (match_dup 1) (match_dup 2)))]
""
[(parallel
[(set (reg:CC CC_REG) (compare:CC (match_dup 1) (match_dup 2)))
(set (match_dup 0) (minus:SI (match_dup 1) (match_dup 2)))])])
(define_peephole2
[(set (reg:CC CC_REG)
(compare:CC (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))
(set (match_operand 3 "" "") (match_operand 4 "" ""))
(set (match_operand:SI 0 "dest_reg_operand" "")
(minus:SI (match_dup 1) (match_dup 2)))]
"!reg_overlap_mentioned_p (operands[3], operands[1])
&& !reg_overlap_mentioned_p (operands[3], operands[2])
&& !reg_overlap_mentioned_p (operands[0], operands[4])
&& !reg_overlap_mentioned_p (operands[0], operands[3])"
[(parallel
[(set (reg:CC CC_REG) (compare:CC (match_dup 1) (match_dup 2)))
(set (match_dup 0) (minus:SI (match_dup 1) (match_dup 2)))])
(set (match_dup 3) (match_dup 4))])
(define_insn "*add_n"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,r,r")
(plus:SI (mult:SI (match_operand:SI 1 "register_operand" "q,r,r")
(match_operand:SI 2 "_2_4_8_operand" ""))
(match_operand:SI 3 "arc_nonmemory_operand" "0,r,Csz")))]
""
"add%z2%?\\t%0,%3,%1"
[(set_attr "type" "shift")
(set_attr "length" "*,4,8")
(set_attr "predicable" "yes,no,no")
(set_attr "cond" "canuse,nocond,nocond")
(set_attr "iscompact" "maybe,false,false")])
;; N.B. sub[123] has the operands of the MINUS in the opposite order from
;; what synth_mult likes.
(define_insn "*sub_n"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r")
(minus:SI (match_operand:SI 1 "nonmemory_operand" "0,r,?Cal")
(ashift:SI (match_operand:SI 2 "register_operand" "r,r,r")
(match_operand:SI 3 "_1_2_3_operand" ""))))]
""
"sub%c3%?\\t%0,%1,%2"
[(set_attr "type" "shift")
(set_attr "length" "4,4,8")
(set_attr "predicable" "yes,no,no")
(set_attr "cond" "canuse,nocond,nocond")
(set_attr "iscompact" "false")])
(define_insn "*sub_n"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r")
(minus:SI (match_operand:SI 1 "nonmemory_operand" "0,r,?Cal")
(mult:SI (match_operand:SI 2 "register_operand" "r,r,r")
(match_operand:SI 3 "_2_4_8_operand" ""))))]
""
"sub%z3%?\\t%0,%1,%2"
[(set_attr "type" "shift")
(set_attr "length" "4,4,8")
(set_attr "predicable" "yes,no,no")
(set_attr "cond" "canuse,nocond,nocond")
(set_attr "iscompact" "false")])
; ??? check if combine matches this.
(define_insn "*bset"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r")
(ior:SI (ashift:SI (const_int 1)
(match_operand:SI 1 "nonmemory_operand" "rL,rL,r"))
(match_operand:SI 2 "nonmemory_operand" "0,r,Cal")))]
""
"bset%?\\t%0,%2,%1"
[(set_attr "length" "4,4,8")
(set_attr "predicable" "yes,no,no")
(set_attr "cond" "canuse,nocond,nocond")]
)
; ??? check if combine matches this.
(define_insn "*bxor"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r")
(xor:SI (ashift:SI (const_int 1)
(match_operand:SI 1 "nonmemory_operand" "rL,rL,r"))
(match_operand:SI 2 "nonmemory_operand" "0,r,Cal")))]
""
"bxor%?\\t%0,%2,%1"
[(set_attr "length" "4,4,8")
(set_attr "predicable" "yes,no,no")
(set_attr "cond" "canuse,nocond,nocond")]
)
; ??? check if combine matches this.
(define_insn "*bclr"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r")
(and:SI (not:SI (ashift:SI (const_int 1)
(match_operand:SI 1 "nonmemory_operand" "rL,rL,r")))
(match_operand:SI 2 "nonmemory_operand" "0,r,Cal")))]
""
"bclr%?\\t%0,%2,%1"
[(set_attr "length" "4,4,8")
(set_attr "predicable" "yes,no,no")
(set_attr "cond" "canuse,nocond,nocond")]
)
; ??? FIXME: find combine patterns for bmsk.
;;Following are the define_insns added for the purpose of peephole2's
; see also iorsi3 for use with constant bit number.
(define_insn "*bset_insn"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r")
(ior:SI (match_operand:SI 1 "nonmemory_operand" "0,r,Cal")
(ashift:SI (const_int 1)
(match_operand:SI 2 "nonmemory_operand" "rL,rL,r"))) ) ]
""
"@
bset%?\\t%0,%1,%2 ;;peep2, constr 1
bset\\t%0,%1,%2 ;;peep2, constr 2
bset\\t%0,%1,%2 ;;peep2, constr 3"
[(set_attr "length" "4,4,8")
(set_attr "predicable" "yes,no,no")
(set_attr "cond" "canuse,nocond,nocond")]
)
; see also xorsi3 for use with constant bit number.
(define_insn "*bxor_insn"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r")
(xor:SI (match_operand:SI 1 "nonmemory_operand" "0,r,Cal")
(ashift:SI (const_int 1)
(match_operand:SI 2 "nonmemory_operand" "rL,rL,r"))) ) ]
""
"@
bxor%?\\t%0,%1,%2
bxor\\t%0,%1,%2
bxor\\t%0,%1,%2"
[(set_attr "length" "4,4,8")
(set_attr "predicable" "yes,no,no")
(set_attr "cond" "canuse,nocond,nocond")]
)
; see also andsi3 for use with constant bit number.
(define_insn "*bclr_insn"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r")
(and:SI (not:SI (ashift:SI (const_int 1)
(match_operand:SI 2 "nonmemory_operand" "rL,rL,r")))
(match_operand:SI 1 "nonmemory_operand" "0,r,Cal")))]
""
"@
bclr%?\\t%0,%1,%2
bclr\\t%0,%1,%2
bclr\\t%0,%1,%2"
[(set_attr "length" "4,4,8")
(set_attr "predicable" "yes,no,no")
(set_attr "cond" "canuse,nocond,nocond")]
)
; see also andsi3 for use with constant bit number.
(define_insn "*bmsk_insn"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r")
(and:SI (match_operand:SI 1 "nonmemory_operand" "0,r,Cal")
(plus:SI (ashift:SI (const_int 1)
(plus:SI (match_operand:SI 2 "nonmemory_operand" "rL,rL,r")
(const_int 1)))
(const_int -1))))]
""
"@
bmsk%?\\t%0,%1,%2
bmsk\\t%0,%1,%2
bmsk\\t%0,%1,%2"
[(set_attr "length" "4,4,8")
(set_attr "predicable" "yes,no,no")
(set_attr "cond" "canuse,nocond,nocond")]
)
;;Instructions added for peephole2s end
;; Boolean instructions.
(define_expand "andsi3"
[(set (match_operand:SI 0 "dest_reg_operand" "")
(and:SI (match_operand:SI 1 "nonimmediate_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))]
""
"if (!satisfies_constraint_Cux (operands[2]))
operands[1] = force_reg (SImode, operands[1]);
")
(define_insn "andsi3_i" ;0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
[(set (match_operand:SI 0 "dest_reg_operand" "=q,q, q, q, q, r,r, r, r, r,r, r, r, r, r, q,r, r, r, W")
(and:SI (match_operand:SI 1 "nonimmediate_operand" "%0,q, 0, 0, q, 0,r, 0, 0, 0,0, r, r, r, r, q,0, 0, r, o")
(match_operand:SI 2 "nonmemory_operand" "q,0,C1p,Ccp,Cux,rL,0,C2pC1p,Ccp,CnL,I,rL,C2pC1p,Ccp,CnL,Cbf,I,Cal,Cal,Cux")))]
"(register_operand (operands[1], SImode)
&& nonmemory_operand (operands[2], SImode))
|| (memory_operand (operands[1], SImode)
&& satisfies_constraint_Cux (operands[2]))"
{
switch (which_alternative)
{
case 0: case 5: case 10: case 11: case 16: case 17: case 18:
return "and%? %0,%1,%2%&";
case 1: case 6:
return "and%? %0,%2,%1%&";
case 2:
return "bmsk%? %0,%1,%Z2%&";
case 7: case 12:
if (satisfies_constraint_C2p (operands[2]))
{
operands[2] = GEN_INT ((~INTVAL (operands[2])));
return "bmskn%? %0,%1,%Z2%&";
}
else
{
return "bmsk%? %0,%1,%Z2%&";
}
case 3: case 8: case 13:
return "bclr%? %0,%1,%M2%&";
case 4:
return (INTVAL (operands[2]) == 0xff
? "extb%? %0,%1%&" : "ext%_%? %0,%1%&");
case 9: case 14: return \"bic%? %0,%1,%n2-1\";
case 15:
return "movb.cl %0,%1,%p2,%p2,%s2";
case 19:
const char *tmpl;
if (satisfies_constraint_Ucm (operands[1]))
tmpl = (INTVAL (operands[2]) == 0xff
? "xldb%U1 %0,%1" : "xld%_%U1 %0,%1");
else
tmpl = INTVAL (operands[2]) == 0xff ? "ldb %0,%1" : "ld%_ %0,%1";
if (TARGET_BIG_ENDIAN)
{
rtx xop[2];
xop[0] = operands[0];
xop[1] = adjust_address (operands[1], QImode,
INTVAL (operands[2]) == 0xff ? 3 : 2);
output_asm_insn (tmpl, xop);
return "";
}
return tmpl;
default:
gcc_unreachable ();
}
}
[(set_attr "iscompact" "maybe,maybe,maybe,maybe,true,false,false,false,false,false,false,false,false,false,false,false,false,false,false,false")
(set_attr "type" "binary,binary,binary,binary,binary,binary,binary,binary,binary,binary,binary,binary,binary,binary,binary,shift,binary,binary,binary,load")
(set_attr "length" "*,*,*,*,*,4,4,4,4,4,4,4,4,4,4,4,4,8,8,*")
(set_attr "predicable" "no,no,no,no,no,yes,yes,yes,yes,yes,no,no,no,no,no,no,no,yes,no,no")
(set_attr "cond" "canuse,canuse,canuse,canuse,nocond,canuse,canuse,canuse,canuse,canuse,canuse_limm,nocond,nocond,nocond,nocond,nocond,canuse_limm,canuse,nocond,nocond")])
; combiner splitter, pattern found in ldtoa.c .
; and op3,op0,op1 / cmp op3,op2 -> add op3,op0,op4 / bmsk.f 0,op3,op1
(define_split
[(set (reg:CC_Z CC_REG)
(compare:CC_Z (and:SI (match_operand:SI 0 "register_operand" "")
(match_operand 1 "const_int_operand" ""))
(match_operand 2 "const_int_operand" "")))
(clobber (match_operand:SI 3 "register_operand" ""))]
"((INTVAL (operands[1]) + 1) & INTVAL (operands[1])) == 0"
[(set (match_dup 3)
(plus:SI (match_dup 0) (match_dup 4)))
(set (reg:CC_Z CC_REG)
(compare:CC_Z (and:SI (match_dup 3) (match_dup 1))
(const_int 0)))]
"operands[4] = GEN_INT ( -(~INTVAL (operands[1]) | INTVAL (operands[2])));")
;;bic define_insn that allows limm to be the first operand
(define_insn "*bicsi3_insn"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,r,r,r,r,r,r")
(and:SI (not:SI (match_operand:SI 1 "nonmemory_operand" "q,Lr,I,Cal,Lr,Cal,r"))
(match_operand:SI 2 "nonmemory_operand" "0,0,0,0,r,r,Cal")))]
""
"@
bic%?\\t%0, %2, %1%& ;;constraint 0
bic%?\\t%0,%2,%1 ;;constraint 1
bic\\t%0,%2,%1 ;;constraint 2, FIXME: will it ever get generated ???
bic%?\\t%0,%2,%1 ;;constraint 3, FIXME: will it ever get generated ???
bic\\t%0,%2,%1 ;;constraint 4
bic\\t%0,%2,%1 ;;constraint 5, FIXME: will it ever get generated ???
bic\\t%0,%2,%1 ;;constraint 6"
[(set_attr "length" "*,4,4,8,4,8,8")
(set_attr "iscompact" "maybe, false, false, false, false, false, false")
(set_attr "predicable" "no,yes,no,yes,no,no,no")
(set_attr "cond" "canuse,canuse,canuse_limm,canuse,nocond,nocond,nocond")])
(define_insn_and_split "iorsi3"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,q, q, r,r, r,r, r, r,r, q, r, r")
(ior:SI (match_operand:SI 1 "register_operand" "%0,q, 0, 0,r, 0,0, r, r,0, r, 0, r")
(match_operand:SI 2 "nonmemory_operand" "q,0,C0p,rL,0,C0p,I,rL,C0p,I,C0x,Cal,Cal")))]
""
"@
or%?\\t%0,%1,%2
or%?\\t%0,%2,%1
bset%?\\t%0,%1,%z2
or%?\\t%0,%1,%2
or%?\\t%0,%2,%1
bset%?\\t%0,%1,%z2
or%?\\t%0,%1,%2
or%?\\t%0,%1,%2
bset%?\\t%0,%1,%z2
or%?\\t%0,%1,%2
#
or%?\\t%0,%1,%2
or%?\\t%0,%1,%2"
"reload_completed
&& GET_CODE (PATTERN (insn)) != COND_EXEC
&& register_operand (operands[0], SImode)
&& IN_RANGE (REGNO (operands[0]) ^ 4, 4, 11)
&& satisfies_constraint_C0x (operands[2])"
[(const_int 0)]
"
arc_split_ior (operands);
DONE;
"
[(set_attr "iscompact" "maybe,maybe,maybe,false,false,false,false,false,false,false,false,false,false")
(set_attr "length" "*,*,*,4,4,4,4,4,4,4,*,8,8")
(set_attr "predicable" "no,no,no,yes,yes,yes,no,no,no,no,no,yes,no")
(set_attr "cond" "canuse,canuse,canuse,canuse,canuse,canuse,canuse_limm,nocond,nocond,canuse_limm,nocond,canuse,nocond")])
(define_insn "xorsi3"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,q, r,r, r,r, r, r,r, r, r")
(xor:SI (match_operand:SI 1 "register_operand" "%0,q, 0,r, 0,0, r, r,0, 0, r")
(match_operand:SI 2 "nonmemory_operand" "q,0,rL,0,C0p,I,rL,C0p,I,Cal,Cal")))]
""
"*
switch (which_alternative)
{
case 0: case 2: case 5: case 6: case 8: case 9: case 10:
return \"xor%?\\t%0,%1,%2%&\";
case 1: case 3:
return \"xor%?\\t%0,%2,%1%&\";
case 4: case 7:
return \"bxor%?\\t%0,%1,%z2\";
default:
gcc_unreachable ();
}
"
[(set_attr "iscompact" "maybe,maybe,false,false,false,false,false,false,false,false,false")
(set_attr "type" "binary")
(set_attr "length" "*,*,4,4,4,4,4,4,4,8,8")
(set_attr "predicable" "no,no,yes,yes,yes,no,no,no,no,yes,no")
(set_attr "cond" "canuse,canuse,canuse,canuse,canuse,canuse_limm,nocond,nocond,canuse_limm,canuse,nocond")])
(define_insn "negsi2"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,q,r,r")
(neg:SI (match_operand:SI 1 "register_operand" "0,q,0,r")))]
""
"neg%?\\t%0,%1%&"
[(set_attr "type" "unary")
(set_attr "iscompact" "maybe,true,false,false")
(set_attr "predicable" "no,no,yes,no")])
(define_insn "one_cmplsi2"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,w")
(not:SI (match_operand:SI 1 "register_operand" "q,c")))]
""
"not%? %0,%1%&"
[(set_attr "type" "unary,unary")
(set_attr "iscompact" "true,false")])
(define_insn_and_split "one_cmpldi2"
[(set (match_operand:DI 0 "dest_reg_operand" "=q,w")
(not:DI (match_operand:DI 1 "register_operand" "q,c")))]
""
"#"
"&& reload_completed"
[(set (match_dup 2) (not:SI (match_dup 3)))
(set (match_dup 4) (not:SI (match_dup 5)))]
{
int swap = (true_regnum (operands[0]) == true_regnum (operands[1]) + 1);
operands[2] = operand_subword (operands[0], 0+swap, 0, DImode);
operands[3] = operand_subword (operands[1], 0+swap, 0, DImode);
operands[4] = operand_subword (operands[0], 1-swap, 0, DImode);
operands[5] = operand_subword (operands[1], 1-swap, 0, DImode);
}
[(set_attr "type" "unary,unary")
(set_attr "cond" "nocond,nocond")
(set_attr "length" "4,8")])
;; Shift instructions.
(define_expand "ashlsi3"
[(set (match_operand:SI 0 "dest_reg_operand" "")
(ashift:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))]
""
"
{
if (!TARGET_BARREL_SHIFTER)
{
emit_shift (ASHIFT, operands[0], operands[1], operands[2]);
DONE;
}
}")
(define_expand "ashrsi3"
[(set (match_operand:SI 0 "dest_reg_operand" "")
(ashiftrt:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))]
""
"
{
if (!TARGET_BARREL_SHIFTER)
{
emit_shift (ASHIFTRT, operands[0], operands[1], operands[2]);
DONE;
}
}")
(define_expand "lshrsi3"
[(set (match_operand:SI 0 "dest_reg_operand" "")
(lshiftrt:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))]
""
"
{
if (!TARGET_BARREL_SHIFTER)
{
emit_shift (LSHIFTRT, operands[0], operands[1], operands[2]);
DONE;
}
}")
(define_insn "shift_si3"
[(set (match_operand:SI 0 "dest_reg_operand" "=r")
(match_operator:SI 3 "shift4_operator"
[(match_operand:SI 1 "register_operand" "0")
(match_operand:SI 2 "const_int_operand" "n")]))
(clobber (match_scratch:SI 4 "=&r"))
(clobber (reg:CC CC_REG))
]
"!TARGET_BARREL_SHIFTER"
"* return output_shift (operands);"
[(set_attr "type" "shift")
(set_attr "length" "16")])
(define_insn "shift_si3_loop"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r")
(match_operator:SI 3 "shift_operator"
[(match_operand:SI 1 "register_operand" "0,0")
(match_operand:SI 2 "nonmemory_operand" "rn,Cal")]))
(clobber (match_scratch:SI 4 "=X,X"))
(clobber (reg:SI LP_COUNT))
(clobber (reg:CC CC_REG))
]
"!TARGET_BARREL_SHIFTER"
"* return output_shift (operands);"
[(set_attr "type" "shift")
(set_attr "length" "16,20")])
; asl, asr, lsr patterns:
; There is no point in including an 'I' alternative since only the lowest 5
; bits are used for the shift. OTOH Cal can be useful if the shift amount
; is defined in an external symbol, as we don't have special relocations
; to truncate a symbol in a u6 immediate; but that's rather exotic, so only
; provide one alternatice for this, without condexec support.
(define_insn "*ashlsi3_insn"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,q, q, r, r, r")
(ashift:SI (match_operand:SI 1 "arc_nonmemory_operand" "!0,q, 0, 0, r,rCsz")
(match_operand:SI 2 "nonmemory_operand" "K,K,qM,rL,rL,rCal")))]
"TARGET_BARREL_SHIFTER
&& (register_operand (operands[1], SImode)
|| register_operand (operands[2], SImode))"
"asl%?\\t%0,%1,%2"
[(set_attr "type" "shift")
(set_attr "iscompact" "maybe,maybe,maybe,false,false,false")
(set_attr "predicable" "no,no,no,yes,no,no")
(set_attr "cond" "canuse,nocond,canuse,canuse,nocond,nocond")])
(define_insn "*ashrsi3_insn"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,q, q, r, r, r")
(ashiftrt:SI (match_operand:SI 1 "arc_nonmemory_operand" "!0,q, 0, 0, r,rCsz")
(match_operand:SI 2 "nonmemory_operand" "K,K,qM,rL,rL,rCal")))]
"TARGET_BARREL_SHIFTER
&& (register_operand (operands[1], SImode)
|| register_operand (operands[2], SImode))"
"asr%?\\t%0,%1,%2"
[(set_attr "type" "shift")
(set_attr "iscompact" "maybe,maybe,maybe,false,false,false")
(set_attr "predicable" "no,no,no,yes,no,no")
(set_attr "cond" "canuse,nocond,canuse,canuse,nocond,nocond")])
(define_insn "*lshrsi3_insn"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,q, q, r, r, r")
(lshiftrt:SI (match_operand:SI 1 "nonmemory_operand" "!0,q, 0, 0, r,rCal")
(match_operand:SI 2 "nonmemory_operand" "N,N,qM,rL,rL,rCal")))]
"TARGET_BARREL_SHIFTER
&& (register_operand (operands[1], SImode)
|| register_operand (operands[2], SImode))"
"*return (which_alternative <= 1 && !arc_ccfsm_cond_exec_p ()
? \"lsr%?\\t%0,%1%&\" : \"lsr%?\\t%0,%1,%2%&\");"
[(set_attr "type" "shift")
(set_attr "iscompact" "maybe,maybe,maybe,false,false,false")
(set_attr "predicable" "no,no,no,yes,no,no")
(set_attr "cond" "canuse,nocond,canuse,canuse,nocond,nocond")])
(define_insn "rotrsi3"
[(set (match_operand:SI 0 "dest_reg_operand" "=r, r, r")
(rotatert:SI (match_operand:SI 1 "arc_nonmemory_operand" " 0,rL,rCsz")
(match_operand:SI 2 "nonmemory_operand" "rL,rL,rCal")))]
"TARGET_BARREL_SHIFTER"
"ror%?\\t%0,%1,%2"
[(set_attr "type" "shift,shift,shift")
(set_attr "predicable" "yes,no,no")
(set_attr "length" "4,4,8")])
;; Compare / branch instructions.
(define_expand "cbranchsi4"
[(set (reg:CC CC_REG)
(compare:CC (match_operand:SI 1 "nonmemory_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))
(set (pc)
(if_then_else
(match_operator 0 "ordered_comparison_operator" [(reg CC_REG)
(const_int 0)])
(label_ref (match_operand 3 "" ""))
(pc)))]
""
{
gcc_assert (XEXP (operands[0], 0) == operands[1]);
gcc_assert (XEXP (operands[0], 1) == operands[2]);
operands[0] = gen_compare_reg (operands[0], VOIDmode);
emit_jump_insn (gen_branch_insn (operands[3], operands[0]));
DONE;
})
;; ??? Could add a peephole to generate compare with swapped operands and
;; modifed cc user if second, but not first operand is a compact register.
(define_insn "cmpsi_cc_insn_mixed"
[(set (reg:CC CC_REG)
(compare:CC (match_operand:SI 0 "register_operand" "q, q, h, c, c, q,c")
(match_operand:SI 1 "nonmemory_operand" "cO,hO,Cm1,cI,cL,Cal,Cal")))]
""
"cmp%? %0,%B1%&"
[(set_attr "type" "compare")
(set_attr "iscompact" "true,true,true,false,false,true_limm,false")
(set_attr "predicable" "no,no,no,no,yes,no,yes")
(set_attr "cond" "set")
(set_attr "length" "*,*,*,4,4,*,8")
(set_attr "cpu_facility" "av1,av2,*,*,*,*,*")])
(define_insn "*cmpsi_cc_zn_insn"
[(set (reg:CC_ZN CC_REG)
(compare:CC_ZN (match_operand:SI 0 "register_operand" "q,c")
(const_int 0)))]
""
"tst%? %0,%0%&"
[(set_attr "type" "compare,compare")
(set_attr "iscompact" "true,false")
(set_attr "predicable" "no,yes")
(set_attr "cond" "set_zn")
(set_attr "length" "*,4")])
; combiner pattern observed for unwind-dw2-fde.c:linear_search_fdes.
(define_insn "*btst"
[(set (reg:CC_ZN CC_REG)
(compare:CC_ZN
(zero_extract:SI (match_operand:SI 0 "register_operand" "q,c")
(const_int 1)
(match_operand:SI 1 "nonmemory_operand" "L,Lc"))
(const_int 0)))]
""
"btst%? %0,%1"
[(set_attr "iscompact" "true,false")
(set_attr "predicable" "no,yes")
(set_attr "cond" "set")
(set_attr "type" "compare")
(set_attr "length" "*,4")])
; combine suffers from 'simplifications' that replace a one-bit zero
; extract with a shift if it can prove that the upper bits are zero.
; arc_reorg sees the code after sched2, which can have caused our
; inputs to be clobbered even if they were not clobbered before.
; Therefore, add a third way to convert btst / b{eq,ne} to bbit{0,1}
; OTOH, this is somewhat marginal, and can leat to out-of-range
; bbit (i.e. bad scheduling) and missed conditional execution,
; so make this an option.
(define_peephole2
[(set (reg:CC_ZN CC_REG)
(compare:CC_ZN
(zero_extract:SI (match_operand:SI 0 "register_operand" "")
(const_int 1)
(match_operand:SI 1 "nonmemory_operand" ""))
(const_int 0)))
(set (pc)
(if_then_else (match_operator 3 "equality_comparison_operator"
[(reg:CC_ZN CC_REG) (const_int 0)])
(label_ref (match_operand 2 "" ""))
(pc)))]
"TARGET_BBIT_PEEPHOLE && peep2_regno_dead_p (2, CC_REG)"
[(parallel [(set (pc)
(if_then_else
(match_op_dup 3
[(zero_extract:SI (match_dup 0)
(const_int 1) (match_dup 1))
(const_int 0)])
(label_ref (match_dup 2))
(pc)))
(clobber (reg:CC_ZN CC_REG))])])
(define_insn "*cmpsi_cc_z_insn"
[(set (reg:CC_Z CC_REG)
(compare:CC_Z (match_operand:SI 0 "register_operand" "q,c")
(match_operand:SI 1 "p2_immediate_operand" "O,n")))]
""
"@
cmp%? %0,%1%&
bxor.f 0,%0,%z1"
[(set_attr "type" "compare,compare")
(set_attr "iscompact" "true,false")
(set_attr "cond" "set,set_zn")
(set_attr "length" "*,4")])
(define_insn "*cmpsi_cc_c_insn"
[(set (reg:CC_C CC_REG)
(compare:CC_C (match_operand:SI 0 "register_operand" "q, q, h, c, q, c")
(match_operand:SI 1 "nonmemory_operand" "cO,hO,Cm1,cI,Cal,Cal")))]
""
"cmp%? %0,%1%&"
[(set_attr "type" "compare")
(set_attr "iscompact" "true,true,true,false,true_limm,false")
(set_attr "cond" "set")
(set_attr "length" "*,*,*,4,*,8")
(set_attr "cpu_facility" "av1,av2,*,*,*,*")])
;; Next come the scc insns.
(define_expand "cstoresi4"
[(set (match_operand:SI 0 "dest_reg_operand" "")
(match_operator:SI 1 "ordered_comparison_operator"
[(match_operand:SI 2 "nonmemory_operand" "")
(match_operand:SI 3 "nonmemory_operand" "")]))]
""
{
if (!TARGET_CODE_DENSITY)
{
gcc_assert (XEXP (operands[1], 0) == operands[2]);
gcc_assert (XEXP (operands[1], 1) == operands[3]);
operands[1] = gen_compare_reg (operands[1], SImode);
emit_insn (gen_scc_insn (operands[0], operands[1]));
DONE;
}
if (!register_operand (operands[2], SImode))
operands[2] = force_reg (SImode, operands[2]);
})
(define_mode_iterator SDF [(SF "TARGET_FP_SP_BASE || TARGET_OPTFPE")
(DF "TARGET_FP_DP_BASE || TARGET_OPTFPE")])
(define_expand "cstore4"
[(set (reg:CC CC_REG)
(compare:CC (match_operand:SDF 2 "register_operand" "")
(match_operand:SDF 3 "register_operand" "")))
(set (match_operand:SI 0 "dest_reg_operand" "")
(match_operator:SI 1 "comparison_operator" [(reg CC_REG)
(const_int 0)]))]
"TARGET_HARD_FLOAT || TARGET_OPTFPE"
{
gcc_assert (XEXP (operands[1], 0) == operands[2]);
gcc_assert (XEXP (operands[1], 1) == operands[3]);
operands[1] = gen_compare_reg (operands[1], SImode);
emit_insn (gen_scc_insn (operands[0], operands[1]));
DONE;
})
; We need a separate expander for this lest we loose the mode of CC_REG
; when match_operator substitutes the literal operand into the comparison.
(define_expand "scc_insn"
[(set (match_operand:SI 0 "dest_reg_operand" "=w") (match_operand:SI 1 ""))])
(define_insn_and_split "*scc_insn"
[(set (match_operand:SI 0 "dest_reg_operand" "=w")
(match_operator:SI 1 "proper_comparison_operator" [(reg CC_REG) (const_int 0)]))]
""
"#"
"reload_completed"
[(set (match_dup 0) (const_int 1))
(cond_exec
(match_dup 1)
(set (match_dup 0) (const_int 0)))]
{
operands[1]
= gen_rtx_fmt_ee (REVERSE_CONDITION (GET_CODE (operands[1]),
GET_MODE (XEXP (operands[1], 0))),
VOIDmode,
XEXP (operands[1], 0), XEXP (operands[1], 1));
}
[(set_attr "type" "unary")])
; cond_exec patterns
(define_insn "*movsi_ne"
[(cond_exec
(ne (match_operand:CC_Z 2 "cc_use_register" "Rcc,Rcc,Rcc,Rcc,Rcc") (const_int 0))
(set (match_operand:SI 0 "dest_reg_operand" "=q, q, r, q, r")
(match_operand:SI 1 "nonmemory_operand" "C_0, h, Lr,Cal,Cal")))]
""
"@
* current_insn_predicate = 0; return \"sub%?.ne\\t%0,%0,%0\";
* current_insn_predicate = 0; return \"mov%?.ne\\t%0,%1\";
mov.ne\\t%0,%1
* current_insn_predicate = 0; return \"mov%?.ne\\t%0,%1\";
mov.ne\\t%0,%1"
[(set_attr "type" "cmove")
(set_attr "iscompact" "true,true,false,true_limm,false")
(set_attr "length" "2,2,4,6,8")
(set_attr "cpu_facility" "*,av2,*,av2,*")])
(define_insn "*movsi_cond_exec"
[(cond_exec
(match_operator 3 "proper_comparison_operator"
[(match_operand 2 "cc_register" "Rcc,Rcc") (const_int 0)])
(set (match_operand:SI 0 "dest_reg_operand" "=w,w")
(match_operand:SI 1 "nonmemory_operand" "LRac,?Cal")))]
""
"mov.%d3 %0,%1"
[(set_attr "type" "cmove")
(set_attr "length" "4,8")])
(define_insn "*commutative_cond_exec"
[(cond_exec
(match_operator 5 "proper_comparison_operator"
[(match_operand 4 "cc_register" "Rcc,Rcc") (const_int 0)])
(set (match_operand:SI 0 "dest_reg_operand" "=w,w")
(match_operator:SI 3 "commutative_operator"
[(match_operand:SI 1 "register_operand" "%0,0")
(match_operand:SI 2 "nonmemory_operand" "cL,?Cal")])))]
""
{
arc_output_commutative_cond_exec (operands, true);
return "";
}
[(set_attr "cond" "use")
(set_attr "type" "cmove")
(set_attr_alternative "length"
[(const_int 4)
(cond
[(eq (symbol_ref "arc_output_commutative_cond_exec (operands, false)")
(const_int 4))
(const_int 4)]
(const_int 8))])])
(define_insn "*sub_cond_exec"
[(cond_exec
(match_operator 4 "proper_comparison_operator"
[(match_operand 3 "cc_register" "Rcc,Rcc,Rcc") (const_int 0)])
(set (match_operand:SI 0 "dest_reg_operand" "=w,w,w")
(minus:SI (match_operand:SI 1 "nonmemory_operand" "0,cL,Cal")
(match_operand:SI 2 "nonmemory_operand" "cL,0,0"))))]
""
"@
sub.%d4 %0,%1,%2
rsub.%d4 %0,%2,%1
rsub.%d4 %0,%2,%1"
[(set_attr "cond" "use")
(set_attr "type" "cmove")
(set_attr "length" "4,4,8")])
(define_insn "*noncommutative_cond_exec"
[(cond_exec
(match_operator 5 "proper_comparison_operator"
[(match_operand 4 "cc_register" "Rcc,Rcc") (const_int 0)])
(set (match_operand:SI 0 "dest_reg_operand" "=w,w")
(match_operator:SI 3 "noncommutative_operator"
[(match_operand:SI 1 "register_operand" "0,0")
(match_operand:SI 2 "nonmemory_operand" "cL,Cal")])))]
""
"%O3.%d5 %0,%1,%2"
[(set_attr "cond" "use")
(set_attr "type" "cmove")
(set_attr "length" "4,8")])
;; These control RTL generation for conditional jump insns
;; Match both normal and inverted jump.
; We need a separate expander for this lest we loose the mode of CC_REG
; when match_operator substitutes the literal operand into the comparison.
(define_expand "branch_insn"
[(set (pc)
(if_then_else (match_operand 1 "" "")
(label_ref (match_operand 0 "" ""))
(pc)))])
; When estimating sizes during arc_reorg, when optimizing for speed, there
; are three reasons why we need to consider branches to be length 6:
; - annull-false delay slot insns are implemented using conditional execution,
; thus preventing short insn formation where used.
; - for ARC600: annull-true delay slot isnns are implemented where possile
; using conditional execution, preventing short insn formation where used.
; - for ARC700: likely or somewhat likely taken branches are made long and
; unaligned if possible to avoid branch penalty.
(define_insn "*branch_insn"
[(set (pc)
(if_then_else (match_operator 1 "proper_comparison_operator"
[(reg CC_REG) (const_int 0)])
(label_ref (match_operand 0 "" ""))
(pc)))]
""
"*
{
if (arc_ccfsm_branch_deleted_p ())
{
arc_ccfsm_record_branch_deleted ();
return \"; branch deleted, next insns conditionalized\";
}
else
{
arc_ccfsm_record_condition (operands[1], false, insn, 0);
if (get_attr_length (insn) == 2)
return \"b%d1%? %^%l0%&\";
else
return \"b%d1%# %^%l0\";
}
}"
[(set_attr "type" "branch")
(set
(attr "length")
(cond [
(eq_attr "delay_slot_filled" "yes")
(const_int 4)
(ne
(if_then_else
(match_operand 1 "equality_comparison_operator" "")
(ior (lt (minus (match_dup 0) (pc)) (const_int -512))
(gt (minus (match_dup 0) (pc))
(minus (const_int 506)
(symbol_ref "get_attr_delay_slot_length (insn)"))))
(ior (match_test "!arc_short_comparison_p (operands[1], -64)")
(lt (minus (match_dup 0) (pc)) (const_int -64))
(gt (minus (match_dup 0) (pc))
(minus (const_int 58)
(symbol_ref "get_attr_delay_slot_length (insn)")))))
(const_int 0))
(const_int 4)]
(const_int 2)))
(set (attr "iscompact")
(cond [(match_test "get_attr_length (insn) == 2") (const_string "true")]
(const_string "false")))])
(define_insn "*rev_branch_insn"
[(set (pc)
(if_then_else (match_operator 1 "proper_comparison_operator"
[(reg CC_REG) (const_int 0)])
(pc)
(label_ref (match_operand 0 "" ""))))]
"REVERSIBLE_CC_MODE (GET_MODE (XEXP (operands[1], 0)))"
"*
{
if (arc_ccfsm_branch_deleted_p ())
{
arc_ccfsm_record_branch_deleted ();
return \"; branch deleted, next insns conditionalized\";
}
else
{
arc_ccfsm_record_condition (operands[1], true, insn, 0);
if (get_attr_length (insn) == 2)
return \"b%D1%? %^%l0\";
else
return \"b%D1%# %^%l0\";
}
}"
[(set_attr "type" "branch")
(set
(attr "length")
(cond [
(eq_attr "delay_slot_filled" "yes")
(const_int 4)
(ne
(if_then_else
(match_operand 1 "equality_comparison_operator" "")
(ior (lt (minus (match_dup 0) (pc)) (const_int -512))
(gt (minus (match_dup 0) (pc))
(minus (const_int 506)
(symbol_ref "get_attr_delay_slot_length (insn)"))))
(ior (match_test "!arc_short_comparison_p (operands[1], -64)")
(lt (minus (match_dup 0) (pc)) (const_int -64))
(gt (minus (match_dup 0) (pc))
(minus (const_int 58)
(symbol_ref "get_attr_delay_slot_length (insn)")))))
(const_int 0))
(const_int 4)]
(const_int 2)))
(set (attr "iscompact")
(cond [(match_test "get_attr_length (insn) == 2") (const_string "true")]
(const_string "false")))])
;; Unconditional and other jump instructions.
(define_expand "jump"
[(set (pc) (label_ref (match_operand 0 "" "")))]
""
"")
(define_insn "jump_i"
[(set (pc) (label_ref (match_operand 0 "" "")))]
"!TARGET_LONG_CALLS_SET || !CROSSING_JUMP_P (insn)"
"b%!%* %^%l0%&"
[(set_attr "type" "uncond_branch")
(set (attr "iscompact")
(if_then_else (match_test "get_attr_length (insn) == 2")
(const_string "true") (const_string "false")))
(set_attr "cond" "canuse")
(set (attr "length")
(cond [
; In arc_reorg we just guesstimate; might be more or less than 4.
(match_test "arc_branch_size_unknown_p ()")
(const_int 4)
(eq_attr "delay_slot_filled" "yes")
(const_int 4)
(match_test "CROSSING_JUMP_P (insn)")
(const_int 4)
(ior (lt (minus (match_dup 0) (pc)) (const_int -512))
(gt (minus (match_dup 0) (pc))
(minus (const_int 506)
(symbol_ref "get_attr_delay_slot_length (insn)"))))
(const_int 4)]
(const_int 2)))])
(define_insn "indirect_jump"
[(set (pc) (match_operand:SI 0 "nonmemory_operand" "L,I,Cal,q,r"))]
""
"@
j%!%* %0%&
j%!%* %0%&
j%!%* %0%&
j%!%* [%0]%&
j%!%* [%0]%&"
[(set_attr "type" "jump")
(set_attr "iscompact" "false,false,false,maybe,false")
(set_attr "cond" "canuse,canuse_limm,canuse,canuse,canuse")])
;; Implement a switch statement.
(define_expand "casesi"
[(match_operand:SI 0 "register_operand" "") ; Index
(match_operand:SI 1 "const_int_operand" "") ; Lower bound
(match_operand:SI 2 "const_int_operand" "") ; Total range
(match_operand:SI 3 "" "") ; Table label
(match_operand:SI 4 "" "")] ; Out of range label
""
{
if (operands[1] != const0_rtx)
{
rtx reg = gen_reg_rtx (SImode);
emit_insn (gen_subsi3 (reg, operands[0], operands[1]));
operands[0] = reg;
}
emit_jump_insn (gen_cbranchsi4 (gen_rtx_GTU (SImode, operands[0],
operands[2]),
operands[0], operands[2], operands[4]));
if (TARGET_BI_BIH)
emit_jump_insn (gen_casesi_dispatch (operands[0], operands[3]));
else
{
rtx reg = gen_reg_rtx (SImode);
rtx lbl = operands[3];
operands[3] = gen_rtx_LABEL_REF (VOIDmode, operands[3]);
if (flag_pic)
operands[3] = force_reg (Pmode, operands[3]);
emit_insn (gen_casesi_load (reg,
operands[3], operands[0], lbl));
if (CASE_VECTOR_PC_RELATIVE || flag_pic)
emit_insn (gen_addsi3 (reg, reg, operands[3]));
emit_jump_insn (gen_casesi_jump (reg, lbl));
}
DONE;
})
(define_insn "casesi_dispatch"
[(set (pc)
(unspec:SI [(match_operand:SI 0 "register_operand" "r")
(label_ref (match_operand 1 "" ""))]
UNSPEC_ARC_CASESI))]
"TARGET_BI_BIH"
{
rtx diff_vec = PATTERN (next_nonnote_insn (as_a (operands[1])));
gcc_assert (GET_CODE (diff_vec) == ADDR_DIFF_VEC);
switch (GET_MODE (diff_vec))
{
case E_SImode:
return \"bi\\t[%0]\";
case E_HImode:
case E_QImode:
return \"bih\\t[%0]\";
default: gcc_unreachable ();
}
}
[(set_attr "type" "brcc_no_delay_slot")
(set_attr "iscompact" "false")
(set_attr "length" "4")])
(define_insn "casesi_load"
[(set (match_operand:SI 0 "register_operand" "=q,r,r")
(mem:SI (unspec:SI [(match_operand:SI 1 "nonmemory_operand" "q,r,Cal")
(match_operand:SI 2 "register_operand" "q,r,r")]
UNSPEC_ARC_CASESI)))
(use (label_ref (match_operand 3 "" "")))]
""
"*
{
rtx diff_vec = PATTERN (next_nonnote_insn (as_a (operands[3])));
if (GET_CODE (diff_vec) != ADDR_DIFF_VEC)
{
gcc_assert (GET_CODE (diff_vec) == ADDR_VEC);
gcc_assert (GET_MODE (diff_vec) == SImode);
gcc_assert (!CASE_VECTOR_PC_RELATIVE && !flag_pic);
}
switch (GET_MODE (diff_vec))
{
case E_SImode:
return \"ld.as\\t%0,[%1,%2]%&\";
case E_HImode:
if (ADDR_DIFF_VEC_FLAGS (diff_vec).offset_unsigned)
return \"ld%_.as\\t%0,[%1,%2]\";
return \"ld%_.x.as\\t%0,[%1,%2]\";
case E_QImode:
if (ADDR_DIFF_VEC_FLAGS (diff_vec).offset_unsigned)
return \"ldb%?\\t%0,[%1,%2]%&\";
return \"ldb.x\\t%0,[%1,%2]\";
default:
gcc_unreachable ();
}
}"
[(set_attr "type" "load")
(set_attr_alternative "iscompact"
[(cond
[(ne (symbol_ref "GET_MODE (PATTERN (next_nonnote_insn
(as_a (operands[3]))))")
(symbol_ref "QImode"))
(const_string "false")
(match_test "!ADDR_DIFF_VEC_FLAGS (PATTERN (next_nonnote_insn
(as_a (operands[3])))).offset_unsigned")
(const_string "false")]
(const_string "true"))
(const_string "false")
(const_string "false")])
(set_attr_alternative "length"
[(cond
[(eq_attr "iscompact" "false") (const_int 4)
; We have to mention (match_dup 3) to convince genattrtab.cc that this
; is a varying length insn.
(eq (symbol_ref "1+1") (const_int 2)) (const_int 2)
(gt (minus (match_dup 3) (pc)) (const_int 42)) (const_int 4)]
(const_int 2))
(const_int 4)
(const_int 8)])])
; Unlike the canonical tablejump, this pattern always uses a jump address,
; even for CASE_VECTOR_PC_RELATIVE.
(define_insn "casesi_jump"
[(set (pc) (match_operand:SI 0 "register_operand" "Cal,q,c"))
(use (label_ref (match_operand 1 "" "")))]
""
"j%!%* [%0]%&"
[(set_attr "type" "jump")
(set_attr "iscompact" "false,maybe,false")
(set_attr "cond" "canuse")])
(define_expand "call"
;; operands[1] is stack_size_rtx
;; operands[2] is next_arg_register
[(parallel [(call (match_operand:SI 0 "call_operand" "")
(match_operand 1 "" ""))
(clobber (reg:SI 31))])]
""
"{
rtx callee;
gcc_assert (MEM_P (operands[0]));
callee = XEXP (operands[0], 0);
/* This is to decide if we should generate indirect calls by loading the
32 bit address of the callee into a register before performing the
branch and link - this exposes cse opportunities.
Also, in weird cases like compile/20010107-1.c, we may get a PLUS. */
if (GET_CODE (callee) != REG
&& (GET_CODE (callee) == PLUS || arc_is_longcall_p (callee)))
XEXP (operands[0], 0) = force_reg (Pmode, callee);
}
")
; At instruction output time, if it doesn't match and we end up with
; alternative 1 ("q"), that means that we can't use the short form.
(define_insn "*call_i"
[(call (mem:SI (match_operand:SI 0
"call_address_operand" "q,c,Cji,Csc,Cbp,Cbr,L,I,Cal"))
(match_operand 1 "" ""))
(clobber (reg:SI 31))]
""
"@
jl%!%* [%0]%&
jl%!%* [%0]
jli_s %S0
sjli %S0
bl%!%* %P0
bl%!%* %P0
jl%!%* %0
jl%* %0
jl%! %0"
[(set_attr "type" "call,call,call_no_delay_slot,call_no_delay_slot,call,call,call,call,call_no_delay_slot")
(set_attr "iscompact" "maybe,*,true,*,*,*,*,*,*")
(set_attr "predicable" "no,yes,no,no,yes,no,yes,no,yes")
(set_attr "length" "*,4,2,4,4,4,4,4,8")])
(define_expand "call_value"
;; operand 2 is stack_size_rtx
;; operand 3 is next_arg_register
[(parallel [(set (match_operand 0 "dest_reg_operand" "=r")
(call (match_operand:SI 1 "call_operand" "")
(match_operand 2 "" "")))
(clobber (reg:SI 31))])]
""
"
{
rtx callee;
gcc_assert (MEM_P (operands[1]));
callee = XEXP (operands[1], 0);
/* See the comment in define_expand \"call\". */
if (GET_CODE (callee) != REG
&& (GET_CODE (callee) == PLUS || arc_is_longcall_p (callee)))
XEXP (operands[1], 0) = force_reg (Pmode, callee);
}")
; At instruction output time, if it doesn't match and we end up with
; alternative 1 ("q"), that means that we can't use the short form.
(define_insn "*call_value_i"
[(set (match_operand 0 "dest_reg_operand" "=q,w, w, w, w, w,w,w, w")
(call (mem:SI (match_operand:SI 1
"call_address_operand" "q,c,Cji,Csc,Cbp,Cbr,L,I,Cal"))
(match_operand 2 "" "")))
(clobber (reg:SI 31))]
""
"@
jl%!%* [%1]%&
jl%!%* [%1]
jli_s %S1
sjli %S1
bl%!%* %P1;1
bl%!%* %P1;1
jl%!%* %1
jl%* %1
jl%! %1"
[(set_attr "type" "call,call,call_no_delay_slot,call_no_delay_slot,call,call,call,call,call_no_delay_slot")
(set_attr "iscompact" "maybe,*,true,false,*,*,*,*,*")
(set_attr "predicable" "no,yes,no,no,yes,no,yes,no,yes")
(set_attr "length" "*,4,2,4,4,4,4,4,8")])
; There is a bl_s instruction (16 bit opcode branch-and-link), but we can't
; use it for lack of inter-procedural branch shortening.
; Link-time relaxation would help...
(define_insn "trap"
[(trap_if (const_int 1) (const_int 0))]
"!TARGET_ARC600_FAMILY"
"trap_s\\t5"
[(set_attr "type" "misc")
(set_attr "length" "2")])
(define_insn "nop"
[(const_int 0)]
""
"nop%?"
[(set_attr "type" "misc")
(set_attr "iscompact" "true")
(set_attr "cond" "canuse")
(set_attr "length" "2")])
(define_insn "nopv"
[(unspec_volatile [(const_int 0)] VUNSPEC_ARC_NOP)]
""
"nop%?"
[(set_attr "type" "misc")
(set_attr "iscompact" "maybe")
(set_attr "length" "*")])
(define_insn "blockage"
[(unspec_volatile [(const_int 0)] VUNSPEC_ARC_BLOCKAGE)]
""
""
[(set_attr "length" "0")
(set_attr "type" "block")]
)
(define_insn "arc600_stall"
[(unspec_volatile [(const_int 0)] VUNSPEC_ARC_ARC600_STALL)]
"TARGET_MUL64_SET"
"mov\\t0,mlo\t;wait until multiply complete."
[(set_attr "length" "4")
(set_attr "type" "move")]
)
;; Split up troublesome insns for better scheduling.
;; Peepholes go at the end.
;;asl followed by add can be replaced by an add{1,2,3}
;; Three define_peepholes have been added for this optimization
;; ??? This used to target non-canonical rtl. Now we use add_n, which
;; can be generated by combine. Check if these peepholes still provide
;; any benefit.
;; -------------------------------------------------------------
;; Pattern 1 : r0 = r1 << {i}
;; r3 = r4/INT + r0 ;;and commutative
;; ||
;; \/
;; add{i} r3,r4/INT,r1
;; -------------------------------------------------------------
;; ??? This should be covered by combine, alas, at times combine gets
;; too clever for it's own good: when the shifted input is known to be
;; either 0 or 1, the operation will be made into an if-then-else, and
;; thus fail to match the add_n pattern. Example: _mktm_r, line 85 in
;; newlib/libc/time/mktm_r.c .
(define_peephole2
[(set (match_operand:SI 0 "dest_reg_operand" "")
(ashift:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "_1_2_3_operand" "")))
(set (match_operand:SI 3 "dest_reg_operand" "")
(plus:SI (match_operand:SI 4 "arc_nonmemory_operand" "")
(match_operand:SI 5 "arc_nonmemory_operand" "")))]
"(true_regnum (operands[4]) == true_regnum (operands[0])
|| true_regnum (operands[5]) == true_regnum (operands[0]))
&& (peep2_reg_dead_p (2, operands[0])
|| (true_regnum (operands[3]) == true_regnum (operands[0])))"
;; the preparation statements take care to put proper operand in
;; operands[4] operands[4] will always contain the correct
;; operand. This is added to satisfy commutativity
[(set (match_dup 3)
(plus:SI (mult:SI (match_dup 1)
(match_dup 2))
(match_dup 4)))]
"if (true_regnum (operands[4]) == true_regnum (operands[0]))
operands[4] = operands[5];
operands[2] = GEN_INT (1 << INTVAL (operands[2]));"
)
;; -------------------------------------------------------------
;; Pattern 1 : r0 = r1 << {i}
;; r3 = r4 - r0
;; ||
;; \/
;; sub{i} r3,r4,r1
;; -------------------------------------------------------------
;; ??? This should be covered by combine, alas, at times combine gets
;; too clever for it's own good: when the shifted input is known to be
;; either 0 or 1, the operation will be made into an if-then-else, and
;; thus fail to match the sub_n pattern. Example: __ieee754_yn, line 239 in
;; newlib/libm/math/e_jn.c .
(define_peephole2
[(set (match_operand:SI 0 "dest_reg_operand" "")
(ashift:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "const_int_operand" "")))
(set (match_operand:SI 3 "dest_reg_operand" "")
(minus:SI (match_operand:SI 4 "nonmemory_operand" "")
(match_dup 0)))]
"(INTVAL (operands[2]) == 1
|| INTVAL (operands[2]) == 2
|| INTVAL (operands[2]) == 3)
&& (peep2_reg_dead_p (2, operands[0])
|| (true_regnum (operands[3]) == true_regnum (operands[0])))"
[(set (match_dup 3)
(minus:SI (match_dup 4)
(mult:SI (match_dup 1)
(match_dup 2))))]
"operands[2] = GEN_INT (1 << INTVAL (operands[2]));"
)
; When using the high single bit, the result of a multiply is either
; the original number or zero. But MPY costs 4 cycles, which we
; can replace with the 2 cycles for the pair of TST_S and ADD.NE.
(define_peephole2
[(set (match_operand:SI 0 "dest_reg_operand" "")
(lshiftrt:SI (match_operand:SI 1 "register_operand" "")
(const_int 31)))
(set (match_operand:SI 4 "register_operand" "")
(mult:SI (match_operand:SI 2 "register_operand")
(match_operand:SI 3 "nonmemory_operand" "")))]
"TARGET_ARC700_MPY
&& (rtx_equal_p (operands[0], operands[2])
|| rtx_equal_p (operands[0], operands[3]))
&& peep2_regno_dead_p (0, CC_REG)
&& (rtx_equal_p (operands[0], operands[4])
|| (peep2_reg_dead_p (2, operands[0])
&& peep2_reg_dead_p (1, operands[4])))"
[(parallel [(set (reg:CC_Z CC_REG)
(compare:CC_Z (lshiftrt:SI (match_dup 1) (const_int 31))
(const_int 0)))
(set (match_dup 4) (lshiftrt:SI (match_dup 1) (const_int 31)))])
(cond_exec
(ne (reg:CC_Z CC_REG) (const_int 0))
(set (match_dup 4) (match_dup 5)))]
{
if (!rtx_equal_p (operands[0], operands[2]))
operands[5] = operands[2];
else if (!rtx_equal_p (operands[0], operands[3]))
operands[5] = operands[3];
else
operands[5] = operands[4]; /* Actually a no-op... presumably rare. */
})
(define_peephole2
[(set (match_operand:SI 0 "dest_reg_operand" "")
(lshiftrt:SI (match_operand:SI 1 "register_operand" "")
(const_int 31)))
(set (match_operand:SI 4 "register_operand" "")
(mult:SI (match_operand:SI 2 "register_operand")
(match_operand:SI 3 "nonmemory_operand" "")))]
"TARGET_ARC700_MPY
&& (rtx_equal_p (operands[0], operands[2])
|| rtx_equal_p (operands[0], operands[3]))
&& peep2_regno_dead_p (2, CC_REG)"
[(parallel [(set (reg:CC_Z CC_REG)
(compare:CC_Z (lshiftrt:SI (match_dup 1) (const_int 31))
(const_int 0)))
(set (match_dup 0) (lshiftrt:SI (match_dup 1) (const_int 31)))])
(set (match_dup 4) (match_dup 5))
(cond_exec
(eq (reg:CC_Z CC_REG) (const_int 0))
(set (match_dup 4) (const_int 0)))]
"operands[5] = operands[rtx_equal_p (operands[0], operands[2]) ? 3 : 2];")
;; Instructions generated through builtins
(define_insn "clrsbsi2"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r")
(clrsb:SI (match_operand:SI 1 "general_operand" "rL,Cal")))]
"TARGET_NORM"
"norm\\t%0,%1"
[(set_attr "length" "4,8")
(set_attr "type" "two_cycle_core,two_cycle_core")])
(define_insn "norm_f"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r")
(clrsb:SI (match_operand:SI 1 "general_operand" "rL,Cal")))
(set (reg:CC_ZN CC_REG)
(compare:CC_ZN (match_dup 1) (const_int 0)))]
"TARGET_NORM"
"norm.f\\t%0,%1"
[(set_attr "length" "4,8")
(set_attr "type" "two_cycle_core,two_cycle_core")])
(define_insn_and_split "clrsbhi2"
[(set (match_operand:HI 0 "dest_reg_operand" "=w,w")
(clrsb:HI (match_operand:HI 1 "general_operand" "cL,Cal")))]
"TARGET_NORM"
"#"
"reload_completed"
[(set (match_dup 0) (zero_extend:SI (clrsb:HI (match_dup 1))))]
"operands[0] = simplify_gen_subreg (SImode, operands[0], HImode, 0);")
(define_insn "normw"
[(set (match_operand:SI 0 "dest_reg_operand" "=w,w")
(zero_extend:SI
(clrsb:HI (match_operand:HI 1 "general_operand" "cL,Cal"))))]
"TARGET_NORM"
"@
norm%_ \t%0, %1
norm%_ \t%0, %1"
[(set_attr "length" "4,8")
(set_attr "type" "two_cycle_core,two_cycle_core")])
(define_expand "clzsi2"
[(parallel
[(set (match_operand:SI 0 "register_operand" "")
(clz:SI (match_operand:SI 1 "register_operand" "")))
(clobber (match_dup 2))])]
"TARGET_NORM"
"
if (TARGET_V2)
{
/* ARCv2's FLS is a bit more optimal than using norm. */
rtx tmp = gen_reg_rtx (SImode);
emit_insn (gen_fls (tmp, operands[1]));
emit_insn (gen_subsi3 (operands[0], GEN_INT (31), tmp));
DONE;
}
operands[2] = gen_rtx_REG (CC_ZNmode, CC_REG);
")
(define_insn_and_split "*arc_clzsi2"
[(set (match_operand:SI 0 "register_operand" "=r")
(clz:SI (match_operand:SI 1 "register_operand" "r")))
(clobber (reg:CC_ZN CC_REG))]
"TARGET_NORM"
"#"
"reload_completed"
[(const_int 0)]
{
emit_insn (gen_norm_f (operands[0], operands[1]));
emit_insn
(gen_rtx_COND_EXEC
(VOIDmode,
gen_rtx_LT (VOIDmode, gen_rtx_REG (CC_ZNmode, CC_REG), const0_rtx),
gen_rtx_SET (operands[0], const0_rtx)));
emit_insn
(gen_rtx_COND_EXEC
(VOIDmode,
gen_rtx_GE (VOIDmode, gen_rtx_REG (CC_ZNmode, CC_REG), const0_rtx),
gen_rtx_SET (operands[0], plus_constant (SImode, operands[0], 1))));
DONE;
}
[(set_attr "type" "unary")
(set_attr "length" "12")])
(define_expand "ctzsi2"
[(match_operand:SI 0 "register_operand" "")
(match_operand:SI 1 "register_operand" "")]
"TARGET_NORM"
"
if (TARGET_V2)
{
emit_insn (gen_ffs (operands[0], operands[1]));
DONE;
}
emit_insn (gen_arc_ctzsi2 (operands[0], operands[1]));
DONE;
")
(define_insn_and_split "arc_ctzsi2"
[(set (match_operand:SI 0 "register_operand" "=r")
(ctz:SI (match_operand:SI 1 "register_operand" "r")))
(clobber (reg:CC_ZN CC_REG))
(clobber (match_scratch:SI 2 "=&r"))]
"TARGET_NORM"
"#"
"reload_completed"
[(const_int 0)]
{
rtx temp = operands[0];
if (reg_overlap_mentioned_p (temp, operands[1])
|| (REGNO (temp) < FIRST_PSEUDO_REGISTER
&& !TEST_HARD_REG_BIT (reg_class_contents[GENERAL_REGS],
REGNO (temp))))
temp = operands[2];
emit_insn (gen_addsi3 (temp, operands[1], constm1_rtx));
emit_insn (gen_bic_f_zn (temp, temp, operands[1]));
emit_insn (gen_clrsbsi2 (operands[0], temp));
emit_insn
(gen_rtx_COND_EXEC
(VOIDmode,
gen_rtx_LT (VOIDmode, gen_rtx_REG (CC_ZNmode, CC_REG), const0_rtx),
gen_rtx_SET (operands[0], GEN_INT (32))));
emit_insn
(gen_rtx_COND_EXEC
(VOIDmode,
gen_rtx_GE (VOIDmode, gen_rtx_REG (CC_ZNmode, CC_REG), const0_rtx),
gen_rtx_SET (operands[0], gen_rtx_MINUS (SImode, GEN_INT (31),
operands[0]))));
DONE;
}
[(set_attr "type" "unary")
(set_attr "length" "20")])
(define_insn "swap"
[(set (match_operand:SI 0 "dest_reg_operand" "=w,w,w")
(unspec:SI [(match_operand:SI 1 "general_operand" "L,Cal,c")]
UNSPEC_ARC_SWAP))]
"TARGET_SWAP"
"@
swap \t%0, %1
swap \t%0, %1
swap \t%0, %1"
[(set_attr "length" "4,8,4")
(set_attr "type" "two_cycle_core,two_cycle_core,two_cycle_core")])
(define_insn "divaw"
[(set (match_operand:SI 0 "dest_reg_operand" "=&w,&w,&w")
(unspec:SI [(div:SI (match_operand:SI 1 "general_operand" "r,Cal,r")
(match_operand:SI 2 "general_operand" "r,r,Cal"))]
UNSPEC_ARC_DIVAW))]
"TARGET_ARC700 || TARGET_EA_SET"
"@
divaw \t%0, %1, %2
divaw \t%0, %1, %2
divaw \t%0, %1, %2"
[(set_attr "length" "4,8,8")
(set_attr "type" "divaw,divaw,divaw")])
(define_insn "flag"
[(unspec_volatile [(match_operand:SI 0 "nonmemory_operand" "rL,I,Cal")]
VUNSPEC_ARC_FLAG)]
""
"@
flag%? %0
flag %0
flag%? %0"
[(set_attr "length" "4,4,8")
(set_attr "type" "misc,misc,misc")
(set_attr "predicable" "yes,no,yes")
(set_attr "cond" "clob,clob,clob")])
(define_insn "brk"
[(unspec_volatile [(match_operand:SI 0 "immediate_operand" "N")]
VUNSPEC_ARC_BRK)]
""
"brk"
[(set_attr "length" "4")
(set_attr "type" "misc")])
(define_insn "rtie"
[(return)
(unspec_volatile [(const_int 0)] VUNSPEC_ARC_RTIE)]
"!TARGET_ARC600_FAMILY"
"rtie"
[(set_attr "length" "4")
(set_attr "type" "rtie")
(set_attr "cond" "clob")])
(define_insn "sync"
[(unspec_volatile [(match_operand:SI 0 "immediate_operand" "N")]
VUNSPEC_ARC_SYNC)]
""
"sync"
[(set_attr "length" "4")
(set_attr "type" "misc")])
(define_insn "swi"
[(unspec_volatile [(match_operand:SI 0 "immediate_operand" "N")]
VUNSPEC_ARC_SWI)]
""
"*
{
if(TARGET_ARC700)
return \"trap0\";
else
return \"swi\";
}"
[(set_attr "length" "4")
(set_attr "type" "misc")])
(define_insn "sleep"
[(unspec_volatile [(match_operand:SI 0 "nonmemory_operand" "Lr")]
VUNSPEC_ARC_SLEEP)]
""
"sleep %0"
[(set_attr "length" "4")
(set_attr "type" "misc")])
(define_insn "core_read"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r")
(unspec_volatile:SI [(match_operand:SI 1 "general_operand" "Hn,!r")]
VUNSPEC_ARC_CORE_READ))]
""
"*
if (check_if_valid_regno_const (operands, 1))
return \"mov \t%0, r%1\";
return \"mov \t%0, r%1\";
"
[(set_attr "length" "4")
(set_attr "type" "unary")])
(define_insn "core_write"
[(unspec_volatile [(match_operand:SI 0 "general_operand" "r,r")
(match_operand:SI 1 "general_operand" "Hn,!r")]
VUNSPEC_ARC_CORE_WRITE)]
""
"*
if (check_if_valid_regno_const (operands, 1))
return \"mov \tr%1, %0\";
return \"mov \tr%1, %0\";
"
[(set_attr "length" "4")
(set_attr "type" "unary")])
(define_insn "lr"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r,r,r")
(unspec_volatile:SI [(match_operand:SI 1 "general_operand" "I,HCal,r,D")]
VUNSPEC_ARC_LR))]
""
"lr\t%0, [%1]"
[(set_attr "length" "4,8,4,8")
(set_attr "type" "lr,lr,lr,lr")])
(define_insn "sr"
[(unspec_volatile [(match_operand:SI 0 "general_operand" "Cal,r,r,r")
(match_operand:SI 1 "general_operand" "Ir,I,HCal,r")]
VUNSPEC_ARC_SR)]
""
"sr\t%0, [%1]"
[(set_attr "length" "8,4,8,4")
(set_attr "type" "sr,sr,sr,sr")])
(define_mode_iterator ALLI [QI HI SI (DI "TARGET_LL64")])
(define_mode_attr mALLI [(QI "b") (HI "%_") (SI "") (DI "d")])
(define_insn "lddi"
[(set (match_operand:ALLI 0 "register_operand" "=r")
(unspec_volatile:ALLI [(match_operand:ALLI 1 "memory_operand" "m")]
VUNSPEC_ARC_LDDI))]
""
"ld%U1.di\\t%0,%1"
[(set_attr "type" "load")])
(define_insn "stdi"
[(unspec_volatile [(match_operand:ALLI 0 "memory_operand" "m,m,Usc")
(match_operand:ALLI 1 "nonmemory_operand" "r,Cm3,i")]
VUNSPEC_ARC_STDI)]
""
"st%U0.di\\t%1,%0"
[(set_attr "length" "*,*,8")
(set_attr "type" "store")])
(define_insn_and_split "*stdidi_split"
[(unspec_volatile [(match_operand:DI 0 "memory_operand" "m")
(match_operand:DI 1 "register_operand" "r")]
VUNSPEC_ARC_STDI)]
"!TARGET_LL64"
"#"
"&& reload_completed"
[(unspec_volatile:SI [(match_dup 2) (match_dup 3)] VUNSPEC_ARC_STDI)
(unspec_volatile:SI [(match_dup 4) (match_dup 5)] VUNSPEC_ARC_STDI)]
"
{
operands[3] = gen_lowpart (SImode, operands[1]);
operands[5] = gen_highpart_mode (SImode, DImode, operands[1]);
operands[2] = gen_lowpart (SImode, operands[0]);
operands[4] = gen_highpart (SImode, operands[0]);
}
"
)
(define_insn_and_split "*lddidi_split"
[(set (match_operand:DI 0 "register_operand" "=r")
(unspec_volatile:DI [(match_operand:DI 1 "memory_operand" "m")]
VUNSPEC_ARC_LDDI))]
"!TARGET_LL64"
"#"
"&& reload_completed"
[(set (match_dup 2) (unspec_volatile:SI [(match_dup 3)] VUNSPEC_ARC_LDDI))
(set (match_dup 4) (unspec_volatile:SI [(match_dup 5)] VUNSPEC_ARC_LDDI))]
"
{
operands[3] = gen_lowpart (SImode, operands[1]);
operands[5] = gen_highpart (SImode, operands[1]);
operands[2] = gen_lowpart (SImode, operands[0]);
operands[4] = gen_highpart (SImode, operands[0]);
}
"
)
(define_insn "trap_s"
[(unspec_volatile [(match_operand:SI 0 "immediate_operand" "L,Cal")]
VUNSPEC_ARC_TRAP_S)]
"!TARGET_ARC600_FAMILY"
{
if (which_alternative == 0)
{
arc_toggle_unalign ();
return \"trap_s %0\";
}
/* Keep this message in sync with the one in arc.cc:arc_expand_builtin,
because *.md files do not get scanned by exgettext. */
fatal_error (input_location,
\"operand to % should be an unsigned 6-bit value\");
}
[(set_attr "length" "2")
(set_attr "type" "misc")])
(define_insn "unimp_s"
[(unspec_volatile [(match_operand:SI 0 "immediate_operand" "N")]
VUNSPEC_ARC_UNIMP_S)]
"!TARGET_ARC600_FAMILY"
"unimp_s"
[(set_attr "length" "4")
(set_attr "type" "misc")])
;; End of instructions generated through builtins
; Since the demise of REG_N_SETS as reliable data readily available to the
; target, it is no longer possible to find out
; in the prologue / epilogue expanders how many times blink is set.
; Using df_regs_ever_live_p to decide if blink needs saving means that
; any explicit use of blink will cause it to be saved; hence we cannot
; represent the blink use in return / sibcall instructions themselves, and
; instead have to show it in EPILOGUE_USES and must explicitly
; forbid instructions that change blink in the return / sibcall delay slot.
(define_expand "sibcall"
[(parallel [(call (match_operand 0 "memory_operand" "")
(match_operand 1 "general_operand" ""))
(simple_return)
(use (match_operand 2 "" ""))])]
""
"
{
rtx callee = XEXP (operands[0], 0);
if (operands[2] == NULL_RTX)
operands[2] = const0_rtx;
if (GET_CODE (callee) != REG
&& (GET_CODE (callee) == PLUS || arc_is_longcall_p (callee)))
XEXP (operands[0], 0) = force_reg (Pmode, callee);
}"
)
(define_expand "sibcall_value"
[(parallel [(set (match_operand 0 "dest_reg_operand" "")
(call (match_operand 1 "memory_operand" "")
(match_operand 2 "general_operand" "")))
(simple_return)
(use (match_operand 3 "" ""))])]
""
"
{
rtx callee = XEXP (operands[1], 0);
if (operands[3] == NULL_RTX)
operands[3] = const0_rtx;
if (GET_CODE (callee) != REG && arc_is_longcall_p (callee))
XEXP (operands[1], 0) = force_reg (Pmode, callee);
}"
)
(define_insn "*sibcall_insn"
[(call (mem:SI (match_operand:SI 0 "call_address_operand"
"Cbp,Cbr,!Rcd,Rsc,Cal"))
(match_operand 1 "" ""))
(simple_return)
(use (match_operand 2 "" ""))]
""
"@
b%!%*\\t%P0
b%!%*\\t%P0
j%!%*\\t[%0]
j%!%*\\t[%0]
j%!\\t%P0"
[(set_attr "type" "call,call,call,call,call_no_delay_slot")
(set_attr "predicable" "yes,no,no,yes,yes")
(set_attr "iscompact" "false,false,maybe,false,false")
(set_attr "is_SIBCALL" "yes")]
)
(define_insn "*sibcall_value_insn"
[(set (match_operand 0 "dest_reg_operand" "")
(call (mem:SI (match_operand:SI 1 "call_address_operand"
"Cbp,Cbr,!Rcd,Rsc,Cal"))
(match_operand 2 "" "")))
(simple_return)
(use (match_operand 3 "" ""))]
""
"@
b%!%*\\t%P1
b%!%*\\t%P1
j%!%*\\t[%1]
j%!%*\\t[%1]
j%!\\t%P1"
[(set_attr "type" "call,call,call,call,call_no_delay_slot")
(set_attr "predicable" "yes,no,no,yes,yes")
(set_attr "iscompact" "false,false,maybe,false,false")
(set_attr "is_SIBCALL" "yes")]
)
(define_expand "prologue"
[(pc)]
""
{
arc_expand_prologue ();
DONE;
})
(define_expand "epilogue"
[(pc)]
""
{
arc_expand_epilogue (0);
DONE;
})
(define_expand "sibcall_epilogue"
[(pc)]
""
{
arc_expand_epilogue (1);
DONE;
})
; Since the demise of REG_N_SETS, it is no longer possible to find out
; in the prologue / epilogue expanders how many times blink is set.
; Using df_regs_ever_live_p to decide if blink needs saving means that
; any explicit use of blink will cause it to be saved; hence we cannot
; represent the blink use in return / sibcall instructions themselves, and
; instead have to show it in EPILOGUE_USES and must explicitly
; forbid instructions that change blink in the return / sibcall delay slot.
(define_insn "simple_return"
[(simple_return)]
""
"j%!%*\\t[blink]"
[(set_attr "type" "return")
(set_attr "cond" "canuse")
(set_attr "iscompact" "maybe")
(set_attr "length" "*")])
(define_insn "arc600_rtie"
[(return)
(unspec_volatile [(match_operand 0 "pmode_register_operand" "")]
VUNSPEC_ARC_ARC600_RTIE)]
"TARGET_ARC600_FAMILY"
"j.f\\t[%0]"
[(set_attr "length" "4")
(set_attr "type" "rtie")
(set_attr "cond" "clob")])
(define_insn "p_return_i"
[(set (pc)
(if_then_else (match_operator 0 "proper_comparison_operator"
[(reg CC_REG) (const_int 0)])
(simple_return) (pc)))]
"reload_completed"
{
output_asm_insn (\"j%d0%!%#\\t[blink]\", operands);
/* record the condition in case there is a delay insn. */
arc_ccfsm_record_condition (operands[0], false, insn, 0);
return \"\";
}
[(set_attr "type" "return")
(set_attr "cond" "use")
(set_attr "iscompact" "maybe" )
(set (attr "length")
(cond [(not (match_operand 0 "equality_comparison_operator" ""))
(const_int 4)
(eq_attr "delay_slot_filled" "yes")
(const_int 4)]
(const_int 2)))])
;; Return nonzero if this function is known to have a null epilogue.
;; This allows the optimizer to omit jumps to jumps if no stack
;; was created.
(define_expand "return"
[(return)]
"arc_can_use_return_insn ()"
"")
;; Comment in final.cc (insn_current_reference_address) says
;; forward branch addresses are calculated from the next insn after branch
;; and for backward branches, it is calculated from the branch insn start.
;; The shortening logic here is tuned to accomodate this behavior
;; ??? This should be grokked by the ccfsm machinery.
(define_insn "cbranchsi4_scratch"
[(set (pc)
(if_then_else (match_operator 0 "proper_comparison_operator"
[(match_operand:SI 1 "register_operand" "c,c, c")
(match_operand:SI 2 "nonmemory_operand" "L,c,?Cal")])
(label_ref (match_operand 3 "" ""))
(pc)))
(clobber (match_operand 4 "cc_register" ""))]
"(reload_completed
|| (TARGET_EARLY_CBRANCHSI
&& brcc_nolimm_operator (operands[0], VOIDmode)))
&& !CROSSING_JUMP_P (insn)"
"*
switch (get_attr_length (insn))
{
case 2: return \"br%d0%? %1, %2, %^%l3%&\";
case 4: return \"br%d0%* %1, %B2, %^%l3\";
case 8: if (!brcc_nolimm_operator (operands[0], VOIDmode))
return \"br%d0%* %1, %B2, %^%l3\";
/* FALLTHRU */
case 6: case 10:
case 12:return \"cmp%? %1, %B2\\n\\tb%d0%* %^%l3%& ;br%d0 out of range\";
default: fprintf (stderr, \"unexpected length %d\\n\", get_attr_length (insn)); fflush (stderr); gcc_unreachable ();
}
"
[(set_attr "cond" "clob, clob, clob")
(set (attr "type")
(if_then_else
(match_test "valid_brcc_with_delay_p (operands)")
(const_string "brcc")
(const_string "brcc_no_delay_slot")))
; For forward branches, we need to account not only for the distance to
; the target, but also the difference between pcl and pc, the instruction
; length, and any delay insn, if present.
(set
(attr "length")
(cond ; the outer cond does a test independent of branch shortening.
[(match_operand 0 "brcc_nolimm_operator" "")
(cond
[(and (match_operand:CC_Z 4 "cc_register")
(eq_attr "delay_slot_filled" "no")
(ge (minus (match_dup 3) (pc)) (const_int -128))
(le (minus (match_dup 3) (pc))
(minus (const_int 122)
(symbol_ref "get_attr_delay_slot_length (insn)"))))
(const_int 2)
(and (ge (minus (match_dup 3) (pc)) (const_int -256))
(le (minus (match_dup 3) (pc))
(minus (const_int 244)
(symbol_ref "get_attr_delay_slot_length (insn)"))))
(const_int 4)
(and (match_operand:SI 1 "compact_register_operand" "")
(match_operand:SI 2 "compact_hreg_operand" ""))
(const_int 6)]
(const_int 8))]
(cond [(and (ge (minus (match_dup 3) (pc)) (const_int -256))
(le (minus (match_dup 3) (pc)) (const_int 244)))
(const_int 8)
(and (match_operand:SI 1 "compact_register_operand" "")
(match_operand:SI 2 "compact_hreg_operand" ""))
(const_int 10)]
(const_int 12))))
(set (attr "iscompact")
(if_then_else (match_test "get_attr_length (insn) & 2")
(const_string "true") (const_string "false")))])
; combiner pattern observed for unwind-dw2-fde.c:linear_search_fdes.
(define_insn "*bbit"
[(set (pc)
(if_then_else
(match_operator 3 "equality_comparison_operator"
[(zero_extract:SI (match_operand:SI 1 "register_operand" "q,c")
(const_int 1)
(match_operand:SI 2 "nonmemory_operand" "L,Lc"))
(const_int 0)])
(label_ref (match_operand 0 "" ""))
(pc)))
(clobber (reg:CC_ZN CC_REG))]
"!CROSSING_JUMP_P (insn)"
{
switch (get_attr_length (insn))
{
case 4: return (GET_CODE (operands[3]) == EQ
? \"bbit0%* %1,%2,%0\" : \"bbit1%* %1,%2,%0\");
case 6:
case 8: return \"btst%? %1,%2\n\tb%d3%* %0; bbit out of range\";
default: gcc_unreachable ();
}
}
[(set_attr "type" "brcc")
(set_attr "cond" "clob")
(set (attr "length")
(cond [(and (ge (minus (match_dup 0) (pc)) (const_int -254))
(le (minus (match_dup 0) (pc))
(minus (const_int 248)
(symbol_ref "get_attr_delay_slot_length (insn)"))))
(const_int 4)
(eq (symbol_ref "which_alternative") (const_int 0))
(const_int 6)]
(const_int 8)))
(set (attr "iscompact")
(if_then_else (match_test "get_attr_length (insn) == 6")
(const_string "true") (const_string "false")))])
;; -------------------------------------------------------------------
;; Hardware loop
;; -------------------------------------------------------------------
; operand 0 is the loop count pseudo register
; operand 1 is the label to jump to at the top of the loop
(define_expand "doloop_end"
[(parallel [(set (pc)
(if_then_else
(ne (match_operand 0 "nonimmediate_operand")
(const_int 1))
(label_ref (match_operand 1 "" ""))
(pc)))
(set (match_dup 0) (plus:SI (match_dup 0) (const_int -1)))
(unspec:SI [(const_int 0)] UNSPEC_ARC_LP)
(clobber (match_dup 2))])]
""
{
if (GET_MODE (operands[0]) != SImode)
FAIL;
operands[2] = gen_rtx_SCRATCH (SImode);
})
(define_insn "arc_lp"
[(unspec:SI [(reg:SI LP_COUNT)]
UNSPEC_ARC_LP)
(use (label_ref (match_operand 0 "" "")))
(use (label_ref (match_operand 1 "" "")))]
""
"lp\\t@%l1\\t; lp_count:@%l0->@%l1"
[(set_attr "type" "loop_setup")
(set_attr "length" "4")])
;; if by any chance the lp_count is not used, then use an 'r'
;; register, instead of going to memory.
;; split pattern for the very slim chance when the loop register is
;; memory.
(define_insn_and_split "loop_end"
[(set (pc)
(if_then_else (ne (match_operand:SI 0 "nonimmediate_operand" "+r,!m")
(const_int 1))
(label_ref (match_operand 1 "" ""))
(pc)))
(set (match_dup 0) (plus:SI (match_dup 0) (const_int -1)))
(unspec:SI [(const_int 0)] UNSPEC_ARC_LP)
(clobber (match_scratch:SI 2 "=X,&r"))]
""
"@
; ZOL_END, begins @%l1
#"
"reload_completed && memory_operand (operands[0], Pmode)"
[(set (match_dup 2) (match_dup 0))
(parallel
[(set (reg:CC_ZN CC_REG)
(compare:CC_ZN (plus:SI (match_dup 2) (const_int -1))
(const_int 0)))
(set (match_dup 2) (plus:SI (match_dup 2) (const_int -1)))])
(set (match_dup 0) (match_dup 2))
(set (pc)
(if_then_else (ne (reg:CC_ZN CC_REG)
(const_int 0))
(label_ref (match_dup 1))
(pc)))]
""
[(set_attr "length" "0,24")
(set_attr "predicable" "no")
(set_attr "type" "loop_end")])
(define_insn "loop_fail"
[(set (reg:SI LP_COUNT)
(plus:SI (reg:SI LP_COUNT) (const_int -1)))
(set (reg:CC_ZN CC_REG)
(compare:CC_ZN (plus:SI (reg:SI LP_COUNT) (const_int -1))
(const_int 0)))]
""
"sub.f%?\\tlp_count,lp_count,1"
[(set_attr "iscompact" "false")
(set_attr "type" "compare")
(set_attr "cond" "set_zn")
(set_attr "length" "4")
(set_attr "predicable" "yes")])
(define_insn_and_split "dbnz"
[(set (pc)
(if_then_else
(ne (plus:SI (match_operand:SI 0 "nonimmediate_operand" "+rl,m")
(const_int -1))
(const_int 0))
(label_ref (match_operand 1 "" ""))
(pc)))
(set (match_dup 0)
(plus:SI (match_dup 0)
(const_int -1)))
(clobber (match_scratch:SI 2 "=X,r"))]
"TARGET_DBNZ"
"@
dbnz%#\\t%0,%l1
#"
"TARGET_DBNZ && reload_completed && memory_operand (operands[0], SImode)"
[(set (match_dup 2) (match_dup 0))
(set (match_dup 2) (plus:SI (match_dup 2) (const_int -1)))
(set (reg:CC CC_REG) (compare:CC (match_dup 2) (const_int 0)))
(set (match_dup 0) (match_dup 2))
(set (pc) (if_then_else (ge (reg:CC CC_REG)
(const_int 0))
(label_ref (match_dup 1))
(pc)))]
""
[(set_attr "iscompact" "false")
(set_attr "type" "loop_end")
(set_attr "length" "4,20")])
(define_expand "cpymemsi"
[(match_operand:BLK 0 "" "")
(match_operand:BLK 1 "" "")
(match_operand:SI 2 "nonmemory_operand" "")
(match_operand 3 "immediate_operand" "")]
""
"if (arc_expand_cpymem (operands)) DONE; else FAIL;")
;; Close http://gcc.gnu.org/bugzilla/show_bug.cgi?id=35803 if this works
;; to the point that we can generate cmove instructions.
(define_expand "cbranch4"
[(set (reg:CC CC_REG)
(compare:CC (match_operand:SDF 1 "register_operand" "")
(match_operand:SDF 2 "register_operand" "")))
(set (pc)
(if_then_else
(match_operator 0 "comparison_operator" [(reg CC_REG)
(const_int 0)])
(label_ref (match_operand 3 "" ""))
(pc)))]
"TARGET_FP_SP_BASE || TARGET_OPTFPE"
{
gcc_assert (XEXP (operands[0], 0) == operands[1]);
gcc_assert (XEXP (operands[0], 1) == operands[2]);
operands[0] = gen_compare_reg (operands[0], VOIDmode);
emit_jump_insn (gen_branch_insn (operands[3], operands[0]));
DONE;
})
(define_expand "cmp_float"
[(parallel [(set (match_operand 0 "") (match_operand 1 ""))
(clobber (reg:SI RETURN_ADDR_REGNUM))
(clobber (reg:SI R12_REG))])]
""
"")
(define_mode_iterator OPTFPE_CMP [CC_Z CC_FP_GT CC_FP_GE CC_FP_UNEQ CC_FP_ORD])
(define_mode_attr cmp [(CC_Z "eq") (CC_FP_GT "gt") (CC_FP_GE "ge")
(CC_FP_UNEQ "uneq") (CC_FP_ORD "ord")])
(define_insn "*cmpsf_"
[(set (reg:OPTFPE_CMP CC_REG) (compare:OPTFPE_CMP (reg:SF 0) (reg:SF 1)))
(clobber (reg:SI RETURN_ADDR_REGNUM))
(clobber (reg:SI R12_REG))]
"TARGET_OPTFPE && (!TARGET_ARGONAUT_SET || !TARGET_SPFP)
&& SFUNC_CHECK_PREDICABLE"
"*return arc_output_libcall (\"__sf2\");"
[(set_attr "is_sfunc" "yes")
(set_attr "predicable" "yes")])
;; N.B. for "*cmpdf_ord":
;; double precision fpx sets bit 31 for NaNs. We need bit 51 set
;; for the floating point emulation to recognize the NaN.
(define_insn "*cmpdf_"
[(set (reg:OPTFPE_CMP CC_REG) (compare:OPTFPE_CMP (reg:DF 0) (reg:DF 2)))
(clobber (reg:SI RETURN_ADDR_REGNUM))
(clobber (reg:SI R12_REG))]
"TARGET_OPTFPE && (!TARGET_ARGONAUT_SET || !TARGET_DPFP)
&& SFUNC_CHECK_PREDICABLE"
"*return arc_output_libcall (\"__df2\");"
[(set_attr "is_sfunc" "yes")
(set_attr "predicable" "yes")])
(define_insn "abssf2"
[(set (match_operand:SF 0 "dest_reg_operand" "=q,r,r")
(abs:SF (match_operand:SF 1 "register_operand" "0,0,r")))]
""
"bclr%?\\t%0,%1,31%&"
[(set_attr "type" "unary")
(set_attr "iscompact" "maybe,false,false")
(set_attr "length" "2,4,4")
(set_attr "predicable" "no,yes,no")])
(define_insn "negsf2"
[(set (match_operand:SF 0 "dest_reg_operand" "=r,r")
(neg:SF (match_operand:SF 1 "register_operand" "0,r")))]
""
"bxor%?\\t%0,%1,31"
[(set_attr "type" "unary")
(set_attr "predicable" "yes,no")])
;; ??? Should this use arc_output_libcall and set is_sfunc?
(define_insn "*millicode_thunk_st"
[(match_parallel 0 "millicode_store_operation"
[(set (mem:SI (reg:SI SP_REG)) (reg:SI 13))])]
""
{
output_asm_insn ("bl%* __st_r13_to_%0",
&SET_SRC (XVECEXP (operands[0], 0,
XVECLEN (operands[0], 0) - 2)));
return "";
}
[(set_attr "type" "call")])
(define_insn "*millicode_thunk_ld"
[(match_parallel 0 "millicode_load_clob_operation"
[(set (reg:SI 13) (mem:SI (reg:SI SP_REG)))])]
""
{
output_asm_insn ("bl%* __ld_r13_to_%0",
&SET_DEST (XVECEXP (operands[0], 0,
XVECLEN (operands[0], 0) - 2)));
return "";
}
[(set_attr "type" "call")])
; the sibthunk restores blink, so we use the return rtx.
(define_insn "*millicode_sibthunk_ld"
[(match_parallel 0 "millicode_load_operation"
[(return)
(set (reg:SI SP_REG) (plus:SI (reg:SI SP_REG) (reg:SI 12)))
(set (reg:SI 13) (mem:SI (reg:SI SP_REG)))])]
""
{
output_asm_insn ("b%* __ld_r13_to_%0_ret",
&SET_DEST (XVECEXP (operands[0], 0,
XVECLEN (operands[0], 0) - 1)));
return "";
}
[(set_attr "type" "call")
(set_attr "is_SIBCALL" "yes")])
;; For thread pointer builtins
(define_expand "get_thread_pointersi"
[(set (match_operand:SI 0 "register_operand") (match_dup 1))]
""
"operands[1] = gen_rtx_REG (Pmode, arc_tp_regno);")
(define_expand "set_thread_pointersi"
[(set (match_dup 1) (match_operand:SI 0 "register_operand"))]
""
"operands[1] = gen_rtx_REG (Pmode, arc_tp_regno);")
;; If hardware floating point is available, don't define a negdf pattern;
;; it would be something like:
;;(define_insn "negdf2"
;; [(set (match_operand:DF 0 "register_operand" "=w,w,D,?r")
;; (neg:DF (match_operand:DF 1 "register_operand" "0,c,D,D")))
;; (clobber (match_scratch:DF 2 "=X,X,X,X,D1"))]
;; ""
;; "@
;; bxor%? %H0,%H1,31
;; bxor %H0,%H1,31 ` mov %L0,%L1
;; drsubh%F0%F1 0,0,0
;; drsubh%F2%F1 %H0,0,0 ` dexcl%F2 %L0,%H0,%L0"
;; [(set_attr "type" "unary,unary,dpfp_addsub,dpfp_addsub")
;; (set_attr "iscompact" "false,false,false,false")
;; (set_attr "length" "4,4,8,12")
;; (set_attr "cond" "canuse,nocond,nocond,nocond")])
;; and this suffers from always requiring a long immediate when using
;; the floating point hardware.
;; We then want the sub[sd]f patterns to be used, so that we can load the
;; constant zero efficiently into a register when we want to do the
;; computation using the floating point hardware. There should be a special
;; subdf alternative that matches a zero operand 1, which then can allow
;; to use bxor to flip the high bit of an integer register.
;; ??? we actually can't use the floating point hardware for neg, because
;; this would not work right for -0. OTOH optabs.cc has already code
;; to synthesyze negate by flipping the sign bit.
;;V2 instructions
(define_insn "bswapsi2"
[(set (match_operand:SI 0 "register_operand" "= r,r")
(bswap:SI (match_operand:SI 1 "nonmemory_operand" "rL,Cal")))]
"TARGET_V2 && TARGET_SWAP"
"swape %0, %1"
[(set_attr "length" "4,8")
(set_attr "type" "two_cycle_core")])
(define_expand "prefetch"
[(prefetch (match_operand:SI 0 "address_operand" "")
(match_operand:SI 1 "const_int_operand" "")
(match_operand:SI 2 "const_int_operand" ""))]
"TARGET_HS"
"")
(define_insn "prefetch_1"
[(prefetch (match_operand:SI 0 "register_operand" "r")
(match_operand:SI 1 "const_int_operand" "n")
(match_operand:SI 2 "const_int_operand" "n"))]
"TARGET_HS"
{
if (INTVAL (operands[1]))
return "prefetchw [%0]";
else
return "prefetch [%0]";
}
[(set_attr "type" "load")
(set_attr "length" "4")])
(define_insn "prefetch_2"
[(prefetch (plus:SI (match_operand:SI 0 "register_operand" "r,r,r")
(match_operand:SI 1 "nonmemory_operand" "r,Cm2,Cal"))
(match_operand:SI 2 "const_int_operand" "n,n,n")
(match_operand:SI 3 "const_int_operand" "n,n,n"))]
"TARGET_HS"
{
if (INTVAL (operands[2]))
return "prefetchw [%0, %1]";
else
return "prefetch [%0, %1]";
}
[(set_attr "type" "load")
(set_attr "length" "4,4,8")])
(define_insn "prefetch_3"
[(prefetch (match_operand:SI 0 "address_operand" "p")
(match_operand:SI 1 "const_int_operand" "n")
(match_operand:SI 2 "const_int_operand" "n"))]
"TARGET_HS"
{
operands[0] = gen_rtx_MEM (SImode, operands[0]);
if (INTVAL (operands[1]))
return "prefetchw%U0 %0";
else
return "prefetch%U0 %0";
}
[(set_attr "type" "load")
(set_attr "length" "8")])
(define_insn "divsi3"
[(set (match_operand:SI 0 "register_operand" "=r,r, r,r,r,r, r, r")
(div:SI (match_operand:SI 1 "nonmemory_operand" "0,r,Cal,0,r,0, 0, r")
(match_operand:SI 2 "nonmemory_operand" "r,r, r,L,L,I,Cal,Cal")))]
"TARGET_DIVREM"
"div%? %0, %1, %2"
[(set_attr "length" "4,4,8,4,4,4,8,8")
(set_attr "iscompact" "false")
(set_attr "type" "div_rem")
(set_attr "predicable" "yes,no,no,yes,no,no,yes,no")
(set_attr "cond" "canuse,nocond,nocond,canuse,nocond,nocond,canuse,nocond")
])
(define_insn "udivsi3"
[(set (match_operand:SI 0 "register_operand" "=r,r, r,r,r,r, r, r")
(udiv:SI (match_operand:SI 1 "nonmemory_operand" "0,r,Cal,0,r,0, 0, r")
(match_operand:SI 2 "nonmemory_operand" "r,r, r,L,L,I,Cal,Cal")))]
"TARGET_DIVREM"
"divu%? %0, %1, %2"
[(set_attr "length" "4,4,8,4,4,4,8,8")
(set_attr "iscompact" "false")
(set_attr "type" "div_rem")
(set_attr "predicable" "yes,no,no,yes,no,no,yes,no")
(set_attr "cond" "canuse,nocond,nocond,canuse,nocond,nocond,canuse,nocond")
])
(define_insn "modsi3"
[(set (match_operand:SI 0 "register_operand" "=r,r, r,r,r,r, r, r")
(mod:SI (match_operand:SI 1 "nonmemory_operand" "0,r,Cal,0,r,0, 0, r")
(match_operand:SI 2 "nonmemory_operand" "r,r, r,L,L,I,Cal,Cal")))]
"TARGET_DIVREM"
"rem%? %0, %1, %2"
[(set_attr "length" "4,4,8,4,4,4,8,8")
(set_attr "iscompact" "false")
(set_attr "type" "div_rem")
(set_attr "predicable" "yes,no,no,yes,no,no,yes,no")
(set_attr "cond" "canuse,nocond,nocond,canuse,nocond,nocond,canuse,nocond")
])
(define_insn "umodsi3"
[(set (match_operand:SI 0 "register_operand" "=r,r, r,r,r,r, r, r")
(umod:SI (match_operand:SI 1 "nonmemory_operand" "0,r,Cal,0,r,0, 0, r")
(match_operand:SI 2 "nonmemory_operand" "r,r, r,L,L,I,Cal,Cal")))]
"TARGET_DIVREM"
"remu%? %0, %1, %2"
[(set_attr "length" "4,4,8,4,4,4,8,8")
(set_attr "iscompact" "false")
(set_attr "type" "div_rem")
(set_attr "predicable" "yes,no,no,yes,no,no,yes,no")
(set_attr "cond" "canuse,nocond,nocond,canuse,nocond,nocond,canuse,nocond")
])
;; SETcc instructions
(define_code_iterator arcCC_cond [eq ne gt lt ge le])
(define_insn "arcset"
[(set (match_operand:SI 0 "register_operand" "=r,r,r,r,r,r,r")
(arcCC_cond:SI (match_operand:SI 1 "register_operand" "0,r,0,r,0,0,r")
(match_operand:SI 2 "nonmemory_operand" "r,r,L,L,I,n,n")))]
"TARGET_V2 && TARGET_CODE_DENSITY"
"set%? %0, %1, %2"
[(set_attr "length" "4,4,4,4,4,8,8")
(set_attr "iscompact" "false")
(set_attr "type" "compare")
(set_attr "predicable" "yes,no,yes,no,no,yes,no")
(set_attr "cond" "canuse,nocond,canuse,nocond,nocond,canuse,nocond")
])
(define_insn "arcsetltu"
[(set (match_operand:SI 0 "register_operand" "=r,r,r,r,r, r, r")
(ltu:SI (match_operand:SI 1 "register_operand" "0,r,0,r,0, 0, r")
(match_operand:SI 2 "nonmemory_operand" "r,r,L,L,I, n, n")))]
"TARGET_V2 && TARGET_CODE_DENSITY"
"setlo%? %0, %1, %2"
[(set_attr "length" "4,4,4,4,4,8,8")
(set_attr "iscompact" "false")
(set_attr "type" "compare")
(set_attr "predicable" "yes,no,yes,no,no,yes,no")
(set_attr "cond" "canuse,nocond,canuse,nocond,nocond,canuse,nocond")
])
(define_insn "arcsetgeu"
[(set (match_operand:SI 0 "register_operand" "=r,r,r,r,r, r, r")
(geu:SI (match_operand:SI 1 "register_operand" "0,r,0,r,0, 0, r")
(match_operand:SI 2 "nonmemory_operand" "r,r,L,L,I, n, n")))]
"TARGET_V2 && TARGET_CODE_DENSITY"
"seths%? %0, %1, %2"
[(set_attr "length" "4,4,4,4,4,8,8")
(set_attr "iscompact" "false")
(set_attr "type" "compare")
(set_attr "predicable" "yes,no,yes,no,no,yes,no")
(set_attr "cond" "canuse,nocond,canuse,nocond,nocond,canuse,nocond")
])
;; Special cases of SETCC
(define_insn_and_split "arcsethi"
[(set (match_operand:SI 0 "register_operand" "=r,r, r,r")
(gtu:SI (match_operand:SI 1 "register_operand" "r,r, r,r")
(match_operand:SI 2 "nonmemory_operand" "0,r,C62,n")))]
"TARGET_V2 && TARGET_CODE_DENSITY"
"setlo%? %0, %2, %1"
"reload_completed
&& CONST_INT_P (operands[2])
&& satisfies_constraint_C62 (operands[2])"
[(const_int 0)]
"{
/* sethi a,b,u6 => seths a,b,u6 + 1. */
operands[2] = GEN_INT (INTVAL (operands[2]) + 1);
emit_insn (gen_arcsetgeu (operands[0], operands[1], operands[2]));
DONE;
}"
[(set_attr "length" "4,4,4,8")
(set_attr "iscompact" "false")
(set_attr "type" "compare")
(set_attr "predicable" "yes,no,no,no")
(set_attr "cond" "canuse,nocond,nocond,nocond")]
)
(define_insn_and_split "arcsetls"
[(set (match_operand:SI 0 "register_operand" "=r,r, r,r")
(leu:SI (match_operand:SI 1 "register_operand" "r,r, r,r")
(match_operand:SI 2 "nonmemory_operand" "0,r,C62,n")))]
"TARGET_V2 && TARGET_CODE_DENSITY"
"seths%? %0, %2, %1"
"reload_completed
&& CONST_INT_P (operands[2])
&& satisfies_constraint_C62 (operands[2])"
[(const_int 0)]
"{
/* setls a,b,u6 => setlo a,b,u6 + 1. */
operands[2] = GEN_INT (INTVAL (operands[2]) + 1);
emit_insn (gen_arcsetltu (operands[0], operands[1], operands[2]));
DONE;
}"
[(set_attr "length" "4,4,4,8")
(set_attr "iscompact" "false")
(set_attr "type" "compare")
(set_attr "predicable" "yes,no,no,no")
(set_attr "cond" "canuse,nocond,nocond,nocond")]
)
; Any mode that needs to be solved by secondary reload
(define_mode_iterator SRI [QI HI])
(define_expand "reload__load"
[(parallel [(match_operand:SRI 0 "register_operand" "=r")
(match_operand:SRI 1 "memory_operand" "m")
(match_operand:SI 2 "register_operand" "=&r")])]
""
{
arc_secondary_reload_conv (operands[0], operands[1], operands[2], false);
DONE;
})
(define_expand "reload__store"
[(parallel [(match_operand:SRI 0 "memory_operand" "=m")
(match_operand:SRI 1 "register_operand" "r")
(match_operand:SI 2 "register_operand" "=&r")])]
""
{
arc_secondary_reload_conv (operands[1], operands[0], operands[2], true);
DONE;
})
(define_insn "extzvsi"
[(set (match_operand:SI 0 "register_operand" "=r , r,r,r")
(zero_extract:SI (match_operand:SI 1 "register_operand" "0 , r,r,0")
(match_operand:SI 2 "const_int_operand" "C3p,C3p,n,n")
(match_operand:SI 3 "const_int_operand" "n , n,n,n")))]
"TARGET_HS && TARGET_BARREL_SHIFTER"
{
int assemble_op2 = (((INTVAL (operands[2]) - 1) & 0x1f) << 5) | (INTVAL (operands[3]) & 0x1f);
operands[2] = GEN_INT (assemble_op2);
return "xbfu%?\\t%0,%1,%2";
}
[(set_attr "type" "shift")
(set_attr "iscompact" "false")
(set_attr "length" "4,4,8,8")
(set_attr "predicable" "yes,no,no,yes")
(set_attr "cond" "canuse,nocond,nocond,canuse_limm")])
(define_insn "kflag"
[(unspec_volatile [(match_operand:SI 0 "nonmemory_operand" "rL,I,Cal")]
VUNSPEC_ARC_KFLAG)]
"TARGET_V2"
"@
kflag%? %0
kflag %0
kflag%? %0"
[(set_attr "length" "4,4,8")
(set_attr "type" "misc,misc,misc")
(set_attr "predicable" "yes,no,yes")
(set_attr "cond" "clob,clob,clob")])
(define_insn "clri"
[(set (match_operand:SI 0 "dest_reg_operand" "=r")
(unspec_volatile:SI [(match_operand:SI 1 "immediate_operand" "N")]
VUNSPEC_ARC_CLRI))]
"TARGET_V2"
"clri %0"
[(set_attr "length" "4")
(set_attr "type" "misc")])
(define_insn "ffs"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r")
(unspec:SI [(match_operand:SI 1 "general_operand" "rL,Cal")]
UNSPEC_ARC_FFS))]
"TARGET_NORM && TARGET_V2"
"ffs\\t%0,%1"
[(set_attr "length" "4,8")
(set_attr "type" "two_cycle_core,two_cycle_core")])
(define_insn "ffs_f"
[(set (match_operand:SI 0 "dest_reg_operand" "=r,r")
(unspec:SI [(match_operand:SI 1 "general_operand" "rL,Cal")]
UNSPEC_ARC_FFS))
(set (reg:CC_ZN CC_REG)
(compare:CC_ZN (match_dup 1) (const_int 0)))]
"TARGET_NORM && TARGET_V2"
"ffs.f\\t%0,%1"
[(set_attr "length" "4,8")
(set_attr "type" "two_cycle_core,two_cycle_core")])
(define_expand "ffssi2"
[(parallel [(set (match_dup 2)
(unspec:SI [(match_operand:SI 1 "register_operand" "")]
UNSPEC_ARC_FFS))
(set (reg:CC_ZN CC_REG)
(compare:CC_ZN (match_dup 1) (const_int 0)))])
(set (match_dup 2) (plus:SI (match_dup 2) (const_int 1)))
(set (match_operand:SI 0 "dest_reg_operand" "")
(if_then_else:SI (eq:SI (reg:CC_ZN CC_REG) (const_int 0))
(const_int 0)
(match_dup 2)))]
"TARGET_NORM && TARGET_V2"
{
operands[2] = gen_reg_rtx (SImode);
})
(define_insn "fls"
[(set (match_operand:SI 0 "register_operand" "=r,r")
(unspec:SI [(match_operand:SI 1 "nonmemory_operand" "rL,Cal")]
UNSPEC_ARC_FLS))]
"TARGET_NORM && TARGET_V2"
"fls\\t%0,%1"
[(set_attr "length" "4,8")
(set_attr "type" "two_cycle_core,two_cycle_core")])
(define_insn "seti"
[(unspec_volatile:SI [(match_operand:SI 0 "nonmemory_operand" "rL")]
VUNSPEC_ARC_SETI)]
"TARGET_V2"
"seti\\t%0"
[(set_attr "length" "4")
(set_attr "type" "misc")])
;; FPU/FPX expands
;;add
(define_expand "addsf3"
[(set (match_operand:SF 0 "register_operand" "")
(plus:SF (match_operand:SF 1 "nonmemory_operand" "")
(match_operand:SF 2 "nonmemory_operand" "")))]
"TARGET_FP_SP_BASE || TARGET_SPFP"
"
if (!register_operand (operands[1], SFmode)
&& !register_operand (operands[2], SFmode))
operands[1] = force_reg (SFmode, operands[1]);
")
;;sub
(define_expand "subsf3"
[(set (match_operand:SF 0 "register_operand" "")
(minus:SF (match_operand:SF 1 "nonmemory_operand" "")
(match_operand:SF 2 "nonmemory_operand" "")))]
"TARGET_FP_SP_BASE || TARGET_SPFP"
"
if (!register_operand (operands[1], SFmode)
&& !register_operand (operands[2], SFmode))
operands[1] = force_reg (SFmode, operands[1]);
")
;;mul
(define_expand "mulsf3"
[(set (match_operand:SF 0 "register_operand" "")
(mult:SF (match_operand:SF 1 "nonmemory_operand" "")
(match_operand:SF 2 "nonmemory_operand" "")))]
"TARGET_FP_SP_BASE || TARGET_SPFP"
"
if (!register_operand (operands[1], SFmode)
&& !register_operand (operands[2], SFmode))
operands[1] = force_reg (SFmode, operands[1]);
")
;;add
(define_expand "adddf3"
[(set (match_operand:DF 0 "double_register_operand" "")
(plus:DF (match_operand:DF 1 "double_register_operand" "")
(match_operand:DF 2 "nonmemory_operand" "")))]
"TARGET_FP_DP_BASE || TARGET_DPFP"
"
if (TARGET_DPFP)
{
if (GET_CODE (operands[2]) == CONST_DOUBLE)
{
rtx first, second, tmp;
split_double (operands[2], &first, &second);
tmp = force_reg (SImode, TARGET_BIG_ENDIAN ? first : second);
emit_insn (gen_adddf3_insn (operands[0], operands[1],
operands[2], tmp, const0_rtx));
}
else
emit_insn (gen_adddf3_insn (operands[0], operands[1],
operands[2], const1_rtx, const1_rtx));
DONE;
}
else if (TARGET_FP_DP_BASE)
{
if (!even_register_operand (operands[2], DFmode))
operands[2] = force_reg (DFmode, operands[2]);
if (!even_register_operand (operands[1], DFmode))
operands[1] = force_reg (DFmode, operands[1]);
}
else
gcc_unreachable ();
")
;;sub
(define_expand "subdf3"
[(set (match_operand:DF 0 "double_register_operand" "")
(minus:DF (match_operand:DF 1 "nonmemory_operand" "")
(match_operand:DF 2 "nonmemory_operand" "")))]
"TARGET_FP_DP_BASE || TARGET_DPFP"
"
if (TARGET_DPFP)
{
if (TARGET_FP_DP_AX && (GET_CODE (operands[1]) == CONST_DOUBLE))
operands[1] = force_reg (DFmode, operands[1]);
if ((GET_CODE (operands[1]) == CONST_DOUBLE)
|| GET_CODE (operands[2]) == CONST_DOUBLE)
{
rtx first, second, tmp;
int const_index = ((GET_CODE (operands[1]) == CONST_DOUBLE) ? 1 : 2);
split_double (operands[const_index], &first, &second);
tmp = force_reg (SImode, TARGET_BIG_ENDIAN ? first : second);
emit_insn (gen_subdf3_insn (operands[0], operands[1],
operands[2], tmp, const0_rtx));
}
else
emit_insn (gen_subdf3_insn (operands[0], operands[1],
operands[2], const1_rtx, const1_rtx));
DONE;
}
else if (TARGET_FP_DP_BASE)
{
if (!even_register_operand (operands[2], DFmode))
operands[2] = force_reg (DFmode, operands[2]);
if (!even_register_operand (operands[1], DFmode))
operands[1] = force_reg (DFmode, operands[1]);
}
else
gcc_unreachable ();
")
;;mul
(define_expand "muldf3"
[(set (match_operand:DF 0 "double_register_operand" "")
(mult:DF (match_operand:DF 1 "double_register_operand" "")
(match_operand:DF 2 "nonmemory_operand" "")))]
"TARGET_FP_DP_BASE || TARGET_DPFP"
"
if (TARGET_DPFP)
{
if (GET_CODE (operands[2]) == CONST_DOUBLE)
{
rtx first, second, tmp;
split_double (operands[2], &first, &second);
tmp = force_reg (SImode, TARGET_BIG_ENDIAN ? first : second);
emit_insn (gen_muldf3_insn (operands[0], operands[1],
operands[2], tmp, const0_rtx));
}
else
emit_insn (gen_muldf3_insn (operands[0], operands[1],
operands[2], const1_rtx, const1_rtx));
DONE;
}
else if (TARGET_FP_DP_BASE)
{
if (!even_register_operand (operands[2], DFmode))
operands[2] = force_reg (DFmode, operands[2]);
if (!even_register_operand (operands[1], DFmode))
operands[1] = force_reg (DFmode, operands[1]);
}
else
gcc_unreachable ();
")
;;div
(define_expand "divsf3"
[(set (match_operand:SF 0 "register_operand" "")
(div:SF (match_operand:SF 1 "nonmemory_operand" "")
(match_operand:SF 2 "nonmemory_operand" "")))]
"TARGET_FPX_QUARK || TARGET_FP_SP_SQRT"
"
if (TARGET_FPX_QUARK)
{
operands[1] = force_reg (SFmode, operands[1]);
operands[2] = force_reg (SFmode, operands[2]);
}
else
{
if (!register_operand (operands[1], SFmode)
&& !register_operand (operands[2], SFmode))
operands[1] = force_reg (SFmode, operands[1]);
}
")
;; Square root
(define_expand "sqrtsf2"
[(set (match_operand:SF 0 "register_operand" "")
(sqrt:SF (match_operand:SF 1 "nonmemory_operand" "")))]
"TARGET_FPX_QUARK || TARGET_FP_SP_SQRT"
"
if (TARGET_FPX_QUARK)
{
operands[1] = force_reg (SFmode, operands[1]);
}
")
;; SF->SI (using rounding towards zero)
(define_expand "fix_truncsfsi2"
[(set (match_operand:SI 0 "register_operand" "")
(fix:SI (fix:SF (match_operand:SF 1 "register_operand" ""))))]
"TARGET_FPX_QUARK || TARGET_FP_SP_CONV"
"")
;; SI->SF
(define_expand "floatsisf2"
[(set (match_operand:SF 0 "register_operand" "")
(float:SF (match_operand:SI 1 "register_operand" "")))]
"TARGET_FPX_QUARK || TARGET_FP_SP_CONV"
"")
(define_expand "extzv"
[(set (match_operand:SI 0 "register_operand" "")
(zero_extract:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "const_int_operand" "")
(match_operand:SI 3 "const_int_operand" "")))]
"TARGET_NPS_BITOPS")
; We need a sanity check in the instuction predicate because combine
; will throw any old rubbish at us and see what sticks.
(define_insn "*extzv_i"
[(set (match_operand:SI 0 "register_operand" "=Rrq")
(zero_extract:SI (match_operand:SI 1 "register_operand" "Rrq")
(match_operand:SI 2 "const_int_operand" "n")
(match_operand:SI 3 "const_int_operand" "n")))]
"TARGET_NPS_BITOPS && INTVAL (operands[2]) + INTVAL (operands[3]) <= 32"
"movb.cl %0,%1,0,%3,%2"
[(set_attr "type" "shift")
(set_attr "length" "4")])
(define_expand "insv"
[(set (zero_extract:SI (match_operand:SI 0 "register_operand" "")
(match_operand:SI 1 "const_int_operand" "")
(match_operand:SI 2 "const_int_operand" ""))
(match_operand:SI 3 "nonmemory_operand" ""))]
"TARGET_NPS_BITOPS"
{
int size = INTVAL (operands[1]);
if (size != 1 && size != 2 && size != 4 && size != 8)
operands[3] = force_reg (SImode, operands[3]);
})
(define_insn "*insv_i"
[(set (zero_extract:SI (match_operand:SI 0 "register_operand" "+w,Rrq")
(match_operand:SI 1 "const_int_operand" "C18,n")
(match_operand:SI 2 "const_int_operand" "n,n"))
(match_operand:SI 3 "nonmemory_operand" "P,Rrq"))]
"TARGET_NPS_BITOPS
&& (register_operand (operands[3], SImode)
|| satisfies_constraint_C18 (operands[1]))"
"@
movbi %0,%0,%3,%2,%1
movb %0,%0,%3,%2,0,%1"
[(set_attr "type" "shift")
(set_attr "length" "4")])
(define_insn "*movb"
[(set (zero_extract:SI (match_operand:SI 0 "register_operand" "+Rrq")
(match_operand:SI 1 "const_int_operand" "n")
(match_operand:SI 2 "const_int_operand" "n"))
(zero_extract:SI (match_operand:SI 3 "register_operand" "Rrq")
(match_dup 1)
(match_operand:SI 4 "const_int_operand" "n")))]
"TARGET_NPS_BITOPS"
"movb %0,%0,%3,%2,%4,%1"
[(set_attr "type" "shift")
(set_attr "length" "4")])
(define_insn "*movb_signed"
[(set (zero_extract:SI (match_operand:SI 0 "register_operand" "+Rrq")
(match_operand:SI 1 "const_int_operand" "n")
(match_operand:SI 2 "const_int_operand" "n"))
(sign_extract:SI (match_operand:SI 3 "register_operand" "Rrq")
(match_dup 1)
(match_operand:SI 4 "const_int_operand" "n")))]
"TARGET_NPS_BITOPS"
"movb %0,%0,%3,%2,%4,%1"
[(set_attr "type" "shift")
(set_attr "length" "4")])
(define_insn "*movb_high"
[(set (zero_extract:SI (match_operand:SI 0 "register_operand" "+Rrq")
(match_operand:SI 1 "const_int_operand" "n")
(match_operand:SI 2 "const_int_operand" "n"))
(lshiftrt:SI (match_operand:SI 3 "register_operand" "Rrq")
(match_operand:SI 4 "const_int_operand" "n")))]
"TARGET_NPS_BITOPS
&& INTVAL (operands[4]) + INTVAL (operands[1]) <= 32"
"movb %0,%0,%3,%2,%4,%1"
[(set_attr "type" "shift")
(set_attr "length" "4")])
; N.B.: when processing signed bitfields that fit in the top half of
; a word, gcc will use a narrow sign extending load, and in this case
; we will see INTVAL (operands[4]) + INTVAL (operands[1]) == 16 (or 8)
(define_insn "*movb_high_signed"
[(set (zero_extract:SI (match_operand:SI 0 "register_operand" "+Rrq")
(match_operand:SI 1 "const_int_operand" "n")
(match_operand:SI 2 "const_int_operand" "n"))
(ashiftrt:SI (match_operand:SI 3 "register_operand" "Rrq")
(match_operand:SI 4 "const_int_operand" "n")))]
"TARGET_NPS_BITOPS
&& INTVAL (operands[4]) + INTVAL (operands[1]) <= 32"
"movb %0,%0,%3,%2,%4,%1"
[(set_attr "type" "shift")
(set_attr "length" "4")])
(define_split
[(set (match_operand:SI 0 "register_operand" "")
(ior:SI (ashift:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "const_int_operand" ""))
(subreg:SI (match_operand 3 "") 0)))]
"TARGET_NPS_BITOPS
&& GET_MODE_BITSIZE (GET_MODE (operands[3])) <= INTVAL (operands[2])
&& !reg_overlap_mentioned_p (operands[0], operands[1])"
[(set (match_dup 0) (zero_extend:SI (match_dup 3)))
(set (zero_extract:SI (match_dup 0) (match_dup 4) (match_dup 2))
(match_dup 1))]
"operands[4] = GEN_INT (32 - INTVAL (operands[2]));")
(define_insn "*mrgb"
[(set (zero_extract:SI (match_operand:SI 0 "register_operand" "+Rrq")
(match_operand:SI 1 "const_int_operand" "n")
(match_operand:SI 2 "const_int_operand" "n"))
(zero_extract:SI (match_dup 0) (match_dup 1)
(match_operand:SI 3 "const_int_operand" "n")))
(set (zero_extract:SI (match_dup 0)
(match_operand:SI 4 "const_int_operand" "n")
(match_operand:SI 5 "const_int_operand" "n"))
(zero_extract:SI (match_operand:SI 6 "register_operand" "Rrq")
(match_dup 4)
(match_operand:SI 7 "const_int_operand" "n")))]
"TARGET_NPS_BITOPS"
{
output_asm_insn ("mrgb %0,%0,%6,%2,%3,%1,%5,%7,%4", operands);
/* The ;%? updates the known unalignment. */
return arc_short_long (insn, ";%?", "nop_s");
}
[(set_attr "type" "shift")
(set_attr "length" "6")
(set_attr "iscompact" "true")])
;; combine fumbles combination of two movb patterns, and then the
;; combination is rejected by combinable_i3pat.
;; Thus, we can only use a peephole2 to combine two such insns.
(define_peephole2
[(set (match_operand:SI 0 "register_operand" "")
(match_operand:SI 1 "register_operand" ""))
(set (zero_extract:SI (match_dup 0)
(match_operand:SI 2 "const_int_operand" "")
(match_operand:SI 3 "const_int_operand" ""))
(zero_extract:SI (match_dup 1)
(match_dup 2)
(match_operand:SI 4 "const_int_operand" "")))
(match_operand 9) ; unrelated insn scheduled here
(set (zero_extract:SI (match_dup 0)
(match_operand:SI 5 "const_int_operand" "")
(match_operand:SI 6 "const_int_operand" ""))
(zero_extract:SI (match_operand:SI 7 "register_operand" "")
(match_dup 5)
(match_operand:SI 8 "const_int_operand" "")))]
"TARGET_NPS_BITOPS
// Check that the second movb doesn't clobber an input of the extra insn.
&& !reg_overlap_mentioned_p (operands[0], operands[9])
// And vice versa.
&& !reg_set_p (operands[0], operands[9])
&& !reg_set_p (operands[7], operands[9])"
[(set (match_dup 0) (match_dup 1))
(parallel [(set (zero_extract:SI (match_dup 0) (match_dup 1) (match_dup 2))
(zero_extract:SI (match_dup 3) (match_dup 1) (match_dup 4)))
(set (zero_extract:SI (match_dup 0) (match_dup 1) (match_dup 2))
(zero_extract:SI (match_dup 3) (match_dup 1) (match_dup 4)))])
(match_dup 9)])
(define_peephole2
[(set (match_operand:SI 0 "register_operand" "")
(match_operand:SI 1 "register_operand" ""))
(set (zero_extract:SI (match_dup 0)
(match_operand:SI 2 "const_int_operand" "")
(match_operand:SI 3 "const_int_operand" ""))
(zero_extract:SI (match_dup 1)
(match_dup 2)
(match_operand:SI 4 "const_int_operand" "")))
(set (match_dup 1) (match_operand 8))
(set (zero_extract:SI (match_dup 0)
(match_operand:SI 5 "const_int_operand" "")
(match_operand:SI 6 "const_int_operand" ""))
(zero_extract:SI (match_dup 1) (match_dup 5)
(match_operand:SI 7 "const_int_operand" "")))]
"TARGET_NPS_BITOPS
&& !reg_overlap_mentioned_p (operands[0], operands[8])"
[(set (match_dup 0) (match_dup 1))
(set (match_dup 1) (match_dup 8))
(parallel [(set (zero_extract:SI (match_dup 0) (match_dup 2) (match_dup 3))
(zero_extract:SI (match_dup 0) (match_dup 2) (match_dup 4)))
(set (zero_extract:SI (match_dup 0) (match_dup 5) (match_dup 6))
(zero_extract:SI (match_dup 1) (match_dup 5) (match_dup 7)))])
(match_dup 1)])
(define_insn "*rotrsi3_cnt1"
[(set (match_operand:SI 0 "dest_reg_operand" "=r")
(rotatert:SI (match_operand:SI 1 "nonmemory_operand" "rL")
(const_int 1)))]
""
"ror\\t%0,%1"
[(set_attr "type" "shift")
(set_attr "predicable" "no")
(set_attr "length" "4")])
(define_insn "*rotrsi3_cnt8"
[(set (match_operand:SI 0 "register_operand" "=r")
(rotatert:SI (match_operand:SI 1 "nonmemory_operand" "rL")
(const_int 8)))]
"TARGET_BARREL_SHIFTER && TARGET_V2"
"ror8\\t%0,%1"
[(set_attr "type" "shift")
(set_attr "predicable" "no")
(set_attr "length" "4")])
(define_insn "*ashlsi2_cnt1"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,w")
(ashift:SI (match_operand:SI 1 "register_operand" "q,c")
(const_int 1)))]
""
"asl%? %0,%1%&"
[(set_attr "type" "shift")
(set_attr "iscompact" "maybe,false")
(set_attr "length" "4")
(set_attr "predicable" "no,no")])
(define_insn "*ashlsi2_cnt8"
[(set (match_operand:SI 0 "register_operand" "=r")
(ashift:SI (match_operand:SI 1 "nonmemory_operand" "rL")
(const_int 8)))]
"TARGET_BARREL_SHIFTER && TARGET_V2"
"lsl8\\t%0,%1"
[(set_attr "type" "shift")
(set_attr "iscompact" "false")
(set_attr "length" "4")
(set_attr "predicable" "no")])
(define_insn "*ashlsi2_cnt16"
[(set (match_operand:SI 0 "register_operand" "=r")
(ashift:SI (match_operand:SI 1 "nonmemory_operand" "rL")
(const_int 16)))]
"TARGET_BARREL_SHIFTER && TARGET_V2"
"lsl16\\t%0,%1"
[(set_attr "type" "shift")
(set_attr "iscompact" "false")
(set_attr "length" "4")
(set_attr "predicable" "no")])
(define_insn "*lshrsi3_cnt1"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,w")
(lshiftrt:SI (match_operand:SI 1 "register_operand" "q,c")
(const_int 1)))]
""
"lsr%? %0,%1%&"
[(set_attr "type" "shift")
(set_attr "iscompact" "maybe,false")
(set_attr "predicable" "no,no")])
(define_insn "*ashrsi3_cnt1"
[(set (match_operand:SI 0 "dest_reg_operand" "=q,w")
(ashiftrt:SI (match_operand:SI 1 "register_operand" "q,c")
(const_int 1)))]
""
"asr%? %0,%1%&"
[(set_attr "type" "shift")
(set_attr "iscompact" "maybe,false")
(set_attr "predicable" "no,no")])
(define_peephole2
[(set (match_operand:SI 0 "register_operand" "")
(zero_extract:SI (match_dup 0)
(match_operand:SI 1 "const_int_operand" "")
(match_operand:SI 2 "const_int_operand" "")))
(set (zero_extract:SI (match_operand:SI 3 "register_operand" "")
(match_dup 1)
(match_dup 2))
(match_dup 0))]
"TARGET_NPS_BITOPS
&& !reg_overlap_mentioned_p (operands[0], operands[3])"
[(set (zero_extract:SI (match_dup 3) (match_dup 1) (match_dup 2))
(zero_extract:SI (match_dup 0) (match_dup 1) (match_dup 2)))])
;; Dummy pattern used as a place holder for automatically saved
;; registers.
(define_insn "stack_irq_dwarf"
[(unspec_volatile [(const_int 1)] VUNSPEC_ARC_STACK_IRQ)]
""
""
[(set_attr "length" "0")])
;; MAC and DMPY instructions
; Use VMAC2H(U) instruction to emulate scalar 16bit mac.
(define_expand "maddhisi4"
[(match_operand:SI 0 "register_operand" "")
(match_operand:HI 1 "register_operand" "")
(match_operand:HI 2 "register_operand" "")
(match_operand:SI 3 "register_operand" "")]
"TARGET_PLUS_MACD"
"{
rtx acc_reg = gen_rtx_REG (SImode, ACCL_REGNO);
emit_move_insn (acc_reg, operands[3]);
emit_insn (gen_machi (operands[0], operands[1], operands[2], acc_reg));
DONE;
}")
(define_insn "machi"
[(set (match_operand:SI 0 "register_operand" "=Ral,r")
(plus:SI
(mult:SI (sign_extend:SI (match_operand:HI 1 "register_operand" "%r,r"))
(sign_extend:SI (match_operand:HI 2 "register_operand" "r,r")))
(match_operand:SI 3 "accl_operand" "")))
(clobber (reg:DI ARCV2_ACC))]
"TARGET_PLUS_MACD"
"dmach\\t%0,%1,%2"
[(set_attr "length" "4")
(set_attr "type" "multi")
(set_attr "predicable" "no")
(set_attr "cond" "nocond")])
; The same for the unsigned variant, but using VMAC2HU instruction.
(define_expand "umaddhisi4"
[(match_operand:SI 0 "register_operand" "")
(match_operand:HI 1 "register_operand" "")
(match_operand:HI 2 "register_operand" "")
(match_operand:SI 3 "register_operand" "")]
"TARGET_PLUS_MACD"
"{
rtx acc_reg = gen_rtx_REG (SImode, ACCL_REGNO);
emit_move_insn (acc_reg, operands[3]);
emit_insn (gen_umachi (operands[0], operands[1], operands[2], acc_reg));
DONE;
}")
(define_insn "umachi"
[(set (match_operand:SI 0 "register_operand" "=Ral,r")
(plus:SI
(mult:SI (zero_extend:SI (match_operand:HI 1 "register_operand" "%r,r"))
(zero_extend:SI (match_operand:HI 2 "register_operand" "r,r")))
(match_operand:SI 3 "accl_operand" "")))
(clobber (reg:DI ARCV2_ACC))]
"TARGET_PLUS_MACD"
"dmachu\\t%0,%1,%2"
[(set_attr "length" "4")
(set_attr "type" "multi")
(set_attr "predicable" "no")
(set_attr "cond" "nocond")])
(define_expand "maddsidi4"
[(match_operand:DI 0 "register_operand" "")
(match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "extend_operand" "")
(match_operand:DI 3 "register_operand" "")]
"TARGET_PLUS_DMPY"
"{
emit_insn (gen_maddsidi4_split (operands[0], operands[1], operands[2], operands[3]));
DONE;
}")
(define_insn_and_split "maddsidi4_split"
[(set (match_operand:DI 0 "register_operand" "=r")
(plus:DI
(mult:DI
(sign_extend:DI (match_operand:SI 1 "register_operand" "%r"))
(sign_extend:DI (match_operand:SI 2 "extend_operand" "ri")))
(match_operand:DI 3 "register_operand" "r")))
(clobber (reg:DI ARCV2_ACC))]
"TARGET_PLUS_DMPY"
"#"
"TARGET_PLUS_DMPY && reload_completed"
[(const_int 0)]
"{
rtx acc_reg = gen_rtx_REG (DImode, ACC_REG_FIRST);
emit_move_insn (acc_reg, operands[3]);
if (TARGET_PLUS_MACD && even_register_operand (operands[0], DImode)
&& REGNO (operands[0]) != ACC_REG_FIRST)
emit_insn (gen_macd (operands[0], operands[1], operands[2]));
else
{
emit_insn (gen_mac (operands[1], operands[2]));
if (REGNO (operands[0]) != ACC_REG_FIRST)
emit_move_insn (operands[0], acc_reg);
}
DONE;
}"
[(set_attr "type" "multi")
(set_attr "length" "36")])
(define_insn "macd"
[(set (match_operand:DI 0 "even_register_operand" "=r,r,r")
(plus:DI
(mult:DI
(sign_extend:DI (match_operand:SI 1 "register_operand" "%0,r,r"))
(sign_extend:DI (match_operand:SI 2 "extend_operand" "r,rI,Cal")))
(reg:DI ARCV2_ACC)))
(set (reg:DI ARCV2_ACC)
(plus:DI
(mult:DI (sign_extend:DI (match_dup 1))
(sign_extend:DI (match_dup 2)))
(reg:DI ARCV2_ACC)))]
"TARGET_PLUS_MACD"
"macd %0,%1,%2"
[(set_attr "length" "4,4,8")
(set_attr "type" "multi")
(set_attr "predicable" "yes,no,no")
(set_attr "cond" "canuse,nocond,nocond")])
(define_insn "mac"
[(set (reg:DI ARCV2_ACC)
(plus:DI
(mult:DI (sign_extend:DI (match_operand:SI 0 "register_operand" "%r,r"))
(sign_extend:DI (match_operand:SI 1 "extend_operand" "rI,i")))
(reg:DI ARCV2_ACC)))]
"TARGET_PLUS_DMPY"
"mac 0,%0,%1"
[(set_attr "length" "4,8")
(set_attr "type" "multi")
(set_attr "predicable" "no")
(set_attr "cond" "nocond")])
(define_peephole2
[(set (reg:DI ARCV2_ACC)
(plus:DI
(mult:DI (sign_extend:DI (match_operand:SI 0 "register_operand" ""))
(sign_extend:DI (match_operand:SI 1 "extend_operand" "")))
(reg:DI ARCV2_ACC)))
(set (match_operand:SI 2 "register_operand" "")
(match_operand:SI 3 "accl_operand" ""))]
"TARGET_PLUS_DMPY"
[(const_int 0)]
{
emit_insn (gen_mac_r (operands[2], operands[0], operands[1]));
DONE;
})
(define_insn "mac_r"
[(set (match_operand:SI 0 "register_operand" "=r,r")
(truncate:SI
(plus:DI
(mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "%r,r"))
(sign_extend:DI (match_operand:SI 2 "extend_operand" "rI,i")))
(reg:DI ARCV2_ACC))))
(clobber (reg:DI ARCV2_ACC))]
"TARGET_PLUS_DMPY"
"mac %0,%1,%2"
[(set_attr "length" "4,8")
(set_attr "type" "multi")
(set_attr "predicable" "no")
(set_attr "cond" "nocond")])
(define_expand "umaddsidi4"
[(match_operand:DI 0 "register_operand" "")
(match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "extend_operand" "")
(match_operand:DI 3 "register_operand" "")]
"TARGET_PLUS_DMPY"
"{
emit_insn (gen_umaddsidi4_split (operands[0], operands[1], operands[2], operands[3]));
DONE;
}")
(define_insn_and_split "umaddsidi4_split"
[(set (match_operand:DI 0 "register_operand" "=r")
(plus:DI
(mult:DI
(zero_extend:DI (match_operand:SI 1 "register_operand" "%r"))
(zero_extend:DI (match_operand:SI 2 "extend_operand" "ri")))
(match_operand:DI 3 "register_operand" "r")))
(clobber (reg:DI ARCV2_ACC))]
"TARGET_PLUS_DMPY"
"#"
"TARGET_PLUS_DMPY && reload_completed"
[(const_int 0)]
"{
rtx acc_reg = gen_rtx_REG (DImode, ACC_REG_FIRST);
emit_move_insn (acc_reg, operands[3]);
if (TARGET_PLUS_MACD && even_register_operand (operands[0], DImode)
&& REGNO (operands[0]) != ACC_REG_FIRST)
emit_insn (gen_macdu (operands[0], operands[1], operands[2]));
else
{
emit_insn (gen_macu (operands[1], operands[2]));
if (REGNO (operands[0]) != ACC_REG_FIRST)
emit_move_insn (operands[0], acc_reg);
}
DONE;
}"
[(set_attr "type" "multi")
(set_attr "length" "36")])
(define_insn "macdu"
[(set (match_operand:DI 0 "even_register_operand" "=r,r,r")
(plus:DI
(mult:DI
(zero_extend:DI (match_operand:SI 1 "register_operand" "%0,r,r"))
(zero_extend:DI (match_operand:SI 2 "extend_operand" "r,rI,i")))
(reg:DI ARCV2_ACC)))
(set (reg:DI ARCV2_ACC)
(plus:DI
(mult:DI (zero_extend:DI (match_dup 1))
(zero_extend:DI (match_dup 2)))
(reg:DI ARCV2_ACC)))]
"TARGET_PLUS_MACD"
"macdu %0,%1,%2"
[(set_attr "length" "4,4,8")
(set_attr "type" "multi")
(set_attr "predicable" "yes,no,no")
(set_attr "cond" "canuse,nocond,nocond")])
(define_insn "macu"
[(set (reg:DI ARCV2_ACC)
(plus:DI
(mult:DI (zero_extend:DI (match_operand:SI 0 "register_operand" "%r,r"))
(zero_extend:DI (match_operand:SI 1 "extend_operand" "rI,i")))
(reg:DI ARCV2_ACC)))]
"TARGET_PLUS_DMPY"
"macu 0,%0,%1"
[(set_attr "length" "4,8")
(set_attr "type" "multi")
(set_attr "predicable" "no")
(set_attr "cond" "nocond")])
(define_peephole2
[(set (reg:DI ARCV2_ACC)
(plus:DI
(mult:DI (zero_extend:DI (match_operand:SI 0 "register_operand" ""))
(zero_extend:DI (match_operand:SI 1 "extend_operand" "")))
(reg:DI ARCV2_ACC)))
(set (match_operand:SI 2 "register_operand" "")
(match_operand:SI 3 "accl_operand" ""))]
"TARGET_PLUS_DMPY"
[(const_int 0)]
{
emit_insn (gen_macu_r (operands[2], operands[0], operands[1]));
DONE;
})
(define_insn "macu_r"
[(set (match_operand:SI 0 "register_operand" "=r,r")
(truncate:SI
(plus:DI
(mult:DI (zero_extend:DI (match_operand:SI 1 "register_operand" "%r,r"))
(zero_extend:DI (match_operand:SI 2 "extend_operand" "rI,i")))
(reg:DI ARCV2_ACC))))
(clobber (reg:DI ARCV2_ACC))]
"TARGET_PLUS_DMPY"
"macu %0,%1,%2"
[(set_attr "length" "4,8")
(set_attr "type" "multi")
(set_attr "predicable" "no")
(set_attr "cond" "nocond")])
(define_insn "mpyd_arcv2hs"
[(set (match_operand:DI 0 "even_register_operand" "=r")
(mult:DI (SEZ:DI (match_operand:SI 1 "register_operand" "r"))
(SEZ:DI (match_operand:SI 2 "register_operand" "r"))))
(set (reg:DI ARCV2_ACC)
(mult:DI
(SEZ:DI (match_dup 1))
(SEZ:DI (match_dup 2))))]
"TARGET_PLUS_MACD"
"mpyd%?\\t%0,%1,%2"
[(set_attr "length" "4")
(set_attr "iscompact" "false")
(set_attr "type" "multi")
(set_attr "predicable" "no")])
(define_insn "*pmpyd_arcv2hs"
[(set (match_operand:DI 0 "even_register_operand" "=r")
(mult:DI
(SEZ:DI (match_operand:SI 1 "even_register_operand" "%0"))
(SEZ:DI (match_operand:SI 2 "register_operand" "r"))))
(set (reg:DI ARCV2_ACC)
(mult:DI
(SEZ:DI (match_dup 1))
(SEZ:DI (match_dup 2))))]
"TARGET_PLUS_MACD"
"mpyd%?\\t%0,%1,%2"
[(set_attr "length" "4")
(set_attr "iscompact" "false")
(set_attr "type" "multi")
(set_attr "predicable" "yes")])
(define_insn "mpyd_imm_arcv2hs"
[(set (match_operand:DI 0 "even_register_operand" "=r,r, r")
(mult:DI (SEZ:DI (match_operand:SI 1 "register_operand" "r,0, r"))
(match_operand 2 "immediate_operand" "L,I,Cal")))
(set (reg:DI ARCV2_ACC)
(mult:DI (SEZ:DI (match_dup 1))
(match_dup 2)))]
"TARGET_PLUS_MACD"
"mpyd%?\\t%0,%1,%2"
[(set_attr "length" "4,4,8")
(set_attr "iscompact" "false")
(set_attr "type" "multi")
(set_attr "predicable" "no")])
(define_insn "*pmpyd_imm_arcv2hs"
[(set (match_operand:DI 0 "even_register_operand" "=r,r")
(mult:DI
(SEZ:DI (match_operand:SI 1 "even_register_operand" "0,0"))
(match_operand 2 "immediate_operand" "L,Cal")))
(set (reg:DI ARCV2_ACC)
(mult:DI (SEZ:DI (match_dup 1))
(match_dup 2)))]
"TARGET_PLUS_MACD"
"mpyd%?\\t%0,%1,%2"
[(set_attr "length" "4,8")
(set_attr "iscompact" "false")
(set_attr "type" "multi")
(set_attr "predicable" "yes")])
(define_insn "*add_shift"
[(set (match_operand:SI 0 "register_operand" "=q,r,r")
(plus:SI (ashift:SI (match_operand:SI 1 "register_operand" "q,r,r")
(match_operand:SI 2 "_1_2_3_operand" ""))
(match_operand:SI 3 "arc_nonmemory_operand" "0,r,Csz")))]
""
"add%2%?\\t%0,%3,%1"
[(set_attr "length" "*,4,8")
(set_attr "predicable" "yes,no,no")
(set_attr "iscompact" "maybe,false,false")
(set_attr "cond" "canuse,nocond,nocond")])
(define_insn "*add_shift2"
[(set (match_operand:SI 0 "register_operand" "=q,r,r")
(plus:SI (match_operand:SI 1 "nonmemory_operand" "0,r,Cal")
(ashift:SI (match_operand:SI 2 "register_operand" "q,r,r")
(match_operand:SI 3 "_1_2_3_operand" ""))))]
""
"add%3%?\\t%0,%1,%2"
[(set_attr "length" "*,4,8")
(set_attr "predicable" "yes,no,no")
(set_attr "iscompact" "maybe,false,false")
(set_attr "cond" "canuse,nocond,nocond")])
(define_insn "*sub_shift"
[(set (match_operand:SI 0"register_operand" "=r,r,r")
(minus:SI (match_operand:SI 1 "nonmemory_operand" "0,r,Cal")
(ashift:SI (match_operand:SI 2 "register_operand" "r,r,r")
(match_operand:SI 3 "_1_2_3_operand" ""))))]
""
"sub%3\\t%0,%1,%2"
[(set_attr "length" "4,4,8")
(set_attr "cond" "canuse,nocond,nocond")
(set_attr "predicable" "yes,no,no")])
(define_insn "*sub_shift_cmp0_noout"
[(set (match_operand 0 "cc_set_register" "")
(compare:CC
(minus:SI (match_operand:SI 1 "register_operand" "r")
(ashift:SI (match_operand:SI 2 "register_operand" "r")
(match_operand:SI 3 "_1_2_3_operand" "")))
(const_int 0)))]
""
"sub%3.f\\t0,%1,%2"
[(set_attr "length" "4")])
(define_insn "*compare_si_ashiftsi"
[(set (match_operand 0 "cc_set_register" "")
(compare:CC (match_operand:SI 1 "register_operand" "r")
(ashift:SI (match_operand:SI 2 "register_operand" "r")
(match_operand:SI 3 "_1_2_3_operand" ""))))]
""
"sub%3.f\\t0,%1,%2"
[(set_attr "length" "4")])
;; Convert the sequence
;; asl rd,rn,_1_2_3
;; cmp ra,rd
;; into
;; sub{123}.f 0,ra,rn
(define_peephole2
[(set (match_operand:SI 0 "register_operand" "")
(ashift:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "_1_2_3_operand" "")))
(set (reg:CC CC_REG)
(compare:CC (match_operand:SI 3 "register_operand" "")
(match_dup 0)))]
"peep2_reg_dead_p (2, operands[0])"
[(set (reg:CC CC_REG) (compare:CC (match_dup 3)
(ashift:SI (match_dup 1) (match_dup 2))))])
(define_peephole2 ; std
[(set (match_operand:SI 2 "memory_operand" "")
(match_operand:SI 0 "register_operand" ""))
(set (match_operand:SI 3 "memory_operand" "")
(match_operand:SI 1 "register_operand" ""))]
"TARGET_LL64"
[(const_int 0)]
{
if (!gen_operands_ldd_std (operands, false, false))
FAIL;
operands[0] = gen_rtx_REG (DImode, REGNO (operands[0]));
operands[2] = adjust_address (operands[2], DImode, 0);
emit_insn (gen_rtx_SET (operands[2], operands[0]));
DONE;
})
(define_peephole2 ; ldd
[(set (match_operand:SI 0 "register_operand" "")
(match_operand:SI 2 "memory_operand" ""))
(set (match_operand:SI 1 "register_operand" "")
(match_operand:SI 3 "memory_operand" ""))]
"TARGET_LL64"
[(const_int 0)]
{
if (!gen_operands_ldd_std (operands, true, false))
FAIL;
operands[0] = gen_rtx_REG (DImode, REGNO (operands[0]));
operands[2] = adjust_address (operands[2], DImode, 0);
emit_insn (gen_rtx_SET (operands[0], operands[2]));
DONE;
})
;; We require consecutive registers for LDD instruction. Check if we
;; can reorder them and use an LDD.
(define_peephole2 ; swap the destination registers of two loads
; before a commutative operation.
[(set (match_operand:SI 0 "register_operand" "")
(match_operand:SI 2 "memory_operand" ""))
(set (match_operand:SI 1 "register_operand" "")
(match_operand:SI 3 "memory_operand" ""))
(set (match_operand:SI 4 "register_operand" "")
(match_operator:SI 5 "commutative_operator"
[(match_operand 6 "register_operand" "")
(match_operand 7 "register_operand" "") ]))]
"TARGET_LL64
&& (((rtx_equal_p (operands[0], operands[6]))
&& (rtx_equal_p (operands[1], operands[7])))
|| ((rtx_equal_p (operands[0], operands[7]))
&& (rtx_equal_p (operands[1], operands[6]))))
&& (peep2_reg_dead_p (3, operands[0])
|| rtx_equal_p (operands[0], operands[4]))
&& (peep2_reg_dead_p (3, operands[1])
|| rtx_equal_p (operands[1], operands[4]))"
[(set (match_dup 0) (match_dup 2))
(set (match_dup 4) (match_op_dup 5 [(match_dup 6) (match_dup 7)]))]
{
if (!gen_operands_ldd_std (operands, true, true))
{
FAIL;
}
else
{
operands[0] = gen_rtx_REG (DImode, REGNO (operands[0]));
operands[2] = adjust_address (operands[2], DImode, 0);
}
}
)
(define_insn "*push_multi_fp"
[(match_parallel 0 "push_multi_operand"
[(set (reg:SI SP_REG)
(plus:SI (reg:SI SP_REG)
(match_operand 1 "immediate_operand" "")))
(set (mem:SI (plus:SI (reg:SI SP_REG)
(match_operand 2 "immediate_operand"
"")))
(reg:SI 13))])]
"TARGET_CODE_DENSITY"
{
int len = XVECLEN (operands[0], 0);
rtx tmp = XVECEXP (operands[0], 0, len - 1);
if (MEM_P (XEXP (tmp, 0)))
{
operands[3] = XEXP (tmp, 1);
return "enter_s\\t{r13-%3} ; sp=sp+(%1)";
}
else
{
tmp = XVECEXP (operands[0], 0, len - 3);
operands[3] = XEXP (tmp, 1);
return "enter_s\\t{r13-%3, fp} ; sp=sp+(%1)";
}
}
[(set_attr "type" "call_no_delay_slot")
(set_attr "length" "2")])
(define_insn "*push_multi_fp_blink"
[(match_parallel 0 "push_multi_operand"
[(set (reg:SI SP_REG)
(plus:SI (reg:SI SP_REG)
(match_operand 1 "immediate_operand" "")))
(set (mem:SI (plus:SI (reg:SI SP_REG)
(match_operand 2 "immediate_operand"
"")))
(reg:SI RETURN_ADDR_REGNUM))])]
"TARGET_CODE_DENSITY"
{
int len = XVECLEN (operands[0], 0);
rtx tmp = XVECEXP (operands[0], 0, len - 1);
if (MEM_P (XEXP (tmp, 0)))
{
operands[3] = XEXP (tmp, 1);
return "enter_s\\t{r13-%3, blink} ; sp=sp+(%1)";
}
else
{
tmp = XVECEXP (operands[0], 0, len - 3);
operands[3] = XEXP (tmp, 1);
return "enter_s\\t{r13-%3, fp, blink} ; sp=sp+(%1)";
}
}
[(set_attr "type" "call_no_delay_slot")
(set_attr "length" "2")])
(define_insn "*pop_multi_fp"
[(match_parallel 0 "pop_multi_operand"
[(set (reg:SI SP_REG)
(plus:SI (reg:SI SP_REG)
(match_operand 1 "immediate_operand" "")))
(set (reg:SI 13)
(mem:SI
(plus:SI
(reg:SI SP_REG)
(match_operand 2 "immediate_operand" ""))))])]
"TARGET_CODE_DENSITY"
{
int len = XVECLEN (operands[0], 0);
rtx tmp = XVECEXP (operands[0], 0, len - 1);
if (XEXP (tmp, 0) != hard_frame_pointer_rtx)
{
operands[3] = XEXP (tmp, 0);
gcc_assert (INTVAL (operands[1]) == INTVAL (operands[2]));
return "leave_s\\t{r13-%3} ; sp=sp+%1";
}
else
{
tmp = XVECEXP (operands[0], 0, len - 2);
operands[3] = XEXP (tmp, 0);
return "leave_s\\t{r13-%3, fp} ; sp=sp+%1";
}
}
[(set_attr "type" "call_no_delay_slot")
(set_attr "length" "2")])
(define_insn "*pop_multi_fp_blink"
[(match_parallel 0 "pop_multi_operand"
[(set (reg:SI SP_REG)
(plus:SI (reg:SI SP_REG)
(match_operand 1 "immediate_operand" "")))
(set (reg:SI RETURN_ADDR_REGNUM)
(mem:SI
(plus:SI
(reg:SI SP_REG)
(match_operand 2 "immediate_operand" ""))))])]
"TARGET_CODE_DENSITY"
{
int len = XVECLEN (operands[0], 0);
rtx tmp = XVECEXP (operands[0], 0, len - 1);
if (XEXP (tmp, 0) != hard_frame_pointer_rtx)
{
operands[3] = XEXP (tmp, 0);
gcc_assert (INTVAL (operands[1]) == INTVAL (operands[2]));
return "leave_s\\t{r13-%3, blink} ; sp=sp+%1";
}
else
{
tmp = XVECEXP (operands[0], 0, len - 2);
operands[3] = XEXP (tmp, 0);
return "leave_s\\t{r13-%3, fp, blink} ; sp=sp+%1";
}
}
[(set_attr "type" "call_no_delay_slot")
(set_attr "length" "2")])
(define_insn "*pop_multi_fp_ret"
[(match_parallel 0 "pop_multi_operand"
[(return)
(set (reg:SI SP_REG)
(plus:SI (reg:SI SP_REG)
(match_operand 1 "immediate_operand" "")))
(set (reg:SI 13)
(mem:SI
(plus:SI
(reg:SI SP_REG)
(match_operand 2 "immediate_operand" ""))))])]
"TARGET_CODE_DENSITY"
{
int len = XVECLEN (operands[0], 0);
rtx tmp = XVECEXP (operands[0], 0, len - 1);
if (XEXP (tmp, 0) != hard_frame_pointer_rtx)
{
operands[3] = XEXP (tmp, 0);
gcc_assert (INTVAL (operands[1]) == INTVAL (operands[2]));
return "leave_s\\t{r13-%3, pcl} ; sp=sp+%1";
}
else
{
tmp = XVECEXP (operands[0], 0, len - 2);
operands[3] = XEXP (tmp, 0);
return "leave_s\\t{r13-%3, fp, pcl} ; sp=sp+%1";
}
}
[(set_attr "type" "call_no_delay_slot")
(set_attr "length" "2")])
(define_insn "*pop_multi_fp_blink_ret"
[(match_parallel 0 "pop_multi_operand"
[(return)
(set (reg:SI SP_REG)
(plus:SI (reg:SI SP_REG)
(match_operand 1 "immediate_operand" "")))
(set (reg:SI RETURN_ADDR_REGNUM)
(mem:SI
(plus:SI
(reg:SI SP_REG)
(match_operand 2 "immediate_operand" ""))))])]
"TARGET_CODE_DENSITY"
{
int len = XVECLEN (operands[0], 0);
rtx tmp = XVECEXP (operands[0], 0, len - 1);
if (XEXP (tmp, 0) != hard_frame_pointer_rtx)
{
operands[3] = XEXP (tmp, 0);
gcc_assert (INTVAL (operands[1]) == INTVAL (operands[2]));
return "leave_s\\t{r13-%3, blink, pcl} ; sp=sp+%1";
}
else
{
tmp = XVECEXP (operands[0], 0, len - 2);
operands[3] = XEXP (tmp, 0);
return "leave_s\\t{r13-%3, fp, blink, pcl} ; sp=sp+%1";
}
}
[(set_attr "type" "call_no_delay_slot")
(set_attr "length" "2")])
;; Patterns for exception handling
(define_insn_and_split "eh_return"
[(unspec_volatile [(match_operand:SI 0 "register_operand" "r")]
VUNSPEC_ARC_EH_RETURN)]
""
"#"
"reload_completed"
[(const_int 0)]
"
{
arc_eh_return_address_location (operands[0]);
DONE;
}"
[(set_attr "length" "8")]
)
;; include the arc-FPX instructions
(include "fpx.md")
;; include the arc-FPU instructions
(include "fpu.md")
(include "simdext.md")
;; include atomic extensions
(include "atomic.md")