Propagate uniform execution predicates to all Ventus sALU operations.
This commit is contained in:
parent
2a123a636c
commit
5fe0ee73c4
|
@ -340,11 +340,11 @@ def PseudoFROUND_D : PseudoFROUND<FPR64>;
|
|||
|
||||
/// Loads
|
||||
|
||||
defm : LdPat<load, FLD, f64>;
|
||||
defm : UniformLdPat<load, FLD, f64>;
|
||||
|
||||
/// Stores
|
||||
|
||||
defm : StPat<store, FSD, FPR64, f64>;
|
||||
defm : UniformStPat<store, FSD, FPR64, f64>;
|
||||
|
||||
/// Pseudo-instructions needed for the soft-float ABI with RV32D
|
||||
|
||||
|
|
|
@ -587,11 +587,11 @@ def PseudoFROUND_S : PseudoFROUND<FPR32>;
|
|||
|
||||
/// Loads
|
||||
|
||||
defm : LdPat<load, FLW, f32>;
|
||||
defm : UniformLdPat<load, FLW, f32>;
|
||||
|
||||
/// Stores
|
||||
|
||||
defm : StPat<store, FSW, FPR32, f32>;
|
||||
defm : UniformStPat<store, FSW, FPR32, f32>;
|
||||
|
||||
} // Predicates = [HasStdExtF]
|
||||
|
||||
|
|
|
@ -525,8 +525,8 @@ def : Pat<(xor GPR:$rs1, (not GPR:$rs2)), (XNOR GPR:$rs1, GPR:$rs2)>;
|
|||
} // Predicates = [HasStdExtZbbOrZbkb]
|
||||
|
||||
let Predicates = [HasStdExtZbbOrZbkb] in {
|
||||
def : PatGprGpr<shiftop<rotl>, ROL>;
|
||||
def : PatGprGpr<shiftop<rotr>, ROR>;
|
||||
def : PatGprGpr<UniformShiftFrag<rotl>, ROL>;
|
||||
def : PatGprGpr<UniformShiftFrag<rotr>, ROR>;
|
||||
|
||||
def : PatGprImm<rotr, RORI, uimmlog2xlen>;
|
||||
// There's no encoding for roli in the the 'B' extension as it can be
|
||||
|
@ -536,25 +536,25 @@ def : Pat<(rotl GPR:$rs1, uimmlog2xlen:$shamt),
|
|||
} // Predicates = [HasStdExtZbbOrZbkb]
|
||||
|
||||
let Predicates = [HasStdExtZbbOrZbkb, IsRV64] in {
|
||||
def : PatGprGpr<shiftopw<riscv_rolw>, ROLW>;
|
||||
def : PatGprGpr<shiftopw<riscv_rorw>, RORW>;
|
||||
def : PatGprGpr<UniformShiftWFrag<riscv_rolw>, ROLW>;
|
||||
def : PatGprGpr<UniformShiftWFrag<riscv_rorw>, RORW>;
|
||||
def : PatGprImm<riscv_rorw, RORIW, uimm5>;
|
||||
def : Pat<(riscv_rolw GPR:$rs1, uimm5:$rs2),
|
||||
(RORIW GPR:$rs1, (ImmSubFrom32 uimm5:$rs2))>;
|
||||
} // Predicates = [HasStdExtZbbOrZbkb, IsRV64]
|
||||
|
||||
let Predicates = [HasStdExtZbs] in {
|
||||
def : Pat<(and (not (shiftop<shl> 1, GPR:$rs2)), GPR:$rs1),
|
||||
def : Pat<(and (not (UniformShiftFrag<shl> 1, GPR:$rs2)), GPR:$rs1),
|
||||
(BCLR GPR:$rs1, GPR:$rs2)>;
|
||||
def : Pat<(and (rotl -2, GPR:$rs2), GPR:$rs1), (BCLR GPR:$rs1, GPR:$rs2)>;
|
||||
def : Pat<(or (shiftop<shl> 1, GPR:$rs2), GPR:$rs1),
|
||||
def : Pat<(or (UniformShiftFrag<shl> 1, GPR:$rs2), GPR:$rs1),
|
||||
(BSET GPR:$rs1, GPR:$rs2)>;
|
||||
def : Pat<(xor (shiftop<shl> 1, GPR:$rs2), GPR:$rs1),
|
||||
def : Pat<(xor (UniformShiftFrag<shl> 1, GPR:$rs2), GPR:$rs1),
|
||||
(BINV GPR:$rs1, GPR:$rs2)>;
|
||||
def : Pat<(and (shiftop<srl> GPR:$rs1, GPR:$rs2), 1),
|
||||
def : Pat<(and (UniformShiftFrag<srl> GPR:$rs1, GPR:$rs2), 1),
|
||||
(BEXT GPR:$rs1, GPR:$rs2)>;
|
||||
|
||||
def : Pat<(shiftop<shl> 1, GPR:$rs2),
|
||||
def : Pat<(UniformShiftFrag<shl> 1, GPR:$rs2),
|
||||
(BSET X0, GPR:$rs2)>;
|
||||
|
||||
def : Pat<(and GPR:$rs1, BCLRMask:$mask),
|
||||
|
@ -659,7 +659,7 @@ let Predicates = [HasStdExtZbkb, IsRV64] in {
|
|||
def : Pat<(i64 (or (zexti32 GPR:$rs1), (shl GPR:$rs2, (i64 32)))),
|
||||
(PACK GPR:$rs1, GPR:$rs2)>;
|
||||
|
||||
def : Pat<(binop_allwusers<or> (shl GPR:$rs2, (i64 16)),
|
||||
def : Pat<(UniformWBinFrag<or> (shl GPR:$rs2, (i64 16)),
|
||||
(zexti16 GPR:$rs1)),
|
||||
(PACKW GPR:$rs1, GPR:$rs2)>;
|
||||
def : Pat<(i64 (or (sext_inreg (shl GPR:$rs2, (i64 16)), i32),
|
||||
|
|
|
@ -344,11 +344,11 @@ def PseudoFROUND_H : PseudoFROUND<FPR16>;
|
|||
let Predicates = [HasStdExtZfhOrZfhmin] in {
|
||||
/// Loads
|
||||
|
||||
defm : LdPat<load, FLH, f16>;
|
||||
defm : UniformLdPat<load, FLH, f16>;
|
||||
|
||||
/// Stores
|
||||
|
||||
defm : StPat<store, FSH, FPR16, f16>;
|
||||
defm : UniformStPat<store, FSH, FPR16, f16>;
|
||||
|
||||
/// Float conversion operations
|
||||
|
||||
|
|
|
@ -48,7 +48,7 @@ class UniformTernaryFrag<SDPatternOperator Op> : PatFrag<
|
|||
[{ return !N->isDivergent(); }]>;
|
||||
|
||||
class UniformTetradFrag<SDPatternOperator Op> : PatFrag<
|
||||
(ops node:$src0, node:$src1, node:$src2, node:$sec3),
|
||||
(ops node:$src0, node:$src1, node:$src2, node:$src3),
|
||||
(Op $src0, $src1, $src2, $src3),
|
||||
[{ return !N->isDivergent(); }]>;
|
||||
|
||||
|
@ -1418,18 +1418,6 @@ multiclass SelectCC_GPR_rrirr<RegisterClass valty> {
|
|||
let Predicates = [NoShortForwardBranchOpt] in
|
||||
defm Select_GPR : SelectCC_GPR_rrirr<GPR>;
|
||||
|
||||
class SelectCompressOpt<CondCode Cond>: Pat<(riscv_selectcc_frag:$select GPR:$lhs, simm12_no6:$Constant, Cond,
|
||||
GPR:$truev, GPR:$falsev),
|
||||
(Select_GPR_Using_CC_GPR (ADDI GPR:$lhs, (NegImm simm12:$Constant)), X0,
|
||||
(IntCCtoRISCVCC $select), GPR:$truev, GPR:$falsev)>;
|
||||
|
||||
def OptForMinSize : Predicate<"MF ? MF->getFunction().hasMinSize() : false">;
|
||||
|
||||
let Predicates = [HasStdExtC, OptForMinSize] in {
|
||||
def : SelectCompressOpt<SETEQ>;
|
||||
def : SelectCompressOpt<SETNE>;
|
||||
}
|
||||
|
||||
/// Branches and jumps
|
||||
|
||||
// Match `riscv_brcc` and lower to the appropriate RISC-V branch instruction.
|
||||
|
@ -1441,10 +1429,6 @@ multiclass UniformBccPat<CondCode Cond, RVInstB Inst> {
|
|||
(Inst GPR:$rs1, X0, simm13_lsb0:$imm12)>;
|
||||
}
|
||||
|
||||
class UniBrccCompressOpt<CondCode Cond, RVInstB Inst> : Pat<
|
||||
(UniformTetradFrag<riscv_brcc> GPR:$lhs, simm12_no6:$Constant, Cond, bb:$place),
|
||||
(Inst (ADDI GPR:$lhs, (NegImm simm12:$Constant)), X0, bb:$place)>;
|
||||
|
||||
defm : UniformBccPat<SETEQ, BEQ>;
|
||||
defm : UniformBccPat<SETNE, BNE>;
|
||||
defm : UniformBccPat<SETLT, BLT>;
|
||||
|
@ -1452,11 +1436,6 @@ defm : UniformBccPat<SETGE, BGE>;
|
|||
defm : UniformBccPat<SETULT, BLTU>;
|
||||
defm : UniformBccPat<SETUGE, BGEU>;
|
||||
|
||||
let Predicates = [HasStdExtC, OptForMinSize] in {
|
||||
def : UniBrccCompressOpt<SETEQ, BEQ>;
|
||||
def : UniBrccCompressOpt<SETNE, BNE>;
|
||||
}
|
||||
|
||||
let isBarrier = 1, isBranch = 1, isTerminator = 1 in
|
||||
def PseudoBR : Pseudo<(outs), (ins simm21_lsb0_jal:$imm20), [(br bb:$imm20)]>,
|
||||
PseudoInstExpansion<(JAL X0, simm21_lsb0_jal:$imm20)>;
|
||||
|
@ -1590,7 +1569,7 @@ def PseudoZEXT_W : Pseudo<(outs GPR:$rd), (ins GPR:$rs), [], "zext.w", "$rd, $rs
|
|||
/// Loads
|
||||
|
||||
multiclass UniformLdPat<PatFrag LoadOp, RVInst Inst, ValueType vt = XLenVT> {
|
||||
def : Pat<(vt (UniformBinFrag<LoadOp> (AddrRegImm GPR:$rs1, simm12:$imm12))),
|
||||
def : Pat<(vt (UniformUnaryFrag<LoadOp> (AddrRegImm GPR:$rs1, simm12:$imm12))),
|
||||
(Inst GPR:$rs1, simm12:$imm12)>;
|
||||
}
|
||||
|
||||
|
@ -1605,7 +1584,7 @@ defm : UniformLdPat<zextloadi16, LHU>;
|
|||
/// Stores
|
||||
|
||||
multiclass UniformStPat<PatFrag StoreOp, RVInst Inst, RegisterClass StTy,
|
||||
ValueType vt> {
|
||||
ValueType vt> {
|
||||
def : Pat<(UniformBinFrag<StoreOp> (vt StTy:$rs2), (AddrRegImm GPR:$rs1, simm12:$imm12)),
|
||||
(Inst StTy:$rs2, GPR:$rs1, simm12:$imm12)>;
|
||||
}
|
||||
|
@ -1829,7 +1808,7 @@ def : Pat<(UniformWBinFrag<add> GPR:$rs1, (AddiPair:$rs2)),
|
|||
//===----------------------------------------------------------------------===//
|
||||
|
||||
include "VentusInstrInfoM.td"
|
||||
include "RISCVInstrInfoA.td"
|
||||
include "VentusInstrInfoA.td"
|
||||
include "RISCVInstrInfoF.td"
|
||||
include "RISCVInstrInfoD.td"
|
||||
include "VentusInstrInfoC.td"
|
||||
|
|
|
@ -0,0 +1,378 @@
|
|||
//===-- VentusInstrInfoA.td - Ventus 'A' instructions ------*- tablegen -*-===//
|
||||
//
|
||||
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
||||
// See https://llvm.org/LICENSE.txt for license information.
|
||||
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
//
|
||||
// This file describes the Ventus instructions from the standard 'A', Atomic
|
||||
// Instructions extension.
|
||||
//
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Instruction class templates
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
let hasSideEffects = 0, mayLoad = 1, mayStore = 0 in
|
||||
class LR_r<bit aq, bit rl, bits<3> funct3, string opcodestr>
|
||||
: RVInstRAtomic<0b00010, aq, rl, funct3, OPC_AMO,
|
||||
(outs GPR:$rd), (ins GPRMemZeroOffset:$rs1),
|
||||
opcodestr, "$rd, $rs1"> {
|
||||
let rs2 = 0;
|
||||
}
|
||||
|
||||
multiclass LR_r_aq_rl<bits<3> funct3, string opcodestr> {
|
||||
def "" : LR_r<0, 0, funct3, opcodestr>;
|
||||
def _AQ : LR_r<1, 0, funct3, opcodestr # ".aq">;
|
||||
def _RL : LR_r<0, 1, funct3, opcodestr # ".rl">;
|
||||
def _AQ_RL : LR_r<1, 1, funct3, opcodestr # ".aqrl">;
|
||||
}
|
||||
|
||||
let hasSideEffects = 0, mayLoad = 1, mayStore = 1 in
|
||||
class AMO_rr<bits<5> funct5, bit aq, bit rl, bits<3> funct3, string opcodestr>
|
||||
: RVInstRAtomic<funct5, aq, rl, funct3, OPC_AMO,
|
||||
(outs GPR:$rd), (ins GPRMemZeroOffset:$rs1, GPR:$rs2),
|
||||
opcodestr, "$rd, $rs2, $rs1">;
|
||||
|
||||
multiclass AMO_rr_aq_rl<bits<5> funct5, bits<3> funct3, string opcodestr> {
|
||||
def "" : AMO_rr<funct5, 0, 0, funct3, opcodestr>;
|
||||
def _AQ : AMO_rr<funct5, 1, 0, funct3, opcodestr # ".aq">;
|
||||
def _RL : AMO_rr<funct5, 0, 1, funct3, opcodestr # ".rl">;
|
||||
def _AQ_RL : AMO_rr<funct5, 1, 1, funct3, opcodestr # ".aqrl">;
|
||||
}
|
||||
|
||||
multiclass AtomicStPat<PatFrag StoreOp, RVInst Inst, RegisterClass StTy,
|
||||
ValueType vt = XLenVT> {
|
||||
def : Pat<(StoreOp (AddrRegImm GPR:$rs1, simm12:$imm12), (vt StTy:$rs2)),
|
||||
(Inst StTy:$rs2, GPR:$rs1, simm12:$imm12)>;
|
||||
}
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Instructions
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
let Predicates = [HasStdExtA] in {
|
||||
defm LR_W : LR_r_aq_rl<0b010, "lr.w">, Sched<[WriteAtomicLDW, ReadAtomicLDW]>;
|
||||
defm SC_W : AMO_rr_aq_rl<0b00011, 0b010, "sc.w">,
|
||||
Sched<[WriteAtomicSTW, ReadAtomicSTW, ReadAtomicSTW]>;
|
||||
defm AMOSWAP_W : AMO_rr_aq_rl<0b00001, 0b010, "amoswap.w">,
|
||||
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
||||
defm AMOADD_W : AMO_rr_aq_rl<0b00000, 0b010, "amoadd.w">,
|
||||
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
||||
defm AMOXOR_W : AMO_rr_aq_rl<0b00100, 0b010, "amoxor.w">,
|
||||
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
||||
defm AMOAND_W : AMO_rr_aq_rl<0b01100, 0b010, "amoand.w">,
|
||||
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
||||
defm AMOOR_W : AMO_rr_aq_rl<0b01000, 0b010, "amoor.w">,
|
||||
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
||||
defm AMOMIN_W : AMO_rr_aq_rl<0b10000, 0b010, "amomin.w">,
|
||||
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
||||
defm AMOMAX_W : AMO_rr_aq_rl<0b10100, 0b010, "amomax.w">,
|
||||
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
||||
defm AMOMINU_W : AMO_rr_aq_rl<0b11000, 0b010, "amominu.w">,
|
||||
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
||||
defm AMOMAXU_W : AMO_rr_aq_rl<0b11100, 0b010, "amomaxu.w">,
|
||||
Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>;
|
||||
} // Predicates = [HasStdExtA]
|
||||
|
||||
let Predicates = [HasStdExtA, IsRV64] in {
|
||||
defm LR_D : LR_r_aq_rl<0b011, "lr.d">, Sched<[WriteAtomicLDD, ReadAtomicLDD]>;
|
||||
defm SC_D : AMO_rr_aq_rl<0b00011, 0b011, "sc.d">,
|
||||
Sched<[WriteAtomicSTD, ReadAtomicSTD, ReadAtomicSTD]>;
|
||||
defm AMOSWAP_D : AMO_rr_aq_rl<0b00001, 0b011, "amoswap.d">,
|
||||
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
||||
defm AMOADD_D : AMO_rr_aq_rl<0b00000, 0b011, "amoadd.d">,
|
||||
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
||||
defm AMOXOR_D : AMO_rr_aq_rl<0b00100, 0b011, "amoxor.d">,
|
||||
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
||||
defm AMOAND_D : AMO_rr_aq_rl<0b01100, 0b011, "amoand.d">,
|
||||
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
||||
defm AMOOR_D : AMO_rr_aq_rl<0b01000, 0b011, "amoor.d">,
|
||||
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
||||
defm AMOMIN_D : AMO_rr_aq_rl<0b10000, 0b011, "amomin.d">,
|
||||
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
||||
defm AMOMAX_D : AMO_rr_aq_rl<0b10100, 0b011, "amomax.d">,
|
||||
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
||||
defm AMOMINU_D : AMO_rr_aq_rl<0b11000, 0b011, "amominu.d">,
|
||||
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
||||
defm AMOMAXU_D : AMO_rr_aq_rl<0b11100, 0b011, "amomaxu.d">,
|
||||
Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>;
|
||||
} // Predicates = [HasStdExtA, IsRV64]
|
||||
|
||||
//===----------------------------------------------------------------------===//
|
||||
// Pseudo-instructions and codegen patterns
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
// Atomic load/store are available under both +a and +force-atomics.
|
||||
// Fences will be inserted for atomic load/stores according to the logic in
|
||||
// RISCVTargetLowering::{emitLeadingFence,emitTrailingFence}.
|
||||
let Predicates = [HasAtomicLdSt] in {
|
||||
defm : UniformLdPat<atomic_load_8, LB>;
|
||||
defm : UniformLdPat<atomic_load_16, LH>;
|
||||
defm : UniformLdPat<atomic_load_32, LW>;
|
||||
|
||||
defm : AtomicStPat<atomic_store_8, SB, GPR>;
|
||||
defm : AtomicStPat<atomic_store_16, SH, GPR>;
|
||||
defm : AtomicStPat<atomic_store_32, SW, GPR>;
|
||||
}
|
||||
|
||||
let Predicates = [HasAtomicLdSt, IsRV64] in {
|
||||
defm : UniformLdPat<atomic_load_64, LD, i64>;
|
||||
defm : AtomicStPat<atomic_store_64, SD, GPR, i64>;
|
||||
}
|
||||
|
||||
let Predicates = [HasStdExtA] in {
|
||||
|
||||
/// AMOs
|
||||
|
||||
multiclass AMOPat<string AtomicOp, string BaseInst> {
|
||||
def : PatGprGpr<!cast<PatFrag>(AtomicOp#"_monotonic"),
|
||||
!cast<RVInst>(BaseInst)>;
|
||||
def : PatGprGpr<!cast<PatFrag>(AtomicOp#"_acquire"),
|
||||
!cast<RVInst>(BaseInst#"_AQ")>;
|
||||
def : PatGprGpr<!cast<PatFrag>(AtomicOp#"_release"),
|
||||
!cast<RVInst>(BaseInst#"_RL")>;
|
||||
def : PatGprGpr<!cast<PatFrag>(AtomicOp#"_acq_rel"),
|
||||
!cast<RVInst>(BaseInst#"_AQ_RL")>;
|
||||
def : PatGprGpr<!cast<PatFrag>(AtomicOp#"_seq_cst"),
|
||||
!cast<RVInst>(BaseInst#"_AQ_RL")>;
|
||||
}
|
||||
|
||||
defm : AMOPat<"atomic_swap_32", "AMOSWAP_W">;
|
||||
defm : AMOPat<"atomic_load_add_32", "AMOADD_W">;
|
||||
defm : AMOPat<"atomic_load_and_32", "AMOAND_W">;
|
||||
defm : AMOPat<"atomic_load_or_32", "AMOOR_W">;
|
||||
defm : AMOPat<"atomic_load_xor_32", "AMOXOR_W">;
|
||||
defm : AMOPat<"atomic_load_max_32", "AMOMAX_W">;
|
||||
defm : AMOPat<"atomic_load_min_32", "AMOMIN_W">;
|
||||
defm : AMOPat<"atomic_load_umax_32", "AMOMAXU_W">;
|
||||
defm : AMOPat<"atomic_load_umin_32", "AMOMINU_W">;
|
||||
|
||||
def : Pat<(atomic_load_sub_32_monotonic GPR:$addr, GPR:$incr),
|
||||
(AMOADD_W GPR:$addr, (SUB X0, GPR:$incr))>;
|
||||
def : Pat<(atomic_load_sub_32_acquire GPR:$addr, GPR:$incr),
|
||||
(AMOADD_W_AQ GPR:$addr, (SUB X0, GPR:$incr))>;
|
||||
def : Pat<(atomic_load_sub_32_release GPR:$addr, GPR:$incr),
|
||||
(AMOADD_W_RL GPR:$addr, (SUB X0, GPR:$incr))>;
|
||||
def : Pat<(atomic_load_sub_32_acq_rel GPR:$addr, GPR:$incr),
|
||||
(AMOADD_W_AQ_RL GPR:$addr, (SUB X0, GPR:$incr))>;
|
||||
def : Pat<(atomic_load_sub_32_seq_cst GPR:$addr, GPR:$incr),
|
||||
(AMOADD_W_AQ_RL GPR:$addr, (SUB X0, GPR:$incr))>;
|
||||
|
||||
/// Pseudo AMOs
|
||||
|
||||
class PseudoAMO : Pseudo<(outs GPR:$res, GPR:$scratch),
|
||||
(ins GPR:$addr, GPR:$incr, ixlenimm:$ordering), []> {
|
||||
let Constraints = "@earlyclobber $res,@earlyclobber $scratch";
|
||||
let mayLoad = 1;
|
||||
let mayStore = 1;
|
||||
let hasSideEffects = 0;
|
||||
}
|
||||
|
||||
let Size = 20 in
|
||||
def PseudoAtomicLoadNand32 : PseudoAMO;
|
||||
// Ordering constants must be kept in sync with the AtomicOrdering enum in
|
||||
// AtomicOrdering.h.
|
||||
def : Pat<(atomic_load_nand_32_monotonic GPR:$addr, GPR:$incr),
|
||||
(PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 2)>;
|
||||
def : Pat<(atomic_load_nand_32_acquire GPR:$addr, GPR:$incr),
|
||||
(PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 4)>;
|
||||
def : Pat<(atomic_load_nand_32_release GPR:$addr, GPR:$incr),
|
||||
(PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 5)>;
|
||||
def : Pat<(atomic_load_nand_32_acq_rel GPR:$addr, GPR:$incr),
|
||||
(PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 6)>;
|
||||
def : Pat<(atomic_load_nand_32_seq_cst GPR:$addr, GPR:$incr),
|
||||
(PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 7)>;
|
||||
|
||||
class PseudoMaskedAMO
|
||||
: Pseudo<(outs GPR:$res, GPR:$scratch),
|
||||
(ins GPR:$addr, GPR:$incr, GPR:$mask, ixlenimm:$ordering), []> {
|
||||
let Constraints = "@earlyclobber $res,@earlyclobber $scratch";
|
||||
let mayLoad = 1;
|
||||
let mayStore = 1;
|
||||
let hasSideEffects = 0;
|
||||
}
|
||||
|
||||
class PseudoMaskedAMOMinMax
|
||||
: Pseudo<(outs GPR:$res, GPR:$scratch1, GPR:$scratch2),
|
||||
(ins GPR:$addr, GPR:$incr, GPR:$mask, ixlenimm:$sextshamt,
|
||||
ixlenimm:$ordering), []> {
|
||||
let Constraints = "@earlyclobber $res,@earlyclobber $scratch1,"
|
||||
"@earlyclobber $scratch2";
|
||||
let mayLoad = 1;
|
||||
let mayStore = 1;
|
||||
let hasSideEffects = 0;
|
||||
}
|
||||
|
||||
class PseudoMaskedAMOUMinUMax
|
||||
: Pseudo<(outs GPR:$res, GPR:$scratch1, GPR:$scratch2),
|
||||
(ins GPR:$addr, GPR:$incr, GPR:$mask, ixlenimm:$ordering), []> {
|
||||
let Constraints = "@earlyclobber $res,@earlyclobber $scratch1,"
|
||||
"@earlyclobber $scratch2";
|
||||
let mayLoad = 1;
|
||||
let mayStore = 1;
|
||||
let hasSideEffects = 0;
|
||||
}
|
||||
|
||||
class PseudoMaskedAMOPat<Intrinsic intrin, Pseudo AMOInst>
|
||||
: Pat<(intrin GPR:$addr, GPR:$incr, GPR:$mask, timm:$ordering),
|
||||
(AMOInst GPR:$addr, GPR:$incr, GPR:$mask, timm:$ordering)>;
|
||||
|
||||
class PseudoMaskedAMOMinMaxPat<Intrinsic intrin, Pseudo AMOInst>
|
||||
: Pat<(intrin GPR:$addr, GPR:$incr, GPR:$mask, GPR:$shiftamt,
|
||||
timm:$ordering),
|
||||
(AMOInst GPR:$addr, GPR:$incr, GPR:$mask, GPR:$shiftamt,
|
||||
timm:$ordering)>;
|
||||
|
||||
let Size = 28 in
|
||||
def PseudoMaskedAtomicSwap32 : PseudoMaskedAMO;
|
||||
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_xchg_i32,
|
||||
PseudoMaskedAtomicSwap32>;
|
||||
let Size = 28 in
|
||||
def PseudoMaskedAtomicLoadAdd32 : PseudoMaskedAMO;
|
||||
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_add_i32,
|
||||
PseudoMaskedAtomicLoadAdd32>;
|
||||
let Size = 28 in
|
||||
def PseudoMaskedAtomicLoadSub32 : PseudoMaskedAMO;
|
||||
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_sub_i32,
|
||||
PseudoMaskedAtomicLoadSub32>;
|
||||
let Size = 32 in
|
||||
def PseudoMaskedAtomicLoadNand32 : PseudoMaskedAMO;
|
||||
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_nand_i32,
|
||||
PseudoMaskedAtomicLoadNand32>;
|
||||
let Size = 44 in
|
||||
def PseudoMaskedAtomicLoadMax32 : PseudoMaskedAMOMinMax;
|
||||
def : PseudoMaskedAMOMinMaxPat<int_riscv_masked_atomicrmw_max_i32,
|
||||
PseudoMaskedAtomicLoadMax32>;
|
||||
let Size = 44 in
|
||||
def PseudoMaskedAtomicLoadMin32 : PseudoMaskedAMOMinMax;
|
||||
def : PseudoMaskedAMOMinMaxPat<int_riscv_masked_atomicrmw_min_i32,
|
||||
PseudoMaskedAtomicLoadMin32>;
|
||||
let Size = 36 in
|
||||
def PseudoMaskedAtomicLoadUMax32 : PseudoMaskedAMOUMinUMax;
|
||||
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_umax_i32,
|
||||
PseudoMaskedAtomicLoadUMax32>;
|
||||
let Size = 36 in
|
||||
def PseudoMaskedAtomicLoadUMin32 : PseudoMaskedAMOUMinUMax;
|
||||
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_umin_i32,
|
||||
PseudoMaskedAtomicLoadUMin32>;
|
||||
|
||||
/// Compare and exchange
|
||||
|
||||
class PseudoCmpXchg
|
||||
: Pseudo<(outs GPR:$res, GPR:$scratch),
|
||||
(ins GPR:$addr, GPR:$cmpval, GPR:$newval, ixlenimm:$ordering), []> {
|
||||
let Constraints = "@earlyclobber $res,@earlyclobber $scratch";
|
||||
let mayLoad = 1;
|
||||
let mayStore = 1;
|
||||
let hasSideEffects = 0;
|
||||
let Size = 16;
|
||||
}
|
||||
|
||||
// Ordering constants must be kept in sync with the AtomicOrdering enum in
|
||||
// AtomicOrdering.h.
|
||||
multiclass PseudoCmpXchgPat<string Op, Pseudo CmpXchgInst> {
|
||||
def : Pat<(!cast<PatFrag>(Op#"_monotonic") GPR:$addr, GPR:$cmp, GPR:$new),
|
||||
(CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 2)>;
|
||||
def : Pat<(!cast<PatFrag>(Op#"_acquire") GPR:$addr, GPR:$cmp, GPR:$new),
|
||||
(CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 4)>;
|
||||
def : Pat<(!cast<PatFrag>(Op#"_release") GPR:$addr, GPR:$cmp, GPR:$new),
|
||||
(CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 5)>;
|
||||
def : Pat<(!cast<PatFrag>(Op#"_acq_rel") GPR:$addr, GPR:$cmp, GPR:$new),
|
||||
(CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 6)>;
|
||||
def : Pat<(!cast<PatFrag>(Op#"_seq_cst") GPR:$addr, GPR:$cmp, GPR:$new),
|
||||
(CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 7)>;
|
||||
}
|
||||
|
||||
def PseudoCmpXchg32 : PseudoCmpXchg;
|
||||
defm : PseudoCmpXchgPat<"atomic_cmp_swap_32", PseudoCmpXchg32>;
|
||||
|
||||
def PseudoMaskedCmpXchg32
|
||||
: Pseudo<(outs GPR:$res, GPR:$scratch),
|
||||
(ins GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask,
|
||||
ixlenimm:$ordering), []> {
|
||||
let Constraints = "@earlyclobber $res,@earlyclobber $scratch";
|
||||
let mayLoad = 1;
|
||||
let mayStore = 1;
|
||||
let hasSideEffects = 0;
|
||||
let Size = 32;
|
||||
}
|
||||
|
||||
def : Pat<(int_riscv_masked_cmpxchg_i32
|
||||
GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask, timm:$ordering),
|
||||
(PseudoMaskedCmpXchg32
|
||||
GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask, timm:$ordering)>;
|
||||
|
||||
} // Predicates = [HasStdExtA]
|
||||
|
||||
let Predicates = [HasStdExtA, IsRV64] in {
|
||||
|
||||
defm : AMOPat<"atomic_swap_64", "AMOSWAP_D">;
|
||||
defm : AMOPat<"atomic_load_add_64", "AMOADD_D">;
|
||||
defm : AMOPat<"atomic_load_and_64", "AMOAND_D">;
|
||||
defm : AMOPat<"atomic_load_or_64", "AMOOR_D">;
|
||||
defm : AMOPat<"atomic_load_xor_64", "AMOXOR_D">;
|
||||
defm : AMOPat<"atomic_load_max_64", "AMOMAX_D">;
|
||||
defm : AMOPat<"atomic_load_min_64", "AMOMIN_D">;
|
||||
defm : AMOPat<"atomic_load_umax_64", "AMOMAXU_D">;
|
||||
defm : AMOPat<"atomic_load_umin_64", "AMOMINU_D">;
|
||||
|
||||
/// 64-bit AMOs
|
||||
|
||||
def : Pat<(i64 (atomic_load_sub_64_monotonic GPR:$addr, GPR:$incr)),
|
||||
(AMOADD_D GPR:$addr, (SUB X0, GPR:$incr))>;
|
||||
def : Pat<(i64 (atomic_load_sub_64_acquire GPR:$addr, GPR:$incr)),
|
||||
(AMOADD_D_AQ GPR:$addr, (SUB X0, GPR:$incr))>;
|
||||
def : Pat<(i64 (atomic_load_sub_64_release GPR:$addr, GPR:$incr)),
|
||||
(AMOADD_D_RL GPR:$addr, (SUB X0, GPR:$incr))>;
|
||||
def : Pat<(i64 (atomic_load_sub_64_acq_rel GPR:$addr, GPR:$incr)),
|
||||
(AMOADD_D_AQ_RL GPR:$addr, (SUB X0, GPR:$incr))>;
|
||||
def : Pat<(i64 (atomic_load_sub_64_seq_cst GPR:$addr, GPR:$incr)),
|
||||
(AMOADD_D_AQ_RL GPR:$addr, (SUB X0, GPR:$incr))>;
|
||||
|
||||
/// 64-bit pseudo AMOs
|
||||
|
||||
let Size = 20 in
|
||||
def PseudoAtomicLoadNand64 : PseudoAMO;
|
||||
// Ordering constants must be kept in sync with the AtomicOrdering enum in
|
||||
// AtomicOrdering.h.
|
||||
def : Pat<(i64 (atomic_load_nand_64_monotonic GPR:$addr, GPR:$incr)),
|
||||
(PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 2)>;
|
||||
def : Pat<(i64 (atomic_load_nand_64_acquire GPR:$addr, GPR:$incr)),
|
||||
(PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 4)>;
|
||||
def : Pat<(i64 (atomic_load_nand_64_release GPR:$addr, GPR:$incr)),
|
||||
(PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 5)>;
|
||||
def : Pat<(i64 (atomic_load_nand_64_acq_rel GPR:$addr, GPR:$incr)),
|
||||
(PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 6)>;
|
||||
def : Pat<(i64 (atomic_load_nand_64_seq_cst GPR:$addr, GPR:$incr)),
|
||||
(PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 7)>;
|
||||
|
||||
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_xchg_i64,
|
||||
PseudoMaskedAtomicSwap32>;
|
||||
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_add_i64,
|
||||
PseudoMaskedAtomicLoadAdd32>;
|
||||
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_sub_i64,
|
||||
PseudoMaskedAtomicLoadSub32>;
|
||||
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_nand_i64,
|
||||
PseudoMaskedAtomicLoadNand32>;
|
||||
def : PseudoMaskedAMOMinMaxPat<int_riscv_masked_atomicrmw_max_i64,
|
||||
PseudoMaskedAtomicLoadMax32>;
|
||||
def : PseudoMaskedAMOMinMaxPat<int_riscv_masked_atomicrmw_min_i64,
|
||||
PseudoMaskedAtomicLoadMin32>;
|
||||
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_umax_i64,
|
||||
PseudoMaskedAtomicLoadUMax32>;
|
||||
def : PseudoMaskedAMOPat<int_riscv_masked_atomicrmw_umin_i64,
|
||||
PseudoMaskedAtomicLoadUMin32>;
|
||||
|
||||
/// 64-bit compare and exchange
|
||||
|
||||
def PseudoCmpXchg64 : PseudoCmpXchg;
|
||||
defm : PseudoCmpXchgPat<"atomic_cmp_swap_64", PseudoCmpXchg64>;
|
||||
|
||||
def : Pat<(int_riscv_masked_cmpxchg_i64
|
||||
GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask, timm:$ordering),
|
||||
(PseudoMaskedCmpXchg32
|
||||
GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask, timm:$ordering)>;
|
||||
} // Predicates = [HasStdExtA, IsRV64]
|
|
@ -752,6 +752,27 @@ def : InstAlias<"c.fsdsp $rs2, (${rs1})", (C_FSDSP FPR64C:$rs2, SPMem:$rs1, 0)>;
|
|||
// Compress Instruction tablegen backend.
|
||||
//===----------------------------------------------------------------------===//
|
||||
|
||||
class SelectCompressOpt<CondCode Cond>: Pat<
|
||||
(riscv_selectcc_frag:$select GPR:$lhs, simm12_no6:$Constant, Cond,
|
||||
GPR:$truev, GPR:$falsev),
|
||||
(Select_GPR_Using_CC_GPR (ADDI GPR:$lhs, (NegImm simm12:$Constant)), X0,
|
||||
(IntCCtoRISCVCC $select), GPR:$truev, GPR:$falsev)>;
|
||||
|
||||
class UniBrccCompressOpt<CondCode Cond, RVInstB Inst> : Pat<
|
||||
(UniformTetradFrag<riscv_brcc> GPR:$lhs, simm12_no6:$Constant, Cond, bb:$place),
|
||||
(Inst (ADDI GPR:$lhs, (NegImm simm12:$Constant)), X0, bb:$place)>;
|
||||
|
||||
def OptForMinSize : Predicate<"MF ? MF->getFunction().hasMinSize() : false">;
|
||||
|
||||
let Predicates = [HasStdExtC, OptForMinSize] in {
|
||||
def : SelectCompressOpt<SETEQ>;
|
||||
def : SelectCompressOpt<SETNE>;
|
||||
|
||||
def : UniBrccCompressOpt<SETEQ, BEQ>;
|
||||
def : UniBrccCompressOpt<SETNE, BNE>;
|
||||
}
|
||||
|
||||
|
||||
// Patterns are defined in the same order the compressed instructions appear
|
||||
// on page 82 of the ISA manual.
|
||||
|
||||
|
|
|
@ -67,42 +67,42 @@ def REMUW : ALUW_rr<0b0000001, 0b111, "remuw">,
|
|||
//===----------------------------------------------------------------------===//
|
||||
|
||||
let Predicates = [HasStdExtMOrZmmul] in {
|
||||
def : PatGprGpr<mul, MUL>;
|
||||
def : PatGprGpr<mulhs, MULH>;
|
||||
def : PatGprGpr<mulhu, MULHU>;
|
||||
def : PatGprGpr<riscv_mulhsu, MULHSU>;
|
||||
def : PatGprGpr<UniformBinFrag<mul>, MUL>;
|
||||
def : PatGprGpr<UniformBinFrag<mulhs>, MULH>;
|
||||
def : PatGprGpr<UniformBinFrag<mulhu>, MULHU>;
|
||||
def : PatGprGpr<UniformBinFrag<riscv_mulhsu>, MULHSU>;
|
||||
} // Predicates = [HasStdExtMOrZmmul]
|
||||
|
||||
let Predicates = [HasStdExtM] in {
|
||||
def : PatGprGpr<sdiv, DIV>;
|
||||
def : PatGprGpr<udiv, DIVU>;
|
||||
def : PatGprGpr<srem, REM>;
|
||||
def : PatGprGpr<urem, REMU>;
|
||||
def : PatGprGpr<UniformBinFrag<sdiv>, DIV>;
|
||||
def : PatGprGpr<UniformBinFrag<udiv>, DIVU>;
|
||||
def : PatGprGpr<UniformBinFrag<srem>, REM>;
|
||||
def : PatGprGpr<UniformBinFrag<urem>, REMU>;
|
||||
} // Predicates = [HasStdExtM]
|
||||
|
||||
// Select W instructions if only the lower 32-bits of the result are used.
|
||||
let Predicates = [HasStdExtMOrZmmul, IsRV64] in
|
||||
def : PatGprGpr<binop_allwusers<mul>, MULW>;
|
||||
def : PatGprGpr<UniformWBinFrag<mul>, MULW>;
|
||||
|
||||
let Predicates = [HasStdExtM, IsRV64] in {
|
||||
def : PatGprGpr<riscv_divw, DIVW>;
|
||||
def : PatGprGpr<riscv_divuw, DIVUW>;
|
||||
def : PatGprGpr<riscv_remuw, REMUW>;
|
||||
def : PatGprGpr<UniformWBinFrag<riscv_divw>, DIVW>;
|
||||
def : PatGprGpr<UniformWBinFrag<riscv_divuw>, DIVUW>;
|
||||
def : PatGprGpr<UniformWBinFrag<riscv_remuw>, REMUW>;
|
||||
|
||||
// Handle the specific cases where using DIVU/REMU would be correct and result
|
||||
// in fewer instructions than emitting DIVUW/REMUW then zero-extending the
|
||||
// result.
|
||||
def : Pat<(and (riscv_divuw (assertzexti32 GPR:$rs1),
|
||||
(assertzexti32 GPR:$rs2)), 0xffffffff),
|
||||
def : Pat<(UniformBinFrag<and> (riscv_divuw (assertzexti32 GPR:$rs1),
|
||||
(assertzexti32 GPR:$rs2)), 0xffffffff),
|
||||
(DIVU GPR:$rs1, GPR:$rs2)>;
|
||||
def : Pat<(and (riscv_remuw (assertzexti32 GPR:$rs1),
|
||||
(assertzexti32 GPR:$rs2)), 0xffffffff),
|
||||
def : Pat<(UniformBinFrag<and> (riscv_remuw (assertzexti32 GPR:$rs1),
|
||||
(assertzexti32 GPR:$rs2)), 0xffffffff),
|
||||
(REMU GPR:$rs1, GPR:$rs2)>;
|
||||
|
||||
// Although the sexti32 operands may not have originated from an i32 srem,
|
||||
// this pattern is safe as it is impossible for two sign extended inputs to
|
||||
// produce a result where res[63:32]=0 and res[31]=1.
|
||||
def : Pat<(srem (sexti32 (i64 GPR:$rs1)), (sexti32 (i64 GPR:$rs2))),
|
||||
def : Pat<(UniformBinFrag<srem> (sexti32 (i64 GPR:$rs1)), (sexti32 (i64 GPR:$rs2))),
|
||||
(REMW GPR:$rs1, GPR:$rs2)>;
|
||||
} // Predicates = [HasStdExtM, IsRV64]
|
||||
|
||||
|
@ -111,6 +111,6 @@ let Predicates = [HasStdExtMOrZmmul, IsRV64, NotHasStdExtZba] in {
|
|||
// multiply where the inputs aren't known to be zero extended. We can shift the
|
||||
// inputs left by 32 and use a MULHU. This saves two SRLIs needed to finish
|
||||
// zeroing the upper 32 bits.
|
||||
def : Pat<(i64 (mul (and GPR:$rs1, 0xffffffff), (and GPR:$rs2, 0xffffffff))),
|
||||
def : Pat<(i64 (UniformBinFrag<mul> (and GPR:$rs1, 0xffffffff), (and GPR:$rs2, 0xffffffff))),
|
||||
(MULHU (SLLI GPR:$rs1, 32), (SLLI GPR:$rs2, 32))>;
|
||||
} // Predicates = [HasStdExtMOrZmmul, IsRV64, NotHasStdExtZba]
|
||||
|
|
Loading…
Reference in New Issue