//===-- RISCVInstrInfoA.td - RISC-V 'A' instructions -------*- tablegen -*-===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// // // This file describes the RISC-V instructions from the standard 'A', Atomic // Instructions extension. // //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // Operand and SDNode transformation definitions. //===----------------------------------------------------------------------===// // A parse method for (${gpr}) or 0(${gpr}), where the 0 is be silently ignored. // Used for GNU as Compatibility. def AtomicMemOpOperand : AsmOperandClass { let Name = "AtomicMemOpOperand"; let RenderMethod = "addRegOperands"; let PredicateMethod = "isGPR"; let ParserMethod = "parseAtomicMemOp"; } def GPRMemAtomic : RegisterOperand { let ParserMatchClass = AtomicMemOpOperand; let PrintMethod = "printAtomicMemOp"; } //===----------------------------------------------------------------------===// // Instruction class templates //===----------------------------------------------------------------------===// let hasSideEffects = 0, mayLoad = 1, mayStore = 0 in class LR_r funct3, string opcodestr> : RVInstRAtomic<0b00010, aq, rl, funct3, OPC_AMO, (outs GPR:$rd), (ins GPRMemAtomic:$rs1), opcodestr, "$rd, $rs1"> { let rs2 = 0; } multiclass LR_r_aq_rl funct3, string opcodestr> { def "" : LR_r<0, 0, funct3, opcodestr>; def _AQ : LR_r<1, 0, funct3, opcodestr # ".aq">; def _RL : LR_r<0, 1, funct3, opcodestr # ".rl">; def _AQ_RL : LR_r<1, 1, funct3, opcodestr # ".aqrl">; } let hasSideEffects = 0, mayLoad = 1, mayStore = 1 in class AMO_rr funct5, bit aq, bit rl, bits<3> funct3, string opcodestr> : RVInstRAtomic; multiclass AMO_rr_aq_rl funct5, bits<3> funct3, string opcodestr> { def "" : AMO_rr; def _AQ : AMO_rr; def _RL : AMO_rr; def _AQ_RL : AMO_rr; } multiclass AtomicStPat { def : Pat<(StoreOp BaseAddr:$rs1, (vt StTy:$rs2)), (Inst StTy:$rs2, BaseAddr:$rs1, 0)>; def : Pat<(StoreOp (add BaseAddr:$rs1, simm12:$imm12), (vt StTy:$rs2)), (Inst StTy:$rs2, BaseAddr:$rs1, simm12:$imm12)>; def : Pat<(StoreOp (IsOrAdd AddrFI:$rs1, simm12:$imm12), (vt StTy:$rs2)), (Inst StTy:$rs2, AddrFI:$rs1, simm12:$imm12)>; } //===----------------------------------------------------------------------===// // Instructions //===----------------------------------------------------------------------===// let Predicates = [HasStdExtA] in { defm LR_W : LR_r_aq_rl<0b010, "lr.w">, Sched<[WriteAtomicLDW, ReadAtomicLDW]>; defm SC_W : AMO_rr_aq_rl<0b00011, 0b010, "sc.w">, Sched<[WriteAtomicSTW, ReadAtomicSTW, ReadAtomicSTW]>; defm AMOSWAP_W : AMO_rr_aq_rl<0b00001, 0b010, "amoswap.w">, Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>; defm AMOADD_W : AMO_rr_aq_rl<0b00000, 0b010, "amoadd.w">, Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>; defm AMOXOR_W : AMO_rr_aq_rl<0b00100, 0b010, "amoxor.w">, Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>; defm AMOAND_W : AMO_rr_aq_rl<0b01100, 0b010, "amoand.w">, Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>; defm AMOOR_W : AMO_rr_aq_rl<0b01000, 0b010, "amoor.w">, Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>; defm AMOMIN_W : AMO_rr_aq_rl<0b10000, 0b010, "amomin.w">, Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>; defm AMOMAX_W : AMO_rr_aq_rl<0b10100, 0b010, "amomax.w">, Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>; defm AMOMINU_W : AMO_rr_aq_rl<0b11000, 0b010, "amominu.w">, Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>; defm AMOMAXU_W : AMO_rr_aq_rl<0b11100, 0b010, "amomaxu.w">, Sched<[WriteAtomicW, ReadAtomicWA, ReadAtomicWD]>; } // Predicates = [HasStdExtA] let Predicates = [HasStdExtA, IsRV64] in { defm LR_D : LR_r_aq_rl<0b011, "lr.d">, Sched<[WriteAtomicLDD, ReadAtomicLDD]>; defm SC_D : AMO_rr_aq_rl<0b00011, 0b011, "sc.d">, Sched<[WriteAtomicSTD, ReadAtomicSTD, ReadAtomicSTD]>; defm AMOSWAP_D : AMO_rr_aq_rl<0b00001, 0b011, "amoswap.d">, Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>; defm AMOADD_D : AMO_rr_aq_rl<0b00000, 0b011, "amoadd.d">, Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>; defm AMOXOR_D : AMO_rr_aq_rl<0b00100, 0b011, "amoxor.d">, Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>; defm AMOAND_D : AMO_rr_aq_rl<0b01100, 0b011, "amoand.d">, Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>; defm AMOOR_D : AMO_rr_aq_rl<0b01000, 0b011, "amoor.d">, Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>; defm AMOMIN_D : AMO_rr_aq_rl<0b10000, 0b011, "amomin.d">, Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>; defm AMOMAX_D : AMO_rr_aq_rl<0b10100, 0b011, "amomax.d">, Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>; defm AMOMINU_D : AMO_rr_aq_rl<0b11000, 0b011, "amominu.d">, Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>; defm AMOMAXU_D : AMO_rr_aq_rl<0b11100, 0b011, "amomaxu.d">, Sched<[WriteAtomicD, ReadAtomicDA, ReadAtomicDD]>; } // Predicates = [HasStdExtA, IsRV64] //===----------------------------------------------------------------------===// // Pseudo-instructions and codegen patterns //===----------------------------------------------------------------------===// let Predicates = [HasStdExtA] in { /// Atomic loads and stores // Fences will be inserted for atomic load/stores according to the logic in // RISCVTargetLowering::{emitLeadingFence,emitTrailingFence}. defm : LdPat; defm : LdPat; defm : LdPat; defm : AtomicStPat; defm : AtomicStPat; defm : AtomicStPat; /// AMOs multiclass AMOPat { def : PatGprGpr(AtomicOp#"_monotonic"), !cast(BaseInst)>; def : PatGprGpr(AtomicOp#"_acquire"), !cast(BaseInst#"_AQ")>; def : PatGprGpr(AtomicOp#"_release"), !cast(BaseInst#"_RL")>; def : PatGprGpr(AtomicOp#"_acq_rel"), !cast(BaseInst#"_AQ_RL")>; def : PatGprGpr(AtomicOp#"_seq_cst"), !cast(BaseInst#"_AQ_RL")>; } defm : AMOPat<"atomic_swap_32", "AMOSWAP_W">; defm : AMOPat<"atomic_load_add_32", "AMOADD_W">; defm : AMOPat<"atomic_load_and_32", "AMOAND_W">; defm : AMOPat<"atomic_load_or_32", "AMOOR_W">; defm : AMOPat<"atomic_load_xor_32", "AMOXOR_W">; defm : AMOPat<"atomic_load_max_32", "AMOMAX_W">; defm : AMOPat<"atomic_load_min_32", "AMOMIN_W">; defm : AMOPat<"atomic_load_umax_32", "AMOMAXU_W">; defm : AMOPat<"atomic_load_umin_32", "AMOMINU_W">; def : Pat<(atomic_load_sub_32_monotonic GPR:$addr, GPR:$incr), (AMOADD_W GPR:$addr, (SUB X0, GPR:$incr))>; def : Pat<(atomic_load_sub_32_acquire GPR:$addr, GPR:$incr), (AMOADD_W_AQ GPR:$addr, (SUB X0, GPR:$incr))>; def : Pat<(atomic_load_sub_32_release GPR:$addr, GPR:$incr), (AMOADD_W_RL GPR:$addr, (SUB X0, GPR:$incr))>; def : Pat<(atomic_load_sub_32_acq_rel GPR:$addr, GPR:$incr), (AMOADD_W_AQ_RL GPR:$addr, (SUB X0, GPR:$incr))>; def : Pat<(atomic_load_sub_32_seq_cst GPR:$addr, GPR:$incr), (AMOADD_W_AQ_RL GPR:$addr, (SUB X0, GPR:$incr))>; /// Pseudo AMOs class PseudoAMO : Pseudo<(outs GPR:$res, GPR:$scratch), (ins GPR:$addr, GPR:$incr, ixlenimm:$ordering), []> { let Constraints = "@earlyclobber $res,@earlyclobber $scratch"; let mayLoad = 1; let mayStore = 1; let hasSideEffects = 0; } def PseudoAtomicLoadNand32 : PseudoAMO; // Ordering constants must be kept in sync with the AtomicOrdering enum in // AtomicOrdering.h. def : Pat<(atomic_load_nand_32_monotonic GPR:$addr, GPR:$incr), (PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 2)>; def : Pat<(atomic_load_nand_32_acquire GPR:$addr, GPR:$incr), (PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 4)>; def : Pat<(atomic_load_nand_32_release GPR:$addr, GPR:$incr), (PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 5)>; def : Pat<(atomic_load_nand_32_acq_rel GPR:$addr, GPR:$incr), (PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 6)>; def : Pat<(atomic_load_nand_32_seq_cst GPR:$addr, GPR:$incr), (PseudoAtomicLoadNand32 GPR:$addr, GPR:$incr, 7)>; class PseudoMaskedAMO : Pseudo<(outs GPR:$res, GPR:$scratch), (ins GPR:$addr, GPR:$incr, GPR:$mask, ixlenimm:$ordering), []> { let Constraints = "@earlyclobber $res,@earlyclobber $scratch"; let mayLoad = 1; let mayStore = 1; let hasSideEffects = 0; } class PseudoMaskedAMOMinMax : Pseudo<(outs GPR:$res, GPR:$scratch1, GPR:$scratch2), (ins GPR:$addr, GPR:$incr, GPR:$mask, ixlenimm:$sextshamt, ixlenimm:$ordering), []> { let Constraints = "@earlyclobber $res,@earlyclobber $scratch1," "@earlyclobber $scratch2"; let mayLoad = 1; let mayStore = 1; let hasSideEffects = 0; } class PseudoMaskedAMOUMinUMax : Pseudo<(outs GPR:$res, GPR:$scratch1, GPR:$scratch2), (ins GPR:$addr, GPR:$incr, GPR:$mask, ixlenimm:$ordering), []> { let Constraints = "@earlyclobber $res,@earlyclobber $scratch1," "@earlyclobber $scratch2"; let mayLoad = 1; let mayStore = 1; let hasSideEffects = 0; } class PseudoMaskedAMOPat : Pat<(intrin GPR:$addr, GPR:$incr, GPR:$mask, timm:$ordering), (AMOInst GPR:$addr, GPR:$incr, GPR:$mask, timm:$ordering)>; class PseudoMaskedAMOMinMaxPat : Pat<(intrin GPR:$addr, GPR:$incr, GPR:$mask, GPR:$shiftamt, timm:$ordering), (AMOInst GPR:$addr, GPR:$incr, GPR:$mask, GPR:$shiftamt, timm:$ordering)>; def PseudoMaskedAtomicSwap32 : PseudoMaskedAMO; def : PseudoMaskedAMOPat; def PseudoMaskedAtomicLoadAdd32 : PseudoMaskedAMO; def : PseudoMaskedAMOPat; def PseudoMaskedAtomicLoadSub32 : PseudoMaskedAMO; def : PseudoMaskedAMOPat; def PseudoMaskedAtomicLoadNand32 : PseudoMaskedAMO; def : PseudoMaskedAMOPat; def PseudoMaskedAtomicLoadMax32 : PseudoMaskedAMOMinMax; def : PseudoMaskedAMOMinMaxPat; def PseudoMaskedAtomicLoadMin32 : PseudoMaskedAMOMinMax; def : PseudoMaskedAMOMinMaxPat; def PseudoMaskedAtomicLoadUMax32 : PseudoMaskedAMOUMinUMax; def : PseudoMaskedAMOPat; def PseudoMaskedAtomicLoadUMin32 : PseudoMaskedAMOUMinUMax; def : PseudoMaskedAMOPat; /// Compare and exchange class PseudoCmpXchg : Pseudo<(outs GPR:$res, GPR:$scratch), (ins GPR:$addr, GPR:$cmpval, GPR:$newval, ixlenimm:$ordering), []> { let Constraints = "@earlyclobber $res,@earlyclobber $scratch"; let mayLoad = 1; let mayStore = 1; let hasSideEffects = 0; } // Ordering constants must be kept in sync with the AtomicOrdering enum in // AtomicOrdering.h. multiclass PseudoCmpXchgPat { def : Pat<(!cast(Op#"_monotonic") GPR:$addr, GPR:$cmp, GPR:$new), (CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 2)>; def : Pat<(!cast(Op#"_acquire") GPR:$addr, GPR:$cmp, GPR:$new), (CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 4)>; def : Pat<(!cast(Op#"_release") GPR:$addr, GPR:$cmp, GPR:$new), (CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 5)>; def : Pat<(!cast(Op#"_acq_rel") GPR:$addr, GPR:$cmp, GPR:$new), (CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 6)>; def : Pat<(!cast(Op#"_seq_cst") GPR:$addr, GPR:$cmp, GPR:$new), (CmpXchgInst GPR:$addr, GPR:$cmp, GPR:$new, 7)>; } def PseudoCmpXchg32 : PseudoCmpXchg; defm : PseudoCmpXchgPat<"atomic_cmp_swap_32", PseudoCmpXchg32>; def PseudoMaskedCmpXchg32 : Pseudo<(outs GPR:$res, GPR:$scratch), (ins GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask, ixlenimm:$ordering), []> { let Constraints = "@earlyclobber $res,@earlyclobber $scratch"; let mayLoad = 1; let mayStore = 1; let hasSideEffects = 0; } def : Pat<(int_riscv_masked_cmpxchg_i32 GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask, timm:$ordering), (PseudoMaskedCmpXchg32 GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask, timm:$ordering)>; } // Predicates = [HasStdExtA] let Predicates = [HasStdExtA, IsRV64] in { /// 64-bit atomic loads and stores // Fences will be inserted for atomic load/stores according to the logic in // RISCVTargetLowering::{emitLeadingFence,emitTrailingFence}. defm : LdPat; defm : AtomicStPat; defm : AMOPat<"atomic_swap_64", "AMOSWAP_D">; defm : AMOPat<"atomic_load_add_64", "AMOADD_D">; defm : AMOPat<"atomic_load_and_64", "AMOAND_D">; defm : AMOPat<"atomic_load_or_64", "AMOOR_D">; defm : AMOPat<"atomic_load_xor_64", "AMOXOR_D">; defm : AMOPat<"atomic_load_max_64", "AMOMAX_D">; defm : AMOPat<"atomic_load_min_64", "AMOMIN_D">; defm : AMOPat<"atomic_load_umax_64", "AMOMAXU_D">; defm : AMOPat<"atomic_load_umin_64", "AMOMINU_D">; /// 64-bit AMOs def : Pat<(i64 (atomic_load_sub_64_monotonic GPR:$addr, GPR:$incr)), (AMOADD_D GPR:$addr, (SUB X0, GPR:$incr))>; def : Pat<(i64 (atomic_load_sub_64_acquire GPR:$addr, GPR:$incr)), (AMOADD_D_AQ GPR:$addr, (SUB X0, GPR:$incr))>; def : Pat<(i64 (atomic_load_sub_64_release GPR:$addr, GPR:$incr)), (AMOADD_D_RL GPR:$addr, (SUB X0, GPR:$incr))>; def : Pat<(i64 (atomic_load_sub_64_acq_rel GPR:$addr, GPR:$incr)), (AMOADD_D_AQ_RL GPR:$addr, (SUB X0, GPR:$incr))>; def : Pat<(i64 (atomic_load_sub_64_seq_cst GPR:$addr, GPR:$incr)), (AMOADD_D_AQ_RL GPR:$addr, (SUB X0, GPR:$incr))>; /// 64-bit pseudo AMOs def PseudoAtomicLoadNand64 : PseudoAMO; // Ordering constants must be kept in sync with the AtomicOrdering enum in // AtomicOrdering.h. def : Pat<(i64 (atomic_load_nand_64_monotonic GPR:$addr, GPR:$incr)), (PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 2)>; def : Pat<(i64 (atomic_load_nand_64_acquire GPR:$addr, GPR:$incr)), (PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 4)>; def : Pat<(i64 (atomic_load_nand_64_release GPR:$addr, GPR:$incr)), (PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 5)>; def : Pat<(i64 (atomic_load_nand_64_acq_rel GPR:$addr, GPR:$incr)), (PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 6)>; def : Pat<(i64 (atomic_load_nand_64_seq_cst GPR:$addr, GPR:$incr)), (PseudoAtomicLoadNand64 GPR:$addr, GPR:$incr, 7)>; def : PseudoMaskedAMOPat; def : PseudoMaskedAMOPat; def : PseudoMaskedAMOPat; def : PseudoMaskedAMOPat; def : PseudoMaskedAMOMinMaxPat; def : PseudoMaskedAMOMinMaxPat; def : PseudoMaskedAMOPat; def : PseudoMaskedAMOPat; /// 64-bit compare and exchange def PseudoCmpXchg64 : PseudoCmpXchg; defm : PseudoCmpXchgPat<"atomic_cmp_swap_64", PseudoCmpXchg64>; def : Pat<(int_riscv_masked_cmpxchg_i64 GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask, timm:$ordering), (PseudoMaskedCmpXchg32 GPR:$addr, GPR:$cmpval, GPR:$newval, GPR:$mask, timm:$ordering)>; } // Predicates = [HasStdExtA, IsRV64]