//==-- AArch64InstPrinter.cpp - Convert AArch64 MCInst to assembly syntax --==// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// // // This class prints an AArch64 MCInst to a .s file. // //===----------------------------------------------------------------------===// #include "AArch64InstPrinter.h" #include "MCTargetDesc/AArch64AddressingModes.h" #include "Utils/AArch64BaseInfo.h" #include "llvm/ADT/STLExtras.h" #include "llvm/ADT/StringExtras.h" #include "llvm/ADT/StringRef.h" #include "llvm/MC/MCAsmInfo.h" #include "llvm/MC/MCExpr.h" #include "llvm/MC/MCInst.h" #include "llvm/MC/MCRegisterInfo.h" #include "llvm/MC/MCSubtargetInfo.h" #include "llvm/Support/Casting.h" #include "llvm/Support/ErrorHandling.h" #include "llvm/Support/Format.h" #include "llvm/Support/MathExtras.h" #include "llvm/Support/raw_ostream.h" #include #include #include using namespace llvm; #define DEBUG_TYPE "asm-printer" #define GET_INSTRUCTION_NAME #define PRINT_ALIAS_INSTR #include "AArch64GenAsmWriter.inc" #define GET_INSTRUCTION_NAME #define PRINT_ALIAS_INSTR #include "AArch64GenAsmWriter1.inc" AArch64InstPrinter::AArch64InstPrinter(const MCAsmInfo &MAI, const MCInstrInfo &MII, const MCRegisterInfo &MRI) : MCInstPrinter(MAI, MII, MRI) {} AArch64AppleInstPrinter::AArch64AppleInstPrinter(const MCAsmInfo &MAI, const MCInstrInfo &MII, const MCRegisterInfo &MRI) : AArch64InstPrinter(MAI, MII, MRI) {} bool AArch64InstPrinter::applyTargetSpecificCLOption(StringRef Opt) { if (Opt == "no-aliases") { PrintAliases = false; return true; } return false; } void AArch64InstPrinter::printRegName(raw_ostream &OS, MCRegister Reg) const { OS << markup(""); } void AArch64InstPrinter::printRegName(raw_ostream &OS, MCRegister Reg, unsigned AltIdx) const { OS << markup(""); } StringRef AArch64InstPrinter::getRegName(MCRegister Reg) const { return getRegisterName(Reg); } void AArch64InstPrinter::printInst(const MCInst *MI, uint64_t Address, StringRef Annot, const MCSubtargetInfo &STI, raw_ostream &O) { // Check for special encodings and print the canonical alias instead. unsigned Opcode = MI->getOpcode(); if (Opcode == AArch64::SYSxt) if (printSysAlias(MI, STI, O)) { printAnnotation(O, Annot); return; } if (Opcode == AArch64::SYSPxt || Opcode == AArch64::SYSPxt_XZR) if (printSyspAlias(MI, STI, O)) { printAnnotation(O, Annot); return; } // RPRFM overlaps PRFM (reg), so try to print it as RPRFM here. if ((Opcode == AArch64::PRFMroX) || (Opcode == AArch64::PRFMroW)) { if (printRangePrefetchAlias(MI, STI, O, Annot)) return; } // SBFM/UBFM should print to a nicer aliased form if possible. if (Opcode == AArch64::SBFMXri || Opcode == AArch64::SBFMWri || Opcode == AArch64::UBFMXri || Opcode == AArch64::UBFMWri) { const MCOperand &Op0 = MI->getOperand(0); const MCOperand &Op1 = MI->getOperand(1); const MCOperand &Op2 = MI->getOperand(2); const MCOperand &Op3 = MI->getOperand(3); bool IsSigned = (Opcode == AArch64::SBFMXri || Opcode == AArch64::SBFMWri); bool Is64Bit = (Opcode == AArch64::SBFMXri || Opcode == AArch64::UBFMXri); if (Op2.isImm() && Op2.getImm() == 0 && Op3.isImm()) { const char *AsmMnemonic = nullptr; switch (Op3.getImm()) { default: break; case 7: if (IsSigned) AsmMnemonic = "sxtb"; else if (!Is64Bit) AsmMnemonic = "uxtb"; break; case 15: if (IsSigned) AsmMnemonic = "sxth"; else if (!Is64Bit) AsmMnemonic = "uxth"; break; case 31: // *xtw is only valid for signed 64-bit operations. if (Is64Bit && IsSigned) AsmMnemonic = "sxtw"; break; } if (AsmMnemonic) { O << '\t' << AsmMnemonic << '\t'; printRegName(O, Op0.getReg()); O << ", "; printRegName(O, getWRegFromXReg(Op1.getReg())); printAnnotation(O, Annot); return; } } // All immediate shifts are aliases, implemented using the Bitfield // instruction. In all cases the immediate shift amount shift must be in // the range 0 to (reg.size -1). if (Op2.isImm() && Op3.isImm()) { const char *AsmMnemonic = nullptr; int shift = 0; int64_t immr = Op2.getImm(); int64_t imms = Op3.getImm(); if (Opcode == AArch64::UBFMWri && imms != 0x1F && ((imms + 1) == immr)) { AsmMnemonic = "lsl"; shift = 31 - imms; } else if (Opcode == AArch64::UBFMXri && imms != 0x3f && ((imms + 1 == immr))) { AsmMnemonic = "lsl"; shift = 63 - imms; } else if (Opcode == AArch64::UBFMWri && imms == 0x1f) { AsmMnemonic = "lsr"; shift = immr; } else if (Opcode == AArch64::UBFMXri && imms == 0x3f) { AsmMnemonic = "lsr"; shift = immr; } else if (Opcode == AArch64::SBFMWri && imms == 0x1f) { AsmMnemonic = "asr"; shift = immr; } else if (Opcode == AArch64::SBFMXri && imms == 0x3f) { AsmMnemonic = "asr"; shift = immr; } if (AsmMnemonic) { O << '\t' << AsmMnemonic << '\t'; printRegName(O, Op0.getReg()); O << ", "; printRegName(O, Op1.getReg()); O << ", " << markup(""); printAnnotation(O, Annot); return; } } // SBFIZ/UBFIZ aliases if (Op2.getImm() > Op3.getImm()) { O << '\t' << (IsSigned ? "sbfiz" : "ubfiz") << '\t'; printRegName(O, Op0.getReg()); O << ", "; printRegName(O, Op1.getReg()); O << ", " << markup("") << ", " << markup(""); printAnnotation(O, Annot); return; } // Otherwise SBFX/UBFX is the preferred form O << '\t' << (IsSigned ? "sbfx" : "ubfx") << '\t'; printRegName(O, Op0.getReg()); O << ", "; printRegName(O, Op1.getReg()); O << ", " << markup("") << ", " << markup(""); printAnnotation(O, Annot); return; } if (Opcode == AArch64::BFMXri || Opcode == AArch64::BFMWri) { const MCOperand &Op0 = MI->getOperand(0); // Op1 == Op0 const MCOperand &Op2 = MI->getOperand(2); int ImmR = MI->getOperand(3).getImm(); int ImmS = MI->getOperand(4).getImm(); if ((Op2.getReg() == AArch64::WZR || Op2.getReg() == AArch64::XZR) && (ImmR == 0 || ImmS < ImmR) && STI.getFeatureBits()[AArch64::HasV8_2aOps]) { // BFC takes precedence over its entire range, sligtly differently to BFI. int BitWidth = Opcode == AArch64::BFMXri ? 64 : 32; int LSB = (BitWidth - ImmR) % BitWidth; int Width = ImmS + 1; O << "\tbfc\t"; printRegName(O, Op0.getReg()); O << ", " << markup("") << ", " << markup(""); printAnnotation(O, Annot); return; } else if (ImmS < ImmR) { // BFI alias int BitWidth = Opcode == AArch64::BFMXri ? 64 : 32; int LSB = (BitWidth - ImmR) % BitWidth; int Width = ImmS + 1; O << "\tbfi\t"; printRegName(O, Op0.getReg()); O << ", "; printRegName(O, Op2.getReg()); O << ", " << markup("") << ", " << markup(""); printAnnotation(O, Annot); return; } int LSB = ImmR; int Width = ImmS - ImmR + 1; // Otherwise BFXIL the preferred form O << "\tbfxil\t"; printRegName(O, Op0.getReg()); O << ", "; printRegName(O, Op2.getReg()); O << ", " << markup("") << ", " << markup(""); printAnnotation(O, Annot); return; } // Symbolic operands for MOVZ, MOVN and MOVK already imply a shift // (e.g. :gottprel_g1: is always going to be "lsl #16") so it should not be // printed. if ((Opcode == AArch64::MOVZXi || Opcode == AArch64::MOVZWi || Opcode == AArch64::MOVNXi || Opcode == AArch64::MOVNWi) && MI->getOperand(1).isExpr()) { if (Opcode == AArch64::MOVZXi || Opcode == AArch64::MOVZWi) O << "\tmovz\t"; else O << "\tmovn\t"; printRegName(O, MI->getOperand(0).getReg()); O << ", " << markup("getOperand(1).getExpr()->print(O, &MAI); O << markup(">"); return; } if ((Opcode == AArch64::MOVKXi || Opcode == AArch64::MOVKWi) && MI->getOperand(2).isExpr()) { O << "\tmovk\t"; printRegName(O, MI->getOperand(0).getReg()); O << ", " << markup("getOperand(2).getExpr()->print(O, &MAI); O << markup(">"); return; } // MOVZ, MOVN and "ORR wzr, #imm" instructions are aliases for MOV, but their // domains overlap so they need to be prioritized. The chain is "MOVZ lsl #0 > // MOVZ lsl #N > MOVN lsl #0 > MOVN lsl #N > ORR". The highest instruction // that can represent the move is the MOV alias, and the rest get printed // normally. if ((Opcode == AArch64::MOVZXi || Opcode == AArch64::MOVZWi) && MI->getOperand(1).isImm() && MI->getOperand(2).isImm()) { int RegWidth = Opcode == AArch64::MOVZXi ? 64 : 32; int Shift = MI->getOperand(2).getImm(); uint64_t Value = (uint64_t)MI->getOperand(1).getImm() << Shift; if (AArch64_AM::isMOVZMovAlias(Value, Shift, Opcode == AArch64::MOVZXi ? 64 : 32)) { O << "\tmov\t"; printRegName(O, MI->getOperand(0).getReg()); O << ", " << markup(""); return; } } if ((Opcode == AArch64::MOVNXi || Opcode == AArch64::MOVNWi) && MI->getOperand(1).isImm() && MI->getOperand(2).isImm()) { int RegWidth = Opcode == AArch64::MOVNXi ? 64 : 32; int Shift = MI->getOperand(2).getImm(); uint64_t Value = ~((uint64_t)MI->getOperand(1).getImm() << Shift); if (RegWidth == 32) Value = Value & 0xffffffff; if (AArch64_AM::isMOVNMovAlias(Value, Shift, RegWidth)) { O << "\tmov\t"; printRegName(O, MI->getOperand(0).getReg()); O << ", " << markup(""); return; } } if ((Opcode == AArch64::ORRXri || Opcode == AArch64::ORRWri) && (MI->getOperand(1).getReg() == AArch64::XZR || MI->getOperand(1).getReg() == AArch64::WZR) && MI->getOperand(2).isImm()) { int RegWidth = Opcode == AArch64::ORRXri ? 64 : 32; uint64_t Value = AArch64_AM::decodeLogicalImmediate( MI->getOperand(2).getImm(), RegWidth); if (!AArch64_AM::isAnyMOVWMovAlias(Value, RegWidth)) { O << "\tmov\t"; printRegName(O, MI->getOperand(0).getReg()); O << ", " << markup(""); return; } } if (Opcode == AArch64::SPACE) { O << '\t' << MAI.getCommentString() << " SPACE " << MI->getOperand(1).getImm(); printAnnotation(O, Annot); return; } // Instruction TSB is specified as a one operand instruction, but 'csync' is // not encoded, so for printing it is treated as a special case here: if (Opcode == AArch64::TSB) { O << "\ttsb\tcsync"; return; } if (!PrintAliases || !printAliasInstr(MI, Address, STI, O)) printInstruction(MI, Address, STI, O); printAnnotation(O, Annot); if (atomicBarrierDroppedOnZero(Opcode) && (MI->getOperand(0).getReg() == AArch64::XZR || MI->getOperand(0).getReg() == AArch64::WZR)) { printAnnotation(O, "acquire semantics dropped since destination is zero"); } } static bool isTblTbxInstruction(unsigned Opcode, StringRef &Layout, bool &IsTbx) { switch (Opcode) { case AArch64::TBXv8i8One: case AArch64::TBXv8i8Two: case AArch64::TBXv8i8Three: case AArch64::TBXv8i8Four: IsTbx = true; Layout = ".8b"; return true; case AArch64::TBLv8i8One: case AArch64::TBLv8i8Two: case AArch64::TBLv8i8Three: case AArch64::TBLv8i8Four: IsTbx = false; Layout = ".8b"; return true; case AArch64::TBXv16i8One: case AArch64::TBXv16i8Two: case AArch64::TBXv16i8Three: case AArch64::TBXv16i8Four: IsTbx = true; Layout = ".16b"; return true; case AArch64::TBLv16i8One: case AArch64::TBLv16i8Two: case AArch64::TBLv16i8Three: case AArch64::TBLv16i8Four: IsTbx = false; Layout = ".16b"; return true; default: return false; } } struct LdStNInstrDesc { unsigned Opcode; const char *Mnemonic; const char *Layout; int ListOperand; bool HasLane; int NaturalOffset; }; static const LdStNInstrDesc LdStNInstInfo[] = { { AArch64::LD1i8, "ld1", ".b", 1, true, 0 }, { AArch64::LD1i16, "ld1", ".h", 1, true, 0 }, { AArch64::LD1i32, "ld1", ".s", 1, true, 0 }, { AArch64::LD1i64, "ld1", ".d", 1, true, 0 }, { AArch64::LD1i8_POST, "ld1", ".b", 2, true, 1 }, { AArch64::LD1i16_POST, "ld1", ".h", 2, true, 2 }, { AArch64::LD1i32_POST, "ld1", ".s", 2, true, 4 }, { AArch64::LD1i64_POST, "ld1", ".d", 2, true, 8 }, { AArch64::LD1Rv16b, "ld1r", ".16b", 0, false, 0 }, { AArch64::LD1Rv8h, "ld1r", ".8h", 0, false, 0 }, { AArch64::LD1Rv4s, "ld1r", ".4s", 0, false, 0 }, { AArch64::LD1Rv2d, "ld1r", ".2d", 0, false, 0 }, { AArch64::LD1Rv8b, "ld1r", ".8b", 0, false, 0 }, { AArch64::LD1Rv4h, "ld1r", ".4h", 0, false, 0 }, { AArch64::LD1Rv2s, "ld1r", ".2s", 0, false, 0 }, { AArch64::LD1Rv1d, "ld1r", ".1d", 0, false, 0 }, { AArch64::LD1Rv16b_POST, "ld1r", ".16b", 1, false, 1 }, { AArch64::LD1Rv8h_POST, "ld1r", ".8h", 1, false, 2 }, { AArch64::LD1Rv4s_POST, "ld1r", ".4s", 1, false, 4 }, { AArch64::LD1Rv2d_POST, "ld1r", ".2d", 1, false, 8 }, { AArch64::LD1Rv8b_POST, "ld1r", ".8b", 1, false, 1 }, { AArch64::LD1Rv4h_POST, "ld1r", ".4h", 1, false, 2 }, { AArch64::LD1Rv2s_POST, "ld1r", ".2s", 1, false, 4 }, { AArch64::LD1Rv1d_POST, "ld1r", ".1d", 1, false, 8 }, { AArch64::LD1Onev16b, "ld1", ".16b", 0, false, 0 }, { AArch64::LD1Onev8h, "ld1", ".8h", 0, false, 0 }, { AArch64::LD1Onev4s, "ld1", ".4s", 0, false, 0 }, { AArch64::LD1Onev2d, "ld1", ".2d", 0, false, 0 }, { AArch64::LD1Onev8b, "ld1", ".8b", 0, false, 0 }, { AArch64::LD1Onev4h, "ld1", ".4h", 0, false, 0 }, { AArch64::LD1Onev2s, "ld1", ".2s", 0, false, 0 }, { AArch64::LD1Onev1d, "ld1", ".1d", 0, false, 0 }, { AArch64::LD1Onev16b_POST, "ld1", ".16b", 1, false, 16 }, { AArch64::LD1Onev8h_POST, "ld1", ".8h", 1, false, 16 }, { AArch64::LD1Onev4s_POST, "ld1", ".4s", 1, false, 16 }, { AArch64::LD1Onev2d_POST, "ld1", ".2d", 1, false, 16 }, { AArch64::LD1Onev8b_POST, "ld1", ".8b", 1, false, 8 }, { AArch64::LD1Onev4h_POST, "ld1", ".4h", 1, false, 8 }, { AArch64::LD1Onev2s_POST, "ld1", ".2s", 1, false, 8 }, { AArch64::LD1Onev1d_POST, "ld1", ".1d", 1, false, 8 }, { AArch64::LD1Twov16b, "ld1", ".16b", 0, false, 0 }, { AArch64::LD1Twov8h, "ld1", ".8h", 0, false, 0 }, { AArch64::LD1Twov4s, "ld1", ".4s", 0, false, 0 }, { AArch64::LD1Twov2d, "ld1", ".2d", 0, false, 0 }, { AArch64::LD1Twov8b, "ld1", ".8b", 0, false, 0 }, { AArch64::LD1Twov4h, "ld1", ".4h", 0, false, 0 }, { AArch64::LD1Twov2s, "ld1", ".2s", 0, false, 0 }, { AArch64::LD1Twov1d, "ld1", ".1d", 0, false, 0 }, { AArch64::LD1Twov16b_POST, "ld1", ".16b", 1, false, 32 }, { AArch64::LD1Twov8h_POST, "ld1", ".8h", 1, false, 32 }, { AArch64::LD1Twov4s_POST, "ld1", ".4s", 1, false, 32 }, { AArch64::LD1Twov2d_POST, "ld1", ".2d", 1, false, 32 }, { AArch64::LD1Twov8b_POST, "ld1", ".8b", 1, false, 16 }, { AArch64::LD1Twov4h_POST, "ld1", ".4h", 1, false, 16 }, { AArch64::LD1Twov2s_POST, "ld1", ".2s", 1, false, 16 }, { AArch64::LD1Twov1d_POST, "ld1", ".1d", 1, false, 16 }, { AArch64::LD1Threev16b, "ld1", ".16b", 0, false, 0 }, { AArch64::LD1Threev8h, "ld1", ".8h", 0, false, 0 }, { AArch64::LD1Threev4s, "ld1", ".4s", 0, false, 0 }, { AArch64::LD1Threev2d, "ld1", ".2d", 0, false, 0 }, { AArch64::LD1Threev8b, "ld1", ".8b", 0, false, 0 }, { AArch64::LD1Threev4h, "ld1", ".4h", 0, false, 0 }, { AArch64::LD1Threev2s, "ld1", ".2s", 0, false, 0 }, { AArch64::LD1Threev1d, "ld1", ".1d", 0, false, 0 }, { AArch64::LD1Threev16b_POST, "ld1", ".16b", 1, false, 48 }, { AArch64::LD1Threev8h_POST, "ld1", ".8h", 1, false, 48 }, { AArch64::LD1Threev4s_POST, "ld1", ".4s", 1, false, 48 }, { AArch64::LD1Threev2d_POST, "ld1", ".2d", 1, false, 48 }, { AArch64::LD1Threev8b_POST, "ld1", ".8b", 1, false, 24 }, { AArch64::LD1Threev4h_POST, "ld1", ".4h", 1, false, 24 }, { AArch64::LD1Threev2s_POST, "ld1", ".2s", 1, false, 24 }, { AArch64::LD1Threev1d_POST, "ld1", ".1d", 1, false, 24 }, { AArch64::LD1Fourv16b, "ld1", ".16b", 0, false, 0 }, { AArch64::LD1Fourv8h, "ld1", ".8h", 0, false, 0 }, { AArch64::LD1Fourv4s, "ld1", ".4s", 0, false, 0 }, { AArch64::LD1Fourv2d, "ld1", ".2d", 0, false, 0 }, { AArch64::LD1Fourv8b, "ld1", ".8b", 0, false, 0 }, { AArch64::LD1Fourv4h, "ld1", ".4h", 0, false, 0 }, { AArch64::LD1Fourv2s, "ld1", ".2s", 0, false, 0 }, { AArch64::LD1Fourv1d, "ld1", ".1d", 0, false, 0 }, { AArch64::LD1Fourv16b_POST, "ld1", ".16b", 1, false, 64 }, { AArch64::LD1Fourv8h_POST, "ld1", ".8h", 1, false, 64 }, { AArch64::LD1Fourv4s_POST, "ld1", ".4s", 1, false, 64 }, { AArch64::LD1Fourv2d_POST, "ld1", ".2d", 1, false, 64 }, { AArch64::LD1Fourv8b_POST, "ld1", ".8b", 1, false, 32 }, { AArch64::LD1Fourv4h_POST, "ld1", ".4h", 1, false, 32 }, { AArch64::LD1Fourv2s_POST, "ld1", ".2s", 1, false, 32 }, { AArch64::LD1Fourv1d_POST, "ld1", ".1d", 1, false, 32 }, { AArch64::LD2i8, "ld2", ".b", 1, true, 0 }, { AArch64::LD2i16, "ld2", ".h", 1, true, 0 }, { AArch64::LD2i32, "ld2", ".s", 1, true, 0 }, { AArch64::LD2i64, "ld2", ".d", 1, true, 0 }, { AArch64::LD2i8_POST, "ld2", ".b", 2, true, 2 }, { AArch64::LD2i16_POST, "ld2", ".h", 2, true, 4 }, { AArch64::LD2i32_POST, "ld2", ".s", 2, true, 8 }, { AArch64::LD2i64_POST, "ld2", ".d", 2, true, 16 }, { AArch64::LD2Rv16b, "ld2r", ".16b", 0, false, 0 }, { AArch64::LD2Rv8h, "ld2r", ".8h", 0, false, 0 }, { AArch64::LD2Rv4s, "ld2r", ".4s", 0, false, 0 }, { AArch64::LD2Rv2d, "ld2r", ".2d", 0, false, 0 }, { AArch64::LD2Rv8b, "ld2r", ".8b", 0, false, 0 }, { AArch64::LD2Rv4h, "ld2r", ".4h", 0, false, 0 }, { AArch64::LD2Rv2s, "ld2r", ".2s", 0, false, 0 }, { AArch64::LD2Rv1d, "ld2r", ".1d", 0, false, 0 }, { AArch64::LD2Rv16b_POST, "ld2r", ".16b", 1, false, 2 }, { AArch64::LD2Rv8h_POST, "ld2r", ".8h", 1, false, 4 }, { AArch64::LD2Rv4s_POST, "ld2r", ".4s", 1, false, 8 }, { AArch64::LD2Rv2d_POST, "ld2r", ".2d", 1, false, 16 }, { AArch64::LD2Rv8b_POST, "ld2r", ".8b", 1, false, 2 }, { AArch64::LD2Rv4h_POST, "ld2r", ".4h", 1, false, 4 }, { AArch64::LD2Rv2s_POST, "ld2r", ".2s", 1, false, 8 }, { AArch64::LD2Rv1d_POST, "ld2r", ".1d", 1, false, 16 }, { AArch64::LD2Twov16b, "ld2", ".16b", 0, false, 0 }, { AArch64::LD2Twov8h, "ld2", ".8h", 0, false, 0 }, { AArch64::LD2Twov4s, "ld2", ".4s", 0, false, 0 }, { AArch64::LD2Twov2d, "ld2", ".2d", 0, false, 0 }, { AArch64::LD2Twov8b, "ld2", ".8b", 0, false, 0 }, { AArch64::LD2Twov4h, "ld2", ".4h", 0, false, 0 }, { AArch64::LD2Twov2s, "ld2", ".2s", 0, false, 0 }, { AArch64::LD2Twov16b_POST, "ld2", ".16b", 1, false, 32 }, { AArch64::LD2Twov8h_POST, "ld2", ".8h", 1, false, 32 }, { AArch64::LD2Twov4s_POST, "ld2", ".4s", 1, false, 32 }, { AArch64::LD2Twov2d_POST, "ld2", ".2d", 1, false, 32 }, { AArch64::LD2Twov8b_POST, "ld2", ".8b", 1, false, 16 }, { AArch64::LD2Twov4h_POST, "ld2", ".4h", 1, false, 16 }, { AArch64::LD2Twov2s_POST, "ld2", ".2s", 1, false, 16 }, { AArch64::LD3i8, "ld3", ".b", 1, true, 0 }, { AArch64::LD3i16, "ld3", ".h", 1, true, 0 }, { AArch64::LD3i32, "ld3", ".s", 1, true, 0 }, { AArch64::LD3i64, "ld3", ".d", 1, true, 0 }, { AArch64::LD3i8_POST, "ld3", ".b", 2, true, 3 }, { AArch64::LD3i16_POST, "ld3", ".h", 2, true, 6 }, { AArch64::LD3i32_POST, "ld3", ".s", 2, true, 12 }, { AArch64::LD3i64_POST, "ld3", ".d", 2, true, 24 }, { AArch64::LD3Rv16b, "ld3r", ".16b", 0, false, 0 }, { AArch64::LD3Rv8h, "ld3r", ".8h", 0, false, 0 }, { AArch64::LD3Rv4s, "ld3r", ".4s", 0, false, 0 }, { AArch64::LD3Rv2d, "ld3r", ".2d", 0, false, 0 }, { AArch64::LD3Rv8b, "ld3r", ".8b", 0, false, 0 }, { AArch64::LD3Rv4h, "ld3r", ".4h", 0, false, 0 }, { AArch64::LD3Rv2s, "ld3r", ".2s", 0, false, 0 }, { AArch64::LD3Rv1d, "ld3r", ".1d", 0, false, 0 }, { AArch64::LD3Rv16b_POST, "ld3r", ".16b", 1, false, 3 }, { AArch64::LD3Rv8h_POST, "ld3r", ".8h", 1, false, 6 }, { AArch64::LD3Rv4s_POST, "ld3r", ".4s", 1, false, 12 }, { AArch64::LD3Rv2d_POST, "ld3r", ".2d", 1, false, 24 }, { AArch64::LD3Rv8b_POST, "ld3r", ".8b", 1, false, 3 }, { AArch64::LD3Rv4h_POST, "ld3r", ".4h", 1, false, 6 }, { AArch64::LD3Rv2s_POST, "ld3r", ".2s", 1, false, 12 }, { AArch64::LD3Rv1d_POST, "ld3r", ".1d", 1, false, 24 }, { AArch64::LD3Threev16b, "ld3", ".16b", 0, false, 0 }, { AArch64::LD3Threev8h, "ld3", ".8h", 0, false, 0 }, { AArch64::LD3Threev4s, "ld3", ".4s", 0, false, 0 }, { AArch64::LD3Threev2d, "ld3", ".2d", 0, false, 0 }, { AArch64::LD3Threev8b, "ld3", ".8b", 0, false, 0 }, { AArch64::LD3Threev4h, "ld3", ".4h", 0, false, 0 }, { AArch64::LD3Threev2s, "ld3", ".2s", 0, false, 0 }, { AArch64::LD3Threev16b_POST, "ld3", ".16b", 1, false, 48 }, { AArch64::LD3Threev8h_POST, "ld3", ".8h", 1, false, 48 }, { AArch64::LD3Threev4s_POST, "ld3", ".4s", 1, false, 48 }, { AArch64::LD3Threev2d_POST, "ld3", ".2d", 1, false, 48 }, { AArch64::LD3Threev8b_POST, "ld3", ".8b", 1, false, 24 }, { AArch64::LD3Threev4h_POST, "ld3", ".4h", 1, false, 24 }, { AArch64::LD3Threev2s_POST, "ld3", ".2s", 1, false, 24 }, { AArch64::LD4i8, "ld4", ".b", 1, true, 0 }, { AArch64::LD4i16, "ld4", ".h", 1, true, 0 }, { AArch64::LD4i32, "ld4", ".s", 1, true, 0 }, { AArch64::LD4i64, "ld4", ".d", 1, true, 0 }, { AArch64::LD4i8_POST, "ld4", ".b", 2, true, 4 }, { AArch64::LD4i16_POST, "ld4", ".h", 2, true, 8 }, { AArch64::LD4i32_POST, "ld4", ".s", 2, true, 16 }, { AArch64::LD4i64_POST, "ld4", ".d", 2, true, 32 }, { AArch64::LD4Rv16b, "ld4r", ".16b", 0, false, 0 }, { AArch64::LD4Rv8h, "ld4r", ".8h", 0, false, 0 }, { AArch64::LD4Rv4s, "ld4r", ".4s", 0, false, 0 }, { AArch64::LD4Rv2d, "ld4r", ".2d", 0, false, 0 }, { AArch64::LD4Rv8b, "ld4r", ".8b", 0, false, 0 }, { AArch64::LD4Rv4h, "ld4r", ".4h", 0, false, 0 }, { AArch64::LD4Rv2s, "ld4r", ".2s", 0, false, 0 }, { AArch64::LD4Rv1d, "ld4r", ".1d", 0, false, 0 }, { AArch64::LD4Rv16b_POST, "ld4r", ".16b", 1, false, 4 }, { AArch64::LD4Rv8h_POST, "ld4r", ".8h", 1, false, 8 }, { AArch64::LD4Rv4s_POST, "ld4r", ".4s", 1, false, 16 }, { AArch64::LD4Rv2d_POST, "ld4r", ".2d", 1, false, 32 }, { AArch64::LD4Rv8b_POST, "ld4r", ".8b", 1, false, 4 }, { AArch64::LD4Rv4h_POST, "ld4r", ".4h", 1, false, 8 }, { AArch64::LD4Rv2s_POST, "ld4r", ".2s", 1, false, 16 }, { AArch64::LD4Rv1d_POST, "ld4r", ".1d", 1, false, 32 }, { AArch64::LD4Fourv16b, "ld4", ".16b", 0, false, 0 }, { AArch64::LD4Fourv8h, "ld4", ".8h", 0, false, 0 }, { AArch64::LD4Fourv4s, "ld4", ".4s", 0, false, 0 }, { AArch64::LD4Fourv2d, "ld4", ".2d", 0, false, 0 }, { AArch64::LD4Fourv8b, "ld4", ".8b", 0, false, 0 }, { AArch64::LD4Fourv4h, "ld4", ".4h", 0, false, 0 }, { AArch64::LD4Fourv2s, "ld4", ".2s", 0, false, 0 }, { AArch64::LD4Fourv16b_POST, "ld4", ".16b", 1, false, 64 }, { AArch64::LD4Fourv8h_POST, "ld4", ".8h", 1, false, 64 }, { AArch64::LD4Fourv4s_POST, "ld4", ".4s", 1, false, 64 }, { AArch64::LD4Fourv2d_POST, "ld4", ".2d", 1, false, 64 }, { AArch64::LD4Fourv8b_POST, "ld4", ".8b", 1, false, 32 }, { AArch64::LD4Fourv4h_POST, "ld4", ".4h", 1, false, 32 }, { AArch64::LD4Fourv2s_POST, "ld4", ".2s", 1, false, 32 }, { AArch64::ST1i8, "st1", ".b", 0, true, 0 }, { AArch64::ST1i16, "st1", ".h", 0, true, 0 }, { AArch64::ST1i32, "st1", ".s", 0, true, 0 }, { AArch64::ST1i64, "st1", ".d", 0, true, 0 }, { AArch64::ST1i8_POST, "st1", ".b", 1, true, 1 }, { AArch64::ST1i16_POST, "st1", ".h", 1, true, 2 }, { AArch64::ST1i32_POST, "st1", ".s", 1, true, 4 }, { AArch64::ST1i64_POST, "st1", ".d", 1, true, 8 }, { AArch64::ST1Onev16b, "st1", ".16b", 0, false, 0 }, { AArch64::ST1Onev8h, "st1", ".8h", 0, false, 0 }, { AArch64::ST1Onev4s, "st1", ".4s", 0, false, 0 }, { AArch64::ST1Onev2d, "st1", ".2d", 0, false, 0 }, { AArch64::ST1Onev8b, "st1", ".8b", 0, false, 0 }, { AArch64::ST1Onev4h, "st1", ".4h", 0, false, 0 }, { AArch64::ST1Onev2s, "st1", ".2s", 0, false, 0 }, { AArch64::ST1Onev1d, "st1", ".1d", 0, false, 0 }, { AArch64::ST1Onev16b_POST, "st1", ".16b", 1, false, 16 }, { AArch64::ST1Onev8h_POST, "st1", ".8h", 1, false, 16 }, { AArch64::ST1Onev4s_POST, "st1", ".4s", 1, false, 16 }, { AArch64::ST1Onev2d_POST, "st1", ".2d", 1, false, 16 }, { AArch64::ST1Onev8b_POST, "st1", ".8b", 1, false, 8 }, { AArch64::ST1Onev4h_POST, "st1", ".4h", 1, false, 8 }, { AArch64::ST1Onev2s_POST, "st1", ".2s", 1, false, 8 }, { AArch64::ST1Onev1d_POST, "st1", ".1d", 1, false, 8 }, { AArch64::ST1Twov16b, "st1", ".16b", 0, false, 0 }, { AArch64::ST1Twov8h, "st1", ".8h", 0, false, 0 }, { AArch64::ST1Twov4s, "st1", ".4s", 0, false, 0 }, { AArch64::ST1Twov2d, "st1", ".2d", 0, false, 0 }, { AArch64::ST1Twov8b, "st1", ".8b", 0, false, 0 }, { AArch64::ST1Twov4h, "st1", ".4h", 0, false, 0 }, { AArch64::ST1Twov2s, "st1", ".2s", 0, false, 0 }, { AArch64::ST1Twov1d, "st1", ".1d", 0, false, 0 }, { AArch64::ST1Twov16b_POST, "st1", ".16b", 1, false, 32 }, { AArch64::ST1Twov8h_POST, "st1", ".8h", 1, false, 32 }, { AArch64::ST1Twov4s_POST, "st1", ".4s", 1, false, 32 }, { AArch64::ST1Twov2d_POST, "st1", ".2d", 1, false, 32 }, { AArch64::ST1Twov8b_POST, "st1", ".8b", 1, false, 16 }, { AArch64::ST1Twov4h_POST, "st1", ".4h", 1, false, 16 }, { AArch64::ST1Twov2s_POST, "st1", ".2s", 1, false, 16 }, { AArch64::ST1Twov1d_POST, "st1", ".1d", 1, false, 16 }, { AArch64::ST1Threev16b, "st1", ".16b", 0, false, 0 }, { AArch64::ST1Threev8h, "st1", ".8h", 0, false, 0 }, { AArch64::ST1Threev4s, "st1", ".4s", 0, false, 0 }, { AArch64::ST1Threev2d, "st1", ".2d", 0, false, 0 }, { AArch64::ST1Threev8b, "st1", ".8b", 0, false, 0 }, { AArch64::ST1Threev4h, "st1", ".4h", 0, false, 0 }, { AArch64::ST1Threev2s, "st1", ".2s", 0, false, 0 }, { AArch64::ST1Threev1d, "st1", ".1d", 0, false, 0 }, { AArch64::ST1Threev16b_POST, "st1", ".16b", 1, false, 48 }, { AArch64::ST1Threev8h_POST, "st1", ".8h", 1, false, 48 }, { AArch64::ST1Threev4s_POST, "st1", ".4s", 1, false, 48 }, { AArch64::ST1Threev2d_POST, "st1", ".2d", 1, false, 48 }, { AArch64::ST1Threev8b_POST, "st1", ".8b", 1, false, 24 }, { AArch64::ST1Threev4h_POST, "st1", ".4h", 1, false, 24 }, { AArch64::ST1Threev2s_POST, "st1", ".2s", 1, false, 24 }, { AArch64::ST1Threev1d_POST, "st1", ".1d", 1, false, 24 }, { AArch64::ST1Fourv16b, "st1", ".16b", 0, false, 0 }, { AArch64::ST1Fourv8h, "st1", ".8h", 0, false, 0 }, { AArch64::ST1Fourv4s, "st1", ".4s", 0, false, 0 }, { AArch64::ST1Fourv2d, "st1", ".2d", 0, false, 0 }, { AArch64::ST1Fourv8b, "st1", ".8b", 0, false, 0 }, { AArch64::ST1Fourv4h, "st1", ".4h", 0, false, 0 }, { AArch64::ST1Fourv2s, "st1", ".2s", 0, false, 0 }, { AArch64::ST1Fourv1d, "st1", ".1d", 0, false, 0 }, { AArch64::ST1Fourv16b_POST, "st1", ".16b", 1, false, 64 }, { AArch64::ST1Fourv8h_POST, "st1", ".8h", 1, false, 64 }, { AArch64::ST1Fourv4s_POST, "st1", ".4s", 1, false, 64 }, { AArch64::ST1Fourv2d_POST, "st1", ".2d", 1, false, 64 }, { AArch64::ST1Fourv8b_POST, "st1", ".8b", 1, false, 32 }, { AArch64::ST1Fourv4h_POST, "st1", ".4h", 1, false, 32 }, { AArch64::ST1Fourv2s_POST, "st1", ".2s", 1, false, 32 }, { AArch64::ST1Fourv1d_POST, "st1", ".1d", 1, false, 32 }, { AArch64::ST2i8, "st2", ".b", 0, true, 0 }, { AArch64::ST2i16, "st2", ".h", 0, true, 0 }, { AArch64::ST2i32, "st2", ".s", 0, true, 0 }, { AArch64::ST2i64, "st2", ".d", 0, true, 0 }, { AArch64::ST2i8_POST, "st2", ".b", 1, true, 2 }, { AArch64::ST2i16_POST, "st2", ".h", 1, true, 4 }, { AArch64::ST2i32_POST, "st2", ".s", 1, true, 8 }, { AArch64::ST2i64_POST, "st2", ".d", 1, true, 16 }, { AArch64::ST2Twov16b, "st2", ".16b", 0, false, 0 }, { AArch64::ST2Twov8h, "st2", ".8h", 0, false, 0 }, { AArch64::ST2Twov4s, "st2", ".4s", 0, false, 0 }, { AArch64::ST2Twov2d, "st2", ".2d", 0, false, 0 }, { AArch64::ST2Twov8b, "st2", ".8b", 0, false, 0 }, { AArch64::ST2Twov4h, "st2", ".4h", 0, false, 0 }, { AArch64::ST2Twov2s, "st2", ".2s", 0, false, 0 }, { AArch64::ST2Twov16b_POST, "st2", ".16b", 1, false, 32 }, { AArch64::ST2Twov8h_POST, "st2", ".8h", 1, false, 32 }, { AArch64::ST2Twov4s_POST, "st2", ".4s", 1, false, 32 }, { AArch64::ST2Twov2d_POST, "st2", ".2d", 1, false, 32 }, { AArch64::ST2Twov8b_POST, "st2", ".8b", 1, false, 16 }, { AArch64::ST2Twov4h_POST, "st2", ".4h", 1, false, 16 }, { AArch64::ST2Twov2s_POST, "st2", ".2s", 1, false, 16 }, { AArch64::ST3i8, "st3", ".b", 0, true, 0 }, { AArch64::ST3i16, "st3", ".h", 0, true, 0 }, { AArch64::ST3i32, "st3", ".s", 0, true, 0 }, { AArch64::ST3i64, "st3", ".d", 0, true, 0 }, { AArch64::ST3i8_POST, "st3", ".b", 1, true, 3 }, { AArch64::ST3i16_POST, "st3", ".h", 1, true, 6 }, { AArch64::ST3i32_POST, "st3", ".s", 1, true, 12 }, { AArch64::ST3i64_POST, "st3", ".d", 1, true, 24 }, { AArch64::ST3Threev16b, "st3", ".16b", 0, false, 0 }, { AArch64::ST3Threev8h, "st3", ".8h", 0, false, 0 }, { AArch64::ST3Threev4s, "st3", ".4s", 0, false, 0 }, { AArch64::ST3Threev2d, "st3", ".2d", 0, false, 0 }, { AArch64::ST3Threev8b, "st3", ".8b", 0, false, 0 }, { AArch64::ST3Threev4h, "st3", ".4h", 0, false, 0 }, { AArch64::ST3Threev2s, "st3", ".2s", 0, false, 0 }, { AArch64::ST3Threev16b_POST, "st3", ".16b", 1, false, 48 }, { AArch64::ST3Threev8h_POST, "st3", ".8h", 1, false, 48 }, { AArch64::ST3Threev4s_POST, "st3", ".4s", 1, false, 48 }, { AArch64::ST3Threev2d_POST, "st3", ".2d", 1, false, 48 }, { AArch64::ST3Threev8b_POST, "st3", ".8b", 1, false, 24 }, { AArch64::ST3Threev4h_POST, "st3", ".4h", 1, false, 24 }, { AArch64::ST3Threev2s_POST, "st3", ".2s", 1, false, 24 }, { AArch64::ST4i8, "st4", ".b", 0, true, 0 }, { AArch64::ST4i16, "st4", ".h", 0, true, 0 }, { AArch64::ST4i32, "st4", ".s", 0, true, 0 }, { AArch64::ST4i64, "st4", ".d", 0, true, 0 }, { AArch64::ST4i8_POST, "st4", ".b", 1, true, 4 }, { AArch64::ST4i16_POST, "st4", ".h", 1, true, 8 }, { AArch64::ST4i32_POST, "st4", ".s", 1, true, 16 }, { AArch64::ST4i64_POST, "st4", ".d", 1, true, 32 }, { AArch64::ST4Fourv16b, "st4", ".16b", 0, false, 0 }, { AArch64::ST4Fourv8h, "st4", ".8h", 0, false, 0 }, { AArch64::ST4Fourv4s, "st4", ".4s", 0, false, 0 }, { AArch64::ST4Fourv2d, "st4", ".2d", 0, false, 0 }, { AArch64::ST4Fourv8b, "st4", ".8b", 0, false, 0 }, { AArch64::ST4Fourv4h, "st4", ".4h", 0, false, 0 }, { AArch64::ST4Fourv2s, "st4", ".2s", 0, false, 0 }, { AArch64::ST4Fourv16b_POST, "st4", ".16b", 1, false, 64 }, { AArch64::ST4Fourv8h_POST, "st4", ".8h", 1, false, 64 }, { AArch64::ST4Fourv4s_POST, "st4", ".4s", 1, false, 64 }, { AArch64::ST4Fourv2d_POST, "st4", ".2d", 1, false, 64 }, { AArch64::ST4Fourv8b_POST, "st4", ".8b", 1, false, 32 }, { AArch64::ST4Fourv4h_POST, "st4", ".4h", 1, false, 32 }, { AArch64::ST4Fourv2s_POST, "st4", ".2s", 1, false, 32 }, }; static const LdStNInstrDesc *getLdStNInstrDesc(unsigned Opcode) { for (const auto &Info : LdStNInstInfo) if (Info.Opcode == Opcode) return &Info; return nullptr; } void AArch64AppleInstPrinter::printInst(const MCInst *MI, uint64_t Address, StringRef Annot, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Opcode = MI->getOpcode(); StringRef Layout; bool IsTbx; if (isTblTbxInstruction(MI->getOpcode(), Layout, IsTbx)) { O << "\t" << (IsTbx ? "tbx" : "tbl") << Layout << '\t'; printRegName(O, MI->getOperand(0).getReg(), AArch64::vreg); O << ", "; unsigned ListOpNum = IsTbx ? 2 : 1; printVectorList(MI, ListOpNum, STI, O, ""); O << ", "; printRegName(O, MI->getOperand(ListOpNum + 1).getReg(), AArch64::vreg); printAnnotation(O, Annot); return; } if (const LdStNInstrDesc *LdStDesc = getLdStNInstrDesc(Opcode)) { O << "\t" << LdStDesc->Mnemonic << LdStDesc->Layout << '\t'; // Now onto the operands: first a vector list with possible lane // specifier. E.g. { v0 }[2] int OpNum = LdStDesc->ListOperand; printVectorList(MI, OpNum++, STI, O, ""); if (LdStDesc->HasLane) O << '[' << MI->getOperand(OpNum++).getImm() << ']'; // Next the address: [xN] unsigned AddrReg = MI->getOperand(OpNum++).getReg(); O << ", ["; printRegName(O, AddrReg); O << ']'; // Finally, there might be a post-indexed offset. if (LdStDesc->NaturalOffset != 0) { unsigned Reg = MI->getOperand(OpNum++).getReg(); if (Reg != AArch64::XZR) { O << ", "; printRegName(O, Reg); } else { assert(LdStDesc->NaturalOffset && "no offset on post-inc instruction?"); O << ", " << markup("NaturalOffset << markup(">"); } } printAnnotation(O, Annot); return; } AArch64InstPrinter::printInst(MI, Address, Annot, STI, O); } StringRef AArch64AppleInstPrinter::getRegName(MCRegister Reg) const { return getRegisterName(Reg); } bool AArch64InstPrinter::printRangePrefetchAlias(const MCInst *MI, const MCSubtargetInfo &STI, raw_ostream &O, StringRef Annot) { unsigned Opcode = MI->getOpcode(); #ifndef NDEBUG assert(((Opcode == AArch64::PRFMroX) || (Opcode == AArch64::PRFMroW)) && "Invalid opcode for RPRFM alias!"); #endif unsigned PRFOp = MI->getOperand(0).getImm(); unsigned Mask = 0x18; // 0b11000 if ((PRFOp & Mask) != Mask) return false; // Rt != '11xxx', it's a PRFM instruction. unsigned Rm = MI->getOperand(2).getReg(); // "Rm" must be a 64-bit GPR for RPRFM. if (MRI.getRegClass(AArch64::GPR32RegClassID).contains(Rm)) Rm = MRI.getMatchingSuperReg(Rm, AArch64::sub_32, &MRI.getRegClass(AArch64::GPR64RegClassID)); unsigned SignExtend = MI->getOperand(3).getImm(); // encoded in "option<2>". unsigned Shift = MI->getOperand(4).getImm(); // encoded in "S". assert((SignExtend <= 1) && "sign extend should be a single bit!"); assert((Shift <= 1) && "Shift should be a single bit!"); unsigned Option0 = (Opcode == AArch64::PRFMroX) ? 1 : 0; // encoded in "option<2>:option<0>:S:Rt<2:0>". unsigned RPRFOp = (SignExtend << 5) | (Option0 << 4) | (Shift << 3) | (PRFOp & 0x7); O << "\trprfm "; if (auto RPRFM = AArch64RPRFM::lookupRPRFMByEncoding(RPRFOp)) O << RPRFM->Name << ", "; else O << "#" << formatImm(RPRFOp) << ", "; O << getRegisterName(Rm); O << ", ["; printOperand(MI, 1, STI, O); // "Rn". O << "]"; printAnnotation(O, Annot); return true; } bool AArch64InstPrinter::printSysAlias(const MCInst *MI, const MCSubtargetInfo &STI, raw_ostream &O) { #ifndef NDEBUG unsigned Opcode = MI->getOpcode(); assert(Opcode == AArch64::SYSxt && "Invalid opcode for SYS alias!"); #endif const MCOperand &Op1 = MI->getOperand(0); const MCOperand &Cn = MI->getOperand(1); const MCOperand &Cm = MI->getOperand(2); const MCOperand &Op2 = MI->getOperand(3); unsigned Op1Val = Op1.getImm(); unsigned CnVal = Cn.getImm(); unsigned CmVal = Cm.getImm(); unsigned Op2Val = Op2.getImm(); uint16_t Encoding = Op2Val; Encoding |= CmVal << 3; Encoding |= CnVal << 7; Encoding |= Op1Val << 11; bool NeedsReg; std::string Ins; std::string Name; if (CnVal == 7) { switch (CmVal) { default: return false; // Maybe IC, maybe Prediction Restriction case 1: switch (Op1Val) { default: return false; case 0: goto Search_IC; case 3: goto Search_PRCTX; } // Prediction Restriction aliases case 3: { Search_PRCTX: if (Op1Val != 3 || CnVal != 7 || CmVal != 3) return false; const auto Requires = Op2Val == 6 ? AArch64::FeatureSPECRES2 : AArch64::FeaturePredRes; if (!(STI.hasFeature(AArch64::FeatureAll) || STI.hasFeature(Requires))) return false; NeedsReg = true; switch (Op2Val) { default: return false; case 4: Ins = "cfp\t"; break; case 5: Ins = "dvp\t"; break; case 6: Ins = "cosp\t"; break; case 7: Ins = "cpp\t"; break; } Name = "RCTX"; } break; // IC aliases case 5: { Search_IC: const AArch64IC::IC *IC = AArch64IC::lookupICByEncoding(Encoding); if (!IC || !IC->haveFeatures(STI.getFeatureBits())) return false; NeedsReg = IC->NeedsReg; Ins = "ic\t"; Name = std::string(IC->Name); } break; // DC aliases case 4: case 6: case 10: case 11: case 12: case 13: case 14: { const AArch64DC::DC *DC = AArch64DC::lookupDCByEncoding(Encoding); if (!DC || !DC->haveFeatures(STI.getFeatureBits())) return false; NeedsReg = true; Ins = "dc\t"; Name = std::string(DC->Name); } break; // AT aliases case 8: case 9: { const AArch64AT::AT *AT = AArch64AT::lookupATByEncoding(Encoding); if (!AT || !AT->haveFeatures(STI.getFeatureBits())) return false; NeedsReg = true; Ins = "at\t"; Name = std::string(AT->Name); } break; } } else if (CnVal == 8 || CnVal == 9) { // TLBI aliases const AArch64TLBI::TLBI *TLBI = AArch64TLBI::lookupTLBIByEncoding(Encoding); if (!TLBI || !TLBI->haveFeatures(STI.getFeatureBits())) return false; NeedsReg = TLBI->NeedsReg; Ins = "tlbi\t"; Name = std::string(TLBI->Name); } else return false; std::string Str = Ins + Name; std::transform(Str.begin(), Str.end(), Str.begin(), ::tolower); O << '\t' << Str; if (NeedsReg) { O << ", "; printRegName(O, MI->getOperand(4).getReg()); } return true; } bool AArch64InstPrinter::printSyspAlias(const MCInst *MI, const MCSubtargetInfo &STI, raw_ostream &O) { #ifndef NDEBUG unsigned Opcode = MI->getOpcode(); assert((Opcode == AArch64::SYSPxt || Opcode == AArch64::SYSPxt_XZR) && "Invalid opcode for SYSP alias!"); #endif const MCOperand &Op1 = MI->getOperand(0); const MCOperand &Cn = MI->getOperand(1); const MCOperand &Cm = MI->getOperand(2); const MCOperand &Op2 = MI->getOperand(3); unsigned Op1Val = Op1.getImm(); unsigned CnVal = Cn.getImm(); unsigned CmVal = Cm.getImm(); unsigned Op2Val = Op2.getImm(); uint16_t Encoding = Op2Val; Encoding |= CmVal << 3; Encoding |= CnVal << 7; Encoding |= Op1Val << 11; std::string Ins; std::string Name; if (CnVal == 8 || CnVal == 9) { // TLBIP aliases if (CnVal == 9) { if (!STI.hasFeature(AArch64::FeatureXS)) return false; Encoding &= ~(1 << 7); } const AArch64TLBI::TLBI *TLBI = AArch64TLBI::lookupTLBIByEncoding(Encoding); if (!TLBI || !TLBI->haveFeatures(STI.getFeatureBits())) return false; Ins = "tlbip\t"; Name = std::string(TLBI->Name); if (CnVal == 9) Name += "nXS"; } else return false; std::string Str = Ins + Name; std::transform(Str.begin(), Str.end(), Str.begin(), ::tolower); O << '\t' << Str; O << ", "; if (MI->getOperand(4).getReg() == AArch64::XZR) printSyspXzrPair(MI, 4, STI, O); else printGPRSeqPairsClassOperand<64>(MI, 4, STI, O); return true; } template void AArch64InstPrinter::printMatrix(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &RegOp = MI->getOperand(OpNum); assert(RegOp.isReg() && "Unexpected operand type!"); printRegName(O, RegOp.getReg()); switch (EltSize) { case 0: break; case 8: O << ".b"; break; case 16: O << ".h"; break; case 32: O << ".s"; break; case 64: O << ".d"; break; case 128: O << ".q"; break; default: llvm_unreachable("Unsupported element size"); } } template void AArch64InstPrinter::printMatrixTileVector(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &RegOp = MI->getOperand(OpNum); assert(RegOp.isReg() && "Unexpected operand type!"); StringRef RegName = getRegisterName(RegOp.getReg()); // Insert the horizontal/vertical flag before the suffix. StringRef Base, Suffix; std::tie(Base, Suffix) = RegName.split('.'); O << Base << (IsVertical ? "v" : "h") << '.' << Suffix; } void AArch64InstPrinter::printMatrixTile(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &RegOp = MI->getOperand(OpNum); assert(RegOp.isReg() && "Unexpected operand type!"); printRegName(O, RegOp.getReg()); } void AArch64InstPrinter::printSVCROp(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &MO = MI->getOperand(OpNum); assert(MO.isImm() && "Unexpected operand type!"); unsigned svcrop = MO.getImm(); const auto *SVCR = AArch64SVCR::lookupSVCRByEncoding(svcrop); assert(SVCR && "Unexpected SVCR operand!"); O << SVCR->Name; } void AArch64InstPrinter::printOperand(const MCInst *MI, unsigned OpNo, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &Op = MI->getOperand(OpNo); if (Op.isReg()) { unsigned Reg = Op.getReg(); printRegName(O, Reg); } else if (Op.isImm()) { printImm(MI, OpNo, STI, O); } else { assert(Op.isExpr() && "unknown operand kind in printOperand"); Op.getExpr()->print(O, &MAI); } } void AArch64InstPrinter::printImm(const MCInst *MI, unsigned OpNo, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &Op = MI->getOperand(OpNo); O << markup(""); } void AArch64InstPrinter::printImmHex(const MCInst *MI, unsigned OpNo, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &Op = MI->getOperand(OpNo); O << markup(""); } template void AArch64InstPrinter::printSImm(const MCInst *MI, unsigned OpNo, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &Op = MI->getOperand(OpNo); if (Size == 8) O << markup(""); else if (Size == 16) O << markup(""); else O << markup(""); } void AArch64InstPrinter::printPostIncOperand(const MCInst *MI, unsigned OpNo, unsigned Imm, raw_ostream &O) { const MCOperand &Op = MI->getOperand(OpNo); if (Op.isReg()) { unsigned Reg = Op.getReg(); if (Reg == AArch64::XZR) O << markup(""); else printRegName(O, Reg); } else llvm_unreachable("unknown operand kind in printPostIncOperand64"); } void AArch64InstPrinter::printVRegOperand(const MCInst *MI, unsigned OpNo, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &Op = MI->getOperand(OpNo); assert(Op.isReg() && "Non-register vreg operand!"); unsigned Reg = Op.getReg(); printRegName(O, Reg, AArch64::vreg); } void AArch64InstPrinter::printSysCROperand(const MCInst *MI, unsigned OpNo, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &Op = MI->getOperand(OpNo); assert(Op.isImm() && "System instruction C[nm] operands must be immediates!"); O << "c" << Op.getImm(); } void AArch64InstPrinter::printAddSubImm(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &MO = MI->getOperand(OpNum); if (MO.isImm()) { unsigned Val = (MO.getImm() & 0xfff); assert(Val == MO.getImm() && "Add/sub immediate out of range!"); unsigned Shift = AArch64_AM::getShiftValue(MI->getOperand(OpNum + 1).getImm()); O << markup(""); if (Shift != 0) { printShifter(MI, OpNum + 1, STI, O); if (CommentStream) *CommentStream << '=' << formatImm(Val << Shift) << '\n'; } } else { assert(MO.isExpr() && "Unexpected operand type!"); MO.getExpr()->print(O, &MAI); printShifter(MI, OpNum + 1, STI, O); } } template void AArch64InstPrinter::printLogicalImm(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { uint64_t Val = MI->getOperand(OpNum).getImm(); O << markup(""); } void AArch64InstPrinter::printShifter(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Val = MI->getOperand(OpNum).getImm(); // LSL #0 should not be printed. if (AArch64_AM::getShiftType(Val) == AArch64_AM::LSL && AArch64_AM::getShiftValue(Val) == 0) return; O << ", " << AArch64_AM::getShiftExtendName(AArch64_AM::getShiftType(Val)) << " " << markup(""); } void AArch64InstPrinter::printShiftedRegister(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { printRegName(O, MI->getOperand(OpNum).getReg()); printShifter(MI, OpNum + 1, STI, O); } void AArch64InstPrinter::printExtendedRegister(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { printRegName(O, MI->getOperand(OpNum).getReg()); printArithExtend(MI, OpNum + 1, STI, O); } void AArch64InstPrinter::printArithExtend(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Val = MI->getOperand(OpNum).getImm(); AArch64_AM::ShiftExtendType ExtType = AArch64_AM::getArithExtendType(Val); unsigned ShiftVal = AArch64_AM::getArithShiftValue(Val); // If the destination or first source register operand is [W]SP, print // UXTW/UXTX as LSL, and if the shift amount is also zero, print nothing at // all. if (ExtType == AArch64_AM::UXTW || ExtType == AArch64_AM::UXTX) { unsigned Dest = MI->getOperand(0).getReg(); unsigned Src1 = MI->getOperand(1).getReg(); if ( ((Dest == AArch64::SP || Src1 == AArch64::SP) && ExtType == AArch64_AM::UXTX) || ((Dest == AArch64::WSP || Src1 == AArch64::WSP) && ExtType == AArch64_AM::UXTW) ) { if (ShiftVal != 0) O << ", lsl " << markup(""); return; } } O << ", " << AArch64_AM::getShiftExtendName(ExtType); if (ShiftVal != 0) O << " " << markup(""); } static void printMemExtendImpl(bool SignExtend, bool DoShift, unsigned Width, char SrcRegKind, raw_ostream &O, bool UseMarkup) { // sxtw, sxtx, uxtw or lsl (== uxtx) bool IsLSL = !SignExtend && SrcRegKind == 'x'; if (IsLSL) O << "lsl"; else O << (SignExtend ? 's' : 'u') << "xt" << SrcRegKind; if (DoShift || IsLSL) { O << " "; if (UseMarkup) O << ""; } } void AArch64InstPrinter::printMemExtend(const MCInst *MI, unsigned OpNum, raw_ostream &O, char SrcRegKind, unsigned Width) { bool SignExtend = MI->getOperand(OpNum).getImm(); bool DoShift = MI->getOperand(OpNum + 1).getImm(); printMemExtendImpl(SignExtend, DoShift, Width, SrcRegKind, O, UseMarkup); } template void AArch64InstPrinter::printRegWithShiftExtend(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { printOperand(MI, OpNum, STI, O); if (Suffix == 's' || Suffix == 'd') O << '.' << Suffix; else assert(Suffix == 0 && "Unsupported suffix size"); bool DoShift = ExtWidth != 8; if (SignExtend || DoShift || SrcRegKind == 'w') { O << ", "; printMemExtendImpl(SignExtend, DoShift, ExtWidth, SrcRegKind, O, UseMarkup); } } template void AArch64InstPrinter::printPredicateAsCounter(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Reg = MI->getOperand(OpNum).getReg(); assert(Reg <= AArch64::P15 && "Unsupported predicate register"); O << "pn" << (Reg - AArch64::P0); switch (EltSize) { case 0: break; case 8: O << ".b"; break; case 16: O << ".h"; break; case 32: O << ".s"; break; case 64: O << ".d"; break; default: llvm_unreachable("Unsupported element size"); } } void AArch64InstPrinter::printCondCode(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { AArch64CC::CondCode CC = (AArch64CC::CondCode)MI->getOperand(OpNum).getImm(); O << AArch64CC::getCondCodeName(CC); } void AArch64InstPrinter::printInverseCondCode(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { AArch64CC::CondCode CC = (AArch64CC::CondCode)MI->getOperand(OpNum).getImm(); O << AArch64CC::getCondCodeName(AArch64CC::getInvertedCondCode(CC)); } void AArch64InstPrinter::printAMNoIndex(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { O << '['; printRegName(O, MI->getOperand(OpNum).getReg()); O << ']'; } template void AArch64InstPrinter::printImmScale(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { O << markup("getOperand(OpNum).getImm()) << markup(">"); } template void AArch64InstPrinter::printImmRangeScale(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned FirstImm = Scale * MI->getOperand(OpNum).getImm(); O << formatImm(FirstImm); O << ":" << formatImm(FirstImm + Offset); } void AArch64InstPrinter::printUImm12Offset(const MCInst *MI, unsigned OpNum, unsigned Scale, raw_ostream &O) { const MCOperand MO = MI->getOperand(OpNum); if (MO.isImm()) { O << markup(""); } else { assert(MO.isExpr() && "Unexpected operand type!"); MO.getExpr()->print(O, &MAI); } } void AArch64InstPrinter::printAMIndexedWB(const MCInst *MI, unsigned OpNum, unsigned Scale, raw_ostream &O) { const MCOperand MO1 = MI->getOperand(OpNum + 1); O << '['; printRegName(O, MI->getOperand(OpNum).getReg()); if (MO1.isImm()) { O << ", " << markup(""); } else { assert(MO1.isExpr() && "Unexpected operand type!"); O << ", "; MO1.getExpr()->print(O, &MAI); } O << ']'; } void AArch64InstPrinter::printRPRFMOperand(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned prfop = MI->getOperand(OpNum).getImm(); if (auto PRFM = AArch64RPRFM::lookupRPRFMByEncoding(prfop)) { O << PRFM->Name; return; } O << '#' << formatImm(prfop); } template void AArch64InstPrinter::printPrefetchOp(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned prfop = MI->getOperand(OpNum).getImm(); if (IsSVEPrefetch) { if (auto PRFM = AArch64SVEPRFM::lookupSVEPRFMByEncoding(prfop)) { O << PRFM->Name; return; } } else { auto PRFM = AArch64PRFM::lookupPRFMByEncoding(prfop); if (PRFM && PRFM->haveFeatures(STI.getFeatureBits())) { O << PRFM->Name; return; } } O << markup(""); } void AArch64InstPrinter::printPSBHintOp(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned psbhintop = MI->getOperand(OpNum).getImm(); auto PSB = AArch64PSBHint::lookupPSBByEncoding(psbhintop); if (PSB) O << PSB->Name; else O << markup(""); } void AArch64InstPrinter::printBTIHintOp(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned btihintop = MI->getOperand(OpNum).getImm() ^ 32; auto BTI = AArch64BTIHint::lookupBTIByEncoding(btihintop); if (BTI) O << BTI->Name; else O << markup(""); } void AArch64InstPrinter::printFPImmOperand(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &MO = MI->getOperand(OpNum); float FPImm = MO.isDFPImm() ? bit_cast(MO.getDFPImm()) : AArch64_AM::getFPImmFloat(MO.getImm()); // 8 decimal places are enough to perfectly represent permitted floats. O << markup(""); } static unsigned getNextVectorRegister(unsigned Reg, unsigned Stride = 1) { while (Stride--) { switch (Reg) { default: llvm_unreachable("Vector register expected!"); case AArch64::Q0: Reg = AArch64::Q1; break; case AArch64::Q1: Reg = AArch64::Q2; break; case AArch64::Q2: Reg = AArch64::Q3; break; case AArch64::Q3: Reg = AArch64::Q4; break; case AArch64::Q4: Reg = AArch64::Q5; break; case AArch64::Q5: Reg = AArch64::Q6; break; case AArch64::Q6: Reg = AArch64::Q7; break; case AArch64::Q7: Reg = AArch64::Q8; break; case AArch64::Q8: Reg = AArch64::Q9; break; case AArch64::Q9: Reg = AArch64::Q10; break; case AArch64::Q10: Reg = AArch64::Q11; break; case AArch64::Q11: Reg = AArch64::Q12; break; case AArch64::Q12: Reg = AArch64::Q13; break; case AArch64::Q13: Reg = AArch64::Q14; break; case AArch64::Q14: Reg = AArch64::Q15; break; case AArch64::Q15: Reg = AArch64::Q16; break; case AArch64::Q16: Reg = AArch64::Q17; break; case AArch64::Q17: Reg = AArch64::Q18; break; case AArch64::Q18: Reg = AArch64::Q19; break; case AArch64::Q19: Reg = AArch64::Q20; break; case AArch64::Q20: Reg = AArch64::Q21; break; case AArch64::Q21: Reg = AArch64::Q22; break; case AArch64::Q22: Reg = AArch64::Q23; break; case AArch64::Q23: Reg = AArch64::Q24; break; case AArch64::Q24: Reg = AArch64::Q25; break; case AArch64::Q25: Reg = AArch64::Q26; break; case AArch64::Q26: Reg = AArch64::Q27; break; case AArch64::Q27: Reg = AArch64::Q28; break; case AArch64::Q28: Reg = AArch64::Q29; break; case AArch64::Q29: Reg = AArch64::Q30; break; case AArch64::Q30: Reg = AArch64::Q31; break; // Vector lists can wrap around. case AArch64::Q31: Reg = AArch64::Q0; break; case AArch64::Z0: Reg = AArch64::Z1; break; case AArch64::Z1: Reg = AArch64::Z2; break; case AArch64::Z2: Reg = AArch64::Z3; break; case AArch64::Z3: Reg = AArch64::Z4; break; case AArch64::Z4: Reg = AArch64::Z5; break; case AArch64::Z5: Reg = AArch64::Z6; break; case AArch64::Z6: Reg = AArch64::Z7; break; case AArch64::Z7: Reg = AArch64::Z8; break; case AArch64::Z8: Reg = AArch64::Z9; break; case AArch64::Z9: Reg = AArch64::Z10; break; case AArch64::Z10: Reg = AArch64::Z11; break; case AArch64::Z11: Reg = AArch64::Z12; break; case AArch64::Z12: Reg = AArch64::Z13; break; case AArch64::Z13: Reg = AArch64::Z14; break; case AArch64::Z14: Reg = AArch64::Z15; break; case AArch64::Z15: Reg = AArch64::Z16; break; case AArch64::Z16: Reg = AArch64::Z17; break; case AArch64::Z17: Reg = AArch64::Z18; break; case AArch64::Z18: Reg = AArch64::Z19; break; case AArch64::Z19: Reg = AArch64::Z20; break; case AArch64::Z20: Reg = AArch64::Z21; break; case AArch64::Z21: Reg = AArch64::Z22; break; case AArch64::Z22: Reg = AArch64::Z23; break; case AArch64::Z23: Reg = AArch64::Z24; break; case AArch64::Z24: Reg = AArch64::Z25; break; case AArch64::Z25: Reg = AArch64::Z26; break; case AArch64::Z26: Reg = AArch64::Z27; break; case AArch64::Z27: Reg = AArch64::Z28; break; case AArch64::Z28: Reg = AArch64::Z29; break; case AArch64::Z29: Reg = AArch64::Z30; break; case AArch64::Z30: Reg = AArch64::Z31; break; // Vector lists can wrap around. case AArch64::Z31: Reg = AArch64::Z0; break; case AArch64::P0: Reg = AArch64::P1; break; case AArch64::P1: Reg = AArch64::P2; break; case AArch64::P2: Reg = AArch64::P3; break; case AArch64::P3: Reg = AArch64::P4; break; case AArch64::P4: Reg = AArch64::P5; break; case AArch64::P5: Reg = AArch64::P6; break; case AArch64::P6: Reg = AArch64::P7; break; case AArch64::P7: Reg = AArch64::P8; break; case AArch64::P8: Reg = AArch64::P9; break; case AArch64::P9: Reg = AArch64::P10; break; case AArch64::P10: Reg = AArch64::P11; break; case AArch64::P11: Reg = AArch64::P12; break; case AArch64::P12: Reg = AArch64::P13; break; case AArch64::P13: Reg = AArch64::P14; break; case AArch64::P14: Reg = AArch64::P15; break; // Vector lists can wrap around. case AArch64::P15: Reg = AArch64::P0; break; } } return Reg; } template void AArch64InstPrinter::printGPRSeqPairsClassOperand(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { static_assert(size == 64 || size == 32, "Template parameter must be either 32 or 64"); unsigned Reg = MI->getOperand(OpNum).getReg(); unsigned Sube = (size == 32) ? AArch64::sube32 : AArch64::sube64; unsigned Subo = (size == 32) ? AArch64::subo32 : AArch64::subo64; unsigned Even = MRI.getSubReg(Reg, Sube); unsigned Odd = MRI.getSubReg(Reg, Subo); printRegName(O, Even); O << ", "; printRegName(O, Odd); } void AArch64InstPrinter::printMatrixTileList(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned MaxRegs = 8; unsigned RegMask = MI->getOperand(OpNum).getImm(); unsigned NumRegs = 0; for (unsigned I = 0; I < MaxRegs; ++I) if ((RegMask & (1 << I)) != 0) ++NumRegs; O << "{"; unsigned Printed = 0; for (unsigned I = 0; I < MaxRegs; ++I) { unsigned Reg = RegMask & (1 << I); if (Reg == 0) continue; printRegName(O, AArch64::ZAD0 + I); if (Printed + 1 != NumRegs) O << ", "; ++Printed; } O << "}"; } void AArch64InstPrinter::printVectorList(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O, StringRef LayoutSuffix) { unsigned Reg = MI->getOperand(OpNum).getReg(); O << "{ "; // Work out how many registers there are in the list (if there is an actual // list). unsigned NumRegs = 1; if (MRI.getRegClass(AArch64::DDRegClassID).contains(Reg) || MRI.getRegClass(AArch64::ZPR2RegClassID).contains(Reg) || MRI.getRegClass(AArch64::QQRegClassID).contains(Reg) || MRI.getRegClass(AArch64::PPR2RegClassID).contains(Reg) || MRI.getRegClass(AArch64::ZPR2StridedRegClassID).contains(Reg)) NumRegs = 2; else if (MRI.getRegClass(AArch64::DDDRegClassID).contains(Reg) || MRI.getRegClass(AArch64::ZPR3RegClassID).contains(Reg) || MRI.getRegClass(AArch64::QQQRegClassID).contains(Reg)) NumRegs = 3; else if (MRI.getRegClass(AArch64::DDDDRegClassID).contains(Reg) || MRI.getRegClass(AArch64::ZPR4RegClassID).contains(Reg) || MRI.getRegClass(AArch64::QQQQRegClassID).contains(Reg) || MRI.getRegClass(AArch64::ZPR4StridedRegClassID).contains(Reg)) NumRegs = 4; unsigned Stride = 1; if (MRI.getRegClass(AArch64::ZPR2StridedRegClassID).contains(Reg)) Stride = 8; else if (MRI.getRegClass(AArch64::ZPR4StridedRegClassID).contains(Reg)) Stride = 4; // Now forget about the list and find out what the first register is. if (unsigned FirstReg = MRI.getSubReg(Reg, AArch64::dsub0)) Reg = FirstReg; else if (unsigned FirstReg = MRI.getSubReg(Reg, AArch64::qsub0)) Reg = FirstReg; else if (unsigned FirstReg = MRI.getSubReg(Reg, AArch64::zsub0)) Reg = FirstReg; else if (unsigned FirstReg = MRI.getSubReg(Reg, AArch64::psub0)) Reg = FirstReg; // If it's a D-reg, we need to promote it to the equivalent Q-reg before // printing (otherwise getRegisterName fails). if (MRI.getRegClass(AArch64::FPR64RegClassID).contains(Reg)) { const MCRegisterClass &FPR128RC = MRI.getRegClass(AArch64::FPR128RegClassID); Reg = MRI.getMatchingSuperReg(Reg, AArch64::dsub, &FPR128RC); } if ((MRI.getRegClass(AArch64::ZPRRegClassID).contains(Reg) || MRI.getRegClass(AArch64::PPRRegClassID).contains(Reg)) && NumRegs > 1 && Stride == 1 && // Do not print the range when the last register is lower than the first. // Because it is a wrap-around register. Reg < getNextVectorRegister(Reg, NumRegs - 1)) { printRegName(O, Reg); O << LayoutSuffix; if (NumRegs > 1) { // Set of two sve registers should be separated by ',' StringRef split_char = NumRegs == 2 ? ", " : " - "; O << split_char; printRegName(O, (getNextVectorRegister(Reg, NumRegs - 1))); O << LayoutSuffix; } } else { for (unsigned i = 0; i < NumRegs; ++i, Reg = getNextVectorRegister(Reg, Stride)) { // wrap-around sve register if (MRI.getRegClass(AArch64::ZPRRegClassID).contains(Reg) || MRI.getRegClass(AArch64::PPRRegClassID).contains(Reg)) printRegName(O, Reg); else printRegName(O, Reg, AArch64::vreg); O << LayoutSuffix; if (i + 1 != NumRegs) O << ", "; } } O << " }"; } void AArch64InstPrinter::printImplicitlyTypedVectorList(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { printVectorList(MI, OpNum, STI, O, ""); } template void AArch64InstPrinter::printTypedVectorList(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { std::string Suffix("."); if (NumLanes) Suffix += itostr(NumLanes) + LaneKind; else Suffix += LaneKind; printVectorList(MI, OpNum, STI, O, Suffix); } template void AArch64InstPrinter::printVectorIndex(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { O << "[" << Scale * MI->getOperand(OpNum).getImm() << "]"; } void AArch64InstPrinter::printMatrixIndex(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { O << MI->getOperand(OpNum).getImm(); } void AArch64InstPrinter::printAlignedLabel(const MCInst *MI, uint64_t Address, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &Op = MI->getOperand(OpNum); // If the label has already been resolved to an immediate offset (say, when // we're running the disassembler), just print the immediate. if (Op.isImm()) { O << markup(""); return; } // If the branch target is simply an address then print it in hex. const MCConstantExpr *BranchTarget = dyn_cast(MI->getOperand(OpNum).getExpr()); int64_t TargetAddress; if (BranchTarget && BranchTarget->evaluateAsAbsolute(TargetAddress)) { O << formatHex((uint64_t)TargetAddress); } else { // Otherwise, just print the expression. MI->getOperand(OpNum).getExpr()->print(O, &MAI); } } void AArch64InstPrinter::printAdrpLabel(const MCInst *MI, uint64_t Address, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { const MCOperand &Op = MI->getOperand(OpNum); // If the label has already been resolved to an immediate offset (say, when // we're running the disassembler), just print the immediate. if (Op.isImm()) { const int64_t Offset = Op.getImm() * 4096; O << markup(""); return; } // Otherwise, just print the expression. MI->getOperand(OpNum).getExpr()->print(O, &MAI); } void AArch64InstPrinter::printBarrierOption(const MCInst *MI, unsigned OpNo, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Val = MI->getOperand(OpNo).getImm(); unsigned Opcode = MI->getOpcode(); StringRef Name; if (Opcode == AArch64::ISB) { auto ISB = AArch64ISB::lookupISBByEncoding(Val); Name = ISB ? ISB->Name : ""; } else if (Opcode == AArch64::TSB) { auto TSB = AArch64TSB::lookupTSBByEncoding(Val); Name = TSB ? TSB->Name : ""; } else { auto DB = AArch64DB::lookupDBByEncoding(Val); Name = DB ? DB->Name : ""; } if (!Name.empty()) O << Name; else O << markup(""); } void AArch64InstPrinter::printBarriernXSOption(const MCInst *MI, unsigned OpNo, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Val = MI->getOperand(OpNo).getImm(); assert(MI->getOpcode() == AArch64::DSBnXS); StringRef Name; auto DB = AArch64DBnXS::lookupDBnXSByEncoding(Val); Name = DB ? DB->Name : ""; if (!Name.empty()) O << Name; else O << markup(""); } static bool isValidSysReg(const AArch64SysReg::SysReg *Reg, bool Read, const MCSubtargetInfo &STI) { return (Reg && (Read ? Reg->Readable : Reg->Writeable) && Reg->haveFeatures(STI.getFeatureBits())); } // Looks up a system register either by encoding or by name. Some system // registers share the same encoding between different architectures, // therefore a tablegen lookup by encoding will return an entry regardless // of the register's predication on a specific subtarget feature. To work // around this problem we keep an alternative name for such registers and // look them up by that name if the first lookup was unsuccessful. static const AArch64SysReg::SysReg *lookupSysReg(unsigned Val, bool Read, const MCSubtargetInfo &STI) { const AArch64SysReg::SysReg *Reg = AArch64SysReg::lookupSysRegByEncoding(Val); if (Reg && !isValidSysReg(Reg, Read, STI)) Reg = AArch64SysReg::lookupSysRegByName(Reg->AltName); return Reg; } void AArch64InstPrinter::printMRSSystemRegister(const MCInst *MI, unsigned OpNo, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Val = MI->getOperand(OpNo).getImm(); // Horrible hack for the one register that has identical encodings but // different names in MSR and MRS. Because of this, one of MRS and MSR is // going to get the wrong entry if (Val == AArch64SysReg::DBGDTRRX_EL0) { O << "DBGDTRRX_EL0"; return; } // Horrible hack for two different registers having the same encoding. if (Val == AArch64SysReg::TRCEXTINSELR) { O << "TRCEXTINSELR"; return; } const AArch64SysReg::SysReg *Reg = lookupSysReg(Val, true /*Read*/, STI); if (isValidSysReg(Reg, true /*Read*/, STI)) O << Reg->Name; else O << AArch64SysReg::genericRegisterString(Val); } void AArch64InstPrinter::printMSRSystemRegister(const MCInst *MI, unsigned OpNo, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Val = MI->getOperand(OpNo).getImm(); // Horrible hack for the one register that has identical encodings but // different names in MSR and MRS. Because of this, one of MRS and MSR is // going to get the wrong entry if (Val == AArch64SysReg::DBGDTRTX_EL0) { O << "DBGDTRTX_EL0"; return; } // Horrible hack for two different registers having the same encoding. if (Val == AArch64SysReg::TRCEXTINSELR) { O << "TRCEXTINSELR"; return; } const AArch64SysReg::SysReg *Reg = lookupSysReg(Val, false /*Read*/, STI); if (isValidSysReg(Reg, false /*Read*/, STI)) O << Reg->Name; else O << AArch64SysReg::genericRegisterString(Val); } void AArch64InstPrinter::printSystemPStateField(const MCInst *MI, unsigned OpNo, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Val = MI->getOperand(OpNo).getImm(); auto PStateImm15 = AArch64PState::lookupPStateImm0_15ByEncoding(Val); auto PStateImm1 = AArch64PState::lookupPStateImm0_1ByEncoding(Val); if (PStateImm15 && PStateImm15->haveFeatures(STI.getFeatureBits())) O << PStateImm15->Name; else if (PStateImm1 && PStateImm1->haveFeatures(STI.getFeatureBits())) O << PStateImm1->Name; else O << "#" << formatImm(Val); } void AArch64InstPrinter::printSIMDType10Operand(const MCInst *MI, unsigned OpNo, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned RawVal = MI->getOperand(OpNo).getImm(); uint64_t Val = AArch64_AM::decodeAdvSIMDModImmType10(RawVal); O << markup(""); } template void AArch64InstPrinter::printComplexRotationOp(const MCInst *MI, unsigned OpNo, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Val = MI->getOperand(OpNo).getImm(); O << markup(""); } void AArch64InstPrinter::printSVEPattern(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Val = MI->getOperand(OpNum).getImm(); if (auto Pat = AArch64SVEPredPattern::lookupSVEPREDPATByEncoding(Val)) O << Pat->Name; else O << markup(""); } void AArch64InstPrinter::printSVEVecLenSpecifier(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Val = MI->getOperand(OpNum).getImm(); // Pattern has only 1 bit if (Val > 1) llvm_unreachable("Invalid vector length specifier"); if (auto Pat = AArch64SVEVecLenSpecifier::lookupSVEVECLENSPECIFIERByEncoding(Val)) O << Pat->Name; else llvm_unreachable("Invalid vector length specifier"); } template void AArch64InstPrinter::printSVERegOp(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { switch (suffix) { case 0: case 'b': case 'h': case 's': case 'd': case 'q': break; default: llvm_unreachable("Invalid kind specifier."); } unsigned Reg = MI->getOperand(OpNum).getReg(); printRegName(O, Reg); if (suffix != 0) O << '.' << suffix; } template void AArch64InstPrinter::printImmSVE(T Value, raw_ostream &O) { std::make_unsigned_t HexValue = Value; if (getPrintImmHex()) O << markup(""); else O << markup(""); if (CommentStream) { // Do the opposite to that used for instruction operands. if (getPrintImmHex()) *CommentStream << '=' << formatDec(HexValue) << '\n'; else *CommentStream << '=' << formatHex((uint64_t)Value) << '\n'; } } template void AArch64InstPrinter::printImm8OptLsl(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned UnscaledVal = MI->getOperand(OpNum).getImm(); unsigned Shift = MI->getOperand(OpNum + 1).getImm(); assert(AArch64_AM::getShiftType(Shift) == AArch64_AM::LSL && "Unexepected shift type!"); // #0 lsl #8 is never pretty printed if ((UnscaledVal == 0) && (AArch64_AM::getShiftValue(Shift) != 0)) { O << markup(""); printShifter(MI, OpNum + 1, STI, O); return; } T Val; if (std::is_signed()) Val = (int8_t)UnscaledVal * (1 << AArch64_AM::getShiftValue(Shift)); else Val = (uint8_t)UnscaledVal * (1 << AArch64_AM::getShiftValue(Shift)); printImmSVE(Val, O); } template void AArch64InstPrinter::printSVELogicalImm(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { typedef std::make_signed_t SignedT; typedef std::make_unsigned_t UnsignedT; uint64_t Val = MI->getOperand(OpNum).getImm(); UnsignedT PrintVal = AArch64_AM::decodeLogicalImmediate(Val, 64); // Prefer the default format for 16bit values, hex otherwise. if ((int16_t)PrintVal == (SignedT)PrintVal) printImmSVE((T)PrintVal, O); else if ((uint16_t)PrintVal == PrintVal) printImmSVE(PrintVal, O); else O << markup(""); } template void AArch64InstPrinter::printZPRasFPR(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Base; switch (Width) { case 8: Base = AArch64::B0; break; case 16: Base = AArch64::H0; break; case 32: Base = AArch64::S0; break; case 64: Base = AArch64::D0; break; case 128: Base = AArch64::Q0; break; default: llvm_unreachable("Unsupported width"); } unsigned Reg = MI->getOperand(OpNum).getReg(); printRegName(O, Reg - AArch64::Z0 + Base); } template void AArch64InstPrinter::printExactFPImm(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { auto *Imm0Desc = AArch64ExactFPImm::lookupExactFPImmByEnum(ImmIs0); auto *Imm1Desc = AArch64ExactFPImm::lookupExactFPImmByEnum(ImmIs1); unsigned Val = MI->getOperand(OpNum).getImm(); O << markup("Repr : Imm0Desc->Repr) << markup(">"); } void AArch64InstPrinter::printGPR64as32(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Reg = MI->getOperand(OpNum).getReg(); printRegName(O, getWRegFromXReg(Reg)); } void AArch64InstPrinter::printGPR64x8(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Reg = MI->getOperand(OpNum).getReg(); printRegName(O, MRI.getSubReg(Reg, AArch64::x8sub_0)); } void AArch64InstPrinter::printSyspXzrPair(const MCInst *MI, unsigned OpNum, const MCSubtargetInfo &STI, raw_ostream &O) { unsigned Reg = MI->getOperand(OpNum).getReg(); assert(Reg == AArch64::XZR && "MC representation of SyspXzrPair should be XZR"); O << getRegisterName(Reg) << ", " << getRegisterName(Reg); }