/netbsd/external/apache2/llvm/dist/llvm/include/llvm/CodeGen/GlobalISel/ |
H A D | MIPatternMatch.h | 387 SrcTy L; 402 template <typename SrcTy> 408 template <typename SrcTy> 409 inline UnaryOp_match<SrcTy, TargetOpcode::G_SEXT> m_GSExt(const SrcTy &Src) { 414 inline UnaryOp_match<SrcTy, TargetOpcode::G_ZEXT> m_GZExt(const SrcTy &Src) { 419 inline UnaryOp_match<SrcTy, TargetOpcode::G_FPEXT> m_GFPExt(const SrcTy &Src) { 424 inline UnaryOp_match<SrcTy, TargetOpcode::G_TRUNC> m_GTrunc(const SrcTy &Src) { 453 inline UnaryOp_match<SrcTy, TargetOpcode::G_FABS> m_GFabs(const SrcTy &Src) { 458 inline UnaryOp_match<SrcTy, TargetOpcode::G_FNEG> m_GFNeg(const SrcTy &Src) { 463 inline UnaryOp_match<SrcTy, TargetOpcode::COPY> m_Copy(SrcTy &&Src) { [all …]
|
H A D | LegalizationArtifactCombiner.h | 120 LLT SrcTy = MRI.getType(SrcReg); in tryCombineZExt() local 121 APInt MaskVal = APInt::getAllOnesValue(SrcTy.getScalarSizeInBits()); in tryCombineZExt() 175 LLT SrcTy = MRI.getType(SrcReg); in tryCombineSExt() local 176 uint64_t SizeInBits = SrcTy.getScalarSizeInBits(); in tryCombineSExt() 369 const LLT SrcTy = MRI.getType(MI.getOperand(NumDefs).getReg()); in tryFoldUnmergeCast() local 375 if (SrcTy.isVector() && SrcTy.getScalarType() == DestTy.getScalarType()) { in tryFoldUnmergeCast() 406 if (CastSrcTy.isScalar() && SrcTy.isScalar() && !DestTy.isVector()) { in tryFoldUnmergeCast() 765 LLT SrcTy = MRI.getType(SrcReg); in tryCombineExtract() local 771 unsigned MergeSrcSize = SrcTy.getSizeInBits() / NumMergeSrcs; in tryCombineExtract()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Transforms/InstCombine/ |
H A D | InstCombineCasts.cpp | 252 Type *SrcTy = CI1->getSrcTy(); in isEliminableCastPair() local 259 SrcTy->isPtrOrPtrVectorTy() ? DL.getIntPtrType(SrcTy) : nullptr; in isEliminableCastPair() 623 Type *SrcTy = Trunc.getSrcTy(); in narrowBinOp() local 625 if (!isa<VectorType>(SrcTy) && !shouldChangeType(SrcTy, DestTy)) in narrowBinOp() 885 (isa<VectorType>(SrcTy) || shouldChangeType(SrcTy, DestTy))) { in visitTrunc() 1640 Type *SrcTy = Src->getType(); in isKnownExactCastIntToFP() local 1986 Type *SrcTy = SrcOp->getType(); in visitPtrToInt() local 2045 SrcTy = in optimizeVectorResizeWithIntegerBitCasts() 2073 V2 = UndefValue::get(SrcTy); in optimizeVectorResizeWithIntegerBitCasts() 2556 Type *SrcTy = Src->getType(); in visitBitCast() local [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/CodeGen/GlobalISel/ |
H A D | LegalizerHelper.cpp | 1518 SrcTy = LLT::scalar(SrcTy.getSizeInBits()); in widenScalarUnmergeValues() 3991 SrcTy = LLT::scalar(SrcTy.getSizeInBits()); in reduceOperationWidth() 5141 if (SrcTy.isScalar() && SrcTy.getSizeInBits() == 2 * NarrowSize) { in narrowScalarCTLZ() 5177 if (SrcTy.isScalar() && SrcTy.getSizeInBits() == 2 * NarrowSize) { in narrowScalarCTTZ() 5212 if (SrcTy.isScalar() && SrcTy.getSizeInBits() == 2 * NarrowSize) { in narrowScalarCTPOP() 5278 SrcTy, Op, MIRBuilder.buildLShr(SrcTy, Op, MIBShiftAmt)); in lowerBitCount() 5321 if (!isSupported({TargetOpcode::G_CTPOP, {SrcTy, SrcTy}}) && in lowerBitCount() 5322 isSupported({TargetOpcode::G_CTLZ, {SrcTy, SrcTy}})) { in lowerBitCount() 5697 if (SrcTy != S64) in lowerSITOFP() 5733 if (SrcTy != S64 && SrcTy != S32) in lowerFPTOUI() [all …]
|
H A D | CallLowering.cpp | 460 if (PartTy.isVector() == SrcTy.isVector() && in buildCopyToRegs() 467 if (SrcTy.isVector() && !PartTy.isVector() && in buildCopyToRegs() 476 LLT GCDTy = getGCDType(SrcTy, PartTy); in buildCopyToRegs() 485 LLT LCMTy = getLCMType(SrcTy, PartTy); in buildCopyToRegs() 489 const unsigned SrcSize = SrcTy.getSizeInBits(); in buildCopyToRegs() 494 Register Undef = B.buildUndef(SrcTy).getReg(0); in buildCopyToRegs() 1123 if (SrcTy == DstTy) in isCopyCompatibleType() 1126 if (SrcTy.getSizeInBits() != DstTy.getSizeInBits()) in isCopyCompatibleType() 1129 SrcTy = SrcTy.getScalarType(); in isCopyCompatibleType() 1132 return (SrcTy.isPointer() && DstTy.isScalar()) || in isCopyCompatibleType() [all …]
|
H A D | GISelKnownBits.cpp | 435 LLT SrcTy = MRI.getType(SrcReg); in computeKnownBitsImpl() local 442 SrcBitWidth = SrcTy.isPointer() in computeKnownBitsImpl() 443 ? DL.getIndexSizeInBits(SrcTy.getAddressSpace()) in computeKnownBitsImpl() 444 : SrcTy.getSizeInBits(); in computeKnownBitsImpl() 557 LLT SrcTy = MRI.getType(Src); in computeNumSignBits() local 558 unsigned Tmp = DstTy.getScalarSizeInBits() - SrcTy.getScalarSizeInBits(); in computeNumSignBits() 589 LLT SrcTy = MRI.getType(Src); in computeNumSignBits() local 593 unsigned NumSrcBits = SrcTy.getScalarSizeInBits(); in computeNumSignBits()
|
H A D | MachineIRBuilder.cpp | 500 LLT SrcTy = Src.getLLTTy(*getMRI()); in buildCast() local 502 if (SrcTy == DstTy) in buildCast() 506 if (SrcTy.isPointer() && DstTy.isScalar()) in buildCast() 508 else if (DstTy.isPointer() && SrcTy.isScalar()) in buildCast() 521 LLT SrcTy = Src.getLLTTy(*getMRI()); in buildExtract() local 525 assert(SrcTy.isValid() && "invalid operand type"); in buildExtract() 527 assert(Index + DstTy.getSizeInBits() <= SrcTy.getSizeInBits() && in buildExtract() 531 if (DstTy.getSizeInBits() == SrcTy.getSizeInBits()) { in buildExtract() 965 assert(SrcTy.getNumElements() == DstTy.getNumElements() && in validateTruncExt() 971 assert(DstTy.getSizeInBits() > SrcTy.getSizeInBits() && in validateTruncExt() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/AArch64/GISel/ |
H A D | AArch64RegisterBankInfo.cpp | 597 LLT SrcTy = MRI.getType(MI.getOperand(1).getReg()); in getInstrMapping() local 608 LLT SrcTy = MRI.getType(MI.getOperand(1).getReg()); in getInstrMapping() local 643 LLT SrcTy = MRI.getType(MI.getOperand(1).getReg()); in getInstrMapping() local 646 bool SrcIsGPR = !SrcTy.isVector() && SrcTy.getSizeInBits() <= 64; in getInstrMapping() 701 LLT SrcTy = MRI.getType(MI.getOperand(1).getReg()); in getInstrMapping() local 702 if (!SrcTy.isVector() && SrcTy.getSizeInBits() == 128) in getInstrMapping() 788 if (SrcTy.isVector()) { in getInstrMapping() 845 if (SrcTy.isVector() || SrcTy == LLT::scalar(128) || in getInstrMapping() 880 if (SrcTy.getSizeInBits() != 128) in getInstrMapping() 905 const LLT SrcTy = MRI.getType(VReg); in getInstrMapping() local [all …]
|
H A D | AArch64LegalizerInfo.cpp | 125 const auto &SrcTy = Query.Types[0]; in AArch64LegalizerInfo() local 127 return !SrcTy.isVector() && SrcTy.getSizeInBits() == 32 && in AArch64LegalizerInfo() 374 const LLT &SrcTy = Query.Types[1]; in AArch64LegalizerInfo() local 400 const LLT &SrcTy = Query.Types[1]; in AArch64LegalizerInfo() local 403 if (SrcTy == s1) in AArch64LegalizerInfo() 409 unsigned SrcSize = SrcTy.getSizeInBits(); in AArch64LegalizerInfo() 425 LLT SrcTy = Query.Types[1]; in AArch64LegalizerInfo() local 664 const LLT &SrcTy = Query.Types[1]; in AArch64LegalizerInfo() local 667 if (DstTy != SrcTy) in AArch64LegalizerInfo() 824 LLT SrcTy = MRI.getType(SrcReg); in legalizeVectorTrunc() local [all …]
|
H A D | AArch64InstructionSelector.cpp | 1855 if (SrcTy.isVector()) in preISelLower() 1919 LLT SrcTy = MRI.getType(SrcReg); in preISelLower() local 1921 if (SrcTy.isVector() || SrcTy.getSizeInBits() != DstTy.getSizeInBits()) in preISelLower() 2459 LLT SrcTy = MRI.getType(SrcReg); in select() local 2869 if (!SrcTy.isVector() && SrcTy.getSizeInBits() == 128) { in select() 3354 if (!SrcTy.isVector()) { in selectIntrinsicTrunc() 3355 switch (SrcTy.getSizeInBits()) { in selectIntrinsicTrunc() 3409 if (!SrcTy.isVector()) { in selectIntrinsicRound() 3410 switch (SrcTy.getSizeInBits()) { in selectIntrinsicRound() 3464 LLT SrcTy = MRI.getType(SrcReg); in selectVectorICmp() local [all …]
|
H A D | AArch64PostLegalizerLowering.cpp | 648 const LLT SrcTy = MRI.getType(Src1Reg); in matchDupLane() local 656 if (*LaneIdx >= SrcTy.getNumElements()) in matchDupLane() 659 if (DstTy != SrcTy) in matchDupLane() 662 LLT ScalarTy = SrcTy.getElementType(); in matchDupLane() 666 switch (SrcTy.getNumElements()) { in matchDupLane() 700 const LLT SrcTy = MRI.getType(Src1Reg); in applyDupLane() local 708 if (SrcTy == LLT::vector(2, LLT::scalar(32))) { in applyDupLane() 711 auto Undef = B.buildUndef(SrcTy); in applyDupLane() 712 DupSrc = B.buildConcatVectors(SrcTy.changeNumElements(4), in applyDupLane()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Linker/ |
H A D | IRMover.cpp | 57 void addTypeMapping(Type *DstTy, Type *SrcTy); 65 Type *get(Type *SrcTy); 75 Type *remapType(Type *SrcTy) override { return get(SrcTy); } in remapType() argument 87 if (!areTypesIsomorphic(DstTy, SrcTy)) { in addTypeMapping() 117 if (DstTy->getTypeID() != SrcTy->getTypeID()) in areTypesIsomorphic() 121 Type *&Entry = MappedTypes[SrcTy]; in areTypesIsomorphic() 127 if (DstTy == SrcTy) { in areTypesIsomorphic() 139 SpeculativeTypes.push_back(SrcTy); in areTypesIsomorphic() 152 SpeculativeTypes.push_back(SrcTy); in areTypesIsomorphic() 173 StructType *SSTy = cast<StructType>(SrcTy); in areTypesIsomorphic() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/CodeGen/ |
H A D | MachineVerifier.cpp | 961 LLT SrcTy = MRI->getType(Src); in verifyPreISelGenericInstruction() local 1077 if (SrcTy == DstTy) in verifyPreISelGenericInstruction() 1093 SrcTy = SrcTy.getScalarType(); in verifyPreISelGenericInstruction() 1098 if (SrcTy.isPointer()) in verifyPreISelGenericInstruction() 1101 if (!SrcTy.isPointer()) in verifyPreISelGenericInstruction() 1167 LLT SrcElTy = SrcTy.getScalarType(); in verifyPreISelGenericInstruction() 1229 if (SrcTy.getSizeInBits() != in verifyPreISelGenericInstruction() 1418 if (Imm >= SrcTy.getScalarSizeInBits()) in verifyPreISelGenericInstruction() 1567 if (!SrcTy.isVector()) in verifyPreISelGenericInstruction() 1685 LLT SrcTy = MRI->getType(SrcReg); in visitMachineInstrBefore() local [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/ARM/ |
H A D | ARMTargetTransformInfo.cpp | 417 EVT SrcTy = TLI->getValueType(DL, Src); in getCastInstrCost() local 420 if (!SrcTy.isSimple() || !DstTy.isSimple()) in getCastInstrCost() 653 if (SrcTy.isVector() && ST->hasNEON()) { in getCastInstrCost() 714 if (SrcTy.isInteger() && ST->hasNEON()) { in getCastInstrCost() 753 if (SrcTy.isFixedLengthVector()) in getCastInstrCost() 754 Lanes = SrcTy.getVectorNumElements(); in getCastInstrCost() 763 SrcTy.isFixedLengthVector()) { in getCastInstrCost() 766 if ((SrcTy.getScalarType() == MVT::i8 || in getCastInstrCost() 767 SrcTy.getScalarType() == MVT::i16 || in getCastInstrCost() 769 SrcTy.getSizeInBits() > 128 && in getCastInstrCost() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/IR/ |
H A D | Instructions.cpp | 2670 if (SrcTy == DstTy) in isLosslessCast() 2674 if (SrcTy->isPointerTy()) in isLosslessCast() 2828 if (SrcTy->isIntegerTy()) in isEliminableCastPair() 2834 if (SrcTy->isFloatingPointTy()) in isEliminableCastPair() 3170 if (SrcTy == DestTy) in isBitCastable() 3230 Type *SrcTy = Src->getType(); in getCastOpcode() local 3235 if (SrcTy == DestTy) in getCastOpcode() 3270 } else if (SrcTy->isVectorTy()) { in getCastOpcode() 3275 assert(SrcTy->isPointerTy() && in getCastOpcode() 3304 if (SrcTy->isPointerTy()) { in getCastOpcode() [all …]
|
H A D | Verifier.cpp | 2925 bool SrcVec = SrcTy->isVectorTy(); in visitUIToFPInst() 2930 Assert(SrcTy->isIntOrIntVectorTy(), in visitUIToFPInst() 2948 bool SrcVec = SrcTy->isVectorTy(); in visitSIToFPInst() 2953 Assert(SrcTy->isIntOrIntVectorTy(), in visitSIToFPInst() 2971 bool SrcVec = SrcTy->isVectorTy(); in visitFPToUIInst() 2994 bool SrcVec = SrcTy->isVectorTy(); in visitFPToSIInst() 3027 if (SrcTy->isVectorTy()) { in visitPtrToIntInst() 3028 auto *VSrc = cast<VectorType>(SrcTy); in visitPtrToIntInst() 3042 Assert(SrcTy->isIntOrIntVectorTy(), in visitIntToPtrInst() 3052 if (SrcTy->isVectorTy()) { in visitIntToPtrInst() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/X86/ |
H A D | X86InstructionSelector.cpp | 717 const LLT SrcTy = MRI.getType(SrcReg); in selectTruncOrPtrToInt() local 781 const LLT SrcTy = MRI.getType(SrcReg); in selectZext() local 796 if (SrcTy != LLT::scalar(1)) in selectZext() 846 const LLT SrcTy = MRI.getType(SrcReg); in selectAnyext() local 1104 const LLT SrcTy = MRI.getType(SrcReg); in selectExtract() local 1155 const LLT SrcTy = MRI.getType(SrcReg); in emitExtractSubreg() local 1193 const LLT SrcTy = MRI.getType(SrcReg); in emitInsertSubreg() local 1203 if (SrcTy.getSizeInBits() == 128) in emitInsertSubreg() 1205 else if (SrcTy.getSizeInBits() == 256) in emitInsertSubreg() 1320 const LLT SrcTy = MRI.getType(SrcReg0); in selectMergeValues() local [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPUPostLegalizerCombiner.cpp | 188 LLT SrcTy = MRI.getType(SrcReg); in applyUCharToFloat() local 189 if (SrcTy != S32) in applyUCharToFloat() 239 LLT SrcTy = MRI.getType(MatchInfo.CvtVal); in applyCvtF32UByteN() local 240 if (SrcTy != S32) { in applyCvtF32UByteN() 241 assert(SrcTy.isScalar() && SrcTy.getSizeInBits() >= 8); in applyCvtF32UByteN()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/SystemZ/ |
H A D | SystemZTargetTransformInfo.cpp | 611 getVectorTruncCost(Type *SrcTy, Type *DstTy) { in getVectorTruncCost() argument 612 assert (SrcTy->isVectorTy() && DstTy->isVectorTy()); in getVectorTruncCost() 613 assert(SrcTy->getPrimitiveSizeInBits().getFixedSize() > in getVectorTruncCost() 616 assert(cast<FixedVectorType>(SrcTy)->getNumElements() == in getVectorTruncCost() 622 unsigned NumParts = getNumVectorRegs(SrcTy); in getVectorTruncCost() 632 unsigned Log2Diff = getElSizeLog2Diff(SrcTy, DstTy); in getVectorTruncCost() 643 if (VF == 8 && SrcTy->getScalarSizeInBits() == 64 && in getVectorTruncCost() 654 assert (SrcTy->isVectorTy() && DstTy->isVectorTy() && in getVectorBitmaskConversionCost() 658 unsigned SrcScalarBits = SrcTy->getScalarSizeInBits(); in getVectorBitmaskConversionCost() 660 unsigned Log2Diff = getElSizeLog2Diff(SrcTy, DstTy); in getVectorBitmaskConversionCost() [all …]
|
H A D | SystemZTargetTransformInfo.h | 96 unsigned getVectorTruncCost(Type *SrcTy, Type *DstTy); 97 unsigned getVectorBitmaskConversionCost(Type *SrcTy, Type *DstTy);
|
/netbsd/external/apache2/llvm/dist/clang/lib/CodeGen/ |
H A D | CGExprScalar.cpp | 1215 SrcElementTy = SrcTy; in EmitScalarCast() 1322 SrcTy = CGF.FloatTy; in EmitScalarConversion() 1327 if (SrcTy == DstTy) { in EmitScalarConversion() 1471 if (SrcTy->isRealFloatingType()) in EmitFixedPointConversion() 1486 else if (SrcTy->isIntegerType()) in EmitFixedPointConversion() 1501 SrcTy = SrcTy->castAs<ComplexType>()->getElementType(); in EmitComplexToScalarConversion() 1657 if (SrcTy == DstTy) in VisitConvertVectorExpr() 1663 assert(SrcTy->isVectorTy() && in VisitConvertVectorExpr() 4709 auto SrcTy = Src->getType(); in createCastsForTypeOfSameSize() local 4729 if (!SrcTy->isIntegerTy()) in createCastsForTypeOfSameSize() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/ExecutionEngine/Interpreter/ |
H A D | Execution.cpp | 1270 Type *SrcTy = SrcVal->getType(); in executeTruncInst() local 1271 if (SrcTy->isVectorTy()) { in executeTruncInst() 1291 if (SrcTy->isVectorTy()) { in executeSExtInst() 1311 if (SrcTy->isVectorTy()) { in executeZExtInst() 1378 if (isa<VectorType>(SrcTy)) { in executeFPToUIInst() 1416 if (isa<VectorType>(SrcTy)) { in executeFPToSIInst() 1557 if (isa<VectorType>(SrcTy)) { in executeBitCastInst() 1564 SrcElemTy = SrcTy; in executeBitCastInst() 1673 if (SrcTy->isFloatTy()) in executeBitCastInst() 1683 if (SrcTy->isIntegerTy()) in executeBitCastInst() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/Hexagon/ |
H A D | HexagonTargetTransformInfo.cpp | 284 Type *SrcTy, in getCastInstrCost() argument 288 if (SrcTy->isFPOrFPVectorTy() || DstTy->isFPOrFPVectorTy()) { in getCastInstrCost() 289 unsigned SrcN = SrcTy->isFPOrFPVectorTy() ? getTypeNumElements(SrcTy) : 0; in getCastInstrCost() 293 TLI.getTypeLegalizationCost(DL, SrcTy); in getCastInstrCost()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Transforms/IPO/ |
H A D | MergeFunctions.cpp | 485 Type *SrcTy = V->getType(); in createCast() local 486 if (SrcTy->isStructTy()) { in createCast() 488 assert(SrcTy->getStructNumElements() == DestTy->getStructNumElements()); in createCast() 490 for (unsigned int I = 0, E = SrcTy->getStructNumElements(); I < E; ++I) { in createCast() 501 if (SrcTy->isIntegerTy() && DestTy->isPointerTy()) in createCast() 503 else if (SrcTy->isPointerTy() && DestTy->isIntegerTy()) in createCast()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/NVPTX/ |
H A D | NVPTXISelLowering.h | 463 bool isTruncateFree(Type *SrcTy, Type *DstTy) const override { in isTruncateFree() argument 465 if (!SrcTy->isIntegerTy() || !DstTy->isIntegerTy()) in isTruncateFree() 467 return SrcTy->getPrimitiveSizeInBits() == 64 && in isTruncateFree()
|