/netbsd/external/apache2/llvm/dist/llvm/lib/Target/Mips/ |
H A D | MipsInstructionSelector.cpp | 123 const unsigned TySize = Ty.getSizeInBits(); in getRegClassForTypeOnBank() local 126 assert((Ty.isScalar() || Ty.isPointer()) && TySize == 32 && in getRegClassForTypeOnBank() 133 assert((TySize == 32 || TySize == 64) && in getRegClassForTypeOnBank() 135 if (TySize == 32) in getRegClassForTypeOnBank() 186 const unsigned TySize = Ty.getSizeInBits(); in selectLoadStoreOpCode() local 192 assert(((Ty.isScalar() && TySize == 32) || in selectLoadStoreOpCode() 193 (Ty.isPointer() && TySize == 32 && MemSizeInBytes == 4)) && in selectLoadStoreOpCode() 195 (void)TySize; in selectLoadStoreOpCode() 223 assert(((TySize == 32 && MemSizeInBytes == 4) || in selectLoadStoreOpCode() 224 (TySize == 64 && MemSizeInBytes == 8)) && in selectLoadStoreOpCode() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Transforms/Utils/ |
H A D | AutoInitRemark.cpp | 177 Optional<TypeSize> TySize = AI->getAllocationSizeInBits(DL); in inspectVariable() local 179 TySize ? getSizeInBytes(TySize->getFixedSize()) : None; in inspectVariable()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Transforms/Vectorize/ |
H A D | LoadStoreVectorizer.cpp | 814 unsigned TySize = DL.getTypeSizeInBits(Ty); in collectInstructions() local 815 if ((TySize % 8) != 0) in collectInstructions() 829 unsigned VF = VecRegSize / TySize; in collectInstructions() 833 if (TySize > VecRegSize / 2 || in collectInstructions() 834 (VecTy && TTI.getLoadVectorFactor(VF, TySize, TySize / 8, VecTy) == 0)) in collectInstructions() 868 unsigned TySize = DL.getTypeSizeInBits(Ty); in collectInstructions() local 869 if ((TySize % 8) != 0) in collectInstructions() 876 unsigned VF = VecRegSize / TySize; in collectInstructions() 880 if (TySize > VecRegSize / 2 || in collectInstructions() 881 (VecTy && TTI.getStoreVectorFactor(VF, TySize, TySize / 8, VecTy) == 0)) in collectInstructions()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/CodeGen/SelectionDAG/ |
H A D | FunctionLoweringInfo.cpp | 159 uint64_t TySize = in set() local 162 TySize *= CUI->getZExtValue(); // Get total allocated size. in set() 163 if (TySize == 0) TySize = 1; // Don't create zero-sized stack objects. in set() 168 TySize, 0, /*IsImmutable=*/false, /*isAliased=*/true); in set() 171 FrameIndex = MF->getFrameInfo().CreateStackObject(TySize, Alignment, in set()
|
H A D | SelectionDAGBuilder.cpp | 3990 uint64_t TySize = DL.getTypeAllocSize(Ty); in visitAlloca() local 4001 DAG.getConstant(TySize, dl, IntPtr)); in visitAlloca() 8184 uint64_t TySize = DL.getTypeAllocSize(Ty); in getAddressForMemoryInput() local 8187 TySize, DL.getPrefTypeAlign(Ty), false); in getAddressForMemoryInput() 9390 uint64_t TySize = DL.getTypeAllocSize(CLI.RetTy); in LowerCallTo() local 9394 MF.getFrameInfo().CreateStackObject(TySize, Alignment, false); in LowerCallTo()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPULateCodeGenPrepare.cpp | 119 unsigned TySize = DL->getTypeStoreSize(Ty); in canWidenScalarExtLoad() local 121 if (TySize >= 4) in canWidenScalarExtLoad()
|
H A D | AMDGPUCodeGenPrepare.cpp | 307 int TySize = DL.getTypeSizeInBits(Ty); in canWidenScalarExtLoad() local 310 return I.isSimple() && TySize < 32 && Alignment >= 4 && DA->isUniform(&I); in canWidenScalarExtLoad() 1304 int TySize = Mod->getDataLayout().getTypeSizeInBits(I.getType()); in visitLoadInst() local 1305 Type *IntNTy = Builder.getIntNTy(TySize); in visitLoadInst()
|
/netbsd/external/apache2/llvm/dist/clang/lib/CodeGen/ |
H A D | TargetInfo.cpp | 3299 if (TySize <= StartBit) in BitsContainNoUserData() 6041 TySize < SlotSize) { in EmitAAPCSVAArg() 6042 CharUnits Offset = SlotSize - TySize; in EmitAAPCSVAArg() 6084 StackSize = TySize.alignTo(StackSlotSize); in EmitAAPCSVAArg() 6094 TySize < StackSlotSize) { in EmitAAPCSVAArg() 7705 void CoerceToIntArgs(uint64_t TySize, 7813 CoerceToIntArgs(TySize, ArgList); in HandleAggregates() 7824 CoerceToIntArgs(TySize, ArgList); in HandleAggregates() 7888 if (TySize == 0) in classifyArgumentType() 8082 int TySize = getContext().getTypeSize(Ty); in extendType() local [all …]
|
H A D | CGOpenMPRuntimeGPU.cpp | 2719 for (unsigned TySize = 4; TySize > 0 && RealTySize > 0; TySize /=2) { in emitInterWarpCopyFunction() local 2720 unsigned NumIters = RealTySize / TySize; in emitInterWarpCopyFunction() 2724 C.toBits(CharUnits::fromQuantity(TySize)), /*Signed=*/1); in emitInterWarpCopyFunction() 2726 CharUnits Align = CharUnits::fromQuantity(TySize); in emitInterWarpCopyFunction() 2860 RealTySize %= TySize; in emitInterWarpCopyFunction()
|
H A D | CGExpr.cpp | 737 uint64_t TySize = CGM.getMinimumObjectSize(Ty).getQuantity(); in EmitTypeCheck() local 738 llvm::Value *Size = llvm::ConstantInt::get(IntPtrTy, TySize); in EmitTypeCheck()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/CodeGen/GlobalISel/ |
H A D | CombinerHelper.cpp | 1141 unsigned TySize = Ty.getSizeInBytes(); in findGISelOptimalMemOpLowering() local 1142 while (TySize > Size) { in findGISelOptimalMemOpLowering() 1163 TySize = Size; in findGISelOptimalMemOpLowering() 1166 TySize = NewTySize; in findGISelOptimalMemOpLowering() 1174 Size -= TySize; in findGISelOptimalMemOpLowering() 1291 unsigned TySize = Ty.getSizeInBytes(); in optimizeMemset() local 1292 if (TySize > Size) { in optimizeMemset() 1296 DstOff -= TySize - Size; in optimizeMemset() 1326 Size -= TySize; in optimizeMemset()
|
H A D | IRTranslator.cpp | 2594 Register TySize = in translateAlloca() local 2596 MIRBuilder.buildMul(AllocSize, NumElts, TySize); in translateAlloca()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/CodeGen/ |
H A D | SafeStack.cpp | 668 uint64_t TySize = DL.getTypeAllocSize(Ty); in moveDynamicAllocasToUnsafeStack() local 669 Value *Size = IRB.CreateMul(ArraySize, ConstantInt::get(IntPtrTy, TySize)); in moveDynamicAllocasToUnsafeStack()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Transforms/InstCombine/ |
H A D | InstructionCombining.cpp | 1241 if (int64_t TySize = DL.getTypeAllocSize(Ty)) { in FindElementAtOffset() local 1242 FirstIdx = Offset/TySize; in FindElementAtOffset() 1243 Offset -= FirstIdx*TySize; in FindElementAtOffset() 1248 Offset += TySize; in FindElementAtOffset() 1251 assert((uint64_t)Offset < (uint64_t)TySize && "Out of range offset"); in FindElementAtOffset()
|
H A D | InstCombineCasts.cpp | 1989 unsigned TySize = Ty->getScalarSizeInBits(); in visitPtrToInt() local 1991 if (TySize != PtrSize) { in visitPtrToInt()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/ARM/ |
H A D | ARMISelDAGToDAG.cpp | 260 size_t Stride, size_t TySize); 2669 size_t Stride, size_t TySize) { in SelectBaseMVE_VMLLDAV() argument 2670 assert(TySize < Stride && "Invalid TySize"); in SelectBaseMVE_VMLLDAV() 2699 uint16_t Opcode = Opcodes[TySize]; in SelectBaseMVE_VMLLDAV()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/AArch64/ |
H A D | AArch64ISelLowering.cpp | 16987 const TypeSize &TySize = Ty->getPrimitiveSizeInBits(); in functionArgumentNeedsConsecutiveRegisters() local 16988 if (TySize.isScalable() && TySize.getKnownMinSize() > 128) in functionArgumentNeedsConsecutiveRegisters()
|