/openbsd/gnu/llvm/llvm/lib/CodeGen/ |
H A D | ExpandMemCmp.cpp | 340 const LoadPair Loads = in emitLoadCompareByteBlock() local 343 Value *Diff = Builder.CreateSub(Loads.Lhs, Loads.Rhs); in emitLoadCompareByteBlock() 396 const LoadPair Loads = getLoadPair( in getCompareLoadPairs() local 403 Diff = Builder.CreateXor(Loads.Lhs, Loads.Rhs); in getCompareLoadPairs() 408 Cmp = Builder.CreateICmpNE(Loads.Lhs, Loads.Rhs); in getCompareLoadPairs() 489 const LoadPair Loads = in emitLoadCompareBlock() local 500 Value *Cmp = Builder.CreateICmp(ICmpInst::ICMP_EQ, Loads.Lhs, Loads.Rhs); in emitLoadCompareBlock() 602 const LoadPair Loads = in getMemCmpOneBlock() local 605 return Builder.CreateSub(Loads.Lhs, Loads.Rhs); in getMemCmpOneBlock() 616 Value *CmpUGT = Builder.CreateICmpUGT(Loads.Lhs, Loads.Rhs); in getMemCmpOneBlock() [all …]
|
H A D | ScheduleDAGInstrs.cpp | 757 Value2SUsMap Stores, Loads(1 /*TrueMemOrderLatency*/); in buildSchedGraph() local 901 addBarrierChain(Loads); in buildSchedGraph() 945 addChainDependencies(SU, Loads); in buildSchedGraph() 959 addChainDependencies(SU, (ThisMayAlias ? Loads : NonAliasLoads), V); in buildSchedGraph() 972 addChainDependencies(SU, Loads, UnknownValue); in buildSchedGraph() 981 Loads.insert(SU, UnknownValue); in buildSchedGraph() 992 (ThisMayAlias ? Loads : NonAliasLoads).insert(SU, V); in buildSchedGraph() 1000 if (Stores.size() + Loads.size() >= HugeRegion) { in buildSchedGraph() 1002 reduceHugeMemNodeMaps(Stores, Loads, getReductionSize()); in buildSchedGraph()
|
H A D | WinEHPrepare.cpp | 83 DenseMap<BasicBlock *, Value *> &Loads, Function &F); 1102 DenseMap<BasicBlock *, Value *> Loads; in insertPHILoads() local 1110 replaceUseWithLoad(PN, U, SpillSlot, Loads, F); in insertPHILoads() 1171 DenseMap<BasicBlock *, Value *> &Loads, in replaceUseWithLoad() argument 1231 Value *&Load = Loads[IncomingBlock]; in replaceUseWithLoad()
|
/openbsd/gnu/llvm/llvm/lib/Target/X86/ |
H A D | X86PreAMXConfig.cpp | 109 bool checkVolatileModel(SmallSet<Value *, 4> &Loads, IntrinsicInst *Store, 207 bool X86PreAMXConfig::checkVolatileModel(SmallSet<Value *, 4> &Loads, in checkVolatileModel() argument 214 return (Loads.size() == 1) && Loads.contains(ST); in checkVolatileModel() 220 if (!Loads.erase(Op)) in checkVolatileModel() 226 return Loads.empty() && (ST == cast<Value>(KeyAMX)); in checkVolatileModel() 265 SmallSet<Value *, 4> Loads; in getShapesAndConfigPosEnd() local 275 Loads.insert(II); in getShapesAndConfigPosEnd() 277 if (!checkVolatileModel(Loads, II, KeyAMX)) in getShapesAndConfigPosEnd()
|
/openbsd/gnu/llvm/llvm/lib/Target/ARM/ |
H A D | ARMParallelDSP.cpp | 200 SmallVector<LoadInst*, 4> Loads; member in __anonfbc564360111::WidenedLoad 205 append_range(Loads, Lds); in WidenedLoad() 343 SmallVector<LoadInst*, 8> Loads; in RecordMemoryOps() local 358 Loads.push_back(Ld); in RecordMemoryOps() 361 if (Loads.empty() || Loads.size() > NumLoadLimit) in RecordMemoryOps() 371 for (auto *Read : Loads) { in RecordMemoryOps() 403 for (auto *Base : Loads) { in RecordMemoryOps() 404 for (auto *Offset : Loads) { in RecordMemoryOps() 732 LoadInst* ARMParallelDSP::CreateWideLoad(MemInstList &Loads, in CreateWideLoad() argument 736 LoadInst *Base = Loads[0]; in CreateWideLoad() [all …]
|
H A D | ARMSelectionDAGInfo.cpp | 205 SDValue Loads[6]; in EmitTargetCodeForMemcpy() local 260 Loads[i] = DAG.getLoad(VT, dl, Chain, in EmitTargetCodeForMemcpy() 264 TFOps[i] = Loads[i].getValue(1); in EmitTargetCodeForMemcpy() 276 TFOps[i] = DAG.getStore(Chain, dl, Loads[i], in EmitTargetCodeForMemcpy()
|
H A D | ARMScheduleM4.td | 47 // Loads, MAC's and DIV all get a higher latency of 2 106 // Loads still take 2 cycles.
|
/openbsd/gnu/llvm/llvm/lib/Transforms/Utils/ |
H A D | DemoteRegToStack.cpp | 68 DenseMap<BasicBlock*, Value*> Loads; in DemoteRegToStack() local 71 Value *&V = Loads[PN->getIncomingBlock(i)]; in DemoteRegToStack()
|
/openbsd/gnu/llvm/llvm/lib/Target/NVPTX/ |
H A D | NVPTXLowerArgs.cpp | 267 SmallVector<Load> Loads; in adjustByValArgAlignment() local 277 Loads.push_back({I, Ctx.Offset}); in adjustByValArgAlignment() 306 for (Load &CurLoad : Loads) { in adjustByValArgAlignment()
|
/openbsd/gnu/llvm/llvm/lib/Analysis/ |
H A D | AliasAnalysisEvaluator.cpp | 110 SetVector<Value *> Loads; in runInternal() local 116 Loads.insert(LI); in runInternal() 160 for (Value *Load : Loads) { in runInternal()
|
H A D | CMakeLists.txt | 88 Loads.cpp
|
/openbsd/gnu/llvm/llvm/lib/CodeGen/SelectionDAG/ |
H A D | ScheduleDAGSDNodes.cpp | 272 SmallVector<SDNode*, 4> Loads; in ClusterNeighboringLoads() local 276 Loads.push_back(BaseLoad); in ClusterNeighboringLoads() 282 Loads.push_back(Load); in ClusterNeighboringLoads() 291 SDNode *Lead = Loads[0]; in ClusterNeighboringLoads() 295 for (unsigned I = 1, E = Loads.size(); I != E; ++I) { in ClusterNeighboringLoads() 297 SDNode *Load = Loads[I]; in ClusterNeighboringLoads()
|
/openbsd/gnu/llvm/llvm/lib/Transforms/Instrumentation/ |
H A D | SanitizerCoverage.cpp | 234 void InjectTraceForLoadsAndStores(Function &F, ArrayRef<LoadInst *> Loads, 657 SmallVector<LoadInst *, 8> Loads; in instrumentFunction() local 690 Loads.push_back(LI); in instrumentFunction() 710 InjectTraceForLoadsAndStores(F, Loads, Stores); in instrumentFunction() 894 Function &, ArrayRef<LoadInst *> Loads, ArrayRef<StoreInst *> Stores) { in InjectTraceForLoadsAndStores() argument 906 for (auto *LI : Loads) { in InjectTraceForLoadsAndStores()
|
/openbsd/gnu/llvm/llvm/lib/Transforms/IPO/ |
H A D | GlobalOpt.cpp | 1330 SmallVector<LoadInst *, 4> Loads; in isPointerValueDeadOnEntryToFunction() local 1336 Loads.push_back(LI); in isPointerValueDeadOnEntryToFunction() 1351 Loads.push_back(LI); in isPointerValueDeadOnEntryToFunction() 1374 if (Loads.size() * Stores.size() > Threshold) in isPointerValueDeadOnEntryToFunction() 1377 for (auto *L : Loads) { in isPointerValueDeadOnEntryToFunction() 1463 SmallVector<LoadInst *> Loads; in forwardStoredOnceStore() local 1468 Loads.push_back(LI); in forwardStoredOnceStore() 1473 if (!Loads.empty()) { in forwardStoredOnceStore() 1475 for (auto *LI : Loads) { in forwardStoredOnceStore()
|
H A D | ArgumentPromotion.cpp | 573 SmallVector<LoadInst *, 16> Loads; in findArgParts() local 598 Loads.push_back(LI); in findArgParts() 661 for (LoadInst *Load : Loads) { in findArgParts()
|
/openbsd/gnu/usr.bin/clang/libLLVMAnalysis/ |
H A D | Makefile | 69 Loads.cpp \
|
/openbsd/gnu/llvm/llvm/utils/gn/secondary/llvm/lib/Analysis/ |
H A D | BUILD.gn | 77 "Loads.cpp",
|
/openbsd/gnu/llvm/llvm/lib/Target/RISCV/ |
H A D | RISCVScheduleV.td | 116 // 7. Vector Loads and Stores 149 // 7.7. Vector Unit-stride Fault-Only-First Loads 359 // 7. Vector Loads and Stores 590 // 7. Vector Loads and Stores 629 // Vector Segment Loads and Stores 768 // 7. Vector Loads and Stores
|
/openbsd/gnu/gcc/gcc/config/sh/ |
H A D | sh1.md | 33 ;; Loads have a latency of two.
|
/openbsd/gnu/llvm/llvm/lib/Target/PowerPC/ |
H A D | PPCScheduleP8.td | 29 def P8_LU1 : FuncUnit; // Loads or fixed-point operations 1 30 def P8_LU2 : FuncUnit; // Loads or fixed-point operations 2
|
/openbsd/gnu/usr.bin/gcc/gcc/config/alpha/ |
H A D | ev4.md | 38 ; Loads from L0 completes in three cycles. adjust_cost still factors
|
H A D | ev5.md | 49 ; Loads from L0 complete in two cycles. adjust_cost still factors
|
/openbsd/gnu/gcc/gcc/config/alpha/ |
H A D | ev4.md | 38 ; Loads from L0 completes in three cycles. adjust_cost still factors
|
H A D | ev5.md | 49 ; Loads from L0 complete in two cycles. adjust_cost still factors
|
/openbsd/gnu/gcc/gcc/config/arm/ |
H A D | arm926ejs.md | 122 ;; Loads with a shifted offset take 3 cycles, and are (a) probably the
|