/netbsd/external/apache2/llvm/dist/llvm/lib/Transforms/Scalar/ |
H A D | MergedLoadStoreMotion.cpp | 118 StoreInst *canSinkFromBlock(BasicBlock *BB, StoreInst *SI); 119 PHINode *getPHIOperand(BasicBlock *BB, StoreInst *S0, StoreInst *S1); 122 bool canSinkStoresAndGEPs(StoreInst *S0, StoreInst *S1) const; 123 void sinkStoresAndGEPs(BasicBlock *BB, StoreInst *SinkCand, 124 StoreInst *ElseInst); 192 auto *Store1 = dyn_cast<StoreInst>(&Inst); in canSinkFromBlock() 211 StoreInst *S1) { in getPHIOperand() 244 StoreInst *S1) { in sinkStoresAndGEPs() 258 StoreInst *SNew = cast<StoreInst>(S0->clone()); in sinkStoresAndGEPs() 314 auto *S0 = dyn_cast<StoreInst>(I); in mergeStores() [all …]
|
H A D | LowerAtomic.cpp | 113 static bool LowerStoreInst(StoreInst *SI) { in LowerStoreInst() 130 } else if (StoreInst *SI = dyn_cast<StoreInst>(&Inst)) { in runOnBasicBlock()
|
H A D | DeadStoreElimination.cpp | 168 if (isa<StoreInst>(I)) in hasAnalyzableMemoryWrite() 209 if (StoreInst *SI = dyn_cast<StoreInst>(Inst)) in getLocForWrite() 241 if (StoreInst *SI = dyn_cast<StoreInst>(I)) in isRemovable() 279 if (isa<StoreInst>(I)) in isShortenableAtTheEnd() 742 StoreInst *Earlier, StoreInst *Later, int64_t InstWriteOffset, in tryToMergePartialOverlappingStores() 1236 if (auto SI = dyn_cast<StoreInst>(UseInst)) in isReadClobber() 1713 if (auto *SI = dyn_cast<StoreInst>(NI)) in isDSEBarrier() 1768 StoreInst *Store = dyn_cast<StoreInst>(Def->getMemoryInst()); in storeIsNoop() 1952 auto *Later = dyn_cast<StoreInst>(SI); in eliminateDeadStores() 2023 NumRemainingStores += isa<StoreInst>(&I); in run() [all …]
|
H A D | SROA.cpp | 980 if (StoreInst *SI = dyn_cast<StoreInst>(I)) { in hasUnsafePHIOrSelectUse() 1150 } else if (StoreInst *SI = dyn_cast<StoreInst>(U->getUser())) { in findCommonType() 1866 } else if (StoreInst *SI = dyn_cast<StoreInst>(U->getUser())) { in isVectorPromotionViableForSlice() 2050 } else if (StoreInst *SI = dyn_cast<StoreInst>(U->getUser())) { in isIntegerWideningViableForSlice() 2699 StoreInst *NewSI; in visitStoreInst() 2891 StoreInst *New = in visitMemSetInst() 3086 StoreInst *Store = cast<StoreInst>( in visitMemTransferInst() 3155 if (StoreInst *SI = dyn_cast<StoreInst>(I)) { in fixLoadStoreAlign() 3440 StoreInst *Store = in emitFunc() 4051 StoreInst *SI = cast<StoreInst>(LU); in presplitLoadsAndStores() [all …]
|
H A D | LoopDataPrefetch.cpp | 254 Writes = isa<StoreInst>(I); in addInstruction() 264 if (isa<StoreInst>(I) && PtrDiff == 0) in addInstruction() 326 } else if (StoreInst *SMemI = dyn_cast<StoreInst>(&I)) { in runOnLoop()
|
H A D | MemCpyOptimizer.cpp | 122 if (!isa<StoreInst>(SI)) in isProfitableToUseMemset() 174 if (StoreInst *SI = dyn_cast<StoreInst>(Inst)) in addInst() 180 void addStore(int64_t OffsetFromFirst, StoreInst *SI) { in addStore() 409 if (!isa<StoreInst>(BI) && !isa<MemSetInst>(BI)) { in tryMergingIntoMemset() 418 if (StoreInst *NextStore = dyn_cast<StoreInst>(BI)) { in tryMergingIntoMemset() 525 bool MemCpyOptPass::moveUp(StoreInst *SI, Instruction *P, const LoadInst *LI) { in moveUp() 587 } else if (isa<LoadInst>(C) || isa<StoreInst>(C) || isa<VAArgInst>(C)) { in moveUp() 649 bool MemCpyOptPass::processStore(StoreInst *SI, BasicBlock::iterator &BBI) { in processStore() 1684 if (StoreInst *SI = dyn_cast<StoreInst>(I)) in iterateOnFunction()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/X86/ |
H A D | X86AvoidStoreForwardingBlocks.cpp | 391 MachineOperand &StoreBase = getBaseOperand(StoreInst); in buildCopy() 413 MachineInstr *StInst = StoreInst; in buildCopy() 446 StoreInst, getYMMtoXMMStoreOpcode(StoreInst->getOpcode()), in buildCopies() 500 MachineOperand &StoreBase = getBaseOperand(StoreInst); in updateKillStatus() 516 MachineInstr *StInst = StoreInst; in updateKillStatus() 569 MachineInstr *LoadInst, MachineInstr *StoreInst, in breakBlockedCopies() argument 572 int64_t StDispImm = getDispOperand(StoreInst).getImm(); in breakBlockedCopies() 715 MachineInstr *StoreInst = LoadStoreInstPair.second; in runOnMachineFunction() local 718 LLVM_DEBUG(StoreInst->dump()); in runOnMachineFunction() 722 updateKillStatus(LoadInst, StoreInst); in runOnMachineFunction() [all …]
|
H A D | X86PreAMXConfig.cpp | 164 new StoreInst(PaletteValue, PalettePos, Pos); in preWriteTileCfg() 178 new StoreInst(Row, RowPos, Pos); in preWriteTileCfg() 179 new StoreInst(Col, ColPos, Pos); in preWriteTileCfg() 205 Instruction *Init0 = new StoreInst(Val0, Addr, false, Alignment, Cfg); in addTileConfig()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Transforms/Utils/ |
H A D | PromoteMemoryToRegister.cpp | 72 } else if (const StoreInst *SI = dyn_cast<StoreInst>(U)) { in isAllocaPromotable() 109 StoreInst *OnlyStore; 135 if (StoreInst *SI = dyn_cast<StoreInst>(User)) { in AnalyzeAlloca() 190 (isa<StoreInst>(I) && isa<AllocaInst>(I->getOperand(1))); in isInterestingInstruction() 318 if (isa<LoadInst>(I) || isa<StoreInst>(I)) in removeIntrinsicUsers() 357 StoreInst *OnlyStore = Info.OnlyStore; in rewriteSingleStoreAlloca() 470 if (StoreInst *SI = dyn_cast<StoreInst>(U)) in promoteSingleBlockAlloca() 489 std::make_pair(LoadIdx, static_cast<StoreInst *>(nullptr)), in promoteSingleBlockAlloca() 522 StoreInst *SI = cast<StoreInst>(AI->user_back()); in promoteSingleBlockAlloca() 805 if (StoreInst *SI = dyn_cast<StoreInst>(I)) { in ComputeLiveInBlocks() [all …]
|
H A D | SSAUpdater.cpp | 330 SomeVal = cast<StoreInst>(Insts[0])->getOperand(0); in LoadAndStorePromoter() 363 if (StoreInst *SI = dyn_cast<StoreInst>(User)) { in run() 376 if (isa<StoreInst>(I)) { in run() 415 if (StoreInst *SI = dyn_cast<StoreInst>(&I)) { in run()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPURewriteOutArguments.cpp | 122 StoreInst *SI = dyn_cast<StoreInst>(U.getUser()); in checkArgumentUses() 160 U.getOperandNo() != StoreInst::getPointerOperandIndex()) in checkArgumentUses() 269 SmallVector<std::pair<ReturnInst *, StoreInst *>, 4> ReplaceableStores; in runOnFunction() 289 StoreInst *SI = nullptr; in runOnFunction() 291 SI = dyn_cast<StoreInst>(Q.getInst()); in runOnFunction() 305 for (std::pair<ReturnInst *, StoreInst *> Store : ReplaceableStores) { in runOnFunction()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Transforms/InstCombine/ |
H A D | InstCombineLoadStoreAlloca.cpp | 482 static StoreInst *combineStoreToNewValue(InstCombinerImpl &IC, StoreInst &SI, in combineStoreToNewValue() 492 StoreInst *NewStore = IC.Builder.CreateAlignedStore( in combineStoreToNewValue() 912 static bool canSimplifyNullStoreOrGEP(StoreInst &SI) { in canSimplifyNullStoreOrGEP() 984 StoreInst *SI = new StoreInst(UndefValue::get(LI.getType()), in visitLoadInst() 1319 auto *SI = dyn_cast<StoreInst>(U); in removeBitcastsFromLoadStoreOnMinMax() 1331 auto *USI = cast<StoreInst>(UI); in removeBitcastsFromLoadStoreOnMinMax() 1405 if (StoreInst *PrevSI = dyn_cast<StoreInst>(BBI)) { in visitStoreInst() 1488 StoreInst *OtherStore = nullptr; in mergeStoreIntoSuccessor() 1500 OtherStore = dyn_cast<StoreInst>(BBI); in mergeStoreIntoSuccessor() 1516 if ((OtherStore = dyn_cast<StoreInst>(BBI))) { in mergeStoreIntoSuccessor() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/include/llvm/Transforms/Scalar/ |
H A D | MemCpyOptimizer.h | 38 class StoreInst; variable 63 bool processStore(StoreInst *SI, BasicBlock::iterator &BBI); 76 bool moveUp(StoreInst *SI, Instruction *P, const LoadInst *LI);
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Transforms/Instrumentation/ |
H A D | ThreadSanitizer.cpp | 450 const bool IsWrite = isa<StoreInst>(*I); in chooseInstructionsToInstrument() 451 Value *Addr = IsWrite ? cast<StoreInst>(I)->getPointerOperand() in chooseInstructionsToInstrument() 465 cast<StoreInst>(WI.Inst)->isVolatile()); in chooseInstructionsToInstrument() 505 if (StoreInst *SI = dyn_cast<StoreInst>(I)) in isAtomic() 551 else if (isa<LoadInst>(Inst) || isa<StoreInst>(Inst)) in sanitizeFunction() 614 const bool IsWrite = isa<StoreInst>(*II.Inst); in instrumentLoadOrStore() 615 Value *Addr = IsWrite ? cast<StoreInst>(II.Inst)->getPointerOperand() in instrumentLoadOrStore() 629 Value *StoredValue = cast<StoreInst>(II.Inst)->getValueOperand(); in instrumentLoadOrStore() 652 const unsigned Alignment = IsWrite ? cast<StoreInst>(II.Inst)->getAlignment() in instrumentLoadOrStore() 657 (IsWrite ? cast<StoreInst>(II.Inst)->isVolatile() in instrumentLoadOrStore() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/CodeGen/ |
H A D | GCRootLowering.cpp | 128 if (isa<AllocaInst>(I) || isa<GetElementPtrInst>(I) || isa<StoreInst>(I) || in CouldBecomeSafePoint() 151 if (StoreInst *SI = dyn_cast<StoreInst>(IP)) in InsertRootInitializers() 161 new StoreInst( in InsertRootInitializers() 205 Value *St = new StoreInst(CI->getArgOperand(0), in DoLowering()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Transforms/Vectorize/ |
H A D | LoadStoreVectorizer.cpp | 298 if (StoreInst *S = dyn_cast<StoreInst>(I)) in getPointerAddressSpace() 659 assert(isa<StoreInst>(I) && in getVectorizablePrefix() 665 if (isa<LoadInst>(I) || isa<StoreInst>(I)) { in getVectorizablePrefix() 726 if (isa<StoreInst>(MemInstr) && ChainLoad && in getVectorizablePrefix() 731 if (MemLoad && isa<StoreInst>(ChainInstr) && in getVectorizablePrefix() 847 } else if (StoreInst *SI = dyn_cast<StoreInst>(&I)) { in collectInstructions() 992 StoreInst *S0 = cast<StoreInst>(Chain[0]); in vectorizeStoreChain() 997 StoreTy = cast<StoreInst>(I)->getValueOperand()->getType(); in vectorizeStoreChain() 1103 StoreInst *Store = cast<StoreInst>(Chain[I]); in vectorizeStoreChain() 1118 StoreInst *Store = cast<StoreInst>(Chain[I]); in vectorizeStoreChain() [all …]
|
H A D | VPlanTransforms.cpp | 66 } else if (StoreInst *Store = dyn_cast<StoreInst>(Inst)) { in VPInstructionsToVPRecipes()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/NVPTX/ |
H A D | NVPTXLowerAggrCopies.cpp | 79 if (StoreInst *SI = dyn_cast<StoreInst>(LI->user_back())) { in runOnFunction() 106 auto *SI = cast<StoreInst>(*LI->user_begin()); in runOnFunction()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Transforms/IPO/ |
H A D | GlobalOpt.cpp | 214 if (StoreInst *SI = dyn_cast<StoreInst>(U)) { in CleanupPointerRootUsers() 306 } else if (StoreInst *SI = dyn_cast<StoreInst>(U)) { in CleanupConstantGlobalUsers() 420 if (StoreInst *SI = dyn_cast<StoreInst>(I)) in isSafeSROAElementUse() 678 } else if (const StoreInst *SI = dyn_cast<StoreInst>(U)) { in AllUsesOfValueWillTrapIfNull() 719 } else if (isa<StoreInst>(U)) { in AllUsesOfLoadedValueWillTrapIfNull() 740 } else if (StoreInst *SI = dyn_cast<StoreInst>(I)) { in OptimizeAwayTrappingUsesOfValue() 942 if (StoreInst *SI = dyn_cast<StoreInst>(GV->user_back())) { in OptimizeGlobalAddressOfMalloc() 1025 if (const StoreInst *SI = dyn_cast<StoreInst>(Inst)) { in valueIsOnlyUsedLocallyOrStoredToOneGlobal() 1223 if (StoreInst *SI = dyn_cast<StoreInst>(UI)) { in TryToShrinkGlobalToBoolean() 1253 StoreInst *NSI = in TryToShrinkGlobalToBoolean() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/include/llvm/Transforms/Vectorize/ |
H A D | SLPVectorizer.h | 45 class StoreInst; variable 59 using StoreList = SmallVector<StoreInst *, 8>; 138 bool vectorizeStores(ArrayRef<StoreInst *> Stores, slpvectorizer::BoUpSLP &R);
|
/netbsd/external/apache2/llvm/dist/llvm/lib/IR/ |
H A D | Instruction.cpp | 420 if (const StoreInst *SI = dyn_cast<StoreInst>(I1)) in haveSameSpecialState() 421 return SI->isVolatile() == cast<StoreInst>(I2)->isVolatile() && in haveSameSpecialState() 422 (SI->getAlignment() == cast<StoreInst>(I2)->getAlignment() || in haveSameSpecialState() 424 SI->getOrdering() == cast<StoreInst>(I2)->getOrdering() && in haveSameSpecialState() 425 SI->getSyncScopeID() == cast<StoreInst>(I2)->getSyncScopeID(); in haveSameSpecialState() 559 return !cast<StoreInst>(this)->isUnordered(); in mayReadFromMemory() 594 return cast<StoreInst>(this)->getOrdering() != AtomicOrdering::NotAtomic; in isAtomic() 629 return cast<StoreInst>(this)->isVolatile(); in isVolatile()
|
/netbsd/external/apache2/llvm/dist/llvm/include/llvm/Transforms/Utils/ |
H A D | AutoInitRemark.h | 30 class StoreInst; variable 45 void inspectStore(StoreInst &SI);
|
H A D | VNCoercion.h | 26 class StoreInst; variable 55 StoreInst *DepSI, const DataLayout &DL);
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Transforms/ObjCARC/ |
H A D | ObjCARCContract.cpp | 197 static StoreInst *findSafeStoreForStoreStrongContraction(LoadInst *Load, in findSafeStoreForStoreStrongContraction() 201 StoreInst *Store = nullptr; in findSafeStoreForStoreStrongContraction() 256 Store = dyn_cast<StoreInst>(Inst); in findSafeStoreForStoreStrongContraction() 284 findRetainForStoreStrongContraction(Value *New, StoreInst *Store, in findRetainForStoreStrongContraction() 350 StoreInst *Store = in tryToContractReleaseIntoStoreStrong() 481 new StoreInst(Null, CI->getArgOperand(0), CI); in tryToPeepholeInstruction()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Analysis/ |
H A D | Loads.cpp | 385 } else if (StoreInst *SI = dyn_cast<StoreInst>(BBI)) { in isSafeToLoadUnconditionally() 497 if (StoreInst *SI = dyn_cast<StoreInst>(Inst)) { in getAvailableLoadStore() 555 if (StoreInst *SI = dyn_cast<StoreInst>(Inst)) { in findAvailablePtrLoadStore()
|