/netbsd/external/apache2/llvm/dist/llvm/lib/Target/ARM/ |
H A D | MVEGatherScatterLowering.cpp | 228 Offsets = GEP->getOperand(1); in checkGEP() 238 Offsets = GEP->getOperand(1); in checkGEP() 259 if (Ty != Offsets->getType()) { in checkGEP() 262 Offsets = Builder.CreateTrunc(Offsets, Ty); in checkGEP() 264 Offsets = Builder.CreateZExt(Offsets, VectorType::getInteger(Ty)); in checkGEP() 480 Value *Offsets; in tryCreateMaskedGatherOffset() local 622 Value *Offsets; in tryCreateMaskedScatterOffset() local 1110 if (!isa<Constant>(Offsets)) in foldGEP() 1118 Offsets = in foldGEP() 1120 if (Offsets == nullptr) in foldGEP() [all …]
|
/netbsd/external/apache2/llvm/dist/clang/tools/clang-format/ |
H A D | ClangFormat.cpp | 37 Offsets("offset", variable 212 if (!Offsets.empty() || !Lengths.empty()) { in fillRanges() 238 if (Offsets.empty()) in fillRanges() 239 Offsets.push_back(0); in fillRanges() 240 if (Offsets.size() != Lengths.size() && in fillRanges() 241 !(Offsets.size() == 1 && Lengths.empty())) { in fillRanges() 245 for (unsigned i = 0, e = Offsets.size(); i != e; ++i) { in fillRanges() 246 if (Offsets[i] >= Code->getBufferSize()) { in fillRanges() 251 Sources.getLocForStartOfFile(ID).getLocWithOffset(Offsets[i]); in fillRanges() 254 if (Offsets[i] + Lengths[i] > Code->getBufferSize()) { in fillRanges() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Remarks/ |
H A D | RemarkParser.cpp | 31 Offsets.push_back(Split.first.data() - Buffer.data()); in ParsedStringTable() 37 if (Index >= Offsets.size()) in operator []() 41 Offsets.size()); in operator []() 43 size_t Offset = Offsets[Index]; in operator []() 47 (Index == Offsets.size() - 1) ? Buffer.size() : Offsets[Index + 1]; in operator []()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/CodeGen/ |
H A D | Analysis.cpp | 85 SmallVectorImpl<uint64_t> *Offsets, in ComputeValueVTs() argument 99 ComputeValueVTs(TLI, DL, *EI, ValueVTs, MemVTs, Offsets, in ComputeValueVTs() 109 ComputeValueVTs(TLI, DL, EltTy, ValueVTs, MemVTs, Offsets, in ComputeValueVTs() 120 if (Offsets) in ComputeValueVTs() 121 Offsets->push_back(StartingOffset); in ComputeValueVTs() 126 SmallVectorImpl<uint64_t> *Offsets, in ComputeValueVTs() argument 134 SmallVectorImpl<uint64_t> *Offsets, in computeValueLLTs() argument 144 computeValueLLTs(DL, *STy->getElementType(I), ValueTys, Offsets, in computeValueLLTs() 154 computeValueLLTs(DL, *EltTy, ValueTys, Offsets, in computeValueLLTs() 163 if (Offsets != nullptr) in computeValueLLTs() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Support/ |
H A D | Signals.cpp | 100 const char **Modules, intptr_t *Offsets, 149 std::vector<intptr_t> Offsets(Depth, 0); in printSymbolizedStackTrace() local 150 if (!findModulesAndOffsets(StackTrace, Depth, Modules.data(), Offsets.data(), in printSymbolizedStackTrace() 164 Input << Modules[i] << " " << (void*)Offsets[i] << "\n"; in printSymbolizedStackTrace() 221 OS << "(" << Modules[i] << '+' << format_hex(Offsets[i], 0) << ")"; in printSymbolizedStackTrace()
|
H A D | SourceMgr.cpp | 78 auto *Offsets = new std::vector<T>(); in GetOrCreateOffsetCache() local 84 Offsets->push_back(static_cast<T>(N)); in GetOrCreateOffsetCache() 87 OffsetCache = Offsets; in GetOrCreateOffsetCache() 88 return *Offsets; in GetOrCreateOffsetCache() 93 std::vector<T> &Offsets = in getLineNumberSpecialized() local 105 return llvm::lower_bound(Offsets, PtrOffset) - Offsets.begin() + 1; in getLineNumberSpecialized() 125 std::vector<T> &Offsets = in getPointerForLineNumberSpecialized() local 138 if (LineNo > Offsets.size()) in getPointerForLineNumberSpecialized() 140 return BufStart + Offsets[LineNo - 1] + 1; in getPointerForLineNumberSpecialized()
|
H A D | IntervalMap.cpp | 19 void Path::replaceRoot(void *Root, unsigned Size, IdxPair Offsets) { in replaceRoot() argument 21 path.front() = Entry(Root, Size, Offsets.first); in replaceRoot() 22 path.insert(path.begin() + 1, Entry(subtree(0), Offsets.second)); in replaceRoot()
|
/netbsd/external/apache2/llvm/dist/llvm/include/llvm/CodeGen/ |
H A D | Analysis.h | 70 SmallVectorImpl<uint64_t> *Offsets = nullptr, 77 SmallVectorImpl<uint64_t> *Offsets = nullptr, 89 SmallVectorImpl<uint64_t> *Offsets = nullptr,
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/VE/ |
H A D | VEFrameLowering.h | 59 static const SpillSlot Offsets[] = { in getCalleeSavedSpillSlots() local 65 NumEntries = array_lengthof(Offsets); in getCalleeSavedSpillSlots() 66 return Offsets; in getCalleeSavedSpillSlots()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/AMDGPU/MCTargetDesc/ |
H A D | R600MCCodeEmitter.cpp | 124 int64_t Offsets[3] = { in encodeInstruction() local 133 SrcSelect[ELEMENT_W] << 29 | Offsets[0] << 0 | Offsets[1] << 5 | in encodeInstruction() 134 Offsets[2] << 10; in encodeInstruction()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Analysis/ |
H A D | StackSafetyAnalysis.cpp | 281 ConstantRange Offsets = offsetFrom(Addr, Base); in getAccessRange() local 282 if (isUnsafe(Offsets)) in getAccessRange() 285 Offsets = addOverflowNever(Offsets, SizeRange); in getAccessRange() 286 if (isUnsafe(Offsets)) in getAccessRange() 288 return Offsets; in getAccessRange() 422 ConstantRange Offsets = offsetFrom(UI, Ptr); in analyzeAllUses() local 426 Insert.first->second = Insert.first->second.unionWith(Offsets); in analyzeAllUses() 511 const ConstantRange &Offsets) const { in getArgumentAccessRange() 525 return addOverflowNever(Access, Offsets); in getArgumentAccessRange() 1037 assert(!Call.Offsets.isFullSet()); in generateParamAccessSummary() [all …]
|
/netbsd/external/apache2/llvm/dist/clang/lib/Lex/ |
H A D | DependencyDirectivesSourceMinimizer.cpp | 912 llvm::SmallVector<Directive, 32> Offsets; in computeSkippedRanges() local 918 Offsets.push_back({T.Offset, Directive::If}); in computeSkippedRanges() 923 if (Offsets.empty()) in computeSkippedRanges() 925 int PreviousOffset = Offsets.back().Offset; in computeSkippedRanges() 927 Offsets.push_back({T.Offset, Directive::Else}); in computeSkippedRanges() 932 if (Offsets.empty()) in computeSkippedRanges() 934 int PreviousOffset = Offsets.back().Offset; in computeSkippedRanges() 937 Directive::DirectiveKind Kind = Offsets.pop_back_val().Kind; in computeSkippedRanges() 940 } while (!Offsets.empty()); in computeSkippedRanges()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/CodeGen/GlobalISel/ |
H A D | CallLowering.cpp | 208 SmallVector<uint64_t, 4> Offsets; in splitToValueTypes() local 209 ComputeValueVTs(*TLI, DL, OrigArg.Ty, SplitVTs, &Offsets, 0); in splitToValueTypes() 247 SmallVector<uint64_t, 8> Offsets; in unpackRegs() local 248 computeValueLLTs(DL, *PackedTy, LLTs, &Offsets); in unpackRegs() 252 MIRBuilder.buildExtract(DstRegs[i], SrcReg, Offsets[i]); in unpackRegs() 773 SmallVector<uint64_t, 4> Offsets; in insertSRetLoads() local 774 ComputeValueVTs(*TLI, DL, RetTy, SplitVTs, &Offsets, 0); in insertSRetLoads() 787 MIRBuilder.materializePtrAdd(Addr, DemoteReg, OffsetLLTy, Offsets[I]); in insertSRetLoads() 803 SmallVector<uint64_t, 4> Offsets; in insertSRetStores() local 804 ComputeValueVTs(*TLI, DL, RetTy, SplitVTs, &Offsets, 0); in insertSRetStores() [all …]
|
H A D | IRTranslator.cpp | 177 auto *Offsets = VMap.getOffsets(Val); in allocateVRegs() local 180 Offsets->empty() ? Offsets : nullptr); in allocateVRegs() 196 auto *Offsets = VMap.getOffsets(Val); in getOrCreateVRegs() local 203 Offsets->empty() ? Offsets : nullptr); in getOrCreateVRegs() 1271 ArrayRef<uint64_t> Offsets = *VMap.getOffsets(LI); in translateLoad() local 1294 MachinePointerInfo Ptr(LI.getPointerOperand(), Offsets[i] / 8); in translateLoad() 1378 ArrayRef<uint64_t> Offsets = *VMap.getOffsets(*Src); in translateExtractValue() local 1379 unsigned Idx = llvm::lower_bound(Offsets, Offset) - Offsets.begin(); in translateExtractValue() 2860 if (Offsets && !Offsets->empty()) in valueIsSplit() 2861 Offsets->clear(); in valueIsSplit() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/ObjectYAML/ |
H A D | DWARFEmitter.cpp | 680 4 + Table.Offsets.size() * (Table.Format == dwarf::DWARF64 ? 8 : 4); in emitDebugStrOffsets() 686 for (uint64_t Offset : Table.Offsets) in emitDebugStrOffsets() 934 std::vector<uint64_t> Offsets; in writeDWARFLists() local 937 Offsets.push_back(ListBufferOS.tell()); in writeDWARFLists() 960 OffsetEntryCount = Table.Offsets ? Table.Offsets->size() : Offsets.size(); in writeDWARFLists() 976 auto EmitOffsets = [&](ArrayRef<uint64_t> Offsets, uint64_t OffsetsSize) { in writeDWARFLists() argument 977 for (uint64_t Offset : Offsets) in writeDWARFLists() 982 if (Table.Offsets) in writeDWARFLists() 983 EmitOffsets(ArrayRef<uint64_t>((const uint64_t *)Table.Offsets->data(), in writeDWARFLists() 984 Table.Offsets->size()), in writeDWARFLists() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/include/llvm/Remarks/ |
H A D | RemarkParser.h | 64 std::vector<size_t> Offsets; member 74 size_t size() const { return Offsets.size(); } in size()
|
/netbsd/external/apache2/llvm/dist/clang/lib/CodeGen/ |
H A D | CGExprConstant.cpp | 85 llvm::SmallVector<CharUnits, 32> Offsets; member in __anon20ee5db20111::ConstantAggregateBuilder 152 Offsets.push_back(Size); in add() 155 Offsets.push_back(Offset); in add() 289 return Offsets.size(); in splitAt() 293 if (FirstAfterPos == Offsets.begin()) in splitAt() 302 if (Offsets[LastAtOrBeforePosIndex] + in splitAt() 318 CharUnits Offset = Offsets[Index]; in split() 333 Offsets, Index, Index + 1, in split() 341 replace(Offsets, Index, Index + 1, in split() 361 replace(Offsets, Index, Index + 1, in split() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/include/llvm/Transforms/IPO/ |
H A D | LowerTypeTests.h | 62 SmallVector<uint64_t, 16> Offsets; member 74 Offsets.push_back(Offset); in addOffset()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/Hexagon/ |
H A D | HexagonFrameLowering.h | 92 static const SpillSlot Offsets[] = { in getCalleeSavedSpillSlots() local 100 NumEntries = array_lengthof(Offsets); in getCalleeSavedSpillSlots() 101 return Offsets; in getCalleeSavedSpillSlots()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Object/ |
H A D | Archive.cpp | 797 const char *Offsets = Buf; in getMember() local 799 Offsets += sizeof(uint64_t); in getMember() 801 Offsets += sizeof(uint32_t); in getMember() 804 Offset = read32be(Offsets + SymbolIndex * 4); in getMember() 806 Offset = read64be(Offsets + SymbolIndex * 8); in getMember() 814 Offset = read32le(Offsets + SymbolIndex * 8 + 4); in getMember() 822 Offset = read64le(Offsets + SymbolIndex * 16 + 8); in getMember() 844 Offset = read32le(Offsets + OffsetIndex * 4); in getMember()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/NVPTX/ |
H A D | NVPTXISelLowering.cpp | 160 if (Offsets) { in ComputePTXValueVTs() 198 if (Offsets) in ComputePTXValueVTs() 203 if (Offsets) in ComputePTXValueVTs() 204 Offsets->push_back(Off); in ComputePTXValueVTs() 227 if (Offsets[Idx] & (AccessSize - 1)) in CanMergeParamLoadStoresStartingAt() 256 if (Offsets[j] - Offsets[j - 1] != EltSize) in CanMergeParamLoadStoresStartingAt() 1450 SmallVector<uint64_t, 16> Offsets; in LowerCall() local 1564 SmallVector<uint64_t, 16> Offsets; in LowerCall() local 1591 int curOffset = Offsets[j]; in LowerCall() 1752 SmallVector<uint64_t, 16> Offsets; in LowerCall() local [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Transforms/Scalar/ |
H A D | SROA.cpp | 3880 assert(Offsets.Splits.empty() && in presplitLoadsAndStores() 3882 Offsets.S = &S; in presplitLoadsAndStores() 3896 assert(!Offsets.Splits.empty() && in presplitLoadsAndStores() 3898 assert(Offsets.Splits.back() == in presplitLoadsAndStores() 3905 Offsets.Splits.push_back(P.endOffset() - Offsets.S->beginOffset()); in presplitLoadsAndStores() 3996 assert(LoadSize == Offsets.S->endOffset() - Offsets.S->beginOffset() && in presplitLoadsAndStores() 4041 PartOffset = Offsets.Splits[Idx]; in presplitLoadsAndStores() 4105 Offsets.S->kill(); in presplitLoadsAndStores() 4120 assert(StoreSize == Offsets.S->endOffset() - Offsets.S->beginOffset() && in presplitLoadsAndStores() 4193 PartOffset = Offsets.Splits[Idx]; in presplitLoadsAndStores() [all …]
|
/netbsd/external/apache2/llvm/dist/llvm/lib/CodeGen/SelectionDAG/ |
H A D | ScheduleDAGSDNodes.cpp | 224 SmallVector<int64_t, 4> Offsets; in ClusterNeighboringLoads() local 252 Offsets.push_back(Offset1); in ClusterNeighboringLoads() 254 Offsets.push_back(Offset2); in ClusterNeighboringLoads() 266 llvm::sort(Offsets); in ClusterNeighboringLoads() 271 int64_t BaseOff = Offsets[0]; in ClusterNeighboringLoads() 274 for (unsigned i = 1, e = Offsets.size(); i != e; ++i) { in ClusterNeighboringLoads() 275 int64_t Offset = Offsets[i]; in ClusterNeighboringLoads()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/CodeGen/AsmPrinter/ |
H A D | DbgEntityHistoryCalculator.cpp | 140 SmallVector<size_t, 4> Offsets; in trimLocationRanges() local 234 Offsets.assign(HistoryMapEntries.size(), 0); in trimLocationRanges() 244 Offsets[EntryIdx] = CurOffset; in trimLocationRanges() 251 Entry.EndIndex -= Offsets[Entry.EndIndex]; in trimLocationRanges()
|
/netbsd/external/bsd/openldap/dist/libraries/liblunicode/ucdata/ |
H A D | format.txt | 85 unsigned short Offsets[OffsetArraySize + 1] 86 unsigned long Ranges[N], N = value of Offsets[OffsetArraySize] 88 The Bytes field provides the total byte count used for the Offsets[] and 89 Ranges[] arrays. The Offsets[] array is aligned on a 4-byte boundary and
|