1 //===- InstrProfWriter.cpp - Instrumented profiling writer ----------------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // This file contains support for writing profiling data for clang's
10 // instrumentation based PGO and coverage.
11 //
12 //===----------------------------------------------------------------------===//
13 
14 #include "llvm/ProfileData/InstrProfWriter.h"
15 #include "llvm/ADT/STLExtras.h"
16 #include "llvm/ADT/SetVector.h"
17 #include "llvm/ADT/StringRef.h"
18 #include "llvm/IR/ProfileSummary.h"
19 #include "llvm/ProfileData/InstrProf.h"
20 #include "llvm/ProfileData/MemProf.h"
21 #include "llvm/ProfileData/ProfileCommon.h"
22 #include "llvm/Support/Endian.h"
23 #include "llvm/Support/EndianStream.h"
24 #include "llvm/Support/Error.h"
25 #include "llvm/Support/MemoryBuffer.h"
26 #include "llvm/Support/OnDiskHashTable.h"
27 #include "llvm/Support/raw_ostream.h"
28 #include <cstdint>
29 #include <memory>
30 #include <string>
31 #include <tuple>
32 #include <utility>
33 #include <vector>
34 
35 using namespace llvm;
36 
37 // A struct to define how the data stream should be patched. For Indexed
38 // profiling, only uint64_t data type is needed.
39 struct PatchItem {
40   uint64_t Pos; // Where to patch.
41   uint64_t *D;  // Pointer to an array of source data.
42   int N;        // Number of elements in \c D array.
43 };
44 
45 namespace llvm {
46 
47 // A wrapper class to abstract writer stream with support of bytes
48 // back patching.
49 class ProfOStream {
50 public:
51   ProfOStream(raw_fd_ostream &FD)
52       : IsFDOStream(true), OS(FD), LE(FD, support::little) {}
53   ProfOStream(raw_string_ostream &STR)
54       : IsFDOStream(false), OS(STR), LE(STR, support::little) {}
55 
56   uint64_t tell() { return OS.tell(); }
57   void write(uint64_t V) { LE.write<uint64_t>(V); }
58   void writeByte(uint8_t V) { LE.write<uint8_t>(V); }
59 
60   // \c patch can only be called when all data is written and flushed.
61   // For raw_string_ostream, the patch is done on the target string
62   // directly and it won't be reflected in the stream's internal buffer.
63   void patch(PatchItem *P, int NItems) {
64     using namespace support;
65 
66     if (IsFDOStream) {
67       raw_fd_ostream &FDOStream = static_cast<raw_fd_ostream &>(OS);
68       const uint64_t LastPos = FDOStream.tell();
69       for (int K = 0; K < NItems; K++) {
70         FDOStream.seek(P[K].Pos);
71         for (int I = 0; I < P[K].N; I++)
72           write(P[K].D[I]);
73       }
74       // Reset the stream to the last position after patching so that users
75       // don't accidentally overwrite data. This makes it consistent with
76       // the string stream below which replaces the data directly.
77       FDOStream.seek(LastPos);
78     } else {
79       raw_string_ostream &SOStream = static_cast<raw_string_ostream &>(OS);
80       std::string &Data = SOStream.str(); // with flush
81       for (int K = 0; K < NItems; K++) {
82         for (int I = 0; I < P[K].N; I++) {
83           uint64_t Bytes = endian::byte_swap<uint64_t, little>(P[K].D[I]);
84           Data.replace(P[K].Pos + I * sizeof(uint64_t), sizeof(uint64_t),
85                        (const char *)&Bytes, sizeof(uint64_t));
86         }
87       }
88     }
89   }
90 
91   // If \c OS is an instance of \c raw_fd_ostream, this field will be
92   // true. Otherwise, \c OS will be an raw_string_ostream.
93   bool IsFDOStream;
94   raw_ostream &OS;
95   support::endian::Writer LE;
96 };
97 
98 class InstrProfRecordWriterTrait {
99 public:
100   using key_type = StringRef;
101   using key_type_ref = StringRef;
102 
103   using data_type = const InstrProfWriter::ProfilingData *const;
104   using data_type_ref = const InstrProfWriter::ProfilingData *const;
105 
106   using hash_value_type = uint64_t;
107   using offset_type = uint64_t;
108 
109   support::endianness ValueProfDataEndianness = support::little;
110   InstrProfSummaryBuilder *SummaryBuilder;
111   InstrProfSummaryBuilder *CSSummaryBuilder;
112 
113   InstrProfRecordWriterTrait() = default;
114 
115   static hash_value_type ComputeHash(key_type_ref K) {
116     return IndexedInstrProf::ComputeHash(K);
117   }
118 
119   static std::pair<offset_type, offset_type>
120   EmitKeyDataLength(raw_ostream &Out, key_type_ref K, data_type_ref V) {
121     using namespace support;
122 
123     endian::Writer LE(Out, little);
124 
125     offset_type N = K.size();
126     LE.write<offset_type>(N);
127 
128     offset_type M = 0;
129     for (const auto &ProfileData : *V) {
130       const InstrProfRecord &ProfRecord = ProfileData.second;
131       M += sizeof(uint64_t); // The function hash
132       M += sizeof(uint64_t); // The size of the Counts vector
133       M += ProfRecord.Counts.size() * sizeof(uint64_t);
134 
135       // Value data
136       M += ValueProfData::getSize(ProfileData.second);
137     }
138     LE.write<offset_type>(M);
139 
140     return std::make_pair(N, M);
141   }
142 
143   void EmitKey(raw_ostream &Out, key_type_ref K, offset_type N) {
144     Out.write(K.data(), N);
145   }
146 
147   void EmitData(raw_ostream &Out, key_type_ref, data_type_ref V, offset_type) {
148     using namespace support;
149 
150     endian::Writer LE(Out, little);
151     for (const auto &ProfileData : *V) {
152       const InstrProfRecord &ProfRecord = ProfileData.second;
153       if (NamedInstrProfRecord::hasCSFlagInHash(ProfileData.first))
154         CSSummaryBuilder->addRecord(ProfRecord);
155       else
156         SummaryBuilder->addRecord(ProfRecord);
157 
158       LE.write<uint64_t>(ProfileData.first); // Function hash
159       LE.write<uint64_t>(ProfRecord.Counts.size());
160       for (uint64_t I : ProfRecord.Counts)
161         LE.write<uint64_t>(I);
162 
163       // Write value data
164       std::unique_ptr<ValueProfData> VDataPtr =
165           ValueProfData::serializeFrom(ProfileData.second);
166       uint32_t S = VDataPtr->getSize();
167       VDataPtr->swapBytesFromHost(ValueProfDataEndianness);
168       Out.write((const char *)VDataPtr.get(), S);
169     }
170   }
171 };
172 
173 } // end namespace llvm
174 
175 InstrProfWriter::InstrProfWriter(bool Sparse,
176                                  uint64_t TemporalProfTraceReservoirSize,
177                                  uint64_t MaxTemporalProfTraceLength)
178     : Sparse(Sparse), MaxTemporalProfTraceLength(MaxTemporalProfTraceLength),
179       TemporalProfTraceReservoirSize(TemporalProfTraceReservoirSize),
180       InfoObj(new InstrProfRecordWriterTrait()) {}
181 
182 InstrProfWriter::~InstrProfWriter() { delete InfoObj; }
183 
184 // Internal interface for testing purpose only.
185 void InstrProfWriter::setValueProfDataEndianness(
186     support::endianness Endianness) {
187   InfoObj->ValueProfDataEndianness = Endianness;
188 }
189 
190 void InstrProfWriter::setOutputSparse(bool Sparse) {
191   this->Sparse = Sparse;
192 }
193 
194 void InstrProfWriter::addRecord(NamedInstrProfRecord &&I, uint64_t Weight,
195                                 function_ref<void(Error)> Warn) {
196   auto Name = I.Name;
197   auto Hash = I.Hash;
198   addRecord(Name, Hash, std::move(I), Weight, Warn);
199 }
200 
201 void InstrProfWriter::overlapRecord(NamedInstrProfRecord &&Other,
202                                     OverlapStats &Overlap,
203                                     OverlapStats &FuncLevelOverlap,
204                                     const OverlapFuncFilters &FuncFilter) {
205   auto Name = Other.Name;
206   auto Hash = Other.Hash;
207   Other.accumulateCounts(FuncLevelOverlap.Test);
208   if (!FunctionData.contains(Name)) {
209     Overlap.addOneUnique(FuncLevelOverlap.Test);
210     return;
211   }
212   if (FuncLevelOverlap.Test.CountSum < 1.0f) {
213     Overlap.Overlap.NumEntries += 1;
214     return;
215   }
216   auto &ProfileDataMap = FunctionData[Name];
217   bool NewFunc;
218   ProfilingData::iterator Where;
219   std::tie(Where, NewFunc) =
220       ProfileDataMap.insert(std::make_pair(Hash, InstrProfRecord()));
221   if (NewFunc) {
222     Overlap.addOneMismatch(FuncLevelOverlap.Test);
223     return;
224   }
225   InstrProfRecord &Dest = Where->second;
226 
227   uint64_t ValueCutoff = FuncFilter.ValueCutoff;
228   if (!FuncFilter.NameFilter.empty() && Name.contains(FuncFilter.NameFilter))
229     ValueCutoff = 0;
230 
231   Dest.overlap(Other, Overlap, FuncLevelOverlap, ValueCutoff);
232 }
233 
234 void InstrProfWriter::addRecord(StringRef Name, uint64_t Hash,
235                                 InstrProfRecord &&I, uint64_t Weight,
236                                 function_ref<void(Error)> Warn) {
237   auto &ProfileDataMap = FunctionData[Name];
238 
239   bool NewFunc;
240   ProfilingData::iterator Where;
241   std::tie(Where, NewFunc) =
242       ProfileDataMap.insert(std::make_pair(Hash, InstrProfRecord()));
243   InstrProfRecord &Dest = Where->second;
244 
245   auto MapWarn = [&](instrprof_error E) {
246     Warn(make_error<InstrProfError>(E));
247   };
248 
249   if (NewFunc) {
250     // We've never seen a function with this name and hash, add it.
251     Dest = std::move(I);
252     if (Weight > 1)
253       Dest.scale(Weight, 1, MapWarn);
254   } else {
255     // We're updating a function we've seen before.
256     Dest.merge(I, Weight, MapWarn);
257   }
258 
259   Dest.sortValueData();
260 }
261 
262 void InstrProfWriter::addMemProfRecord(
263     const Function::GUID Id, const memprof::IndexedMemProfRecord &Record) {
264   auto Result = MemProfRecordData.insert({Id, Record});
265   // If we inserted a new record then we are done.
266   if (Result.second) {
267     return;
268   }
269   memprof::IndexedMemProfRecord &Existing = Result.first->second;
270   Existing.merge(Record);
271 }
272 
273 bool InstrProfWriter::addMemProfFrame(const memprof::FrameId Id,
274                                       const memprof::Frame &Frame,
275                                       function_ref<void(Error)> Warn) {
276   auto Result = MemProfFrameData.insert({Id, Frame});
277   // If a mapping already exists for the current frame id and it does not
278   // match the new mapping provided then reset the existing contents and bail
279   // out. We don't support the merging of memprof data whose Frame -> Id
280   // mapping across profiles is inconsistent.
281   if (!Result.second && Result.first->second != Frame) {
282     Warn(make_error<InstrProfError>(instrprof_error::malformed,
283                                     "frame to id mapping mismatch"));
284     return false;
285   }
286   return true;
287 }
288 
289 void InstrProfWriter::addBinaryIds(ArrayRef<llvm::object::BuildID> BIs) {
290   llvm::append_range(BinaryIds, BIs);
291 }
292 
293 void InstrProfWriter::addTemporalProfileTrace(TemporalProfTraceTy Trace) {
294   if (Trace.FunctionNameRefs.size() > MaxTemporalProfTraceLength)
295     Trace.FunctionNameRefs.resize(MaxTemporalProfTraceLength);
296   if (Trace.FunctionNameRefs.empty())
297     return;
298 
299   if (TemporalProfTraceStreamSize < TemporalProfTraceReservoirSize) {
300     // Simply append the trace if we have not yet hit our reservoir size limit.
301     TemporalProfTraces.push_back(std::move(Trace));
302   } else {
303     // Otherwise, replace a random trace in the stream.
304     std::uniform_int_distribution<uint64_t> Distribution(
305         0, TemporalProfTraceStreamSize);
306     uint64_t RandomIndex = Distribution(RNG);
307     if (RandomIndex < TemporalProfTraces.size())
308       TemporalProfTraces[RandomIndex] = std::move(Trace);
309   }
310   ++TemporalProfTraceStreamSize;
311 }
312 
313 void InstrProfWriter::addTemporalProfileTraces(
314     SmallVectorImpl<TemporalProfTraceTy> &SrcTraces, uint64_t SrcStreamSize) {
315   // Assume that the source has the same reservoir size as the destination to
316   // avoid needing to record it in the indexed profile format.
317   bool IsDestSampled =
318       (TemporalProfTraceStreamSize > TemporalProfTraceReservoirSize);
319   bool IsSrcSampled = (SrcStreamSize > TemporalProfTraceReservoirSize);
320   if (!IsDestSampled && IsSrcSampled) {
321     // If one of the traces are sampled, ensure that it belongs to Dest.
322     std::swap(TemporalProfTraces, SrcTraces);
323     std::swap(TemporalProfTraceStreamSize, SrcStreamSize);
324     std::swap(IsDestSampled, IsSrcSampled);
325   }
326   if (!IsSrcSampled) {
327     // If the source stream is not sampled, we add each source trace normally.
328     for (auto &Trace : SrcTraces)
329       addTemporalProfileTrace(std::move(Trace));
330     return;
331   }
332   // Otherwise, we find the traces that would have been removed if we added
333   // the whole source stream.
334   SmallSetVector<uint64_t, 8> IndicesToReplace;
335   for (uint64_t I = 0; I < SrcStreamSize; I++) {
336     std::uniform_int_distribution<uint64_t> Distribution(
337         0, TemporalProfTraceStreamSize);
338     uint64_t RandomIndex = Distribution(RNG);
339     if (RandomIndex < TemporalProfTraces.size())
340       IndicesToReplace.insert(RandomIndex);
341     ++TemporalProfTraceStreamSize;
342   }
343   // Then we insert a random sample of the source traces.
344   llvm::shuffle(SrcTraces.begin(), SrcTraces.end(), RNG);
345   for (const auto &[Index, Trace] : llvm::zip(IndicesToReplace, SrcTraces))
346     TemporalProfTraces[Index] = std::move(Trace);
347 }
348 
349 void InstrProfWriter::mergeRecordsFromWriter(InstrProfWriter &&IPW,
350                                              function_ref<void(Error)> Warn) {
351   for (auto &I : IPW.FunctionData)
352     for (auto &Func : I.getValue())
353       addRecord(I.getKey(), Func.first, std::move(Func.second), 1, Warn);
354 
355   BinaryIds.reserve(BinaryIds.size() + IPW.BinaryIds.size());
356   for (auto &I : IPW.BinaryIds)
357     addBinaryIds(I);
358 
359   addTemporalProfileTraces(IPW.TemporalProfTraces,
360                            IPW.TemporalProfTraceStreamSize);
361 
362   MemProfFrameData.reserve(IPW.MemProfFrameData.size());
363   for (auto &I : IPW.MemProfFrameData) {
364     // If we weren't able to add the frame mappings then it doesn't make sense
365     // to try to merge the records from this profile.
366     if (!addMemProfFrame(I.first, I.second, Warn))
367       return;
368   }
369 
370   MemProfRecordData.reserve(IPW.MemProfRecordData.size());
371   for (auto &I : IPW.MemProfRecordData) {
372     addMemProfRecord(I.first, I.second);
373   }
374 }
375 
376 bool InstrProfWriter::shouldEncodeData(const ProfilingData &PD) {
377   if (!Sparse)
378     return true;
379   for (const auto &Func : PD) {
380     const InstrProfRecord &IPR = Func.second;
381     if (llvm::any_of(IPR.Counts, [](uint64_t Count) { return Count > 0; }))
382       return true;
383   }
384   return false;
385 }
386 
387 static void setSummary(IndexedInstrProf::Summary *TheSummary,
388                        ProfileSummary &PS) {
389   using namespace IndexedInstrProf;
390 
391   const std::vector<ProfileSummaryEntry> &Res = PS.getDetailedSummary();
392   TheSummary->NumSummaryFields = Summary::NumKinds;
393   TheSummary->NumCutoffEntries = Res.size();
394   TheSummary->set(Summary::MaxFunctionCount, PS.getMaxFunctionCount());
395   TheSummary->set(Summary::MaxBlockCount, PS.getMaxCount());
396   TheSummary->set(Summary::MaxInternalBlockCount, PS.getMaxInternalCount());
397   TheSummary->set(Summary::TotalBlockCount, PS.getTotalCount());
398   TheSummary->set(Summary::TotalNumBlocks, PS.getNumCounts());
399   TheSummary->set(Summary::TotalNumFunctions, PS.getNumFunctions());
400   for (unsigned I = 0; I < Res.size(); I++)
401     TheSummary->setEntry(I, Res[I]);
402 }
403 
404 Error InstrProfWriter::writeImpl(ProfOStream &OS) {
405   using namespace IndexedInstrProf;
406   using namespace support;
407 
408   OnDiskChainedHashTableGenerator<InstrProfRecordWriterTrait> Generator;
409 
410   InstrProfSummaryBuilder ISB(ProfileSummaryBuilder::DefaultCutoffs);
411   InfoObj->SummaryBuilder = &ISB;
412   InstrProfSummaryBuilder CSISB(ProfileSummaryBuilder::DefaultCutoffs);
413   InfoObj->CSSummaryBuilder = &CSISB;
414 
415   // Populate the hash table generator.
416   SmallVector<std::pair<StringRef, const ProfilingData *>, 0> OrderedData;
417   for (const auto &I : FunctionData)
418     if (shouldEncodeData(I.getValue()))
419       OrderedData.emplace_back((I.getKey()), &I.getValue());
420   llvm::sort(OrderedData, less_first());
421   for (const auto &I : OrderedData)
422     Generator.insert(I.first, I.second);
423 
424   // Write the header.
425   IndexedInstrProf::Header Header;
426   Header.Magic = IndexedInstrProf::Magic;
427   Header.Version = IndexedInstrProf::ProfVersion::CurrentVersion;
428   if (static_cast<bool>(ProfileKind & InstrProfKind::IRInstrumentation))
429     Header.Version |= VARIANT_MASK_IR_PROF;
430   if (static_cast<bool>(ProfileKind & InstrProfKind::ContextSensitive))
431     Header.Version |= VARIANT_MASK_CSIR_PROF;
432   if (static_cast<bool>(ProfileKind &
433                         InstrProfKind::FunctionEntryInstrumentation))
434     Header.Version |= VARIANT_MASK_INSTR_ENTRY;
435   if (static_cast<bool>(ProfileKind & InstrProfKind::SingleByteCoverage))
436     Header.Version |= VARIANT_MASK_BYTE_COVERAGE;
437   if (static_cast<bool>(ProfileKind & InstrProfKind::FunctionEntryOnly))
438     Header.Version |= VARIANT_MASK_FUNCTION_ENTRY_ONLY;
439   if (static_cast<bool>(ProfileKind & InstrProfKind::MemProf))
440     Header.Version |= VARIANT_MASK_MEMPROF;
441   if (static_cast<bool>(ProfileKind & InstrProfKind::TemporalProfile))
442     Header.Version |= VARIANT_MASK_TEMPORAL_PROF;
443 
444   Header.Unused = 0;
445   Header.HashType = static_cast<uint64_t>(IndexedInstrProf::HashType);
446   Header.HashOffset = 0;
447   Header.MemProfOffset = 0;
448   Header.BinaryIdOffset = 0;
449   Header.TemporalProfTracesOffset = 0;
450   int N = sizeof(IndexedInstrProf::Header) / sizeof(uint64_t);
451 
452   // Only write out all the fields except 'HashOffset', 'MemProfOffset',
453   // 'BinaryIdOffset' and `TemporalProfTracesOffset`. We need to remember the
454   // offset of these fields to allow back patching later.
455   for (int I = 0; I < N - 4; I++)
456     OS.write(reinterpret_cast<uint64_t *>(&Header)[I]);
457 
458   // Save the location of Header.HashOffset field in \c OS.
459   uint64_t HashTableStartFieldOffset = OS.tell();
460   // Reserve the space for HashOffset field.
461   OS.write(0);
462 
463   // Save the location of MemProf profile data. This is stored in two parts as
464   // the schema and as a separate on-disk chained hashtable.
465   uint64_t MemProfSectionOffset = OS.tell();
466   // Reserve space for the MemProf table field to be patched later if this
467   // profile contains memory profile information.
468   OS.write(0);
469 
470   // Save the location of binary ids section.
471   uint64_t BinaryIdSectionOffset = OS.tell();
472   // Reserve space for the BinaryIdOffset field to be patched later if this
473   // profile contains binary ids.
474   OS.write(0);
475 
476   uint64_t TemporalProfTracesOffset = OS.tell();
477   OS.write(0);
478 
479   // Reserve space to write profile summary data.
480   uint32_t NumEntries = ProfileSummaryBuilder::DefaultCutoffs.size();
481   uint32_t SummarySize = Summary::getSize(Summary::NumKinds, NumEntries);
482   // Remember the summary offset.
483   uint64_t SummaryOffset = OS.tell();
484   for (unsigned I = 0; I < SummarySize / sizeof(uint64_t); I++)
485     OS.write(0);
486   uint64_t CSSummaryOffset = 0;
487   uint64_t CSSummarySize = 0;
488   if (static_cast<bool>(ProfileKind & InstrProfKind::ContextSensitive)) {
489     CSSummaryOffset = OS.tell();
490     CSSummarySize = SummarySize / sizeof(uint64_t);
491     for (unsigned I = 0; I < CSSummarySize; I++)
492       OS.write(0);
493   }
494 
495   // Write the hash table.
496   uint64_t HashTableStart = Generator.Emit(OS.OS, *InfoObj);
497 
498   // Write the MemProf profile data if we have it. This includes a simple schema
499   // with the format described below followed by the hashtable:
500   // uint64_t RecordTableOffset = RecordTableGenerator.Emit
501   // uint64_t FramePayloadOffset = Stream offset before emitting the frame table
502   // uint64_t FrameTableOffset = FrameTableGenerator.Emit
503   // uint64_t Num schema entries
504   // uint64_t Schema entry 0
505   // uint64_t Schema entry 1
506   // ....
507   // uint64_t Schema entry N - 1
508   // OnDiskChainedHashTable MemProfRecordData
509   // OnDiskChainedHashTable MemProfFrameData
510   uint64_t MemProfSectionStart = 0;
511   if (static_cast<bool>(ProfileKind & InstrProfKind::MemProf)) {
512     MemProfSectionStart = OS.tell();
513     OS.write(0ULL); // Reserve space for the memprof record table offset.
514     OS.write(0ULL); // Reserve space for the memprof frame payload offset.
515     OS.write(0ULL); // Reserve space for the memprof frame table offset.
516 
517     auto Schema = memprof::PortableMemInfoBlock::getSchema();
518     OS.write(static_cast<uint64_t>(Schema.size()));
519     for (const auto Id : Schema) {
520       OS.write(static_cast<uint64_t>(Id));
521     }
522 
523     auto RecordWriter = std::make_unique<memprof::RecordWriterTrait>();
524     RecordWriter->Schema = &Schema;
525     OnDiskChainedHashTableGenerator<memprof::RecordWriterTrait>
526         RecordTableGenerator;
527     for (auto &I : MemProfRecordData) {
528       // Insert the key (func hash) and value (memprof record).
529       RecordTableGenerator.insert(I.first, I.second);
530     }
531 
532     uint64_t RecordTableOffset =
533         RecordTableGenerator.Emit(OS.OS, *RecordWriter);
534 
535     uint64_t FramePayloadOffset = OS.tell();
536 
537     auto FrameWriter = std::make_unique<memprof::FrameWriterTrait>();
538     OnDiskChainedHashTableGenerator<memprof::FrameWriterTrait>
539         FrameTableGenerator;
540     for (auto &I : MemProfFrameData) {
541       // Insert the key (frame id) and value (frame contents).
542       FrameTableGenerator.insert(I.first, I.second);
543     }
544 
545     uint64_t FrameTableOffset = FrameTableGenerator.Emit(OS.OS, *FrameWriter);
546 
547     PatchItem PatchItems[] = {
548         {MemProfSectionStart, &RecordTableOffset, 1},
549         {MemProfSectionStart + sizeof(uint64_t), &FramePayloadOffset, 1},
550         {MemProfSectionStart + 2 * sizeof(uint64_t), &FrameTableOffset, 1},
551     };
552     OS.patch(PatchItems, 3);
553   }
554 
555   // BinaryIdSection has two parts:
556   // 1. uint64_t BinaryIdsSectionSize
557   // 2. list of binary ids that consist of:
558   //    a. uint64_t BinaryIdLength
559   //    b. uint8_t  BinaryIdData
560   //    c. uint8_t  Padding (if necessary)
561   uint64_t BinaryIdSectionStart = OS.tell();
562   // Calculate size of binary section.
563   uint64_t BinaryIdsSectionSize = 0;
564 
565   // Remove duplicate binary ids.
566   llvm::sort(BinaryIds);
567   BinaryIds.erase(std::unique(BinaryIds.begin(), BinaryIds.end()),
568                   BinaryIds.end());
569 
570   for (auto BI : BinaryIds) {
571     // Increment by binary id length data type size.
572     BinaryIdsSectionSize += sizeof(uint64_t);
573     // Increment by binary id data length, aligned to 8 bytes.
574     BinaryIdsSectionSize += alignToPowerOf2(BI.size(), sizeof(uint64_t));
575   }
576   // Write binary ids section size.
577   OS.write(BinaryIdsSectionSize);
578 
579   for (auto BI : BinaryIds) {
580     uint64_t BILen = BI.size();
581     // Write binary id length.
582     OS.write(BILen);
583     // Write binary id data.
584     for (unsigned K = 0; K < BILen; K++)
585       OS.writeByte(BI[K]);
586     // Write padding if necessary.
587     uint64_t PaddingSize = alignToPowerOf2(BILen, sizeof(uint64_t)) - BILen;
588     for (unsigned K = 0; K < PaddingSize; K++)
589       OS.writeByte(0);
590   }
591 
592   uint64_t TemporalProfTracesSectionStart = 0;
593   if (static_cast<bool>(ProfileKind & InstrProfKind::TemporalProfile)) {
594     TemporalProfTracesSectionStart = OS.tell();
595     OS.write(TemporalProfTraces.size());
596     OS.write(TemporalProfTraceStreamSize);
597     for (auto &Trace : TemporalProfTraces) {
598       OS.write(Trace.Weight);
599       OS.write(Trace.FunctionNameRefs.size());
600       for (auto &NameRef : Trace.FunctionNameRefs)
601         OS.write(NameRef);
602     }
603   }
604 
605   // Allocate space for data to be serialized out.
606   std::unique_ptr<IndexedInstrProf::Summary> TheSummary =
607       IndexedInstrProf::allocSummary(SummarySize);
608   // Compute the Summary and copy the data to the data
609   // structure to be serialized out (to disk or buffer).
610   std::unique_ptr<ProfileSummary> PS = ISB.getSummary();
611   setSummary(TheSummary.get(), *PS);
612   InfoObj->SummaryBuilder = nullptr;
613 
614   // For Context Sensitive summary.
615   std::unique_ptr<IndexedInstrProf::Summary> TheCSSummary = nullptr;
616   if (static_cast<bool>(ProfileKind & InstrProfKind::ContextSensitive)) {
617     TheCSSummary = IndexedInstrProf::allocSummary(SummarySize);
618     std::unique_ptr<ProfileSummary> CSPS = CSISB.getSummary();
619     setSummary(TheCSSummary.get(), *CSPS);
620   }
621   InfoObj->CSSummaryBuilder = nullptr;
622 
623   // Now do the final patch:
624   PatchItem PatchItems[] = {
625       // Patch the Header.HashOffset field.
626       {HashTableStartFieldOffset, &HashTableStart, 1},
627       // Patch the Header.MemProfOffset (=0 for profiles without MemProf
628       // data).
629       {MemProfSectionOffset, &MemProfSectionStart, 1},
630       // Patch the Header.BinaryIdSectionOffset.
631       {BinaryIdSectionOffset, &BinaryIdSectionStart, 1},
632       // Patch the Header.TemporalProfTracesOffset (=0 for profiles without
633       // traces).
634       {TemporalProfTracesOffset, &TemporalProfTracesSectionStart, 1},
635       // Patch the summary data.
636       {SummaryOffset, reinterpret_cast<uint64_t *>(TheSummary.get()),
637        (int)(SummarySize / sizeof(uint64_t))},
638       {CSSummaryOffset, reinterpret_cast<uint64_t *>(TheCSSummary.get()),
639        (int)CSSummarySize}};
640 
641   OS.patch(PatchItems, std::size(PatchItems));
642 
643   for (const auto &I : FunctionData)
644     for (const auto &F : I.getValue())
645       if (Error E = validateRecord(F.second))
646         return E;
647 
648   return Error::success();
649 }
650 
651 Error InstrProfWriter::write(raw_fd_ostream &OS) {
652   // Write the hash table.
653   ProfOStream POS(OS);
654   return writeImpl(POS);
655 }
656 
657 Error InstrProfWriter::write(raw_string_ostream &OS) {
658   ProfOStream POS(OS);
659   return writeImpl(POS);
660 }
661 
662 std::unique_ptr<MemoryBuffer> InstrProfWriter::writeBuffer() {
663   std::string Data;
664   raw_string_ostream OS(Data);
665   // Write the hash table.
666   if (Error E = write(OS))
667     return nullptr;
668   // Return this in an aligned memory buffer.
669   return MemoryBuffer::getMemBufferCopy(Data);
670 }
671 
672 static const char *ValueProfKindStr[] = {
673 #define VALUE_PROF_KIND(Enumerator, Value, Descr) #Enumerator,
674 #include "llvm/ProfileData/InstrProfData.inc"
675 };
676 
677 Error InstrProfWriter::validateRecord(const InstrProfRecord &Func) {
678   for (uint32_t VK = 0; VK <= IPVK_Last; VK++) {
679     uint32_t NS = Func.getNumValueSites(VK);
680     if (!NS)
681       continue;
682     for (uint32_t S = 0; S < NS; S++) {
683       uint32_t ND = Func.getNumValueDataForSite(VK, S);
684       std::unique_ptr<InstrProfValueData[]> VD = Func.getValueForSite(VK, S);
685       DenseSet<uint64_t> SeenValues;
686       for (uint32_t I = 0; I < ND; I++)
687         if ((VK != IPVK_IndirectCallTarget) && !SeenValues.insert(VD[I].Value).second)
688           return make_error<InstrProfError>(instrprof_error::invalid_prof);
689     }
690   }
691 
692   return Error::success();
693 }
694 
695 void InstrProfWriter::writeRecordInText(StringRef Name, uint64_t Hash,
696                                         const InstrProfRecord &Func,
697                                         InstrProfSymtab &Symtab,
698                                         raw_fd_ostream &OS) {
699   OS << Name << "\n";
700   OS << "# Func Hash:\n" << Hash << "\n";
701   OS << "# Num Counters:\n" << Func.Counts.size() << "\n";
702   OS << "# Counter Values:\n";
703   for (uint64_t Count : Func.Counts)
704     OS << Count << "\n";
705 
706   uint32_t NumValueKinds = Func.getNumValueKinds();
707   if (!NumValueKinds) {
708     OS << "\n";
709     return;
710   }
711 
712   OS << "# Num Value Kinds:\n" << Func.getNumValueKinds() << "\n";
713   for (uint32_t VK = 0; VK < IPVK_Last + 1; VK++) {
714     uint32_t NS = Func.getNumValueSites(VK);
715     if (!NS)
716       continue;
717     OS << "# ValueKind = " << ValueProfKindStr[VK] << ":\n" << VK << "\n";
718     OS << "# NumValueSites:\n" << NS << "\n";
719     for (uint32_t S = 0; S < NS; S++) {
720       uint32_t ND = Func.getNumValueDataForSite(VK, S);
721       OS << ND << "\n";
722       std::unique_ptr<InstrProfValueData[]> VD = Func.getValueForSite(VK, S);
723       for (uint32_t I = 0; I < ND; I++) {
724         if (VK == IPVK_IndirectCallTarget)
725           OS << Symtab.getFuncNameOrExternalSymbol(VD[I].Value) << ":"
726              << VD[I].Count << "\n";
727         else
728           OS << VD[I].Value << ":" << VD[I].Count << "\n";
729       }
730     }
731   }
732 
733   OS << "\n";
734 }
735 
736 Error InstrProfWriter::writeText(raw_fd_ostream &OS) {
737   // Check CS first since it implies an IR level profile.
738   if (static_cast<bool>(ProfileKind & InstrProfKind::ContextSensitive))
739     OS << "# CSIR level Instrumentation Flag\n:csir\n";
740   else if (static_cast<bool>(ProfileKind & InstrProfKind::IRInstrumentation))
741     OS << "# IR level Instrumentation Flag\n:ir\n";
742 
743   if (static_cast<bool>(ProfileKind &
744                         InstrProfKind::FunctionEntryInstrumentation))
745     OS << "# Always instrument the function entry block\n:entry_first\n";
746   InstrProfSymtab Symtab;
747 
748   using FuncPair = detail::DenseMapPair<uint64_t, InstrProfRecord>;
749   using RecordType = std::pair<StringRef, FuncPair>;
750   SmallVector<RecordType, 4> OrderedFuncData;
751 
752   for (const auto &I : FunctionData) {
753     if (shouldEncodeData(I.getValue())) {
754       if (Error E = Symtab.addFuncName(I.getKey()))
755         return E;
756       for (const auto &Func : I.getValue())
757         OrderedFuncData.push_back(std::make_pair(I.getKey(), Func));
758     }
759   }
760 
761   if (static_cast<bool>(ProfileKind & InstrProfKind::TemporalProfile))
762     writeTextTemporalProfTraceData(OS, Symtab);
763 
764   llvm::sort(OrderedFuncData, [](const RecordType &A, const RecordType &B) {
765     return std::tie(A.first, A.second.first) <
766            std::tie(B.first, B.second.first);
767   });
768 
769   for (const auto &record : OrderedFuncData) {
770     const StringRef &Name = record.first;
771     const FuncPair &Func = record.second;
772     writeRecordInText(Name, Func.first, Func.second, Symtab, OS);
773   }
774 
775   for (const auto &record : OrderedFuncData) {
776     const FuncPair &Func = record.second;
777     if (Error E = validateRecord(Func.second))
778       return E;
779   }
780 
781   return Error::success();
782 }
783 
784 void InstrProfWriter::writeTextTemporalProfTraceData(raw_fd_ostream &OS,
785                                                      InstrProfSymtab &Symtab) {
786   OS << ":temporal_prof_traces\n";
787   OS << "# Num Temporal Profile Traces:\n" << TemporalProfTraces.size() << "\n";
788   OS << "# Temporal Profile Trace Stream Size:\n"
789      << TemporalProfTraceStreamSize << "\n";
790   for (auto &Trace : TemporalProfTraces) {
791     OS << "# Weight:\n" << Trace.Weight << "\n";
792     for (auto &NameRef : Trace.FunctionNameRefs)
793       OS << Symtab.getFuncName(NameRef) << ",";
794     OS << "\n";
795   }
796   OS << "\n";
797 }
798