/netbsd/external/apache2/llvm/dist/llvm/lib/Object/ |
H A D | ELF.cpp | 565 ELFFile<ELFT>::toMappedAddr(uint64_t VAddr, WarningHandler WarnHandler) const { in toMappedAddr() argument 588 LoadSegments, VAddr, [](uint64_t VAddr, const Elf_Phdr_Impl<ELFT> *Phdr) { in toMappedAddr() argument 589 return VAddr < Phdr->p_vaddr; in toMappedAddr() 594 Twine::utohexstr(VAddr)); in toMappedAddr() 597 uint64_t Delta = VAddr - Phdr.p_vaddr; in toMappedAddr() 600 Twine::utohexstr(VAddr)); in toMappedAddr() 605 Twine::utohexstr(VAddr) + " to the segment with index " + in toMappedAddr()
|
/netbsd/external/apache2/llvm/dist/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPUISelDAGToDAG.cpp | 1426 VAddr = N0; in SelectMUBUF() 1430 VAddr = N2; in SelectMUBUF() 1435 VAddr = N3; in SelectMUBUF() 1442 VAddr = N0; in SelectMUBUF() 1539 VAddr = SDValue(MovHighBits, 0); in SelectMUBUFScratchOffen() 1578 std::tie(VAddr, SOffset) = foldFrameIndex(N0); in SelectMUBUFScratchOffen() 1585 std::tie(VAddr, SOffset) = foldFrameIndex(Addr); in SelectMUBUFScratchOffen() 1620 SDValue Ptr, VAddr, Offen, Idxen, Addr64; in SelectMUBUFOffset() local 1745 VAddr = Addr; in SelectFlatOffsetImpl() 1765 return SelectFlatOffsetImpl(N, Addr, VAddr, Offset, in SelectScratchOffset() [all …]
|
H A D | AMDGPUInstructionSelector.cpp | 2404 if (VAddr) in selectG_AMDGPU_ATOMIC_CMPXCHG() 2405 MIB.addReg(VAddr); in selectG_AMDGPU_ATOMIC_CMPXCHG() 3752 VAddr = PtrBase; in selectMUBUFScratchOffen() 4104 VAddr = N0; in selectMUBUFAddr64Impl() 4107 VAddr = N2; in selectMUBUFAddr64Impl() 4112 VAddr = N3; in selectMUBUFAddr64Impl() 4116 VAddr = N0; in selectMUBUFAddr64Impl() 4156 Register VAddr; in selectMUBUFAddr64() local 4171 MIB.addReg(VAddr); in selectMUBUFAddr64() 4216 Register VAddr; in selectMUBUFAddr64Atomic() local [all …]
|
H A D | SILoadStoreOptimizer.cpp | 89 bool VAddr = false; member 431 Result.VAddr = true; in getRegs() 446 Result.VAddr = true; in getRegs() 457 Result.VAddr = true; in getRegs() 554 if (Regs.VAddr) in setMI() 1336 if (Regs.VAddr) in mergeBufferLoadPair() 1397 if (Regs.VAddr) in mergeTBufferLoadPair() 1475 if (Regs.VAddr) in mergeTBufferStorePair() 1630 if (Regs.VAddr) in mergeBufferStorePair()
|
H A D | AMDGPUInstructionSelector.h | 255 bool selectMUBUFAddr64Impl(MachineOperand &Root, Register &VAddr,
|
H A D | SIInstrInfo.cpp | 5011 MachineOperand &VAddr = Inst.getOperand(OldVAddrIdx); in moveFlatAddrToVGPR() local 5012 VAddrDef = MRI.getUniqueVRegDef(VAddr.getReg()); in moveFlatAddrToVGPR() 5572 MachineOperand *VAddr = getNamedOperand(MI, AMDGPU::OpName::vaddr); in legalizeOperands() local 5573 if (VAddr && AMDGPU::getIfAddr64Inst(MI.getOpcode()) != -1) { in legalizeOperands() 5592 .addReg(VAddr->getReg(), 0, AMDGPU::sub0) in legalizeOperands() 5599 .addReg(VAddr->getReg(), 0, AMDGPU::sub1) in legalizeOperands() 5610 VAddr->setReg(NewVAddr); in legalizeOperands() 5612 } else if (!VAddr && ST.hasAddr64()) { in legalizeOperands()
|
H A D | AMDGPULegalizerInfo.cpp | 4088 auto VAddr = B.buildBuildVector(LLT::vector(NumAddrRegs, 32), AddrRegs); in convertImageAddrToPacked() local 4089 MI.getOperand(DimIdx).setReg(VAddr.getReg(0)); in convertImageAddrToPacked()
|
H A D | SIISelLowering.cpp | 6162 SDValue VAddr; in lowerImage() local 6164 VAddr = getBuildDwordsVector(DAG, DL, VAddrs); in lowerImage() 6235 Ops.push_back(VAddr); in lowerImage() 6262 UseNSA ? VAddrs.size() : VAddr.getValueType().getSizeInBits() / 32; in lowerImage()
|
/netbsd/external/apache2/llvm/dist/llvm/tools/llvm-objcopy/ELF/ |
H A D | Object.cpp | 46 Phdr.p_vaddr = Seg.VAddr; in writePhdr() 1221 return Seg.VAddr <= Sec.Addr && in sectionWithinSegment() 1222 Seg.VAddr + Seg.MemSize >= Sec.Addr + SecSize; in sectionWithinSegment() 1438 Seg.VAddr = Phdr.p_vaddr; in readProgramHeaders() 1464 PrHdr.OriginalOffset = PrHdr.Offset = PrHdr.VAddr = EhdrOffset + Ehdr.e_phoff; in readProgramHeaders() 2239 alignTo(Offset, std::max<uint64_t>(Seg->Align, 1), Seg->VAddr); in layoutSegments() 2361 ElfHdr.VAddr = 0; in initEhdrSegment()
|
H A D | Object.h | 453 uint64_t VAddr = 0; variable
|
/netbsd/external/apache2/llvm/dist/llvm/include/llvm/ObjectYAML/ |
H A D | ELFYAML.h | 682 llvm::yaml::Hex64 VAddr; member
|
/netbsd/external/apache2/llvm/dist/llvm/lib/ObjectYAML/ |
H A D | ELFYAML.cpp | 1027 IO.mapOptional("VAddr", Phdr.VAddr, Hex64(0)); in mapping() 1028 IO.mapOptional("PAddr", Phdr.PAddr, Phdr.VAddr); in mapping()
|
H A D | ELFEmitter.cpp | 508 Phdr.p_vaddr = YamlPhdr.VAddr; in initProgramHeaders()
|
/netbsd/external/apache2/llvm/dist/llvm/include/llvm/Object/ |
H A D | ELF.h | 258 toMappedAddr(uint64_t VAddr,
|
/netbsd/external/apache2/llvm/dist/llvm/tools/obj2yaml/ |
H A D | elf2yaml.cpp | 452 PH.VAddr = Phdr.p_vaddr; in dumpProgramHeaders()
|
/netbsd/external/apache2/llvm/dist/llvm/tools/llvm-readobj/ |
H A D | ELFDumper.cpp | 1846 auto toMappedAddr = [&](uint64_t Tag, uint64_t VAddr) -> const uint8_t * { in parseDynamicTable() argument 1847 auto MappedAddrOrError = Obj.toMappedAddr(VAddr, [&](const Twine &Msg) { in parseDynamicTable()
|