/freebsd/contrib/llvm-project/llvm/lib/Target/Mips/ |
H A D | MipsLegalizerInfo.cpp | 25 unsigned MemSize; member 34 if (MemSize > AlignInBits) in isUnalignedMemmoryAccess() 53 if (Val.MemSize != QueryMemSize) in CheckTy0Ty1MemSizeAlign() 347 unsigned MemSize = (**MI.memoperands_begin()).getSize(); in legalizeCustom() local 353 assert(MemSize <= 8 && "MemSize is too large"); in legalizeCustom() 359 if (isPowerOf2_64(MemSize)) { in legalizeCustom() 360 P2HalfMemSize = RemMemSize = MemSize / 2; in legalizeCustom() 362 P2HalfMemSize = 1 << Log2_32(MemSize); in legalizeCustom() 363 RemMemSize = MemSize - P2HalfMemSize; in legalizeCustom() 383 if (MI.getOpcode() == G_STORE && MemSize <= 4) { in legalizeCustom() [all …]
|
/freebsd/sys/dev/pms/freebsd/driver/common/ |
H A D | lxproto.h | 52 U32 MemSize,
|
H A D | lxutil.c | 628 U32 MemSize, in agtiapi_MemAlloc() argument 640 *VirtAlloc = malloc( MemSize + Align, M_PMC_MMAL, M_ZERO | M_NOWAIT ); in agtiapi_MemAlloc() 651 pmsc->typhIdx += residAlign + MemSize; // update index in agtiapi_MemAlloc() 659 pmsc->tyPhsIx += residAlign + MemSize; // update index in agtiapi_MemAlloc()
|
/freebsd/contrib/llvm-project/llvm/include/llvm/CodeGen/ |
H A D | RuntimeLibcalls.h | 89 AtomicOrdering Order, uint64_t MemSize);
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/GlobalISel/ |
H A D | CallLowering.cpp | 812 uint64_t MemSize = Flags.getByValSize(); in handleAssignments() local 817 Handler.getStackAddress(MemSize, Offset, DstMPO, Flags); in handleAssignments() 835 MemSize, VA); in handleAssignments() 1157 const MachinePointerInfo &SrcPtrInfo, Align SrcAlign, uint64_t MemSize, in copyArgumentMemory() argument 1162 MachineMemOperand::MOLoad | MachineMemOperand::MODereferenceable, MemSize, in copyArgumentMemory() 1168 MemSize, DstAlign); in copyArgumentMemory() 1173 auto SizeConst = MIRBuilder.buildConstant(SizeTy, MemSize); in copyArgumentMemory()
|
H A D | LegalizerHelper.cpp | 715 uint64_t MemSize = MemType.getSizeInBytes(); in getOutlineAtomicLibcall() local 727 return getOutlineAtomicHelper(LC, Ordering, MemSize); in getOutlineAtomicLibcall() 731 return getOutlineAtomicHelper(LC, Ordering, MemSize); in getOutlineAtomicLibcall() 736 return getOutlineAtomicHelper(LC, Ordering, MemSize); in getOutlineAtomicLibcall() 740 return getOutlineAtomicHelper(LC, Ordering, MemSize); in getOutlineAtomicLibcall() 744 return getOutlineAtomicHelper(LC, Ordering, MemSize); in getOutlineAtomicLibcall() 748 return getOutlineAtomicHelper(LC, Ordering, MemSize); in getOutlineAtomicLibcall() 1338 unsigned MemSize = MMO.getSizeInBits(); in narrowScalar() local 1340 if (MemSize == NarrowSize) { in narrowScalar() 1342 } else if (MemSize < NarrowSize) { in narrowScalar() [all …]
|
/freebsd/contrib/llvm-project/llvm/include/llvm/CodeGen/GlobalISel/ |
H A D | CallLowering.h | 254 virtual Register getStackAddress(uint64_t MemSize, int64_t Offset, 310 uint64_t MemSize, CCValAssign &VA) const;
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/ |
H A D | TargetLoweringBase.cpp | 525 uint64_t MemSize) { in getOutlineAtomicHelper() argument 527 switch (MemSize) { in getOutlineAtomicHelper() 572 uint64_t MemSize = VT.getScalarSizeInBits() / 8; in getOUTLINE_ATOMIC() local 581 return getOutlineAtomicHelper(LC, Order, MemSize); in getOUTLINE_ATOMIC() 585 return getOutlineAtomicHelper(LC, Order, MemSize); in getOUTLINE_ATOMIC() 589 return getOutlineAtomicHelper(LC, Order, MemSize); in getOUTLINE_ATOMIC() 593 return getOutlineAtomicHelper(LC, Order, MemSize); in getOUTLINE_ATOMIC() 597 return getOutlineAtomicHelper(LC, Order, MemSize); in getOUTLINE_ATOMIC() 601 return getOutlineAtomicHelper(LC, Order, MemSize); in getOUTLINE_ATOMIC()
|
H A D | TargetInstrInfo.cpp | 650 int64_t MemSize = 0; in foldMemoryOperand() local 655 MemSize = MFI.getObjectSize(FI); in foldMemoryOperand() 666 MemSize = std::max(MemSize, OpSize); in foldMemoryOperand() 670 assert(MemSize && "Did not expect a zero-sized stack slot"); in foldMemoryOperand() 700 Flags, MemSize, MFI.getObjectAlign(FI)); in foldMemoryOperand()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/RISCV/GISel/ |
H A D | RISCVCallLowering.cpp | 67 Register getStackAddress(uint64_t MemSize, int64_t Offset, in getStackAddress() 208 Register getStackAddress(uint64_t MemSize, int64_t Offset, in getStackAddress() 213 int FI = MFI.CreateFixedObject(MemSize, Offset, /*Immutable=*/true); in getStackAddress()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPULegalizerInfo.cpp | 343 if (IsLoad && MemSize < Size) in isLoadStoreSizeLegal() 344 MemSize = std::max(MemSize, Align); in isLoadStoreSizeLegal() 356 switch (MemSize) { in isLoadStoreSizeLegal() 375 assert(RegSize >= MemSize); in isLoadStoreSizeLegal() 377 if (AlignBits < MemSize) { in isLoadStoreSizeLegal() 1470 if (DstSize > MemSize) in AMDGPULegalizerInfo() 1476 if (MemSize > MaxSize) in AMDGPULegalizerInfo() 1502 if (MemSize > MaxSize) { in AMDGPULegalizerInfo() 5616 const int MemSize = MMO->getSize(); in legalizeBufferStore() local 5655 switch (MemSize) { in legalizeBufferStore() [all …]
|
H A D | AMDGPURegisterBankInfo.cpp | 452 const unsigned MemSize = 8 * MMO->getSize(); in isScalarLoadLegal() local 457 ((MemSize == 16 && MMO->getAlign() >= Align(2)) || in isScalarLoadLegal() 458 (MemSize == 8 && MMO->getAlign() >= Align(1))))) && in isScalarLoadLegal() 1073 const unsigned MemSize = 8 * MMO->getSize(); in applyMappingLoad() local 1083 ((MemSize == 8 && MMO->getAlign() >= Align(1)) || in applyMappingLoad() 1084 (MemSize == 16 && MMO->getAlign() >= Align(2))) && in applyMappingLoad() 1100 B.buildSExtInReg(MI.getOperand(0), WideLoad, MemSize); in applyMappingLoad() 1104 B.buildZExtInReg(MI.getOperand(0), WideLoad, MemSize); in applyMappingLoad() 1372 const unsigned MemSize = (Ty.getSizeInBits() + 7) / 8; in applyMappingSBufferLoad() local 1378 MemSize, MemAlign); in applyMappingSBufferLoad() [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/ |
H A D | AArch64LoadStoreOptimizer.cpp | 993 int MemSize = TII->getMemScale(*Paired); in mergePairedInsns() local 999 PairedOffset /= MemSize; in mergePairedInsns() 1001 PairedOffset *= MemSize; in mergePairedInsns() 1745 int MemSize = TII->getMemScale(MI); in findMatchingInsn() local 1749 if (MIOffset % MemSize) { in findMatchingInsn() 1755 MIOffset /= MemSize; in findMatchingInsn() 1757 MIOffset *= MemSize; in findMatchingInsn()
|
/freebsd/contrib/llvm-project/llvm/lib/ObjCopy/ELF/ |
H A D | ELFObject.cpp | 47 Phdr.p_memsz = Seg.MemSize; in writePhdr() 1214 Seg.VAddr + Seg.MemSize >= Sec.Addr + SecSize; in sectionWithinSegment() 1444 Seg.MemSize = Phdr.p_memsz; in readProgramHeaders() 1469 PrHdr.FileSize = PrHdr.MemSize = Ehdr.e_phentsize * Ehdr.e_phnum; in readProgramHeaders() 2438 ElfHdr.FileSize = ElfHdr.MemSize = sizeof(Elf_Ehdr); in initEhdrSegment()
|
H A D | ELFObject.h | 462 uint64_t MemSize = 0; variable
|
/freebsd/contrib/llvm-project/llvm/lib/Target/Hexagon/ |
H A D | HexagonFrameLowering.cpp | 2465 unsigned MemSize = HII.getMemAccessSize(MI); in optimizeSpillSlots() local 2469 CopyOpc = (MemSize == 1) ? Hexagon::A2_sxtb : Hexagon::A2_sxth; in optimizeSpillSlots() 2471 CopyOpc = (MemSize == 1) ? Hexagon::A2_zxtb : Hexagon::A2_zxth; in optimizeSpillSlots()
|
H A D | HexagonISelLoweringHVX.cpp | 2984 uint64_t MemSize = (MemOpc == ISD::MLOAD || MemOpc == ISD::MSTORE) in SplitHvxMemOp() local 2987 MOp0 = MF.getMachineMemOperand(MMO, 0, MemSize); in SplitHvxMemOp() 2988 MOp1 = MF.getMachineMemOperand(MMO, HwLen, MemSize); in SplitHvxMemOp()
|
/freebsd/contrib/llvm-project/llvm/include/llvm/ObjectYAML/ |
H A D | ELFYAML.h | 701 std::optional<llvm::yaml::Hex64> MemSize; member
|
/freebsd/contrib/llvm-project/llvm/lib/ObjectYAML/ |
H A D | ELFEmitter.cpp | 1207 PHeader.p_memsz = YamlPhdr.MemSize ? uint64_t(*YamlPhdr.MemSize) in setProgramHeaderLayout()
|
H A D | ELFYAML.cpp | 1137 IO.mapOptional("MemSize", Phdr.MemSize); in mapping()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/GISel/ |
H A D | AArch64InstructionSelector.cpp | 5707 unsigned MemSize = Ld.getMMO().getMemoryType().getSizeInBytes(); in selectIndexedLoad() local 5709 if (MemSize < MRI.getType(Dst).getSizeInBytes()) in selectIndexedLoad() 5721 Opc = FPROpcodes[Log2_32(MemSize)]; in selectIndexedLoad() 5723 Opc = GPROpcodes[Log2_32(MemSize)]; in selectIndexedLoad() 5732 Opc = FPROpcodes[Log2_32(MemSize)]; in selectIndexedLoad() 5734 Opc = GPROpcodes[Log2_32(MemSize)]; in selectIndexedLoad()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/X86/ |
H A D | X86FastISel.cpp | 3618 unsigned MemSize = ResVT.getSizeInBits()/8; in fastLowerCall() local 3619 int FI = MFI.CreateStackObject(MemSize, Align(MemSize), false); in fastLowerCall()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/ARM/ |
H A D | ARMISelDAGToDAG.cpp | 1071 unsigned MemSize = MemN->getMemoryVT().getSizeInBits() / 8; in SelectAddrMode6() local 1072 if (MMOAlign.value() >= MemSize && MemSize > 1) in SelectAddrMode6() 1073 Alignment = MemSize; in SelectAddrMode6()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/X86/AsmParser/ |
H A D | X86AsmParser.cpp | 4134 const char *MemSize = Base[0] != 'f' ? "\x08\x10\x20\x40" : "\x20\x40\x50\0"; in MatchAndEmitATTInstruction() local 4164 MemOp->Mem.Size = MemSize[I]; in MatchAndEmitATTInstruction()
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/SelectionDAG/ |
H A D | SelectionDAGBuilder.cpp | 11040 uint64_t MemSize = DL.getTypeAllocSize(ArgMemTy); in LowerArguments() local 11052 Flags.setByRefSize(MemSize); in LowerArguments() 11054 Flags.setByValSize(MemSize); in LowerArguments()
|