1 //===-- SystemZFrameLowering.cpp - Frame lowering for SystemZ -------------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 
9 #include "SystemZFrameLowering.h"
10 #include "SystemZCallingConv.h"
11 #include "SystemZInstrBuilder.h"
12 #include "SystemZInstrInfo.h"
13 #include "SystemZMachineFunctionInfo.h"
14 #include "SystemZRegisterInfo.h"
15 #include "SystemZSubtarget.h"
16 #include "llvm/CodeGen/LivePhysRegs.h"
17 #include "llvm/CodeGen/MachineModuleInfo.h"
18 #include "llvm/CodeGen/MachineRegisterInfo.h"
19 #include "llvm/CodeGen/RegisterScavenging.h"
20 #include "llvm/IR/Function.h"
21 #include "llvm/Target/TargetMachine.h"
22 
23 using namespace llvm;
24 
25 namespace {
26 // The ABI-defined register save slots, relative to the CFA (i.e.
27 // incoming stack pointer + SystemZMC::ELFCallFrameSize).
28 static const TargetFrameLowering::SpillSlot ELFSpillOffsetTable[] = {
29   { SystemZ::R2D,  0x10 },
30   { SystemZ::R3D,  0x18 },
31   { SystemZ::R4D,  0x20 },
32   { SystemZ::R5D,  0x28 },
33   { SystemZ::R6D,  0x30 },
34   { SystemZ::R7D,  0x38 },
35   { SystemZ::R8D,  0x40 },
36   { SystemZ::R9D,  0x48 },
37   { SystemZ::R10D, 0x50 },
38   { SystemZ::R11D, 0x58 },
39   { SystemZ::R12D, 0x60 },
40   { SystemZ::R13D, 0x68 },
41   { SystemZ::R14D, 0x70 },
42   { SystemZ::R15D, 0x78 },
43   { SystemZ::F0D,  0x80 },
44   { SystemZ::F2D,  0x88 },
45   { SystemZ::F4D,  0x90 },
46   { SystemZ::F6D,  0x98 }
47 };
48 
49 static const TargetFrameLowering::SpillSlot XPLINKSpillOffsetTable[] = {
50     {SystemZ::R4D, 0x00},  {SystemZ::R5D, 0x08},  {SystemZ::R6D, 0x10},
51     {SystemZ::R7D, 0x18},  {SystemZ::R8D, 0x20},  {SystemZ::R9D, 0x28},
52     {SystemZ::R10D, 0x30}, {SystemZ::R11D, 0x38}, {SystemZ::R12D, 0x40},
53     {SystemZ::R13D, 0x48}, {SystemZ::R14D, 0x50}, {SystemZ::R15D, 0x58}};
54 } // end anonymous namespace
55 
56 SystemZFrameLowering::SystemZFrameLowering(StackDirection D, Align StackAl,
57                                            int LAO, Align TransAl,
58                                            bool StackReal)
59     : TargetFrameLowering(D, StackAl, LAO, TransAl, StackReal) {}
60 
61 std::unique_ptr<SystemZFrameLowering>
62 SystemZFrameLowering::create(const SystemZSubtarget &STI) {
63   if (STI.isTargetXPLINK64())
64     return std::make_unique<SystemZXPLINKFrameLowering>();
65   return std::make_unique<SystemZELFFrameLowering>();
66 }
67 
68 MachineBasicBlock::iterator SystemZFrameLowering::eliminateCallFramePseudoInstr(
69     MachineFunction &MF, MachineBasicBlock &MBB,
70     MachineBasicBlock::iterator MI) const {
71   switch (MI->getOpcode()) {
72   case SystemZ::ADJCALLSTACKDOWN:
73   case SystemZ::ADJCALLSTACKUP:
74     assert(hasReservedCallFrame(MF) &&
75            "ADJSTACKDOWN and ADJSTACKUP should be no-ops");
76     return MBB.erase(MI);
77     break;
78 
79   default:
80     llvm_unreachable("Unexpected call frame instruction");
81   }
82 }
83 
84 namespace {
85 struct SZFrameSortingObj {
86   bool IsValid = false;     // True if we care about this Object.
87   uint32_t ObjectIndex = 0; // Index of Object into MFI list.
88   uint64_t ObjectSize = 0;  // Size of Object in bytes.
89   uint32_t D12Count = 0;    // 12-bit displacement only.
90   uint32_t DPairCount = 0;  // 12 or 20 bit displacement.
91 };
92 typedef std::vector<SZFrameSortingObj> SZFrameObjVec;
93 } // namespace
94 
95 // TODO: Move to base class.
96 void SystemZELFFrameLowering::orderFrameObjects(
97     const MachineFunction &MF, SmallVectorImpl<int> &ObjectsToAllocate) const {
98   const MachineFrameInfo &MFI = MF.getFrameInfo();
99   auto *TII = MF.getSubtarget<SystemZSubtarget>().getInstrInfo();
100 
101   // Make a vector of sorting objects to track all MFI objects and mark those
102   // to be sorted as valid.
103   if (ObjectsToAllocate.size() <= 1)
104     return;
105   SZFrameObjVec SortingObjects(MFI.getObjectIndexEnd());
106   for (auto &Obj : ObjectsToAllocate) {
107     SortingObjects[Obj].IsValid = true;
108     SortingObjects[Obj].ObjectIndex = Obj;
109     SortingObjects[Obj].ObjectSize = MFI.getObjectSize(Obj);
110   }
111 
112   // Examine uses for each object and record short (12-bit) and "pair"
113   // displacement types.
114   for (auto &MBB : MF)
115     for (auto &MI : MBB) {
116       if (MI.isDebugInstr())
117         continue;
118       for (unsigned I = 0, E = MI.getNumOperands(); I != E; ++I) {
119         const MachineOperand &MO = MI.getOperand(I);
120         if (!MO.isFI())
121           continue;
122         int Index = MO.getIndex();
123         if (Index >= 0 && Index < MFI.getObjectIndexEnd() &&
124             SortingObjects[Index].IsValid) {
125           if (TII->hasDisplacementPairInsn(MI.getOpcode()))
126             SortingObjects[Index].DPairCount++;
127           else if (!(MI.getDesc().TSFlags & SystemZII::Has20BitOffset))
128             SortingObjects[Index].D12Count++;
129         }
130       }
131     }
132 
133   // Sort all objects for short/paired displacements, which should be
134   // sufficient as it seems like all frame objects typically are within the
135   // long displacement range.  Sorting works by computing the "density" as
136   // Count / ObjectSize. The comparisons of two such fractions are refactored
137   // by multiplying both sides with A.ObjectSize * B.ObjectSize, in order to
138   // eliminate the (fp) divisions.  A higher density object needs to go after
139   // in the list in order for it to end up lower on the stack.
140   auto CmpD12 = [](const SZFrameSortingObj &A, const SZFrameSortingObj &B) {
141     // Put all invalid and variable sized objects at the end.
142     if (!A.IsValid || !B.IsValid)
143       return A.IsValid;
144     if (!A.ObjectSize || !B.ObjectSize)
145       return A.ObjectSize > 0;
146     uint64_t ADensityCmp = A.D12Count * B.ObjectSize;
147     uint64_t BDensityCmp = B.D12Count * A.ObjectSize;
148     if (ADensityCmp != BDensityCmp)
149       return ADensityCmp < BDensityCmp;
150     return A.DPairCount * B.ObjectSize < B.DPairCount * A.ObjectSize;
151   };
152   std::stable_sort(SortingObjects.begin(), SortingObjects.end(), CmpD12);
153 
154   // Now modify the original list to represent the final order that
155   // we want.
156   unsigned Idx = 0;
157   for (auto &Obj : SortingObjects) {
158     // All invalid items are sorted at the end, so it's safe to stop.
159     if (!Obj.IsValid)
160       break;
161     ObjectsToAllocate[Idx++] = Obj.ObjectIndex;
162   }
163 }
164 
165 bool SystemZFrameLowering::hasReservedCallFrame(
166     const MachineFunction &MF) const {
167   // The ELF ABI requires us to allocate 160 bytes of stack space for the
168   // callee, with any outgoing stack arguments being placed above that. It
169   // seems better to make that area a permanent feature of the frame even if
170   // we're using a frame pointer. Similarly, 64-bit XPLINK requires 96 bytes
171   // of stack space for the register save area.
172   return true;
173 }
174 
175 bool SystemZELFFrameLowering::assignCalleeSavedSpillSlots(
176     MachineFunction &MF, const TargetRegisterInfo *TRI,
177     std::vector<CalleeSavedInfo> &CSI) const {
178   SystemZMachineFunctionInfo *ZFI = MF.getInfo<SystemZMachineFunctionInfo>();
179   MachineFrameInfo &MFFrame = MF.getFrameInfo();
180   bool IsVarArg = MF.getFunction().isVarArg();
181   if (CSI.empty())
182     return true; // Early exit if no callee saved registers are modified!
183 
184   unsigned LowGPR = 0;
185   unsigned HighGPR = SystemZ::R15D;
186   int StartSPOffset = SystemZMC::ELFCallFrameSize;
187   for (auto &CS : CSI) {
188     Register Reg = CS.getReg();
189     int Offset = getRegSpillOffset(MF, Reg);
190     if (Offset) {
191       if (SystemZ::GR64BitRegClass.contains(Reg) && StartSPOffset > Offset) {
192         LowGPR = Reg;
193         StartSPOffset = Offset;
194       }
195       Offset -= SystemZMC::ELFCallFrameSize;
196       int FrameIdx = MFFrame.CreateFixedSpillStackObject(8, Offset);
197       CS.setFrameIdx(FrameIdx);
198     } else
199       CS.setFrameIdx(INT32_MAX);
200   }
201 
202   // Save the range of call-saved registers, for use by the
203   // prologue/epilogue inserters.
204   ZFI->setRestoreGPRRegs(LowGPR, HighGPR, StartSPOffset);
205   if (IsVarArg) {
206     // Also save the GPR varargs, if any.  R6D is call-saved, so would
207     // already be included, but we also need to handle the call-clobbered
208     // argument registers.
209     Register FirstGPR = ZFI->getVarArgsFirstGPR();
210     if (FirstGPR < SystemZ::ELFNumArgGPRs) {
211       unsigned Reg = SystemZ::ELFArgGPRs[FirstGPR];
212       int Offset = getRegSpillOffset(MF, Reg);
213       if (StartSPOffset > Offset) {
214         LowGPR = Reg; StartSPOffset = Offset;
215       }
216     }
217   }
218   ZFI->setSpillGPRRegs(LowGPR, HighGPR, StartSPOffset);
219 
220   // Create fixed stack objects for the remaining registers.
221   int CurrOffset = -SystemZMC::ELFCallFrameSize;
222   if (usePackedStack(MF))
223     CurrOffset += StartSPOffset;
224 
225   for (auto &CS : CSI) {
226     if (CS.getFrameIdx() != INT32_MAX)
227       continue;
228     Register Reg = CS.getReg();
229     const TargetRegisterClass *RC = TRI->getMinimalPhysRegClass(Reg);
230     unsigned Size = TRI->getSpillSize(*RC);
231     CurrOffset -= Size;
232     assert(CurrOffset % 8 == 0 &&
233            "8-byte alignment required for for all register save slots");
234     int FrameIdx = MFFrame.CreateFixedSpillStackObject(Size, CurrOffset);
235     CS.setFrameIdx(FrameIdx);
236   }
237 
238   return true;
239 }
240 
241 void SystemZELFFrameLowering::determineCalleeSaves(MachineFunction &MF,
242                                                    BitVector &SavedRegs,
243                                                    RegScavenger *RS) const {
244   TargetFrameLowering::determineCalleeSaves(MF, SavedRegs, RS);
245 
246   MachineFrameInfo &MFFrame = MF.getFrameInfo();
247   const TargetRegisterInfo *TRI = MF.getSubtarget().getRegisterInfo();
248   bool HasFP = hasFP(MF);
249   SystemZMachineFunctionInfo *MFI = MF.getInfo<SystemZMachineFunctionInfo>();
250   bool IsVarArg = MF.getFunction().isVarArg();
251 
252   // va_start stores incoming FPR varargs in the normal way, but delegates
253   // the saving of incoming GPR varargs to spillCalleeSavedRegisters().
254   // Record these pending uses, which typically include the call-saved
255   // argument register R6D.
256   if (IsVarArg)
257     for (unsigned I = MFI->getVarArgsFirstGPR(); I < SystemZ::ELFNumArgGPRs; ++I)
258       SavedRegs.set(SystemZ::ELFArgGPRs[I]);
259 
260   // If there are any landing pads, entering them will modify r6/r7.
261   if (!MF.getLandingPads().empty()) {
262     SavedRegs.set(SystemZ::R6D);
263     SavedRegs.set(SystemZ::R7D);
264   }
265 
266   // If the function requires a frame pointer, record that the hard
267   // frame pointer will be clobbered.
268   if (HasFP)
269     SavedRegs.set(SystemZ::R11D);
270 
271   // If the function calls other functions, record that the return
272   // address register will be clobbered.
273   if (MFFrame.hasCalls())
274     SavedRegs.set(SystemZ::R14D);
275 
276   // If we are saving GPRs other than the stack pointer, we might as well
277   // save and restore the stack pointer at the same time, via STMG and LMG.
278   // This allows the deallocation to be done by the LMG, rather than needing
279   // a separate %r15 addition.
280   const MCPhysReg *CSRegs = TRI->getCalleeSavedRegs(&MF);
281   for (unsigned I = 0; CSRegs[I]; ++I) {
282     unsigned Reg = CSRegs[I];
283     if (SystemZ::GR64BitRegClass.contains(Reg) && SavedRegs.test(Reg)) {
284       SavedRegs.set(SystemZ::R15D);
285       break;
286     }
287   }
288 }
289 
290 SystemZELFFrameLowering::SystemZELFFrameLowering()
291     : SystemZFrameLowering(TargetFrameLowering::StackGrowsDown, Align(8), 0,
292                            Align(8), /* StackRealignable */ false),
293       RegSpillOffsets(0) {
294 
295   // Due to the SystemZ ABI, the DWARF CFA (Canonical Frame Address) is not
296   // equal to the incoming stack pointer, but to incoming stack pointer plus
297   // 160.  Instead of using a Local Area Offset, the Register save area will
298   // be occupied by fixed frame objects, and all offsets are actually
299   // relative to CFA.
300 
301   // Create a mapping from register number to save slot offset.
302   // These offsets are relative to the start of the register save area.
303   RegSpillOffsets.grow(SystemZ::NUM_TARGET_REGS);
304   for (unsigned I = 0, E = array_lengthof(ELFSpillOffsetTable); I != E; ++I)
305     RegSpillOffsets[ELFSpillOffsetTable[I].Reg] = ELFSpillOffsetTable[I].Offset;
306 }
307 
308 // Add GPR64 to the save instruction being built by MIB, which is in basic
309 // block MBB.  IsImplicit says whether this is an explicit operand to the
310 // instruction, or an implicit one that comes between the explicit start
311 // and end registers.
312 static void addSavedGPR(MachineBasicBlock &MBB, MachineInstrBuilder &MIB,
313                         unsigned GPR64, bool IsImplicit) {
314   const TargetRegisterInfo *RI =
315       MBB.getParent()->getSubtarget().getRegisterInfo();
316   Register GPR32 = RI->getSubReg(GPR64, SystemZ::subreg_l32);
317   bool IsLive = MBB.isLiveIn(GPR64) || MBB.isLiveIn(GPR32);
318   if (!IsLive || !IsImplicit) {
319     MIB.addReg(GPR64, getImplRegState(IsImplicit) | getKillRegState(!IsLive));
320     if (!IsLive)
321       MBB.addLiveIn(GPR64);
322   }
323 }
324 
325 bool SystemZELFFrameLowering::spillCalleeSavedRegisters(
326     MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI,
327     ArrayRef<CalleeSavedInfo> CSI, const TargetRegisterInfo *TRI) const {
328   if (CSI.empty())
329     return false;
330 
331   MachineFunction &MF = *MBB.getParent();
332   const TargetInstrInfo *TII = MF.getSubtarget().getInstrInfo();
333   SystemZMachineFunctionInfo *ZFI = MF.getInfo<SystemZMachineFunctionInfo>();
334   bool IsVarArg = MF.getFunction().isVarArg();
335   DebugLoc DL;
336 
337   // Save GPRs
338   SystemZ::GPRRegs SpillGPRs = ZFI->getSpillGPRRegs();
339   if (SpillGPRs.LowGPR) {
340     assert(SpillGPRs.LowGPR != SpillGPRs.HighGPR &&
341            "Should be saving %r15 and something else");
342 
343     // Build an STMG instruction.
344     MachineInstrBuilder MIB = BuildMI(MBB, MBBI, DL, TII->get(SystemZ::STMG));
345 
346     // Add the explicit register operands.
347     addSavedGPR(MBB, MIB, SpillGPRs.LowGPR, false);
348     addSavedGPR(MBB, MIB, SpillGPRs.HighGPR, false);
349 
350     // Add the address.
351     MIB.addReg(SystemZ::R15D).addImm(SpillGPRs.GPROffset);
352 
353     // Make sure all call-saved GPRs are included as operands and are
354     // marked as live on entry.
355     for (const CalleeSavedInfo &I : CSI) {
356       Register Reg = I.getReg();
357       if (SystemZ::GR64BitRegClass.contains(Reg))
358         addSavedGPR(MBB, MIB, Reg, true);
359     }
360 
361     // ...likewise GPR varargs.
362     if (IsVarArg)
363       for (unsigned I = ZFI->getVarArgsFirstGPR(); I < SystemZ::ELFNumArgGPRs; ++I)
364         addSavedGPR(MBB, MIB, SystemZ::ELFArgGPRs[I], true);
365   }
366 
367   // Save FPRs/VRs in the normal TargetInstrInfo way.
368   for (const CalleeSavedInfo &I : CSI) {
369     Register Reg = I.getReg();
370     if (SystemZ::FP64BitRegClass.contains(Reg)) {
371       MBB.addLiveIn(Reg);
372       TII->storeRegToStackSlot(MBB, MBBI, Reg, true, I.getFrameIdx(),
373                                &SystemZ::FP64BitRegClass, TRI);
374     }
375     if (SystemZ::VR128BitRegClass.contains(Reg)) {
376       MBB.addLiveIn(Reg);
377       TII->storeRegToStackSlot(MBB, MBBI, Reg, true, I.getFrameIdx(),
378                                &SystemZ::VR128BitRegClass, TRI);
379     }
380   }
381 
382   return true;
383 }
384 
385 bool SystemZELFFrameLowering::restoreCalleeSavedRegisters(
386     MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI,
387     MutableArrayRef<CalleeSavedInfo> CSI, const TargetRegisterInfo *TRI) const {
388   if (CSI.empty())
389     return false;
390 
391   MachineFunction &MF = *MBB.getParent();
392   const TargetInstrInfo *TII = MF.getSubtarget().getInstrInfo();
393   SystemZMachineFunctionInfo *ZFI = MF.getInfo<SystemZMachineFunctionInfo>();
394   bool HasFP = hasFP(MF);
395   DebugLoc DL = MBBI != MBB.end() ? MBBI->getDebugLoc() : DebugLoc();
396 
397   // Restore FPRs/VRs in the normal TargetInstrInfo way.
398   for (const CalleeSavedInfo &I : CSI) {
399     Register Reg = I.getReg();
400     if (SystemZ::FP64BitRegClass.contains(Reg))
401       TII->loadRegFromStackSlot(MBB, MBBI, Reg, I.getFrameIdx(),
402                                 &SystemZ::FP64BitRegClass, TRI);
403     if (SystemZ::VR128BitRegClass.contains(Reg))
404       TII->loadRegFromStackSlot(MBB, MBBI, Reg, I.getFrameIdx(),
405                                 &SystemZ::VR128BitRegClass, TRI);
406   }
407 
408   // Restore call-saved GPRs (but not call-clobbered varargs, which at
409   // this point might hold return values).
410   SystemZ::GPRRegs RestoreGPRs = ZFI->getRestoreGPRRegs();
411   if (RestoreGPRs.LowGPR) {
412     // If we saved any of %r2-%r5 as varargs, we should also be saving
413     // and restoring %r6.  If we're saving %r6 or above, we should be
414     // restoring it too.
415     assert(RestoreGPRs.LowGPR != RestoreGPRs.HighGPR &&
416            "Should be loading %r15 and something else");
417 
418     // Build an LMG instruction.
419     MachineInstrBuilder MIB = BuildMI(MBB, MBBI, DL, TII->get(SystemZ::LMG));
420 
421     // Add the explicit register operands.
422     MIB.addReg(RestoreGPRs.LowGPR, RegState::Define);
423     MIB.addReg(RestoreGPRs.HighGPR, RegState::Define);
424 
425     // Add the address.
426     MIB.addReg(HasFP ? SystemZ::R11D : SystemZ::R15D);
427     MIB.addImm(RestoreGPRs.GPROffset);
428 
429     // Do a second scan adding regs as being defined by instruction
430     for (const CalleeSavedInfo &I : CSI) {
431       Register Reg = I.getReg();
432       if (Reg != RestoreGPRs.LowGPR && Reg != RestoreGPRs.HighGPR &&
433           SystemZ::GR64BitRegClass.contains(Reg))
434         MIB.addReg(Reg, RegState::ImplicitDefine);
435     }
436   }
437 
438   return true;
439 }
440 
441 void SystemZELFFrameLowering::processFunctionBeforeFrameFinalized(
442     MachineFunction &MF, RegScavenger *RS) const {
443   MachineFrameInfo &MFFrame = MF.getFrameInfo();
444   SystemZMachineFunctionInfo *ZFI = MF.getInfo<SystemZMachineFunctionInfo>();
445   MachineRegisterInfo *MRI = &MF.getRegInfo();
446   bool BackChain = MF.getFunction().hasFnAttribute("backchain");
447 
448   if (!usePackedStack(MF) || BackChain)
449     // Create the incoming register save area.
450     getOrCreateFramePointerSaveIndex(MF);
451 
452   // Get the size of our stack frame to be allocated ...
453   uint64_t StackSize = (MFFrame.estimateStackSize(MF) +
454                         SystemZMC::ELFCallFrameSize);
455   // ... and the maximum offset we may need to reach into the
456   // caller's frame to access the save area or stack arguments.
457   int64_t MaxArgOffset = 0;
458   for (int I = MFFrame.getObjectIndexBegin(); I != 0; ++I)
459     if (MFFrame.getObjectOffset(I) >= 0) {
460       int64_t ArgOffset = MFFrame.getObjectOffset(I) +
461                           MFFrame.getObjectSize(I);
462       MaxArgOffset = std::max(MaxArgOffset, ArgOffset);
463     }
464 
465   uint64_t MaxReach = StackSize + MaxArgOffset;
466   if (!isUInt<12>(MaxReach)) {
467     // We may need register scavenging slots if some parts of the frame
468     // are outside the reach of an unsigned 12-bit displacement.
469     // Create 2 for the case where both addresses in an MVC are
470     // out of range.
471     RS->addScavengingFrameIndex(MFFrame.CreateStackObject(8, Align(8), false));
472     RS->addScavengingFrameIndex(MFFrame.CreateStackObject(8, Align(8), false));
473   }
474 
475   // If R6 is used as an argument register it is still callee saved. If it in
476   // this case is not clobbered (and restored) it should never be marked as
477   // killed.
478   if (MF.front().isLiveIn(SystemZ::R6D) &&
479       ZFI->getRestoreGPRRegs().LowGPR != SystemZ::R6D)
480     for (auto &MO : MRI->use_nodbg_operands(SystemZ::R6D))
481       MO.setIsKill(false);
482 }
483 
484 // Emit instructions before MBBI (in MBB) to add NumBytes to Reg.
485 static void emitIncrement(MachineBasicBlock &MBB,
486                           MachineBasicBlock::iterator &MBBI, const DebugLoc &DL,
487                           Register Reg, int64_t NumBytes,
488                           const TargetInstrInfo *TII) {
489   while (NumBytes) {
490     unsigned Opcode;
491     int64_t ThisVal = NumBytes;
492     if (isInt<16>(NumBytes))
493       Opcode = SystemZ::AGHI;
494     else {
495       Opcode = SystemZ::AGFI;
496       // Make sure we maintain 8-byte stack alignment.
497       int64_t MinVal = -uint64_t(1) << 31;
498       int64_t MaxVal = (int64_t(1) << 31) - 8;
499       if (ThisVal < MinVal)
500         ThisVal = MinVal;
501       else if (ThisVal > MaxVal)
502         ThisVal = MaxVal;
503     }
504     MachineInstr *MI = BuildMI(MBB, MBBI, DL, TII->get(Opcode), Reg)
505       .addReg(Reg).addImm(ThisVal);
506     // The CC implicit def is dead.
507     MI->getOperand(3).setIsDead();
508     NumBytes -= ThisVal;
509   }
510 }
511 
512 // Add CFI for the new CFA offset.
513 static void buildCFAOffs(MachineBasicBlock &MBB,
514                          MachineBasicBlock::iterator MBBI,
515                          const DebugLoc &DL, int Offset,
516                          const SystemZInstrInfo *ZII) {
517   unsigned CFIIndex = MBB.getParent()->addFrameInst(
518     MCCFIInstruction::cfiDefCfaOffset(nullptr, -Offset));
519   BuildMI(MBB, MBBI, DL, ZII->get(TargetOpcode::CFI_INSTRUCTION))
520     .addCFIIndex(CFIIndex);
521 }
522 
523 // Add CFI for the new frame location.
524 static void buildDefCFAReg(MachineBasicBlock &MBB,
525                            MachineBasicBlock::iterator MBBI,
526                            const DebugLoc &DL, unsigned Reg,
527                            const SystemZInstrInfo *ZII) {
528   MachineFunction &MF = *MBB.getParent();
529   MachineModuleInfo &MMI = MF.getMMI();
530   const MCRegisterInfo *MRI = MMI.getContext().getRegisterInfo();
531   unsigned RegNum = MRI->getDwarfRegNum(Reg, true);
532   unsigned CFIIndex = MF.addFrameInst(
533                         MCCFIInstruction::createDefCfaRegister(nullptr, RegNum));
534   BuildMI(MBB, MBBI, DL, ZII->get(TargetOpcode::CFI_INSTRUCTION))
535     .addCFIIndex(CFIIndex);
536 }
537 
538 void SystemZELFFrameLowering::emitPrologue(MachineFunction &MF,
539                                            MachineBasicBlock &MBB) const {
540   assert(&MF.front() == &MBB && "Shrink-wrapping not yet supported");
541   const SystemZSubtarget &STI = MF.getSubtarget<SystemZSubtarget>();
542   const SystemZTargetLowering &TLI = *STI.getTargetLowering();
543   MachineFrameInfo &MFFrame = MF.getFrameInfo();
544   auto *ZII = static_cast<const SystemZInstrInfo *>(STI.getInstrInfo());
545   SystemZMachineFunctionInfo *ZFI = MF.getInfo<SystemZMachineFunctionInfo>();
546   MachineBasicBlock::iterator MBBI = MBB.begin();
547   MachineModuleInfo &MMI = MF.getMMI();
548   const MCRegisterInfo *MRI = MMI.getContext().getRegisterInfo();
549   const std::vector<CalleeSavedInfo> &CSI = MFFrame.getCalleeSavedInfo();
550   bool HasFP = hasFP(MF);
551 
552   // In GHC calling convention C stack space, including the ABI-defined
553   // 160-byte base area, is (de)allocated by GHC itself.  This stack space may
554   // be used by LLVM as spill slots for the tail recursive GHC functions.  Thus
555   // do not allocate stack space here, too.
556   if (MF.getFunction().getCallingConv() == CallingConv::GHC) {
557     if (MFFrame.getStackSize() > 2048 * sizeof(long)) {
558       report_fatal_error(
559           "Pre allocated stack space for GHC function is too small");
560     }
561     if (HasFP) {
562       report_fatal_error(
563           "In GHC calling convention a frame pointer is not supported");
564     }
565     MFFrame.setStackSize(MFFrame.getStackSize() + SystemZMC::ELFCallFrameSize);
566     return;
567   }
568 
569   // Debug location must be unknown since the first debug location is used
570   // to determine the end of the prologue.
571   DebugLoc DL;
572 
573   // The current offset of the stack pointer from the CFA.
574   int64_t SPOffsetFromCFA = -SystemZMC::ELFCFAOffsetFromInitialSP;
575 
576   if (ZFI->getSpillGPRRegs().LowGPR) {
577     // Skip over the GPR saves.
578     if (MBBI != MBB.end() && MBBI->getOpcode() == SystemZ::STMG)
579       ++MBBI;
580     else
581       llvm_unreachable("Couldn't skip over GPR saves");
582 
583     // Add CFI for the GPR saves.
584     for (auto &Save : CSI) {
585       Register Reg = Save.getReg();
586       if (SystemZ::GR64BitRegClass.contains(Reg)) {
587         int FI = Save.getFrameIdx();
588         int64_t Offset = MFFrame.getObjectOffset(FI);
589         unsigned CFIIndex = MF.addFrameInst(MCCFIInstruction::createOffset(
590             nullptr, MRI->getDwarfRegNum(Reg, true), Offset));
591         BuildMI(MBB, MBBI, DL, ZII->get(TargetOpcode::CFI_INSTRUCTION))
592             .addCFIIndex(CFIIndex);
593       }
594     }
595   }
596 
597   uint64_t StackSize = MFFrame.getStackSize();
598   // We need to allocate the ABI-defined 160-byte base area whenever
599   // we allocate stack space for our own use and whenever we call another
600   // function.
601   bool HasStackObject = false;
602   for (unsigned i = 0, e = MFFrame.getObjectIndexEnd(); i != e; ++i)
603     if (!MFFrame.isDeadObjectIndex(i)) {
604       HasStackObject = true;
605       break;
606     }
607   if (HasStackObject || MFFrame.hasCalls())
608     StackSize += SystemZMC::ELFCallFrameSize;
609   // Don't allocate the incoming reg save area.
610   StackSize = StackSize > SystemZMC::ELFCallFrameSize
611                   ? StackSize - SystemZMC::ELFCallFrameSize
612                   : 0;
613   MFFrame.setStackSize(StackSize);
614 
615   if (StackSize) {
616     // Allocate StackSize bytes.
617     int64_t Delta = -int64_t(StackSize);
618     const unsigned ProbeSize = TLI.getStackProbeSize(MF);
619     bool FreeProbe = (ZFI->getSpillGPRRegs().GPROffset &&
620            (ZFI->getSpillGPRRegs().GPROffset + StackSize) < ProbeSize);
621     if (!FreeProbe &&
622         MF.getSubtarget().getTargetLowering()->hasInlineStackProbe(MF)) {
623       // Stack probing may involve looping, but splitting the prologue block
624       // is not possible at this point since it would invalidate the
625       // SaveBlocks / RestoreBlocks sets of PEI in the single block function
626       // case. Build a pseudo to be handled later by inlineStackProbe().
627       BuildMI(MBB, MBBI, DL, ZII->get(SystemZ::PROBED_STACKALLOC))
628         .addImm(StackSize);
629     }
630     else {
631       bool StoreBackchain = MF.getFunction().hasFnAttribute("backchain");
632       // If we need backchain, save current stack pointer.  R1 is free at
633       // this point.
634       if (StoreBackchain)
635         BuildMI(MBB, MBBI, DL, ZII->get(SystemZ::LGR))
636           .addReg(SystemZ::R1D, RegState::Define).addReg(SystemZ::R15D);
637       emitIncrement(MBB, MBBI, DL, SystemZ::R15D, Delta, ZII);
638       buildCFAOffs(MBB, MBBI, DL, SPOffsetFromCFA + Delta, ZII);
639       if (StoreBackchain)
640         BuildMI(MBB, MBBI, DL, ZII->get(SystemZ::STG))
641           .addReg(SystemZ::R1D, RegState::Kill).addReg(SystemZ::R15D)
642           .addImm(getBackchainOffset(MF)).addReg(0);
643     }
644     SPOffsetFromCFA += Delta;
645   }
646 
647   if (HasFP) {
648     // Copy the base of the frame to R11.
649     BuildMI(MBB, MBBI, DL, ZII->get(SystemZ::LGR), SystemZ::R11D)
650       .addReg(SystemZ::R15D);
651 
652     // Add CFI for the new frame location.
653     buildDefCFAReg(MBB, MBBI, DL, SystemZ::R11D, ZII);
654 
655     // Mark the FramePtr as live at the beginning of every block except
656     // the entry block.  (We'll have marked R11 as live on entry when
657     // saving the GPRs.)
658     for (MachineBasicBlock &MBBJ : llvm::drop_begin(MF))
659       MBBJ.addLiveIn(SystemZ::R11D);
660   }
661 
662   // Skip over the FPR/VR saves.
663   SmallVector<unsigned, 8> CFIIndexes;
664   for (auto &Save : CSI) {
665     Register Reg = Save.getReg();
666     if (SystemZ::FP64BitRegClass.contains(Reg)) {
667       if (MBBI != MBB.end() &&
668           (MBBI->getOpcode() == SystemZ::STD ||
669            MBBI->getOpcode() == SystemZ::STDY))
670         ++MBBI;
671       else
672         llvm_unreachable("Couldn't skip over FPR save");
673     } else if (SystemZ::VR128BitRegClass.contains(Reg)) {
674       if (MBBI != MBB.end() &&
675           MBBI->getOpcode() == SystemZ::VST)
676         ++MBBI;
677       else
678         llvm_unreachable("Couldn't skip over VR save");
679     } else
680       continue;
681 
682     // Add CFI for the this save.
683     unsigned DwarfReg = MRI->getDwarfRegNum(Reg, true);
684     Register IgnoredFrameReg;
685     int64_t Offset =
686         getFrameIndexReference(MF, Save.getFrameIdx(), IgnoredFrameReg)
687             .getFixed();
688 
689     unsigned CFIIndex = MF.addFrameInst(MCCFIInstruction::createOffset(
690           nullptr, DwarfReg, SPOffsetFromCFA + Offset));
691     CFIIndexes.push_back(CFIIndex);
692   }
693   // Complete the CFI for the FPR/VR saves, modelling them as taking effect
694   // after the last save.
695   for (auto CFIIndex : CFIIndexes) {
696     BuildMI(MBB, MBBI, DL, ZII->get(TargetOpcode::CFI_INSTRUCTION))
697         .addCFIIndex(CFIIndex);
698   }
699 }
700 
701 void SystemZELFFrameLowering::emitEpilogue(MachineFunction &MF,
702                                            MachineBasicBlock &MBB) const {
703   MachineBasicBlock::iterator MBBI = MBB.getLastNonDebugInstr();
704   auto *ZII =
705       static_cast<const SystemZInstrInfo *>(MF.getSubtarget().getInstrInfo());
706   SystemZMachineFunctionInfo *ZFI = MF.getInfo<SystemZMachineFunctionInfo>();
707   MachineFrameInfo &MFFrame = MF.getFrameInfo();
708 
709   // See SystemZELFFrameLowering::emitPrologue
710   if (MF.getFunction().getCallingConv() == CallingConv::GHC)
711     return;
712 
713   // Skip the return instruction.
714   assert(MBBI->isReturn() && "Can only insert epilogue into returning blocks");
715 
716   uint64_t StackSize = MFFrame.getStackSize();
717   if (ZFI->getRestoreGPRRegs().LowGPR) {
718     --MBBI;
719     unsigned Opcode = MBBI->getOpcode();
720     if (Opcode != SystemZ::LMG)
721       llvm_unreachable("Expected to see callee-save register restore code");
722 
723     unsigned AddrOpNo = 2;
724     DebugLoc DL = MBBI->getDebugLoc();
725     uint64_t Offset = StackSize + MBBI->getOperand(AddrOpNo + 1).getImm();
726     unsigned NewOpcode = ZII->getOpcodeForOffset(Opcode, Offset);
727 
728     // If the offset is too large, use the largest stack-aligned offset
729     // and add the rest to the base register (the stack or frame pointer).
730     if (!NewOpcode) {
731       uint64_t NumBytes = Offset - 0x7fff8;
732       emitIncrement(MBB, MBBI, DL, MBBI->getOperand(AddrOpNo).getReg(),
733                     NumBytes, ZII);
734       Offset -= NumBytes;
735       NewOpcode = ZII->getOpcodeForOffset(Opcode, Offset);
736       assert(NewOpcode && "No restore instruction available");
737     }
738 
739     MBBI->setDesc(ZII->get(NewOpcode));
740     MBBI->getOperand(AddrOpNo + 1).ChangeToImmediate(Offset);
741   } else if (StackSize) {
742     DebugLoc DL = MBBI->getDebugLoc();
743     emitIncrement(MBB, MBBI, DL, SystemZ::R15D, StackSize, ZII);
744   }
745 }
746 
747 void SystemZELFFrameLowering::inlineStackProbe(
748     MachineFunction &MF, MachineBasicBlock &PrologMBB) const {
749   auto *ZII =
750     static_cast<const SystemZInstrInfo *>(MF.getSubtarget().getInstrInfo());
751   const SystemZSubtarget &STI = MF.getSubtarget<SystemZSubtarget>();
752   const SystemZTargetLowering &TLI = *STI.getTargetLowering();
753 
754   MachineInstr *StackAllocMI = nullptr;
755   for (MachineInstr &MI : PrologMBB)
756     if (MI.getOpcode() == SystemZ::PROBED_STACKALLOC) {
757       StackAllocMI = &MI;
758       break;
759     }
760   if (StackAllocMI == nullptr)
761     return;
762   uint64_t StackSize = StackAllocMI->getOperand(0).getImm();
763   const unsigned ProbeSize = TLI.getStackProbeSize(MF);
764   uint64_t NumFullBlocks = StackSize / ProbeSize;
765   uint64_t Residual = StackSize % ProbeSize;
766   int64_t SPOffsetFromCFA = -SystemZMC::ELFCFAOffsetFromInitialSP;
767   MachineBasicBlock *MBB = &PrologMBB;
768   MachineBasicBlock::iterator MBBI = StackAllocMI;
769   const DebugLoc DL = StackAllocMI->getDebugLoc();
770 
771   // Allocate a block of Size bytes on the stack and probe it.
772   auto allocateAndProbe = [&](MachineBasicBlock &InsMBB,
773                               MachineBasicBlock::iterator InsPt, unsigned Size,
774                               bool EmitCFI) -> void {
775     emitIncrement(InsMBB, InsPt, DL, SystemZ::R15D, -int64_t(Size), ZII);
776     if (EmitCFI) {
777       SPOffsetFromCFA -= Size;
778       buildCFAOffs(InsMBB, InsPt, DL, SPOffsetFromCFA, ZII);
779     }
780     // Probe by means of a volatile compare.
781     MachineMemOperand *MMO = MF.getMachineMemOperand(MachinePointerInfo(),
782       MachineMemOperand::MOVolatile | MachineMemOperand::MOLoad, 8, Align(1));
783     BuildMI(InsMBB, InsPt, DL, ZII->get(SystemZ::CG))
784       .addReg(SystemZ::R0D, RegState::Undef)
785       .addReg(SystemZ::R15D).addImm(Size - 8).addReg(0)
786       .addMemOperand(MMO);
787   };
788 
789   bool StoreBackchain = MF.getFunction().hasFnAttribute("backchain");
790   if (StoreBackchain)
791     BuildMI(*MBB, MBBI, DL, ZII->get(SystemZ::LGR))
792       .addReg(SystemZ::R1D, RegState::Define).addReg(SystemZ::R15D);
793 
794   MachineBasicBlock *DoneMBB = nullptr;
795   MachineBasicBlock *LoopMBB = nullptr;
796   if (NumFullBlocks < 3) {
797     // Emit unrolled probe statements.
798     for (unsigned int i = 0; i < NumFullBlocks; i++)
799       allocateAndProbe(*MBB, MBBI, ProbeSize, true/*EmitCFI*/);
800   } else {
801     // Emit a loop probing the pages.
802     uint64_t LoopAlloc = ProbeSize * NumFullBlocks;
803     SPOffsetFromCFA -= LoopAlloc;
804 
805     // Use R0D to hold the exit value.
806     BuildMI(*MBB, MBBI, DL, ZII->get(SystemZ::LGR), SystemZ::R0D)
807       .addReg(SystemZ::R15D);
808     buildDefCFAReg(*MBB, MBBI, DL, SystemZ::R0D, ZII);
809     emitIncrement(*MBB, MBBI, DL, SystemZ::R0D, -int64_t(LoopAlloc), ZII);
810     buildCFAOffs(*MBB, MBBI, DL, -int64_t(SystemZMC::ELFCallFrameSize + LoopAlloc),
811                  ZII);
812 
813     DoneMBB = SystemZ::splitBlockBefore(MBBI, MBB);
814     LoopMBB = SystemZ::emitBlockAfter(MBB);
815     MBB->addSuccessor(LoopMBB);
816     LoopMBB->addSuccessor(LoopMBB);
817     LoopMBB->addSuccessor(DoneMBB);
818 
819     MBB = LoopMBB;
820     allocateAndProbe(*MBB, MBB->end(), ProbeSize, false/*EmitCFI*/);
821     BuildMI(*MBB, MBB->end(), DL, ZII->get(SystemZ::CLGR))
822       .addReg(SystemZ::R15D).addReg(SystemZ::R0D);
823     BuildMI(*MBB, MBB->end(), DL, ZII->get(SystemZ::BRC))
824       .addImm(SystemZ::CCMASK_ICMP).addImm(SystemZ::CCMASK_CMP_GT).addMBB(MBB);
825 
826     MBB = DoneMBB;
827     MBBI = DoneMBB->begin();
828     buildDefCFAReg(*MBB, MBBI, DL, SystemZ::R15D, ZII);
829   }
830 
831   if (Residual)
832     allocateAndProbe(*MBB, MBBI, Residual, true/*EmitCFI*/);
833 
834   if (StoreBackchain)
835     BuildMI(*MBB, MBBI, DL, ZII->get(SystemZ::STG))
836       .addReg(SystemZ::R1D, RegState::Kill).addReg(SystemZ::R15D)
837       .addImm(getBackchainOffset(MF)).addReg(0);
838 
839   StackAllocMI->eraseFromParent();
840   if (DoneMBB != nullptr) {
841     // Compute the live-in lists for the new blocks.
842     recomputeLiveIns(*DoneMBB);
843     recomputeLiveIns(*LoopMBB);
844   }
845 }
846 
847 bool SystemZELFFrameLowering::hasFP(const MachineFunction &MF) const {
848   return (MF.getTarget().Options.DisableFramePointerElim(MF) ||
849           MF.getFrameInfo().hasVarSizedObjects());
850 }
851 
852 StackOffset SystemZELFFrameLowering::getFrameIndexReference(
853     const MachineFunction &MF, int FI, Register &FrameReg) const {
854   // Our incoming SP is actually SystemZMC::ELFCallFrameSize below the CFA, so
855   // add that difference here.
856   StackOffset Offset =
857       TargetFrameLowering::getFrameIndexReference(MF, FI, FrameReg);
858   return Offset + StackOffset::getFixed(SystemZMC::ELFCallFrameSize);
859 }
860 
861 unsigned SystemZELFFrameLowering::getRegSpillOffset(MachineFunction &MF,
862                                                     Register Reg) const {
863   bool IsVarArg = MF.getFunction().isVarArg();
864   bool BackChain = MF.getFunction().hasFnAttribute("backchain");
865   bool SoftFloat = MF.getSubtarget<SystemZSubtarget>().hasSoftFloat();
866   unsigned Offset = RegSpillOffsets[Reg];
867   if (usePackedStack(MF) && !(IsVarArg && !SoftFloat)) {
868     if (SystemZ::GR64BitRegClass.contains(Reg))
869       // Put all GPRs at the top of the Register save area with packed
870       // stack. Make room for the backchain if needed.
871       Offset += BackChain ? 24 : 32;
872     else
873       Offset = 0;
874   }
875   return Offset;
876 }
877 
878 int SystemZELFFrameLowering::getOrCreateFramePointerSaveIndex(
879     MachineFunction &MF) const {
880   SystemZMachineFunctionInfo *ZFI = MF.getInfo<SystemZMachineFunctionInfo>();
881   int FI = ZFI->getFramePointerSaveIndex();
882   if (!FI) {
883     MachineFrameInfo &MFFrame = MF.getFrameInfo();
884     int Offset = getBackchainOffset(MF) - SystemZMC::ELFCallFrameSize;
885     FI = MFFrame.CreateFixedObject(8, Offset, false);
886     ZFI->setFramePointerSaveIndex(FI);
887   }
888   return FI;
889 }
890 
891 bool SystemZELFFrameLowering::usePackedStack(MachineFunction &MF) const {
892   bool HasPackedStackAttr = MF.getFunction().hasFnAttribute("packed-stack");
893   bool BackChain = MF.getFunction().hasFnAttribute("backchain");
894   bool SoftFloat = MF.getSubtarget<SystemZSubtarget>().hasSoftFloat();
895   if (HasPackedStackAttr && BackChain && !SoftFloat)
896     report_fatal_error("packed-stack + backchain + hard-float is unsupported.");
897   bool CallConv = MF.getFunction().getCallingConv() != CallingConv::GHC;
898   return HasPackedStackAttr && CallConv;
899 }
900 
901 SystemZXPLINKFrameLowering::SystemZXPLINKFrameLowering()
902     : SystemZFrameLowering(TargetFrameLowering::StackGrowsDown, Align(32), 0,
903                            Align(32), /* StackRealignable */ false),
904       RegSpillOffsets(-1) {
905 
906   // Create a mapping from register number to save slot offset.
907   // These offsets are relative to the start of the local are area.
908   RegSpillOffsets.grow(SystemZ::NUM_TARGET_REGS);
909   for (unsigned I = 0, E = array_lengthof(XPLINKSpillOffsetTable); I != E; ++I)
910     RegSpillOffsets[XPLINKSpillOffsetTable[I].Reg] =
911         XPLINKSpillOffsetTable[I].Offset;
912 }
913 
914 // Checks if the function is a potential candidate for being a XPLeaf routine.
915 static bool isXPLeafCandidate(const MachineFunction &MF) {
916   const MachineFrameInfo &MFFrame = MF.getFrameInfo();
917   const MachineRegisterInfo &MRI = MF.getRegInfo();
918   const SystemZSubtarget &Subtarget = MF.getSubtarget<SystemZSubtarget>();
919   auto *Regs =
920       static_cast<SystemZXPLINK64Registers *>(Subtarget.getSpecialRegisters());
921 
922   // If function calls other functions including alloca, then it is not a XPLeaf
923   // routine.
924   if (MFFrame.hasCalls())
925     return false;
926 
927   // If the function has var Sized Objects, then it is not a XPLeaf routine.
928   if (MFFrame.hasVarSizedObjects())
929     return false;
930 
931   // If the function adjusts the stack, then it is not a XPLeaf routine.
932   if (MFFrame.adjustsStack())
933     return false;
934 
935   // If function modifies the stack pointer register, then it is not a XPLeaf
936   // routine.
937   if (MRI.isPhysRegModified(Regs->getStackPointerRegister()))
938     return false;
939 
940   // If function modifies the ADA register, then it is not a XPLeaf routine.
941   if (MRI.isPhysRegModified(Regs->getAddressOfCalleeRegister()))
942     return false;
943 
944   // If function modifies the return address register, then it is not a XPLeaf
945   // routine.
946   if (MRI.isPhysRegModified(Regs->getReturnFunctionAddressRegister()))
947     return false;
948 
949   // If the backchain pointer should be stored, then it is not a XPLeaf routine.
950   if (MF.getFunction().hasFnAttribute("backchain"))
951     return false;
952 
953   // If function acquires its own stack frame, then it is not a XPLeaf routine.
954   // At the time this function is called, only slots for local variables are
955   // allocated, so this is a very rough estimate.
956   if (MFFrame.estimateStackSize(MF) > 0)
957     return false;
958 
959   return true;
960 }
961 
962 bool SystemZXPLINKFrameLowering::assignCalleeSavedSpillSlots(
963     MachineFunction &MF, const TargetRegisterInfo *TRI,
964     std::vector<CalleeSavedInfo> &CSI) const {
965   MachineFrameInfo &MFFrame = MF.getFrameInfo();
966   SystemZMachineFunctionInfo *MFI = MF.getInfo<SystemZMachineFunctionInfo>();
967   const SystemZSubtarget &Subtarget = MF.getSubtarget<SystemZSubtarget>();
968   auto &Regs = Subtarget.getSpecialRegisters<SystemZXPLINK64Registers>();
969   auto &GRRegClass = SystemZ::GR64BitRegClass;
970 
971   // At this point, the result of isXPLeafCandidate() is not accurate because
972   // the size of the save area has not yet been determined. If
973   // isXPLeafCandidate() indicates a potential leaf function, and there are no
974   // callee-save registers, then it is indeed a leaf function, and we can early
975   // exit.
976   // TODO: It is possible for leaf functions to use callee-saved registers.
977   // It can use the 0-2k range between R4 and the caller's stack frame without
978   // acquiring its own stack frame.
979   bool IsLeaf = CSI.empty() && isXPLeafCandidate(MF);
980   if (IsLeaf)
981     return true;
982 
983   // For non-leaf functions:
984   // - the address of callee (entry point) register R6 must be saved
985   CSI.push_back(CalleeSavedInfo(Regs.getAddressOfCalleeRegister()));
986   CSI.back().setRestored(false);
987 
988   // The return address register R7 must be saved and restored.
989   CSI.push_back(CalleeSavedInfo(Regs.getReturnFunctionAddressRegister()));
990 
991   // If the function needs a frame pointer, or if the backchain pointer should
992   // be stored, then save the stack pointer register R4.
993   if (hasFP(MF) || MF.getFunction().hasFnAttribute("backchain"))
994     CSI.push_back(CalleeSavedInfo(Regs.getStackPointerRegister()));
995 
996   // Scan the call-saved GPRs and find the bounds of the register spill area.
997   Register LowRestoreGPR = 0;
998   int LowRestoreOffset = INT32_MAX;
999   Register LowSpillGPR = 0;
1000   int LowSpillOffset = INT32_MAX;
1001   Register HighGPR = 0;
1002   int HighOffset = -1;
1003 
1004   for (auto &CS : CSI) {
1005     Register Reg = CS.getReg();
1006     int Offset = RegSpillOffsets[Reg];
1007     if (Offset >= 0) {
1008       if (GRRegClass.contains(Reg)) {
1009         if (LowSpillOffset > Offset) {
1010           LowSpillOffset = Offset;
1011           LowSpillGPR = Reg;
1012         }
1013         if (CS.isRestored() && LowRestoreOffset > Offset) {
1014           LowRestoreOffset = Offset;
1015           LowRestoreGPR = Reg;
1016         }
1017 
1018         if (Offset > HighOffset) {
1019           HighOffset = Offset;
1020           HighGPR = Reg;
1021         }
1022         // Non-volatile GPRs are saved in the dedicated register save area at
1023         // the bottom of the stack and are not truly part of the "normal" stack
1024         // frame. Mark the frame index as NoAlloc to indicate it as such.
1025         unsigned RegSize = 8;
1026         int FrameIdx = MFFrame.CreateFixedSpillStackObject(RegSize, Offset);
1027         CS.setFrameIdx(FrameIdx);
1028         MFFrame.setStackID(FrameIdx, TargetStackID::NoAlloc);
1029       }
1030     } else {
1031       Register Reg = CS.getReg();
1032       const TargetRegisterClass *RC = TRI->getMinimalPhysRegClass(Reg);
1033       Align Alignment = TRI->getSpillAlign(*RC);
1034       unsigned Size = TRI->getSpillSize(*RC);
1035       Alignment = std::min(Alignment, getStackAlign());
1036       int FrameIdx = MFFrame.CreateStackObject(Size, Alignment, true);
1037       CS.setFrameIdx(FrameIdx);
1038     }
1039   }
1040 
1041   // Save the range of call-saved registers, for use by the
1042   // prologue/epilogue inserters.
1043   if (LowRestoreGPR)
1044     MFI->setRestoreGPRRegs(LowRestoreGPR, HighGPR, LowRestoreOffset);
1045 
1046   // Save the range of call-saved registers, for use by the epilogue inserter.
1047   assert(LowSpillGPR && "Expected registers to spill");
1048   MFI->setSpillGPRRegs(LowSpillGPR, HighGPR, LowSpillOffset);
1049 
1050   return true;
1051 }
1052 
1053 void SystemZXPLINKFrameLowering::determineCalleeSaves(MachineFunction &MF,
1054                                                       BitVector &SavedRegs,
1055                                                       RegScavenger *RS) const {
1056   TargetFrameLowering::determineCalleeSaves(MF, SavedRegs, RS);
1057 
1058   bool HasFP = hasFP(MF);
1059   const SystemZSubtarget &Subtarget = MF.getSubtarget<SystemZSubtarget>();
1060   auto &Regs = Subtarget.getSpecialRegisters<SystemZXPLINK64Registers>();
1061 
1062   // If the function requires a frame pointer, record that the hard
1063   // frame pointer will be clobbered.
1064   if (HasFP)
1065     SavedRegs.set(Regs.getFramePointerRegister());
1066 }
1067 
1068 bool SystemZXPLINKFrameLowering::spillCalleeSavedRegisters(
1069     MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI,
1070     ArrayRef<CalleeSavedInfo> CSI, const TargetRegisterInfo *TRI) const {
1071   if (CSI.empty())
1072     return true;
1073 
1074   MachineFunction &MF = *MBB.getParent();
1075   SystemZMachineFunctionInfo *ZFI = MF.getInfo<SystemZMachineFunctionInfo>();
1076   const SystemZSubtarget &Subtarget = MF.getSubtarget<SystemZSubtarget>();
1077   const TargetInstrInfo *TII = Subtarget.getInstrInfo();
1078   auto &Regs = Subtarget.getSpecialRegisters<SystemZXPLINK64Registers>();
1079   SystemZ::GPRRegs SpillGPRs = ZFI->getSpillGPRRegs();
1080   DebugLoc DL;
1081 
1082   // Save GPRs
1083   if (SpillGPRs.LowGPR) {
1084     assert(SpillGPRs.LowGPR != SpillGPRs.HighGPR &&
1085            "Should be saving multiple registers");
1086 
1087     // Build an STM/STMG instruction.
1088     MachineInstrBuilder MIB = BuildMI(MBB, MBBI, DL, TII->get(SystemZ::STMG));
1089 
1090     // Add the explicit register operands.
1091     addSavedGPR(MBB, MIB, SpillGPRs.LowGPR, false);
1092     addSavedGPR(MBB, MIB, SpillGPRs.HighGPR, false);
1093 
1094     // Add the address r4
1095     MIB.addReg(Regs.getStackPointerRegister());
1096 
1097     // Add the partial offset
1098     // We cannot add the actual offset as, at the stack is not finalized
1099     MIB.addImm(SpillGPRs.GPROffset);
1100 
1101     // Make sure all call-saved GPRs are included as operands and are
1102     // marked as live on entry.
1103     auto &GRRegClass = SystemZ::GR64BitRegClass;
1104     for (const CalleeSavedInfo &I : CSI) {
1105       Register Reg = I.getReg();
1106       if (GRRegClass.contains(Reg))
1107         addSavedGPR(MBB, MIB, Reg, true);
1108     }
1109   }
1110 
1111   // Spill FPRs to the stack in the normal TargetInstrInfo way
1112   for (const CalleeSavedInfo &I : CSI) {
1113     Register Reg = I.getReg();
1114     if (SystemZ::FP64BitRegClass.contains(Reg)) {
1115       MBB.addLiveIn(Reg);
1116       TII->storeRegToStackSlot(MBB, MBBI, Reg, true, I.getFrameIdx(),
1117                                &SystemZ::FP64BitRegClass, TRI);
1118     }
1119     if (SystemZ::VR128BitRegClass.contains(Reg)) {
1120       MBB.addLiveIn(Reg);
1121       TII->storeRegToStackSlot(MBB, MBBI, Reg, true, I.getFrameIdx(),
1122                                &SystemZ::VR128BitRegClass, TRI);
1123     }
1124   }
1125 
1126   return true;
1127 }
1128 
1129 bool SystemZXPLINKFrameLowering::restoreCalleeSavedRegisters(
1130     MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI,
1131     MutableArrayRef<CalleeSavedInfo> CSI, const TargetRegisterInfo *TRI) const {
1132 
1133   if (CSI.empty())
1134     return false;
1135 
1136   MachineFunction &MF = *MBB.getParent();
1137   SystemZMachineFunctionInfo *ZFI = MF.getInfo<SystemZMachineFunctionInfo>();
1138   const SystemZSubtarget &Subtarget = MF.getSubtarget<SystemZSubtarget>();
1139   const TargetInstrInfo *TII = Subtarget.getInstrInfo();
1140   auto &Regs = Subtarget.getSpecialRegisters<SystemZXPLINK64Registers>();
1141 
1142   DebugLoc DL = MBBI != MBB.end() ? MBBI->getDebugLoc() : DebugLoc();
1143 
1144   // Restore FPRs in the normal TargetInstrInfo way.
1145   for (unsigned I = 0, E = CSI.size(); I != E; ++I) {
1146     Register Reg = CSI[I].getReg();
1147     if (SystemZ::FP64BitRegClass.contains(Reg))
1148       TII->loadRegFromStackSlot(MBB, MBBI, Reg, CSI[I].getFrameIdx(),
1149                                 &SystemZ::FP64BitRegClass, TRI);
1150     if (SystemZ::VR128BitRegClass.contains(Reg))
1151       TII->loadRegFromStackSlot(MBB, MBBI, Reg, CSI[I].getFrameIdx(),
1152                                 &SystemZ::VR128BitRegClass, TRI);
1153   }
1154 
1155   // Restore call-saved GPRs (but not call-clobbered varargs, which at
1156   // this point might hold return values).
1157   SystemZ::GPRRegs RestoreGPRs = ZFI->getRestoreGPRRegs();
1158   if (RestoreGPRs.LowGPR) {
1159     assert(isInt<20>(Regs.getStackPointerBias() + RestoreGPRs.GPROffset));
1160     if (RestoreGPRs.LowGPR == RestoreGPRs.HighGPR)
1161       // Build an LG/L instruction.
1162       BuildMI(MBB, MBBI, DL, TII->get(SystemZ::LG), RestoreGPRs.LowGPR)
1163           .addReg(Regs.getStackPointerRegister())
1164           .addImm(Regs.getStackPointerBias() + RestoreGPRs.GPROffset)
1165           .addReg(0);
1166     else {
1167       // Build an LMG/LM instruction.
1168       MachineInstrBuilder MIB = BuildMI(MBB, MBBI, DL, TII->get(SystemZ::LMG));
1169 
1170       // Add the explicit register operands.
1171       MIB.addReg(RestoreGPRs.LowGPR, RegState::Define);
1172       MIB.addReg(RestoreGPRs.HighGPR, RegState::Define);
1173 
1174       // Add the address.
1175       MIB.addReg(Regs.getStackPointerRegister());
1176       MIB.addImm(Regs.getStackPointerBias() + RestoreGPRs.GPROffset);
1177 
1178       // Do a second scan adding regs as being defined by instruction
1179       for (unsigned I = 0, E = CSI.size(); I != E; ++I) {
1180         Register Reg = CSI[I].getReg();
1181         if (Reg > RestoreGPRs.LowGPR && Reg < RestoreGPRs.HighGPR)
1182           MIB.addReg(Reg, RegState::ImplicitDefine);
1183       }
1184     }
1185   }
1186 
1187   return true;
1188 }
1189 
1190 void SystemZXPLINKFrameLowering::emitPrologue(MachineFunction &MF,
1191                                               MachineBasicBlock &MBB) const {
1192   assert(&MF.front() == &MBB && "Shrink-wrapping not yet supported");
1193   const SystemZSubtarget &Subtarget = MF.getSubtarget<SystemZSubtarget>();
1194   SystemZMachineFunctionInfo *ZFI = MF.getInfo<SystemZMachineFunctionInfo>();
1195   MachineBasicBlock::iterator MBBI = MBB.begin();
1196   auto *ZII = static_cast<const SystemZInstrInfo *>(Subtarget.getInstrInfo());
1197   auto &Regs = Subtarget.getSpecialRegisters<SystemZXPLINK64Registers>();
1198   MachineFrameInfo &MFFrame = MF.getFrameInfo();
1199   MachineInstr *StoreInstr = nullptr;
1200 
1201   determineFrameLayout(MF);
1202 
1203   bool HasFP = hasFP(MF);
1204   // Debug location must be unknown since the first debug location is used
1205   // to determine the end of the prologue.
1206   DebugLoc DL;
1207   uint64_t Offset = 0;
1208 
1209   const uint64_t StackSize = MFFrame.getStackSize();
1210 
1211   if (ZFI->getSpillGPRRegs().LowGPR) {
1212     // Skip over the GPR saves.
1213     if ((MBBI != MBB.end()) && ((MBBI->getOpcode() == SystemZ::STMG))) {
1214       const int Operand = 3;
1215       // Now we can set the offset for the operation, since now the Stack
1216       // has been finalized.
1217       Offset = Regs.getStackPointerBias() + MBBI->getOperand(Operand).getImm();
1218       // Maximum displacement for STMG instruction.
1219       if (isInt<20>(Offset - StackSize))
1220         Offset -= StackSize;
1221       else
1222         StoreInstr = &*MBBI;
1223       MBBI->getOperand(Operand).setImm(Offset);
1224       ++MBBI;
1225     } else
1226       llvm_unreachable("Couldn't skip over GPR saves");
1227   }
1228 
1229   if (StackSize) {
1230     MachineBasicBlock::iterator InsertPt = StoreInstr ? StoreInstr : MBBI;
1231     // Allocate StackSize bytes.
1232     int64_t Delta = -int64_t(StackSize);
1233 
1234     // In case the STM(G) instruction also stores SP (R4), but the displacement
1235     // is too large, the SP register is manipulated first before storing,
1236     // resulting in the wrong value stored and retrieved later. In this case, we
1237     // need to temporarily save the value of SP, and store it later to memory.
1238     if (StoreInstr && HasFP) {
1239       // Insert LR r0,r4 before STMG instruction.
1240       BuildMI(MBB, InsertPt, DL, ZII->get(SystemZ::LGR))
1241           .addReg(SystemZ::R0D, RegState::Define)
1242           .addReg(SystemZ::R4D);
1243       // Insert ST r0,xxx(,r4) after STMG instruction.
1244       BuildMI(MBB, MBBI, DL, ZII->get(SystemZ::STG))
1245           .addReg(SystemZ::R0D, RegState::Kill)
1246           .addReg(SystemZ::R4D)
1247           .addImm(Offset)
1248           .addReg(0);
1249     }
1250 
1251     emitIncrement(MBB, InsertPt, DL, Regs.getStackPointerRegister(), Delta,
1252                   ZII);
1253 
1254     // If the requested stack size is larger than the guard page, then we need
1255     // to check if we need to call the stack extender. This requires adding a
1256     // conditional branch, but splitting the prologue block is not possible at
1257     // this point since it would invalidate the SaveBlocks / RestoreBlocks sets
1258     // of PEI in the single block function case. Build a pseudo to be handled
1259     // later by inlineStackProbe().
1260     const uint64_t GuardPageSize = 1024 * 1024;
1261     if (StackSize > GuardPageSize) {
1262       assert(StoreInstr && "Wrong insertion point");
1263       BuildMI(MBB, InsertPt, DL, ZII->get(SystemZ::XPLINK_STACKALLOC));
1264     }
1265   }
1266 
1267   if (HasFP) {
1268     // Copy the base of the frame to Frame Pointer Register.
1269     BuildMI(MBB, MBBI, DL, ZII->get(SystemZ::LGR),
1270             Regs.getFramePointerRegister())
1271         .addReg(Regs.getStackPointerRegister());
1272 
1273     // Mark the FramePtr as live at the beginning of every block except
1274     // the entry block.  (We'll have marked R8 as live on entry when
1275     // saving the GPRs.)
1276     for (MachineBasicBlock &B : llvm::drop_begin(MF))
1277       B.addLiveIn(Regs.getFramePointerRegister());
1278   }
1279 }
1280 
1281 void SystemZXPLINKFrameLowering::emitEpilogue(MachineFunction &MF,
1282                                               MachineBasicBlock &MBB) const {
1283   const SystemZSubtarget &Subtarget = MF.getSubtarget<SystemZSubtarget>();
1284   MachineBasicBlock::iterator MBBI = MBB.getLastNonDebugInstr();
1285   SystemZMachineFunctionInfo *ZFI = MF.getInfo<SystemZMachineFunctionInfo>();
1286   MachineFrameInfo &MFFrame = MF.getFrameInfo();
1287   auto *ZII = static_cast<const SystemZInstrInfo *>(Subtarget.getInstrInfo());
1288   auto &Regs = Subtarget.getSpecialRegisters<SystemZXPLINK64Registers>();
1289 
1290   // Skip the return instruction.
1291   assert(MBBI->isReturn() && "Can only insert epilogue into returning blocks");
1292 
1293   uint64_t StackSize = MFFrame.getStackSize();
1294   if (StackSize) {
1295     unsigned SPReg = Regs.getStackPointerRegister();
1296     if (ZFI->getRestoreGPRRegs().LowGPR != SPReg) {
1297       DebugLoc DL = MBBI->getDebugLoc();
1298       emitIncrement(MBB, MBBI, DL, SPReg, StackSize, ZII);
1299     }
1300   }
1301 }
1302 
1303 // Emit a compare of the stack pointer against the stack floor, and a call to
1304 // the LE stack extender if needed.
1305 void SystemZXPLINKFrameLowering::inlineStackProbe(
1306     MachineFunction &MF, MachineBasicBlock &PrologMBB) const {
1307   auto *ZII =
1308       static_cast<const SystemZInstrInfo *>(MF.getSubtarget().getInstrInfo());
1309 
1310   MachineInstr *StackAllocMI = nullptr;
1311   for (MachineInstr &MI : PrologMBB)
1312     if (MI.getOpcode() == SystemZ::XPLINK_STACKALLOC) {
1313       StackAllocMI = &MI;
1314       break;
1315     }
1316   if (StackAllocMI == nullptr)
1317     return;
1318 
1319   MachineBasicBlock &MBB = PrologMBB;
1320   const DebugLoc DL = StackAllocMI->getDebugLoc();
1321 
1322   // The 2nd half of block MBB after split.
1323   MachineBasicBlock *NextMBB;
1324 
1325   // Add new basic block for the call to the stack overflow function.
1326   MachineBasicBlock *StackExtMBB =
1327       MF.CreateMachineBasicBlock(MBB.getBasicBlock());
1328   MF.push_back(StackExtMBB);
1329 
1330   // LG r3,72(,r3)
1331   BuildMI(StackExtMBB, DL, ZII->get(SystemZ::LG), SystemZ::R3D)
1332       .addReg(SystemZ::R3D)
1333       .addImm(72)
1334       .addReg(0);
1335   // BASR r3,r3
1336   BuildMI(StackExtMBB, DL, ZII->get(SystemZ::CallBASR_STACKEXT))
1337       .addReg(SystemZ::R3D);
1338 
1339   // LLGT r3,1208
1340   BuildMI(MBB, StackAllocMI, DL, ZII->get(SystemZ::LLGT), SystemZ::R3D)
1341       .addReg(0)
1342       .addImm(1208)
1343       .addReg(0);
1344   // CG r4,64(,r3)
1345   BuildMI(MBB, StackAllocMI, DL, ZII->get(SystemZ::CG))
1346       .addReg(SystemZ::R4D)
1347       .addReg(SystemZ::R3D)
1348       .addImm(64)
1349       .addReg(0);
1350   // JLL b'0100',F'37'
1351   BuildMI(MBB, StackAllocMI, DL, ZII->get(SystemZ::BRC))
1352       .addImm(SystemZ::CCMASK_ICMP)
1353       .addImm(SystemZ::CCMASK_CMP_LT)
1354       .addMBB(StackExtMBB);
1355 
1356   NextMBB = SystemZ::splitBlockBefore(StackAllocMI, &MBB);
1357   MBB.addSuccessor(NextMBB);
1358   MBB.addSuccessor(StackExtMBB);
1359 
1360   // Add jump back from stack extension BB.
1361   BuildMI(StackExtMBB, DL, ZII->get(SystemZ::J)).addMBB(NextMBB);
1362   StackExtMBB->addSuccessor(NextMBB);
1363 
1364   StackAllocMI->eraseFromParent();
1365 
1366   // Compute the live-in lists for the new blocks.
1367   recomputeLiveIns(*NextMBB);
1368   recomputeLiveIns(*StackExtMBB);
1369 }
1370 
1371 bool SystemZXPLINKFrameLowering::hasFP(const MachineFunction &MF) const {
1372   return (MF.getFrameInfo().hasVarSizedObjects());
1373 }
1374 
1375 void SystemZXPLINKFrameLowering::processFunctionBeforeFrameFinalized(
1376     MachineFunction &MF, RegScavenger *RS) const {
1377   MachineFrameInfo &MFFrame = MF.getFrameInfo();
1378   const SystemZSubtarget &Subtarget = MF.getSubtarget<SystemZSubtarget>();
1379   auto &Regs = Subtarget.getSpecialRegisters<SystemZXPLINK64Registers>();
1380 
1381   // Setup stack frame offset
1382   MFFrame.setOffsetAdjustment(Regs.getStackPointerBias());
1383 }
1384 
1385 // Determines the size of the frame, and creates the deferred spill objects.
1386 void SystemZXPLINKFrameLowering::determineFrameLayout(
1387     MachineFunction &MF) const {
1388   MachineFrameInfo &MFFrame = MF.getFrameInfo();
1389   const SystemZSubtarget &Subtarget = MF.getSubtarget<SystemZSubtarget>();
1390   auto *Regs =
1391       static_cast<SystemZXPLINK64Registers *>(Subtarget.getSpecialRegisters());
1392 
1393   uint64_t StackSize = MFFrame.getStackSize();
1394   if (StackSize == 0)
1395     return;
1396 
1397   // Add the size of the register save area and the reserved area to the size.
1398   StackSize += Regs->getCallFrameSize();
1399   MFFrame.setStackSize(StackSize);
1400 
1401   // We now know the stack size. Create the fixed spill stack objects for the
1402   // register save area now. This has no impact on the stack frame layout, as
1403   // this is already computed. However, it makes sure that all callee saved
1404   // registers have a valid frame index assigned.
1405   const unsigned RegSize = MF.getDataLayout().getPointerSize();
1406   for (auto &CS : MFFrame.getCalleeSavedInfo()) {
1407     int Offset = RegSpillOffsets[CS.getReg()];
1408     if (Offset >= 0)
1409       CS.setFrameIdx(
1410           MFFrame.CreateFixedSpillStackObject(RegSize, Offset - StackSize));
1411   }
1412 }
1413