1 //===-- X86RegisterInfo.cpp - X86 Register Information --------------------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // This file contains the X86 implementation of the TargetRegisterInfo class.
10 // This file is responsible for the frame pointer elimination optimization
11 // on X86.
12 //
13 //===----------------------------------------------------------------------===//
14 
15 #include "X86RegisterInfo.h"
16 #include "X86FrameLowering.h"
17 #include "X86MachineFunctionInfo.h"
18 #include "X86Subtarget.h"
19 #include "llvm/ADT/BitVector.h"
20 #include "llvm/ADT/STLExtras.h"
21 #include "llvm/CodeGen/MachineFrameInfo.h"
22 #include "llvm/CodeGen/MachineFunction.h"
23 #include "llvm/CodeGen/MachineFunctionPass.h"
24 #include "llvm/CodeGen/MachineRegisterInfo.h"
25 #include "llvm/CodeGen/TargetFrameLowering.h"
26 #include "llvm/CodeGen/TargetInstrInfo.h"
27 #include "llvm/IR/Constants.h"
28 #include "llvm/IR/Function.h"
29 #include "llvm/IR/Type.h"
30 #include "llvm/Support/CommandLine.h"
31 #include "llvm/Support/ErrorHandling.h"
32 #include "llvm/Target/TargetMachine.h"
33 #include "llvm/Target/TargetOptions.h"
34 
35 using namespace llvm;
36 
37 #define GET_REGINFO_TARGET_DESC
38 #include "X86GenRegisterInfo.inc"
39 
40 static cl::opt<bool>
41 EnableBasePointer("x86-use-base-pointer", cl::Hidden, cl::init(true),
42           cl::desc("Enable use of a base pointer for complex stack frames"));
43 
44 X86RegisterInfo::X86RegisterInfo(const Triple &TT)
45     : X86GenRegisterInfo((TT.isArch64Bit() ? X86::RIP : X86::EIP),
46                          X86_MC::getDwarfRegFlavour(TT, false),
47                          X86_MC::getDwarfRegFlavour(TT, true),
48                          (TT.isArch64Bit() ? X86::RIP : X86::EIP)) {
49   X86_MC::initLLVMToSEHAndCVRegMapping(this);
50 
51   // Cache some information.
52   Is64Bit = TT.isArch64Bit();
53   IsWin64 = Is64Bit && TT.isOSWindows();
54 
55   // Use a callee-saved register as the base pointer.  These registers must
56   // not conflict with any ABI requirements.  For example, in 32-bit mode PIC
57   // requires GOT in the EBX register before function calls via PLT GOT pointer.
58   if (Is64Bit) {
59     SlotSize = 8;
60     // This matches the simplified 32-bit pointer code in the data layout
61     // computation.
62     // FIXME: Should use the data layout?
63     bool Use64BitReg = TT.getEnvironment() != Triple::GNUX32;
64     StackPtr = Use64BitReg ? X86::RSP : X86::ESP;
65     FramePtr = Use64BitReg ? X86::RBP : X86::EBP;
66     BasePtr = Use64BitReg ? X86::RBX : X86::EBX;
67   } else {
68     SlotSize = 4;
69     StackPtr = X86::ESP;
70     FramePtr = X86::EBP;
71     BasePtr = X86::ESI;
72   }
73 }
74 
75 bool
76 X86RegisterInfo::trackLivenessAfterRegAlloc(const MachineFunction &MF) const {
77   // ExecutionDomainFix, BreakFalseDeps and PostRAScheduler require liveness.
78   return true;
79 }
80 
81 int
82 X86RegisterInfo::getSEHRegNum(unsigned i) const {
83   return getEncodingValue(i);
84 }
85 
86 const TargetRegisterClass *
87 X86RegisterInfo::getSubClassWithSubReg(const TargetRegisterClass *RC,
88                                        unsigned Idx) const {
89   // The sub_8bit sub-register index is more constrained in 32-bit mode.
90   // It behaves just like the sub_8bit_hi index.
91   if (!Is64Bit && Idx == X86::sub_8bit)
92     Idx = X86::sub_8bit_hi;
93 
94   // Forward to TableGen's default version.
95   return X86GenRegisterInfo::getSubClassWithSubReg(RC, Idx);
96 }
97 
98 const TargetRegisterClass *
99 X86RegisterInfo::getMatchingSuperRegClass(const TargetRegisterClass *A,
100                                           const TargetRegisterClass *B,
101                                           unsigned SubIdx) const {
102   // The sub_8bit sub-register index is more constrained in 32-bit mode.
103   if (!Is64Bit && SubIdx == X86::sub_8bit) {
104     A = X86GenRegisterInfo::getSubClassWithSubReg(A, X86::sub_8bit_hi);
105     if (!A)
106       return nullptr;
107   }
108   return X86GenRegisterInfo::getMatchingSuperRegClass(A, B, SubIdx);
109 }
110 
111 const TargetRegisterClass *
112 X86RegisterInfo::getLargestLegalSuperClass(const TargetRegisterClass *RC,
113                                            const MachineFunction &MF) const {
114   // Don't allow super-classes of GR8_NOREX.  This class is only used after
115   // extracting sub_8bit_hi sub-registers.  The H sub-registers cannot be copied
116   // to the full GR8 register class in 64-bit mode, so we cannot allow the
117   // reigster class inflation.
118   //
119   // The GR8_NOREX class is always used in a way that won't be constrained to a
120   // sub-class, so sub-classes like GR8_ABCD_L are allowed to expand to the
121   // full GR8 class.
122   if (RC == &X86::GR8_NOREXRegClass)
123     return RC;
124 
125   const X86Subtarget &Subtarget = MF.getSubtarget<X86Subtarget>();
126 
127   const TargetRegisterClass *Super = RC;
128   TargetRegisterClass::sc_iterator I = RC->getSuperClasses();
129   do {
130     switch (Super->getID()) {
131     case X86::FR32RegClassID:
132     case X86::FR64RegClassID:
133       // If AVX-512 isn't supported we should only inflate to these classes.
134       if (!Subtarget.hasAVX512() &&
135           getRegSizeInBits(*Super) == getRegSizeInBits(*RC))
136         return Super;
137       break;
138     case X86::VR128RegClassID:
139     case X86::VR256RegClassID:
140       // If VLX isn't supported we should only inflate to these classes.
141       if (!Subtarget.hasVLX() &&
142           getRegSizeInBits(*Super) == getRegSizeInBits(*RC))
143         return Super;
144       break;
145     case X86::VR128XRegClassID:
146     case X86::VR256XRegClassID:
147       // If VLX isn't support we shouldn't inflate to these classes.
148       if (Subtarget.hasVLX() &&
149           getRegSizeInBits(*Super) == getRegSizeInBits(*RC))
150         return Super;
151       break;
152     case X86::FR32XRegClassID:
153     case X86::FR64XRegClassID:
154       // If AVX-512 isn't support we shouldn't inflate to these classes.
155       if (Subtarget.hasAVX512() &&
156           getRegSizeInBits(*Super) == getRegSizeInBits(*RC))
157         return Super;
158       break;
159     case X86::GR8RegClassID:
160     case X86::GR16RegClassID:
161     case X86::GR32RegClassID:
162     case X86::GR64RegClassID:
163     case X86::RFP32RegClassID:
164     case X86::RFP64RegClassID:
165     case X86::RFP80RegClassID:
166     case X86::VR512_0_15RegClassID:
167     case X86::VR512RegClassID:
168       // Don't return a super-class that would shrink the spill size.
169       // That can happen with the vector and float classes.
170       if (getRegSizeInBits(*Super) == getRegSizeInBits(*RC))
171         return Super;
172     }
173     Super = *I++;
174   } while (Super);
175   return RC;
176 }
177 
178 const TargetRegisterClass *
179 X86RegisterInfo::getPointerRegClass(const MachineFunction &MF,
180                                     unsigned Kind) const {
181   const X86Subtarget &Subtarget = MF.getSubtarget<X86Subtarget>();
182   switch (Kind) {
183   default: llvm_unreachable("Unexpected Kind in getPointerRegClass!");
184   case 0: // Normal GPRs.
185     if (Subtarget.isTarget64BitLP64())
186       return &X86::GR64RegClass;
187     // If the target is 64bit but we have been told to use 32bit addresses,
188     // we can still use 64-bit register as long as we know the high bits
189     // are zeros.
190     // Reflect that in the returned register class.
191     if (Is64Bit) {
192       // When the target also allows 64-bit frame pointer and we do have a
193       // frame, this is fine to use it for the address accesses as well.
194       const X86FrameLowering *TFI = getFrameLowering(MF);
195       return TFI->hasFP(MF) && TFI->Uses64BitFramePtr
196                  ? &X86::LOW32_ADDR_ACCESS_RBPRegClass
197                  : &X86::LOW32_ADDR_ACCESSRegClass;
198     }
199     return &X86::GR32RegClass;
200   case 1: // Normal GPRs except the stack pointer (for encoding reasons).
201     if (Subtarget.isTarget64BitLP64())
202       return &X86::GR64_NOSPRegClass;
203     // NOSP does not contain RIP, so no special case here.
204     return &X86::GR32_NOSPRegClass;
205   case 2: // NOREX GPRs.
206     if (Subtarget.isTarget64BitLP64())
207       return &X86::GR64_NOREXRegClass;
208     return &X86::GR32_NOREXRegClass;
209   case 3: // NOREX GPRs except the stack pointer (for encoding reasons).
210     if (Subtarget.isTarget64BitLP64())
211       return &X86::GR64_NOREX_NOSPRegClass;
212     // NOSP does not contain RIP, so no special case here.
213     return &X86::GR32_NOREX_NOSPRegClass;
214   case 4: // Available for tailcall (not callee-saved GPRs).
215     return getGPRsForTailCall(MF);
216   }
217 }
218 
219 bool X86RegisterInfo::shouldRewriteCopySrc(const TargetRegisterClass *DefRC,
220                                            unsigned DefSubReg,
221                                            const TargetRegisterClass *SrcRC,
222                                            unsigned SrcSubReg) const {
223   // Prevent rewriting a copy where the destination size is larger than the
224   // input size. See PR41619.
225   // FIXME: Should this be factored into the base implementation somehow.
226   if (DefRC->hasSuperClassEq(&X86::GR64RegClass) && DefSubReg == 0 &&
227       SrcRC->hasSuperClassEq(&X86::GR64RegClass) && SrcSubReg == X86::sub_32bit)
228     return false;
229 
230   return TargetRegisterInfo::shouldRewriteCopySrc(DefRC, DefSubReg,
231                                                   SrcRC, SrcSubReg);
232 }
233 
234 const TargetRegisterClass *
235 X86RegisterInfo::getGPRsForTailCall(const MachineFunction &MF) const {
236   const Function &F = MF.getFunction();
237   if (IsWin64 || (F.getCallingConv() == CallingConv::Win64))
238     return &X86::GR64_TCW64RegClass;
239   else if (Is64Bit)
240     return &X86::GR64_TCRegClass;
241 
242   bool hasHipeCC = (F.getCallingConv() == CallingConv::HiPE);
243   if (hasHipeCC)
244     return &X86::GR32RegClass;
245   return &X86::GR32_TCRegClass;
246 }
247 
248 const TargetRegisterClass *
249 X86RegisterInfo::getCrossCopyRegClass(const TargetRegisterClass *RC) const {
250   if (RC == &X86::CCRRegClass) {
251     if (Is64Bit)
252       return &X86::GR64RegClass;
253     else
254       return &X86::GR32RegClass;
255   }
256   return RC;
257 }
258 
259 unsigned
260 X86RegisterInfo::getRegPressureLimit(const TargetRegisterClass *RC,
261                                      MachineFunction &MF) const {
262   const X86FrameLowering *TFI = getFrameLowering(MF);
263 
264   unsigned FPDiff = TFI->hasFP(MF) ? 1 : 0;
265   switch (RC->getID()) {
266   default:
267     return 0;
268   case X86::GR32RegClassID:
269     return 4 - FPDiff;
270   case X86::GR64RegClassID:
271     return 12 - FPDiff;
272   case X86::VR128RegClassID:
273     return Is64Bit ? 10 : 4;
274   case X86::VR64RegClassID:
275     return 4;
276   }
277 }
278 
279 const MCPhysReg *
280 X86RegisterInfo::getCalleeSavedRegs(const MachineFunction *MF) const {
281   assert(MF && "MachineFunction required");
282 
283   const X86Subtarget &Subtarget = MF->getSubtarget<X86Subtarget>();
284   const Function &F = MF->getFunction();
285   bool HasSSE = Subtarget.hasSSE1();
286   bool HasAVX = Subtarget.hasAVX();
287   bool HasAVX512 = Subtarget.hasAVX512();
288   bool CallsEHReturn = MF->callsEHReturn();
289 
290   CallingConv::ID CC = F.getCallingConv();
291 
292   // If attribute NoCallerSavedRegisters exists then we set X86_INTR calling
293   // convention because it has the CSR list.
294   if (MF->getFunction().hasFnAttribute("no_caller_saved_registers"))
295     CC = CallingConv::X86_INTR;
296 
297   switch (CC) {
298   case CallingConv::GHC:
299   case CallingConv::HiPE:
300     return CSR_NoRegs_SaveList;
301   case CallingConv::AnyReg:
302     if (HasAVX)
303       return CSR_64_AllRegs_AVX_SaveList;
304     return CSR_64_AllRegs_SaveList;
305   case CallingConv::PreserveMost:
306     return CSR_64_RT_MostRegs_SaveList;
307   case CallingConv::PreserveAll:
308     if (HasAVX)
309       return CSR_64_RT_AllRegs_AVX_SaveList;
310     return CSR_64_RT_AllRegs_SaveList;
311   case CallingConv::CXX_FAST_TLS:
312     if (Is64Bit)
313       return MF->getInfo<X86MachineFunctionInfo>()->isSplitCSR() ?
314              CSR_64_CXX_TLS_Darwin_PE_SaveList : CSR_64_TLS_Darwin_SaveList;
315     break;
316   case CallingConv::Intel_OCL_BI: {
317     if (HasAVX512 && IsWin64)
318       return CSR_Win64_Intel_OCL_BI_AVX512_SaveList;
319     if (HasAVX512 && Is64Bit)
320       return CSR_64_Intel_OCL_BI_AVX512_SaveList;
321     if (HasAVX && IsWin64)
322       return CSR_Win64_Intel_OCL_BI_AVX_SaveList;
323     if (HasAVX && Is64Bit)
324       return CSR_64_Intel_OCL_BI_AVX_SaveList;
325     if (!HasAVX && !IsWin64 && Is64Bit)
326       return CSR_64_Intel_OCL_BI_SaveList;
327     break;
328   }
329   case CallingConv::HHVM:
330     return CSR_64_HHVM_SaveList;
331   case CallingConv::X86_RegCall:
332     if (Is64Bit) {
333       if (IsWin64) {
334         return (HasSSE ? CSR_Win64_RegCall_SaveList :
335                          CSR_Win64_RegCall_NoSSE_SaveList);
336       } else {
337         return (HasSSE ? CSR_SysV64_RegCall_SaveList :
338                          CSR_SysV64_RegCall_NoSSE_SaveList);
339       }
340     } else {
341       return (HasSSE ? CSR_32_RegCall_SaveList :
342                        CSR_32_RegCall_NoSSE_SaveList);
343     }
344   case CallingConv::CFGuard_Check:
345     assert(!Is64Bit && "CFGuard check mechanism only used on 32-bit X86");
346     return (HasSSE ? CSR_Win32_CFGuard_Check_SaveList
347                    : CSR_Win32_CFGuard_Check_NoSSE_SaveList);
348   case CallingConv::Cold:
349     if (Is64Bit)
350       return CSR_64_MostRegs_SaveList;
351     break;
352   case CallingConv::Win64:
353     if (!HasSSE)
354       return CSR_Win64_NoSSE_SaveList;
355     return CSR_Win64_SaveList;
356   case CallingConv::X86_64_SysV:
357     if (CallsEHReturn)
358       return CSR_64EHRet_SaveList;
359     return CSR_64_SaveList;
360   case CallingConv::X86_INTR:
361     if (Is64Bit) {
362       if (HasAVX512)
363         return CSR_64_AllRegs_AVX512_SaveList;
364       if (HasAVX)
365         return CSR_64_AllRegs_AVX_SaveList;
366       if (HasSSE)
367         return CSR_64_AllRegs_SaveList;
368       return CSR_64_AllRegs_NoSSE_SaveList;
369     } else {
370       if (HasAVX512)
371         return CSR_32_AllRegs_AVX512_SaveList;
372       if (HasAVX)
373         return CSR_32_AllRegs_AVX_SaveList;
374       if (HasSSE)
375         return CSR_32_AllRegs_SSE_SaveList;
376       return CSR_32_AllRegs_SaveList;
377     }
378   default:
379     break;
380   }
381 
382   if (Is64Bit) {
383     bool IsSwiftCC = Subtarget.getTargetLowering()->supportSwiftError() &&
384                      F.getAttributes().hasAttrSomewhere(Attribute::SwiftError);
385     if (IsSwiftCC)
386       return IsWin64 ? CSR_Win64_SwiftError_SaveList
387                      : CSR_64_SwiftError_SaveList;
388 
389     if (IsWin64)
390       return HasSSE ? CSR_Win64_SaveList : CSR_Win64_NoSSE_SaveList;
391     if (CallsEHReturn)
392       return CSR_64EHRet_SaveList;
393     return CSR_64_SaveList;
394   }
395 
396   return CallsEHReturn ? CSR_32EHRet_SaveList : CSR_32_SaveList;
397 }
398 
399 const MCPhysReg *X86RegisterInfo::getCalleeSavedRegsViaCopy(
400     const MachineFunction *MF) const {
401   assert(MF && "Invalid MachineFunction pointer.");
402   if (MF->getFunction().getCallingConv() == CallingConv::CXX_FAST_TLS &&
403       MF->getInfo<X86MachineFunctionInfo>()->isSplitCSR())
404     return CSR_64_CXX_TLS_Darwin_ViaCopy_SaveList;
405   return nullptr;
406 }
407 
408 const uint32_t *
409 X86RegisterInfo::getCallPreservedMask(const MachineFunction &MF,
410                                       CallingConv::ID CC) const {
411   const X86Subtarget &Subtarget = MF.getSubtarget<X86Subtarget>();
412   bool HasSSE = Subtarget.hasSSE1();
413   bool HasAVX = Subtarget.hasAVX();
414   bool HasAVX512 = Subtarget.hasAVX512();
415 
416   switch (CC) {
417   case CallingConv::GHC:
418   case CallingConv::HiPE:
419     return CSR_NoRegs_RegMask;
420   case CallingConv::AnyReg:
421     if (HasAVX)
422       return CSR_64_AllRegs_AVX_RegMask;
423     return CSR_64_AllRegs_RegMask;
424   case CallingConv::PreserveMost:
425     return CSR_64_RT_MostRegs_RegMask;
426   case CallingConv::PreserveAll:
427     if (HasAVX)
428       return CSR_64_RT_AllRegs_AVX_RegMask;
429     return CSR_64_RT_AllRegs_RegMask;
430   case CallingConv::CXX_FAST_TLS:
431     if (Is64Bit)
432       return CSR_64_TLS_Darwin_RegMask;
433     break;
434   case CallingConv::Intel_OCL_BI: {
435     if (HasAVX512 && IsWin64)
436       return CSR_Win64_Intel_OCL_BI_AVX512_RegMask;
437     if (HasAVX512 && Is64Bit)
438       return CSR_64_Intel_OCL_BI_AVX512_RegMask;
439     if (HasAVX && IsWin64)
440       return CSR_Win64_Intel_OCL_BI_AVX_RegMask;
441     if (HasAVX && Is64Bit)
442       return CSR_64_Intel_OCL_BI_AVX_RegMask;
443     if (!HasAVX && !IsWin64 && Is64Bit)
444       return CSR_64_Intel_OCL_BI_RegMask;
445     break;
446   }
447   case CallingConv::HHVM:
448     return CSR_64_HHVM_RegMask;
449   case CallingConv::X86_RegCall:
450     if (Is64Bit) {
451       if (IsWin64) {
452         return (HasSSE ? CSR_Win64_RegCall_RegMask :
453                          CSR_Win64_RegCall_NoSSE_RegMask);
454       } else {
455         return (HasSSE ? CSR_SysV64_RegCall_RegMask :
456                          CSR_SysV64_RegCall_NoSSE_RegMask);
457       }
458     } else {
459       return (HasSSE ? CSR_32_RegCall_RegMask :
460                        CSR_32_RegCall_NoSSE_RegMask);
461     }
462   case CallingConv::CFGuard_Check:
463     assert(!Is64Bit && "CFGuard check mechanism only used on 32-bit X86");
464     return (HasSSE ? CSR_Win32_CFGuard_Check_RegMask
465                    : CSR_Win32_CFGuard_Check_NoSSE_RegMask);
466   case CallingConv::Cold:
467     if (Is64Bit)
468       return CSR_64_MostRegs_RegMask;
469     break;
470   case CallingConv::Win64:
471     return CSR_Win64_RegMask;
472   case CallingConv::X86_64_SysV:
473     return CSR_64_RegMask;
474   case CallingConv::X86_INTR:
475     if (Is64Bit) {
476       if (HasAVX512)
477         return CSR_64_AllRegs_AVX512_RegMask;
478       if (HasAVX)
479         return CSR_64_AllRegs_AVX_RegMask;
480       if (HasSSE)
481         return CSR_64_AllRegs_RegMask;
482       return CSR_64_AllRegs_NoSSE_RegMask;
483     } else {
484       if (HasAVX512)
485         return CSR_32_AllRegs_AVX512_RegMask;
486       if (HasAVX)
487         return CSR_32_AllRegs_AVX_RegMask;
488       if (HasSSE)
489         return CSR_32_AllRegs_SSE_RegMask;
490       return CSR_32_AllRegs_RegMask;
491     }
492   default:
493     break;
494   }
495 
496   // Unlike getCalleeSavedRegs(), we don't have MMI so we can't check
497   // callsEHReturn().
498   if (Is64Bit) {
499     const Function &F = MF.getFunction();
500     bool IsSwiftCC = Subtarget.getTargetLowering()->supportSwiftError() &&
501                      F.getAttributes().hasAttrSomewhere(Attribute::SwiftError);
502     if (IsSwiftCC)
503       return IsWin64 ? CSR_Win64_SwiftError_RegMask : CSR_64_SwiftError_RegMask;
504     return IsWin64 ? CSR_Win64_RegMask : CSR_64_RegMask;
505   }
506 
507   return CSR_32_RegMask;
508 }
509 
510 const uint32_t*
511 X86RegisterInfo::getNoPreservedMask() const {
512   return CSR_NoRegs_RegMask;
513 }
514 
515 const uint32_t *X86RegisterInfo::getDarwinTLSCallPreservedMask() const {
516   return CSR_64_TLS_Darwin_RegMask;
517 }
518 
519 BitVector X86RegisterInfo::getReservedRegs(const MachineFunction &MF) const {
520   BitVector Reserved(getNumRegs());
521   const X86FrameLowering *TFI = getFrameLowering(MF);
522 
523   // Set the floating point control register as reserved.
524   Reserved.set(X86::FPCW);
525 
526   // Set the floating point status register as reserved.
527   Reserved.set(X86::FPSW);
528 
529   // Set the SIMD floating point control register as reserved.
530   Reserved.set(X86::MXCSR);
531 
532   // Set the stack-pointer register and its aliases as reserved.
533   for (const MCPhysReg &SubReg : subregs_inclusive(X86::RSP))
534     Reserved.set(SubReg);
535 
536   // Set the Shadow Stack Pointer as reserved.
537   Reserved.set(X86::SSP);
538 
539   // Set the instruction pointer register and its aliases as reserved.
540   for (const MCPhysReg &SubReg : subregs_inclusive(X86::RIP))
541     Reserved.set(SubReg);
542 
543   // Set the frame-pointer register and its aliases as reserved if needed.
544   if (TFI->hasFP(MF)) {
545     for (const MCPhysReg &SubReg : subregs_inclusive(X86::RBP))
546       Reserved.set(SubReg);
547   }
548 
549   // Set the base-pointer register and its aliases as reserved if needed.
550   if (hasBasePointer(MF)) {
551     CallingConv::ID CC = MF.getFunction().getCallingConv();
552     const uint32_t *RegMask = getCallPreservedMask(MF, CC);
553     if (MachineOperand::clobbersPhysReg(RegMask, getBaseRegister()))
554       report_fatal_error(
555         "Stack realignment in presence of dynamic allocas is not supported with"
556         "this calling convention.");
557 
558     Register BasePtr = getX86SubSuperRegister(getBaseRegister(), 64);
559     for (const MCPhysReg &SubReg : subregs_inclusive(BasePtr))
560       Reserved.set(SubReg);
561   }
562 
563   // Mark the segment registers as reserved.
564   Reserved.set(X86::CS);
565   Reserved.set(X86::SS);
566   Reserved.set(X86::DS);
567   Reserved.set(X86::ES);
568   Reserved.set(X86::FS);
569   Reserved.set(X86::GS);
570 
571   // Mark the floating point stack registers as reserved.
572   for (unsigned n = 0; n != 8; ++n)
573     Reserved.set(X86::ST0 + n);
574 
575   // Reserve the registers that only exist in 64-bit mode.
576   if (!Is64Bit) {
577     // These 8-bit registers are part of the x86-64 extension even though their
578     // super-registers are old 32-bits.
579     Reserved.set(X86::SIL);
580     Reserved.set(X86::DIL);
581     Reserved.set(X86::BPL);
582     Reserved.set(X86::SPL);
583     Reserved.set(X86::SIH);
584     Reserved.set(X86::DIH);
585     Reserved.set(X86::BPH);
586     Reserved.set(X86::SPH);
587 
588     for (unsigned n = 0; n != 8; ++n) {
589       // R8, R9, ...
590       for (MCRegAliasIterator AI(X86::R8 + n, this, true); AI.isValid(); ++AI)
591         Reserved.set(*AI);
592 
593       // XMM8, XMM9, ...
594       for (MCRegAliasIterator AI(X86::XMM8 + n, this, true); AI.isValid(); ++AI)
595         Reserved.set(*AI);
596     }
597   }
598   if (!Is64Bit || !MF.getSubtarget<X86Subtarget>().hasAVX512()) {
599     for (unsigned n = 16; n != 32; ++n) {
600       for (MCRegAliasIterator AI(X86::XMM0 + n, this, true); AI.isValid(); ++AI)
601         Reserved.set(*AI);
602     }
603   }
604 
605   assert(checkAllSuperRegsMarked(Reserved,
606                                  {X86::SIL, X86::DIL, X86::BPL, X86::SPL,
607                                   X86::SIH, X86::DIH, X86::BPH, X86::SPH}));
608   return Reserved;
609 }
610 
611 void X86RegisterInfo::adjustStackMapLiveOutMask(uint32_t *Mask) const {
612   // Check if the EFLAGS register is marked as live-out. This shouldn't happen,
613   // because the calling convention defines the EFLAGS register as NOT
614   // preserved.
615   //
616   // Unfortunatelly the EFLAGS show up as live-out after branch folding. Adding
617   // an assert to track this and clear the register afterwards to avoid
618   // unnecessary crashes during release builds.
619   assert(!(Mask[X86::EFLAGS / 32] & (1U << (X86::EFLAGS % 32))) &&
620          "EFLAGS are not live-out from a patchpoint.");
621 
622   // Also clean other registers that don't need preserving (IP).
623   for (auto Reg : {X86::EFLAGS, X86::RIP, X86::EIP, X86::IP})
624     Mask[Reg / 32] &= ~(1U << (Reg % 32));
625 }
626 
627 //===----------------------------------------------------------------------===//
628 // Stack Frame Processing methods
629 //===----------------------------------------------------------------------===//
630 
631 static bool CantUseSP(const MachineFrameInfo &MFI) {
632   return MFI.hasVarSizedObjects() || MFI.hasOpaqueSPAdjustment();
633 }
634 
635 bool X86RegisterInfo::hasBasePointer(const MachineFunction &MF) const {
636    const MachineFrameInfo &MFI = MF.getFrameInfo();
637 
638    if (!EnableBasePointer)
639      return false;
640 
641    // When we need stack realignment, we can't address the stack from the frame
642    // pointer.  When we have dynamic allocas or stack-adjusting inline asm, we
643    // can't address variables from the stack pointer.  MS inline asm can
644    // reference locals while also adjusting the stack pointer.  When we can't
645    // use both the SP and the FP, we need a separate base pointer register.
646    bool CantUseFP = needsStackRealignment(MF);
647    return CantUseFP && CantUseSP(MFI);
648 }
649 
650 bool X86RegisterInfo::canRealignStack(const MachineFunction &MF) const {
651   if (!TargetRegisterInfo::canRealignStack(MF))
652     return false;
653 
654   const MachineFrameInfo &MFI = MF.getFrameInfo();
655   const MachineRegisterInfo *MRI = &MF.getRegInfo();
656 
657   // Stack realignment requires a frame pointer.  If we already started
658   // register allocation with frame pointer elimination, it is too late now.
659   if (!MRI->canReserveReg(FramePtr))
660     return false;
661 
662   // If a base pointer is necessary.  Check that it isn't too late to reserve
663   // it.
664   if (CantUseSP(MFI))
665     return MRI->canReserveReg(BasePtr);
666   return true;
667 }
668 
669 bool X86RegisterInfo::hasReservedSpillSlot(const MachineFunction &MF,
670                                            unsigned Reg, int &FrameIdx) const {
671   // Since X86 defines assignCalleeSavedSpillSlots which always return true
672   // this function neither used nor tested.
673   llvm_unreachable("Unused function on X86. Otherwise need a test case.");
674 }
675 
676 // tryOptimizeLEAtoMOV - helper function that tries to replace a LEA instruction
677 // of the form 'lea (%esp), %ebx' --> 'mov %esp, %ebx'.
678 // TODO: In this case we should be really trying first to entirely eliminate
679 // this instruction which is a plain copy.
680 static bool tryOptimizeLEAtoMOV(MachineBasicBlock::iterator II) {
681   MachineInstr &MI = *II;
682   unsigned Opc = II->getOpcode();
683   // Check if this is a LEA of the form 'lea (%esp), %ebx'
684   if ((Opc != X86::LEA32r && Opc != X86::LEA64r && Opc != X86::LEA64_32r) ||
685       MI.getOperand(2).getImm() != 1 ||
686       MI.getOperand(3).getReg() != X86::NoRegister ||
687       MI.getOperand(4).getImm() != 0 ||
688       MI.getOperand(5).getReg() != X86::NoRegister)
689     return false;
690   Register BasePtr = MI.getOperand(1).getReg();
691   // In X32 mode, ensure the base-pointer is a 32-bit operand, so the LEA will
692   // be replaced with a 32-bit operand MOV which will zero extend the upper
693   // 32-bits of the super register.
694   if (Opc == X86::LEA64_32r)
695     BasePtr = getX86SubSuperRegister(BasePtr, 32);
696   Register NewDestReg = MI.getOperand(0).getReg();
697   const X86InstrInfo *TII =
698       MI.getParent()->getParent()->getSubtarget<X86Subtarget>().getInstrInfo();
699   TII->copyPhysReg(*MI.getParent(), II, MI.getDebugLoc(), NewDestReg, BasePtr,
700                    MI.getOperand(1).isKill());
701   MI.eraseFromParent();
702   return true;
703 }
704 
705 static bool isFuncletReturnInstr(MachineInstr &MI) {
706   switch (MI.getOpcode()) {
707   case X86::CATCHRET:
708   case X86::CLEANUPRET:
709     return true;
710   default:
711     return false;
712   }
713   llvm_unreachable("impossible");
714 }
715 
716 void
717 X86RegisterInfo::eliminateFrameIndex(MachineBasicBlock::iterator II,
718                                      int SPAdj, unsigned FIOperandNum,
719                                      RegScavenger *RS) const {
720   MachineInstr &MI = *II;
721   MachineBasicBlock &MBB = *MI.getParent();
722   MachineFunction &MF = *MBB.getParent();
723   MachineBasicBlock::iterator MBBI = MBB.getFirstTerminator();
724   bool IsEHFuncletEpilogue = MBBI == MBB.end() ? false
725                                                : isFuncletReturnInstr(*MBBI);
726   const X86FrameLowering *TFI = getFrameLowering(MF);
727   int FrameIndex = MI.getOperand(FIOperandNum).getIndex();
728 
729   // Determine base register and offset.
730   int FIOffset;
731   unsigned BasePtr;
732   if (MI.isReturn()) {
733     assert((!needsStackRealignment(MF) ||
734            MF.getFrameInfo().isFixedObjectIndex(FrameIndex)) &&
735            "Return instruction can only reference SP relative frame objects");
736     FIOffset = TFI->getFrameIndexReferenceSP(MF, FrameIndex, BasePtr, 0);
737   } else if (TFI->Is64Bit && (MBB.isEHFuncletEntry() || IsEHFuncletEpilogue)) {
738     FIOffset = TFI->getWin64EHFrameIndexRef(MF, FrameIndex, BasePtr);
739   } else {
740     FIOffset = TFI->getFrameIndexReference(MF, FrameIndex, BasePtr);
741   }
742 
743   // LOCAL_ESCAPE uses a single offset, with no register. It only works in the
744   // simple FP case, and doesn't work with stack realignment. On 32-bit, the
745   // offset is from the traditional base pointer location.  On 64-bit, the
746   // offset is from the SP at the end of the prologue, not the FP location. This
747   // matches the behavior of llvm.frameaddress.
748   unsigned Opc = MI.getOpcode();
749   if (Opc == TargetOpcode::LOCAL_ESCAPE) {
750     MachineOperand &FI = MI.getOperand(FIOperandNum);
751     FI.ChangeToImmediate(FIOffset);
752     return;
753   }
754 
755   // For LEA64_32r when BasePtr is 32-bits (X32) we can use full-size 64-bit
756   // register as source operand, semantic is the same and destination is
757   // 32-bits. It saves one byte per lea in code since 0x67 prefix is avoided.
758   // Don't change BasePtr since it is used later for stack adjustment.
759   Register MachineBasePtr = BasePtr;
760   if (Opc == X86::LEA64_32r && X86::GR32RegClass.contains(BasePtr))
761     MachineBasePtr = getX86SubSuperRegister(BasePtr, 64);
762 
763   // This must be part of a four operand memory reference.  Replace the
764   // FrameIndex with base register.  Add an offset to the offset.
765   MI.getOperand(FIOperandNum).ChangeToRegister(MachineBasePtr, false);
766 
767   if (BasePtr == StackPtr)
768     FIOffset += SPAdj;
769 
770   // The frame index format for stackmaps and patchpoints is different from the
771   // X86 format. It only has a FI and an offset.
772   if (Opc == TargetOpcode::STACKMAP || Opc == TargetOpcode::PATCHPOINT) {
773     assert(BasePtr == FramePtr && "Expected the FP as base register");
774     int64_t Offset = MI.getOperand(FIOperandNum + 1).getImm() + FIOffset;
775     MI.getOperand(FIOperandNum + 1).ChangeToImmediate(Offset);
776     return;
777   }
778 
779   if (MI.getOperand(FIOperandNum+3).isImm()) {
780     // Offset is a 32-bit integer.
781     int Imm = (int)(MI.getOperand(FIOperandNum + 3).getImm());
782     int Offset = FIOffset + Imm;
783     assert((!Is64Bit || isInt<32>((long long)FIOffset + Imm)) &&
784            "Requesting 64-bit offset in 32-bit immediate!");
785     if (Offset != 0 || !tryOptimizeLEAtoMOV(II))
786       MI.getOperand(FIOperandNum + 3).ChangeToImmediate(Offset);
787   } else {
788     // Offset is symbolic. This is extremely rare.
789     uint64_t Offset = FIOffset +
790       (uint64_t)MI.getOperand(FIOperandNum+3).getOffset();
791     MI.getOperand(FIOperandNum + 3).setOffset(Offset);
792   }
793 }
794 
795 Register X86RegisterInfo::getFrameRegister(const MachineFunction &MF) const {
796   const X86FrameLowering *TFI = getFrameLowering(MF);
797   return TFI->hasFP(MF) ? FramePtr : StackPtr;
798 }
799 
800 unsigned
801 X86RegisterInfo::getPtrSizedFrameRegister(const MachineFunction &MF) const {
802   const X86Subtarget &Subtarget = MF.getSubtarget<X86Subtarget>();
803   Register FrameReg = getFrameRegister(MF);
804   if (Subtarget.isTarget64BitILP32())
805     FrameReg = getX86SubSuperRegister(FrameReg, 32);
806   return FrameReg;
807 }
808 
809 unsigned
810 X86RegisterInfo::getPtrSizedStackRegister(const MachineFunction &MF) const {
811   const X86Subtarget &Subtarget = MF.getSubtarget<X86Subtarget>();
812   Register StackReg = getStackRegister();
813   if (Subtarget.isTarget64BitILP32())
814     StackReg = getX86SubSuperRegister(StackReg, 32);
815   return StackReg;
816 }
817