diff options
Diffstat (limited to 'contrib/llvm-project/llvm/lib/Target/Sparc/SparcInstrInfo.cpp')
| -rw-r--r-- | contrib/llvm-project/llvm/lib/Target/Sparc/SparcInstrInfo.cpp | 509 | 
1 files changed, 509 insertions, 0 deletions
| diff --git a/contrib/llvm-project/llvm/lib/Target/Sparc/SparcInstrInfo.cpp b/contrib/llvm-project/llvm/lib/Target/Sparc/SparcInstrInfo.cpp new file mode 100644 index 000000000000..3d3d314a26bb --- /dev/null +++ b/contrib/llvm-project/llvm/lib/Target/Sparc/SparcInstrInfo.cpp @@ -0,0 +1,509 @@ +//===-- SparcInstrInfo.cpp - Sparc Instruction Information ----------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +// +// This file contains the Sparc implementation of the TargetInstrInfo class. +// +//===----------------------------------------------------------------------===// + +#include "SparcInstrInfo.h" +#include "Sparc.h" +#include "SparcMachineFunctionInfo.h" +#include "SparcSubtarget.h" +#include "llvm/ADT/STLExtras.h" +#include "llvm/ADT/SmallVector.h" +#include "llvm/CodeGen/MachineFrameInfo.h" +#include "llvm/CodeGen/MachineInstrBuilder.h" +#include "llvm/CodeGen/MachineMemOperand.h" +#include "llvm/CodeGen/MachineRegisterInfo.h" +#include "llvm/Support/ErrorHandling.h" +#include "llvm/Support/TargetRegistry.h" + +using namespace llvm; + +#define GET_INSTRINFO_CTOR_DTOR +#include "SparcGenInstrInfo.inc" + +// Pin the vtable to this file. +void SparcInstrInfo::anchor() {} + +SparcInstrInfo::SparcInstrInfo(SparcSubtarget &ST) +    : SparcGenInstrInfo(SP::ADJCALLSTACKDOWN, SP::ADJCALLSTACKUP), RI(), +      Subtarget(ST) {} + +/// isLoadFromStackSlot - If the specified machine instruction is a direct +/// load from a stack slot, return the virtual or physical register number of +/// the destination along with the FrameIndex of the loaded stack slot.  If +/// not, return 0.  This predicate must return 0 if the instruction has +/// any side effects other than loading from the stack slot. +unsigned SparcInstrInfo::isLoadFromStackSlot(const MachineInstr &MI, +                                             int &FrameIndex) const { +  if (MI.getOpcode() == SP::LDri || MI.getOpcode() == SP::LDXri || +      MI.getOpcode() == SP::LDFri || MI.getOpcode() == SP::LDDFri || +      MI.getOpcode() == SP::LDQFri) { +    if (MI.getOperand(1).isFI() && MI.getOperand(2).isImm() && +        MI.getOperand(2).getImm() == 0) { +      FrameIndex = MI.getOperand(1).getIndex(); +      return MI.getOperand(0).getReg(); +    } +  } +  return 0; +} + +/// isStoreToStackSlot - If the specified machine instruction is a direct +/// store to a stack slot, return the virtual or physical register number of +/// the source reg along with the FrameIndex of the loaded stack slot.  If +/// not, return 0.  This predicate must return 0 if the instruction has +/// any side effects other than storing to the stack slot. +unsigned SparcInstrInfo::isStoreToStackSlot(const MachineInstr &MI, +                                            int &FrameIndex) const { +  if (MI.getOpcode() == SP::STri || MI.getOpcode() == SP::STXri || +      MI.getOpcode() == SP::STFri || MI.getOpcode() == SP::STDFri || +      MI.getOpcode() == SP::STQFri) { +    if (MI.getOperand(0).isFI() && MI.getOperand(1).isImm() && +        MI.getOperand(1).getImm() == 0) { +      FrameIndex = MI.getOperand(0).getIndex(); +      return MI.getOperand(2).getReg(); +    } +  } +  return 0; +} + +static bool IsIntegerCC(unsigned CC) +{ +  return  (CC <= SPCC::ICC_VC); +} + +static SPCC::CondCodes GetOppositeBranchCondition(SPCC::CondCodes CC) +{ +  switch(CC) { +  case SPCC::ICC_A:    return SPCC::ICC_N; +  case SPCC::ICC_N:    return SPCC::ICC_A; +  case SPCC::ICC_NE:   return SPCC::ICC_E; +  case SPCC::ICC_E:    return SPCC::ICC_NE; +  case SPCC::ICC_G:    return SPCC::ICC_LE; +  case SPCC::ICC_LE:   return SPCC::ICC_G; +  case SPCC::ICC_GE:   return SPCC::ICC_L; +  case SPCC::ICC_L:    return SPCC::ICC_GE; +  case SPCC::ICC_GU:   return SPCC::ICC_LEU; +  case SPCC::ICC_LEU:  return SPCC::ICC_GU; +  case SPCC::ICC_CC:   return SPCC::ICC_CS; +  case SPCC::ICC_CS:   return SPCC::ICC_CC; +  case SPCC::ICC_POS:  return SPCC::ICC_NEG; +  case SPCC::ICC_NEG:  return SPCC::ICC_POS; +  case SPCC::ICC_VC:   return SPCC::ICC_VS; +  case SPCC::ICC_VS:   return SPCC::ICC_VC; + +  case SPCC::FCC_A:    return SPCC::FCC_N; +  case SPCC::FCC_N:    return SPCC::FCC_A; +  case SPCC::FCC_U:    return SPCC::FCC_O; +  case SPCC::FCC_O:    return SPCC::FCC_U; +  case SPCC::FCC_G:    return SPCC::FCC_ULE; +  case SPCC::FCC_LE:   return SPCC::FCC_UG; +  case SPCC::FCC_UG:   return SPCC::FCC_LE; +  case SPCC::FCC_ULE:  return SPCC::FCC_G; +  case SPCC::FCC_L:    return SPCC::FCC_UGE; +  case SPCC::FCC_GE:   return SPCC::FCC_UL; +  case SPCC::FCC_UL:   return SPCC::FCC_GE; +  case SPCC::FCC_UGE:  return SPCC::FCC_L; +  case SPCC::FCC_LG:   return SPCC::FCC_UE; +  case SPCC::FCC_UE:   return SPCC::FCC_LG; +  case SPCC::FCC_NE:   return SPCC::FCC_E; +  case SPCC::FCC_E:    return SPCC::FCC_NE; + +  case SPCC::CPCC_A:   return SPCC::CPCC_N; +  case SPCC::CPCC_N:   return SPCC::CPCC_A; +  case SPCC::CPCC_3:   LLVM_FALLTHROUGH; +  case SPCC::CPCC_2:   LLVM_FALLTHROUGH; +  case SPCC::CPCC_23:  LLVM_FALLTHROUGH; +  case SPCC::CPCC_1:   LLVM_FALLTHROUGH; +  case SPCC::CPCC_13:  LLVM_FALLTHROUGH; +  case SPCC::CPCC_12:  LLVM_FALLTHROUGH; +  case SPCC::CPCC_123: LLVM_FALLTHROUGH; +  case SPCC::CPCC_0:   LLVM_FALLTHROUGH; +  case SPCC::CPCC_03:  LLVM_FALLTHROUGH; +  case SPCC::CPCC_02:  LLVM_FALLTHROUGH; +  case SPCC::CPCC_023: LLVM_FALLTHROUGH; +  case SPCC::CPCC_01:  LLVM_FALLTHROUGH; +  case SPCC::CPCC_013: LLVM_FALLTHROUGH; +  case SPCC::CPCC_012: +      // "Opposite" code is not meaningful, as we don't know +      // what the CoProc condition means here. The cond-code will +      // only be used in inline assembler, so this code should +      // not be reached in a normal compilation pass. +      llvm_unreachable("Meaningless inversion of co-processor cond code"); +  } +  llvm_unreachable("Invalid cond code"); +} + +static bool isUncondBranchOpcode(int Opc) { return Opc == SP::BA; } + +static bool isCondBranchOpcode(int Opc) { +  return Opc == SP::FBCOND || Opc == SP::BCOND; +} + +static bool isIndirectBranchOpcode(int Opc) { +  return Opc == SP::BINDrr || Opc == SP::BINDri; +} + +static void parseCondBranch(MachineInstr *LastInst, MachineBasicBlock *&Target, +                            SmallVectorImpl<MachineOperand> &Cond) { +  Cond.push_back(MachineOperand::CreateImm(LastInst->getOperand(1).getImm())); +  Target = LastInst->getOperand(0).getMBB(); +} + +bool SparcInstrInfo::analyzeBranch(MachineBasicBlock &MBB, +                                   MachineBasicBlock *&TBB, +                                   MachineBasicBlock *&FBB, +                                   SmallVectorImpl<MachineOperand> &Cond, +                                   bool AllowModify) const { +  MachineBasicBlock::iterator I = MBB.getLastNonDebugInstr(); +  if (I == MBB.end()) +    return false; + +  if (!isUnpredicatedTerminator(*I)) +    return false; + +  // Get the last instruction in the block. +  MachineInstr *LastInst = &*I; +  unsigned LastOpc = LastInst->getOpcode(); + +  // If there is only one terminator instruction, process it. +  if (I == MBB.begin() || !isUnpredicatedTerminator(*--I)) { +    if (isUncondBranchOpcode(LastOpc)) { +      TBB = LastInst->getOperand(0).getMBB(); +      return false; +    } +    if (isCondBranchOpcode(LastOpc)) { +      // Block ends with fall-through condbranch. +      parseCondBranch(LastInst, TBB, Cond); +      return false; +    } +    return true; // Can't handle indirect branch. +  } + +  // Get the instruction before it if it is a terminator. +  MachineInstr *SecondLastInst = &*I; +  unsigned SecondLastOpc = SecondLastInst->getOpcode(); + +  // If AllowModify is true and the block ends with two or more unconditional +  // branches, delete all but the first unconditional branch. +  if (AllowModify && isUncondBranchOpcode(LastOpc)) { +    while (isUncondBranchOpcode(SecondLastOpc)) { +      LastInst->eraseFromParent(); +      LastInst = SecondLastInst; +      LastOpc = LastInst->getOpcode(); +      if (I == MBB.begin() || !isUnpredicatedTerminator(*--I)) { +        // Return now the only terminator is an unconditional branch. +        TBB = LastInst->getOperand(0).getMBB(); +        return false; +      } else { +        SecondLastInst = &*I; +        SecondLastOpc = SecondLastInst->getOpcode(); +      } +    } +  } + +  // If there are three terminators, we don't know what sort of block this is. +  if (SecondLastInst && I != MBB.begin() && isUnpredicatedTerminator(*--I)) +    return true; + +  // If the block ends with a B and a Bcc, handle it. +  if (isCondBranchOpcode(SecondLastOpc) && isUncondBranchOpcode(LastOpc)) { +    parseCondBranch(SecondLastInst, TBB, Cond); +    FBB = LastInst->getOperand(0).getMBB(); +    return false; +  } + +  // If the block ends with two unconditional branches, handle it.  The second +  // one is not executed. +  if (isUncondBranchOpcode(SecondLastOpc) && isUncondBranchOpcode(LastOpc)) { +    TBB = SecondLastInst->getOperand(0).getMBB(); +    return false; +  } + +  // ...likewise if it ends with an indirect branch followed by an unconditional +  // branch. +  if (isIndirectBranchOpcode(SecondLastOpc) && isUncondBranchOpcode(LastOpc)) { +    I = LastInst; +    if (AllowModify) +      I->eraseFromParent(); +    return true; +  } + +  // Otherwise, can't handle this. +  return true; +} + +unsigned SparcInstrInfo::insertBranch(MachineBasicBlock &MBB, +                                      MachineBasicBlock *TBB, +                                      MachineBasicBlock *FBB, +                                      ArrayRef<MachineOperand> Cond, +                                      const DebugLoc &DL, +                                      int *BytesAdded) const { +  assert(TBB && "insertBranch must not be told to insert a fallthrough"); +  assert((Cond.size() == 1 || Cond.size() == 0) && +         "Sparc branch conditions should have one component!"); +  assert(!BytesAdded && "code size not handled"); + +  if (Cond.empty()) { +    assert(!FBB && "Unconditional branch with multiple successors!"); +    BuildMI(&MBB, DL, get(SP::BA)).addMBB(TBB); +    return 1; +  } + +  // Conditional branch +  unsigned CC = Cond[0].getImm(); + +  if (IsIntegerCC(CC)) +    BuildMI(&MBB, DL, get(SP::BCOND)).addMBB(TBB).addImm(CC); +  else +    BuildMI(&MBB, DL, get(SP::FBCOND)).addMBB(TBB).addImm(CC); +  if (!FBB) +    return 1; + +  BuildMI(&MBB, DL, get(SP::BA)).addMBB(FBB); +  return 2; +} + +unsigned SparcInstrInfo::removeBranch(MachineBasicBlock &MBB, +                                      int *BytesRemoved) const { +  assert(!BytesRemoved && "code size not handled"); + +  MachineBasicBlock::iterator I = MBB.end(); +  unsigned Count = 0; +  while (I != MBB.begin()) { +    --I; + +    if (I->isDebugInstr()) +      continue; + +    if (I->getOpcode() != SP::BA +        && I->getOpcode() != SP::BCOND +        && I->getOpcode() != SP::FBCOND) +      break; // Not a branch + +    I->eraseFromParent(); +    I = MBB.end(); +    ++Count; +  } +  return Count; +} + +bool SparcInstrInfo::reverseBranchCondition( +    SmallVectorImpl<MachineOperand> &Cond) const { +  assert(Cond.size() == 1); +  SPCC::CondCodes CC = static_cast<SPCC::CondCodes>(Cond[0].getImm()); +  Cond[0].setImm(GetOppositeBranchCondition(CC)); +  return false; +} + +void SparcInstrInfo::copyPhysReg(MachineBasicBlock &MBB, +                                 MachineBasicBlock::iterator I, +                                 const DebugLoc &DL, unsigned DestReg, +                                 unsigned SrcReg, bool KillSrc) const { +  unsigned numSubRegs = 0; +  unsigned movOpc     = 0; +  const unsigned *subRegIdx = nullptr; +  bool ExtraG0 = false; + +  const unsigned DW_SubRegsIdx[]  = { SP::sub_even, SP::sub_odd }; +  const unsigned DFP_FP_SubRegsIdx[]  = { SP::sub_even, SP::sub_odd }; +  const unsigned QFP_DFP_SubRegsIdx[] = { SP::sub_even64, SP::sub_odd64 }; +  const unsigned QFP_FP_SubRegsIdx[]  = { SP::sub_even, SP::sub_odd, +                                          SP::sub_odd64_then_sub_even, +                                          SP::sub_odd64_then_sub_odd }; + +  if (SP::IntRegsRegClass.contains(DestReg, SrcReg)) +    BuildMI(MBB, I, DL, get(SP::ORrr), DestReg).addReg(SP::G0) +      .addReg(SrcReg, getKillRegState(KillSrc)); +  else if (SP::IntPairRegClass.contains(DestReg, SrcReg)) { +    subRegIdx  = DW_SubRegsIdx; +    numSubRegs = 2; +    movOpc     = SP::ORrr; +    ExtraG0 = true; +  } else if (SP::FPRegsRegClass.contains(DestReg, SrcReg)) +    BuildMI(MBB, I, DL, get(SP::FMOVS), DestReg) +      .addReg(SrcReg, getKillRegState(KillSrc)); +  else if (SP::DFPRegsRegClass.contains(DestReg, SrcReg)) { +    if (Subtarget.isV9()) { +      BuildMI(MBB, I, DL, get(SP::FMOVD), DestReg) +        .addReg(SrcReg, getKillRegState(KillSrc)); +    } else { +      // Use two FMOVS instructions. +      subRegIdx  = DFP_FP_SubRegsIdx; +      numSubRegs = 2; +      movOpc     = SP::FMOVS; +    } +  } else if (SP::QFPRegsRegClass.contains(DestReg, SrcReg)) { +    if (Subtarget.isV9()) { +      if (Subtarget.hasHardQuad()) { +        BuildMI(MBB, I, DL, get(SP::FMOVQ), DestReg) +          .addReg(SrcReg, getKillRegState(KillSrc)); +      } else { +        // Use two FMOVD instructions. +        subRegIdx  = QFP_DFP_SubRegsIdx; +        numSubRegs = 2; +        movOpc     = SP::FMOVD; +      } +    } else { +      // Use four FMOVS instructions. +      subRegIdx  = QFP_FP_SubRegsIdx; +      numSubRegs = 4; +      movOpc     = SP::FMOVS; +    } +  } else if (SP::ASRRegsRegClass.contains(DestReg) && +             SP::IntRegsRegClass.contains(SrcReg)) { +    BuildMI(MBB, I, DL, get(SP::WRASRrr), DestReg) +        .addReg(SP::G0) +        .addReg(SrcReg, getKillRegState(KillSrc)); +  } else if (SP::IntRegsRegClass.contains(DestReg) && +             SP::ASRRegsRegClass.contains(SrcReg)) { +    BuildMI(MBB, I, DL, get(SP::RDASR), DestReg) +        .addReg(SrcReg, getKillRegState(KillSrc)); +  } else +    llvm_unreachable("Impossible reg-to-reg copy"); + +  if (numSubRegs == 0 || subRegIdx == nullptr || movOpc == 0) +    return; + +  const TargetRegisterInfo *TRI = &getRegisterInfo(); +  MachineInstr *MovMI = nullptr; + +  for (unsigned i = 0; i != numSubRegs; ++i) { +    Register Dst = TRI->getSubReg(DestReg, subRegIdx[i]); +    Register Src = TRI->getSubReg(SrcReg, subRegIdx[i]); +    assert(Dst && Src && "Bad sub-register"); + +    MachineInstrBuilder MIB = BuildMI(MBB, I, DL, get(movOpc), Dst); +    if (ExtraG0) +      MIB.addReg(SP::G0); +    MIB.addReg(Src); +    MovMI = MIB.getInstr(); +  } +  // Add implicit super-register defs and kills to the last MovMI. +  MovMI->addRegisterDefined(DestReg, TRI); +  if (KillSrc) +    MovMI->addRegisterKilled(SrcReg, TRI); +} + +void SparcInstrInfo:: +storeRegToStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, +                    unsigned SrcReg, bool isKill, int FI, +                    const TargetRegisterClass *RC, +                    const TargetRegisterInfo *TRI) const { +  DebugLoc DL; +  if (I != MBB.end()) DL = I->getDebugLoc(); + +  MachineFunction *MF = MBB.getParent(); +  const MachineFrameInfo &MFI = MF->getFrameInfo(); +  MachineMemOperand *MMO = MF->getMachineMemOperand( +      MachinePointerInfo::getFixedStack(*MF, FI), MachineMemOperand::MOStore, +      MFI.getObjectSize(FI), MFI.getObjectAlignment(FI)); + +  // On the order of operands here: think "[FrameIdx + 0] = SrcReg". +  if (RC == &SP::I64RegsRegClass) +    BuildMI(MBB, I, DL, get(SP::STXri)).addFrameIndex(FI).addImm(0) +      .addReg(SrcReg, getKillRegState(isKill)).addMemOperand(MMO); +  else if (RC == &SP::IntRegsRegClass) +    BuildMI(MBB, I, DL, get(SP::STri)).addFrameIndex(FI).addImm(0) +      .addReg(SrcReg, getKillRegState(isKill)).addMemOperand(MMO); +  else if (RC == &SP::IntPairRegClass) +    BuildMI(MBB, I, DL, get(SP::STDri)).addFrameIndex(FI).addImm(0) +      .addReg(SrcReg, getKillRegState(isKill)).addMemOperand(MMO); +  else if (RC == &SP::FPRegsRegClass) +    BuildMI(MBB, I, DL, get(SP::STFri)).addFrameIndex(FI).addImm(0) +      .addReg(SrcReg,  getKillRegState(isKill)).addMemOperand(MMO); +  else if (SP::DFPRegsRegClass.hasSubClassEq(RC)) +    BuildMI(MBB, I, DL, get(SP::STDFri)).addFrameIndex(FI).addImm(0) +      .addReg(SrcReg,  getKillRegState(isKill)).addMemOperand(MMO); +  else if (SP::QFPRegsRegClass.hasSubClassEq(RC)) +    // Use STQFri irrespective of its legality. If STQ is not legal, it will be +    // lowered into two STDs in eliminateFrameIndex. +    BuildMI(MBB, I, DL, get(SP::STQFri)).addFrameIndex(FI).addImm(0) +      .addReg(SrcReg,  getKillRegState(isKill)).addMemOperand(MMO); +  else +    llvm_unreachable("Can't store this register to stack slot"); +} + +void SparcInstrInfo:: +loadRegFromStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, +                     unsigned DestReg, int FI, +                     const TargetRegisterClass *RC, +                     const TargetRegisterInfo *TRI) const { +  DebugLoc DL; +  if (I != MBB.end()) DL = I->getDebugLoc(); + +  MachineFunction *MF = MBB.getParent(); +  const MachineFrameInfo &MFI = MF->getFrameInfo(); +  MachineMemOperand *MMO = MF->getMachineMemOperand( +      MachinePointerInfo::getFixedStack(*MF, FI), MachineMemOperand::MOLoad, +      MFI.getObjectSize(FI), MFI.getObjectAlignment(FI)); + +  if (RC == &SP::I64RegsRegClass) +    BuildMI(MBB, I, DL, get(SP::LDXri), DestReg).addFrameIndex(FI).addImm(0) +      .addMemOperand(MMO); +  else if (RC == &SP::IntRegsRegClass) +    BuildMI(MBB, I, DL, get(SP::LDri), DestReg).addFrameIndex(FI).addImm(0) +      .addMemOperand(MMO); +  else if (RC == &SP::IntPairRegClass) +    BuildMI(MBB, I, DL, get(SP::LDDri), DestReg).addFrameIndex(FI).addImm(0) +      .addMemOperand(MMO); +  else if (RC == &SP::FPRegsRegClass) +    BuildMI(MBB, I, DL, get(SP::LDFri), DestReg).addFrameIndex(FI).addImm(0) +      .addMemOperand(MMO); +  else if (SP::DFPRegsRegClass.hasSubClassEq(RC)) +    BuildMI(MBB, I, DL, get(SP::LDDFri), DestReg).addFrameIndex(FI).addImm(0) +      .addMemOperand(MMO); +  else if (SP::QFPRegsRegClass.hasSubClassEq(RC)) +    // Use LDQFri irrespective of its legality. If LDQ is not legal, it will be +    // lowered into two LDDs in eliminateFrameIndex. +    BuildMI(MBB, I, DL, get(SP::LDQFri), DestReg).addFrameIndex(FI).addImm(0) +      .addMemOperand(MMO); +  else +    llvm_unreachable("Can't load this register from stack slot"); +} + +unsigned SparcInstrInfo::getGlobalBaseReg(MachineFunction *MF) const +{ +  SparcMachineFunctionInfo *SparcFI = MF->getInfo<SparcMachineFunctionInfo>(); +  unsigned GlobalBaseReg = SparcFI->getGlobalBaseReg(); +  if (GlobalBaseReg != 0) +    return GlobalBaseReg; + +  // Insert the set of GlobalBaseReg into the first MBB of the function +  MachineBasicBlock &FirstMBB = MF->front(); +  MachineBasicBlock::iterator MBBI = FirstMBB.begin(); +  MachineRegisterInfo &RegInfo = MF->getRegInfo(); + +  const TargetRegisterClass *PtrRC = +    Subtarget.is64Bit() ? &SP::I64RegsRegClass : &SP::IntRegsRegClass; +  GlobalBaseReg = RegInfo.createVirtualRegister(PtrRC); + +  DebugLoc dl; + +  BuildMI(FirstMBB, MBBI, dl, get(SP::GETPCX), GlobalBaseReg); +  SparcFI->setGlobalBaseReg(GlobalBaseReg); +  return GlobalBaseReg; +} + +bool SparcInstrInfo::expandPostRAPseudo(MachineInstr &MI) const { +  switch (MI.getOpcode()) { +  case TargetOpcode::LOAD_STACK_GUARD: { +    assert(Subtarget.isTargetLinux() && +           "Only Linux target is expected to contain LOAD_STACK_GUARD"); +    // offsetof(tcbhead_t, stack_guard) from sysdeps/sparc/nptl/tls.h in glibc. +    const int64_t Offset = Subtarget.is64Bit() ? 0x28 : 0x14; +    MI.setDesc(get(Subtarget.is64Bit() ? SP::LDXri : SP::LDri)); +    MachineInstrBuilder(*MI.getParent()->getParent(), MI) +        .addReg(SP::G7) +        .addImm(Offset); +    return true; +  } +  } +  return false; +} | 
