Subversion Repositories QNX 8.QNX8 LLVM/Clang compiler suite

Rev

Blame | Last modification | View Log | Download | RSS feed

  1. //===-- llvm/CodeGen/GlobalISel/CombinerHelper.h --------------*- C++ -*-===//
  2. //
  3. // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
  4. // See https://llvm.org/LICENSE.txt for license information.
  5. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
  6. //
  7. //===--------------------------------------------------------------------===//
  8. /// \file
  9. /// This contains common combine transformations that may be used in a combine
  10. /// pass,or by the target elsewhere.
  11. /// Targets can pick individual opcode transformations from the helper or use
  12. /// tryCombine which invokes all transformations. All of the transformations
  13. /// return true if the MachineInstruction changed and false otherwise.
  14. ///
  15. //===--------------------------------------------------------------------===//
  16.  
  17. #ifndef LLVM_CODEGEN_GLOBALISEL_COMBINERHELPER_H
  18. #define LLVM_CODEGEN_GLOBALISEL_COMBINERHELPER_H
  19.  
  20. #include "llvm/ADT/DenseMap.h"
  21. #include "llvm/ADT/SmallVector.h"
  22. #include "llvm/CodeGen/Register.h"
  23. #include "llvm/Support/LowLevelTypeImpl.h"
  24. #include "llvm/IR/InstrTypes.h"
  25. #include <functional>
  26.  
  27. namespace llvm {
  28.  
  29. class GISelChangeObserver;
  30. class APFloat;
  31. class APInt;
  32. class GPtrAdd;
  33. class GStore;
  34. class GZExtLoad;
  35. class MachineIRBuilder;
  36. class MachineInstrBuilder;
  37. class MachineRegisterInfo;
  38. class MachineInstr;
  39. class MachineOperand;
  40. class GISelKnownBits;
  41. class MachineDominatorTree;
  42. class LegalizerInfo;
  43. struct LegalityQuery;
  44. class RegisterBank;
  45. class RegisterBankInfo;
  46. class TargetLowering;
  47. class TargetRegisterInfo;
  48.  
  49. struct PreferredTuple {
  50.   LLT Ty;                // The result type of the extend.
  51.   unsigned ExtendOpcode; // G_ANYEXT/G_SEXT/G_ZEXT
  52.   MachineInstr *MI;
  53. };
  54.  
  55. struct IndexedLoadStoreMatchInfo {
  56.   Register Addr;
  57.   Register Base;
  58.   Register Offset;
  59.   bool IsPre;
  60. };
  61.  
  62. struct PtrAddChain {
  63.   int64_t Imm;
  64.   Register Base;
  65.   const RegisterBank *Bank;
  66. };
  67.  
  68. struct RegisterImmPair {
  69.   Register Reg;
  70.   int64_t Imm;
  71. };
  72.  
  73. struct ShiftOfShiftedLogic {
  74.   MachineInstr *Logic;
  75.   MachineInstr *Shift2;
  76.   Register LogicNonShiftReg;
  77.   uint64_t ValSum;
  78. };
  79.  
  80. using BuildFnTy = std::function<void(MachineIRBuilder &)>;
  81.  
  82. struct MergeTruncStoresInfo {
  83.   SmallVector<GStore *> FoundStores;
  84.   GStore *LowestIdxStore = nullptr;
  85.   Register WideSrcVal;
  86.   bool NeedBSwap = false;
  87.   bool NeedRotate = false;
  88. };
  89.  
  90. using OperandBuildSteps =
  91.     SmallVector<std::function<void(MachineInstrBuilder &)>, 4>;
  92. struct InstructionBuildSteps {
  93.   unsigned Opcode = 0;          /// The opcode for the produced instruction.
  94.   OperandBuildSteps OperandFns; /// Operands to be added to the instruction.
  95.   InstructionBuildSteps() = default;
  96.   InstructionBuildSteps(unsigned Opcode, const OperandBuildSteps &OperandFns)
  97.       : Opcode(Opcode), OperandFns(OperandFns) {}
  98. };
  99.  
  100. struct InstructionStepsMatchInfo {
  101.   /// Describes instructions to be built during a combine.
  102.   SmallVector<InstructionBuildSteps, 2> InstrsToBuild;
  103.   InstructionStepsMatchInfo() = default;
  104.   InstructionStepsMatchInfo(
  105.       std::initializer_list<InstructionBuildSteps> InstrsToBuild)
  106.       : InstrsToBuild(InstrsToBuild) {}
  107. };
  108.  
  109. class CombinerHelper {
  110. protected:
  111.   MachineIRBuilder &Builder;
  112.   MachineRegisterInfo &MRI;
  113.   GISelChangeObserver &Observer;
  114.   GISelKnownBits *KB;
  115.   MachineDominatorTree *MDT;
  116.   bool IsPreLegalize;
  117.   const LegalizerInfo *LI;
  118.   const RegisterBankInfo *RBI;
  119.   const TargetRegisterInfo *TRI;
  120.  
  121. public:
  122.   CombinerHelper(GISelChangeObserver &Observer, MachineIRBuilder &B,
  123.                  bool IsPreLegalize,
  124.                  GISelKnownBits *KB = nullptr,
  125.                  MachineDominatorTree *MDT = nullptr,
  126.                  const LegalizerInfo *LI = nullptr);
  127.  
  128.   GISelKnownBits *getKnownBits() const {
  129.     return KB;
  130.   }
  131.  
  132.   MachineIRBuilder &getBuilder() const {
  133.     return Builder;
  134.   }
  135.  
  136.   const TargetLowering &getTargetLowering() const;
  137.  
  138.   /// \returns true if the combiner is running pre-legalization.
  139.   bool isPreLegalize() const;
  140.  
  141.   /// \returns true if \p Query is legal on the target.
  142.   bool isLegal(const LegalityQuery &Query) const;
  143.  
  144.   /// \return true if the combine is running prior to legalization, or if \p
  145.   /// Query is legal on the target.
  146.   bool isLegalOrBeforeLegalizer(const LegalityQuery &Query) const;
  147.  
  148.   /// \return true if the combine is running prior to legalization, or if \p Ty
  149.   /// is a legal integer constant type on the target.
  150.   bool isConstantLegalOrBeforeLegalizer(const LLT Ty) const;
  151.  
  152.   /// MachineRegisterInfo::replaceRegWith() and inform the observer of the changes
  153.   void replaceRegWith(MachineRegisterInfo &MRI, Register FromReg, Register ToReg) const;
  154.  
  155.   /// Replace a single register operand with a new register and inform the
  156.   /// observer of the changes.
  157.   void replaceRegOpWith(MachineRegisterInfo &MRI, MachineOperand &FromRegOp,
  158.                         Register ToReg) const;
  159.  
  160.   /// Replace the opcode in instruction with a new opcode and inform the
  161.   /// observer of the changes.
  162.   void replaceOpcodeWith(MachineInstr &FromMI, unsigned ToOpcode) const;
  163.  
  164.   /// Get the register bank of \p Reg.
  165.   /// If Reg has not been assigned a register, a register class,
  166.   /// or a register bank, then this returns nullptr.
  167.   ///
  168.   /// \pre Reg.isValid()
  169.   const RegisterBank *getRegBank(Register Reg) const;
  170.  
  171.   /// Set the register bank of \p Reg.
  172.   /// Does nothing if the RegBank is null.
  173.   /// This is the counterpart to getRegBank.
  174.   void setRegBank(Register Reg, const RegisterBank *RegBank);
  175.  
  176.   /// If \p MI is COPY, try to combine it.
  177.   /// Returns true if MI changed.
  178.   bool tryCombineCopy(MachineInstr &MI);
  179.   bool matchCombineCopy(MachineInstr &MI);
  180.   void applyCombineCopy(MachineInstr &MI);
  181.  
  182.   /// Returns true if \p DefMI precedes \p UseMI or they are the same
  183.   /// instruction. Both must be in the same basic block.
  184.   bool isPredecessor(const MachineInstr &DefMI, const MachineInstr &UseMI);
  185.  
  186.   /// Returns true if \p DefMI dominates \p UseMI. By definition an
  187.   /// instruction dominates itself.
  188.   ///
  189.   /// If we haven't been provided with a MachineDominatorTree during
  190.   /// construction, this function returns a conservative result that tracks just
  191.   /// a single basic block.
  192.   bool dominates(const MachineInstr &DefMI, const MachineInstr &UseMI);
  193.  
  194.   /// If \p MI is extend that consumes the result of a load, try to combine it.
  195.   /// Returns true if MI changed.
  196.   bool tryCombineExtendingLoads(MachineInstr &MI);
  197.   bool matchCombineExtendingLoads(MachineInstr &MI, PreferredTuple &MatchInfo);
  198.   void applyCombineExtendingLoads(MachineInstr &MI, PreferredTuple &MatchInfo);
  199.  
  200.   /// Match (and (load x), mask) -> zextload x
  201.   bool matchCombineLoadWithAndMask(MachineInstr &MI, BuildFnTy &MatchInfo);
  202.  
  203.   /// Combine \p MI into a pre-indexed or post-indexed load/store operation if
  204.   /// legal and the surrounding code makes it useful.
  205.   bool tryCombineIndexedLoadStore(MachineInstr &MI);
  206.   bool matchCombineIndexedLoadStore(MachineInstr &MI, IndexedLoadStoreMatchInfo &MatchInfo);
  207.   void applyCombineIndexedLoadStore(MachineInstr &MI, IndexedLoadStoreMatchInfo &MatchInfo);
  208.  
  209.   bool matchSextTruncSextLoad(MachineInstr &MI);
  210.   void applySextTruncSextLoad(MachineInstr &MI);
  211.  
  212.   /// Match sext_inreg(load p), imm -> sextload p
  213.   bool matchSextInRegOfLoad(MachineInstr &MI, std::tuple<Register, unsigned> &MatchInfo);
  214.   void applySextInRegOfLoad(MachineInstr &MI, std::tuple<Register, unsigned> &MatchInfo);
  215.  
  216.   /// Try to combine G_[SU]DIV and G_[SU]REM into a single G_[SU]DIVREM
  217.   /// when their source operands are identical.
  218.   bool matchCombineDivRem(MachineInstr &MI, MachineInstr *&OtherMI);
  219.   void applyCombineDivRem(MachineInstr &MI, MachineInstr *&OtherMI);
  220.  
  221.   /// If a brcond's true block is not the fallthrough, make it so by inverting
  222.   /// the condition and swapping operands.
  223.   bool matchOptBrCondByInvertingCond(MachineInstr &MI, MachineInstr *&BrCond);
  224.   void applyOptBrCondByInvertingCond(MachineInstr &MI, MachineInstr *&BrCond);
  225.  
  226.   /// If \p MI is G_CONCAT_VECTORS, try to combine it.
  227.   /// Returns true if MI changed.
  228.   /// Right now, we support:
  229.   /// - concat_vector(undef, undef) => undef
  230.   /// - concat_vector(build_vector(A, B), build_vector(C, D)) =>
  231.   ///   build_vector(A, B, C, D)
  232.   ///
  233.   /// \pre MI.getOpcode() == G_CONCAT_VECTORS.
  234.   bool tryCombineConcatVectors(MachineInstr &MI);
  235.   /// Check if the G_CONCAT_VECTORS \p MI is undef or if it
  236.   /// can be flattened into a build_vector.
  237.   /// In the first case \p IsUndef will be true.
  238.   /// In the second case \p Ops will contain the operands needed
  239.   /// to produce the flattened build_vector.
  240.   ///
  241.   /// \pre MI.getOpcode() == G_CONCAT_VECTORS.
  242.   bool matchCombineConcatVectors(MachineInstr &MI, bool &IsUndef,
  243.                                  SmallVectorImpl<Register> &Ops);
  244.   /// Replace \p MI with a flattened build_vector with \p Ops or an
  245.   /// implicit_def if IsUndef is true.
  246.   void applyCombineConcatVectors(MachineInstr &MI, bool IsUndef,
  247.                                  const ArrayRef<Register> Ops);
  248.  
  249.   /// Try to combine G_SHUFFLE_VECTOR into G_CONCAT_VECTORS.
  250.   /// Returns true if MI changed.
  251.   ///
  252.   /// \pre MI.getOpcode() == G_SHUFFLE_VECTOR.
  253.   bool tryCombineShuffleVector(MachineInstr &MI);
  254.   /// Check if the G_SHUFFLE_VECTOR \p MI can be replaced by a
  255.   /// concat_vectors.
  256.   /// \p Ops will contain the operands needed to produce the flattened
  257.   /// concat_vectors.
  258.   ///
  259.   /// \pre MI.getOpcode() == G_SHUFFLE_VECTOR.
  260.   bool matchCombineShuffleVector(MachineInstr &MI,
  261.                                  SmallVectorImpl<Register> &Ops);
  262.   /// Replace \p MI with a concat_vectors with \p Ops.
  263.   void applyCombineShuffleVector(MachineInstr &MI,
  264.                                  const ArrayRef<Register> Ops);
  265.  
  266.   /// Optimize memcpy intrinsics et al, e.g. constant len calls.
  267.   /// /p MaxLen if non-zero specifies the max length of a mem libcall to inline.
  268.   ///
  269.   /// For example (pre-indexed):
  270.   ///
  271.   ///     $addr = G_PTR_ADD $base, $offset
  272.   ///     [...]
  273.   ///     $val = G_LOAD $addr
  274.   ///     [...]
  275.   ///     $whatever = COPY $addr
  276.   ///
  277.   /// -->
  278.   ///
  279.   ///     $val, $addr = G_INDEXED_LOAD $base, $offset, 1 (IsPre)
  280.   ///     [...]
  281.   ///     $whatever = COPY $addr
  282.   ///
  283.   /// or (post-indexed):
  284.   ///
  285.   ///     G_STORE $val, $base
  286.   ///     [...]
  287.   ///     $addr = G_PTR_ADD $base, $offset
  288.   ///     [...]
  289.   ///     $whatever = COPY $addr
  290.   ///
  291.   /// -->
  292.   ///
  293.   ///     $addr = G_INDEXED_STORE $val, $base, $offset
  294.   ///     [...]
  295.   ///     $whatever = COPY $addr
  296.   bool tryCombineMemCpyFamily(MachineInstr &MI, unsigned MaxLen = 0);
  297.  
  298.   bool matchPtrAddImmedChain(MachineInstr &MI, PtrAddChain &MatchInfo);
  299.   void applyPtrAddImmedChain(MachineInstr &MI, PtrAddChain &MatchInfo);
  300.  
  301.   /// Fold (shift (shift base, x), y) -> (shift base (x+y))
  302.   bool matchShiftImmedChain(MachineInstr &MI, RegisterImmPair &MatchInfo);
  303.   void applyShiftImmedChain(MachineInstr &MI, RegisterImmPair &MatchInfo);
  304.  
  305.   /// If we have a shift-by-constant of a bitwise logic op that itself has a
  306.   /// shift-by-constant operand with identical opcode, we may be able to convert
  307.   /// that into 2 independent shifts followed by the logic op.
  308.   bool matchShiftOfShiftedLogic(MachineInstr &MI,
  309.                                 ShiftOfShiftedLogic &MatchInfo);
  310.   void applyShiftOfShiftedLogic(MachineInstr &MI,
  311.                                 ShiftOfShiftedLogic &MatchInfo);
  312.  
  313.   /// Transform a multiply by a power-of-2 value to a left shift.
  314.   bool matchCombineMulToShl(MachineInstr &MI, unsigned &ShiftVal);
  315.   void applyCombineMulToShl(MachineInstr &MI, unsigned &ShiftVal);
  316.  
  317.   // Transform a G_SHL with an extended source into a narrower shift if
  318.   // possible.
  319.   bool matchCombineShlOfExtend(MachineInstr &MI, RegisterImmPair &MatchData);
  320.   void applyCombineShlOfExtend(MachineInstr &MI,
  321.                                const RegisterImmPair &MatchData);
  322.  
  323.   /// Fold away a merge of an unmerge of the corresponding values.
  324.   bool matchCombineMergeUnmerge(MachineInstr &MI, Register &MatchInfo);
  325.  
  326.   /// Reduce a shift by a constant to an unmerge and a shift on a half sized
  327.   /// type. This will not produce a shift smaller than \p TargetShiftSize.
  328.   bool matchCombineShiftToUnmerge(MachineInstr &MI, unsigned TargetShiftSize,
  329.                                  unsigned &ShiftVal);
  330.   void applyCombineShiftToUnmerge(MachineInstr &MI, const unsigned &ShiftVal);
  331.   bool tryCombineShiftToUnmerge(MachineInstr &MI, unsigned TargetShiftAmount);
  332.  
  333.   /// Transform <ty,...> G_UNMERGE(G_MERGE ty X, Y, Z) -> ty X, Y, Z.
  334.   bool
  335.   matchCombineUnmergeMergeToPlainValues(MachineInstr &MI,
  336.                                         SmallVectorImpl<Register> &Operands);
  337.   void
  338.   applyCombineUnmergeMergeToPlainValues(MachineInstr &MI,
  339.                                         SmallVectorImpl<Register> &Operands);
  340.  
  341.   /// Transform G_UNMERGE Constant -> Constant1, Constant2, ...
  342.   bool matchCombineUnmergeConstant(MachineInstr &MI,
  343.                                    SmallVectorImpl<APInt> &Csts);
  344.   void applyCombineUnmergeConstant(MachineInstr &MI,
  345.                                    SmallVectorImpl<APInt> &Csts);
  346.  
  347.   /// Transform G_UNMERGE G_IMPLICIT_DEF -> G_IMPLICIT_DEF, G_IMPLICIT_DEF, ...
  348.   bool
  349.   matchCombineUnmergeUndef(MachineInstr &MI,
  350.                            std::function<void(MachineIRBuilder &)> &MatchInfo);
  351.  
  352.   /// Transform X, Y<dead> = G_UNMERGE Z -> X = G_TRUNC Z.
  353.   bool matchCombineUnmergeWithDeadLanesToTrunc(MachineInstr &MI);
  354.   void applyCombineUnmergeWithDeadLanesToTrunc(MachineInstr &MI);
  355.  
  356.   /// Transform X, Y = G_UNMERGE(G_ZEXT(Z)) -> X = G_ZEXT(Z); Y = G_CONSTANT 0
  357.   bool matchCombineUnmergeZExtToZExt(MachineInstr &MI);
  358.   void applyCombineUnmergeZExtToZExt(MachineInstr &MI);
  359.  
  360.   /// Transform fp_instr(cst) to constant result of the fp operation.
  361.   bool matchCombineConstantFoldFpUnary(MachineInstr &MI,
  362.                                        std::optional<APFloat> &Cst);
  363.   void applyCombineConstantFoldFpUnary(MachineInstr &MI,
  364.                                        std::optional<APFloat> &Cst);
  365.  
  366.   /// Transform IntToPtr(PtrToInt(x)) to x if cast is in the same address space.
  367.   bool matchCombineI2PToP2I(MachineInstr &MI, Register &Reg);
  368.   void applyCombineI2PToP2I(MachineInstr &MI, Register &Reg);
  369.  
  370.   /// Transform PtrToInt(IntToPtr(x)) to x.
  371.   void applyCombineP2IToI2P(MachineInstr &MI, Register &Reg);
  372.  
  373.   /// Transform G_ADD (G_PTRTOINT x), y -> G_PTRTOINT (G_PTR_ADD x, y)
  374.   /// Transform G_ADD y, (G_PTRTOINT x) -> G_PTRTOINT (G_PTR_ADD x, y)
  375.   bool matchCombineAddP2IToPtrAdd(MachineInstr &MI,
  376.                                   std::pair<Register, bool> &PtrRegAndCommute);
  377.   void applyCombineAddP2IToPtrAdd(MachineInstr &MI,
  378.                                   std::pair<Register, bool> &PtrRegAndCommute);
  379.  
  380.   // Transform G_PTR_ADD (G_PTRTOINT C1), C2 -> C1 + C2
  381.   bool matchCombineConstPtrAddToI2P(MachineInstr &MI, APInt &NewCst);
  382.   void applyCombineConstPtrAddToI2P(MachineInstr &MI, APInt &NewCst);
  383.  
  384.   /// Transform anyext(trunc(x)) to x.
  385.   bool matchCombineAnyExtTrunc(MachineInstr &MI, Register &Reg);
  386.   void applyCombineAnyExtTrunc(MachineInstr &MI, Register &Reg);
  387.  
  388.   /// Transform zext(trunc(x)) to x.
  389.   bool matchCombineZextTrunc(MachineInstr &MI, Register &Reg);
  390.  
  391.   /// Transform [asz]ext([asz]ext(x)) to [asz]ext x.
  392.   bool matchCombineExtOfExt(MachineInstr &MI,
  393.                             std::tuple<Register, unsigned> &MatchInfo);
  394.   void applyCombineExtOfExt(MachineInstr &MI,
  395.                             std::tuple<Register, unsigned> &MatchInfo);
  396.  
  397.   /// Transform fabs(fabs(x)) to fabs(x).
  398.   void applyCombineFAbsOfFAbs(MachineInstr &MI, Register &Src);
  399.  
  400.   /// Transform fabs(fneg(x)) to fabs(x).
  401.   bool matchCombineFAbsOfFNeg(MachineInstr &MI, BuildFnTy &MatchInfo);
  402.  
  403.   /// Transform trunc ([asz]ext x) to x or ([asz]ext x) or (trunc x).
  404.   bool matchCombineTruncOfExt(MachineInstr &MI,
  405.                               std::pair<Register, unsigned> &MatchInfo);
  406.   void applyCombineTruncOfExt(MachineInstr &MI,
  407.                               std::pair<Register, unsigned> &MatchInfo);
  408.  
  409.   /// Transform trunc (shl x, K) to shl (trunc x), K
  410.   ///    if K < VT.getScalarSizeInBits().
  411.   ///
  412.   /// Transforms trunc ([al]shr x, K) to (trunc ([al]shr (MidVT (trunc x)), K))
  413.   ///    if K <= (MidVT.getScalarSizeInBits() - VT.getScalarSizeInBits())
  414.   /// MidVT is obtained by finding a legal type between the trunc's src and dst
  415.   /// types.
  416.   bool matchCombineTruncOfShift(MachineInstr &MI,
  417.                                 std::pair<MachineInstr *, LLT> &MatchInfo);
  418.   void applyCombineTruncOfShift(MachineInstr &MI,
  419.                                 std::pair<MachineInstr *, LLT> &MatchInfo);
  420.  
  421.   /// Transform G_MUL(x, -1) to G_SUB(0, x)
  422.   void applyCombineMulByNegativeOne(MachineInstr &MI);
  423.  
  424.   /// Return true if any explicit use operand on \p MI is defined by a
  425.   /// G_IMPLICIT_DEF.
  426.   bool matchAnyExplicitUseIsUndef(MachineInstr &MI);
  427.  
  428.   /// Return true if all register explicit use operands on \p MI are defined by
  429.   /// a G_IMPLICIT_DEF.
  430.   bool matchAllExplicitUsesAreUndef(MachineInstr &MI);
  431.  
  432.   /// Return true if a G_SHUFFLE_VECTOR instruction \p MI has an undef mask.
  433.   bool matchUndefShuffleVectorMask(MachineInstr &MI);
  434.  
  435.   /// Return true if a G_STORE instruction \p MI is storing an undef value.
  436.   bool matchUndefStore(MachineInstr &MI);
  437.  
  438.   /// Return true if a G_SELECT instruction \p MI has an undef comparison.
  439.   bool matchUndefSelectCmp(MachineInstr &MI);
  440.  
  441.   /// Return true if a G_{EXTRACT,INSERT}_VECTOR_ELT has an out of range index.
  442.   bool matchInsertExtractVecEltOutOfBounds(MachineInstr &MI);
  443.  
  444.   /// Return true if a G_SELECT instruction \p MI has a constant comparison. If
  445.   /// true, \p OpIdx will store the operand index of the known selected value.
  446.   bool matchConstantSelectCmp(MachineInstr &MI, unsigned &OpIdx);
  447.  
  448.   /// Replace an instruction with a G_FCONSTANT with value \p C.
  449.   bool replaceInstWithFConstant(MachineInstr &MI, double C);
  450.  
  451.   /// Replace an instruction with a G_CONSTANT with value \p C.
  452.   bool replaceInstWithConstant(MachineInstr &MI, int64_t C);
  453.  
  454.   /// Replace an instruction with a G_CONSTANT with value \p C.
  455.   bool replaceInstWithConstant(MachineInstr &MI, APInt C);
  456.  
  457.   /// Replace an instruction with a G_IMPLICIT_DEF.
  458.   bool replaceInstWithUndef(MachineInstr &MI);
  459.  
  460.   /// Delete \p MI and replace all of its uses with its \p OpIdx-th operand.
  461.   bool replaceSingleDefInstWithOperand(MachineInstr &MI, unsigned OpIdx);
  462.  
  463.   /// Delete \p MI and replace all of its uses with \p Replacement.
  464.   bool replaceSingleDefInstWithReg(MachineInstr &MI, Register Replacement);
  465.  
  466.   /// Return true if \p MOP1 and \p MOP2 are register operands are defined by
  467.   /// equivalent instructions.
  468.   bool matchEqualDefs(const MachineOperand &MOP1, const MachineOperand &MOP2);
  469.  
  470.   /// Return true if \p MOP is defined by a G_CONSTANT with a value equal to
  471.   /// \p C.
  472.   bool matchConstantOp(const MachineOperand &MOP, int64_t C);
  473.  
  474.   /// Optimize (cond ? x : x) -> x
  475.   bool matchSelectSameVal(MachineInstr &MI);
  476.  
  477.   /// Optimize (x op x) -> x
  478.   bool matchBinOpSameVal(MachineInstr &MI);
  479.  
  480.   /// Check if operand \p OpIdx is zero.
  481.   bool matchOperandIsZero(MachineInstr &MI, unsigned OpIdx);
  482.  
  483.   /// Check if operand \p OpIdx is undef.
  484.   bool matchOperandIsUndef(MachineInstr &MI, unsigned OpIdx);
  485.  
  486.   /// Check if operand \p OpIdx is known to be a power of 2.
  487.   bool matchOperandIsKnownToBeAPowerOfTwo(MachineInstr &MI, unsigned OpIdx);
  488.  
  489.   /// Erase \p MI
  490.   bool eraseInst(MachineInstr &MI);
  491.  
  492.   /// Return true if MI is a G_ADD which can be simplified to a G_SUB.
  493.   bool matchSimplifyAddToSub(MachineInstr &MI,
  494.                              std::tuple<Register, Register> &MatchInfo);
  495.   void applySimplifyAddToSub(MachineInstr &MI,
  496.                              std::tuple<Register, Register> &MatchInfo);
  497.  
  498.   /// Match (logic_op (op x...), (op y...)) -> (op (logic_op x, y))
  499.   bool
  500.   matchHoistLogicOpWithSameOpcodeHands(MachineInstr &MI,
  501.                                        InstructionStepsMatchInfo &MatchInfo);
  502.  
  503.   /// Replace \p MI with a series of instructions described in \p MatchInfo.
  504.   void applyBuildInstructionSteps(MachineInstr &MI,
  505.                                   InstructionStepsMatchInfo &MatchInfo);
  506.  
  507.   /// Match ashr (shl x, C), C -> sext_inreg (C)
  508.   bool matchAshrShlToSextInreg(MachineInstr &MI,
  509.                                std::tuple<Register, int64_t> &MatchInfo);
  510.   void applyAshShlToSextInreg(MachineInstr &MI,
  511.                               std::tuple<Register, int64_t> &MatchInfo);
  512.  
  513.   /// Fold and(and(x, C1), C2) -> C1&C2 ? and(x, C1&C2) : 0
  514.   bool matchOverlappingAnd(MachineInstr &MI,
  515.                            BuildFnTy &MatchInfo);
  516.  
  517.   /// \return true if \p MI is a G_AND instruction whose operands are x and y
  518.   /// where x & y == x or x & y == y. (E.g., one of operands is all-ones value.)
  519.   ///
  520.   /// \param [in] MI - The G_AND instruction.
  521.   /// \param [out] Replacement - A register the G_AND should be replaced with on
  522.   /// success.
  523.   bool matchRedundantAnd(MachineInstr &MI, Register &Replacement);
  524.  
  525.   /// \return true if \p MI is a G_OR instruction whose operands are x and y
  526.   /// where x | y == x or x | y == y. (E.g., one of operands is all-zeros
  527.   /// value.)
  528.   ///
  529.   /// \param [in] MI - The G_OR instruction.
  530.   /// \param [out] Replacement - A register the G_OR should be replaced with on
  531.   /// success.
  532.   bool matchRedundantOr(MachineInstr &MI, Register &Replacement);
  533.  
  534.   /// \return true if \p MI is a G_SEXT_INREG that can be erased.
  535.   bool matchRedundantSExtInReg(MachineInstr &MI);
  536.  
  537.   /// Combine inverting a result of a compare into the opposite cond code.
  538.   bool matchNotCmp(MachineInstr &MI, SmallVectorImpl<Register> &RegsToNegate);
  539.   void applyNotCmp(MachineInstr &MI, SmallVectorImpl<Register> &RegsToNegate);
  540.  
  541.   /// Fold (xor (and x, y), y) -> (and (not x), y)
  542.   ///{
  543.   bool matchXorOfAndWithSameReg(MachineInstr &MI,
  544.                                 std::pair<Register, Register> &MatchInfo);
  545.   void applyXorOfAndWithSameReg(MachineInstr &MI,
  546.                                 std::pair<Register, Register> &MatchInfo);
  547.   ///}
  548.  
  549.   /// Combine G_PTR_ADD with nullptr to G_INTTOPTR
  550.   bool matchPtrAddZero(MachineInstr &MI);
  551.   void applyPtrAddZero(MachineInstr &MI);
  552.  
  553.   /// Combine G_UREM x, (known power of 2) to an add and bitmasking.
  554.   void applySimplifyURemByPow2(MachineInstr &MI);
  555.  
  556.   /// Push a binary operator through a select on constants.
  557.   ///
  558.   /// binop (select cond, K0, K1), K2 ->
  559.   ///   select cond, (binop K0, K2), (binop K1, K2)
  560.   bool matchFoldBinOpIntoSelect(MachineInstr &MI, unsigned &SelectOpNo);
  561.   bool applyFoldBinOpIntoSelect(MachineInstr &MI, const unsigned &SelectOpNo);
  562.  
  563.   bool matchCombineInsertVecElts(MachineInstr &MI,
  564.                                  SmallVectorImpl<Register> &MatchInfo);
  565.  
  566.   void applyCombineInsertVecElts(MachineInstr &MI,
  567.                              SmallVectorImpl<Register> &MatchInfo);
  568.  
  569.   /// Match expression trees of the form
  570.   ///
  571.   /// \code
  572.   ///  sN *a = ...
  573.   ///  sM val = a[0] | (a[1] << N) | (a[2] << 2N) | (a[3] << 3N) ...
  574.   /// \endcode
  575.   ///
  576.   /// And check if the tree can be replaced with a M-bit load + possibly a
  577.   /// bswap.
  578.   bool matchLoadOrCombine(MachineInstr &MI, BuildFnTy &MatchInfo);
  579.  
  580.   bool matchTruncStoreMerge(MachineInstr &MI, MergeTruncStoresInfo &MatchInfo);
  581.   void applyTruncStoreMerge(MachineInstr &MI, MergeTruncStoresInfo &MatchInfo);
  582.  
  583.   bool matchExtendThroughPhis(MachineInstr &MI, MachineInstr *&ExtMI);
  584.   void applyExtendThroughPhis(MachineInstr &MI, MachineInstr *&ExtMI);
  585.  
  586.   bool matchExtractVecEltBuildVec(MachineInstr &MI, Register &Reg);
  587.   void applyExtractVecEltBuildVec(MachineInstr &MI, Register &Reg);
  588.  
  589.   bool matchExtractAllEltsFromBuildVector(
  590.       MachineInstr &MI,
  591.       SmallVectorImpl<std::pair<Register, MachineInstr *>> &MatchInfo);
  592.   void applyExtractAllEltsFromBuildVector(
  593.       MachineInstr &MI,
  594.       SmallVectorImpl<std::pair<Register, MachineInstr *>> &MatchInfo);
  595.  
  596.   /// Use a function which takes in a MachineIRBuilder to perform a combine.
  597.   /// By default, it erases the instruction \p MI from the function.
  598.   void applyBuildFn(MachineInstr &MI, BuildFnTy &MatchInfo);
  599.   /// Use a function which takes in a MachineIRBuilder to perform a combine.
  600.   /// This variant does not erase \p MI after calling the build function.
  601.   void applyBuildFnNoErase(MachineInstr &MI, BuildFnTy &MatchInfo);
  602.  
  603.   bool matchOrShiftToFunnelShift(MachineInstr &MI, BuildFnTy &MatchInfo);
  604.   bool matchFunnelShiftToRotate(MachineInstr &MI);
  605.   void applyFunnelShiftToRotate(MachineInstr &MI);
  606.   bool matchRotateOutOfRange(MachineInstr &MI);
  607.   void applyRotateOutOfRange(MachineInstr &MI);
  608.  
  609.   /// \returns true if a G_ICMP instruction \p MI can be replaced with a true
  610.   /// or false constant based off of KnownBits information.
  611.   bool matchICmpToTrueFalseKnownBits(MachineInstr &MI, int64_t &MatchInfo);
  612.  
  613.   /// \returns true if a G_ICMP \p MI can be replaced with its LHS based off of
  614.   /// KnownBits information.
  615.   bool
  616.   matchICmpToLHSKnownBits(MachineInstr &MI,
  617.                           BuildFnTy &MatchInfo);
  618.  
  619.   /// \returns true if (and (or x, c1), c2) can be replaced with (and x, c2)
  620.   bool matchAndOrDisjointMask(MachineInstr &MI, BuildFnTy &MatchInfo);
  621.  
  622.   bool matchBitfieldExtractFromSExtInReg(MachineInstr &MI,
  623.                                          BuildFnTy &MatchInfo);
  624.   /// Match: and (lshr x, cst), mask -> ubfx x, cst, width
  625.   bool matchBitfieldExtractFromAnd(MachineInstr &MI, BuildFnTy &MatchInfo);
  626.  
  627.   /// Match: shr (shl x, n), k -> sbfx/ubfx x, pos, width
  628.   bool matchBitfieldExtractFromShr(MachineInstr &MI, BuildFnTy &MatchInfo);
  629.  
  630.   /// Match: shr (and x, n), k -> ubfx x, pos, width
  631.   bool matchBitfieldExtractFromShrAnd(MachineInstr &MI, BuildFnTy &MatchInfo);
  632.  
  633.   // Helpers for reassociation:
  634.   bool matchReassocConstantInnerRHS(GPtrAdd &MI, MachineInstr *RHS,
  635.                                     BuildFnTy &MatchInfo);
  636.   bool matchReassocFoldConstantsInSubTree(GPtrAdd &MI, MachineInstr *LHS,
  637.                                           MachineInstr *RHS,
  638.                                           BuildFnTy &MatchInfo);
  639.   bool matchReassocConstantInnerLHS(GPtrAdd &MI, MachineInstr *LHS,
  640.                                     MachineInstr *RHS, BuildFnTy &MatchInfo);
  641.   /// Reassociate pointer calculations with G_ADD involved, to allow better
  642.   /// addressing mode usage.
  643.   bool matchReassocPtrAdd(MachineInstr &MI, BuildFnTy &MatchInfo);
  644.  
  645.   /// Do constant folding when opportunities are exposed after MIR building.
  646.   bool matchConstantFold(MachineInstr &MI, APInt &MatchInfo);
  647.  
  648.   /// \returns true if it is possible to narrow the width of a scalar binop
  649.   /// feeding a G_AND instruction \p MI.
  650.   bool matchNarrowBinopFeedingAnd(MachineInstr &MI, BuildFnTy &MatchInfo);
  651.  
  652.   /// Given an G_UDIV \p MI expressing a divide by constant, return an
  653.   /// expression that implements it by multiplying by a magic number.
  654.   /// Ref: "Hacker's Delight" or "The PowerPC Compiler Writer's Guide".
  655.   MachineInstr *buildUDivUsingMul(MachineInstr &MI);
  656.   /// Combine G_UDIV by constant into a multiply by magic constant.
  657.   bool matchUDivByConst(MachineInstr &MI);
  658.   void applyUDivByConst(MachineInstr &MI);
  659.  
  660.   /// Given an G_SDIV \p MI expressing a signed divide by constant, return an
  661.   /// expression that implements it by multiplying by a magic number.
  662.   /// Ref: "Hacker's Delight" or "The PowerPC Compiler Writer's Guide".
  663.   MachineInstr *buildSDivUsingMul(MachineInstr &MI);
  664.   bool matchSDivByConst(MachineInstr &MI);
  665.   void applySDivByConst(MachineInstr &MI);
  666.  
  667.   // G_UMULH x, (1 << c)) -> x >> (bitwidth - c)
  668.   bool matchUMulHToLShr(MachineInstr &MI);
  669.   void applyUMulHToLShr(MachineInstr &MI);
  670.  
  671.   /// Try to transform \p MI by using all of the above
  672.   /// combine functions. Returns true if changed.
  673.   bool tryCombine(MachineInstr &MI);
  674.  
  675.   /// Emit loads and stores that perform the given memcpy.
  676.   /// Assumes \p MI is a G_MEMCPY_INLINE
  677.   /// TODO: implement dynamically sized inline memcpy,
  678.   ///       and rename: s/bool tryEmit/void emit/
  679.   bool tryEmitMemcpyInline(MachineInstr &MI);
  680.  
  681.   /// Match:
  682.   ///   (G_UMULO x, 2) -> (G_UADDO x, x)
  683.   ///   (G_SMULO x, 2) -> (G_SADDO x, x)
  684.   bool matchMulOBy2(MachineInstr &MI, BuildFnTy &MatchInfo);
  685.  
  686.   /// Match:
  687.   /// (G_*MULO x, 0) -> 0 + no carry out
  688.   bool matchMulOBy0(MachineInstr &MI, BuildFnTy &MatchInfo);
  689.  
  690.   /// Match:
  691.   /// (G_*ADDO x, 0) -> x + no carry out
  692.   bool matchAddOBy0(MachineInstr &MI, BuildFnTy &MatchInfo);
  693.  
  694.   /// Match:
  695.   /// (G_*ADDE x, y, 0) -> (G_*ADDO x, y)
  696.   /// (G_*SUBE x, y, 0) -> (G_*SUBO x, y)
  697.   bool matchAddEToAddO(MachineInstr &MI, BuildFnTy &MatchInfo);
  698.  
  699.   /// Transform (fadd x, fneg(y)) -> (fsub x, y)
  700.   ///           (fadd fneg(x), y) -> (fsub y, x)
  701.   ///           (fsub x, fneg(y)) -> (fadd x, y)
  702.   ///           (fmul fneg(x), fneg(y)) -> (fmul x, y)
  703.   ///           (fdiv fneg(x), fneg(y)) -> (fdiv x, y)
  704.   ///           (fmad fneg(x), fneg(y), z) -> (fmad x, y, z)
  705.   ///           (fma fneg(x), fneg(y), z) -> (fma x, y, z)
  706.   bool matchRedundantNegOperands(MachineInstr &MI, BuildFnTy &MatchInfo);
  707.  
  708.   bool matchFsubToFneg(MachineInstr &MI, Register &MatchInfo);
  709.   void applyFsubToFneg(MachineInstr &MI, Register &MatchInfo);
  710.  
  711.   bool canCombineFMadOrFMA(MachineInstr &MI, bool &AllowFusionGlobally,
  712.                            bool &HasFMAD, bool &Aggressive,
  713.                            bool CanReassociate = false);
  714.  
  715.   /// Transform (fadd (fmul x, y), z) -> (fma x, y, z)
  716.   ///           (fadd (fmul x, y), z) -> (fmad x, y, z)
  717.   bool matchCombineFAddFMulToFMadOrFMA(MachineInstr &MI, BuildFnTy &MatchInfo);
  718.  
  719.   /// Transform (fadd (fpext (fmul x, y)), z) -> (fma (fpext x), (fpext y), z)
  720.   ///           (fadd (fpext (fmul x, y)), z) -> (fmad (fpext x), (fpext y), z)
  721.   bool matchCombineFAddFpExtFMulToFMadOrFMA(MachineInstr &MI,
  722.                                             BuildFnTy &MatchInfo);
  723.  
  724.   /// Transform (fadd (fma x, y, (fmul u, v)), z) -> (fma x, y, (fma u, v, z))
  725.   ///          (fadd (fmad x, y, (fmul u, v)), z) -> (fmad x, y, (fmad u, v, z))
  726.   bool matchCombineFAddFMAFMulToFMadOrFMA(MachineInstr &MI,
  727.                                           BuildFnTy &MatchInfo);
  728.  
  729.   // Transform (fadd (fma x, y, (fpext (fmul u, v))), z)
  730.   //            -> (fma x, y, (fma (fpext u), (fpext v), z))
  731.   //           (fadd (fmad x, y, (fpext (fmul u, v))), z)
  732.   //            -> (fmad x, y, (fmad (fpext u), (fpext v), z))
  733.   bool matchCombineFAddFpExtFMulToFMadOrFMAAggressive(MachineInstr &MI,
  734.                                                       BuildFnTy &MatchInfo);
  735.  
  736.   /// Transform (fsub (fmul x, y), z) -> (fma x, y, -z)
  737.   ///           (fsub (fmul x, y), z) -> (fmad x, y, -z)
  738.   bool matchCombineFSubFMulToFMadOrFMA(MachineInstr &MI, BuildFnTy &MatchInfo);
  739.  
  740.   /// Transform (fsub (fneg (fmul, x, y)), z) -> (fma (fneg x), y, (fneg z))
  741.   ///           (fsub (fneg (fmul, x, y)), z) -> (fmad (fneg x), y, (fneg z))
  742.   bool matchCombineFSubFNegFMulToFMadOrFMA(MachineInstr &MI,
  743.                                            BuildFnTy &MatchInfo);
  744.  
  745.   /// Transform (fsub (fpext (fmul x, y)), z)
  746.   ///           -> (fma (fpext x), (fpext y), (fneg z))
  747.   ///           (fsub (fpext (fmul x, y)), z)
  748.   ///           -> (fmad (fpext x), (fpext y), (fneg z))
  749.   bool matchCombineFSubFpExtFMulToFMadOrFMA(MachineInstr &MI,
  750.                                             BuildFnTy &MatchInfo);
  751.  
  752.   /// Transform (fsub (fpext (fneg (fmul x, y))), z)
  753.   ///           -> (fneg (fma (fpext x), (fpext y), z))
  754.   ///           (fsub (fpext (fneg (fmul x, y))), z)
  755.   ///           -> (fneg (fmad (fpext x), (fpext y), z))
  756.   bool matchCombineFSubFpExtFNegFMulToFMadOrFMA(MachineInstr &MI,
  757.                                                 BuildFnTy &MatchInfo);
  758.  
  759.   /// Fold boolean selects to logical operations.
  760.   bool matchSelectToLogical(MachineInstr &MI, BuildFnTy &MatchInfo);
  761.  
  762.   bool matchCombineFMinMaxNaN(MachineInstr &MI, unsigned &Info);
  763.  
  764.   /// Transform G_ADD(x, G_SUB(y, x)) to y.
  765.   /// Transform G_ADD(G_SUB(y, x), x) to y.
  766.   bool matchAddSubSameReg(MachineInstr &MI, Register &Src);
  767.  
  768.   bool matchBuildVectorIdentityFold(MachineInstr &MI, Register &MatchInfo);
  769.   bool matchTruncBuildVectorFold(MachineInstr &MI, Register &MatchInfo);
  770.   bool matchTruncLshrBuildVectorFold(MachineInstr &MI, Register &MatchInfo);
  771.  
  772.   /// Transform:
  773.   ///   (x + y) - y -> x
  774.   ///   (x + y) - x -> y
  775.   ///   x - (y + x) -> 0 - y
  776.   ///   x - (x + z) -> 0 - z
  777.   bool matchSubAddSameReg(MachineInstr &MI, BuildFnTy &MatchInfo);
  778.  
  779.   /// \returns true if it is possible to simplify a select instruction \p MI
  780.   /// to a min/max instruction of some sort.
  781.   bool matchSimplifySelectToMinMax(MachineInstr &MI, BuildFnTy &MatchInfo);
  782.  
  783.   /// Transform:
  784.   ///   (X + Y) == X -> Y == 0
  785.   ///   (X - Y) == X -> Y == 0
  786.   ///   (X ^ Y) == X -> Y == 0
  787.   ///   (X + Y) != X -> Y != 0
  788.   ///   (X - Y) != X -> Y != 0
  789.   ///   (X ^ Y) != X -> Y != 0
  790.   bool matchRedundantBinOpInEquality(MachineInstr &MI, BuildFnTy &MatchInfo);
  791.  
  792. private:
  793.   /// Given a non-indexed load or store instruction \p MI, find an offset that
  794.   /// can be usefully and legally folded into it as a post-indexing operation.
  795.   ///
  796.   /// \returns true if a candidate is found.
  797.   bool findPostIndexCandidate(MachineInstr &MI, Register &Addr, Register &Base,
  798.                               Register &Offset);
  799.  
  800.   /// Given a non-indexed load or store instruction \p MI, find an offset that
  801.   /// can be usefully and legally folded into it as a pre-indexing operation.
  802.   ///
  803.   /// \returns true if a candidate is found.
  804.   bool findPreIndexCandidate(MachineInstr &MI, Register &Addr, Register &Base,
  805.                              Register &Offset);
  806.  
  807.   /// Helper function for matchLoadOrCombine. Searches for Registers
  808.   /// which may have been produced by a load instruction + some arithmetic.
  809.   ///
  810.   /// \param [in] Root - The search root.
  811.   ///
  812.   /// \returns The Registers found during the search.
  813.   std::optional<SmallVector<Register, 8>>
  814.   findCandidatesForLoadOrCombine(const MachineInstr *Root) const;
  815.  
  816.   /// Helper function for matchLoadOrCombine.
  817.   ///
  818.   /// Checks if every register in \p RegsToVisit is defined by a load
  819.   /// instruction + some arithmetic.
  820.   ///
  821.   /// \param [out] MemOffset2Idx - Maps the byte positions each load ends up
  822.   /// at to the index of the load.
  823.   /// \param [in] MemSizeInBits - The number of bits each load should produce.
  824.   ///
  825.   /// \returns On success, a 3-tuple containing lowest-index load found, the
  826.   /// lowest index, and the last load in the sequence.
  827.   std::optional<std::tuple<GZExtLoad *, int64_t, GZExtLoad *>>
  828.   findLoadOffsetsForLoadOrCombine(
  829.       SmallDenseMap<int64_t, int64_t, 8> &MemOffset2Idx,
  830.       const SmallVector<Register, 8> &RegsToVisit,
  831.       const unsigned MemSizeInBits);
  832.  
  833.   /// Examines the G_PTR_ADD instruction \p PtrAdd and determines if performing
  834.   /// a re-association of its operands would break an existing legal addressing
  835.   /// mode that the address computation currently represents.
  836.   bool reassociationCanBreakAddressingModePattern(MachineInstr &PtrAdd);
  837.  
  838.   /// Behavior when a floating point min/max is given one NaN and one
  839.   /// non-NaN as input.
  840.   enum class SelectPatternNaNBehaviour {
  841.     NOT_APPLICABLE = 0, /// NaN behavior not applicable.
  842.     RETURNS_NAN,        /// Given one NaN input, returns the NaN.
  843.     RETURNS_OTHER,      /// Given one NaN input, returns the non-NaN.
  844.     RETURNS_ANY /// Given one NaN input, can return either (or both operands are
  845.                 /// known non-NaN.)
  846.   };
  847.  
  848.   /// \returns which of \p LHS and \p RHS would be the result of a non-equality
  849.   /// floating point comparison where one of \p LHS and \p RHS may be NaN.
  850.   ///
  851.   /// If both \p LHS and \p RHS may be NaN, returns
  852.   /// SelectPatternNaNBehaviour::NOT_APPLICABLE.
  853.   SelectPatternNaNBehaviour
  854.   computeRetValAgainstNaN(Register LHS, Register RHS,
  855.                           bool IsOrderedComparison) const;
  856.  
  857.   /// Determines the floating point min/max opcode which should be used for
  858.   /// a G_SELECT fed by a G_FCMP with predicate \p Pred.
  859.   ///
  860.   /// \returns 0 if this G_SELECT should not be combined to a floating point
  861.   /// min or max. If it should be combined, returns one of
  862.   ///
  863.   /// * G_FMAXNUM
  864.   /// * G_FMAXIMUM
  865.   /// * G_FMINNUM
  866.   /// * G_FMINIMUM
  867.   ///
  868.   /// Helper function for matchFPSelectToMinMax.
  869.   unsigned getFPMinMaxOpcForSelect(CmpInst::Predicate Pred, LLT DstTy,
  870.                                    SelectPatternNaNBehaviour VsNaNRetVal) const;
  871.  
  872.   /// Handle floating point cases for matchSimplifySelectToMinMax.
  873.   ///
  874.   /// E.g.
  875.   ///
  876.   /// select (fcmp uge x, 1.0) x, 1.0 -> fmax x, 1.0
  877.   /// select (fcmp uge x, 1.0) 1.0, x -> fminnm x, 1.0
  878.   bool matchFPSelectToMinMax(Register Dst, Register Cond, Register TrueVal,
  879.                              Register FalseVal, BuildFnTy &MatchInfo);
  880. };
  881. } // namespace llvm
  882.  
  883. #endif
  884.