Home | History | Annotate | Line # | Download | only in CodeGen
      1 //===-- llvm/CodeGen/ISDOpcodes.h - CodeGen opcodes -------------*- C++ -*-===//
      2 //
      3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
      4 // See https://llvm.org/LICENSE.txt for license information.
      5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
      6 //
      7 //===----------------------------------------------------------------------===//
      8 //
      9 // This file declares codegen opcodes and related utilities.
     10 //
     11 //===----------------------------------------------------------------------===//
     12 
     13 #ifndef LLVM_CODEGEN_ISDOPCODES_H
     14 #define LLVM_CODEGEN_ISDOPCODES_H
     15 
     16 #include "llvm/CodeGen/ValueTypes.h"
     17 
     18 namespace llvm {
     19 
     20 /// ISD namespace - This namespace contains an enum which represents all of the
     21 /// SelectionDAG node types and value types.
     22 ///
     23 namespace ISD {
     24 
     25 //===--------------------------------------------------------------------===//
     26 /// ISD::NodeType enum - This enum defines the target-independent operators
     27 /// for a SelectionDAG.
     28 ///
     29 /// Targets may also define target-dependent operator codes for SDNodes. For
     30 /// example, on x86, these are the enum values in the X86ISD namespace.
     31 /// Targets should aim to use target-independent operators to model their
     32 /// instruction sets as much as possible, and only use target-dependent
     33 /// operators when they have special requirements.
     34 ///
     35 /// Finally, during and after selection proper, SNodes may use special
     36 /// operator codes that correspond directly with MachineInstr opcodes. These
     37 /// are used to represent selected instructions. See the isMachineOpcode()
     38 /// and getMachineOpcode() member functions of SDNode.
     39 ///
     40 enum NodeType {
     41 
     42   /// DELETED_NODE - This is an illegal value that is used to catch
     43   /// errors.  This opcode is not a legal opcode for any node.
     44   DELETED_NODE,
     45 
     46   /// EntryToken - This is the marker used to indicate the start of a region.
     47   EntryToken,
     48 
     49   /// TokenFactor - This node takes multiple tokens as input and produces a
     50   /// single token result. This is used to represent the fact that the operand
     51   /// operators are independent of each other.
     52   TokenFactor,
     53 
     54   /// AssertSext, AssertZext - These nodes record if a register contains a
     55   /// value that has already been zero or sign extended from a narrower type.
     56   /// These nodes take two operands.  The first is the node that has already
     57   /// been extended, and the second is a value type node indicating the width
     58   /// of the extension
     59   AssertSext,
     60   AssertZext,
     61   AssertAlign,
     62 
     63   /// Various leaf nodes.
     64   BasicBlock,
     65   VALUETYPE,
     66   CONDCODE,
     67   Register,
     68   RegisterMask,
     69   Constant,
     70   ConstantFP,
     71   GlobalAddress,
     72   GlobalTLSAddress,
     73   FrameIndex,
     74   JumpTable,
     75   ConstantPool,
     76   ExternalSymbol,
     77   BlockAddress,
     78 
     79   /// The address of the GOT
     80   GLOBAL_OFFSET_TABLE,
     81 
     82   /// FRAMEADDR, RETURNADDR - These nodes represent llvm.frameaddress and
     83   /// llvm.returnaddress on the DAG.  These nodes take one operand, the index
     84   /// of the frame or return address to return.  An index of zero corresponds
     85   /// to the current function's frame or return address, an index of one to
     86   /// the parent's frame or return address, and so on.
     87   FRAMEADDR,
     88   RETURNADDR,
     89 
     90   /// ADDROFRETURNADDR - Represents the llvm.addressofreturnaddress intrinsic.
     91   /// This node takes no operand, returns a target-specific pointer to the
     92   /// place in the stack frame where the return address of the current
     93   /// function is stored.
     94   ADDROFRETURNADDR,
     95 
     96   /// SPONENTRY - Represents the llvm.sponentry intrinsic. Takes no argument
     97   /// and returns the stack pointer value at the entry of the current
     98   /// function calling this intrinsic.
     99   SPONENTRY,
    100 
    101   /// LOCAL_RECOVER - Represents the llvm.localrecover intrinsic.
    102   /// Materializes the offset from the local object pointer of another
    103   /// function to a particular local object passed to llvm.localescape. The
    104   /// operand is the MCSymbol label used to represent this offset, since
    105   /// typically the offset is not known until after code generation of the
    106   /// parent.
    107   LOCAL_RECOVER,
    108 
    109   /// READ_REGISTER, WRITE_REGISTER - This node represents llvm.register on
    110   /// the DAG, which implements the named register global variables extension.
    111   READ_REGISTER,
    112   WRITE_REGISTER,
    113 
    114   /// FRAME_TO_ARGS_OFFSET - This node represents offset from frame pointer to
    115   /// first (possible) on-stack argument. This is needed for correct stack
    116   /// adjustment during unwind.
    117   FRAME_TO_ARGS_OFFSET,
    118 
    119   /// EH_DWARF_CFA - This node represents the pointer to the DWARF Canonical
    120   /// Frame Address (CFA), generally the value of the stack pointer at the
    121   /// call site in the previous frame.
    122   EH_DWARF_CFA,
    123 
    124   /// OUTCHAIN = EH_RETURN(INCHAIN, OFFSET, HANDLER) - This node represents
    125   /// 'eh_return' gcc dwarf builtin, which is used to return from
    126   /// exception. The general meaning is: adjust stack by OFFSET and pass
    127   /// execution to HANDLER. Many platform-related details also :)
    128   EH_RETURN,
    129 
    130   /// RESULT, OUTCHAIN = EH_SJLJ_SETJMP(INCHAIN, buffer)
    131   /// This corresponds to the eh.sjlj.setjmp intrinsic.
    132   /// It takes an input chain and a pointer to the jump buffer as inputs
    133   /// and returns an outchain.
    134   EH_SJLJ_SETJMP,
    135 
    136   /// OUTCHAIN = EH_SJLJ_LONGJMP(INCHAIN, buffer)
    137   /// This corresponds to the eh.sjlj.longjmp intrinsic.
    138   /// It takes an input chain and a pointer to the jump buffer as inputs
    139   /// and returns an outchain.
    140   EH_SJLJ_LONGJMP,
    141 
    142   /// OUTCHAIN = EH_SJLJ_SETUP_DISPATCH(INCHAIN)
    143   /// The target initializes the dispatch table here.
    144   EH_SJLJ_SETUP_DISPATCH,
    145 
    146   /// TargetConstant* - Like Constant*, but the DAG does not do any folding,
    147   /// simplification, or lowering of the constant. They are used for constants
    148   /// which are known to fit in the immediate fields of their users, or for
    149   /// carrying magic numbers which are not values which need to be
    150   /// materialized in registers.
    151   TargetConstant,
    152   TargetConstantFP,
    153 
    154   /// TargetGlobalAddress - Like GlobalAddress, but the DAG does no folding or
    155   /// anything else with this node, and this is valid in the target-specific
    156   /// dag, turning into a GlobalAddress operand.
    157   TargetGlobalAddress,
    158   TargetGlobalTLSAddress,
    159   TargetFrameIndex,
    160   TargetJumpTable,
    161   TargetConstantPool,
    162   TargetExternalSymbol,
    163   TargetBlockAddress,
    164 
    165   MCSymbol,
    166 
    167   /// TargetIndex - Like a constant pool entry, but with completely
    168   /// target-dependent semantics. Holds target flags, a 32-bit index, and a
    169   /// 64-bit index. Targets can use this however they like.
    170   TargetIndex,
    171 
    172   /// RESULT = INTRINSIC_WO_CHAIN(INTRINSICID, arg1, arg2, ...)
    173   /// This node represents a target intrinsic function with no side effects.
    174   /// The first operand is the ID number of the intrinsic from the
    175   /// llvm::Intrinsic namespace.  The operands to the intrinsic follow.  The
    176   /// node returns the result of the intrinsic.
    177   INTRINSIC_WO_CHAIN,
    178 
    179   /// RESULT,OUTCHAIN = INTRINSIC_W_CHAIN(INCHAIN, INTRINSICID, arg1, ...)
    180   /// This node represents a target intrinsic function with side effects that
    181   /// returns a result.  The first operand is a chain pointer.  The second is
    182   /// the ID number of the intrinsic from the llvm::Intrinsic namespace.  The
    183   /// operands to the intrinsic follow.  The node has two results, the result
    184   /// of the intrinsic and an output chain.
    185   INTRINSIC_W_CHAIN,
    186 
    187   /// OUTCHAIN = INTRINSIC_VOID(INCHAIN, INTRINSICID, arg1, arg2, ...)
    188   /// This node represents a target intrinsic function with side effects that
    189   /// does not return a result.  The first operand is a chain pointer.  The
    190   /// second is the ID number of the intrinsic from the llvm::Intrinsic
    191   /// namespace.  The operands to the intrinsic follow.
    192   INTRINSIC_VOID,
    193 
    194   /// CopyToReg - This node has three operands: a chain, a register number to
    195   /// set to this value, and a value.
    196   CopyToReg,
    197 
    198   /// CopyFromReg - This node indicates that the input value is a virtual or
    199   /// physical register that is defined outside of the scope of this
    200   /// SelectionDAG.  The register is available from the RegisterSDNode object.
    201   CopyFromReg,
    202 
    203   /// UNDEF - An undefined node.
    204   UNDEF,
    205 
    206   // FREEZE - FREEZE(VAL) returns an arbitrary value if VAL is UNDEF (or
    207   // is evaluated to UNDEF), or returns VAL otherwise. Note that each
    208   // read of UNDEF can yield different value, but FREEZE(UNDEF) cannot.
    209   FREEZE,
    210 
    211   /// EXTRACT_ELEMENT - This is used to get the lower or upper (determined by
    212   /// a Constant, which is required to be operand #1) half of the integer or
    213   /// float value specified as operand #0.  This is only for use before
    214   /// legalization, for values that will be broken into multiple registers.
    215   EXTRACT_ELEMENT,
    216 
    217   /// BUILD_PAIR - This is the opposite of EXTRACT_ELEMENT in some ways.
    218   /// Given two values of the same integer value type, this produces a value
    219   /// twice as big.  Like EXTRACT_ELEMENT, this can only be used before
    220   /// legalization. The lower part of the composite value should be in
    221   /// element 0 and the upper part should be in element 1.
    222   BUILD_PAIR,
    223 
    224   /// MERGE_VALUES - This node takes multiple discrete operands and returns
    225   /// them all as its individual results.  This nodes has exactly the same
    226   /// number of inputs and outputs. This node is useful for some pieces of the
    227   /// code generator that want to think about a single node with multiple
    228   /// results, not multiple nodes.
    229   MERGE_VALUES,
    230 
    231   /// Simple integer binary arithmetic operators.
    232   ADD,
    233   SUB,
    234   MUL,
    235   SDIV,
    236   UDIV,
    237   SREM,
    238   UREM,
    239 
    240   /// SMUL_LOHI/UMUL_LOHI - Multiply two integers of type iN, producing
    241   /// a signed/unsigned value of type i[2*N], and return the full value as
    242   /// two results, each of type iN.
    243   SMUL_LOHI,
    244   UMUL_LOHI,
    245 
    246   /// SDIVREM/UDIVREM - Divide two integers and produce both a quotient and
    247   /// remainder result.
    248   SDIVREM,
    249   UDIVREM,
    250 
    251   /// CARRY_FALSE - This node is used when folding other nodes,
    252   /// like ADDC/SUBC, which indicate the carry result is always false.
    253   CARRY_FALSE,
    254 
    255   /// Carry-setting nodes for multiple precision addition and subtraction.
    256   /// These nodes take two operands of the same value type, and produce two
    257   /// results.  The first result is the normal add or sub result, the second
    258   /// result is the carry flag result.
    259   /// FIXME: These nodes are deprecated in favor of ADDCARRY and SUBCARRY.
    260   /// They are kept around for now to provide a smooth transition path
    261   /// toward the use of ADDCARRY/SUBCARRY and will eventually be removed.
    262   ADDC,
    263   SUBC,
    264 
    265   /// Carry-using nodes for multiple precision addition and subtraction. These
    266   /// nodes take three operands: The first two are the normal lhs and rhs to
    267   /// the add or sub, and the third is the input carry flag.  These nodes
    268   /// produce two results; the normal result of the add or sub, and the output
    269   /// carry flag.  These nodes both read and write a carry flag to allow them
    270   /// to them to be chained together for add and sub of arbitrarily large
    271   /// values.
    272   ADDE,
    273   SUBE,
    274 
    275   /// Carry-using nodes for multiple precision addition and subtraction.
    276   /// These nodes take three operands: The first two are the normal lhs and
    277   /// rhs to the add or sub, and the third is a boolean indicating if there
    278   /// is an incoming carry. These nodes produce two results: the normal
    279   /// result of the add or sub, and the output carry so they can be chained
    280   /// together. The use of this opcode is preferable to adde/sube if the
    281   /// target supports it, as the carry is a regular value rather than a
    282   /// glue, which allows further optimisation.
    283   ADDCARRY,
    284   SUBCARRY,
    285 
    286   /// Carry-using overflow-aware nodes for multiple precision addition and
    287   /// subtraction. These nodes take three operands: The first two are normal lhs
    288   /// and rhs to the add or sub, and the third is a boolean indicating if there
    289   /// is an incoming carry. They produce two results: the normal result of the
    290   /// add or sub, and a boolean that indicates if an overflow occured (*not*
    291   /// flag, because it may be a store to memory, etc.). If the type of the
    292   /// boolean is not i1 then the high bits conform to getBooleanContents.
    293   SADDO_CARRY,
    294   SSUBO_CARRY,
    295 
    296   /// RESULT, BOOL = [SU]ADDO(LHS, RHS) - Overflow-aware nodes for addition.
    297   /// These nodes take two operands: the normal LHS and RHS to the add. They
    298   /// produce two results: the normal result of the add, and a boolean that
    299   /// indicates if an overflow occurred (*not* a flag, because it may be store
    300   /// to memory, etc.).  If the type of the boolean is not i1 then the high
    301   /// bits conform to getBooleanContents.
    302   /// These nodes are generated from llvm.[su]add.with.overflow intrinsics.
    303   SADDO,
    304   UADDO,
    305 
    306   /// Same for subtraction.
    307   SSUBO,
    308   USUBO,
    309 
    310   /// Same for multiplication.
    311   SMULO,
    312   UMULO,
    313 
    314   /// RESULT = [US]ADDSAT(LHS, RHS) - Perform saturation addition on 2
    315   /// integers with the same bit width (W). If the true value of LHS + RHS
    316   /// exceeds the largest value that can be represented by W bits, the
    317   /// resulting value is this maximum value. Otherwise, if this value is less
    318   /// than the smallest value that can be represented by W bits, the
    319   /// resulting value is this minimum value.
    320   SADDSAT,
    321   UADDSAT,
    322 
    323   /// RESULT = [US]SUBSAT(LHS, RHS) - Perform saturation subtraction on 2
    324   /// integers with the same bit width (W). If the true value of LHS - RHS
    325   /// exceeds the largest value that can be represented by W bits, the
    326   /// resulting value is this maximum value. Otherwise, if this value is less
    327   /// than the smallest value that can be represented by W bits, the
    328   /// resulting value is this minimum value.
    329   SSUBSAT,
    330   USUBSAT,
    331 
    332   /// RESULT = [US]SHLSAT(LHS, RHS) - Perform saturation left shift. The first
    333   /// operand is the value to be shifted, and the second argument is the amount
    334   /// to shift by. Both must be integers of the same bit width (W). If the true
    335   /// value of LHS << RHS exceeds the largest value that can be represented by
    336   /// W bits, the resulting value is this maximum value, Otherwise, if this
    337   /// value is less than the smallest value that can be represented by W bits,
    338   /// the resulting value is this minimum value.
    339   SSHLSAT,
    340   USHLSAT,
    341 
    342   /// RESULT = [US]MULFIX(LHS, RHS, SCALE) - Perform fixed point multiplication
    343   /// on
    344   /// 2 integers with the same width and scale. SCALE represents the scale of
    345   /// both operands as fixed point numbers. This SCALE parameter must be a
    346   /// constant integer. A scale of zero is effectively performing
    347   /// multiplication on 2 integers.
    348   SMULFIX,
    349   UMULFIX,
    350 
    351   /// Same as the corresponding unsaturated fixed point instructions, but the
    352   /// result is clamped between the min and max values representable by the
    353   /// bits of the first 2 operands.
    354   SMULFIXSAT,
    355   UMULFIXSAT,
    356 
    357   /// RESULT = [US]DIVFIX(LHS, RHS, SCALE) - Perform fixed point division on
    358   /// 2 integers with the same width and scale. SCALE represents the scale
    359   /// of both operands as fixed point numbers. This SCALE parameter must be a
    360   /// constant integer.
    361   SDIVFIX,
    362   UDIVFIX,
    363 
    364   /// Same as the corresponding unsaturated fixed point instructions, but the
    365   /// result is clamped between the min and max values representable by the
    366   /// bits of the first 2 operands.
    367   SDIVFIXSAT,
    368   UDIVFIXSAT,
    369 
    370   /// Simple binary floating point operators.
    371   FADD,
    372   FSUB,
    373   FMUL,
    374   FDIV,
    375   FREM,
    376 
    377   /// Constrained versions of the binary floating point operators.
    378   /// These will be lowered to the simple operators before final selection.
    379   /// They are used to limit optimizations while the DAG is being
    380   /// optimized.
    381   STRICT_FADD,
    382   STRICT_FSUB,
    383   STRICT_FMUL,
    384   STRICT_FDIV,
    385   STRICT_FREM,
    386   STRICT_FMA,
    387 
    388   /// Constrained versions of libm-equivalent floating point intrinsics.
    389   /// These will be lowered to the equivalent non-constrained pseudo-op
    390   /// (or expanded to the equivalent library call) before final selection.
    391   /// They are used to limit optimizations while the DAG is being optimized.
    392   STRICT_FSQRT,
    393   STRICT_FPOW,
    394   STRICT_FPOWI,
    395   STRICT_FSIN,
    396   STRICT_FCOS,
    397   STRICT_FEXP,
    398   STRICT_FEXP2,
    399   STRICT_FLOG,
    400   STRICT_FLOG10,
    401   STRICT_FLOG2,
    402   STRICT_FRINT,
    403   STRICT_FNEARBYINT,
    404   STRICT_FMAXNUM,
    405   STRICT_FMINNUM,
    406   STRICT_FCEIL,
    407   STRICT_FFLOOR,
    408   STRICT_FROUND,
    409   STRICT_FROUNDEVEN,
    410   STRICT_FTRUNC,
    411   STRICT_LROUND,
    412   STRICT_LLROUND,
    413   STRICT_LRINT,
    414   STRICT_LLRINT,
    415   STRICT_FMAXIMUM,
    416   STRICT_FMINIMUM,
    417 
    418   /// STRICT_FP_TO_[US]INT - Convert a floating point value to a signed or
    419   /// unsigned integer. These have the same semantics as fptosi and fptoui
    420   /// in IR.
    421   /// They are used to limit optimizations while the DAG is being optimized.
    422   STRICT_FP_TO_SINT,
    423   STRICT_FP_TO_UINT,
    424 
    425   /// STRICT_[US]INT_TO_FP - Convert a signed or unsigned integer to
    426   /// a floating point value. These have the same semantics as sitofp and
    427   /// uitofp in IR.
    428   /// They are used to limit optimizations while the DAG is being optimized.
    429   STRICT_SINT_TO_FP,
    430   STRICT_UINT_TO_FP,
    431 
    432   /// X = STRICT_FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating
    433   /// point type down to the precision of the destination VT.  TRUNC is a
    434   /// flag, which is always an integer that is zero or one.  If TRUNC is 0,
    435   /// this is a normal rounding, if it is 1, this FP_ROUND is known to not
    436   /// change the value of Y.
    437   ///
    438   /// The TRUNC = 1 case is used in cases where we know that the value will
    439   /// not be modified by the node, because Y is not using any of the extra
    440   /// precision of source type.  This allows certain transformations like
    441   /// STRICT_FP_EXTEND(STRICT_FP_ROUND(X,1)) -> X which are not safe for
    442   /// STRICT_FP_EXTEND(STRICT_FP_ROUND(X,0)) because the extra bits aren't
    443   /// removed.
    444   /// It is used to limit optimizations while the DAG is being optimized.
    445   STRICT_FP_ROUND,
    446 
    447   /// X = STRICT_FP_EXTEND(Y) - Extend a smaller FP type into a larger FP
    448   /// type.
    449   /// It is used to limit optimizations while the DAG is being optimized.
    450   STRICT_FP_EXTEND,
    451 
    452   /// STRICT_FSETCC/STRICT_FSETCCS - Constrained versions of SETCC, used
    453   /// for floating-point operands only.  STRICT_FSETCC performs a quiet
    454   /// comparison operation, while STRICT_FSETCCS performs a signaling
    455   /// comparison operation.
    456   STRICT_FSETCC,
    457   STRICT_FSETCCS,
    458 
    459   /// FMA - Perform a * b + c with no intermediate rounding step.
    460   FMA,
    461 
    462   /// FMAD - Perform a * b + c, while getting the same result as the
    463   /// separately rounded operations.
    464   FMAD,
    465 
    466   /// FCOPYSIGN(X, Y) - Return the value of X with the sign of Y.  NOTE: This
    467   /// DAG node does not require that X and Y have the same type, just that
    468   /// they are both floating point.  X and the result must have the same type.
    469   /// FCOPYSIGN(f32, f64) is allowed.
    470   FCOPYSIGN,
    471 
    472   /// INT = FGETSIGN(FP) - Return the sign bit of the specified floating point
    473   /// value as an integer 0/1 value.
    474   FGETSIGN,
    475 
    476   /// Returns platform specific canonical encoding of a floating point number.
    477   FCANONICALIZE,
    478 
    479   /// BUILD_VECTOR(ELT0, ELT1, ELT2, ELT3,...) - Return a fixed-width vector
    480   /// with the specified, possibly variable, elements. The types of the
    481   /// operands must match the vector element type, except that integer types
    482   /// are allowed to be larger than the element type, in which case the
    483   /// operands are implicitly truncated. The types of the operands must all
    484   /// be the same.
    485   BUILD_VECTOR,
    486 
    487   /// INSERT_VECTOR_ELT(VECTOR, VAL, IDX) - Returns VECTOR with the element
    488   /// at IDX replaced with VAL. If the type of VAL is larger than the vector
    489   /// element type then VAL is truncated before replacement.
    490   ///
    491   /// If VECTOR is a scalable vector, then IDX may be larger than the minimum
    492   /// vector width. IDX is not first scaled by the runtime scaling factor of
    493   /// VECTOR.
    494   INSERT_VECTOR_ELT,
    495 
    496   /// EXTRACT_VECTOR_ELT(VECTOR, IDX) - Returns a single element from VECTOR
    497   /// identified by the (potentially variable) element number IDX. If the return
    498   /// type is an integer type larger than the element type of the vector, the
    499   /// result is extended to the width of the return type. In that case, the high
    500   /// bits are undefined.
    501   ///
    502   /// If VECTOR is a scalable vector, then IDX may be larger than the minimum
    503   /// vector width. IDX is not first scaled by the runtime scaling factor of
    504   /// VECTOR.
    505   EXTRACT_VECTOR_ELT,
    506 
    507   /// CONCAT_VECTORS(VECTOR0, VECTOR1, ...) - Given a number of values of
    508   /// vector type with the same length and element type, this produces a
    509   /// concatenated vector result value, with length equal to the sum of the
    510   /// lengths of the input vectors. If VECTOR0 is a fixed-width vector, then
    511   /// VECTOR1..VECTORN must all be fixed-width vectors. Similarly, if VECTOR0
    512   /// is a scalable vector, then VECTOR1..VECTORN must all be scalable vectors.
    513   CONCAT_VECTORS,
    514 
    515   /// INSERT_SUBVECTOR(VECTOR1, VECTOR2, IDX) - Returns a vector with VECTOR2
    516   /// inserted into VECTOR1. IDX represents the starting element number at which
    517   /// VECTOR2 will be inserted. IDX must be a constant multiple of T's known
    518   /// minimum vector length. Let the type of VECTOR2 be T, then if T is a
    519   /// scalable vector, IDX is first scaled by the runtime scaling factor of T.
    520   /// The elements of VECTOR1 starting at IDX are overwritten with VECTOR2.
    521   /// Elements IDX through (IDX + num_elements(T) - 1) must be valid VECTOR1
    522   /// indices. If this condition cannot be determined statically but is false at
    523   /// runtime, then the result vector is undefined. The IDX parameter must be a
    524   /// vector index constant type, which for most targets will be an integer
    525   /// pointer type.
    526   ///
    527   /// This operation supports inserting a fixed-width vector into a scalable
    528   /// vector, but not the other way around.
    529   INSERT_SUBVECTOR,
    530 
    531   /// EXTRACT_SUBVECTOR(VECTOR, IDX) - Returns a subvector from VECTOR.
    532   /// Let the result type be T, then IDX represents the starting element number
    533   /// from which a subvector of type T is extracted. IDX must be a constant
    534   /// multiple of T's known minimum vector length. If T is a scalable vector,
    535   /// IDX is first scaled by the runtime scaling factor of T. Elements IDX
    536   /// through (IDX + num_elements(T) - 1) must be valid VECTOR indices. If this
    537   /// condition cannot be determined statically but is false at runtime, then
    538   /// the result vector is undefined. The IDX parameter must be a vector index
    539   /// constant type, which for most targets will be an integer pointer type.
    540   ///
    541   /// This operation supports extracting a fixed-width vector from a scalable
    542   /// vector, but not the other way around.
    543   EXTRACT_SUBVECTOR,
    544 
    545   /// VECTOR_REVERSE(VECTOR) - Returns a vector, of the same type as VECTOR,
    546   /// whose elements are shuffled using the following algorithm:
    547   ///   RESULT[i] = VECTOR[VECTOR.ElementCount - 1 - i]
    548   VECTOR_REVERSE,
    549 
    550   /// VECTOR_SHUFFLE(VEC1, VEC2) - Returns a vector, of the same type as
    551   /// VEC1/VEC2.  A VECTOR_SHUFFLE node also contains an array of constant int
    552   /// values that indicate which value (or undef) each result element will
    553   /// get.  These constant ints are accessible through the
    554   /// ShuffleVectorSDNode class.  This is quite similar to the Altivec
    555   /// 'vperm' instruction, except that the indices must be constants and are
    556   /// in terms of the element size of VEC1/VEC2, not in terms of bytes.
    557   VECTOR_SHUFFLE,
    558 
    559   /// VECTOR_SPLICE(VEC1, VEC2, IMM) - Returns a subvector of the same type as
    560   /// VEC1/VEC2 from CONCAT_VECTORS(VEC1, VEC2), based on the IMM in two ways.
    561   /// Let the result type be T, if IMM is positive it represents the starting
    562   /// element number (an index) from which a subvector of type T is extracted
    563   /// from CONCAT_VECTORS(VEC1, VEC2). If IMM is negative it represents a count
    564   /// specifying the number of trailing elements to extract from VEC1, where the
    565   /// elements of T are selected using the following algorithm:
    566   ///   RESULT[i] = CONCAT_VECTORS(VEC1,VEC2)[VEC1.ElementCount - ABS(IMM) + i]
    567   /// If IMM is not in the range [-VL, VL-1] the result vector is undefined. IMM
    568   /// is a constant integer.
    569   VECTOR_SPLICE,
    570 
    571   /// SCALAR_TO_VECTOR(VAL) - This represents the operation of loading a
    572   /// scalar value into element 0 of the resultant vector type.  The top
    573   /// elements 1 to N-1 of the N-element vector are undefined.  The type
    574   /// of the operand must match the vector element type, except when they
    575   /// are integer types.  In this case the operand is allowed to be wider
    576   /// than the vector element type, and is implicitly truncated to it.
    577   SCALAR_TO_VECTOR,
    578 
    579   /// SPLAT_VECTOR(VAL) - Returns a vector with the scalar value VAL
    580   /// duplicated in all lanes. The type of the operand must match the vector
    581   /// element type, except when they are integer types.  In this case the
    582   /// operand is allowed to be wider than the vector element type, and is
    583   /// implicitly truncated to it.
    584   SPLAT_VECTOR,
    585 
    586   /// SPLAT_VECTOR_PARTS(SCALAR1, SCALAR2, ...) - Returns a vector with the
    587   /// scalar values joined together and then duplicated in all lanes. This
    588   /// represents a SPLAT_VECTOR that has had its scalar operand expanded. This
    589   /// allows representing a 64-bit splat on a target with 32-bit integers. The
    590   /// total width of the scalars must cover the element width. SCALAR1 contains
    591   /// the least significant bits of the value regardless of endianness and all
    592   /// scalars should have the same type.
    593   SPLAT_VECTOR_PARTS,
    594 
    595   /// STEP_VECTOR(IMM) - Returns a scalable vector whose lanes are comprised
    596   /// of a linear sequence of unsigned values starting from 0 with a step of
    597   /// IMM, where IMM must be a vector index constant integer value which must
    598   /// fit in the vector element type.
    599   /// Note that IMM may be a smaller type than the vector element type, in
    600   /// which case the step is implicitly sign-extended to the vector element
    601   /// type. IMM may also be a larger type than the vector element type, in
    602   /// which case the step is implicitly truncated to the vector element type.
    603   /// The operation does not support returning fixed-width vectors or
    604   /// non-constant operands. If the sequence value exceeds the limit allowed
    605   /// for the element type then the values for those lanes are undefined.
    606   STEP_VECTOR,
    607 
    608   /// MULHU/MULHS - Multiply high - Multiply two integers of type iN,
    609   /// producing an unsigned/signed value of type i[2*N], then return the top
    610   /// part.
    611   MULHU,
    612   MULHS,
    613 
    614   /// [US]{MIN/MAX} - Binary minimum or maximum of signed or unsigned
    615   /// integers.
    616   SMIN,
    617   SMAX,
    618   UMIN,
    619   UMAX,
    620 
    621   /// Bitwise operators - logical and, logical or, logical xor.
    622   AND,
    623   OR,
    624   XOR,
    625 
    626   /// ABS - Determine the unsigned absolute value of a signed integer value of
    627   /// the same bitwidth.
    628   /// Note: A value of INT_MIN will return INT_MIN, no saturation or overflow
    629   /// is performed.
    630   ABS,
    631 
    632   /// Shift and rotation operations.  After legalization, the type of the
    633   /// shift amount is known to be TLI.getShiftAmountTy().  Before legalization
    634   /// the shift amount can be any type, but care must be taken to ensure it is
    635   /// large enough.  TLI.getShiftAmountTy() is i8 on some targets, but before
    636   /// legalization, types like i1024 can occur and i8 doesn't have enough bits
    637   /// to represent the shift amount.
    638   /// When the 1st operand is a vector, the shift amount must be in the same
    639   /// type. (TLI.getShiftAmountTy() will return the same type when the input
    640   /// type is a vector.)
    641   /// For rotates and funnel shifts, the shift amount is treated as an unsigned
    642   /// amount modulo the element size of the first operand.
    643   ///
    644   /// Funnel 'double' shifts take 3 operands, 2 inputs and the shift amount.
    645   /// fshl(X,Y,Z): (X << (Z % BW)) | (Y >> (BW - (Z % BW)))
    646   /// fshr(X,Y,Z): (X << (BW - (Z % BW))) | (Y >> (Z % BW))
    647   SHL,
    648   SRA,
    649   SRL,
    650   ROTL,
    651   ROTR,
    652   FSHL,
    653   FSHR,
    654 
    655   /// Byte Swap and Counting operators.
    656   BSWAP,
    657   CTTZ,
    658   CTLZ,
    659   CTPOP,
    660   BITREVERSE,
    661   PARITY,
    662 
    663   /// Bit counting operators with an undefined result for zero inputs.
    664   CTTZ_ZERO_UNDEF,
    665   CTLZ_ZERO_UNDEF,
    666 
    667   /// Select(COND, TRUEVAL, FALSEVAL).  If the type of the boolean COND is not
    668   /// i1 then the high bits must conform to getBooleanContents.
    669   SELECT,
    670 
    671   /// Select with a vector condition (op #0) and two vector operands (ops #1
    672   /// and #2), returning a vector result.  All vectors have the same length.
    673   /// Much like the scalar select and setcc, each bit in the condition selects
    674   /// whether the corresponding result element is taken from op #1 or op #2.
    675   /// At first, the VSELECT condition is of vXi1 type. Later, targets may
    676   /// change the condition type in order to match the VSELECT node using a
    677   /// pattern. The condition follows the BooleanContent format of the target.
    678   VSELECT,
    679 
    680   /// Select with condition operator - This selects between a true value and
    681   /// a false value (ops #2 and #3) based on the boolean result of comparing
    682   /// the lhs and rhs (ops #0 and #1) of a conditional expression with the
    683   /// condition code in op #4, a CondCodeSDNode.
    684   SELECT_CC,
    685 
    686   /// SetCC operator - This evaluates to a true value iff the condition is
    687   /// true.  If the result value type is not i1 then the high bits conform
    688   /// to getBooleanContents.  The operands to this are the left and right
    689   /// operands to compare (ops #0, and #1) and the condition code to compare
    690   /// them with (op #2) as a CondCodeSDNode. If the operands are vector types
    691   /// then the result type must also be a vector type.
    692   SETCC,
    693 
    694   /// Like SetCC, ops #0 and #1 are the LHS and RHS operands to compare, but
    695   /// op #2 is a boolean indicating if there is an incoming carry. This
    696   /// operator checks the result of "LHS - RHS - Carry", and can be used to
    697   /// compare two wide integers:
    698   /// (setcccarry lhshi rhshi (subcarry lhslo rhslo) cc).
    699   /// Only valid for integers.
    700   SETCCCARRY,
    701 
    702   /// SHL_PARTS/SRA_PARTS/SRL_PARTS - These operators are used for expanded
    703   /// integer shift operations.  The operation ordering is:
    704   ///       [Lo,Hi] = op [LoLHS,HiLHS], Amt
    705   SHL_PARTS,
    706   SRA_PARTS,
    707   SRL_PARTS,
    708 
    709   /// Conversion operators.  These are all single input single output
    710   /// operations.  For all of these, the result type must be strictly
    711   /// wider or narrower (depending on the operation) than the source
    712   /// type.
    713 
    714   /// SIGN_EXTEND - Used for integer types, replicating the sign bit
    715   /// into new bits.
    716   SIGN_EXTEND,
    717 
    718   /// ZERO_EXTEND - Used for integer types, zeroing the new bits.
    719   ZERO_EXTEND,
    720 
    721   /// ANY_EXTEND - Used for integer types.  The high bits are undefined.
    722   ANY_EXTEND,
    723 
    724   /// TRUNCATE - Completely drop the high bits.
    725   TRUNCATE,
    726 
    727   /// [SU]INT_TO_FP - These operators convert integers (whose interpreted sign
    728   /// depends on the first letter) to floating point.
    729   SINT_TO_FP,
    730   UINT_TO_FP,
    731 
    732   /// SIGN_EXTEND_INREG - This operator atomically performs a SHL/SRA pair to
    733   /// sign extend a small value in a large integer register (e.g. sign
    734   /// extending the low 8 bits of a 32-bit register to fill the top 24 bits
    735   /// with the 7th bit).  The size of the smaller type is indicated by the 1th
    736   /// operand, a ValueType node.
    737   SIGN_EXTEND_INREG,
    738 
    739   /// ANY_EXTEND_VECTOR_INREG(Vector) - This operator represents an
    740   /// in-register any-extension of the low lanes of an integer vector. The
    741   /// result type must have fewer elements than the operand type, and those
    742   /// elements must be larger integer types such that the total size of the
    743   /// operand type is less than or equal to the size of the result type. Each
    744   /// of the low operand elements is any-extended into the corresponding,
    745   /// wider result elements with the high bits becoming undef.
    746   /// NOTE: The type legalizer prefers to make the operand and result size
    747   /// the same to allow expansion to shuffle vector during op legalization.
    748   ANY_EXTEND_VECTOR_INREG,
    749 
    750   /// SIGN_EXTEND_VECTOR_INREG(Vector) - This operator represents an
    751   /// in-register sign-extension of the low lanes of an integer vector. The
    752   /// result type must have fewer elements than the operand type, and those
    753   /// elements must be larger integer types such that the total size of the
    754   /// operand type is less than or equal to the size of the result type. Each
    755   /// of the low operand elements is sign-extended into the corresponding,
    756   /// wider result elements.
    757   /// NOTE: The type legalizer prefers to make the operand and result size
    758   /// the same to allow expansion to shuffle vector during op legalization.
    759   SIGN_EXTEND_VECTOR_INREG,
    760 
    761   /// ZERO_EXTEND_VECTOR_INREG(Vector) - This operator represents an
    762   /// in-register zero-extension of the low lanes of an integer vector. The
    763   /// result type must have fewer elements than the operand type, and those
    764   /// elements must be larger integer types such that the total size of the
    765   /// operand type is less than or equal to the size of the result type. Each
    766   /// of the low operand elements is zero-extended into the corresponding,
    767   /// wider result elements.
    768   /// NOTE: The type legalizer prefers to make the operand and result size
    769   /// the same to allow expansion to shuffle vector during op legalization.
    770   ZERO_EXTEND_VECTOR_INREG,
    771 
    772   /// FP_TO_[US]INT - Convert a floating point value to a signed or unsigned
    773   /// integer. These have the same semantics as fptosi and fptoui in IR. If
    774   /// the FP value cannot fit in the integer type, the results are undefined.
    775   FP_TO_SINT,
    776   FP_TO_UINT,
    777 
    778   /// FP_TO_[US]INT_SAT - Convert floating point value in operand 0 to a
    779   /// signed or unsigned scalar integer type given in operand 1 with the
    780   /// following semantics:
    781   ///
    782   ///  * If the value is NaN, zero is returned.
    783   ///  * If the value is larger/smaller than the largest/smallest integer,
    784   ///    the largest/smallest integer is returned (saturation).
    785   ///  * Otherwise the result of rounding the value towards zero is returned.
    786   ///
    787   /// The scalar width of the type given in operand 1 must be equal to, or
    788   /// smaller than, the scalar result type width. It may end up being smaller
    789   /// than the result width as a result of integer type legalization.
    790   FP_TO_SINT_SAT,
    791   FP_TO_UINT_SAT,
    792 
    793   /// X = FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type
    794   /// down to the precision of the destination VT.  TRUNC is a flag, which is
    795   /// always an integer that is zero or one.  If TRUNC is 0, this is a
    796   /// normal rounding, if it is 1, this FP_ROUND is known to not change the
    797   /// value of Y.
    798   ///
    799   /// The TRUNC = 1 case is used in cases where we know that the value will
    800   /// not be modified by the node, because Y is not using any of the extra
    801   /// precision of source type.  This allows certain transformations like
    802   /// FP_EXTEND(FP_ROUND(X,1)) -> X which are not safe for
    803   /// FP_EXTEND(FP_ROUND(X,0)) because the extra bits aren't removed.
    804   FP_ROUND,
    805 
    806   /// Returns current rounding mode:
    807   /// -1 Undefined
    808   ///  0 Round to 0
    809   ///  1 Round to nearest, ties to even
    810   ///  2 Round to +inf
    811   ///  3 Round to -inf
    812   ///  4 Round to nearest, ties to zero
    813   /// Result is rounding mode and chain. Input is a chain.
    814   /// TODO: Rename this node to GET_ROUNDING.
    815   FLT_ROUNDS_,
    816 
    817   /// Set rounding mode.
    818   /// The first operand is a chain pointer. The second specifies the required
    819   /// rounding mode, encoded in the same way as used in '``FLT_ROUNDS_``'.
    820   SET_ROUNDING,
    821 
    822   /// X = FP_EXTEND(Y) - Extend a smaller FP type into a larger FP type.
    823   FP_EXTEND,
    824 
    825   /// BITCAST - This operator converts between integer, vector and FP
    826   /// values, as if the value was stored to memory with one type and loaded
    827   /// from the same address with the other type (or equivalently for vector
    828   /// format conversions, etc).  The source and result are required to have
    829   /// the same bit size (e.g.  f32 <-> i32).  This can also be used for
    830   /// int-to-int or fp-to-fp conversions, but that is a noop, deleted by
    831   /// getNode().
    832   ///
    833   /// This operator is subtly different from the bitcast instruction from
    834   /// LLVM-IR since this node may change the bits in the register. For
    835   /// example, this occurs on big-endian NEON and big-endian MSA where the
    836   /// layout of the bits in the register depends on the vector type and this
    837   /// operator acts as a shuffle operation for some vector type combinations.
    838   BITCAST,
    839 
    840   /// ADDRSPACECAST - This operator converts between pointers of different
    841   /// address spaces.
    842   ADDRSPACECAST,
    843 
    844   /// FP16_TO_FP, FP_TO_FP16 - These operators are used to perform promotions
    845   /// and truncation for half-precision (16 bit) floating numbers. These nodes
    846   /// form a semi-softened interface for dealing with f16 (as an i16), which
    847   /// is often a storage-only type but has native conversions.
    848   FP16_TO_FP,
    849   FP_TO_FP16,
    850   STRICT_FP16_TO_FP,
    851   STRICT_FP_TO_FP16,
    852 
    853   /// Perform various unary floating-point operations inspired by libm. For
    854   /// FPOWI, the result is undefined if if the integer operand doesn't fit
    855   /// into 32 bits.
    856   FNEG,
    857   FABS,
    858   FSQRT,
    859   FCBRT,
    860   FSIN,
    861   FCOS,
    862   FPOWI,
    863   FPOW,
    864   FLOG,
    865   FLOG2,
    866   FLOG10,
    867   FEXP,
    868   FEXP2,
    869   FCEIL,
    870   FTRUNC,
    871   FRINT,
    872   FNEARBYINT,
    873   FROUND,
    874   FROUNDEVEN,
    875   FFLOOR,
    876   LROUND,
    877   LLROUND,
    878   LRINT,
    879   LLRINT,
    880 
    881   /// FMINNUM/FMAXNUM - Perform floating-point minimum or maximum on two
    882   /// values.
    883   //
    884   /// In the case where a single input is a NaN (either signaling or quiet),
    885   /// the non-NaN input is returned.
    886   ///
    887   /// The return value of (FMINNUM 0.0, -0.0) could be either 0.0 or -0.0.
    888   FMINNUM,
    889   FMAXNUM,
    890 
    891   /// FMINNUM_IEEE/FMAXNUM_IEEE - Perform floating-point minimum or maximum on
    892   /// two values, following the IEEE-754 2008 definition. This differs from
    893   /// FMINNUM/FMAXNUM in the handling of signaling NaNs. If one input is a
    894   /// signaling NaN, returns a quiet NaN.
    895   FMINNUM_IEEE,
    896   FMAXNUM_IEEE,
    897 
    898   /// FMINIMUM/FMAXIMUM - NaN-propagating minimum/maximum that also treat -0.0
    899   /// as less than 0.0. While FMINNUM_IEEE/FMAXNUM_IEEE follow IEEE 754-2008
    900   /// semantics, FMINIMUM/FMAXIMUM follow IEEE 754-2018 draft semantics.
    901   FMINIMUM,
    902   FMAXIMUM,
    903 
    904   /// FSINCOS - Compute both fsin and fcos as a single operation.
    905   FSINCOS,
    906 
    907   /// LOAD and STORE have token chains as their first operand, then the same
    908   /// operands as an LLVM load/store instruction, then an offset node that
    909   /// is added / subtracted from the base pointer to form the address (for
    910   /// indexed memory ops).
    911   LOAD,
    912   STORE,
    913 
    914   /// DYNAMIC_STACKALLOC - Allocate some number of bytes on the stack aligned
    915   /// to a specified boundary.  This node always has two return values: a new
    916   /// stack pointer value and a chain. The first operand is the token chain,
    917   /// the second is the number of bytes to allocate, and the third is the
    918   /// alignment boundary.  The size is guaranteed to be a multiple of the
    919   /// stack alignment, and the alignment is guaranteed to be bigger than the
    920   /// stack alignment (if required) or 0 to get standard stack alignment.
    921   DYNAMIC_STACKALLOC,
    922 
    923   /// Control flow instructions.  These all have token chains.
    924 
    925   /// BR - Unconditional branch.  The first operand is the chain
    926   /// operand, the second is the MBB to branch to.
    927   BR,
    928 
    929   /// BRIND - Indirect branch.  The first operand is the chain, the second
    930   /// is the value to branch to, which must be of the same type as the
    931   /// target's pointer type.
    932   BRIND,
    933 
    934   /// BR_JT - Jumptable branch. The first operand is the chain, the second
    935   /// is the jumptable index, the last one is the jumptable entry index.
    936   BR_JT,
    937 
    938   /// BRCOND - Conditional branch.  The first operand is the chain, the
    939   /// second is the condition, the third is the block to branch to if the
    940   /// condition is true.  If the type of the condition is not i1, then the
    941   /// high bits must conform to getBooleanContents. If the condition is undef,
    942   /// it nondeterministically jumps to the block.
    943   /// TODO: Its semantics w.r.t undef requires further discussion; we need to
    944   /// make it sure that it is consistent with optimizations in MIR & the
    945   /// meaning of IMPLICIT_DEF. See https://reviews.llvm.org/D92015
    946   BRCOND,
    947 
    948   /// BR_CC - Conditional branch.  The behavior is like that of SELECT_CC, in
    949   /// that the condition is represented as condition code, and two nodes to
    950   /// compare, rather than as a combined SetCC node.  The operands in order
    951   /// are chain, cc, lhs, rhs, block to branch to if condition is true. If
    952   /// condition is undef, it nondeterministically jumps to the block.
    953   BR_CC,
    954 
    955   /// INLINEASM - Represents an inline asm block.  This node always has two
    956   /// return values: a chain and a flag result.  The inputs are as follows:
    957   ///   Operand #0  : Input chain.
    958   ///   Operand #1  : a ExternalSymbolSDNode with a pointer to the asm string.
    959   ///   Operand #2  : a MDNodeSDNode with the !srcloc metadata.
    960   ///   Operand #3  : HasSideEffect, IsAlignStack bits.
    961   ///   After this, it is followed by a list of operands with this format:
    962   ///     ConstantSDNode: Flags that encode whether it is a mem or not, the
    963   ///                     of operands that follow, etc.  See InlineAsm.h.
    964   ///     ... however many operands ...
    965   ///   Operand #last: Optional, an incoming flag.
    966   ///
    967   /// The variable width operands are required to represent target addressing
    968   /// modes as a single "operand", even though they may have multiple
    969   /// SDOperands.
    970   INLINEASM,
    971 
    972   /// INLINEASM_BR - Branching version of inline asm. Used by asm-goto.
    973   INLINEASM_BR,
    974 
    975   /// EH_LABEL - Represents a label in mid basic block used to track
    976   /// locations needed for debug and exception handling tables.  These nodes
    977   /// take a chain as input and return a chain.
    978   EH_LABEL,
    979 
    980   /// ANNOTATION_LABEL - Represents a mid basic block label used by
    981   /// annotations. This should remain within the basic block and be ordered
    982   /// with respect to other call instructions, but loads and stores may float
    983   /// past it.
    984   ANNOTATION_LABEL,
    985 
    986   /// CATCHRET - Represents a return from a catch block funclet. Used for
    987   /// MSVC compatible exception handling. Takes a chain operand and a
    988   /// destination basic block operand.
    989   CATCHRET,
    990 
    991   /// CLEANUPRET - Represents a return from a cleanup block funclet.  Used for
    992   /// MSVC compatible exception handling. Takes only a chain operand.
    993   CLEANUPRET,
    994 
    995   /// STACKSAVE - STACKSAVE has one operand, an input chain.  It produces a
    996   /// value, the same type as the pointer type for the system, and an output
    997   /// chain.
    998   STACKSAVE,
    999 
   1000   /// STACKRESTORE has two operands, an input chain and a pointer to restore
   1001   /// to it returns an output chain.
   1002   STACKRESTORE,
   1003 
   1004   /// CALLSEQ_START/CALLSEQ_END - These operators mark the beginning and end
   1005   /// of a call sequence, and carry arbitrary information that target might
   1006   /// want to know.  The first operand is a chain, the rest are specified by
   1007   /// the target and not touched by the DAG optimizers.
   1008   /// Targets that may use stack to pass call arguments define additional
   1009   /// operands:
   1010   /// - size of the call frame part that must be set up within the
   1011   ///   CALLSEQ_START..CALLSEQ_END pair,
   1012   /// - part of the call frame prepared prior to CALLSEQ_START.
   1013   /// Both these parameters must be constants, their sum is the total call
   1014   /// frame size.
   1015   /// CALLSEQ_START..CALLSEQ_END pairs may not be nested.
   1016   CALLSEQ_START, // Beginning of a call sequence
   1017   CALLSEQ_END,   // End of a call sequence
   1018 
   1019   /// VAARG - VAARG has four operands: an input chain, a pointer, a SRCVALUE,
   1020   /// and the alignment. It returns a pair of values: the vaarg value and a
   1021   /// new chain.
   1022   VAARG,
   1023 
   1024   /// VACOPY - VACOPY has 5 operands: an input chain, a destination pointer,
   1025   /// a source pointer, a SRCVALUE for the destination, and a SRCVALUE for the
   1026   /// source.
   1027   VACOPY,
   1028 
   1029   /// VAEND, VASTART - VAEND and VASTART have three operands: an input chain,
   1030   /// pointer, and a SRCVALUE.
   1031   VAEND,
   1032   VASTART,
   1033 
   1034   // PREALLOCATED_SETUP - This has 2 operands: an input chain and a SRCVALUE
   1035   // with the preallocated call Value.
   1036   PREALLOCATED_SETUP,
   1037   // PREALLOCATED_ARG - This has 3 operands: an input chain, a SRCVALUE
   1038   // with the preallocated call Value, and a constant int.
   1039   PREALLOCATED_ARG,
   1040 
   1041   /// SRCVALUE - This is a node type that holds a Value* that is used to
   1042   /// make reference to a value in the LLVM IR.
   1043   SRCVALUE,
   1044 
   1045   /// MDNODE_SDNODE - This is a node that holdes an MDNode*, which is used to
   1046   /// reference metadata in the IR.
   1047   MDNODE_SDNODE,
   1048 
   1049   /// PCMARKER - This corresponds to the pcmarker intrinsic.
   1050   PCMARKER,
   1051 
   1052   /// READCYCLECOUNTER - This corresponds to the readcyclecounter intrinsic.
   1053   /// It produces a chain and one i64 value. The only operand is a chain.
   1054   /// If i64 is not legal, the result will be expanded into smaller values.
   1055   /// Still, it returns an i64, so targets should set legality for i64.
   1056   /// The result is the content of the architecture-specific cycle
   1057   /// counter-like register (or other high accuracy low latency clock source).
   1058   READCYCLECOUNTER,
   1059 
   1060   /// HANDLENODE node - Used as a handle for various purposes.
   1061   HANDLENODE,
   1062 
   1063   /// INIT_TRAMPOLINE - This corresponds to the init_trampoline intrinsic.  It
   1064   /// takes as input a token chain, the pointer to the trampoline, the pointer
   1065   /// to the nested function, the pointer to pass for the 'nest' parameter, a
   1066   /// SRCVALUE for the trampoline and another for the nested function
   1067   /// (allowing targets to access the original Function*).
   1068   /// It produces a token chain as output.
   1069   INIT_TRAMPOLINE,
   1070 
   1071   /// ADJUST_TRAMPOLINE - This corresponds to the adjust_trampoline intrinsic.
   1072   /// It takes a pointer to the trampoline and produces a (possibly) new
   1073   /// pointer to the same trampoline with platform-specific adjustments
   1074   /// applied.  The pointer it returns points to an executable block of code.
   1075   ADJUST_TRAMPOLINE,
   1076 
   1077   /// TRAP - Trapping instruction
   1078   TRAP,
   1079 
   1080   /// DEBUGTRAP - Trap intended to get the attention of a debugger.
   1081   DEBUGTRAP,
   1082 
   1083   /// UBSANTRAP - Trap with an immediate describing the kind of sanitizer
   1084   /// failure.
   1085   UBSANTRAP,
   1086 
   1087   /// PREFETCH - This corresponds to a prefetch intrinsic. The first operand
   1088   /// is the chain.  The other operands are the address to prefetch,
   1089   /// read / write specifier, locality specifier and instruction / data cache
   1090   /// specifier.
   1091   PREFETCH,
   1092 
   1093   /// OUTCHAIN = ATOMIC_FENCE(INCHAIN, ordering, scope)
   1094   /// This corresponds to the fence instruction. It takes an input chain, and
   1095   /// two integer constants: an AtomicOrdering and a SynchronizationScope.
   1096   ATOMIC_FENCE,
   1097 
   1098   /// Val, OUTCHAIN = ATOMIC_LOAD(INCHAIN, ptr)
   1099   /// This corresponds to "load atomic" instruction.
   1100   ATOMIC_LOAD,
   1101 
   1102   /// OUTCHAIN = ATOMIC_STORE(INCHAIN, ptr, val)
   1103   /// This corresponds to "store atomic" instruction.
   1104   ATOMIC_STORE,
   1105 
   1106   /// Val, OUTCHAIN = ATOMIC_CMP_SWAP(INCHAIN, ptr, cmp, swap)
   1107   /// For double-word atomic operations:
   1108   /// ValLo, ValHi, OUTCHAIN = ATOMIC_CMP_SWAP(INCHAIN, ptr, cmpLo, cmpHi,
   1109   ///                                          swapLo, swapHi)
   1110   /// This corresponds to the cmpxchg instruction.
   1111   ATOMIC_CMP_SWAP,
   1112 
   1113   /// Val, Success, OUTCHAIN
   1114   ///     = ATOMIC_CMP_SWAP_WITH_SUCCESS(INCHAIN, ptr, cmp, swap)
   1115   /// N.b. this is still a strong cmpxchg operation, so
   1116   /// Success == "Val == cmp".
   1117   ATOMIC_CMP_SWAP_WITH_SUCCESS,
   1118 
   1119   /// Val, OUTCHAIN = ATOMIC_SWAP(INCHAIN, ptr, amt)
   1120   /// Val, OUTCHAIN = ATOMIC_LOAD_[OpName](INCHAIN, ptr, amt)
   1121   /// For double-word atomic operations:
   1122   /// ValLo, ValHi, OUTCHAIN = ATOMIC_SWAP(INCHAIN, ptr, amtLo, amtHi)
   1123   /// ValLo, ValHi, OUTCHAIN = ATOMIC_LOAD_[OpName](INCHAIN, ptr, amtLo, amtHi)
   1124   /// These correspond to the atomicrmw instruction.
   1125   ATOMIC_SWAP,
   1126   ATOMIC_LOAD_ADD,
   1127   ATOMIC_LOAD_SUB,
   1128   ATOMIC_LOAD_AND,
   1129   ATOMIC_LOAD_CLR,
   1130   ATOMIC_LOAD_OR,
   1131   ATOMIC_LOAD_XOR,
   1132   ATOMIC_LOAD_NAND,
   1133   ATOMIC_LOAD_MIN,
   1134   ATOMIC_LOAD_MAX,
   1135   ATOMIC_LOAD_UMIN,
   1136   ATOMIC_LOAD_UMAX,
   1137   ATOMIC_LOAD_FADD,
   1138   ATOMIC_LOAD_FSUB,
   1139 
   1140   // Masked load and store - consecutive vector load and store operations
   1141   // with additional mask operand that prevents memory accesses to the
   1142   // masked-off lanes.
   1143   //
   1144   // Val, OutChain = MLOAD(BasePtr, Mask, PassThru)
   1145   // OutChain = MSTORE(Value, BasePtr, Mask)
   1146   MLOAD,
   1147   MSTORE,
   1148 
   1149   // Masked gather and scatter - load and store operations for a vector of
   1150   // random addresses with additional mask operand that prevents memory
   1151   // accesses to the masked-off lanes.
   1152   //
   1153   // Val, OutChain = GATHER(InChain, PassThru, Mask, BasePtr, Index, Scale)
   1154   // OutChain = SCATTER(InChain, Value, Mask, BasePtr, Index, Scale)
   1155   //
   1156   // The Index operand can have more vector elements than the other operands
   1157   // due to type legalization. The extra elements are ignored.
   1158   MGATHER,
   1159   MSCATTER,
   1160 
   1161   /// This corresponds to the llvm.lifetime.* intrinsics. The first operand
   1162   /// is the chain and the second operand is the alloca pointer.
   1163   LIFETIME_START,
   1164   LIFETIME_END,
   1165 
   1166   /// GC_TRANSITION_START/GC_TRANSITION_END - These operators mark the
   1167   /// beginning and end of GC transition  sequence, and carry arbitrary
   1168   /// information that target might need for lowering.  The first operand is
   1169   /// a chain, the rest are specified by the target and not touched by the DAG
   1170   /// optimizers. GC_TRANSITION_START..GC_TRANSITION_END pairs may not be
   1171   /// nested.
   1172   GC_TRANSITION_START,
   1173   GC_TRANSITION_END,
   1174 
   1175   /// GET_DYNAMIC_AREA_OFFSET - get offset from native SP to the address of
   1176   /// the most recent dynamic alloca. For most targets that would be 0, but
   1177   /// for some others (e.g. PowerPC, PowerPC64) that would be compile-time
   1178   /// known nonzero constant. The only operand here is the chain.
   1179   GET_DYNAMIC_AREA_OFFSET,
   1180 
   1181   /// Pseudo probe for AutoFDO, as a place holder in a basic block to improve
   1182   /// the sample counts quality.
   1183   PSEUDO_PROBE,
   1184 
   1185   /// VSCALE(IMM) - Returns the runtime scaling factor used to calculate the
   1186   /// number of elements within a scalable vector. IMM is a constant integer
   1187   /// multiplier that is applied to the runtime value.
   1188   VSCALE,
   1189 
   1190   /// Generic reduction nodes. These nodes represent horizontal vector
   1191   /// reduction operations, producing a scalar result.
   1192   /// The SEQ variants perform reductions in sequential order. The first
   1193   /// operand is an initial scalar accumulator value, and the second operand
   1194   /// is the vector to reduce.
   1195   /// E.g. RES = VECREDUCE_SEQ_FADD f32 ACC, <4 x f32> SRC_VEC
   1196   ///  ... is equivalent to
   1197   /// RES = (((ACC + SRC_VEC[0]) + SRC_VEC[1]) + SRC_VEC[2]) + SRC_VEC[3]
   1198   VECREDUCE_SEQ_FADD,
   1199   VECREDUCE_SEQ_FMUL,
   1200 
   1201   /// These reductions have relaxed evaluation order semantics, and have a
   1202   /// single vector operand. The order of evaluation is unspecified. For
   1203   /// pow-of-2 vectors, one valid legalizer expansion is to use a tree
   1204   /// reduction, i.e.:
   1205   /// For RES = VECREDUCE_FADD <8 x f16> SRC_VEC
   1206   ///   PART_RDX = FADD SRC_VEC[0:3], SRC_VEC[4:7]
   1207   ///   PART_RDX2 = FADD PART_RDX[0:1], PART_RDX[2:3]
   1208   ///   RES = FADD PART_RDX2[0], PART_RDX2[1]
   1209   /// For non-pow-2 vectors, this can be computed by extracting each element
   1210   /// and performing the operation as if it were scalarized.
   1211   VECREDUCE_FADD,
   1212   VECREDUCE_FMUL,
   1213   /// FMIN/FMAX nodes can have flags, for NaN/NoNaN variants.
   1214   VECREDUCE_FMAX,
   1215   VECREDUCE_FMIN,
   1216   /// Integer reductions may have a result type larger than the vector element
   1217   /// type. However, the reduction is performed using the vector element type
   1218   /// and the value in the top bits is unspecified.
   1219   VECREDUCE_ADD,
   1220   VECREDUCE_MUL,
   1221   VECREDUCE_AND,
   1222   VECREDUCE_OR,
   1223   VECREDUCE_XOR,
   1224   VECREDUCE_SMAX,
   1225   VECREDUCE_SMIN,
   1226   VECREDUCE_UMAX,
   1227   VECREDUCE_UMIN,
   1228 
   1229 // Vector Predication
   1230 #define BEGIN_REGISTER_VP_SDNODE(VPSDID, ...) VPSDID,
   1231 #include "llvm/IR/VPIntrinsics.def"
   1232 
   1233   /// BUILTIN_OP_END - This must be the last enum value in this list.
   1234   /// The target-specific pre-isel opcode values start here.
   1235   BUILTIN_OP_END
   1236 };
   1237 
   1238 /// FIRST_TARGET_STRICTFP_OPCODE - Target-specific pre-isel operations
   1239 /// which cannot raise FP exceptions should be less than this value.
   1240 /// Those that do must not be less than this value.
   1241 static const int FIRST_TARGET_STRICTFP_OPCODE = BUILTIN_OP_END + 400;
   1242 
   1243 /// FIRST_TARGET_MEMORY_OPCODE - Target-specific pre-isel operations
   1244 /// which do not reference a specific memory location should be less than
   1245 /// this value. Those that do must not be less than this value, and can
   1246 /// be used with SelectionDAG::getMemIntrinsicNode.
   1247 static const int FIRST_TARGET_MEMORY_OPCODE = BUILTIN_OP_END + 500;
   1248 
   1249 /// Get underlying scalar opcode for VECREDUCE opcode.
   1250 /// For example ISD::AND for ISD::VECREDUCE_AND.
   1251 NodeType getVecReduceBaseOpcode(unsigned VecReduceOpcode);
   1252 
   1253 /// Whether this is a vector-predicated Opcode.
   1254 bool isVPOpcode(unsigned Opcode);
   1255 
   1256 /// The operand position of the vector mask.
   1257 Optional<unsigned> getVPMaskIdx(unsigned Opcode);
   1258 
   1259 /// The operand position of the explicit vector length parameter.
   1260 Optional<unsigned> getVPExplicitVectorLengthIdx(unsigned Opcode);
   1261 
   1262 //===--------------------------------------------------------------------===//
   1263 /// MemIndexedMode enum - This enum defines the load / store indexed
   1264 /// addressing modes.
   1265 ///
   1266 /// UNINDEXED    "Normal" load / store. The effective address is already
   1267 ///              computed and is available in the base pointer. The offset
   1268 ///              operand is always undefined. In addition to producing a
   1269 ///              chain, an unindexed load produces one value (result of the
   1270 ///              load); an unindexed store does not produce a value.
   1271 ///
   1272 /// PRE_INC      Similar to the unindexed mode where the effective address is
   1273 /// PRE_DEC      the value of the base pointer add / subtract the offset.
   1274 ///              It considers the computation as being folded into the load /
   1275 ///              store operation (i.e. the load / store does the address
   1276 ///              computation as well as performing the memory transaction).
   1277 ///              The base operand is always undefined. In addition to
   1278 ///              producing a chain, pre-indexed load produces two values
   1279 ///              (result of the load and the result of the address
   1280 ///              computation); a pre-indexed store produces one value (result
   1281 ///              of the address computation).
   1282 ///
   1283 /// POST_INC     The effective address is the value of the base pointer. The
   1284 /// POST_DEC     value of the offset operand is then added to / subtracted
   1285 ///              from the base after memory transaction. In addition to
   1286 ///              producing a chain, post-indexed load produces two values
   1287 ///              (the result of the load and the result of the base +/- offset
   1288 ///              computation); a post-indexed store produces one value (the
   1289 ///              the result of the base +/- offset computation).
   1290 enum MemIndexedMode { UNINDEXED = 0, PRE_INC, PRE_DEC, POST_INC, POST_DEC };
   1291 
   1292 static const int LAST_INDEXED_MODE = POST_DEC + 1;
   1293 
   1294 //===--------------------------------------------------------------------===//
   1295 /// MemIndexType enum - This enum defines how to interpret MGATHER/SCATTER's
   1296 /// index parameter when calculating addresses.
   1297 ///
   1298 /// SIGNED_SCALED     Addr = Base + ((signed)Index * sizeof(element))
   1299 /// SIGNED_UNSCALED   Addr = Base + (signed)Index
   1300 /// UNSIGNED_SCALED   Addr = Base + ((unsigned)Index * sizeof(element))
   1301 /// UNSIGNED_UNSCALED Addr = Base + (unsigned)Index
   1302 enum MemIndexType {
   1303   SIGNED_SCALED = 0,
   1304   SIGNED_UNSCALED,
   1305   UNSIGNED_SCALED,
   1306   UNSIGNED_UNSCALED
   1307 };
   1308 
   1309 static const int LAST_MEM_INDEX_TYPE = UNSIGNED_UNSCALED + 1;
   1310 
   1311 //===--------------------------------------------------------------------===//
   1312 /// LoadExtType enum - This enum defines the three variants of LOADEXT
   1313 /// (load with extension).
   1314 ///
   1315 /// SEXTLOAD loads the integer operand and sign extends it to a larger
   1316 ///          integer result type.
   1317 /// ZEXTLOAD loads the integer operand and zero extends it to a larger
   1318 ///          integer result type.
   1319 /// EXTLOAD  is used for two things: floating point extending loads and
   1320 ///          integer extending loads [the top bits are undefined].
   1321 enum LoadExtType { NON_EXTLOAD = 0, EXTLOAD, SEXTLOAD, ZEXTLOAD };
   1322 
   1323 static const int LAST_LOADEXT_TYPE = ZEXTLOAD + 1;
   1324 
   1325 NodeType getExtForLoadExtType(bool IsFP, LoadExtType);
   1326 
   1327 //===--------------------------------------------------------------------===//
   1328 /// ISD::CondCode enum - These are ordered carefully to make the bitfields
   1329 /// below work out, when considering SETFALSE (something that never exists
   1330 /// dynamically) as 0.  "U" -> Unsigned (for integer operands) or Unordered
   1331 /// (for floating point), "L" -> Less than, "G" -> Greater than, "E" -> Equal
   1332 /// to.  If the "N" column is 1, the result of the comparison is undefined if
   1333 /// the input is a NAN.
   1334 ///
   1335 /// All of these (except for the 'always folded ops') should be handled for
   1336 /// floating point.  For integer, only the SETEQ,SETNE,SETLT,SETLE,SETGT,
   1337 /// SETGE,SETULT,SETULE,SETUGT, and SETUGE opcodes are used.
   1338 ///
   1339 /// Note that these are laid out in a specific order to allow bit-twiddling
   1340 /// to transform conditions.
   1341 enum CondCode {
   1342   // Opcode       N U L G E       Intuitive operation
   1343   SETFALSE, //      0 0 0 0       Always false (always folded)
   1344   SETOEQ,   //      0 0 0 1       True if ordered and equal
   1345   SETOGT,   //      0 0 1 0       True if ordered and greater than
   1346   SETOGE,   //      0 0 1 1       True if ordered and greater than or equal
   1347   SETOLT,   //      0 1 0 0       True if ordered and less than
   1348   SETOLE,   //      0 1 0 1       True if ordered and less than or equal
   1349   SETONE,   //      0 1 1 0       True if ordered and operands are unequal
   1350   SETO,     //      0 1 1 1       True if ordered (no nans)
   1351   SETUO,    //      1 0 0 0       True if unordered: isnan(X) | isnan(Y)
   1352   SETUEQ,   //      1 0 0 1       True if unordered or equal
   1353   SETUGT,   //      1 0 1 0       True if unordered or greater than
   1354   SETUGE,   //      1 0 1 1       True if unordered, greater than, or equal
   1355   SETULT,   //      1 1 0 0       True if unordered or less than
   1356   SETULE,   //      1 1 0 1       True if unordered, less than, or equal
   1357   SETUNE,   //      1 1 1 0       True if unordered or not equal
   1358   SETTRUE,  //      1 1 1 1       Always true (always folded)
   1359   // Don't care operations: undefined if the input is a nan.
   1360   SETFALSE2, //   1 X 0 0 0       Always false (always folded)
   1361   SETEQ,     //   1 X 0 0 1       True if equal
   1362   SETGT,     //   1 X 0 1 0       True if greater than
   1363   SETGE,     //   1 X 0 1 1       True if greater than or equal
   1364   SETLT,     //   1 X 1 0 0       True if less than
   1365   SETLE,     //   1 X 1 0 1       True if less than or equal
   1366   SETNE,     //   1 X 1 1 0       True if not equal
   1367   SETTRUE2,  //   1 X 1 1 1       Always true (always folded)
   1368 
   1369   SETCC_INVALID // Marker value.
   1370 };
   1371 
   1372 /// Return true if this is a setcc instruction that performs a signed
   1373 /// comparison when used with integer operands.
   1374 inline bool isSignedIntSetCC(CondCode Code) {
   1375   return Code == SETGT || Code == SETGE || Code == SETLT || Code == SETLE;
   1376 }
   1377 
   1378 /// Return true if this is a setcc instruction that performs an unsigned
   1379 /// comparison when used with integer operands.
   1380 inline bool isUnsignedIntSetCC(CondCode Code) {
   1381   return Code == SETUGT || Code == SETUGE || Code == SETULT || Code == SETULE;
   1382 }
   1383 
   1384 /// Return true if this is a setcc instruction that performs an equality
   1385 /// comparison when used with integer operands.
   1386 inline bool isIntEqualitySetCC(CondCode Code) {
   1387   return Code == SETEQ || Code == SETNE;
   1388 }
   1389 
   1390 /// Return true if the specified condition returns true if the two operands to
   1391 /// the condition are equal. Note that if one of the two operands is a NaN,
   1392 /// this value is meaningless.
   1393 inline bool isTrueWhenEqual(CondCode Cond) { return ((int)Cond & 1) != 0; }
   1394 
   1395 /// This function returns 0 if the condition is always false if an operand is
   1396 /// a NaN, 1 if the condition is always true if the operand is a NaN, and 2 if
   1397 /// the condition is undefined if the operand is a NaN.
   1398 inline unsigned getUnorderedFlavor(CondCode Cond) {
   1399   return ((int)Cond >> 3) & 3;
   1400 }
   1401 
   1402 /// Return the operation corresponding to !(X op Y), where 'op' is a valid
   1403 /// SetCC operation.
   1404 CondCode getSetCCInverse(CondCode Operation, EVT Type);
   1405 
   1406 namespace GlobalISel {
   1407 /// Return the operation corresponding to !(X op Y), where 'op' is a valid
   1408 /// SetCC operation. The U bit of the condition code has different meanings
   1409 /// between floating point and integer comparisons and LLT's don't provide
   1410 /// this distinction. As such we need to be told whether the comparison is
   1411 /// floating point or integer-like. Pointers should use integer-like
   1412 /// comparisons.
   1413 CondCode getSetCCInverse(CondCode Operation, bool isIntegerLike);
   1414 } // end namespace GlobalISel
   1415 
   1416 /// Return the operation corresponding to (Y op X) when given the operation
   1417 /// for (X op Y).
   1418 CondCode getSetCCSwappedOperands(CondCode Operation);
   1419 
   1420 /// Return the result of a logical OR between different comparisons of
   1421 /// identical values: ((X op1 Y) | (X op2 Y)). This function returns
   1422 /// SETCC_INVALID if it is not possible to represent the resultant comparison.
   1423 CondCode getSetCCOrOperation(CondCode Op1, CondCode Op2, EVT Type);
   1424 
   1425 /// Return the result of a logical AND between different comparisons of
   1426 /// identical values: ((X op1 Y) & (X op2 Y)). This function returns
   1427 /// SETCC_INVALID if it is not possible to represent the resultant comparison.
   1428 CondCode getSetCCAndOperation(CondCode Op1, CondCode Op2, EVT Type);
   1429 
   1430 } // namespace ISD
   1431 
   1432 } // namespace llvm
   1433 
   1434 #endif
   1435