HomeSort by: relevance | last modified time | path
    Searched defs:SubReg (Results 1 - 25 of 35) sorted by relevancy

1 2

  /src/external/apache2/llvm/dist/llvm/lib/CodeGen/
LiveIntervalCalc.cpp 68 unsigned SubReg = MO.getSubReg();
69 if (LI.hasSubRanges() || (SubReg != 0 && TrackSubRegs)) {
70 LaneBitmask SubMask = SubReg != 0 ? TRI.getSubRegIndexLaneMask(SubReg)
168 unsigned SubReg = MO.getSubReg();
169 if (SubReg != 0) {
170 LaneBitmask SLM = TRI.getSubRegIndexLaneMask(SubReg);
LiveRangeEdit.cpp 201 // FIXME: Targets don't know how to fold subreg uses.
251 unsigned SubReg = MO.getSubReg();
252 LaneBitmask LaneMask = TRI.getSubRegIndexLaneMask(SubReg);
MachineInstrBundle.cpp 198 unsigned SubReg = *SubRegs;
199 if (LocalDefSet.insert(SubReg).second)
200 LocalDefs.push_back(SubReg);
DetectDeadLanes.cpp 174 unsigned SubReg = MI.getOperand(2).getImm();
175 SrcSubIdx = TRI.composeSubRegIndices(SubReg, SrcSubIdx);
424 unsigned SubReg = MO.getSubReg();
447 if (SubReg == 0)
450 UsedLanes |= TRI->getSubRegIndexLaneMask(SubReg);
457 unsigned SubReg = MO.getSubReg();
458 LaneBitmask Mask = TRI->getSubRegIndexLaneMask(SubReg);
LiveVariables.cpp 196 unsigned SubReg = *SubRegs;
197 MachineInstr *Def = PhysRegDef[SubReg];
202 LastDefReg = SubReg;
250 unsigned SubReg = *SubRegs;
251 if (Processed.count(SubReg))
253 if (PartDefRegs.count(SubReg))
257 LastPartialDef->addOperand(MachineOperand::CreateReg(SubReg,
260 PhysRegDef[SubReg] = LastPartialDef;
261 for (MCSubRegIterator SS(SubReg, TRI); SS.isValid(); ++SS)
289 unsigned SubReg = *SubRegs
    [all...]
VirtRegMap.cpp 359 // Sort and unique MBB LiveIns as we've not checked if SubReg/PhysReg were in
554 unsigned SubReg = MO.getSubReg();
555 if (SubReg != 0) {
593 PhysReg = TRI->getSubReg(PhysReg, SubReg);
594 assert(PhysReg.isValid() && "Invalid SubReg for physical register");
MachineOperand.cpp 518 OS << "%subreg.";
773 if (unsigned SubReg = getSubReg()) {
775 OS << '.' << TRI->getSubRegIndexName(SubReg);
777 OS << ".subreg" << SubReg;
LiveIntervals.cpp 577 unsigned SubReg = MO.getSubReg();
578 if (SubReg != 0) {
579 LaneBitmask LaneMask = TRI->getSubRegIndexLaneMask(SubReg);
791 unsigned SubReg = MO.getSubReg();
792 LaneBitmask UseMask = SubReg ? TRI->getSubRegIndexLaneMask(SubReg)
1031 unsigned SubReg = MO.getSubReg();
1032 LaneBitmask LaneMask = SubReg ? TRI.getSubRegIndexLaneMask(SubReg)
1193 // We merge OldIdxOut and its successor. As we're dealing with subreg
    [all...]
ScheduleDAGInstrs.cpp 340 for (MCSubRegIterator SubReg(Reg, TRI, true); SubReg.isValid(); ++SubReg) {
341 if (Uses.contains(*SubReg))
342 Uses.eraseAll(*SubReg);
344 Defs.eraseAll(*SubReg);
376 unsigned SubReg = MO.getSubReg();
377 if (SubReg == 0)
379 return TRI->getSubRegIndexLaneMask(SubReg);
  /src/external/apache2/llvm/dist/llvm/lib/Target/AMDGPU/
SIFixSGPRCopies.cpp 247 unsigned SubReg = CopyUse.getOperand(1).getSubReg();
248 if (SubReg != AMDGPU::NoSubRegister)
SIPreAllocateWWMRegs.cpp 132 const unsigned SubReg = MO.getSubReg();
133 if (SubReg != 0) {
134 PhysReg = TRI->getSubReg(PhysReg, SubReg);
R600OptimizeVectorRegisters.cpp 11 /// common data and/or have enough undef subreg using swizzle abilities.
196 unsigned SubReg = (*It).first;
203 .addReg(SubReg)
205 UpdatedRegToChan[SubReg] = Chan;
SIRegisterInfo.cpp 282 "getNumCoveredRegs() will not work with generated subreg masks!");
1144 Register SubReg = e == 1
1190 // current SubReg has been already spilled into AGPRs by the loop above.
1197 SubReg = Register(getSubReg(ValueReg,
1203 unsigned FinalReg = SubReg;
1217 .addReg(SubReg, getKillRegState(IsKill));
1222 SubReg = TmpReg;
1232 .addReg(SubReg, getDefRegState(!IsStore) | getKillRegState(IsKill));
1321 Register SubReg =
1333 .addReg(SubReg, getKillRegState(UseKill)
    [all...]
SIWholeQuadMode.cpp 187 unsigned SubReg, char Flag, std::vector<WorkItem> &Worklist);
316 Register Reg, unsigned SubReg, char Flag,
328 SubReg ? TRI->getSubRegIndexLaneMask(SubReg)
1428 const unsigned SubReg = MI->getOperand(0).getSubReg();
1433 if (SubReg)
1434 regClass = TRI->getSubRegClass(regClass, SubReg);
  /src/external/apache2/llvm/dist/llvm/lib/Target/AArch64/
AArch64AdvSIMDScalarPass.cpp 104 static bool isGPR64(unsigned Reg, unsigned SubReg,
106 if (SubReg)
113 static bool isFPR64(unsigned Reg, unsigned SubReg,
117 SubReg == 0) ||
119 SubReg == AArch64::dsub);
121 return (AArch64::FPR64RegClass.contains(Reg) && SubReg == 0) ||
122 (AArch64::FPR128RegClass.contains(Reg) && SubReg == AArch64::dsub);
129 unsigned &SubReg) {
130 SubReg = 0;
138 SubReg = AArch64::dsub
    [all...]
AArch64InstrInfo.cpp 3186 int SubReg = 0, End = NumRegs, Incr = 1;
3188 SubReg = NumRegs - 1;
3193 for (; SubReg != End; SubReg += Incr) {
3195 AddSubReg(MIB, DestReg, Indices[SubReg], RegState::Define, TRI);
3196 AddSubReg(MIB, SrcReg, Indices[SubReg], 0, TRI);
3197 AddSubReg(MIB, SrcReg, Indices[SubReg], getKillRegState(KillSrc), TRI);
3217 for (unsigned SubReg = 0; SubReg != NumRegs; ++SubReg) {
    [all...]
  /src/external/apache2/llvm/dist/llvm/lib/Target/ARM/
ARMAsmPrinter.cpp 417 Register SubReg =
419 O << ARMInstPrinter::getRegisterName(SubReg);
ARMExpandPseudoInsts.cpp 553 Register SubReg = TRI->getSubReg(DstReg, SubRegIndex);
554 unsigned DstRegPair = TRI->getMatchingSuperReg(SubReg, ARM::dsub_0,
  /src/external/apache2/llvm/dist/llvm/lib/Target/BPF/
BPFMIPeephole.cpp 219 Register SubReg = MovMI->getOperand(1).getReg();
227 .addImm(0).addReg(SubReg).addImm(BPF::sub_32);
  /src/external/apache2/llvm/dist/llvm/include/llvm/CodeGen/GlobalISel/
InstructionSelectorImpl.h 941 unsigned SubReg = 0;
943 SubReg = MatchTable[CurrentIdx++];
947 OutMIs[InsnID].addReg(State.TempRegisters[TempRegID], TempRegFlags, SubReg);
952 if (SubReg)
953 dbgs() << '.' << TRI.getSubRegIndexName(SubReg);
  /src/external/apache2/llvm/dist/llvm/lib/Target/AArch64/GISel/
AArch64InstructionSelector.cpp 553 const TargetRegisterInfo &TRI, unsigned &SubReg) {
556 SubReg = AArch64::bsub;
559 SubReg = AArch64::hsub;
563 SubReg = AArch64::sub_32;
565 SubReg = AArch64::ssub;
568 SubReg = AArch64::dsub;
803 /// E.g "To = COPY SrcReg:SubReg"
806 const TargetRegisterClass *To, unsigned SubReg) {
809 assert(SubReg && "Expected a valid subregister");
813 MIB.buildInstr(TargetOpcode::COPY, {To}, {}).addReg(SrcReg, 0, SubReg);
    [all...]
  /src/external/apache2/llvm/dist/llvm/lib/CodeGen/MIRParser/
MIParser.cpp 445 bool parseSubRegisterIndex(unsigned &SubReg);
1428 bool MIParser::parseSubRegisterIndex(unsigned &SubReg) {
1434 SubReg = PFS.Target.getSubRegIndex(Name);
1435 if (!SubReg)
1506 unsigned SubReg = 0;
1508 if (parseSubRegisterIndex(SubReg))
1570 Flags & RegState::EarlyClobber, SubReg, Flags & RegState::Debug,
  /src/external/apache2/llvm/dist/llvm/include/llvm/CodeGen/
TargetInstrInfo.h 244 /// expected the pre-extension value is available as a subreg of the result
382 /// DestReg:SubIdx. Any existing subreg index is preserved or composed with
463 /// Used to give some type checking when modeling Reg:SubReg.
466 unsigned SubReg;
468 RegSubRegPair(Register Reg = Register(), unsigned SubReg = 0)
469 : Reg(Reg), SubReg(SubReg) {}
472 return Reg == P.Reg && SubReg == P.SubReg;
485 RegSubRegPairAndIdx(Register Reg = Register(), unsigned SubReg = 0
    [all...]
  /src/external/apache2/llvm/dist/llvm/utils/TableGen/
CodeGenRegisters.cpp 254 for (const auto &SubReg : SubRegs) {
255 CodeGenRegister *SR = SubReg.second;
301 // Expand any composed subreg indices.
303 // qsub_1 subreg, add a dsub_2 subreg. Keep growing Indices and process
304 // expanded subreg indices recursively.
318 // Add I->second as a name for the subreg SRI->second, assuming it is
331 // Consider this subreg sequence:
342 // dsub_1, dsub_2] subregs without necessarily having a qsub_1 subreg. The
348 for (const auto &SubReg : Map
    [all...]
  /src/external/apache2/llvm/dist/llvm/lib/CodeGen/GlobalISel/
IRTranslator.cpp 1013 Register SubReg = RangeSub.getReg(0);
1015 SubReg = MIB.buildZExtOrTrunc(MaskTy, SubReg).getReg(0);
1018 B.Reg = SubReg;

Completed in 98 milliseconds

1 2