33#include "llvm/IR/IntrinsicsAArch64.h"
38#define GET_TARGET_REGBANK_IMPL
39#include "AArch64GenRegisterBank.inc"
42#include "AArch64GenRegisterBankInfo.def"
51 static auto InitializeRegisterBankOnce = [&]() {
60 assert(&AArch64::GPRRegBank == &RBGPR &&
61 "The order in RegBanks is messed up");
65 assert(&AArch64::FPRRegBank == &RBFPR &&
66 "The order in RegBanks is messed up");
70 assert(&AArch64::CCRegBank == &RBCCR &&
71 "The order in RegBanks is messed up");
76 "Subclass not added?");
78 "GPRs should hold up to 128-bit");
83 "Subclass not added?");
85 "Subclass not added?");
87 "FPRs should hold up to 512-bit via QQQQ sequence");
92 "CCR should hold up to 32-bit");
98 "PartialMappingIdx's are incorrectly ordered");
102 "PartialMappingIdx's are incorrectly ordered");
105#define CHECK_PARTIALMAP(Idx, ValStartIdx, ValLength, RB) \
108 checkPartialMap(PartialMappingIdx::Idx, ValStartIdx, ValLength, RB) && \
109 #Idx " is incorrectly initialized"); \
123#define CHECK_VALUEMAP_IMPL(RBName, Size, Offset) \
125 assert(checkValueMapImpl(PartialMappingIdx::PMI_##RBName##Size, \
126 PartialMappingIdx::PMI_First##RBName, Size, \
128 #RBName #Size " " #Offset " is incorrectly initialized"); \
131#define CHECK_VALUEMAP(RBName, Size) CHECK_VALUEMAP_IMPL(RBName, Size, 0)
145#define CHECK_VALUEMAP_3OPS(RBName, Size) \
147 CHECK_VALUEMAP_IMPL(RBName, Size, 0); \
148 CHECK_VALUEMAP_IMPL(RBName, Size, 1); \
149 CHECK_VALUEMAP_IMPL(RBName, Size, 2); \
161#define CHECK_VALUEMAP_CROSSREGCPY(RBNameDst, RBNameSrc, Size) \
163 unsigned PartialMapDstIdx = PMI_##RBNameDst##Size - PMI_Min; \
164 unsigned PartialMapSrcIdx = PMI_##RBNameSrc##Size - PMI_Min; \
165 (void)PartialMapDstIdx; \
166 (void)PartialMapSrcIdx; \
167 const ValueMapping *Map = getCopyMapping(AArch64::RBNameDst##RegBankID, \
168 AArch64::RBNameSrc##RegBankID, \
169 TypeSize::getFixed(Size)); \
171 assert(Map[0].BreakDown == \
172 &AArch64GenRegisterBankInfo::PartMappings[PartialMapDstIdx] && \
173 Map[0].NumBreakDowns == 1 && \
174 #RBNameDst #Size " Dst is incorrectly initialized"); \
175 assert(Map[1].BreakDown == \
176 &AArch64GenRegisterBankInfo::PartMappings[PartialMapSrcIdx] && \
177 Map[1].NumBreakDowns == 1 && \
178 #RBNameSrc #Size " Src is incorrectly initialized"); \
191#define CHECK_VALUEMAP_FPEXT(DstSize, SrcSize) \
193 unsigned PartialMapDstIdx = PMI_FPR##DstSize - PMI_Min; \
194 unsigned PartialMapSrcIdx = PMI_FPR##SrcSize - PMI_Min; \
195 (void)PartialMapDstIdx; \
196 (void)PartialMapSrcIdx; \
197 const ValueMapping *Map = getFPExtMapping(DstSize, SrcSize); \
199 assert(Map[0].BreakDown == \
200 &AArch64GenRegisterBankInfo::PartMappings[PartialMapDstIdx] && \
201 Map[0].NumBreakDowns == 1 && "FPR" #DstSize \
202 " Dst is incorrectly initialized"); \
203 assert(Map[1].BreakDown == \
204 &AArch64GenRegisterBankInfo::PartMappings[PartialMapSrcIdx] && \
205 Map[1].NumBreakDowns == 1 && "FPR" #SrcSize \
206 " Src is incorrectly initialized"); \
218 llvm::call_once(InitializeRegisterBankFlag, InitializeRegisterBankOnce);
232 if (&
A == &AArch64::GPRRegBank && &
B == &AArch64::FPRRegBank)
235 if (&
A == &AArch64::FPRRegBank && &
B == &AArch64::GPRRegBank)
245 switch (RC.
getID()) {
246 case AArch64::GPR64sponlyRegClassID:
261 switch (
MI.getOpcode()) {
262 case TargetOpcode::G_OR: {
271 if (
MI.getNumOperands() != 3)
285 case TargetOpcode::G_BITCAST: {
292 if (
MI.getNumOperands() != 2)
307 copyCost(AArch64::GPRRegBank, AArch64::FPRRegBank,
314 copyCost(AArch64::GPRRegBank, AArch64::FPRRegBank,
325 case TargetOpcode::G_LOAD: {
332 if (
MI.getNumOperands() != 2)
361void AArch64RegisterBankInfo::applyMappingImpl(
366 switch (
MI.getOpcode()) {
367 case TargetOpcode::G_OR:
368 case TargetOpcode::G_BITCAST:
369 case TargetOpcode::G_LOAD:
371 assert((OpdMapper.getInstrMapping().getID() >= 1 &&
372 OpdMapper.getInstrMapping().getID() <= 4) &&
373 "Don't know how to handle that ID");
375 case TargetOpcode::G_INSERT_VECTOR_ELT: {
377 Builder.setInsertPt(*
MI.getParent(),
MI.getIterator());
378 auto Ext = Builder.buildAnyExt(
LLT::scalar(32),
MI.getOperand(2).getReg());
379 MRI.setRegBank(Ext.getReg(0),
getRegBank(AArch64::GPRRegBankID));
380 MI.getOperand(2).setReg(Ext.getReg(0));
383 case AArch64::G_DUP: {
385 assert(
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits() < 32 &&
386 "Expected sources smaller than 32-bits");
387 Builder.setInsertPt(*
MI.getParent(),
MI.getIterator());
390 auto ConstMI =
MRI.getVRegDef(
MI.getOperand(1).getReg());
391 if (ConstMI->getOpcode() == TargetOpcode::G_CONSTANT) {
392 auto CstVal = ConstMI->getOperand(1).getCImm()->getValue();
394 Builder.buildConstant(
LLT::scalar(32), CstVal.sext(32)).getReg(0);
400 MI.getOperand(1).setReg(ConstReg);
409AArch64RegisterBankInfo::getSameKindOfOperandsMapping(
411 const unsigned Opc =
MI.getOpcode();
412 const MachineFunction &MF = *
MI.getParent()->getParent();
415 unsigned NumOperands =
MI.getNumOperands();
416 assert(NumOperands <= 3 &&
417 "This code is for instructions with 3 or less operands");
419 LLT Ty =
MRI.getType(
MI.getOperand(0).getReg());
434 for (
unsigned Idx = 1; Idx != NumOperands; ++Idx) {
435 LLT OpTy =
MRI.getType(
MI.getOperand(Idx).getReg());
440 "Operand has incompatible size");
443 assert(IsFPR == OpIsFPR &&
"Operand has incompatible type");
458 case Intrinsic::aarch64_neon_uaddlv:
459 case Intrinsic::aarch64_neon_uaddv:
460 case Intrinsic::aarch64_neon_saddv:
461 case Intrinsic::aarch64_neon_umaxv:
462 case Intrinsic::aarch64_neon_smaxv:
463 case Intrinsic::aarch64_neon_uminv:
464 case Intrinsic::aarch64_neon_sminv:
465 case Intrinsic::aarch64_neon_faddv:
466 case Intrinsic::aarch64_neon_fmaxv:
467 case Intrinsic::aarch64_neon_fminv:
468 case Intrinsic::aarch64_neon_fmaxnmv:
469 case Intrinsic::aarch64_neon_fminnmv:
470 case Intrinsic::aarch64_neon_fmulx:
471 case Intrinsic::aarch64_neon_frecpe:
472 case Intrinsic::aarch64_neon_frecps:
473 case Intrinsic::aarch64_neon_frecpx:
474 case Intrinsic::aarch64_neon_frsqrte:
475 case Intrinsic::aarch64_neon_frsqrts:
476 case Intrinsic::aarch64_neon_facge:
477 case Intrinsic::aarch64_neon_facgt:
478 case Intrinsic::aarch64_neon_fabd:
479 case Intrinsic::aarch64_neon_sqrdmlah:
480 case Intrinsic::aarch64_neon_sqrdmlsh:
481 case Intrinsic::aarch64_neon_sqrdmulh:
482 case Intrinsic::aarch64_neon_sqadd:
483 case Intrinsic::aarch64_neon_sqsub:
484 case Intrinsic::aarch64_neon_srshl:
485 case Intrinsic::aarch64_neon_urshl:
486 case Intrinsic::aarch64_neon_sqshl:
487 case Intrinsic::aarch64_neon_uqshl:
488 case Intrinsic::aarch64_neon_sqrshl:
489 case Intrinsic::aarch64_neon_uqrshl:
490 case Intrinsic::aarch64_neon_ushl:
491 case Intrinsic::aarch64_neon_sshl:
492 case Intrinsic::aarch64_neon_sqshrn:
493 case Intrinsic::aarch64_neon_sqshrun:
494 case Intrinsic::aarch64_neon_sqrshrn:
495 case Intrinsic::aarch64_neon_sqrshrun:
496 case Intrinsic::aarch64_neon_uqshrn:
497 case Intrinsic::aarch64_neon_uqrshrn:
498 case Intrinsic::aarch64_crypto_sha1h:
499 case Intrinsic::aarch64_crypto_sha1c:
500 case Intrinsic::aarch64_crypto_sha1p:
501 case Intrinsic::aarch64_crypto_sha1m:
502 case Intrinsic::aarch64_sisd_fcvtxn:
503 case Intrinsic::aarch64_sisd_fabd:
505 case Intrinsic::aarch64_neon_saddlv: {
506 const LLT SrcTy =
MRI.getType(
MI.getOperand(2).getReg());
507 return SrcTy.getElementType().getSizeInBits() >= 16 &&
508 SrcTy.getElementCount().getFixedValue() >= 4;
513bool AArch64RegisterBankInfo::isPHIWithFPConstraints(
516 if (!
MI.isPHI() ||
Depth > MaxFPRSearchDepth)
519 return any_of(
MRI.use_nodbg_instructions(
MI.getOperand(0).getReg()),
520 [&](
const MachineInstr &
UseMI) {
521 if (onlyUsesFP(UseMI, MRI, TRI, Depth + 1))
523 return isPHIWithFPConstraints(UseMI, MRI, TRI, Depth + 1);
527bool AArch64RegisterBankInfo::hasFPConstraints(
const MachineInstr &
MI,
530 unsigned Depth)
const {
531 unsigned Op =
MI.getOpcode();
541 if (
Op != TargetOpcode::COPY && !
MI.isPHI() &&
547 if (RB == &AArch64::FPRRegBank)
549 if (RB == &AArch64::GPRRegBank)
556 if (!
MI.isPHI() ||
Depth > MaxFPRSearchDepth)
559 return any_of(
MI.explicit_uses(), [&](
const MachineOperand &
Op) {
561 onlyDefinesFP(*MRI.getVRegDef(Op.getReg()), MRI, TRI, Depth + 1);
568 unsigned Depth)
const {
569 switch (
MI.getOpcode()) {
570 case TargetOpcode::G_FPTOSI:
571 case TargetOpcode::G_FPTOUI:
572 case TargetOpcode::G_FPTOSI_SAT:
573 case TargetOpcode::G_FPTOUI_SAT:
574 case TargetOpcode::G_FCMP:
575 case TargetOpcode::G_LROUND:
576 case TargetOpcode::G_LLROUND:
577 case AArch64::G_PMULL:
581 case TargetOpcode::G_INTRINSIC:
583 case Intrinsic::aarch64_neon_fcvtas:
584 case Intrinsic::aarch64_neon_fcvtau:
585 case Intrinsic::aarch64_neon_fcvtzs:
586 case Intrinsic::aarch64_neon_fcvtzu:
587 case Intrinsic::aarch64_neon_fcvtms:
588 case Intrinsic::aarch64_neon_fcvtmu:
589 case Intrinsic::aarch64_neon_fcvtns:
590 case Intrinsic::aarch64_neon_fcvtnu:
591 case Intrinsic::aarch64_neon_fcvtps:
592 case Intrinsic::aarch64_neon_fcvtpu:
604bool AArch64RegisterBankInfo::onlyDefinesFP(
const MachineInstr &
MI,
607 unsigned Depth)
const {
608 switch (
MI.getOpcode()) {
610 case AArch64::G_SADDLP:
611 case AArch64::G_UADDLP:
612 case TargetOpcode::G_SITOFP:
613 case TargetOpcode::G_UITOFP:
614 case TargetOpcode::G_EXTRACT_VECTOR_ELT:
615 case TargetOpcode::G_INSERT_VECTOR_ELT:
616 case TargetOpcode::G_BUILD_VECTOR:
617 case TargetOpcode::G_BUILD_VECTOR_TRUNC:
621 case TargetOpcode::G_INTRINSIC_W_SIDE_EFFECTS:
623 case Intrinsic::aarch64_neon_ld1x2:
624 case Intrinsic::aarch64_neon_ld1x3:
625 case Intrinsic::aarch64_neon_ld1x4:
626 case Intrinsic::aarch64_neon_ld2:
627 case Intrinsic::aarch64_neon_ld2lane:
628 case Intrinsic::aarch64_neon_ld2r:
629 case Intrinsic::aarch64_neon_ld3:
630 case Intrinsic::aarch64_neon_ld3lane:
631 case Intrinsic::aarch64_neon_ld3r:
632 case Intrinsic::aarch64_neon_ld4:
633 case Intrinsic::aarch64_neon_ld4lane:
634 case Intrinsic::aarch64_neon_ld4r:
649 unsigned Depth)
const {
650 switch (
MI.getOpcode()) {
651 case TargetOpcode::G_SITOFP:
652 case TargetOpcode::G_UITOFP:
653 return MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits() ==
654 MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
659bool AArch64RegisterBankInfo::isLoadFromFPType(
const MachineInstr &
MI)
const {
662 const Value *LdVal = MemOp->getMMO().getValue();
666 Type *EltTy =
nullptr;
668 EltTy = GV->getValueType();
672 if (StructEltTy->getNumElements() == 0)
674 EltTy = StructEltTy->getTypeAtIndex(0U);
682 for (
const auto *LdUser : LdVal->
users()) {
684 EltTy = LdUser->getType();
688 EltTy = LdUser->getOperand(0)->getType();
698 const unsigned Opc =
MI.getOpcode();
703 Opc == TargetOpcode::G_PHI) {
718 case TargetOpcode::G_ADD:
719 case TargetOpcode::G_SUB:
720 case TargetOpcode::G_PTR_ADD:
721 case TargetOpcode::G_MUL:
722 case TargetOpcode::G_SDIV:
723 case TargetOpcode::G_UDIV:
725 case TargetOpcode::G_AND:
726 case TargetOpcode::G_OR:
727 case TargetOpcode::G_XOR:
729 case TargetOpcode::G_FADD:
730 case TargetOpcode::G_FSUB:
731 case TargetOpcode::G_FMUL:
732 case TargetOpcode::G_FDIV:
733 case TargetOpcode::G_FMAXIMUM:
734 case TargetOpcode::G_FMINIMUM:
735 return getSameKindOfOperandsMapping(
MI);
736 case TargetOpcode::G_FPEXT: {
737 LLT DstTy =
MRI.getType(
MI.getOperand(0).getReg());
738 LLT SrcTy =
MRI.getType(
MI.getOperand(1).getReg());
745 case TargetOpcode::G_SHL:
746 case TargetOpcode::G_LSHR:
747 case TargetOpcode::G_ASHR: {
748 LLT ShiftAmtTy =
MRI.getType(
MI.getOperand(2).getReg());
749 LLT SrcTy =
MRI.getType(
MI.getOperand(1).getReg());
750 if (ShiftAmtTy.
getSizeInBits() == 64 && SrcTy.getSizeInBits() == 32)
753 return getSameKindOfOperandsMapping(
MI);
755 case TargetOpcode::COPY: {
759 if ((DstReg.
isPhysical() || !
MRI.getType(DstReg).isValid()) ||
769 assert(DstRB && SrcRB &&
"Both RegBank were nullptr");
780 case TargetOpcode::G_BITCAST: {
781 LLT DstTy =
MRI.getType(
MI.getOperand(0).getReg());
782 LLT SrcTy =
MRI.getType(
MI.getOperand(1).getReg());
785 bool SrcIsGPR = !SrcTy.isVector() && SrcTy.getSizeInBits() <= 64;
787 DstIsGPR ? AArch64::GPRRegBank : AArch64::FPRRegBank;
789 SrcIsGPR ? AArch64::GPRRegBank : AArch64::FPRRegBank;
794 Opc == TargetOpcode::G_BITCAST ? 2 : 1);
800 unsigned NumOperands =
MI.getNumOperands();
806 for (
unsigned Idx = 0; Idx < NumOperands; ++Idx) {
807 auto &MO =
MI.getOperand(Idx);
808 if (!MO.isReg() || !MO.getReg())
811 LLT Ty =
MRI.getType(MO.getReg());
814 OpSize[Idx] = Ty.getSizeInBits().getKnownMinValue();
822 (MO.isDef() && onlyDefinesFP(
MI,
MRI,
TRI)) ||
823 (MO.isUse() && onlyUsesFP(
MI,
MRI,
TRI)) ||
824 Ty.getSizeInBits() > 64)
834 case AArch64::G_DUP: {
835 Register ScalarReg =
MI.getOperand(1).getReg();
836 LLT ScalarTy =
MRI.getType(ScalarReg);
837 auto ScalarDef =
MRI.getVRegDef(ScalarReg);
839 if (ScalarDef->getOpcode() == TargetOpcode::G_LOAD)
844 onlyDefinesFP(*ScalarDef,
MRI,
TRI)))
856 case TargetOpcode::G_TRUNC: {
857 LLT SrcTy =
MRI.getType(
MI.getOperand(1).getReg());
858 if (!SrcTy.isVector() && SrcTy.getSizeInBits() == 128)
862 case TargetOpcode::G_SITOFP:
863 case TargetOpcode::G_UITOFP: {
864 if (
MRI.getType(
MI.getOperand(0).getReg()).isVector())
870 MRI.getType(SrcReg).getSizeInBits() ==
871 MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits())
877 case TargetOpcode::G_FPTOSI_SAT:
878 case TargetOpcode::G_FPTOUI_SAT:
879 case TargetOpcode::G_FPTOSI:
880 case TargetOpcode::G_FPTOUI:
881 case TargetOpcode::G_INTRINSIC_LRINT:
882 case TargetOpcode::G_INTRINSIC_LLRINT:
883 case TargetOpcode::G_LROUND:
884 case TargetOpcode::G_LLROUND: {
885 LLT DstType =
MRI.getType(
MI.getOperand(0).getReg());
886 if (DstType.isVector())
894 if (((DstSize == SrcSize) || STI.hasFeature(AArch64::FeatureFPRCVT)) &&
895 all_of(
MRI.use_nodbg_instructions(
MI.getOperand(0).getReg()),
897 return onlyUsesFP(UseMI, MRI, TRI) ||
898 prefersFPUse(UseMI, MRI, TRI);
905 case TargetOpcode::G_FCMP: {
910 OpRegBankIdx = {Idx0,
914 case TargetOpcode::G_BITCAST:
916 if (OpRegBankIdx[0] != OpRegBankIdx[1])
922 case TargetOpcode::G_LOAD: {
941 if (isLoadFromFPType(
MI)) {
949 if (
any_of(
MRI.use_nodbg_instructions(
MI.getOperand(0).getReg()),
960 if (isPHIWithFPConstraints(UseMI, MRI, TRI))
963 return onlyUsesFP(UseMI, MRI, TRI) ||
964 prefersFPUse(UseMI, MRI, TRI);
969 case TargetOpcode::G_STORE:
981 case TargetOpcode::G_INDEXED_STORE:
992 case TargetOpcode::G_INDEXED_SEXTLOAD:
993 case TargetOpcode::G_INDEXED_ZEXTLOAD:
997 case TargetOpcode::G_INDEXED_LOAD: {
998 if (isLoadFromFPType(
MI))
1002 case TargetOpcode::G_SELECT: {
1009 LLT SrcTy =
MRI.getType(
MI.getOperand(2).getReg());
1010 if (SrcTy.isVector()) {
1026 if (
any_of(
MRI.use_nodbg_instructions(
MI.getOperand(0).getReg()),
1043 for (
unsigned Idx = 2; Idx < 4; ++Idx) {
1058 case TargetOpcode::G_UNMERGE_VALUES: {
1064 LLT SrcTy =
MRI.getType(
MI.getOperand(
MI.getNumOperands()-1).getReg());
1067 if (SrcTy.isVector() || SrcTy ==
LLT::scalar(128) ||
1068 any_of(
MRI.use_nodbg_instructions(
MI.getOperand(0).getReg()),
1071 for (
unsigned Idx = 0, NumOperands =
MI.getNumOperands();
1072 Idx < NumOperands; ++Idx)
1077 case TargetOpcode::G_EXTRACT_VECTOR_ELT:
1085 case AArch64::G_SQSHLU_I:
1094 case TargetOpcode::G_INSERT_VECTOR_ELT:
1104 LLT Ty =
MRI.getType(
MI.getOperand(2).getReg());
1105 if (Ty.getSizeInBits() == 8 || Ty.getSizeInBits() == 16) {
1115 case TargetOpcode::G_EXTRACT: {
1117 auto Src =
MI.getOperand(1).getReg();
1118 LLT SrcTy =
MRI.getType(
MI.getOperand(1).getReg());
1119 if (SrcTy.getSizeInBits() != 128)
1121 auto Idx =
MRI.getRegClassOrNull(Src) == &AArch64::XSeqPairsClassRegClass
1124 OpRegBankIdx[0] = Idx;
1125 OpRegBankIdx[1] = Idx;
1128 case TargetOpcode::G_BUILD_VECTOR: {
1143 unsigned DefOpc =
DefMI->getOpcode();
1144 const LLT SrcTy =
MRI.getType(VReg);
1146 return Op.isDef() || MRI.getVRegDef(Op.getReg())->getOpcode() ==
1147 TargetOpcode::G_CONSTANT;
1151 SrcTy.getSizeInBits() < 32 ||
1155 unsigned NumOperands =
MI.getNumOperands();
1156 for (
unsigned Idx = 0; Idx < NumOperands; ++Idx)
1161 case TargetOpcode::G_VECREDUCE_FADD:
1162 case TargetOpcode::G_VECREDUCE_FMUL:
1163 case TargetOpcode::G_VECREDUCE_FMAX:
1164 case TargetOpcode::G_VECREDUCE_FMIN:
1165 case TargetOpcode::G_VECREDUCE_FMAXIMUM:
1166 case TargetOpcode::G_VECREDUCE_FMINIMUM:
1167 case TargetOpcode::G_VECREDUCE_ADD:
1168 case TargetOpcode::G_VECREDUCE_MUL:
1169 case TargetOpcode::G_VECREDUCE_AND:
1170 case TargetOpcode::G_VECREDUCE_OR:
1171 case TargetOpcode::G_VECREDUCE_XOR:
1172 case TargetOpcode::G_VECREDUCE_SMAX:
1173 case TargetOpcode::G_VECREDUCE_SMIN:
1174 case TargetOpcode::G_VECREDUCE_UMAX:
1175 case TargetOpcode::G_VECREDUCE_UMIN:
1180 case TargetOpcode::G_VECREDUCE_SEQ_FADD:
1181 case TargetOpcode::G_VECREDUCE_SEQ_FMUL:
1186 case TargetOpcode::G_INTRINSIC:
1187 case TargetOpcode::G_INTRINSIC_W_SIDE_EFFECTS: {
1189 case Intrinsic::aarch64_neon_fcvtas:
1190 case Intrinsic::aarch64_neon_fcvtau:
1191 case Intrinsic::aarch64_neon_fcvtzs:
1192 case Intrinsic::aarch64_neon_fcvtzu:
1193 case Intrinsic::aarch64_neon_fcvtms:
1194 case Intrinsic::aarch64_neon_fcvtmu:
1195 case Intrinsic::aarch64_neon_fcvtns:
1196 case Intrinsic::aarch64_neon_fcvtnu:
1197 case Intrinsic::aarch64_neon_fcvtps:
1198 case Intrinsic::aarch64_neon_fcvtpu: {
1200 if (
MRI.getType(
MI.getOperand(0).getReg()).isVector()) {
1206 if (((DstSize == SrcSize) || STI.hasFeature(AArch64::FeatureFPRCVT)) &&
1207 all_of(
MRI.use_nodbg_instructions(
MI.getOperand(0).getReg()),
1209 return onlyUsesFP(UseMI, MRI, TRI) ||
1210 prefersFPUse(UseMI, MRI, TRI);
1217 case Intrinsic::aarch64_neon_vcvtfxs2fp:
1218 case Intrinsic::aarch64_neon_vcvtfxu2fp:
1219 case Intrinsic::aarch64_neon_vcvtfp2fxs:
1220 case Intrinsic::aarch64_neon_vcvtfp2fxu:
1234 for (
const auto &
Op :
MI.defs()) {
1240 Idx +=
MI.getNumExplicitDefs();
1243 for (
const auto &
Op :
MI.explicit_uses()) {
1257 for (
unsigned Idx = 0; Idx < NumOperands; ++Idx) {
1258 if (
MI.getOperand(Idx).isReg() &&
MI.getOperand(Idx).getReg()) {
1259 LLT Ty =
MRI.getType(
MI.getOperand(Idx).getReg());
1264 if (!Mapping->isValid())
1267 OpdsMapping[Idx] = Mapping;
unsigned const MachineRegisterInfo * MRI
MachineInstrBuilder & UseMI
MachineInstrBuilder MachineInstrBuilder & DefMI
static unsigned getIntrinsicID(const SDNode *N)
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
#define CHECK_VALUEMAP(RBName, Size)
static bool isFPIntrinsic(const MachineRegisterInfo &MRI, const MachineInstr &MI)
#define CHECK_VALUEMAP_3OPS(RBName, Size)
static const unsigned CustomMappingID
#define CHECK_PARTIALMAP(Idx, ValStartIdx, ValLength, RB)
#define CHECK_VALUEMAP_CROSSREGCPY(RBNameDst, RBNameSrc, Size)
#define CHECK_VALUEMAP_FPEXT(DstSize, SrcSize)
This file declares the targeting of the RegisterBankInfo class for AArch64.
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
Declares convenience wrapper classes for interpreting MachineInstr instances as specific generic oper...
Implement a low-level type suitable for MachineInstr level instruction selection.
This file declares the MachineIRBuilder class.
Register const TargetRegisterInfo * TRI
static const MCPhysReg FPR[]
FPR - The set of FP registers that should be allocated for arguments on Darwin and AIX.
This file defines the SmallVector class.
static unsigned getRegBankBaseIdxOffset(unsigned RBIdx, TypeSize Size)
static const RegisterBankInfo::ValueMapping * getCopyMapping(unsigned DstBankID, unsigned SrcBankID, TypeSize Size)
Get the pointer to the ValueMapping of the operands of a copy instruction from the SrcBankID register...
static bool checkPartialMappingIdx(PartialMappingIdx FirstAlias, PartialMappingIdx LastAlias, ArrayRef< PartialMappingIdx > Order)
static const RegisterBankInfo::PartialMapping PartMappings[]
static const RegisterBankInfo::ValueMapping * getFPExtMapping(unsigned DstSize, unsigned SrcSize)
Get the instruction mapping for G_FPEXT.
static const RegisterBankInfo::ValueMapping * getValueMapping(PartialMappingIdx RBIdx, TypeSize Size)
Get the pointer to the ValueMapping representing the RegisterBank at RBIdx with a size of Size.
static const RegisterBankInfo::ValueMapping ValMappings[]
InstructionMappings getInstrAlternativeMappings(const MachineInstr &MI) const override
Get the alternative mappings for MI.
unsigned copyCost(const RegisterBank &A, const RegisterBank &B, TypeSize Size) const override
Get the cost of a copy from B to A, or put differently, get the cost of A = COPY B.
const RegisterBank & getRegBankFromRegClass(const TargetRegisterClass &RC, LLT Ty) const override
Get a register bank that covers RC.
AArch64RegisterBankInfo(const TargetRegisterInfo &TRI)
const InstructionMapping & getInstrMapping(const MachineInstr &MI) const override
Get the mapping of the different operands of MI on the register bank.
const AArch64RegisterInfo * getRegisterInfo() const override
static constexpr LLT scalar(unsigned SizeInBits)
Get a low-level scalar or aggregate "bag of bits".
constexpr bool isVector() const
constexpr TypeSize getSizeInBits() const
Returns the total size of the type. Must only be called on sized types.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Helper class to build MachineInstr.
MachineInstrBuilder buildAnyExt(const DstOp &Res, const SrcOp &Op)
Build and insert Res = G_ANYEXT Op0.
Register getReg(unsigned Idx) const
Get the register for the operand index.
Representation of each machine instruction.
MachineOperand class - Representation of each machine instruction operand.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
Helper class that represents how the value of an instruction may be mapped and what is the related co...
bool isValid() const
Check whether this object is valid.
virtual InstructionMappings getInstrAlternativeMappings(const MachineInstr &MI) const
Get the alternative mappings for MI.
const InstructionMapping & getInstructionMapping(unsigned ID, unsigned Cost, const ValueMapping *OperandsMapping, unsigned NumOperands) const
Method to get a uniquely generated InstructionMapping.
static void applyDefaultMapping(const OperandsMapper &OpdMapper)
Helper method to apply something that is like the default mapping.
const InstructionMapping & getInvalidInstructionMapping() const
Method to get a uniquely generated invalid InstructionMapping.
const RegisterBank & getRegBank(unsigned ID)
Get the register bank identified by ID.
unsigned getMaximumSize(unsigned RegBankID) const
Get the maximum size in bits that fits in the given register bank.
TypeSize getSizeInBits(Register Reg, const MachineRegisterInfo &MRI, const TargetRegisterInfo &TRI) const
Get the size in bits of Reg.
virtual const RegisterBank & getRegBankFromRegClass(const TargetRegisterClass &RC, LLT Ty) const
Get a register bank that covers RC.
const ValueMapping * getOperandsMapping(Iterator Begin, Iterator End) const
Get the uniquely generated array of ValueMapping for the elements of between Begin and End.
static const unsigned DefaultMappingID
Identifier used when the related instruction mapping instance is generated by target independent code...
SmallVector< const InstructionMapping *, 4 > InstructionMappings
Convenient type to represent the alternatives for mapping an instruction.
virtual unsigned copyCost(const RegisterBank &A, const RegisterBank &B, TypeSize Size) const
Get the cost of a copy from B to A, or put differently, get the cost of A = COPY B.
const InstructionMapping & getInstrMappingImpl(const MachineInstr &MI) const
Try to get the mapping of MI.
This class implements the register bank concept.
LLVM_ABI bool covers(const TargetRegisterClass &RC) const
Check whether this register bank covers RC.
unsigned getID() const
Get the identifier of this register bank.
Wrapper class representing virtual and physical registers.
constexpr bool isPhysical() const
Return true if the specified register number is in the physical register namespace.
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
unsigned getID() const
Return the register class ID number.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
TargetSubtargetInfo - Generic base class for all target subtargets.
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
Type * getArrayElementType() const
bool isFPOrFPVectorTy() const
Return true if this is a FP type or a vector of FP.
iterator_range< user_iterator > users()
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
This is an optimization pass for GlobalISel generic memory operations.
FunctionAddr VTableAddr Value
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
bool isPreISelGenericOpcode(unsigned Opcode)
Check whether the given Opcode is a generic opcode that is not supposed to appear after ISel.
bool isPreISelGenericOptimizationHint(unsigned Opcode)
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
DWARFExpression::Operation Op
void call_once(once_flag &flag, Function &&F, Args &&... ArgList)
Execute the function specified as a parameter once.
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
LLVM_ABI bool isPreISelGenericFloatingPointOpcode(unsigned Opc)
Returns whether opcode Opc is a pre-isel generic floating-point opcode, having only floating-point op...
The llvm::once_flag structure.