28#include "llvm/IR/IntrinsicsAMDGPU.h"
36#define DEBUG_TYPE "AMDGPUtti"
40struct AMDGPUImageDMaskIntrinsic {
44#define GET_AMDGPUImageDMaskIntrinsicTable_IMPL
45#include "AMDGPUGenSearchableTables.inc"
75 Type *VTy = V.getType();
84 APFloat FloatValue(ConstFloat->getValueAPF());
85 bool LosesInfo =
true;
94 APInt IntValue(ConstInt->getValue());
113 Type *VTy = V.getType();
139 Func(Args, OverloadTys);
151 bool RemoveOldIntr = &OldIntr != &InstToReplace;
160static std::optional<Instruction *>
165 if (
const auto *LZMappingInfo =
167 if (
auto *ConstantLod =
169 if (ConstantLod->isZero() || ConstantLod->isNegative()) {
174 II,
II, NewImageDimIntr->
Intr, IC, [&](
auto &Args,
auto &ArgTys) {
175 Args.erase(Args.begin() + ImageDimIntr->LodIndex);
182 if (
const auto *MIPMappingInfo =
184 if (
auto *ConstantMip =
186 if (ConstantMip->isZero()) {
191 II,
II, NewImageDimIntr->
Intr, IC, [&](
auto &Args,
auto &ArgTys) {
192 Args.erase(Args.begin() + ImageDimIntr->MipIndex);
199 if (
const auto *BiasMappingInfo =
201 if (
auto *ConstantBias =
203 if (ConstantBias->isZero()) {
208 II,
II, NewImageDimIntr->
Intr, IC, [&](
auto &Args,
auto &ArgTys) {
209 Args.erase(Args.begin() + ImageDimIntr->BiasIndex);
210 ArgTys.erase(ArgTys.begin() + ImageDimIntr->BiasTyArg);
217 if (
const auto *OffsetMappingInfo =
219 if (
auto *ConstantOffset =
221 if (ConstantOffset->isZero()) {
224 OffsetMappingInfo->NoOffset, ImageDimIntr->
Dim);
226 II,
II, NewImageDimIntr->
Intr, IC, [&](
auto &Args,
auto &ArgTys) {
227 Args.erase(Args.begin() + ImageDimIntr->OffsetIndex);
234 if (ST->hasD16Images()) {
244 if (
II.hasOneUse()) {
247 if (
User->getOpcode() == Instruction::FPTrunc &&
251 [&](
auto &Args,
auto &ArgTys) {
254 ArgTys[0] = User->getType();
263 bool AllHalfExtracts =
true;
265 for (
User *U :
II.users()) {
267 if (!Ext || !Ext->hasOneUse()) {
268 AllHalfExtracts =
false;
273 if (!Tr || !Tr->getType()->isHalfTy()) {
274 AllHalfExtracts =
false;
281 if (!ExtractTruncPairs.
empty() && AllHalfExtracts) {
290 OverloadTys[0] = HalfVecTy;
294 M, ImageDimIntr->
Intr, OverloadTys);
296 II.mutateType(HalfVecTy);
297 II.setCalledFunction(HalfDecl);
300 for (
auto &[Ext, Tr] : ExtractTruncPairs) {
301 Value *Idx = Ext->getIndexOperand();
303 Builder.SetInsertPoint(Tr);
305 Value *HalfExtract = Builder.CreateExtractElement(&
II, Idx);
308 Tr->replaceAllUsesWith(HalfExtract);
311 for (
auto &[Ext, Tr] : ExtractTruncPairs) {
322 if (!ST->hasA16() && !ST->hasG16())
329 bool FloatCoord =
false;
331 bool OnlyDerivatives =
false;
334 OperandIndex < ImageDimIntr->VAddrEnd; OperandIndex++) {
335 Value *Coord =
II.getOperand(OperandIndex);
338 if (OperandIndex < ImageDimIntr->CoordStart ||
343 OnlyDerivatives =
true;
352 if (!OnlyDerivatives && !ST->hasA16())
353 OnlyDerivatives =
true;
356 if (!OnlyDerivatives && ImageDimIntr->
NumBiasArgs != 0) {
359 "Only image instructions with a sampler can have a bias");
361 OnlyDerivatives =
true;
364 if (OnlyDerivatives && (!ST->hasG16() || ImageDimIntr->
GradientStart ==
372 II,
II,
II.getIntrinsicID(), IC, [&](
auto &Args,
auto &ArgTys) {
373 ArgTys[ImageDimIntr->GradientTyArg] = CoordType;
374 if (!OnlyDerivatives) {
375 ArgTys[ImageDimIntr->CoordTyArg] = CoordType;
378 if (ImageDimIntr->NumBiasArgs != 0)
379 ArgTys[ImageDimIntr->BiasTyArg] = Type::getHalfTy(II.getContext());
385 OperandIndex < EndIndex; OperandIndex++) {
387 convertTo16Bit(*II.getOperand(OperandIndex), IC.Builder);
392 Value *Bias = II.getOperand(ImageDimIntr->BiasIndex);
393 Args[ImageDimIntr->BiasIndex] = convertTo16Bit(*Bias, IC.Builder);
422 Value *Src =
nullptr;
425 if (Src->getType()->isHalfTy())
442 unsigned VWidth = VTy->getNumElements();
445 for (
int i = VWidth - 1; i > 0; --i) {
467 unsigned VWidth = VTy->getNumElements();
473 SVI->getShuffleMask(ShuffleMask);
475 for (
int I = VWidth - 1;
I > 0; --
I) {
476 if (ShuffleMask.empty()) {
527 unsigned LaneArgIdx)
const {
528 unsigned MaskBits = ST->getWavefrontSizeLog2();
542 Value *LaneArg =
II.getArgOperand(LaneArgIdx);
545 if (MaskedConst != LaneArg) {
546 II.getOperandUse(LaneArgIdx).set(MaskedConst);
558 CallInst *NewCall =
B.CreateCall(&NewCallee,
Ops, OpBundles);
574 if (ST.isWave32() &&
match(V, W32Pred))
576 if (ST.isWave64() &&
match(V, W64Pred))
585 const auto IID =
II.getIntrinsicID();
586 assert(IID == Intrinsic::amdgcn_readlane ||
587 IID == Intrinsic::amdgcn_readfirstlane ||
588 IID == Intrinsic::amdgcn_permlane64);
598 const bool IsReadLane = (IID == Intrinsic::amdgcn_readlane);
602 Value *LaneID =
nullptr;
604 LaneID =
II.getOperand(1);
618 const auto DoIt = [&](
unsigned OpIdx,
622 Ops.push_back(LaneID);
638 return DoIt(0,
II.getCalledFunction());
642 Type *SrcTy = Src->getType();
648 return DoIt(0, Remangled);
656 return DoIt(1,
II.getCalledFunction());
658 return DoIt(0,
II.getCalledFunction());
669 unsigned Depth = 0) {
679 return CI->getZExtValue();
688 std::optional<unsigned>
LHS =
692 std::optional<unsigned>
RHS =
701 return CI ? std::optional<unsigned>(CI->getZExtValue()) : std::nullopt;
709 unsigned WaveSize = ST.getWavefrontSize();
711 for (
unsigned Lane :
seq(WaveSize)) {
713 if (!Val || *Val >= WaveSize)
722template <
unsigned Period>
724 static_assert(
isPowerOf2_32(Period),
"Period must be a power of two");
725 for (
unsigned I = Period,
E = Ids.
size();
I <
E; ++
I)
726 if (Ids[
I] != Ids[
I % Period] + (
I & ~(Period - 1)))
734 for (
unsigned I = 0;
I <
N; ++
I)
750 return Ids[3] << 6 | Ids[2] << 4 | Ids[1] << 2 | Ids[0];
757 for (
unsigned J = 0; J <
N; ++J)
758 if (Ids[J] != (
N - 1) - J)
770 for (
unsigned J = 1; J < 16; ++J)
771 if (Ids[J] != (Ids[0] + J) % 16)
789 unsigned Mask = Ids[0];
792 for (
unsigned J = 0; J < 16; ++J)
793 if (Ids[J] != (Mask ^ J))
803 unsigned Selector = 0;
804 for (
unsigned J = 0; J < 8; ++J)
805 Selector |= Ids[J] << (J * 3);
814 for (
unsigned J = 0; J < 16; ++J)
815 Sel |=
static_cast<uint64_t>(Ids[J] & 0xF) << (J * 4);
822 if (Ids.
size() != 64)
824 for (
unsigned J = 0; J < 64; ++J)
825 if (Ids[J] != (J ^ 32))
836 for (
unsigned J = 0; J < 16; ++J) {
837 if (Ids[J] < 16 || Ids[J] >= 32)
839 if (Ids[J + 16] != Ids[J] - 16)
850static std::optional<unsigned>
859 unsigned AndMask = 0, OrMask = 0, XorMask = 0;
860 for (
unsigned B = 0;
B < 5; ++
B) {
861 unsigned Bit0 = (Ids[0] >>
B) & 1;
862 unsigned Bit1 = (Ids[1u <<
B] >>
B) & 1;
865 XorMask |= Bit0 <<
B;
873 for (
unsigned I :
seq(32u)) {
874 unsigned Expected = ((
I & AndMask) | OrMask) ^ XorMask;
890 return B.CreateIntrinsic(Intrinsic::amdgcn_update_dpp, {Ty},
892 B.getInt32(0xF),
B.getInt32(0xF),
B.getTrue()});
897 return B.CreateIntrinsic(Intrinsic::amdgcn_mov_dpp8, {Val->
getType()},
898 {Val,
B.getInt32(Selector)});
905 return B.CreateIntrinsic(Intrinsic::amdgcn_permlane16, {Ty},
907 B.getInt32(
Hi),
B.getFalse(),
B.getFalse()});
915 return B.CreateIntrinsic(Intrinsic::amdgcn_permlanex16, {Ty},
917 B.getInt32(
Hi),
B.getFalse(),
B.getFalse()});
925 assert(
DL.getTypeSizeInBits(OrigTy) == 32 &&
926 "ds_swizzle only supports 32-bit operands");
930 Src =
B.CreatePtrToInt(Src, I32Ty);
931 else if (OrigTy != I32Ty)
932 Src =
B.CreateBitCast(Src, I32Ty);
933 Value *Result =
B.CreateIntrinsic(Intrinsic::amdgcn_ds_swizzle, {},
936 return B.CreateIntToPtr(Result, OrigTy);
938 return B.CreateBitCast(Result, OrigTy);
944 return B.CreateIntrinsic(Intrinsic::amdgcn_permlane64, {Val->
getType()},
974 if (ST.hasDPPRowShare()) {
979 if (ST.hasDPP() && ST.hasGFX10Insts()) {
989 if (ST.hasPermLaneX16()) {
1015static std::optional<Instruction *>
1019 if (
DL.getTypeSizeInBits(
II.getType()) != 32)
1020 return std::nullopt;
1022 if (!ST.isWaveSizeKnown())
1023 return std::nullopt;
1025 unsigned WaveSize = ST.getWavefrontSize();
1026 bool IsBpermute =
II.getIntrinsicID() == Intrinsic::amdgcn_ds_bpermute;
1027 Value *Src =
II.getArgOperand(IsBpermute ? 1 : 0);
1028 Value *Index =
II.getArgOperand(IsBpermute ? 0 : 1);
1033 for (
unsigned Lane :
seq(WaveSize)) {
1035 if (!Val || (*Val & 3) || (*Val >> 2) >= WaveSize)
1036 return std::nullopt;
1037 Ids[Lane] = *Val >> 2;
1041 return std::nullopt;
1046 return std::nullopt;
1051std::optional<Instruction *>
1055 case Intrinsic::amdgcn_implicitarg_ptr: {
1056 if (
II.getFunction()->hasFnAttribute(
"amdgpu-no-implicitarg-ptr"))
1058 uint64_t ImplicitArgBytes = ST->getImplicitArgNumBytes(*
II.getFunction());
1061 II.getAttributes().getRetDereferenceableOrNullBytes();
1062 if (CurrentOrNullBytes != 0) {
1065 uint64_t NewBytes = std::max(CurrentOrNullBytes, ImplicitArgBytes);
1068 II.removeRetAttr(Attribute::DereferenceableOrNull);
1072 uint64_t CurrentBytes =
II.getAttributes().getRetDereferenceableBytes();
1073 uint64_t NewBytes = std::max(CurrentBytes, ImplicitArgBytes);
1074 if (NewBytes != CurrentBytes) {
1080 return std::nullopt;
1082 case Intrinsic::amdgcn_rcp: {
1083 Value *Src =
II.getArgOperand(0);
1094 if (
II.isStrictFP())
1098 const APFloat &ArgVal =
C->getValueAPF();
1116 auto IID = SrcCI->getIntrinsicID();
1121 if (IID == Intrinsic::amdgcn_sqrt || IID == Intrinsic::sqrt) {
1131 SrcCI->getModule(), Intrinsic::amdgcn_rsq, {SrcCI->getType()});
1134 II.setFastMathFlags(InnerFMF);
1136 II.setCalledFunction(NewDecl);
1142 case Intrinsic::amdgcn_sqrt:
1143 case Intrinsic::amdgcn_rsq:
1144 case Intrinsic::amdgcn_tanh: {
1145 Value *Src =
II.getArgOperand(0);
1157 if (IID == Intrinsic::amdgcn_sqrt && Src->getType()->isHalfTy()) {
1159 II.getModule(), Intrinsic::sqrt, {II.getType()});
1160 II.setCalledFunction(NewDecl);
1166 case Intrinsic::amdgcn_log:
1167 case Intrinsic::amdgcn_exp2: {
1168 const bool IsLog = IID == Intrinsic::amdgcn_log;
1169 const bool IsExp = IID == Intrinsic::amdgcn_exp2;
1170 Value *Src =
II.getArgOperand(0);
1180 if (
C->isInfinity()) {
1183 if (!
C->isNegative())
1187 if (IsExp &&
C->isNegative())
1191 if (
II.isStrictFP())
1195 Constant *Quieted = ConstantFP::get(Ty,
C->getValue().makeQuiet());
1200 if (
C->isZero() || (
C->getValue().isDenormal() && Ty->isFloatTy())) {
1202 : ConstantFP::get(Ty, 1.0);
1206 if (IsLog &&
C->isNegative())
1214 case Intrinsic::amdgcn_frexp_mant:
1215 case Intrinsic::amdgcn_frexp_exp: {
1216 Value *Src =
II.getArgOperand(0);
1222 if (IID == Intrinsic::amdgcn_frexp_mant) {
1224 II, ConstantFP::get(
II.getContext(), Significand));
1244 case Intrinsic::amdgcn_class: {
1245 Value *Src0 =
II.getArgOperand(0);
1246 Value *Src1 =
II.getArgOperand(1);
1250 II.getModule(), Intrinsic::is_fpclass, Src0->
getType()));
1253 II.setArgOperand(1, ConstantInt::get(Src1->
getType(),
1274 case Intrinsic::amdgcn_cvt_pkrtz: {
1275 auto foldFPTruncToF16RTZ = [](
Value *Arg) ->
Value * {
1288 return ConstantFP::get(HalfTy, Val);
1291 Value *Src =
nullptr;
1293 if (Src->getType()->isHalfTy())
1300 if (
Value *Src0 = foldFPTruncToF16RTZ(
II.getArgOperand(0))) {
1301 if (
Value *Src1 = foldFPTruncToF16RTZ(
II.getArgOperand(1))) {
1311 case Intrinsic::amdgcn_cvt_pknorm_i16:
1312 case Intrinsic::amdgcn_cvt_pknorm_u16:
1313 case Intrinsic::amdgcn_cvt_pk_i16:
1314 case Intrinsic::amdgcn_cvt_pk_u16: {
1315 Value *Src0 =
II.getArgOperand(0);
1316 Value *Src1 =
II.getArgOperand(1);
1328 case Intrinsic::amdgcn_cvt_off_f32_i4: {
1329 Value* Arg =
II.getArgOperand(0);
1343 constexpr size_t ResValsSize = 16;
1344 static constexpr float ResVals[ResValsSize] = {
1345 0.0, 0.0625, 0.125, 0.1875, 0.25, 0.3125, 0.375, 0.4375,
1346 -0.5, -0.4375, -0.375, -0.3125, -0.25, -0.1875, -0.125, -0.0625};
1348 ConstantFP::get(Ty, ResVals[CArg->
getZExtValue() & (ResValsSize - 1)]);
1351 case Intrinsic::amdgcn_ubfe:
1352 case Intrinsic::amdgcn_sbfe: {
1354 Value *Src =
II.getArgOperand(0);
1361 unsigned IntSize = Ty->getIntegerBitWidth();
1366 if ((Width & (IntSize - 1)) == 0) {
1371 if (Width >= IntSize) {
1373 II, 2, ConstantInt::get(CWidth->
getType(), Width & (IntSize - 1)));
1384 ConstantInt::get(COffset->
getType(),
Offset & (IntSize - 1)));
1388 bool Signed = IID == Intrinsic::amdgcn_sbfe;
1390 if (!CWidth || !COffset)
1400 if (
Offset + Width < IntSize) {
1404 RightShift->takeName(&
II);
1411 RightShift->takeName(&
II);
1414 case Intrinsic::amdgcn_exp:
1415 case Intrinsic::amdgcn_exp_row:
1416 case Intrinsic::amdgcn_exp_compr: {
1422 bool IsCompr = IID == Intrinsic::amdgcn_exp_compr;
1424 for (
int I = 0;
I < (IsCompr ? 2 : 4); ++
I) {
1425 if ((!IsCompr && (EnBits & (1 <<
I)) == 0) ||
1426 (IsCompr && ((EnBits & (0x3 << (2 *
I))) == 0))) {
1427 Value *Src =
II.getArgOperand(
I + 2);
1441 case Intrinsic::amdgcn_fmed3: {
1442 Value *Src0 =
II.getArgOperand(0);
1443 Value *Src1 =
II.getArgOperand(1);
1444 Value *Src2 =
II.getArgOperand(2);
1446 for (
Value *Src : {Src0, Src1, Src2}) {
1451 if (
II.isStrictFP())
1488 const APFloat *ConstSrc0 =
nullptr;
1489 const APFloat *ConstSrc1 =
nullptr;
1490 const APFloat *ConstSrc2 =
nullptr;
1495 const bool IsPosInfinity = ConstSrc0 && ConstSrc0->
isPosInfinity();
1515 const bool IsPosInfinity = ConstSrc1 && ConstSrc1->
isPosInfinity();
1538 auto *Quieted = ConstantFP::get(
II.getType(), ConstSrc2->
makeQuiet());
1558 CI->copyFastMathFlags(&
II);
1584 II.setArgOperand(0, Src0);
1585 II.setArgOperand(1, Src1);
1586 II.setArgOperand(2, Src2);
1596 ConstantFP::get(
II.getType(), Result));
1601 if (!ST->hasMed3_16())
1610 IID, {
X->getType()}, {
X,
Y, Z}, &
II,
II.getName());
1618 case Intrinsic::amdgcn_icmp:
1619 case Intrinsic::amdgcn_fcmp: {
1623 bool IsInteger = IID == Intrinsic::amdgcn_icmp;
1630 Value *Src0 =
II.getArgOperand(0);
1631 Value *Src1 =
II.getArgOperand(1);
1651 II.getType(), Args);
1652 NewCall->
addFnAttr(Attribute::Convergent);
1660 II.setArgOperand(0, Src1);
1661 II.setArgOperand(1, Src0);
1663 2, ConstantInt::get(CC->
getType(),
static_cast<int>(SwapPred)));
1710 ? Intrinsic::amdgcn_fcmp
1711 : Intrinsic::amdgcn_icmp;
1716 unsigned Width = CmpType->getBitWidth();
1717 unsigned NewWidth = Width;
1725 else if (Width <= 32)
1727 else if (Width <= 64)
1732 if (Width != NewWidth) {
1742 }
else if (!Ty->isFloatTy() && !Ty->isDoubleTy() && !Ty->isHalfTy())
1745 Value *Args[] = {SrcLHS, SrcRHS,
1746 ConstantInt::get(CC->
getType(), SrcPred)};
1748 NewIID, {
II.getType(), SrcLHS->
getType()}, Args);
1755 case Intrinsic::amdgcn_mbcnt_hi:
1760 case Intrinsic::amdgcn_mbcnt_lo: {
1773 if (std::optional<ConstantRange> ExistingRange =
II.getRange()) {
1774 ComputedRange = ComputedRange.
intersectWith(*ExistingRange);
1775 if (ComputedRange == *ExistingRange)
1779 II.addRangeRetAttr(ComputedRange);
1782 case Intrinsic::amdgcn_ballot: {
1783 Value *Arg =
II.getArgOperand(0);
1788 if (Src->isZero()) {
1793 if (ST->isWave32() &&
II.getType()->getIntegerBitWidth() == 64) {
1800 {IC.Builder.getInt32Ty()},
1801 {II.getArgOperand(0)}),
1808 case Intrinsic::amdgcn_wavefrontsize: {
1809 if (ST->isWaveSizeKnown())
1811 II, ConstantInt::get(
II.getType(), ST->getWavefrontSize()));
1814 case Intrinsic::amdgcn_wqm_vote: {
1821 case Intrinsic::amdgcn_kill: {
1823 if (!
C || !
C->getZExtValue())
1829 case Intrinsic::amdgcn_s_sendmsg:
1830 case Intrinsic::amdgcn_s_sendmsghalt: {
1836 Value *M0Val =
II.getArgOperand(1);
1842 decodeMsg(MsgImm->getZExtValue(), MsgId, OpId, StreamId, *ST);
1844 if (!msgDoesNotUseM0(MsgId, *ST))
1848 II.dropUBImplyingAttrsAndMetadata();
1852 case Intrinsic::amdgcn_update_dpp: {
1853 Value *Old =
II.getArgOperand(0);
1858 if (BC->isNullValue() || RM->getZExtValue() != 0xF ||
1865 case Intrinsic::amdgcn_permlane16:
1866 case Intrinsic::amdgcn_permlane16_var:
1867 case Intrinsic::amdgcn_permlanex16:
1868 case Intrinsic::amdgcn_permlanex16_var: {
1870 Value *VDstIn =
II.getArgOperand(0);
1875 unsigned int FiIdx = (IID == Intrinsic::amdgcn_permlane16 ||
1876 IID == Intrinsic::amdgcn_permlanex16)
1883 unsigned int BcIdx = FiIdx + 1;
1892 case Intrinsic::amdgcn_wave_shuffle:
1894 case Intrinsic::amdgcn_permlane64:
1895 case Intrinsic::amdgcn_readfirstlane:
1896 case Intrinsic::amdgcn_readlane:
1897 case Intrinsic::amdgcn_ds_bpermute: {
1899 unsigned SrcIdx = IID == Intrinsic::amdgcn_ds_bpermute ? 1 : 0;
1900 const Use &Src =
II.getArgOperandUse(SrcIdx);
1904 if (IID == Intrinsic::amdgcn_readlane &&
1911 if (IID == Intrinsic::amdgcn_ds_bpermute) {
1912 const Use &Lane =
II.getArgOperandUse(0);
1916 II.getModule(), Intrinsic::amdgcn_readlane,
II.getType());
1917 II.setCalledFunction(NewDecl);
1918 II.setOperand(0, Src);
1919 II.setOperand(1, NewLane);
1924 if (IID == Intrinsic::amdgcn_ds_bpermute)
1930 return std::nullopt;
1932 case Intrinsic::amdgcn_writelane: {
1936 return std::nullopt;
1938 case Intrinsic::amdgcn_trig_preop: {
1941 if (!
II.getType()->isDoubleTy())
1944 Value *Src =
II.getArgOperand(0);
1945 Value *Segment =
II.getArgOperand(1);
1954 if (StrippedSign != Src)
1957 if (
II.isStrictFP())
1979 unsigned Shift = SegmentVal * 53;
1984 static const uint32_t TwoByPi[] = {
1985 0xa2f9836e, 0x4e441529, 0xfc2757d1, 0xf534ddc0, 0xdb629599, 0x3c439041,
1986 0xfe5163ab, 0xdebbc561, 0xb7246e3a, 0x424dd2e0, 0x06492eea, 0x09d1921c,
1987 0xfe1deb1c, 0xb129a73e, 0xe88235f5, 0x2ebb4484, 0xe99c7026, 0xb45f7e41,
1988 0x3991d639, 0x835339f4, 0x9c845f8b, 0xbdf9283b, 0x1ff897ff, 0xde05980f,
1989 0xef2f118b, 0x5a0a6d1f, 0x6d367ecf, 0x27cb09b7, 0x4f463f66, 0x9e5fea2d,
1990 0x7527bac7, 0xebe5f17b, 0x3d0739f7, 0x8a5292ea, 0x6bfb5fb1, 0x1f8d5d08,
1994 unsigned Idx = Shift >> 5;
1995 if (Idx + 2 >= std::size(TwoByPi)) {
2000 unsigned BShift = Shift & 0x1f;
2004 Thi = (Thi << BShift) | (Tlo >> (64 - BShift));
2008 int Scale = -53 - Shift;
2015 case Intrinsic::amdgcn_fmul_legacy: {
2016 Value *Op0 =
II.getArgOperand(0);
2017 Value *Op1 =
II.getArgOperand(1);
2019 for (
Value *Src : {Op0, Op1}) {
2040 case Intrinsic::amdgcn_fma_legacy: {
2041 Value *Op0 =
II.getArgOperand(0);
2042 Value *Op1 =
II.getArgOperand(1);
2043 Value *Op2 =
II.getArgOperand(2);
2045 for (
Value *Src : {Op0, Op1, Op2}) {
2067 II.getModule(), Intrinsic::fma,
II.getType()));
2072 case Intrinsic::amdgcn_is_shared:
2073 case Intrinsic::amdgcn_is_private: {
2074 Value *Src =
II.getArgOperand(0);
2084 case Intrinsic::amdgcn_make_buffer_rsrc: {
2085 Value *Src =
II.getArgOperand(0);
2088 return std::nullopt;
2090 case Intrinsic::amdgcn_raw_buffer_store_format:
2091 case Intrinsic::amdgcn_struct_buffer_store_format:
2092 case Intrinsic::amdgcn_raw_tbuffer_store:
2093 case Intrinsic::amdgcn_struct_tbuffer_store:
2094 case Intrinsic::amdgcn_image_store_1d:
2095 case Intrinsic::amdgcn_image_store_1darray:
2096 case Intrinsic::amdgcn_image_store_2d:
2097 case Intrinsic::amdgcn_image_store_2darray:
2098 case Intrinsic::amdgcn_image_store_2darraymsaa:
2099 case Intrinsic::amdgcn_image_store_2dmsaa:
2100 case Intrinsic::amdgcn_image_store_3d:
2101 case Intrinsic::amdgcn_image_store_cube:
2102 case Intrinsic::amdgcn_image_store_mip_1d:
2103 case Intrinsic::amdgcn_image_store_mip_1darray:
2104 case Intrinsic::amdgcn_image_store_mip_2d:
2105 case Intrinsic::amdgcn_image_store_mip_2darray:
2106 case Intrinsic::amdgcn_image_store_mip_3d:
2107 case Intrinsic::amdgcn_image_store_mip_cube: {
2112 if (ST->hasDefaultComponentBroadcast())
2114 else if (ST->hasDefaultComponentZero())
2119 int DMaskIdx = getAMDGPUImageDMaskIntrinsic(
II.getIntrinsicID()) ? 1 : -1;
2127 case Intrinsic::amdgcn_prng_b32: {
2128 auto *Src =
II.getArgOperand(0);
2132 return std::nullopt;
2134 case Intrinsic::amdgcn_mfma_scale_f32_16x16x128_f8f6f4:
2135 case Intrinsic::amdgcn_mfma_scale_f32_32x32x64_f8f6f4: {
2136 Value *Src0 =
II.getArgOperand(0);
2137 Value *Src1 =
II.getArgOperand(1);
2143 auto getFormatNumRegs = [](
unsigned FormatVal) {
2144 switch (FormatVal) {
2158 bool MadeChange =
false;
2159 unsigned Src0NumElts = getFormatNumRegs(CBSZ);
2160 unsigned Src1NumElts = getFormatNumRegs(BLGP);
2164 if (Src0Ty->getNumElements() > Src0NumElts) {
2171 if (Src1Ty->getNumElements() > Src1NumElts) {
2179 return std::nullopt;
2190 case Intrinsic::amdgcn_wmma_f32_16x16x128_f8f6f4:
2191 case Intrinsic::amdgcn_wmma_scale_f32_16x16x128_f8f6f4:
2192 case Intrinsic::amdgcn_wmma_scale16_f32_16x16x128_f8f6f4: {
2193 Value *Src0 =
II.getArgOperand(1);
2194 Value *Src1 =
II.getArgOperand(3);
2200 bool MadeChange =
false;
2206 if (Src0Ty->getNumElements() > Src0NumElts) {
2213 if (Src1Ty->getNumElements() > Src1NumElts) {
2221 return std::nullopt;
2238 return std::nullopt;
2251 int DMaskIdx,
bool IsLoad) {
2254 :
II.getOperand(0)->getType());
2255 unsigned VWidth = IIVTy->getNumElements();
2258 Type *EltTy = IIVTy->getElementType();
2270 const unsigned UnusedComponentsAtFront = DemandedElts.
countr_zero();
2275 DemandedElts = (1 << ActiveBits) - 1;
2277 if (UnusedComponentsAtFront > 0) {
2278 static const unsigned InvalidOffsetIdx = 0xf;
2281 switch (
II.getIntrinsicID()) {
2282 case Intrinsic::amdgcn_raw_buffer_load:
2283 case Intrinsic::amdgcn_raw_ptr_buffer_load:
2286 case Intrinsic::amdgcn_s_buffer_load:
2290 if (ActiveBits == 4 && UnusedComponentsAtFront == 1)
2291 OffsetIdx = InvalidOffsetIdx;
2295 case Intrinsic::amdgcn_struct_buffer_load:
2296 case Intrinsic::amdgcn_struct_ptr_buffer_load:
2301 OffsetIdx = InvalidOffsetIdx;
2305 if (OffsetIdx != InvalidOffsetIdx) {
2307 DemandedElts &= ~((1 << UnusedComponentsAtFront) - 1);
2308 auto *
Offset = Args[OffsetIdx];
2309 unsigned SingleComponentSizeInBits =
2311 unsigned OffsetAdd =
2312 UnusedComponentsAtFront * SingleComponentSizeInBits / 8;
2313 auto *OffsetAddVal = ConstantInt::get(
Offset->getType(), OffsetAdd);
2330 unsigned NewDMaskVal = 0;
2331 unsigned OrigLdStIdx = 0;
2332 for (
unsigned SrcIdx = 0; SrcIdx < 4; ++SrcIdx) {
2333 const unsigned Bit = 1 << SrcIdx;
2334 if (!!(DMaskVal & Bit)) {
2335 if (!!DemandedElts[OrigLdStIdx])
2341 if (DMaskVal != NewDMaskVal)
2342 Args[DMaskIdx] = ConstantInt::get(DMask->
getType(), NewDMaskVal);
2345 unsigned NewNumElts = DemandedElts.
popcount();
2349 if (NewNumElts >= VWidth && DemandedElts.
isMask()) {
2351 II.setArgOperand(DMaskIdx, Args[DMaskIdx]);
2363 OverloadTys[0] = NewTy;
2367 for (
unsigned OrigStoreIdx = 0; OrigStoreIdx < VWidth; ++OrigStoreIdx)
2368 if (DemandedElts[OrigStoreIdx])
2371 if (NewNumElts == 1)
2383 if (NewNumElts == 1) {
2389 unsigned NewLoadIdx = 0;
2390 for (
unsigned OrigLoadIdx = 0; OrigLoadIdx < VWidth; ++OrigLoadIdx) {
2391 if (!!DemandedElts[OrigLoadIdx])
2407 APInt &UndefElts)
const {
2412 const unsigned FirstElt = DemandedElts.
countr_zero();
2414 const unsigned MaskLen = LastElt - FirstElt + 1;
2416 unsigned OldNumElts = VT->getNumElements();
2417 if (MaskLen == OldNumElts && MaskLen != 1)
2420 Type *EltTy = VT->getElementType();
2428 Value *Src =
II.getArgOperand(0);
2433 II.getOperandBundlesAsDefs(OpBundles);
2450 for (
unsigned I = 0;
I != MaskLen; ++
I) {
2451 if (DemandedElts[FirstElt +
I])
2452 ExtractMask[
I] = FirstElt +
I;
2461 for (
unsigned I = 0;
I != MaskLen; ++
I) {
2462 if (DemandedElts[FirstElt +
I])
2463 InsertMask[FirstElt +
I] =
I;
2475 SimplifyAndSetOp)
const {
2476 switch (
II.getIntrinsicID()) {
2477 case Intrinsic::amdgcn_readfirstlane:
2478 SimplifyAndSetOp(&
II, 0, DemandedElts, UndefElts);
2480 case Intrinsic::amdgcn_raw_buffer_load:
2481 case Intrinsic::amdgcn_raw_ptr_buffer_load:
2482 case Intrinsic::amdgcn_raw_buffer_load_format:
2483 case Intrinsic::amdgcn_raw_ptr_buffer_load_format:
2484 case Intrinsic::amdgcn_raw_tbuffer_load:
2485 case Intrinsic::amdgcn_raw_ptr_tbuffer_load:
2486 case Intrinsic::amdgcn_s_buffer_load:
2487 case Intrinsic::amdgcn_struct_buffer_load:
2488 case Intrinsic::amdgcn_struct_ptr_buffer_load:
2489 case Intrinsic::amdgcn_struct_buffer_load_format:
2490 case Intrinsic::amdgcn_struct_ptr_buffer_load_format:
2491 case Intrinsic::amdgcn_struct_tbuffer_load:
2492 case Intrinsic::amdgcn_struct_ptr_tbuffer_load:
2495 if (getAMDGPUImageDMaskIntrinsic(
II.getIntrinsicID())) {
2501 return std::nullopt;
for(const MachineOperand &MO :llvm::drop_begin(OldMI.operands(), Desc.getNumOperands()))
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
static Value * createPermlane16(IRBuilderBase &B, Value *Val, uint32_t Lo, uint32_t Hi)
Emit v_permlane16 with the precomputed lane-select halves.
static std::optional< unsigned > matchRowSharePattern(ArrayRef< uint8_t > Ids)
Match a row-share pattern: all 16 lanes of each row read the same source lane.
static bool matchMirrorPattern(ArrayRef< uint8_t > Ids)
Match an N-lane reversal (mirror) pattern.
static bool tryBuildShuffleMap(Value *Index, const GCNSubtarget &ST, SmallVectorImpl< uint8_t > &Ids, const DataLayout &DL)
Build the per-lane shuffle map by evaluating Index for every lane in the wave.
static std::optional< unsigned > matchQuadPermPattern(ArrayRef< uint8_t > Ids)
Match a 4-lane (quad) permutation, encoded as the v_mov_b32_dpp QUAD_PERM control word: bits[1:0]=Ids...
static std::optional< unsigned > matchHalfRowPermPattern(ArrayRef< uint8_t > Ids)
Match an 8-lane arbitrary permutation, encoded as the v_mov_b32_dpp8 24-bit selector (three bits per ...
static std::optional< unsigned > matchRowXMaskPattern(ArrayRef< uint8_t > Ids)
Match an XOR mask pattern within each 16-lane row: Ids[J] == Mask ^ J, with Mask in [1,...
static constexpr auto matchHalfRowMirrorPattern
static Value * createPermlaneX16(IRBuilderBase &B, Value *Val, uint32_t Lo, uint32_t Hi)
Emit v_permlanex16 with the precomputed lane-select halves.
static bool isRowPattern(ArrayRef< uint8_t > Ids)
Match an N-lane row pattern: each lane in [0, N) reads from a source lane in the same N-lane row,...
static bool canContractSqrtToRsq(const FPMathOperator *SqrtOp)
Return true if it's legal to contract llvm.amdgcn.rcp(llvm.sqrt)
static bool isTriviallyUniform(const Use &U)
Return true if we can easily prove that use U is uniform.
static CallInst * rewriteCall(IRBuilderBase &B, CallInst &Old, Function &NewCallee, ArrayRef< Value * > Ops)
static Value * convertTo16Bit(Value &V, InstCombiner::BuilderTy &Builder)
static constexpr auto isFullRowPattern
static constexpr auto isQuadPattern
static APInt trimTrailingZerosInVector(InstCombiner &IC, Value *UseV, Instruction *I)
static uint64_t computePermlane16Masks(ArrayRef< uint8_t > Ids)
Pack a 16-lane permutation into a single 64-bit value: four bits per output lane, lane J in bits [J*4...
static bool matchHalfWaveSwapPattern(ArrayRef< uint8_t > Ids)
Match a half-wave swap: lane J reads from lane J ^ 32.
static bool hasPeriodicLayout(ArrayRef< uint8_t > Ids)
Lanes are partitioned into groups of Period; each group is a translated copy of the first: Ids[I] = I...
static std::optional< Instruction * > tryOptimizeShufflePattern(InstCombiner &IC, IntrinsicInst &II, const GCNSubtarget &ST)
Try to fold a wave_shuffle/ds_bpermute whose lane index is a constant function of the lane ID into a ...
static constexpr auto isHalfRowPattern
static APInt defaultComponentBroadcast(Value *V)
static std::optional< unsigned > matchDsSwizzleBitmaskPattern(ArrayRef< uint8_t > Ids)
Match a DS_SWIZZLE bitmask-mode permutation: dst_lane = ((src_lane & AND) | OR) ^ XOR with each mask ...
static Value * createDsSwizzle(IRBuilderBase &B, Value *Val, unsigned Offset, const DataLayout &DL)
Emit ds_swizzle with the given immediate, bitcasting/converting between pointer/float types and i32 a...
static std::optional< Instruction * > modifyIntrinsicCall(IntrinsicInst &OldIntr, Instruction &InstToReplace, unsigned NewIntr, InstCombiner &IC, std::function< void(SmallVectorImpl< Value * > &, SmallVectorImpl< Type * > &)> Func)
Applies Func(OldIntr.Args, OldIntr.ArgTys), creates intrinsic call with modified arguments (based on ...
static Value * matchShuffleToHWIntrinsic(IRBuilderBase &B, Value *Src, ArrayRef< uint8_t > Ids, const GCNSubtarget &ST, const DataLayout &DL)
Given a shuffle map, try to emit the best hardware intrinsic.
static std::optional< unsigned > matchRowRotatePattern(ArrayRef< uint8_t > Ids)
Match a 16-lane cyclic rotation; returns the rotation amount in [1, 15].
static bool isCrossRowPattern(ArrayRef< uint8_t > Ids)
Match a cross-row permutation suitable for v_permlanex16: every lane in the low 16-lane half reads fr...
static bool isThreadID(const GCNSubtarget &ST, Value *V)
static Value * createUpdateDpp(IRBuilderBase &B, Value *Val, unsigned Ctrl)
Emit v_mov_b32_dpp with the given control word, row/bank masks 0xF, and bound_ctrl=1 so out-of-bounds...
static APFloat fmed3AMDGCN(const APFloat &Src0, const APFloat &Src1, const APFloat &Src2)
static Value * simplifyAMDGCNMemoryIntrinsicDemanded(InstCombiner &IC, IntrinsicInst &II, APInt DemandedElts, int DMaskIdx=-1, bool IsLoad=true)
Implement SimplifyDemandedVectorElts for amdgcn buffer and image intrinsics.
static std::optional< Instruction * > simplifyAMDGCNImageIntrinsic(const GCNSubtarget *ST, const AMDGPU::ImageDimIntrinsicInfo *ImageDimIntr, IntrinsicInst &II, InstCombiner &IC)
static bool canSafelyConvertTo16Bit(Value &V, bool IsFloat)
static Value * createMovDpp8(IRBuilderBase &B, Value *Val, unsigned Selector)
Emit v_mov_b32_dpp8 with the given 24-bit lane selector.
static Value * matchFPExtFromF16(Value *Arg)
Match an fpext from half to float, or a constant we can convert.
static constexpr auto matchFullRowMirrorPattern
static std::optional< unsigned > evalLaneExpr(Value *V, unsigned Lane, const GCNSubtarget &ST, const DataLayout &DL, unsigned Depth=0)
Evaluate V as a function of the lane ID and return its value on Lane, or std::nullopt if V is not a c...
static Value * createPermlane64(IRBuilderBase &B, Value *Val)
Emit v_permlane64 (swap of the two 32-lane halves of a wave64).
Contains the definition of a TargetInstrInfo class that is common to all AMD GPUs.
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
This file contains the declarations for the subclasses of Constant, which represent the different fla...
Utilities for dealing with flags related to floating point properties and mode controls.
AMD GCN specific subclass of TargetSubtarget.
This file provides the interface for the instcombine pass implementation.
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
MachineInstr unsigned OpIdx
uint64_t IntrinsicInst * II
Provides some synthesis utilities to produce sequences of values.
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
cmpResult
IEEE-754R 5.11: Floating Point Comparison Relations.
static constexpr roundingMode rmTowardZero
static constexpr roundingMode rmNearestTiesToEven
static const fltSemantics & IEEEhalf()
static APFloat getQNaN(const fltSemantics &Sem, bool Negative=false, const APInt *payload=nullptr)
Factory for QNaN values.
opStatus divide(const APFloat &RHS, roundingMode RM)
LLVM_ABI opStatus convert(const fltSemantics &ToSemantics, roundingMode RM, bool *losesInfo)
bool isPosInfinity() const
const fltSemantics & getSemantics() const
APFloat makeQuiet() const
Assuming this is an IEEE-754 NaN value, quiet its signaling bit.
APInt bitcastToAPInt() const
bool isNegInfinity() const
static APFloat getZero(const fltSemantics &Sem, bool Negative=false)
Factory for Positive and Negative Zero.
cmpResult compare(const APFloat &RHS) const
Class for arbitrary precision integers.
static APInt getAllOnes(unsigned numBits)
Return an APInt of a specified width with all bits set.
void clearBit(unsigned BitPosition)
Set a given bit to 0.
uint64_t getZExtValue() const
Get zero extended value.
unsigned popcount() const
Count the number of bits set.
LLVM_ABI uint64_t extractBitsAsZExtValue(unsigned numBits, unsigned bitPosition) const
unsigned getActiveBits() const
Compute the number of active bits in the value.
LLVM_ABI APInt trunc(unsigned width) const
Truncate to new width.
unsigned countr_zero() const
Count the number of trailing zero bits.
bool isMask(unsigned numBits) const
Represent a constant reference to an array (0 or more elements consecutively in memory),...
ArrayRef< T > take_front(size_t N=1) const
Return a copy of *this with only the first N elements.
size_t size() const
Get the array size.
static LLVM_ABI Attribute getWithDereferenceableBytes(LLVMContext &Context, uint64_t Bytes)
LLVM_ABI const Module * getModule() const
Return the module owning the function this basic block belongs to, or nullptr if the function does no...
bool isTypeLegal(Type *Ty) const override
void addFnAttr(Attribute::AttrKind Kind)
Adds the attribute to the function.
LLVM_ABI void getOperandBundlesAsDefs(SmallVectorImpl< OperandBundleDef > &Defs) const
Return the list of operand bundles attached to this instruction as a vector of OperandBundleDefs.
Function * getCalledFunction() const
Returns the function called, or null if this is an indirect function invocation or the function signa...
iterator_range< User::op_iterator > args()
Iteration adapter for range-for loops.
This class represents a function call, abstracting a target machine's calling convention.
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Predicate getSwappedPredicate() const
For example, EQ->EQ, SLE->SGE, ULT->UGT, OEQ->OEQ, ULE->UGE, OLT->OGT, etc.
bool isFPPredicate() const
Predicate getInversePredicate() const
For example, EQ -> NE, UGT -> ULE, SLT -> SGE, OEQ -> UNE, UGT -> OLE, OLT -> UGE,...
An abstraction over a floating-point predicate, and a pack of an integer predicate with samesign info...
ConstantFP - Floating Point Values [float, double].
const APFloat & getValueAPF() const
static LLVM_ABI Constant * getInfinity(Type *Ty, bool Negative=false)
static LLVM_ABI Constant * getZero(Type *Ty, bool Negative=false)
static LLVM_ABI Constant * getNaN(Type *Ty, bool Negative=false, uint64_t Payload=0)
This is the shared class of boolean and integer constants.
static ConstantInt * getSigned(IntegerType *Ty, int64_t V, bool ImplicitTrunc=false)
Return a ConstantInt with the specified value for the specified type.
static LLVM_ABI ConstantInt * getFalse(LLVMContext &Context)
uint64_t getZExtValue() const
Return the constant as a 64-bit unsigned integer value after it has been zero extended as appropriate...
const APInt & getValue() const
Return the constant as an APInt value reference.
This class represents a range of values.
LLVM_ABI ConstantRange add(const ConstantRange &Other) const
Return a new range representing the possible values resulting from an addition of a value in this ran...
LLVM_ABI bool isFullSet() const
Return true if this set contains all of the elements possible for this data-type.
LLVM_ABI ConstantRange intersectWith(const ConstantRange &CR, PreferredRangeType Type=Smallest) const
Return the range that results from the intersection of this range with another range.
This is an important base class in LLVM.
static LLVM_ABI Constant * getNullValue(Type *Ty)
Constructor to create a '0' constant of arbitrary type.
LLVM_ABI bool isNullValue() const
Return true if this is the value that would be returned by getNullValue.
A parsed version of the target data layout string in and methods for querying it.
TypeSize getTypeSizeInBits(Type *Ty) const
Size examples:
LLVM_ABI bool dominates(const BasicBlock *BB, const Use &U) const
Return true if the (end of the) basic block BB dominates the use U.
Tagged union holding either a T or a Error.
This class represents an extension of floating point types.
Utility class for floating point operations which can have information about relaxed accuracy require...
FastMathFlags getFastMathFlags() const
Convenience function for getting all the fast-math flags.
bool hasApproxFunc() const
Test if this operation allows approximations of math library functions or intrinsics.
LLVM_ABI float getFPAccuracy() const
Get the maximum error permitted by this operation in ULPs.
Convenience struct for specifying and reasoning about fast-math flags.
bool allowContract() const
static LLVM_ABI FixedVectorType * get(Type *ElementType, unsigned NumElts)
bool simplifyDemandedLaneMaskArg(InstCombiner &IC, IntrinsicInst &II, unsigned LaneAgIdx) const
Simplify a lane index operand (e.g.
std::optional< Instruction * > instCombineIntrinsic(InstCombiner &IC, IntrinsicInst &II) const override
Instruction * hoistLaneIntrinsicThroughOperand(InstCombiner &IC, IntrinsicInst &II) const
std::optional< Value * > simplifyDemandedVectorEltsIntrinsic(InstCombiner &IC, IntrinsicInst &II, APInt DemandedElts, APInt &UndefElts, APInt &UndefElts2, APInt &UndefElts3, std::function< void(Instruction *, unsigned, APInt, APInt &)> SimplifyAndSetOp) const override
KnownIEEEMode fpenvIEEEMode(const Instruction &I) const
Return KnownIEEEMode::On if we know if the use context can assume "amdgpu-ieee"="true" and KnownIEEEM...
Value * simplifyAMDGCNLaneIntrinsicDemanded(InstCombiner &IC, IntrinsicInst &II, const APInt &DemandedElts, APInt &UndefElts) const
bool canSimplifyLegacyMulToMul(const Instruction &I, const Value *Op0, const Value *Op1, InstCombiner &IC) const
Common base class shared among various IRBuilders.
CallInst * CreateExtractVector(Type *DstType, Value *SrcVec, Value *Idx, const Twine &Name="")
Create a call to the vector.extract intrinsic.
Value * CreateInsertElement(Type *VecTy, Value *NewElt, Value *Idx, const Twine &Name="")
Value * CreateExtractElement(Value *Vec, Value *Idx, const Twine &Name="")
IntegerType * getIntNTy(unsigned N)
Fetch the type representing an N-bit integer.
LLVM_ABI CallInst * CreateIntrinsic(Intrinsic::ID ID, ArrayRef< Type * > OverloadTypes, ArrayRef< Value * > Args, FMFSource FMFSource={}, const Twine &Name="", ArrayRef< OperandBundleDef > OpBundles={})
Create a call to intrinsic ID with Args, mangled using OverloadTypes.
Value * CreateSExt(Value *V, Type *DestTy, const Twine &Name="")
Value * CreateLShr(Value *LHS, Value *RHS, const Twine &Name="", bool isExact=false)
BasicBlock * GetInsertBlock() const
Value * CreateICmpNE(Value *LHS, Value *RHS, const Twine &Name="")
ConstantInt * getInt64(uint64_t C)
Get a constant 64-bit value.
Value * CreateMaxNum(Value *LHS, Value *RHS, FMFSource FMFSource={}, const Twine &Name="")
Create call to the maxnum intrinsic.
Value * CreateShl(Value *LHS, Value *RHS, const Twine &Name="", bool HasNUW=false, bool HasNSW=false)
Value * CreateZExt(Value *V, Type *DestTy, const Twine &Name="", bool IsNonNeg=false)
Value * CreateShuffleVector(Value *V1, Value *V2, Value *Mask, const Twine &Name="")
Value * CreateMaximumNum(Value *LHS, Value *RHS, const Twine &Name="")
Create call to the maximum intrinsic.
Value * CreateMinNum(Value *LHS, Value *RHS, FMFSource FMFSource={}, const Twine &Name="")
Create call to the minnum intrinsic.
Value * CreateAdd(Value *LHS, Value *RHS, const Twine &Name="", bool HasNUW=false, bool HasNSW=false)
CallInst * CreateCall(FunctionType *FTy, Value *Callee, ArrayRef< Value * > Args={}, const Twine &Name="", MDNode *FPMathTag=nullptr)
void SetInsertPoint(BasicBlock *TheBB)
This specifies that created instructions should be appended to the end of the specified block.
Value * CreateFAddFMF(Value *L, Value *R, FMFSource FMFSource, const Twine &Name="", MDNode *FPMD=nullptr)
Value * CreateMinimumNum(Value *LHS, Value *RHS, const Twine &Name="")
Create call to the minimumnum intrinsic.
Value * CreateAShr(Value *LHS, Value *RHS, const Twine &Name="", bool isExact=false)
Value * CreateFMulFMF(Value *L, Value *R, FMFSource FMFSource, const Twine &Name="", MDNode *FPMD=nullptr)
This provides a uniform API for creating instructions and inserting them into a basic block: either a...
The core instruction combiner logic.
const DataLayout & getDataLayout() const
virtual Instruction * eraseInstFromFunction(Instruction &I)=0
Combiner aware instruction erasure.
IRBuilder< TargetFolder, IRBuilderCallbackInserter > BuilderTy
An IRBuilder that automatically inserts new instructions into the worklist.
DominatorTree & getDominatorTree() const
Instruction * replaceInstUsesWith(Instruction &I, Value *V)
A combiner-aware RAUW-like routine.
virtual bool SimplifyDemandedBits(Instruction *I, unsigned OpNo, const APInt &DemandedMask, KnownBits &Known, const SimplifyQuery &Q, unsigned Depth=0)=0
static Value * stripSignOnlyFPOps(Value *Val)
Ignore all operations which only change the sign of a value, returning the underlying magnitude value...
Instruction * replaceOperand(Instruction &I, unsigned OpNum, Value *V)
Replace operand of instruction and add old operand to the worklist.
const SimplifyQuery & getSimplifyQuery() const
LLVM_ABI Instruction * clone() const
Create a copy of 'this' instruction that is identical in all ways except the following:
LLVM_ABI void copyFastMathFlags(FastMathFlags FMF)
Convenience function for transferring all fast-math flag values to this instruction,...
LLVM_ABI void copyMetadata(const Instruction &SrcInst, ArrayRef< unsigned > WL=ArrayRef< unsigned >())
Copy metadata from SrcInst to this instruction.
Class to represent integer types.
A wrapper class for inspecting calls to intrinsic functions.
static MDTuple * get(LLVMContext &Context, ArrayRef< Metadata * > MDs)
static LLVM_ABI MDString * get(LLVMContext &Context, StringRef Str)
A Module instance is used to store all the information related to an LLVM module.
static LLVM_ABI PoisonValue * get(Type *T)
Static factory methods - Return an 'poison' object of the specified type.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
reference emplace_back(ArgTypes &&... Args)
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
The instances of the Type class are immutable: once they are created, they are never changed.
bool isPointerTy() const
True if this is an instance of PointerType.
bool isFloatTy() const
Return true if this is 'float', a 32-bit IEEE fp type.
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
static LLVM_ABI IntegerType * getInt16Ty(LLVMContext &C)
bool isHalfTy() const
Return true if this is 'half', a 16-bit IEEE fp type.
LLVM_ABI Type * getWithNewType(Type *EltTy) const
Given vector type, change the element type, whilst keeping the old number of elements.
bool isFloatingPointTy() const
Return true if this is one of the floating-point types.
bool isIntegerTy() const
True if this is an instance of IntegerType.
static LLVM_ABI Type * getHalfTy(LLVMContext &C)
bool isVoidTy() const
Return true if this is 'void'.
static LLVM_ABI UndefValue * get(Type *T)
Static factory methods - Return an 'undef' object of the specified type.
A Use represents the edge between a Value definition and its users.
const Use & getOperandUse(unsigned i) const
void setOperand(unsigned i, Value *Val)
Value * getOperand(unsigned i) const
LLVM Value Representation.
Type * getType() const
All values are typed, get the type of this value.
LLVM_ABI bool hasOneUser() const
Return true if there is exactly one user of this value.
LLVMContext & getContext() const
All values hold a context through their type.
LLVM_ABI void takeName(Value *V)
Transfer the name from V to this value.
const ParentTy * getParent() const
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
LLVM_READONLY const MIMGOffsetMappingInfo * getMIMGOffsetMappingInfo(unsigned Offset)
uint8_t wmmaScaleF8F6F4FormatToNumRegs(unsigned Fmt)
const ImageDimIntrinsicInfo * getImageDimIntrinsicByBaseOpcode(unsigned BaseOpcode, unsigned Dim)
LLVM_READONLY const MIMGMIPMappingInfo * getMIMGMIPMappingInfo(unsigned MIP)
bool isArgPassedInSGPR(const Argument *A)
bool isIntrinsicAlwaysUniform(unsigned IntrID)
LLVM_READONLY const MIMGBiasMappingInfo * getMIMGBiasMappingInfo(unsigned Bias)
LLVM_READONLY const MIMGLZMappingInfo * getMIMGLZMappingInfo(unsigned L)
LLVM_READONLY const MIMGBaseOpcodeInfo * getMIMGBaseOpcodeInfo(unsigned BaseOpcode)
const ImageDimIntrinsicInfo * getImageDimIntrinsicInfo(unsigned Intr)
@ C
The default llvm calling convention, compatible with C.
LLVM_ABI Function * getOrInsertDeclaration(Module *M, ID id, ArrayRef< Type * > OverloadTys={})
Look up the Function declaration of the intrinsic id in the Module M.
LLVM_ABI bool isSignatureValid(Intrinsic::ID ID, FunctionType *FT, SmallVectorImpl< Type * > &OverloadTys, raw_ostream &OS=nulls())
Returns true if FT is a valid function type for intrinsic ID.
OneUse_match< SubPat > m_OneUse(const SubPat &SP)
cst_pred_ty< is_all_ones > m_AllOnes()
Match an integer or vector with all bits set.
auto m_Cmp()
Matches any compare instruction and ignore it.
bool match(Val *V, const Pattern &P)
cstfp_pred_ty< is_any_zero_fp > m_AnyZeroFP()
Match a floating-point negative zero or positive zero.
ap_match< APFloat > m_APFloat(const APFloat *&Res)
Match a ConstantFP or splatted ConstantVector, binding the specified pointer to the contained APFloat...
cst_pred_ty< is_one > m_One()
Match an integer 1 or a vector with all elements equal to 1.
IntrinsicID_match m_Intrinsic()
Match intrinsic calls like this: m_Intrinsic<Intrinsic::fabs>(m_Value(X))
auto m_Value()
Match an arbitrary value and ignore it.
CastInst_match< OpTy, FPExtInst > m_FPExt(const OpTy &Op)
CastInst_match< OpTy, ZExtInst > m_ZExt(const OpTy &Op)
Matches ZExt.
cstfp_pred_ty< is_finitenonzero > m_FiniteNonZero()
Match a finite non-zero FP constant.
match_combine_or< CastInst_match< OpTy, ZExtInst >, CastInst_match< OpTy, SExtInst > > m_ZExtOrSExt(const OpTy &Op)
auto m_ConstantFP()
Match an arbitrary ConstantFP and ignore it.
CastInst_match< OpTy, SExtInst > m_SExt(const OpTy &Op)
Matches SExt.
is_zero m_Zero()
Match any null constant or a vector with all elements equal to 0.
auto m_ConstantInt()
Match an arbitrary ConstantInt and ignore it.
This is an optimization pass for GlobalISel generic memory operations.
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
LLVM_ABI Constant * ConstantFoldCompareInstOperands(unsigned Predicate, Constant *LHS, Constant *RHS, const DataLayout &DL, const TargetLibraryInfo *TLI=nullptr, const Instruction *I=nullptr)
Attempt to constant fold a compare instruction (icmp/fcmp) with the specified operands.
constexpr int popcount(T Value) noexcept
Count the number of set bits in a value.
APFloat frexp(const APFloat &X, int &Exp, APFloat::roundingMode RM)
Equivalent of C standard library function.
auto dyn_cast_or_null(const Y &Val)
LLVM_READONLY APFloat maxnum(const APFloat &A, const APFloat &B)
Implements IEEE-754 2008 maxNum semantics.
constexpr unsigned MaxAnalysisRecursionDepth
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
APFloat scalbn(APFloat X, int Exp, APFloat::roundingMode RM)
Returns: X * 2^Exp for integral exponents.
constexpr uint32_t Hi_32(uint64_t Value)
Return the high 32 bits of a 64 bit value.
constexpr uint32_t Lo_32(uint64_t Value)
Return the low 32 bits of a 64 bit value.
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
constexpr int PoisonMaskElem
LLVM_ABI Value * findScalarElement(Value *V, unsigned EltNo)
Given a vector and an element number, see if the scalar value is already around as a register,...
@ NearestTiesToEven
roundTiesToEven.
LLVM_ABI bool isKnownNeverInfOrNaN(const Value *V, const SimplifyQuery &SQ, unsigned Depth=0)
Return true if the floating-point value can never contain a NaN or infinity.
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
bool all_equal(std::initializer_list< T > Values)
Returns true if all Values in the initializer lists are equal or the list.
auto seq(T Begin, T End)
Iterate over an integral type from Begin up to - but not including - End.
constexpr T maskTrailingOnes(unsigned N)
Create a bitmask with the N right-most bits set to 1, and all other bits set to 0.
LLVM_ABI Constant * ConstantFoldInstOperands(const Instruction *I, ArrayRef< Constant * > Ops, const DataLayout &DL, const TargetLibraryInfo *TLI=nullptr, bool AllowNonDeterministic=true)
ConstantFoldInstOperands - Attempt to constant fold an instruction with the specified operands.
constexpr uint64_t Make_64(uint32_t High, uint32_t Low)
Make a 64-bit integer from a high / low pair of 32-bit integers.
LLVM_ABI ConstantRange computeConstantRange(const Value *V, bool ForSigned, const SimplifyQuery &SQ, unsigned Depth=0)
Determine the possible constant range of an integer or vector of integer value.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
bool isConstant() const
Returns true if we know the value of all bits.
const APInt & getConstant() const
Returns the value when all bits have a known value.
SimplifyQuery getWithInstruction(const Instruction *I) const
LLVM_ABI bool isUndefValue(Value *V) const
If CanUseUndef is true, returns whether V is undef.