LLVM 23.0.0git
AArch64CallLowering.cpp
Go to the documentation of this file.
1//===--- AArch64CallLowering.cpp - Call lowering --------------------------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8///
9/// \file
10/// This file implements the lowering of LLVM calls to machine code calls for
11/// GlobalISel.
12///
13//===----------------------------------------------------------------------===//
14
15#include "AArch64CallLowering.h"
17#include "AArch64ISelLowering.h"
19#include "AArch64RegisterInfo.h"
21#include "AArch64Subtarget.h"
23#include "llvm/ADT/ArrayRef.h"
45#include "llvm/IR/Argument.h"
46#include "llvm/IR/Attributes.h"
47#include "llvm/IR/Function.h"
48#include "llvm/IR/Type.h"
49#include "llvm/IR/Value.h"
50#include <algorithm>
51#include <cassert>
52#include <cstdint>
53
54#define DEBUG_TYPE "aarch64-call-lowering"
55
56using namespace llvm;
57using namespace AArch64GISelUtils;
58
60
63
64static void applyStackPassedSmallTypeDAGHack(EVT OrigVT, MVT &ValVT,
65 MVT &LocVT) {
66 // If ValVT is i1/i8/i16, we should set LocVT to i8/i8/i16. This is a legacy
67 // hack because the DAG calls the assignment function with pre-legalized
68 // register typed values, not the raw type.
69 //
70 // This hack is not applied to return values which are not passed on the
71 // stack.
72 if (OrigVT == MVT::i1 || OrigVT == MVT::i8)
73 ValVT = LocVT = MVT::i8;
74 else if (OrigVT == MVT::i16)
75 ValVT = LocVT = MVT::i16;
76}
77
78// Account for i1/i8/i16 stack passed value hack
80 const MVT ValVT = VA.getValVT();
81 return (ValVT == MVT::i8 || ValVT == MVT::i16) ? LLT(ValVT)
82 : LLT(VA.getLocVT());
83}
84
85namespace {
86
87struct AArch64IncomingValueAssigner
89 AArch64IncomingValueAssigner(CCAssignFn *AssignFn_,
90 CCAssignFn *AssignFnVarArg_)
91 : IncomingValueAssigner(AssignFn_, AssignFnVarArg_) {}
92
93 bool assignArg(unsigned ValNo, EVT OrigVT, MVT ValVT, MVT LocVT,
95 const CallLowering::ArgInfo &Info, ISD::ArgFlagsTy Flags,
96 CCState &State) override {
97 applyStackPassedSmallTypeDAGHack(OrigVT, ValVT, LocVT);
98 return IncomingValueAssigner::assignArg(ValNo, OrigVT, ValVT, LocVT,
99 LocInfo, Info, Flags, State);
100 }
101};
102
103struct AArch64OutgoingValueAssigner
105 const AArch64Subtarget &Subtarget;
106
107 /// Track if this is used for a return instead of function argument
108 /// passing. We apply a hack to i1/i8/i16 stack passed values, but do not use
109 /// stack passed returns for them and cannot apply the type adjustment.
110 bool IsReturn;
111
112 AArch64OutgoingValueAssigner(CCAssignFn *AssignFn_,
113 CCAssignFn *AssignFnVarArg_,
114 const AArch64Subtarget &Subtarget_,
115 bool IsReturn)
116 : OutgoingValueAssigner(AssignFn_, AssignFnVarArg_),
117 Subtarget(Subtarget_), IsReturn(IsReturn) {}
118
119 bool assignArg(unsigned ValNo, EVT OrigVT, MVT ValVT, MVT LocVT,
120 CCValAssign::LocInfo LocInfo,
121 const CallLowering::ArgInfo &Info, ISD::ArgFlagsTy Flags,
122 CCState &State) override {
123 const Function &F = State.getMachineFunction().getFunction();
124 bool IsCalleeWin =
125 Subtarget.isCallingConvWin64(State.getCallingConv(), F.isVarArg());
126 bool UseVarArgsCCForFixed = IsCalleeWin && State.isVarArg();
127
128 bool Res;
129 if (!Flags.isVarArg() && !UseVarArgsCCForFixed) {
130 if (!IsReturn)
131 applyStackPassedSmallTypeDAGHack(OrigVT, ValVT, LocVT);
132 Res = AssignFn(ValNo, ValVT, LocVT, LocInfo, Flags, Info.Ty, State);
133 } else
134 Res = AssignFnVarArg(ValNo, ValVT, LocVT, LocInfo, Flags, Info.Ty, State);
135
136 StackSize = State.getStackSize();
137 return Res;
138 }
139};
140
141struct IncomingArgHandler : public CallLowering::IncomingValueHandler {
142 IncomingArgHandler(MachineIRBuilder &MIRBuilder, MachineRegisterInfo &MRI)
143 : IncomingValueHandler(MIRBuilder, MRI) {}
144
145 Register getStackAddress(uint64_t Size, int64_t Offset,
146 MachinePointerInfo &MPO,
147 ISD::ArgFlagsTy Flags) override {
148 auto &MFI = MIRBuilder.getMF().getFrameInfo();
149
150 // Byval is assumed to be writable memory, but other stack passed arguments
151 // are not.
152 const bool IsImmutable = !Flags.isByVal();
153
154 int FI = MFI.CreateFixedObject(Size, Offset, IsImmutable);
155 MPO = MachinePointerInfo::getFixedStack(MIRBuilder.getMF(), FI);
156 auto AddrReg = MIRBuilder.buildFrameIndex(LLT::pointer(0, 64), FI);
157 return AddrReg.getReg(0);
158 }
159
160 LLT getStackValueStoreType(const DataLayout &DL, const CCValAssign &VA,
161 ISD::ArgFlagsTy Flags) const override {
162 // For pointers, we just need to fixup the integer types reported in the
163 // CCValAssign.
164 if (Flags.isPointer())
167 }
168
169 void assignValueToReg(Register ValVReg, Register PhysReg,
170 const CCValAssign &VA,
171 ISD::ArgFlagsTy Flags = {}) override {
172 markRegUsed(PhysReg);
173 IncomingValueHandler::assignValueToReg(ValVReg, PhysReg, VA);
174 }
175
176 void assignValueToAddress(Register ValVReg, Register Addr, LLT MemTy,
177 const MachinePointerInfo &MPO,
178 const CCValAssign &VA) override {
179 MachineFunction &MF = MIRBuilder.getMF();
180
181 LLT ValTy(VA.getValVT());
182 LLT LocTy(VA.getLocVT());
183
184 // Fixup the types for the DAG compatibility hack.
185 if (VA.getValVT() == MVT::i8 || VA.getValVT() == MVT::i16)
186 std::swap(ValTy, LocTy);
187 else {
188 // The calling code knows if this is a pointer or not, we're only touching
189 // the LocTy for the i8/i16 hack.
190 assert(LocTy.getSizeInBits() == MemTy.getSizeInBits());
191 LocTy = MemTy;
192 }
193
194 auto MMO = MF.getMachineMemOperand(
196 inferAlignFromPtrInfo(MF, MPO));
197
198 switch (VA.getLocInfo()) {
199 case CCValAssign::LocInfo::ZExt:
200 MIRBuilder.buildLoadInstr(TargetOpcode::G_ZEXTLOAD, ValVReg, Addr, *MMO);
201 return;
202 case CCValAssign::LocInfo::SExt:
203 MIRBuilder.buildLoadInstr(TargetOpcode::G_SEXTLOAD, ValVReg, Addr, *MMO);
204 return;
205 default:
206 MIRBuilder.buildLoad(ValVReg, Addr, *MMO);
207 return;
208 }
209 }
210
211 /// How the physical register gets marked varies between formal
212 /// parameters (it's a basic-block live-in), and a call instruction
213 /// (it's an implicit-def of the BL).
214 virtual void markRegUsed(Register Reg) = 0;
215};
216
217struct FormalArgHandler : public IncomingArgHandler {
218 FormalArgHandler(MachineIRBuilder &MIRBuilder, MachineRegisterInfo &MRI)
219 : IncomingArgHandler(MIRBuilder, MRI) {}
220
221 void markRegUsed(Register Reg) override {
222 MIRBuilder.getMRI()->addLiveIn(Reg.asMCReg());
223 MIRBuilder.getMBB().addLiveIn(Reg.asMCReg());
224 }
225};
226
227struct CallReturnHandler : public IncomingArgHandler {
228 CallReturnHandler(MachineIRBuilder &MIRBuilder, MachineRegisterInfo &MRI,
229 MachineInstrBuilder MIB)
230 : IncomingArgHandler(MIRBuilder, MRI), MIB(MIB) {}
231
232 void markRegUsed(Register Reg) override {
233 MIB.addDef(Reg, RegState::Implicit);
234 }
235
236 MachineInstrBuilder MIB;
237};
238
239/// A special return arg handler for "returned" attribute arg calls.
240struct ReturnedArgCallReturnHandler : public CallReturnHandler {
241 ReturnedArgCallReturnHandler(MachineIRBuilder &MIRBuilder,
242 MachineRegisterInfo &MRI,
243 MachineInstrBuilder MIB)
244 : CallReturnHandler(MIRBuilder, MRI, MIB) {}
245
246 void markRegUsed(Register Reg) override {}
247};
248
249struct OutgoingArgHandler : public CallLowering::OutgoingValueHandler {
250 OutgoingArgHandler(MachineIRBuilder &MIRBuilder, MachineRegisterInfo &MRI,
251 MachineInstrBuilder MIB, bool IsTailCall = false,
252 int FPDiff = 0)
253 : OutgoingValueHandler(MIRBuilder, MRI), MIB(MIB), IsTailCall(IsTailCall),
254 FPDiff(FPDiff),
255 Subtarget(MIRBuilder.getMF().getSubtarget<AArch64Subtarget>()) {}
256
257 Register getStackAddress(uint64_t Size, int64_t Offset,
258 MachinePointerInfo &MPO,
259 ISD::ArgFlagsTy Flags) override {
260 MachineFunction &MF = MIRBuilder.getMF();
261 LLT p0 = LLT::pointer(0, 64);
262 LLT s64 = LLT::integer(64);
263
264 if (IsTailCall) {
265 assert(!Flags.isByVal() && "byval unhandled with tail calls");
266
267 Offset += FPDiff;
268 int FI = MF.getFrameInfo().CreateFixedObject(Size, Offset, true);
269 auto FIReg = MIRBuilder.buildFrameIndex(p0, FI);
271 return FIReg.getReg(0);
272 }
273
274 if (!SPReg)
275 SPReg = MIRBuilder.buildCopy(p0, Register(AArch64::SP)).getReg(0);
276
277 auto OffsetReg = MIRBuilder.buildConstant(s64, Offset);
278
279 auto AddrReg = MIRBuilder.buildPtrAdd(p0, SPReg, OffsetReg);
280
282 return AddrReg.getReg(0);
283 }
284
285 /// We need to fixup the reported store size for certain value types because
286 /// we invert the interpretation of ValVT and LocVT in certain cases. This is
287 /// for compatibility with the DAG call lowering implementation, which we're
288 /// currently building on top of.
289 LLT getStackValueStoreType(const DataLayout &DL, const CCValAssign &VA,
290 ISD::ArgFlagsTy Flags) const override {
291 if (Flags.isPointer())
294 }
295
296 void assignValueToReg(Register ValVReg, Register PhysReg,
297 const CCValAssign &VA, ISD::ArgFlagsTy Flags) override {
298 MIB.addUse(PhysReg, RegState::Implicit);
299 Register ExtReg = extendRegister(ValVReg, VA);
300 MIRBuilder.buildCopy(PhysReg, ExtReg);
301 }
302
303 /// Check whether a stack argument requires lowering in a tail call.
304 static bool shouldLowerTailCallStackArg(const MachineFunction &MF,
305 const CCValAssign &VA,
306 Register ValVReg,
307 Register StoreAddr) {
308 const MachineRegisterInfo &MRI = MF.getRegInfo();
309 // Print the defining instruction for the value.
310 auto *DefMI = MRI.getVRegDef(ValVReg);
311 assert(DefMI && "No defining instruction");
312 for (;;) {
313 // Look through nodes that don't alter the bits of the incoming value.
314 unsigned Op = DefMI->getOpcode();
315 if (Op == TargetOpcode::G_ZEXT || Op == TargetOpcode::G_ANYEXT ||
316 Op == TargetOpcode::G_BITCAST || isAssertMI(*DefMI)) {
318 continue;
319 }
320 break;
321 }
322
323 auto *Load = dyn_cast<GLoad>(DefMI);
324 if (!Load)
325 return true;
326 Register LoadReg = Load->getPointerReg();
327 auto *LoadAddrDef = MRI.getVRegDef(LoadReg);
328 if (LoadAddrDef->getOpcode() != TargetOpcode::G_FRAME_INDEX)
329 return true;
330 const MachineFrameInfo &MFI = MF.getFrameInfo();
331 int LoadFI = LoadAddrDef->getOperand(1).getIndex();
332
333 auto *StoreAddrDef = MRI.getVRegDef(StoreAddr);
334 if (StoreAddrDef->getOpcode() != TargetOpcode::G_FRAME_INDEX)
335 return true;
336 int StoreFI = StoreAddrDef->getOperand(1).getIndex();
337
338 if (!MFI.isImmutableObjectIndex(LoadFI))
339 return true;
340 if (MFI.getObjectOffset(LoadFI) != MFI.getObjectOffset(StoreFI))
341 return true;
342 if (Load->getMemSize() != MFI.getObjectSize(StoreFI))
343 return true;
344
345 return false;
346 }
347
348 void assignValueToAddress(Register ValVReg, Register Addr, LLT MemTy,
349 const MachinePointerInfo &MPO,
350 const CCValAssign &VA) override {
351 MachineFunction &MF = MIRBuilder.getMF();
352 if (!FPDiff && !shouldLowerTailCallStackArg(MF, VA, ValVReg, Addr))
353 return;
354 auto MMO = MF.getMachineMemOperand(MPO, MachineMemOperand::MOStore, MemTy,
355 inferAlignFromPtrInfo(MF, MPO));
356 MIRBuilder.buildStore(ValVReg, Addr, *MMO);
357 }
358
359 void assignValueToAddress(const CallLowering::ArgInfo &Arg, unsigned RegIndex,
360 Register Addr, LLT MemTy,
361 const MachinePointerInfo &MPO,
362 const CCValAssign &VA) override {
363 unsigned MaxSize = MemTy.getSizeInBytes() * 8;
364 // For varargs, we always want to extend them to 8 bytes, in which case
365 // we disable setting a max.
366 if (Arg.Flags[0].isVarArg())
367 MaxSize = 0;
368
369 Register ValVReg = Arg.Regs[RegIndex];
370 if (VA.getLocInfo() != CCValAssign::LocInfo::FPExt) {
371 MVT LocVT = VA.getLocVT();
372 MVT ValVT = VA.getValVT();
373
374 if (VA.getValVT() == MVT::i8 || VA.getValVT() == MVT::i16) {
375 std::swap(ValVT, LocVT);
376 MemTy = LLT(VA.getValVT());
377 }
378
379 ValVReg = extendRegister(ValVReg, VA, MaxSize);
380 } else {
381 // The store does not cover the full allocated stack slot.
382 MemTy = LLT(VA.getValVT());
383 }
384
385 assignValueToAddress(ValVReg, Addr, MemTy, MPO, VA);
386 }
387
388 MachineInstrBuilder MIB;
389
390 bool IsTailCall;
391
392 /// For tail calls, the byte offset of the call's argument area from the
393 /// callee's. Unused elsewhere.
394 int FPDiff;
395
396 // Cache the SP register vreg if we need it more than once in this call site.
398
399 const AArch64Subtarget &Subtarget;
400};
401} // namespace
402
403static bool doesCalleeRestoreStack(CallingConv::ID CallConv, bool TailCallOpt) {
404 return (CallConv == CallingConv::Fast && TailCallOpt) ||
405 CallConv == CallingConv::Tail || CallConv == CallingConv::SwiftTail;
406}
407
409 const Value *Val,
410 ArrayRef<Register> VRegs,
412 Register SwiftErrorVReg) const {
413 auto MIB = MIRBuilder.buildInstrNoInsert(AArch64::RET_ReallyLR);
414 assert(((Val && !VRegs.empty()) || (!Val && VRegs.empty())) &&
415 "Return value without a vreg");
416
417 bool Success = true;
418 if (!FLI.CanLowerReturn) {
419 insertSRetStores(MIRBuilder, Val->getType(), VRegs, FLI.DemoteRegister);
420 } else if (!VRegs.empty()) {
421 MachineFunction &MF = MIRBuilder.getMF();
422 const Function &F = MF.getFunction();
423 const AArch64Subtarget &Subtarget = MF.getSubtarget<AArch64Subtarget>();
424
427 CCAssignFn *AssignFn = TLI.CCAssignFnForReturn(F.getCallingConv());
428 auto &DL = F.getDataLayout();
429 LLVMContext &Ctx = Val->getType()->getContext();
430
431 SmallVector<EVT, 4> SplitEVTs;
432 ComputeValueVTs(TLI, DL, Val->getType(), SplitEVTs);
433 assert(VRegs.size() == SplitEVTs.size() &&
434 "For each split Type there should be exactly one VReg.");
435
436 SmallVector<ArgInfo, 8> SplitArgs;
437 CallingConv::ID CC = F.getCallingConv();
438
439 for (unsigned i = 0; i < SplitEVTs.size(); ++i) {
440 Register CurVReg = VRegs[i];
441 ArgInfo CurArgInfo = ArgInfo{CurVReg, SplitEVTs[i].getTypeForEVT(Ctx), 0};
442 setArgFlags(CurArgInfo, AttributeList::ReturnIndex, DL, F);
443
444 // i1 is a special case because SDAG i1 true is naturally zero extended
445 // when widened using ANYEXT. We need to do it explicitly here.
446 auto &Flags = CurArgInfo.Flags[0];
447 if (MRI.getType(CurVReg).getSizeInBits() == TypeSize::getFixed(1) &&
448 !Flags.isSExt() && !Flags.isZExt()) {
449 CurVReg = MIRBuilder.buildZExt(LLT::integer(8), CurVReg).getReg(0);
450 } else if (TLI.getNumRegistersForCallingConv(Ctx, CC, SplitEVTs[i]) ==
451 1) {
452 // Some types will need extending as specified by the CC.
453 MVT NewVT = TLI.getRegisterTypeForCallingConv(Ctx, CC, SplitEVTs[i]);
454 if (EVT(NewVT) != SplitEVTs[i]) {
455 unsigned ExtendOp = TargetOpcode::G_ANYEXT;
456 if (F.getAttributes().hasRetAttr(Attribute::SExt))
457 ExtendOp = TargetOpcode::G_SEXT;
458 else if (F.getAttributes().hasRetAttr(Attribute::ZExt))
459 ExtendOp = TargetOpcode::G_ZEXT;
460
461 LLT NewLLT(NewVT);
462 LLT OldLLT = getLLTForType(*CurArgInfo.Ty, DL);
463 CurArgInfo.Ty = EVT(NewVT).getTypeForEVT(Ctx);
464 // Instead of an extend, we might have a vector type which needs
465 // padding with more elements, e.g. <2 x half> -> <4 x half>.
466 if (NewVT.isVector()) {
467 if (OldLLT.isVector()) {
468 if (NewLLT.getNumElements() > OldLLT.getNumElements()) {
469 CurVReg =
470 MIRBuilder.buildPadVectorWithUndefElements(NewLLT, CurVReg)
471 .getReg(0);
472 } else {
473 // Just do a vector extend.
474 CurVReg = MIRBuilder.buildInstr(ExtendOp, {NewLLT}, {CurVReg})
475 .getReg(0);
476 }
477 } else if (NewLLT.getNumElements() >= 2 &&
478 NewLLT.getNumElements() <= 8) {
479 // We need to pad a <1 x S> type to <2/4/8 x S>. Since we don't
480 // have <1 x S> vector types in GISel we use a build_vector
481 // instead of a vector merge/concat.
482 CurVReg =
483 MIRBuilder.buildPadVectorWithUndefElements(NewLLT, CurVReg)
484 .getReg(0);
485 } else {
486 LLVM_DEBUG(dbgs() << "Could not handle ret ty\n");
487 return false;
488 }
489 } else {
490 // If the split EVT was a <1 x T> vector, and NewVT is T, then we
491 // don't have to do anything since we don't distinguish between the
492 // two.
493 if (NewLLT.getScalarSizeInBits() !=
494 MRI.getType(CurVReg).getScalarSizeInBits()) {
495 // A scalar extend.
496 CurVReg = MIRBuilder.buildInstr(ExtendOp, {NewLLT}, {CurVReg})
497 .getReg(0);
498 }
499 }
500 }
501 }
502 if (CurVReg != CurArgInfo.Regs[0]) {
503 CurArgInfo.Regs[0] = CurVReg;
504 // Reset the arg flags after modifying CurVReg.
505 setArgFlags(CurArgInfo, AttributeList::ReturnIndex, DL, F);
506 }
507 splitToValueTypes(CurArgInfo, SplitArgs, DL, CC);
508 }
509
510 AArch64OutgoingValueAssigner Assigner(AssignFn, AssignFn, Subtarget,
511 /*IsReturn*/ true);
512 OutgoingArgHandler Handler(MIRBuilder, MRI, MIB);
513 Success = determineAndHandleAssignments(Handler, Assigner, SplitArgs,
514 MIRBuilder, CC, F.isVarArg());
515 }
516
517 if (SwiftErrorVReg) {
518 MIB.addUse(AArch64::X21, RegState::Implicit);
519 MIRBuilder.buildCopy(AArch64::X21, SwiftErrorVReg);
520 }
521
522 MIRBuilder.insertInstr(MIB);
523 return Success;
524}
525
527 CallingConv::ID CallConv,
529 bool IsVarArg) const {
531 const auto &TLI = *getTLI<AArch64TargetLowering>();
532 CCState CCInfo(CallConv, IsVarArg, MF, ArgLocs,
533 MF.getFunction().getContext());
534
535 return checkReturn(CCInfo, Outs, TLI.CCAssignFnForReturn(CallConv));
536}
537
538/// Helper function to compute forwarded registers for musttail calls. Computes
539/// the forwarded registers, sets MBB liveness, and emits COPY instructions that
540/// can be used to save + restore registers later.
542 CCAssignFn *AssignFn) {
543 MachineBasicBlock &MBB = MIRBuilder.getMBB();
544 MachineFunction &MF = MIRBuilder.getMF();
545 MachineFrameInfo &MFI = MF.getFrameInfo();
546
547 if (!MFI.hasMustTailInVarArgFunc())
548 return;
549
551 const Function &F = MF.getFunction();
552 assert(F.isVarArg() && "Expected F to be vararg?");
553
554 // Compute the set of forwarded registers. The rest are scratch.
556 CCState CCInfo(F.getCallingConv(), /*IsVarArg=*/true, MF, ArgLocs,
557 F.getContext());
558 SmallVector<MVT, 2> RegParmTypes;
559 RegParmTypes.push_back(MVT::i64);
560 RegParmTypes.push_back(MVT::f128);
561
562 // Later on, we can use this vector to restore the registers if necessary.
565 CCInfo.analyzeMustTailForwardedRegisters(Forwards, RegParmTypes, AssignFn);
566
567 // Conservatively forward X8, since it might be used for an aggregate
568 // return.
569 if (!CCInfo.isAllocated(AArch64::X8)) {
570 Register X8VReg = MF.addLiveIn(AArch64::X8, &AArch64::GPR64RegClass);
571 Forwards.push_back(ForwardedRegister(X8VReg, AArch64::X8, MVT::i64));
572 }
573
574 // Add the forwards to the MachineBasicBlock and MachineFunction.
575 for (const auto &F : Forwards) {
576 MBB.addLiveIn(F.PReg);
577 MIRBuilder.buildCopy(Register(F.VReg), Register(F.PReg));
578 }
579}
580
582 auto &F = MF.getFunction();
583 const auto &TM = static_cast<const AArch64TargetMachine &>(MF.getTarget());
584
585 if (!EnableSVEGISel && (F.getReturnType()->isScalableTy() ||
586 llvm::any_of(F.args(), [](const Argument &A) {
587 return A.getType()->isScalableTy();
588 })))
589 return true;
590 const auto &ST = MF.getSubtarget<AArch64Subtarget>();
591 if (!ST.hasNEON() || !ST.hasFPARMv8()) {
592 LLVM_DEBUG(dbgs() << "Falling back to SDAG because we don't support no-NEON\n");
593 return true;
594 }
595
596 SMEAttrs Attrs = MF.getInfo<AArch64FunctionInfo>()->getSMEFnAttrs();
597 if (Attrs.hasZAState() || Attrs.hasZT0State() ||
598 Attrs.hasStreamingInterfaceOrBody() ||
599 Attrs.hasStreamingCompatibleInterface())
600 return true;
601
602 auto OptLevel = MF.getTarget().getOptLevel();
603 bool IsGlobalISelPreferred =
605 static_cast<unsigned>(OptLevel) <= TM.getEnableGlobalISelAtO() ||
606 F.hasOptNone();
607 return !IsGlobalISelPreferred;
608}
609
610void AArch64CallLowering::saveVarArgRegisters(
612 CCState &CCInfo) const {
615
616 MachineFunction &MF = MIRBuilder.getMF();
618 MachineFrameInfo &MFI = MF.getFrameInfo();
620 auto &Subtarget = MF.getSubtarget<AArch64Subtarget>();
621 bool IsWin64CC = Subtarget.isCallingConvWin64(CCInfo.getCallingConv(),
622 MF.getFunction().isVarArg());
623 const LLT p0 = LLT::pointer(0, 64);
624 const LLT s64 = LLT::integer(64);
625
626 unsigned FirstVariadicGPR = CCInfo.getFirstUnallocated(GPRArgRegs);
627 unsigned NumVariadicGPRArgRegs = GPRArgRegs.size() - FirstVariadicGPR + 1;
628
629 unsigned GPRSaveSize = 8 * (GPRArgRegs.size() - FirstVariadicGPR);
630 int GPRIdx = 0;
631 if (GPRSaveSize != 0) {
632 if (IsWin64CC) {
633 GPRIdx = MFI.CreateFixedObject(GPRSaveSize,
634 -static_cast<int>(GPRSaveSize), false);
635 if (GPRSaveSize & 15)
636 // The extra size here, if triggered, will always be 8.
637 MFI.CreateFixedObject(16 - (GPRSaveSize & 15),
638 -static_cast<int>(alignTo(GPRSaveSize, 16)),
639 false);
640 } else
641 GPRIdx = MFI.CreateStackObject(GPRSaveSize, Align(8), false);
642
643 auto FIN = MIRBuilder.buildFrameIndex(p0, GPRIdx);
644 auto Offset =
645 MIRBuilder.buildConstant(MRI.createGenericVirtualRegister(s64), 8);
646
647 for (unsigned i = FirstVariadicGPR; i < GPRArgRegs.size(); ++i) {
649 Handler.assignValueToReg(
650 Val, GPRArgRegs[i],
652 GPRArgRegs[i], MVT::i64, CCValAssign::Full));
653 auto MPO = IsWin64CC ? MachinePointerInfo::getFixedStack(
654 MF, GPRIdx, (i - FirstVariadicGPR) * 8)
655 : MachinePointerInfo::getStack(MF, i * 8);
656 MIRBuilder.buildStore(Val, FIN, MPO, inferAlignFromPtrInfo(MF, MPO));
657
658 FIN = MIRBuilder.buildPtrAdd(MRI.createGenericVirtualRegister(p0),
659 FIN.getReg(0), Offset);
660 }
661 }
662 FuncInfo->setVarArgsGPRIndex(GPRIdx);
663 FuncInfo->setVarArgsGPRSize(GPRSaveSize);
664
665 if (Subtarget.hasFPARMv8() && !IsWin64CC) {
666 unsigned FirstVariadicFPR = CCInfo.getFirstUnallocated(FPRArgRegs);
667
668 unsigned FPRSaveSize = 16 * (FPRArgRegs.size() - FirstVariadicFPR);
669 int FPRIdx = 0;
670 if (FPRSaveSize != 0) {
671 FPRIdx = MFI.CreateStackObject(FPRSaveSize, Align(16), false);
672
673 auto FIN = MIRBuilder.buildFrameIndex(p0, FPRIdx);
674 auto Offset =
675 MIRBuilder.buildConstant(MRI.createGenericVirtualRegister(s64), 16);
676
677 for (unsigned i = FirstVariadicFPR; i < FPRArgRegs.size(); ++i) {
679 Handler.assignValueToReg(
680 Val, FPRArgRegs[i],
682 i + MF.getFunction().getNumOperands() + NumVariadicGPRArgRegs,
683 MVT::f128, FPRArgRegs[i], MVT::f128, CCValAssign::Full));
684
685 auto MPO = MachinePointerInfo::getStack(MF, i * 16);
686 MIRBuilder.buildStore(Val, FIN, MPO, inferAlignFromPtrInfo(MF, MPO));
687
688 FIN = MIRBuilder.buildPtrAdd(MRI.createGenericVirtualRegister(p0),
689 FIN.getReg(0), Offset);
690 }
691 }
692 FuncInfo->setVarArgsFPRIndex(FPRIdx);
693 FuncInfo->setVarArgsFPRSize(FPRSaveSize);
694 }
695}
696
698 MachineIRBuilder &MIRBuilder, const Function &F,
700 MachineFunction &MF = MIRBuilder.getMF();
701 MachineBasicBlock &MBB = MIRBuilder.getMBB();
703 auto &DL = F.getDataLayout();
704 auto &Subtarget = MF.getSubtarget<AArch64Subtarget>();
705
706 // Arm64EC has extra requirements for varargs calls which are only implemented
707 // in SelectionDAG; bail out for now.
708 if (F.isVarArg() && Subtarget.isWindowsArm64EC())
709 return false;
710
711 // Arm64EC thunks have a special calling convention which is only implemented
712 // in SelectionDAG; bail out for now.
713 if (F.getCallingConv() == CallingConv::ARM64EC_Thunk_Native ||
714 F.getCallingConv() == CallingConv::ARM64EC_Thunk_X64)
715 return false;
716
717 bool IsWin64 =
718 Subtarget.isCallingConvWin64(F.getCallingConv(), F.isVarArg()) &&
719 !Subtarget.isWindowsArm64EC();
720
721 SmallVector<ArgInfo, 8> SplitArgs;
723
724 // Insert the hidden sret parameter if the return value won't fit in the
725 // return registers.
726 if (!FLI.CanLowerReturn)
727 insertSRetIncomingArgument(F, SplitArgs, FLI.DemoteRegister, MRI, DL);
728
729 unsigned i = 0;
730 for (auto &Arg : F.args()) {
731 if (DL.getTypeStoreSize(Arg.getType()).isZero())
732 continue;
733
734 ArgInfo OrigArg{VRegs[i], Arg, i};
735 setArgFlags(OrigArg, i + AttributeList::FirstArgIndex, DL, F);
736
737 // i1 arguments are zero-extended to i8 by the caller. Emit a
738 // hint to reflect this.
739 if (OrigArg.Ty->isIntegerTy(1)) {
740 assert(OrigArg.Regs.size() == 1 &&
741 MRI.getType(OrigArg.Regs[0]).getSizeInBits() == 1 &&
742 "Unexpected registers used for i1 arg");
743
744 auto &Flags = OrigArg.Flags[0];
745 if (!Flags.isZExt() && !Flags.isSExt()) {
746 // Lower i1 argument as i8, and insert AssertZExt + Trunc later.
747 Register OrigReg = OrigArg.Regs[0];
749 OrigArg.Regs[0] = WideReg;
750 BoolArgs.push_back({OrigReg, WideReg});
751 }
752 }
753
754 if (Arg.hasAttribute(Attribute::SwiftAsync))
755 MF.getInfo<AArch64FunctionInfo>()->setHasSwiftAsyncContext(true);
756
757 splitToValueTypes(OrigArg, SplitArgs, DL, F.getCallingConv());
758 ++i;
759 }
760
761 if (!MBB.empty())
762 MIRBuilder.setInstr(*MBB.begin());
763
765 CCAssignFn *AssignFn = TLI.CCAssignFnForCall(F.getCallingConv(), IsWin64 && F.isVarArg());
766
767 AArch64IncomingValueAssigner Assigner(AssignFn, AssignFn);
768 FormalArgHandler Handler(MIRBuilder, MRI);
770 CCState CCInfo(F.getCallingConv(), F.isVarArg(), MF, ArgLocs, F.getContext());
771 if (!determineAssignments(Assigner, SplitArgs, CCInfo) ||
772 !handleAssignments(Handler, SplitArgs, CCInfo, ArgLocs, MIRBuilder))
773 return false;
774
775 if (!BoolArgs.empty()) {
776 for (auto &KV : BoolArgs) {
777 Register OrigReg = KV.first;
778 Register WideReg = KV.second;
779 LLT WideTy = MRI.getType(WideReg);
780 assert(MRI.getType(OrigReg).getScalarSizeInBits() == 1 &&
781 "Unexpected bit size of a bool arg");
782 MIRBuilder.buildTrunc(
783 OrigReg, MIRBuilder.buildAssertZExt(WideTy, WideReg, 1).getReg(0));
784 }
785 }
786
788 uint64_t StackSize = Assigner.StackSize;
789 if (F.isVarArg()) {
790 if ((!Subtarget.isTargetDarwin() && !Subtarget.isWindowsArm64EC()) || IsWin64) {
791 // The AAPCS variadic function ABI is identical to the non-variadic
792 // one. As a result there may be more arguments in registers and we should
793 // save them for future reference.
794 // Win64 variadic functions also pass arguments in registers, but all
795 // float arguments are passed in integer registers.
796 saveVarArgRegisters(MIRBuilder, Handler, CCInfo);
797 } else if (Subtarget.isWindowsArm64EC()) {
798 return false;
799 }
800
801 // We currently pass all varargs at 8-byte alignment, or 4 in ILP32.
802 StackSize = alignTo(Assigner.StackSize, Subtarget.isTargetILP32() ? 4 : 8);
803
804 auto &MFI = MIRBuilder.getMF().getFrameInfo();
805 FuncInfo->setVarArgsStackIndex(MFI.CreateFixedObject(4, StackSize, true));
806 }
807
808 if (doesCalleeRestoreStack(F.getCallingConv(),
810 // We have a non-standard ABI, so why not make full use of the stack that
811 // we're going to pop? It must be aligned to 16 B in any case.
812 StackSize = alignTo(StackSize, 16);
813
814 // If we're expected to restore the stack (e.g. fastcc), then we'll be
815 // adding a multiple of 16.
816 FuncInfo->setArgumentStackToRestore(StackSize);
817
818 // Our own callers will guarantee that the space is free by giving an
819 // aligned value to CALLSEQ_START.
820 }
821
822 // When we tail call, we need to check if the callee's arguments
823 // will fit on the caller's stack. So, whenever we lower formal arguments,
824 // we should keep track of this information, since we might lower a tail call
825 // in this function later.
826 FuncInfo->setBytesInStackArgArea(StackSize);
827
828 if (Subtarget.hasCustomCallingConv())
829 Subtarget.getRegisterInfo()->UpdateCustomCalleeSavedRegs(MF);
830
831 handleMustTailForwardedRegisters(MIRBuilder, AssignFn);
832
833 // Move back to the end of the basic block.
834 MIRBuilder.setMBB(MBB);
835
836 return true;
837}
838
839/// Return true if the calling convention is one that we can guarantee TCO for.
840static bool canGuaranteeTCO(CallingConv::ID CC, bool GuaranteeTailCalls) {
841 return (CC == CallingConv::Fast && GuaranteeTailCalls) ||
843}
844
845/// Return true if we might ever do TCO for calls with this calling convention.
847 switch (CC) {
848 case CallingConv::C:
856 return true;
857 default:
858 return false;
859 }
860}
861
862/// Returns a pair containing the fixed CCAssignFn and the vararg CCAssignFn for
863/// CC.
864static std::pair<CCAssignFn *, CCAssignFn *>
866 return {TLI.CCAssignFnForCall(CC, false), TLI.CCAssignFnForCall(CC, true)};
867}
868
869bool AArch64CallLowering::doCallerAndCalleePassArgsTheSameWay(
870 CallLoweringInfo &Info, MachineFunction &MF,
871 SmallVectorImpl<ArgInfo> &InArgs) const {
872 const Function &CallerF = MF.getFunction();
873 CallingConv::ID CalleeCC = Info.CallConv;
874 CallingConv::ID CallerCC = CallerF.getCallingConv();
875
876 // If the calling conventions match, then everything must be the same.
877 if (CalleeCC == CallerCC)
878 return true;
879
880 // Check if the caller and callee will handle arguments in the same way.
881 const AArch64TargetLowering &TLI = *getTLI<AArch64TargetLowering>();
882 CCAssignFn *CalleeAssignFnFixed;
883 CCAssignFn *CalleeAssignFnVarArg;
884 std::tie(CalleeAssignFnFixed, CalleeAssignFnVarArg) =
885 getAssignFnsForCC(CalleeCC, TLI);
886
887 CCAssignFn *CallerAssignFnFixed;
888 CCAssignFn *CallerAssignFnVarArg;
889 std::tie(CallerAssignFnFixed, CallerAssignFnVarArg) =
890 getAssignFnsForCC(CallerCC, TLI);
891
892 AArch64IncomingValueAssigner CalleeAssigner(CalleeAssignFnFixed,
893 CalleeAssignFnVarArg);
894 AArch64IncomingValueAssigner CallerAssigner(CallerAssignFnFixed,
895 CallerAssignFnVarArg);
896
897 if (!resultsCompatible(Info, MF, InArgs, CalleeAssigner, CallerAssigner))
898 return false;
899
900 // Make sure that the caller and callee preserve all of the same registers.
901 auto TRI = MF.getSubtarget<AArch64Subtarget>().getRegisterInfo();
902 const uint32_t *CallerPreserved = TRI->getCallPreservedMask(MF, CallerCC);
903 const uint32_t *CalleePreserved = TRI->getCallPreservedMask(MF, CalleeCC);
904 if (MF.getSubtarget<AArch64Subtarget>().hasCustomCallingConv()) {
905 TRI->UpdateCustomCallPreservedMask(MF, &CallerPreserved);
906 TRI->UpdateCustomCallPreservedMask(MF, &CalleePreserved);
907 }
908
909 return TRI->regmaskSubsetEqual(CallerPreserved, CalleePreserved);
910}
911
912bool AArch64CallLowering::areCalleeOutgoingArgsTailCallable(
913 CallLoweringInfo &Info, MachineFunction &MF,
914 SmallVectorImpl<ArgInfo> &OrigOutArgs) const {
915 // If there are no outgoing arguments, then we are done.
916 if (OrigOutArgs.empty())
917 return true;
918
919 const Function &CallerF = MF.getFunction();
920 LLVMContext &Ctx = CallerF.getContext();
921 CallingConv::ID CalleeCC = Info.CallConv;
922 CallingConv::ID CallerCC = CallerF.getCallingConv();
923 const AArch64TargetLowering &TLI = *getTLI<AArch64TargetLowering>();
924 const AArch64Subtarget &Subtarget = MF.getSubtarget<AArch64Subtarget>();
925
926 CCAssignFn *AssignFnFixed;
927 CCAssignFn *AssignFnVarArg;
928 std::tie(AssignFnFixed, AssignFnVarArg) = getAssignFnsForCC(CalleeCC, TLI);
929
930 // We have outgoing arguments. Make sure that we can tail call with them.
932 CCState OutInfo(CalleeCC, false, MF, OutLocs, Ctx);
933
934 AArch64OutgoingValueAssigner CalleeAssigner(AssignFnFixed, AssignFnVarArg,
935 Subtarget, /*IsReturn*/ false);
936 // determineAssignments() may modify argument flags, so make a copy.
938 append_range(OutArgs, OrigOutArgs);
939 if (!determineAssignments(CalleeAssigner, OutArgs, OutInfo)) {
940 LLVM_DEBUG(dbgs() << "... Could not analyze call operands.\n");
941 return false;
942 }
943
944 // Make sure that they can fit on the caller's stack.
945 const AArch64FunctionInfo *FuncInfo = MF.getInfo<AArch64FunctionInfo>();
946 if (OutInfo.getStackSize() > FuncInfo->getBytesInStackArgArea()) {
947 LLVM_DEBUG(dbgs() << "... Cannot fit call operands on caller's stack.\n");
948 return false;
949 }
950
951 // Verify that the parameters in callee-saved registers match.
952 // TODO: Port this over to CallLowering as general code once swiftself is
953 // supported.
954 auto TRI = MF.getSubtarget<AArch64Subtarget>().getRegisterInfo();
955 const uint32_t *CallerPreservedMask = TRI->getCallPreservedMask(MF, CallerCC);
956 MachineRegisterInfo &MRI = MF.getRegInfo();
957
958 if (Info.IsVarArg) {
959 // Be conservative and disallow variadic memory operands to match SDAG's
960 // behaviour.
961 // FIXME: If the caller's calling convention is C, then we can
962 // potentially use its argument area. However, for cases like fastcc,
963 // we can't do anything.
964 for (unsigned i = 0; i < OutLocs.size(); ++i) {
965 auto &ArgLoc = OutLocs[i];
966 if (ArgLoc.isRegLoc())
967 continue;
968
970 dbgs()
971 << "... Cannot tail call vararg function with stack arguments\n");
972 return false;
973 }
974 }
975
976 return parametersInCSRMatch(MRI, CallerPreservedMask, OutLocs, OutArgs);
977}
978
980 MachineIRBuilder &MIRBuilder, CallLoweringInfo &Info,
982 SmallVectorImpl<ArgInfo> &OutArgs) const {
983
984 // Must pass all target-independent checks in order to tail call optimize.
985 if (!Info.IsTailCall)
986 return false;
987
988 CallingConv::ID CalleeCC = Info.CallConv;
989 MachineFunction &MF = MIRBuilder.getMF();
990 const Function &CallerF = MF.getFunction();
991
992 LLVM_DEBUG(dbgs() << "Attempting to lower call as tail call\n");
993
994 if (Info.SwiftErrorVReg) {
995 // TODO: We should handle this.
996 // Note that this is also handled by the check for no outgoing arguments.
997 // Proactively disabling this though, because the swifterror handling in
998 // lowerCall inserts a COPY *after* the location of the call.
999 LLVM_DEBUG(dbgs() << "... Cannot handle tail calls with swifterror yet.\n");
1000 return false;
1001 }
1002
1003 if (!mayTailCallThisCC(CalleeCC)) {
1004 LLVM_DEBUG(dbgs() << "... Calling convention cannot be tail called.\n");
1005 return false;
1006 }
1007
1008 // Byval parameters hand the function a pointer directly into the stack area
1009 // we want to reuse during a tail call. Working around this *is* possible (see
1010 // X86).
1011 //
1012 // FIXME: In AArch64ISelLowering, this isn't worked around. Can/should we try
1013 // it?
1014 //
1015 // On Windows, "inreg" attributes signify non-aggregate indirect returns.
1016 // In this case, it is necessary to save/restore X0 in the callee. Tail
1017 // call opt interferes with this. So we disable tail call opt when the
1018 // caller has an argument with "inreg" attribute.
1019 //
1020 // FIXME: Check whether the callee also has an "inreg" argument.
1021 //
1022 // When the caller has a swifterror argument, we don't want to tail call
1023 // because would have to move into the swifterror register before the
1024 // tail call.
1025 if (any_of(CallerF.args(), [](const Argument &A) {
1026 return A.hasByValAttr() || A.hasInRegAttr() || A.hasSwiftErrorAttr();
1027 })) {
1028 LLVM_DEBUG(dbgs() << "... Cannot tail call from callers with byval, "
1029 "inreg, or swifterror arguments\n");
1030 return false;
1031 }
1032
1033 // Externally-defined functions with weak linkage should not be
1034 // tail-called on AArch64 when the OS does not support dynamic
1035 // pre-emption of symbols, as the AAELF spec requires normal calls
1036 // to undefined weak functions to be replaced with a NOP or jump to the
1037 // next instruction. The behaviour of branch instructions in this
1038 // situation (as used for tail calls) is implementation-defined, so we
1039 // cannot rely on the linker replacing the tail call with a return.
1040 if (Info.Callee.isGlobal()) {
1041 const GlobalValue *GV = Info.Callee.getGlobal();
1042 const Triple &TT = MF.getTarget().getTargetTriple();
1043 if (GV->hasExternalWeakLinkage() &&
1044 (!TT.isOSWindows() || TT.isOSBinFormatELF() ||
1045 TT.isOSBinFormatMachO())) {
1046 LLVM_DEBUG(dbgs() << "... Cannot tail call externally-defined function "
1047 "with weak linkage for this OS.\n");
1048 return false;
1049 }
1050 }
1051
1052 // If we have -tailcallopt, then we're done.
1054 return CalleeCC == CallerF.getCallingConv();
1055
1056 // We don't have -tailcallopt, so we're allowed to change the ABI (sibcall).
1057 // Try to find cases where we can do that.
1058
1059 // I want anyone implementing a new calling convention to think long and hard
1060 // about this assert.
1061 assert((!Info.IsVarArg || CalleeCC == CallingConv::C) &&
1062 "Unexpected variadic calling convention");
1063
1064 // Verify that the incoming and outgoing arguments from the callee are
1065 // safe to tail call.
1066 if (!doCallerAndCalleePassArgsTheSameWay(Info, MF, InArgs)) {
1067 LLVM_DEBUG(
1068 dbgs()
1069 << "... Caller and callee have incompatible calling conventions.\n");
1070 return false;
1071 }
1072
1073 if (!areCalleeOutgoingArgsTailCallable(Info, MF, OutArgs))
1074 return false;
1075
1076 LLVM_DEBUG(
1077 dbgs() << "... Call is eligible for tail call optimization.\n");
1078 return true;
1079}
1080
1081static unsigned getCallOpcode(const MachineFunction &CallerF, bool IsIndirect,
1082 bool IsTailCall,
1083 std::optional<CallLowering::PtrAuthInfo> &PAI,
1084 MachineRegisterInfo &MRI) {
1085 const AArch64FunctionInfo *FuncInfo = CallerF.getInfo<AArch64FunctionInfo>();
1086
1087 if (!IsTailCall) {
1088 if (!PAI)
1089 return IsIndirect ? getBLRCallOpcode(CallerF) : (unsigned)AArch64::BL;
1090
1091 assert(IsIndirect && "Direct call should not be authenticated");
1092 assert((PAI->Key == AArch64PACKey::IA || PAI->Key == AArch64PACKey::IB) &&
1093 "Invalid auth call key");
1094 return AArch64::BLRA;
1095 }
1096
1097 if (!IsIndirect)
1098 return AArch64::TCRETURNdi;
1099
1100 // When BTI or PAuthLR are enabled, there are restrictions on using x16 and
1101 // x17 to hold the function pointer.
1102 if (FuncInfo->branchTargetEnforcement()) {
1103 if (FuncInfo->branchProtectionPAuthLR()) {
1104 assert(!PAI && "ptrauth tail-calls not yet supported with PAuthLR");
1105 return AArch64::TCRETURNrix17;
1106 }
1107 if (PAI)
1108 return AArch64::AUTH_TCRETURN_BTI;
1109 return AArch64::TCRETURNrix16x17;
1110 }
1111
1112 if (FuncInfo->branchProtectionPAuthLR()) {
1113 assert(!PAI && "ptrauth tail-calls not yet supported with PAuthLR");
1114 return AArch64::TCRETURNrinotx16;
1115 }
1116
1117 if (PAI)
1118 return AArch64::AUTH_TCRETURN;
1119 return AArch64::TCRETURNri;
1120}
1121
1122static const uint32_t *
1126 const uint32_t *Mask;
1127 if (!OutArgs.empty() && OutArgs[0].Flags[0].isReturned()) {
1128 // For 'this' returns, use the X0-preserving mask if applicable
1129 Mask = TRI.getThisReturnPreservedMask(MF, Info.CallConv);
1130 if (!Mask) {
1131 OutArgs[0].Flags[0].setReturned(false);
1132 Mask = TRI.getCallPreservedMask(MF, Info.CallConv);
1133 }
1134 } else {
1135 Mask = TRI.getCallPreservedMask(MF, Info.CallConv);
1136 }
1137 return Mask;
1138}
1139
1140bool AArch64CallLowering::lowerTailCall(
1141 MachineIRBuilder &MIRBuilder, CallLoweringInfo &Info,
1142 SmallVectorImpl<ArgInfo> &OutArgs) const {
1143 MachineFunction &MF = MIRBuilder.getMF();
1144 const Function &F = MF.getFunction();
1145 MachineRegisterInfo &MRI = MF.getRegInfo();
1146 const AArch64TargetLowering &TLI = *getTLI<AArch64TargetLowering>();
1147 AArch64FunctionInfo *FuncInfo = MF.getInfo<AArch64FunctionInfo>();
1148
1149 // True when we're tail calling, but without -tailcallopt.
1150 bool IsSibCall = !MF.getTarget().Options.GuaranteedTailCallOpt &&
1151 Info.CallConv != CallingConv::Tail &&
1152 Info.CallConv != CallingConv::SwiftTail;
1153
1154 // Find out which ABI gets to decide where things go.
1155 CallingConv::ID CalleeCC = Info.CallConv;
1156 CCAssignFn *AssignFnFixed;
1157 CCAssignFn *AssignFnVarArg;
1158 std::tie(AssignFnFixed, AssignFnVarArg) = getAssignFnsForCC(CalleeCC, TLI);
1159
1160 MachineInstrBuilder CallSeqStart;
1161 if (!IsSibCall)
1162 CallSeqStart = MIRBuilder.buildInstr(AArch64::ADJCALLSTACKDOWN);
1163
1164 unsigned Opc = getCallOpcode(MF, Info.Callee.isReg(), true, Info.PAI, MRI);
1165 auto MIB = MIRBuilder.buildInstrNoInsert(Opc);
1166 MIB.add(Info.Callee);
1167
1168 // Tell the call which registers are clobbered.
1169 const AArch64Subtarget &Subtarget = MF.getSubtarget<AArch64Subtarget>();
1170 auto TRI = Subtarget.getRegisterInfo();
1171
1172 // Byte offset for the tail call. When we are sibcalling, this will always
1173 // be 0.
1174 MIB.addImm(0);
1175
1176 // Authenticated tail calls always take key/discriminator arguments.
1177 if (Opc == AArch64::AUTH_TCRETURN || Opc == AArch64::AUTH_TCRETURN_BTI) {
1178 assert((Info.PAI->Key == AArch64PACKey::IA ||
1179 Info.PAI->Key == AArch64PACKey::IB) &&
1180 "Invalid auth call key");
1181 MIB.addImm(Info.PAI->Key);
1182
1183 Register AddrDisc = 0;
1184 uint16_t IntDisc = 0;
1185 std::tie(IntDisc, AddrDisc) =
1186 extractPtrauthBlendDiscriminators(Info.PAI->Discriminator, MRI);
1187
1188 MIB.addImm(IntDisc);
1189 MIB.addUse(AddrDisc);
1190 if (AddrDisc != AArch64::NoRegister) {
1191 MIB->getOperand(4).setReg(constrainOperandRegClass(
1192 MF, *TRI, MRI, *MF.getSubtarget().getInstrInfo(),
1193 *MF.getSubtarget().getRegBankInfo(), *MIB, MIB->getDesc(),
1194 MIB->getOperand(4), 4));
1195 }
1196 }
1197
1198 // Tell the call which registers are clobbered.
1199 const uint32_t *Mask = TRI->getCallPreservedMask(MF, CalleeCC);
1200 if (Subtarget.hasCustomCallingConv())
1201 TRI->UpdateCustomCallPreservedMask(MF, &Mask);
1202 MIB.addRegMask(Mask);
1203
1204 if (Info.CFIType)
1205 MIB->setCFIType(MF, Info.CFIType->getZExtValue());
1206
1207 if (TRI->isAnyArgRegReserved(MF))
1208 TRI->emitReservedArgRegCallError(MF);
1209
1210 // FPDiff is the byte offset of the call's argument area from the callee's.
1211 // Stores to callee stack arguments will be placed in FixedStackSlots offset
1212 // by this amount for a tail call. In a sibling call it must be 0 because the
1213 // caller will deallocate the entire stack and the callee still expects its
1214 // arguments to begin at SP+0.
1215 int FPDiff = 0;
1216
1217 // This will be 0 for sibcalls, potentially nonzero for tail calls produced
1218 // by -tailcallopt. For sibcalls, the memory operands for the call are
1219 // already available in the caller's incoming argument space.
1220 unsigned NumBytes = 0;
1221 if (!IsSibCall) {
1222 // We aren't sibcalling, so we need to compute FPDiff. We need to do this
1223 // before handling assignments, because FPDiff must be known for memory
1224 // arguments.
1225 unsigned NumReusableBytes = FuncInfo->getBytesInStackArgArea();
1227 CCState OutInfo(CalleeCC, false, MF, OutLocs, F.getContext());
1228
1229 AArch64OutgoingValueAssigner CalleeAssigner(AssignFnFixed, AssignFnVarArg,
1230 Subtarget, /*IsReturn*/ false);
1231 if (!determineAssignments(CalleeAssigner, OutArgs, OutInfo))
1232 return false;
1233
1234 // The callee will pop the argument stack as a tail call. Thus, we must
1235 // keep it 16-byte aligned.
1236 NumBytes = alignTo(OutInfo.getStackSize(), 16);
1237
1238 // FPDiff will be negative if this tail call requires more space than we
1239 // would automatically have in our incoming argument space. Positive if we
1240 // actually shrink the stack.
1241 FPDiff = NumReusableBytes - NumBytes;
1242
1243 // Update the required reserved area if this is the tail call requiring the
1244 // most argument stack space.
1245 if (FPDiff < 0 && FuncInfo->getTailCallReservedStack() < (unsigned)-FPDiff)
1246 FuncInfo->setTailCallReservedStack(-FPDiff);
1247
1248 // The stack pointer must be 16-byte aligned at all times it's used for a
1249 // memory operation, which in practice means at *all* times and in
1250 // particular across call boundaries. Therefore our own arguments started at
1251 // a 16-byte aligned SP and the delta applied for the tail call should
1252 // satisfy the same constraint.
1253 assert(FPDiff % 16 == 0 && "unaligned stack on tail call");
1254 }
1255
1256 const auto &Forwards = FuncInfo->getForwardedMustTailRegParms();
1257
1258 AArch64OutgoingValueAssigner Assigner(AssignFnFixed, AssignFnVarArg,
1259 Subtarget, /*IsReturn*/ false);
1260
1261 // Do the actual argument marshalling.
1262 OutgoingArgHandler Handler(MIRBuilder, MRI, MIB,
1263 /*IsTailCall*/ true, FPDiff);
1264 if (!determineAndHandleAssignments(Handler, Assigner, OutArgs, MIRBuilder,
1265 CalleeCC, Info.IsVarArg))
1266 return false;
1267
1268 Mask = getMaskForArgs(OutArgs, Info, *TRI, MF);
1269
1270 if (Info.IsVarArg && Info.IsMustTailCall) {
1271 // Now we know what's being passed to the function. Add uses to the call for
1272 // the forwarded registers that we *aren't* passing as parameters. This will
1273 // preserve the copies we build earlier.
1274 for (const auto &F : Forwards) {
1275 Register ForwardedReg = F.PReg;
1276 // If the register is already passed, or aliases a register which is
1277 // already being passed, then skip it.
1278 if (any_of(MIB->uses(), [&ForwardedReg, &TRI](const MachineOperand &Use) {
1279 if (!Use.isReg())
1280 return false;
1281 return TRI->regsOverlap(Use.getReg(), ForwardedReg);
1282 }))
1283 continue;
1284
1285 // We aren't passing it already, so we should add it to the call.
1286 MIRBuilder.buildCopy(ForwardedReg, Register(F.VReg));
1287 MIB.addReg(ForwardedReg, RegState::Implicit);
1288 }
1289 }
1290
1291 // If we have -tailcallopt, we need to adjust the stack. We'll do the call
1292 // sequence start and end here.
1293 if (!IsSibCall) {
1294 MIB->getOperand(1).setImm(FPDiff);
1295 CallSeqStart.addImm(0).addImm(0);
1296 // End the call sequence *before* emitting the call. Normally, we would
1297 // tidy the frame up after the call. However, here, we've laid out the
1298 // parameters so that when SP is reset, they will be in the correct
1299 // location.
1300 MIRBuilder.buildInstr(AArch64::ADJCALLSTACKUP).addImm(0).addImm(0);
1301 }
1302
1303 // Now we can add the actual call instruction to the correct basic block.
1304 MIRBuilder.insertInstr(MIB);
1305
1306 // If Callee is a reg, since it is used by a target specific instruction,
1307 // it must have a register class matching the constraint of that instruction.
1308 if (MIB->getOperand(0).isReg())
1310 *MF.getSubtarget().getRegBankInfo(), *MIB,
1311 MIB->getDesc(), MIB->getOperand(0), 0);
1312
1314 Info.LoweredTailCall = true;
1315 return true;
1316}
1317
1319 CallLoweringInfo &Info) const {
1320 MachineFunction &MF = MIRBuilder.getMF();
1321 const Function &F = MF.getFunction();
1322 MachineRegisterInfo &MRI = MF.getRegInfo();
1323 auto &DL = F.getDataLayout();
1325 const AArch64Subtarget &Subtarget = MF.getSubtarget<AArch64Subtarget>();
1326
1327 // Arm64EC has extra requirements for varargs calls; bail out for now.
1328 //
1329 // Arm64EC has special mangling rules for calls; bail out on all calls for
1330 // now.
1331 if (Subtarget.isWindowsArm64EC())
1332 return false;
1333
1334 // Arm64EC thunks have a special calling convention which is only implemented
1335 // in SelectionDAG; bail out for now.
1336 if (Info.CallConv == CallingConv::ARM64EC_Thunk_Native ||
1337 Info.CallConv == CallingConv::ARM64EC_Thunk_X64)
1338 return false;
1339
1341 for (auto &OrigArg : Info.OrigArgs) {
1342 splitToValueTypes(OrigArg, OutArgs, DL, Info.CallConv);
1343 // AAPCS requires that we zero-extend i1 to 8 bits by the caller.
1344 auto &Flags = OrigArg.Flags[0];
1345 if (OrigArg.Ty->isIntegerTy(1) && !Flags.isSExt() && !Flags.isZExt()) {
1346 ArgInfo &OutArg = OutArgs.back();
1347 assert(OutArg.Regs.size() == 1 &&
1348 MRI.getType(OutArg.Regs[0]).getSizeInBits() == 1 &&
1349 "Unexpected registers used for i1 arg");
1350
1351 // We cannot use a ZExt ArgInfo flag here, because it will
1352 // zero-extend the argument to i32 instead of just i8.
1353 OutArg.Regs[0] =
1354 MIRBuilder.buildZExt(LLT::integer(8), OutArg.Regs[0]).getReg(0);
1355 LLVMContext &Ctx = MF.getFunction().getContext();
1356 OutArg.Ty = Type::getInt8Ty(Ctx);
1357 }
1358 }
1359
1361 if (!Info.OrigRet.Ty->isVoidTy())
1362 splitToValueTypes(Info.OrigRet, InArgs, DL, Info.CallConv);
1363
1364 // If we can lower as a tail call, do that instead.
1365 bool CanTailCallOpt =
1366 isEligibleForTailCallOptimization(MIRBuilder, Info, InArgs, OutArgs);
1367
1368 // We must emit a tail call if we have musttail.
1369 if (Info.IsMustTailCall && !CanTailCallOpt) {
1370 // There are types of incoming/outgoing arguments we can't handle yet, so
1371 // it doesn't make sense to actually die here like in ISelLowering. Instead,
1372 // fall back to SelectionDAG and let it try to handle this.
1373 LLVM_DEBUG(dbgs() << "Failed to lower musttail call as tail call\n");
1374 return false;
1375 }
1376
1377 Info.IsTailCall = CanTailCallOpt;
1378 if (CanTailCallOpt)
1379 return lowerTailCall(MIRBuilder, Info, OutArgs);
1380
1381 // Find out which ABI gets to decide where things go.
1382 CCAssignFn *AssignFnFixed;
1383 CCAssignFn *AssignFnVarArg;
1384 std::tie(AssignFnFixed, AssignFnVarArg) =
1385 getAssignFnsForCC(Info.CallConv, TLI);
1386
1387 MachineInstrBuilder CallSeqStart;
1388 CallSeqStart = MIRBuilder.buildInstr(AArch64::ADJCALLSTACKDOWN);
1389
1390 // Create a temporarily-floating call instruction so we can add the implicit
1391 // uses of arg registers.
1392
1393 unsigned Opc = 0;
1394 // Calls with operand bundle "clang.arc.attachedcall" are special. They should
1395 // be expanded to the call, directly followed by a special marker sequence and
1396 // a call to an ObjC library function.
1397 if (Info.CB && objcarc::hasAttachedCallOpBundle(Info.CB))
1398 Opc = Info.PAI ? AArch64::BLRA_RVMARKER : AArch64::BLR_RVMARKER;
1399 // A call to a returns twice function like setjmp must be followed by a bti
1400 // instruction.
1401 else if (Info.CB && Info.CB->hasFnAttr(Attribute::ReturnsTwice) &&
1402 !Subtarget.noBTIAtReturnTwice() &&
1404 Opc = AArch64::BLR_BTI;
1405 else {
1406 // For an intrinsic call (e.g. memset), use GOT if "RtLibUseGOT" (-fno-plt)
1407 // is set.
1408 if (Info.Callee.isSymbol() && F.getParent()->getRtLibUseGOT()) {
1409 auto MIB = MIRBuilder.buildInstr(TargetOpcode::G_GLOBAL_VALUE);
1410 DstOp(getLLTForType(*F.getType(), DL)).addDefToMIB(MRI, MIB);
1411 MIB.addExternalSymbol(Info.Callee.getSymbolName(), AArch64II::MO_GOT);
1412 Info.Callee = MachineOperand::CreateReg(MIB.getReg(0), false);
1413 }
1414 Opc = getCallOpcode(MF, Info.Callee.isReg(), false, Info.PAI, MRI);
1415 }
1416
1417 auto MIB = MIRBuilder.buildInstrNoInsert(Opc);
1418 unsigned CalleeOpNo = 0;
1419
1420 if (Opc == AArch64::BLR_RVMARKER || Opc == AArch64::BLRA_RVMARKER) {
1421 // Add a target global address for the retainRV/claimRV runtime function
1422 // just before the call target.
1423 Function *ARCFn = *objcarc::getAttachedARCFunction(Info.CB);
1424 MIB.addGlobalAddress(ARCFn);
1425 ++CalleeOpNo;
1426
1427 // We may or may not need to emit both the marker and the retain/claim call.
1428 // Tell the pseudo expansion using an additional boolean op.
1429 MIB.addImm(objcarc::attachedCallOpBundleNeedsMarker(Info.CB));
1430 ++CalleeOpNo;
1431 } else if (Info.CFIType) {
1432 MIB->setCFIType(MF, Info.CFIType->getZExtValue());
1433 }
1434 MIB->setDeactivationSymbol(MF, Info.DeactivationSymbol);
1435
1436 MIB.add(Info.Callee);
1437
1438 // Tell the call which registers are clobbered.
1439 const uint32_t *Mask;
1440 const auto *TRI = Subtarget.getRegisterInfo();
1441
1442 AArch64OutgoingValueAssigner Assigner(AssignFnFixed, AssignFnVarArg,
1443 Subtarget, /*IsReturn*/ false);
1444 // Do the actual argument marshalling.
1445 OutgoingArgHandler Handler(MIRBuilder, MRI, MIB, /*IsReturn*/ false);
1446 if (!determineAndHandleAssignments(Handler, Assigner, OutArgs, MIRBuilder,
1447 Info.CallConv, Info.IsVarArg))
1448 return false;
1449
1450 Mask = getMaskForArgs(OutArgs, Info, *TRI, MF);
1451
1452 if (Opc == AArch64::BLRA || Opc == AArch64::BLRA_RVMARKER) {
1453 assert((Info.PAI->Key == AArch64PACKey::IA ||
1454 Info.PAI->Key == AArch64PACKey::IB) &&
1455 "Invalid auth call key");
1456 MIB.addImm(Info.PAI->Key);
1457
1458 Register AddrDisc = 0;
1459 uint16_t IntDisc = 0;
1460 std::tie(IntDisc, AddrDisc) =
1461 extractPtrauthBlendDiscriminators(Info.PAI->Discriminator, MRI);
1462
1463 MIB.addImm(IntDisc);
1464 MIB.addUse(AddrDisc);
1465 if (AddrDisc != AArch64::NoRegister) {
1467 *MF.getSubtarget().getRegBankInfo(), *MIB,
1468 MIB->getDesc(), MIB->getOperand(CalleeOpNo + 3),
1469 CalleeOpNo + 3);
1470 }
1471 }
1472
1473 // Tell the call which registers are clobbered.
1475 TRI->UpdateCustomCallPreservedMask(MF, &Mask);
1476 MIB.addRegMask(Mask);
1477
1478 if (TRI->isAnyArgRegReserved(MF))
1479 TRI->emitReservedArgRegCallError(MF);
1480
1481 // Now we can add the actual call instruction to the correct basic block.
1482 MIRBuilder.insertInstr(MIB);
1483
1484 uint64_t CalleePopBytes =
1485 doesCalleeRestoreStack(Info.CallConv,
1487 ? alignTo(Assigner.StackSize, 16)
1488 : 0;
1489
1490 CallSeqStart.addImm(Assigner.StackSize).addImm(0);
1491 MIRBuilder.buildInstr(AArch64::ADJCALLSTACKUP)
1492 .addImm(Assigner.StackSize)
1493 .addImm(CalleePopBytes);
1494
1495 // If Callee is a reg, since it is used by a target specific
1496 // instruction, it must have a register class matching the
1497 // constraint of that instruction.
1498 if (MIB->getOperand(CalleeOpNo).isReg())
1499 constrainOperandRegClass(MF, *TRI, MRI, *Subtarget.getInstrInfo(),
1500 *Subtarget.getRegBankInfo(), *MIB, MIB->getDesc(),
1501 MIB->getOperand(CalleeOpNo), CalleeOpNo);
1502
1503 // Finally we can copy the returned value back into its virtual-register. In
1504 // symmetry with the arguments, the physical register must be an
1505 // implicit-define of the call instruction.
1506 if (Info.CanLowerReturn && !Info.OrigRet.Ty->isVoidTy()) {
1507 CCAssignFn *RetAssignFn = TLI.CCAssignFnForReturn(Info.CallConv);
1508 CallReturnHandler Handler(MIRBuilder, MRI, MIB);
1509 bool UsingReturnedArg =
1510 !OutArgs.empty() && OutArgs[0].Flags[0].isReturned();
1511
1512 AArch64OutgoingValueAssigner Assigner(RetAssignFn, RetAssignFn, Subtarget,
1513 /*IsReturn*/ false);
1514 ReturnedArgCallReturnHandler ReturnedArgHandler(MIRBuilder, MRI, MIB);
1516 UsingReturnedArg ? ReturnedArgHandler : Handler, Assigner, InArgs,
1517 MIRBuilder, Info.CallConv, Info.IsVarArg,
1518 UsingReturnedArg ? ArrayRef(OutArgs[0].Regs)
1519 : ArrayRef<Register>()))
1520 return false;
1521 }
1522
1523 if (Info.SwiftErrorVReg) {
1524 MIB.addDef(AArch64::X21, RegState::Implicit);
1525 MIRBuilder.buildCopy(Info.SwiftErrorVReg, Register(AArch64::X21));
1526 }
1527
1528 if (!Info.CanLowerReturn) {
1529 insertSRetLoads(MIRBuilder, Info.OrigRet.Ty, Info.OrigRet.Regs,
1530 Info.DemoteRegister, Info.DemoteStackIndex);
1531 }
1532 return true;
1533}
1534
1536 return Ty.getSizeInBits() == 64;
1537}
static void handleMustTailForwardedRegisters(MachineIRBuilder &MIRBuilder, CCAssignFn *AssignFn)
Helper function to compute forwarded registers for musttail calls.
static unsigned getCallOpcode(const MachineFunction &CallerF, bool IsIndirect, bool IsTailCall, std::optional< CallLowering::PtrAuthInfo > &PAI, MachineRegisterInfo &MRI)
static LLT getStackValueStoreTypeHack(const CCValAssign &VA)
static const uint32_t * getMaskForArgs(SmallVectorImpl< AArch64CallLowering::ArgInfo > &OutArgs, AArch64CallLowering::CallLoweringInfo &Info, const AArch64RegisterInfo &TRI, MachineFunction &MF)
static void applyStackPassedSmallTypeDAGHack(EVT OrigVT, MVT &ValVT, MVT &LocVT)
static std::pair< CCAssignFn *, CCAssignFn * > getAssignFnsForCC(CallingConv::ID CC, const AArch64TargetLowering &TLI)
Returns a pair containing the fixed CCAssignFn and the vararg CCAssignFn for CC.
static bool doesCalleeRestoreStack(CallingConv::ID CallConv, bool TailCallOpt)
This file describes how to lower LLVM calls to machine code calls.
MachineInstrBuilder MachineInstrBuilder & DefMI
static std::tuple< SDValue, SDValue > extractPtrauthBlendDiscriminators(SDValue Disc, SelectionDAG *DAG)
static bool shouldLowerTailCallStackArg(const MachineFunction &MF, const CCValAssign &VA, SDValue Arg, ISD::ArgFlagsTy Flags, int CallOffset)
Check whether a stack argument requires lowering in a tail call.
static const MCPhysReg GPRArgRegs[]
static const MCPhysReg FPRArgRegs[]
cl::opt< bool > EnableSVEGISel("aarch64-enable-gisel-sve", cl::Hidden, cl::desc("Enable / disable SVE scalable vectors in Global ISel"), cl::init(false))
static bool canGuaranteeTCO(CallingConv::ID CC, bool GuaranteeTailCalls)
Return true if the calling convention is one that we can guarantee TCO for.
static bool mayTailCallThisCC(CallingConv::ID CC)
Return true if we might ever do TCO for calls with this calling convention.
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock & MBB
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
This file contains the simple types necessary to represent the attributes associated with functions a...
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
Declares convenience wrapper classes for interpreting MachineInstr instances as specific generic oper...
Implement a low-level type suitable for MachineInstr level instruction selection.
#define F(x, y, z)
Definition MD5.cpp:54
This file declares the MachineIRBuilder class.
Register Reg
Register const TargetRegisterInfo * TRI
Promote Memory to Register
Definition Mem2Reg.cpp:110
static MCRegister getReg(const MCDisassembler *D, unsigned RC, unsigned RegNo)
This file defines ARC utility functions which are used by various parts of the compiler.
static constexpr MCPhysReg SPReg
This file defines the SmallVector class.
#define LLVM_DEBUG(...)
Definition Debug.h:119
bool lowerReturn(MachineIRBuilder &MIRBuilder, const Value *Val, ArrayRef< Register > VRegs, FunctionLoweringInfo &FLI, Register SwiftErrorVReg) const override
This hook must be implemented to lower outgoing return values, described by Val, into the specified v...
bool canLowerReturn(MachineFunction &MF, CallingConv::ID CallConv, SmallVectorImpl< BaseArgInfo > &Outs, bool IsVarArg) const override
This hook must be implemented to check whether the return values described by Outs can fit into the r...
bool fallBackToDAGISel(const MachineFunction &MF) const override
bool isTypeIsValidForThisReturn(EVT Ty) const override
For targets which support the "returned" parameter attribute, returns true if the given type is a val...
bool isEligibleForTailCallOptimization(MachineIRBuilder &MIRBuilder, CallLoweringInfo &Info, SmallVectorImpl< ArgInfo > &InArgs, SmallVectorImpl< ArgInfo > &OutArgs) const
Returns true if the call can be lowered as a tail call.
AArch64CallLowering(const AArch64TargetLowering &TLI)
bool lowerCall(MachineIRBuilder &MIRBuilder, CallLoweringInfo &Info) const override
This hook must be implemented to lower the given call instruction, including argument and return valu...
bool lowerFormalArguments(MachineIRBuilder &MIRBuilder, const Function &F, ArrayRef< ArrayRef< Register > > VRegs, FunctionLoweringInfo &FLI) const override
This hook must be implemented to lower the incoming (formal) arguments, described by VRegs,...
AArch64FunctionInfo - This class is derived from MachineFunctionInfo and contains private AArch64-spe...
void setTailCallReservedStack(unsigned bytes)
SmallVectorImpl< ForwardedRegister > & getForwardedMustTailRegParms()
void setBytesInStackArgArea(unsigned bytes)
void setArgumentStackToRestore(unsigned bytes)
const AArch64RegisterInfo * getRegisterInfo() const override
const AArch64InstrInfo * getInstrInfo() const override
bool isCallingConvWin64(CallingConv::ID CC, bool IsVarArg) const
const RegisterBankInfo * getRegBankInfo() const override
bool hasCustomCallingConv() const
CCAssignFn * CCAssignFnForCall(CallingConv::ID CC, bool IsVarArg) const
Selects the correct CCAssignFn for a given CallingConvention value.
This class represents an incoming formal argument to a Function.
Definition Argument.h:32
Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
size_t size() const
Get the array size.
Definition ArrayRef.h:141
bool empty() const
Check if the array is empty.
Definition ArrayRef.h:136
CCState - This class holds information needed while lowering arguments and return values.
MachineFunction & getMachineFunction() const
unsigned getFirstUnallocated(ArrayRef< MCPhysReg > Regs) const
getFirstUnallocated - Return the index of the first unallocated register in the set,...
LLVM_ABI void analyzeMustTailForwardedRegisters(SmallVectorImpl< ForwardedRegister > &Forwards, ArrayRef< MVT > RegParmTypes, CCAssignFn Fn)
Compute the set of registers that need to be preserved and forwarded to any musttail calls.
CallingConv::ID getCallingConv() const
uint64_t getStackSize() const
Returns the size of the currently allocated portion of the stack.
bool isVarArg() const
bool isAllocated(MCRegister Reg) const
isAllocated - Return true if the specified register (or an alias) is allocated.
CCValAssign - Represent assignment of one arg/retval to a location.
LocInfo getLocInfo() const
static CCValAssign getReg(unsigned ValNo, MVT ValVT, MCRegister Reg, MVT LocVT, LocInfo HTP, bool IsCustom=false)
void insertSRetLoads(MachineIRBuilder &MIRBuilder, Type *RetTy, ArrayRef< Register > VRegs, Register DemoteReg, int FI) const
Load the returned value from the stack into virtual registers in VRegs.
bool handleAssignments(ValueHandler &Handler, SmallVectorImpl< ArgInfo > &Args, CCState &CCState, SmallVectorImpl< CCValAssign > &ArgLocs, MachineIRBuilder &MIRBuilder, ArrayRef< Register > ThisReturnRegs={}) const
Use Handler to insert code to handle the argument/return values represented by Args.
bool resultsCompatible(CallLoweringInfo &Info, MachineFunction &MF, SmallVectorImpl< ArgInfo > &InArgs, ValueAssigner &CalleeAssigner, ValueAssigner &CallerAssigner) const
void insertSRetIncomingArgument(const Function &F, SmallVectorImpl< ArgInfo > &SplitArgs, Register &DemoteReg, MachineRegisterInfo &MRI, const DataLayout &DL) const
Insert the hidden sret ArgInfo to the beginning of SplitArgs.
void splitToValueTypes(const ArgInfo &OrigArgInfo, SmallVectorImpl< ArgInfo > &SplitArgs, const DataLayout &DL, CallingConv::ID CallConv, SmallVectorImpl< TypeSize > *Offsets=nullptr) const
Break OrigArgInfo into one or more pieces the calling convention can process, returned in SplitArgs.
bool determineAndHandleAssignments(ValueHandler &Handler, ValueAssigner &Assigner, SmallVectorImpl< ArgInfo > &Args, MachineIRBuilder &MIRBuilder, CallingConv::ID CallConv, bool IsVarArg, ArrayRef< Register > ThisReturnRegs={}) const
Invoke ValueAssigner::assignArg on each of the given Args and then use Handler to move them to the as...
void insertSRetStores(MachineIRBuilder &MIRBuilder, Type *RetTy, ArrayRef< Register > VRegs, Register DemoteReg) const
Store the return value given by VRegs into stack starting at the offset specified in DemoteReg.
bool parametersInCSRMatch(const MachineRegisterInfo &MRI, const uint32_t *CallerPreservedMask, const SmallVectorImpl< CCValAssign > &ArgLocs, const SmallVectorImpl< ArgInfo > &OutVals) const
Check whether parameters to a call that are passed in callee saved registers are the same as from the...
bool determineAssignments(ValueAssigner &Assigner, SmallVectorImpl< ArgInfo > &Args, CCState &CCInfo) const
Analyze the argument list in Args, using Assigner to populate CCInfo.
bool checkReturn(CCState &CCInfo, SmallVectorImpl< BaseArgInfo > &Outs, CCAssignFn *Fn) const
CallLowering(const TargetLowering *TLI)
const TargetLowering * getTLI() const
Getter for generic TargetLowering class.
void setArgFlags(ArgInfo &Arg, unsigned OpIdx, const DataLayout &DL, const FuncInfoTy &FuncInfo) const
void addDefToMIB(MachineRegisterInfo &MRI, MachineInstrBuilder &MIB) const
FormalArgHandler(MachineIRBuilder &MIRBuilder, MachineRegisterInfo &MRI)
FunctionLoweringInfo - This contains information that is global to a function that is used when lower...
Register DemoteRegister
DemoteRegister - if CanLowerReturn is false, DemoteRegister is a vreg allocated to hold a pointer to ...
bool CanLowerReturn
CanLowerReturn - true iff the function's return value can be lowered to registers.
iterator_range< arg_iterator > args()
Definition Function.h:892
CallingConv::ID getCallingConv() const
getCallingConv()/setCallingConv(CC) - These method get and set the calling convention of this functio...
Definition Function.h:272
LLVMContext & getContext() const
getContext - Return a reference to the LLVMContext associated with this function.
Definition Function.cpp:358
bool isVarArg() const
isVarArg - Return true if this function takes a variable number of arguments.
Definition Function.h:229
bool hasExternalWeakLinkage() const
constexpr unsigned getScalarSizeInBits() const
static constexpr LLT scalar(unsigned SizeInBits)
Get a low-level scalar or aggregate "bag of bits".
constexpr uint16_t getNumElements() const
Returns the number of elements in a vector LLT.
constexpr bool isVector() const
static constexpr LLT pointer(unsigned AddressSpace, unsigned SizeInBits)
Get a low-level pointer in the given address space.
constexpr TypeSize getSizeInBits() const
Returns the total size of the type. Must only be called on sized types.
static LLT integer(unsigned SizeInBits)
constexpr TypeSize getSizeInBytes() const
Returns the total size of the type in bytes, i.e.
This is an important class for using LLVM in a threaded context.
Definition LLVMContext.h:68
Machine Value Type.
bool isVector() const
Return true if this is a vector value type.
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
LLVM_ABI int CreateFixedObject(uint64_t Size, int64_t SPOffset, bool IsImmutable, bool isAliased=false)
Create a new object at a fixed location on the stack.
LLVM_ABI int CreateStackObject(uint64_t Size, Align Alignment, bool isSpillSlot, const AllocaInst *Alloca=nullptr, uint8_t ID=0)
Create a new statically sized stack object, returning a nonnegative identifier to represent it.
bool isImmutableObjectIndex(int ObjectIdx) const
Returns true if the specified index corresponds to an immutable object.
void setHasTailCall(bool V=true)
bool hasMustTailInVarArgFunc() const
Returns true if the function is variadic and contains a musttail call.
int64_t getObjectSize(int ObjectIdx) const
Return the size of the specified object.
int64_t getObjectOffset(int ObjectIdx) const
Return the assigned stack offset of the specified object from the incoming stack pointer.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Function & getFunction()
Return the LLVM function that this machine code represents.
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
Register addLiveIn(MCRegister PReg, const TargetRegisterClass *RC)
addLiveIn - Add the specified physical register as a live-in value and create a corresponding virtual...
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
Helper class to build MachineInstr.
MachineInstrBuilder insertInstr(MachineInstrBuilder MIB)
Insert an existing instruction at the insertion point.
MachineInstrBuilder buildZExt(const DstOp &Res, const SrcOp &Op, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_ZEXT Op.
void setInstr(MachineInstr &MI)
Set the insertion point to before MI.
MachineInstrBuilder buildAssertZExt(const DstOp &Res, const SrcOp &Op, unsigned Size)
Build and insert Res = G_ASSERT_ZEXT Op, Size.
MachineInstrBuilder buildPtrAdd(const DstOp &Res, const SrcOp &Op0, const SrcOp &Op1, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_PTR_ADD Op0, Op1.
MachineInstrBuilder buildStore(const SrcOp &Val, const SrcOp &Addr, MachineMemOperand &MMO)
Build and insert G_STORE Val, Addr, MMO.
MachineInstrBuilder buildInstr(unsigned Opcode)
Build and insert <empty> = Opcode <empty>.
MachineInstrBuilder buildPadVectorWithUndefElements(const DstOp &Res, const SrcOp &Op0)
Build and insert a, b, ..., x = G_UNMERGE_VALUES Op0 Res = G_BUILD_VECTOR a, b, .....
MachineInstrBuilder buildFrameIndex(const DstOp &Res, int Idx)
Build and insert Res = G_FRAME_INDEX Idx.
MachineFunction & getMF()
Getter for the function we currently build.
MachineInstrBuilder buildTrunc(const DstOp &Res, const SrcOp &Op, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_TRUNC Op.
const MachineBasicBlock & getMBB() const
Getter for the basic block we currently build.
void setMBB(MachineBasicBlock &MBB)
Set the insertion point to the end of MBB.
MachineInstrBuilder buildInstrNoInsert(unsigned Opcode)
Build but don't insert <empty> = Opcode <empty>.
MachineInstrBuilder buildCopy(const DstOp &Res, const SrcOp &Op)
Build and insert Res = COPY Op.
virtual MachineInstrBuilder buildConstant(const DstOp &Res, const ConstantInt &Val)
Build and insert Res = G_CONSTANT Val.
Register getReg(unsigned Idx) const
Get the register for the operand index.
const MachineInstrBuilder & addReg(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addDef(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a virtual register definition operand.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
const MachineOperand & getOperand(unsigned i) const
LLVM_ABI void setDeactivationSymbol(MachineFunction &MF, Value *DS)
@ MOLoad
The memory access reads data.
@ MOInvariant
The memory access always returns the same value (or traps).
@ MOStore
The memory access writes data.
Register getReg() const
getReg - Returns the register number.
static MachineOperand CreateReg(Register Reg, bool isDef, bool isImp=false, bool isKill=false, bool isDead=false, bool isUndef=false, bool isEarlyClobber=false, unsigned SubReg=0, bool isDebug=false, bool isInternalRead=false, bool isRenamable=false)
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
LLVM_ABI MachineInstr * getVRegDef(Register Reg) const
getVRegDef - Return the machine instr that defines the specified virtual register or null if none is ...
LLT getType(Register Reg) const
Get the low-level type of Reg or LLT{} if Reg is not a generic (target independent) virtual register.
LLVM_ABI Register createGenericVirtualRegister(LLT Ty, StringRef Name="")
Create and return a new generic virtual register with low-level type Ty.
Wrapper class representing virtual and physical registers.
Definition Register.h:20
MCRegister asMCReg() const
Utility to check-convert this value to a MCRegister.
Definition Register.h:107
SMEAttrs is a utility class to parse the SME ACLE attributes on functions.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
CodeGenOptLevel getOptLevel() const
Returns the optimization level: None, Less, Default, or Aggressive.
const Triple & getTargetTriple() const
TargetOptions Options
unsigned GuaranteedTailCallOpt
GuaranteedTailCallOpt - This flag is enabled when -tailcallopt is specified on the commandline.
virtual const RegisterBankInfo * getRegBankInfo() const
If the information for the register banks is available, return it.
virtual const TargetInstrInfo * getInstrInfo() const
Triple - Helper class for working with autoconf configuration names.
Definition Triple.h:47
static constexpr TypeSize getFixed(ScalarTy ExactSize)
Definition TypeSize.h:343
static LLVM_ABI IntegerType * getInt8Ty(LLVMContext &C)
Definition Type.cpp:311
LLVMContext & getContext() const
Return the LLVMContext in which this type was uniqued.
Definition Type.h:130
bool isIntegerTy() const
True if this is an instance of IntegerType.
Definition Type.h:257
unsigned getNumOperands() const
Definition User.h:229
LLVM Value Representation.
Definition Value.h:75
Type * getType() const
All values are typed, get the type of this value.
Definition Value.h:255
@ MO_GOT
MO_GOT - This flag indicates that a symbol operand represents the address of the GOT entry for the sy...
ArrayRef< MCPhysReg > getFPRArgRegs()
ArrayRef< MCPhysReg > getGPRArgRegs()
constexpr char Align[]
Key for Kernel::Arg::Metadata::mAlign.
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
Definition CallingConv.h:24
@ ARM64EC_Thunk_Native
Calling convention used in the ARM64EC ABI to implement calls between ARM64 code and thunks.
@ Swift
Calling convention for Swift.
Definition CallingConv.h:69
@ PreserveMost
Used for runtime calls that preserves most registers.
Definition CallingConv.h:63
@ PreserveAll
Used for runtime calls that preserves (almost) all registers.
Definition CallingConv.h:66
@ Fast
Attempts to make calls as fast as possible (e.g.
Definition CallingConv.h:41
@ PreserveNone
Used for runtime calls that preserves none general registers.
Definition CallingConv.h:90
@ Tail
Attemps to make calls as fast as possible while guaranteeing that tail call optimization can always b...
Definition CallingConv.h:76
@ SwiftTail
This follows the Swift calling convention in how arguments are passed but guarantees tail calls will ...
Definition CallingConv.h:87
@ ARM64EC_Thunk_X64
Calling convention used in the ARM64EC ABI to implement calls between x64 code and thunks.
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
std::optional< Function * > getAttachedARCFunction(const CallBase *CB)
This function returns operand bundle clang_arc_attachedcall's argument, which is the address of the A...
Definition ObjCARCUtil.h:43
bool attachedCallOpBundleNeedsMarker(const CallBase *CB)
This function determines whether the clang_arc_attachedcall should be emitted with or without the mar...
Definition ObjCARCUtil.h:58
bool hasAttachedCallOpBundle(const CallBase *CB)
Definition ObjCARCUtil.h:29
This is an optimization pass for GlobalISel generic memory operations.
@ Offset
Definition DWP.cpp:557
LLVM_ABI Register constrainOperandRegClass(const MachineFunction &MF, const TargetRegisterInfo &TRI, MachineRegisterInfo &MRI, const TargetInstrInfo &TII, const RegisterBankInfo &RBI, MachineInstr &InsertPt, const TargetRegisterClass &RegClass, MachineOperand &RegMO)
Constrain the Register operand OpIdx, so that it is now constrained to the TargetRegisterClass passed...
Definition Utils.cpp:57
@ Implicit
Not emitted register (e.g. carry, or temporary result).
void ComputeValueVTs(const TargetLowering &TLI, const DataLayout &DL, Type *Ty, SmallVectorImpl< EVT > &ValueVTs, SmallVectorImpl< EVT > *MemVTs=nullptr, SmallVectorImpl< TypeSize > *Offsets=nullptr, TypeSize StartingOffset=TypeSize::getZero())
ComputeValueVTs - Given an LLVM IR type, compute a sequence of EVTs that represent all the individual...
Definition Analysis.cpp:119
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:643
bool CCAssignFn(unsigned ValNo, MVT ValVT, MVT LocVT, CCValAssign::LocInfo LocInfo, ISD::ArgFlagsTy ArgFlags, Type *OrigTy, CCState &State)
CCAssignFn - This function assigns a location for Val, updating State to reflect the change.
void append_range(Container &C, Range &&R)
Wrapper function to append range R to container C.
Definition STLExtras.h:2207
unsigned getBLRCallOpcode(const MachineFunction &MF)
Return opcode to be used for indirect calls.
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1745
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition Debug.cpp:209
constexpr uint64_t alignTo(uint64_t Size, Align A)
Returns a multiple of A needed to store Size bytes.
Definition Alignment.h:144
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
@ Success
The lock was released successfully.
DWARFExpression::Operation Op
LLVM_ABI bool isAssertMI(const MachineInstr &MI)
Returns true if the instruction MI is one of the assert instructions.
Definition Utils.cpp:1991
LLVM_ABI LLT getLLTForType(Type &Ty, const DataLayout &DL)
Construct a low-level type based on an LLVM type.
LLVM_ABI CGPassBuilderOption getCGPassBuilderOption()
LLVM_ABI Align inferAlignFromPtrInfo(MachineFunction &MF, const MachinePointerInfo &MPO)
Definition Utils.cpp:841
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition BitVector.h:876
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition Alignment.h:39
std::optional< bool > EnableGlobalISelOption
SmallVector< Register, 4 > Regs
SmallVector< ISD::ArgFlagsTy, 4 > Flags
Base class for ValueHandlers used for arguments coming into the current function, or for return value...
void assignValueToReg(Register ValVReg, Register PhysReg, const CCValAssign &VA, ISD::ArgFlagsTy Flags={}) override
Provides a default implementation for argument handling.
Base class for ValueHandlers used for arguments passed to a function call, or for return values.
virtual LLT getStackValueStoreType(const DataLayout &DL, const CCValAssign &VA, ISD::ArgFlagsTy Flags) const
Return the in-memory size to write for the argument at VA.
Extended Value Type.
Definition ValueTypes.h:35
LLVM_ABI Type * getTypeForEVT(LLVMContext &Context) const
This method returns an LLVM type corresponding to the specified EVT.
Describes a register that needs to be forwarded from the prologue to a musttail call.
This class contains a discriminated union of information about pointers in memory operands,...
static LLVM_ABI MachinePointerInfo getStack(MachineFunction &MF, int64_t Offset, uint8_t ID=0)
Stack pointer relative access.
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.