LLVM 23.0.0git
TargetLoweringBase.cpp
Go to the documentation of this file.
1//===- TargetLoweringBase.cpp - Implement the TargetLoweringBase class ----===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This implements the TargetLoweringBase class.
10//
11//===----------------------------------------------------------------------===//
12
13#include "llvm/ADT/BitVector.h"
14#include "llvm/ADT/DenseMap.h"
15#include "llvm/ADT/STLExtras.h"
18#include "llvm/ADT/StringRef.h"
19#include "llvm/ADT/Twine.h"
20#include "llvm/Analysis/Loads.h"
39#include "llvm/IR/Attributes.h"
40#include "llvm/IR/CallingConv.h"
41#include "llvm/IR/DataLayout.h"
43#include "llvm/IR/Function.h"
44#include "llvm/IR/GlobalValue.h"
46#include "llvm/IR/IRBuilder.h"
47#include "llvm/IR/Module.h"
48#include "llvm/IR/Type.h"
58#include <algorithm>
59#include <cassert>
60#include <cstdint>
61#include <cstring>
62#include <string>
63#include <tuple>
64#include <utility>
65
66using namespace llvm;
67
69 "jump-is-expensive", cl::init(false),
70 cl::desc("Do not create extra branches to split comparison logic."),
72
74 ("min-jump-table-entries", cl::init(4), cl::Hidden,
75 cl::desc("Set minimum number of entries to use a jump table."));
76
78 ("max-jump-table-size", cl::init(UINT_MAX), cl::Hidden,
79 cl::desc("Set maximum size of jump tables."));
80
81/// Minimum jump table density for normal functions.
83 JumpTableDensity("jump-table-density", cl::init(10), cl::Hidden,
84 cl::desc("Minimum density for building a jump table in "
85 "a normal function"));
86
87/// Minimum jump table density for -Os or -Oz functions.
89 "optsize-jump-table-density", cl::init(40), cl::Hidden,
90 cl::desc("Minimum density for building a jump table in "
91 "an optsize function"));
92
94 "min-bit-test-cmps", cl::init(2), cl::Hidden,
95 cl::desc("Set minimum of largest number of comparisons "
96 "to use bit test for switch."));
97
99 "max-store-memset", cl::init(0), cl::Hidden,
100 cl::desc("Override target's MaxStoresPerMemset and "
101 "MaxStoresPerMemsetOptSize. "
102 "Set to 0 to use the target default."));
103
105 "max-store-memcpy", cl::init(0), cl::Hidden,
106 cl::desc("Override target's MaxStoresPerMemcpy and "
107 "MaxStoresPerMemcpyOptSize. "
108 "Set to 0 to use the target default."));
109
111 "max-store-memmove", cl::init(0), cl::Hidden,
112 cl::desc("Override target's MaxStoresPerMemmove and "
113 "MaxStoresPerMemmoveOptSize. "
114 "Set to 0 to use the target default."));
115
116// FIXME: This option is only to test if the strict fp operation processed
117// correctly by preventing mutating strict fp operation to normal fp operation
118// during development. When the backend supports strict float operation, this
119// option will be meaningless.
120static cl::opt<bool> DisableStrictNodeMutation("disable-strictnode-mutation",
121 cl::desc("Don't mutate strict-float node to a legalize node"),
122 cl::init(false), cl::Hidden);
123
124LLVM_ABI RTLIB::Libcall RTLIB::getSHL(EVT VT) {
125 if (VT == MVT::i16)
126 return RTLIB::SHL_I16;
127 if (VT == MVT::i32)
128 return RTLIB::SHL_I32;
129 if (VT == MVT::i64)
130 return RTLIB::SHL_I64;
131 if (VT == MVT::i128)
132 return RTLIB::SHL_I128;
133
134 return RTLIB::UNKNOWN_LIBCALL;
135}
136
137LLVM_ABI RTLIB::Libcall RTLIB::getSRL(EVT VT) {
138 if (VT == MVT::i16)
139 return RTLIB::SRL_I16;
140 if (VT == MVT::i32)
141 return RTLIB::SRL_I32;
142 if (VT == MVT::i64)
143 return RTLIB::SRL_I64;
144 if (VT == MVT::i128)
145 return RTLIB::SRL_I128;
146
147 return RTLIB::UNKNOWN_LIBCALL;
148}
149
150LLVM_ABI RTLIB::Libcall RTLIB::getSRA(EVT VT) {
151 if (VT == MVT::i16)
152 return RTLIB::SRA_I16;
153 if (VT == MVT::i32)
154 return RTLIB::SRA_I32;
155 if (VT == MVT::i64)
156 return RTLIB::SRA_I64;
157 if (VT == MVT::i128)
158 return RTLIB::SRA_I128;
159
160 return RTLIB::UNKNOWN_LIBCALL;
161}
162
163LLVM_ABI RTLIB::Libcall RTLIB::getMUL(EVT VT) {
164 if (VT == MVT::i16)
165 return RTLIB::MUL_I16;
166 if (VT == MVT::i32)
167 return RTLIB::MUL_I32;
168 if (VT == MVT::i64)
169 return RTLIB::MUL_I64;
170 if (VT == MVT::i128)
171 return RTLIB::MUL_I128;
172 return RTLIB::UNKNOWN_LIBCALL;
173}
174
175LLVM_ABI RTLIB::Libcall RTLIB::getMULO(EVT VT) {
176 if (VT == MVT::i32)
177 return RTLIB::MULO_I32;
178 if (VT == MVT::i64)
179 return RTLIB::MULO_I64;
180 if (VT == MVT::i128)
181 return RTLIB::MULO_I128;
182 return RTLIB::UNKNOWN_LIBCALL;
183}
184
185LLVM_ABI RTLIB::Libcall RTLIB::getSDIV(EVT VT) {
186 if (VT == MVT::i16)
187 return RTLIB::SDIV_I16;
188 if (VT == MVT::i32)
189 return RTLIB::SDIV_I32;
190 if (VT == MVT::i64)
191 return RTLIB::SDIV_I64;
192 if (VT == MVT::i128)
193 return RTLIB::SDIV_I128;
194 return RTLIB::UNKNOWN_LIBCALL;
195}
196
197LLVM_ABI RTLIB::Libcall RTLIB::getUDIV(EVT VT) {
198 if (VT == MVT::i16)
199 return RTLIB::UDIV_I16;
200 if (VT == MVT::i32)
201 return RTLIB::UDIV_I32;
202 if (VT == MVT::i64)
203 return RTLIB::UDIV_I64;
204 if (VT == MVT::i128)
205 return RTLIB::UDIV_I128;
206 return RTLIB::UNKNOWN_LIBCALL;
207}
208
209LLVM_ABI RTLIB::Libcall RTLIB::getSREM(EVT VT) {
210 if (VT == MVT::i16)
211 return RTLIB::SREM_I16;
212 if (VT == MVT::i32)
213 return RTLIB::SREM_I32;
214 if (VT == MVT::i64)
215 return RTLIB::SREM_I64;
216 if (VT == MVT::i128)
217 return RTLIB::SREM_I128;
218 return RTLIB::UNKNOWN_LIBCALL;
219}
220
221LLVM_ABI RTLIB::Libcall RTLIB::getUREM(EVT VT) {
222 if (VT == MVT::i16)
223 return RTLIB::UREM_I16;
224 if (VT == MVT::i32)
225 return RTLIB::UREM_I32;
226 if (VT == MVT::i64)
227 return RTLIB::UREM_I64;
228 if (VT == MVT::i128)
229 return RTLIB::UREM_I128;
230 return RTLIB::UNKNOWN_LIBCALL;
231}
232
233LLVM_ABI RTLIB::Libcall RTLIB::getCTPOP(EVT VT) {
234 if (VT == MVT::i32)
235 return RTLIB::CTPOP_I32;
236 if (VT == MVT::i64)
237 return RTLIB::CTPOP_I64;
238 if (VT == MVT::i128)
239 return RTLIB::CTPOP_I128;
240 return RTLIB::UNKNOWN_LIBCALL;
241}
242
243/// GetFPLibCall - Helper to return the right libcall for the given floating
244/// point type, or UNKNOWN_LIBCALL if there is none.
245RTLIB::Libcall RTLIB::getFPLibCall(EVT VT,
246 RTLIB::Libcall Call_F32,
247 RTLIB::Libcall Call_F64,
248 RTLIB::Libcall Call_F80,
249 RTLIB::Libcall Call_F128,
250 RTLIB::Libcall Call_PPCF128) {
251 return
252 VT == MVT::f32 ? Call_F32 :
253 VT == MVT::f64 ? Call_F64 :
254 VT == MVT::f80 ? Call_F80 :
255 VT == MVT::f128 ? Call_F128 :
256 VT == MVT::ppcf128 ? Call_PPCF128 :
257 RTLIB::UNKNOWN_LIBCALL;
258}
259
260/// getFPEXT - Return the FPEXT_*_* value for the given types, or
261/// UNKNOWN_LIBCALL if there is none.
262RTLIB::Libcall RTLIB::getFPEXT(EVT OpVT, EVT RetVT) {
263 if (OpVT == MVT::f16) {
264 if (RetVT == MVT::f32)
265 return FPEXT_F16_F32;
266 if (RetVT == MVT::f64)
267 return FPEXT_F16_F64;
268 if (RetVT == MVT::f80)
269 return FPEXT_F16_F80;
270 if (RetVT == MVT::f128)
271 return FPEXT_F16_F128;
272 } else if (OpVT == MVT::f32) {
273 if (RetVT == MVT::f64)
274 return FPEXT_F32_F64;
275 if (RetVT == MVT::f128)
276 return FPEXT_F32_F128;
277 if (RetVT == MVT::ppcf128)
278 return FPEXT_F32_PPCF128;
279 } else if (OpVT == MVT::f64) {
280 if (RetVT == MVT::f128)
281 return FPEXT_F64_F128;
282 else if (RetVT == MVT::ppcf128)
283 return FPEXT_F64_PPCF128;
284 } else if (OpVT == MVT::f80) {
285 if (RetVT == MVT::f128)
286 return FPEXT_F80_F128;
287 } else if (OpVT == MVT::bf16) {
288 if (RetVT == MVT::f32)
289 return FPEXT_BF16_F32;
290 }
291
292 return UNKNOWN_LIBCALL;
293}
294
295/// getFPROUND - Return the FPROUND_*_* value for the given types, or
296/// UNKNOWN_LIBCALL if there is none.
297RTLIB::Libcall RTLIB::getFPROUND(EVT OpVT, EVT RetVT) {
298 if (RetVT == MVT::f16) {
299 if (OpVT == MVT::f32)
300 return FPROUND_F32_F16;
301 if (OpVT == MVT::f64)
302 return FPROUND_F64_F16;
303 if (OpVT == MVT::f80)
304 return FPROUND_F80_F16;
305 if (OpVT == MVT::f128)
306 return FPROUND_F128_F16;
307 if (OpVT == MVT::ppcf128)
308 return FPROUND_PPCF128_F16;
309 } else if (RetVT == MVT::bf16) {
310 if (OpVT == MVT::f32)
311 return FPROUND_F32_BF16;
312 if (OpVT == MVT::f64)
313 return FPROUND_F64_BF16;
314 if (OpVT == MVT::f80)
315 return FPROUND_F80_BF16;
316 if (OpVT == MVT::f128)
317 return FPROUND_F128_BF16;
318 } else if (RetVT == MVT::f32) {
319 if (OpVT == MVT::f64)
320 return FPROUND_F64_F32;
321 if (OpVT == MVT::f80)
322 return FPROUND_F80_F32;
323 if (OpVT == MVT::f128)
324 return FPROUND_F128_F32;
325 if (OpVT == MVT::ppcf128)
326 return FPROUND_PPCF128_F32;
327 } else if (RetVT == MVT::f64) {
328 if (OpVT == MVT::f80)
329 return FPROUND_F80_F64;
330 if (OpVT == MVT::f128)
331 return FPROUND_F128_F64;
332 if (OpVT == MVT::ppcf128)
333 return FPROUND_PPCF128_F64;
334 } else if (RetVT == MVT::f80) {
335 if (OpVT == MVT::f128)
336 return FPROUND_F128_F80;
337 }
338
339 return UNKNOWN_LIBCALL;
340}
341
342/// getFPTOSINT - Return the FPTOSINT_*_* value for the given types, or
343/// UNKNOWN_LIBCALL if there is none.
344RTLIB::Libcall RTLIB::getFPTOSINT(EVT OpVT, EVT RetVT) {
345 if (OpVT == MVT::f16) {
346 if (RetVT == MVT::i32)
347 return FPTOSINT_F16_I32;
348 if (RetVT == MVT::i64)
349 return FPTOSINT_F16_I64;
350 if (RetVT == MVT::i128)
351 return FPTOSINT_F16_I128;
352 } else if (OpVT == MVT::f32) {
353 if (RetVT == MVT::i32)
354 return FPTOSINT_F32_I32;
355 if (RetVT == MVT::i64)
356 return FPTOSINT_F32_I64;
357 if (RetVT == MVT::i128)
358 return FPTOSINT_F32_I128;
359 } else if (OpVT == MVT::f64) {
360 if (RetVT == MVT::i32)
361 return FPTOSINT_F64_I32;
362 if (RetVT == MVT::i64)
363 return FPTOSINT_F64_I64;
364 if (RetVT == MVT::i128)
365 return FPTOSINT_F64_I128;
366 } else if (OpVT == MVT::f80) {
367 if (RetVT == MVT::i32)
368 return FPTOSINT_F80_I32;
369 if (RetVT == MVT::i64)
370 return FPTOSINT_F80_I64;
371 if (RetVT == MVT::i128)
372 return FPTOSINT_F80_I128;
373 } else if (OpVT == MVT::f128) {
374 if (RetVT == MVT::i32)
375 return FPTOSINT_F128_I32;
376 if (RetVT == MVT::i64)
377 return FPTOSINT_F128_I64;
378 if (RetVT == MVT::i128)
379 return FPTOSINT_F128_I128;
380 } else if (OpVT == MVT::ppcf128) {
381 if (RetVT == MVT::i32)
382 return FPTOSINT_PPCF128_I32;
383 if (RetVT == MVT::i64)
384 return FPTOSINT_PPCF128_I64;
385 if (RetVT == MVT::i128)
386 return FPTOSINT_PPCF128_I128;
387 }
388 return UNKNOWN_LIBCALL;
389}
390
391/// getFPTOUINT - Return the FPTOUINT_*_* value for the given types, or
392/// UNKNOWN_LIBCALL if there is none.
393RTLIB::Libcall RTLIB::getFPTOUINT(EVT OpVT, EVT RetVT) {
394 if (OpVT == MVT::f16) {
395 if (RetVT == MVT::i32)
396 return FPTOUINT_F16_I32;
397 if (RetVT == MVT::i64)
398 return FPTOUINT_F16_I64;
399 if (RetVT == MVT::i128)
400 return FPTOUINT_F16_I128;
401 } else if (OpVT == MVT::f32) {
402 if (RetVT == MVT::i32)
403 return FPTOUINT_F32_I32;
404 if (RetVT == MVT::i64)
405 return FPTOUINT_F32_I64;
406 if (RetVT == MVT::i128)
407 return FPTOUINT_F32_I128;
408 } else if (OpVT == MVT::f64) {
409 if (RetVT == MVT::i32)
410 return FPTOUINT_F64_I32;
411 if (RetVT == MVT::i64)
412 return FPTOUINT_F64_I64;
413 if (RetVT == MVT::i128)
414 return FPTOUINT_F64_I128;
415 } else if (OpVT == MVT::f80) {
416 if (RetVT == MVT::i32)
417 return FPTOUINT_F80_I32;
418 if (RetVT == MVT::i64)
419 return FPTOUINT_F80_I64;
420 if (RetVT == MVT::i128)
421 return FPTOUINT_F80_I128;
422 } else if (OpVT == MVT::f128) {
423 if (RetVT == MVT::i32)
424 return FPTOUINT_F128_I32;
425 if (RetVT == MVT::i64)
426 return FPTOUINT_F128_I64;
427 if (RetVT == MVT::i128)
428 return FPTOUINT_F128_I128;
429 } else if (OpVT == MVT::ppcf128) {
430 if (RetVT == MVT::i32)
431 return FPTOUINT_PPCF128_I32;
432 if (RetVT == MVT::i64)
433 return FPTOUINT_PPCF128_I64;
434 if (RetVT == MVT::i128)
435 return FPTOUINT_PPCF128_I128;
436 }
437 return UNKNOWN_LIBCALL;
438}
439
440/// getSINTTOFP - Return the SINTTOFP_*_* value for the given types, or
441/// UNKNOWN_LIBCALL if there is none.
442RTLIB::Libcall RTLIB::getSINTTOFP(EVT OpVT, EVT RetVT) {
443 if (OpVT == MVT::i32) {
444 if (RetVT == MVT::f16)
445 return SINTTOFP_I32_F16;
446 if (RetVT == MVT::f32)
447 return SINTTOFP_I32_F32;
448 if (RetVT == MVT::f64)
449 return SINTTOFP_I32_F64;
450 if (RetVT == MVT::f80)
451 return SINTTOFP_I32_F80;
452 if (RetVT == MVT::f128)
453 return SINTTOFP_I32_F128;
454 if (RetVT == MVT::ppcf128)
455 return SINTTOFP_I32_PPCF128;
456 } else if (OpVT == MVT::i64) {
457 if (RetVT == MVT::bf16)
458 return SINTTOFP_I64_BF16;
459 if (RetVT == MVT::f16)
460 return SINTTOFP_I64_F16;
461 if (RetVT == MVT::f32)
462 return SINTTOFP_I64_F32;
463 if (RetVT == MVT::f64)
464 return SINTTOFP_I64_F64;
465 if (RetVT == MVT::f80)
466 return SINTTOFP_I64_F80;
467 if (RetVT == MVT::f128)
468 return SINTTOFP_I64_F128;
469 if (RetVT == MVT::ppcf128)
470 return SINTTOFP_I64_PPCF128;
471 } else if (OpVT == MVT::i128) {
472 if (RetVT == MVT::f16)
473 return SINTTOFP_I128_F16;
474 if (RetVT == MVT::f32)
475 return SINTTOFP_I128_F32;
476 if (RetVT == MVT::f64)
477 return SINTTOFP_I128_F64;
478 if (RetVT == MVT::f80)
479 return SINTTOFP_I128_F80;
480 if (RetVT == MVT::f128)
481 return SINTTOFP_I128_F128;
482 if (RetVT == MVT::ppcf128)
483 return SINTTOFP_I128_PPCF128;
484 }
485 return UNKNOWN_LIBCALL;
486}
487
488/// getUINTTOFP - Return the UINTTOFP_*_* value for the given types, or
489/// UNKNOWN_LIBCALL if there is none.
490RTLIB::Libcall RTLIB::getUINTTOFP(EVT OpVT, EVT RetVT) {
491 if (OpVT == MVT::i32) {
492 if (RetVT == MVT::f16)
493 return UINTTOFP_I32_F16;
494 if (RetVT == MVT::f32)
495 return UINTTOFP_I32_F32;
496 if (RetVT == MVT::f64)
497 return UINTTOFP_I32_F64;
498 if (RetVT == MVT::f80)
499 return UINTTOFP_I32_F80;
500 if (RetVT == MVT::f128)
501 return UINTTOFP_I32_F128;
502 if (RetVT == MVT::ppcf128)
503 return UINTTOFP_I32_PPCF128;
504 } else if (OpVT == MVT::i64) {
505 if (RetVT == MVT::bf16)
506 return UINTTOFP_I64_BF16;
507 if (RetVT == MVT::f16)
508 return UINTTOFP_I64_F16;
509 if (RetVT == MVT::f32)
510 return UINTTOFP_I64_F32;
511 if (RetVT == MVT::f64)
512 return UINTTOFP_I64_F64;
513 if (RetVT == MVT::f80)
514 return UINTTOFP_I64_F80;
515 if (RetVT == MVT::f128)
516 return UINTTOFP_I64_F128;
517 if (RetVT == MVT::ppcf128)
518 return UINTTOFP_I64_PPCF128;
519 } else if (OpVT == MVT::i128) {
520 if (RetVT == MVT::f16)
521 return UINTTOFP_I128_F16;
522 if (RetVT == MVT::f32)
523 return UINTTOFP_I128_F32;
524 if (RetVT == MVT::f64)
525 return UINTTOFP_I128_F64;
526 if (RetVT == MVT::f80)
527 return UINTTOFP_I128_F80;
528 if (RetVT == MVT::f128)
529 return UINTTOFP_I128_F128;
530 if (RetVT == MVT::ppcf128)
531 return UINTTOFP_I128_PPCF128;
532 }
533 return UNKNOWN_LIBCALL;
534}
535
536RTLIB::Libcall RTLIB::getPOWI(EVT RetVT) {
537 return getFPLibCall(RetVT, POWI_F32, POWI_F64, POWI_F80, POWI_F128,
538 POWI_PPCF128);
539}
540
541RTLIB::Libcall RTLIB::getPOW(EVT RetVT) {
542 // TODO: Tablegen should generate this function
543 if (RetVT.isVector()) {
544 if (!RetVT.isSimple())
545 return RTLIB::UNKNOWN_LIBCALL;
546 switch (RetVT.getSimpleVT().SimpleTy) {
547 case MVT::v4f32:
548 return RTLIB::POW_V4F32;
549 case MVT::v2f64:
550 return RTLIB::POW_V2F64;
551 case MVT::nxv4f32:
552 return RTLIB::POW_NXV4F32;
553 case MVT::nxv2f64:
554 return RTLIB::POW_NXV2F64;
555 default:
556 return RTLIB::UNKNOWN_LIBCALL;
557 }
558 }
559
560 return getFPLibCall(RetVT, POW_F32, POW_F64, POW_F80, POW_F128, POW_PPCF128);
561}
562
563RTLIB::Libcall RTLIB::getLDEXP(EVT RetVT) {
564 return getFPLibCall(RetVT, LDEXP_F32, LDEXP_F64, LDEXP_F80, LDEXP_F128,
565 LDEXP_PPCF128);
566}
567
568RTLIB::Libcall RTLIB::getFREXP(EVT RetVT) {
569 return getFPLibCall(RetVT, FREXP_F32, FREXP_F64, FREXP_F80, FREXP_F128,
570 FREXP_PPCF128);
571}
572
573RTLIB::Libcall RTLIB::getSIN(EVT RetVT) {
574 return getFPLibCall(RetVT, SIN_F32, SIN_F64, SIN_F80, SIN_F128, SIN_PPCF128);
575}
576
577RTLIB::Libcall RTLIB::getCOS(EVT RetVT) {
578 return getFPLibCall(RetVT, COS_F32, COS_F64, COS_F80, COS_F128, COS_PPCF128);
579}
580
581RTLIB::Libcall RTLIB::getSINCOS(EVT RetVT) {
582 // TODO: Tablegen should generate this function
583 if (RetVT.isVector()) {
584 if (!RetVT.isSimple())
585 return RTLIB::UNKNOWN_LIBCALL;
586 switch (RetVT.getSimpleVT().SimpleTy) {
587 case MVT::v4f32:
588 return RTLIB::SINCOS_V4F32;
589 case MVT::v8f32:
590 return RTLIB::SINCOS_V8F32;
591 case MVT::v16f32:
592 return RTLIB::SINCOS_V16F32;
593 case MVT::v2f64:
594 return RTLIB::SINCOS_V2F64;
595 case MVT::v4f64:
596 return RTLIB::SINCOS_V4F64;
597 case MVT::v8f64:
598 return RTLIB::SINCOS_V8F64;
599 case MVT::nxv4f32:
600 return RTLIB::SINCOS_NXV4F32;
601 case MVT::nxv2f64:
602 return RTLIB::SINCOS_NXV2F64;
603 default:
604 return RTLIB::UNKNOWN_LIBCALL;
605 }
606 }
607
608 return getFPLibCall(RetVT, SINCOS_F32, SINCOS_F64, SINCOS_F80, SINCOS_F128,
609 SINCOS_PPCF128);
610}
611
612RTLIB::Libcall RTLIB::getSINCOSPI(EVT RetVT) {
613 // TODO: Tablegen should generate this function
614 if (RetVT.isVector()) {
615 if (!RetVT.isSimple())
616 return RTLIB::UNKNOWN_LIBCALL;
617 switch (RetVT.getSimpleVT().SimpleTy) {
618 case MVT::v4f32:
619 return RTLIB::SINCOSPI_V4F32;
620 case MVT::v2f64:
621 return RTLIB::SINCOSPI_V2F64;
622 case MVT::nxv4f32:
623 return RTLIB::SINCOSPI_NXV4F32;
624 case MVT::nxv2f64:
625 return RTLIB::SINCOSPI_NXV2F64;
626 default:
627 return RTLIB::UNKNOWN_LIBCALL;
628 }
629 }
630
631 return getFPLibCall(RetVT, SINCOSPI_F32, SINCOSPI_F64, SINCOSPI_F80,
632 SINCOSPI_F128, SINCOSPI_PPCF128);
633}
634
635RTLIB::Libcall RTLIB::getSINCOS_STRET(EVT RetVT) {
636 return getFPLibCall(RetVT, SINCOS_STRET_F32, SINCOS_STRET_F64,
637 UNKNOWN_LIBCALL, UNKNOWN_LIBCALL, UNKNOWN_LIBCALL);
638}
639
640RTLIB::Libcall RTLIB::getREM(EVT VT) {
641 // TODO: Tablegen should generate this function
642 if (VT.isVector()) {
643 if (!VT.isSimple())
644 return RTLIB::UNKNOWN_LIBCALL;
645 switch (VT.getSimpleVT().SimpleTy) {
646 case MVT::v4f32:
647 return RTLIB::REM_V4F32;
648 case MVT::v2f64:
649 return RTLIB::REM_V2F64;
650 case MVT::nxv4f32:
651 return RTLIB::REM_NXV4F32;
652 case MVT::nxv2f64:
653 return RTLIB::REM_NXV2F64;
654 default:
655 return RTLIB::UNKNOWN_LIBCALL;
656 }
657 }
658
659 return getFPLibCall(VT, REM_F32, REM_F64, REM_F80, REM_F128, REM_PPCF128);
660}
661
662RTLIB::Libcall RTLIB::getCBRT(EVT VT) {
663 // TODO: Tablegen should generate this function
664 if (VT.isVector()) {
665 if (!VT.isSimple())
666 return RTLIB::UNKNOWN_LIBCALL;
667 switch (VT.getSimpleVT().SimpleTy) {
668 case MVT::v4f32:
669 return RTLIB::CBRT_V4F32;
670 case MVT::v2f64:
671 return RTLIB::CBRT_V2F64;
672 case MVT::nxv4f32:
673 return RTLIB::CBRT_NXV4F32;
674 case MVT::nxv2f64:
675 return RTLIB::CBRT_NXV2F64;
676 default:
677 return RTLIB::UNKNOWN_LIBCALL;
678 }
679 }
680
681 return getFPLibCall(VT, CBRT_F32, CBRT_F64, CBRT_F80, CBRT_F128,
682 CBRT_PPCF128);
683}
684
685RTLIB::Libcall RTLIB::getMODF(EVT RetVT) {
686 // TODO: Tablegen should generate this function
687 if (RetVT.isVector()) {
688 if (!RetVT.isSimple())
689 return RTLIB::UNKNOWN_LIBCALL;
690 switch (RetVT.getSimpleVT().SimpleTy) {
691 case MVT::v4f32:
692 return RTLIB::MODF_V4F32;
693 case MVT::v2f64:
694 return RTLIB::MODF_V2F64;
695 case MVT::nxv4f32:
696 return RTLIB::MODF_NXV4F32;
697 case MVT::nxv2f64:
698 return RTLIB::MODF_NXV2F64;
699 default:
700 return RTLIB::UNKNOWN_LIBCALL;
701 }
702 }
703
704 return getFPLibCall(RetVT, MODF_F32, MODF_F64, MODF_F80, MODF_F128,
705 MODF_PPCF128);
706}
707
708RTLIB::Libcall RTLIB::getLROUND(EVT VT) {
709 if (VT == MVT::f32)
710 return RTLIB::LROUND_F32;
711 if (VT == MVT::f64)
712 return RTLIB::LROUND_F64;
713 if (VT == MVT::f80)
714 return RTLIB::LROUND_F80;
715 if (VT == MVT::f128)
716 return RTLIB::LROUND_F128;
717 if (VT == MVT::ppcf128)
718 return RTLIB::LROUND_PPCF128;
719
720 return RTLIB::UNKNOWN_LIBCALL;
721}
722
723RTLIB::Libcall RTLIB::getLLROUND(EVT VT) {
724 if (VT == MVT::f32)
725 return RTLIB::LLROUND_F32;
726 if (VT == MVT::f64)
727 return RTLIB::LLROUND_F64;
728 if (VT == MVT::f80)
729 return RTLIB::LLROUND_F80;
730 if (VT == MVT::f128)
731 return RTLIB::LLROUND_F128;
732 if (VT == MVT::ppcf128)
733 return RTLIB::LLROUND_PPCF128;
734
735 return RTLIB::UNKNOWN_LIBCALL;
736}
737
738RTLIB::Libcall RTLIB::getLRINT(EVT VT) {
739 if (VT == MVT::f32)
740 return RTLIB::LRINT_F32;
741 if (VT == MVT::f64)
742 return RTLIB::LRINT_F64;
743 if (VT == MVT::f80)
744 return RTLIB::LRINT_F80;
745 if (VT == MVT::f128)
746 return RTLIB::LRINT_F128;
747 if (VT == MVT::ppcf128)
748 return RTLIB::LRINT_PPCF128;
749 return RTLIB::UNKNOWN_LIBCALL;
750}
751
752RTLIB::Libcall RTLIB::getLLRINT(EVT VT) {
753 if (VT == MVT::f32)
754 return RTLIB::LLRINT_F32;
755 if (VT == MVT::f64)
756 return RTLIB::LLRINT_F64;
757 if (VT == MVT::f80)
758 return RTLIB::LLRINT_F80;
759 if (VT == MVT::f128)
760 return RTLIB::LLRINT_F128;
761 if (VT == MVT::ppcf128)
762 return RTLIB::LLRINT_PPCF128;
763 return RTLIB::UNKNOWN_LIBCALL;
764}
765
766RTLIB::Libcall RTLIB::getOutlineAtomicHelper(const Libcall (&LC)[5][4],
767 AtomicOrdering Order,
768 uint64_t MemSize) {
769 unsigned ModeN, ModelN;
770 switch (MemSize) {
771 case 1:
772 ModeN = 0;
773 break;
774 case 2:
775 ModeN = 1;
776 break;
777 case 4:
778 ModeN = 2;
779 break;
780 case 8:
781 ModeN = 3;
782 break;
783 case 16:
784 ModeN = 4;
785 break;
786 default:
787 return RTLIB::UNKNOWN_LIBCALL;
788 }
789
790 switch (Order) {
792 ModelN = 0;
793 break;
795 ModelN = 1;
796 break;
798 ModelN = 2;
799 break;
802 ModelN = 3;
803 break;
804 default:
805 return UNKNOWN_LIBCALL;
806 }
807
808 return LC[ModeN][ModelN];
809}
810
811RTLIB::Libcall RTLIB::getOUTLINE_ATOMIC(unsigned Opc, AtomicOrdering Order,
812 MVT VT) {
813 if (!VT.isScalarInteger())
814 return UNKNOWN_LIBCALL;
815 uint64_t MemSize = VT.getScalarSizeInBits() / 8;
816
817#define LCALLS(A, B) \
818 { A##B##_RELAX, A##B##_ACQ, A##B##_REL, A##B##_ACQ_REL }
819#define LCALL5(A) \
820 LCALLS(A, 1), LCALLS(A, 2), LCALLS(A, 4), LCALLS(A, 8), LCALLS(A, 16)
821 switch (Opc) {
823 const Libcall LC[5][4] = {LCALL5(OUTLINE_ATOMIC_CAS)};
824 return getOutlineAtomicHelper(LC, Order, MemSize);
825 }
826 case ISD::ATOMIC_SWAP: {
827 const Libcall LC[5][4] = {LCALL5(OUTLINE_ATOMIC_SWP)};
828 return getOutlineAtomicHelper(LC, Order, MemSize);
829 }
831 const Libcall LC[5][4] = {LCALL5(OUTLINE_ATOMIC_LDADD)};
832 return getOutlineAtomicHelper(LC, Order, MemSize);
833 }
834 case ISD::ATOMIC_LOAD_OR: {
835 const Libcall LC[5][4] = {LCALL5(OUTLINE_ATOMIC_LDSET)};
836 return getOutlineAtomicHelper(LC, Order, MemSize);
837 }
839 const Libcall LC[5][4] = {LCALL5(OUTLINE_ATOMIC_LDCLR)};
840 return getOutlineAtomicHelper(LC, Order, MemSize);
841 }
843 const Libcall LC[5][4] = {LCALL5(OUTLINE_ATOMIC_LDEOR)};
844 return getOutlineAtomicHelper(LC, Order, MemSize);
845 }
846 default:
847 return UNKNOWN_LIBCALL;
848 }
849#undef LCALLS
850#undef LCALL5
851}
852
853RTLIB::Libcall RTLIB::getSYNC(unsigned Opc, MVT VT) {
854#define OP_TO_LIBCALL(Name, Enum) \
855 case Name: \
856 switch (VT.SimpleTy) { \
857 default: \
858 return UNKNOWN_LIBCALL; \
859 case MVT::i8: \
860 return Enum##_1; \
861 case MVT::i16: \
862 return Enum##_2; \
863 case MVT::i32: \
864 return Enum##_4; \
865 case MVT::i64: \
866 return Enum##_8; \
867 case MVT::i128: \
868 return Enum##_16; \
869 }
870
871 switch (Opc) {
872 OP_TO_LIBCALL(ISD::ATOMIC_SWAP, SYNC_LOCK_TEST_AND_SET)
873 OP_TO_LIBCALL(ISD::ATOMIC_CMP_SWAP, SYNC_VAL_COMPARE_AND_SWAP)
874 OP_TO_LIBCALL(ISD::ATOMIC_LOAD_ADD, SYNC_FETCH_AND_ADD)
875 OP_TO_LIBCALL(ISD::ATOMIC_LOAD_SUB, SYNC_FETCH_AND_SUB)
876 OP_TO_LIBCALL(ISD::ATOMIC_LOAD_AND, SYNC_FETCH_AND_AND)
877 OP_TO_LIBCALL(ISD::ATOMIC_LOAD_OR, SYNC_FETCH_AND_OR)
878 OP_TO_LIBCALL(ISD::ATOMIC_LOAD_XOR, SYNC_FETCH_AND_XOR)
879 OP_TO_LIBCALL(ISD::ATOMIC_LOAD_NAND, SYNC_FETCH_AND_NAND)
880 OP_TO_LIBCALL(ISD::ATOMIC_LOAD_MAX, SYNC_FETCH_AND_MAX)
881 OP_TO_LIBCALL(ISD::ATOMIC_LOAD_UMAX, SYNC_FETCH_AND_UMAX)
882 OP_TO_LIBCALL(ISD::ATOMIC_LOAD_MIN, SYNC_FETCH_AND_MIN)
883 OP_TO_LIBCALL(ISD::ATOMIC_LOAD_UMIN, SYNC_FETCH_AND_UMIN)
884 }
885
886#undef OP_TO_LIBCALL
887
888 return UNKNOWN_LIBCALL;
889}
890
892 switch (ElementSize) {
893 case 1:
894 return MEMCPY_ELEMENT_UNORDERED_ATOMIC_1;
895 case 2:
896 return MEMCPY_ELEMENT_UNORDERED_ATOMIC_2;
897 case 4:
898 return MEMCPY_ELEMENT_UNORDERED_ATOMIC_4;
899 case 8:
900 return MEMCPY_ELEMENT_UNORDERED_ATOMIC_8;
901 case 16:
902 return MEMCPY_ELEMENT_UNORDERED_ATOMIC_16;
903 default:
904 return UNKNOWN_LIBCALL;
905 }
906}
907
909 switch (ElementSize) {
910 case 1:
911 return MEMMOVE_ELEMENT_UNORDERED_ATOMIC_1;
912 case 2:
913 return MEMMOVE_ELEMENT_UNORDERED_ATOMIC_2;
914 case 4:
915 return MEMMOVE_ELEMENT_UNORDERED_ATOMIC_4;
916 case 8:
917 return MEMMOVE_ELEMENT_UNORDERED_ATOMIC_8;
918 case 16:
919 return MEMMOVE_ELEMENT_UNORDERED_ATOMIC_16;
920 default:
921 return UNKNOWN_LIBCALL;
922 }
923}
924
926 switch (ElementSize) {
927 case 1:
928 return MEMSET_ELEMENT_UNORDERED_ATOMIC_1;
929 case 2:
930 return MEMSET_ELEMENT_UNORDERED_ATOMIC_2;
931 case 4:
932 return MEMSET_ELEMENT_UNORDERED_ATOMIC_4;
933 case 8:
934 return MEMSET_ELEMENT_UNORDERED_ATOMIC_8;
935 case 16:
936 return MEMSET_ELEMENT_UNORDERED_ATOMIC_16;
937 default:
938 return UNKNOWN_LIBCALL;
939 }
940}
941
943 RTLIB::LibcallImpl Impl) const {
944 switch (Impl) {
945 case RTLIB::impl___aeabi_dcmpeq__une:
946 case RTLIB::impl___aeabi_fcmpeq__une:
947 // Usage in the eq case, so we have to invert the comparison.
948 return ISD::SETEQ;
949 case RTLIB::impl___aeabi_dcmpeq__oeq:
950 case RTLIB::impl___aeabi_fcmpeq__oeq:
951 // Normal comparison to boolean value.
952 return ISD::SETNE;
953 case RTLIB::impl___aeabi_dcmplt:
954 case RTLIB::impl___aeabi_dcmple:
955 case RTLIB::impl___aeabi_dcmpge:
956 case RTLIB::impl___aeabi_dcmpgt:
957 case RTLIB::impl___aeabi_dcmpun:
958 case RTLIB::impl___aeabi_fcmplt:
959 case RTLIB::impl___aeabi_fcmple:
960 case RTLIB::impl___aeabi_fcmpge:
961 case RTLIB::impl___aeabi_fcmpgt:
962 /// The AEABI versions return a typical boolean value, so we can compare
963 /// against the integer result as simply != 0.
964 return ISD::SETNE;
965 default:
966 break;
967 }
968
969 // Assume libgcc/compiler-rt behavior. Most of the cases are really aliases of
970 // each other, and return a 3-way comparison style result of -1, 0, or 1
971 // depending on lt/eq/gt.
972 //
973 // FIXME: It would be cleaner to directly express this as a 3-way comparison
974 // soft FP libcall instead of individual compares.
975 RTLIB::Libcall LC = RTLIB::RuntimeLibcallsInfo::getLibcallFromImpl(Impl);
976 switch (LC) {
977 case RTLIB::OEQ_F32:
978 case RTLIB::OEQ_F64:
979 case RTLIB::OEQ_F128:
980 case RTLIB::OEQ_PPCF128:
981 return ISD::SETEQ;
982 case RTLIB::UNE_F32:
983 case RTLIB::UNE_F64:
984 case RTLIB::UNE_F128:
985 case RTLIB::UNE_PPCF128:
986 return ISD::SETNE;
987 case RTLIB::OGE_F32:
988 case RTLIB::OGE_F64:
989 case RTLIB::OGE_F128:
990 case RTLIB::OGE_PPCF128:
991 return ISD::SETGE;
992 case RTLIB::OLT_F32:
993 case RTLIB::OLT_F64:
994 case RTLIB::OLT_F128:
995 case RTLIB::OLT_PPCF128:
996 return ISD::SETLT;
997 case RTLIB::OLE_F32:
998 case RTLIB::OLE_F64:
999 case RTLIB::OLE_F128:
1000 case RTLIB::OLE_PPCF128:
1001 return ISD::SETLE;
1002 case RTLIB::OGT_F32:
1003 case RTLIB::OGT_F64:
1004 case RTLIB::OGT_F128:
1005 case RTLIB::OGT_PPCF128:
1006 return ISD::SETGT;
1007 case RTLIB::UO_F32:
1008 case RTLIB::UO_F64:
1009 case RTLIB::UO_F128:
1010 case RTLIB::UO_PPCF128:
1011 return ISD::SETNE;
1012 default:
1013 llvm_unreachable("not a compare libcall");
1014 }
1015}
1016
1017/// NOTE: The TargetMachine owns TLOF.
1019 const TargetSubtargetInfo &STI)
1020 : TM(tm),
1021 RuntimeLibcallInfo(TM.getTargetTriple(), TM.Options.ExceptionModel,
1022 TM.Options.FloatABIType, TM.Options.EABIVersion,
1023 TM.Options.MCOptions.getABIName(), TM.Options.VecLib),
1024 Libcalls(RuntimeLibcallInfo, STI) {
1025 initActions();
1026
1027 // Perform these initializations only once.
1033 HasExtractBitsInsn = false;
1034 JumpIsExpensive = JumpIsExpensiveOverride;
1036 EnableExtLdPromotion = false;
1037 StackPointerRegisterToSaveRestore = 0;
1038 BooleanContents = UndefinedBooleanContent;
1039 BooleanFloatContents = UndefinedBooleanContent;
1040 BooleanVectorContents = UndefinedBooleanContent;
1041 SchedPreferenceInfo = Sched::ILP;
1044 MaxBytesForAlignment = 0;
1045 MaxAtomicSizeInBitsSupported = 0;
1046
1047 // Assume that even with libcalls, no target supports wider than 128 bit
1048 // division.
1049 MaxDivRemBitWidthSupported = 128;
1050
1051 MaxLargeFPConvertBitWidthSupported = 128;
1052
1053 MinCmpXchgSizeInBits = 0;
1054 SupportsUnalignedAtomics = false;
1055
1056 MinimumBitTestCmps = MinimumBitTestCmpsOverride;
1057}
1058
1059// Define the virtual destructor out-of-line to act as a key method to anchor
1060// debug info (see coding standards).
1062
1064 // All operations default to being supported.
1065 memset(OpActions, 0, sizeof(OpActions));
1066 memset(LoadExtActions, 0, sizeof(LoadExtActions));
1067 memset(AtomicLoadExtActions, 0, sizeof(AtomicLoadExtActions));
1068 memset(TruncStoreActions, 0, sizeof(TruncStoreActions));
1069 memset(IndexedModeActions, 0, sizeof(IndexedModeActions));
1070 memset(CondCodeActions, 0, sizeof(CondCodeActions));
1071 llvm::fill(RegClassForVT, nullptr);
1072 llvm::fill(TargetDAGCombineArray, 0);
1073
1074 // Let extending atomic loads be unsupported by default.
1075 for (MVT ValVT : MVT::all_valuetypes())
1076 for (MVT MemVT : MVT::all_valuetypes())
1078 Expand);
1079
1080 // We're somewhat special casing MVT::i2 and MVT::i4. Ideally we want to
1081 // remove this and targets should individually set these types if not legal.
1084 for (MVT VT : {MVT::i2, MVT::i4})
1085 OpActions[(unsigned)VT.SimpleTy][NT] = Expand;
1086 }
1087 for (MVT AVT : MVT::all_valuetypes()) {
1088 for (MVT VT : {MVT::i2, MVT::i4, MVT::v128i2, MVT::v64i4}) {
1089 setTruncStoreAction(AVT, VT, Expand);
1092 }
1093 }
1094 for (unsigned IM = (unsigned)ISD::PRE_INC;
1095 IM != (unsigned)ISD::LAST_INDEXED_MODE; ++IM) {
1096 for (MVT VT : {MVT::i2, MVT::i4}) {
1101 }
1102 }
1103
1104 for (MVT VT : MVT::fp_valuetypes()) {
1105 MVT IntVT = MVT::getIntegerVT(VT.getFixedSizeInBits());
1106 if (IntVT.isValid()) {
1109 }
1110 }
1111
1112 // If f16 fma is not natively supported, the value must be promoted to an f64
1113 // (and not to f32!) to prevent double rounding issues.
1114 AddPromotedToType(ISD::FMA, MVT::f16, MVT::f64);
1115 AddPromotedToType(ISD::STRICT_FMA, MVT::f16, MVT::f64);
1116
1117 // Set default actions for various operations.
1118 for (MVT VT : MVT::all_valuetypes()) {
1119 // Default all indexed load / store to expand.
1120 for (unsigned IM = (unsigned)ISD::PRE_INC;
1121 IM != (unsigned)ISD::LAST_INDEXED_MODE; ++IM) {
1126 }
1127
1128 // Most backends expect to see the node which just returns the value loaded.
1130
1131 // These operations default to expand.
1161 VT, Expand);
1162
1163 // Overflow operations default to expand
1166 VT, Expand);
1167
1168 // Carry-using overflow operations default to expand.
1171 VT, Expand);
1172
1173 // ADDC/ADDE/SUBC/SUBE default to expand.
1175 Expand);
1176
1177 // [US]CMP default to expand
1179
1180 // Halving adds
1183 Expand);
1184
1185 // Absolute difference
1187
1188 // Carry-less multiply
1190
1191 // Saturated trunc
1195
1196 // These default to Expand so they will be expanded to CTLZ/CTTZ by default.
1198 Expand);
1199
1200 // This defaults to Expand so it will be expanded to ABS by default.
1203
1205
1206 // These library functions default to expand.
1209 VT, Expand);
1210
1211 // These operations default to expand for vector types.
1212 if (VT.isVector())
1218 VT, Expand);
1219
1220 // Constrained floating-point operations default to expand.
1221#define DAG_INSTRUCTION(NAME, NARG, ROUND_MODE, INTRINSIC, DAGN) \
1222 setOperationAction(ISD::STRICT_##DAGN, VT, Expand);
1223#include "llvm/IR/ConstrainedOps.def"
1224
1225 // For most targets @llvm.get.dynamic.area.offset just returns 0.
1227
1228 // Vector reduction default to expand.
1236 VT, Expand);
1237
1238 // Named vector shuffles default to expand.
1240 Expand);
1241
1242 // Only some target support this vector operation. Most need to expand it.
1244
1245 // cttz.elts defaults to expand.
1247 Expand);
1248
1249 // VP operations default to expand.
1250#define BEGIN_REGISTER_VP_SDNODE(SDOPC, ...) \
1251 setOperationAction(ISD::SDOPC, VT, Expand);
1252#include "llvm/IR/VPIntrinsics.def"
1253
1254 // Masked vector extracts default to expand.
1256
1259
1260 // FP environment operations default to expand.
1264
1266
1271 }
1272
1273 // Most targets ignore the @llvm.prefetch intrinsic.
1275
1276 // Most targets also ignore the @llvm.readcyclecounter intrinsic.
1278
1279 // Most targets also ignore the @llvm.readsteadycounter intrinsic.
1281
1282 // ConstantFP nodes default to expand. Targets can either change this to
1283 // Legal, in which case all fp constants are legal, or use isFPImmLegal()
1284 // to optimize expansions for certain constants.
1286 {MVT::bf16, MVT::f16, MVT::f32, MVT::f64, MVT::f80, MVT::f128},
1287 Expand);
1288
1289 // Insert custom handling default for llvm.canonicalize.*.
1291 {MVT::f16, MVT::f32, MVT::f64, MVT::f128}, Expand);
1292
1293 // FIXME: Query RuntimeLibCalls to make the decision.
1295 {MVT::f32, MVT::f64, MVT::f128}, LibCall);
1296
1299 MVT::f16, Promote);
1300 // Default ISD::TRAP to expand (which turns it into abort).
1301 setOperationAction(ISD::TRAP, MVT::Other, Expand);
1302
1303 // On most systems, DEBUGTRAP and TRAP have no difference. The "Expand"
1304 // here is to inform DAG Legalizer to replace DEBUGTRAP with TRAP.
1306
1308
1311
1312 for (MVT VT : {MVT::i8, MVT::i16, MVT::i32, MVT::i64}) {
1315 }
1317
1318 // This one by default will call __clear_cache unless the target
1319 // wants something different.
1321
1322 // By default, STACKADDRESS nodes are expanded like STACKSAVE nodes.
1323 // On SPARC targets, custom lowering is required.
1325}
1326
1328 EVT) const {
1329 return MVT::getIntegerVT(DL.getPointerSizeInBits(0));
1330}
1331
1333 const DataLayout &DL) const {
1334 assert(LHSTy.isInteger() && "Shift amount is not an integer type!");
1335 if (LHSTy.isVector())
1336 return LHSTy;
1337 MVT ShiftVT = getScalarShiftAmountTy(DL, LHSTy);
1338 // If any possible shift value won't fit in the prefered type, just use
1339 // something safe. Assume it will be legalized when the shift is expanded.
1340 if (ShiftVT.getSizeInBits() < Log2_32_Ceil(LHSTy.getSizeInBits()))
1341 ShiftVT = MVT::i32;
1342 assert(ShiftVT.getSizeInBits() >= Log2_32_Ceil(LHSTy.getSizeInBits()) &&
1343 "ShiftVT is still too small!");
1344 return ShiftVT;
1345}
1346
1347bool TargetLoweringBase::canOpTrap(unsigned Op, EVT VT) const {
1348 assert(isTypeLegal(VT));
1349 switch (Op) {
1350 default:
1351 return false;
1352 case ISD::SDIV:
1353 case ISD::UDIV:
1354 case ISD::SREM:
1355 case ISD::UREM:
1356 return true;
1357 }
1358}
1359
1361 unsigned DestAS) const {
1362 return TM.isNoopAddrSpaceCast(SrcAS, DestAS);
1363}
1364
1366 EVT RetVT, ElementCount EC, bool ZeroIsPoison,
1367 const ConstantRange *VScaleRange) const {
1368 // Find the smallest "sensible" element type to use for the expansion.
1369 ConstantRange CR(APInt(64, EC.getKnownMinValue()));
1370 if (EC.isScalable())
1371 CR = CR.umul_sat(*VScaleRange);
1372
1373 if (ZeroIsPoison)
1374 CR = CR.subtract(APInt(64, 1));
1375
1376 unsigned EltWidth = RetVT.getScalarSizeInBits();
1377 EltWidth = std::min(EltWidth, CR.getActiveBits());
1378 EltWidth = std::max(llvm::bit_ceil(EltWidth), (unsigned)8);
1379
1380 return EltWidth;
1381}
1382
1384 // If the command-line option was specified, ignore this request.
1385 if (!JumpIsExpensiveOverride.getNumOccurrences())
1386 JumpIsExpensive = isExpensive;
1387}
1388
1391 // If this is a simple type, use the ComputeRegisterProp mechanism.
1392 if (VT.isSimple()) {
1393 MVT SVT = VT.getSimpleVT();
1394 assert((unsigned)SVT.SimpleTy < std::size(TransformToType));
1395 MVT NVT = TransformToType[SVT.SimpleTy];
1396 LegalizeTypeAction LA = ValueTypeActions.getTypeAction(SVT);
1397
1398 assert((LA == TypeLegal || LA == TypeSoftenFloat ||
1399 LA == TypeSoftPromoteHalf ||
1400 (NVT.isVector() ||
1401 ValueTypeActions.getTypeAction(NVT) != TypePromoteInteger)) &&
1402 "Promote may not follow Expand or Promote");
1403
1404 if (LA == TypeSplitVector)
1405 return LegalizeKind(LA, EVT(SVT).getHalfNumVectorElementsVT(Context));
1406 if (LA == TypeScalarizeVector)
1407 return LegalizeKind(LA, SVT.getVectorElementType());
1408 return LegalizeKind(LA, NVT);
1409 }
1410
1411 // Handle Extended Scalar Types.
1412 if (!VT.isVector()) {
1413 assert(VT.isInteger() && "Float types must be simple");
1414 unsigned BitSize = VT.getSizeInBits();
1415 // First promote to a power-of-two size, then expand if necessary.
1416 if (BitSize < 8 || !isPowerOf2_32(BitSize)) {
1417 EVT NVT = VT.getRoundIntegerType(Context);
1418 assert(NVT != VT && "Unable to round integer VT");
1419 LegalizeKind NextStep = getTypeConversion(Context, NVT);
1420 // Avoid multi-step promotion.
1421 if (NextStep.first == TypePromoteInteger)
1422 return NextStep;
1423 // Return rounded integer type.
1424 return LegalizeKind(TypePromoteInteger, NVT);
1425 }
1426
1428 EVT::getIntegerVT(Context, VT.getSizeInBits() / 2));
1429 }
1430
1431 // Handle vector types.
1432 ElementCount NumElts = VT.getVectorElementCount();
1433 EVT EltVT = VT.getVectorElementType();
1434
1435 // Vectors with only one element are always scalarized.
1436 if (NumElts.isScalar())
1437 return LegalizeKind(TypeScalarizeVector, EltVT);
1438
1439 // Try to widen vector elements until the element type is a power of two and
1440 // promote it to a legal type later on, for example:
1441 // <3 x i8> -> <4 x i8> -> <4 x i32>
1442 if (EltVT.isInteger()) {
1443 // Vectors with a number of elements that is not a power of two are always
1444 // widened, for example <3 x i8> -> <4 x i8>.
1445 if (!VT.isPow2VectorType()) {
1446 NumElts = NumElts.coefficientNextPowerOf2();
1447 EVT NVT = EVT::getVectorVT(Context, EltVT, NumElts);
1448 return LegalizeKind(TypeWidenVector, NVT);
1449 }
1450
1451 // Examine the element type.
1452 LegalizeKind LK = getTypeConversion(Context, EltVT);
1453
1454 // If type is to be expanded, split the vector.
1455 // <4 x i140> -> <2 x i140>
1456 if (LK.first == TypeExpandInteger) {
1457 if (NumElts.isScalable() && NumElts.getKnownMinValue() == 1)
1460 VT.getHalfNumVectorElementsVT(Context));
1461 }
1462
1463 // Promote the integer element types until a legal vector type is found
1464 // or until the element integer type is too big. If a legal type was not
1465 // found, fallback to the usual mechanism of widening/splitting the
1466 // vector.
1467 EVT OldEltVT = EltVT;
1468 while (true) {
1469 // Increase the bitwidth of the element to the next pow-of-two
1470 // (which is greater than 8 bits).
1471 EltVT = EVT::getIntegerVT(Context, 1 + EltVT.getSizeInBits())
1472 .getRoundIntegerType(Context);
1473
1474 // Stop trying when getting a non-simple element type.
1475 // Note that vector elements may be greater than legal vector element
1476 // types. Example: X86 XMM registers hold 64bit element on 32bit
1477 // systems.
1478 if (!EltVT.isSimple())
1479 break;
1480
1481 // Build a new vector type and check if it is legal.
1482 MVT NVT = MVT::getVectorVT(EltVT.getSimpleVT(), NumElts);
1483 // Found a legal promoted vector type.
1484 if (NVT != MVT() && ValueTypeActions.getTypeAction(NVT) == TypeLegal)
1486 EVT::getVectorVT(Context, EltVT, NumElts));
1487 }
1488
1489 // Reset the type to the unexpanded type if we did not find a legal vector
1490 // type with a promoted vector element type.
1491 EltVT = OldEltVT;
1492 }
1493
1494 // Try to widen the vector until a legal type is found.
1495 // If there is no wider legal type, split the vector.
1496 while (true) {
1497 // Round up to the next power of 2.
1498 NumElts = NumElts.coefficientNextPowerOf2();
1499
1500 // If there is no simple vector type with this many elements then there
1501 // cannot be a larger legal vector type. Note that this assumes that
1502 // there are no skipped intermediate vector types in the simple types.
1503 if (!EltVT.isSimple())
1504 break;
1505 MVT LargerVector = MVT::getVectorVT(EltVT.getSimpleVT(), NumElts);
1506 if (LargerVector == MVT())
1507 break;
1508
1509 // If this type is legal then widen the vector.
1510 if (ValueTypeActions.getTypeAction(LargerVector) == TypeLegal)
1511 return LegalizeKind(TypeWidenVector, LargerVector);
1512 }
1513
1514 // Widen odd vectors to next power of two.
1515 if (!VT.isPow2VectorType()) {
1516 EVT NVT = VT.getPow2VectorType(Context);
1517 return LegalizeKind(TypeWidenVector, NVT);
1518 }
1519
1522
1523 // Vectors with illegal element types are expanded.
1524 EVT NVT = EVT::getVectorVT(Context, EltVT,
1526 return LegalizeKind(TypeSplitVector, NVT);
1527}
1528
1529static unsigned getVectorTypeBreakdownMVT(MVT VT, MVT &IntermediateVT,
1530 unsigned &NumIntermediates,
1531 MVT &RegisterVT,
1532 TargetLoweringBase *TLI) {
1533 // Figure out the right, legal destination reg to copy into.
1535 MVT EltTy = VT.getVectorElementType();
1536
1537 unsigned NumVectorRegs = 1;
1538
1539 // Scalable vectors cannot be scalarized, so splitting or widening is
1540 // required.
1541 if (VT.isScalableVector() && !isPowerOf2_32(EC.getKnownMinValue()))
1543 "Splitting or widening of non-power-of-2 MVTs is not implemented.");
1544
1545 // FIXME: We don't support non-power-of-2-sized vectors for now.
1546 // Ideally we could break down into LHS/RHS like LegalizeDAG does.
1547 if (!isPowerOf2_32(EC.getKnownMinValue())) {
1548 // Split EC to unit size (scalable property is preserved).
1549 NumVectorRegs = EC.getKnownMinValue();
1550 EC = ElementCount::getFixed(1);
1551 }
1552
1553 // Divide the input until we get to a supported size. This will
1554 // always end up with an EC that represent a scalar or a scalable
1555 // scalar.
1556 while (EC.getKnownMinValue() > 1 &&
1557 !TLI->isTypeLegal(MVT::getVectorVT(EltTy, EC))) {
1558 EC = EC.divideCoefficientBy(2);
1559 NumVectorRegs <<= 1;
1560 }
1561
1562 NumIntermediates = NumVectorRegs;
1563
1564 MVT NewVT = MVT::getVectorVT(EltTy, EC);
1565 if (!TLI->isTypeLegal(NewVT))
1566 NewVT = EltTy;
1567 IntermediateVT = NewVT;
1568
1569 unsigned LaneSizeInBits = NewVT.getScalarSizeInBits();
1570
1571 // Convert sizes such as i33 to i64.
1572 LaneSizeInBits = llvm::bit_ceil(LaneSizeInBits);
1573
1574 MVT DestVT = TLI->getRegisterType(NewVT);
1575 RegisterVT = DestVT;
1576 if (EVT(DestVT).bitsLT(NewVT)) // Value is expanded, e.g. i64 -> i16.
1577 return NumVectorRegs * (LaneSizeInBits / DestVT.getScalarSizeInBits());
1578
1579 // Otherwise, promotion or legal types use the same number of registers as
1580 // the vector decimated to the appropriate level.
1581 return NumVectorRegs;
1582}
1583
1584/// isLegalRC - Return true if the value types that can be represented by the
1585/// specified register class are all legal.
1587 const TargetRegisterClass &RC) const {
1588 for (const auto *I = TRI.legalclasstypes_begin(RC); *I != MVT::Other; ++I)
1589 if (isTypeLegal(*I))
1590 return true;
1591 return false;
1592}
1593
1594/// Replace/modify any TargetFrameIndex operands with a targte-dependent
1595/// sequence of memory operands that is recognized by PrologEpilogInserter.
1598 MachineBasicBlock *MBB) const {
1599 MachineInstr *MI = &InitialMI;
1600 MachineFunction &MF = *MI->getMF();
1601 MachineFrameInfo &MFI = MF.getFrameInfo();
1602
1603 // We're handling multiple types of operands here:
1604 // PATCHPOINT MetaArgs - live-in, read only, direct
1605 // STATEPOINT Deopt Spill - live-through, read only, indirect
1606 // STATEPOINT Deopt Alloca - live-through, read only, direct
1607 // (We're currently conservative and mark the deopt slots read/write in
1608 // practice.)
1609 // STATEPOINT GC Spill - live-through, read/write, indirect
1610 // STATEPOINT GC Alloca - live-through, read/write, direct
1611 // The live-in vs live-through is handled already (the live through ones are
1612 // all stack slots), but we need to handle the different type of stackmap
1613 // operands and memory effects here.
1614
1615 if (llvm::none_of(MI->operands(),
1616 [](MachineOperand &Operand) { return Operand.isFI(); }))
1617 return MBB;
1618
1619 MachineInstrBuilder MIB = BuildMI(MF, MI->getDebugLoc(), MI->getDesc());
1620
1621 // Inherit previous memory operands.
1622 MIB.cloneMemRefs(*MI);
1623
1624 for (unsigned i = 0; i < MI->getNumOperands(); ++i) {
1625 MachineOperand &MO = MI->getOperand(i);
1626 if (!MO.isFI()) {
1627 // Index of Def operand this Use it tied to.
1628 // Since Defs are coming before Uses, if Use is tied, then
1629 // index of Def must be smaller that index of that Use.
1630 // Also, Defs preserve their position in new MI.
1631 unsigned TiedTo = i;
1632 if (MO.isReg() && MO.isTied())
1633 TiedTo = MI->findTiedOperandIdx(i);
1634 MIB.add(MO);
1635 if (TiedTo < i)
1636 MIB->tieOperands(TiedTo, MIB->getNumOperands() - 1);
1637 continue;
1638 }
1639
1640 // foldMemoryOperand builds a new MI after replacing a single FI operand
1641 // with the canonical set of five x86 addressing-mode operands.
1642 int FI = MO.getIndex();
1643
1644 // Add frame index operands recognized by stackmaps.cpp
1646 // indirect-mem-ref tag, size, #FI, offset.
1647 // Used for spills inserted by StatepointLowering. This codepath is not
1648 // used for patchpoints/stackmaps at all, for these spilling is done via
1649 // foldMemoryOperand callback only.
1650 assert(MI->getOpcode() == TargetOpcode::STATEPOINT && "sanity");
1651 MIB.addImm(StackMaps::IndirectMemRefOp);
1652 MIB.addImm(MFI.getObjectSize(FI));
1653 MIB.add(MO);
1654 MIB.addImm(0);
1655 } else {
1656 // direct-mem-ref tag, #FI, offset.
1657 // Used by patchpoint, and direct alloca arguments to statepoints
1658 MIB.addImm(StackMaps::DirectMemRefOp);
1659 MIB.add(MO);
1660 MIB.addImm(0);
1661 }
1662
1663 assert(MIB->mayLoad() && "Folded a stackmap use to a non-load!");
1664
1665 // Add a new memory operand for this FI.
1666 assert(MFI.getObjectOffset(FI) != -1);
1667
1668 // Note: STATEPOINT MMOs are added during SelectionDAG. STACKMAP, and
1669 // PATCHPOINT should be updated to do the same. (TODO)
1670 if (MI->getOpcode() != TargetOpcode::STATEPOINT) {
1671 auto Flags = MachineMemOperand::MOLoad;
1673 MachinePointerInfo::getFixedStack(MF, FI), Flags,
1675 MIB->addMemOperand(MF, MMO);
1676 }
1677 }
1678 MBB->insert(MachineBasicBlock::iterator(MI), MIB);
1679 MI->eraseFromParent();
1680 return MBB;
1681}
1682
1683/// findRepresentativeClass - Return the largest legal super-reg register class
1684/// of the register class for the specified type and its associated "cost".
1685// This function is in TargetLowering because it uses RegClassForVT which would
1686// need to be moved to TargetRegisterInfo and would necessitate moving
1687// isTypeLegal over as well - a massive change that would just require
1688// TargetLowering having a TargetRegisterInfo class member that it would use.
1689std::pair<const TargetRegisterClass *, uint8_t>
1691 MVT VT) const {
1692 const TargetRegisterClass *RC = RegClassForVT[VT.SimpleTy];
1693 if (!RC)
1694 return std::make_pair(RC, 0);
1695
1696 // Compute the set of all super-register classes.
1697 BitVector SuperRegRC(TRI->getNumRegClasses());
1698 for (SuperRegClassIterator RCI(RC, TRI); RCI.isValid(); ++RCI)
1699 SuperRegRC.setBitsInMask(RCI.getMask());
1700
1701 // Find the first legal register class with the largest spill size.
1702 const TargetRegisterClass *BestRC = RC;
1703 for (unsigned i : SuperRegRC.set_bits()) {
1704 const TargetRegisterClass *SuperRC = TRI->getRegClass(i);
1705 // We want the largest possible spill size.
1706 if (TRI->getSpillSize(*SuperRC) <= TRI->getSpillSize(*BestRC))
1707 continue;
1708 if (!isLegalRC(*TRI, *SuperRC))
1709 continue;
1710 BestRC = SuperRC;
1711 }
1712 return std::make_pair(BestRC, 1);
1713}
1714
1715/// computeRegisterProperties - Once all of the register classes are added,
1716/// this allows us to compute derived properties we expose.
1718 const TargetRegisterInfo *TRI) {
1719 // Everything defaults to needing one register.
1720 for (unsigned i = 0; i != MVT::VALUETYPE_SIZE; ++i) {
1721 NumRegistersForVT[i] = 1;
1722 RegisterTypeForVT[i] = TransformToType[i] = (MVT::SimpleValueType)i;
1723 }
1724 // ...except isVoid, which doesn't need any registers.
1725 NumRegistersForVT[MVT::isVoid] = 0;
1726
1727 // Find the largest integer register class.
1728 unsigned LargestIntReg = MVT::LAST_INTEGER_VALUETYPE;
1729 for (; RegClassForVT[LargestIntReg] == nullptr; --LargestIntReg)
1730 assert(LargestIntReg != MVT::i1 && "No integer registers defined!");
1731
1732 // Every integer value type larger than this largest register takes twice as
1733 // many registers to represent as the previous ValueType.
1734 for (unsigned ExpandedReg = LargestIntReg + 1;
1735 ExpandedReg <= MVT::LAST_INTEGER_VALUETYPE; ++ExpandedReg) {
1736 NumRegistersForVT[ExpandedReg] = 2*NumRegistersForVT[ExpandedReg-1];
1737 RegisterTypeForVT[ExpandedReg] = (MVT::SimpleValueType)LargestIntReg;
1738 TransformToType[ExpandedReg] = (MVT::SimpleValueType)(ExpandedReg - 1);
1739 ValueTypeActions.setTypeAction((MVT::SimpleValueType)ExpandedReg,
1741 }
1742
1743 // Inspect all of the ValueType's smaller than the largest integer
1744 // register to see which ones need promotion.
1745 unsigned LegalIntReg = LargestIntReg;
1746 for (unsigned IntReg = LargestIntReg - 1;
1747 IntReg >= (unsigned)MVT::i1; --IntReg) {
1748 MVT IVT = (MVT::SimpleValueType)IntReg;
1749 if (isTypeLegal(IVT)) {
1750 LegalIntReg = IntReg;
1751 } else {
1752 RegisterTypeForVT[IntReg] = TransformToType[IntReg] =
1753 (MVT::SimpleValueType)LegalIntReg;
1754 ValueTypeActions.setTypeAction(IVT, TypePromoteInteger);
1755 }
1756 }
1757
1758 // ppcf128 type is really two f64's.
1759 if (!isTypeLegal(MVT::ppcf128)) {
1760 if (isTypeLegal(MVT::f64)) {
1761 NumRegistersForVT[MVT::ppcf128] = 2*NumRegistersForVT[MVT::f64];
1762 RegisterTypeForVT[MVT::ppcf128] = MVT::f64;
1763 TransformToType[MVT::ppcf128] = MVT::f64;
1764 ValueTypeActions.setTypeAction(MVT::ppcf128, TypeExpandFloat);
1765 } else {
1766 NumRegistersForVT[MVT::ppcf128] = NumRegistersForVT[MVT::i128];
1767 RegisterTypeForVT[MVT::ppcf128] = RegisterTypeForVT[MVT::i128];
1768 TransformToType[MVT::ppcf128] = MVT::i128;
1769 ValueTypeActions.setTypeAction(MVT::ppcf128, TypeSoftenFloat);
1770 }
1771 }
1772
1773 // Decide how to handle f128. If the target does not have native f128 support,
1774 // expand it to i128 and we will be generating soft float library calls.
1775 if (!isTypeLegal(MVT::f128)) {
1776 NumRegistersForVT[MVT::f128] = NumRegistersForVT[MVT::i128];
1777 RegisterTypeForVT[MVT::f128] = RegisterTypeForVT[MVT::i128];
1778 TransformToType[MVT::f128] = MVT::i128;
1779 ValueTypeActions.setTypeAction(MVT::f128, TypeSoftenFloat);
1780 }
1781
1782 // Decide how to handle f80. If the target does not have native f80 support,
1783 // expand it to i96 and we will be generating soft float library calls.
1784 if (!isTypeLegal(MVT::f80)) {
1785 NumRegistersForVT[MVT::f80] = 3*NumRegistersForVT[MVT::i32];
1786 RegisterTypeForVT[MVT::f80] = RegisterTypeForVT[MVT::i32];
1787 TransformToType[MVT::f80] = MVT::i32;
1788 ValueTypeActions.setTypeAction(MVT::f80, TypeSoftenFloat);
1789 }
1790
1791 // Decide how to handle f64. If the target does not have native f64 support,
1792 // expand it to i64 and we will be generating soft float library calls.
1793 if (!isTypeLegal(MVT::f64)) {
1794 NumRegistersForVT[MVT::f64] = NumRegistersForVT[MVT::i64];
1795 RegisterTypeForVT[MVT::f64] = RegisterTypeForVT[MVT::i64];
1796 TransformToType[MVT::f64] = MVT::i64;
1797 ValueTypeActions.setTypeAction(MVT::f64, TypeSoftenFloat);
1798 }
1799
1800 // Decide how to handle f32. If the target does not have native f32 support,
1801 // expand it to i32 and we will be generating soft float library calls.
1802 if (!isTypeLegal(MVT::f32)) {
1803 NumRegistersForVT[MVT::f32] = NumRegistersForVT[MVT::i32];
1804 RegisterTypeForVT[MVT::f32] = RegisterTypeForVT[MVT::i32];
1805 TransformToType[MVT::f32] = MVT::i32;
1806 ValueTypeActions.setTypeAction(MVT::f32, TypeSoftenFloat);
1807 }
1808
1809 // Decide how to handle f16. If the target does not have native f16 support,
1810 // promote it to f32, because there are no f16 library calls (except for
1811 // conversions).
1812 if (!isTypeLegal(MVT::f16)) {
1813 // Allow targets to control how we legalize half.
1814 bool UseFPRegsForHalfType = useFPRegsForHalfType();
1815
1816 if (!UseFPRegsForHalfType) {
1817 NumRegistersForVT[MVT::f16] = NumRegistersForVT[MVT::i16];
1818 RegisterTypeForVT[MVT::f16] = RegisterTypeForVT[MVT::i16];
1819 } else {
1820 NumRegistersForVT[MVT::f16] = NumRegistersForVT[MVT::f32];
1821 RegisterTypeForVT[MVT::f16] = RegisterTypeForVT[MVT::f32];
1822 }
1823 TransformToType[MVT::f16] = MVT::f32;
1824 ValueTypeActions.setTypeAction(MVT::f16, TypeSoftPromoteHalf);
1825 }
1826
1827 // Decide how to handle bf16. If the target does not have native bf16 support,
1828 // promote it to f32, because there are no bf16 library calls (except for
1829 // converting from f32 to bf16).
1830 if (!isTypeLegal(MVT::bf16)) {
1831 NumRegistersForVT[MVT::bf16] = NumRegistersForVT[MVT::f32];
1832 RegisterTypeForVT[MVT::bf16] = RegisterTypeForVT[MVT::f32];
1833 TransformToType[MVT::bf16] = MVT::f32;
1834 ValueTypeActions.setTypeAction(MVT::bf16, TypeSoftPromoteHalf);
1835 }
1836
1837 // Loop over all of the vector value types to see which need transformations.
1838 for (unsigned i = MVT::FIRST_VECTOR_VALUETYPE;
1839 i <= (unsigned)MVT::LAST_VECTOR_VALUETYPE; ++i) {
1840 MVT VT = (MVT::SimpleValueType) i;
1841 if (isTypeLegal(VT))
1842 continue;
1843
1844 MVT EltVT = VT.getVectorElementType();
1846 bool IsLegalWiderType = false;
1847 bool IsScalable = VT.isScalableVector();
1848 LegalizeTypeAction PreferredAction = getPreferredVectorAction(VT);
1849 switch (PreferredAction) {
1850 case TypePromoteInteger: {
1851 MVT::SimpleValueType EndVT = IsScalable ?
1852 MVT::LAST_INTEGER_SCALABLE_VECTOR_VALUETYPE :
1853 MVT::LAST_INTEGER_FIXEDLEN_VECTOR_VALUETYPE;
1854 // Try to promote the elements of integer vectors. If no legal
1855 // promotion was found, fall through to the widen-vector method.
1856 for (unsigned nVT = i + 1;
1857 (MVT::SimpleValueType)nVT <= EndVT; ++nVT) {
1858 MVT SVT = (MVT::SimpleValueType) nVT;
1859 // Promote vectors of integers to vectors with the same number
1860 // of elements, with a wider element type.
1861 if (SVT.getScalarSizeInBits() > EltVT.getFixedSizeInBits() &&
1862 SVT.getVectorElementCount() == EC && isTypeLegal(SVT)) {
1863 TransformToType[i] = SVT;
1864 RegisterTypeForVT[i] = SVT;
1865 NumRegistersForVT[i] = 1;
1866 ValueTypeActions.setTypeAction(VT, TypePromoteInteger);
1867 IsLegalWiderType = true;
1868 break;
1869 }
1870 }
1871 if (IsLegalWiderType)
1872 break;
1873 [[fallthrough]];
1874 }
1875
1876 case TypeWidenVector:
1877 if (isPowerOf2_32(EC.getKnownMinValue())) {
1878 // Try to widen the vector.
1879 for (unsigned nVT = i + 1; nVT <= MVT::LAST_VECTOR_VALUETYPE; ++nVT) {
1880 MVT SVT = (MVT::SimpleValueType) nVT;
1881 if (SVT.getVectorElementType() == EltVT &&
1882 SVT.isScalableVector() == IsScalable &&
1884 EC.getKnownMinValue() &&
1885 isTypeLegal(SVT)) {
1886 TransformToType[i] = SVT;
1887 RegisterTypeForVT[i] = SVT;
1888 NumRegistersForVT[i] = 1;
1889 ValueTypeActions.setTypeAction(VT, TypeWidenVector);
1890 IsLegalWiderType = true;
1891 break;
1892 }
1893 }
1894 if (IsLegalWiderType)
1895 break;
1896 } else {
1897 // Only widen to the next power of 2 to keep consistency with EVT.
1898 MVT NVT = VT.getPow2VectorType();
1899 if (isTypeLegal(NVT)) {
1900 TransformToType[i] = NVT;
1901 ValueTypeActions.setTypeAction(VT, TypeWidenVector);
1902 RegisterTypeForVT[i] = NVT;
1903 NumRegistersForVT[i] = 1;
1904 break;
1905 }
1906 }
1907 [[fallthrough]];
1908
1909 case TypeSplitVector:
1910 case TypeScalarizeVector: {
1911 MVT IntermediateVT;
1912 MVT RegisterVT;
1913 unsigned NumIntermediates;
1914 unsigned NumRegisters = getVectorTypeBreakdownMVT(VT, IntermediateVT,
1915 NumIntermediates, RegisterVT, this);
1916 NumRegistersForVT[i] = NumRegisters;
1917 assert(NumRegistersForVT[i] == NumRegisters &&
1918 "NumRegistersForVT size cannot represent NumRegisters!");
1919 RegisterTypeForVT[i] = RegisterVT;
1920
1921 MVT NVT = VT.getPow2VectorType();
1922 if (NVT == VT) {
1923 // Type is already a power of 2. The default action is to split.
1924 TransformToType[i] = MVT::Other;
1925 if (PreferredAction == TypeScalarizeVector)
1926 ValueTypeActions.setTypeAction(VT, TypeScalarizeVector);
1927 else if (PreferredAction == TypeSplitVector)
1928 ValueTypeActions.setTypeAction(VT, TypeSplitVector);
1929 else if (EC.getKnownMinValue() > 1)
1930 ValueTypeActions.setTypeAction(VT, TypeSplitVector);
1931 else
1932 ValueTypeActions.setTypeAction(VT, EC.isScalable()
1935 } else {
1936 TransformToType[i] = NVT;
1937 ValueTypeActions.setTypeAction(VT, TypeWidenVector);
1938 }
1939 break;
1940 }
1941 default:
1942 llvm_unreachable("Unknown vector legalization action!");
1943 }
1944 }
1945
1946 // Determine the 'representative' register class for each value type.
1947 // An representative register class is the largest (meaning one which is
1948 // not a sub-register class / subreg register class) legal register class for
1949 // a group of value types. For example, on i386, i8, i16, and i32
1950 // representative would be GR32; while on x86_64 it's GR64.
1951 for (unsigned i = 0; i != MVT::VALUETYPE_SIZE; ++i) {
1952 const TargetRegisterClass* RRC;
1953 uint8_t Cost;
1955 RepRegClassForVT[i] = RRC;
1956 RepRegClassCostForVT[i] = Cost;
1957 }
1958
1959 // Compute minimum known-legal store size.
1960 MaximumLegalStoreInBits = 0;
1961 for (MVT VT : MVT::all_valuetypes())
1962 if (VT != MVT::Other && isTypeLegal(VT) &&
1963 VT.getSizeInBits().getKnownMinValue() >= MaximumLegalStoreInBits)
1964 MaximumLegalStoreInBits = VT.getSizeInBits().getKnownMinValue();
1965}
1966
1968 EVT VT) const {
1969 assert(!VT.isVector() && "No default SetCC type for vectors!");
1970 return getPointerTy(DL).SimpleTy;
1971}
1972
1973/// getVectorTypeBreakdown - Vector types are broken down into some number of
1974/// legal first class types. For example, MVT::v8f32 maps to 2 MVT::v4f32
1975/// with Altivec or SSE1, or 8 promoted MVT::f64 values with the X86 FP stack.
1976/// Similarly, MVT::v2i64 turns into 4 MVT::i32 values with both PPC and X86.
1977///
1978/// This method returns the number of registers needed, and the VT for each
1979/// register. It also returns the VT and quantity of the intermediate values
1980/// before they are promoted/expanded.
1982 EVT VT, EVT &IntermediateVT,
1983 unsigned &NumIntermediates,
1984 MVT &RegisterVT) const {
1985 ElementCount EltCnt = VT.getVectorElementCount();
1986
1987 // If there is a wider vector type with the same element type as this one,
1988 // or a promoted vector type that has the same number of elements which
1989 // are wider, then we should convert to that legal vector type.
1990 // This handles things like <2 x float> -> <4 x float> and
1991 // <4 x i1> -> <4 x i32>.
1992 LegalizeTypeAction TA = getTypeAction(Context, VT);
1993 if (!EltCnt.isScalar() &&
1994 (TA == TypeWidenVector || TA == TypePromoteInteger)) {
1995 EVT RegisterEVT = getTypeToTransformTo(Context, VT);
1996 if (isTypeLegal(RegisterEVT)) {
1997 IntermediateVT = RegisterEVT;
1998 RegisterVT = RegisterEVT.getSimpleVT();
1999 NumIntermediates = 1;
2000 return 1;
2001 }
2002 }
2003
2004 // Figure out the right, legal destination reg to copy into.
2005 EVT EltTy = VT.getVectorElementType();
2006
2007 unsigned NumVectorRegs = 1;
2008
2009 // Scalable vectors cannot be scalarized, so handle the legalisation of the
2010 // types like done elsewhere in SelectionDAG.
2011 if (EltCnt.isScalable()) {
2012 LegalizeKind LK;
2013 EVT PartVT = VT;
2014 do {
2015 // Iterate until we've found a legal (part) type to hold VT.
2016 LK = getTypeConversion(Context, PartVT);
2017 PartVT = LK.second;
2018 } while (LK.first != TypeLegal);
2019
2020 if (!PartVT.isVector()) {
2022 "Don't know how to legalize this scalable vector type");
2023 }
2024
2025 NumIntermediates =
2028 IntermediateVT = PartVT;
2029 RegisterVT = getRegisterType(Context, IntermediateVT);
2030 return NumIntermediates;
2031 }
2032
2033 // FIXME: We don't support non-power-of-2-sized vectors for now. Ideally
2034 // we could break down into LHS/RHS like LegalizeDAG does.
2035 if (!isPowerOf2_32(EltCnt.getKnownMinValue())) {
2036 NumVectorRegs = EltCnt.getKnownMinValue();
2037 EltCnt = ElementCount::getFixed(1);
2038 }
2039
2040 // Divide the input until we get to a supported size. This will always
2041 // end with a scalar if the target doesn't support vectors.
2042 while (EltCnt.getKnownMinValue() > 1 &&
2043 !isTypeLegal(EVT::getVectorVT(Context, EltTy, EltCnt))) {
2044 EltCnt = EltCnt.divideCoefficientBy(2);
2045 NumVectorRegs <<= 1;
2046 }
2047
2048 NumIntermediates = NumVectorRegs;
2049
2050 EVT NewVT = EVT::getVectorVT(Context, EltTy, EltCnt);
2051 if (!isTypeLegal(NewVT))
2052 NewVT = EltTy;
2053 IntermediateVT = NewVT;
2054
2055 MVT DestVT = getRegisterType(Context, NewVT);
2056 RegisterVT = DestVT;
2057
2058 if (EVT(DestVT).bitsLT(NewVT)) { // Value is expanded, e.g. i64 -> i16.
2059 TypeSize NewVTSize = NewVT.getSizeInBits();
2060 // Convert sizes such as i33 to i64.
2062 NewVTSize = NewVTSize.coefficientNextPowerOf2();
2063 return NumVectorRegs*(NewVTSize/DestVT.getSizeInBits());
2064 }
2065
2066 // Otherwise, promotion or legal types use the same number of registers as
2067 // the vector decimated to the appropriate level.
2068 return NumVectorRegs;
2069}
2070
2072 uint64_t NumCases,
2074 ProfileSummaryInfo *PSI,
2075 BlockFrequencyInfo *BFI) const {
2076 // FIXME: This function check the maximum table size and density, but the
2077 // minimum size is not checked. It would be nice if the minimum size is
2078 // also combined within this function. Currently, the minimum size check is
2079 // performed in findJumpTable() in SelectionDAGBuiler and
2080 // getEstimatedNumberOfCaseClusters() in BasicTTIImpl.
2081 const bool OptForSize =
2082 llvm::shouldOptimizeForSize(SI->getParent(), PSI, BFI);
2083 const unsigned MinDensity = getMinimumJumpTableDensity(OptForSize);
2084 const unsigned MaxJumpTableSize = getMaximumJumpTableSize();
2085
2086 // Check whether the number of cases is small enough and
2087 // the range is dense enough for a jump table.
2088 return (OptForSize || Range <= MaxJumpTableSize) &&
2089 (NumCases * 100 >= Range * MinDensity);
2090}
2091
2093 EVT ConditionVT) const {
2094 return getRegisterType(Context, ConditionVT);
2095}
2096
2097/// Get the EVTs and ArgFlags collections that represent the legalized return
2098/// type of the given function. This does not require a DAG or a return value,
2099/// and is suitable for use before any DAGs for the function are constructed.
2100/// TODO: Move this out of TargetLowering.cpp.
2102 AttributeList attr,
2104 const TargetLowering &TLI, const DataLayout &DL) {
2106 ComputeValueTypes(DL, ReturnType, Types);
2107 unsigned NumValues = Types.size();
2108 if (NumValues == 0) return;
2109
2110 for (Type *Ty : Types) {
2111 EVT VT = TLI.getValueType(DL, Ty);
2112 ISD::NodeType ExtendKind = ISD::ANY_EXTEND;
2113
2114 if (attr.hasRetAttr(Attribute::SExt))
2115 ExtendKind = ISD::SIGN_EXTEND;
2116 else if (attr.hasRetAttr(Attribute::ZExt))
2117 ExtendKind = ISD::ZERO_EXTEND;
2118
2119 if (ExtendKind != ISD::ANY_EXTEND && VT.isInteger())
2120 VT = TLI.getTypeForExtReturn(ReturnType->getContext(), VT, ExtendKind);
2121
2122 unsigned NumParts =
2123 TLI.getNumRegistersForCallingConv(ReturnType->getContext(), CC, VT);
2124 MVT PartVT =
2125 TLI.getRegisterTypeForCallingConv(ReturnType->getContext(), CC, VT);
2126
2127 // 'inreg' on function refers to return value
2129 if (attr.hasRetAttr(Attribute::InReg))
2130 Flags.setInReg();
2131
2132 // Propagate extension type if any
2133 if (attr.hasRetAttr(Attribute::SExt))
2134 Flags.setSExt();
2135 else if (attr.hasRetAttr(Attribute::ZExt))
2136 Flags.setZExt();
2137
2138 for (unsigned i = 0; i < NumParts; ++i)
2139 Outs.push_back(ISD::OutputArg(Flags, PartVT, VT, Ty, 0, 0));
2140 }
2141}
2142
2144 const DataLayout &DL) const {
2145 return DL.getABITypeAlign(Ty);
2146}
2147
2149 LLVMContext &Context, const DataLayout &DL, EVT VT, unsigned AddrSpace,
2150 Align Alignment, MachineMemOperand::Flags Flags, unsigned *Fast) const {
2151 // Check if the specified alignment is sufficient based on the data layout.
2152 // TODO: While using the data layout works in practice, a better solution
2153 // would be to implement this check directly (make this a virtual function).
2154 // For example, the ABI alignment may change based on software platform while
2155 // this function should only be affected by hardware implementation.
2156 Type *Ty = VT.getTypeForEVT(Context);
2157 if (VT.isZeroSized() || Alignment >= DL.getABITypeAlign(Ty)) {
2158 // Assume that an access that meets the ABI-specified alignment is fast.
2159 if (Fast != nullptr)
2160 *Fast = 1;
2161 return true;
2162 }
2163
2164 // This is a misaligned access.
2165 return allowsMisalignedMemoryAccesses(VT, AddrSpace, Alignment, Flags, Fast);
2166}
2167
2169 LLVMContext &Context, const DataLayout &DL, EVT VT,
2170 const MachineMemOperand &MMO, unsigned *Fast) const {
2171 return allowsMemoryAccessForAlignment(Context, DL, VT, MMO.getAddrSpace(),
2172 MMO.getAlign(), MMO.getFlags(), Fast);
2173}
2174
2176 const DataLayout &DL, EVT VT,
2177 unsigned AddrSpace, Align Alignment,
2179 unsigned *Fast) const {
2180 return allowsMemoryAccessForAlignment(Context, DL, VT, AddrSpace, Alignment,
2181 Flags, Fast);
2182}
2183
2185 const DataLayout &DL, EVT VT,
2186 const MachineMemOperand &MMO,
2187 unsigned *Fast) const {
2188 return allowsMemoryAccess(Context, DL, VT, MMO.getAddrSpace(), MMO.getAlign(),
2189 MMO.getFlags(), Fast);
2190}
2191
2193 const DataLayout &DL, LLT Ty,
2194 const MachineMemOperand &MMO,
2195 unsigned *Fast) const {
2196 EVT VT = getApproximateEVTForLLT(Ty, Context);
2197 return allowsMemoryAccess(Context, DL, VT, MMO.getAddrSpace(), MMO.getAlign(),
2198 MMO.getFlags(), Fast);
2199}
2200
2201unsigned TargetLoweringBase::getMaxStoresPerMemset(bool OptSize) const {
2204
2206}
2207
2208unsigned TargetLoweringBase::getMaxStoresPerMemcpy(bool OptSize) const {
2211
2213}
2214
2218
2220}
2221
2222//===----------------------------------------------------------------------===//
2223// TargetTransformInfo Helpers
2224//===----------------------------------------------------------------------===//
2225
2227 enum InstructionOpcodes {
2228#define HANDLE_INST(NUM, OPCODE, CLASS) OPCODE = NUM,
2229#define LAST_OTHER_INST(NUM) InstructionOpcodesCount = NUM
2230#include "llvm/IR/Instruction.def"
2231 };
2232 switch (static_cast<InstructionOpcodes>(Opcode)) {
2233 case Ret: return 0;
2234 case UncondBr: return 0;
2235 case CondBr: return 0;
2236 case Switch: return 0;
2237 case IndirectBr: return 0;
2238 case Invoke: return 0;
2239 case CallBr: return 0;
2240 case Resume: return 0;
2241 case Unreachable: return 0;
2242 case CleanupRet: return 0;
2243 case CatchRet: return 0;
2244 case CatchPad: return 0;
2245 case CatchSwitch: return 0;
2246 case CleanupPad: return 0;
2247 case FNeg: return ISD::FNEG;
2248 case Add: return ISD::ADD;
2249 case FAdd: return ISD::FADD;
2250 case Sub: return ISD::SUB;
2251 case FSub: return ISD::FSUB;
2252 case Mul: return ISD::MUL;
2253 case FMul: return ISD::FMUL;
2254 case UDiv: return ISD::UDIV;
2255 case SDiv: return ISD::SDIV;
2256 case FDiv: return ISD::FDIV;
2257 case URem: return ISD::UREM;
2258 case SRem: return ISD::SREM;
2259 case FRem: return ISD::FREM;
2260 case Shl: return ISD::SHL;
2261 case LShr: return ISD::SRL;
2262 case AShr: return ISD::SRA;
2263 case And: return ISD::AND;
2264 case Or: return ISD::OR;
2265 case Xor: return ISD::XOR;
2266 case Alloca: return 0;
2267 case Load: return ISD::LOAD;
2268 case Store: return ISD::STORE;
2269 case GetElementPtr: return 0;
2270 case Fence: return 0;
2271 case AtomicCmpXchg: return 0;
2272 case AtomicRMW: return 0;
2273 case Trunc: return ISD::TRUNCATE;
2274 case ZExt: return ISD::ZERO_EXTEND;
2275 case SExt: return ISD::SIGN_EXTEND;
2276 case FPToUI: return ISD::FP_TO_UINT;
2277 case FPToSI: return ISD::FP_TO_SINT;
2278 case UIToFP: return ISD::UINT_TO_FP;
2279 case SIToFP: return ISD::SINT_TO_FP;
2280 case FPTrunc: return ISD::FP_ROUND;
2281 case FPExt: return ISD::FP_EXTEND;
2282 case PtrToAddr: return ISD::BITCAST;
2283 case PtrToInt: return ISD::BITCAST;
2284 case IntToPtr: return ISD::BITCAST;
2285 case BitCast: return ISD::BITCAST;
2286 case AddrSpaceCast: return ISD::ADDRSPACECAST;
2287 case ICmp: return ISD::SETCC;
2288 case FCmp: return ISD::SETCC;
2289 case PHI: return 0;
2290 case Call: return 0;
2291 case Select: return ISD::SELECT;
2292 case UserOp1: return 0;
2293 case UserOp2: return 0;
2294 case VAArg: return 0;
2295 case ExtractElement: return ISD::EXTRACT_VECTOR_ELT;
2296 case InsertElement: return ISD::INSERT_VECTOR_ELT;
2297 case ShuffleVector: return ISD::VECTOR_SHUFFLE;
2298 case ExtractValue: return ISD::MERGE_VALUES;
2299 case InsertValue: return ISD::MERGE_VALUES;
2300 case LandingPad: return 0;
2301 case Freeze: return ISD::FREEZE;
2302 }
2303
2304 llvm_unreachable("Unknown instruction type encountered!");
2305}
2306
2308 switch (ID) {
2309 case Intrinsic::acos:
2310 return ISD::FACOS;
2311 case Intrinsic::asin:
2312 return ISD::FASIN;
2313 case Intrinsic::atan:
2314 return ISD::FATAN;
2315 case Intrinsic::cos:
2316 return ISD::FCOS;
2317 case Intrinsic::cosh:
2318 return ISD::FCOSH;
2319 case Intrinsic::exp:
2320 return ISD::FEXP;
2321 case Intrinsic::exp2:
2322 return ISD::FEXP2;
2323 case Intrinsic::exp10:
2324 return ISD::FEXP10;
2325 case Intrinsic::log:
2326 return ISD::FLOG;
2327 case Intrinsic::log2:
2328 return ISD::FLOG2;
2329 case Intrinsic::log10:
2330 return ISD::FLOG10;
2331 case Intrinsic::sin:
2332 return ISD::FSIN;
2333 case Intrinsic::sinh:
2334 return ISD::FSINH;
2335 case Intrinsic::tan:
2336 return ISD::FTAN;
2337 case Intrinsic::tanh:
2338 return ISD::FTANH;
2339 default:
2340 return ISD::DELETED_NODE;
2341 }
2342}
2343
2344Value *
2346 bool UseTLS) const {
2347 // compiler-rt provides a variable with a magic name. Targets that do not
2348 // link with compiler-rt may also provide such a variable.
2349 Module *M = IRB.GetInsertBlock()->getParent()->getParent();
2350
2351 RTLIB::LibcallImpl UnsafeStackPtrImpl =
2352 Libcalls.getLibcallImpl(RTLIB::SAFESTACK_UNSAFE_STACK_PTR);
2353 if (UnsafeStackPtrImpl == RTLIB::Unsupported)
2354 return nullptr;
2355
2356 StringRef UnsafeStackPtrVar =
2358 auto UnsafeStackPtr =
2359 dyn_cast_or_null<GlobalVariable>(M->getNamedValue(UnsafeStackPtrVar));
2360
2361 const DataLayout &DL = M->getDataLayout();
2362 PointerType *StackPtrTy = DL.getAllocaPtrType(M->getContext());
2363
2364 if (!UnsafeStackPtr) {
2365 auto TLSModel = UseTLS ?
2368 // The global variable is not defined yet, define it ourselves.
2369 // We use the initial-exec TLS model because we do not support the
2370 // variable living anywhere other than in the main executable.
2371 UnsafeStackPtr = new GlobalVariable(
2372 *M, StackPtrTy, false, GlobalValue::ExternalLinkage, nullptr,
2373 UnsafeStackPtrVar, nullptr, TLSModel);
2374 } else {
2375 // The variable exists, check its type and attributes.
2376 //
2377 // FIXME: Move to IR verifier.
2378 if (UnsafeStackPtr->getValueType() != StackPtrTy)
2379 report_fatal_error(Twine(UnsafeStackPtrVar) + " must have void* type");
2380 if (UseTLS != UnsafeStackPtr->isThreadLocal())
2381 report_fatal_error(Twine(UnsafeStackPtrVar) + " must " +
2382 (UseTLS ? "" : "not ") + "be thread-local");
2383 }
2384 return UnsafeStackPtr;
2385}
2386
2388 IRBuilderBase &IRB, const LibcallLoweringInfo &Libcalls) const {
2389 RTLIB::LibcallImpl SafestackPointerAddressImpl =
2390 Libcalls.getLibcallImpl(RTLIB::SAFESTACK_POINTER_ADDRESS);
2391 if (SafestackPointerAddressImpl == RTLIB::Unsupported)
2392 return getDefaultSafeStackPointerLocation(IRB, true);
2393
2394 Module *M = IRB.GetInsertBlock()->getParent()->getParent();
2395 auto *PtrTy = PointerType::getUnqual(M->getContext());
2396
2397 // Android provides a libc function to retrieve the address of the current
2398 // thread's unsafe stack pointer.
2399 FunctionCallee Fn =
2401 SafestackPointerAddressImpl),
2402 PtrTy);
2403 return IRB.CreateCall(Fn);
2404}
2405
2406//===----------------------------------------------------------------------===//
2407// Loop Strength Reduction hooks
2408//===----------------------------------------------------------------------===//
2409
2410/// isLegalAddressingMode - Return true if the addressing mode represented
2411/// by AM is legal for this target, for a load/store of the specified type.
2413 const AddrMode &AM, Type *Ty,
2414 unsigned AS, Instruction *I) const {
2415 // The default implementation of this implements a conservative RISCy, r+r and
2416 // r+i addr mode.
2417
2418 // Scalable offsets not supported
2419 if (AM.ScalableOffset)
2420 return false;
2421
2422 // Allows a sign-extended 16-bit immediate field.
2423 if (AM.BaseOffs <= -(1LL << 16) || AM.BaseOffs >= (1LL << 16)-1)
2424 return false;
2425
2426 // No global is ever allowed as a base.
2427 if (AM.BaseGV)
2428 return false;
2429
2430 // Only support r+r,
2431 switch (AM.Scale) {
2432 case 0: // "r+i" or just "i", depending on HasBaseReg.
2433 break;
2434 case 1:
2435 if (AM.HasBaseReg && AM.BaseOffs) // "r+r+i" is not allowed.
2436 return false;
2437 // Otherwise we have r+r or r+i.
2438 break;
2439 case 2:
2440 if (AM.HasBaseReg || AM.BaseOffs) // 2*r+r or 2*r+i is not allowed.
2441 return false;
2442 // Allow 2*r as r+r.
2443 break;
2444 default: // Don't allow n * r
2445 return false;
2446 }
2447
2448 return true;
2449}
2450
2451//===----------------------------------------------------------------------===//
2452// Stack Protector
2453//===----------------------------------------------------------------------===//
2454
2455// For OpenBSD return its special guard variable. Otherwise return nullptr,
2456// so that SelectionDAG handle SSP.
2457Value *
2459 const LibcallLoweringInfo &Libcalls) const {
2460 RTLIB::LibcallImpl GuardLocalImpl =
2461 Libcalls.getLibcallImpl(RTLIB::STACK_CHECK_GUARD);
2462 if (GuardLocalImpl != RTLIB::impl___guard_local)
2463 return nullptr;
2464
2465 Module &M = *IRB.GetInsertBlock()->getParent()->getParent();
2466 const DataLayout &DL = M.getDataLayout();
2467 PointerType *PtrTy =
2468 PointerType::get(M.getContext(), DL.getDefaultGlobalsAddressSpace());
2469 GlobalVariable *G =
2470 M.getOrInsertGlobal(getLibcallImplName(GuardLocalImpl), PtrTy);
2471 G->setVisibility(GlobalValue::HiddenVisibility);
2472 return G;
2473}
2474
2475// Currently only support "standard" __stack_chk_guard.
2476// TODO: add LOAD_STACK_GUARD support.
2478 Module &M, const LibcallLoweringInfo &Libcalls) const {
2479 RTLIB::LibcallImpl StackGuardImpl =
2480 Libcalls.getLibcallImpl(RTLIB::STACK_CHECK_GUARD);
2481 if (StackGuardImpl == RTLIB::Unsupported)
2482 return;
2483
2484 StringRef StackGuardVarName = getLibcallImplName(StackGuardImpl);
2485 M.getOrInsertGlobal(
2486 StackGuardVarName, PointerType::getUnqual(M.getContext()), [=, &M]() {
2487 auto *GV = new GlobalVariable(M, PointerType::getUnqual(M.getContext()),
2488 false, GlobalVariable::ExternalLinkage,
2489 nullptr, StackGuardVarName);
2490
2491 // FreeBSD has "__stack_chk_guard" defined externally on libc.so
2492 if (M.getDirectAccessExternalData() &&
2493 !TM.getTargetTriple().isOSCygMing() &&
2494 !(TM.getTargetTriple().isPPC64() &&
2495 TM.getTargetTriple().isOSFreeBSD()) &&
2496 (!TM.getTargetTriple().isOSDarwin() ||
2497 TM.getRelocationModel() == Reloc::Static))
2498 GV->setDSOLocal(true);
2499
2500 return GV;
2501 });
2502}
2503
2504// Currently only support "standard" __stack_chk_guard.
2505// TODO: add LOAD_STACK_GUARD support.
2507 const Module &M, const LibcallLoweringInfo &Libcalls) const {
2508 RTLIB::LibcallImpl GuardVarImpl =
2509 Libcalls.getLibcallImpl(RTLIB::STACK_CHECK_GUARD);
2510 if (GuardVarImpl == RTLIB::Unsupported)
2511 return nullptr;
2512 return M.getNamedValue(getLibcallImplName(GuardVarImpl));
2513}
2514
2516 const Module &M, const LibcallLoweringInfo &Libcalls) const {
2517 // MSVC CRT has a function to validate security cookie.
2518 RTLIB::LibcallImpl SecurityCheckCookieLibcall =
2519 Libcalls.getLibcallImpl(RTLIB::SECURITY_CHECK_COOKIE);
2520 if (SecurityCheckCookieLibcall != RTLIB::Unsupported)
2521 return M.getFunction(getLibcallImplName(SecurityCheckCookieLibcall));
2522 return nullptr;
2523}
2524
2528
2532
2533unsigned TargetLoweringBase::getMinimumJumpTableDensity(bool OptForSize) const {
2534 return OptForSize ? OptsizeJumpTableDensity : JumpTableDensity;
2535}
2536
2540
2544
2548
2550 return MinimumBitTestCmps;
2551}
2552
2554 MinimumBitTestCmps = Val;
2555}
2556
2558 if (TM.Options.LoopAlignment)
2559 return Align(TM.Options.LoopAlignment);
2560 return PrefLoopAlignment;
2561}
2562
2564 MachineBasicBlock *MBB) const {
2565 return MaxBytesForAlignment;
2566}
2567
2568//===----------------------------------------------------------------------===//
2569// Reciprocal Estimates
2570//===----------------------------------------------------------------------===//
2571
2572/// Get the reciprocal estimate attribute string for a function that will
2573/// override the target defaults.
2575 const Function &F = MF.getFunction();
2576 return F.getFnAttribute("reciprocal-estimates").getValueAsString();
2577}
2578
2579/// Construct a string for the given reciprocal operation of the given type.
2580/// This string should match the corresponding option to the front-end's
2581/// "-mrecip" flag assuming those strings have been passed through in an
2582/// attribute string. For example, "vec-divf" for a division of a vXf32.
2583static std::string getReciprocalOpName(bool IsSqrt, EVT VT) {
2584 std::string Name = VT.isVector() ? "vec-" : "";
2585
2586 Name += IsSqrt ? "sqrt" : "div";
2587
2588 // TODO: Handle other float types?
2589 if (VT.getScalarType() == MVT::f64) {
2590 Name += "d";
2591 } else if (VT.getScalarType() == MVT::f16) {
2592 Name += "h";
2593 } else {
2594 assert(VT.getScalarType() == MVT::f32 &&
2595 "Unexpected FP type for reciprocal estimate");
2596 Name += "f";
2597 }
2598
2599 return Name;
2600}
2601
2602/// Return the character position and value (a single numeric character) of a
2603/// customized refinement operation in the input string if it exists. Return
2604/// false if there is no customized refinement step count.
2605static bool parseRefinementStep(StringRef In, size_t &Position,
2606 uint8_t &Value) {
2607 const char RefStepToken = ':';
2608 Position = In.find(RefStepToken);
2609 if (Position == StringRef::npos)
2610 return false;
2611
2612 StringRef RefStepString = In.substr(Position + 1);
2613 // Allow exactly one numeric character for the additional refinement
2614 // step parameter.
2615 if (RefStepString.size() == 1) {
2616 char RefStepChar = RefStepString[0];
2617 if (isDigit(RefStepChar)) {
2618 Value = RefStepChar - '0';
2619 return true;
2620 }
2621 }
2622 report_fatal_error("Invalid refinement step for -recip.");
2623}
2624
2625/// For the input attribute string, return one of the ReciprocalEstimate enum
2626/// status values (enabled, disabled, or not specified) for this operation on
2627/// the specified data type.
2628static int getOpEnabled(bool IsSqrt, EVT VT, StringRef Override) {
2629 if (Override.empty())
2631
2632 SmallVector<StringRef, 4> OverrideVector;
2633 Override.split(OverrideVector, ',');
2634 unsigned NumArgs = OverrideVector.size();
2635
2636 // Check if "all", "none", or "default" was specified.
2637 if (NumArgs == 1) {
2638 // Look for an optional setting of the number of refinement steps needed
2639 // for this type of reciprocal operation.
2640 size_t RefPos;
2641 uint8_t RefSteps;
2642 if (parseRefinementStep(Override, RefPos, RefSteps)) {
2643 // Split the string for further processing.
2644 Override = Override.substr(0, RefPos);
2645 }
2646
2647 // All reciprocal types are enabled.
2648 if (Override == "all")
2650
2651 // All reciprocal types are disabled.
2652 if (Override == "none")
2654
2655 // Target defaults for enablement are used.
2656 if (Override == "default")
2658 }
2659
2660 // The attribute string may omit the size suffix ('f'/'d').
2661 std::string VTName = getReciprocalOpName(IsSqrt, VT);
2662 std::string VTNameNoSize = VTName;
2663 VTNameNoSize.pop_back();
2664 static const char DisabledPrefix = '!';
2665
2666 for (StringRef RecipType : OverrideVector) {
2667 size_t RefPos;
2668 uint8_t RefSteps;
2669 if (parseRefinementStep(RecipType, RefPos, RefSteps))
2670 RecipType = RecipType.substr(0, RefPos);
2671
2672 // Ignore the disablement token for string matching.
2673 bool IsDisabled = RecipType[0] == DisabledPrefix;
2674 if (IsDisabled)
2675 RecipType = RecipType.substr(1);
2676
2677 if (RecipType == VTName || RecipType == VTNameNoSize)
2680 }
2681
2683}
2684
2685/// For the input attribute string, return the customized refinement step count
2686/// for this operation on the specified data type. If the step count does not
2687/// exist, return the ReciprocalEstimate enum value for unspecified.
2688static int getOpRefinementSteps(bool IsSqrt, EVT VT, StringRef Override) {
2689 if (Override.empty())
2691
2692 SmallVector<StringRef, 4> OverrideVector;
2693 Override.split(OverrideVector, ',');
2694 unsigned NumArgs = OverrideVector.size();
2695
2696 // Check if "all", "default", or "none" was specified.
2697 if (NumArgs == 1) {
2698 // Look for an optional setting of the number of refinement steps needed
2699 // for this type of reciprocal operation.
2700 size_t RefPos;
2701 uint8_t RefSteps;
2702 if (!parseRefinementStep(Override, RefPos, RefSteps))
2704
2705 // Split the string for further processing.
2706 Override = Override.substr(0, RefPos);
2707 assert(Override != "none" &&
2708 "Disabled reciprocals, but specifed refinement steps?");
2709
2710 // If this is a general override, return the specified number of steps.
2711 if (Override == "all" || Override == "default")
2712 return RefSteps;
2713 }
2714
2715 // The attribute string may omit the size suffix ('f'/'d').
2716 std::string VTName = getReciprocalOpName(IsSqrt, VT);
2717 std::string VTNameNoSize = VTName;
2718 VTNameNoSize.pop_back();
2719
2720 for (StringRef RecipType : OverrideVector) {
2721 size_t RefPos;
2722 uint8_t RefSteps;
2723 if (!parseRefinementStep(RecipType, RefPos, RefSteps))
2724 continue;
2725
2726 RecipType = RecipType.substr(0, RefPos);
2727 if (RecipType == VTName || RecipType == VTNameNoSize)
2728 return RefSteps;
2729 }
2730
2732}
2733
2738
2743
2748
2753
2755 EVT LoadVT, EVT BitcastVT, const SelectionDAG &DAG,
2756 const MachineMemOperand &MMO) const {
2757 // Single-element vectors are scalarized, so we should generally avoid having
2758 // any memory operations on such types, as they would get scalarized too.
2759 if (LoadVT.isFixedLengthVector() && BitcastVT.isFixedLengthVector() &&
2760 BitcastVT.getVectorNumElements() == 1)
2761 return false;
2762
2763 // Don't do if we could do an indexed load on the original type, but not on
2764 // the new one.
2765 if (!LoadVT.isSimple() || !BitcastVT.isSimple())
2766 return true;
2767
2768 MVT LoadMVT = LoadVT.getSimpleVT();
2769
2770 // Don't bother doing this if it's just going to be promoted again later, as
2771 // doing so might interfere with other combines.
2772 if (getOperationAction(ISD::LOAD, LoadMVT) == Promote &&
2773 getTypeToPromoteTo(ISD::LOAD, LoadMVT) == BitcastVT.getSimpleVT())
2774 return false;
2775
2776 unsigned Fast = 0;
2777 return allowsMemoryAccess(*DAG.getContext(), DAG.getDataLayout(), BitcastVT,
2778 MMO, &Fast) &&
2779 Fast;
2780}
2781
2785
2787 const LoadInst &LI, const DataLayout &DL, AssumptionCache *AC,
2788 const TargetLibraryInfo *LibInfo, CodeGenOptLevel OptLevel) const {
2790 if (LI.isVolatile())
2792
2793 if (LI.hasMetadata(LLVMContext::MD_nontemporal))
2795
2796 if (LI.hasMetadata(LLVMContext::MD_invariant_load))
2798
2799 // Dereferenceability analysis is expensive, skip at O0.
2800 if (OptLevel != CodeGenOptLevel::None &&
2802 LI.getAlign(), DL, &LI, AC,
2803 /*DT=*/nullptr, LibInfo)) {
2805 } else if (LI.hasMetadata(LLVMContext::MD_dereferenceable)) {
2807 }
2808
2809 Flags |= getTargetMMOFlags(LI);
2810 return Flags;
2811}
2812
2815 const DataLayout &DL) const {
2817
2818 if (SI.isVolatile())
2820
2821 if (SI.hasMetadata(LLVMContext::MD_nontemporal))
2823
2824 // FIXME: Not preserving dereferenceable
2825 Flags |= getTargetMMOFlags(SI);
2826 return Flags;
2827}
2828
2831 const DataLayout &DL) const {
2833
2834 if (const AtomicRMWInst *RMW = dyn_cast<AtomicRMWInst>(&AI)) {
2835 if (RMW->isVolatile())
2837 } else if (const AtomicCmpXchgInst *CmpX = dyn_cast<AtomicCmpXchgInst>(&AI)) {
2838 if (CmpX->isVolatile())
2840 } else
2841 llvm_unreachable("not an atomic instruction");
2842
2843 // FIXME: Not preserving dereferenceable
2844 Flags |= getTargetMMOFlags(AI);
2845 return Flags;
2846}
2847
2849 const VPIntrinsic &VPIntrin) const {
2851 Intrinsic::ID IntrinID = VPIntrin.getIntrinsicID();
2852
2853 switch (IntrinID) {
2854 default:
2855 llvm_unreachable("unexpected intrinsic. Existing code may be appropriate "
2856 "for it, but support must be explicitly enabled");
2857 case Intrinsic::vp_load:
2858 case Intrinsic::vp_gather:
2859 case Intrinsic::experimental_vp_strided_load:
2861 break;
2862 case Intrinsic::vp_store:
2863 case Intrinsic::vp_scatter:
2864 case Intrinsic::experimental_vp_strided_store:
2866 break;
2867 }
2868
2869 if (VPIntrin.hasMetadata(LLVMContext::MD_nontemporal))
2871
2872 Flags |= getTargetMMOFlags(VPIntrin);
2873 return Flags;
2874}
2875
2877 Instruction *Inst,
2878 AtomicOrdering Ord) const {
2879 if (isReleaseOrStronger(Ord) && Inst->hasAtomicStore())
2880 return Builder.CreateFence(Ord);
2881 else
2882 return nullptr;
2883}
2884
2886 Instruction *Inst,
2887 AtomicOrdering Ord) const {
2888 if (isAcquireOrStronger(Ord))
2889 return Builder.CreateFence(Ord);
2890 else
2891 return nullptr;
2892}
2893
2894//===----------------------------------------------------------------------===//
2895// GlobalISel Hooks
2896//===----------------------------------------------------------------------===//
2897
2899 const TargetTransformInfo *TTI) const {
2900 auto &MF = *MI.getMF();
2901 auto &MRI = MF.getRegInfo();
2902 // Assuming a spill and reload of a value has a cost of 1 instruction each,
2903 // this helper function computes the maximum number of uses we should consider
2904 // for remat. E.g. on arm64 global addresses take 2 insts to materialize. We
2905 // break even in terms of code size when the original MI has 2 users vs
2906 // choosing to potentially spill. Any more than 2 users we we have a net code
2907 // size increase. This doesn't take into account register pressure though.
2908 auto maxUses = [](unsigned RematCost) {
2909 // A cost of 1 means remats are basically free.
2910 if (RematCost == 1)
2911 return std::numeric_limits<unsigned>::max();
2912 if (RematCost == 2)
2913 return 2U;
2914
2915 // Remat is too expensive, only sink if there's one user.
2916 if (RematCost > 2)
2917 return 1U;
2918 llvm_unreachable("Unexpected remat cost");
2919 };
2920
2921 switch (MI.getOpcode()) {
2922 default:
2923 return false;
2924 // Constants-like instructions should be close to their users.
2925 // We don't want long live-ranges for them.
2926 case TargetOpcode::G_CONSTANT:
2927 case TargetOpcode::G_FCONSTANT:
2928 case TargetOpcode::G_FRAME_INDEX:
2929 case TargetOpcode::G_INTTOPTR:
2930 return true;
2931 case TargetOpcode::G_GLOBAL_VALUE: {
2932 unsigned RematCost = TTI->getGISelRematGlobalCost();
2933 Register Reg = MI.getOperand(0).getReg();
2934 unsigned MaxUses = maxUses(RematCost);
2935 if (MaxUses == UINT_MAX)
2936 return true; // Remats are "free" so always localize.
2937 return MRI.hasAtMostUserInstrs(Reg, MaxUses);
2938 }
2939 }
2940}
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
AMDGPU Register Bank Select
Rewrite undef for PHI
MachineBasicBlock & MBB
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
This file contains the simple types necessary to represent the attributes associated with functions a...
This file implements the BitVector class.
#define LLVM_ABI
Definition Compiler.h:213
This file defines the DenseMap class.
IRTranslator LLVM IR MI
Module.h This file contains the declarations for the Module class.
static LVOptions Options
Definition LVOptions.cpp:25
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
#define G(x, y, z)
Definition MD5.cpp:55
Register const TargetRegisterInfo * TRI
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
This file contains some templates that are useful if you are working with the STL at all.
This file defines the SmallVector class.
This file contains some functions that are useful when dealing with strings.
static cl::opt< unsigned > MinimumBitTestCmpsOverride("min-bit-test-cmps", cl::init(2), cl::Hidden, cl::desc("Set minimum of largest number of comparisons " "to use bit test for switch."))
static cl::opt< bool > JumpIsExpensiveOverride("jump-is-expensive", cl::init(false), cl::desc("Do not create extra branches to split comparison logic."), cl::Hidden)
#define OP_TO_LIBCALL(Name, Enum)
static cl::opt< unsigned > MinimumJumpTableEntries("min-jump-table-entries", cl::init(4), cl::Hidden, cl::desc("Set minimum number of entries to use a jump table."))
static cl::opt< bool > DisableStrictNodeMutation("disable-strictnode-mutation", cl::desc("Don't mutate strict-float node to a legalize node"), cl::init(false), cl::Hidden)
static bool parseRefinementStep(StringRef In, size_t &Position, uint8_t &Value)
Return the character position and value (a single numeric character) of a customized refinement opera...
static cl::opt< unsigned > MaximumJumpTableSize("max-jump-table-size", cl::init(UINT_MAX), cl::Hidden, cl::desc("Set maximum size of jump tables."))
static cl::opt< unsigned > JumpTableDensity("jump-table-density", cl::init(10), cl::Hidden, cl::desc("Minimum density for building a jump table in " "a normal function"))
Minimum jump table density for normal functions.
static unsigned getVectorTypeBreakdownMVT(MVT VT, MVT &IntermediateVT, unsigned &NumIntermediates, MVT &RegisterVT, TargetLoweringBase *TLI)
static cl::opt< unsigned > MaxStoresPerMemmoveOverride("max-store-memmove", cl::init(0), cl::Hidden, cl::desc("Override target's MaxStoresPerMemmove and " "MaxStoresPerMemmoveOptSize. " "Set to 0 to use the target default."))
static std::string getReciprocalOpName(bool IsSqrt, EVT VT)
Construct a string for the given reciprocal operation of the given type.
#define LCALL5(A)
static cl::opt< unsigned > MaxStoresPerMemsetOverride("max-store-memset", cl::init(0), cl::Hidden, cl::desc("Override target's MaxStoresPerMemset and " "MaxStoresPerMemsetOptSize. " "Set to 0 to use the target default."))
static int getOpRefinementSteps(bool IsSqrt, EVT VT, StringRef Override)
For the input attribute string, return the customized refinement step count for this operation on the...
static int getOpEnabled(bool IsSqrt, EVT VT, StringRef Override)
For the input attribute string, return one of the ReciprocalEstimate enum status values (enabled,...
static StringRef getRecipEstimateForFunc(MachineFunction &MF)
Get the reciprocal estimate attribute string for a function that will override the target defaults.
static cl::opt< unsigned > MaxStoresPerMemcpyOverride("max-store-memcpy", cl::init(0), cl::Hidden, cl::desc("Override target's MaxStoresPerMemcpy and " "MaxStoresPerMemcpyOptSize. " "Set to 0 to use the target default."))
static cl::opt< unsigned > OptsizeJumpTableDensity("optsize-jump-table-density", cl::init(40), cl::Hidden, cl::desc("Minimum density for building a jump table in " "an optsize function"))
Minimum jump table density for -Os or -Oz functions.
This file describes how to lower LLVM code to machine code.
This pass exposes codegen information to IR-level passes.
Class for arbitrary precision integers.
Definition APInt.h:78
A cache of @llvm.assume calls within a function.
An instruction that atomically checks whether a specified value is in a memory location,...
an instruction that atomically reads a memory location, combines it with another value,...
const Function * getParent() const
Return the enclosing method, or null if none.
Definition BasicBlock.h:213
void setBitsInMask(const uint32_t *Mask, unsigned MaskWords=~0u)
Add '1' bits from Mask to this vector.
Definition BitVector.h:730
iterator_range< const_set_bits_iterator > set_bits() const
Definition BitVector.h:159
BlockFrequencyInfo pass uses BlockFrequencyInfoImpl implementation to estimate IR basic block frequen...
This class represents a range of values.
LLVM_ABI unsigned getActiveBits() const
Compute the maximal number of active bits needed to represent every value in this range.
LLVM_ABI ConstantRange umul_sat(const ConstantRange &Other) const
Perform an unsigned saturating multiplication of two constant ranges.
LLVM_ABI ConstantRange subtract(const APInt &CI) const
Subtract the specified constant from the endpoints of this constant range.
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:64
LLVM_ABI unsigned getPointerSize(unsigned AS=0) const
The pointer representation size in bytes, rounded up to a whole number of bytes.
static constexpr ElementCount getScalable(ScalarTy MinVal)
Definition TypeSize.h:312
static constexpr ElementCount getFixed(ScalarTy MinVal)
Definition TypeSize.h:309
constexpr bool isScalar() const
Exactly one element.
Definition TypeSize.h:320
A handy container for a FunctionType+Callee-pointer pair, which can be passed around as a single enti...
const Function & getFunction() const
Definition Function.h:166
Module * getParent()
Get the module that this global value is contained inside of...
@ HiddenVisibility
The GV is hidden.
Definition GlobalValue.h:69
@ ExternalLinkage
Externally visible function.
Definition GlobalValue.h:53
Common base class shared among various IRBuilders.
Definition IRBuilder.h:114
BasicBlock * GetInsertBlock() const
Definition IRBuilder.h:201
CallInst * CreateCall(FunctionType *FTy, Value *Callee, ArrayRef< Value * > Args={}, const Twine &Name="", MDNode *FPMathTag=nullptr)
Definition IRBuilder.h:2553
LLVM_ABI bool hasAtomicStore() const LLVM_READONLY
Return true if this atomic instruction stores to memory.
bool hasMetadata() const
Return true if this instruction has any metadata attached to it.
Intrinsic::ID getIntrinsicID() const
Return the intrinsic ID of this intrinsic.
This is an important class for using LLVM in a threaded context.
Definition LLVMContext.h:68
Tracks which library functions to use for a particular subtarget.
An instruction for reading from memory.
Value * getPointerOperand()
bool isVolatile() const
Return true if this is a load from a volatile memory location.
Align getAlign() const
Return the alignment of the access that is being performed.
Machine Value Type.
SimpleValueType SimpleTy
uint64_t getScalarSizeInBits() const
bool isVector() const
Return true if this is a vector value type.
bool isScalableVector() const
Return true if this is a vector value type where the runtime length is machine dependent.
static auto all_valuetypes()
SimpleValueType Iteration.
TypeSize getSizeInBits() const
Returns the size of the specified MVT in bits.
uint64_t getFixedSizeInBits() const
Return the size of the specified fixed width value type in bits.
ElementCount getVectorElementCount() const
bool isScalarInteger() const
Return true if this is an integer, not including vectors.
static MVT getVectorVT(MVT VT, unsigned NumElements)
MVT getVectorElementType() const
bool isValid() const
Return true if this is a valid simple valuetype.
static MVT getIntegerVT(unsigned BitWidth)
static auto fp_valuetypes()
MVT getPow2VectorType() const
Widens the length of the given vector MVT up to the nearest power of 2 and returns that type.
MachineInstrBundleIterator< MachineInstr > iterator
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
bool isStatepointSpillSlotObjectIndex(int ObjectIdx) const
Align getObjectAlign(int ObjectIdx) const
Return the alignment of the specified stack object.
int64_t getObjectSize(int ObjectIdx) const
Return the size of the specified object.
int64_t getObjectOffset(int ObjectIdx) const
Return the assigned stack offset of the specified object from the incoming stack pointer.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
const DataLayout & getDataLayout() const
Return the DataLayout attached to the Module associated to this MF.
Function & getFunction()
Return the LLVM function that this machine code represents.
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & cloneMemRefs(const MachineInstr &OtherMI) const
Representation of each machine instruction.
unsigned getNumOperands() const
Retuns the total number of operands.
bool mayLoad(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read memory.
LLVM_ABI void tieOperands(unsigned DefIdx, unsigned UseIdx)
Add a tie between the register operands at DefIdx and UseIdx.
LLVM_ABI void addMemOperand(MachineFunction &MF, MachineMemOperand *MO)
Add a MachineMemOperand to the machine instruction.
A description of a memory reference used in the backend.
unsigned getAddrSpace() const
Flags
Flags values. These may be or'd together.
@ MOVolatile
The memory access is volatile.
@ MODereferenceable
The memory access is dereferenceable (i.e., doesn't trap).
@ MOLoad
The memory access reads data.
@ MONonTemporal
The memory access is non-temporal.
@ MOInvariant
The memory access always returns the same value (or traps).
@ MOStore
The memory access writes data.
Flags getFlags() const
Return the raw flags of the source value,.
LLVM_ABI Align getAlign() const
Return the minimum known alignment in bytes of the actual memory reference.
MachineOperand class - Representation of each machine instruction operand.
bool isReg() const
isReg - Tests if this is a MO_Register operand.
bool isFI() const
isFI - Tests if this is a MO_FrameIndex operand.
LLVM_ABI void freezeReservedRegs()
freezeReservedRegs - Called by the register allocator to freeze the set of reserved registers before ...
A Module instance is used to store all the information related to an LLVM module.
Definition Module.h:67
Class to represent pointers.
static PointerType * getUnqual(Type *ElementType)
This constructs a pointer to an object of the specified type in the default address space (address sp...
static LLVM_ABI PointerType * get(Type *ElementType, unsigned AddressSpace)
This constructs a pointer to an object of the specified type in a numbered address space.
Analysis providing profile information.
Wrapper class representing virtual and physical registers.
Definition Register.h:20
This is used to represent a portion of an LLVM function in a low-level Data Dependence DAG representa...
const DataLayout & getDataLayout() const
LLVMContext * getContext() const
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
An instruction for storing to memory.
Represent a constant reference to a string, i.e.
Definition StringRef.h:56
std::pair< StringRef, StringRef > split(char Separator) const
Split into two substrings around the first occurrence of a separator character.
Definition StringRef.h:730
static constexpr size_t npos
Definition StringRef.h:58
constexpr StringRef substr(size_t Start, size_t N=npos) const
Return a reference to the substring from [Start, Start + N).
Definition StringRef.h:591
constexpr bool empty() const
Check if the string is empty.
Definition StringRef.h:141
constexpr size_t size() const
Get the string size.
Definition StringRef.h:144
bool isValid() const
Returns true if this iterator is still pointing at a valid entry.
Multiway switch.
Provides information about what library functions are available for the current target.
This base class for TargetLowering contains the SelectionDAG-independent parts that can be used from ...
virtual Align getByValTypeAlignment(Type *Ty, const DataLayout &DL) const
Returns the desired alignment for ByVal or InAlloca aggregate function arguments in the caller parame...
int InstructionOpcodeToISD(unsigned Opcode) const
Get the ISD node that corresponds to the Instruction class opcode.
unsigned getBitWidthForCttzElements(EVT RetVT, ElementCount EC, bool ZeroIsPoison, const ConstantRange *VScaleRange) const
Return the minimum number of bits required to hold the maximum possible number of trailing zero vecto...
void setOperationAction(unsigned Op, MVT VT, LegalizeAction Action)
Indicate that the specified operation does not work with the specified type and indicate what to do a...
virtual void finalizeLowering(MachineFunction &MF) const
Execute target specific actions to finalize target lowering.
void initActions()
Initialize all of the actions to default values.
bool PredictableSelectIsExpensive
Tells the code generator that select is more expensive than a branch if the branch is usually predict...
Function * getSSPStackGuardCheck(const Module &M, const LibcallLoweringInfo &Libcalls) const
If the target has a standard stack protection check function that performs validation and error handl...
EVT getValueType(const DataLayout &DL, Type *Ty, bool AllowUnknown=false) const
Return the EVT corresponding to this LLVM type.
void setMinimumBitTestCmps(unsigned Val)
Set the minimum of largest of number of comparisons to generate BitTest.
unsigned MaxStoresPerMemcpyOptSize
Likewise for functions with the OptSize attribute.
MachineBasicBlock * emitPatchPoint(MachineInstr &MI, MachineBasicBlock *MBB) const
Replace/modify any TargetFrameIndex operands with a targte-dependent sequence of memory operands that...
int getRecipEstimateSqrtEnabled(EVT VT, MachineFunction &MF) const
Return a ReciprocalEstimate enum value for a square root of the given type based on the function's at...
virtual bool canOpTrap(unsigned Op, EVT VT) const
Returns true if the operation can trap for the value type.
virtual Value * getIRStackGuard(IRBuilderBase &IRB, const LibcallLoweringInfo &Libcalls) const
If the target has a standard location for the stack protector guard, returns the address of that loca...
virtual bool shouldLocalize(const MachineInstr &MI, const TargetTransformInfo *TTI) const
Check whether or not MI needs to be moved close to its uses.
virtual unsigned getMaxPermittedBytesForAlignment(MachineBasicBlock *MBB) const
Return the maximum amount of bytes allowed to be emitted when padding for alignment.
void setMaximumJumpTableSize(unsigned)
Indicate the maximum number of entries in jump tables.
virtual unsigned getMinimumJumpTableEntries() const
Return lower limit for number of blocks in a jump table.
const TargetMachine & getTargetMachine() const
unsigned MaxLoadsPerMemcmp
Specify maximum number of load instructions per memcmp call.
virtual unsigned getNumRegistersForCallingConv(LLVMContext &Context, CallingConv::ID CC, EVT VT) const
Certain targets require unusual breakdowns of certain types.
virtual MachineMemOperand::Flags getTargetMMOFlags(const Instruction &I) const
This callback is used to inspect load/store instructions and add target-specific MachineMemOperand fl...
unsigned MaxGluedStoresPerMemcpy
Specify max number of store instructions to glue in inlined memcpy.
virtual MVT getRegisterTypeForCallingConv(LLVMContext &Context, CallingConv::ID CC, EVT VT) const
Certain combinations of ABIs, Targets and features require that types are legal for some operations a...
LegalizeTypeAction
This enum indicates whether a types are legal for a target, and if not, what action should be used to...
virtual void insertSSPDeclarations(Module &M, const LibcallLoweringInfo &Libcalls) const
Inserts necessary declarations for SSP (stack protection) purpose.
virtual bool isSuitableForJumpTable(const SwitchInst *SI, uint64_t NumCases, uint64_t Range, ProfileSummaryInfo *PSI, BlockFrequencyInfo *BFI) const
Return true if lowering to a jump table is suitable for a set of case clusters which may contain NumC...
void setIndexedMaskedLoadAction(unsigned IdxMode, MVT VT, LegalizeAction Action)
Indicate that the specified indexed masked load does or does not work with the specified type and ind...
unsigned getMaxStoresPerMemcpy(bool OptSize) const
Get maximum # of store operations permitted for llvm.memcpy.
unsigned getMinimumBitTestCmps() const
Retuen the minimum of largest number of comparisons in BitTest.
virtual bool useFPRegsForHalfType() const
virtual bool isLoadBitCastBeneficial(EVT LoadVT, EVT BitcastVT, const SelectionDAG &DAG, const MachineMemOperand &MMO) const
Return true if the following transform is beneficial: fold (conv (load x)) -> (load (conv*)x) On arch...
void setIndexedLoadAction(ArrayRef< unsigned > IdxModes, MVT VT, LegalizeAction Action)
Indicate that the specified indexed load does or does not work with the specified type and indicate w...
unsigned getMaximumJumpTableSize() const
Return upper limit for number of entries in a jump table.
MachineMemOperand::Flags getLoadMemOperandFlags(const LoadInst &LI, const DataLayout &DL, AssumptionCache *AC=nullptr, const TargetLibraryInfo *LibInfo=nullptr, CodeGenOptLevel OptLevel=CodeGenOptLevel::Default) const
bool isLegalRC(const TargetRegisterInfo &TRI, const TargetRegisterClass &RC) const
Return true if the value types that can be represented by the specified register class are all legal.
virtual TargetLoweringBase::LegalizeTypeAction getPreferredVectorAction(MVT VT) const
Return the preferred vector type legalization action.
void setAtomicLoadExtAction(unsigned ExtType, MVT ValVT, MVT MemVT, LegalizeAction Action)
Let target indicate that an extending atomic load of the specified type is legal.
Value * getDefaultSafeStackPointerLocation(IRBuilderBase &IRB, bool UseTLS) const
unsigned getMaxStoresPerMemset(bool OptSize) const
Get maximum # of store operations permitted for llvm.memset.
MachineMemOperand::Flags getAtomicMemOperandFlags(const Instruction &AI, const DataLayout &DL) const
virtual bool allowsMisalignedMemoryAccesses(EVT, unsigned AddrSpace=0, Align Alignment=Align(1), MachineMemOperand::Flags Flags=MachineMemOperand::MONone, unsigned *=nullptr) const
Determine if the target supports unaligned memory accesses.
unsigned MaxStoresPerMemsetOptSize
Likewise for functions with the OptSize attribute.
EVT getShiftAmountTy(EVT LHSTy, const DataLayout &DL) const
Returns the type for the shift amount of a shift opcode.
unsigned MaxStoresPerMemmove
Specify maximum number of store instructions per memmove call.
virtual Align getPrefLoopAlignment(MachineLoop *ML=nullptr) const
Return the preferred loop alignment.
void computeRegisterProperties(const TargetRegisterInfo *TRI)
Once all of the register classes are added, this allows us to compute derived properties we expose.
MachineMemOperand::Flags getVPIntrinsicMemOperandFlags(const VPIntrinsic &VPIntrin) const
int getDivRefinementSteps(EVT VT, MachineFunction &MF) const
Return the refinement step count for a division of the given type based on the function's attributes.
virtual EVT getSetCCResultType(const DataLayout &DL, LLVMContext &Context, EVT VT) const
Return the ValueType of the result of SETCC operations.
virtual EVT getTypeToTransformTo(LLVMContext &Context, EVT VT) const
For types supported by the target, this is an identity function.
unsigned MaxStoresPerMemmoveOptSize
Likewise for functions with the OptSize attribute.
virtual MVT getPreferredSwitchConditionType(LLVMContext &Context, EVT ConditionVT) const
Returns preferred type for switch condition.
bool isTypeLegal(EVT VT) const
Return true if the target has native support for the specified value type.
int getRecipEstimateDivEnabled(EVT VT, MachineFunction &MF) const
Return a ReciprocalEstimate enum value for a division of the given type based on the function's attri...
void setIndexedStoreAction(ArrayRef< unsigned > IdxModes, MVT VT, LegalizeAction Action)
Indicate that the specified indexed store does or does not work with the specified type and indicate ...
virtual bool isJumpTableRelative() const
virtual MVT getScalarShiftAmountTy(const DataLayout &, EVT) const
Return the type to use for a scalar shift opcode, given the shifted amount type.
virtual MVT getPointerTy(const DataLayout &DL, uint32_t AS=0) const
Return the pointer type for the given address space, defaults to the pointer type from the data layou...
virtual bool isFreeAddrSpaceCast(unsigned SrcAS, unsigned DestAS) const
Returns true if a cast from SrcAS to DestAS is "cheap", such that e.g.
ISD::CondCode getSoftFloatCmpLibcallPredicate(RTLIB::LibcallImpl Call) const
Get the comparison predicate that's to be used to test the result of the comparison libcall against z...
void setIndexedMaskedStoreAction(unsigned IdxMode, MVT VT, LegalizeAction Action)
Indicate that the specified indexed masked store does or does not work with the specified type and in...
TargetLoweringBase(const TargetMachine &TM, const TargetSubtargetInfo &STI)
NOTE: The TargetMachine owns TLOF.
unsigned MaxStoresPerMemset
Specify maximum number of store instructions per memset call.
void setMinimumJumpTableEntries(unsigned Val)
Indicate the minimum number of blocks to generate jump tables.
void setTruncStoreAction(MVT ValVT, MVT MemVT, LegalizeAction Action)
Indicate that the specified truncating store does not work with the specified type and indicate what ...
virtual bool allowsMemoryAccess(LLVMContext &Context, const DataLayout &DL, EVT VT, unsigned AddrSpace=0, Align Alignment=Align(1), MachineMemOperand::Flags Flags=MachineMemOperand::MONone, unsigned *Fast=nullptr) const
Return true if the target supports a memory access of this type for the given address space and align...
unsigned MaxLoadsPerMemcmpOptSize
Likewise for functions with the OptSize attribute.
MachineMemOperand::Flags getStoreMemOperandFlags(const StoreInst &SI, const DataLayout &DL) const
void AddPromotedToType(unsigned Opc, MVT OrigVT, MVT DestVT)
If Opc/OrigVT is specified as being promoted, the promotion code defaults to trying a larger integer/...
unsigned getMinimumJumpTableDensity(bool OptForSize) const
Return lower limit of the density in a jump table.
virtual Value * getSDagStackGuard(const Module &M, const LibcallLoweringInfo &Libcalls) const
Return the variable that's previously inserted by insertSSPDeclarations, if any, otherwise return nul...
virtual std::pair< const TargetRegisterClass *, uint8_t > findRepresentativeClass(const TargetRegisterInfo *TRI, MVT VT) const
Return the largest legal super-reg register class of the register class for the specified type and it...
static StringRef getLibcallImplName(RTLIB::LibcallImpl Call)
Get the libcall routine name for the specified libcall implementation.
virtual Value * getSafeStackPointerLocation(IRBuilderBase &IRB, const LibcallLoweringInfo &Libcalls) const
Returns the target-specific address of the unsafe stack pointer.
LegalizeKind getTypeConversion(LLVMContext &Context, EVT VT) const
Return pair that represents the legalization kind (first) that needs to happen to EVT (second) in ord...
void setLoadExtAction(unsigned ExtType, MVT ValVT, MVT MemVT, LegalizeAction Action)
Indicate that the specified load with extension does not work with the specified type and indicate wh...
unsigned GatherAllAliasesMaxDepth
Depth that GatherAllAliases should continue looking for chain dependencies when trying to find a more...
int IntrinsicIDToISD(Intrinsic::ID ID) const
Get the ISD node that corresponds to the Intrinsic ID.
LegalizeTypeAction getTypeAction(LLVMContext &Context, EVT VT) const
Return how we should legalize values of this type, either it is already legal (return 'Legal') or we ...
int getSqrtRefinementSteps(EVT VT, MachineFunction &MF) const
Return the refinement step count for a square root of the given type based on the function's attribut...
bool allowsMemoryAccessForAlignment(LLVMContext &Context, const DataLayout &DL, EVT VT, unsigned AddrSpace=0, Align Alignment=Align(1), MachineMemOperand::Flags Flags=MachineMemOperand::MONone, unsigned *Fast=nullptr) const
This function returns true if the memory access is aligned or if the target allows this specific unal...
virtual Instruction * emitTrailingFence(IRBuilderBase &Builder, Instruction *Inst, AtomicOrdering Ord) const
virtual Instruction * emitLeadingFence(IRBuilderBase &Builder, Instruction *Inst, AtomicOrdering Ord) const
Inserts in the IR a target-specific intrinsic specifying a fence.
unsigned MaxStoresPerMemcpy
Specify maximum number of store instructions per memcpy call.
unsigned getMaxStoresPerMemmove(bool OptSize) const
Get maximum # of store operations permitted for llvm.memmove.
MVT getRegisterType(MVT VT) const
Return the type of registers that this ValueType will eventually require.
void setJumpIsExpensive(bool isExpensive=true)
Tells the code generator not to expand logic operations on comparison predicates into separate sequen...
LegalizeAction getOperationAction(unsigned Op, EVT VT) const
Return how this operation should be treated: either it is legal, needs to be promoted to a larger siz...
MVT getTypeToPromoteTo(unsigned Op, MVT VT) const
If the action for this operation is to promote, this method returns the ValueType to promote to.
virtual bool isLegalAddressingMode(const DataLayout &DL, const AddrMode &AM, Type *Ty, unsigned AddrSpace, Instruction *I=nullptr) const
Return true if the addressing mode represented by AM is legal for this target, for a load/store of th...
unsigned getVectorTypeBreakdown(LLVMContext &Context, EVT VT, EVT &IntermediateVT, unsigned &NumIntermediates, MVT &RegisterVT) const
Vector types are broken down into some number of legal first class types.
std::pair< LegalizeTypeAction, EVT > LegalizeKind
LegalizeKind holds the legalization kind that needs to happen to EVT in order to type-legalize it.
This class defines information used to lower LLVM code to legal SelectionDAG operators that the targe...
virtual EVT getTypeForExtReturn(LLVMContext &Context, EVT VT, ISD::NodeType) const
Return the type that should be used to zero or sign extend a zeroext/signext integer return value.
Primary interface to the complete machine description for the target machine.
bool isPositionIndependent() const
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
TargetSubtargetInfo - Generic base class for all target subtargets.
This pass provides access to the codegen interfaces that are needed for IR-level transformations.
Twine - A lightweight data structure for efficiently representing the concatenation of temporary valu...
Definition Twine.h:82
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:46
This is the common base class for vector predication intrinsics.
LLVM Value Representation.
Definition Value.h:75
Type * getType() const
All values are typed, get the type of this value.
Definition Value.h:255
constexpr LeafTy coefficientNextPowerOf2() const
Definition TypeSize.h:260
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
Definition TypeSize.h:168
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
Definition TypeSize.h:165
constexpr LeafTy divideCoefficientBy(ScalarTy RHS) const
We do not provide the '/' operator here because division for polynomial types does not work in the sa...
Definition TypeSize.h:252
CallInst * Call
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
Definition CallingConv.h:24
@ Fast
Attempts to make calls as fast as possible (e.g.
Definition CallingConv.h:41
NodeType
ISD::NodeType enum - This enum defines the target-independent operators for a SelectionDAG.
Definition ISDOpcodes.h:41
@ SETCC
SetCC operator - This evaluates to a true value iff the condition is true.
Definition ISDOpcodes.h:823
@ MERGE_VALUES
MERGE_VALUES - This node takes multiple discrete operands and returns them all as its individual resu...
Definition ISDOpcodes.h:261
@ DELETED_NODE
DELETED_NODE - This is an illegal value that is used to catch errors.
Definition ISDOpcodes.h:45
@ SET_FPENV
Sets the current floating-point environment.
@ LOOP_DEPENDENCE_RAW_MASK
@ VECREDUCE_SEQ_FADD
Generic reduction nodes.
@ FGETSIGN
INT = FGETSIGN(FP) - Return the sign bit of the specified floating point value as an integer 0/1 valu...
Definition ISDOpcodes.h:538
@ STACKADDRESS
STACKADDRESS - Represents the llvm.stackaddress intrinsic.
Definition ISDOpcodes.h:127
@ SMULFIX
RESULT = [US]MULFIX(LHS, RHS, SCALE) - Perform fixed point multiplication on 2 integers with the same...
Definition ISDOpcodes.h:394
@ ADDC
Carry-setting nodes for multiple precision addition and subtraction.
Definition ISDOpcodes.h:294
@ RESET_FPENV
Set floating-point environment to default state.
@ FMAD
FMAD - Perform a * b + c, while getting the same result as the separately rounded operations.
Definition ISDOpcodes.h:522
@ ADD
Simple integer binary arithmetic operators.
Definition ISDOpcodes.h:264
@ LOAD
LOAD and STORE have token chains as their first operand, then the same operands as an LLVM load/store...
@ SMULFIXSAT
Same as the corresponding unsaturated fixed point instructions, but the result is clamped between the...
Definition ISDOpcodes.h:400
@ SET_FPMODE
Sets the current dynamic floating-point control modes.
@ ANY_EXTEND
ANY_EXTEND - Used for integer types. The high bits are undefined.
Definition ISDOpcodes.h:857
@ CTTZ_ELTS
Returns the number of number of trailing (least significant) zero elements in a vector.
@ FMA
FMA - Perform a * b + c with no intermediate rounding step.
Definition ISDOpcodes.h:518
@ VECTOR_FIND_LAST_ACTIVE
Finds the index of the last active mask element Operands: Mask.
@ FMODF
FMODF - Decomposes the operand into integral and fractional parts, each having the same type and sign...
@ FATAN2
FATAN2 - atan2, inspired by libm.
@ FSINCOSPI
FSINCOSPI - Compute both the sine and cosine times pi more accurately than FSINCOS(pi*x),...
@ ATOMIC_CMP_SWAP_WITH_SUCCESS
Val, Success, OUTCHAIN = ATOMIC_CMP_SWAP_WITH_SUCCESS(INCHAIN, ptr, cmp, swap) N.b.
@ SINT_TO_FP
[SU]INT_TO_FP - These operators convert integers (whose interpreted sign depends on the first letter)...
Definition ISDOpcodes.h:884
@ CONCAT_VECTORS
CONCAT_VECTORS(VECTOR0, VECTOR1, ...) - Given a number of values of vector type with the same length ...
Definition ISDOpcodes.h:584
@ VECREDUCE_FMAX
FMIN/FMAX nodes can have flags, for NaN/NoNaN variants.
@ FADD
Simple binary floating point operators.
Definition ISDOpcodes.h:417
@ VECREDUCE_FMAXIMUM
FMINIMUM/FMAXIMUM nodes propatate NaNs and signed zeroes using the llvm.minimum and llvm....
@ ABS
ABS - Determine the unsigned absolute value of a signed integer value of the same bitwidth.
Definition ISDOpcodes.h:747
@ RESET_FPMODE
Sets default dynamic floating-point control modes.
@ SIGN_EXTEND_VECTOR_INREG
SIGN_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register sign-extension of the low ...
Definition ISDOpcodes.h:914
@ FMULADD
FMULADD - Performs a * b + c, with, or without, intermediate rounding.
Definition ISDOpcodes.h:528
@ BITCAST
BITCAST - This operator converts between integer, vector and FP values, as if the value was stored to...
Definition ISDOpcodes.h:997
@ CLMUL
Carry-less multiplication operations.
Definition ISDOpcodes.h:778
@ FLDEXP
FLDEXP - ldexp, inspired by libm (op0 * 2**op1).
@ SDIVFIX
RESULT = [US]DIVFIX(LHS, RHS, SCALE) - Perform fixed point division on 2 integers with the same width...
Definition ISDOpcodes.h:407
@ BUILTIN_OP_END
BUILTIN_OP_END - This must be the last enum value in this list.
@ CONVERT_FROM_ARBITRARY_FP
CONVERT_FROM_ARBITRARY_FP - This operator converts from an arbitrary floating-point represented as an...
@ CTLZ_ZERO_POISON
Definition ISDOpcodes.h:792
@ SIGN_EXTEND
Conversion operators.
Definition ISDOpcodes.h:848
@ AVGCEILS
AVGCEILS/AVGCEILU - Rounding averaging add - Add two integers using an integer of type i[N+2],...
Definition ISDOpcodes.h:715
@ READSTEADYCOUNTER
READSTEADYCOUNTER - This corresponds to the readfixedcounter intrinsic.
@ VECREDUCE_FADD
These reductions have relaxed evaluation order semantics, and have a single vector operand.
@ PREFETCH
PREFETCH - This corresponds to a prefetch intrinsic.
@ TRUNCATE_SSAT_U
Definition ISDOpcodes.h:877
@ FSINCOS
FSINCOS - Compute both fsin and fcos as a single operation.
@ SETCCCARRY
Like SetCC, ops #0 and #1 are the LHS and RHS operands to compare, but op #2 is a boolean indicating ...
Definition ISDOpcodes.h:831
@ FNEG
Perform various unary floating-point operations inspired by libm.
@ SSUBO
Same for subtraction.
Definition ISDOpcodes.h:352
@ FCANONICALIZE
Returns platform specific canonical encoding of a floating point number.
Definition ISDOpcodes.h:541
@ IS_FPCLASS
Performs a check of floating point class property, defined by IEEE-754.
Definition ISDOpcodes.h:548
@ SSUBSAT
RESULT = [US]SUBSAT(LHS, RHS) - Perform saturation subtraction on 2 integers with the same bit width ...
Definition ISDOpcodes.h:374
@ SELECT
Select(COND, TRUEVAL, FALSEVAL).
Definition ISDOpcodes.h:800
@ SPLAT_VECTOR
SPLAT_VECTOR(VAL) - Returns a vector with the scalar value VAL duplicated in all lanes.
Definition ISDOpcodes.h:672
@ SADDO
RESULT, BOOL = [SU]ADDO(LHS, RHS) - Overflow-aware nodes for addition.
Definition ISDOpcodes.h:348
@ CTLS
Count leading redundant sign bits.
Definition ISDOpcodes.h:796
@ VECREDUCE_ADD
Integer reductions may have a result type larger than the vector element type.
@ GET_FPMODE
Reads the current dynamic floating-point control modes.
@ GET_FPENV
Gets the current floating-point environment.
@ SHL
Shift and rotation operations.
Definition ISDOpcodes.h:769
@ VECTOR_SHUFFLE
VECTOR_SHUFFLE(VEC1, VEC2) - Returns a vector, of the same type as VEC1/VEC2.
Definition ISDOpcodes.h:649
@ FMINNUM_IEEE
FMINNUM_IEEE/FMAXNUM_IEEE - Perform floating-point minimumNumber or maximumNumber on two values,...
@ EXTRACT_VECTOR_ELT
EXTRACT_VECTOR_ELT(VECTOR, IDX) - Returns a single element from VECTOR identified by the (potentially...
Definition ISDOpcodes.h:576
@ ZERO_EXTEND
ZERO_EXTEND - Used for integer types, zeroing the new bits.
Definition ISDOpcodes.h:854
@ DEBUGTRAP
DEBUGTRAP - Trap intended to get the attention of a debugger.
@ ATOMIC_CMP_SWAP
Val, OUTCHAIN = ATOMIC_CMP_SWAP(INCHAIN, ptr, cmp, swap) For double-word atomic operations: ValLo,...
@ FMINNUM
FMINNUM/FMAXNUM - Perform floating-point minimum maximum on two values, following IEEE-754 definition...
@ UBSANTRAP
UBSANTRAP - Trap with an immediate describing the kind of sanitizer failure.
@ SSHLSAT
RESULT = [US]SHLSAT(LHS, RHS) - Perform saturation left shift.
Definition ISDOpcodes.h:386
@ SMULO
Same for multiplication.
Definition ISDOpcodes.h:356
@ VECTOR_SPLICE_LEFT
VECTOR_SPLICE_LEFT(VEC1, VEC2, OFFSET) - Shifts CONCAT_VECTORS(VEC1, VEC2) left by OFFSET elements an...
Definition ISDOpcodes.h:653
@ ANY_EXTEND_VECTOR_INREG
ANY_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register any-extension of the low la...
Definition ISDOpcodes.h:903
@ SIGN_EXTEND_INREG
SIGN_EXTEND_INREG - This operator atomically performs a SHL/SRA pair to sign extend a small value in ...
Definition ISDOpcodes.h:892
@ SMIN
[US]{MIN/MAX} - Binary minimum or maximum of signed or unsigned integers.
Definition ISDOpcodes.h:727
@ MASKED_UDIV
Masked vector arithmetic that returns poison on disabled lanes.
@ SDIVFIXSAT
Same as the corresponding unsaturated fixed point instructions, but the result is clamped between the...
Definition ISDOpcodes.h:413
@ FP_EXTEND
X = FP_EXTEND(Y) - Extend a smaller FP type into a larger FP type.
Definition ISDOpcodes.h:982
@ UADDO_CARRY
Carry-using nodes for multiple precision addition and subtraction.
Definition ISDOpcodes.h:328
@ FMINIMUM
FMINIMUM/FMAXIMUM - NaN-propagating minimum/maximum that also treat -0.0 as less than 0....
@ FP_TO_SINT
FP_TO_[US]INT - Convert a floating point value to a signed or unsigned integer.
Definition ISDOpcodes.h:930
@ READCYCLECOUNTER
READCYCLECOUNTER - This corresponds to the readcyclecounter intrinsic.
@ AND
Bitwise operators - logical and, logical or, logical xor.
Definition ISDOpcodes.h:739
@ TRAP
TRAP - Trapping instruction.
@ GET_FPENV_MEM
Gets the current floating-point environment.
@ SCMP
[US]CMP - 3-way comparison of signed or unsigned integers.
Definition ISDOpcodes.h:735
@ AVGFLOORS
AVGFLOORS/AVGFLOORU - Averaging add - Add two integers using an integer of type i[N+1],...
Definition ISDOpcodes.h:710
@ VECTOR_SPLICE_RIGHT
VECTOR_SPLICE_RIGHT(VEC1, VEC2, OFFSET) - Shifts CONCAT_VECTORS(VEC1,VEC2) right by OFFSET elements a...
Definition ISDOpcodes.h:657
@ ADDE
Carry-using nodes for multiple precision addition and subtraction.
Definition ISDOpcodes.h:304
@ FREEZE
FREEZE - FREEZE(VAL) returns an arbitrary value if VAL is UNDEF (or is evaluated to UNDEF),...
Definition ISDOpcodes.h:241
@ INSERT_VECTOR_ELT
INSERT_VECTOR_ELT(VECTOR, VAL, IDX) - Returns VECTOR with the element at IDX replaced with VAL.
Definition ISDOpcodes.h:565
@ ATOMIC_SWAP
Val, OUTCHAIN = ATOMIC_SWAP(INCHAIN, ptr, amt) Val, OUTCHAIN = ATOMIC_LOAD_[OpName](INCHAIN,...
@ CTTZ_ZERO_POISON
Bit counting operators with a poisoned result for zero inputs.
Definition ISDOpcodes.h:791
@ FFREXP
FFREXP - frexp, extract fractional and exponent component of a floating-point value.
@ FP_ROUND
X = FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision of the ...
Definition ISDOpcodes.h:963
@ VECTOR_COMPRESS
VECTOR_COMPRESS(Vec, Mask, Passthru) consecutively place vector elements based on mask e....
Definition ISDOpcodes.h:699
@ CLEAR_CACHE
llvm.clear_cache intrinsic Operands: Input Chain, Start Addres, End Address Outputs: Output Chain
@ ZERO_EXTEND_VECTOR_INREG
ZERO_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register zero-extension of the low ...
Definition ISDOpcodes.h:925
@ ADDRSPACECAST
ADDRSPACECAST - This operator converts between pointers of different address spaces.
@ FP_TO_SINT_SAT
FP_TO_[US]INT_SAT - Convert floating point value in operand 0 to a signed or unsigned scalar integer ...
Definition ISDOpcodes.h:949
@ VECREDUCE_FMINIMUM
@ TRUNCATE
TRUNCATE - Completely drop the high bits.
Definition ISDOpcodes.h:860
@ VECREDUCE_SEQ_FMUL
@ FCOPYSIGN
FCOPYSIGN(X, Y) - Return the value of X with the sign of Y.
Definition ISDOpcodes.h:534
@ SADDSAT
RESULT = [US]ADDSAT(LHS, RHS) - Perform saturation addition on 2 integers with the same bit width (W)...
Definition ISDOpcodes.h:365
@ GET_DYNAMIC_AREA_OFFSET
GET_DYNAMIC_AREA_OFFSET - get offset from native SP to the address of the most recent dynamic alloca.
@ CTTZ_ELTS_ZERO_POISON
@ SET_FPENV_MEM
Sets the current floating point environment.
@ FMINIMUMNUM
FMINIMUMNUM/FMAXIMUMNUM - minimumnum/maximumnum that is same with FMINNUM_IEEE and FMAXNUM_IEEE besid...
@ TRUNCATE_SSAT_S
TRUNCATE_[SU]SAT_[SU] - Truncate for saturated operand [SU] located in middle, prefix for SAT means i...
Definition ISDOpcodes.h:875
@ ABDS
ABDS/ABDU - Absolute difference - Return the absolute difference between two numbers interpreted as s...
Definition ISDOpcodes.h:722
@ TRUNCATE_USAT_U
Definition ISDOpcodes.h:879
@ SADDO_CARRY
Carry-using overflow-aware nodes for multiple precision addition and subtraction.
Definition ISDOpcodes.h:338
@ ABS_MIN_POISON
ABS with a poison result for INT_MIN.
Definition ISDOpcodes.h:751
@ LOOP_DEPENDENCE_WAR_MASK
The llvm.loop.dependence.
CondCode
ISD::CondCode enum - These are ordered carefully to make the bitfields below work out,...
static const int LAST_INDEXED_MODE
LLVM_ABI Libcall getPOWI(EVT RetVT)
getPOWI - Return the POWI_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getSINTTOFP(EVT OpVT, EVT RetVT)
getSINTTOFP - Return the SINTTOFP_*_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getUREM(EVT VT)
LLVM_ABI Libcall getSHL(EVT VT)
LLVM_ABI Libcall getSYNC(unsigned Opc, MVT VT)
Return the SYNC_FETCH_AND_* value for the given opcode and type, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getLDEXP(EVT RetVT)
getLDEXP - Return the LDEXP_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getUINTTOFP(EVT OpVT, EVT RetVT)
getUINTTOFP - Return the UINTTOFP_*_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getFREXP(EVT RetVT)
getFREXP - Return the FREXP_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getREM(EVT VT)
LLVM_ABI Libcall getSINCOSPI(EVT RetVT)
getSINCOSPI - Return the SINCOSPI_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getSDIV(EVT VT)
LLVM_ABI Libcall getSRL(EVT VT)
LLVM_ABI Libcall getMEMCPY_ELEMENT_UNORDERED_ATOMIC(uint64_t ElementSize)
getMEMCPY_ELEMENT_UNORDERED_ATOMIC - Return MEMCPY_ELEMENT_UNORDERED_ATOMIC_* value for the given ele...
LLVM_ABI Libcall getSRA(EVT VT)
LLVM_ABI Libcall getUDIV(EVT VT)
LLVM_ABI Libcall getFPLibCall(EVT VT, Libcall Call_F32, Libcall Call_F64, Libcall Call_F80, Libcall Call_F128, Libcall Call_PPCF128)
GetFPLibCall - Helper to return the right libcall for the given floating point type,...
LLVM_ABI Libcall getFPTOUINT(EVT OpVT, EVT RetVT)
getFPTOUINT - Return the FPTOUINT_*_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getLLROUND(EVT VT)
LLVM_ABI Libcall getCOS(EVT RetVT)
Return the COS_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getLROUND(EVT VT)
LLVM_ABI Libcall getMODF(EVT VT)
getMODF - Return the MODF_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getFPTOSINT(EVT OpVT, EVT RetVT)
getFPTOSINT - Return the FPTOSINT_*_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getLRINT(EVT RetVT)
LLVM_ABI Libcall getCBRT(EVT RetVT)
getCBRT - Return the CBRT_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getOUTLINE_ATOMIC(unsigned Opc, AtomicOrdering Order, MVT VT)
Return the outline atomics value for the given opcode, atomic ordering and type, or UNKNOWN_LIBCALL i...
LLVM_ABI Libcall getLLRINT(EVT RetVT)
LLVM_ABI Libcall getFPEXT(EVT OpVT, EVT RetVT)
getFPEXT - Return the FPEXT_*_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getFPROUND(EVT OpVT, EVT RetVT)
getFPROUND - Return the FPROUND_*_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getSREM(EVT VT)
LLVM_ABI Libcall getSIN(EVT RetVT)
Return the SIN_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getMEMSET_ELEMENT_UNORDERED_ATOMIC(uint64_t ElementSize)
getMEMSET_ELEMENT_UNORDERED_ATOMIC - Return MEMSET_ELEMENT_UNORDERED_ATOMIC_* value for the given ele...
LLVM_ABI Libcall getSINCOS_STRET(EVT RetVT)
Return the SINCOS_STRET_ value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getPOW(EVT RetVT)
getPOW - Return the POW_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getOutlineAtomicHelper(const Libcall(&LC)[5][4], AtomicOrdering Order, uint64_t MemSize)
Return the outline atomics value for the given atomic ordering, access size and set of libcalls for a...
LLVM_ABI Libcall getMUL(EVT VT)
LLVM_ABI Libcall getCTPOP(EVT VT)
LLVM_ABI Libcall getSINCOS(EVT RetVT)
getSINCOS - Return the SINCOS_* value for the given types, or UNKNOWN_LIBCALL if there is none.
LLVM_ABI Libcall getMULO(EVT VT)
LLVM_ABI Libcall getMEMMOVE_ELEMENT_UNORDERED_ATOMIC(uint64_t ElementSize)
getMEMMOVE_ELEMENT_UNORDERED_ATOMIC - Return MEMMOVE_ELEMENT_UNORDERED_ATOMIC_* value for the given e...
initializer< Ty > init(const Ty &Val)
This is an optimization pass for GlobalISel generic memory operations.
unsigned Log2_32_Ceil(uint32_t Value)
Return the ceil log base 2 of the specified value, 32 if the value is zero.
Definition MathExtras.h:344
void fill(R &&Range, T &&Value)
Provide wrappers to std::fill which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1758
LLVM_ABI void GetReturnInfo(CallingConv::ID CC, Type *ReturnType, AttributeList attr, SmallVectorImpl< ISD::OutputArg > &Outs, const TargetLowering &TLI, const DataLayout &DL)
Given an LLVM IR type and return type attributes, compute the return value EVTs and flags,...
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
InstructionCost Cost
auto enum_seq(EnumT Begin, EnumT End)
Iterate over an enum type from Begin up to - but not including - End.
Definition Sequence.h:337
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:643
LLVM_ABI bool isDereferenceableAndAlignedPointer(const Value *V, Type *Ty, Align Alignment, const DataLayout &DL, const Instruction *CtxI=nullptr, AssumptionCache *AC=nullptr, const DominatorTree *DT=nullptr, const TargetLibraryInfo *TLI=nullptr)
Returns true if V is always a dereferenceable pointer with alignment greater or equal than requested.
Definition Loads.cpp:230
LLVM_ABI bool shouldOptimizeForSize(const MachineFunction *MF, ProfileSummaryInfo *PSI, const MachineBlockFrequencyInfo *BFI, PGSOQueryType QueryType=PGSOQueryType::Other)
Returns true if machine function MF is suggested to be size-optimized based on the profile.
constexpr force_iteration_on_noniterable_enum_t force_iteration_on_noniterable_enum
Definition Sequence.h:109
T bit_ceil(T Value)
Returns the smallest integral power of two no smaller than Value if Value is nonzero.
Definition bit.h:362
void ComputeValueTypes(const DataLayout &DL, Type *Ty, SmallVectorImpl< Type * > &Types, SmallVectorImpl< TypeSize > *Offsets=nullptr, TypeSize StartingOffset=TypeSize::getZero())
Given an LLVM IR type, compute non-aggregate subtypes.
Definition Analysis.cpp:72
bool isReleaseOrStronger(AtomicOrdering AO)
auto dyn_cast_or_null(const Y &Val)
Definition Casting.h:753
constexpr bool has_single_bit(T Value) noexcept
Definition bit.h:149
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
Definition MathExtras.h:279
bool none_of(R &&Range, UnaryPredicate P)
Provide wrappers to std::none_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1752
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
Definition Error.cpp:163
bool isDigit(char C)
Checks if character C is one of the 10 decimal digits.
CodeGenOptLevel
Code generation optimization level.
Definition CodeGen.h:82
AtomicOrdering
Atomic ordering for LLVM's memory model.
LLVM_ABI EVT getApproximateEVTForLLT(LLT Ty, LLVMContext &Ctx)
constexpr T divideCeil(U Numerator, V Denominator)
Returns the integer ceil(Numerator / Denominator).
Definition MathExtras.h:394
TargetTransformInfo TTI
@ Mul
Product of integers.
@ FSub
Subtraction of floats.
@ Xor
Bitwise or logical XOR of integers.
@ FMul
Product of floats.
@ Sub
Subtraction of integers.
@ Add
Sum of integers.
@ FAdd
Sum of floats.
DWARFExpression::Operation Op
bool isAcquireOrStronger(AtomicOrdering AO)
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition Alignment.h:39
Extended Value Type.
Definition ValueTypes.h:35
EVT getPow2VectorType(LLVMContext &Context) const
Widens the length of the given vector EVT up to the nearest power of 2 and returns that type.
Definition ValueTypes.h:493
bool isSimple() const
Test if the given EVT is simple (as opposed to being extended).
Definition ValueTypes.h:145
static EVT getVectorVT(LLVMContext &Context, EVT VT, unsigned NumElements, bool IsScalable=false)
Returns the EVT that represents a vector NumElements in length, where each element is of type VT.
Definition ValueTypes.h:70
ElementCount getVectorElementCount() const
Definition ValueTypes.h:358
TypeSize getSizeInBits() const
Return the size of the specified value type in bits.
Definition ValueTypes.h:381
uint64_t getScalarSizeInBits() const
Definition ValueTypes.h:393
bool isPow2VectorType() const
Returns true if the given vector is a power of 2.
Definition ValueTypes.h:486
MVT getSimpleVT() const
Return the SimpleValueType held in the specified simple EVT.
Definition ValueTypes.h:324
static EVT getIntegerVT(LLVMContext &Context, unsigned BitWidth)
Returns the EVT that represents an integer with the given number of bits.
Definition ValueTypes.h:61
bool isFixedLengthVector() const
Definition ValueTypes.h:189
EVT getRoundIntegerType(LLVMContext &Context) const
Rounds the bit-width of the given integer EVT up to the nearest power of two (and at least to eight),...
Definition ValueTypes.h:427
bool isVector() const
Return true if this is a vector value type.
Definition ValueTypes.h:176
EVT getScalarType() const
If this is a vector type, return the element type, otherwise return this.
Definition ValueTypes.h:331
LLVM_ABI Type * getTypeForEVT(LLVMContext &Context) const
This method returns an LLVM type corresponding to the specified EVT.
EVT getVectorElementType() const
Given a vector type, return the type of each element.
Definition ValueTypes.h:336
unsigned getVectorNumElements() const
Given a vector type, return the number of elements it contains.
Definition ValueTypes.h:344
bool isZeroSized() const
Test if the given EVT has zero size, this will fail if called on a scalable type.
Definition ValueTypes.h:140
EVT getHalfNumVectorElementsVT(LLVMContext &Context) const
Definition ValueTypes.h:469
bool isInteger() const
Return true if this is an integer or a vector integer type.
Definition ValueTypes.h:160
OutputArg - This struct carries flags and a value for a single outgoing (actual) argument or outgoing...
Matching combinators.
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
static StringRef getLibcallImplName(RTLIB::LibcallImpl CallImpl)
Get the libcall routine name for the specified libcall implementation.
static RTLIB::Libcall getLibcallFromImpl(RTLIB::LibcallImpl Impl)
Return the libcall provided by Impl.
This represents an addressing mode of: BaseGV + BaseOffs + BaseReg + Scale*ScaleReg + ScalableOffset*...