45 cl::desc(
"Force a specific generic_v<N> flag to be "
46 "added. For testing purposes only."),
51 if (!HSAMetadataDoc.
fromYAML(HSAMetadataString))
61 case ELF::EF_AMDGPU_MACH_R600_R600: AK = GK_R600;
break;
62 case ELF::EF_AMDGPU_MACH_R600_R630: AK = GK_R630;
break;
63 case ELF::EF_AMDGPU_MACH_R600_RS880: AK = GK_RS880;
break;
64 case ELF::EF_AMDGPU_MACH_R600_RV670: AK = GK_RV670;
break;
65 case ELF::EF_AMDGPU_MACH_R600_RV710: AK = GK_RV710;
break;
66 case ELF::EF_AMDGPU_MACH_R600_RV730: AK = GK_RV730;
break;
67 case ELF::EF_AMDGPU_MACH_R600_RV770: AK = GK_RV770;
break;
68 case ELF::EF_AMDGPU_MACH_R600_CEDAR: AK = GK_CEDAR;
break;
69 case ELF::EF_AMDGPU_MACH_R600_CYPRESS: AK = GK_CYPRESS;
break;
70 case ELF::EF_AMDGPU_MACH_R600_JUNIPER: AK = GK_JUNIPER;
break;
71 case ELF::EF_AMDGPU_MACH_R600_REDWOOD: AK = GK_REDWOOD;
break;
72 case ELF::EF_AMDGPU_MACH_R600_SUMO: AK = GK_SUMO;
break;
73 case ELF::EF_AMDGPU_MACH_R600_BARTS: AK = GK_BARTS;
break;
74 case ELF::EF_AMDGPU_MACH_R600_CAICOS: AK = GK_CAICOS;
break;
75 case ELF::EF_AMDGPU_MACH_R600_CAYMAN: AK = GK_CAYMAN;
break;
76 case ELF::EF_AMDGPU_MACH_R600_TURKS: AK = GK_TURKS;
break;
77 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX600: AK = GK_GFX600;
break;
78 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX601: AK = GK_GFX601;
break;
79 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX602: AK = GK_GFX602;
break;
80 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX700: AK = GK_GFX700;
break;
81 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX701: AK = GK_GFX701;
break;
82 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX702: AK = GK_GFX702;
break;
83 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX703: AK = GK_GFX703;
break;
84 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX704: AK = GK_GFX704;
break;
85 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX705: AK = GK_GFX705;
break;
86 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX801: AK = GK_GFX801;
break;
87 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX802: AK = GK_GFX802;
break;
88 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX803: AK = GK_GFX803;
break;
89 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX805: AK = GK_GFX805;
break;
90 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX810: AK = GK_GFX810;
break;
91 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX900: AK = GK_GFX900;
break;
92 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX902: AK = GK_GFX902;
break;
93 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX904: AK = GK_GFX904;
break;
94 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX906: AK = GK_GFX906;
break;
95 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX908: AK = GK_GFX908;
break;
96 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX909: AK = GK_GFX909;
break;
97 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX90A: AK = GK_GFX90A;
break;
98 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX90C: AK = GK_GFX90C;
break;
99 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX942: AK = GK_GFX942;
break;
100 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX950: AK = GK_GFX950;
break;
101 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1010: AK = GK_GFX1010;
break;
102 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1011: AK = GK_GFX1011;
break;
103 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1012: AK = GK_GFX1012;
break;
104 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1013: AK = GK_GFX1013;
break;
105 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1030: AK = GK_GFX1030;
break;
106 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1031: AK = GK_GFX1031;
break;
107 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1032: AK = GK_GFX1032;
break;
108 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1033: AK = GK_GFX1033;
break;
109 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1034: AK = GK_GFX1034;
break;
110 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1035: AK = GK_GFX1035;
break;
111 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1036: AK = GK_GFX1036;
break;
112 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1100: AK = GK_GFX1100;
break;
113 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1101: AK = GK_GFX1101;
break;
114 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1102: AK = GK_GFX1102;
break;
115 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1103: AK = GK_GFX1103;
break;
116 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1150: AK = GK_GFX1150;
break;
117 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1151: AK = GK_GFX1151;
break;
118 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1152: AK = GK_GFX1152;
break;
119 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1153: AK = GK_GFX1153;
break;
120 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1170: AK = GK_GFX1170;
break;
121 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1171: AK = GK_GFX1171;
break;
122 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1172: AK = GK_GFX1172;
break;
123 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1200: AK = GK_GFX1200;
break;
124 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1201: AK = GK_GFX1201;
break;
125 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1250: AK = GK_GFX1250;
break;
126 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1251: AK = GK_GFX1251;
break;
127 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX1310: AK = GK_GFX1310;
break;
128 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX9_GENERIC: AK = GK_GFX9_GENERIC;
break;
129 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX9_4_GENERIC: AK = GK_GFX9_4_GENERIC;
break;
130 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX10_1_GENERIC: AK = GK_GFX10_1_GENERIC;
break;
131 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX10_3_GENERIC: AK = GK_GFX10_3_GENERIC;
break;
132 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX11_GENERIC: AK = GK_GFX11_GENERIC;
break;
133 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX12_GENERIC: AK = GK_GFX12_GENERIC;
break;
134 case ELF::EF_AMDGPU_MACH_AMDGCN_GFX12_5_GENERIC: AK = GK_GFX12_5_GENERIC;
break;
153 case GK_R600:
return ELF::EF_AMDGPU_MACH_R600_R600;
154 case GK_R630:
return ELF::EF_AMDGPU_MACH_R600_R630;
155 case GK_RS880:
return ELF::EF_AMDGPU_MACH_R600_RS880;
156 case GK_RV670:
return ELF::EF_AMDGPU_MACH_R600_RV670;
157 case GK_RV710:
return ELF::EF_AMDGPU_MACH_R600_RV710;
158 case GK_RV730:
return ELF::EF_AMDGPU_MACH_R600_RV730;
159 case GK_RV770:
return ELF::EF_AMDGPU_MACH_R600_RV770;
160 case GK_CEDAR:
return ELF::EF_AMDGPU_MACH_R600_CEDAR;
161 case GK_CYPRESS:
return ELF::EF_AMDGPU_MACH_R600_CYPRESS;
162 case GK_JUNIPER:
return ELF::EF_AMDGPU_MACH_R600_JUNIPER;
163 case GK_REDWOOD:
return ELF::EF_AMDGPU_MACH_R600_REDWOOD;
164 case GK_SUMO:
return ELF::EF_AMDGPU_MACH_R600_SUMO;
165 case GK_BARTS:
return ELF::EF_AMDGPU_MACH_R600_BARTS;
166 case GK_CAICOS:
return ELF::EF_AMDGPU_MACH_R600_CAICOS;
167 case GK_CAYMAN:
return ELF::EF_AMDGPU_MACH_R600_CAYMAN;
168 case GK_TURKS:
return ELF::EF_AMDGPU_MACH_R600_TURKS;
169 case GK_GFX600:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX600;
170 case GK_GFX601:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX601;
171 case GK_GFX602:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX602;
172 case GK_GFX700:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX700;
173 case GK_GFX701:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX701;
174 case GK_GFX702:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX702;
175 case GK_GFX703:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX703;
176 case GK_GFX704:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX704;
177 case GK_GFX705:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX705;
178 case GK_GFX801:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX801;
179 case GK_GFX802:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX802;
180 case GK_GFX803:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX803;
181 case GK_GFX805:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX805;
182 case GK_GFX810:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX810;
183 case GK_GFX900:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX900;
184 case GK_GFX902:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX902;
185 case GK_GFX904:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX904;
186 case GK_GFX906:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX906;
187 case GK_GFX908:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX908;
188 case GK_GFX909:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX909;
189 case GK_GFX90A:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX90A;
190 case GK_GFX90C:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX90C;
191 case GK_GFX942:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX942;
192 case GK_GFX950:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX950;
193 case GK_GFX1010:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1010;
194 case GK_GFX1011:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1011;
195 case GK_GFX1012:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1012;
196 case GK_GFX1013:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1013;
197 case GK_GFX1030:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1030;
198 case GK_GFX1031:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1031;
199 case GK_GFX1032:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1032;
200 case GK_GFX1033:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1033;
201 case GK_GFX1034:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1034;
202 case GK_GFX1035:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1035;
203 case GK_GFX1036:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1036;
204 case GK_GFX1100:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1100;
205 case GK_GFX1101:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1101;
206 case GK_GFX1102:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1102;
207 case GK_GFX1103:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1103;
208 case GK_GFX1150:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1150;
209 case GK_GFX1151:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1151;
210 case GK_GFX1152:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1152;
211 case GK_GFX1153:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1153;
212 case GK_GFX1170:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1170;
213 case GK_GFX1171:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1171;
214 case GK_GFX1172:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1172;
215 case GK_GFX1200:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1200;
216 case GK_GFX1201:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1201;
217 case GK_GFX1250:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1250;
218 case GK_GFX1251:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1251;
219 case GK_GFX1310:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX1310;
220 case GK_GFX9_GENERIC:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX9_GENERIC;
221 case GK_GFX9_4_GENERIC:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX9_4_GENERIC;
222 case GK_GFX10_1_GENERIC:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX10_1_GENERIC;
223 case GK_GFX10_3_GENERIC:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX10_3_GENERIC;
224 case GK_GFX11_GENERIC:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX11_GENERIC;
225 case GK_GFX12_GENERIC:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX12_GENERIC;
226 case GK_GFX12_5_GENERIC:
return ELF::EF_AMDGPU_MACH_AMDGCN_GFX12_5_GENERIC;
256 OS <<
"\t.amdgcn_target \"" << *
getTargetID() <<
"\"\n";
262 OS <<
"\t.amdhsa_code_object_version " << COV <<
'\n';
271 OS <<
"\t.amd_kernel_code_t\n";
272 Header.EmitKernelCodeT(OS,
getContext(), FoldAndPrint);
273 OS <<
"\t.end_amd_kernel_code_t\n";
282 OS <<
"\t.amdgpu_hsa_kernel " << SymbolName <<
'\n';
289 OS <<
"\t.amdgpu_lds " << Symbol->getName() <<
", " <<
Size <<
", "
290 << Alignment.
value() <<
'\n';
299#define PRINT_RES_INFO(ARG) \
301 ARG->print(OS, &getContext().getAsmInfo()); \
303 getContext().getAsmInfo().printExpr(OS, *ARG->getVariableValue()); \
304 Streamer.addBlankLine();
322#define PRINT_RES_INFO(ARG) \
324 ARG->print(OS, &getContext().getAsmInfo()); \
326 getContext().getAsmInfo().printExpr(OS, *ARG->getVariableValue()); \
327 Streamer.addBlankLine();
337 OS <<
"\t.amd_amdgpu_isa \"" <<
getTargetID() <<
"\"\n";
344 if (!Verifier.verify(HSAMetadataDoc.
getRoot()))
347 std::string HSAMetadataString;
349 HSAMetadataDoc.
toYAML(StrOS);
352 OS << StrOS.
str() <<
'\n';
358 const uint32_t Encoded_s_code_end = 0xbf9f0000;
359 const uint32_t Encoded_s_nop = 0xbf800000;
360 uint32_t Encoded_pad = Encoded_s_code_end;
370 Encoded_pad = Encoded_s_nop;
374 OS <<
"\t.p2alignl " << Log2CacheLineSize <<
", " << Encoded_pad <<
'\n';
375 OS <<
"\t.fill " << (FillSize / 4) <<
", 4, " << Encoded_pad <<
'\n';
383 const MCExpr *ReserveFlatScr) {
387 OS <<
"\t.amdhsa_kernel " << KernelName <<
'\n';
392 const MCExpr *ShiftedAndMaskedExpr =
404 OS <<
"\t\t.amdhsa_group_segment_fixed_size ";
408 OS <<
"\t\t.amdhsa_private_segment_fixed_size ";
412 OS <<
"\t\t.amdhsa_kernarg_size ";
418 amdhsa::COMPUTE_PGM_RSRC2_GFX125_USER_SGPR_COUNT_SHIFT,
419 amdhsa::COMPUTE_PGM_RSRC2_GFX125_USER_SGPR_COUNT,
420 ".amdhsa_user_sgpr_count");
423 amdhsa::COMPUTE_PGM_RSRC2_GFX6_GFX120_USER_SGPR_COUNT_SHIFT,
424 amdhsa::COMPUTE_PGM_RSRC2_GFX6_GFX120_USER_SGPR_COUNT,
425 ".amdhsa_user_sgpr_count");
431 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_PRIVATE_SEGMENT_BUFFER_SHIFT,
432 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_PRIVATE_SEGMENT_BUFFER,
433 ".amdhsa_user_sgpr_private_segment_buffer");
435 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_DISPATCH_PTR_SHIFT,
436 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_DISPATCH_PTR,
437 ".amdhsa_user_sgpr_dispatch_ptr");
439 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_QUEUE_PTR_SHIFT,
440 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_QUEUE_PTR,
441 ".amdhsa_user_sgpr_queue_ptr");
443 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_KERNARG_SEGMENT_PTR_SHIFT,
444 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_KERNARG_SEGMENT_PTR,
445 ".amdhsa_user_sgpr_kernarg_segment_ptr");
447 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_DISPATCH_ID_SHIFT,
448 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_DISPATCH_ID,
449 ".amdhsa_user_sgpr_dispatch_id");
452 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_FLAT_SCRATCH_INIT_SHIFT,
453 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_FLAT_SCRATCH_INIT,
454 ".amdhsa_user_sgpr_flat_scratch_init");
457 amdhsa::KERNARG_PRELOAD_SPEC_LENGTH,
458 ".amdhsa_user_sgpr_kernarg_preload_length");
460 amdhsa::KERNARG_PRELOAD_SPEC_OFFSET,
461 ".amdhsa_user_sgpr_kernarg_preload_offset");
465 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_PRIVATE_SEGMENT_SIZE_SHIFT,
466 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_SGPR_PRIVATE_SEGMENT_SIZE,
467 ".amdhsa_user_sgpr_private_segment_size");
468 if (IVersion.
Major >= 10)
470 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_WAVEFRONT_SIZE32_SHIFT,
471 amdhsa::KERNEL_CODE_PROPERTY_ENABLE_WAVEFRONT_SIZE32,
472 ".amdhsa_wavefront_size32");
475 amdhsa::KERNEL_CODE_PROPERTY_USES_DYNAMIC_STACK_SHIFT,
476 amdhsa::KERNEL_CODE_PROPERTY_USES_DYNAMIC_STACK,
477 ".amdhsa_uses_dynamic_stack");
479 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_PRIVATE_SEGMENT_SHIFT,
480 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_PRIVATE_SEGMENT,
482 ?
".amdhsa_enable_private_segment"
483 :
".amdhsa_system_sgpr_private_segment_wavefront_offset"));
485 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_SGPR_WORKGROUP_ID_X_SHIFT,
486 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_SGPR_WORKGROUP_ID_X,
487 ".amdhsa_system_sgpr_workgroup_id_x");
489 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_SGPR_WORKGROUP_ID_Y_SHIFT,
490 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_SGPR_WORKGROUP_ID_Y,
491 ".amdhsa_system_sgpr_workgroup_id_y");
493 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_SGPR_WORKGROUP_ID_Z_SHIFT,
494 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_SGPR_WORKGROUP_ID_Z,
495 ".amdhsa_system_sgpr_workgroup_id_z");
497 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_SGPR_WORKGROUP_INFO_SHIFT,
498 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_SGPR_WORKGROUP_INFO,
499 ".amdhsa_system_sgpr_workgroup_info");
501 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_VGPR_WORKITEM_ID_SHIFT,
502 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_VGPR_WORKITEM_ID,
503 ".amdhsa_system_vgpr_workitem_id");
506 OS <<
"\t\t.amdhsa_next_free_vgpr ";
507 EmitMCExpr(NextVGPR);
510 OS <<
"\t\t.amdhsa_next_free_sgpr ";
511 EmitMCExpr(NextSGPR);
518 amdhsa::COMPUTE_PGM_RSRC3_GFX90A_ACCUM_OFFSET_SHIFT,
519 amdhsa::COMPUTE_PGM_RSRC3_GFX90A_ACCUM_OFFSET,
getContext());
524 OS <<
"\t\t.amdhsa_accum_offset ";
532 amdhsa::COMPUTE_PGM_RSRC3_GFX125_NAMED_BAR_CNT_SHIFT,
533 amdhsa::COMPUTE_PGM_RSRC3_GFX125_NAMED_BAR_CNT,
534 ".amdhsa_named_barrier_count");
536 OS <<
"\t\t.amdhsa_reserve_vcc ";
537 EmitMCExpr(ReserveVCC);
541 OS <<
"\t\t.amdhsa_reserve_flat_scratch ";
542 EmitMCExpr(ReserveFlatScr);
552 OS <<
"\t\t.amdhsa_reserve_xnack_mask " <<
getTargetID()->isXnackOnOrAny()
558 amdhsa::COMPUTE_PGM_RSRC1_FLOAT_ROUND_MODE_32_SHIFT,
559 amdhsa::COMPUTE_PGM_RSRC1_FLOAT_ROUND_MODE_32,
560 ".amdhsa_float_round_mode_32");
562 amdhsa::COMPUTE_PGM_RSRC1_FLOAT_ROUND_MODE_16_64_SHIFT,
563 amdhsa::COMPUTE_PGM_RSRC1_FLOAT_ROUND_MODE_16_64,
564 ".amdhsa_float_round_mode_16_64");
566 amdhsa::COMPUTE_PGM_RSRC1_FLOAT_DENORM_MODE_32_SHIFT,
567 amdhsa::COMPUTE_PGM_RSRC1_FLOAT_DENORM_MODE_32,
568 ".amdhsa_float_denorm_mode_32");
570 amdhsa::COMPUTE_PGM_RSRC1_FLOAT_DENORM_MODE_16_64_SHIFT,
571 amdhsa::COMPUTE_PGM_RSRC1_FLOAT_DENORM_MODE_16_64,
572 ".amdhsa_float_denorm_mode_16_64");
573 if (STI.
hasFeature(AMDGPU::FeatureDX10ClampAndIEEEMode)) {
575 amdhsa::COMPUTE_PGM_RSRC1_GFX6_GFX11_ENABLE_DX10_CLAMP_SHIFT,
576 amdhsa::COMPUTE_PGM_RSRC1_GFX6_GFX11_ENABLE_DX10_CLAMP,
577 ".amdhsa_dx10_clamp");
579 amdhsa::COMPUTE_PGM_RSRC1_GFX6_GFX11_ENABLE_IEEE_MODE_SHIFT,
580 amdhsa::COMPUTE_PGM_RSRC1_GFX6_GFX11_ENABLE_IEEE_MODE,
581 ".amdhsa_ieee_mode");
583 if (IVersion.
Major >= 9) {
585 amdhsa::COMPUTE_PGM_RSRC1_GFX9_PLUS_FP16_OVFL_SHIFT,
586 amdhsa::COMPUTE_PGM_RSRC1_GFX9_PLUS_FP16_OVFL,
587 ".amdhsa_fp16_overflow");
591 amdhsa::COMPUTE_PGM_RSRC3_GFX90A_TG_SPLIT_SHIFT,
592 amdhsa::COMPUTE_PGM_RSRC3_GFX90A_TG_SPLIT,
".amdhsa_tg_split");
595 amdhsa::COMPUTE_PGM_RSRC1_GFX10_PLUS_WGP_MODE_SHIFT,
596 amdhsa::COMPUTE_PGM_RSRC1_GFX10_PLUS_WGP_MODE,
597 ".amdhsa_workgroup_processor_mode");
598 if (IVersion.
Major >= 10) {
600 amdhsa::COMPUTE_PGM_RSRC1_GFX10_PLUS_MEM_ORDERED_SHIFT,
601 amdhsa::COMPUTE_PGM_RSRC1_GFX10_PLUS_MEM_ORDERED,
602 ".amdhsa_memory_ordered");
604 amdhsa::COMPUTE_PGM_RSRC1_GFX10_PLUS_FWD_PROGRESS_SHIFT,
605 amdhsa::COMPUTE_PGM_RSRC1_GFX10_PLUS_FWD_PROGRESS,
606 ".amdhsa_forward_progress");
608 if (IVersion.
Major >= 10 && IVersion.
Major < 12) {
610 amdhsa::COMPUTE_PGM_RSRC3_GFX10_GFX11_SHARED_VGPR_COUNT_SHIFT,
611 amdhsa::COMPUTE_PGM_RSRC3_GFX10_GFX11_SHARED_VGPR_COUNT,
612 ".amdhsa_shared_vgpr_count");
614 if (IVersion.
Major == 11) {
616 amdhsa::COMPUTE_PGM_RSRC3_GFX11_INST_PREF_SIZE_SHIFT,
617 amdhsa::COMPUTE_PGM_RSRC3_GFX11_INST_PREF_SIZE,
618 ".amdhsa_inst_pref_size");
620 if (IVersion.
Major >= 12) {
622 amdhsa::COMPUTE_PGM_RSRC3_GFX12_PLUS_INST_PREF_SIZE_SHIFT,
623 amdhsa::COMPUTE_PGM_RSRC3_GFX12_PLUS_INST_PREF_SIZE,
624 ".amdhsa_inst_pref_size");
626 amdhsa::COMPUTE_PGM_RSRC1_GFX12_PLUS_ENABLE_WG_RR_EN_SHIFT,
627 amdhsa::COMPUTE_PGM_RSRC1_GFX12_PLUS_ENABLE_WG_RR_EN,
628 ".amdhsa_round_robin_scheduling");
633 COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_IEEE_754_FP_INVALID_OPERATION_SHIFT,
634 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_IEEE_754_FP_INVALID_OPERATION,
635 ".amdhsa_exception_fp_ieee_invalid_op");
638 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_FP_DENORMAL_SOURCE_SHIFT,
639 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_FP_DENORMAL_SOURCE,
640 ".amdhsa_exception_fp_denorm_src");
644 COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_IEEE_754_FP_DIVISION_BY_ZERO_SHIFT,
645 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_IEEE_754_FP_DIVISION_BY_ZERO,
646 ".amdhsa_exception_fp_ieee_div_zero");
649 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_IEEE_754_FP_OVERFLOW_SHIFT,
650 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_IEEE_754_FP_OVERFLOW,
651 ".amdhsa_exception_fp_ieee_overflow");
654 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_IEEE_754_FP_UNDERFLOW_SHIFT,
655 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_IEEE_754_FP_UNDERFLOW,
656 ".amdhsa_exception_fp_ieee_underflow");
659 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_IEEE_754_FP_INEXACT_SHIFT,
660 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_IEEE_754_FP_INEXACT,
661 ".amdhsa_exception_fp_ieee_inexact");
664 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_INT_DIVIDE_BY_ZERO_SHIFT,
665 amdhsa::COMPUTE_PGM_RSRC2_ENABLE_EXCEPTION_INT_DIVIDE_BY_ZERO,
666 ".amdhsa_exception_int_div_zero");
668 OS <<
"\t.end_amdhsa_kernel\n";
686 InfoScopeEmitter Emit) {
691 for (
const auto &[Func, Res] :
Data.Uses)
692 FuncUses[Func].push_back(Res);
693 for (
const auto &[Src, Dst] :
Data.Calls)
694 FuncCalls[Src].push_back(Dst);
695 for (
const auto &[Func, TypeId] :
Data.IndirectCalls)
696 FuncIndirectCalls[Func].push_back(TypeId);
697 for (
const auto &[Sym, TypeId] :
Data.TypeIds)
698 FuncTypeIds[Sym].push_back(TypeId);
702 if (!Emitted.insert(Sym).second)
706 if (
auto It = FuncUses.
find(Sym); It != FuncUses.
end())
708 if (
auto It = FuncCalls.
find(Sym); It != FuncCalls.
end())
710 if (
auto It = FuncIndirectCalls.
find(Sym); It != FuncIndirectCalls.
end())
711 IndirectCallTypeIds = It->second;
712 if (
auto It = FuncTypeIds.
find(Sym); It != FuncTypeIds.
end())
713 TypeIds = It->second;
714 Emit(Sym, Info,
Uses, Calls, IndirectCallTypeIds, TypeIds);
718 EmitIfNew(Func.Sym, &Func);
721 for (
const auto &[Sym, TypeId] :
Data.TypeIds)
722 EmitIfNew(Sym,
nullptr);
723 for (
const auto &[Sym, Res] :
Data.Uses)
724 EmitIfNew(Sym,
nullptr);
725 for (
const auto &[Sym, Dst] :
Data.Calls)
726 EmitIfNew(Sym,
nullptr);
727 for (
const auto &[Sym, TypeId] :
Data.IndirectCalls)
728 EmitIfNew(Sym,
nullptr);
739 OS <<
"\t.amdgpu_info " << Sym->
getName() <<
'\n';
744 if (Info->UsesFlatScratch)
746 if (Info->HasDynStack)
749 OS <<
"\t\t.amdgpu_num_sgpr " << Info->NumSGPR <<
'\n';
750 OS <<
"\t\t.amdgpu_num_vgpr " << Info->NumArchVGPR <<
'\n';
751 if (Info->NumAccVGPR)
752 OS <<
"\t\t.amdgpu_num_agpr " << Info->NumAccVGPR <<
'\n';
753 OS <<
"\t\t.amdgpu_private_segment_size " << Info->PrivateSegmentSize
757 OS <<
"\t\t.amdgpu_use " << Res->getName() <<
'\n';
759 OS <<
"\t\t.amdgpu_call " << Dst->getName() <<
'\n';
760 for (
StringRef TypeId : IndirectCallTypeIds)
761 OS <<
"\t\t.amdgpu_indirect_call \"" << TypeId <<
"\"\n";
763 OS <<
"\t\t.amdgpu_typeid \"" << TypeId <<
"\"\n";
764 OS <<
"\t.end_amdgpu_info\n\n";
785 W.setELFHeaderEFlags(getEFlags());
786 W.setOverrideABIVersion(
803void AMDGPUTargetELFStreamer::EmitNote(
807 auto &Context = S.getContext();
809 auto NameSZ = Name.size() + 1;
811 unsigned NoteFlags = 0;
821 S.emitValue(DescSZ, 4);
822 S.emitInt32(NoteType);
824 S.emitValueToAlignment(
Align(4), 0, 1, 0);
826 S.emitValueToAlignment(
Align(4), 0, 1, 0);
830unsigned AMDGPUTargetELFStreamer::getEFlags() {
835 return getEFlagsR600();
837 return getEFlagsAMDGCN();
841unsigned AMDGPUTargetELFStreamer::getEFlagsR600() {
847unsigned AMDGPUTargetELFStreamer::getEFlagsAMDGCN() {
848 assert(STI.getTargetTriple().isAMDGCN());
850 switch (STI.getTargetTriple().getOS()) {
855 return getEFlagsUnknownOS();
857 return getEFlagsAMDHSA();
859 return getEFlagsAMDPAL();
861 return getEFlagsMesa3D();
865unsigned AMDGPUTargetELFStreamer::getEFlagsUnknownOS() {
869 return getEFlagsV3();
872unsigned AMDGPUTargetELFStreamer::getEFlagsAMDHSA() {
876 return getEFlagsV6();
877 return getEFlagsV4();
880unsigned AMDGPUTargetELFStreamer::getEFlagsAMDPAL() {
883 return getEFlagsV3();
886unsigned AMDGPUTargetELFStreamer::getEFlagsMesa3D() {
889 return getEFlagsV3();
892unsigned AMDGPUTargetELFStreamer::getEFlagsV3() {
893 unsigned EFlagsV3 = 0;
908unsigned AMDGPUTargetELFStreamer::getEFlagsV4() {
909 unsigned EFlagsV4 = 0;
948unsigned AMDGPUTargetELFStreamer::getEFlagsV6() {
949 unsigned Flags = getEFlagsV4();
954 case AMDGPU::GK_GFX9_GENERIC:
957 case AMDGPU::GK_GFX9_4_GENERIC:
960 case AMDGPU::GK_GFX10_1_GENERIC:
963 case AMDGPU::GK_GFX10_3_GENERIC:
966 case AMDGPU::GK_GFX11_GENERIC:
969 case AMDGPU::GK_GFX12_GENERIC:
972 case AMDGPU::GK_GFX12_5_GENERIC:
985 " - no ELF flag can represent this version!");
1010 auto *SymbolELF =
static_cast<MCSymbolELF *
>(Symbol);
1013 if (!SymbolELF->isBindingSet())
1016 if (SymbolELF->declareCommon(
Size, Alignment)) {
1018 " redeclared as different type");
1029 auto *DescBegin = Context.createTempSymbol();
1030 auto *DescEnd = Context.createTempSymbol();
1052 if (!Verifier.verify(HSAMetadataDoc.
getRoot()))
1055 std::string HSAMetadataString;
1061 auto *DescBegin = Context.createTempSymbol();
1062 auto *DescEnd = Context.createTempSymbol();
1077 const uint32_t Encoded_s_code_end = 0xbf9f0000;
1078 const uint32_t Encoded_s_nop = 0xbf800000;
1079 uint32_t Encoded_pad = Encoded_s_code_end;
1089 Encoded_pad = Encoded_s_nop;
1096 for (
unsigned I = 0;
I < FillSize;
I += 4)
1106 const MCExpr *ReserveFlatScr) {
1108 auto &Context = Streamer.getContext();
1110 auto *KernelCodeSymbol =
1112 auto *KernelDescriptorSymbol =
static_cast<MCSymbolELF *
>(
1113 Context.getOrCreateSymbol(
Twine(KernelName) +
Twine(
".kd")));
1117 KernelDescriptorSymbol->
setBinding(KernelCodeSymbol->getBinding());
1118 KernelDescriptorSymbol->setOther(KernelCodeSymbol->getOther());
1119 KernelDescriptorSymbol->setVisibility(KernelCodeSymbol->getVisibility());
1122 KernelDescriptorSymbol->setSize(
1130 Streamer.emitLabel(KernelDescriptorSymbol);
1141 Streamer.emitInt8(0u);
1154 Streamer.emitInt8(0u);
1167 Streamer.emitInt8(0u);
1179 return StrTab.
add(Str);
1209 if (Info->UsesFlatScratch)
1211 if (Info->HasDynStack)
1218 if (Info->NumAccVGPR)
1221 Info->PrivateSegmentSize);
1228 for (
StringRef TypeId : IndirectCallTypeIds) {
1230 getOrAddString(TypeId));
1236 if (!StrTab.
empty()) {
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
AMDHSA kernel descriptor MCExpr struct for use in MC layer.
Provides AMDGPU specific target descriptions.
Enums shared between the AMDGPU backend (LLVM) and the ELF linker (LLD) for the .amdgpu....
Enums and constants for AMDGPU PT_NOTE sections.
static cl::opt< unsigned > ForceGenericVersion("amdgpu-force-generic-version", cl::desc("Force a specific generic_v<N> flag to be " "added. For testing purposes only."), cl::ReallyHidden, cl::init(0))
#define PRINT_RES_INFO(ARG)
AMDHSA kernel descriptor definitions.
MC layer struct for AMDGPUMCKernelCodeT, provides MCExpr functionality where required.
Remove Loads Into Fake Uses
verify safepoint Safepoint IR Verifier
void emitAMDGPUInfo(const AMDGPU::InfoSectionData &Data) override
AMDGPUTargetAsmStreamer(MCStreamer &S, formatted_raw_ostream &OS)
bool EmitHSAMetadata(msgpack::Document &HSAMetadata, bool Strict) override
void EmitAMDGPUSymbolType(StringRef SymbolName, unsigned Type) override
bool EmitISAVersion() override
void EmitDirectiveAMDHSACodeObjectVersion(unsigned COV) override
void EmitDirectiveAMDGCNTarget() override
void EmitMCResourceMaximums(const MCSymbol *MaxVGPR, const MCSymbol *MaxAGPR, const MCSymbol *MaxSGPR, const MCSymbol *MaxNamedBarrier) override
void EmitAMDKernelCodeT(AMDGPU::AMDGPUMCKernelCodeT &Header) override
void EmitAmdhsaKernelDescriptor(const MCSubtargetInfo &STI, StringRef KernelName, const AMDGPU::MCKernelDescriptor &KernelDescriptor, const MCExpr *NextVGPR, const MCExpr *NextSGPR, const MCExpr *ReserveVCC, const MCExpr *ReserveFlatScr) override
void EmitMCResourceInfo(const MCSymbol *NumVGPR, const MCSymbol *NumAGPR, const MCSymbol *NumExplicitSGPR, const MCSymbol *NumNamedBarrier, const MCSymbol *PrivateSegmentSize, const MCSymbol *UsesVCC, const MCSymbol *UsesFlatScratch, const MCSymbol *HasDynamicallySizedStack, const MCSymbol *HasRecursion, const MCSymbol *HasIndirectCall) override
bool EmitCodeEnd(const MCSubtargetInfo &STI) override
void emitAMDGPULDS(MCSymbol *Sym, unsigned Size, Align Alignment) override
void EmitDirectiveAMDGCNTarget() override
bool EmitCodeEnd(const MCSubtargetInfo &STI) override
void EmitAMDKernelCodeT(AMDGPU::AMDGPUMCKernelCodeT &Header) override
bool EmitHSAMetadata(msgpack::Document &HSAMetadata, bool Strict) override
AMDGPUTargetELFStreamer(MCStreamer &S, const MCSubtargetInfo &STI)
void emitAMDGPULDS(MCSymbol *Sym, unsigned Size, Align Alignment) override
void EmitAmdhsaKernelDescriptor(const MCSubtargetInfo &STI, StringRef KernelName, const AMDGPU::MCKernelDescriptor &KernelDescriptor, const MCExpr *NextVGPR, const MCExpr *NextSGPR, const MCExpr *ReserveVCC, const MCExpr *ReserveFlatScr) override
MCELFStreamer & getStreamer()
void EmitAMDGPUSymbolType(StringRef SymbolName, unsigned Type) override
void emitAMDGPUInfo(const AMDGPU::InfoSectionData &Data) override
bool EmitISAVersion() override
virtual bool EmitHSAMetadata(msgpack::Document &HSAMetadata, bool Strict)
Emit HSA Metadata.
AMDGPUPALMetadata * getPALMetadata()
AMDGPUTargetStreamer(MCStreamer &S)
virtual void EmitDirectiveAMDHSACodeObjectVersion(unsigned COV)
virtual bool EmitHSAMetadataV3(StringRef HSAMetadataString)
static unsigned getElfMach(StringRef GPU)
MCContext & getContext() const
static StringRef getArchNameFromElfMach(unsigned ElfMach)
const std::optional< AMDGPU::IsaInfo::AMDGPUTargetID > & getTargetID() const
unsigned CodeObjectVersion
Represent a constant reference to an array (0 or more elements consecutively in memory),...
iterator find(const_arg_type_t< KeyT > Val)
Implements a dense probed hash-table based set.
This class is intended to be used as a base class for asm properties and features specific to the tar...
static const MCBinaryExpr * createAdd(const MCExpr *LHS, const MCExpr *RHS, MCContext &Ctx, SMLoc Loc=SMLoc())
static const MCBinaryExpr * createMul(const MCExpr *LHS, const MCExpr *RHS, MCContext &Ctx)
static const MCBinaryExpr * createSub(const MCExpr *LHS, const MCExpr *RHS, MCContext &Ctx)
static LLVM_ABI const MCConstantExpr * create(int64_t Value, MCContext &Ctx, bool PrintInHex=false, unsigned SizeInBytes=0)
Context object for machine code objects.
LLVM_ABI MCSymbol * getOrCreateSymbol(const Twine &Name)
Lookup the symbol inside with the specified Name.
const MCAsmInfo & getAsmInfo() const
ELFObjectWriter & getWriter()
void emitLabel(MCSymbol *Symbol, SMLoc Loc=SMLoc()) override
Emit a label for Symbol into the current section.
Base class for the full range of assembler expressions which are needed for parsing.
void emitBytes(StringRef Data) override
Emit the bytes in Data into the output.
This represents a section on linux, lots of unix variants and some bare metal systems.
Streaming machine code generation interface.
virtual bool popSection()
Restore the current and previous section from the section stack.
MCContext & getContext() const
void emitValue(const MCExpr *Value, unsigned Size, SMLoc Loc=SMLoc())
virtual void emitValueToAlignment(Align Alignment, int64_t Fill=0, uint8_t FillLen=1, unsigned MaxBytesToEmit=0)
Emit some number of copies of Value until the byte alignment ByteAlignment is reached.
void pushSection()
Save the current and previous section on the section stack.
virtual void switchSection(MCSection *Section, uint32_t Subsec=0)
Set the current section where code is being emitted to Section.
void emitInt32(uint64_t Value)
void emitInt8(uint64_t Value)
Generic base class for all target subtargets.
bool hasFeature(unsigned Feature) const
const Triple & getTargetTriple() const
void setBinding(unsigned Binding) const
void setType(unsigned Type) const
static const MCSymbolRefExpr * create(const MCSymbol *Symbol, MCContext &Ctx, SMLoc Loc=SMLoc())
MCSymbol - Instances of this class represent a symbol name in the MC file, and MCSymbols are created ...
StringRef getName() const
getName - Get the symbol name.
SmallString - A SmallString is just a SmallVector with methods and accessors that make it work better...
Represent a constant reference to a string, i.e.
Utility for building string tables with deduplicated suffixes.
LLVM_ABI void finalizeInOrder()
Finalize the string table without reording it.
LLVM_ABI size_t add(CachedHashStringRef S, uint8_t Priority=0)
Add a string to the builder.
LLVM_ABI void write(raw_ostream &OS) const
ArchType getArch() const
Get the parsed architecture type of this triple.
Twine - A lightweight data structure for efficiently representing the concatenation of temporary valu...
The instances of the Type class are immutable: once they are created, they are never changed.
LLVM Value Representation.
An efficient, type-erasing, non-owning reference to a callable.
Simple in-memory representation of a document of msgpack objects with ability to find and create arra...
DocNode & getRoot()
Get ref to the document's root element.
LLVM_ABI void toYAML(raw_ostream &OS)
Convert MsgPack Document to YAML text.
LLVM_ABI void writeToBlob(std::string &Blob)
Write a MsgPack document to a binary MsgPack blob.
LLVM_ABI bool fromYAML(StringRef S)
Read YAML text into the MsgPack document. Returns false on failure.
This class implements an extremely fast bulk output stream that can only output to a stream.
A raw_ostream that writes to an std::string.
std::string & str()
Returns the string's reference.
A raw_ostream that writes to an SmallVector or SmallString.
StringRef str() const
Return a StringRef for the vector contents.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
static constexpr unsigned GFX12_5
static constexpr unsigned GFX9_4
static constexpr unsigned GFX10_1
static constexpr unsigned GFX10_3
static constexpr unsigned GFX11
static constexpr unsigned GFX9
static constexpr unsigned GFX12
constexpr char AssemblerDirectiveBegin[]
HSA metadata beginning assembler directive.
constexpr char AssemblerDirectiveEnd[]
HSA metadata ending assembler directive.
LLVM_ABI StringRef getArchNameR600(GPUKind AK)
FuncInfoFlags
Per-function flags packed into INFO_FLAGS entries.
GPUKind
GPU kinds supported by the AMDGPU target.
void printAMDGPUMCExpr(const MCExpr *Expr, raw_ostream &OS, const MCAsmInfo *MAI)
bool isHsaAbi(const MCSubtargetInfo &STI)
LLVM_ABI IsaVersion getIsaVersion(StringRef GPU)
bool isGFX90A(const MCSubtargetInfo &STI)
LLVM_ABI GPUKind parseArchAMDGCN(StringRef CPU)
bool hasArchitectedFlatScratch(const MCSubtargetInfo &STI)
bool isGFX11Plus(const MCSubtargetInfo &STI)
const MCExpr * foldAMDGPUMCExpr(const MCExpr *Expr, MCContext &Ctx)
InfoKind
Entry kind values for the .amdgpu.info section.
@ INFO_INDIRECT_CALL
Indirect call edge: the function contains an indirect call whose callee is expected to match the type...
@ INFO_FLAGS
Bitfield of FuncInfoFlags properties for the function. [u32].
@ INFO_FUNC
Opens a new function scope.
@ INFO_NUM_SGPR
Number of SGPRs explicitly used by the function. [u32].
@ INFO_NUM_VGPR
Number of architectural VGPRs used by the function. [u32].
@ INFO_CALL
Direct call edge: the function calls the callee identified by the 8-byte relocated symbol.
@ INFO_NUM_AGPR
Number of accumulator VGPRs (AGPRs) used by the function. [u32].
@ INFO_TYPEID
Function type ID: tags an address-taken function with a type-ID string (at the given ....
@ INFO_PRIVATE_SEGMENT_SIZE
Private (scratch) memory size in bytes required by the function. [u32].
@ INFO_USE
Dependency edge: the function uses the resource identified by the 8-byte relocated symbol (e....
LLVM_ABI StringRef getArchNameAMDGCN(GPUKind AK)
unsigned hasKernargPreload(const MCSubtargetInfo &STI)
bool supportsWGP(const MCSubtargetInfo &STI)
bool isGFX1250Plus(const MCSubtargetInfo &STI)
uint8_t getELFABIVersion(const Triple &T, unsigned CodeObjectVersion)
LLVM_ABI GPUKind parseArchR600(StringRef CPU)
@ EF_AMDGPU_GENERIC_VERSION_MAX
@ EF_AMDGPU_FEATURE_XNACK_ANY_V4
@ EF_AMDGPU_FEATURE_SRAMECC_V3
@ EF_AMDGPU_GENERIC_VERSION_OFFSET
@ EF_AMDGPU_FEATURE_SRAMECC_UNSUPPORTED_V4
@ EF_AMDGPU_FEATURE_SRAMECC_OFF_V4
@ EF_AMDGPU_FEATURE_XNACK_UNSUPPORTED_V4
@ EF_AMDGPU_FEATURE_XNACK_OFF_V4
@ EF_AMDGPU_FEATURE_XNACK_V3
@ EF_AMDGPU_FEATURE_XNACK_ON_V4
@ EF_AMDGPU_FEATURE_SRAMECC_ANY_V4
@ EF_AMDGPU_FEATURE_SRAMECC_ON_V4
initializer< Ty > init(const Ty &Val)
This is an optimization pass for GlobalISel generic memory operations.
FunctionAddr VTableAddr uintptr_t uintptr_t Version
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
constexpr std::underlying_type_t< Enum > to_underlying(Enum E)
Returns underlying integer value of an enum.
FunctionAddr VTableAddr uintptr_t uintptr_t Data
Instruction set architecture version.
const MCExpr * compute_pgm_rsrc2
const MCExpr * kernarg_size
const MCExpr * kernarg_preload
const MCExpr * compute_pgm_rsrc3
const MCExpr * private_segment_fixed_size
static const MCExpr * bits_get(const MCExpr *Src, uint32_t Shift, uint32_t Mask, MCContext &Ctx)
const MCExpr * compute_pgm_rsrc1
const MCExpr * group_segment_fixed_size
const MCExpr * kernel_code_properties
This struct is a compact representation of a valid (non-zero power of two) alignment.
constexpr uint64_t value() const
This is a hole in the type system and should not be abused.
uint32_t group_segment_fixed_size
uint32_t compute_pgm_rsrc1
uint32_t private_segment_fixed_size
uint32_t compute_pgm_rsrc2
uint16_t kernel_code_properties
uint32_t compute_pgm_rsrc3
int64_t kernel_code_entry_byte_offset