@@ -201,10 +201,10 @@ define amdgpu_kernel void @fast_add_fmuladd_fmul_multi_use_fmuladd_commute() #0
201
201
202
202
; GCN-SLOWFMA-DAG: v_mul_f32_e32 v{{[0-9]+}}, [[X]], [[Y]]
203
203
; GCN-SLOWFMA: v_add_f32_e32
204
- ; GCN-SLOWFMA: v_sub_f32_e32 [[MAD :v[0-9]+]]
204
+ ; GCN-SLOWFMA: v_sub_f32_e32 [[SUB :v[0-9]+]]
205
205
206
206
; GCN: buffer_store_dword [[MUL]]
207
- ; GCN: buffer_store_dword [[MAD ]]
207
+ ; GCN: buffer_store_dword [[SUB ]]
208
208
define amdgpu_kernel void @fast_sub_fmuladd_fmul_multi_use_mul () #0 {
209
209
%x = load volatile float , float addrspace (1 )* undef
210
210
%y = load volatile float , float addrspace (1 )* undef
@@ -213,9 +213,9 @@ define amdgpu_kernel void @fast_sub_fmuladd_fmul_multi_use_mul() #0 {
213
213
%v = load volatile float , float addrspace (1 )* undef
214
214
%mul.u.v = fmul fast float %u , %v
215
215
%fma = call fast float @llvm.fmuladd.f32 (float %x , float %y , float %mul.u.v )
216
- %add = fsub fast float %fma , %z
216
+ %sub = fsub fast float %fma , %z
217
217
store volatile float %mul.u.v , float addrspace (1 )* undef
218
- store volatile float %add , float addrspace (1 )* undef
218
+ store volatile float %sub , float addrspace (1 )* undef
219
219
ret void
220
220
}
221
221
0 commit comments