|
1 | 1 | ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
2 |
| -; RUN: llc -mtriple=amdgcn-mesa-mesa3d -mcpu=gfx908 -amdgpu-atomic-optimizer-strategy=None -verify-machineinstrs < %s | FileCheck %s -check-prefix=CHECK |
| 2 | +; RUN: llc -mtriple=amdgcn-mesa-mesa3d -mcpu=gfx908 < %s | FileCheck -check-prefix=CHECK %s |
3 | 3 |
|
4 |
| -define amdgpu_ps void @raw_ptr_buffer_atomic_add_f32_noret__vgpr_val__sgpr_rsrc__vgpr_voffset__sgpr_soffset(float %val, ptr addrspace(8) inreg %rsrc, i32 %voffset, i32 inreg %soffset) { |
| 4 | +define void @raw_ptr_buffer_atomic_add_f32_noret__vgpr_val__sgpr_rsrc__vgpr_voffset__sgpr_soffset(float %val, ptr addrspace(8) inreg %rsrc, i32 %voffset, i32 inreg %soffset) { |
5 | 5 | ; CHECK-LABEL: raw_ptr_buffer_atomic_add_f32_noret__vgpr_val__sgpr_rsrc__vgpr_voffset__sgpr_soffset:
|
6 | 6 | ; CHECK: ; %bb.0:
|
7 |
| -; CHECK-NEXT: s_mov_b32 s11, s5 |
8 |
| -; CHECK-NEXT: s_mov_b32 s10, s4 |
9 |
| -; CHECK-NEXT: s_mov_b32 s9, s3 |
10 |
| -; CHECK-NEXT: s_mov_b32 s8, s2 |
11 |
| -; CHECK-NEXT: buffer_atomic_add_f32 v0, v1, s[8:11], s6 offen |
12 |
| -; CHECK-NEXT: s_endpgm |
| 7 | +; CHECK-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) |
| 8 | +; CHECK-NEXT: buffer_atomic_add_f32 v0, v1, s[4:7], s8 offen |
| 9 | +; CHECK-NEXT: s_waitcnt vmcnt(0) |
| 10 | +; CHECK-NEXT: s_setpc_b64 s[30:31] |
13 | 11 | %ret = call float @llvm.amdgcn.raw.ptr.buffer.atomic.fadd.f32(float %val, ptr addrspace(8) %rsrc, i32 %voffset, i32 %soffset, i32 24)
|
14 | 12 | ret void
|
15 | 13 | }
|
16 | 14 |
|
17 |
| -define amdgpu_ps void @raw_ptr_buffer_atomic_add_f32_noret__vgpr_val__sgpr_rsrc__0_voffset__sgpr_soffset(float %val, ptr addrspace(8) inreg %rsrc, i32 inreg %soffset) { |
| 15 | +define void @raw_ptr_buffer_atomic_add_f32_noret__vgpr_val__sgpr_rsrc__0_voffset__sgpr_soffset(float %val, ptr addrspace(8) inreg %rsrc, i32 inreg %soffset) { |
18 | 16 | ; CHECK-LABEL: raw_ptr_buffer_atomic_add_f32_noret__vgpr_val__sgpr_rsrc__0_voffset__sgpr_soffset:
|
19 | 17 | ; CHECK: ; %bb.0:
|
20 |
| -; CHECK-NEXT: s_mov_b32 s11, s5 |
21 |
| -; CHECK-NEXT: s_mov_b32 s10, s4 |
22 |
| -; CHECK-NEXT: s_mov_b32 s9, s3 |
23 |
| -; CHECK-NEXT: s_mov_b32 s8, s2 |
24 |
| -; CHECK-NEXT: buffer_atomic_add_f32 v0, off, s[8:11], s6 |
25 |
| -; CHECK-NEXT: s_endpgm |
| 18 | +; CHECK-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) |
| 19 | +; CHECK-NEXT: buffer_atomic_add_f32 v0, off, s[4:7], s8 |
| 20 | +; CHECK-NEXT: s_waitcnt vmcnt(0) |
| 21 | +; CHECK-NEXT: s_setpc_b64 s[30:31] |
26 | 22 | %ret = call float @llvm.amdgcn.raw.ptr.buffer.atomic.fadd.f32(float %val, ptr addrspace(8) %rsrc, i32 0, i32 %soffset, i32 0)
|
27 | 23 | ret void
|
28 | 24 | }
|
29 | 25 |
|
30 |
| -define amdgpu_ps void @raw_ptr_buffer_atomic_add_v2f16_noret__vgpr_val__sgpr_rsrc__vgpr_voffset__sgpr_soffset(<2 x half> %val, ptr addrspace(8) inreg %rsrc, i32 %voffset, i32 inreg %soffset) { |
| 26 | +define void @raw_ptr_buffer_atomic_add_v2f16_noret__vgpr_val__sgpr_rsrc__vgpr_voffset__sgpr_soffset(<2 x half> %val, ptr addrspace(8) inreg %rsrc, i32 %voffset, i32 inreg %soffset) { |
31 | 27 | ; CHECK-LABEL: raw_ptr_buffer_atomic_add_v2f16_noret__vgpr_val__sgpr_rsrc__vgpr_voffset__sgpr_soffset:
|
32 | 28 | ; CHECK: ; %bb.0:
|
33 |
| -; CHECK-NEXT: s_mov_b32 s11, s5 |
34 |
| -; CHECK-NEXT: s_mov_b32 s10, s4 |
35 |
| -; CHECK-NEXT: s_mov_b32 s9, s3 |
36 |
| -; CHECK-NEXT: s_mov_b32 s8, s2 |
37 |
| -; CHECK-NEXT: buffer_atomic_pk_add_f16 v0, v1, s[8:11], s6 offen |
38 |
| -; CHECK-NEXT: s_endpgm |
| 29 | +; CHECK-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) |
| 30 | +; CHECK-NEXT: buffer_atomic_pk_add_f16 v0, v1, s[4:7], s8 offen |
| 31 | +; CHECK-NEXT: s_waitcnt vmcnt(0) |
| 32 | +; CHECK-NEXT: s_setpc_b64 s[30:31] |
39 | 33 | %ret = call <2 x half> @llvm.amdgcn.raw.ptr.buffer.atomic.fadd.v2f16(<2 x half> %val, ptr addrspace(8) %rsrc, i32 %voffset, i32 %soffset, i32 0)
|
40 | 34 | ret void
|
41 | 35 | }
|
42 | 36 |
|
43 |
| -define amdgpu_ps void @raw_ptr_buffer_atomic_add_v2f16_noret__vgpr_val__sgpr_rsrc__0_voffset__sgpr_soffset(<2 x half> %val, ptr addrspace(8) inreg %rsrc, i32 %voffset, i32 inreg %soffset) { |
| 37 | +define void @raw_ptr_buffer_atomic_add_v2f16_noret__vgpr_val__sgpr_rsrc__0_voffset__sgpr_soffset(<2 x half> %val, ptr addrspace(8) inreg %rsrc, i32 %voffset, i32 inreg %soffset) { |
44 | 38 | ; CHECK-LABEL: raw_ptr_buffer_atomic_add_v2f16_noret__vgpr_val__sgpr_rsrc__0_voffset__sgpr_soffset:
|
45 | 39 | ; CHECK: ; %bb.0:
|
46 |
| -; CHECK-NEXT: s_mov_b32 s11, s5 |
47 |
| -; CHECK-NEXT: s_mov_b32 s10, s4 |
48 |
| -; CHECK-NEXT: s_mov_b32 s9, s3 |
49 |
| -; CHECK-NEXT: s_mov_b32 s8, s2 |
50 |
| -; CHECK-NEXT: buffer_atomic_pk_add_f16 v0, off, s[8:11], s6 offset:92 |
51 |
| -; CHECK-NEXT: s_endpgm |
| 40 | +; CHECK-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) |
| 41 | +; CHECK-NEXT: buffer_atomic_pk_add_f16 v0, off, s[4:7], s8 offset:92 |
| 42 | +; CHECK-NEXT: s_waitcnt vmcnt(0) |
| 43 | +; CHECK-NEXT: s_setpc_b64 s[30:31] |
52 | 44 | %ret = call <2 x half> @llvm.amdgcn.raw.ptr.buffer.atomic.fadd.v2f16(<2 x half> %val, ptr addrspace(8) %rsrc, i32 92, i32 %soffset, i32 0)
|
53 | 45 | ret void
|
54 | 46 | }
|
55 | 47 |
|
56 |
| -define amdgpu_ps void @raw_ptr_buffer_atomic_add_f32_noret__vgpr_val__sgpr_rsrc__vgpr_voffset__sgpr_soffset_slc(float %val, ptr addrspace(8) inreg %rsrc, i32 %voffset, i32 inreg %soffset) { |
| 48 | +define void @raw_ptr_buffer_atomic_add_f32_noret__vgpr_val__sgpr_rsrc__vgpr_voffset__sgpr_soffset_slc(float %val, ptr addrspace(8) inreg %rsrc, i32 %voffset, i32 inreg %soffset) { |
57 | 49 | ; CHECK-LABEL: raw_ptr_buffer_atomic_add_f32_noret__vgpr_val__sgpr_rsrc__vgpr_voffset__sgpr_soffset_slc:
|
58 | 50 | ; CHECK: ; %bb.0:
|
59 |
| -; CHECK-NEXT: s_mov_b32 s11, s5 |
60 |
| -; CHECK-NEXT: s_mov_b32 s10, s4 |
61 |
| -; CHECK-NEXT: s_mov_b32 s9, s3 |
62 |
| -; CHECK-NEXT: s_mov_b32 s8, s2 |
63 |
| -; CHECK-NEXT: buffer_atomic_add_f32 v0, v1, s[8:11], s6 offen slc |
64 |
| -; CHECK-NEXT: s_endpgm |
| 51 | +; CHECK-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) |
| 52 | +; CHECK-NEXT: buffer_atomic_add_f32 v0, v1, s[4:7], s8 offen slc |
| 53 | +; CHECK-NEXT: s_waitcnt vmcnt(0) |
| 54 | +; CHECK-NEXT: s_setpc_b64 s[30:31] |
65 | 55 | %ret = call float @llvm.amdgcn.raw.ptr.buffer.atomic.fadd.f32(float %val, ptr addrspace(8) %rsrc, i32 %voffset, i32 %soffset, i32 2)
|
66 | 56 | ret void
|
67 | 57 | }
|
|
0 commit comments