26
26
27
27
28
28
; VMEM: v_mov_b32_e32 v[[V_SAVEEXEC_LO:[0-9]+]], s[[SAVEEXEC_LO]]
29
- ; VMEM: buffer_store_dword v[[V_SAVEEXEC_LO]], off, s[8:11 ], s12 ; 8-byte Folded Spill
29
+ ; VMEM: buffer_store_dword v[[V_SAVEEXEC_LO]], off, s[0:3 ], s7 ; 8-byte Folded Spill
30
30
; VMEM: v_mov_b32_e32 v[[V_SAVEEXEC_HI:[0-9]+]], s[[SAVEEXEC_HI]]
31
- ; VMEM: buffer_store_dword v[[V_SAVEEXEC_HI]], off, s[8:11 ], s12 offset:4 ; 8-byte Folded Spill
31
+ ; VMEM: buffer_store_dword v[[V_SAVEEXEC_HI]], off, s[0:3 ], s7 offset:4 ; 8-byte Folded Spill
32
32
33
33
; Spill load
34
- ; GCN: buffer_store_dword [[LOAD0]], off, s[8:11 ], s12 offset:[[LOAD0_OFFSET:[0-9]+]] ; 4-byte Folded Spill
34
+ ; GCN: buffer_store_dword [[LOAD0]], off, s[0:3 ], s7 offset:[[LOAD0_OFFSET:[0-9]+]] ; 4-byte Folded Spill
35
35
; GCN: s_mov_b64 exec, s{{\[}}[[ANDEXEC_LO]]:[[ANDEXEC_HI]]{{\]}}
36
36
37
37
; GCN: s_waitcnt vmcnt(0) expcnt(0)
40
40
; GCN: {{^}}BB{{[0-9]+}}_1: ; %if
41
41
; GCN: s_mov_b32 m0, -1
42
42
; GCN: ds_read_b32 [[LOAD1:v[0-9]+]]
43
- ; GCN: buffer_load_dword [[RELOAD_LOAD0:v[0-9]+]], off, s[8:11 ], s12 offset:[[LOAD0_OFFSET]] ; 4-byte Folded Reload
43
+ ; GCN: buffer_load_dword [[RELOAD_LOAD0:v[0-9]+]], off, s[0:3 ], s7 offset:[[LOAD0_OFFSET]] ; 4-byte Folded Reload
44
44
; GCN: s_waitcnt vmcnt(0)
45
45
46
46
; Spill val register
47
47
; GCN: v_add_i32_e32 [[VAL:v[0-9]+]], vcc, [[LOAD1]], [[RELOAD_LOAD0]]
48
- ; GCN: buffer_store_dword [[VAL]], off, s[8:11 ], s12 offset:[[VAL_OFFSET:[0-9]+]] ; 4-byte Folded Spill
48
+ ; GCN: buffer_store_dword [[VAL]], off, s[0:3 ], s7 offset:[[VAL_OFFSET:[0-9]+]] ; 4-byte Folded Spill
49
49
; GCN: s_waitcnt vmcnt(0)
50
50
51
51
; VMEM: [[ENDIF]]:
55
55
56
56
57
57
58
- ; VMEM: buffer_load_dword v[[V_RELOAD_SAVEEXEC_LO:[0-9]+]], off, s[8:11 ], s12 ; 8-byte Folded Reload
58
+ ; VMEM: buffer_load_dword v[[V_RELOAD_SAVEEXEC_LO:[0-9]+]], off, s[0:3 ], s7 ; 8-byte Folded Reload
59
59
; VMEM: s_waitcnt vmcnt(0)
60
60
; VMEM: v_readfirstlane_b32 s[[S_RELOAD_SAVEEXEC_LO:[0-9]+]], v[[V_RELOAD_SAVEEXEC_LO]]
61
61
62
- ; VMEM: buffer_load_dword v[[V_RELOAD_SAVEEXEC_HI:[0-9]+]], off, s[8:11 ], s12 offset:4 ; 8-byte Folded Reload
62
+ ; VMEM: buffer_load_dword v[[V_RELOAD_SAVEEXEC_HI:[0-9]+]], off, s[0:3 ], s7 offset:4 ; 8-byte Folded Reload
63
63
; VMEM: s_waitcnt vmcnt(0)
64
64
; VMEM: v_readfirstlane_b32 s[[S_RELOAD_SAVEEXEC_HI:[0-9]+]], v[[V_RELOAD_SAVEEXEC_HI]]
65
65
66
66
; GCN: s_or_b64 exec, exec, s{{\[}}[[S_RELOAD_SAVEEXEC_LO]]:[[S_RELOAD_SAVEEXEC_HI]]{{\]}}
67
67
68
68
; Restore val
69
- ; GCN: buffer_load_dword [[RELOAD_VAL:v[0-9]+]], off, s[8:11 ], s12 offset:[[VAL_OFFSET]] ; 4-byte Folded Reload
69
+ ; GCN: buffer_load_dword [[RELOAD_VAL:v[0-9]+]], off, s[0:3 ], s7 offset:[[VAL_OFFSET]] ; 4-byte Folded Reload
70
70
71
71
; GCN: flat_store_dword v{{\[[0-9]+:[0-9]+\]}}, [[RELOAD_VAL]]
72
72
define void @divergent_if_endif (i32 addrspace (1 )* %out ) #0 {
@@ -105,12 +105,12 @@ endif:
105
105
106
106
107
107
; VMEM: v_mov_b32_e32 v[[V_SAVEEXEC_LO:[0-9]+]], s[[SAVEEXEC_LO]]
108
- ; VMEM: buffer_store_dword v[[V_SAVEEXEC_LO]], off, s[8:11 ], s12 ; 8-byte Folded Spill
108
+ ; VMEM: buffer_store_dword v[[V_SAVEEXEC_LO]], off, s[0:3 ], s7 ; 8-byte Folded Spill
109
109
; VMEM: v_mov_b32_e32 v[[V_SAVEEXEC_HI:[0-9]+]], s[[SAVEEXEC_HI]]
110
- ; VMEM: buffer_store_dword v[[V_SAVEEXEC_HI]], off, s[8:11 ], s12 offset:4 ; 8-byte Folded Spill
110
+ ; VMEM: buffer_store_dword v[[V_SAVEEXEC_HI]], off, s[0:3 ], s7 offset:4 ; 8-byte Folded Spill
111
111
112
112
; Spill load
113
- ; GCN: buffer_store_dword [[LOAD0]], off, s[8:11 ], s12 offset:[[VAL_OFFSET:[0-9]+]] ; 4-byte Folded Spill
113
+ ; GCN: buffer_store_dword [[LOAD0]], off, s[0:3 ], s7 offset:[[VAL_OFFSET:[0-9]+]] ; 4-byte Folded Spill
114
114
115
115
; GCN: s_mov_b64 exec, s{{\[}}[[ANDEXEC_LO]]:[[ANDEXEC_HI]]{{\]}}
116
116
@@ -120,11 +120,11 @@ endif:
120
120
121
121
122
122
; GCN: [[LOOP:BB[0-9]+_[0-9]+]]:
123
- ; GCN: buffer_load_dword v[[VAL_LOOP_RELOAD:[0-9]+]], off, s[8:11 ], s12 offset:[[VAL_OFFSET]] ; 4-byte Folded Reload
123
+ ; GCN: buffer_load_dword v[[VAL_LOOP_RELOAD:[0-9]+]], off, s[0:3 ], s7 offset:[[VAL_OFFSET]] ; 4-byte Folded Reload
124
124
; GCN: v_subrev_i32_e32 [[VAL_LOOP:v[0-9]+]], vcc, v{{[0-9]+}}, v[[VAL_LOOP_RELOAD]]
125
125
; GCN: v_cmp_ne_u32_e32 vcc,
126
126
; GCN: s_and_b64 vcc, exec, vcc
127
- ; GCN: buffer_store_dword [[VAL_LOOP]], off, s[8:11 ], s12 offset:[[VAL_SUB_OFFSET:[0-9]+]] ; 4-byte Folded Spill
127
+ ; GCN: buffer_store_dword [[VAL_LOOP]], off, s[0:3 ], s7 offset:[[VAL_SUB_OFFSET:[0-9]+]] ; 4-byte Folded Spill
128
128
; GCN: s_waitcnt vmcnt(0) expcnt(0)
129
129
; GCN-NEXT: s_cbranch_vccnz [[LOOP]]
130
130
@@ -133,16 +133,16 @@ endif:
133
133
; VGPR: v_readlane_b32 s[[S_RELOAD_SAVEEXEC_LO:[0-9]+]], [[SPILL_VGPR]], [[SAVEEXEC_LO_LANE]]
134
134
; VGPR: v_readlane_b32 s[[S_RELOAD_SAVEEXEC_HI:[0-9]+]], [[SPILL_VGPR]], [[SAVEEXEC_HI_LANE]]
135
135
136
- ; VMEM: buffer_load_dword v[[V_RELOAD_SAVEEXEC_LO:[0-9]+]], off, s[8:11 ], s12 ; 8-byte Folded Reload
136
+ ; VMEM: buffer_load_dword v[[V_RELOAD_SAVEEXEC_LO:[0-9]+]], off, s[0:3 ], s7 ; 8-byte Folded Reload
137
137
; VMEM: s_waitcnt vmcnt(0)
138
138
; VMEM: v_readfirstlane_b32 s[[S_RELOAD_SAVEEXEC_LO:[0-9]+]], v[[V_RELOAD_SAVEEXEC_LO]]
139
139
140
- ; VMEM: buffer_load_dword v[[V_RELOAD_SAVEEXEC_HI:[0-9]+]], off, s[8:11 ], s12 offset:4 ; 8-byte Folded Reload
140
+ ; VMEM: buffer_load_dword v[[V_RELOAD_SAVEEXEC_HI:[0-9]+]], off, s[0:3 ], s7 offset:4 ; 8-byte Folded Reload
141
141
; VMEM: s_waitcnt vmcnt(0)
142
142
; VMEM: v_readfirstlane_b32 s[[S_RELOAD_SAVEEXEC_HI:[0-9]+]], v[[V_RELOAD_SAVEEXEC_HI]]
143
143
144
144
; GCN: s_or_b64 exec, exec, s{{\[}}[[S_RELOAD_SAVEEXEC_LO]]:[[S_RELOAD_SAVEEXEC_HI]]{{\]}}
145
- ; GCN: buffer_load_dword v[[VAL_END:[0-9]+]], off, s[8:11 ], s12 offset:[[VAL_SUB_OFFSET]] ; 4-byte Folded Reload
145
+ ; GCN: buffer_load_dword v[[VAL_END:[0-9]+]], off, s[0:3 ], s7 offset:[[VAL_SUB_OFFSET]] ; 4-byte Folded Reload
146
146
147
147
; GCN: flat_store_dword v{{\[[0-9]+:[0-9]+\]}}, v[[VAL_END]]
148
148
define void @divergent_loop (i32 addrspace (1 )* %out ) #0 {
@@ -180,16 +180,16 @@ end:
180
180
; GCN: s_xor_b64 s{{\[}}[[SAVEEXEC_LO]]:[[SAVEEXEC_HI]]{{\]}}, s{{\[}}[[ANDEXEC_LO]]:[[ANDEXEC_HI]]{{\]}}, s{{\[}}[[SAVEEXEC_LO]]:[[SAVEEXEC_HI]]{{\]}}
181
181
182
182
; Spill load
183
- ; GCN: buffer_store_dword [[LOAD0]], off, s[8:11 ], s12 ; 4-byte Folded Spill
183
+ ; GCN: buffer_store_dword [[LOAD0]], off, s[0:3 ], s7 ; 4-byte Folded Spill
184
184
185
185
; Spill saved exec
186
186
; VGPR: v_writelane_b32 [[SPILL_VGPR:v[0-9]+]], s[[SAVEEXEC_LO]], [[SAVEEXEC_LO_LANE:[0-9]+]]
187
187
; VGPR: v_writelane_b32 [[SPILL_VGPR]], s[[SAVEEXEC_HI]], [[SAVEEXEC_HI_LANE:[0-9]+]]
188
188
189
189
; VMEM: v_mov_b32_e32 v[[V_SAVEEXEC_LO:[0-9]+]], s[[SAVEEXEC_LO]]
190
- ; VMEM: buffer_store_dword v[[V_SAVEEXEC_LO]], off, s[8:11 ], s12 offset:[[SAVEEXEC_LO_OFFSET:[0-9]+]] ; 8-byte Folded Spill
190
+ ; VMEM: buffer_store_dword v[[V_SAVEEXEC_LO]], off, s[0:3 ], s7 offset:[[SAVEEXEC_LO_OFFSET:[0-9]+]] ; 8-byte Folded Spill
191
191
; VMEM: v_mov_b32_e32 v[[V_SAVEEXEC_HI:[0-9]+]], s[[SAVEEXEC_HI]]
192
- ; VMEM: buffer_store_dword v[[V_SAVEEXEC_HI]], off, s[8:11 ], s12 offset:[[SAVEEXEC_HI_OFFSET:[0-9]+]] ; 8-byte Folded Spill
192
+ ; VMEM: buffer_store_dword v[[V_SAVEEXEC_HI]], off, s[0:3 ], s7 offset:[[SAVEEXEC_HI_OFFSET:[0-9]+]] ; 8-byte Folded Spill
193
193
194
194
; GCN: s_mov_b64 exec, [[CMP0]]
195
195
; GCN: s_waitcnt vmcnt(0) expcnt(0)
@@ -204,18 +204,18 @@ end:
204
204
; VGPR: v_readlane_b32 s[[FLOW_S_RELOAD_SAVEEXEC_HI:[0-9]+]], [[SPILL_VGPR]], [[SAVEEXEC_HI_LANE]]
205
205
206
206
207
- ; VMEM: buffer_load_dword v[[FLOW_V_RELOAD_SAVEEXEC_LO:[0-9]+]], off, s[8:11 ], s12 offset:[[SAVEEXEC_LO_OFFSET]]
207
+ ; VMEM: buffer_load_dword v[[FLOW_V_RELOAD_SAVEEXEC_LO:[0-9]+]], off, s[0:3 ], s7 offset:[[SAVEEXEC_LO_OFFSET]]
208
208
; VMEM: s_waitcnt vmcnt(0)
209
209
; VMEM: v_readfirstlane_b32 s[[FLOW_S_RELOAD_SAVEEXEC_LO:[0-9]+]], v[[FLOW_V_RELOAD_SAVEEXEC_LO]]
210
210
211
- ; VMEM: buffer_load_dword v[[FLOW_V_RELOAD_SAVEEXEC_HI:[0-9]+]], off, s[8:11 ], s12 offset:[[SAVEEXEC_HI_OFFSET]] ; 8-byte Folded Reload
211
+ ; VMEM: buffer_load_dword v[[FLOW_V_RELOAD_SAVEEXEC_HI:[0-9]+]], off, s[0:3 ], s7 offset:[[SAVEEXEC_HI_OFFSET]] ; 8-byte Folded Reload
212
212
; VMEM: s_waitcnt vmcnt(0)
213
213
; VMEM: v_readfirstlane_b32 s[[FLOW_S_RELOAD_SAVEEXEC_HI:[0-9]+]], v[[FLOW_V_RELOAD_SAVEEXEC_HI]]
214
214
215
215
; GCN: s_or_saveexec_b64 s{{\[}}[[FLOW_S_RELOAD_SAVEEXEC_LO]]:[[FLOW_S_RELOAD_SAVEEXEC_HI]]{{\]}}, s{{\[}}[[FLOW_S_RELOAD_SAVEEXEC_LO]]:[[FLOW_S_RELOAD_SAVEEXEC_HI]]{{\]}}
216
216
217
217
; Regular spill value restored after exec modification
218
- ; GCN: buffer_load_dword [[FLOW_VAL:v[0-9]+]], off, s[8:11 ], s12 offset:[[FLOW_VAL_OFFSET:[0-9]+]] ; 4-byte Folded Reload
218
+ ; GCN: buffer_load_dword [[FLOW_VAL:v[0-9]+]], off, s[0:3 ], s7 offset:[[FLOW_VAL_OFFSET:[0-9]+]] ; 4-byte Folded Reload
219
219
220
220
221
221
; Spill saved exec
@@ -224,11 +224,11 @@ end:
224
224
225
225
226
226
; VMEM: v_mov_b32_e32 v[[FLOW_V_SAVEEXEC_LO:[0-9]+]], s[[FLOW_S_RELOAD_SAVEEXEC_LO]]
227
- ; VMEM: buffer_store_dword v[[FLOW_V_SAVEEXEC_LO]], off, s[8:11 ], s12 offset:[[FLOW_SAVEEXEC_LO_OFFSET:[0-9]+]] ; 8-byte Folded Spill
227
+ ; VMEM: buffer_store_dword v[[FLOW_V_SAVEEXEC_LO]], off, s[0:3 ], s7 offset:[[FLOW_SAVEEXEC_LO_OFFSET:[0-9]+]] ; 8-byte Folded Spill
228
228
; VMEM: v_mov_b32_e32 v[[FLOW_V_SAVEEXEC_HI:[0-9]+]], s[[FLOW_S_RELOAD_SAVEEXEC_HI]]
229
- ; VMEM: buffer_store_dword v[[FLOW_V_SAVEEXEC_HI]], off, s[8:11 ], s12 offset:[[FLOW_SAVEEXEC_HI_OFFSET:[0-9]+]] ; 8-byte Folded Spill
229
+ ; VMEM: buffer_store_dword v[[FLOW_V_SAVEEXEC_HI]], off, s[0:3 ], s7 offset:[[FLOW_SAVEEXEC_HI_OFFSET:[0-9]+]] ; 8-byte Folded Spill
230
230
231
- ; GCN: buffer_store_dword [[FLOW_VAL]], off, s[8:11 ], s12 offset:[[RESULT_OFFSET:[0-9]+]] ; 4-byte Folded Spill
231
+ ; GCN: buffer_store_dword [[FLOW_VAL]], off, s[0:3 ], s7 offset:[[RESULT_OFFSET:[0-9]+]] ; 4-byte Folded Spill
232
232
; GCN: s_xor_b64 exec, exec, s{{\[}}[[FLOW_S_RELOAD_SAVEEXEC_LO]]:[[FLOW_S_RELOAD_SAVEEXEC_HI]]{{\]}}
233
233
; GCN-NEXT: s_waitcnt vmcnt(0) expcnt(0)
234
234
; GCN-NEXT: ; mask branch [[ENDIF:BB[0-9]+_[0-9]+]]
@@ -237,16 +237,16 @@ end:
237
237
238
238
; GCN: BB{{[0-9]+}}_2: ; %if
239
239
; GCN: ds_read_b32
240
- ; GCN: buffer_load_dword v[[LOAD0_RELOAD:[0-9]+]], off, s[8:11 ], s12 ; 4-byte Folded Reload
240
+ ; GCN: buffer_load_dword v[[LOAD0_RELOAD:[0-9]+]], off, s[0:3 ], s7 ; 4-byte Folded Reload
241
241
; GCN: v_add_i32_e32 [[ADD:v[0-9]+]], vcc, v{{[0-9]+}}, v[[LOAD0_RELOAD]]
242
- ; GCN: buffer_store_dword [[ADD]], off, s[8:11 ], s12 offset:[[RESULT_OFFSET]] ; 4-byte Folded Spill
242
+ ; GCN: buffer_store_dword [[ADD]], off, s[0:3 ], s7 offset:[[RESULT_OFFSET]] ; 4-byte Folded Spill
243
243
; GCN: s_waitcnt vmcnt(0) expcnt(0)
244
244
; GCN-NEXT: s_branch [[ENDIF:BB[0-9]+_[0-9]+]]
245
245
246
246
; GCN: [[ELSE]]: ; %else
247
- ; GCN: buffer_load_dword v[[LOAD0_RELOAD:[0-9]+]], off, s[8:11 ], s12 ; 4-byte Folded Reload
247
+ ; GCN: buffer_load_dword v[[LOAD0_RELOAD:[0-9]+]], off, s[0:3 ], s7 ; 4-byte Folded Reload
248
248
; GCN: v_subrev_i32_e32 [[SUB:v[0-9]+]], vcc, v{{[0-9]+}}, v[[LOAD0_RELOAD]]
249
- ; GCN: buffer_store_dword [[ADD]], off, s[8:11 ], s12 offset:[[FLOW_RESULT_OFFSET:[0-9]+]] ; 4-byte Folded Spill
249
+ ; GCN: buffer_store_dword [[ADD]], off, s[0:3 ], s7 offset:[[FLOW_RESULT_OFFSET:[0-9]+]] ; 4-byte Folded Spill
250
250
; GCN: s_waitcnt vmcnt(0) expcnt(0)
251
251
; GCN-NEXT: s_branch [[FLOW]]
252
252
@@ -255,17 +255,17 @@ end:
255
255
; VGPR: v_readlane_b32 s[[S_RELOAD_SAVEEXEC_HI:[0-9]+]], [[SPILL_VGPR]], [[FLOW_SAVEEXEC_HI_LANE]]
256
256
257
257
258
- ; VMEM: buffer_load_dword v[[V_RELOAD_SAVEEXEC_LO:[0-9]+]], off, s[8:11 ], s12 offset:[[FLOW_SAVEEXEC_LO_OFFSET]] ; 8-byte Folded Reload
258
+ ; VMEM: buffer_load_dword v[[V_RELOAD_SAVEEXEC_LO:[0-9]+]], off, s[0:3 ], s7 offset:[[FLOW_SAVEEXEC_LO_OFFSET]] ; 8-byte Folded Reload
259
259
; VMEM: s_waitcnt vmcnt(0)
260
260
; VMEM: v_readfirstlane_b32 s[[S_RELOAD_SAVEEXEC_LO:[0-9]+]], v[[V_RELOAD_SAVEEXEC_LO]]
261
261
262
- ; VMEM: buffer_load_dword v[[V_RELOAD_SAVEEXEC_HI:[0-9]+]], off, s[8:11 ], s12 offset:[[FLOW_SAVEEXEC_HI_OFFSET]] ; 8-byte Folded Reload
262
+ ; VMEM: buffer_load_dword v[[V_RELOAD_SAVEEXEC_HI:[0-9]+]], off, s[0:3 ], s7 offset:[[FLOW_SAVEEXEC_HI_OFFSET]] ; 8-byte Folded Reload
263
263
; VMEM: s_waitcnt vmcnt(0)
264
264
; VMEM: v_readfirstlane_b32 s[[S_RELOAD_SAVEEXEC_HI:[0-9]+]], v[[V_RELOAD_SAVEEXEC_HI]]
265
265
266
266
; GCN: s_or_b64 exec, exec, s{{\[}}[[S_RELOAD_SAVEEXEC_LO]]:[[S_RELOAD_SAVEEXEC_HI]]{{\]}}
267
267
268
- ; GCN: buffer_load_dword v[[RESULT:[0-9]+]], off, s[8:11 ], s12 offset:[[RESULT_OFFSET]] ; 4-byte Folded Reload
268
+ ; GCN: buffer_load_dword v[[RESULT:[0-9]+]], off, s[0:3 ], s7 offset:[[RESULT_OFFSET]] ; 4-byte Folded Reload
269
269
; GCN: flat_store_dword v{{\[[0-9]+:[0-9]+\]}}, v[[RESULT]]
270
270
define void @divergent_if_else_endif (i32 addrspace (1 )* %out ) #0 {
271
271
entry:
0 commit comments