@@ -9,21 +9,22 @@ body: |
9
9
liveins: $x0, $x1, $x2, $d0, $d1, $d2, $d3, $d4, $d5, $d6, $d7, $x0
10
10
; CHECK-LABEL: name: combine_unmerge_from_unmerge_of_concat_tree
11
11
; CHECK: liveins: $x0, $x1, $x2, $d0, $d1, $d2, $d3, $d4, $d5, $d6, $d7, $x0
12
- ; CHECK: [[COPY:%[0-9]+]]:_(s64) = COPY $x0
13
- ; CHECK: [[COPY1:%[0-9]+]]:_(s64) = COPY $x1
14
- ; CHECK: [[COPY2:%[0-9]+]]:_(p0) = COPY $x2
15
- ; CHECK: [[COPY3:%[0-9]+]]:_(s64) = COPY $d0
16
- ; CHECK: [[COPY4:%[0-9]+]]:_(s64) = COPY $d1
17
- ; CHECK: [[COPY5:%[0-9]+]]:_(s64) = COPY $d2
18
- ; CHECK: [[COPY6:%[0-9]+]]:_(s64) = COPY $d3
19
- ; CHECK: [[COPY7:%[0-9]+]]:_(s64) = COPY $d4
20
- ; CHECK: [[COPY8:%[0-9]+]]:_(s64) = COPY $d5
21
- ; CHECK: %v2s64_val:_(<2 x s64>) = G_BUILD_VECTOR [[COPY5]](s64), [[COPY6]](s64)
22
- ; CHECK: %v2s64_val2:_(<2 x s64>) = G_BUILD_VECTOR [[COPY6]](s64), [[COPY8]](s64)
23
- ; CHECK: G_STORE %v2s64_val(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
24
- ; CHECK: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
25
- ; CHECK: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
26
- ; CHECK: RET_ReallyLR
12
+ ; CHECK-NEXT: {{ $}}
13
+ ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(s64) = COPY $x0
14
+ ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(s64) = COPY $x1
15
+ ; CHECK-NEXT: [[COPY2:%[0-9]+]]:_(p0) = COPY $x2
16
+ ; CHECK-NEXT: [[COPY3:%[0-9]+]]:_(s64) = COPY $d0
17
+ ; CHECK-NEXT: [[COPY4:%[0-9]+]]:_(s64) = COPY $d1
18
+ ; CHECK-NEXT: [[COPY5:%[0-9]+]]:_(s64) = COPY $d2
19
+ ; CHECK-NEXT: [[COPY6:%[0-9]+]]:_(s64) = COPY $d3
20
+ ; CHECK-NEXT: [[COPY7:%[0-9]+]]:_(s64) = COPY $d4
21
+ ; CHECK-NEXT: [[COPY8:%[0-9]+]]:_(s64) = COPY $d5
22
+ ; CHECK-NEXT: %v2s64_val:_(<2 x s64>) = G_BUILD_VECTOR [[COPY5]](s64), [[COPY6]](s64)
23
+ ; CHECK-NEXT: %v2s64_val2:_(<2 x s64>) = G_BUILD_VECTOR [[COPY6]](s64), [[COPY8]](s64)
24
+ ; CHECK-NEXT: G_STORE %v2s64_val(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
25
+ ; CHECK-NEXT: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
26
+ ; CHECK-NEXT: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
27
+ ; CHECK-NEXT: RET_ReallyLR
27
28
%0:_(s64) = COPY $x0
28
29
%1:_(s64) = COPY $x1
29
30
%2:_(p0) = COPY $x2
@@ -60,21 +61,22 @@ body: |
60
61
liveins: $x0, $x1, $x2, $d0, $d1, $d2, $d3, $d4, $d5, $d6, $d7, $x0
61
62
; CHECK-LABEL: name: combine_unmerge_from_unmerge_of_concat_tree_high_bits
62
63
; CHECK: liveins: $x0, $x1, $x2, $d0, $d1, $d2, $d3, $d4, $d5, $d6, $d7, $x0
63
- ; CHECK: [[COPY:%[0-9]+]]:_(s64) = COPY $x0
64
- ; CHECK: [[COPY1:%[0-9]+]]:_(s64) = COPY $x1
65
- ; CHECK: [[COPY2:%[0-9]+]]:_(p0) = COPY $x2
66
- ; CHECK: [[COPY3:%[0-9]+]]:_(s64) = COPY $d0
67
- ; CHECK: [[COPY4:%[0-9]+]]:_(s64) = COPY $d1
68
- ; CHECK: [[COPY5:%[0-9]+]]:_(s64) = COPY $d2
69
- ; CHECK: [[COPY6:%[0-9]+]]:_(s64) = COPY $d3
70
- ; CHECK: [[COPY7:%[0-9]+]]:_(s64) = COPY $d4
71
- ; CHECK: [[COPY8:%[0-9]+]]:_(s64) = COPY $d5
72
- ; CHECK: %v2s64_val:_(<2 x s64>) = G_BUILD_VECTOR [[COPY5]](s64), [[COPY6]](s64)
73
- ; CHECK: %v2s64_val2:_(<2 x s64>) = G_BUILD_VECTOR [[COPY6]](s64), [[COPY8]](s64)
74
- ; CHECK: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
75
- ; CHECK: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
76
- ; CHECK: G_STORE %v2s64_val(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
77
- ; CHECK: RET_ReallyLR
64
+ ; CHECK-NEXT: {{ $}}
65
+ ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(s64) = COPY $x0
66
+ ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(s64) = COPY $x1
67
+ ; CHECK-NEXT: [[COPY2:%[0-9]+]]:_(p0) = COPY $x2
68
+ ; CHECK-NEXT: [[COPY3:%[0-9]+]]:_(s64) = COPY $d0
69
+ ; CHECK-NEXT: [[COPY4:%[0-9]+]]:_(s64) = COPY $d1
70
+ ; CHECK-NEXT: [[COPY5:%[0-9]+]]:_(s64) = COPY $d2
71
+ ; CHECK-NEXT: [[COPY6:%[0-9]+]]:_(s64) = COPY $d3
72
+ ; CHECK-NEXT: [[COPY7:%[0-9]+]]:_(s64) = COPY $d4
73
+ ; CHECK-NEXT: [[COPY8:%[0-9]+]]:_(s64) = COPY $d5
74
+ ; CHECK-NEXT: %v2s64_val:_(<2 x s64>) = G_BUILD_VECTOR [[COPY5]](s64), [[COPY6]](s64)
75
+ ; CHECK-NEXT: %v2s64_val2:_(<2 x s64>) = G_BUILD_VECTOR [[COPY6]](s64), [[COPY8]](s64)
76
+ ; CHECK-NEXT: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
77
+ ; CHECK-NEXT: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
78
+ ; CHECK-NEXT: G_STORE %v2s64_val(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
79
+ ; CHECK-NEXT: RET_ReallyLR
78
80
%0:_(s64) = COPY $x0
79
81
%1:_(s64) = COPY $x1
80
82
%2:_(p0) = COPY $x2
@@ -110,20 +112,21 @@ body: |
110
112
liveins: $x0, $x1, $x2, $d0, $d1, $d2, $d3, $d4, $d5, $d6, $d7, $x0
111
113
; CHECK-LABEL: name: combine_unmerge_from_insert_into_low
112
114
; CHECK: liveins: $x0, $x1, $x2, $d0, $d1, $d2, $d3, $d4, $d5, $d6, $d7, $x0
113
- ; CHECK: [[COPY:%[0-9]+]]:_(s64) = COPY $x0
114
- ; CHECK: [[COPY1:%[0-9]+]]:_(s64) = COPY $x1
115
- ; CHECK: [[COPY2:%[0-9]+]]:_(p0) = COPY $x2
116
- ; CHECK: [[COPY3:%[0-9]+]]:_(s64) = COPY $d0
117
- ; CHECK: [[COPY4:%[0-9]+]]:_(s64) = COPY $d1
118
- ; CHECK: [[COPY5:%[0-9]+]]:_(s64) = COPY $d2
119
- ; CHECK: [[COPY6:%[0-9]+]]:_(s64) = COPY $d3
120
- ; CHECK: [[COPY7:%[0-9]+]]:_(s64) = COPY $d4
121
- ; CHECK: [[COPY8:%[0-9]+]]:_(s64) = COPY $d5
122
- ; CHECK: %v2s64_val:_(<2 x s64>) = G_BUILD_VECTOR [[COPY5]](s64), [[COPY6]](s64)
123
- ; CHECK: %v2s64_val2:_(<2 x s64>) = G_BUILD_VECTOR [[COPY6]](s64), [[COPY8]](s64)
124
- ; CHECK: G_STORE %v2s64_val(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
125
- ; CHECK: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
126
- ; CHECK: RET_ReallyLR
115
+ ; CHECK-NEXT: {{ $}}
116
+ ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(s64) = COPY $x0
117
+ ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(s64) = COPY $x1
118
+ ; CHECK-NEXT: [[COPY2:%[0-9]+]]:_(p0) = COPY $x2
119
+ ; CHECK-NEXT: [[COPY3:%[0-9]+]]:_(s64) = COPY $d0
120
+ ; CHECK-NEXT: [[COPY4:%[0-9]+]]:_(s64) = COPY $d1
121
+ ; CHECK-NEXT: [[COPY5:%[0-9]+]]:_(s64) = COPY $d2
122
+ ; CHECK-NEXT: [[COPY6:%[0-9]+]]:_(s64) = COPY $d3
123
+ ; CHECK-NEXT: [[COPY7:%[0-9]+]]:_(s64) = COPY $d4
124
+ ; CHECK-NEXT: [[COPY8:%[0-9]+]]:_(s64) = COPY $d5
125
+ ; CHECK-NEXT: %v2s64_val:_(<2 x s64>) = G_BUILD_VECTOR [[COPY5]](s64), [[COPY6]](s64)
126
+ ; CHECK-NEXT: %v2s64_val2:_(<2 x s64>) = G_BUILD_VECTOR [[COPY6]](s64), [[COPY8]](s64)
127
+ ; CHECK-NEXT: G_STORE %v2s64_val(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
128
+ ; CHECK-NEXT: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
129
+ ; CHECK-NEXT: RET_ReallyLR
127
130
%0:_(s64) = COPY $x0
128
131
%1:_(s64) = COPY $x1
129
132
%2:_(p0) = COPY $x2
@@ -156,20 +159,21 @@ body: |
156
159
liveins: $x0, $x1, $x2, $d0, $d1, $d2, $d3, $d4, $d5, $d6, $d7, $x0
157
160
; CHECK-LABEL: name: combine_unmerge_from_insert_into_high
158
161
; CHECK: liveins: $x0, $x1, $x2, $d0, $d1, $d2, $d3, $d4, $d5, $d6, $d7, $x0
159
- ; CHECK: [[COPY:%[0-9]+]]:_(s64) = COPY $x0
160
- ; CHECK: [[COPY1:%[0-9]+]]:_(s64) = COPY $x1
161
- ; CHECK: [[COPY2:%[0-9]+]]:_(p0) = COPY $x2
162
- ; CHECK: [[COPY3:%[0-9]+]]:_(s64) = COPY $d0
163
- ; CHECK: [[COPY4:%[0-9]+]]:_(s64) = COPY $d1
164
- ; CHECK: [[COPY5:%[0-9]+]]:_(s64) = COPY $d2
165
- ; CHECK: [[COPY6:%[0-9]+]]:_(s64) = COPY $d3
166
- ; CHECK: [[COPY7:%[0-9]+]]:_(s64) = COPY $d4
167
- ; CHECK: [[COPY8:%[0-9]+]]:_(s64) = COPY $d5
168
- ; CHECK: %v2s64_val:_(<2 x s64>) = G_BUILD_VECTOR [[COPY5]](s64), [[COPY6]](s64)
169
- ; CHECK: %v2s64_val2:_(<2 x s64>) = G_BUILD_VECTOR [[COPY6]](s64), [[COPY8]](s64)
170
- ; CHECK: G_STORE %v2s64_val(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
171
- ; CHECK: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
172
- ; CHECK: RET_ReallyLR
162
+ ; CHECK-NEXT: {{ $}}
163
+ ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(s64) = COPY $x0
164
+ ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(s64) = COPY $x1
165
+ ; CHECK-NEXT: [[COPY2:%[0-9]+]]:_(p0) = COPY $x2
166
+ ; CHECK-NEXT: [[COPY3:%[0-9]+]]:_(s64) = COPY $d0
167
+ ; CHECK-NEXT: [[COPY4:%[0-9]+]]:_(s64) = COPY $d1
168
+ ; CHECK-NEXT: [[COPY5:%[0-9]+]]:_(s64) = COPY $d2
169
+ ; CHECK-NEXT: [[COPY6:%[0-9]+]]:_(s64) = COPY $d3
170
+ ; CHECK-NEXT: [[COPY7:%[0-9]+]]:_(s64) = COPY $d4
171
+ ; CHECK-NEXT: [[COPY8:%[0-9]+]]:_(s64) = COPY $d5
172
+ ; CHECK-NEXT: %v2s64_val:_(<2 x s64>) = G_BUILD_VECTOR [[COPY5]](s64), [[COPY6]](s64)
173
+ ; CHECK-NEXT: %v2s64_val2:_(<2 x s64>) = G_BUILD_VECTOR [[COPY6]](s64), [[COPY8]](s64)
174
+ ; CHECK-NEXT: G_STORE %v2s64_val(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
175
+ ; CHECK-NEXT: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
176
+ ; CHECK-NEXT: RET_ReallyLR
173
177
%0:_(s64) = COPY $x0
174
178
%1:_(s64) = COPY $x1
175
179
%2:_(p0) = COPY $x2
@@ -202,20 +206,21 @@ body: |
202
206
liveins: $x0, $x1, $x2, $d0, $d1, $d2, $d3, $d4, $d5, $d6, $d7, $x0
203
207
; CHECK-LABEL: name: combine_unmerge_from_insert_look_into_container
204
208
; CHECK: liveins: $x0, $x1, $x2, $d0, $d1, $d2, $d3, $d4, $d5, $d6, $d7, $x0
205
- ; CHECK: [[COPY:%[0-9]+]]:_(s64) = COPY $x0
206
- ; CHECK: [[COPY1:%[0-9]+]]:_(s64) = COPY $x1
207
- ; CHECK: [[COPY2:%[0-9]+]]:_(p0) = COPY $x2
208
- ; CHECK: [[COPY3:%[0-9]+]]:_(s64) = COPY $d0
209
- ; CHECK: [[COPY4:%[0-9]+]]:_(s64) = COPY $d1
210
- ; CHECK: [[COPY5:%[0-9]+]]:_(s64) = COPY $d2
211
- ; CHECK: [[COPY6:%[0-9]+]]:_(s64) = COPY $d3
212
- ; CHECK: [[COPY7:%[0-9]+]]:_(s64) = COPY $d4
213
- ; CHECK: [[COPY8:%[0-9]+]]:_(s64) = COPY $d5
214
- ; CHECK: %v2s64_val:_(<2 x s64>) = G_BUILD_VECTOR [[COPY5]](s64), [[COPY6]](s64)
215
- ; CHECK: %v2s64_val2:_(<2 x s64>) = G_BUILD_VECTOR [[COPY6]](s64), [[COPY8]](s64)
216
- ; CHECK: G_STORE %v2s64_val(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
217
- ; CHECK: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
218
- ; CHECK: RET_ReallyLR
209
+ ; CHECK-NEXT: {{ $}}
210
+ ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(s64) = COPY $x0
211
+ ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(s64) = COPY $x1
212
+ ; CHECK-NEXT: [[COPY2:%[0-9]+]]:_(p0) = COPY $x2
213
+ ; CHECK-NEXT: [[COPY3:%[0-9]+]]:_(s64) = COPY $d0
214
+ ; CHECK-NEXT: [[COPY4:%[0-9]+]]:_(s64) = COPY $d1
215
+ ; CHECK-NEXT: [[COPY5:%[0-9]+]]:_(s64) = COPY $d2
216
+ ; CHECK-NEXT: [[COPY6:%[0-9]+]]:_(s64) = COPY $d3
217
+ ; CHECK-NEXT: [[COPY7:%[0-9]+]]:_(s64) = COPY $d4
218
+ ; CHECK-NEXT: [[COPY8:%[0-9]+]]:_(s64) = COPY $d5
219
+ ; CHECK-NEXT: %v2s64_val:_(<2 x s64>) = G_BUILD_VECTOR [[COPY5]](s64), [[COPY6]](s64)
220
+ ; CHECK-NEXT: %v2s64_val2:_(<2 x s64>) = G_BUILD_VECTOR [[COPY6]](s64), [[COPY8]](s64)
221
+ ; CHECK-NEXT: G_STORE %v2s64_val(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
222
+ ; CHECK-NEXT: G_STORE %v2s64_val2(<2 x s64>), [[COPY2]](p0) :: (store (<2 x s64>))
223
+ ; CHECK-NEXT: RET_ReallyLR
219
224
%0:_(s64) = COPY $x0
220
225
%1:_(s64) = COPY $x1
221
226
%2:_(p0) = COPY $x2
0 commit comments