@@ -37,10 +37,10 @@ define void @test_dmxvi8gerx4(ptr %vpp, ptr %vcp, ptr %resp) {
37
37
; CHECK-BE-NEXT: stxvp vsp34, 0(r5)
38
38
; CHECK-BE-NEXT: blr
39
39
entry:
40
- %0 = load <256 x i1 >, ptr %vpp , align 32
41
- %1 = load <16 x i8 >, ptr %vcp , align 32
42
- %2 = tail call <1024 x i1 > @llvm.ppc.mma.dmxvi8gerx4 (<256 x i1 > %0 , <16 x i8 > %1 )
43
- store <1024 x i1 > %2 , ptr %resp , align 64
40
+ %v1 = load <256 x i1 >, ptr %vpp , align 32
41
+ %v2 = load <16 x i8 >, ptr %vcp , align 32
42
+ %call = tail call <1024 x i1 > @llvm.ppc.mma.dmxvi8gerx4 (<256 x i1 > %v1 , <16 x i8 > %v2 )
43
+ store <1024 x i1 > %call , ptr %resp , align 64
44
44
ret void
45
45
}
46
46
@@ -87,11 +87,11 @@ define void @test_dmxvi8gerx4pp(ptr %vop, ptr %vpp, ptr %vcp, ptr %resp) {
87
87
; CHECK-BE-NEXT: stxvp vsp34, 0(r6)
88
88
; CHECK-BE-NEXT: blr
89
89
entry:
90
- %0 = load <1024 x i1 >, ptr %vop , align 64
91
- %1 = load <256 x i1 >, ptr %vpp , align 32
92
- %2 = load <16 x i8 >, ptr %vcp , align 32
93
- %3 = tail call <1024 x i1 > @llvm.ppc.mma.dmxvi8gerx4pp (<1024 x i1 > %0 , <256 x i1 > %1 , <16 x i8 > %2 )
94
- store <1024 x i1 > %3 , ptr %resp , align 64
90
+ %v.dmr = load <1024 x i1 >, ptr %vop , align 64
91
+ %v1 = load <256 x i1 >, ptr %vpp , align 32
92
+ %v2 = load <16 x i8 >, ptr %vcp , align 32
93
+ %call = tail call <1024 x i1 > @llvm.ppc.mma.dmxvi8gerx4pp (<1024 x i1 > %v.dmr , <256 x i1 > %v1 , <16 x i8 > %v2 )
94
+ store <1024 x i1 > %call , ptr %resp , align 64
95
95
ret void
96
96
}
97
97
@@ -138,11 +138,11 @@ define void @test_dmxvi8gerx4spp(ptr %vop, ptr %vpp, ptr %vcp, ptr %resp) {
138
138
; CHECK-BE-NEXT: stxvp vsp34, 0(r6)
139
139
; CHECK-BE-NEXT: blr
140
140
entry:
141
- %0 = load <1024 x i1 >, ptr %vop , align 64
142
- %1 = load <256 x i1 >, ptr %vpp , align 32
143
- %2 = load <16 x i8 >, ptr %vcp , align 32
144
- %3 = tail call <1024 x i1 > @llvm.ppc.mma.dmxvi8gerx4spp (<1024 x i1 > %0 , <256 x i1 > %1 , <16 x i8 > %2 )
145
- store <1024 x i1 > %3 , ptr %resp , align 64
141
+ %v.dmr = load <1024 x i1 >, ptr %vop , align 64
142
+ %v1 = load <256 x i1 >, ptr %vpp , align 32
143
+ %v2 = load <16 x i8 >, ptr %vcp , align 32
144
+ %call = tail call <1024 x i1 > @llvm.ppc.mma.dmxvi8gerx4spp (<1024 x i1 > %v.dmr , <256 x i1 > %v1 , <16 x i8 > %v2 )
145
+ store <1024 x i1 > %call , ptr %resp , align 64
146
146
ret void
147
147
}
148
148
@@ -160,7 +160,7 @@ define void @test_pmdmxvi8gerx4pp(ptr %vop, ptr %vpp, ptr %vcp, ptr %resp) {
160
160
; CHECK-NEXT: lxv v3, 0(r4)
161
161
; CHECK-NEXT: lxv vs0, 0(r5)
162
162
; CHECK-NEXT: lxv v2, 16(r4)
163
- ; CHECK-NEXT: pmdmxvi8gerx4pp dmr0, vsp34, vs0, 0, 0, 0
163
+ ; CHECK-NEXT: pmdmxvi8gerx4pp dmr0, vsp34, vs0, 42, 7, 9
164
164
; CHECK-NEXT: dmxxextfdmr512 wacc0, vsp34, vsp36, 0
165
165
; CHECK-NEXT: stxvp vsp34, 96(r6)
166
166
; CHECK-NEXT: stxvp vsp36, 64(r6)
@@ -180,7 +180,7 @@ define void @test_pmdmxvi8gerx4pp(ptr %vop, ptr %vpp, ptr %vcp, ptr %resp) {
180
180
; CHECK-BE-NEXT: lxv v3, 16(r4)
181
181
; CHECK-BE-NEXT: lxv vs0, 0(r5)
182
182
; CHECK-BE-NEXT: lxv v2, 0(r4)
183
- ; CHECK-BE-NEXT: pmdmxvi8gerx4pp dmr0, vsp34, vs0, 0, 0, 0
183
+ ; CHECK-BE-NEXT: pmdmxvi8gerx4pp dmr0, vsp34, vs0, 42, 7, 9
184
184
; CHECK-BE-NEXT: dmxxextfdmr512 wacc_hi0, vsp34, vsp36, 1
185
185
; CHECK-BE-NEXT: stxvp vsp36, 96(r6)
186
186
; CHECK-BE-NEXT: stxvp vsp34, 64(r6)
@@ -189,11 +189,11 @@ define void @test_pmdmxvi8gerx4pp(ptr %vop, ptr %vpp, ptr %vcp, ptr %resp) {
189
189
; CHECK-BE-NEXT: stxvp vsp34, 0(r6)
190
190
; CHECK-BE-NEXT: blr
191
191
entry:
192
- %0 = load <1024 x i1 >, ptr %vop , align 64
193
- %1 = load <256 x i1 >, ptr %vpp , align 32
194
- %2 = load <16 x i8 >, ptr %vcp , align 32
195
- %3 = tail call <1024 x i1 > @llvm.ppc.mma.pmdmxvi8gerx4pp (<1024 x i1 > %0 , <256 x i1 > %1 , <16 x i8 > %2 , i32 0 , i32 0 , i32 0 )
196
- store <1024 x i1 > %3 , ptr %resp , align 64
192
+ %v.dmr = load <1024 x i1 >, ptr %vop , align 64
193
+ %v1 = load <256 x i1 >, ptr %vpp , align 32
194
+ %v2 = load <16 x i8 >, ptr %vcp , align 32
195
+ %call = tail call <1024 x i1 > @llvm.ppc.mma.pmdmxvi8gerx4pp (<1024 x i1 > %v.dmr , <256 x i1 > %v1 , <16 x i8 > %v2 , i32 42 , i32 7 , i32 9 )
196
+ store <1024 x i1 > %call , ptr %resp , align 64
197
197
ret void
198
198
}
199
199
@@ -205,7 +205,7 @@ define void @test_pmdmxvi8gerx4(ptr %vpp, ptr %vcp, ptr %resp) {
205
205
; CHECK-NEXT: lxv v3, 0(r3)
206
206
; CHECK-NEXT: lxv vs0, 0(r4)
207
207
; CHECK-NEXT: lxv v2, 16(r3)
208
- ; CHECK-NEXT: pmdmxvi8gerx4 dmr0, vsp34, vs0, 0, 0, 0
208
+ ; CHECK-NEXT: pmdmxvi8gerx4 dmr0, vsp34, vs0, 55, 5, 10
209
209
; CHECK-NEXT: dmxxextfdmr512 wacc0, vsp34, vsp36, 0
210
210
; CHECK-NEXT: stxvp vsp34, 96(r5)
211
211
; CHECK-NEXT: stxvp vsp36, 64(r5)
@@ -219,7 +219,7 @@ define void @test_pmdmxvi8gerx4(ptr %vpp, ptr %vcp, ptr %resp) {
219
219
; CHECK-BE-NEXT: lxv v3, 16(r3)
220
220
; CHECK-BE-NEXT: lxv vs0, 0(r4)
221
221
; CHECK-BE-NEXT: lxv v2, 0(r3)
222
- ; CHECK-BE-NEXT: pmdmxvi8gerx4 dmr0, vsp34, vs0, 0, 0, 0
222
+ ; CHECK-BE-NEXT: pmdmxvi8gerx4 dmr0, vsp34, vs0, 55, 5, 10
223
223
; CHECK-BE-NEXT: dmxxextfdmr512 wacc_hi0, vsp34, vsp36, 1
224
224
; CHECK-BE-NEXT: stxvp vsp36, 96(r5)
225
225
; CHECK-BE-NEXT: stxvp vsp34, 64(r5)
@@ -228,10 +228,10 @@ define void @test_pmdmxvi8gerx4(ptr %vpp, ptr %vcp, ptr %resp) {
228
228
; CHECK-BE-NEXT: stxvp vsp34, 0(r5)
229
229
; CHECK-BE-NEXT: blr
230
230
entry:
231
- %0 = load <256 x i1 >, ptr %vpp , align 32
232
- %1 = load <16 x i8 >, ptr %vcp , align 32
233
- %2 = tail call <1024 x i1 > @llvm.ppc.mma.pmdmxvi8gerx4 (<256 x i1 > %0 , <16 x i8 > %1 , i32 0 , i32 0 , i32 0 )
234
- store <1024 x i1 > %2 , ptr %resp , align 64
231
+ %v1 = load <256 x i1 >, ptr %vpp , align 32
232
+ %v2 = load <16 x i8 >, ptr %vcp , align 32
233
+ %call = tail call <1024 x i1 > @llvm.ppc.mma.pmdmxvi8gerx4 (<256 x i1 > %v1 , <16 x i8 > %v2 , i32 55 , i32 5 , i32 10 )
234
+ store <1024 x i1 > %call , ptr %resp , align 64
235
235
ret void
236
236
}
237
237
@@ -249,7 +249,7 @@ define dso_local void @test_pmdmxvi8gerx4spp(ptr %vop, ptr %vpp, ptr %vcp, ptr %
249
249
; CHECK-NEXT: lxv v3, 0(r4)
250
250
; CHECK-NEXT: lxv vs0, 0(r5)
251
251
; CHECK-NEXT: lxv v2, 16(r4)
252
- ; CHECK-NEXT: pmdmxvi8gerx4spp dmr0, vsp34, vs0, 0, 0, 0
252
+ ; CHECK-NEXT: pmdmxvi8gerx4spp dmr0, vsp34, vs0, 100, 6, 12
253
253
; CHECK-NEXT: dmxxextfdmr512 wacc0, vsp34, vsp36, 0
254
254
; CHECK-NEXT: stxvp vsp34, 96(r6)
255
255
; CHECK-NEXT: stxvp vsp36, 64(r6)
@@ -269,7 +269,7 @@ define dso_local void @test_pmdmxvi8gerx4spp(ptr %vop, ptr %vpp, ptr %vcp, ptr %
269
269
; CHECK-BE-NEXT: lxv v3, 16(r4)
270
270
; CHECK-BE-NEXT: lxv vs0, 0(r5)
271
271
; CHECK-BE-NEXT: lxv v2, 0(r4)
272
- ; CHECK-BE-NEXT: pmdmxvi8gerx4spp dmr0, vsp34, vs0, 0, 0, 0
272
+ ; CHECK-BE-NEXT: pmdmxvi8gerx4spp dmr0, vsp34, vs0, 100, 6, 12
273
273
; CHECK-BE-NEXT: dmxxextfdmr512 wacc_hi0, vsp34, vsp36, 1
274
274
; CHECK-BE-NEXT: stxvp vsp36, 96(r6)
275
275
; CHECK-BE-NEXT: stxvp vsp34, 64(r6)
@@ -278,10 +278,10 @@ define dso_local void @test_pmdmxvi8gerx4spp(ptr %vop, ptr %vpp, ptr %vcp, ptr %
278
278
; CHECK-BE-NEXT: stxvp vsp34, 0(r6)
279
279
; CHECK-BE-NEXT: blr
280
280
entry:
281
- %0 = load <1024 x i1 >, ptr %vop , align 64
282
- %1 = load <256 x i1 >, ptr %vpp , align 32
283
- %2 = load <16 x i8 >, ptr %vcp , align 32
284
- %3 = tail call <1024 x i1 > @llvm.ppc.mma.pmdmxvi8gerx4spp (<1024 x i1 > %0 , <256 x i1 > %1 , <16 x i8 > %2 , i32 0 , i32 0 , i32 0 )
285
- store <1024 x i1 > %3 , ptr %resp , align 64
281
+ %v.dmr = load <1024 x i1 >, ptr %vop , align 64
282
+ %v1 = load <256 x i1 >, ptr %vpp , align 32
283
+ %v2 = load <16 x i8 >, ptr %vcp , align 32
284
+ %call = tail call <1024 x i1 > @llvm.ppc.mma.pmdmxvi8gerx4spp (<1024 x i1 > %v.dmr , <256 x i1 > %v1 , <16 x i8 > %v2 , i32 100 , i32 6 , i32 12 )
285
+ store <1024 x i1 > %call , ptr %resp , align 64
286
286
ret void
287
287
}
0 commit comments