@@ -113,7 +113,6 @@ define <2 x i32> @ustest_f64i32(<2 x double> %x) {
113
113
; CHECK-V: # %bb.0: # %entry
114
114
; CHECK-V-NEXT: vsetivli zero, 2, e64, m1, ta, ma
115
115
; CHECK-V-NEXT: vfcvt.rtz.x.f.v v8, v8
116
- ; CHECK-V-NEXT: vmax.vx v8, v8, zero
117
116
; CHECK-V-NEXT: vsetvli zero, zero, e32, mf2, ta, ma
118
117
; CHECK-V-NEXT: vnclipu.wi v8, v8, 0
119
118
; CHECK-V-NEXT: ret
@@ -304,9 +303,6 @@ define <4 x i32> @ustest_f32i32(<4 x float> %x) {
304
303
; CHECK-V: # %bb.0: # %entry
305
304
; CHECK-V-NEXT: vsetivli zero, 4, e32, m1, ta, ma
306
305
; CHECK-V-NEXT: vfwcvt.rtz.x.f.v v10, v8
307
- ; CHECK-V-NEXT: vsetvli zero, zero, e64, m2, ta, ma
308
- ; CHECK-V-NEXT: vmax.vx v10, v10, zero
309
- ; CHECK-V-NEXT: vsetvli zero, zero, e32, m1, ta, ma
310
306
; CHECK-V-NEXT: vnclipu.wi v8, v10, 0
311
307
; CHECK-V-NEXT: ret
312
308
entry:
@@ -801,17 +797,16 @@ define <4 x i32> @ustest_f16i32(<4 x half> %x) {
801
797
; CHECK-V-NEXT: call __extendhfsf2
802
798
; CHECK-V-NEXT: fcvt.l.s a0, fa0, rtz
803
799
; CHECK-V-NEXT: vsetivli zero, 2, e64, m1, ta, ma
804
- ; CHECK-V-NEXT: vmv.s.x v8 , a0
800
+ ; CHECK-V-NEXT: vmv.s.x v10 , a0
805
801
; CHECK-V-NEXT: addi a0, sp, 16
806
- ; CHECK-V-NEXT: vl1r.v v9 , (a0) # Unknown-size Folded Reload
807
- ; CHECK-V-NEXT: vslideup.vi v8, v9 , 1
802
+ ; CHECK-V-NEXT: vl1r.v v8 , (a0) # Unknown-size Folded Reload
803
+ ; CHECK-V-NEXT: vslideup.vi v10, v8 , 1
808
804
; CHECK-V-NEXT: csrr a0, vlenb
809
805
; CHECK-V-NEXT: add a0, sp, a0
810
806
; CHECK-V-NEXT: addi a0, a0, 16
811
- ; CHECK-V-NEXT: vl2r.v v10 , (a0) # Unknown-size Folded Reload
807
+ ; CHECK-V-NEXT: vl2r.v v8 , (a0) # Unknown-size Folded Reload
812
808
; CHECK-V-NEXT: vsetivli zero, 4, e64, m2, ta, ma
813
- ; CHECK-V-NEXT: vslideup.vi v8, v10, 2
814
- ; CHECK-V-NEXT: vmax.vx v10, v8, zero
809
+ ; CHECK-V-NEXT: vslideup.vi v10, v8, 2
815
810
; CHECK-V-NEXT: vsetvli zero, zero, e32, m1, ta, ma
816
811
; CHECK-V-NEXT: vnclipu.wi v8, v10, 0
817
812
; CHECK-V-NEXT: csrr a0, vlenb
@@ -944,9 +939,8 @@ define <2 x i16> @ustest_f64i16(<2 x double> %x) {
944
939
; CHECK-V: # %bb.0: # %entry
945
940
; CHECK-V-NEXT: vsetivli zero, 2, e32, mf2, ta, ma
946
941
; CHECK-V-NEXT: vfncvt.rtz.x.f.w v9, v8
947
- ; CHECK-V-NEXT: vmax.vx v8, v9, zero
948
942
; CHECK-V-NEXT: vsetvli zero, zero, e16, mf4, ta, ma
949
- ; CHECK-V-NEXT: vnclipu.wi v8, v8 , 0
943
+ ; CHECK-V-NEXT: vnclipu.wi v8, v9 , 0
950
944
; CHECK-V-NEXT: ret
951
945
entry:
952
946
%conv = fptosi <2 x double > %x to <2 x i32 >
@@ -1139,7 +1133,6 @@ define <4 x i16> @ustest_f32i16(<4 x float> %x) {
1139
1133
; CHECK-V: # %bb.0: # %entry
1140
1134
; CHECK-V-NEXT: vsetivli zero, 4, e32, m1, ta, ma
1141
1135
; CHECK-V-NEXT: vfcvt.rtz.x.f.v v8, v8
1142
- ; CHECK-V-NEXT: vmax.vx v8, v8, zero
1143
1136
; CHECK-V-NEXT: vsetvli zero, zero, e16, mf2, ta, ma
1144
1137
; CHECK-V-NEXT: vnclipu.wi v8, v8, 0
1145
1138
; CHECK-V-NEXT: ret
@@ -2114,24 +2107,23 @@ define <8 x i16> @ustest_f16i16(<8 x half> %x) {
2114
2107
; CHECK-V-NEXT: call __extendhfsf2
2115
2108
; CHECK-V-NEXT: fcvt.l.s a0, fa0, rtz
2116
2109
; CHECK-V-NEXT: vsetivli zero, 2, e32, mf2, ta, ma
2117
- ; CHECK-V-NEXT: vmv.s.x v8 , a0
2110
+ ; CHECK-V-NEXT: vmv.s.x v10 , a0
2118
2111
; CHECK-V-NEXT: addi a0, sp, 16
2119
- ; CHECK-V-NEXT: vl1r.v v9 , (a0) # Unknown-size Folded Reload
2120
- ; CHECK-V-NEXT: vslideup.vi v8, v9 , 1
2112
+ ; CHECK-V-NEXT: vl1r.v v8 , (a0) # Unknown-size Folded Reload
2113
+ ; CHECK-V-NEXT: vslideup.vi v10, v8 , 1
2121
2114
; CHECK-V-NEXT: csrr a0, vlenb
2122
2115
; CHECK-V-NEXT: add a0, sp, a0
2123
2116
; CHECK-V-NEXT: addi a0, a0, 16
2124
- ; CHECK-V-NEXT: vl1r.v v9 , (a0) # Unknown-size Folded Reload
2117
+ ; CHECK-V-NEXT: vl1r.v v8 , (a0) # Unknown-size Folded Reload
2125
2118
; CHECK-V-NEXT: vsetivli zero, 4, e32, m1, ta, ma
2126
- ; CHECK-V-NEXT: vslideup.vi v8, v9 , 2
2119
+ ; CHECK-V-NEXT: vslideup.vi v10, v8 , 2
2127
2120
; CHECK-V-NEXT: csrr a0, vlenb
2128
2121
; CHECK-V-NEXT: slli a0, a0, 1
2129
2122
; CHECK-V-NEXT: add a0, sp, a0
2130
2123
; CHECK-V-NEXT: addi a0, a0, 16
2131
- ; CHECK-V-NEXT: vl2r.v v10 , (a0) # Unknown-size Folded Reload
2124
+ ; CHECK-V-NEXT: vl2r.v v8 , (a0) # Unknown-size Folded Reload
2132
2125
; CHECK-V-NEXT: vsetivli zero, 8, e32, m2, ta, ma
2133
- ; CHECK-V-NEXT: vslideup.vi v8, v10, 4
2134
- ; CHECK-V-NEXT: vmax.vx v10, v8, zero
2126
+ ; CHECK-V-NEXT: vslideup.vi v10, v8, 4
2135
2127
; CHECK-V-NEXT: vsetvli zero, zero, e16, m1, ta, ma
2136
2128
; CHECK-V-NEXT: vnclipu.wi v8, v10, 0
2137
2129
; CHECK-V-NEXT: csrr a0, vlenb
@@ -3473,7 +3465,6 @@ define <2 x i32> @ustest_f64i32_mm(<2 x double> %x) {
3473
3465
; CHECK-V: # %bb.0: # %entry
3474
3466
; CHECK-V-NEXT: vsetivli zero, 2, e64, m1, ta, ma
3475
3467
; CHECK-V-NEXT: vfcvt.rtz.x.f.v v8, v8
3476
- ; CHECK-V-NEXT: vmax.vx v8, v8, zero
3477
3468
; CHECK-V-NEXT: vsetvli zero, zero, e32, mf2, ta, ma
3478
3469
; CHECK-V-NEXT: vnclipu.wi v8, v8, 0
3479
3470
; CHECK-V-NEXT: ret
@@ -3659,9 +3650,6 @@ define <4 x i32> @ustest_f32i32_mm(<4 x float> %x) {
3659
3650
; CHECK-V: # %bb.0: # %entry
3660
3651
; CHECK-V-NEXT: vsetivli zero, 4, e32, m1, ta, ma
3661
3652
; CHECK-V-NEXT: vfwcvt.rtz.x.f.v v10, v8
3662
- ; CHECK-V-NEXT: vsetvli zero, zero, e64, m2, ta, ma
3663
- ; CHECK-V-NEXT: vmax.vx v10, v10, zero
3664
- ; CHECK-V-NEXT: vsetvli zero, zero, e32, m1, ta, ma
3665
3653
; CHECK-V-NEXT: vnclipu.wi v8, v10, 0
3666
3654
; CHECK-V-NEXT: ret
3667
3655
entry:
@@ -4151,17 +4139,16 @@ define <4 x i32> @ustest_f16i32_mm(<4 x half> %x) {
4151
4139
; CHECK-V-NEXT: call __extendhfsf2
4152
4140
; CHECK-V-NEXT: fcvt.l.s a0, fa0, rtz
4153
4141
; CHECK-V-NEXT: vsetivli zero, 2, e64, m1, ta, ma
4154
- ; CHECK-V-NEXT: vmv.s.x v8 , a0
4142
+ ; CHECK-V-NEXT: vmv.s.x v10 , a0
4155
4143
; CHECK-V-NEXT: addi a0, sp, 16
4156
- ; CHECK-V-NEXT: vl1r.v v9 , (a0) # Unknown-size Folded Reload
4157
- ; CHECK-V-NEXT: vslideup.vi v8, v9 , 1
4144
+ ; CHECK-V-NEXT: vl1r.v v8 , (a0) # Unknown-size Folded Reload
4145
+ ; CHECK-V-NEXT: vslideup.vi v10, v8 , 1
4158
4146
; CHECK-V-NEXT: csrr a0, vlenb
4159
4147
; CHECK-V-NEXT: add a0, sp, a0
4160
4148
; CHECK-V-NEXT: addi a0, a0, 16
4161
- ; CHECK-V-NEXT: vl2r.v v10 , (a0) # Unknown-size Folded Reload
4149
+ ; CHECK-V-NEXT: vl2r.v v8 , (a0) # Unknown-size Folded Reload
4162
4150
; CHECK-V-NEXT: vsetivli zero, 4, e64, m2, ta, ma
4163
- ; CHECK-V-NEXT: vslideup.vi v8, v10, 2
4164
- ; CHECK-V-NEXT: vmax.vx v10, v8, zero
4151
+ ; CHECK-V-NEXT: vslideup.vi v10, v8, 2
4165
4152
; CHECK-V-NEXT: vsetvli zero, zero, e32, m1, ta, ma
4166
4153
; CHECK-V-NEXT: vnclipu.wi v8, v10, 0
4167
4154
; CHECK-V-NEXT: csrr a0, vlenb
@@ -4289,9 +4276,8 @@ define <2 x i16> @ustest_f64i16_mm(<2 x double> %x) {
4289
4276
; CHECK-V: # %bb.0: # %entry
4290
4277
; CHECK-V-NEXT: vsetivli zero, 2, e32, mf2, ta, ma
4291
4278
; CHECK-V-NEXT: vfncvt.rtz.x.f.w v9, v8
4292
- ; CHECK-V-NEXT: vmax.vx v8, v9, zero
4293
4279
; CHECK-V-NEXT: vsetvli zero, zero, e16, mf4, ta, ma
4294
- ; CHECK-V-NEXT: vnclipu.wi v8, v8 , 0
4280
+ ; CHECK-V-NEXT: vnclipu.wi v8, v9 , 0
4295
4281
; CHECK-V-NEXT: ret
4296
4282
entry:
4297
4283
%conv = fptosi <2 x double > %x to <2 x i32 >
@@ -4479,7 +4465,6 @@ define <4 x i16> @ustest_f32i16_mm(<4 x float> %x) {
4479
4465
; CHECK-V: # %bb.0: # %entry
4480
4466
; CHECK-V-NEXT: vsetivli zero, 4, e32, m1, ta, ma
4481
4467
; CHECK-V-NEXT: vfcvt.rtz.x.f.v v8, v8
4482
- ; CHECK-V-NEXT: vmax.vx v8, v8, zero
4483
4468
; CHECK-V-NEXT: vsetvli zero, zero, e16, mf2, ta, ma
4484
4469
; CHECK-V-NEXT: vnclipu.wi v8, v8, 0
4485
4470
; CHECK-V-NEXT: ret
@@ -5449,24 +5434,23 @@ define <8 x i16> @ustest_f16i16_mm(<8 x half> %x) {
5449
5434
; CHECK-V-NEXT: call __extendhfsf2
5450
5435
; CHECK-V-NEXT: fcvt.l.s a0, fa0, rtz
5451
5436
; CHECK-V-NEXT: vsetivli zero, 2, e32, mf2, ta, ma
5452
- ; CHECK-V-NEXT: vmv.s.x v8 , a0
5437
+ ; CHECK-V-NEXT: vmv.s.x v10 , a0
5453
5438
; CHECK-V-NEXT: addi a0, sp, 16
5454
- ; CHECK-V-NEXT: vl1r.v v9 , (a0) # Unknown-size Folded Reload
5455
- ; CHECK-V-NEXT: vslideup.vi v8, v9 , 1
5439
+ ; CHECK-V-NEXT: vl1r.v v8 , (a0) # Unknown-size Folded Reload
5440
+ ; CHECK-V-NEXT: vslideup.vi v10, v8 , 1
5456
5441
; CHECK-V-NEXT: csrr a0, vlenb
5457
5442
; CHECK-V-NEXT: add a0, sp, a0
5458
5443
; CHECK-V-NEXT: addi a0, a0, 16
5459
- ; CHECK-V-NEXT: vl1r.v v9 , (a0) # Unknown-size Folded Reload
5444
+ ; CHECK-V-NEXT: vl1r.v v8 , (a0) # Unknown-size Folded Reload
5460
5445
; CHECK-V-NEXT: vsetivli zero, 4, e32, m1, ta, ma
5461
- ; CHECK-V-NEXT: vslideup.vi v8, v9 , 2
5446
+ ; CHECK-V-NEXT: vslideup.vi v10, v8 , 2
5462
5447
; CHECK-V-NEXT: csrr a0, vlenb
5463
5448
; CHECK-V-NEXT: slli a0, a0, 1
5464
5449
; CHECK-V-NEXT: add a0, sp, a0
5465
5450
; CHECK-V-NEXT: addi a0, a0, 16
5466
- ; CHECK-V-NEXT: vl2r.v v10 , (a0) # Unknown-size Folded Reload
5451
+ ; CHECK-V-NEXT: vl2r.v v8 , (a0) # Unknown-size Folded Reload
5467
5452
; CHECK-V-NEXT: vsetivli zero, 8, e32, m2, ta, ma
5468
- ; CHECK-V-NEXT: vslideup.vi v8, v10, 4
5469
- ; CHECK-V-NEXT: vmax.vx v10, v8, zero
5453
+ ; CHECK-V-NEXT: vslideup.vi v10, v8, 4
5470
5454
; CHECK-V-NEXT: vsetvli zero, zero, e16, m1, ta, ma
5471
5455
; CHECK-V-NEXT: vnclipu.wi v8, v10, 0
5472
5456
; CHECK-V-NEXT: csrr a0, vlenb
0 commit comments