@@ -976,24 +976,24 @@ define <vscale x 16 x i64> @fshr_v16i64(<vscale x 16 x i64> %a, <vscale x 16 x i
976
976
; CHECK-NEXT: add a1, sp, a1
977
977
; CHECK-NEXT: addi a1, a1, 16
978
978
; CHECK-NEXT: vs8r.v v8, (a1) # Unknown-size Folded Spill
979
- ; CHECK-NEXT: csrr a1 , vlenb
980
- ; CHECK-NEXT: slli a3, a1 , 3
981
- ; CHECK-NEXT: add a5, a0, a3
979
+ ; CHECK-NEXT: csrr a3 , vlenb
980
+ ; CHECK-NEXT: slli a1, a3 , 3
981
+ ; CHECK-NEXT: add a5, a0, a1
982
982
; CHECK-NEXT: vl8re64.v v8, (a5)
983
983
; CHECK-NEXT: csrr a5, vlenb
984
984
; CHECK-NEXT: slli a5, a5, 3
985
985
; CHECK-NEXT: add a5, sp, a5
986
986
; CHECK-NEXT: addi a5, a5, 16
987
987
; CHECK-NEXT: vs8r.v v8, (a5) # Unknown-size Folded Spill
988
- ; CHECK-NEXT: srli a5, a1 , 3
988
+ ; CHECK-NEXT: srli a5, a3 , 3
989
989
; CHECK-NEXT: vsetvli a6, zero, e8, mf4, ta, ma
990
990
; CHECK-NEXT: vslidedown.vx v0, v0, a5
991
- ; CHECK-NEXT: add a5, a2, a3
992
- ; CHECK-NEXT: sub a3 , a4, a1
993
- ; CHECK-NEXT: sltu a6, a4, a3
991
+ ; CHECK-NEXT: add a5, a2, a1
992
+ ; CHECK-NEXT: sub a1 , a4, a3
993
+ ; CHECK-NEXT: sltu a6, a4, a1
994
994
; CHECK-NEXT: addi a6, a6, -1
995
- ; CHECK-NEXT: and a6, a6, a3
996
- ; CHECK-NEXT: li a3 , 63
995
+ ; CHECK-NEXT: and a6, a6, a1
996
+ ; CHECK-NEXT: li a1 , 63
997
997
; CHECK-NEXT: vl8re64.v v8, (a5)
998
998
; CHECK-NEXT: csrr a5, vlenb
999
999
; CHECK-NEXT: li a7, 40
@@ -1021,7 +1021,7 @@ define <vscale x 16 x i64> @fshr_v16i64(<vscale x 16 x i64> %a, <vscale x 16 x i
1021
1021
; CHECK-NEXT: add a0, sp, a0
1022
1022
; CHECK-NEXT: addi a0, a0, 16
1023
1023
; CHECK-NEXT: vl8r.v v8, (a0) # Unknown-size Folded Reload
1024
- ; CHECK-NEXT: vand.vx v8, v8, a3 , v0.t
1024
+ ; CHECK-NEXT: vand.vx v8, v8, a1 , v0.t
1025
1025
; CHECK-NEXT: addi a0, sp, 16
1026
1026
; CHECK-NEXT: vs8r.v v8, (a0) # Unknown-size Folded Spill
1027
1027
; CHECK-NEXT: csrr a0, vlenb
@@ -1044,7 +1044,7 @@ define <vscale x 16 x i64> @fshr_v16i64(<vscale x 16 x i64> %a, <vscale x 16 x i
1044
1044
; CHECK-NEXT: addi a0, a0, 16
1045
1045
; CHECK-NEXT: vl8r.v v8, (a0) # Unknown-size Folded Reload
1046
1046
; CHECK-NEXT: vnot.v v8, v8, v0.t
1047
- ; CHECK-NEXT: vand.vx v16, v8, a3 , v0.t
1047
+ ; CHECK-NEXT: vand.vx v16, v8, a1 , v0.t
1048
1048
; CHECK-NEXT: csrr a0, vlenb
1049
1049
; CHECK-NEXT: li a2, 24
1050
1050
; CHECK-NEXT: mul a0, a0, a2
@@ -1065,22 +1065,22 @@ define <vscale x 16 x i64> @fshr_v16i64(<vscale x 16 x i64> %a, <vscale x 16 x i
1065
1065
; CHECK-NEXT: add a0, sp, a0
1066
1066
; CHECK-NEXT: addi a0, a0, 16
1067
1067
; CHECK-NEXT: vs8r.v v8, (a0) # Unknown-size Folded Spill
1068
- ; CHECK-NEXT: bltu a4, a1 , .LBB46_2
1068
+ ; CHECK-NEXT: bltu a4, a3 , .LBB46_2
1069
1069
; CHECK-NEXT: # %bb.1:
1070
- ; CHECK-NEXT: mv a4, a1
1070
+ ; CHECK-NEXT: mv a4, a3
1071
1071
; CHECK-NEXT: .LBB46_2:
1072
1072
; CHECK-NEXT: vsetvli zero, a4, e64, m8, ta, ma
1073
1073
; CHECK-NEXT: vmv1r.v v0, v24
1074
1074
; CHECK-NEXT: csrr a0, vlenb
1075
- ; CHECK-NEXT: li a1 , 48
1076
- ; CHECK-NEXT: mul a0, a0, a1
1075
+ ; CHECK-NEXT: li a2 , 48
1076
+ ; CHECK-NEXT: mul a0, a0, a2
1077
1077
; CHECK-NEXT: add a0, sp, a0
1078
1078
; CHECK-NEXT: addi a0, a0, 16
1079
1079
; CHECK-NEXT: vl8r.v v8, (a0) # Unknown-size Folded Reload
1080
- ; CHECK-NEXT: vand.vx v8, v8, a3 , v0.t
1080
+ ; CHECK-NEXT: vand.vx v8, v8, a1 , v0.t
1081
1081
; CHECK-NEXT: csrr a0, vlenb
1082
- ; CHECK-NEXT: li a1 , 24
1083
- ; CHECK-NEXT: mul a0, a0, a1
1082
+ ; CHECK-NEXT: li a2 , 24
1083
+ ; CHECK-NEXT: mul a0, a0, a2
1084
1084
; CHECK-NEXT: add a0, sp, a0
1085
1085
; CHECK-NEXT: addi a0, a0, 16
1086
1086
; CHECK-NEXT: vs8r.v v8, (a0) # Unknown-size Folded Spill
@@ -1090,26 +1090,26 @@ define <vscale x 16 x i64> @fshr_v16i64(<vscale x 16 x i64> %a, <vscale x 16 x i
1090
1090
; CHECK-NEXT: addi a0, a0, 16
1091
1091
; CHECK-NEXT: vl8r.v v8, (a0) # Unknown-size Folded Reload
1092
1092
; CHECK-NEXT: csrr a0, vlenb
1093
- ; CHECK-NEXT: li a1 , 24
1094
- ; CHECK-NEXT: mul a0, a0, a1
1093
+ ; CHECK-NEXT: li a2 , 24
1094
+ ; CHECK-NEXT: mul a0, a0, a2
1095
1095
; CHECK-NEXT: add a0, sp, a0
1096
1096
; CHECK-NEXT: addi a0, a0, 16
1097
1097
; CHECK-NEXT: vl8r.v v16, (a0) # Unknown-size Folded Reload
1098
1098
; CHECK-NEXT: vsrl.vv v8, v8, v16, v0.t
1099
1099
; CHECK-NEXT: csrr a0, vlenb
1100
- ; CHECK-NEXT: li a1 , 24
1101
- ; CHECK-NEXT: mul a0, a0, a1
1100
+ ; CHECK-NEXT: li a2 , 24
1101
+ ; CHECK-NEXT: mul a0, a0, a2
1102
1102
; CHECK-NEXT: add a0, sp, a0
1103
1103
; CHECK-NEXT: addi a0, a0, 16
1104
1104
; CHECK-NEXT: vs8r.v v8, (a0) # Unknown-size Folded Spill
1105
1105
; CHECK-NEXT: csrr a0, vlenb
1106
- ; CHECK-NEXT: li a1 , 48
1107
- ; CHECK-NEXT: mul a0, a0, a1
1106
+ ; CHECK-NEXT: li a2 , 48
1107
+ ; CHECK-NEXT: mul a0, a0, a2
1108
1108
; CHECK-NEXT: add a0, sp, a0
1109
1109
; CHECK-NEXT: addi a0, a0, 16
1110
1110
; CHECK-NEXT: vl8r.v v8, (a0) # Unknown-size Folded Reload
1111
1111
; CHECK-NEXT: vnot.v v16, v8, v0.t
1112
- ; CHECK-NEXT: vand.vx v16, v16, a3 , v0.t
1112
+ ; CHECK-NEXT: vand.vx v16, v16, a1 , v0.t
1113
1113
; CHECK-NEXT: csrr a0, vlenb
1114
1114
; CHECK-NEXT: slli a0, a0, 5
1115
1115
; CHECK-NEXT: add a0, sp, a0
@@ -1162,24 +1162,24 @@ define <vscale x 16 x i64> @fshl_v16i64(<vscale x 16 x i64> %a, <vscale x 16 x i
1162
1162
; CHECK-NEXT: add a1, sp, a1
1163
1163
; CHECK-NEXT: addi a1, a1, 16
1164
1164
; CHECK-NEXT: vs8r.v v8, (a1) # Unknown-size Folded Spill
1165
- ; CHECK-NEXT: csrr a1 , vlenb
1166
- ; CHECK-NEXT: slli a3, a1 , 3
1167
- ; CHECK-NEXT: add a5, a0, a3
1165
+ ; CHECK-NEXT: csrr a3 , vlenb
1166
+ ; CHECK-NEXT: slli a1, a3 , 3
1167
+ ; CHECK-NEXT: add a5, a0, a1
1168
1168
; CHECK-NEXT: vl8re64.v v8, (a5)
1169
1169
; CHECK-NEXT: csrr a5, vlenb
1170
1170
; CHECK-NEXT: slli a5, a5, 4
1171
1171
; CHECK-NEXT: add a5, sp, a5
1172
1172
; CHECK-NEXT: addi a5, a5, 16
1173
1173
; CHECK-NEXT: vs8r.v v8, (a5) # Unknown-size Folded Spill
1174
- ; CHECK-NEXT: srli a5, a1 , 3
1174
+ ; CHECK-NEXT: srli a5, a3 , 3
1175
1175
; CHECK-NEXT: vsetvli a6, zero, e8, mf4, ta, ma
1176
1176
; CHECK-NEXT: vslidedown.vx v0, v0, a5
1177
- ; CHECK-NEXT: add a5, a2, a3
1178
- ; CHECK-NEXT: sub a3 , a4, a1
1179
- ; CHECK-NEXT: sltu a6, a4, a3
1177
+ ; CHECK-NEXT: add a5, a2, a1
1178
+ ; CHECK-NEXT: sub a1 , a4, a3
1179
+ ; CHECK-NEXT: sltu a6, a4, a1
1180
1180
; CHECK-NEXT: addi a6, a6, -1
1181
- ; CHECK-NEXT: and a6, a6, a3
1182
- ; CHECK-NEXT: li a3 , 63
1181
+ ; CHECK-NEXT: and a6, a6, a1
1182
+ ; CHECK-NEXT: li a1 , 63
1183
1183
; CHECK-NEXT: vl8re64.v v8, (a5)
1184
1184
; CHECK-NEXT: csrr a5, vlenb
1185
1185
; CHECK-NEXT: li a7, 40
@@ -1208,7 +1208,7 @@ define <vscale x 16 x i64> @fshl_v16i64(<vscale x 16 x i64> %a, <vscale x 16 x i
1208
1208
; CHECK-NEXT: add a0, sp, a0
1209
1209
; CHECK-NEXT: addi a0, a0, 16
1210
1210
; CHECK-NEXT: vl8r.v v8, (a0) # Unknown-size Folded Reload
1211
- ; CHECK-NEXT: vand.vx v8, v8, a3 , v0.t
1211
+ ; CHECK-NEXT: vand.vx v8, v8, a1 , v0.t
1212
1212
; CHECK-NEXT: addi a0, sp, 16
1213
1213
; CHECK-NEXT: vs8r.v v8, (a0) # Unknown-size Folded Spill
1214
1214
; CHECK-NEXT: csrr a0, vlenb
@@ -1231,7 +1231,7 @@ define <vscale x 16 x i64> @fshl_v16i64(<vscale x 16 x i64> %a, <vscale x 16 x i
1231
1231
; CHECK-NEXT: addi a0, a0, 16
1232
1232
; CHECK-NEXT: vl8r.v v8, (a0) # Unknown-size Folded Reload
1233
1233
; CHECK-NEXT: vnot.v v8, v8, v0.t
1234
- ; CHECK-NEXT: vand.vx v16, v8, a3 , v0.t
1234
+ ; CHECK-NEXT: vand.vx v16, v8, a1 , v0.t
1235
1235
; CHECK-NEXT: csrr a0, vlenb
1236
1236
; CHECK-NEXT: slli a0, a0, 4
1237
1237
; CHECK-NEXT: add a0, sp, a0
@@ -1251,19 +1251,19 @@ define <vscale x 16 x i64> @fshl_v16i64(<vscale x 16 x i64> %a, <vscale x 16 x i
1251
1251
; CHECK-NEXT: add a0, sp, a0
1252
1252
; CHECK-NEXT: addi a0, a0, 16
1253
1253
; CHECK-NEXT: vs8r.v v8, (a0) # Unknown-size Folded Spill
1254
- ; CHECK-NEXT: bltu a4, a1 , .LBB47_2
1254
+ ; CHECK-NEXT: bltu a4, a3 , .LBB47_2
1255
1255
; CHECK-NEXT: # %bb.1:
1256
- ; CHECK-NEXT: mv a4, a1
1256
+ ; CHECK-NEXT: mv a4, a3
1257
1257
; CHECK-NEXT: .LBB47_2:
1258
1258
; CHECK-NEXT: vsetvli zero, a4, e64, m8, ta, ma
1259
1259
; CHECK-NEXT: vmv1r.v v0, v24
1260
1260
; CHECK-NEXT: csrr a0, vlenb
1261
- ; CHECK-NEXT: li a1 , 48
1262
- ; CHECK-NEXT: mul a0, a0, a1
1261
+ ; CHECK-NEXT: li a2 , 48
1262
+ ; CHECK-NEXT: mul a0, a0, a2
1263
1263
; CHECK-NEXT: add a0, sp, a0
1264
1264
; CHECK-NEXT: addi a0, a0, 16
1265
1265
; CHECK-NEXT: vl8r.v v8, (a0) # Unknown-size Folded Reload
1266
- ; CHECK-NEXT: vand.vx v8, v8, a3 , v0.t
1266
+ ; CHECK-NEXT: vand.vx v8, v8, a1 , v0.t
1267
1267
; CHECK-NEXT: csrr a0, vlenb
1268
1268
; CHECK-NEXT: slli a0, a0, 4
1269
1269
; CHECK-NEXT: add a0, sp, a0
@@ -1286,13 +1286,13 @@ define <vscale x 16 x i64> @fshl_v16i64(<vscale x 16 x i64> %a, <vscale x 16 x i
1286
1286
; CHECK-NEXT: addi a0, a0, 16
1287
1287
; CHECK-NEXT: vs8r.v v16, (a0) # Unknown-size Folded Spill
1288
1288
; CHECK-NEXT: csrr a0, vlenb
1289
- ; CHECK-NEXT: li a1 , 48
1290
- ; CHECK-NEXT: mul a0, a0, a1
1289
+ ; CHECK-NEXT: li a2 , 48
1290
+ ; CHECK-NEXT: mul a0, a0, a2
1291
1291
; CHECK-NEXT: add a0, sp, a0
1292
1292
; CHECK-NEXT: addi a0, a0, 16
1293
1293
; CHECK-NEXT: vl8r.v v8, (a0) # Unknown-size Folded Reload
1294
1294
; CHECK-NEXT: vnot.v v16, v8, v0.t
1295
- ; CHECK-NEXT: vand.vx v16, v16, a3 , v0.t
1295
+ ; CHECK-NEXT: vand.vx v16, v16, a1 , v0.t
1296
1296
; CHECK-NEXT: csrr a0, vlenb
1297
1297
; CHECK-NEXT: li a1, 24
1298
1298
; CHECK-NEXT: mul a0, a0, a1
0 commit comments