@@ -781,13 +781,13 @@ define {<8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>} @load_
781
781
; RV64-NEXT: vl8r.v v16, (a1) # Unknown-size Folded Reload
782
782
; RV64-NEXT: vsetivli zero, 16, e64, m8, ta, ma
783
783
; RV64-NEXT: vmerge.vvm v24, v16, v8, v0
784
- ; RV64-NEXT: vmv8r.v v16, v8
785
784
; RV64-NEXT: csrr a1, vlenb
786
785
; RV64-NEXT: li a3, 76
787
786
; RV64-NEXT: mul a1, a1, a3
788
787
; RV64-NEXT: add a1, sp, a1
789
788
; RV64-NEXT: addi a1, a1, 16
790
789
; RV64-NEXT: vs8r.v v8, (a1) # Unknown-size Folded Spill
790
+ ; RV64-NEXT: vmv8r.v v16, v8
791
791
; RV64-NEXT: vrgatherei16.vv v8, v24, v2
792
792
; RV64-NEXT: csrr a1, vlenb
793
793
; RV64-NEXT: slli a1, a1, 5
@@ -904,7 +904,6 @@ define {<8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>} @load_
904
904
; RV64-NEXT: add a1, sp, a1
905
905
; RV64-NEXT: addi a1, a1, 16
906
906
; RV64-NEXT: vs8r.v v24, (a1) # Unknown-size Folded Spill
907
- ; RV64-NEXT: vmv8r.v v16, v8
908
907
; RV64-NEXT: vmv1r.v v0, v7
909
908
; RV64-NEXT: csrr a1, vlenb
910
909
; RV64-NEXT: li a3, 72
@@ -917,10 +916,10 @@ define {<8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>} @load_
917
916
; RV64-NEXT: mul a1, a1, a3
918
917
; RV64-NEXT: add a1, sp, a1
919
918
; RV64-NEXT: addi a1, a1, 16
920
- ; RV64-NEXT: vl8r.v v24 , (a1) # Unknown-size Folded Reload
921
- ; RV64-NEXT: vmv4r.v v8, v24
919
+ ; RV64-NEXT: vl8r.v v16 , (a1) # Unknown-size Folded Reload
920
+ ; RV64-NEXT: vmv4r.v v8, v16
922
921
; RV64-NEXT: vsetivli zero, 8, e64, m4, ta, mu
923
- ; RV64-NEXT: vrgather.vi v12, v24 , 5, v0.t
922
+ ; RV64-NEXT: vrgather.vi v12, v16 , 5, v0.t
924
923
; RV64-NEXT: csrr a1, vlenb
925
924
; RV64-NEXT: li a3, 72
926
925
; RV64-NEXT: mul a1, a1, a3
@@ -939,7 +938,7 @@ define {<8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>} @load_
939
938
; RV64-NEXT: addi a1, a1, 16
940
939
; RV64-NEXT: vl4r.v v24, (a1) # Unknown-size Folded Reload
941
940
; RV64-NEXT: vsetivli zero, 8, e64, m4, ta, mu
942
- ; RV64-NEXT: vrgatherei16.vv v24, v8 , v12, v0.t
941
+ ; RV64-NEXT: vrgatherei16.vv v24, v16 , v12, v0.t
943
942
; RV64-NEXT: csrr a1, vlenb
944
943
; RV64-NEXT: slli a1, a1, 6
945
944
; RV64-NEXT: add a1, sp, a1
@@ -955,7 +954,12 @@ define {<8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>, <8 x i64>} @load_
955
954
; RV64-NEXT: vmv.v.x v12, a4
956
955
; RV64-NEXT: vsetivli zero, 16, e64, m8, ta, ma
957
956
; RV64-NEXT: vle16.v v6, (a1)
958
- ; RV64-NEXT: vmv8r.v v24, v16
957
+ ; RV64-NEXT: csrr a1, vlenb
958
+ ; RV64-NEXT: li a3, 40
959
+ ; RV64-NEXT: mul a1, a1, a3
960
+ ; RV64-NEXT: add a1, sp, a1
961
+ ; RV64-NEXT: addi a1, a1, 16
962
+ ; RV64-NEXT: vl8r.v v24, (a1) # Unknown-size Folded Reload
959
963
; RV64-NEXT: csrr a1, vlenb
960
964
; RV64-NEXT: li a3, 76
961
965
; RV64-NEXT: mul a1, a1, a3
0 commit comments