@@ -549,36 +549,36 @@ define <33 x double> @strided_load_v33f64(ptr %ptr, i64 %stride, <33 x i1> %mask
549
549
; CHECK-RV32-NEXT: # %bb.1:
550
550
; CHECK-RV32-NEXT: li a3, 32
551
551
; CHECK-RV32-NEXT: .LBB42_2:
552
- ; CHECK-RV32-NEXT: mul a5 , a3, a2
553
- ; CHECK-RV32-NEXT: addi a6 , a4, -32
554
- ; CHECK-RV32-NEXT: sltu a4 , a4, a6
555
- ; CHECK-RV32-NEXT: addi a4, a4 , -1
556
- ; CHECK-RV32-NEXT: and a6, a4, a6
557
- ; CHECK-RV32-NEXT: li a4 , 16
558
- ; CHECK-RV32-NEXT: add a5 , a1, a5
559
- ; CHECK-RV32-NEXT: bltu a6, a4 , .LBB42_4
552
+ ; CHECK-RV32-NEXT: mul a6 , a3, a2
553
+ ; CHECK-RV32-NEXT: addi a5 , a4, -32
554
+ ; CHECK-RV32-NEXT: sltu a7 , a4, a5
555
+ ; CHECK-RV32-NEXT: addi a7, a7 , -1
556
+ ; CHECK-RV32-NEXT: and a7, a7, a5
557
+ ; CHECK-RV32-NEXT: li a5 , 16
558
+ ; CHECK-RV32-NEXT: add a6 , a1, a6
559
+ ; CHECK-RV32-NEXT: bltu a7, a5 , .LBB42_4
560
560
; CHECK-RV32-NEXT: # %bb.3:
561
- ; CHECK-RV32-NEXT: li a6 , 16
561
+ ; CHECK-RV32-NEXT: li a7 , 16
562
562
; CHECK-RV32-NEXT: .LBB42_4:
563
563
; CHECK-RV32-NEXT: vsetivli zero, 4, e8, mf2, ta, ma
564
564
; CHECK-RV32-NEXT: vslidedown.vi v0, v8, 4
565
- ; CHECK-RV32-NEXT: vsetvli zero, a6 , e64, m8, ta, ma
566
- ; CHECK-RV32-NEXT: vlse64.v v16, (a5 ), a2, v0.t
567
- ; CHECK-RV32-NEXT: addi a5 , a3, -16
568
- ; CHECK-RV32-NEXT: sltu a6 , a3, a5
569
- ; CHECK-RV32-NEXT: addi a6, a6 , -1
570
- ; CHECK-RV32-NEXT: and a5, a6, a5
571
- ; CHECK-RV32-NEXT: bltu a3, a4 , .LBB42_6
565
+ ; CHECK-RV32-NEXT: vsetvli zero, a7 , e64, m8, ta, ma
566
+ ; CHECK-RV32-NEXT: vlse64.v v16, (a6 ), a2, v0.t
567
+ ; CHECK-RV32-NEXT: addi a6 , a3, -16
568
+ ; CHECK-RV32-NEXT: sltu a3 , a3, a6
569
+ ; CHECK-RV32-NEXT: addi a3, a3 , -1
570
+ ; CHECK-RV32-NEXT: and a3, a3, a6
571
+ ; CHECK-RV32-NEXT: bltu a4, a5 , .LBB42_6
572
572
; CHECK-RV32-NEXT: # %bb.5:
573
- ; CHECK-RV32-NEXT: li a3 , 16
573
+ ; CHECK-RV32-NEXT: li a4 , 16
574
574
; CHECK-RV32-NEXT: .LBB42_6:
575
- ; CHECK-RV32-NEXT: mul a4, a3 , a2
576
- ; CHECK-RV32-NEXT: add a4 , a1, a4
575
+ ; CHECK-RV32-NEXT: mul a5, a4 , a2
576
+ ; CHECK-RV32-NEXT: add a5 , a1, a5
577
577
; CHECK-RV32-NEXT: vsetivli zero, 2, e8, mf4, ta, ma
578
578
; CHECK-RV32-NEXT: vslidedown.vi v0, v8, 2
579
- ; CHECK-RV32-NEXT: vsetvli zero, a5, e64, m8, ta, ma
580
- ; CHECK-RV32-NEXT: vlse64.v v24, (a4), a2, v0.t
581
579
; CHECK-RV32-NEXT: vsetvli zero, a3, e64, m8, ta, ma
580
+ ; CHECK-RV32-NEXT: vlse64.v v24, (a5), a2, v0.t
581
+ ; CHECK-RV32-NEXT: vsetvli zero, a4, e64, m8, ta, ma
582
582
; CHECK-RV32-NEXT: vmv1r.v v0, v8
583
583
; CHECK-RV32-NEXT: vlse64.v v8, (a1), a2, v0.t
584
584
; CHECK-RV32-NEXT: vsetivli zero, 16, e64, m8, ta, ma
@@ -599,36 +599,36 @@ define <33 x double> @strided_load_v33f64(ptr %ptr, i64 %stride, <33 x i1> %mask
599
599
; CHECK-RV64-NEXT: # %bb.1:
600
600
; CHECK-RV64-NEXT: li a4, 32
601
601
; CHECK-RV64-NEXT: .LBB42_2:
602
- ; CHECK-RV64-NEXT: mul a5 , a4, a2
603
- ; CHECK-RV64-NEXT: addi a6 , a3, -32
604
- ; CHECK-RV64-NEXT: sltu a3 , a3, a6
605
- ; CHECK-RV64-NEXT: addi a3, a3 , -1
606
- ; CHECK-RV64-NEXT: and a6, a3, a6
607
- ; CHECK-RV64-NEXT: li a3 , 16
608
- ; CHECK-RV64-NEXT: add a5 , a1, a5
609
- ; CHECK-RV64-NEXT: bltu a6, a3 , .LBB42_4
602
+ ; CHECK-RV64-NEXT: mul a6 , a4, a2
603
+ ; CHECK-RV64-NEXT: addi a5 , a3, -32
604
+ ; CHECK-RV64-NEXT: sltu a7 , a3, a5
605
+ ; CHECK-RV64-NEXT: addi a7, a7 , -1
606
+ ; CHECK-RV64-NEXT: and a7, a7, a5
607
+ ; CHECK-RV64-NEXT: li a5 , 16
608
+ ; CHECK-RV64-NEXT: add a6 , a1, a6
609
+ ; CHECK-RV64-NEXT: bltu a7, a5 , .LBB42_4
610
610
; CHECK-RV64-NEXT: # %bb.3:
611
- ; CHECK-RV64-NEXT: li a6 , 16
611
+ ; CHECK-RV64-NEXT: li a7 , 16
612
612
; CHECK-RV64-NEXT: .LBB42_4:
613
613
; CHECK-RV64-NEXT: vsetivli zero, 4, e8, mf2, ta, ma
614
614
; CHECK-RV64-NEXT: vslidedown.vi v0, v8, 4
615
- ; CHECK-RV64-NEXT: vsetvli zero, a6 , e64, m8, ta, ma
616
- ; CHECK-RV64-NEXT: vlse64.v v16, (a5 ), a2, v0.t
617
- ; CHECK-RV64-NEXT: addi a5 , a4, -16
618
- ; CHECK-RV64-NEXT: sltu a6 , a4, a5
619
- ; CHECK-RV64-NEXT: addi a6, a6 , -1
620
- ; CHECK-RV64-NEXT: and a5, a6, a5
621
- ; CHECK-RV64-NEXT: bltu a4, a3 , .LBB42_6
615
+ ; CHECK-RV64-NEXT: vsetvli zero, a7 , e64, m8, ta, ma
616
+ ; CHECK-RV64-NEXT: vlse64.v v16, (a6 ), a2, v0.t
617
+ ; CHECK-RV64-NEXT: addi a6 , a4, -16
618
+ ; CHECK-RV64-NEXT: sltu a4 , a4, a6
619
+ ; CHECK-RV64-NEXT: addi a4, a4 , -1
620
+ ; CHECK-RV64-NEXT: and a4, a4, a6
621
+ ; CHECK-RV64-NEXT: bltu a3, a5 , .LBB42_6
622
622
; CHECK-RV64-NEXT: # %bb.5:
623
- ; CHECK-RV64-NEXT: li a4 , 16
623
+ ; CHECK-RV64-NEXT: li a3 , 16
624
624
; CHECK-RV64-NEXT: .LBB42_6:
625
- ; CHECK-RV64-NEXT: mul a3, a4 , a2
626
- ; CHECK-RV64-NEXT: add a3 , a1, a3
625
+ ; CHECK-RV64-NEXT: mul a5, a3 , a2
626
+ ; CHECK-RV64-NEXT: add a5 , a1, a5
627
627
; CHECK-RV64-NEXT: vsetivli zero, 2, e8, mf4, ta, ma
628
628
; CHECK-RV64-NEXT: vslidedown.vi v0, v8, 2
629
- ; CHECK-RV64-NEXT: vsetvli zero, a5, e64, m8, ta, ma
630
- ; CHECK-RV64-NEXT: vlse64.v v24, (a3), a2, v0.t
631
629
; CHECK-RV64-NEXT: vsetvli zero, a4, e64, m8, ta, ma
630
+ ; CHECK-RV64-NEXT: vlse64.v v24, (a5), a2, v0.t
631
+ ; CHECK-RV64-NEXT: vsetvli zero, a3, e64, m8, ta, ma
632
632
; CHECK-RV64-NEXT: vmv1r.v v0, v8
633
633
; CHECK-RV64-NEXT: vlse64.v v8, (a1), a2, v0.t
634
634
; CHECK-RV64-NEXT: vsetivli zero, 16, e64, m8, ta, ma
0 commit comments