7
7
define void @vadd_vint16m1 (<vscale x 4 x i16 > *%pc , <vscale x 4 x i16 > *%pa , <vscale x 4 x i16 > *%pb ) nounwind {
8
8
; CHECK-LABEL: vadd_vint16m1:
9
9
; CHECK: # %bb.0:
10
- ; CHECK-NEXT: vsetvli a3, zero, e16,m1,tu ,mu
10
+ ; CHECK-NEXT: vsetvli a3, zero, e16,m1,ta ,mu
11
11
; CHECK-NEXT: vle16.v v25, (a1)
12
- ; CHECK-NEXT: vsetvli a1, zero, e16,m1,tu ,mu
12
+ ; CHECK-NEXT: vsetvli a1, zero, e16,m1,ta ,mu
13
13
; CHECK-NEXT: vle16.v v26, (a2)
14
- ; CHECK-NEXT: vsetvli a1, zero, e16,m1,tu ,mu
14
+ ; CHECK-NEXT: vsetvli a1, zero, e16,m1,ta ,mu
15
15
; CHECK-NEXT: vadd.vv v25, v25, v26
16
- ; CHECK-NEXT: vsetvli a1, zero, e16,m1,tu ,mu
16
+ ; CHECK-NEXT: vsetvli a1, zero, e16,m1,ta ,mu
17
17
; CHECK-NEXT: vse16.v v25, (a0)
18
18
; CHECK-NEXT: ret
19
19
%va = load <vscale x 4 x i16 >, <vscale x 4 x i16 >* %pa
@@ -26,13 +26,13 @@ define void @vadd_vint16m1(<vscale x 4 x i16> *%pc, <vscale x 4 x i16> *%pa, <vs
26
26
define void @vadd_vint16m2 (<vscale x 8 x i16 > *%pc , <vscale x 8 x i16 > *%pa , <vscale x 8 x i16 > *%pb ) nounwind {
27
27
; CHECK-LABEL: vadd_vint16m2:
28
28
; CHECK: # %bb.0:
29
- ; CHECK-NEXT: vsetvli a3, zero, e16,m2,tu ,mu
29
+ ; CHECK-NEXT: vsetvli a3, zero, e16,m2,ta ,mu
30
30
; CHECK-NEXT: vle16.v v26, (a1)
31
- ; CHECK-NEXT: vsetvli a1, zero, e16,m2,tu ,mu
31
+ ; CHECK-NEXT: vsetvli a1, zero, e16,m2,ta ,mu
32
32
; CHECK-NEXT: vle16.v v28, (a2)
33
- ; CHECK-NEXT: vsetvli a1, zero, e16,m2,tu ,mu
33
+ ; CHECK-NEXT: vsetvli a1, zero, e16,m2,ta ,mu
34
34
; CHECK-NEXT: vadd.vv v26, v26, v28
35
- ; CHECK-NEXT: vsetvli a1, zero, e16,m2,tu ,mu
35
+ ; CHECK-NEXT: vsetvli a1, zero, e16,m2,ta ,mu
36
36
; CHECK-NEXT: vse16.v v26, (a0)
37
37
; CHECK-NEXT: ret
38
38
%va = load <vscale x 8 x i16 >, <vscale x 8 x i16 >* %pa
@@ -45,13 +45,13 @@ define void @vadd_vint16m2(<vscale x 8 x i16> *%pc, <vscale x 8 x i16> *%pa, <vs
45
45
define void @vadd_vint16m4 (<vscale x 16 x i16 > *%pc , <vscale x 16 x i16 > *%pa , <vscale x 16 x i16 > *%pb ) nounwind {
46
46
; CHECK-LABEL: vadd_vint16m4:
47
47
; CHECK: # %bb.0:
48
- ; CHECK-NEXT: vsetvli a3, zero, e16,m4,tu ,mu
48
+ ; CHECK-NEXT: vsetvli a3, zero, e16,m4,ta ,mu
49
49
; CHECK-NEXT: vle16.v v28, (a1)
50
- ; CHECK-NEXT: vsetvli a1, zero, e16,m4,tu ,mu
50
+ ; CHECK-NEXT: vsetvli a1, zero, e16,m4,ta ,mu
51
51
; CHECK-NEXT: vle16.v v8, (a2)
52
- ; CHECK-NEXT: vsetvli a1, zero, e16,m4,tu ,mu
52
+ ; CHECK-NEXT: vsetvli a1, zero, e16,m4,ta ,mu
53
53
; CHECK-NEXT: vadd.vv v28, v28, v8
54
- ; CHECK-NEXT: vsetvli a1, zero, e16,m4,tu ,mu
54
+ ; CHECK-NEXT: vsetvli a1, zero, e16,m4,ta ,mu
55
55
; CHECK-NEXT: vse16.v v28, (a0)
56
56
; CHECK-NEXT: ret
57
57
%va = load <vscale x 16 x i16 >, <vscale x 16 x i16 >* %pa
@@ -64,13 +64,13 @@ define void @vadd_vint16m4(<vscale x 16 x i16> *%pc, <vscale x 16 x i16> *%pa, <
64
64
define void @vadd_vint16m8 (<vscale x 32 x i16 > *%pc , <vscale x 32 x i16 > *%pa , <vscale x 32 x i16 > *%pb ) nounwind {
65
65
; CHECK-LABEL: vadd_vint16m8:
66
66
; CHECK: # %bb.0:
67
- ; CHECK-NEXT: vsetvli a3, zero, e16,m8,tu ,mu
67
+ ; CHECK-NEXT: vsetvli a3, zero, e16,m8,ta ,mu
68
68
; CHECK-NEXT: vle16.v v8, (a1)
69
- ; CHECK-NEXT: vsetvli a1, zero, e16,m8,tu ,mu
69
+ ; CHECK-NEXT: vsetvli a1, zero, e16,m8,ta ,mu
70
70
; CHECK-NEXT: vle16.v v16, (a2)
71
- ; CHECK-NEXT: vsetvli a1, zero, e16,m8,tu ,mu
71
+ ; CHECK-NEXT: vsetvli a1, zero, e16,m8,ta ,mu
72
72
; CHECK-NEXT: vadd.vv v8, v8, v16
73
- ; CHECK-NEXT: vsetvli a1, zero, e16,m8,tu ,mu
73
+ ; CHECK-NEXT: vsetvli a1, zero, e16,m8,ta ,mu
74
74
; CHECK-NEXT: vse16.v v8, (a0)
75
75
; CHECK-NEXT: ret
76
76
%va = load <vscale x 32 x i16 >, <vscale x 32 x i16 >* %pa
@@ -83,13 +83,13 @@ define void @vadd_vint16m8(<vscale x 32 x i16> *%pc, <vscale x 32 x i16> *%pa, <
83
83
define void @vadd_vint16mf2 (<vscale x 2 x i16 > *%pc , <vscale x 2 x i16 > *%pa , <vscale x 2 x i16 > *%pb ) nounwind {
84
84
; CHECK-LABEL: vadd_vint16mf2:
85
85
; CHECK: # %bb.0:
86
- ; CHECK-NEXT: vsetvli a3, zero, e16,mf2,tu ,mu
86
+ ; CHECK-NEXT: vsetvli a3, zero, e16,mf2,ta ,mu
87
87
; CHECK-NEXT: vle16.v v25, (a1)
88
- ; CHECK-NEXT: vsetvli a1, zero, e16,mf2,tu ,mu
88
+ ; CHECK-NEXT: vsetvli a1, zero, e16,mf2,ta ,mu
89
89
; CHECK-NEXT: vle16.v v26, (a2)
90
- ; CHECK-NEXT: vsetvli a1, zero, e16,mf2,tu ,mu
90
+ ; CHECK-NEXT: vsetvli a1, zero, e16,mf2,ta ,mu
91
91
; CHECK-NEXT: vadd.vv v25, v25, v26
92
- ; CHECK-NEXT: vsetvli a1, zero, e16,mf2,tu ,mu
92
+ ; CHECK-NEXT: vsetvli a1, zero, e16,mf2,ta ,mu
93
93
; CHECK-NEXT: vse16.v v25, (a0)
94
94
; CHECK-NEXT: ret
95
95
%va = load <vscale x 2 x i16 >, <vscale x 2 x i16 >* %pa
@@ -102,13 +102,13 @@ define void @vadd_vint16mf2(<vscale x 2 x i16> *%pc, <vscale x 2 x i16> *%pa, <v
102
102
define void @vadd_vint16mf4 (<vscale x 1 x i16 > *%pc , <vscale x 1 x i16 > *%pa , <vscale x 1 x i16 > *%pb ) nounwind {
103
103
; CHECK-LABEL: vadd_vint16mf4:
104
104
; CHECK: # %bb.0:
105
- ; CHECK-NEXT: vsetvli a3, zero, e16,mf4,tu ,mu
105
+ ; CHECK-NEXT: vsetvli a3, zero, e16,mf4,ta ,mu
106
106
; CHECK-NEXT: vle16.v v25, (a1)
107
- ; CHECK-NEXT: vsetvli a1, zero, e16,mf4,tu ,mu
107
+ ; CHECK-NEXT: vsetvli a1, zero, e16,mf4,ta ,mu
108
108
; CHECK-NEXT: vle16.v v26, (a2)
109
- ; CHECK-NEXT: vsetvli a1, zero, e16,mf4,tu ,mu
109
+ ; CHECK-NEXT: vsetvli a1, zero, e16,mf4,ta ,mu
110
110
; CHECK-NEXT: vadd.vv v25, v25, v26
111
- ; CHECK-NEXT: vsetvli a1, zero, e16,mf4,tu ,mu
111
+ ; CHECK-NEXT: vsetvli a1, zero, e16,mf4,ta ,mu
112
112
; CHECK-NEXT: vse16.v v25, (a0)
113
113
; CHECK-NEXT: ret
114
114
%va = load <vscale x 1 x i16 >, <vscale x 1 x i16 >* %pa
0 commit comments