@@ -94,7 +94,7 @@ void test_column_major_load_with_stride_template_double(double *Ptr) {
94
94
95
95
// CHECK-LABEL: define linkonce_odr <40 x double> @_Z29column_major_load_with_strideIdLj10ELj4ELj15EEu11matrix_typeIXT0_EXT1_ET_EPS0_(double* %Ptr)
96
96
// CHECK: [[PTR:%.*]] = load double*, double** %Ptr.addr, align 8
97
- // CHECK-NEXT: call <40 x double> @llvm.matrix.column.major.load.v40f64(double* align 8 [[PTR]], i64 15, i1 false, i32 10, i32 4)
97
+ // CHECK-NEXT: call <40 x double> @llvm.matrix.column.major.load.v40f64.i64 (double* align 8 [[PTR]], i64 15, i1 false, i32 10, i32 4)
98
98
99
99
matrix_t <double , 10 , 4 > M1 = column_major_load_with_stride<double , 10 , 4 , 15 >(Ptr);
100
100
}
@@ -106,7 +106,7 @@ void test_column_major_load_with_stride_template_int(int *Ptr) {
106
106
107
107
// CHECK-LABEL: define linkonce_odr <6 x i32> @_Z29column_major_load_with_strideIiLj3ELj2ELj12EEu11matrix_typeIXT0_EXT1_ET_EPS0_(i32* %Ptr)
108
108
// CHECK: [[PTR:%.*]] = load i32*, i32** %Ptr.addr, align 8
109
- // CHECK-NEXT: call <6 x i32> @llvm.matrix.column.major.load.v6i32(i32* align 4 [[PTR]], i64 12, i1 false, i32 3, i32 2)
109
+ // CHECK-NEXT: call <6 x i32> @llvm.matrix.column.major.load.v6i32.i64 (i32* align 4 [[PTR]], i64 12, i1 false, i32 3, i32 2)
110
110
111
111
matrix_t <int , 3 , 2 > M1 = column_major_load_with_stride<int , 3 , 2 , 12 >(Ptr);
112
112
}
@@ -124,7 +124,7 @@ void test_column_major_load_stride_wrapper(int *Ptr, UnsignedWrapper &W) {
124
124
// CHECK-NEXT: [[STRIDE:%.*]] = call i32 @_ZN15UnsignedWrappercvjEv(%struct.UnsignedWrapper* {{[^,]*}} [[W]])
125
125
// CHECK-NEXT: [[STRIDE_EXT:%.*]] = zext i32 [[STRIDE]] to i64
126
126
// CHECK-NEXT: [[PTR:%.*]] = load i32*, i32** %Ptr.addr, align 8
127
- // CHECK-NEXT: call <4 x i32> @llvm.matrix.column.major.load.v4i32(i32* align 4 [[PTR]], i64 [[STRIDE_EXT]], i1 false, i32 2, i32 2)
127
+ // CHECK-NEXT: call <4 x i32> @llvm.matrix.column.major.load.v4i32.i64 (i32* align 4 [[PTR]], i64 [[STRIDE_EXT]], i1 false, i32 2, i32 2)
128
128
matrix_t <int , 2 , 2 > M1 = __builtin_matrix_column_major_load (Ptr, 2 , 2 , W);
129
129
}
130
130
@@ -133,7 +133,7 @@ constexpr int constexpr3() { return 3; }
133
133
void test_column_major_load_constexpr_num_rows (int *Ptr) {
134
134
// CHECK-LABEL: define{{.*}} void @_Z41test_column_major_load_constexpr_num_rowsPi(i32* %Ptr)
135
135
// CHECK: [[PTR:%.*]] = load i32*, i32** %Ptr.addr, align 8
136
- // CHECK-NEXT: call <6 x i32> @llvm.matrix.column.major.load.v6i32(i32* align 4 [[PTR]], i64 3, i1 false, i32 3, i32 2)
136
+ // CHECK-NEXT: call <6 x i32> @llvm.matrix.column.major.load.v6i32.i64 (i32* align 4 [[PTR]], i64 3, i1 false, i32 3, i32 2)
137
137
138
138
matrix_t <int , 3 , 2 > M1 = __builtin_matrix_column_major_load (Ptr, constexpr3 (), 2 , 3 );
139
139
}
@@ -143,7 +143,7 @@ constexpr int constexpr1() { return 1; }
143
143
void test_column_major_load_constexpr_num_columns (int *Ptr) {
144
144
// CHECK-LABEL: define{{.*}} void @_Z44test_column_major_load_constexpr_num_columnsPi(i32* %Ptr)
145
145
// CHECK: [[PTR:%.*]] = load i32*, i32** %Ptr.addr, align 8
146
- // CHECK-NEXT: call <2 x i32> @llvm.matrix.column.major.load.v2i32(i32* align 4 [[PTR]], i64 3, i1 false, i32 2, i32 1)
146
+ // CHECK-NEXT: call <2 x i32> @llvm.matrix.column.major.load.v2i32.i64 (i32* align 4 [[PTR]], i64 3, i1 false, i32 2, i32 1)
147
147
matrix_t <int , 2 , 1 > M1 = __builtin_matrix_column_major_load (Ptr, 2 , constexpr1 (), 3 );
148
148
}
149
149
@@ -153,7 +153,7 @@ constexpr int constexpr_plus1() { return N + 1; }
153
153
void test_column_major_load_constexpr_num_columns_temp (int *Ptr) {
154
154
// CHECK-LABEL: define{{.*}} void @_Z49test_column_major_load_constexpr_num_columns_tempPi(i32* %Ptr)
155
155
// CHECK: [[PTR:%.*]] = load i32*, i32** %Ptr.addr, align 8
156
- // CHECK-NEXT: call <10 x i32> @llvm.matrix.column.major.load.v10i32(i32* align 4 [[PTR]], i64 3, i1 false, i32 2, i32 5)
156
+ // CHECK-NEXT: call <10 x i32> @llvm.matrix.column.major.load.v10i32.i64 (i32* align 4 [[PTR]], i64 3, i1 false, i32 2, i32 5)
157
157
matrix_t <int , 2 , 5 > M1 = __builtin_matrix_column_major_load (Ptr, 2 , constexpr_plus1<4 >(), 3 );
158
158
}
159
159
@@ -162,7 +162,7 @@ void test_column_major_load_constexpr_stride_constexpr(int *Ptr) {
162
162
// CHECK: [[STRIDE:%.*]] = call i32 @_Z10constexpr3v()
163
163
// CHECK-NEXT: [[STRIDE_EXT:%.*]] = sext i32 [[STRIDE]] to i64
164
164
// CHECK-NEXT: [[PTR:%.*]] = load i32*, i32** %Ptr.addr, align 8
165
- // CHECK-NEXT: call <4 x i32> @llvm.matrix.column.major.load.v4i32(i32* align 4 [[PTR]], i64 [[STRIDE_EXT]], i1 false, i32 2, i32 2)
165
+ // CHECK-NEXT: call <4 x i32> @llvm.matrix.column.major.load.v4i32.i64 (i32* align 4 [[PTR]], i64 [[STRIDE_EXT]], i1 false, i32 2, i32 2)
166
166
167
167
matrix_t <int , 2 , 2 > M1 = __builtin_matrix_column_major_load (Ptr, 2 , 2 , constexpr3 ());
168
168
}
@@ -200,7 +200,7 @@ void test_column_major_store_with_stride_template_double(double *Ptr) {
200
200
// CHECK-LABEL: define linkonce_odr void @_Z30column_major_store_with_strideIdLj10ELj4ELj15EEvRu11matrix_typeIXT0_EXT1_ET_EPS0_([40 x double]* nonnull align 8 dereferenceable(320) %m, double* %Ptr)
201
201
// CHECK: [[M:%.*]] = load <40 x double>, <40 x double>* {{.*}}, align 8
202
202
// CHECK-NEXT: [[PTR:%.*]] = load double*, double** %Ptr.addr, align 8
203
- // CHECK-NEXT: call void @llvm.matrix.column.major.store.v40f64(<40 x double> [[M]], double* align 8 [[PTR]], i64 15, i1 false, i32 10, i32 4)
203
+ // CHECK-NEXT: call void @llvm.matrix.column.major.store.v40f64.i64 (<40 x double> [[M]], double* align 8 [[PTR]], i64 15, i1 false, i32 10, i32 4)
204
204
205
205
matrix_t <double , 10 , 4 > M1;
206
206
column_major_store_with_stride<double , 10 , 4 , 15 >(M1, Ptr);
@@ -214,7 +214,7 @@ void test_column_major_store_with_stride_template_int(int *Ptr) {
214
214
// CHECK-LABEL: define linkonce_odr void @_Z30column_major_store_with_strideIiLj3ELj2ELj3EEvRu11matrix_typeIXT0_EXT1_ET_EPS0_([6 x i32]* nonnull align 4 dereferenceable(24) %m, i32* %Ptr)
215
215
// CHECK: [[M:%.*]] = load <6 x i32>, <6 x i32>* {{.*}}, align 4
216
216
// CHECK-NEXT: [[PTR:%.*]] = load i32*, i32** %Ptr.addr, align 8
217
- // CHECK-NEXT: call void @llvm.matrix.column.major.store.v6i32(<6 x i32> [[M]], i32* align 4 [[PTR]], i64 3, i1 false, i32 3, i32 2)
217
+ // CHECK-NEXT: call void @llvm.matrix.column.major.store.v6i32.i64 (<6 x i32> [[M]], i32* align 4 [[PTR]], i64 3, i1 false, i32 3, i32 2)
218
218
219
219
matrix_t <int , 3 , 2 > M1;
220
220
column_major_store_with_stride<int , 3 , 2 , 3 >(M1, Ptr);
@@ -227,7 +227,7 @@ void test_column_major_store_stride_wrapper(int *Ptr, UnsignedWrapper &W) {
227
227
// CHECK-NEXT: [[W:%.*]] = load %struct.UnsignedWrapper*, %struct.UnsignedWrapper** %W.addr, align 8
228
228
// CHECK-NEXT: [[IDX:%.*]] = call i32 @_ZN15UnsignedWrappercvjEv(%struct.UnsignedWrapper* {{[^,]*}} [[W]])
229
229
// CHECK-NEXT: [[IDX_EXT:%.*]] = zext i32 [[IDX]] to i64
230
- // CHECK-NEXT: call void @llvm.matrix.column.major.store.v4i32(<4 x i32> [[M]], i32* align 4 [[PTR]], i64 [[IDX_EXT]], i1 false, i32 2, i32 2)
230
+ // CHECK-NEXT: call void @llvm.matrix.column.major.store.v4i32.i64 (<4 x i32> [[M]], i32* align 4 [[PTR]], i64 [[IDX_EXT]], i1 false, i32 2, i32 2)
231
231
232
232
matrix_t <int , 2 , 2 > M1;
233
233
__builtin_matrix_column_major_store (M1, Ptr, W);
@@ -239,7 +239,7 @@ void test_column_major_store_constexpr_stride_constexpr(int *Ptr) {
239
239
// CHECK-NEXT: [[PTR:%.*]] = load i32*, i32** %Ptr.addr, align 8
240
240
// CHECK-NEXT: [[IDX:%.*]] = call i32 @_Z10constexpr3v()
241
241
// CHECK-NEXT: [[IDX_EXT:%.*]] = sext i32 [[IDX]] to i64
242
- // CHECK-NEXT: call void @llvm.matrix.column.major.store.v4i32(<4 x i32> [[M]], i32* align 4 [[PTR]], i64 [[IDX_EXT]], i1 false, i32 2, i32 2)
242
+ // CHECK-NEXT: call void @llvm.matrix.column.major.store.v4i32.i64 (<4 x i32> [[M]], i32* align 4 [[PTR]], i64 [[IDX_EXT]], i1 false, i32 2, i32 2)
243
243
244
244
matrix_t <int , 2 , 2 > M;
245
245
__builtin_matrix_column_major_store (M, Ptr, constexpr3 ());
0 commit comments