@@ -148,6 +148,8 @@ void ggml_cuda_op_gelu(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
148
148
float * dst_d = (float *)dst->data ;
149
149
cudaStream_t stream = ctx.stream ();
150
150
151
+ GGML_ASSERT (ggml_is_contiguous (src0));
152
+
151
153
GGML_ASSERT (src0->type == GGML_TYPE_F32);
152
154
GGML_ASSERT ( dst->type == GGML_TYPE_F32);
153
155
@@ -160,6 +162,8 @@ void ggml_cuda_op_silu(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
160
162
float * dst_d = (float *)dst->data ;
161
163
cudaStream_t stream = ctx.stream ();
162
164
165
+ GGML_ASSERT (ggml_is_contiguous (src0));
166
+
163
167
GGML_ASSERT (src0->type == GGML_TYPE_F32);
164
168
GGML_ASSERT ( dst->type == GGML_TYPE_F32);
165
169
@@ -172,6 +176,8 @@ void ggml_cuda_op_gelu_quick(ggml_backend_cuda_context & ctx, ggml_tensor * dst)
172
176
float * dst_d = (float *)dst->data ;
173
177
cudaStream_t stream = ctx.stream ();
174
178
179
+ GGML_ASSERT (ggml_is_contiguous (src0));
180
+
175
181
GGML_ASSERT (src0->type == GGML_TYPE_F32);
176
182
GGML_ASSERT ( dst->type == GGML_TYPE_F32);
177
183
@@ -184,6 +190,8 @@ void ggml_cuda_op_tanh(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
184
190
float * dst_d = (float *)dst->data ;
185
191
cudaStream_t stream = ctx.stream ();
186
192
193
+ GGML_ASSERT (ggml_is_contiguous (src0));
194
+
187
195
GGML_ASSERT (src0->type == GGML_TYPE_F32);
188
196
GGML_ASSERT ( dst->type == GGML_TYPE_F32);
189
197
@@ -196,6 +204,8 @@ void ggml_cuda_op_relu(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
196
204
float * dst_d = (float *)dst->data ;
197
205
cudaStream_t stream = ctx.stream ();
198
206
207
+ GGML_ASSERT (ggml_is_contiguous (src0));
208
+
199
209
GGML_ASSERT (src0->type == GGML_TYPE_F32);
200
210
GGML_ASSERT ( dst->type == GGML_TYPE_F32);
201
211
@@ -208,6 +218,8 @@ void ggml_cuda_op_sigmoid(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
208
218
float * dst_d = (float *)dst->data ;
209
219
cudaStream_t stream = ctx.stream ();
210
220
221
+ GGML_ASSERT (ggml_is_contiguous (src0));
222
+
211
223
GGML_ASSERT (src0->type == GGML_TYPE_F32);
212
224
GGML_ASSERT ( dst->type == GGML_TYPE_F32);
213
225
@@ -220,6 +232,8 @@ void ggml_cuda_op_hardsigmoid(ggml_backend_cuda_context & ctx, ggml_tensor * dst
220
232
float * dst_d = (float *)dst->data ;
221
233
cudaStream_t stream = ctx.stream ();
222
234
235
+ GGML_ASSERT (ggml_is_contiguous (src0));
236
+
223
237
GGML_ASSERT (src0->type == GGML_TYPE_F32);
224
238
GGML_ASSERT ( dst->type == GGML_TYPE_F32);
225
239
@@ -232,6 +246,8 @@ void ggml_cuda_op_hardswish(ggml_backend_cuda_context & ctx, ggml_tensor * dst)
232
246
float * dst_d = (float *)dst->data ;
233
247
cudaStream_t stream = ctx.stream ();
234
248
249
+ GGML_ASSERT (ggml_is_contiguous (src0));
250
+
235
251
GGML_ASSERT (src0->type == GGML_TYPE_F32);
236
252
GGML_ASSERT ( dst->type == GGML_TYPE_F32);
237
253
@@ -244,6 +260,8 @@ void ggml_cuda_op_leaky_relu(ggml_backend_cuda_context & ctx, ggml_tensor * dst)
244
260
float * dst_d = (float *)dst->data ;
245
261
cudaStream_t stream = ctx.stream ();
246
262
263
+ GGML_ASSERT (ggml_is_contiguous (src0));
264
+
247
265
GGML_ASSERT (src0->type == GGML_TYPE_F32);
248
266
GGML_ASSERT ( dst->type == GGML_TYPE_F32);
249
267
@@ -259,6 +277,8 @@ void ggml_cuda_op_sqr(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
259
277
float * dst_d = (float *)dst->data ;
260
278
cudaStream_t stream = ctx.stream ();
261
279
280
+ GGML_ASSERT (ggml_is_contiguous (src0));
281
+
262
282
GGML_ASSERT (src0->type == GGML_TYPE_F32);
263
283
GGML_ASSERT ( dst->type == GGML_TYPE_F32);
264
284
0 commit comments