@@ -148,6 +148,8 @@ void ggml_cuda_op_gelu(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
148148 float * dst_d = (float *)dst->data ;
149149 cudaStream_t stream = ctx.stream ();
150150
151+ GGML_ASSERT (ggml_is_contiguous (src0));
152+
151153 GGML_ASSERT (src0->type == GGML_TYPE_F32);
152154 GGML_ASSERT ( dst->type == GGML_TYPE_F32);
153155
@@ -160,6 +162,8 @@ void ggml_cuda_op_silu(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
160162 float * dst_d = (float *)dst->data ;
161163 cudaStream_t stream = ctx.stream ();
162164
165+ GGML_ASSERT (ggml_is_contiguous (src0));
166+
163167 GGML_ASSERT (src0->type == GGML_TYPE_F32);
164168 GGML_ASSERT ( dst->type == GGML_TYPE_F32);
165169
@@ -172,6 +176,8 @@ void ggml_cuda_op_gelu_quick(ggml_backend_cuda_context & ctx, ggml_tensor * dst)
172176 float * dst_d = (float *)dst->data ;
173177 cudaStream_t stream = ctx.stream ();
174178
179+ GGML_ASSERT (ggml_is_contiguous (src0));
180+
175181 GGML_ASSERT (src0->type == GGML_TYPE_F32);
176182 GGML_ASSERT ( dst->type == GGML_TYPE_F32);
177183
@@ -184,6 +190,8 @@ void ggml_cuda_op_tanh(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
184190 float * dst_d = (float *)dst->data ;
185191 cudaStream_t stream = ctx.stream ();
186192
193+ GGML_ASSERT (ggml_is_contiguous (src0));
194+
187195 GGML_ASSERT (src0->type == GGML_TYPE_F32);
188196 GGML_ASSERT ( dst->type == GGML_TYPE_F32);
189197
@@ -196,6 +204,8 @@ void ggml_cuda_op_relu(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
196204 float * dst_d = (float *)dst->data ;
197205 cudaStream_t stream = ctx.stream ();
198206
207+ GGML_ASSERT (ggml_is_contiguous (src0));
208+
199209 GGML_ASSERT (src0->type == GGML_TYPE_F32);
200210 GGML_ASSERT ( dst->type == GGML_TYPE_F32);
201211
@@ -208,6 +218,8 @@ void ggml_cuda_op_sigmoid(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
208218 float * dst_d = (float *)dst->data ;
209219 cudaStream_t stream = ctx.stream ();
210220
221+ GGML_ASSERT (ggml_is_contiguous (src0));
222+
211223 GGML_ASSERT (src0->type == GGML_TYPE_F32);
212224 GGML_ASSERT ( dst->type == GGML_TYPE_F32);
213225
@@ -220,6 +232,8 @@ void ggml_cuda_op_hardsigmoid(ggml_backend_cuda_context & ctx, ggml_tensor * dst
220232 float * dst_d = (float *)dst->data ;
221233 cudaStream_t stream = ctx.stream ();
222234
235+ GGML_ASSERT (ggml_is_contiguous (src0));
236+
223237 GGML_ASSERT (src0->type == GGML_TYPE_F32);
224238 GGML_ASSERT ( dst->type == GGML_TYPE_F32);
225239
@@ -232,6 +246,8 @@ void ggml_cuda_op_hardswish(ggml_backend_cuda_context & ctx, ggml_tensor * dst)
232246 float * dst_d = (float *)dst->data ;
233247 cudaStream_t stream = ctx.stream ();
234248
249+ GGML_ASSERT (ggml_is_contiguous (src0));
250+
235251 GGML_ASSERT (src0->type == GGML_TYPE_F32);
236252 GGML_ASSERT ( dst->type == GGML_TYPE_F32);
237253
@@ -244,6 +260,8 @@ void ggml_cuda_op_leaky_relu(ggml_backend_cuda_context & ctx, ggml_tensor * dst)
244260 float * dst_d = (float *)dst->data ;
245261 cudaStream_t stream = ctx.stream ();
246262
263+ GGML_ASSERT (ggml_is_contiguous (src0));
264+
247265 GGML_ASSERT (src0->type == GGML_TYPE_F32);
248266 GGML_ASSERT ( dst->type == GGML_TYPE_F32);
249267
@@ -259,6 +277,8 @@ void ggml_cuda_op_sqr(ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
259277 float * dst_d = (float *)dst->data ;
260278 cudaStream_t stream = ctx.stream ();
261279
280+ GGML_ASSERT (ggml_is_contiguous (src0));
281+
262282 GGML_ASSERT (src0->type == GGML_TYPE_F32);
263283 GGML_ASSERT ( dst->type == GGML_TYPE_F32);
264284
0 commit comments