@@ -9539,7 +9539,7 @@ struct ggml_sycl_pool_leg : public ggml_sycl_pool {
95399539 }
95409540 void * ptr;
95419541 size_t look_ahead_size = (size_t) (1.05 * size);
9542-
9542+
95439543 SYCL_CHECK(
95449544 CHECK_TRY_ERROR(ptr = (void *)sycl::malloc_device(
95459545 look_ahead_size, *qptr)));
@@ -10335,7 +10335,7 @@ inline void ggml_sycl_op_dequantize_mul_mat_vec(
1033510335}
1033610336
1033710337inline void ggml_sycl_op_mul_mat_sycl(
10338- ggml_backend_sycl_context & ctx,
10338+ ggml_backend_sycl_context & ctx,
1033910339 const ggml_tensor *src0, const ggml_tensor *src1, ggml_tensor *dst,
1034010340 const char *src0_dd_i, const float *src1_ddf_i, const char *src1_ddq_i,
1034110341 float *dst_dd_i, const int64_t row_low, const int64_t row_high,
@@ -11426,7 +11426,7 @@ static void ggml_sycl_mul_mat_vec_nc(ggml_backend_sycl_context & ctx, const ggml
1142611426
1142711427 SYCL_CHECK(ggml_sycl_set_device(ctx.device));
1142811428 queue_ptr main_stream = ctx.stream();
11429-
11429+
1143011430 void * src0_ddq = src0->data;
1143111431 float * src1_ddf = (float *) src1->data;
1143211432 float * dst_ddf = (float *) dst->data;
@@ -12741,7 +12741,7 @@ ggml_backend_sycl_split_buffer_init_tensor(ggml_backend_buffer_t buffer,
1274112741 // FIXME: do not crash if cudaMalloc fails
1274212742 // currently, init_tensor cannot fail, it needs to be fixed in ggml-backend first
1274312743 ggml_sycl_set_device(i);
12744- const queue_ptr stream = ctx->streams[i];
12744+ const queue_ptr stream = ctx->streams[i];
1274512745 char * buf;
1274612746 /*
1274712747 DPCT1009:208: SYCL uses exceptions to report errors and does not use the
@@ -12878,7 +12878,7 @@ ggml_backend_sycl_split_buffer_get_tensor(ggml_backend_buffer_t buffer,
1287812878 was inserted. You need to rewrite this code.
1287912879 */
1288012880 ggml_sycl_set_device(i);
12881- const queue_ptr stream = ctx->streams[i];
12881+ const queue_ptr stream = ctx->streams[i];
1288212882 SYCL_CHECK(CHECK_TRY_ERROR(
1288312883 (*stream)
1288412884 .memcpy(buf_host, extra->data_device[i], original_size)
@@ -12981,7 +12981,7 @@ static ggml_backend_buffer_type_i ggml_backend_sycl_split_buffer_type_interface
1298112981GGML_CALL ggml_backend_buffer_type_t ggml_backend_sycl_split_buffer_type(const float * tensor_split) {
1298212982 static std::mutex mutex;
1298312983 std::lock_guard<std::mutex> lock(mutex);
12984-
12984+
1298512985 GGML_SYCL_DEBUG("[SYCL] call ggml_backend_sycl_split_buffer_type\n");
1298612986 ggml_check_sycl();
1298712987 // FIXME: this is not thread safe
0 commit comments