@@ -1231,8 +1231,6 @@ static std::string format_size(size_t size) {
1231
1231
return oss.str();
1232
1232
}
1233
1233
1234
- static std::mutex log_mutex;
1235
-
1236
1234
class vk_memory_logger {
1237
1235
public:
1238
1236
vk_memory_logger(): total_device(0), total_host(0) {}
@@ -1422,6 +1420,8 @@ struct ggml_backend_vk_buffer_context {
1422
1420
};
1423
1421
1424
1422
#ifdef GGML_VULKAN_MEMORY_DEBUG
1423
+ static std::mutex log_mutex;
1424
+
1425
1425
void vk_memory_logger::log_allocation(vk_buffer_ref buf_ref, size_t size) {
1426
1426
std::lock_guard<std::mutex> guard(log_mutex);
1427
1427
vk_buffer buf = buf_ref.lock();
@@ -13138,16 +13138,16 @@ static void ggml_vk_check_results_0(ggml_backend_vk_context * ctx, ggml_cgraph *
13138
13138
} else if (tensor->op == GGML_OP_IM2COL_3D) {
13139
13139
const int32_t s0 = tensor->op_params[0];
13140
13140
const int32_t s1 = tensor->op_params[1];
13141
- const int32_t s1 = tensor->op_params[2];
13141
+ const int32_t s2 = tensor->op_params[2];
13142
13142
const int32_t p0 = tensor->op_params[3];
13143
13143
const int32_t p1 = tensor->op_params[4];
13144
- const int32_t p1 = tensor->op_params[5];
13144
+ const int32_t p2 = tensor->op_params[5];
13145
13145
const int32_t d0 = tensor->op_params[6];
13146
13146
const int32_t d1 = tensor->op_params[7];
13147
- const int32_t d1 = tensor->op_params[8];
13147
+ const int32_t d2 = tensor->op_params[8];
13148
13148
const int32_t IC = tensor->op_params[9];
13149
13149
13150
- tensor_clone = ggml_im2col (ggml_ctx, src_clone[0], src_clone[1], IC, s0, s1, s2, p0, p1, p2, d0, d1, d2, tensor->type);
13150
+ tensor_clone = ggml_im2col_3d (ggml_ctx, src_clone[0], src_clone[1], IC, s0, s1, s2, p0, p1, p2, d0, d1, d2, tensor->type);
13151
13151
} else if (tensor->op == GGML_OP_TIMESTEP_EMBEDDING) {
13152
13152
const int32_t dim = tensor->op_params[0];
13153
13153
const int32_t max_period = tensor->op_params[1];
0 commit comments