From 4e73b81a67fd76e0ef8cc9e4ee6d8d7b7c7c4fca Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Johannes=20G=C3=A4=C3=9Fler?= Date: Mon, 27 Jan 2025 18:33:34 +0100 Subject: [PATCH] try CI fix --- include/llama.h | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/include/llama.h b/include/llama.h index cdb8a768e..508f94e43 100644 --- a/include/llama.h +++ b/include/llama.h @@ -1441,7 +1441,7 @@ extern "C" { typedef bool (*llama_opt_param_filter)(const struct ggml_tensor * tensor, void * userdata); // always returns true - bool llama_opt_param_filter_all(const struct ggml_tensor * tensor, void * userdata); + LLAMA_API bool llama_opt_param_filter_all(const struct ggml_tensor * tensor, void * userdata); struct llama_opt_params { uint32_t n_ctx_train; // assumed context size post training, use context size specified in llama_context if 0