mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-08-15 20:53:00 -04:00
export separate lora ggufs instead
This commit is contained in:
@@ -588,6 +588,12 @@ extern "C" {
|
||||
struct llama_model * model,
|
||||
const char * path_lora);
|
||||
|
||||
// Get the LoRA task name. Returns a blank string if not applicable
|
||||
LLAMA_API const char * llama_adapter_lora_task_name(struct llama_adapter_lora * adapter);
|
||||
|
||||
// Get the required LoRA prompt prefix. Returns a blank string if not applicable
|
||||
LLAMA_API const char * llama_adapter_lora_prompt_prefix(struct llama_adapter_lora * adapter);
|
||||
|
||||
// Manually free a LoRA adapter
|
||||
// Note: loaded adapters will be free when the associated model is deleted
|
||||
LLAMA_API void llama_adapter_lora_free(struct llama_adapter_lora * adapter);
|
||||
|
Reference in New Issue
Block a user