mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-08-25 09:38:35 -04:00
speculative : update default params (#11954)
* speculative : update default params * speculative : do not discard the last drafted token
This commit is contained in:
@@ -9,7 +9,7 @@ struct common_speculative_params {
|
||||
int n_draft = 16; // max drafted tokens
|
||||
int n_reuse = 256;
|
||||
|
||||
float p_min = 0.9f; // min probability required to accept a token in the draft
|
||||
float p_min = 0.75f; // min probability required to accept a token in the draft
|
||||
};
|
||||
|
||||
struct common_speculative * common_speculative_init(struct llama_context * ctx_dft);
|
||||
|
Reference in New Issue
Block a user