mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-06-30 04:45:17 +00:00
context : do not clear output buffer on reserve (#13152)
Co-authored-by: pockers21 <liyang2@uniontech.com>
This commit is contained in:
@ -1536,8 +1536,6 @@ int32_t llama_context::output_reserve(int32_t n_outputs) {
|
|||||||
// set all ids as invalid (negative)
|
// set all ids as invalid (negative)
|
||||||
std::fill(output_ids.begin(), output_ids.end(), -1);
|
std::fill(output_ids.begin(), output_ids.end(), -1);
|
||||||
|
|
||||||
ggml_backend_buffer_clear(buf_output.get(), 0);
|
|
||||||
|
|
||||||
this->n_outputs = 0;
|
this->n_outputs = 0;
|
||||||
this->n_outputs_max = n_outputs_max;
|
this->n_outputs_max = n_outputs_max;
|
||||||
|
|
||||||
|
Reference in New Issue
Block a user