mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-07-10 05:20:26 +00:00
kv-cache : fix handling when find_slot fails
ggml-ci
This commit is contained in:
@ -573,7 +573,7 @@ llama_kv_cache_unified::slot_info_vec_t llama_kv_cache_unified::prepare(const st
|
||||
apply_ubatch(sinfo_new, ubatch);
|
||||
}
|
||||
|
||||
GGML_ASSERT(!states.empty());
|
||||
GGML_ASSERT(!states.empty() || !success);
|
||||
|
||||
// iterate backwards and restore the cells to their original state
|
||||
for (auto it = states.rbegin(); it != states.rend(); ++it) {
|
||||
|
Reference in New Issue
Block a user