mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-07-12 06:09:18 +00:00
kv-cache : fix handling when find_slot fails
ggml-ci
This commit is contained in:
@ -573,7 +573,7 @@ llama_kv_cache_unified::slot_info_vec_t llama_kv_cache_unified::prepare(const st
|
|||||||
apply_ubatch(sinfo_new, ubatch);
|
apply_ubatch(sinfo_new, ubatch);
|
||||||
}
|
}
|
||||||
|
|
||||||
GGML_ASSERT(!states.empty());
|
GGML_ASSERT(!states.empty() || !success);
|
||||||
|
|
||||||
// iterate backwards and restore the cells to their original state
|
// iterate backwards and restore the cells to their original state
|
||||||
for (auto it = states.rbegin(); it != states.rend(); ++it) {
|
for (auto it = states.rbegin(); it != states.rend(); ++it) {
|
||||||
|
Reference in New Issue
Block a user