Files
llama.cpp/tools
Georgi Gerganov 3600cc2886 llama : use n_swa + n_ubatch cells for SWA cache (#13833)
* llama : use n_swa + n_ubatch cells for SWA cache

ggml-ci

* llama : add warning about multi-sqeuence SWA contexts
2025-05-31 15:57:44 +03:00
..
2025-05-25 15:35:53 +03:00
2025-05-30 16:25:45 +03:00
2025-05-30 16:25:45 +03:00