7b00429295
llama : use "stream" vs "virtual sequence"
...
ggml-ci
2025-07-04 13:53:15 +03:00
38479e2642
llama : add "virtual sequences"
...
ggml-ci
2025-07-04 13:36:28 +03:00
ab8443d0ea
batched-bench : fix oob write
...
ggml-ci
2025-07-04 13:36:19 +03:00
745aa5319b
llama : deprecate llama_kv_self_ API ( #14030 )
...
* llama : deprecate llama_kv_self_ API
ggml-ci
* llama : allow llama_memory_(nullptr)
ggml-ci
* memory : add flag for optional data clear in llama_memory_clear
ggml-ci
2025-06-06 14:11:15 +03:00
b89d605a91
batched-bench : fix pp batch contents ( #13492 )
2025-05-13 18:01:53 +03:00
1d36b3670b
llama : move end-user examples to tools directory ( #13249 )
...
* llama : move end-user examples to tools directory
---------
Co-authored-by: Xuan Son Nguyen <son@huggingface.co >
2025-05-02 20:27:13 +02:00