Files
llama.cpp/examples
Georgi Gerganov 8c70a5ff25 batched : add bench tool (#3545)
* batched : add bench tool

* batched : minor fix table

* batched-bench : add readme + n_kv_max is now configurable

* batched-bench : init warm-up batch

* batched-bench : pass custom set of PP, TG and PL

* batched-bench : add mmq CLI arg
2023-10-11 21:25:33 +03:00
..
2023-10-11 21:25:33 +03:00
2023-09-28 21:40:11 +03:00
2023-10-10 10:31:21 +03:00
2023-10-10 10:31:21 +03:00
2023-08-30 09:29:32 +03:00