Files
llama.cpp/examples
Pierrick Hymbert 2f0ee84b9b server: bench: minor fixes (#10765)
* server/bench:
- support openAI streaming standard output with [DONE]\n\n
- export k6 raw results in csv
- fix too many tcp idle connection in tcp_wait
- add metric time to emit first token

* server/bench:
- fix when prometheus not started
- wait for server to be ready before starting bench
2025-01-02 18:06:12 +01:00
..
2024-12-02 21:22:53 +02:00
2024-12-19 18:47:15 +02:00
2024-12-11 16:16:32 +01:00
2024-12-18 19:27:21 +02:00
2023-03-29 20:21:09 +03:00