Logo
Explore Help
Sign In
tqcq/llama.cpp
0
0
Fork 0
You've already forked llama.cpp
mirror of https://github.com/ggml-org/llama.cpp.git synced 2025-07-31 22:53:52 -04:00
Code Issues Packages Projects Releases Wiki Activity
Files
9065ca71a29009b7a090879e6dbe4117c9632862
llama.cpp/tools/server/tests/unit
History
Olivier Chafik f13847cfb5 server: fix regression on streamed non-chat completion w/ stops (#13785)
* more forgiving message diffs: partial stop words aren't erased, full stops are

* Add (slow) server test for completion + stream + stop
2025-05-26 14:16:37 +01:00
..
test_basic.py
…
test_chat_completion.py
server: streaming of tool calls and thoughts when --jinja is on (#12379)
2025-05-25 01:48:08 +01:00
test_completion.py
server: fix regression on streamed non-chat completion w/ stops (#13785)
2025-05-26 14:16:37 +01:00
test_ctx_shift.py
server : do not return error out of context (with ctx shift disabled) (#13577)
2025-05-16 21:50:00 +02:00
test_embedding.py
…
test_infill.py
…
test_lora.py
…
test_rerank.py
…
test_security.py
…
test_slot_save.py
…
test_speculative.py
…
test_template.py
server: add --reasoning-budget 0 to disable thinking (incl. qwen3 w/ enable_thinking:false) (#13771)
2025-05-26 00:30:51 +01:00
test_tokenize.py
…
test_tool_call.py
server: streaming of tool calls and thoughts when --jinja is on (#12379)
2025-05-25 01:48:08 +01:00
test_vision_api.py
server : support audio input (#13714)
2025-05-23 11:03:47 +02:00
Powered by Gitea Version: 1.24.3 Page: 1687ms Template: 21ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API