This website requires JavaScript.
Explore
Help
Sign In
tqcq
/
llama.cpp
Watch
0
Star
0
Fork
0
You've already forked llama.cpp
mirror of
https://github.com/ggml-org/llama.cpp.git
synced
2025-06-30 12:55:17 +00:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
Files
67fd33132fab93e6c2087bd6fa656a8a57419efa
llama.cpp
/
examples
/
server
/
tests
/
features
History
Pierrick Hymbert
e3965cf35a
server: tests - slow inference causes timeout on the CI (
#5715
)
...
* server: tests - longer inference timeout for CI
2024-02-25 22:48:33 +01:00
..
steps
server: tests - slow inference causes timeout on the CI (
#5715
)
2024-02-25 22:48:33 +01:00
environment.py
server: concurrency fix + monitoring - add /metrics prometheus compatible endpoint (
#5708
)
2024-02-25 13:49:43 +01:00
issues.feature
server: continue to update other slots on embedding concurrent request (
#5699
)
2024-02-24 19:16:04 +01:00
parallel.feature
server: continue to update other slots on embedding concurrent request (
#5699
)
2024-02-24 19:16:04 +01:00
security.feature
server: init functional tests (
#5566
)
2024-02-24 12:28:55 +01:00
server.feature
server: logs - unified format and --log-format option (
#5700
)
2024-02-25 13:50:32 +01:00
wrong_usages.feature
server: init functional tests (
#5566
)
2024-02-24 12:28:55 +01:00