This website requires JavaScript.
Explore
Help
Sign In
tqcq
/
llama.cpp
Watch
0
Star
0
Fork
0
You've already forked llama.cpp
mirror of
https://github.com/ggml-org/llama.cpp.git
synced
2025-07-24 03:31:31 +00:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
Files
3a504d9a0bd7d952d22cd2d707446de2316ec955
llama.cpp
/
common
History
Georgi Gerganov
972f91c7d7
Merge branch 'master' into gg/llama-kv-cache
...
ggml-ci
2025-02-10 14:45:54 +02:00
..
cmake
…
arg.cpp
…
arg.h
…
base64.hpp
…
build-info.cpp.in
…
chat-template.hpp
sync: minja (
a72057e519
) (
#11774
)
2025-02-10 09:34:09 +00:00
chat.cpp
…
chat.hpp
…
CMakeLists.txt
…
common.cpp
…
common.h
…
console.cpp
…
console.h
…
json-schema-to-grammar.cpp
…
json-schema-to-grammar.h
…
json.hpp
…
llguidance.cpp
llama : add llama_sampler_init for safe usage of llama_sampler_free (
#11727
)
2025-02-07 11:33:27 +02:00
log.cpp
…
log.h
There's a better way of clearing lines (
#11756
)
2025-02-09 10:34:49 +00:00
minja.hpp
sync: minja (
a72057e519
) (
#11774
)
2025-02-10 09:34:09 +00:00
ngram-cache.cpp
…
ngram-cache.h
…
sampling.cpp
…
sampling.h
…
speculative.cpp
…
speculative.h
…
stb_image.h
…