This website requires JavaScript.
Explore
Help
Sign In
tqcq
/
llama.cpp
Watch
0
Star
0
Fork
0
You've already forked llama.cpp
mirror of
https://github.com/ggml-org/llama.cpp.git
synced
2025-08-30 11:59:59 -04:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
Files
c104023994d36a8e791fc6a43789b84fd552cefc
llama.cpp
/
common
History
David Huang
7f323a589f
Add
--no-op-offload
to improve
-ot
pp perf in MoE models like llama4 400B (
#13386
)
2025-05-11 14:18:39 +02:00
..
cmake
…
minja
…
arg.cpp
…
arg.h
…
base64.hpp
…
build-info.cpp.in
…
chat.cpp
…
chat.h
…
CMakeLists.txt
…
common.cpp
…
common.h
…
console.cpp
…
console.h
…
json-schema-to-grammar.cpp
…
json-schema-to-grammar.h
…
json.hpp
…
llguidance.cpp
…
log.cpp
…
log.h
…
ngram-cache.cpp
…
ngram-cache.h
…
sampling.cpp
…
sampling.h
…
speculative.cpp
…
speculative.h
…
stb_image.h
…