This website requires JavaScript.
Explore
Help
Sign In
tqcq
/
llama.cpp
Watch
0
Star
0
Fork
0
You've already forked llama.cpp
mirror of
https://github.com/ggml-org/llama.cpp.git
synced
2025-08-19 22:36:13 -04:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
Files
d19101c9a0e38359a303127bb5ccde47395ee083
llama.cpp
/
gguf-py
/
gguf
History
Francis Couture-Harpin
3ad0603c65
Merge branch 'master' into compilade/imatrix-batched-chunks
2024-09-08 10:05:08 -04:00
..
__init__.py
…
constants.py
Merge branch 'master' into compilade/imatrix-batched-chunks
2024-09-08 10:05:08 -04:00
gguf_reader.py
…
gguf_writer.py
llama : support RWKV v6 models (
#8980
)
2024-09-01 17:38:17 +03:00
gguf.py
…
lazy.py
gguf-py : simplify support for quant types (
#8838
)
2024-08-08 13:33:09 -04:00
metadata.py
server : add lora hotswap endpoint (WIP) (
#8857
)
2024-08-06 17:33:39 +02:00
py.typed
…
quants.py
ggml-quants : ternary packing for TriLMs and BitNet b1.58 (
#8151
)
2024-09-05 21:48:47 -04:00
tensor_mapping.py
llama : support RWKV v6 models (
#8980
)
2024-09-01 17:38:17 +03:00
utility.py
…
vocab.py
…