This website requires JavaScript.
Explore
Help
Sign In
tqcq
/
llama.cpp
Watch
0
Star
0
Fork
0
You've already forked llama.cpp
mirror of
https://github.com/ggml-org/llama.cpp.git
synced
2025-08-11 03:03:47 -04:00
Code
Issues
Packages
Projects
Releases
Wiki
Activity
Files
36667c8edcded08063ed51c7d57e9e086bbfc903
llama.cpp
/
gguf-py
/
gguf
History
Jared Van Bortel
2f567611c0
llama-model : support Qwen2 embedding models and pooling_mode_lasttoken (
#13245
)
2025-05-02 11:42:30 -04:00
..
scripts
gguf-py : GGUF Editor GUI - Python + Qt6 (
#12930
)
2025-04-18 20:30:41 +02:00
__init__.py
…
constants.py
llama-model : support Qwen2 embedding models and pooling_mode_lasttoken (
#13245
)
2025-05-02 11:42:30 -04:00
gguf_reader.py
…
gguf_writer.py
convert : converting mmproj for Qwen2/2.5VL from convert_hf_to_gguf (
#13209
)
2025-05-02 17:17:15 +02:00
gguf.py
…
lazy.py
gguf-py : support lazy tensor splitting (
#12809
)
2025-04-08 09:03:07 +02:00
metadata.py
…
py.typed
…
quants.py
…
tensor_mapping.py
convert : converting mmproj for Qwen2/2.5VL from convert_hf_to_gguf (
#13209
)
2025-05-02 17:17:15 +02:00
utility.py
convert : ability to lazy-load safetensors remotely without downloading to disk (
#12820
)
2025-04-10 17:24:44 +02:00
vocab.py
convert : Support chat_template.json (
#12460
)
2025-03-19 08:58:13 +01:00