Xuan Son Nguyen
|
bfdddbc150
|
bring back mistakenly deleted llama_batch_init/free
|
2025-03-14 00:22:28 +01:00 |
|
Xuan Son Nguyen
|
c3dd79007b
|
fix llama_batch_ext_init_from_text
|
2025-03-13 23:09:27 +01:00 |
|
Xuan Son Nguyen
|
4aabf4e8f4
|
return output ID from llama_batch_ext_add/set
|
2025-03-13 17:47:07 +01:00 |
|
Xuan Son Nguyen
|
1d6ba97789
|
remove token_info API
|
2025-03-01 16:21:16 +01:00 |
|
Xuan Son Nguyen
|
1170135dfb
|
llama_batch_ext_add_text
|
2025-03-01 14:00:14 +01:00 |
|
Xuan Son Nguyen
|
17d3658b5f
|
move to llama_batch_ext
|
2025-02-16 00:02:53 +01:00 |
|
Xuan Son Nguyen
|
f2e59a8eb9
|
rework, targeting llama-server
|
2025-02-14 18:16:49 +01:00 |
|
Xuan Son Nguyen
|
4ed4fe75ed
|
first proposal for private llama_batch
|
2025-02-14 00:48:12 +01:00 |
|
Georgi Gerganov
|
f66f582927
|
llama : refactor src/llama.cpp (#10902)
* llama : scatter llama.cpp into multiple modules (wip)
* llama : control-vector -> adapter
* llama : arch
* llama : mmap
ggml-ci
* ci : remove BUILD_SHARED_LIBS=OFF
ggml-ci
* llama : arch (cont)
ggml-ci
* llama : chat
ggml-ci
* llama : model
ggml-ci
* llama : hparams
ggml-ci
* llama : adapter
ggml-ci
* examples : fix
ggml-ci
* rebase
ggml-ci
* minor
* llama : kv cache
ggml-ci
* llama : impl
ggml-ci
* llama : batch
ggml-ci
* cont
ggml-ci
* llama : context
ggml-ci
* minor
* llama : context (cont)
ggml-ci
* llama : model loader
ggml-ci
* common : update lora
ggml-ci
* llama : quant
ggml-ci
* llama : quant (cont)
ggml-ci
* minor [no ci]
|
2025-01-03 10:18:53 +02:00 |
|