mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-06-28 20:25:20 +00:00
llama_batch_ext_ptr::from_text/embd
This commit is contained in:
@ -113,7 +113,7 @@ int main(int argc, char ** argv) {
|
||||
struct common_sampler * smpl = common_sampler_init(model_tgt, params.sampling);
|
||||
|
||||
// eval the prompt
|
||||
llama_batch_ext_ptr batch(llama_batch_ext_init_from_text(inp.data(), inp.size() - 1, 0, 0, true));
|
||||
auto batch = llama_batch_ext_ptr::from_text(inp.data(), inp.size() - 1, 0, 0, true);
|
||||
llama_decode_ext(ctx_tgt, batch.get());
|
||||
|
||||
// note: keep the last token separate!
|
||||
|
Reference in New Issue
Block a user