mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-06-29 12:35:16 +00:00
return output ID from llama_batch_ext_add/set
This commit is contained in:
@ -131,7 +131,7 @@ int main(int argc, char ** argv) {
|
||||
}
|
||||
|
||||
// llama_decode will output logits only for the last token of the prompt
|
||||
llama_batch_ext_set_logits_last(batch);
|
||||
llama_batch_ext_set_output_last(batch);
|
||||
|
||||
if (llama_decode_ext(ctx, batch) != 0) {
|
||||
LOG_ERR("%s: llama_decode() failed\n", __func__);
|
||||
|
Reference in New Issue
Block a user