mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-07-01 05:05:10 +00:00
qwen2vl: use llama_batch_ext_set_pos
This commit is contained in:
@ -950,6 +950,12 @@ extern "C" {
|
||||
int32_t pos0,
|
||||
int32_t seq_id);
|
||||
|
||||
// Set arbitrary token to the embeddings batch
|
||||
// Note: this is only to be used in conjunction with llama_batch_ext_init_from_embd()
|
||||
// n_pos must match the n_tokens of the batch
|
||||
// Returns -1 if n_pos does not match the n_tokens of the batch
|
||||
LLAMA_API int32_t llama_batch_ext_set_pos(struct llama_batch_ext * batch, llama_pos * pos, size_t n_pos);
|
||||
|
||||
// Get the number of tokens in the batch
|
||||
LLAMA_API int32_t llama_batch_ext_get_n_tokens(const struct llama_batch_ext * batch);
|
||||
|
||||
|
Reference in New Issue
Block a user