mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-06-27 03:55:20 +00:00
examples : allow extracting embeddings from decoder contexts (#13797)
ggml-ci
This commit is contained in:
@ -3394,13 +3394,7 @@ struct server_context {
|
||||
batch.logits + i,
|
||||
};
|
||||
|
||||
int ret = 0;
|
||||
|
||||
if (do_encode) {
|
||||
ret = llama_encode(ctx, batch_view);
|
||||
} else {
|
||||
ret = llama_decode(ctx, batch_view);
|
||||
}
|
||||
const int ret = llama_decode(ctx, batch_view);
|
||||
|
||||
metrics.on_decoded(slots);
|
||||
|
||||
|
Reference in New Issue
Block a user