From ed99a8ea04f1eb66c12ad4333c6e8966bd7a25ac Mon Sep 17 00:00:00 2001 From: Georgi Gerganov Date: Thu, 12 Jun 2025 10:43:55 +0300 Subject: [PATCH] cont : fix comments --- src/llama-context.cpp | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/src/llama-context.cpp b/src/llama-context.cpp index a001891da..2e551bf6e 100644 --- a/src/llama-context.cpp +++ b/src/llama-context.cpp @@ -941,7 +941,7 @@ int llama_context::decode(llama_batch & inp_batch) { } } - // this indicates we are doing pooled embedding, so we ignore batch.logits and output all tokens + // this indicates we are doing pooled embedding const bool embd_pooled = cparams.embeddings && cparams.pooling_type != LLAMA_POOLING_TYPE_NONE; int64_t n_outputs_all = 0; @@ -2077,7 +2077,7 @@ void llama_context::opt_epoch_iter( n_queued_tokens += n_tokens_all; - // this indicates we are doing pooled embedding, so we ignore batch.logits and output all tokens + // this indicates we are doing pooled embedding const bool embd_pooled = cparams.embeddings && cparams.pooling_type != LLAMA_POOLING_TYPE_NONE; embd_seq.clear();