Files
llama.cpp/examples
Georgi Gerganov ad19812cda perplexity : faster HellaSwag via batching (#5017)
* perplexity : faster HellaSwag

ggml-ci

* perplexity : clean-up

ggml-ci

* perplexity : no need for decode_helper

ggml-ci

* perplexity : add comments

* perplexity : option to specify max batched tasks via `n_parallel`

* perplexity : remove HellaSwag restruction for n_batch
2024-01-18 15:33:01 +02:00
..
2024-01-14 09:45:56 +02:00
2023-12-21 23:08:14 +02:00
2024-01-13 20:45:45 +02:00