Files
llama.cpp/examples
Pierrick Hymbert 7f5ff558ee server: stop generation at n_ctx_train if n_predict is not set (#6638)
* server: cap n_predict if not set to n_ctx_train

* server: fix infinite loop

* server: infinite loop, move in process_token
server: infinite loop: set stop limit to true

* minor: spaces

* minor: spaces

* server: include prompt tokens in the EOS limit
2024-04-26 12:15:30 +02:00
..
2024-04-09 13:44:08 -04:00
2024-04-25 14:27:20 +03:00
2024-04-09 13:44:08 -04:00