Files
llama.cpp/examples
Xiao-Yong Jin b8ad1b66b2 server : allow json array in prompt or content for direct token input (#2306)
* server: allow json array in prompt or content

We accept an array of strings and numbers representing tokens,
in addition to the current string valued prompt or content.

This allows direct token input, so that any special tokens
can be processed and used at the frontend during the construction
of the json data, before sending to the server. And the server
does not need to know or parse special tokens from textual input.

With this, we can use EOS and BOS used in llama-2-chat models.

* server: use tokenizePrompt(json) and default "" if empty prompt

* server: fix prompt check

* server: tokenize endpoint no longer adds BOS
2023-08-23 15:12:12 +08:00
..
2023-07-19 10:01:11 +03:00
2023-08-22 21:01:57 -04:00
2023-03-29 20:21:09 +03:00
2023-03-25 21:51:41 +02:00
2023-08-08 14:44:48 +03:00
2023-07-28 21:05:08 +03:00