mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-06-27 20:05:20 +00:00
Server: Enable setting default sampling parameters via command-line (#8402)
* Load server sampling parameters from the server context by default. * Wordsmithing comment
This commit is contained in:
@ -884,7 +884,8 @@ struct server_context {
|
|||||||
|
|
||||||
bool launch_slot_with_task(server_slot & slot, const server_task & task) {
|
bool launch_slot_with_task(server_slot & slot, const server_task & task) {
|
||||||
slot_params default_params;
|
slot_params default_params;
|
||||||
llama_sampling_params default_sparams;
|
// Sampling parameter defaults are loaded from the global server context (but individual requests can still override them)
|
||||||
|
llama_sampling_params default_sparams = params.sparams;
|
||||||
auto & data = task.data;
|
auto & data = task.data;
|
||||||
|
|
||||||
if (data.count("__oaicompat") != 0) {
|
if (data.count("__oaicompat") != 0) {
|
||||||
|
Reference in New Issue
Block a user