mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-06-27 03:55:20 +00:00
server : re-enable swa speculative decoding
ggml-ci
This commit is contained in:
@ -2017,11 +2017,6 @@ struct server_context {
|
|||||||
params_base.n_cache_reuse = 0;
|
params_base.n_cache_reuse = 0;
|
||||||
SRV_WRN("%s\n", "cache_reuse is not supported by this context, it will be disabled");
|
SRV_WRN("%s\n", "cache_reuse is not supported by this context, it will be disabled");
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!params_base.speculative.model.path.empty()) {
|
|
||||||
SRV_ERR("%s\n", "err: speculative decode is not supported by this context");
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return true;
|
return true;
|
||||||
|
Reference in New Issue
Block a user