llama-server: fix model params not propagated (#21509)

Signed-off-by: Aaron Teo <aaron.teo1@ibm.com>
This commit is contained in:
Aaron Teo
2026-04-07 21:39:41 +08:00
committed by GitHub
parent 0d049d6a92
commit 69c28f1547
2 changed files with 6 additions and 3 deletions
+1 -1
View File
@@ -56,7 +56,7 @@ struct server_context {
// load the model and initialize llama_context
// returns true on success
bool load_model(const common_params & params);
bool load_model(common_params & params);
// this function will block main thread until termination
void start_loop();