Files
llama.cpp/tools
Radoslav Gerganov c830f99cfa server : support max_completion_tokens request property (#19831)
"max_tokens" is deprectated in favor of "max_completion_tokens" which
sets the upper bound for reasoning+output token.

Closes: #13700
2026-02-24 10:30:00 +02:00
..
2026-02-14 09:47:01 +01:00