We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent d06d354 commit 658451bCopy full SHA for 658451b
examples/server/server.cpp
@@ -1208,7 +1208,7 @@ struct server_context {
1208
}
1209
1210
auto n_ctx_train = llama_n_ctx_train(model);
1211
- if (slot.params.n_predict < 1 && slot.ga_n == 1
+ if (slot.params.n_predict < 1 && slot.n_predict < 1 && slot.ga_n == 1
1212
&& slot.n_prompt_tokens + slot.n_decoded >= n_ctx_train) {
1213
LOG_WARNING("n_predict is not set and self-context extend is disabled."
1214
" Limiting generated tokens to n_ctx_train to avoid EOS-less generation infinite loop", {
0 commit comments