@@ -40,7 +40,7 @@ class LLMBackend:
"messages": messages,
"model": model,
"stream": stream,
- "max_tokens": 100000,
+ "max_tokens": 20000,
#"presence_penalty": 0,
#"frequency_penalty": 0
# "parallel_tool_calls": parallel_tool_calls,