|
@@ -35,7 +35,7 @@ class LLMBackend:
|
|
|
"messages": messages,
|
|
|
"model": model,
|
|
|
"stream": stream,
|
|
|
- "parallel_tool_calls": parallel_tool_calls,
|
|
|
+ # "parallel_tool_calls": parallel_tool_calls,
|
|
|
}
|
|
|
if extra_body:
|
|
|
model_params = extra_body.get("model_params")
|
|
@@ -67,6 +67,7 @@ class LLMBackend:
|
|
|
chat_params["top_p"] = top_p
|
|
|
if tools:
|
|
|
chat_params["tools"] = tools
|
|
|
+ chat_params["parallel_tool_calls"] = parallel_tool_calls
|
|
|
chat_params["tool_choice"] = tool_choice if tool_choice else "auto"
|
|
|
if (
|
|
|
isinstance(response_format, dict)
|