|
@@ -35,7 +35,7 @@ class LLMBackend:
|
|
|
"messages": messages,
|
|
|
"model": model,
|
|
|
"stream": stream,
|
|
|
- "parallel_tool_calls": False,
|
|
|
+ "parallel_tool_calls": parallel_tool_calls,
|
|
|
}
|
|
|
if extra_body:
|
|
|
model_params = extra_body.get("model_params")
|
|
@@ -57,8 +57,6 @@ class LLMBackend:
|
|
|
chat_params["stream_options"] = {
|
|
|
"include_usage": bool(stream_options["include_usage"])
|
|
|
}
|
|
|
- # if parallel_tool_calls:
|
|
|
- chat_params["parallel_tool_calls"] = parallel_tool_calls
|
|
|
if audio:
|
|
|
chat_params["audio"] = audio
|
|
|
if modalities:
|