From ca3e9883d19128691d657040d5c9175a49a9cdf4 Mon Sep 17 00:00:00 2001 From: zxstty Date: Tue, 16 Sep 2025 16:59:24 +0800 Subject: [PATCH] =?UTF-8?q?=E5=B0=86maxtoken=E7=94=A8=E4=BA=8E=E9=99=90?= =?UTF-8?q?=E5=88=B6=E6=A8=A1=E5=9E=8B=E7=9A=84=E8=BE=93=E5=85=A5=E8=80=8C?= =?UTF-8?q?=E4=B8=8D=E6=98=AF=E8=BE=93=E5=87=BA?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- apps/llm/reasoning.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/apps/llm/reasoning.py b/apps/llm/reasoning.py index 79768783..fddc84db 100644 --- a/apps/llm/reasoning.py +++ b/apps/llm/reasoning.py @@ -142,7 +142,7 @@ class ReasoningLLM: return await self._client.chat.completions.create( model=model, messages=messages, # type: ignore[] - max_tokens=max_tokens or self._config.max_tokens, + max_completion_tokens=max_tokens or self._config.max_tokens, temperature=temperature or self._config.temperature, stream=True, stream_options={"include_usage": True}, -- Gitee