feat(core): Support max output tokens parameter (#2218)

This commit is contained in:
Fangyin Cheng
2024-12-18 17:19:18 +08:00
committed by GitHub
parent b05febbf77
commit 7f4b5e79cf
121 changed files with 333 additions and 219 deletions

View File

@@ -112,6 +112,11 @@ class OpenAILLMClient(ProxyLLMClient):
_initialize_openai(self._init_params)
if not self._openai_less_then_v1:
# Prepare openai client and cache default headers
# It will block the main thread in some cases
_ = self.client.default_headers
@classmethod
def new_client(
cls,