Add timeout parameter for better support of openailike LLM tools on local computer (like LM Studio).

Reuse other parameters to improve settings of the OpenAILike object.
This commit is contained in:
jcbonnet-fwd
2024-04-18 15:15:13 +02:00
parent 08c4ab175e
commit 507ae62414
3 changed files with 13 additions and 1 deletions

View File

@@ -105,8 +105,12 @@ class LLMComponent:
api_key=openai_settings.api_key, api_key=openai_settings.api_key,
model=openai_settings.model, model=openai_settings.model,
is_chat_model=True, is_chat_model=True,
max_tokens=None, max_tokens=settings.llm.max_new_tokens,
api_version="", api_version="",
temperature=settings.llm.temperature,
tokenizer=settings.llm.tokenizer,
timeout=openai_settings.request_timeout,
reuse_client=False,
) )
case "ollama": case "ollama":
try: try:

View File

@@ -202,6 +202,10 @@ class OpenAISettings(BaseModel):
"gpt-3.5-turbo", "gpt-3.5-turbo",
description="OpenAI Model to use. Example: 'gpt-4'.", description="OpenAI Model to use. Example: 'gpt-4'.",
) )
request_timeout: float = Field(
120.0,
description="Time elapsed until openailike server times out the request. Default is 120s. Format is float. ",
)
class OllamaSettings(BaseModel): class OllamaSettings(BaseModel):

View File

@@ -3,6 +3,9 @@ server:
llm: llm:
mode: openailike mode: openailike
max_new_tokens: 512
tokenizer: mistralai/Mistral-7B-Instruct-v0.2
temperature: 0.1
embedding: embedding:
mode: huggingface mode: huggingface
@@ -15,3 +18,4 @@ openai:
api_base: http://localhost:8000/v1 api_base: http://localhost:8000/v1
api_key: EMPTY api_key: EMPTY
model: facebook/opt-125m model: facebook/opt-125m
request_timeout: 600.0