Add timeout parameter for better support of openailike LLM tools on local computer (like LM Studio).

Reuse other parameters to improve settings of the OpenAILike object.
This commit is contained in:
jcbonnet-fwd
2024-04-18 15:15:13 +02:00
parent 08c4ab175e
commit 507ae62414
3 changed files with 13 additions and 1 deletions

View File

@@ -105,8 +105,12 @@ class LLMComponent:
api_key=openai_settings.api_key,
model=openai_settings.model,
is_chat_model=True,
max_tokens=None,
max_tokens=settings.llm.max_new_tokens,
api_version="",
temperature=settings.llm.temperature,
tokenizer=settings.llm.tokenizer,
timeout=openai_settings.request_timeout,
reuse_client=False,
)
case "ollama":
try:

View File

@@ -202,6 +202,10 @@ class OpenAISettings(BaseModel):
"gpt-3.5-turbo",
description="OpenAI Model to use. Example: 'gpt-4'.",
)
request_timeout: float = Field(
120.0,
description="Time elapsed until openailike server times out the request. Default is 120s. Format is float. ",
)
class OllamaSettings(BaseModel):

View File

@@ -3,6 +3,9 @@ server:
llm:
mode: openailike
max_new_tokens: 512
tokenizer: mistralai/Mistral-7B-Instruct-v0.2
temperature: 0.1
embedding:
mode: huggingface
@@ -15,3 +18,4 @@ openai:
api_base: http://localhost:8000/v1
api_key: EMPTY
model: facebook/opt-125m
request_timeout: 600.0