mirror of
https://github.com/imartinez/privateGPT.git
synced 2025-09-02 15:55:02 +00:00
Add timeout parameter for better support of openailike LLM tools on local computer (like LM Studio).
Reuse other parameters to improve settings of the OpenAILike object.
This commit is contained in:
@@ -105,8 +105,12 @@ class LLMComponent:
|
||||
api_key=openai_settings.api_key,
|
||||
model=openai_settings.model,
|
||||
is_chat_model=True,
|
||||
max_tokens=None,
|
||||
max_tokens=settings.llm.max_new_tokens,
|
||||
api_version="",
|
||||
temperature=settings.llm.temperature,
|
||||
tokenizer=settings.llm.tokenizer,
|
||||
timeout=openai_settings.request_timeout,
|
||||
reuse_client=False,
|
||||
)
|
||||
case "ollama":
|
||||
try:
|
||||
|
@@ -202,6 +202,10 @@ class OpenAISettings(BaseModel):
|
||||
"gpt-3.5-turbo",
|
||||
description="OpenAI Model to use. Example: 'gpt-4'.",
|
||||
)
|
||||
request_timeout: float = Field(
|
||||
120.0,
|
||||
description="Time elapsed until openailike server times out the request. Default is 120s. Format is float. ",
|
||||
)
|
||||
|
||||
|
||||
class OllamaSettings(BaseModel):
|
||||
|
@@ -3,6 +3,9 @@ server:
|
||||
|
||||
llm:
|
||||
mode: openailike
|
||||
max_new_tokens: 512
|
||||
tokenizer: mistralai/Mistral-7B-Instruct-v0.2
|
||||
temperature: 0.1
|
||||
|
||||
embedding:
|
||||
mode: huggingface
|
||||
@@ -15,3 +18,4 @@ openai:
|
||||
api_base: http://localhost:8000/v1
|
||||
api_key: EMPTY
|
||||
model: facebook/opt-125m
|
||||
request_timeout: 600.0
|
Reference in New Issue
Block a user