chore: Modify ollama config

This commit is contained in:
Fangyin Cheng 2025-03-04 18:45:52 +08:00
parent c8e252c4de
commit a22de08936

View File

@ -9,13 +9,13 @@ from dbgpt.core.awel.flow import (
ResourceCategory, ResourceCategory,
auto_register_resource, auto_register_resource,
) )
from dbgpt.core.interface.parameter import LLMDeployModelParameters
from dbgpt.model.proxy.base import ( from dbgpt.model.proxy.base import (
AsyncGenerateStreamFunction, AsyncGenerateStreamFunction,
GenerateStreamFunction, GenerateStreamFunction,
ProxyLLMClient, ProxyLLMClient,
register_proxy_model_adapter, register_proxy_model_adapter,
) )
from dbgpt.model.proxy.llms.chatgpt import OpenAICompatibleDeployModelParameters
from dbgpt.model.proxy.llms.proxy_model import ProxyModel, parse_model_request from dbgpt.model.proxy.llms.proxy_model import ProxyModel, parse_model_request
from dbgpt.util.i18n_utils import _ from dbgpt.util.i18n_utils import _
@ -31,7 +31,7 @@ logger = logging.getLogger(__name__)
show_in_ui=False, show_in_ui=False,
) )
@dataclass @dataclass
class OllamaDeployModelParameters(OpenAICompatibleDeployModelParameters): class OllamaDeployModelParameters(LLMDeployModelParameters):
"""Deploy model parameters for Ollama.""" """Deploy model parameters for Ollama."""
provider: str = "proxy/ollama" provider: str = "proxy/ollama"