feat(model): support ollama as an optional llm & embedding proxy (#1475)

Signed-off-by: shanhaikang.shk <shanhaikang.shk@oceanbase.com>
Co-authored-by: Fangyin Cheng <staneyffer@gmail.com>
This commit is contained in:
GITHUBear
2024-04-28 18:36:45 +08:00
committed by GitHub
parent 0f8188b152
commit 744b3e4933
10 changed files with 231 additions and 1 deletions

View File

@@ -11,6 +11,7 @@ def __lazy_import(name):
"ZhipuLLMClient": "dbgpt.model.proxy.llms.zhipu",
"YiLLMClient": "dbgpt.model.proxy.llms.yi",
"MoonshotLLMClient": "dbgpt.model.proxy.llms.moonshot",
"OllamaLLMClient": "dbgpt.model.proxy.llms.ollama",
}
if name in module_path:
@@ -33,4 +34,5 @@ __all__ = [
"SparkLLMClient",
"YiLLMClient",
"MoonshotLLMClient",
"OllamaLLMClient",
]