doc:llm deploy docs (#692)

This commit is contained in:
FangYin Cheng 2023-10-19 12:10:55 +08:00 committed by GitHub
commit 9e5a7bea1f
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
4 changed files with 338 additions and 0 deletions

Binary file not shown.

After

Width:  |  Height:  |  Size: 1.1 MiB

View File

@ -0,0 +1,98 @@
Standalone Deployment
==================================
(standalone-index)=
### Install Prepare
```commandline
git clone https://github.com/eosphoros-ai/DB-GPT.git
cd DB-GPT
```
### Create conda environment
```commandline
conda create -n dbgpt_env python=3.10
conda activate dbgpt_env
```
### Install Default Requirements
```commandline
# Install Default Requirements
pip install -e ".[default]"
```
### Download and Prepare LLM Model and Embedding Model
```{tip}
If you don't have high performance hardware server
```
you can use openai api, tongyi api , bard api, etc.
```commandline
mkdir models && cd models
# download embedding model, eg: text2vec-large-chinese
git clone https://huggingface.co/GanymedeNil/text2vec-large-chinese
```
set proxy api in .env
```commandline
#set LLM_MODEL TYPE
LLM_MODEL=proxyllm
#set your Proxy Api key and Proxy Server url
PROXY_API_KEY={your-openai-sk}
PROXY_SERVER_URL=https://api.openai.com/v1/chat/completions
```
```{tip}
If you have high performance hardware server
```
```commandline
mkdir models && cd models
# # download embedding model, eg: vicuna-13b-v1.5 or
git clone https://huggingface.co/lmsys/vicuna-13b-v1.5
# download embedding model, eg: text2vec-large-chinese
git clone https://huggingface.co/GanymedeNil/text2vec-large-chinese
popd
```
### Start all services with a single command.
```commandline
LLM_MODEL=vicuna-13b-v1.5
dbgpt start webserver --port 6006
```
By default, the "dbgpt start webserver" command will start the Webserver, Model Controller, and Model Worker in a single Python process. Here, we specify the service to be started on port 6006.
### View and validate the model service in the command line, you can use the following commands
##### 1.list the started model services and deployed Model Workers, you can use the following command
```commandline
dbgpt model list
```
output is:
```commandline
+-----------------+------------+------------+------+---------+---------+-----------------+----------------------------+
| Model Name | Model Type | Host | Port | Healthy | Enabled | Prompt Template | Last Heartbeat |
+-----------------+------------+------------+------+---------+---------+-----------------+----------------------------+
| vicuna-13b-v1.5 | llm | 172.17.0.9 | 6006 | True | True | | 2023-10-16T19:49:59.201313 |
| WorkerManager | service | 172.17.0.9 | 6006 | True | True | | 2023-10-16T19:49:59.246756 |
+-----------------+------------+------------+------+---------+---------+-----------------+----------------------------+
```
The WorkerManager is the management process for Model Workers
##### validate the deployed model in the command line, you can use the following command
```commandline
dbgpt model chat --model_name vicuna-13b-v1.5
```
Then an interactive page will be launched where you can have a conversation with the deployed LLM in the terminal.
```commandline
Chatbot started with model vicuna-13b-v1.5. Type 'exit' to leave the chat.
You: Hello
Bot: Hello! How can I assist you today?
You:
```

View File

@ -0,0 +1,123 @@
# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2023, csunny
# This file is distributed under the same license as the DB-GPT package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2023.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: DB-GPT 👏👏 0.3.9\n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2023-10-17 17:24+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language: zh_CN\n"
"Language-Team: zh_CN <LL@li.org>\n"
"Plural-Forms: nplurals=1; plural=0;\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Generated-By: Babel 2.12.1\n"
#: ../../getting_started/application/chatagent/chatagent.md:1
#: ../../getting_started/application/chatagent/chatagent.md:3
#: 17aae18073594c7498c8b2434d50b218 77f0177e55e144a89e3bcf735b0d52de
msgid "ChatAgent"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:3
#: 41b42808938245dc946960571c09026a
msgid ""
"ChatAgent can automatically select the right tool to answer and solve "
"user problems through LLM. ![ChatAgent](https://github.com/eosphoros-ai"
"/DB-GPT/assets/13723926/)"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:6
#: bb76bb44442647a091d008c9e1427e4a
msgid "1.Update Plugins From Git"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:7
#: a7f188c3de7e4b85b2d3079128b535b2
msgid ""
"Update the latest plug-in information from the DB-GPT official plug-in "
"repository. ![Update From Github](https://github.com/eosphoros-ai/DB-"
"GPT/assets/13723926/)"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:7
#: 612cb75edb2f48a087fc99e6d0890b80
msgid "Update From Github"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:10
#: 1ad3f7680837410cb8018fff5787366f
msgid "2.Install Plugin"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:11
#: b32e9318a8e347339b6c4cad4b966880
msgid ""
"Install the plugin you want to use from the list of agent hub plugins. "
"![Install Plugin](https://github.com/eosphoros-ai/DB-"
"GPT/assets/13723926/)"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:11
#: bee96f7936564bc6be98655bfc6aae0d
msgid "Install Plugin"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:15
#: 20d562b9ed004336bb6df8d163554924
msgid "3.Upload My Local Plugin File"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:16
#: b43e485ff4f042dd80b5950444c0db08
msgid ""
"If you want to use a private plug-in developed by yourself, you can "
"upload the plug-in first and then use it. ![Upload Private "
"Plugin](https://github.com/eosphoros-ai/DB-GPT/assets/13723926/)"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:16
#: a05724784c0e45528039b8238fd9ed3c
msgid "Upload Private Plugin"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:19
#: 726b5c4cad7d4be9815e9fbeedaecc2b
msgid "4.Select installed plug-ins dialog"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:20
#: e3d9dc892d054cbca29267cd0fffe7be
msgid ""
"Select an installed plugin to have a conversation. ![Select "
"Plugins](https://github.com/eosphoros-ai/DB-GPT/assets/13723926/)"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:20
#: 97f28bacc02444c580f6481b21e6c155
msgid "Select Plugins"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:23
#: d0272e7d94d743289fb2c1613ee85c36
msgid "5.Uninstall My Plugin"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:24
#: 2fee213a12da4674b19fe8891e4318f5
msgid ""
"Uninstall plugins you no longer need. ![Uninstall "
"Plugin](https://github.com/eosphoros-ai/DB-GPT/assets/13723926/)"
msgstr ""
#: ../../getting_started/application/chatagent/chatagent.md:24
#: 32e5405249134863bcaf85b0ef91260f
msgid "Uninstall Plugin"
msgstr ""

View File

@ -0,0 +1,117 @@
# SOME DESCRIPTIVE TITLE.
# Copyright (C) 2023, csunny
# This file is distributed under the same license as the DB-GPT package.
# FIRST AUTHOR <EMAIL@ADDRESS>, 2023.
#
#, fuzzy
msgid ""
msgstr ""
"Project-Id-Version: DB-GPT 👏👏 0.3.9\n"
"Report-Msgid-Bugs-To: \n"
"POT-Creation-Date: 2023-10-17 17:24+0800\n"
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
"Language: zh_CN\n"
"Language-Team: zh_CN <LL@li.org>\n"
"Plural-Forms: nplurals=1; plural=0;\n"
"MIME-Version: 1.0\n"
"Content-Type: text/plain; charset=utf-8\n"
"Content-Transfer-Encoding: 8bit\n"
"Generated-By: Babel 2.12.1\n"
#: ../../getting_started/install/cluster/vms/standalone.md:1
#: 40af95d327574e5ea9a6c29f68a7709c
msgid "Standalone Deployment"
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:5
#: 7ef23971a0d2415a857d5eb9c4d955ac
msgid "Install Prepare"
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:12
#: f3cd2941fa2d49d686fc246237f82e68
msgid "Create conda environment"
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:20
#: 589d80d5d97044ee93747091904583c5
msgid "Install Default Requirements"
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:26
#: 1d151e72660644e7b43084ffccb99598
msgid "Download and Prepare LLM Model and Embedding Model"
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:28
#: 4443c83f8dfa4fdba30929f1e5ecf619
msgid "If you don't have high performance hardware server"
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:30
#: 33a6ee878a4a467d8bcfe5dfbdcbe184
msgid "you can use openai api, tongyi api , bard api, etc."
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:39
#: 7efdb99bc36d42819b1fcd1341b45eef
msgid "set proxy api in .env"
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:48
#: 50154ece70bc421ebe7ac05966369bc8
msgid "If you have high performance hardware server"
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:62
#: 4f3696ed069a4a84b2d794593df23765
msgid "Start all services with a single command."
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:67
#: f5f671962cb14a10aba34c4274d2fc2c
msgid ""
"By default, the \"dbgpt start webserver\" command will start the "
"Webserver, Model Controller, and Model Worker in a single Python process."
" Here, we specify the service to be started on port 6006."
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:69
#: 4883221a356843f6a6335c25847aecd2
msgid ""
"View and validate the model service in the command line, you can use the "
"following commands"
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:70
#: 1346eb55d7c647be90a90a867ba04ec3
msgid ""
"1.list the started model services and deployed Model Workers, you can use"
" the following command"
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:74
#: 8ab332ebce554758952dbe85c68330a9
msgid "output is:"
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:83
#: efb460bbb70842218f5c056cf05ba1a0
msgid "The WorkerManager is the management process for Model Workers"
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:85
#: 10e5b79f3d914ee3974b1a0d07962fe2
msgid ""
"validate the deployed model in the command line, you can use the "
"following command"
msgstr ""
#: ../../getting_started/install/cluster/vms/standalone.md:89
#: 39d062e7fea948949c1c5c099893f308
msgid ""
"Then an interactive page will be launched where you can have a "
"conversation with the deployed LLM in the terminal."
msgstr ""