diff --git a/docs/getting_started/application/chatdb/img.png b/docs/getting_started/application/chatdb/img.png new file mode 100644 index 000000000..b6a3afc5e Binary files /dev/null and b/docs/getting_started/application/chatdb/img.png differ diff --git a/docs/getting_started/install/cluster/vms/standalone.md b/docs/getting_started/install/cluster/vms/standalone.md new file mode 100644 index 000000000..6e75e7d05 --- /dev/null +++ b/docs/getting_started/install/cluster/vms/standalone.md @@ -0,0 +1,98 @@ +Standalone Deployment +================================== +(standalone-index)= + +### Install Prepare +```commandline +git clone https://github.com/eosphoros-ai/DB-GPT.git + +cd DB-GPT +``` + +### Create conda environment +```commandline +conda create -n dbgpt_env python=3.10 + +conda activate dbgpt_env +``` + + +### Install Default Requirements +```commandline +# Install Default Requirements +pip install -e ".[default]" +``` + +### Download and Prepare LLM Model and Embedding Model +```{tip} + If you don't have high performance hardware server +``` +you can use openai api, tongyi api , bard api, etc. +```commandline +mkdir models && cd models + +# download embedding model, eg: text2vec-large-chinese +git clone https://huggingface.co/GanymedeNil/text2vec-large-chinese + +``` + +set proxy api in .env +```commandline +#set LLM_MODEL TYPE +LLM_MODEL=proxyllm +#set your Proxy Api key and Proxy Server url +PROXY_API_KEY={your-openai-sk} +PROXY_SERVER_URL=https://api.openai.com/v1/chat/completions +``` +```{tip} +If you have high performance hardware server +``` + +```commandline +mkdir models && cd models + +# # download embedding model, eg: vicuna-13b-v1.5 or +git clone https://huggingface.co/lmsys/vicuna-13b-v1.5 + +# download embedding model, eg: text2vec-large-chinese +git clone https://huggingface.co/GanymedeNil/text2vec-large-chinese + +popd +``` +### Start all services with a single command. +```commandline +LLM_MODEL=vicuna-13b-v1.5 +dbgpt start webserver --port 6006 +``` +By default, the "dbgpt start webserver" command will start the Webserver, Model Controller, and Model Worker in a single Python process. Here, we specify the service to be started on port 6006. + +### View and validate the model service in the command line, you can use the following commands +##### 1.list the started model services and deployed Model Workers, you can use the following command +```commandline +dbgpt model list +``` +output is: +```commandline ++-----------------+------------+------------+------+---------+---------+-----------------+----------------------------+ +| Model Name | Model Type | Host | Port | Healthy | Enabled | Prompt Template | Last Heartbeat | ++-----------------+------------+------------+------+---------+---------+-----------------+----------------------------+ +| vicuna-13b-v1.5 | llm | 172.17.0.9 | 6006 | True | True | | 2023-10-16T19:49:59.201313 | +| WorkerManager | service | 172.17.0.9 | 6006 | True | True | | 2023-10-16T19:49:59.246756 | ++-----------------+------------+------------+------+---------+---------+-----------------+----------------------------+ +``` +The WorkerManager is the management process for Model Workers + +##### validate the deployed model in the command line, you can use the following command +```commandline +dbgpt model chat --model_name vicuna-13b-v1.5 +``` +Then an interactive page will be launched where you can have a conversation with the deployed LLM in the terminal. +```commandline +Chatbot started with model vicuna-13b-v1.5. Type 'exit' to leave the chat. + + +You: Hello +Bot: Hello! How can I assist you today? + +You: +``` \ No newline at end of file diff --git a/docs/locales/zh_CN/LC_MESSAGES/getting_started/application/chatagent/chatagent.po b/docs/locales/zh_CN/LC_MESSAGES/getting_started/application/chatagent/chatagent.po new file mode 100644 index 000000000..8e43a26e4 --- /dev/null +++ b/docs/locales/zh_CN/LC_MESSAGES/getting_started/application/chatagent/chatagent.po @@ -0,0 +1,123 @@ +# SOME DESCRIPTIVE TITLE. +# Copyright (C) 2023, csunny +# This file is distributed under the same license as the DB-GPT package. +# FIRST AUTHOR , 2023. +# +#, fuzzy +msgid "" +msgstr "" +"Project-Id-Version: DB-GPT 👏👏 0.3.9\n" +"Report-Msgid-Bugs-To: \n" +"POT-Creation-Date: 2023-10-17 17:24+0800\n" +"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" +"Last-Translator: FULL NAME \n" +"Language: zh_CN\n" +"Language-Team: zh_CN \n" +"Plural-Forms: nplurals=1; plural=0;\n" +"MIME-Version: 1.0\n" +"Content-Type: text/plain; charset=utf-8\n" +"Content-Transfer-Encoding: 8bit\n" +"Generated-By: Babel 2.12.1\n" + +#: ../../getting_started/application/chatagent/chatagent.md:1 +#: ../../getting_started/application/chatagent/chatagent.md:3 +#: 17aae18073594c7498c8b2434d50b218 77f0177e55e144a89e3bcf735b0d52de +msgid "ChatAgent" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:3 +#: 41b42808938245dc946960571c09026a +msgid "" +"ChatAgent can automatically select the right tool to answer and solve " +"user problems through LLM. ![ChatAgent](https://github.com/eosphoros-ai" +"/DB-GPT/assets/13723926/)" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:6 +#: bb76bb44442647a091d008c9e1427e4a +msgid "1.Update Plugins From Git" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:7 +#: a7f188c3de7e4b85b2d3079128b535b2 +msgid "" +"Update the latest plug-in information from the DB-GPT official plug-in " +"repository. ![Update From Github](https://github.com/eosphoros-ai/DB-" +"GPT/assets/13723926/)" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:7 +#: 612cb75edb2f48a087fc99e6d0890b80 +msgid "Update From Github" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:10 +#: 1ad3f7680837410cb8018fff5787366f +msgid "2.Install Plugin" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:11 +#: b32e9318a8e347339b6c4cad4b966880 +msgid "" +"Install the plugin you want to use from the list of agent hub plugins. " +"![Install Plugin](https://github.com/eosphoros-ai/DB-" +"GPT/assets/13723926/)" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:11 +#: bee96f7936564bc6be98655bfc6aae0d +msgid "Install Plugin" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:15 +#: 20d562b9ed004336bb6df8d163554924 +msgid "3.Upload My Local Plugin File" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:16 +#: b43e485ff4f042dd80b5950444c0db08 +msgid "" +"If you want to use a private plug-in developed by yourself, you can " +"upload the plug-in first and then use it. ![Upload Private " +"Plugin](https://github.com/eosphoros-ai/DB-GPT/assets/13723926/)" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:16 +#: a05724784c0e45528039b8238fd9ed3c +msgid "Upload Private Plugin" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:19 +#: 726b5c4cad7d4be9815e9fbeedaecc2b +msgid "4.Select installed plug-ins dialog" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:20 +#: e3d9dc892d054cbca29267cd0fffe7be +msgid "" +"Select an installed plugin to have a conversation. ![Select " +"Plugins](https://github.com/eosphoros-ai/DB-GPT/assets/13723926/)" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:20 +#: 97f28bacc02444c580f6481b21e6c155 +msgid "Select Plugins" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:23 +#: d0272e7d94d743289fb2c1613ee85c36 +msgid "5.Uninstall My Plugin" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:24 +#: 2fee213a12da4674b19fe8891e4318f5 +msgid "" +"Uninstall plugins you no longer need. ![Uninstall " +"Plugin](https://github.com/eosphoros-ai/DB-GPT/assets/13723926/)" +msgstr "" + +#: ../../getting_started/application/chatagent/chatagent.md:24 +#: 32e5405249134863bcaf85b0ef91260f +msgid "Uninstall Plugin" +msgstr "" + diff --git a/docs/locales/zh_CN/LC_MESSAGES/getting_started/install/cluster/vms/standalone.po b/docs/locales/zh_CN/LC_MESSAGES/getting_started/install/cluster/vms/standalone.po new file mode 100644 index 000000000..a183fb2c4 --- /dev/null +++ b/docs/locales/zh_CN/LC_MESSAGES/getting_started/install/cluster/vms/standalone.po @@ -0,0 +1,117 @@ +# SOME DESCRIPTIVE TITLE. +# Copyright (C) 2023, csunny +# This file is distributed under the same license as the DB-GPT package. +# FIRST AUTHOR , 2023. +# +#, fuzzy +msgid "" +msgstr "" +"Project-Id-Version: DB-GPT 👏👏 0.3.9\n" +"Report-Msgid-Bugs-To: \n" +"POT-Creation-Date: 2023-10-17 17:24+0800\n" +"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" +"Last-Translator: FULL NAME \n" +"Language: zh_CN\n" +"Language-Team: zh_CN \n" +"Plural-Forms: nplurals=1; plural=0;\n" +"MIME-Version: 1.0\n" +"Content-Type: text/plain; charset=utf-8\n" +"Content-Transfer-Encoding: 8bit\n" +"Generated-By: Babel 2.12.1\n" + +#: ../../getting_started/install/cluster/vms/standalone.md:1 +#: 40af95d327574e5ea9a6c29f68a7709c +msgid "Standalone Deployment" +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:5 +#: 7ef23971a0d2415a857d5eb9c4d955ac +msgid "Install Prepare" +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:12 +#: f3cd2941fa2d49d686fc246237f82e68 +msgid "Create conda environment" +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:20 +#: 589d80d5d97044ee93747091904583c5 +msgid "Install Default Requirements" +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:26 +#: 1d151e72660644e7b43084ffccb99598 +msgid "Download and Prepare LLM Model and Embedding Model" +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:28 +#: 4443c83f8dfa4fdba30929f1e5ecf619 +msgid "If you don't have high performance hardware server" +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:30 +#: 33a6ee878a4a467d8bcfe5dfbdcbe184 +msgid "you can use openai api, tongyi api , bard api, etc." +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:39 +#: 7efdb99bc36d42819b1fcd1341b45eef +msgid "set proxy api in .env" +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:48 +#: 50154ece70bc421ebe7ac05966369bc8 +msgid "If you have high performance hardware server" +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:62 +#: 4f3696ed069a4a84b2d794593df23765 +msgid "Start all services with a single command." +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:67 +#: f5f671962cb14a10aba34c4274d2fc2c +msgid "" +"By default, the \"dbgpt start webserver\" command will start the " +"Webserver, Model Controller, and Model Worker in a single Python process." +" Here, we specify the service to be started on port 6006." +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:69 +#: 4883221a356843f6a6335c25847aecd2 +msgid "" +"View and validate the model service in the command line, you can use the " +"following commands" +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:70 +#: 1346eb55d7c647be90a90a867ba04ec3 +msgid "" +"1.list the started model services and deployed Model Workers, you can use" +" the following command" +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:74 +#: 8ab332ebce554758952dbe85c68330a9 +msgid "output is:" +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:83 +#: efb460bbb70842218f5c056cf05ba1a0 +msgid "The WorkerManager is the management process for Model Workers" +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:85 +#: 10e5b79f3d914ee3974b1a0d07962fe2 +msgid "" +"validate the deployed model in the command line, you can use the " +"following command" +msgstr "" + +#: ../../getting_started/install/cluster/vms/standalone.md:89 +#: 39d062e7fea948949c1c5c099893f308 +msgid "" +"Then an interactive page will be launched where you can have a " +"conversation with the deployed LLM in the terminal." +msgstr "" +