From 23c1487b3422e8a9d4328427c387e7a2b8778a8a Mon Sep 17 00:00:00 2001 From: aries_ckt <916701291@qq.com> Date: Wed, 13 Sep 2023 23:07:25 +0800 Subject: [PATCH] doc:llm faq doc --- .../getting_started/faq/llm/llm_faq.po | 58 ++++++++++++------- 1 file changed, 37 insertions(+), 21 deletions(-) diff --git a/docs/locales/zh_CN/LC_MESSAGES/getting_started/faq/llm/llm_faq.po b/docs/locales/zh_CN/LC_MESSAGES/getting_started/faq/llm/llm_faq.po index 33e6918d7..4670537c2 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/getting_started/faq/llm/llm_faq.po +++ b/docs/locales/zh_CN/LC_MESSAGES/getting_started/faq/llm/llm_faq.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 👏👏 0.3.5\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-08-28 15:05+0800\n" +"POT-Creation-Date: 2023-09-13 23:04+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -19,90 +19,92 @@ msgstr "" "Content-Transfer-Encoding: 8bit\n" "Generated-By: Babel 2.12.1\n" -#: ../../getting_started/faq/llm/llm_faq.md:1 35c867af1c9c4d52a26c35f6a77af4b8 +#: ../../getting_started/faq/llm/llm_faq.md:1 73790502b62745ec88bbe9fe124254f0 msgid "LLM USE FAQ" msgstr "LLM模型使用FAQ" -#: ../../getting_started/faq/llm/llm_faq.md:3 302d2be76da0463db7ef7b5002ca6bb1 +#: ../../getting_started/faq/llm/llm_faq.md:3 473bdd77bbb242f497f514e6e63d0c5f msgid "Q1:how to use openai chatgpt service" msgstr "我怎么使用OPENAI服务" -#: ../../getting_started/faq/llm/llm_faq.md:4 4d7790aa3cdb4cfcb2360bbb2b136d6e +#: ../../getting_started/faq/llm/llm_faq.md:4 6e073181e48e4604a301f3d7359c91ef msgid "change your LLM_MODEL" msgstr "通过在.env文件设置LLM_MODEL" -#: ../../getting_started/faq/llm/llm_faq.md:9 90b7622f45ef4701aab2c0aaaf52f900 +#: ../../getting_started/faq/llm/llm_faq.md:9 a88cd162dca448b198c0551a70e70da3 msgid "set your OPENAPI KEY" msgstr "set your OPENAPI KEY" -#: ../../getting_started/faq/llm/llm_faq.md:15 67c1e41b57fe460c970bde38327db787 +#: ../../getting_started/faq/llm/llm_faq.md:16 ebaa67e9d31f4c70b4bccbd4394d1c27 msgid "make sure your openapi API_KEY is available" msgstr "确认openapi API_KEY是否可用" -#: ../../getting_started/faq/llm/llm_faq.md:17 c7941e083ef246ae9dc70b6e49315184 +#: ../../getting_started/faq/llm/llm_faq.md:18 8e88363a43b9460dae90a772360dcc5a msgid "" "Q2 What difference between `python dbgpt_server --light` and `python " "dbgpt_server`" -msgstr "Q2 `python dbgpt_server --light` 和 `python " -"dbgpt_server`的区别是什么?" +msgstr "Q2 `python dbgpt_server --light` 和 `python dbgpt_server`的区别是什么?" -#: ../../getting_started/faq/llm/llm_faq.md:19 23e2459b5ac74830b22b1b86dfa85297 +#: ../../getting_started/faq/llm/llm_faq.md:21 1bbf3891883b43659b7ef39ce5e91918 msgid "" "`python dbgpt_server --light` dbgpt_server does not start the llm " "service. Users can deploy the llm service separately by using `python " "llmserver`, and dbgpt_server accesses the llm service through set the " "LLM_SERVER environment variable in .env. The purpose is to allow for the " "separate deployment of dbgpt's backend service and llm service." -msgstr "`python dbgpt_server --light` dbgpt_server在启动后台服务的时候不启动模型服务, 用户可以通过`python llmserver`单独部署模型服务,dbgpt_server通过LLM_SERVER环境变量来访问模型服务。目的是为了可以将dbgpt后台服务和大模型服务分离部署。" +msgstr "" +"`python dbgpt_server --light` dbgpt_server在启动后台服务的时候不启动模型服务, " +"用户可以通过`python " +"llmserver`单独部署模型服务,dbgpt_server通过LLM_SERVER环境变量来访问模型服务。目的是为了可以将dbgpt后台服务和大模型服务分离部署。" -#: ../../getting_started/faq/llm/llm_faq.md:21 4169f2294921469383c25cc8ae4ea83d +#: ../../getting_started/faq/llm/llm_faq.md:23 96a6b6be655c4f85a7c18e813f67517e msgid "" "`python dbgpt_server` dbgpt_server service and the llm service are " "deployed on the same instance. when dbgpt_server starts the service, it " "also starts the llm service at the same time." msgstr "`python dbgpt_server` 是将后台服务和模型服务部署在同一台实例上.dbgpt_server在启动服务的时候同时开启模型服务." -#: ../../getting_started/faq/llm/llm_faq.md:25 56b8dcd2d2d44403a0aca00f7452b92c +#: ../../getting_started/faq/llm/llm_faq.md:27 8a0138f4ceab476a97f112776669c7ca #, fuzzy msgid "Q3 how to use MultiGPUs" msgstr "Q2 怎么使用 MultiGPUs" -#: ../../getting_started/faq/llm/llm_faq.md:26 9d02946ea8d54f11a2600278afc07ab6 +#: ../../getting_started/faq/llm/llm_faq.md:29 6b2f25a5a2b243f78c2f96e3b045bf97 msgid "" "DB-GPT will use all available gpu by default. And you can modify the " "setting `CUDA_VISIBLE_DEVICES=0,1` in `.env` file to use the specific gpu" " IDs." msgstr "DB-GPT默认加载可利用的gpu,你也可以通过修改 在`.env`文件 `CUDA_VISIBLE_DEVICES=0,1`来指定gpu IDs" -#: ../../getting_started/faq/llm/llm_faq.md:28 cf26d432a34d4b2482bcc5255f01f9b1 +#: ../../getting_started/faq/llm/llm_faq.md:32 2adf75ffb0ab451999d2f446389eea6c msgid "" "Optionally, you can also specify the gpu ID to use before the starting " "command, as shown below:" msgstr "你也可以指定gpu ID启动" -#: ../../getting_started/faq/llm/llm_faq.md:38 e105136f3fea42c78338f17d671a578c +#: ../../getting_started/faq/llm/llm_faq.md:42 793d6d8503b74323b4997cf2981cc098 msgid "" "You can modify the setting `MAX_GPU_MEMORY=xxGib` in `.env` file to " "configure the maximum memory used by each GPU." msgstr "同时你可以通过在.env文件设置`MAX_GPU_MEMORY=xxGib`修改每个GPU的最大使用内存" -#: ../../getting_started/faq/llm/llm_faq.md:40 69d02c9004ec4b10a5fe0712a6dc6e9b +#: ../../getting_started/faq/llm/llm_faq.md:44 bdfc8eb5bc89460ea3979f61b8aeca7f #, fuzzy msgid "Q4 Not Enough Memory" msgstr "Q3 机器显存不够 " -#: ../../getting_started/faq/llm/llm_faq.md:42 c3333f6cb608451182d4807fcb9c2d78 +#: ../../getting_started/faq/llm/llm_faq.md:46 e0e60f0263d34eec818b72c38d214b8f msgid "DB-GPT supported 8-bit quantization and 4-bit quantization." msgstr "DB-GPT 支持 8-bit quantization 和 4-bit quantization." -#: ../../getting_started/faq/llm/llm_faq.md:44 01568d7dc96a42dd84d1c1e76ea44186 +#: ../../getting_started/faq/llm/llm_faq.md:48 98c954d9fcf449f4b47610fc96091c4f msgid "" "You can modify the setting `QUANTIZE_8bit=True` or `QUANTIZE_4bit=True` " "in `.env` file to use quantization(8-bit quantization is enabled by " "default)." msgstr "你可以通过在.env文件设置`QUANTIZE_8bit=True` or `QUANTIZE_4bit=True`" -#: ../../getting_started/faq/llm/llm_faq.md:46 7e4682b7992b40d7b18801c9ce4c17b6 +#: ../../getting_started/faq/llm/llm_faq.md:50 2568b441f7e54654b405c7791f08036a msgid "" "Llama-2-70b with 8-bit quantization can run with 80 GB of VRAM, and 4-bit" " quantization can run with 48 GB of VRAM." @@ -110,10 +112,24 @@ msgstr "" "Llama-2-70b with 8-bit quantization 可以运行在 80 GB VRAM机器, 4-bit " "quantization可以运行在 48 GB VRAM" -#: ../../getting_started/faq/llm/llm_faq.md:48 9be65b2cac2541f0aeac21cf36441350 +#: ../../getting_started/faq/llm/llm_faq.md:52 f8d1e4312f9743c7b03820b4a8dbf992 msgid "" "Note: you need to install the latest dependencies according to " "[requirements.txt](https://github.com/eosphoros-ai/DB-" "GPT/blob/main/requirements.txt)." msgstr "" +#: ../../getting_started/faq/llm/llm_faq.md:54 5fe0d9ced7e848799f4d7bce92a5c130 +msgid "Q5 How to Add LLM Service dynamic" +msgstr "Q5 怎样动态新增模型服务" + +#: ../../getting_started/faq/llm/llm_faq.md:56 fd921148e3e547beb6c74035a6b6a8b0 +msgid "" +"Now DB-GPT through multi-llm service switch, so how to add llm service " +"dynamic," +msgstr "DB-GPT支持多个模型服务切换, 怎样添加一个模型服务呢" + +#: ../../getting_started/faq/llm/llm_faq.md:68 f8c024339da447ce8160a4eb9f87c125 +msgid "Q6 dbgpt command not found" +msgstr "Q6 dbgpt command not found" +