From 381c7cdbacb477ac4e0740afd97682a56cadaa83 Mon Sep 17 00:00:00 2001 From: xuyuan23 <643854343@qq.com> Date: Wed, 14 Jun 2023 17:54:21 +0800 Subject: [PATCH] Add Chinese documentation. --- .../getting_started/getting_started.po | 103 ++++++++++-------- .../locales/zh_CN/LC_MESSAGES/modules/llms.po | 69 +++++++++--- .../zh_CN/LC_MESSAGES/modules/plugins.po | 80 ++++++++++++-- .../use_cases/tool_use_with_plugin.po | 85 ++++++++++++++- 4 files changed, 261 insertions(+), 76 deletions(-) diff --git a/docs/locales/zh_CN/LC_MESSAGES/getting_started/getting_started.po b/docs/locales/zh_CN/LC_MESSAGES/getting_started/getting_started.po index bc3e499a5..683eb6938 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/getting_started/getting_started.po +++ b/docs/locales/zh_CN/LC_MESSAGES/getting_started/getting_started.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 0.1.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-06-11 14:10+0800\n" +"POT-Creation-Date: 2023-06-14 17:26+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -17,95 +17,94 @@ msgstr "" "MIME-Version: 1.0\n" "Content-Type: text/plain; charset=utf-8\n" "Content-Transfer-Encoding: 8bit\n" -"Generated-By: Babel 2.11.0\n" +"Generated-By: Babel 2.12.1\n" -#: ../../getting_started/getting_started.md:1 cf1947dea9a843dd8b6fff68642f29b1 +#: ../../getting_started/getting_started.md:1 a0477412435c4c569cf71d243d2884c7 msgid "Quickstart Guide" msgstr "使用指南" -#: ../../getting_started/getting_started.md:3 4184879bf5b34521a95e497f4747241a +#: ../../getting_started/getting_started.md:3 331f8c3fbbac44c1b75d2ff595c0235f msgid "" "This tutorial gives you a quick walkthrough about use DB-GPT with you " "environment and data." msgstr "本教程为您提供了关于如何使用DB-GPT的使用指南。" -#: ../../getting_started/getting_started.md:5 7431b72cc1504b8bbcafb7512a6b6c92 +#: ../../getting_started/getting_started.md:5 df41fe97067d4ba680e3231f05a843de msgid "Installation" msgstr "安装" -#: ../../getting_started/getting_started.md:7 b8faf2ec4e034855a2674ffcade8cee2 +#: ../../getting_started/getting_started.md:7 73f72c06a89341f38f4bfbe70ed0d2ae msgid "To get started, install DB-GPT with the following steps." msgstr "请按照以下步骤安装DB-GPT" -#: ../../getting_started/getting_started.md:9 ae0f536a064647cda04ea3d253991d80 +#: ../../getting_started/getting_started.md:9 59c159fd30104ba081e2ebbf1605fe11 msgid "1. Hardware Requirements" msgstr "1. 硬件要求" -#: ../../getting_started/getting_started.md:10 8fa637100e644b478e0d6858f0a5b63d +#: ../../getting_started/getting_started.md:10 126558467caa4b05bea0bb051c864831 msgid "" "As our project has the ability to achieve ChatGPT performance of over " "85%, there are certain hardware requirements. However, overall, the " "project can be deployed and used on consumer-grade graphics cards. The " "specific hardware requirements for deployment are as follows:" -msgstr "由于我们的项目有能力达到85%以上的ChatGPT性能,所以对硬件有一定的要求。" -"但总体来说,我们在消费级的显卡上即可完成项目的部署使用,具体部署的硬件说明如下:" +msgstr "由于我们的项目有能力达到85%以上的ChatGPT性能,所以对硬件有一定的要求。但总体来说,我们在消费级的显卡上即可完成项目的部署使用,具体部署的硬件说明如下:" -#: ../../getting_started/getting_started.md c68539579083407882fb0d28943d40db +#: ../../getting_started/getting_started.md 967e5cde5c6241edae9e0e0e0b217221 msgid "GPU" msgstr "GPU" -#: ../../getting_started/getting_started.md 613fbe77d41a4a20a30c3c9a0b6ec20c +#: ../../getting_started/getting_started.md 05be5479b15b403c8aa3374ea53feff8 msgid "VRAM Size" msgstr "显存大小" -#: ../../getting_started/getting_started.md c0b7f8249d3d4c629ba5deb8188a49b4 +#: ../../getting_started/getting_started.md ea646d956e834a75912780789946bb47 msgid "Performance" msgstr "显存大小" -#: ../../getting_started/getting_started.md 5d103f7e4d1b4b6cb7358c0c717c9f73 +#: ../../getting_started/getting_started.md aba96df6b91441b982128426eb3a2ebb msgid "RTX 4090" msgstr "RTX 4090" -#: ../../getting_started/getting_started.md 48338f6b18dc41efb3613d47b1a762a7 -#: f14d278e083440b58fc7faeed30e2879 +#: ../../getting_started/getting_started.md d2797ffe3d534460ae77a3397fc07c1c +#: e1b5f1a7502e475b872eebd75972a87f msgid "24 GB" msgstr "24 GB" -#: ../../getting_started/getting_started.md dc238037ff3449cdb95cbd882d8de170 +#: ../../getting_started/getting_started.md 4ed99c06dc4643c290b5914142ea8371 msgid "Smooth conversation inference" msgstr "可以流畅的进行对话推理,无卡顿" -#: ../../getting_started/getting_started.md d7f84ac79bf84cb6a453d3bfd26eb935 +#: ../../getting_started/getting_started.md b3dc189389cd41f2883acfe3fad3d6a4 msgid "RTX 3090" msgstr "RTX 3090" -#: ../../getting_started/getting_started.md 511ee322b777476b87a3aa5624609944 +#: ../../getting_started/getting_started.md 0bea7ccef4154e2994144f81af877919 msgid "Smooth conversation inference, better than V100" msgstr "可以流畅进行对话推理,有卡顿感,但好于V100" -#: ../../getting_started/getting_started.md 974b704e8cf84f6483774153df8a8c6c +#: ../../getting_started/getting_started.md 7d1c9e5c16184f2bacf8b12d6f38f629 msgid "V100" msgstr "V100" -#: ../../getting_started/getting_started.md 72008961ce004a0fa24b74db55fcf96e +#: ../../getting_started/getting_started.md e9ddddf92e15420f85852cea1a1bbd8d msgid "16 GB" msgstr "16 GB" -#: ../../getting_started/getting_started.md 2a3b936fe04c4b7789680c26be7f4869 +#: ../../getting_started/getting_started.md f5532da60b99495c8329d674749ae79f msgid "Conversation inference possible, noticeable stutter" msgstr "可以进行对话推理,有明显卡顿" -#: ../../getting_started/getting_started.md:18 fb1dbccb8f804384ade8e171aa40f99c +#: ../../getting_started/getting_started.md:18 3357d10704b94249b8ccdf7fd3645624 msgid "2. Install" msgstr "2. 安装" -#: ../../getting_started/getting_started.md:20 695fdb8858c6488e9a0872d68fb387e5 +#: ../../getting_started/getting_started.md:20 08e5054dd0fa4e07aa92236dca03a1d3 msgid "" "This project relies on a local MySQL database service, which you need to " "install locally. We recommend using Docker for installation." msgstr "本项目依赖一个本地的 MySQL 数据库服务,你需要本地安装,推荐直接使用 Docker 安装。" -#: ../../getting_started/getting_started.md:25 954f3a282ec54b11a55ebfe1f680d1df +#: ../../getting_started/getting_started.md:25 f99a1af073e24b339390400251a50c9b msgid "" "We use [Chroma embedding database](https://github.com/chroma-core/chroma)" " as the default for our vector database, so there is no need for special " @@ -114,66 +113,74 @@ msgid "" "installation process of DB-GPT, we use the miniconda3 virtual " "environment. Create a virtual environment and install the Python " "dependencies." -msgstr "向量数据库我们默认使用的是Chroma内存数据库,所以无需特殊安装,如果有" -"需要连接其他的同学,可以按照我们的教程进行安装配置。整个DB-GPT的" -"安装过程,我们使用的是miniconda3的虚拟环境。创建虚拟环境,并安装python依赖包" +msgstr "" +"向量数据库我们默认使用的是Chroma内存数据库,所以无需特殊安装,如果有需要连接其他的同学,可以按照我们的教程进行安装配置。整个DB-" +"GPT的安装过程,我们使用的是miniconda3的虚拟环境。创建虚拟环境,并安装python依赖包" - -#: ../../getting_started/getting_started.md:35 0314bad0928940fc8e382d289d356c66 +#: ../../getting_started/getting_started.md:35 d8ebdf7c4ac54113be1c94ed879dc93f msgid "" "Once the environment is installed, we have to create a new folder " "\"models\" in the DB-GPT project, and then we can put all the models " "downloaded from huggingface in this directory" -msgstr "环境安装完成后,我们必须在DB-GPT项目中创建一个新文件夹\"models\"," -"然后我们可以把从huggingface下载的所有模型放到这个目录下。" +msgstr "" +"环境安装完成后,我们必须在DB-" +"GPT项目中创建一个新文件夹\"models\",然后我们可以把从huggingface下载的所有模型放到这个目录下。" -#: ../../getting_started/getting_started.md:42 afdf176f72224fd6b8b6e9e23c80c1ef +#: ../../getting_started/getting_started.md:42 cd3ce56a4d644574a0c30dd86148a58c msgid "" "The model files are large and will take a long time to download. During " "the download, let's configure the .env file, which needs to be copied and" " created from the .env.template" -msgstr "模型文件很大,需要很长时间才能下载。在下载过程中,让我们配置.env文件," -"它需要从。env.template中复制和创建。" +msgstr "模型文件很大,需要很长时间才能下载。在下载过程中,让我们配置.env文件,它需要从。env.template中复制和创建。" -#: ../../getting_started/getting_started.md:48 76c87610993f41059c3c0aade5117171 +#: ../../getting_started/getting_started.md:48 96d634498ae04c84b4eae8502d5f65e8 msgid "" "You can configure basic parameters in the .env file, for example setting " "LLM_MODEL to the model to be used" msgstr "您可以在.env文件中配置基本参数,例如将LLM_MODEL设置为要使用的模型。" -#: ../../getting_started/getting_started.md:35 443f5f92e4cd4ce4887bae2556b605b0 +#: ../../getting_started/getting_started.md:50 a78627885d2a41a481131308d2d061a6 msgid "3. Run" msgstr "3. 运行" -#: ../../getting_started/getting_started.md:36 3dab200eceda460b81a096d44de43d21 +#: ../../getting_started/getting_started.md:51 69c1d06867cb4a0ea36ad0f3686109db msgid "" "You can refer to this document to obtain the Vicuna weights: " "[Vicuna](https://github.com/lm-sys/FastChat/blob/main/README.md#model-" "weights) ." -msgstr "关于基础模型, 可以根据[Vicuna](https://github.com/lm-sys/FastChat/b" -"lob/main/README.md#model-weights) 合成教程进行合成。" +msgstr "" +"关于基础模型, 可以根据[Vicuna](https://github.com/lm-" +"sys/FastChat/blob/main/README.md#model-weights) 合成教程进行合成。" - -#: ../../getting_started/getting_started.md:38 b036ca6294f04bceb686187d2d8b6646 +#: ../../getting_started/getting_started.md:53 1dcccc1784e04fa3a3340aa446f1484f msgid "" "If you have difficulty with this step, you can also directly use the " "model from [this link](https://huggingface.co/Tribbiani/vicuna-7b) as a " "replacement." -msgstr "如果此步有困难的同学,也可以直接使用[此链接](https://huggingface.co/Tribbiani/vicuna-7b)上的模型进行替代。" +msgstr "" +"如果此步有困难的同学,也可以直接使用[此链接](https://huggingface.co/Tribbiani/vicuna-" +"7b)上的模型进行替代。" -#: ../../getting_started/getting_started.md:40 35537c13ff6f4bd69951c486274ca1f9 +#: ../../getting_started/getting_started.md:55 45a4c9f50ee04533bd57652505ab4f62 msgid "Run server" msgstr "运行模型服务" -#: ../../getting_started/getting_started.md:45 f7aa3668a6c94fb3a1b8346392d921f3 +#: ../../getting_started/getting_started.md:60 f3482509ad2f4137becd4775768180bd +msgid "" +"Starting `llmserver.py` with the following command will result in a " +"relatively stable Python service with multiple processes." +msgstr "使用以下命令启动llmserver.py将会得到一个相对稳定的Python服务,并且具有多个进程。" + +#: ../../getting_started/getting_started.md:65 ddf718f6863f4e58b8de1232dc8189dd msgid "Run gradio webui" msgstr "运行模型服务" -#: ../../getting_started/getting_started.md:51 d80c908f01144e2c8a15b7f6e8e7f88d +#: ../../getting_started/getting_started.md:71 d3df863c46ae49b1a63e5778d439d336 msgid "" "Notice: the webserver need to connect llmserver, so you need change the" " .env file. change the MODEL_SERVER = \"http://127.0.0.1:8000\" to your " "address. It's very important." -msgstr "注意: 在启动Webserver之前, 需要修改.env 文件中的MODEL_SERVER" -" = "http://127.0.0.1:8000", 将地址设置为你的服务器地址。" +msgstr "" +"注意: 在启动Webserver之前, 需要修改.env 文件中的MODEL_SERVER = " +"\"http://127.0.0.1:8000\", 将地址设置为你的服务器地址。" diff --git a/docs/locales/zh_CN/LC_MESSAGES/modules/llms.po b/docs/locales/zh_CN/LC_MESSAGES/modules/llms.po index bbb05b046..22138dec3 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/modules/llms.po +++ b/docs/locales/zh_CN/LC_MESSAGES/modules/llms.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 0.1.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-06-13 11:38+0800\n" +"POT-Creation-Date: 2023-06-14 17:26+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -19,11 +19,11 @@ msgstr "" "Content-Transfer-Encoding: 8bit\n" "Generated-By: Babel 2.12.1\n" -#: ../../modules/llms.md:1 34386f3fecba48fbbd86718283ba593c +#: ../../modules/llms.md:1 8eef439964b5442d91ad04ff72b3b45b msgid "LLMs" msgstr "大语言模型" -#: ../../modules/llms.md:3 241b39ad980f4cfd90a7f0fdae05a1d2 +#: ../../modules/llms.md:3 a29256f1f39b4bcda29a6811ad1b10f6 #, python-format msgid "" "In the underlying large model integration, we have designed an open " @@ -36,62 +36,95 @@ msgid "" "of use." msgstr "在底层大模型接入中,我们设计了开放的接口,支持对接多种大模型。同时对于接入模型的效果,我们有非常严格的把控与评审机制。对大模型能力上与ChatGPT对比,在准确率上需要满足85%以上的能力对齐。我们用更高的标准筛选模型,是期望在用户使用过程中,可以省去前面繁琐的测试评估环节。" -#: ../../modules/llms.md:5 25175e87a62e41bca86798eb783cefd6 +#: ../../modules/llms.md:5 e899f11399bb45d9990c73a273ed7697 msgid "Multi LLMs Usage" msgstr "多模型使用" -#: ../../modules/llms.md:6 8c35341e9ca94202ba779567813f9973 +#: ../../modules/llms.md:6 a21d6a875b3949b9be512f8ea396f6b3 msgid "" "To use multiple models, modify the LLM_MODEL parameter in the .env " "configuration file to switch between the models." msgstr "如果要使用不同的模型,请修改.env配置文件中的LLM MODEL参数以在模型之间切换。" -#: ../../modules/llms.md:8 2edf3309a6554f39ad74e19faff09cee +#: ../../modules/llms.md:8 dd061d45bb4044dcbc7e5d4b0014ded8 msgid "" "Notice: you can create .env file from .env.template, just use command " "like this:" msgstr "注意:你可以从 .env.template 创建 .env 文件。只需使用如下命令:" -#: ../../modules/llms.md:14 5fa7639ef294425e89e13b7c6617fb4b +#: ../../modules/llms.md:14 c1a789e2de2c4958987370521d46c7cc msgid "" "now we support models vicuna-13b, vicuna-7b, chatglm-6b, flan-t5-base, " "guanaco-33b-merged, falcon-40b, gorilla-7b." -msgstr "现在我们支持的模型有vicuna-13b, vicuna-7b, chatglm-6b, flan-t5-base, " -"guanaco-33b-merged, falcon-40b, gorilla-7b." +msgstr "" +"现在我们支持的模型有vicuna-13b, vicuna-7b, chatglm-6b, flan-t5-base, guanaco-33b-" +"merged, falcon-40b, gorilla-7b." -#: ../../modules/llms.md:16 96c9a5ad00264bd2a07bdbdec87e471e +#: ../../modules/llms.md:16 ddb6f2f638e642a595365a91ffdba8f9 msgid "" "DB-GPT provides a model load adapter and chat adapter. load adapter which" " allows you to easily adapt load different LLM models by inheriting the " "BaseLLMAdapter. You just implement match() and loader() method." -msgstr "DB-GPT提供了多模型适配器load adapter和chat adapter.load adapter通过继承BaseLLMAdapter类, 实现match和loader方法允许你适配不同的LLM." +msgstr "" +"DB-GPT提供了多模型适配器load adapter和chat adapter.load adapter通过继承BaseLLMAdapter类," +" 实现match和loader方法允许你适配不同的LLM." -#: ../../modules/llms.md:18 1033714691464f50900c04c9e1bb5643 +#: ../../modules/llms.md:18 5e32be54895243caa6d44d0b3421e4a0 msgid "vicuna llm load adapter" msgstr "vicuna llm load adapter" -#: ../../modules/llms.md:35 faa6432575be45bcae5deb1cc7fee3fb +#: ../../modules/llms.md:35 0b1f2e7c65164c9584e0c544394e7d57 msgid "chatglm load adapter" msgstr "chatglm load adapter" -#: ../../modules/llms.md:62 61c4189cabf04e628132c2bf5f02bb50 +#: ../../modules/llms.md:62 885b35375c764e29a983b54514e378d2 msgid "" "chat adapter which allows you to easily adapt chat different LLM models " "by inheriting the BaseChatAdpter.you just implement match() and " "get_generate_stream_func() method" -msgstr "chat adapter通过继承BaseChatAdpter允许你通过实现match和get_generate_stream_func方法允许你适配不同的LLM." +msgstr "" +"chat " +"adapter通过继承BaseChatAdpter允许你通过实现match和get_generate_stream_func方法允许你适配不同的LLM." -#: ../../modules/llms.md:64 407a67e4e2c6414b9cde346961d850c0 +#: ../../modules/llms.md:64 e0538e7e0526440085b32add07f5ec7f msgid "vicuna llm chat adapter" msgstr "vicuna llm chat adapter" -#: ../../modules/llms.md:76 53a55238cd90406db58c50dc64465195 +#: ../../modules/llms.md:76 2c97712441874e0d8deedc1d9a1ce5ed msgid "chatglm llm chat adapter" msgstr "chatglm llm chat adapter" -#: ../../modules/llms.md:89 b0c5ff72c05e40b3b301d6b81205fe63 +#: ../../modules/llms.md:89 485e5aa261714146a03a30dbcd612653 msgid "" "if you want to integrate your own model, just need to inheriting " "BaseLLMAdaper and BaseChatAdpter and implement the methods" msgstr "如果你想集成自己的模型,只需要继承BaseLLMAdaper和BaseChatAdpter类,然后实现里面的方法即可" +#: ../../modules/llms.md:92 a63b63022db74d76b743044be178e227 +#, fuzzy +msgid "Multi Proxy LLMs" +msgstr "多模型使用" + +#: ../../modules/llms.md:93 dab3041e90384049872a7f77933b1a1f +msgid "1. Openai proxy" +msgstr "Openai代理" + +#: ../../modules/llms.md:94 e50eae200bf04e4788bbc394e0b3d6b9 +msgid "" +"If you haven't deployed a private infrastructure for a large model, or if" +" you want to use DB-GPT in a low-cost and high-efficiency way, you can " +"also use OpenAI's large model as your underlying model." +msgstr "如果你没有部署私有大模型的资源,或者你想使用低成本启动DB-GPT,你可以使用openai的大模型作为你的底层模型" + +#: ../../modules/llms.md:96 53dd581608d74355ba0ce486a01ef261 +msgid "" +"If your environment deploying DB-GPT has access to OpenAI, then modify " +"the .env configuration file as below will work." +msgstr "如果你的环境能够访问openai,你只需要参考如下修改.env配置文件即可" + +#: ../../modules/llms.md:104 8df2d75af41b4953a73b6b7eae9f0373 +msgid "" +"If you can't access OpenAI locally but have an OpenAI proxy service, you " +"can configure as follows." +msgstr "如果你本地无法访问openai,但是你有一个openai的代理服务,你可以参考如下配置" + diff --git a/docs/locales/zh_CN/LC_MESSAGES/modules/plugins.po b/docs/locales/zh_CN/LC_MESSAGES/modules/plugins.po index 483761bac..dce190384 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/modules/plugins.po +++ b/docs/locales/zh_CN/LC_MESSAGES/modules/plugins.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 0.1.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-06-11 14:10+0800\n" +"POT-Creation-Date: 2023-06-14 17:26+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -17,13 +17,13 @@ msgstr "" "MIME-Version: 1.0\n" "Content-Type: text/plain; charset=utf-8\n" "Content-Transfer-Encoding: 8bit\n" -"Generated-By: Babel 2.11.0\n" +"Generated-By: Babel 2.12.1\n" -#: ../../modules/plugins.md:1 48f1b7ff4099485ba3853c373e64273f +#: ../../modules/plugins.md:1 be587eb6ad384844b83ac740a2f3309e msgid "Plugins" msgstr "插件" -#: ../../modules/plugins.md:3 3d94b3250511468d80aa29359f01128d +#: ../../modules/plugins.md:3 291a6adc34684b28867b7d1adcceadbb msgid "" "The ability of Agent and Plugin is the core of whether large models can " "be automated. In this project, we natively support the plugin mode, and " @@ -31,7 +31,73 @@ msgid "" "order to give full play to the advantages of the community, the plugins " "used in this project natively support the Auto-GPT plugin ecology, that " "is, Auto-GPT plugins can directly run in our project." -msgstr "Agent与插件能力是大模型能否自动化的核心,在本的项目中,原生支持插件模式," -"大模型可以自动化完成目标。 同时为了充分发挥社区的优势,本项目中所用的插件原生支持" -"Auto-GPT插件生态,即Auto-GPT的插件可以直接在我们的项目中运行。" +msgstr "" +"Agent与插件能力是大模型能否自动化的核心,在本的项目中,原生支持插件模式,大模型可以自动化完成目标。 同时为了充分发挥社区的优势" +",本项目中所用的插件原生支持Auto-GPT插件生态,即Auto-GPT的插件可以直接在我们的项目中运行。" + +#: ../../modules/plugins.md:5 72010cbac395488bb72c5ffa79806f61 +#, fuzzy +msgid "Local Plugins" +msgstr "插件" + +#: ../../modules/plugins.md:7 842e405d6748425dab6684c3377e580a +msgid "1.1 How to write local plugins." +msgstr "如何编写一个本地插件" + +#: ../../modules/plugins.md:9 23fa64fa76954000a3f33f5b3205975e +msgid "" +"Local plugins use the Auto-GPT plugin template. A simple example is as " +"follows: first write a plugin file called \"sql_executor.py\"." +msgstr "本地插件使用Auto-GPT插件模板,一个简单的示例如下:首先编写一个插件文件`sql_executor.py`" + +#: ../../modules/plugins.md:39 5044401271fb4cfba53299875886b5b9 +msgid "" +"Then set the \"can_handle_post_prompt\" method of the plugin template to " +"True. In the \"post_prompt\" method, write the prompt information and the" +" mapped plugin function." +msgstr "然后设置can_handle_post_prompt函数为True, 在post_prompt函数中编写prompt信息和插件映射函数" + +#: ../../modules/plugins.md:81 8f2001c2934d4655a1993a98d5e7dd63 +msgid "1.2 How to use local plugins" +msgstr "1.2 如何使用本地插件" + +#: ../../modules/plugins.md:83 aa6234c1531a470db97e076408c70ebc +msgid "" +"Pack your plugin project into `your-plugin.zip` and place it in the " +"`/plugins/` directory of the DB-GPT project. After starting the " +"webserver, you can select and use it in the `Plugin Model` section." +msgstr "将您的插件项目打包成your-plugin.zip,并将其放置在DB-GPT项目的/plugins/目录中。启动Web服务器后,您可以在插件模型部分中选择并使用它。" + +#: ../../modules/plugins.md:86 6d1402451cb44ab2bb9ded2a303d8dd0 +#, fuzzy +msgid "Public Plugins" +msgstr "插件" + +#: ../../modules/plugins.md:88 2bb33cadc7604f529c762939a6225f17 +msgid "1.1 How to use public plugins" +msgstr "1.1 如何编写公共插件" + +#: ../../modules/plugins.md:90 d5f458dcba2e435987ffbc62a1d7a989 +msgid "" +"By default, after launching the webserver, plugins from the public plugin" +" library `DB-GPT-Plugins` will be automatically loaded. For more details," +" please refer to [DB-GPT-Plugins](https://github.com/csunny/DB-GPT-" +"Plugins)" +msgstr "默认情况下,在启动Web服务器后,将自动加载来自公共插件库DB-GPT-Plugins的插件。要了解更多详情,请参阅[DB-GPT-Plugins](https://github.com/csunny/DB-GPT-Plugins)" + +#: ../../modules/plugins.md:92 28fa983769ae44a0b790d977c60ce982 +msgid "1.2 Contribute to the DB-GPT-Plugins repository" +msgstr "1.2 贡献到DB-GPT-Plugins仓库" + +#: ../../modules/plugins.md:94 68573e7cc17f479fa676d0631e011baf +msgid "" +"Please refer to the plugin development process in the public plugin " +"library, and put the configuration parameters in `.plugin_env`" +msgstr "请参考公共插件库开发过程,将插件配置参数写入.plugin_env文件" + +#: ../../modules/plugins.md:96 8d6860026b824b46b2899cbf3dc3b4a0 +msgid "" +"We warmly welcome everyone to contribute plugins to the public plugin " +"library!" +msgstr "非常欢迎大家向我们公共插件库贡献插件!" diff --git a/docs/locales/zh_CN/LC_MESSAGES/use_cases/tool_use_with_plugin.po b/docs/locales/zh_CN/LC_MESSAGES/use_cases/tool_use_with_plugin.po index 73293771d..383b62b85 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/use_cases/tool_use_with_plugin.po +++ b/docs/locales/zh_CN/LC_MESSAGES/use_cases/tool_use_with_plugin.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 0.1.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-06-11 14:10+0800\n" +"POT-Creation-Date: 2023-06-14 17:26+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -17,9 +17,88 @@ msgstr "" "MIME-Version: 1.0\n" "Content-Type: text/plain; charset=utf-8\n" "Content-Transfer-Encoding: 8bit\n" -"Generated-By: Babel 2.11.0\n" +"Generated-By: Babel 2.12.1\n" -#: ../../use_cases/tool_use_with_plugin.md:1 2bd7d79a16a548c4a3872a12c436aa4f +#: ../../use_cases/tool_use_with_plugin.md:1 defad8b142a2408e89b782803149be50 msgid "Tool use with plugin" msgstr "插件工具" +#: ../../use_cases/tool_use_with_plugin.md:3 30ca5f2ce55a4369b48dd8a8b08c3273 +msgid "" +"DB-GPT supports a variety of plug-ins, such as MySQL, MongoDB, ClickHouse" +" and other database tool plug-ins. In addition, some database management " +"platforms can also package their interfaces and package them into plug-" +"ins, and use the model to realize the ability of \"single-sentence " +"requirements\"" +msgstr "DB-GPT支持各种插件,例如MySQL、MongoDB、ClickHouse等数据库工具插件。此外,一些数据库管理平台也可以将它们的接口打包成插件,使用该模型实现"一句话需求"的能力。" + +#: ../../use_cases/tool_use_with_plugin.md:6 96c20363d25842cbbab21819c2567a52 +msgid "DB-GPT-DASHBOARD-PLUGIN" +msgstr "DB-GPT-DASHBOARD-PLUGIN" + +#: ../../use_cases/tool_use_with_plugin.md:8 1a78f9e15c6949019e213f3f88388e9f +msgid "" +"[Db-GPT Chart Plugin](https://github.com/csunny/DB-GPT-" +"Plugins/blob/main/src/dbgpt_plugins/Readme.md)" +msgstr "" +"[Db-GPT Chart Plugin](https://github.com/csunny/DB-GPT-" +"Plugins/blob/main/src/dbgpt_plugins/Readme.md)" + +#: ../../use_cases/tool_use_with_plugin.md:10 5ac6d662797e4694926cb0b44e58ff40 +msgid "" +"This is a DB-GPT plugin to generate data analysis charts, if you want to " +"use the test sample data, please first pull the code of [DB-GPT-" +"Plugins](https://github.com/csunny/DB-GPT-Plugins), run the command to " +"generate test DuckDB data, and then copy the generated data file to the " +"`/pilot/mock_datas` directory of the DB-GPT project." +msgstr "这是一个DB-GPT插件,用于生成数据分析图表。如果您想使用测试样本数据,请先拉取 DB-GPT-Plugins 的代码,运行命令以生成测试 DuckDB 数据,然后将生成的数据文件复制到 DB-GPT 项目的 /pilot/mock_datas 目录中。" + +#: ../../use_cases/tool_use_with_plugin.md:21 1029537ed0ca44e499a8f2098cc72f1a +msgid "" +"Test Case: Use a histogram to analyze the total order amount of users in " +"different cities." +msgstr "测试示例:请使用柱状图分析各个城市的用户数" + +#: ../../use_cases/tool_use_with_plugin.md:26 a8a2103279d641b98948bcb06971c160 +msgid "" +"More detail see: [DB-DASHBOARD](https://github.com/csunny/DB-GPT-" +"Plugins/blob/main/src/dbgpt_plugins/Readme.md)" +msgstr "更多详情请看:[DB-DASHBOARD](https://github.com/csunny/DB-GPT-" +"Plugins/blob/main/src/dbgpt_plugins/Readme.md)" + +#: ../../use_cases/tool_use_with_plugin.md:29 db2906b5f89e4a01a4a47a89ad750804 +msgid "DB-GPT-SQL-Execution-Plugin" +msgstr "DB-GPT-SQL-Execution-Plugin" + +#: ../../use_cases/tool_use_with_plugin.md:32 cd491706a64a4bf3af4459fa0699c82e +msgid "This is an DbGPT plugin to connect Generic Db And Execute SQL." +msgstr "这是一个 DbGPT 插件,用于连接通用数据库并执行 SQL。" + +#: ../../use_cases/tool_use_with_plugin.md:35 73dbe0fb80b44b9d9c2928cbf161f291 +msgid "DB-GPT-Bytebase-Plugin" +msgstr "DB-GPT-Bytebase-Plugin" + +#: ../../use_cases/tool_use_with_plugin.md:37 72f675332ff04297861394e8eb2cf5c4 +msgid "" +"To use a tool or platform plugin, you should first deploy a plugin. " +"Taking the open-source database management platform Bytebase as an " +"example, you can deploy your Bytebase service with one click using Docker" +" and access it at http://127.0.0.1:5678. More details can be found at " +"https://github.com/bytebase/bytebase." +msgstr "要使用一个工具或平台插件,您应该首先部署一个插件。以开源数据库管理平台Bytebase为例,您可以使用Docker一键部署Bytebase服务,并通过http://127.0.0.1:5678进行访问。更多细节可以在 https://github.com/bytebase/bytebase 找到。" + +#: ../../use_cases/tool_use_with_plugin.md:53 0d84f8fabc29479caa0bf630c00283d2 +msgid "" +"Note: If your machine's CPU architecture is `ARM`, please use `--platform" +" linux/arm64` instead." +msgstr "备注:如果你的机器CPU架构是ARM,请使用--platform linux/arm64 代替" + +#: ../../use_cases/tool_use_with_plugin.md:55 d4117072f9504875a765c558f55f2d88 +msgid "" +"Select the plugin on DB-GPT(All built-in plugins are from our repository:" +" https://github.com/csunny/DB-GPT-Plugins),choose DB-GPT-Bytebase-Plugin." +" Supporting functions include creating projects, creating environments, " +"creating database instances, creating databases, database DDL/DML " +"operations, and ticket approval process, etc." +msgstr "在DB-GPT上选择插件(所有内置插件均来自我们的仓库:https://github.com/csunny/DB-GPT-Plugins),选择DB-GPT-Bytebase-Plugin。支持的功能包括创建项目、创建环境、创建数据库实例、创建数据库、数据库DDL/DML操作和审批流程等。" +