diff --git a/docs/locales/zh_CN/LC_MESSAGES/getting_started/getting_started.po b/docs/locales/zh_CN/LC_MESSAGES/getting_started/getting_started.po index 683eb6938..274fe54bd 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/getting_started/getting_started.po +++ b/docs/locales/zh_CN/LC_MESSAGES/getting_started/getting_started.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 0.1.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-06-14 17:26+0800\n" +"POT-Creation-Date: 2023-06-14 21:47+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -19,29 +19,29 @@ msgstr "" "Content-Transfer-Encoding: 8bit\n" "Generated-By: Babel 2.12.1\n" -#: ../../getting_started/getting_started.md:1 a0477412435c4c569cf71d243d2884c7 +#: ../../getting_started/getting_started.md:1 d1c1cb0cdf374e60924001460f369485 msgid "Quickstart Guide" msgstr "使用指南" -#: ../../getting_started/getting_started.md:3 331f8c3fbbac44c1b75d2ff595c0235f +#: ../../getting_started/getting_started.md:3 5c76cdb6530644ed872329ecc1bd51ec msgid "" "This tutorial gives you a quick walkthrough about use DB-GPT with you " "environment and data." msgstr "本教程为您提供了关于如何使用DB-GPT的使用指南。" -#: ../../getting_started/getting_started.md:5 df41fe97067d4ba680e3231f05a843de +#: ../../getting_started/getting_started.md:5 8c442e4870e549359920ec83d4a77083 msgid "Installation" msgstr "安装" -#: ../../getting_started/getting_started.md:7 73f72c06a89341f38f4bfbe70ed0d2ae +#: ../../getting_started/getting_started.md:7 d302009cb3f64959872d278c4aad7cfa msgid "To get started, install DB-GPT with the following steps." msgstr "请按照以下步骤安装DB-GPT" -#: ../../getting_started/getting_started.md:9 59c159fd30104ba081e2ebbf1605fe11 +#: ../../getting_started/getting_started.md:9 7deb38572ec74f5392ba09749a2b350b msgid "1. Hardware Requirements" msgstr "1. 硬件要求" -#: ../../getting_started/getting_started.md:10 126558467caa4b05bea0bb051c864831 +#: ../../getting_started/getting_started.md:10 7ee61b468637478cad173fa4685ef952 msgid "" "As our project has the ability to achieve ChatGPT performance of over " "85%, there are certain hardware requirements. However, overall, the " @@ -49,62 +49,62 @@ msgid "" "specific hardware requirements for deployment are as follows:" msgstr "由于我们的项目有能力达到85%以上的ChatGPT性能,所以对硬件有一定的要求。但总体来说,我们在消费级的显卡上即可完成项目的部署使用,具体部署的硬件说明如下:" -#: ../../getting_started/getting_started.md 967e5cde5c6241edae9e0e0e0b217221 +#: ../../getting_started/getting_started.md e9f0871662384ab9a24711856a27fdfb msgid "GPU" msgstr "GPU" -#: ../../getting_started/getting_started.md 05be5479b15b403c8aa3374ea53feff8 +#: ../../getting_started/getting_started.md 0dca27585b9c4357bd4f7b52ff664673 msgid "VRAM Size" msgstr "显存大小" -#: ../../getting_started/getting_started.md ea646d956e834a75912780789946bb47 +#: ../../getting_started/getting_started.md fe84cb97a226490eb940dfcf6e581272 msgid "Performance" msgstr "显存大小" -#: ../../getting_started/getting_started.md aba96df6b91441b982128426eb3a2ebb +#: ../../getting_started/getting_started.md bfdc630854674c5db6be46114a67542d msgid "RTX 4090" msgstr "RTX 4090" -#: ../../getting_started/getting_started.md d2797ffe3d534460ae77a3397fc07c1c -#: e1b5f1a7502e475b872eebd75972a87f +#: ../../getting_started/getting_started.md 5e38a7184e024b09be3a858084b60344 +#: 629db58b707c48a2ae93d7396bcd0d67 msgid "24 GB" msgstr "24 GB" -#: ../../getting_started/getting_started.md 4ed99c06dc4643c290b5914142ea8371 +#: ../../getting_started/getting_started.md 625714d5b7cb4550a81305b5c2410980 msgid "Smooth conversation inference" msgstr "可以流畅的进行对话推理,无卡顿" -#: ../../getting_started/getting_started.md b3dc189389cd41f2883acfe3fad3d6a4 +#: ../../getting_started/getting_started.md 8334aa5646b84a4ba6c0df7a55e52f6e msgid "RTX 3090" msgstr "RTX 3090" -#: ../../getting_started/getting_started.md 0bea7ccef4154e2994144f81af877919 +#: ../../getting_started/getting_started.md 84771e5190084d6fab19fa8f3b5e2a30 msgid "Smooth conversation inference, better than V100" msgstr "可以流畅进行对话推理,有卡顿感,但好于V100" -#: ../../getting_started/getting_started.md 7d1c9e5c16184f2bacf8b12d6f38f629 +#: ../../getting_started/getting_started.md 0a539471bf6648e5827d9e10549b81e3 msgid "V100" msgstr "V100" -#: ../../getting_started/getting_started.md e9ddddf92e15420f85852cea1a1bbd8d +#: ../../getting_started/getting_started.md 8bed4645111a4001b5967678a54c6037 msgid "16 GB" msgstr "16 GB" -#: ../../getting_started/getting_started.md f5532da60b99495c8329d674749ae79f +#: ../../getting_started/getting_started.md f3ec49d1591d4cdc9f967c9df5bb8245 msgid "Conversation inference possible, noticeable stutter" msgstr "可以进行对话推理,有明显卡顿" -#: ../../getting_started/getting_started.md:18 3357d10704b94249b8ccdf7fd3645624 +#: ../../getting_started/getting_started.md:18 6006a3d8744746dbab615b438eb6234b msgid "2. Install" msgstr "2. 安装" -#: ../../getting_started/getting_started.md:20 08e5054dd0fa4e07aa92236dca03a1d3 +#: ../../getting_started/getting_started.md:20 eea3037d218843b78e56412490ae6a62 msgid "" "This project relies on a local MySQL database service, which you need to " "install locally. We recommend using Docker for installation." msgstr "本项目依赖一个本地的 MySQL 数据库服务,你需要本地安装,推荐直接使用 Docker 安装。" -#: ../../getting_started/getting_started.md:25 f99a1af073e24b339390400251a50c9b +#: ../../getting_started/getting_started.md:25 4a220bf247c549eaa0d059f29e1c3a7d msgid "" "We use [Chroma embedding database](https://github.com/chroma-core/chroma)" " as the default for our vector database, so there is no need for special " @@ -117,7 +117,7 @@ msgstr "" "向量数据库我们默认使用的是Chroma内存数据库,所以无需特殊安装,如果有需要连接其他的同学,可以按照我们的教程进行安装配置。整个DB-" "GPT的安装过程,我们使用的是miniconda3的虚拟环境。创建虚拟环境,并安装python依赖包" -#: ../../getting_started/getting_started.md:35 d8ebdf7c4ac54113be1c94ed879dc93f +#: ../../getting_started/getting_started.md:35 b51c85b6ec0f4c45afc648d98424a79f msgid "" "Once the environment is installed, we have to create a new folder " "\"models\" in the DB-GPT project, and then we can put all the models " @@ -126,24 +126,24 @@ msgstr "" "环境安装完成后,我们必须在DB-" "GPT项目中创建一个新文件夹\"models\",然后我们可以把从huggingface下载的所有模型放到这个目录下。" -#: ../../getting_started/getting_started.md:42 cd3ce56a4d644574a0c30dd86148a58c +#: ../../getting_started/getting_started.md:43 617f2b53e33e4e7a96b3aef879a1ebe7 msgid "" "The model files are large and will take a long time to download. During " "the download, let's configure the .env file, which needs to be copied and" " created from the .env.template" msgstr "模型文件很大,需要很长时间才能下载。在下载过程中,让我们配置.env文件,它需要从。env.template中复制和创建。" -#: ../../getting_started/getting_started.md:48 96d634498ae04c84b4eae8502d5f65e8 +#: ../../getting_started/getting_started.md:49 0f403de4e9574c3a9058495df8c21961 msgid "" "You can configure basic parameters in the .env file, for example setting " "LLM_MODEL to the model to be used" msgstr "您可以在.env文件中配置基本参数,例如将LLM_MODEL设置为要使用的模型。" -#: ../../getting_started/getting_started.md:50 a78627885d2a41a481131308d2d061a6 +#: ../../getting_started/getting_started.md:51 e3dce051a2e64478872433890b06cb5d msgid "3. Run" msgstr "3. 运行" -#: ../../getting_started/getting_started.md:51 69c1d06867cb4a0ea36ad0f3686109db +#: ../../getting_started/getting_started.md:52 b7ec5fab25b249b5bc811d08049307c3 msgid "" "You can refer to this document to obtain the Vicuna weights: " "[Vicuna](https://github.com/lm-sys/FastChat/blob/main/README.md#model-" @@ -152,7 +152,7 @@ msgstr "" "关于基础模型, 可以根据[Vicuna](https://github.com/lm-" "sys/FastChat/blob/main/README.md#model-weights) 合成教程进行合成。" -#: ../../getting_started/getting_started.md:53 1dcccc1784e04fa3a3340aa446f1484f +#: ../../getting_started/getting_started.md:54 6cab130775394122bfb040ea9797c694 msgid "" "If you have difficulty with this step, you can also directly use the " "model from [this link](https://huggingface.co/Tribbiani/vicuna-7b) as a " @@ -161,21 +161,21 @@ msgstr "" "如果此步有困难的同学,也可以直接使用[此链接](https://huggingface.co/Tribbiani/vicuna-" "7b)上的模型进行替代。" -#: ../../getting_started/getting_started.md:55 45a4c9f50ee04533bd57652505ab4f62 +#: ../../getting_started/getting_started.md:56 381fc71965dd44adac2141677e0dd085 msgid "Run server" msgstr "运行模型服务" -#: ../../getting_started/getting_started.md:60 f3482509ad2f4137becd4775768180bd +#: ../../getting_started/getting_started.md:61 47001c4d9f9449fab0dadefd02b76f6d msgid "" "Starting `llmserver.py` with the following command will result in a " "relatively stable Python service with multiple processes." -msgstr "使用以下命令启动llmserver.py将会得到一个相对稳定的Python服务,并且具有多个进程。" +msgstr "" -#: ../../getting_started/getting_started.md:65 ddf718f6863f4e58b8de1232dc8189dd +#: ../../getting_started/getting_started.md:66 ae3b8aa1e8694715a721717ab4bc182e msgid "Run gradio webui" msgstr "运行模型服务" -#: ../../getting_started/getting_started.md:71 d3df863c46ae49b1a63e5778d439d336 +#: ../../getting_started/getting_started.md:72 0ad4bbedf5ed498686dafb8b148bf63c msgid "" "Notice: the webserver need to connect llmserver, so you need change the" " .env file. change the MODEL_SERVER = \"http://127.0.0.1:8000\" to your " @@ -184,3 +184,10 @@ msgstr "" "注意: 在启动Webserver之前, 需要修改.env 文件中的MODEL_SERVER = " "\"http://127.0.0.1:8000\", 将地址设置为你的服务器地址。" +#~ msgid "" +#~ "Starting `llmserver.py` with the following " +#~ "command will result in a relatively " +#~ "stable Python service with multiple " +#~ "processes." +#~ msgstr "使用以下命令启动llmserver.py将会得到一个相对稳定的Python服务,并且具有多个进程。" + diff --git a/docs/locales/zh_CN/LC_MESSAGES/modules/llms.po b/docs/locales/zh_CN/LC_MESSAGES/modules/llms.po index 22138dec3..8d28b1a13 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/modules/llms.po +++ b/docs/locales/zh_CN/LC_MESSAGES/modules/llms.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 0.1.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-06-14 17:26+0800\n" +"POT-Creation-Date: 2023-06-14 21:47+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -19,11 +19,11 @@ msgstr "" "Content-Transfer-Encoding: 8bit\n" "Generated-By: Babel 2.12.1\n" -#: ../../modules/llms.md:1 8eef439964b5442d91ad04ff72b3b45b +#: ../../modules/llms.md:1 d924f8db422d4adaa4cf0e5d17ffe332 msgid "LLMs" msgstr "大语言模型" -#: ../../modules/llms.md:3 a29256f1f39b4bcda29a6811ad1b10f6 +#: ../../modules/llms.md:3 0fba9725074a4a72be1849e822e69365 #, python-format msgid "" "In the underlying large model integration, we have designed an open " @@ -36,23 +36,23 @@ msgid "" "of use." msgstr "在底层大模型接入中,我们设计了开放的接口,支持对接多种大模型。同时对于接入模型的效果,我们有非常严格的把控与评审机制。对大模型能力上与ChatGPT对比,在准确率上需要满足85%以上的能力对齐。我们用更高的标准筛选模型,是期望在用户使用过程中,可以省去前面繁琐的测试评估环节。" -#: ../../modules/llms.md:5 e899f11399bb45d9990c73a273ed7697 +#: ../../modules/llms.md:5 4f65b4e86d8d4b35b95ccac6a96f9b9c msgid "Multi LLMs Usage" msgstr "多模型使用" -#: ../../modules/llms.md:6 a21d6a875b3949b9be512f8ea396f6b3 +#: ../../modules/llms.md:6 246e4dbf1c634d9d9f2ab40b4883de73 msgid "" "To use multiple models, modify the LLM_MODEL parameter in the .env " "configuration file to switch between the models." msgstr "如果要使用不同的模型,请修改.env配置文件中的LLM MODEL参数以在模型之间切换。" -#: ../../modules/llms.md:8 dd061d45bb4044dcbc7e5d4b0014ded8 +#: ../../modules/llms.md:8 379d7756094f4ceda5ef163031b44e97 msgid "" "Notice: you can create .env file from .env.template, just use command " "like this:" msgstr "注意:你可以从 .env.template 创建 .env 文件。只需使用如下命令:" -#: ../../modules/llms.md:14 c1a789e2de2c4958987370521d46c7cc +#: ../../modules/llms.md:14 3b46be1be08b41b0a4f2fa7ca1f09ccf msgid "" "now we support models vicuna-13b, vicuna-7b, chatglm-6b, flan-t5-base, " "guanaco-33b-merged, falcon-40b, gorilla-7b." @@ -60,7 +60,7 @@ msgstr "" "现在我们支持的模型有vicuna-13b, vicuna-7b, chatglm-6b, flan-t5-base, guanaco-33b-" "merged, falcon-40b, gorilla-7b." -#: ../../modules/llms.md:16 ddb6f2f638e642a595365a91ffdba8f9 +#: ../../modules/llms.md:16 f7e26b7e7a8349159b5c2e8ec2a1abc8 msgid "" "DB-GPT provides a model load adapter and chat adapter. load adapter which" " allows you to easily adapt load different LLM models by inheriting the " @@ -69,15 +69,15 @@ msgstr "" "DB-GPT提供了多模型适配器load adapter和chat adapter.load adapter通过继承BaseLLMAdapter类," " 实现match和loader方法允许你适配不同的LLM." -#: ../../modules/llms.md:18 5e32be54895243caa6d44d0b3421e4a0 +#: ../../modules/llms.md:18 83ad29017489428cad336894de9b1076 msgid "vicuna llm load adapter" msgstr "vicuna llm load adapter" -#: ../../modules/llms.md:35 0b1f2e7c65164c9584e0c544394e7d57 +#: ../../modules/llms.md:35 56997156cf994abbb12a586d435b1f6c msgid "chatglm load adapter" msgstr "chatglm load adapter" -#: ../../modules/llms.md:62 885b35375c764e29a983b54514e378d2 +#: ../../modules/llms.md:62 e0d989155ca443079402921fd15c1b75 msgid "" "chat adapter which allows you to easily adapt chat different LLM models " "by inheriting the BaseChatAdpter.you just implement match() and " @@ -86,45 +86,73 @@ msgstr "" "chat " "adapter通过继承BaseChatAdpter允许你通过实现match和get_generate_stream_func方法允许你适配不同的LLM." -#: ../../modules/llms.md:64 e0538e7e0526440085b32add07f5ec7f +#: ../../modules/llms.md:64 1cef2736bdd04dc89a3f9f20d6ab7e4e msgid "vicuna llm chat adapter" msgstr "vicuna llm chat adapter" -#: ../../modules/llms.md:76 2c97712441874e0d8deedc1d9a1ce5ed +#: ../../modules/llms.md:76 0976f3a966b24e458b968c4410e98403 msgid "chatglm llm chat adapter" msgstr "chatglm llm chat adapter" -#: ../../modules/llms.md:89 485e5aa261714146a03a30dbcd612653 +#: ../../modules/llms.md:89 138577d6f1714b4daef8b20108de983d msgid "" "if you want to integrate your own model, just need to inheriting " "BaseLLMAdaper and BaseChatAdpter and implement the methods" msgstr "如果你想集成自己的模型,只需要继承BaseLLMAdaper和BaseChatAdpter类,然后实现里面的方法即可" -#: ../../modules/llms.md:92 a63b63022db74d76b743044be178e227 +#: ../../modules/llms.md:92 23cd5154bf3a4e7bad206d5e06eef51e #, fuzzy msgid "Multi Proxy LLMs" msgstr "多模型使用" -#: ../../modules/llms.md:93 dab3041e90384049872a7f77933b1a1f +#: ../../modules/llms.md:93 db2739caa40342819965856d6fe83677 msgid "1. Openai proxy" -msgstr "Openai代理" +msgstr "" -#: ../../modules/llms.md:94 e50eae200bf04e4788bbc394e0b3d6b9 +#: ../../modules/llms.md:94 5a305d769a144543b14f6e7099d0fc81 msgid "" "If you haven't deployed a private infrastructure for a large model, or if" " you want to use DB-GPT in a low-cost and high-efficiency way, you can " "also use OpenAI's large model as your underlying model." -msgstr "如果你没有部署私有大模型的资源,或者你想使用低成本启动DB-GPT,你可以使用openai的大模型作为你的底层模型" +msgstr "" -#: ../../modules/llms.md:96 53dd581608d74355ba0ce486a01ef261 +#: ../../modules/llms.md:96 64d78d8e95174a329f67e638fc63786c msgid "" "If your environment deploying DB-GPT has access to OpenAI, then modify " "the .env configuration file as below will work." -msgstr "如果你的环境能够访问openai,你只需要参考如下修改.env配置文件即可" +msgstr "" -#: ../../modules/llms.md:104 8df2d75af41b4953a73b6b7eae9f0373 +#: ../../modules/llms.md:104 f2bcf0c04da844e0b8fa2a3712d56439 msgid "" "If you can't access OpenAI locally but have an OpenAI proxy service, you " "can configure as follows." -msgstr "如果你本地无法访问openai,但是你有一个openai的代理服务,你可以参考如下配置" +msgstr "" + +#~ msgid "Multi Proxy LLMs" +#~ msgstr "多模型使用" + +#~ msgid "1. Openai proxy" +#~ msgstr "Openai代理" + +#~ msgid "" +#~ "If you haven't deployed a private " +#~ "infrastructure for a large model, or " +#~ "if you want to use DB-GPT in" +#~ " a low-cost and high-efficiency " +#~ "way, you can also use OpenAI's " +#~ "large model as your underlying model." +#~ msgstr "如果你没有部署私有大模型的资源,或者你想使用低成本启动DB-GPT,你可以使用openai的大模型作为你的底层模型" + +#~ msgid "" +#~ "If your environment deploying DB-GPT " +#~ "has access to OpenAI, then modify " +#~ "the .env configuration file as below " +#~ "will work." +#~ msgstr "如果你的环境能够访问openai,你只需要参考如下修改.env配置文件即可" + +#~ msgid "" +#~ "If you can't access OpenAI locally " +#~ "but have an OpenAI proxy service, " +#~ "you can configure as follows." +#~ msgstr "如果你本地无法访问openai,但是你有一个openai的代理服务,你可以参考如下配置" diff --git a/docs/locales/zh_CN/LC_MESSAGES/modules/plugins.po b/docs/locales/zh_CN/LC_MESSAGES/modules/plugins.po index dce190384..e02da98ea 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/modules/plugins.po +++ b/docs/locales/zh_CN/LC_MESSAGES/modules/plugins.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 0.1.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-06-14 17:26+0800\n" +"POT-Creation-Date: 2023-06-14 21:47+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -19,11 +19,11 @@ msgstr "" "Content-Transfer-Encoding: 8bit\n" "Generated-By: Babel 2.12.1\n" -#: ../../modules/plugins.md:1 be587eb6ad384844b83ac740a2f3309e +#: ../../modules/plugins.md:1 8e0200134cca45b6aead6d05b60ca95a msgid "Plugins" msgstr "插件" -#: ../../modules/plugins.md:3 291a6adc34684b28867b7d1adcceadbb +#: ../../modules/plugins.md:3 d0d532cfe9b44fa0916d7d5b912a744a msgid "" "The ability of Agent and Plugin is the core of whether large models can " "be automated. In this project, we natively support the plugin mode, and " @@ -35,69 +35,138 @@ msgstr "" "Agent与插件能力是大模型能否自动化的核心,在本的项目中,原生支持插件模式,大模型可以自动化完成目标。 同时为了充分发挥社区的优势" ",本项目中所用的插件原生支持Auto-GPT插件生态,即Auto-GPT的插件可以直接在我们的项目中运行。" -#: ../../modules/plugins.md:5 72010cbac395488bb72c5ffa79806f61 +#: ../../modules/plugins.md:5 2f78a6b397a24f34b0d5771ca93efb0b #, fuzzy msgid "Local Plugins" msgstr "插件" -#: ../../modules/plugins.md:7 842e405d6748425dab6684c3377e580a +#: ../../modules/plugins.md:7 54a817a638c3440989191b3bffaca6de msgid "1.1 How to write local plugins." -msgstr "如何编写一个本地插件" +msgstr "" -#: ../../modules/plugins.md:9 23fa64fa76954000a3f33f5b3205975e +#: ../../modules/plugins.md:9 fbdc0a9d327f432aa6a380117dfb2f11 msgid "" "Local plugins use the Auto-GPT plugin template. A simple example is as " "follows: first write a plugin file called \"sql_executor.py\"." -msgstr "本地插件使用Auto-GPT插件模板,一个简单的示例如下:首先编写一个插件文件`sql_executor.py`" +msgstr "" -#: ../../modules/plugins.md:39 5044401271fb4cfba53299875886b5b9 +#: ../../modules/plugins.md:39 dc398ab427bd4d15b3b7c8cb1ff032b3 msgid "" "Then set the \"can_handle_post_prompt\" method of the plugin template to " "True. In the \"post_prompt\" method, write the prompt information and the" " mapped plugin function." -msgstr "然后设置can_handle_post_prompt函数为True, 在post_prompt函数中编写prompt信息和插件映射函数" +msgstr "" -#: ../../modules/plugins.md:81 8f2001c2934d4655a1993a98d5e7dd63 +#: ../../modules/plugins.md:81 c9d4019392bf452e906057cbe9271005 msgid "1.2 How to use local plugins" -msgstr "1.2 如何使用本地插件" +msgstr "" -#: ../../modules/plugins.md:83 aa6234c1531a470db97e076408c70ebc +#: ../../modules/plugins.md:83 9beaed4a71124ecf9544a1dba0d1e722 msgid "" "Pack your plugin project into `your-plugin.zip` and place it in the " "`/plugins/` directory of the DB-GPT project. After starting the " "webserver, you can select and use it in the `Plugin Model` section." -msgstr "将您的插件项目打包成your-plugin.zip,并将其放置在DB-GPT项目的/plugins/目录中。启动Web服务器后,您可以在插件模型部分中选择并使用它。" +msgstr "" -#: ../../modules/plugins.md:86 6d1402451cb44ab2bb9ded2a303d8dd0 +#: ../../modules/plugins.md:86 9a1439c883a947d7acac3fd1196b3c1e #, fuzzy msgid "Public Plugins" msgstr "插件" -#: ../../modules/plugins.md:88 2bb33cadc7604f529c762939a6225f17 +#: ../../modules/plugins.md:88 2ed4c509bf5848adb3fa163752a1cfa1 msgid "1.1 How to use public plugins" -msgstr "1.1 如何编写公共插件" +msgstr "" -#: ../../modules/plugins.md:90 d5f458dcba2e435987ffbc62a1d7a989 +#: ../../modules/plugins.md:90 dd5ba8d582204b2f89ce802a1232b11d msgid "" "By default, after launching the webserver, plugins from the public plugin" " library `DB-GPT-Plugins` will be automatically loaded. For more details," " please refer to [DB-GPT-Plugins](https://github.com/csunny/DB-GPT-" "Plugins)" -msgstr "默认情况下,在启动Web服务器后,将自动加载来自公共插件库DB-GPT-Plugins的插件。要了解更多详情,请参阅[DB-GPT-Plugins](https://github.com/csunny/DB-GPT-Plugins)" +msgstr "" -#: ../../modules/plugins.md:92 28fa983769ae44a0b790d977c60ce982 +#: ../../modules/plugins.md:92 244f0591bc5045eab175754521b414c4 msgid "1.2 Contribute to the DB-GPT-Plugins repository" -msgstr "1.2 贡献到DB-GPT-Plugins仓库" +msgstr "" -#: ../../modules/plugins.md:94 68573e7cc17f479fa676d0631e011baf +#: ../../modules/plugins.md:94 e00bac1a299b46caa19b9cf16709d6ba msgid "" "Please refer to the plugin development process in the public plugin " "library, and put the configuration parameters in `.plugin_env`" -msgstr "请参考公共插件库开发过程,将插件配置参数写入.plugin_env文件" +msgstr "" -#: ../../modules/plugins.md:96 8d6860026b824b46b2899cbf3dc3b4a0 +#: ../../modules/plugins.md:96 315fbf576ea24158adc7b564f53940e0 msgid "" "We warmly welcome everyone to contribute plugins to the public plugin " "library!" -msgstr "非常欢迎大家向我们公共插件库贡献插件!" +msgstr "" + +#~ msgid "Local Plugins" +#~ msgstr "插件" + +#~ msgid "1.1 How to write local plugins." +#~ msgstr "如何编写一个本地插件" + +#~ msgid "" +#~ "Local plugins use the Auto-GPT " +#~ "plugin template. A simple example is " +#~ "as follows: first write a plugin " +#~ "file called \"sql_executor.py\"." +#~ msgstr "本地插件使用Auto-GPT插件模板,一个简单的示例如下:首先编写一个插件文件`sql_executor.py`" + +#~ msgid "" +#~ "Then set the \"can_handle_post_prompt\" method" +#~ " of the plugin template to True. " +#~ "In the \"post_prompt\" method, write the" +#~ " prompt information and the mapped " +#~ "plugin function." +#~ msgstr "然后设置can_handle_post_prompt函数为True, 在post_prompt函数中编写prompt信息和插件映射函数" + +#~ msgid "1.2 How to use local plugins" +#~ msgstr "1.2 如何使用本地插件" + +#~ msgid "" +#~ "Pack your plugin project into `your-" +#~ "plugin.zip` and place it in the " +#~ "`/plugins/` directory of the DB-GPT " +#~ "project. After starting the webserver, " +#~ "you can select and use it in " +#~ "the `Plugin Model` section." +#~ msgstr "" +#~ "将您的插件项目打包成your-plugin.zip,并将其放置在DB-" +#~ "GPT项目的/plugins/目录中。启动Web服务器后,您可以在插件模型部分中选择并使用它。" + +#~ msgid "Public Plugins" +#~ msgstr "插件" + +#~ msgid "1.1 How to use public plugins" +#~ msgstr "1.1 如何编写公共插件" + +#~ msgid "" +#~ "By default, after launching the " +#~ "webserver, plugins from the public " +#~ "plugin library `DB-GPT-Plugins` will " +#~ "be automatically loaded. For more " +#~ "details, please refer to [DB-GPT-" +#~ "Plugins](https://github.com/csunny/DB-GPT-Plugins)" +#~ msgstr "" +#~ "默认情况下,在启动Web服务器后,将自动加载来自公共插件库DB-GPT-Plugins的插件。要了解更多详情,请参阅" +#~ "[DB-GPT-Plugins](https://github.com/csunny/DB-GPT-" +#~ "Plugins)" + +#~ msgid "1.2 Contribute to the DB-GPT-Plugins repository" +#~ msgstr "1.2 贡献到DB-GPT-Plugins仓库" + +#~ msgid "" +#~ "Please refer to the plugin development" +#~ " process in the public plugin " +#~ "library, and put the configuration " +#~ "parameters in `.plugin_env`" +#~ msgstr "请参考公共插件库开发过程,将插件配置参数写入.plugin_env文件" + +#~ msgid "" +#~ "We warmly welcome everyone to contribute" +#~ " plugins to the public plugin " +#~ "library!" +#~ msgstr "非常欢迎大家向我们公共插件库贡献插件!" diff --git a/docs/locales/zh_CN/LC_MESSAGES/use_cases/sql_generation_and_diagnosis.po b/docs/locales/zh_CN/LC_MESSAGES/use_cases/sql_generation_and_diagnosis.po index fd8918782..1cef3f7f8 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/use_cases/sql_generation_and_diagnosis.po +++ b/docs/locales/zh_CN/LC_MESSAGES/use_cases/sql_generation_and_diagnosis.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 0.1.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-06-14 14:51+0800\n" +"POT-Creation-Date: 2023-06-14 21:47+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -37,6 +37,15 @@ msgstr "" #: ../../use_cases/sql_generation_and_diagnosis.md:5 #: ffcbe5278cdc41a7b656f03bf82a8b26 -msgid "[SQL Generation](../../assets/demo_en.gif)" +msgid "![SQL Generation](../../assets/demo_en.gif)" msgstr "" +#: ../../use_cases/sql_generation_and_diagnosis.md:5 +#: 3e5a3a7db4ee4cc8bff28c3f2c7387dd +#, fuzzy +msgid "SQL Generation" +msgstr "SQL生成和诊断" + +#~ msgid "![SQL Generation](../../assets/demo_en.gif)" +#~ msgstr "![SQL 生成](../../assets/demo_en.gif)" + diff --git a/docs/locales/zh_CN/LC_MESSAGES/use_cases/tool_use_with_plugin.po b/docs/locales/zh_CN/LC_MESSAGES/use_cases/tool_use_with_plugin.po index 383b62b85..e5923c0ad 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/use_cases/tool_use_with_plugin.po +++ b/docs/locales/zh_CN/LC_MESSAGES/use_cases/tool_use_with_plugin.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 0.1.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-06-14 17:26+0800\n" +"POT-Creation-Date: 2023-06-14 21:47+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -19,86 +19,181 @@ msgstr "" "Content-Transfer-Encoding: 8bit\n" "Generated-By: Babel 2.12.1\n" -#: ../../use_cases/tool_use_with_plugin.md:1 defad8b142a2408e89b782803149be50 +#: ../../use_cases/tool_use_with_plugin.md:1 584817fdb00047de8f8d7ae02ce86783 msgid "Tool use with plugin" msgstr "插件工具" -#: ../../use_cases/tool_use_with_plugin.md:3 30ca5f2ce55a4369b48dd8a8b08c3273 +#: ../../use_cases/tool_use_with_plugin.md:3 74d688e857ee4afe9237aa959238d3df msgid "" "DB-GPT supports a variety of plug-ins, such as MySQL, MongoDB, ClickHouse" " and other database tool plug-ins. In addition, some database management " "platforms can also package their interfaces and package them into plug-" "ins, and use the model to realize the ability of \"single-sentence " "requirements\"" -msgstr "DB-GPT支持各种插件,例如MySQL、MongoDB、ClickHouse等数据库工具插件。此外,一些数据库管理平台也可以将它们的接口打包成插件,使用该模型实现"一句话需求"的能力。" +msgstr "" -#: ../../use_cases/tool_use_with_plugin.md:6 96c20363d25842cbbab21819c2567a52 +#: ../../use_cases/tool_use_with_plugin.md:6 55754e6c89d149cd9eb5f935fd9dc761 msgid "DB-GPT-DASHBOARD-PLUGIN" -msgstr "DB-GPT-DASHBOARD-PLUGIN" +msgstr "" -#: ../../use_cases/tool_use_with_plugin.md:8 1a78f9e15c6949019e213f3f88388e9f +#: ../../use_cases/tool_use_with_plugin.md:8 d3c0287afa81409f9bda6fc495d63917 msgid "" -"[Db-GPT Chart Plugin](https://github.com/csunny/DB-GPT-" +"[](https://github.com/csunny/DB-GPT-" "Plugins/blob/main/src/dbgpt_plugins/Readme.md)" msgstr "" -"[Db-GPT Chart Plugin](https://github.com/csunny/DB-GPT-" -"Plugins/blob/main/src/dbgpt_plugins/Readme.md)" -#: ../../use_cases/tool_use_with_plugin.md:10 5ac6d662797e4694926cb0b44e58ff40 +#: ../../use_cases/tool_use_with_plugin.md:10 a65c05f21ee94e8da1f14076dbed8123 msgid "" "This is a DB-GPT plugin to generate data analysis charts, if you want to " "use the test sample data, please first pull the code of [DB-GPT-" "Plugins](https://github.com/csunny/DB-GPT-Plugins), run the command to " "generate test DuckDB data, and then copy the generated data file to the " "`/pilot/mock_datas` directory of the DB-GPT project." -msgstr "这是一个DB-GPT插件,用于生成数据分析图表。如果您想使用测试样本数据,请先拉取 DB-GPT-Plugins 的代码,运行命令以生成测试 DuckDB 数据,然后将生成的数据文件复制到 DB-GPT 项目的 /pilot/mock_datas 目录中。" +msgstr "" -#: ../../use_cases/tool_use_with_plugin.md:21 1029537ed0ca44e499a8f2098cc72f1a +#: ../../use_cases/tool_use_with_plugin.md:21 c25ef922010442f5be632f6d8f2e730c msgid "" "Test Case: Use a histogram to analyze the total order amount of users in " "different cities." -msgstr "测试示例:请使用柱状图分析各个城市的用户数" +msgstr "" -#: ../../use_cases/tool_use_with_plugin.md:26 a8a2103279d641b98948bcb06971c160 +#: ../../use_cases/tool_use_with_plugin.md:26 3f07d6e71ced4011998b1f1fda640394 msgid "" "More detail see: [DB-DASHBOARD](https://github.com/csunny/DB-GPT-" "Plugins/blob/main/src/dbgpt_plugins/Readme.md)" -msgstr "更多详情请看:[DB-DASHBOARD](https://github.com/csunny/DB-GPT-" -"Plugins/blob/main/src/dbgpt_plugins/Readme.md)" +msgstr "" -#: ../../use_cases/tool_use_with_plugin.md:29 db2906b5f89e4a01a4a47a89ad750804 +#: ../../use_cases/tool_use_with_plugin.md:29 20e5d3aed30847ccac905d0d5268824f msgid "DB-GPT-SQL-Execution-Plugin" -msgstr "DB-GPT-SQL-Execution-Plugin" +msgstr "" -#: ../../use_cases/tool_use_with_plugin.md:32 cd491706a64a4bf3af4459fa0699c82e +#: ../../use_cases/tool_use_with_plugin.md:32 4ebfd33a77e547edb1de9d3159745cb6 msgid "This is an DbGPT plugin to connect Generic Db And Execute SQL." -msgstr "这是一个 DbGPT 插件,用于连接通用数据库并执行 SQL。" +msgstr "" -#: ../../use_cases/tool_use_with_plugin.md:35 73dbe0fb80b44b9d9c2928cbf161f291 +#: ../../use_cases/tool_use_with_plugin.md:35 8c11ec372d9346e79e5ebba390b15919 msgid "DB-GPT-Bytebase-Plugin" -msgstr "DB-GPT-Bytebase-Plugin" +msgstr "" -#: ../../use_cases/tool_use_with_plugin.md:37 72f675332ff04297861394e8eb2cf5c4 +#: ../../use_cases/tool_use_with_plugin.md:37 b01eb72df51648a293613dbab2bbe4f0 msgid "" "To use a tool or platform plugin, you should first deploy a plugin. " "Taking the open-source database management platform Bytebase as an " "example, you can deploy your Bytebase service with one click using Docker" " and access it at http://127.0.0.1:5678. More details can be found at " "https://github.com/bytebase/bytebase." -msgstr "要使用一个工具或平台插件,您应该首先部署一个插件。以开源数据库管理平台Bytebase为例,您可以使用Docker一键部署Bytebase服务,并通过http://127.0.0.1:5678进行访问。更多细节可以在 https://github.com/bytebase/bytebase 找到。" +msgstr "" -#: ../../use_cases/tool_use_with_plugin.md:53 0d84f8fabc29479caa0bf630c00283d2 +#: ../../use_cases/tool_use_with_plugin.md:53 1cdcd5fc42b6433ba5573fc157328c5c msgid "" "Note: If your machine's CPU architecture is `ARM`, please use `--platform" " linux/arm64` instead." -msgstr "备注:如果你的机器CPU架构是ARM,请使用--platform linux/arm64 代替" +msgstr "" -#: ../../use_cases/tool_use_with_plugin.md:55 d4117072f9504875a765c558f55f2d88 +#: ../../use_cases/tool_use_with_plugin.md:55 179dc86ad25f4498af7c90f570f1a556 msgid "" "Select the plugin on DB-GPT(All built-in plugins are from our repository:" " https://github.com/csunny/DB-GPT-Plugins),choose DB-GPT-Bytebase-Plugin." " Supporting functions include creating projects, creating environments, " "creating database instances, creating databases, database DDL/DML " "operations, and ticket approval process, etc." -msgstr "在DB-GPT上选择插件(所有内置插件均来自我们的仓库:https://github.com/csunny/DB-GPT-Plugins),选择DB-GPT-Bytebase-Plugin。支持的功能包括创建项目、创建环境、创建数据库实例、创建数据库、数据库DDL/DML操作和审批流程等。" +msgstr "" + +#~ msgid "" +#~ "DB-GPT supports a variety of " +#~ "plug-ins, such as MySQL, MongoDB, " +#~ "ClickHouse and other database tool " +#~ "plug-ins. In addition, some database " +#~ "management platforms can also package " +#~ "their interfaces and package them into" +#~ " plug-ins, and use the model to" +#~ " realize the ability of \"single-" +#~ "sentence requirements\"" +#~ msgstr "" +#~ "DB-" +#~ "GPT支持各种插件,例如MySQL、MongoDB、ClickHouse等数据库工具插件。此外,一些数据库管理平台也可以将它们的接口打包成插件,使用该模型实现\"一句话需求\"的能力。" + +#~ msgid "DB-GPT-DASHBOARD-PLUGIN" +#~ msgstr "DB-GPT-DASHBOARD-PLUGIN" + +#~ msgid "" +#~ "[Db-GPT Chart Plugin](https://github.com/csunny" +#~ "/DB-GPT-Plugins/blob/main/src/dbgpt_plugins/Readme.md)" +#~ msgstr "" +#~ "[Db-GPT Chart Plugin](https://github.com/csunny" +#~ "/DB-GPT-Plugins/blob/main/src/dbgpt_plugins/Readme.md)" + +#~ msgid "" +#~ "This is a DB-GPT plugin to " +#~ "generate data analysis charts, if you" +#~ " want to use the test sample " +#~ "data, please first pull the code " +#~ "of [DB-GPT-Plugins](https://github.com/csunny/DB-" +#~ "GPT-Plugins), run the command to " +#~ "generate test DuckDB data, and then " +#~ "copy the generated data file to " +#~ "the `/pilot/mock_datas` directory of the " +#~ "DB-GPT project." +#~ msgstr "" +#~ "这是一个DB-GPT插件,用于生成数据分析图表。如果您想使用测试样本数据,请先拉取 DB-GPT-" +#~ "Plugins 的代码,运行命令以生成测试 DuckDB 数据,然后将生成的数据文件复制到 " +#~ "DB-GPT 项目的 /pilot/mock_datas 目录中。" + +#~ msgid "" +#~ "Test Case: Use a histogram to " +#~ "analyze the total order amount of " +#~ "users in different cities." +#~ msgstr "测试示例:请使用柱状图分析各个城市的用户数" + +#~ msgid "" +#~ "More detail see: [DB-" +#~ "DASHBOARD](https://github.com/csunny/DB-GPT-" +#~ "Plugins/blob/main/src/dbgpt_plugins/Readme.md)" +#~ msgstr "" +#~ "更多详情请看:[DB-DASHBOARD](https://github.com/csunny/DB-GPT-" +#~ "Plugins/blob/main/src/dbgpt_plugins/Readme.md)" + +#~ msgid "DB-GPT-SQL-Execution-Plugin" +#~ msgstr "DB-GPT-SQL-Execution-Plugin" + +#~ msgid "This is an DbGPT plugin to connect Generic Db And Execute SQL." +#~ msgstr "这是一个 DbGPT 插件,用于连接通用数据库并执行 SQL。" + +#~ msgid "DB-GPT-Bytebase-Plugin" +#~ msgstr "DB-GPT-Bytebase-Plugin" + +#~ msgid "" +#~ "To use a tool or platform plugin," +#~ " you should first deploy a plugin." +#~ " Taking the open-source database " +#~ "management platform Bytebase as an " +#~ "example, you can deploy your Bytebase" +#~ " service with one click using Docker" +#~ " and access it at http://127.0.0.1:5678." +#~ " More details can be found at " +#~ "https://github.com/bytebase/bytebase." +#~ msgstr "" +#~ "要使用一个工具或平台插件,您应该首先部署一个插件。以开源数据库管理平台Bytebase为例,您可以使用Docker一键部署Bytebase服务,并通过http://127.0.0.1:5678进行访问。更多细节可以在" +#~ " https://github.com/bytebase/bytebase 找到。" + +#~ msgid "" +#~ "Note: If your machine's CPU architecture" +#~ " is `ARM`, please use `--platform " +#~ "linux/arm64` instead." +#~ msgstr "备注:如果你的机器CPU架构是ARM,请使用--platform linux/arm64 代替" + +#~ msgid "" +#~ "Select the plugin on DB-GPT(All " +#~ "built-in plugins are from our " +#~ "repository: https://github.com/csunny/DB-GPT-" +#~ "Plugins),choose DB-GPT-Bytebase-Plugin. " +#~ "Supporting functions include creating " +#~ "projects, creating environments, creating " +#~ "database instances, creating databases, " +#~ "database DDL/DML operations, and ticket " +#~ "approval process, etc." +#~ msgstr "" +#~ "在DB-GPT上选择插件(所有内置插件均来自我们的仓库:https://github.com/csunny/DB-" +#~ "GPT-Plugins),选择DB-GPT-Bytebase-" +#~ "Plugin。支持的功能包括创建项目、创建环境、创建数据库实例、创建数据库、数据库DDL/DML操作和审批流程等。" diff --git a/docs/use_cases/sql_generation_and_diagnosis.md b/docs/use_cases/sql_generation_and_diagnosis.md index d608ce1ec..b0282989f 100644 --- a/docs/use_cases/sql_generation_and_diagnosis.md +++ b/docs/use_cases/sql_generation_and_diagnosis.md @@ -2,4 +2,4 @@ DB-GPT provides SQL generation and diagnostic capabilities. With advanced natural language processing algorithms and a deep understanding of SQL syntax and semantics, the model can generate complex SQL queries from natural language input, as well as diagnose and fix issues with existing queries. This allows users to save time and streamline their workflow, while also improving the accuracy and efficiency of their SQL operations. -[SQL Generation](../../assets/demo_en.gif) \ No newline at end of file +![SQL Generation](../../assets/demo_en.gif) \ No newline at end of file diff --git a/pilot/common/plugins.py b/pilot/common/plugins.py index 832144d22..e22224399 100644 --- a/pilot/common/plugins.py +++ b/pilot/common/plugins.py @@ -78,6 +78,7 @@ def load_native_plugins(cfg: Config): if not cfg.plugins_auto_load: print("not auto load_native_plugins") return + def load_from_git(cfg: Config): print("async load_native_plugins") branch_name = cfg.plugins_git_branch @@ -85,16 +86,20 @@ def load_native_plugins(cfg: Config): url = "https://github.com/csunny/{repo}/archive/{branch}.zip" try: session = requests.Session() - response = session.get(url.format(repo=native_plugin_repo, branch=branch_name), - headers={'Authorization': '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'}) + response = session.get( + url.format(repo=native_plugin_repo, branch=branch_name), + headers={"Authorization": "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"}, + ) if response.status_code == 200: plugins_path_path = Path(PLUGINS_DIR) - files = glob.glob(os.path.join(plugins_path_path, f'{native_plugin_repo}*')) + files = glob.glob( + os.path.join(plugins_path_path, f"{native_plugin_repo}*") + ) for file in files: os.remove(file) now = datetime.datetime.now() - time_str = now.strftime('%Y%m%d%H%M%S') + time_str = now.strftime("%Y%m%d%H%M%S") file_name = f"{plugins_path_path}/{native_plugin_repo}-{branch_name}-{time_str}.zip" print(file_name) with open(file_name, "wb") as f: @@ -110,7 +115,6 @@ def load_native_plugins(cfg: Config): t.start() - def scan_plugins(cfg: Config, debug: bool = False) -> List[AutoGPTPluginTemplate]: """Scan the plugins directory for plugins and loads them. diff --git a/pilot/configs/config.py b/pilot/configs/config.py index 6db040a9d..9e6542db9 100644 --- a/pilot/configs/config.py +++ b/pilot/configs/config.py @@ -90,7 +90,7 @@ class Config(metaclass=Singleton): ### The associated configuration parameters of the plug-in control the loading and use of the plug-in self.plugins: List[AutoGPTPluginTemplate] = [] self.plugins_openai = [] - self.plugins_auto_load = os.getenv("AUTO_LOAD_PLUGIN", "True") == "True" + self.plugins_auto_load = os.getenv("AUTO_LOAD_PLUGIN", "True") == "True" self.plugins_git_branch = os.getenv("PLUGINS_GIT_BRANCH", "plugin_dashboard")