diff --git a/docs/faq.md b/docs/faq.md index 3b7064311..84e9f579b 100644 --- a/docs/faq.md +++ b/docs/faq.md @@ -51,14 +51,14 @@ Normal: ##### Q4:When I use openai(MODEL_SERVER=proxyllm) to chat

- +

##### A4: make sure your openapi API_KEY is available ##### Q5:When I Chat Data and Chat Meta Data, I found the error

- +

##### A5: you have not create your database and table @@ -103,7 +103,7 @@ VECTOR_STORE_TYPE=Chroma ``` ##### Q7:When I use vicuna-13b, found some illegal character like this.

- +

##### A7: set KNOWLEDGE_SEARCH_TOP_SIZE smaller or set KNOWLEDGE_CHUNK_SIZE smaller, and reboot server. diff --git a/docs/locales/zh_CN/LC_MESSAGES/faq.po b/docs/locales/zh_CN/LC_MESSAGES/faq.po index 074547881..30774fbde 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/faq.po +++ b/docs/locales/zh_CN/LC_MESSAGES/faq.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 👏👏 0.3.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-07-20 10:53+0800\n" +"POT-Creation-Date: 2023-07-31 17:15+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -19,82 +19,108 @@ msgstr "" "Content-Transfer-Encoding: 8bit\n" "Generated-By: Babel 2.12.1\n" -#: ../../faq.md:1 81cedd396b274db9b2a69448df98a28d +#: ../../faq.md:1 45fbeef1eff84435aced21f0a273e44f msgid "FAQ" msgstr "FAQ" -#: ../../faq.md:2 1a3f8c7661e34721a4465a34281416b1 +#: ../../faq.md:2 fa6cad40d7384c9ba61dd0d4e69f23a0 msgid "Q1: text2vec-large-chinese not found" msgstr "Q1: text2vec-large-chinese not found" -#: ../../faq.md:4 5d2844bcbcc843fc97de41a491f914fe +#: ../../faq.md:4 98f2cb936e6d4589bb27d918b6e96100 msgid "" "A1: make sure you have download text2vec-large-chinese embedding model in" " right way" msgstr "按照正确的姿势下载text2vec-large-chinese模型" -#: ../../faq.md:16 82ade01884534030b81be7c3e06f1504 +#: ../../faq.md:16 d9c0c02c44c74c8d9ab4c499548ca3b6 msgid "" "Q2: execute `pip install -r requirements.txt` error, found some package " "cannot find correct version." msgstr "执行`pip install -r requirements.txt`报错" -#: ../../faq.md:19 451ede833da642788e5224811a71ba0f +#: ../../faq.md:19 08cb0691d56e495686c180cadef17380 msgid "A2: change the pip source." msgstr "修改pip源" -#: ../../faq.md:26 ../../faq.md:33 0142abf1050d4fee9caf056322b52247 -#: 0adb87dd67fa4122997871f7ab064637 +#: ../../faq.md:26 ../../faq.md:33 7dd0a07a8e6544d5898c79d3e04e2664 +#: f4ef94c1e50c467f9e575308755338ed msgid "or" msgstr "或" -#: ../../faq.md:41 d6981160ecc6491284fd32e1098fc10e +#: ../../faq.md:41 435cf523354443b89ffa243e67fe8ed3 msgid "Q3:Access denied for user 'root@localhost'(using password :NO)" msgstr "或" -#: ../../faq.md:43 e50b195057804f28b84d0d10859e4f1b +#: ../../faq.md:43 03fd39015cd845f6a79ae03ef2ee5f7e msgid "A3: make sure you have installed mysql instance in right way" msgstr "按照正确姿势安装mysql" -#: ../../faq.md:45 03ba25aa7bd241d3b32cc1916f858a3e +#: ../../faq.md:45 1b6ea088a9bb45d0a92f0b77b6513b6b msgid "Docker:" msgstr "Docker:" -#: ../../faq.md:49 dd5336b44673459c93a1408097cb76f9 +#: ../../faq.md:49 5bf5348377524ced9074207834bdd200 msgid "Normal: [download mysql instance](https://dev.mysql.com/downloads/mysql/)" msgstr "[download mysql instance](https://dev.mysql.com/downloads/mysql/)" -#: ../../faq.md:52 2b290c4653a2410c8d330ed5b0e9a821 +#: ../../faq.md:52 81194eff961c473c939d32d8fdba3c72 msgid "Q4:When I use openai(MODEL_SERVER=proxyllm) to chat" msgstr "使用openai-chatgpt模型时(MODEL_SERVER=proxyllm)" -#: ../../faq.md:57 f4d0e8e8113f4ca4bc55f167b661fd6a +#: ../../faq.md:57 e0fd44c2be10410ea59e660e3efad53e msgid "A4: make sure your openapi API_KEY is available" msgstr "确认openapi API_KEY是否可用" -#: ../../faq.md:59 092ca3dea0c5466ab6e22ab0049f166e +#: ../../faq.md:59 9982f6a98d014189a838c725ef223b46 msgid "Q5:When I Chat Data and Chat Meta Data, I found the error" msgstr "Chat Data and Chat Meta Data报如下错" -#: ../../faq.md:64 dbf61e6ea2c64ecebfdbbde83cb74e3e +#: ../../faq.md:64 b262611f9f4e4e1eaa49b69c8b06fe34 msgid "A5: you have not create your database and table" msgstr "需要创建自己的数据库" -#: ../../faq.md:65 0505bb716e6445c2a7960436d93cb407 +#: ../../faq.md:65 75c40019c7954eeb9db8da479b67b651 msgid "1.create your database." msgstr "1.先创建数据库" -#: ../../faq.md:71 fd689b541ee549bd85385647c219b4cb +#: ../../faq.md:71 c50304bb737549feaebd0109541a00a1 msgid "2.create table {$your_table} and insert your data. eg:" msgstr "然后创建数据表,模拟数据" -#: ../../faq.md:85 de2d78db5fb6450cb08b0f15385ed525 -msgid "Q6:When I use vicuna-13b, found some illegal character like this." +#: ../../faq.md:85 e311370b69f24c90a0cffbb940b261e4 +msgid "Q6:How to change Vector DB Type in DB-GPT." +msgstr "" + +#: ../../faq.md:87 cd3b34421f044cb0922037fb909ac26a +msgid "A6: Update .env file and set VECTOR_STORE_TYPE." +msgstr "" + +#: ../../faq.md:88 d20c3f4583d04cd285dcc2a052851e29 +#, fuzzy +msgid "" +"DB-GPT currently support Chroma(Default), Milvus(>2.1), Weaviate vector " +"database. If you want to change vector db, Update your .env, set your " +"vector store type, VECTOR_STORE_TYPE=Chroma (now only support Chroma and " +"Milvus(>2.1), if you set Milvus, please set MILVUS_URL and MILVUS_PORT) " +"If you want to support more vector db, you can integrate yourself.[how to" +" integrate](https://db-gpt.readthedocs.io/en/latest/modules/vector.html)" +msgstr "" +"DB-GPT当前支持Chroma(默认),如果你想替换向量数据库,需要更新.env文件,VECTOR_STORE_TYPE=Chroma (now" +" only support Chroma, Milvus Weaviate, if you set Milvus(>2.1), please " +"set MILVUS_URL and " +"MILVUS_PORT)。如果当前支持向量数据库无法满足你的需求,可以集成使用自己的向量数据库。[怎样集成](https://db-" +"gpt.readthedocs.io/en/latest/modules/vector.html)" + +#: ../../faq.md:104 955e38d635df47039bfaf48bd8eabc66 +#, fuzzy +msgid "Q7:When I use vicuna-13b, found some illegal character like this." msgstr "使用vicuna-13b,知识库问答出现乱码" -#: ../../faq.md:90 0cb1d0c2ec434763ae80e6f87d4a1665 +#: ../../faq.md:109 6d11192c1f8c4ab5b20fb54a9d29790c +#, fuzzy msgid "" -"A6: set KNOWLEDGE_SEARCH_TOP_SIZE smaller or set KNOWLEDGE_CHUNK_SIZE " +"A7: set KNOWLEDGE_SEARCH_TOP_SIZE smaller or set KNOWLEDGE_CHUNK_SIZE " "smaller, and reboot server." msgstr "将KNOWLEDGE_SEARCH_TOP_SIZE和KNOWLEDGE_CHUNK_SIZE设置小点然后重启" diff --git a/docs/locales/zh_CN/LC_MESSAGES/getting_started/getting_started.po b/docs/locales/zh_CN/LC_MESSAGES/getting_started/getting_started.po index 86277c361..3ceb24257 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/getting_started/getting_started.po +++ b/docs/locales/zh_CN/LC_MESSAGES/getting_started/getting_started.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 0.3.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-07-20 10:53+0800\n" +"POT-Creation-Date: 2023-07-31 17:04+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -19,29 +19,29 @@ msgstr "" "Content-Transfer-Encoding: 8bit\n" "Generated-By: Babel 2.12.1\n" -#: ../../getting_started/getting_started.md:1 7c12b6d7d5be4528be005cdadec568db +#: ../../getting_started/getting_started.md:1 7cfda7573f014955a3d1860a8944c3cd msgid "Quickstart Guide" msgstr "使用指南" -#: ../../getting_started/getting_started.md:3 ba312f5a132541be89dcc09012076784 +#: ../../getting_started/getting_started.md:3 65d35e746cdc4fcdb8e712cd675081c6 msgid "" "This tutorial gives you a quick walkthrough about use DB-GPT with you " "environment and data." msgstr "本教程为您提供了关于如何使用DB-GPT的使用指南。" -#: ../../getting_started/getting_started.md:5 8e56b3f0726740abaaafa57415b10bea +#: ../../getting_started/getting_started.md:5 dd221eaa459044bd805725f4d89fdde2 msgid "Installation" msgstr "安装" -#: ../../getting_started/getting_started.md:7 3af8fe74db1043349e8f784c109b0417 +#: ../../getting_started/getting_started.md:7 4736b4728e4f43409162371955e42987 msgid "To get started, install DB-GPT with the following steps." msgstr "请按照以下步骤安装DB-GPT" -#: ../../getting_started/getting_started.md:9 0b33cf4604f846e781a63d857dde72b2 +#: ../../getting_started/getting_started.md:9 550d49c0d50a4a93867840a15e624912 msgid "1. Hardware Requirements" msgstr "1. 硬件要求" -#: ../../getting_started/getting_started.md:10 f1d4abb176494bcb85cead7f3f8b719d +#: ../../getting_started/getting_started.md:10 695be3b79d4d4f56b5c8de62646a8c41 msgid "" "As our project has the ability to achieve ChatGPT performance of over " "85%, there are certain hardware requirements. However, overall, the " @@ -49,67 +49,67 @@ msgid "" "specific hardware requirements for deployment are as follows:" msgstr "由于我们的项目有能力达到85%以上的ChatGPT性能,所以对硬件有一定的要求。但总体来说,我们在消费级的显卡上即可完成项目的部署使用,具体部署的硬件说明如下:" -#: ../../getting_started/getting_started.md e8516902f29d4ca2bb46f19b5e3deb81 +#: ../../getting_started/getting_started.md 1bfd2621414b4ea780167151df2f80ff msgid "GPU" msgstr "GPU" -#: ../../getting_started/getting_started.md a951ccca67364cf7ad5f0af2ec0ece8d +#: ../../getting_started/getting_started.md 1a0fc8a2721d46dbb59eb7629c02c480 msgid "VRAM Size" msgstr "显存大小" -#: ../../getting_started/getting_started.md 01b7e055ee4543bdb619fbc14fea4d86 +#: ../../getting_started/getting_started.md 0cc7f0bf931340a1851aa376df8a4a99 msgid "Performance" msgstr "显存大小" -#: ../../getting_started/getting_started.md 0b20b224ff8a4e2c890a8b4ff43b6045 +#: ../../getting_started/getting_started.md dfb52838915146bd906b04383c9a65b7 msgid "RTX 4090" msgstr "RTX 4090" -#: ../../getting_started/getting_started.md 17a343c4359d45c987f29de1c73760b4 -#: c9daaf0578434a7e812a8d3f3edde3f0 +#: ../../getting_started/getting_started.md 9477f857f6df45f0843139ba5c23f263 +#: d14dfe72732c462d8d8073a7804af548 msgid "24 GB" msgstr "24 GB" -#: ../../getting_started/getting_started.md 6a35be039a0a43eaaf7d1aa40aece6f7 +#: ../../getting_started/getting_started.md c76a371ed992440a8b7b16ad405c273d msgid "Smooth conversation inference" msgstr "可以流畅的进行对话推理,无卡顿" -#: ../../getting_started/getting_started.md a414f5ae7bdd450f8d171d1f075d6b66 +#: ../../getting_started/getting_started.md 27fea237ff004247a6f5aa4593a5aabc msgid "RTX 3090" msgstr "RTX 3090" -#: ../../getting_started/getting_started.md 109ed9bda6e541b88eb3400a44e15df7 +#: ../../getting_started/getting_started.md 843f16d5af974b82beabba7d0afceae5 msgid "Smooth conversation inference, better than V100" msgstr "可以流畅进行对话推理,有卡顿感,但好于V100" -#: ../../getting_started/getting_started.md 3f2a05dc610f461faa989f3c12750d00 +#: ../../getting_started/getting_started.md c1ddf9dc9e384a44b988a83bb209814c msgid "V100" msgstr "V100" -#: ../../getting_started/getting_started.md 90e24795876546ecbb4796ca5d313514 +#: ../../getting_started/getting_started.md 94cc69a3ce8a4ea58a0b5c28fb972a41 msgid "16 GB" msgstr "16 GB" -#: ../../getting_started/getting_started.md 34061757e69e4691b7a5ff3c2953f1e3 +#: ../../getting_started/getting_started.md 9a8a3304aab84a2687db2185361e2fe0 msgid "Conversation inference possible, noticeable stutter" msgstr "可以进行对话推理,有明显卡顿" -#: ../../getting_started/getting_started.md:18 61d59cb27daf43eb9aa2775fa6dac820 +#: ../../getting_started/getting_started.md:18 8c83ffd018b242dbb626312ac0c69b79 msgid "2. Install" msgstr "2. 安装" -#: ../../getting_started/getting_started.md:20 28a6b66c92464929bed8f04d5b841e8c +#: ../../getting_started/getting_started.md:20 bab57fbe66814f51959c77a55a24005e #, fuzzy msgid "" "1.This project relies on a local MySQL database service, which you need " "to install locally. We recommend using Docker for installation." msgstr "本项目依赖一个本地的 MySQL 数据库服务,你需要本地安装,推荐直接使用 Docker 安装。" -#: ../../getting_started/getting_started.md:24 7793b19e96b44ba481b527a877c81170 +#: ../../getting_started/getting_started.md:24 f0033070a3d5412a94d69deaca81a8a8 msgid "prepare server sql script" msgstr "准备db-gpt server sql脚本" -#: ../../getting_started/getting_started.md:29 e3743699e95e45eab3e140df266bb3b5 +#: ../../getting_started/getting_started.md:29 2d4c2ea402c1448e8372168ee81c7fb7 msgid "" "We use [Chroma embedding database](https://github.com/chroma-core/chroma)" " as the default for our vector database, so there is no need for special " @@ -122,11 +122,11 @@ msgstr "" "向量数据库我们默认使用的是Chroma内存数据库,所以无需特殊安装,如果有需要连接其他的同学,可以按照我们的教程进行安装配置。整个DB-" "GPT的安装过程,我们使用的是miniconda3的虚拟环境。创建虚拟环境,并安装python依赖包" -#: ../../getting_started/getting_started.md:38 be9f388e255c4bc7837029ce8237ef0b +#: ../../getting_started/getting_started.md:38 125c03f282f3449b9f17f025857b10e2 msgid "Before use DB-GPT Knowledge Management" msgstr "使用知识库管理功能之前" -#: ../../getting_started/getting_started.md:44 9175b34ed78c4310b048f53ac07b13cb +#: ../../getting_started/getting_started.md:44 4d1d8f2d4f044fdeb1b622fabb87a0ea msgid "" "Once the environment is installed, we have to create a new folder " "\"models\" in the DB-GPT project, and then we can put all the models " @@ -135,33 +135,33 @@ msgstr "" "环境安装完成后,我们必须在DB-" "GPT项目中创建一个新文件夹\"models\",然后我们可以把从huggingface下载的所有模型放到这个目录下。" -#: ../../getting_started/getting_started.md:47 40715051bd844b0187265d425debfbee +#: ../../getting_started/getting_started.md:47 00b27ca566f242cfbc3d0fac77654063 #, fuzzy msgid "Notice make sure you have install git-lfs" msgstr "确保你已经安装了git-lfs" -#: ../../getting_started/getting_started.md:57 7d7e7b1fe72b4e0c96eba721a8aa2113 +#: ../../getting_started/getting_started.md:57 b8c136cbdbc34f90b09a52362b785563 msgid "" "The model files are large and will take a long time to download. During " "the download, let's configure the .env file, which needs to be copied and" " created from the .env.template" msgstr "模型文件很大,需要很长时间才能下载。在下载过程中,让我们配置.env文件,它需要从。env.template中复制和创建。" -#: ../../getting_started/getting_started.md:60 6a25e5307bdb49a0afc69b9d17395a5a +#: ../../getting_started/getting_started.md:60 a15420a447c443dabb9f60e460a9ab97 msgid "cp .env.template .env" msgstr "cp .env.template .env" -#: ../../getting_started/getting_started.md:63 14567647544f4036beaae158b59833f6 +#: ../../getting_started/getting_started.md:63 134bf228f94c4aec89c34e69e006af35 msgid "" "You can configure basic parameters in the .env file, for example setting " "LLM_MODEL to the model to be used" msgstr "您可以在.env文件中配置基本参数,例如将LLM_MODEL设置为要使用的模型。" -#: ../../getting_started/getting_started.md:65 1b459d413a4d4b7e883d1ec17384ca30 +#: ../../getting_started/getting_started.md:65 c25744934b7142449e09dd57e0c98cc6 msgid "3. Run" msgstr "3. 运行" -#: ../../getting_started/getting_started.md:66 ed15ee15450e4a028bf5aa05a9309697 +#: ../../getting_started/getting_started.md:66 e2bd77426e054a99801faa76ffa1046a msgid "" "You can refer to this document to obtain the Vicuna weights: " "[Vicuna](https://github.com/lm-sys/FastChat/blob/main/README.md#model-" @@ -170,7 +170,7 @@ msgstr "" "关于基础模型, 可以根据[Vicuna](https://github.com/lm-" "sys/FastChat/blob/main/README.md#model-weights) 合成教程进行合成。" -#: ../../getting_started/getting_started.md:68 de15a70920a94192a1f2017cbe3cdb55 +#: ../../getting_started/getting_started.md:68 794d5f1a8afc4667bbc848cc7192e9bd msgid "" "If you have difficulty with this step, you can also directly use the " "model from [this link](https://huggingface.co/Tribbiani/vicuna-7b) as a " @@ -179,7 +179,7 @@ msgstr "" "如果此步有困难的同学,也可以直接使用[此链接](https://huggingface.co/Tribbiani/vicuna-" "7b)上的模型进行替代。" -#: ../../getting_started/getting_started.md:70 763aaed45fd948fab761552a7e06061a +#: ../../getting_started/getting_started.md:70 72be6fe338304f9e8639e16b75859343 msgid "" "set .env configuration set your vector store type, " "eg:VECTOR_STORE_TYPE=Chroma, now we support Chroma and Milvus(version > " @@ -188,17 +188,21 @@ msgstr "" "在.env文件设置向量数据库环境变量,eg:VECTOR_STORE_TYPE=Chroma, 目前我们支持了 Chroma and " "Milvus(version >2.1) " -#: ../../getting_started/getting_started.md:73 a8f0dc3546c54a1098ff10157f980cef +#: ../../getting_started/getting_started.md:73 e36089f05a5a41a2affcea15ecd1ac3d #, fuzzy msgid "1.Run db-gpt server" msgstr "运行模型服务" -#: ../../getting_started/getting_started.md:78 1715948545154c10af585de8960bf853 +#: ../../getting_started/getting_started.md:78 +#: ../../getting_started/getting_started.md:127 +#: ../../getting_started/getting_started.md:181 +#: 4136b297d88e4dedb1616f28885a8cd6 eec0acd695454df59c67358889f971fb +#: f4f47ac8b1c549e0b2a68f16774fac64 #, fuzzy msgid "Open http://localhost:5000 with your browser to see the product." msgstr "打开浏览器访问http://localhost:5000" -#: ../../getting_started/getting_started.md:80 8ea9964df477473e866fe844dcf4be54 +#: ../../getting_started/getting_started.md:80 ef391125ac0d43358e4701bafe780b4a msgid "" "If you want to access an external LLM service, you need to 1.set the " "variables LLM_MODEL=YOUR_MODEL_NAME " @@ -206,7 +210,7 @@ msgid "" "file. 2.execute dbgpt_server.py in light mode" msgstr "如果你想访问外部的大模型服务,1.需要在.env文件设置模型名和外部模型服务地址。2.使用light模式启动服务" -#: ../../getting_started/getting_started.md:87 4c409a0e2a994f428712ab94b475e9bd +#: ../../getting_started/getting_started.md:87 eb22acf6142b4a81bbab0bf1192f577c msgid "" "If you want to learn about dbgpt-webui, read https://github.com/csunny" "/DB-GPT/tree/new-page-framework/datacenter" @@ -214,3 +218,84 @@ msgstr "" "如果你想了解DB-GPT前端服务,访问https://github.com/csunny/DB-GPT/tree/new-page-" "framework/datacenter" +#: ../../getting_started/getting_started.md:89 bc58c998b37242f8b4a37588bccc612b +msgid "4. Docker (Experimental)" +msgstr "" + +#: ../../getting_started/getting_started.md:91 3c95ec21279b43ba83e3a4e845511f26 +msgid "4.1 Building Docker image" +msgstr "" + +#: ../../getting_started/getting_started.md:97 01be8cbb915f4d02a84b25be3005a200 +msgid "Review images by listing them:" +msgstr "" + +#: ../../getting_started/getting_started.md:103 +#: ../../getting_started/getting_started.md:167 +#: 8664141a7b2f4a6a8f770ae72bc5a815 f1be761680804bb9b6a4928b46dfe598 +msgid "Output should look something like the following:" +msgstr "" + +#: ../../getting_started/getting_started.md:110 +#: 5fff19374f1d4b3ba3bb642f987e8b66 +msgid "4.2. Run all in one docker container" +msgstr "" + +#: ../../getting_started/getting_started.md:112 +#: 5b0137ca2d0345bea0a005fdf4037702 +msgid "**Run with local model**" +msgstr "" + +#: ../../getting_started/getting_started.md:130 +#: 51495a6801e94901935481d57399ed20 +msgid "" +"`-e LLM_MODEL=vicuna-13b`, means we use vicuna-13b as llm model, see " +"/pilot/configs/model_config.LLM_MODEL_CONFIG" +msgstr "" + +#: ../../getting_started/getting_started.md:131 +#: 753a8cf0ebbd4c62be9d5fbb7e5cf691 +msgid "" +"`-v /data/models:/app/models`, means we mount the local model file " +"directory `/data/models` to the docker container directory `/app/models`," +" please replace it with your model file directory." +msgstr "" + +#: ../../getting_started/getting_started.md:133 +#: ../../getting_started/getting_started.md:175 +#: 043b3966baf045aca5615214d9be632a 328b58e3a9ba46a39f0511970f43743c +msgid "You can see log with command:" +msgstr "" + +#: ../../getting_started/getting_started.md:139 +#: f6a0a66509f343d4a3306cd884a49dbc +msgid "**Run with openai interface**" +msgstr "" + +#: ../../getting_started/getting_started.md:158 +#: 65fabfc317ac41c6aaeb80e6f2e1d379 +msgid "" +"`-e LLM_MODEL=proxyllm`, means we use proxy llm(openai interface, " +"fastchat interface...)" +msgstr "" + +#: ../../getting_started/getting_started.md:159 +#: fb2230ca032b4a4d9cdab8d37b3ad02c +msgid "" +"`-v /data/models/text2vec-large-chinese:/app/models/text2vec-large-" +"chinese`, means we mount the local text2vec model to the docker " +"container." +msgstr "" + +#: ../../getting_started/getting_started.md:161 +#: 12c7a94038274324a2996873df60db93 +msgid "4.2. Run with docker compose" +msgstr "" + +#: ../../getting_started/getting_started.md:183 +#: 6ec270a68a094d4894abd2ef02739229 +msgid "" +"You can open docker-compose.yml in the project root directory to see more" +" details." +msgstr "" + diff --git a/docs/locales/zh_CN/LC_MESSAGES/modules/knownledge.po b/docs/locales/zh_CN/LC_MESSAGES/modules/knownledge.po index 0fe40d26d..be26b40a9 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/modules/knownledge.po +++ b/docs/locales/zh_CN/LC_MESSAGES/modules/knownledge.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 0.3.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-06-13 18:04+0800\n" +"POT-Creation-Date: 2023-07-31 17:04+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -19,11 +19,11 @@ msgstr "" "Content-Transfer-Encoding: 8bit\n" "Generated-By: Babel 2.12.1\n" -#: ../../modules/knownledge.md:1 8c5aad32a2cc4c97bc988a1f4143097b +#: ../../modules/knownledge.md:1 b18cf12f806941f3b9d1c13b52d0dfe5 msgid "Knownledge" msgstr "知识" -#: ../../modules/knownledge.md:3 d739696a9e6240c78db3906d55329636 +#: ../../modules/knownledge.md:3 3fe78b30d3994e4484df41a677614eb2 msgid "" "As the knowledge base is currently the most significant user demand " "scenario, we natively support the construction and processing of " @@ -31,15 +31,15 @@ msgid "" "base management strategies in this project, such as:" msgstr "由于知识库是当前用户需求最显著的场景,我们原生支持知识库的构建和处理。同时,我们还在本项目中提供了多种知识库管理策略,如:" -#: ../../modules/knownledge.md:4 16e03ee1cd454786a736b6960c668c3a +#: ../../modules/knownledge.md:4 17b2485a12744b5587655201be50e023 msgid "Default built-in knowledge base" msgstr "默认内置知识库" -#: ../../modules/knownledge.md:5 b5c57a8c773b40d18e344862adf7790e +#: ../../modules/knownledge.md:5 e137d9916a0a4a0681dbfed5d5a5065f msgid "Custom addition of knowledge bases" msgstr "自定义新增知识库" -#: ../../modules/knownledge.md:6 7c05adacafe34780a73fa2bc6748f92f +#: ../../modules/knownledge.md:6 0bca133996d4435b84245f2b53f43d72 msgid "" "Various usage scenarios such as constructing knowledge bases through " "plugin capabilities and web crawling. Users only need to organize the " @@ -47,45 +47,51 @@ msgid "" "the knowledge base required for the large model." msgstr "各种使用场景,例如通过插件功能和爬虫构建知识库。用户只需要组织知识文档,并且他们可以使用我们现有的功能来构建大型模型所需的知识库。" -#: ../../modules/knownledge.md:9 8b196a2a9efb435baf648a99d89e1220 +#: ../../modules/knownledge.md:9 7355eed198514efc8e3bc178039b0251 msgid "Create your own knowledge repository" msgstr "创建你自己的知识库" -#: ../../modules/knownledge.md:11 370071fde98c4c59bb18735364602adf +#: ../../modules/knownledge.md:11 96e0276a5d3047fea5410e9b33c33308 msgid "" "1.Place personal knowledge files or folders in the pilot/datasets " "directory." msgstr "1.将个人知识文件或文件夹放在pilot/datasets目录中。" -#: ../../modules/knownledge.md:13 5ac32a1253c4433e87d64dccb2c8b600 +#: ../../modules/knownledge.md:13 8762c0a463094c19924cdd4b7b1b1ede msgid "" "We currently support many document formats: txt, pdf, md, html, doc, ppt," " and url." msgstr "当前支持txt, pdf, md, doc, ppt, html文档格式" -#: ../../modules/knownledge.md:15 1782a135e84f4e9f8cb090f8af935428 +#: ../../modules/knownledge.md:15 752a5e7c623a49439ecf1ce8e6ccca7d msgid "before execution:" msgstr "在执行之前" -#: ../../modules/knownledge.md:22 43791873b7e043239e160790bbfc10e1 +#: ../../modules/knownledge.md:22 fbef967557a94b938f8e47497bb43c20 msgid "" "2.Update your .env, set your vector store type, VECTOR_STORE_TYPE=Chroma " "(now only support Chroma and Milvus, if you set Milvus, please set " "MILVUS_URL and MILVUS_PORT)" msgstr "2.更新你的.env,设置你的向量存储类型,VECTOR_STORE_TYPE=Chroma(现在只支持Chroma和Milvus,如果你设置了Milvus,请设置MILVUS_URL和MILVUS_PORT)" -#: ../../modules/knownledge.md:25 197e043db45e444e9c930f29be808f31 +#: ../../modules/knownledge.md:25 f1745e2f17864711a636ecbdd6cb9833 msgid "2.Run the knowledge repository script in the tools directory." msgstr "2.在tools目录执行知识入库脚本" -#: ../../modules/knownledge.md:34 abeb77ed400c4838b2ca8e14dcd89b29 +#: ../../modules/knownledge.md:34 5ae832d038b245a29a9e089f9e169cb0 +msgid "" +"Optionally, you can run `python tools/knowledge_init.py -h` command to " +"see more usage." +msgstr "" + +#: ../../modules/knownledge.md:36 4aab02276dfd41819dbd218ecc608326 msgid "" "3.Add the knowledge repository in the interface by entering the name of " "your knowledge repository (if not specified, enter \"default\") so you " "can use it for Q&A based on your knowledge base." msgstr "如果选择新增知识库,在界面上新增知识库输入你的知识库名" -#: ../../modules/knownledge.md:36 dcff9efafd9d441b91c1389af2a49780 +#: ../../modules/knownledge.md:38 f990c8495c994aa1beb040ede6b2329a msgid "" "Note that the default vector model used is text2vec-large-chinese (which " "is a large model, so if your personal computer configuration is not " diff --git a/docs/locales/zh_CN/LC_MESSAGES/modules/llms.po b/docs/locales/zh_CN/LC_MESSAGES/modules/llms.po index 3c30cd820..40bca7866 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/modules/llms.po +++ b/docs/locales/zh_CN/LC_MESSAGES/modules/llms.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 0.3.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-06-30 17:16+0800\n" +"POT-Creation-Date: 2023-07-31 17:04+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -19,11 +19,11 @@ msgstr "" "Content-Transfer-Encoding: 8bit\n" "Generated-By: Babel 2.12.1\n" -#: ../../modules/llms.md:1 7fa87bcbc5de40b2902c41545027d8b8 +#: ../../modules/llms.md:1 8322a2795eb74f1c99007fa0536ee7e5 msgid "LLMs" msgstr "大语言模型" -#: ../../modules/llms.md:3 48a7c98128114cf58c57c41575ae53a6 +#: ../../modules/llms.md:3 b8458e9c89b5403099b1eb04c4b7e755 #, python-format msgid "" "In the underlying large model integration, we have designed an open " @@ -36,47 +36,49 @@ msgid "" "of use." msgstr "在底层大模型接入中,我们设计了开放的接口,支持对接多种大模型。同时对于接入模型的效果,我们有非常严格的把控与评审机制。对大模型能力上与ChatGPT对比,在准确率上需要满足85%以上的能力对齐。我们用更高的标准筛选模型,是期望在用户使用过程中,可以省去前面繁琐的测试评估环节。" -#: ../../modules/llms.md:5 8d162757e5934dca9395605a5809193a +#: ../../modules/llms.md:5 3b4af62d13b24ec28b8b0c8128cabd03 msgid "Multi LLMs Usage" msgstr "多模型使用" -#: ../../modules/llms.md:6 b1a5b940bee64aaa8945c078f9788459 +#: ../../modules/llms.md:6 e51a4f2e4651459cab8bdc53114ae624 msgid "" "To use multiple models, modify the LLM_MODEL parameter in the .env " "configuration file to switch between the models." msgstr "如果要使用不同的模型,请修改.env配置文件中的LLM MODEL参数以在模型之间切换。" -#: ../../modules/llms.md:8 77c89d2a77d64b8aa6796c59aad07ed2 +#: ../../modules/llms.md:8 4ed94e391ae04868b6d30a71c73d1f31 msgid "" "Notice: you can create .env file from .env.template, just use command " "like this:" msgstr "注意:你可以从 .env.template 创建 .env 文件。只需使用如下命令:" -#: ../../modules/llms.md:14 4fc870de8fc14861a97a5417a9e886f4 +#: ../../modules/llms.md:14 3d7e9263b09d487ab23ada4f1ba01e02 +#, fuzzy msgid "" "now we support models vicuna-13b, vicuna-7b, chatglm-6b, flan-t5-base, " -"guanaco-33b-merged, falcon-40b, gorilla-7b." +"guanaco-33b-merged, falcon-40b, gorilla-7b, llama-2-7b, llama-2-13b, " +"baichuan-7b, baichuan-13b" msgstr "" "现在我们支持的模型有vicuna-13b, vicuna-7b, chatglm-6b, flan-t5-base, guanaco-33b-" "merged, falcon-40b, gorilla-7b." -#: ../../modules/llms.md:16 76cd62716d864bc48c362a5f095dbe94 +#: ../../modules/llms.md:16 4cf68821b6704730bbe6a98fb1355f77 msgid "" "if you want use other model, such as chatglm-6b, you just need update " ".env config file." msgstr "如果你想使用其他模型,比如chatglm-6b, 仅仅需要修改.env 配置文件" -#: ../../modules/llms.md:20 be990b58d30c469486c21e4a11739274 +#: ../../modules/llms.md:20 c37b0e5ca58f420f91fdb93bf25215e0 msgid "" "or chatglm2-6b, which is the second-generation version of the open-" "source bilingual (Chinese-English) chat model ChatGLM-6B." msgstr "" -#: ../../modules/llms.md:27 52b37848bd8345aea6a5a331019a01cd +#: ../../modules/llms.md:27 044b42a90acc4390b7714a27f38ee362 msgid "Run Model with cpu." msgstr "用CPU运行模型" -#: ../../modules/llms.md:28 e3090ffbb05646e6af1b0eebedabc9e6 +#: ../../modules/llms.md:28 da4226be516b4cb88d6c1dc94bb918e4 msgid "" "we alse support smaller models, like gpt4all. you can use it with " "cpu/mps(M1/M2), Download from [gpt4all model](https://gpt4all.io/models" @@ -85,11 +87,11 @@ msgstr "" "我们也支持一些小模型,你可以通过CPU/MPS(M1、M2)运行, 模型下载[gpt4all](https://gpt4all.io/models" "/ggml-gpt4all-j-v1.3-groovy.bin)" -#: ../../modules/llms.md:30 67b0f633acee493d86033806a5f0db9a +#: ../../modules/llms.md:30 f869669ebb104270bb408c6a2df4f68c msgid "put it in the models path, then change .env config." msgstr "将模型放在models路径, 修改.env 配置文件" -#: ../../modules/llms.md:35 f1103e6c60b44657ba47129f97156251 +#: ../../modules/llms.md:35 88dff0a183ca4052b28d6c06fe46a3ac msgid "" "DB-GPT provides a model load adapter and chat adapter. load adapter which" " allows you to easily adapt load different LLM models by inheriting the " @@ -98,15 +100,15 @@ msgstr "" "DB-GPT提供了多模型适配器load adapter和chat adapter.load adapter通过继承BaseLLMAdapter类," " 实现match和loader方法允许你适配不同的LLM." -#: ../../modules/llms.md:37 768ae456f55b4b1894d7dff43c7a7f01 +#: ../../modules/llms.md:37 c013c4dc67a645ea8b0ae87d0c253e79 msgid "vicuna llm load adapter" msgstr "vicuna llm load adapter" -#: ../../modules/llms.md:54 5364ea2d1b2e4759a98045b4ff7be4fe +#: ../../modules/llms.md:54 b239c68b5d1a44dfbbaa5ff1354b02d2 msgid "chatglm load adapter" msgstr "chatglm load adapter" -#: ../../modules/llms.md:81 eb7e5dc538a14a67b030b7a8b89a8e8a +#: ../../modules/llms.md:81 ceccd36025064a2dbd8578032db00a01 msgid "" "chat adapter which allows you to easily adapt chat different LLM models " "by inheriting the BaseChatAdpter.you just implement match() and " @@ -115,48 +117,59 @@ msgstr "" "chat " "adapter通过继承BaseChatAdpter允许你通过实现match和get_generate_stream_func方法允许你适配不同的LLM." -#: ../../modules/llms.md:83 f8e2cbbbf1e64b07af1a18bc089d268f +#: ../../modules/llms.md:83 668aea86f47f4082b7bf168074d6514c msgid "vicuna llm chat adapter" msgstr "vicuna llm chat adapter" -#: ../../modules/llms.md:95 dba52e8a07354536bbfa38319129b98f +#: ../../modules/llms.md:95 dba103e2e54a447b9543b51cb9876f6b msgid "chatglm llm chat adapter" msgstr "chatglm llm chat adapter" -#: ../../modules/llms.md:108 3b9b667a154f4123ac0c2618ce9a8626 +#: ../../modules/llms.md:108 3edb5d9efb0a4f90af21f37bb9243e34 msgid "" "if you want to integrate your own model, just need to inheriting " "BaseLLMAdaper and BaseChatAdpter and implement the methods" msgstr "如果你想集成自己的模型,只需要继承BaseLLMAdaper和BaseChatAdpter类,然后实现里面的方法即可" -#: ../../modules/llms.md:110 713bb7cbd7644cc3b3734901b46d3ee0 +#: ../../modules/llms.md:110 a39d9f39b42c4757b9df597dcbcc9603 #, fuzzy msgid "Multi Proxy LLMs" msgstr "多模型使用" -#: ../../modules/llms.md:111 125f310663e54a5b9a78f036df2c7062 +#: ../../modules/llms.md:111 d205dedc7e93407da3624aed610a6093 msgid "1. Openai proxy" msgstr "" -#: ../../modules/llms.md:112 56a578a5e1234fe38a61eacd74d52b93 +#: ../../modules/llms.md:112 dc49675aa40f484ba1de16cffbf15e28 msgid "" "If you haven't deployed a private infrastructure for a large model, or if" " you want to use DB-GPT in a low-cost and high-efficiency way, you can " "also use OpenAI's large model as your underlying model." msgstr "" -#: ../../modules/llms.md:114 ce9d8db45c0441a59dc67babedfa1829 +#: ../../modules/llms.md:114 30bfaa0fcf754c6cb51e1b400246dc9b msgid "" "If your environment deploying DB-GPT has access to OpenAI, then modify " "the .env configuration file as below will work." msgstr "" -#: ../../modules/llms.md:122 cc81057c5bb6453e90bd5ab6c5d29c19 +#: ../../modules/llms.md:122 33ef5db4a49241a8ba1240cafece6d48 msgid "" "If you can't access OpenAI locally but have an OpenAI proxy service, you " "can configure as follows." msgstr "" +#: ../../modules/llms.md:130 2dd7c019ee3f4c4f83948f0b006077f9 +msgid "2. Bard Proxy" +msgstr "" + +#: ../../modules/llms.md:131 7d7837484e474b0d936c838f6303723a +msgid "" +"If your environment deploying DB-GPT has access to " +"https://bard.google.com/ (F12-> application-> __Secure-1PSID), then " +"modify the .env configuration file as below will work." +msgstr "" + #~ msgid "Multi Proxy LLMs" #~ msgstr "多模型使用" diff --git a/docs/locales/zh_CN/LC_MESSAGES/modules/vector.po b/docs/locales/zh_CN/LC_MESSAGES/modules/vector.po index 0fc6e9e32..9b5e8bed0 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/modules/vector.po +++ b/docs/locales/zh_CN/LC_MESSAGES/modules/vector.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 👏👏 0.2.2\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-06-19 19:10+0800\n" +"POT-Creation-Date: 2023-07-31 17:15+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -19,16 +19,16 @@ msgstr "" "Content-Transfer-Encoding: 8bit\n" "Generated-By: Babel 2.12.1\n" -#: ../../modules/vector.rst:2 ../../modules/vector.rst:17 -#: 22f05e3df5a14919ba30198e3800fdb9 +#: ../../modules/vector.rst:2 ../../modules/vector.rst:19 +#: 3f1436c219c34b138a4db11ff55e070c msgid "VectorConnector" msgstr "VectorConnector" -#: ../../modules/vector.rst:4 caa4353c84a04157809b694a6a7b46ca +#: ../../modules/vector.rst:4 55445cdb9379484db2b2127620401adf msgid "**VectorConnector Introduce**" msgstr "VectorConnector介绍" -#: ../../modules/vector.rst:6 e2f584d6cf374a10ae4471d3b9a72cef +#: ../../modules/vector.rst:6 b3b361332bf0447c9ed28976e80f61d5 msgid "" "vector knowledge base is a method of mapping words in language to a high-" "dimensional vector space. In the vector space, each word is represented " @@ -43,7 +43,7 @@ msgid "" " requires a large corpus and computing resources to complete." msgstr "向量知识库是一种将语言中的单词映射到一个高维向量空间中的方法。在向量空间中,每个单词都表示为一个向量,这个向量包含了许多数值特征,这些特征代表了该单词与其他单词之间的关系。这种映射是一种聚类技术,通过将语言中的单词映射到向量空间中,单词之间的语义关系可以通过计算它们在向量空间中的差异来计算。向量知识库可以用于自然语言处理任务,如情感分析、文本分类、机器翻译等。常见的向量知识库有Word2Vec、GloVe、FastText等。这些向量知识库的训练通常需要大量的语料库和计算资源来完成。" -#: ../../modules/vector.rst:8 02a16590b6ec49c09a8247f89714eada +#: ../../modules/vector.rst:8 a3abe6fd2c82463993e5deac78e30700 msgid "" "VectorConnector is a vector database connection adapter that allows you " "to connect different vector databases and abstracts away implementation " @@ -54,21 +54,35 @@ msgstr "" "vectorconnector是一个向量数据库连接适配器,你可以通过它来连接不同的向量数据库,并且屏蔽掉了不同的向量数据的是实现差异和底层细节。例如Milvus," " Chroma, Elasticsearch, Weaviate..." -#: ../../modules/vector.rst:10 c38e3540e8844ea2a550620e2afac343 -msgid "DB-GPT VectorConnector currently support milvus and chroma vector database" +#: ../../modules/vector.rst:10 60bf9f9feab34835ada5b1954f4edefd +#, fuzzy +msgid "" +"DB-GPT VectorConnector currently support Chroma(Default), Milvus(>2.1), " +"Weaviate vector database." msgstr "DB-GPT VectorConnector当前支持Milvus和Chroma,未来会越来越多." -#: ../../modules/vector.rst:12 f7e25f67a8764cab9664f8b7208af1aa +#: ../../modules/vector.rst:12 cd057c92691247d486d0acc4b5827731 +#, fuzzy +msgid "" +"If you want to change vector db, Update your .env, set your vector store " +"type, VECTOR_STORE_TYPE=Chroma (now only support Chroma, Milvus(>2.1) and" +" Weaviate, if you set Milvus, please set MILVUS_URL and MILVUS_PORT)" +msgstr "" +"如果你想替换向量数据库,需要更新.env文件,VECTOR_STORE_TYPE=Chroma (now only support Chroma," +" Milvus Weaviate, if you set Milvus(>2.1), please set MILVUS_URL and " +"MILVUS_PORT)" + +#: ../../modules/vector.rst:14 bce04f3cde7b427895034aa5c2899857 #, fuzzy msgid "`chroma <./vector/chroma.html>`_: supported chroma vector database." msgstr "msgid \"`Chroma <./vector/chroma.html>`_: 支持Chroma向量数据库" -#: ../../modules/vector.rst:13 d08ef7653e964af590d9bb99d36204a8 +#: ../../modules/vector.rst:15 f6947639cf0a4e1d8a83d3521c3b2781 #, fuzzy msgid "`milvus <./vector/milvus.html>`_: supported milvus vector database." msgstr "Milvus <./vector/milvus.html>`_: 支持Milvus向量数据库" -#: ../../modules/vector.rst:14 2f95f18ad59b471cbfbc3f4dda4f5982 +#: ../../modules/vector.rst:16 e5880b2c53fe42ad87b2ee8882abe90e #, fuzzy msgid "`weaviate <./vector/weaviate.html>`_: supported weaviate vector database." msgstr "Milvus <./vector/milvus.html>`_: 支持Milvus向量数据库" diff --git a/docs/modules/vector.rst b/docs/modules/vector.rst index 07b885907..2cbfdbb32 100644 --- a/docs/modules/vector.rst +++ b/docs/modules/vector.rst @@ -11,6 +11,20 @@ DB-GPT VectorConnector currently support Chroma(Default), Milvus(>2.1), Weaviate If you want to change vector db, Update your .env, set your vector store type, VECTOR_STORE_TYPE=Chroma (now only support Chroma, Milvus(>2.1) and Weaviate, if you set Milvus, please set MILVUS_URL and MILVUS_PORT) +:: + + #*******************************************************************# + #** VECTOR STORE SETTINGS **# + #*******************************************************************# + VECTOR_STORE_TYPE=Chroma + #MILVUS_URL=127.0.0.1 + #MILVUS_PORT=19530 + #MILVUS_USERNAME + #MILVUS_PASSWORD + #MILVUS_SECURE= + #WEAVIATE_URL=https://kt-region-m8hcy0wc.weaviate.network + + - `chroma <./vector/chroma.html>`_: supported chroma vector database. - `milvus <./vector/milvus.html>`_: supported milvus vector database. - `weaviate <./vector/weaviate.html>`_: supported weaviate vector database.