From 9cc638630119f5c65ad5a401f537317089b3346c Mon Sep 17 00:00:00 2001 From: aries_ckt <916701291@qq.com> Date: Fri, 3 Nov 2023 10:24:49 +0800 Subject: [PATCH] doc:update deploy doc --- docs/conf.py | 18 +- docs/getting_started/install.rst | 2 +- docs/getting_started/install/deploy.rst | 438 ++++++++++++++++++ .../install/environment/environment.po | 136 +++--- .../zh_CN/LC_MESSAGES/modules/knowledge.po | 264 ++++++----- docs/requirements.txt | 2 +- 6 files changed, 682 insertions(+), 178 deletions(-) create mode 100644 docs/getting_started/install/deploy.rst diff --git a/docs/conf.py b/docs/conf.py index 539153e4c..8601627ca 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -30,10 +30,24 @@ extensions = [ "myst_nb", "sphinx_copybutton", "sphinx_panels", + "sphinx_tabs.tabs", "IPython.sphinxext.ipython_console_highlighting", + 'sphinx.ext.autosectionlabel' ] source_suffix = [".ipynb", ".html", ".md", ".rst"] + +myst_enable_extensions = [ + "dollarmath", + "amsmath", + "deflist", + "html_admonition", + "html_image", + "colon_fence", + "smartquotes", + "replacements", +] + # autodoc_pydantic_model_show_json = False # autodoc_pydantic_field_list_validators = False # autodoc_pydantic_config_members = False @@ -56,5 +70,7 @@ gettext_uuid = True # -- Options for HTML output ------------------------------------------------- # https://www.sphinx-doc.org/en/master/usage/configuration.html#options-for-html-output -html_theme = "furo" +html_theme = "sphinx_book_theme" + + html_static_path = ["_static"] diff --git a/docs/getting_started/install.rst b/docs/getting_started/install.rst index d6e6a15f8..645467e9e 100644 --- a/docs/getting_started/install.rst +++ b/docs/getting_started/install.rst @@ -18,7 +18,7 @@ DB-GPT product is a Web application that you can chat database, chat knowledge, :name: deploy :hidden: - ./install/deploy/deploy.md + ./install/deploy.rst ./install/docker/docker.md ./install/docker_compose/docker_compose.md ./install/cluster/cluster.rst diff --git a/docs/getting_started/install/deploy.rst b/docs/getting_started/install/deploy.rst new file mode 100644 index 000000000..2c6b26b84 --- /dev/null +++ b/docs/getting_started/install/deploy.rst @@ -0,0 +1,438 @@ +.. _installation: + +Installation From Source +============== + +To get started, install DB-GPT with the following steps. + +DB-GPT can be deployed on servers with low hardware requirements or on servers with high hardware requirements. +You can install DB-GPT by Using third-part LLM REST API Service OpenAI, Azure. + +And you can also install DB-GPT by deploy LLM Service by download LLM model. + + +1.Preparation +----------------- +**Download DB-GPT** + +.. code-block:: shell + + git clone https://github.com/eosphoros-ai/DB-GPT.git + +**Install Miniconda** + +We use Sqlite as default database, so there is no need for database installation. If you choose to connect to other databases, you can follow our tutorial for installation and configuration. +For the entire installation process of DB-GPT, we use the miniconda3 virtual environment. Create a virtual environment and install the Python dependencies. +:ref: `https://docs.conda.io/en/latest/miniconda.html` + +.. code-block:: shell + + python>=3.10 + conda create -n dbgpt_env python=3.10 + conda activate dbgpt_env + # it will take some minutes + pip install -e ".[default]" + +.. code-block:: shell + + cp .env.template .env + +2.Deploy LLM Service +----------------- +DB-GPT can be deployed on servers with low hardware requirements or on servers with high hardware requirements. + +If you are low hardware requirements you can install DB-GPT by Using third-part LLM REST API Service OpenAI, Azure, tongyi. + +.. tabs:: + + .. tab:: OpenAI + + Download embedding model + + .. code-block:: shell + + cd DB-GPT + mkdir models and cd models + + #### embedding model + git clone https://huggingface.co/GanymedeNil/text2vec-large-chinese + or + git clone https://huggingface.co/moka-ai/m3e-large + + Configure LLM_MODEL and PROXY_API_URL and API_KEY in `.env` file + + .. code-block:: shell + + LLM_MODEL=chatgpt_proxyllm + PROXY_API_KEY={your-openai-sk} + PROXY_SERVER_URL=https://api.openai.com/v1/chat/completions + + .. tip:: + + Make sure your .env configuration is not overwritten + + + .. tab:: Vicuna + ([Vicuna-v1.5](https://huggingface.co/lmsys/vicuna-13b-v1.5) based on llama-2 has been released, we recommend you set `LLM_MODEL=vicuna-13b-v1.5` to try this model) + + .. list-table:: vicuna-v1.5 hardware requirements + :widths: 50 50 50 + :header-rows: 1 + + * - Model + - Quantize + - VRAM Size + * - vicuna-7b-v1.5 + - 4-bit + - 8 GB + * - vicuna-7b-v1.5 + - 8-bit + - 12 GB + * - vicuna-13b-v1.5 + - 4-bit + - 12 GB + * - vicuna-13b-v1.5 + - 8-bit + - 20 GB + + + .. note:: + + Notice make sure you have install git-lfs + + centos:yum install git-lfs + + ubuntu:apt-get install git-lfs + + macos:brew install git-lfs + + + .. code-block:: shell + + cd DB-GPT + mkdir models and cd models + + #### embedding model + git clone https://huggingface.co/GanymedeNil/text2vec-large-chinese + or + git clone https://huggingface.co/moka-ai/m3e-large + + #### llm model, if you use openai or Azure or tongyi llm api service, you don't need to download llm model + git clone https://huggingface.co/lmsys/vicuna-13b-v1.5 + + The model files are large and will take a long time to download. + + **Configure LLM_MODEL in `.env` file** + + + .. code-block:: shell + + LLM_MODEL=vicuna-13b-v1.5 + + .. tab:: Baichuan + + .. list-table:: Baichuan hardware requirements + :widths: 50 50 50 + :header-rows: 1 + + * - Model + - Quantize + - VRAM Size + * - baichuan-7b + - 4-bit + - 8 GB + * - baichuan-7b + - 8-bit + - 12 GB + * - baichuan-13b + - 4-bit + - 12 GB + * - baichuan-13b + - 8-bit + - 20 GB + + .. note:: + + Notice make sure you have install git-lfs + + centos:yum install git-lfs + + ubuntu:apt-get install git-lfs + + macos:brew install git-lfs + + + .. code-block:: shell + + cd DB-GPT + mkdir models and cd models + + #### embedding model + git clone https://huggingface.co/GanymedeNil/text2vec-large-chinese + or + git clone https://huggingface.co/moka-ai/m3e-large + + #### llm model + git clone https://huggingface.co/baichuan-inc/Baichuan2-7B-Chat + or + git clone https://huggingface.co/baichuan-inc/Baichuan2-13B-Chat + + The model files are large and will take a long time to download. + + **Configure LLM_MODEL in `.env` file** + + please rename Baichuan path to "baichuan2-13b" or "baichuan2-7b" + + .. code-block:: shell + + LLM_MODEL=baichuan2-13b + + .. tab:: ChatGLM + + .. note:: + + Notice make sure you have install git-lfs + + centos:yum install git-lfs + + ubuntu:apt-get install git-lfs + + macos:brew install git-lfs + + + .. code-block:: shell + + cd DB-GPT + mkdir models and cd models + + #### embedding model + git clone https://huggingface.co/GanymedeNil/text2vec-large-chinese + or + git clone https://huggingface.co/moka-ai/m3e-large + + #### llm model + git clone https://huggingface.co/THUDM/chatglm2-6b + + The model files are large and will take a long time to download. + + **Configure LLM_MODEL in `.env` file** + + please rename chatglm model path to "chatglm2-6b" + + .. code-block:: shell + + LLM_MODEL=chatglm2-6b + + .. tab:: Other LLM API + + Download embedding model + + .. code-block:: shell + + cd DB-GPT + mkdir models and cd models + + #### embedding model + git clone https://huggingface.co/GanymedeNil/text2vec-large-chinese + or + git clone https://huggingface.co/moka-ai/m3e-large + + .. note:: + + * OpenAI + * Azure + * Aliyun tongyi + * Baidu wenxin + * Zhipu + * Baichuan + * Bard + + Configure LLM_MODEL and PROXY_API_URL and API_KEY in `.env` file + + .. code-block:: shell + + #OpenAI + LLM_MODEL=chatgpt_proxyllm + PROXY_API_KEY={your-openai-sk} + PROXY_SERVER_URL=https://api.openai.com/v1/chat/completions + + #Azure + LLM_MODEL=chatgpt_proxyllm + PROXY_API_KEY={your-azure-sk} + PROXY_API_BASE=https://{your domain}.openai.azure.com/ + PROXY_API_TYPE=azure + PROXY_SERVER_URL=xxxx + PROXY_API_VERSION=2023-05-15 + PROXYLLM_BACKEND=gpt-35-turbo + + #Aliyun tongyi + LLM_MODEL=tongyi_proxyllm + TONGYI_PROXY_API_KEY={your-tongyi-sk} + PROXY_SERVER_URL={your_service_url} + + ## Baidu wenxin + LLM_MODEL=wenxin_proxyllm + PROXY_SERVER_URL={your_service_url} + WEN_XIN_MODEL_VERSION={version} + WEN_XIN_API_KEY={your-wenxin-sk} + WEN_XIN_SECRET_KEY={your-wenxin-sct} + + ## Zhipu + LLM_MODEL=zhipu_proxyllm + PROXY_SERVER_URL={your_service_url} + ZHIPU_MODEL_VERSION={version} + ZHIPU_PROXY_API_KEY={your-zhipu-sk} + + ## Baichuan + LLM_MODEL=bc_proxyllm + PROXY_SERVER_URL={your_service_url} + BAICHUN_MODEL_NAME={version} + BAICHUAN_PROXY_API_KEY={your-baichuan-sk} + BAICHUAN_PROXY_API_SECRET={your-baichuan-sct} + + ## bard + LLM_MODEL=bard_proxyllm + PROXY_SERVER_URL={your_service_url} + # from https://bard.google.com/ f12-> application-> __Secure-1PSID + BARD_PROXY_API_KEY={your-bard-token} + + .. tip:: + + Make sure your .env configuration is not overwritten + + .. tab:: llama.cpp + + DB-GPT already supports [llama.cpp](https://github.com/ggerganov/llama.cpp) via [llama-cpp-python](https://github.com/abetlen/llama-cpp-python). + + **Preparing Model Files** + + To use llama.cpp, you need to prepare a gguf format model file, and there are two common ways to obtain it, you can choose either: + + **1. Download a pre-converted model file.** + + Suppose you want to use [Vicuna 13B v1.5](https://huggingface.co/lmsys/vicuna-13b-v1.5), you can download the file already converted from [TheBloke/vicuna-13B-v1.5-GGUF](https://huggingface.co/TheBloke/vicuna-13B-v1.5-GGUF), only one file is needed. Download it to the `models` directory and rename it to `ggml-model-q4_0.gguf`. + + .. code-block:: + + wget https://huggingface.co/TheBloke/vicuna-13B-v1.5-GGUF/resolve/main/vicuna-13b-v1.5.Q4_K_M.gguf -O models/ggml-model-q4_0.gguf + + **2. Convert It Yourself** + + You can convert the model file yourself according to the instructions in [llama.cpp#prepare-data--run](https://github.com/ggerganov/llama.cpp#prepare-data--run), and put the converted file in the models directory and rename it to `ggml-model-q4_0.gguf`. + + **Installing Dependencies** + + llama.cpp is an optional dependency in DB-GPT, and you can manually install it using the following command: + + .. code-block:: + + pip install -e ".[llama_cpp]" + + + **3.Modifying the Configuration File** + + Next, you can directly modify your `.env` file to enable llama.cpp. + + .. code-block:: + + LLM_MODEL=llama-cpp + llama_cpp_prompt_template=vicuna_v1.1 + + Then you can run it according to [Run](https://db-gpt.readthedocs.io/en/latest/getting_started/install/deploy/deploy.html#run). + + + **More Configurations** + + In DB-GPT, the model configuration can be done through `{model name}_{config key}`. + + .. list-table:: More Configurations + :widths: 50 50 50 + :header-rows: 1 + + * - Environment Variable Key + - Default + - Description + * - llama_cpp_prompt_template + - None + - Prompt template name, now support: zero_shot, vicuna_v1.1,alpaca,llama-2,baichuan-chat,internlm-chat, If None, the prompt template is automatically determined from model path。 + * - llama_cpp_model_path + - None + - Model path + * - llama_cpp_n_gpu_layers + - 1000000000 + - Number of layers to offload to the GPU, Set this to 1000000000 to offload all layers to the GPU. If your GPU VRAM is not enough, you can set a low number, eg: 10 + * - llama_cpp_n_threads + - None + - Number of threads to use. If None, the number of threads is automatically determined + * - llama_cpp_n_batch + - 512 + - Maximum number of prompt tokens to batch together when calling llama_eval + * - llama_cpp_n_gqa + - None + - Grouped-query attention. Must be 8 for llama-2 70b. + * - llama_cpp_rms_norm_eps + - 5e-06 + - 5e-6 is a good value for llama-2 models. + * - llama_cpp_cache_capacity + - None + - Maximum cache capacity. Examples: 2000MiB, 2GiB + * - llama_cpp_prefer_cpu + - False + - If a GPU is available, it will be preferred by default, unless prefer_cpu=False is configured. + + + .. tab:: vllm + + vLLM is a fast and easy-to-use library for LLM inference and serving. + + **Running vLLM** + + **1.Installing Dependencies** + + vLLM is an optional dependency in DB-GPT, and you can manually install it using the following command: + + .. code-block:: + + pip install -e ".[vllm]" + + **2.Modifying the Configuration File** + + Next, you can directly modify your .env file to enable vllm. + + .. code-block:: + + LLM_MODEL=vicuna-13b-v1.5 + MODEL_TYPE=vllm + + + + + + + + +3.Prepare sql example(Optional) +----------------- +**(Optional) load examples into SQLite** + + .. code-block:: shell + + bash ./scripts/examples/load_examples.sh + + + On windows platform: + + .. code-block:: shell + + .\scripts\examples\load_examples.bat + +4.Run db-gpt server +----------------- + + .. code-block:: shell + + + python pilot/server/dbgpt_server.py + +**Open http://localhost:5000 with your browser to see the product.** + diff --git a/docs/locales/zh_CN/LC_MESSAGES/getting_started/install/environment/environment.po b/docs/locales/zh_CN/LC_MESSAGES/getting_started/install/environment/environment.po index 09b3c8fa2..1f448ee4e 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/getting_started/install/environment/environment.po +++ b/docs/locales/zh_CN/LC_MESSAGES/getting_started/install/environment/environment.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 👏👏 0.3.5\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-08-17 13:07+0800\n" +"POT-Creation-Date: 2023-11-02 10:10+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -20,290 +20,292 @@ msgstr "" "Generated-By: Babel 2.12.1\n" #: ../../getting_started/install/environment/environment.md:1 -#: be341d16f7b24bf4ad123ab78a6d855a +#: 28d2f84fc8884e78afad8118cd59c654 #, fuzzy msgid "Environment Parameter" msgstr "环境变量说明" #: ../../getting_started/install/environment/environment.md:4 -#: 46eddb27c90f41548ea9a724bbcebd37 +#: c83fbb5e1aa643cdb09fffe7f3d1a3c5 msgid "LLM MODEL Config" msgstr "模型配置" #: ../../getting_started/install/environment/environment.md:5 -#: 7deaa85df4a04fb098f5994547a8724f +#: eb675965ae57407e8d8bf90fed8e9e2a msgid "LLM Model Name, see /pilot/configs/model_config.LLM_MODEL_CONFIG" msgstr "LLM Model Name, see /pilot/configs/model_config.LLM_MODEL_CONFIG" #: ../../getting_started/install/environment/environment.md:6 -#: 3902801c546547b3a4009df681ef7d52 +#: 5d28d35126d849ea9b0d963fd1ba8699 msgid "LLM_MODEL=vicuna-13b" msgstr "LLM_MODEL=vicuna-13b" #: ../../getting_started/install/environment/environment.md:8 -#: 84b0fdbfa1544ec28751e9b69b00cc02 +#: 01955b2d0fbe4d94939ebf2cbb380bdd msgid "MODEL_SERVER_ADDRESS" msgstr "MODEL_SERVER_ADDRESS" #: ../../getting_started/install/environment/environment.md:9 -#: 0b430bfab77d405989470d00ca3f6fe0 +#: 4eaaa9ab59854c0386b28b3111c82784 msgid "MODEL_SERVER=http://127.0.0.1:8000 LIMIT_MODEL_CONCURRENCY" msgstr "MODEL_SERVER=http://127.0.0.1:8000 LIMIT_MODEL_CONCURRENCY" #: ../../getting_started/install/environment/environment.md:12 -#: b477a25586c546729a93fb6785b7b2ec +#: 5c2dd05e16834443b7451c2541b59757 msgid "LIMIT_MODEL_CONCURRENCY=5" msgstr "LIMIT_MODEL_CONCURRENCY=5" #: ../../getting_started/install/environment/environment.md:14 -#: 1d6ea800af384fff9c265610f71cc94e +#: 7707836c2fb04e7da13d2d59b5f9566f msgid "MAX_POSITION_EMBEDDINGS" msgstr "MAX_POSITION_EMBEDDINGS" #: ../../getting_started/install/environment/environment.md:16 -#: 388e758ce4ea4692a4c34294cebce7f2 +#: ee24a7d3d8384e61b715ef3bd362b965 msgid "MAX_POSITION_EMBEDDINGS=4096" msgstr "MAX_POSITION_EMBEDDINGS=4096" #: ../../getting_started/install/environment/environment.md:18 -#: 16a307dce1294ceba892ff93ae4e81c0 +#: 90b51aa4e46b4d1298c672e0052c2f68 msgid "QUANTIZE_QLORA" msgstr "QUANTIZE_QLORA" #: ../../getting_started/install/environment/environment.md:20 -#: 93ceb2b2fcd5454b82eefb0ae8c7ae77 +#: 7de7a8eb431e4973ae00f68ca0686281 msgid "QUANTIZE_QLORA=True" msgstr "QUANTIZE_QLORA=True" #: ../../getting_started/install/environment/environment.md:22 -#: 15ffa35d023a4530b02a85ee6168dd4b +#: e331ca016a474f4aa4e9182165a2693a msgid "QUANTIZE_8bit" msgstr "QUANTIZE_8bit" #: ../../getting_started/install/environment/environment.md:24 -#: 81df248ac5cb4ab0b13a711505f6a177 +#: 519ccce5a0884778be2719c437a17bd4 msgid "QUANTIZE_8bit=True" msgstr "QUANTIZE_8bit=True" #: ../../getting_started/install/environment/environment.md:27 -#: 15cc7b7d41ad44f0891c1189709f00f1 +#: 1c0586d070f046de8d0f9f94a6b508b4 msgid "LLM PROXY Settings" msgstr "LLM PROXY Settings" #: ../../getting_started/install/environment/environment.md:28 -#: e6c1115a39404f11b193a1593bc51a22 +#: c208c3f4b13f4b39962de814e5be6ab9 msgid "OPENAI Key" msgstr "OPENAI Key" #: ../../getting_started/install/environment/environment.md:30 -#: 8157e0a831fe4506a426822b7565e4f6 +#: 9228bbee2faa4467b1d24f1125faaac8 msgid "PROXY_API_KEY={your-openai-sk}" msgstr "PROXY_API_KEY={your-openai-sk}" #: ../../getting_started/install/environment/environment.md:31 -#: 89b34d00bdb64e738bd9bc8c086b1f02 +#: 759ae581883348019c1ba79e8954728a msgid "PROXY_SERVER_URL=https://api.openai.com/v1/chat/completions" msgstr "PROXY_SERVER_URL=https://api.openai.com/v1/chat/completions" #: ../../getting_started/install/environment/environment.md:33 -#: 7a97df730aeb484daf19c8172e61a290 +#: 83f3952917d34aab80bd34119f7d1e20 msgid "from https://bard.google.com/ f12-> application-> __Secure-1PSID" msgstr "from https://bard.google.com/ f12-> application-> __Secure-1PSID" #: ../../getting_started/install/environment/environment.md:35 -#: d430ddf726a049c0a9e0a9bfd5a6fe0e +#: 1d70707ca82749bb90b2bed1aee44d62 msgid "BARD_PROXY_API_KEY={your-bard-token}" msgstr "BARD_PROXY_API_KEY={your-bard-token}" #: ../../getting_started/install/environment/environment.md:38 -#: 23d6b0da3e7042abb55f6181c4a382d2 +#: 38a2091fa223493ea23cb9bbb33cf58e msgid "DATABASE SETTINGS" msgstr "DATABASE SETTINGS" #: ../../getting_started/install/environment/environment.md:39 -#: dbae0a2d847f41f5be9396a160ef88d0 +#: 5134180d7a5945b48b072a1eb92b27ba msgid "SQLite database (Current default database)" msgstr "SQLite database (Current default database)" #: ../../getting_started/install/environment/environment.md:40 -#: bdb55b7280c341a981e9d338cce53345 +#: 6875e2300e094668a45fa4f2551e0d30 msgid "LOCAL_DB_PATH=data/default_sqlite.db" msgstr "LOCAL_DB_PATH=data/default_sqlite.db" #: ../../getting_started/install/environment/environment.md:41 -#: 739d67927a9d46b28500deba1917916b +#: 034e8f06f24f44af9d8184563f99b4b3 msgid "LOCAL_DB_TYPE=sqlite # Database Type default:sqlite" msgstr "LOCAL_DB_TYPE=sqlite # Database Type default:sqlite" #: ../../getting_started/install/environment/environment.md:43 -#: eb4717bce6a6483b86d9780d924c5ff1 +#: f688149a97f740269f80b79775236ce9 msgid "MYSQL database" msgstr "MYSQL database" #: ../../getting_started/install/environment/environment.md:44 -#: 0f4cdf0ff5dd4ff0b397dfa88541a2e1 +#: 6db0b305137d45a3aa036e4f2262f460 msgid "LOCAL_DB_TYPE=mysql" msgstr "LOCAL_DB_TYPE=mysql" #: ../../getting_started/install/environment/environment.md:45 -#: c971ead492c34487bd766300730a9cba +#: b6d662ce8d5f44f0b54a7f6e7c66f5a5 msgid "LOCAL_DB_USER=root" msgstr "LOCAL_DB_USER=root" #: ../../getting_started/install/environment/environment.md:46 -#: 02828b29ad044eeab890a2f8af0e5907 +#: cd7493d61ac9415283640dc6c018d2f4 msgid "LOCAL_DB_PASSWORD=aa12345678" msgstr "LOCAL_DB_PASSWORD=aa12345678" #: ../../getting_started/install/environment/environment.md:47 -#: 53dc7f15b3934987b1f4c2e2d0b11299 +#: 4ea2a622b23f4342a4c2ab7f8d9c4e8d msgid "LOCAL_DB_HOST=127.0.0.1" msgstr "LOCAL_DB_HOST=127.0.0.1" #: ../../getting_started/install/environment/environment.md:48 -#: 1ac95fc482934247a118bab8dcebeb57 +#: 936db95a0ab246098028f4dbb596cd17 msgid "LOCAL_DB_PORT=3306" msgstr "LOCAL_DB_PORT=3306" #: ../../getting_started/install/environment/environment.md:51 -#: 34e46aa926844be19c7196759b03af63 +#: d9255f25989840ea9c9e7b34f3947c87 msgid "EMBEDDING SETTINGS" msgstr "EMBEDDING SETTINGS" #: ../../getting_started/install/environment/environment.md:52 -#: 2b5aa08cc995495e85a1f7dc4f97b5d7 +#: b09291d32aca43928a981e873476a985 msgid "EMBEDDING MODEL Name, see /pilot/configs/model_config.LLM_MODEL_CONFIG" msgstr "EMBEDDING模型, 参考see /pilot/configs/model_config.LLM_MODEL_CONFIG" #: ../../getting_started/install/environment/environment.md:53 -#: 0de0ca551ed040248406f848feca541d +#: 63de573b03a54413b997f18a1ccee279 msgid "EMBEDDING_MODEL=text2vec" msgstr "EMBEDDING_MODEL=text2vec" #: ../../getting_started/install/environment/environment.md:55 -#: 43019fb570904c9981eb68f33e64569c +#: 0ef8defbab544bd0b9475a036f278489 msgid "Embedding Chunk size, default 500" msgstr "Embedding 切片大小, 默认500" #: ../../getting_started/install/environment/environment.md:57 -#: 7e3f93854873461286e96887e04167aa +#: 33dbc7941d054baa8c6ecfc0bf1ce271 msgid "KNOWLEDGE_CHUNK_SIZE=500" msgstr "KNOWLEDGE_CHUNK_SIZE=500" #: ../../getting_started/install/environment/environment.md:59 -#: 9504f4a59ae74352a524b7741113e2d6 +#: e6ee9f2620ab45ecbc8e9c0642f5ca42 msgid "Embedding Chunk Overlap, default 100" msgstr "Embedding chunk Overlap, 文本块之间的最大重叠量。保留一些重叠可以保持文本块之间的连续性(例如使用滑动窗口),默认100" #: ../../getting_started/install/environment/environment.md:60 -#: 24e6119c2051479bbd9dba71a9c23dbe +#: fcddf64340a04df4ab95176fc2fc67a6 msgid "KNOWLEDGE_CHUNK_OVERLAP=100" msgstr "KNOWLEDGE_CHUNK_OVERLAP=100" #: ../../getting_started/install/environment/environment.md:62 -#: 0d180d7f2230442abee901c19526e442 -msgid "embeding recall top k,5" +#: 61272200194b4461a921581feb1273da +#, fuzzy +msgid "embedding recall top k,5" msgstr "embedding 召回topk, 默认5" #: ../../getting_started/install/environment/environment.md:64 -#: a5bb9ab2ba50411cbbe87f7836bfbb6d +#: b433091f055542b1b89ff2d525ac99e4 msgid "KNOWLEDGE_SEARCH_TOP_SIZE=5" msgstr "KNOWLEDGE_SEARCH_TOP_SIZE=5" #: ../../getting_started/install/environment/environment.md:66 -#: 183b8dd78cba4ae19bd2e08d69d21e0b -msgid "embeding recall max token ,2000" +#: 1db0de41aebd4caa8cc2eaecb4cacd6a +#, fuzzy +msgid "embedding recall max token ,2000" msgstr "embedding向量召回最大token, 默认2000" #: ../../getting_started/install/environment/environment.md:68 -#: ce0c711febcb44c18ae0fc858c3718d1 +#: 81b9d862e58941a4b09680a7520cdabe msgid "KNOWLEDGE_SEARCH_MAX_TOKEN=5" msgstr "KNOWLEDGE_SEARCH_MAX_TOKEN=5" #: ../../getting_started/install/environment/environment.md:71 #: ../../getting_started/install/environment/environment.md:87 -#: 4cab1f399cc245b4a1a1976d2c4fc926 ec9cec667a1c4473bf9a796a26e1ce20 +#: cac73575d54544778bdee09b18532fd9 f78a509949a64f03aa330f31901e2e7a msgid "Vector Store SETTINGS" msgstr "Vector Store SETTINGS" #: ../../getting_started/install/environment/environment.md:72 #: ../../getting_started/install/environment/environment.md:88 -#: 4dd04aadd46948a5b1dcf01fdb0ef074 bab7d512f33e40cf9e10f0da67e699c8 +#: 5ebba1cb047b4b09849000244237dfbb 7e9285e91bcb4b2d9413909c0d0a06a7 msgid "Chroma" msgstr "Chroma" #: ../../getting_started/install/environment/environment.md:73 #: ../../getting_started/install/environment/environment.md:89 -#: 13eec36741b14e028e2d3859a320826e ab3ffbcf9358401993af636ba9ab2e2d +#: 05625cfcc23c4745ae1fa0d94ce5450c 3a8615f1507f4fc49d1adda5100a4edf msgid "VECTOR_STORE_TYPE=Chroma" msgstr "VECTOR_STORE_TYPE=Chroma" #: ../../getting_started/install/environment/environment.md:74 #: ../../getting_started/install/environment/environment.md:90 -#: d15b91e2a2884f23a1dd2d54783b0638 d1f856d571b547098bb0c2a18f9f1979 +#: 5b559376aea44f159262e6d4b75c7ec1 e954782861404b10b4e893e01cf74452 msgid "MILVUS" msgstr "MILVUS" #: ../../getting_started/install/environment/environment.md:75 #: ../../getting_started/install/environment/environment.md:91 -#: 1e165f6c934343c7808459cc7a65bc70 985dd60c2b7d4baaa6601a810a6522d7 +#: 55ee8199c97a4929aeefd32370f2b92d 8f40c02543ea4a2ca9632dd9e8a08c2e msgid "VECTOR_STORE_TYPE=Milvus" msgstr "VECTOR_STORE_TYPE=Milvus" #: ../../getting_started/install/environment/environment.md:76 #: ../../getting_started/install/environment/environment.md:92 -#: a1a53f051cee40ed886346a94babd75a d263e8eaee684935a58f0a4fe61c6f0e +#: 528a01d25720491c8e086bf43a62ad92 ba1386d551d7494a85681a2803081a6f msgid "MILVUS_URL=127.0.0.1" msgstr "MILVUS_URL=127.0.0.1" #: ../../getting_started/install/environment/environment.md:77 #: ../../getting_started/install/environment/environment.md:93 -#: 2741a312db1a4c6a8a1c1d62415c5fba d03bbf921ddd4f4bb715fe5610c3d0aa +#: b031950dafcd4d4783c120dc933c4178 c2e9c8cdd41741e3aba01e59a6ef245d msgid "MILVUS_PORT=19530" msgstr "MILVUS_PORT=19530" #: ../../getting_started/install/environment/environment.md:78 #: ../../getting_started/install/environment/environment.md:94 -#: d0786490d38c4e4f971cc14f62fe1fc8 e9e0854873dc4c209861ee4eb77d25cd +#: 27b0a64af6434cb2840373e2b38c9bd5 d0e4d79af7954b129ffff7303a1ec3ce msgid "MILVUS_USERNAME" msgstr "MILVUS_USERNAME" #: ../../getting_started/install/environment/environment.md:79 #: ../../getting_started/install/environment/environment.md:95 -#: 9a82d07153cc432ebe754b5bc02fde0d a6485c1cfa7d4069a6894c43674c8c2b +#: 27aa1a5b61e64dd6bfe29124e274809e 5c58892498ce4f46a59f54b2887822d4 msgid "MILVUS_PASSWORD" msgstr "MILVUS_PASSWORD" #: ../../getting_started/install/environment/environment.md:80 #: ../../getting_started/install/environment/environment.md:96 -#: 2f233f32b8ba408a9fbadb21fabb99ec 809b3219dd824485bc2cfc898530d708 +#: 009e57d4acc5434da2146f0545911c85 bac8888dcbff47fbb0ea8ae685445aac msgid "MILVUS_SECURE=" msgstr "MILVUS_SECURE=" #: ../../getting_started/install/environment/environment.md:82 #: ../../getting_started/install/environment/environment.md:98 -#: f00603661f2b42e1bd2bca74ad1e3c31 f378e16fdec44c559e34c6929de812e8 +#: a6eeb16ab5274045bee88ecc3d93e09e eb341774403d47658b9b7e94c4c16d5c msgid "WEAVIATE" msgstr "WEAVIATE" #: ../../getting_started/install/environment/environment.md:83 -#: da2049ebc6874cf0a6b562e0e2fd9ec7 +#: fbd97522d8da4824b41b99298fd41069 msgid "VECTOR_STORE_TYPE=Weaviate" msgstr "VECTOR_STORE_TYPE=Weaviate" #: ../../getting_started/install/environment/environment.md:84 #: ../../getting_started/install/environment/environment.md:99 -#: 25f1246629934289aad7ef01c7304097 c9fe0e413d9a4fc8abf86b3ed99e0581 +#: 341785b4abfe42b5af1c2e04497261f4 a81cc2aabc8240f3ac1f674d9350bff4 msgid "WEAVIATE_URL=https://kt-region-m8hcy0wc.weaviate.network" msgstr "WEAVIATE_URL=https://kt-region-m8hcy0wc.weaviate.network" #: ../../getting_started/install/environment/environment.md:102 -#: ba7c9e707f6a4cd6b99e52b58da3ab2d +#: 5bb9e5daa36241d499089c1b1910f729 msgid "Multi-GPU Setting" msgstr "Multi-GPU Setting" #: ../../getting_started/install/environment/environment.md:103 -#: 5ca75fdf2c264b2c844d77f659b4f0b3 +#: 30df45b7f1f7423c9f18c6360f0b7600 msgid "" "See https://developer.nvidia.com/blog/cuda-pro-tip-control-gpu-" "visibility-cuda_visible_devices/ If CUDA_VISIBLE_DEVICES is not " @@ -313,49 +315,49 @@ msgstr "" "cuda_visible_devices/ 如果 CUDA_VISIBLE_DEVICES没有设置, 会使用所有可用的gpu" #: ../../getting_started/install/environment/environment.md:106 -#: de92eb310aff43fbbbf3c5a116c3b2c6 +#: 8631ea968dfb4d90a7ae6bdb2acdfdce msgid "CUDA_VISIBLE_DEVICES=0" msgstr "CUDA_VISIBLE_DEVICES=0" #: ../../getting_started/install/environment/environment.md:108 -#: d2641df6123a442b8e4444ad5f01a9aa +#: 0010422280dd4fe79326ebceb2a66f0e msgid "" "Optionally, you can also specify the gpu ID to use before the starting " "command" msgstr "你也可以通过启动命令设置gpu ID" #: ../../getting_started/install/environment/environment.md:110 -#: 76c66179d11a4e5fa369421378609aae +#: 00106f7341304fbd9425721ea8e6a261 msgid "CUDA_VISIBLE_DEVICES=3,4,5,6" msgstr "CUDA_VISIBLE_DEVICES=3,4,5,6" #: ../../getting_started/install/environment/environment.md:112 -#: 29bd0f01fdf540ad98385ea8473f7647 +#: 720aa8b3478744d78e4b10dfeccb50b4 msgid "You can configure the maximum memory used by each GPU." msgstr "可以设置GPU的最大内存" #: ../../getting_started/install/environment/environment.md:114 -#: 31e5e23838734ba7a2810e2387e6d6a0 +#: f9639ac96a244296832c75bbcbdae2af msgid "MAX_GPU_MEMORY=16Gib" msgstr "MAX_GPU_MEMORY=16Gib" #: ../../getting_started/install/environment/environment.md:117 -#: 99aa63ab1ae049d9b94536d6a96f3443 +#: fc4d955fdb3e4256af5c8f29b042dcd6 msgid "Other Setting" msgstr "Other Setting" #: ../../getting_started/install/environment/environment.md:118 -#: 3168732183874bffb59a3575d3473d62 +#: 66b14a834e884339be2d48392e884933 msgid "Language Settings(influence prompt language)" msgstr "Language Settings(涉及prompt语言以及知识切片方式)" #: ../../getting_started/install/environment/environment.md:119 -#: 73eb0a96f29b4739bd456faa9cb5033d +#: 5c9f05174eb84edd9e1316cc0721a840 msgid "LANGUAGE=en" msgstr "LANGUAGE=en" #: ../../getting_started/install/environment/environment.md:120 -#: c6646b78c6cf4d25a13108232f5b2046 +#: 7f6d62117d024c51bba9255fa4fcf151 msgid "LANGUAGE=zh" msgstr "LANGUAGE=zh" diff --git a/docs/locales/zh_CN/LC_MESSAGES/modules/knowledge.po b/docs/locales/zh_CN/LC_MESSAGES/modules/knowledge.po index 0b3eac094..032e81609 100644 --- a/docs/locales/zh_CN/LC_MESSAGES/modules/knowledge.po +++ b/docs/locales/zh_CN/LC_MESSAGES/modules/knowledge.po @@ -8,7 +8,7 @@ msgid "" msgstr "" "Project-Id-Version: DB-GPT 0.3.0\n" "Report-Msgid-Bugs-To: \n" -"POT-Creation-Date: 2023-07-13 15:39+0800\n" +"POT-Creation-Date: 2023-11-02 10:10+0800\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME \n" "Language: zh_CN\n" @@ -19,103 +19,84 @@ msgstr "" "Content-Transfer-Encoding: 8bit\n" "Generated-By: Babel 2.12.1\n" -#: ../../modules/knowledge.rst:2 ../../modules/knowledge.rst:136 -#: 3cc8fa6e9fbd4d889603d99424e9529a +#: ../../modules/knowledge.md:1 436b94d3a8374ed18feb5c14893a84e6 msgid "Knowledge" msgstr "知识" -#: ../../modules/knowledge.rst:4 0465a393d9d541958c39c1d07c885d1f +#: ../../modules/knowledge.md:3 918a3747cbed42d18b8c9c4547e67b14 #, fuzzy msgid "" "As the knowledge base is currently the most significant user demand " "scenario, we natively support the construction and processing of " "knowledge bases. At the same time, we also provide multiple knowledge " -"base management strategies in this project, such as pdf knowledge,md " -"knowledge, txt knowledge, word knowledge, ppt knowledge:" +"base management strategies in this project, such as:" msgstr "" "由于知识库是当前用户需求最显著的场景,我们原生支持知识库的构建和处理。同时,我们还在本项目中提供了多种知识库管理策略,如:pdf,md , " "txt, word, ppt" -#: ../../modules/knowledge.rst:6 e670cbe14d8e4da88ba935e4120c31e0 -msgid "" -"We currently support many document formats: raw text, txt, pdf, md, html," -" doc, ppt, and url. In the future, we will continue to support more types" -" of knowledge, including audio, video, various databases, and big data " -"sources. Of course, we look forward to your active participation in " -"contributing code." +#: ../../modules/knowledge.md:4 d4d4b5d57918485aafa457bb9fdcf626 +msgid "Default built-in knowledge base" msgstr "" -#: ../../modules/knowledge.rst:9 e0bf601a1a0c458297306db6ff79f931 -msgid "**Create your own knowledge repository**" +#: ../../modules/knowledge.md:5 d4d4b5d57918485aafa457bb9fdcf626 +msgid "Custom addition of knowledge bases" +msgstr "" + +#: ../../modules/knowledge.md:6 984361ce835c4c3492e29e1fb897348a +msgid "" +"Various usage scenarios such as constructing knowledge bases through " +"plugin capabilities and web crawling. Users only need to organize the " +"knowledge documents, and they can use our existing capabilities to build " +"the knowledge base required for the large model." +msgstr "" + +#: ../../modules/knowledge.md:9 746e4fbd3212460198be51b90caee2c8 +#, fuzzy +msgid "Create your own knowledge repository" msgstr "创建你自己的知识库" -#: ../../modules/knowledge.rst:11 bb26708135d44615be3c1824668010f6 -msgid "1.prepare" -msgstr "准备" +#: ../../modules/knowledge.md:11 1c46b33b0532417c824efbaa3e687c3f +msgid "" +"1.Place personal knowledge files or folders in the pilot/datasets " +"directory." +msgstr "" -#: ../../modules/knowledge.rst:13 c150a0378f3e4625908fa0d8a25860e9 +#: ../../modules/knowledge.md:13 3b16f387b5354947a89d6df77bd65bdb #, fuzzy msgid "" -"We currently support many document formats: TEXT(raw text), " -"DOCUMENT(.txt, .pdf, .md, .doc, .ppt, .html), and URL." +"We currently support many document formats: txt, pdf, md, html, doc, ppt," +" and url." msgstr "当前支持txt, pdf, md, html, doc, ppt, url文档格式" -#: ../../modules/knowledge.rst:15 7f9f02a93d5d4325b3d2d976f4bb28a0 +#: ../../modules/knowledge.md:15 09ec337d7da4418db854e58afb6c0980 msgid "before execution:" msgstr "开始前" -#: ../../modules/knowledge.rst:24 59699a8385e04982a992cf0d71f6dcd5 -#, fuzzy +#: ../../modules/knowledge.md:22 c09b3decb018485f8e56830ddc156194 msgid "" -"2.prepare embedding model, you can download from https://huggingface.co/." -" Notice you have installed git-lfs." +"2.Update your .env, set your vector store type, VECTOR_STORE_TYPE=Chroma " +"(now only support Chroma and Milvus, if you set Milvus, please set " +"MILVUS_URL and MILVUS_PORT)" msgstr "" -"提前准备Embedding Model, 你可以在https://huggingface.co/进行下载,注意:你需要先安装git-lfs.eg:" -" git clone https://huggingface.co/THUDM/chatglm2-6b" -#: ../../modules/knowledge.rst:27 2be1a17d0b54476b9dea080d244fd747 -msgid "" -"eg: git clone https://huggingface.co/sentence-transformers/all-" -"MiniLM-L6-v2" -msgstr "eg: git clone https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2" - -#: ../../modules/knowledge.rst:33 d328f6e243624c9488ebd27c9324621b -msgid "" -"3.prepare vector_store instance and vector store config, now we support " -"Chroma, Milvus and Weaviate." -msgstr "提前准备向量数据库环境,目前支持Chroma, Milvus and Weaviate向量数据库" - -#: ../../modules/knowledge.rst:63 44f97154eff647d399fd30b6f9e3b867 -msgid "" -"3.init Url Type EmbeddingEngine api and embedding your document into " -"vector store in your code." -msgstr "初始化 Url类型 EmbeddingEngine api, 将url文档embedding向量化到向量数据库 " - -#: ../../modules/knowledge.rst:75 e2581b414f0148bca88253c7af9cd591 -msgid "If you want to add your source_reader or text_splitter, do this:" -msgstr "如果你想手动添加你自定义的source_reader和text_splitter, 请参考:" - -#: ../../modules/knowledge.rst:95 74c110414f924bbfa3d512e45ba2f30f -#, fuzzy -msgid "" -"4.init Document Type EmbeddingEngine api and embedding your document into" -" vector store in your code. Document type can be .txt, .pdf, .md, .doc, " -".ppt." +#: ../../modules/knowledge.md:25 74460ec7709441d5945ce9f745a26d20 +msgid "2.Run the knowledge repository initialization command" msgstr "" -"初始化 文档型类型 EmbeddingEngine api, 将文档embedding向量化到向量数据库(文档可以是.txt, .pdf, " -".md, .html, .doc, .ppt)" -#: ../../modules/knowledge.rst:108 0afd40098d5f4dfd9e44fe1d8004da25 +#: ../../modules/knowledge.md:31 4498ec4e46ff4e24b45dd855e829bd32 msgid "" -"5.init TEXT Type EmbeddingEngine api and embedding your document into " -"vector store in your code." -msgstr "初始化TEXT类型 EmbeddingEngine api, 将文档embedding向量化到向量数据库" +"Optionally, you can run `dbgpt knowledge load --help` command to see more" +" usage." +msgstr "" -#: ../../modules/knowledge.rst:120 a66961bf3efd41fa8ea938129446f5a5 -msgid "4.similar search based on your knowledge base. ::" -msgstr "在知识库进行相似性搜索" +#: ../../modules/knowledge.md:33 5048ac3289e540f2a2b5fd0e5ed043f5 +msgid "" +"3.Add the knowledge repository in the interface by entering the name of " +"your knowledge repository (if not specified, enter \"default\") so you " +"can use it for Q&A based on your knowledge base." +msgstr "" -#: ../../modules/knowledge.rst:126 b7066f408378450db26770f83fbd2716 +#: ../../modules/knowledge.md:35 deeccff20f7f453dad0881b63dae2a18 msgid "" "Note that the default vector model used is text2vec-large-chinese (which " "is a large model, so if your personal computer configuration is not " @@ -125,48 +106,6 @@ msgstr "" "注意,这里默认向量模型是text2vec-large-chinese(模型比较大,如果个人电脑配置不够建议采用text2vec-base-" "chinese),因此确保需要将模型download下来放到models目录中。" -#: ../../modules/knowledge.rst:128 58481d55cab74936b6e84b24c39b1674 -#, fuzzy -msgid "" -"`pdf_embedding <./knowledge/pdf/pdf_embedding.html>`_: supported pdf " -"embedding." -msgstr "pdf_embedding <./knowledge/pdf_embedding.html>`_: supported pdf embedding." - -#: ../../modules/knowledge.rst:129 fbb013c4f1bc46af910c91292f6690cf -#, fuzzy -msgid "" -"`markdown_embedding <./knowledge/markdown/markdown_embedding.html>`_: " -"supported markdown embedding." -msgstr "pdf_embedding <./knowledge/pdf_embedding.html>`_: supported pdf embedding." - -#: ../../modules/knowledge.rst:130 59d45732f4914d16b4e01aee0992edf7 -#, fuzzy -msgid "" -"`word_embedding <./knowledge/word/word_embedding.html>`_: supported word " -"embedding." -msgstr "pdf_embedding <./knowledge/pdf_embedding.html>`_: supported pdf embedding." - -#: ../../modules/knowledge.rst:131 df0e6f311861423e885b38e020a7c0f0 -#, fuzzy -msgid "" -"`url_embedding <./knowledge/url/url_embedding.html>`_: supported url " -"embedding." -msgstr "pdf_embedding <./knowledge/pdf_embedding.html>`_: supported pdf embedding." - -#: ../../modules/knowledge.rst:132 7c550c1f5bc34fe9986731fb465e12cd -#, fuzzy -msgid "" -"`ppt_embedding <./knowledge/ppt/ppt_embedding.html>`_: supported ppt " -"embedding." -msgstr "pdf_embedding <./knowledge/pdf_embedding.html>`_: supported pdf embedding." - -#: ../../modules/knowledge.rst:133 8648684cb191476faeeb548389f79050 -#, fuzzy -msgid "" -"`string_embedding <./knowledge/string/string_embedding.html>`_: supported" -" raw text embedding." -msgstr "pdf_embedding <./knowledge/pdf_embedding.html>`_: supported pdf embedding." - #~ msgid "before execution: python -m spacy download zh_core_web_sm" #~ msgstr "在执行之前请先执行python -m spacy download zh_core_web_sm" @@ -201,3 +140,112 @@ msgstr "pdf_embedding <./knowledge/pdf_embedding.html>`_: supported pdf embeddin #~ "and MILVUS_PORT)" #~ msgstr "2.更新你的.env,设置你的向量存储类型,VECTOR_STORE_TYPE=Chroma(现在只支持Chroma和Milvus,如果你设置了Milvus,请设置MILVUS_URL和MILVUS_PORT)" +#~ msgid "" +#~ "We currently support many document " +#~ "formats: raw text, txt, pdf, md, " +#~ "html, doc, ppt, and url. In the" +#~ " future, we will continue to support" +#~ " more types of knowledge, including " +#~ "audio, video, various databases, and big" +#~ " data sources. Of course, we look " +#~ "forward to your active participation in" +#~ " contributing code." +#~ msgstr "" + +#~ msgid "1.prepare" +#~ msgstr "准备" + +#~ msgid "" +#~ "2.prepare embedding model, you can " +#~ "download from https://huggingface.co/. Notice " +#~ "you have installed git-lfs." +#~ msgstr "" +#~ "提前准备Embedding Model, 你可以在https://huggingface.co/进行下载,注意" +#~ ":你需要先安装git-lfs.eg: git clone " +#~ "https://huggingface.co/THUDM/chatglm2-6b" + +#~ msgid "" +#~ "eg: git clone https://huggingface.co/sentence-" +#~ "transformers/all-MiniLM-L6-v2" +#~ msgstr "" +#~ "eg: git clone https://huggingface.co/sentence-" +#~ "transformers/all-MiniLM-L6-v2" + +#~ msgid "" +#~ "3.prepare vector_store instance and vector " +#~ "store config, now we support Chroma, " +#~ "Milvus and Weaviate." +#~ msgstr "提前准备向量数据库环境,目前支持Chroma, Milvus and Weaviate向量数据库" + +#~ msgid "" +#~ "3.init Url Type EmbeddingEngine api and" +#~ " embedding your document into vector " +#~ "store in your code." +#~ msgstr "初始化 Url类型 EmbeddingEngine api, 将url文档embedding向量化到向量数据库 " + +#~ msgid "If you want to add your source_reader or text_splitter, do this:" +#~ msgstr "如果你想手动添加你自定义的source_reader和text_splitter, 请参考:" + +#~ msgid "" +#~ "4.init Document Type EmbeddingEngine api " +#~ "and embedding your document into vector" +#~ " store in your code. Document type" +#~ " can be .txt, .pdf, .md, .doc, " +#~ ".ppt." +#~ msgstr "" +#~ "初始化 文档型类型 EmbeddingEngine api, " +#~ "将文档embedding向量化到向量数据库(文档可以是.txt, .pdf, .md, .html," +#~ " .doc, .ppt)" + +#~ msgid "" +#~ "5.init TEXT Type EmbeddingEngine api and" +#~ " embedding your document into vector " +#~ "store in your code." +#~ msgstr "初始化TEXT类型 EmbeddingEngine api, 将文档embedding向量化到向量数据库" + +#~ msgid "4.similar search based on your knowledge base. ::" +#~ msgstr "在知识库进行相似性搜索" + +#~ msgid "" +#~ "`pdf_embedding <./knowledge/pdf/pdf_embedding.html>`_: " +#~ "supported pdf embedding." +#~ msgstr "" +#~ "pdf_embedding <./knowledge/pdf_embedding.html>`_: " +#~ "supported pdf embedding." + +#~ msgid "" +#~ "`markdown_embedding " +#~ "<./knowledge/markdown/markdown_embedding.html>`_: supported " +#~ "markdown embedding." +#~ msgstr "" +#~ "pdf_embedding <./knowledge/pdf_embedding.html>`_: " +#~ "supported pdf embedding." + +#~ msgid "" +#~ "`word_embedding <./knowledge/word/word_embedding.html>`_: " +#~ "supported word embedding." +#~ msgstr "" +#~ "pdf_embedding <./knowledge/pdf_embedding.html>`_: " +#~ "supported pdf embedding." + +#~ msgid "" +#~ "`url_embedding <./knowledge/url/url_embedding.html>`_: " +#~ "supported url embedding." +#~ msgstr "" +#~ "pdf_embedding <./knowledge/pdf_embedding.html>`_: " +#~ "supported pdf embedding." + +#~ msgid "" +#~ "`ppt_embedding <./knowledge/ppt/ppt_embedding.html>`_: " +#~ "supported ppt embedding." +#~ msgstr "" +#~ "pdf_embedding <./knowledge/pdf_embedding.html>`_: " +#~ "supported pdf embedding." + +#~ msgid "" +#~ "`string_embedding <./knowledge/string/string_embedding.html>`_:" +#~ " supported raw text embedding." +#~ msgstr "" +#~ "pdf_embedding <./knowledge/pdf_embedding.html>`_: " +#~ "supported pdf embedding." + diff --git a/docs/requirements.txt b/docs/requirements.txt index 8cee184af..d0a125868 100644 --- a/docs/requirements.txt +++ b/docs/requirements.txt @@ -9,9 +9,9 @@ sphinx_book_theme sphinx_rtd_theme==1.0.0 sphinx-typlog-theme==0.8.0 sphinx-panels +sphinx-tabs==3.4.0 toml myst_nb sphinx_copybutton pydata-sphinx-theme==0.13.1 -pydantic-settings furo \ No newline at end of file