mirror of
https://github.com/csunny/DB-GPT.git
synced 2025-07-31 15:47:05 +00:00
docs: update llms usage
This commit is contained in:
parent
c522f4f0ba
commit
bb9724cbc1
@ -8,7 +8,7 @@ msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: DB-GPT 0.1.0\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2023-06-14 21:47+0800\n"
|
||||
"POT-Creation-Date: 2023-06-14 22:33+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language: zh_CN\n"
|
||||
@ -19,11 +19,11 @@ msgstr ""
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Generated-By: Babel 2.12.1\n"
|
||||
|
||||
#: ../../modules/llms.md:1 d924f8db422d4adaa4cf0e5d17ffe332
|
||||
#: ../../modules/llms.md:1 7737e31f5fcc4dc4be573a0bb73ca419
|
||||
msgid "LLMs"
|
||||
msgstr "大语言模型"
|
||||
|
||||
#: ../../modules/llms.md:3 0fba9725074a4a72be1849e822e69365
|
||||
#: ../../modules/llms.md:3 8a8422f18e5d4c7aa1c1abf3a89f5d27
|
||||
#, python-format
|
||||
msgid ""
|
||||
"In the underlying large model integration, we have designed an open "
|
||||
@ -36,23 +36,23 @@ msgid ""
|
||||
"of use."
|
||||
msgstr "在底层大模型接入中,我们设计了开放的接口,支持对接多种大模型。同时对于接入模型的效果,我们有非常严格的把控与评审机制。对大模型能力上与ChatGPT对比,在准确率上需要满足85%以上的能力对齐。我们用更高的标准筛选模型,是期望在用户使用过程中,可以省去前面繁琐的测试评估环节。"
|
||||
|
||||
#: ../../modules/llms.md:5 4f65b4e86d8d4b35b95ccac6a96f9b9c
|
||||
#: ../../modules/llms.md:5 c48ee3f4c51c49ae9dba40b1854dd483
|
||||
msgid "Multi LLMs Usage"
|
||||
msgstr "多模型使用"
|
||||
|
||||
#: ../../modules/llms.md:6 246e4dbf1c634d9d9f2ab40b4883de73
|
||||
#: ../../modules/llms.md:6 ad44cef629f64bf4a4d72431568149fe
|
||||
msgid ""
|
||||
"To use multiple models, modify the LLM_MODEL parameter in the .env "
|
||||
"configuration file to switch between the models."
|
||||
msgstr "如果要使用不同的模型,请修改.env配置文件中的LLM MODEL参数以在模型之间切换。"
|
||||
|
||||
#: ../../modules/llms.md:8 379d7756094f4ceda5ef163031b44e97
|
||||
#: ../../modules/llms.md:8 b08325fd36af4ef582c8a46685986aaf
|
||||
msgid ""
|
||||
"Notice: you can create .env file from .env.template, just use command "
|
||||
"like this:"
|
||||
msgstr "注意:你可以从 .env.template 创建 .env 文件。只需使用如下命令:"
|
||||
|
||||
#: ../../modules/llms.md:14 3b46be1be08b41b0a4f2fa7ca1f09ccf
|
||||
#: ../../modules/llms.md:14 75ec45409dc84fd9bf3dfa98835d4645
|
||||
msgid ""
|
||||
"now we support models vicuna-13b, vicuna-7b, chatglm-6b, flan-t5-base, "
|
||||
"guanaco-33b-merged, falcon-40b, gorilla-7b."
|
||||
@ -60,7 +60,29 @@ msgstr ""
|
||||
"现在我们支持的模型有vicuna-13b, vicuna-7b, chatglm-6b, flan-t5-base, guanaco-33b-"
|
||||
"merged, falcon-40b, gorilla-7b."
|
||||
|
||||
#: ../../modules/llms.md:16 f7e26b7e7a8349159b5c2e8ec2a1abc8
|
||||
#: ../../modules/llms.md:16 a6e99cf291e049c99e2d6813d03b0427
|
||||
msgid ""
|
||||
"if you want use other model, such as chatglm-6b, you just need update "
|
||||
".env config file."
|
||||
msgstr "如果你想使用其他模型,比如chatglm-6b, 仅仅需要修改.env 配置文件"
|
||||
|
||||
#: ../../modules/llms.md:21 1b9a9aa83dc7420fb5c6c17c982abb20
|
||||
msgid "Run Model with cpu."
|
||||
msgstr "用CPU运行模型"
|
||||
|
||||
#: ../../modules/llms.md:22 e6bdeeca06764ed583154ddc78f0f26e
|
||||
msgid ""
|
||||
"we alse support smaller models, like gpt4all. you can use it with "
|
||||
"cpu/mps(M1/M2), Download from [gpt4all model](https://gpt4all.io/models"
|
||||
"/ggml-gpt4all-j-v1.3-groovy.bin)"
|
||||
msgstr "我们也支持一些小模型,你可以通过CPU/MPS(M1、M2)运行, 模型下载[gpt4all](https://gpt4all.io/models"
|
||||
"/ggml-gpt4all-j-v1.3-groovy.bin)"
|
||||
|
||||
#: ../../modules/llms.md:24 da5b575421ef45e7ad0f56fac151948d
|
||||
msgid "put it in the models path, then change .env config."
|
||||
msgstr "将模型放在models路径, 修改.env 配置文件"
|
||||
|
||||
#: ../../modules/llms.md:29 03453d147e64404ab2c116faf0147b70
|
||||
msgid ""
|
||||
"DB-GPT provides a model load adapter and chat adapter. load adapter which"
|
||||
" allows you to easily adapt load different LLM models by inheriting the "
|
||||
@ -69,15 +91,15 @@ msgstr ""
|
||||
"DB-GPT提供了多模型适配器load adapter和chat adapter.load adapter通过继承BaseLLMAdapter类,"
|
||||
" 实现match和loader方法允许你适配不同的LLM."
|
||||
|
||||
#: ../../modules/llms.md:18 83ad29017489428cad336894de9b1076
|
||||
#: ../../modules/llms.md:31 06b14da2931349859182473cd79abd68
|
||||
msgid "vicuna llm load adapter"
|
||||
msgstr "vicuna llm load adapter"
|
||||
|
||||
#: ../../modules/llms.md:35 56997156cf994abbb12a586d435b1f6c
|
||||
#: ../../modules/llms.md:48 fe7be51e9e2240c1882ef05f94a39d90
|
||||
msgid "chatglm load adapter"
|
||||
msgstr "chatglm load adapter"
|
||||
|
||||
#: ../../modules/llms.md:62 e0d989155ca443079402921fd15c1b75
|
||||
#: ../../modules/llms.md:75 3535d4e0a0b946a49ed13710ae0ae5f3
|
||||
msgid ""
|
||||
"chat adapter which allows you to easily adapt chat different LLM models "
|
||||
"by inheriting the BaseChatAdpter.you just implement match() and "
|
||||
@ -86,43 +108,43 @@ msgstr ""
|
||||
"chat "
|
||||
"adapter通过继承BaseChatAdpter允许你通过实现match和get_generate_stream_func方法允许你适配不同的LLM."
|
||||
|
||||
#: ../../modules/llms.md:64 1cef2736bdd04dc89a3f9f20d6ab7e4e
|
||||
#: ../../modules/llms.md:77 b8dba90d769d45f090086fa044f22a96
|
||||
msgid "vicuna llm chat adapter"
|
||||
msgstr "vicuna llm chat adapter"
|
||||
|
||||
#: ../../modules/llms.md:76 0976f3a966b24e458b968c4410e98403
|
||||
#: ../../modules/llms.md:89 f1d6e8145f704b5bbd1c49224c1e30f9
|
||||
msgid "chatglm llm chat adapter"
|
||||
msgstr "chatglm llm chat adapter"
|
||||
|
||||
#: ../../modules/llms.md:89 138577d6f1714b4daef8b20108de983d
|
||||
#: ../../modules/llms.md:102 295e498cec384d9589431b1d5942f590
|
||||
msgid ""
|
||||
"if you want to integrate your own model, just need to inheriting "
|
||||
"BaseLLMAdaper and BaseChatAdpter and implement the methods"
|
||||
msgstr "如果你想集成自己的模型,只需要继承BaseLLMAdaper和BaseChatAdpter类,然后实现里面的方法即可"
|
||||
|
||||
#: ../../modules/llms.md:92 23cd5154bf3a4e7bad206d5e06eef51e
|
||||
#: ../../modules/llms.md:104 07dd4757b06440fe8e9959446ff05892
|
||||
#, fuzzy
|
||||
msgid "Multi Proxy LLMs"
|
||||
msgstr "多模型使用"
|
||||
|
||||
#: ../../modules/llms.md:93 db2739caa40342819965856d6fe83677
|
||||
#: ../../modules/llms.md:105 1b9fc9ce08b94f6493f4b6ce51878fe2
|
||||
msgid "1. Openai proxy"
|
||||
msgstr ""
|
||||
|
||||
#: ../../modules/llms.md:94 5a305d769a144543b14f6e7099d0fc81
|
||||
#: ../../modules/llms.md:106 64e44b3c7c254034a1f72d2e362f4c4d
|
||||
msgid ""
|
||||
"If you haven't deployed a private infrastructure for a large model, or if"
|
||||
" you want to use DB-GPT in a low-cost and high-efficiency way, you can "
|
||||
"also use OpenAI's large model as your underlying model."
|
||||
msgstr ""
|
||||
|
||||
#: ../../modules/llms.md:96 64d78d8e95174a329f67e638fc63786c
|
||||
#: ../../modules/llms.md:108 bca7d8118cd546ca8160400fe729be89
|
||||
msgid ""
|
||||
"If your environment deploying DB-GPT has access to OpenAI, then modify "
|
||||
"the .env configuration file as below will work."
|
||||
msgstr ""
|
||||
|
||||
#: ../../modules/llms.md:104 f2bcf0c04da844e0b8fa2a3712d56439
|
||||
#: ../../modules/llms.md:116 7b84ddb937954787b4c8422f743afeda
|
||||
msgid ""
|
||||
"If you can't access OpenAI locally but have an OpenAI proxy service, you "
|
||||
"can configure as follows."
|
||||
|
@ -8,7 +8,7 @@ msgid ""
|
||||
msgstr ""
|
||||
"Project-Id-Version: DB-GPT 0.1.0\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2023-06-11 14:10+0800\n"
|
||||
"POT-Creation-Date: 2023-06-14 22:33+0800\n"
|
||||
"PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n"
|
||||
"Last-Translator: FULL NAME <EMAIL@ADDRESS>\n"
|
||||
"Language: zh_CN\n"
|
||||
@ -17,13 +17,13 @@ msgstr ""
|
||||
"MIME-Version: 1.0\n"
|
||||
"Content-Type: text/plain; charset=utf-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Generated-By: Babel 2.11.0\n"
|
||||
"Generated-By: Babel 2.12.1\n"
|
||||
|
||||
#: ../../modules/prompts.md:1 bb9583334e6948b98b59126234ae045f
|
||||
#: ../../modules/prompts.md:1 b9279c238c014a74aecbc75b5d3dc202
|
||||
msgid "Prompts"
|
||||
msgstr ""
|
||||
|
||||
#: ../../modules/prompts.md:3 e6f5129e260c4a739a40115fff82850f
|
||||
#: ../../modules/prompts.md:3 f0c720c1c85b401cbc26ed0eb3f6e70e
|
||||
msgid ""
|
||||
"Prompt is a very important part of the interaction between the large "
|
||||
"model and the user, and to a certain extent, it determines the quality "
|
||||
@ -31,7 +31,9 @@ msgid ""
|
||||
" we will automatically optimize the corresponding prompt according to "
|
||||
"user input and usage scenarios, making it easier and more efficient for "
|
||||
"users to use large language models."
|
||||
msgstr "Prompt是与大模型交互过程中非常重要的部分,一定程度上Prompt决定了"
|
||||
"大模型生成答案的质量与准确性,在本的项目中,我们会根据用户输入与"
|
||||
"使用场景,自动优化对应的Prompt,让用户使用大语言模型变得更简单、更高效。"
|
||||
msgstr "Prompt是与大模型交互过程中非常重要的部分,一定程度上Prompt决定了大模型生成答案的质量与准确性,在本的项目中,我们会根据用户输入与使用场景,自动优化对应的Prompt,让用户使用大语言模型变得更简单、更高效。"
|
||||
|
||||
#: ../../modules/prompts.md:5 6576d32e28a14be6a5d8180eed000aa7
|
||||
msgid "1.DB-GPT Prompt"
|
||||
msgstr ""
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user