diff --git a/gpt4all-chat/CHANGELOG.md b/gpt4all-chat/CHANGELOG.md
index 47926f05..c774a040 100644
--- a/gpt4all-chat/CHANGELOG.md
+++ b/gpt4all-chat/CHANGELOG.md
@@ -23,8 +23,9 @@ The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.1.0/).
- Set the window icon on Linux ([#2880](https://github.com/nomic-ai/gpt4all/pull/2880))
- Corrections to the Romanian translation (by [@SINAPSA-IC](https://github.com/SINAPSA-IC) in [#2890](https://github.com/nomic-ai/gpt4all/pull/2890))
- Fix singular/plural forms of LocalDocs "x Sources" (by [@cosmic-snow](https://github.com/cosmic-snow) in [#2885](https://github.com/nomic-ai/gpt4all/pull/2885))
-- Fixed typo in several files. (by [@3Simplex](https://github.com/3Simplex) in [#2916](https://github.com/nomic-ai/gpt4all/pull/2916))
+- Fix a typo in Model Settings (by [@3Simplex](https://github.com/3Simplex) in [#2916](https://github.com/nomic-ai/gpt4all/pull/2916))
- Fix the antenna icon tooltip when using the local server ([#2922](https://github.com/nomic-ai/gpt4all/pull/2922))
+- Fix a few issues with locating files and handling errors when loading remote models on startup ([#2875](https://github.com/nomic-ai/gpt4all/pull/2875))
## [3.2.1] - 2024-08-13
diff --git a/gpt4all-chat/src/modellist.cpp b/gpt4all-chat/src/modellist.cpp
index 13711057..d53c8fbf 100644
--- a/gpt4all-chat/src/modellist.cpp
+++ b/gpt4all-chat/src/modellist.cpp
@@ -1208,132 +1208,139 @@ bool ModelList::modelExists(const QString &modelFilename) const
return false;
}
+void ModelList::updateOldRemoteModels(const QString &path)
+{
+ QDirIterator it(path, QDir::Files, QDirIterator::Subdirectories);
+ while (it.hasNext()) {
+ QFileInfo info = it.nextFileInfo();
+ QString filename = it.fileName();
+ if (!filename.startsWith("chatgpt-") || !filename.endsWith(".txt"))
+ continue;
+
+ QString apikey;
+ QString modelname(filename);
+ modelname.chop(4); // strip ".txt" extension
+ modelname.remove(0, 8); // strip "chatgpt-" prefix
+ QFile file(info.filePath());
+ if (!file.open(QIODevice::ReadOnly)) {
+ qWarning().noquote() << tr("cannot open \"%1\": %2").arg(file.fileName(), file.errorString());
+ continue;
+ }
+
+ {
+ QTextStream in(&file);
+ apikey = in.readAll();
+ file.close();
+ }
+
+ QFile newfile(u"%1/gpt4all-%2.rmodel"_s.arg(info.dir().path(), modelname));
+ if (!newfile.open(QIODevice::ReadWrite)) {
+ qWarning().noquote() << tr("cannot create \"%1\": %2").arg(newfile.fileName(), file.errorString());
+ continue;
+ }
+
+ QJsonObject obj {
+ { "apiKey", apikey },
+ { "modelName", modelname },
+ };
+
+ QTextStream out(&newfile);
+ out << QJsonDocument(obj).toJson();
+ newfile.close();
+
+ file.remove();
+ }
+}
+
+void ModelList::processModelDirectory(const QString &path)
+{
+ QDirIterator it(path, QDir::Files, QDirIterator::Subdirectories);
+ while (it.hasNext()) {
+ QFileInfo info = it.nextFileInfo();
+
+ QString filename = it.fileName();
+ if (filename.startsWith("incomplete") || FILENAME_BLACKLIST.contains(filename))
+ continue;
+ if (!filename.endsWith(".gguf") && !filename.endsWith(".rmodel"))
+ continue;
+
+ bool isOnline(filename.endsWith(".rmodel"));
+ bool isCompatibleApi(filename.endsWith("-capi.rmodel"));
+
+ QString name;
+ QString description;
+ if (isCompatibleApi) {
+ QJsonObject obj;
+ {
+ QFile file(info.filePath());
+ if (!file.open(QIODeviceBase::ReadOnly)) {
+ qWarning().noquote() << tr("cannot open \"%1\": %2").arg(file.fileName(), file.errorString());
+ continue;
+ }
+ QJsonDocument doc = QJsonDocument::fromJson(file.readAll());
+ obj = doc.object();
+ }
+ {
+ QString apiKey(obj["apiKey"].toString());
+ QString baseUrl(obj["baseUrl"].toString());
+ QString modelName(obj["modelName"].toString());
+ apiKey = apiKey.length() < 10 ? "*****" : apiKey.left(5) + "*****";
+ name = tr("%1 (%2)").arg(modelName, baseUrl);
+ description = tr("OpenAI-Compatible API Model "
+ "
API Key: %1
"
+ "
Base URL: %2
"
+ "
Model Name: %3
")
+ .arg(apiKey, baseUrl, modelName);
+ }
+ }
+
+ QVector modelsById;
+ {
+ QMutexLocker locker(&m_mutex);
+ for (ModelInfo *info : m_models)
+ if (info->filename() == filename)
+ modelsById.append(info->id());
+ }
+
+ if (modelsById.isEmpty()) {
+ if (!contains(filename))
+ addModel(filename);
+ modelsById.append(filename);
+ }
+
+ for (const QString &id : modelsById) {
+ QVector> data {
+ { InstalledRole, true },
+ { FilenameRole, filename },
+ { OnlineRole, isOnline },
+ { CompatibleApiRole, isCompatibleApi },
+ { DirpathRole, info.dir().absolutePath() + "/" },
+ { FilesizeRole, toFileSize(info.size()) },
+ };
+ if (isCompatibleApi) {
+ // The data will be saved to "GPT4All.ini".
+ data.append({ NameRole, name });
+ // The description is hard-coded into "GPT4All.ini" due to performance issue.
+ // If the description goes to be dynamic from its .rmodel file, it will get high I/O usage while using the ModelList.
+ data.append({ DescriptionRole, description });
+ // Prompt template should be clear while using ChatML format which is using in most of OpenAI-Compatible API server.
+ data.append({ PromptTemplateRole, "%1" });
+ }
+ updateData(id, data);
+ }
+ }
+}
+
void ModelList::updateModelsFromDirectory()
{
const QString exePath = QCoreApplication::applicationDirPath() + QDir::separator();
const QString localPath = MySettings::globalInstance()->modelPath();
- auto updateOldRemoteModels = [&](const QString& path) {
- QDirIterator it(path, QDirIterator::Subdirectories);
- while (it.hasNext()) {
- it.next();
- if (!it.fileInfo().isDir()) {
- QString filename = it.fileName();
- if (filename.startsWith("chatgpt-") && filename.endsWith(".txt")) {
- QString apikey;
- QString modelname(filename);
- modelname.chop(4); // strip ".txt" extension
- modelname.remove(0, 8); // strip "chatgpt-" prefix
- QFile file(path + filename);
- if (file.open(QIODevice::ReadWrite)) {
- QTextStream in(&file);
- apikey = in.readAll();
- file.close();
- }
-
- QJsonObject obj;
- obj.insert("apiKey", apikey);
- obj.insert("modelName", modelname);
- QJsonDocument doc(obj);
-
- auto newfilename = u"gpt4all-%1.rmodel"_s.arg(modelname);
- QFile newfile(path + newfilename);
- if (newfile.open(QIODevice::ReadWrite)) {
- QTextStream out(&newfile);
- out << doc.toJson();
- newfile.close();
- }
- file.remove();
- }
- }
- }
- };
-
- auto processDirectory = [&](const QString& path) {
- QDirIterator it(path, QDir::Files, QDirIterator::Subdirectories);
- while (it.hasNext()) {
- it.next();
-
- QString filename = it.fileName();
- if (filename.startsWith("incomplete") || FILENAME_BLACKLIST.contains(filename))
- continue;
- if (!filename.endsWith(".gguf") && !filename.endsWith(".rmodel"))
- continue;
-
- QVector modelsById;
- {
- QMutexLocker locker(&m_mutex);
- for (ModelInfo *info : m_models)
- if (info->filename() == filename)
- modelsById.append(info->id());
- }
-
- if (modelsById.isEmpty()) {
- if (!contains(filename))
- addModel(filename);
- modelsById.append(filename);
- }
-
- QFileInfo info = it.fileInfo();
-
- bool isOnline(filename.endsWith(".rmodel"));
- bool isCompatibleApi(filename.endsWith("-capi.rmodel"));
-
- QString name;
- QString description;
- if (isCompatibleApi) {
- QJsonObject obj;
- {
- QFile file(path + filename);
- bool success = file.open(QIODeviceBase::ReadOnly);
- (void)success;
- Q_ASSERT(success);
- QJsonDocument doc = QJsonDocument::fromJson(file.readAll());
- obj = doc.object();
- }
- {
- QString apiKey(obj["apiKey"].toString());
- QString baseUrl(obj["baseUrl"].toString());
- QString modelName(obj["modelName"].toString());
- apiKey = apiKey.length() < 10 ? "*****" : apiKey.left(5) + "*****";
- name = tr("%1 (%2)").arg(modelName, baseUrl);
- description = tr("OpenAI-Compatible API Model "
- "
API Key: %1
"
- "
Base URL: %2
"
- "
Model Name: %3
")
- .arg(apiKey, baseUrl, modelName);
- }
- }
-
- for (const QString &id : modelsById) {
- QVector> data {
- { InstalledRole, true },
- { FilenameRole, filename },
- { OnlineRole, isOnline },
- { CompatibleApiRole, isCompatibleApi },
- { DirpathRole, info.dir().absolutePath() + "/" },
- { FilesizeRole, toFileSize(info.size()) },
- };
- if (isCompatibleApi) {
- // The data will be saved to "GPT4All.ini".
- data.append({ NameRole, name });
- // The description is hard-coded into "GPT4All.ini" due to performance issue.
- // If the description goes to be dynamic from its .rmodel file, it will get high I/O usage while using the ModelList.
- data.append({ DescriptionRole, description });
- // Prompt template should be clear while using ChatML format which is using in most of OpenAI-Compatible API server.
- data.append({ PromptTemplateRole, "%1" });
- }
- updateData(id, data);
- }
- }
- };
-
-
updateOldRemoteModels(exePath);
- processDirectory(exePath);
+ processModelDirectory(exePath);
if (localPath != exePath) {
updateOldRemoteModels(localPath);
- processDirectory(localPath);
+ processModelDirectory(localPath);
}
}
diff --git a/gpt4all-chat/src/modellist.h b/gpt4all-chat/src/modellist.h
index 7c13da8e..21d9aeef 100644
--- a/gpt4all-chat/src/modellist.h
+++ b/gpt4all-chat/src/modellist.h
@@ -502,6 +502,8 @@ private:
void parseModelsJsonFile(const QByteArray &jsonData, bool save);
void parseDiscoveryJsonFile(const QByteArray &jsonData);
QString uniqueModelName(const ModelInfo &model) const;
+ void updateOldRemoteModels(const QString &path);
+ void processModelDirectory(const QString &path);
private:
mutable QMutex m_mutex;
diff --git a/gpt4all-chat/translations/gpt4all_en_US.ts b/gpt4all-chat/translations/gpt4all_en_US.ts
index be7dee3f..e46ae94c 100644
--- a/gpt4all-chat/translations/gpt4all_en_US.ts
+++ b/gpt4all-chat/translations/gpt4all_en_US.ts
@@ -4,62 +4,62 @@
AddCollectionView
-
+ ← Existing Collections
-
+ Add Document Collection
-
+ Add a folder containing plain text files, PDFs, or Markdown. Configure additional extensions in Settings.
-
+ Please choose a directory
-
+ Name
-
+ Collection name...
-
+ Name of the collection to add (Required)
-
+ Folder
-
+ Folder path...
-
+ Folder path to documents (Required)
-
+ Browse
-
+ Create Collection
@@ -67,288 +67,288 @@
AddModelView
-
+ ← Existing Models
-
+ Explore Models
-
+ Discover and download models by keyword search...
-
+ Text field for discovering and filtering downloadable models
-
+ Initiate model discovery and filtering
-
+ Triggers discovery and filtering of models
-
+ Default
-
+ Likes
-
+ Downloads
-
+ Recent
-
+ Asc
-
+ Desc
-
+ None
-
+ Searching · %1
-
+ Sort by: %1
-
+ Sort dir: %1
-
+ Limit: %1
-
+ Network error: could not retrieve %1
-
-
+
+ Busy indicator
-
+ Displayed when the models request is ongoing
-
+ Model file
-
+ Model file to be downloaded
-
+ Description
-
+ File description
-
+ Cancel
-
+ Resume
-
+ Download
-
+ Stop/restart/start the download
-
+ Remove
-
+ Remove model from filesystem
-
-
+
+ Install
-
+ Install online model
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font>
-
+ ERROR: $API_KEY is empty.
-
+ ERROR: $BASE_URL is empty.
-
+ enter $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.
-
+ enter $MODEL_NAME
-
+ %1 GB
-
-
+
+ ?
-
+ Describes an error that occurred when downloading
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font>
-
+ Error for incompatible hardware
-
+ Download progressBar
-
+ Shows the progress made in the download
-
+ Download speed
-
+ Download speed in bytes/kilobytes/megabytes per second
-
+ Calculating...
-
-
-
-
+
+
+
+ Whether the file hash is being calculated
-
+ Displayed when the file hash is being calculated
-
+ enter $API_KEY
-
+ File size
-
+ RAM required
-
+ Parameters
-
+ Quant
-
+ Type
@@ -356,22 +356,22 @@
ApplicationSettings
-
+ Application
-
+ Network dialog
-
+ opt-in to share feedback/conversations
-
+ ERROR: Update system could not find the MaintenanceTool used<br>
to check for updates!<br><br>
Did you install this application using the online installer? If so,<br>
@@ -382,223 +382,223 @@
-
+ Error dialog
-
+ Application Settings
-
+ General
-
+ Theme
-
+ The application color scheme.
-
+ Dark
-
+ Light
-
+ LegacyDark
-
+ Font Size
-
+ The size of text in the application.
-
+ Small
-
+ Medium
-
+ Large
-
+ Language and Locale
-
+ The language and locale you wish to use.
-
+ System Locale
-
+ Device
-
+ The compute device used for text generation.
-
-
+
+ Application default
-
+ Default Model
-
+ The preferred model for new chats. Also used as the local server fallback.
-
+ Suggestion Mode
-
+ Generate suggested follow-up questions at the end of responses.
-
+ When chatting with LocalDocs
-
+ Whenever possible
-
+ Never
-
+ Download Path
-
+ Where to store local models and the LocalDocs database.
-
+ Browse
-
+ Choose where to save model files
-
+ Enable Datalake
-
+ Send chats and feedback to the GPT4All Open-Source Datalake.
-
+ Advanced
-
+ CPU Threads
-
+ The number of CPU threads used for inference and embedding.
-
+ Save Chat Context
-
+ Save the chat model's state to disk for faster loading. WARNING: Uses ~2GB per chat.
-
+ Enable Local Server
-
+ Expose an OpenAI-Compatible server to localhost. WARNING: Results in increased resource usage.
-
+ API Server Port
-
+ The port to use for the local server. Requires restart.
-
+ Check For Updates
-
+ Manually check for an update to GPT4All.
-
+ Updates
@@ -606,13 +606,13 @@
Chat
-
-
+
+ New Chat
-
+ Server Chat
@@ -620,12 +620,12 @@
ChatAPIWorker
-
+ ERROR: Network error occurred while connecting to the API server
-
+ ChatAPIWorker::handleFinished got HTTP Error %1 %2
@@ -633,62 +633,62 @@
ChatDrawer
-
+ Drawer
-
+ Main navigation drawer
-
+ + New Chat
-
+ Create a new chat
-
+ Select the current chat or edit the chat when in edit mode
-
+ Edit chat name
-
+ Save chat name
-
+ Delete chat
-
+ Confirm chat deletion
-
+ Cancel chat deletion
-
+ List of chats
-
+ List of chats in the drawer dialog
@@ -696,32 +696,32 @@
ChatListModel
-
+ TODAY
-
+ THIS WEEK
-
+ THIS MONTH
-
+ LAST SIX MONTHS
-
+ THIS YEAR
-
+ LAST YEAR
@@ -729,215 +729,215 @@
ChatView
-
+ <h3>Warning</h3><p>%1</p>
-
+ Switch model dialog
-
+ Warn the user if they switch models, then context will be erased
-
+ Conversation copied to clipboard.
-
+ Code copied to clipboard.
-
+ Chat panel
-
+ Chat panel with options
-
+ Reload the currently loaded model
-
+ Eject the currently loaded model
-
+ No model installed.
-
+ Model loading error.
-
+ Waiting for model...
-
+ Switching context...
-
+ Choose a model...
-
+ Not found: %1
-
+ The top item is the current model
-
-
+
+ LocalDocs
-
+ Add documents
-
+ add collections of documents to the chat
-
+ Load the default model
-
+ Loads the default model which can be changed in settings
-
+ No Model Installed
-
+ GPT4All requires that you install at least one
model to get started
-
+ Install a Model
-
+ Shows the add model view
-
+ Conversation with the model
-
+ prompt / response pairs from the conversation
-
+ GPT4All
-
+ You
-
+ response stopped ...
-
+ processing ...
-
+ generating response ...
-
+ generating questions ...
-
-
+
+ Copy
-
+ Copy Message
-
+ Disable markdown
-
+ Enable markdown
-
+ Thumbs up
-
+ Gives a thumbs up to the response
-
+ Thumbs down
-
+ Opens thumbs down dialog
-
+ %n Source(s)%n Source
@@ -945,113 +945,113 @@ model to get started
-
+ Suggested follow-ups
-
+ Erase and reset chat session
-
+ Copy chat session to clipboard
-
+ Redo last chat response
-
+ Stop generating
-
+ Stop the current response generation
-
+ Reloads the model
-
+ <h3>Encountered an error loading model:</h3><br><i>"%1"</i><br><br>Model loading failures can happen for a variety of reasons, but the most common causes include a bad file format, an incomplete or corrupted download, the wrong file type, not enough system RAM or an incompatible model type. Here are some suggestions for resolving the problem:<br><ul><li>Ensure the model file has a compatible format and type<li>Check the model file is complete in the download folder<li>You can find the download folder in the settings dialog<li>If you've sideloaded the model ensure the file is not corrupt by checking md5sum<li>Read more about what models are supported in our <a href="https://docs.gpt4all.io/">documentation</a> for the gui<li>Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help
-
-
+
+ Reload · %1
-
+ Loading · %1
-
+ Load · %1 (default) →
-
+ restoring from text ...
-
+ retrieving localdocs: %1 ...
-
+ searching localdocs: %1 ...
-
+ Send a message...
-
+ Load a model to continue...
-
+ Send messages/prompts to the model
-
+ Cut
-
+ Paste
-
+ Select All
-
+ Send message
-
+ Sends the message/prompt contained in textfield to the model
@@ -1059,12 +1059,12 @@ model to get started
CollectionsDrawer
-
+ Warning: searching collections while indexing can return incomplete results
-
+ %n file(s)%n file
@@ -1072,7 +1072,7 @@ model to get started
-
+ %n word(s)%n word
@@ -1080,17 +1080,17 @@ model to get started
-
+ Updating
-
+ + Add Docs
-
+ Select a collection to make it available to the chat model.
@@ -1098,37 +1098,37 @@ model to get started
Download
-
+ Model "%1" is installed successfully.
-
+ ERROR: $MODEL_NAME is empty.
-
+ ERROR: $API_KEY is empty.
-
+ ERROR: $BASE_URL is invalid.
-
+ ERROR: Model "%1 (%2)" is conflict.
-
+ Model "%1 (%2)" is installed successfully.
-
+ Model "%1" is removed.
@@ -1136,92 +1136,92 @@ model to get started
HomeView
-
+ Welcome to GPT4All
-
+ The privacy-first LLM chat application
-
+ Start chatting
-
+ Start Chatting
-
+ Chat with any LLM
-
+ LocalDocs
-
+ Chat with your local files
-
+ Find Models
-
+ Explore and download models
-
+ Latest news
-
+ Latest news from GPT4All
-
+ Release Notes
-
+ Documentation
-
+ Discord
-
+ X (Twitter)
-
+ Github
-
+ nomic.ai
-
+ Subscribe to Newsletter
@@ -1229,117 +1229,117 @@ model to get started
LocalDocsSettings
-
+ LocalDocs
-
+ LocalDocs Settings
-
+ Indexing
-
+ Allowed File Extensions
-
+ Comma-separated list. LocalDocs will only attempt to process files with these extensions.
-
+ Embedding
-
+ Use Nomic Embed API
-
+ Embed documents using the fast Nomic API instead of a private local model. Requires restart.
-
+ Nomic API Key
-
+ API key to use for Nomic Embed. Get one from the Atlas <a href="https://atlas.nomic.ai/cli-login">API keys page</a>. Requires restart.
-
+ Embeddings Device
-
+ The compute device used for embeddings. Requires restart.
-
+ Application default
-
+ Display
-
+ Show Sources
-
+ Display the sources used for each response.
-
+ Advanced
-
+ Warning: Advanced usage only.
-
+ Values too large may cause localdocs failure, extremely slow responses or failure to respond at all. Roughly speaking, the {N chars x N snippets} are added to the model's context window. More info <a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">here</a>.
-
+ Document snippet size (characters)
-
+ Number of characters per document snippet. Larger numbers increase likelihood of factual responses, but also result in slower generation.
-
+ Max document snippets per prompt
-
+ Max best N matches of retrieved document snippets to add to the context for prompt. Larger numbers increase likelihood of factual responses, but also result in slower generation.
@@ -1347,117 +1347,117 @@ model to get started
LocalDocsView
-
+ LocalDocs
-
+ Chat with your local files
-
+ + Add Collection
-
+ <h3>ERROR: The LocalDocs database cannot be accessed or is not valid.</h3><br><i>Note: You will need to restart after trying any of the following suggested fixes.</i><br><ul><li>Make sure that the folder set as <b>Download Path</b> exists on the file system.</li><li>Check ownership as well as read and write permissions of the <b>Download Path</b>.</li><li>If there is a <b>localdocs_v2.db</b> file, check its ownership and read/write permissions, too.</li></ul><br>If the problem persists and there are any 'localdocs_v*.db' files present, as a last resort you can<br>try backing them up and removing them. You will have to recreate your collections, however.
-
+ No Collections Installed
-
+ Install a collection of local documents to get started using this feature
-
+ + Add Doc Collection
-
+ Shows the add model view
-
+ Indexing progressBar
-
+ Shows the progress made in the indexing
-
+ ERROR
-
+ INDEXING
-
+ EMBEDDING
-
+ REQUIRES UPDATE
-
+ READY
-
+ INSTALLING
-
+ Indexing in progress
-
+ Embedding in progress
-
+ This collection requires an update after version change
-
+ Automatically reindexes upon changes to the folder
-
+ Installation in progress
-
+ %
-
+ %n file(s)%n file
@@ -1465,7 +1465,7 @@ model to get started
-
+ %n word(s)%n word
@@ -1473,27 +1473,27 @@ model to get started
-
+ Remove
-
+ Rebuild
-
+ Reindex this folder from scratch. This is slow and usually not needed.
-
+ Update
-
+ Update the collection to the new version. This is a slow operation.
@@ -1501,67 +1501,78 @@ model to get started
ModelList
-
+ <ul><li>Requires personal OpenAI API key.</li><li>WARNING: Will send your chats to OpenAI!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with OpenAI</li><li>You can apply for an API key <a href="https://platform.openai.com/account/api-keys">here.</a></li>
-
+ <strong>OpenAI's ChatGPT model GPT-3.5 Turbo</strong><br> %1
-
+ <strong>OpenAI's ChatGPT model GPT-4</strong><br> %1 %2
-
+ <strong>Mistral Tiny model</strong><br> %1
-
+ <strong>Mistral Small model</strong><br> %1
-
+ <strong>Mistral Medium model</strong><br> %1
-
+ <br><br><i>* Even if you pay OpenAI for ChatGPT-4 this does not guarantee API key access. Contact OpenAI for more info.
-
+
+
+ cannot open "%1": %2
+
+
+
+
+ cannot create "%1": %2
+
+
+
+ %1 (%2)
-
+ <strong>OpenAI-Compatible API Model</strong><br><ul><li>API Key: %1</li><li>Base URL: %2</li><li>Model Name: %3</li></ul>
-
+ <ul><li>Requires personal Mistral API key.</li><li>WARNING: Will send your chats to Mistral!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with Mistral</li><li>You can apply for an API key <a href="https://console.mistral.ai/user/api-keys">here</a>.</li>
-
+ <ul><li>Requires personal API key and the API base URL.</li><li>WARNING: Will send your chats to the OpenAI-compatible API Server you specified!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with the OpenAI-compatible API Server</li>
-
+ <strong>Connect to OpenAI-compatible API server</strong><br> %1
-
+ <strong>Created by %1.</strong><br><ul><li>Published on %2.<li>This model has %3 likes.<li>This model has %4 downloads.<li>More info can be found <a href="https://huggingface.co/%5">here.</a></ul>
@@ -1569,217 +1580,217 @@ model to get started
ModelSettings
-
+ Model
-
+ Model Settings
-
+ Clone
-
+ Remove
-
+ Name
-
+ Model File
-
+ System Prompt
-
+ Prefixed at the beginning of every conversation. Must contain the appropriate framing tokens.
-
+ Prompt Template
-
+ The template that wraps every prompt.
-
+ Must contain the string "%1" to be replaced with the user's input.
-
+ Chat Name Prompt
-
+ Prompt used to automatically generate chat names.
-
+ Suggested FollowUp Prompt
-
+ Prompt used to generate suggested follow-up questions.
-
+ Context Length
-
+ Number of input and output tokens the model sees.
-
+ Maximum combined prompt/response tokens before information is lost.
Using more context than the model was trained on will yield poor results.
NOTE: Does not take effect until you reload the model.
-
+ Temperature
-
+ Randomness of model output. Higher -> more variation.
-
+ Temperature increases the chances of choosing less likely tokens.
NOTE: Higher temperature gives more creative but less predictable outputs.
-
+ Top-P
-
+ Nucleus Sampling factor. Lower -> more predictable.
-
+ Only the most likely tokens up to a total probability of top_p can be chosen.
NOTE: Prevents choosing highly unlikely tokens.
-
+ Min-P
-
+ Minimum token probability. Higher -> more predictable.
-
+ Sets the minimum relative probability for a token to be considered.
-
+ Top-K
-
+ Size of selection pool for tokens.
-
+ Only the top K most likely tokens will be chosen from.
-
+ Max Length
-
+ Maximum response length, in tokens.
-
+ Prompt Batch Size
-
+ The batch size used for prompt processing.
-
+ Amount of prompt tokens to process at once.
NOTE: Higher values can speed up reading prompts but will use more RAM.
-
+ Repeat Penalty
-
+ Repetition penalty factor. Set to 1 to disable.
-
+ Repeat Penalty Tokens
-
+ Number of previous tokens used for penalty.
-
+ GPU Layers
-
+ Number of model layers to load into VRAM.
-
+ How many model layers to load into VRAM. Decrease this if GPT4All runs out of VRAM while loading this model.
Lower values increase CPU load and RAM usage, and make inference slower.
NOTE: Does not take effect until you reload the model.
@@ -1789,217 +1800,217 @@ NOTE: Does not take effect until you reload the model.
ModelsView
-
+ No Models Installed
-
+ Install a model to get started using GPT4All
-
-
+
+ + Add Model
-
+ Shows the add model view
-
+ Installed Models
-
+ Locally installed chat models
-
+ Model file
-
+ Model file to be downloaded
-
+ Description
-
+ File description
-
+ Cancel
-
+ Resume
-
+ Stop/restart/start the download
-
+ Remove
-
+ Remove model from filesystem
-
-
+
+ Install
-
+ Install online model
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font>
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font>
-
+ ERROR: $API_KEY is empty.
-
+ ERROR: $BASE_URL is empty.
-
+ enter $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.
-
+ enter $MODEL_NAME
-
+ %1 GB
-
+ ?
-
+ Describes an error that occurred when downloading
-
+ Error for incompatible hardware
-
+ Download progressBar
-
+ Shows the progress made in the download
-
+ Download speed
-
+ Download speed in bytes/kilobytes/megabytes per second
-
+ Calculating...
-
-
-
-
+
+
+
+ Whether the file hash is being calculated
-
+ Busy indicator
-
+ Displayed when the file hash is being calculated
-
+ enter $API_KEY
-
+ File size
-
+ RAM required
-
+ Parameters
-
+ Quant
-
+ Type
@@ -2007,12 +2018,12 @@ NOTE: Does not take effect until you reload the model.
MyFancyLink
-
+ Fancy link
-
+ A stylized link
@@ -2020,7 +2031,7 @@ NOTE: Does not take effect until you reload the model.
MySettingsStack
-
+ Please choose a directory
@@ -2028,12 +2039,12 @@ NOTE: Does not take effect until you reload the model.
MySettingsTab
-
+ Restore Defaults
-
+ Restores settings dialog to a default state
@@ -2041,12 +2052,12 @@ NOTE: Does not take effect until you reload the model.
NetworkDialog
-
+ Contribute data to the GPT4All Opensource Datalake.
-
+ By enabling this feature, you will be able to participate in the democratic process of training a large language model by contributing data for future model improvements.
When a GPT4All model responds to you and you have opted-in, your conversation will be sent to the GPT4All Open Source Datalake. Additionally, you can like/dislike its response. If you dislike a response, you can suggest an alternative response. This data will be collected and aggregated in the GPT4All Datalake.
@@ -2055,47 +2066,47 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
-
+ Terms for opt-in
-
+ Describes what will happen when you opt-in
-
+ Please provide a name for attribution (optional)
-
+ Attribution (optional)
-
+ Provide attribution
-
+ Enable
-
+ Enable opt-in
-
+ Cancel
-
+ Cancel opt-in
@@ -2103,17 +2114,17 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
NewVersionDialog
-
+ New version is available
-
+ Update
-
+ Update to new version
@@ -2121,17 +2132,17 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
PopupDialog
-
+ Reveals a shortlived help balloon
-
+ Busy indicator
-
+ Displayed when the popup is showing busy
@@ -2139,28 +2150,28 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
SettingsView
-
-
+
+ Settings
-
+ Contains various application settings
-
+ Application
-
+ Model
-
+ LocalDocs
@@ -2168,29 +2179,29 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
StartupDialog
-
+ Welcome!
-
+ ### Release notes
%1### Contributors
%2
-
+ Release notes
-
+ Release notes for this version
-
+ ### Opt-ins for anonymous usage analytics and datalake
By enabling these features, you will be able to participate in the democratic process of training a
large language model by contributing data for future model improvements.
@@ -2208,71 +2219,71 @@ model release that uses your data!
-
+ Terms for opt-in
-
+ Describes what will happen when you opt-in
-
-
+
+ Opt-in for anonymous usage statistics
-
-
+
+ Yes
-
+ Allow opt-in for anonymous usage statistics
-
-
+
+ No
-
+ Opt-out for anonymous usage statistics
-
+ Allow opt-out for anonymous usage statistics
-
-
+
+ Opt-in for network
-
+ Allow opt-in for network
-
+ Allow opt-in anonymous sharing of chats to the GPT4All Datalake
-
+ Opt-out for network
-
+ Allow opt-out anonymous sharing of chats to the GPT4All Datalake
@@ -2280,23 +2291,23 @@ model release that uses your data!
SwitchModelDialog
-
+ <b>Warning:</b> changing the model will erase the current conversation. Do you wish to continue?
-
+ Continue
-
+ Continue with model loading
-
-
+
+ Cancel
@@ -2304,32 +2315,32 @@ model release that uses your data!
ThumbsDownDialog
-
+ Please edit the text below to provide a better response. (optional)
-
+ Please provide a better response...
-
+ Submit
-
+ Submits the user's response
-
+ Cancel
-
+ Closes the response dialog
@@ -2337,125 +2348,125 @@ model release that uses your data!
main
-
+ <h3>Encountered an error starting up:</h3><br><i>"Incompatible hardware detected."</i><br><br>Unfortunately, your CPU does not meet the minimal requirements to run this program. In particular, it does not support AVX intrinsics which this program requires to successfully run a modern large language model. The only solution at this time is to upgrade your hardware to a more modern CPU.<br><br>See here for more information: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a>
-
+ GPT4All v%1
-
+ <h3>Encountered an error starting up:</h3><br><i>"Inability to access settings file."</i><br><br>Unfortunately, something is preventing the program from accessing the settings file. This could be caused by incorrect permissions in the local app config directory where the settings file is located. Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help.
-
+ Connection to datalake failed.
-
+ Saving chats.
-
+ Network dialog
-
+ opt-in to share feedback/conversations
-
+ Home view
-
+ Home view of application
-
+ Home
-
+ Chat view
-
+ Chat view to interact with models
-
+ Chats
-
-
+
+ Models
-
+ Models view for installed models
-
-
+
+ LocalDocs
-
+ LocalDocs view to configure and use local docs
-
-
+
+ Settings
-
+ Settings view for application configuration
-
+ The datalake is enabled
-
+ Using a network model
-
+ Server mode is enabled
-
+ Installed models
-
+ View of installed models
diff --git a/gpt4all-chat/translations/gpt4all_es_MX.ts b/gpt4all-chat/translations/gpt4all_es_MX.ts
index 0c9f516f..a3290102 100644
--- a/gpt4all-chat/translations/gpt4all_es_MX.ts
+++ b/gpt4all-chat/translations/gpt4all_es_MX.ts
@@ -4,62 +4,62 @@
AddCollectionView
-
+ ← Existing Collections← Colecciones existentes
-
+ Add Document CollectionAgregar colección de documentos
-
+ Add a folder containing plain text files, PDFs, or Markdown. Configure additional extensions in Settings.Agregue una carpeta que contenga archivos de texto plano, PDFs o Markdown. Configure extensiones adicionales en Configuración.
-
+ Please choose a directoryPor favor, elija un directorio
-
+ NameNombre
-
+ Collection name...Nombre de la colección...
-
+ Name of the collection to add (Required)Nombre de la colección a agregar (Requerido)
-
+ FolderCarpeta
-
+ Folder path...Ruta de la carpeta...
-
+ Folder path to documents (Required)Ruta de la carpeta de documentos (Requerido)
-
+ BrowseExplorar
-
+ Create CollectionCrear colección
@@ -67,288 +67,288 @@
AddModelView
-
+ ← Existing Models← Modelos existentes
-
+ Explore ModelsExplorar modelos
-
+ Discover and download models by keyword search...Descubre y descarga modelos mediante búsqueda por palabras clave...
-
+ Text field for discovering and filtering downloadable modelsCampo de texto para descubrir y filtrar modelos descargables
-
+ Initiate model discovery and filteringIniciar descubrimiento y filtrado de modelos
-
+ Triggers discovery and filtering of modelsActiva el descubrimiento y filtrado de modelos
-
+ DefaultPredeterminado
-
+ LikesMe gusta
-
+ DownloadsDescargas
-
+ RecentReciente
-
+ AscAsc
-
+ DescDesc
-
+ NoneNinguno
-
+ Searching · %1Buscando · %1
-
+ Sort by: %1Ordenar por: %1
-
+ Sort dir: %1Dirección de ordenamiento: %1
-
+ Limit: %1Límite: %1
-
+ Network error: could not retrieve %1Error de red: no se pudo recuperar %1
-
-
+
+ Busy indicatorIndicador de ocupado
-
+ Displayed when the models request is ongoingSe muestra cuando la solicitud de modelos está en curso
-
+ Model fileArchivo del modelo
-
+ Model file to be downloadedArchivo del modelo a descargar
-
+ DescriptionDescripción
-
+ File descriptionDescripción del archivo
-
+ CancelCancelar
-
+ ResumeReanudar
-
+ DownloadDescargar
-
+ Stop/restart/start the downloadDetener/reiniciar/iniciar la descarga
-
+ RemoveEliminar
-
+ Remove model from filesystemEliminar modelo del sistema de archivos
-
-
+
+ InstallInstalar
-
+ Install online modelInstalar modelo en línea
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font><strong><font size="1"><a href="#error">Error</a></strong></font>
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font><strong><font size="2">ADVERTENCIA: No recomendado para tu hardware. El modelo requiere más memoria (%1 GB) de la que tu sistema tiene disponible (%2).</strong></font>
-
+ %1 GB%1 GB
-
-
+
+ ??
-
+ Describes an error that occurred when downloadingDescribe un error que ocurrió durante la descarga
-
+ Error for incompatible hardwareError por hardware incompatible
-
+ Download progressBarBarra de progreso de descarga
-
+ Shows the progress made in the downloadMuestra el progreso realizado en la descarga
-
+ Download speedVelocidad de descarga
-
+ Download speed in bytes/kilobytes/megabytes per secondVelocidad de descarga en bytes/kilobytes/megabytes por segundo
-
+ Calculating...Calculando...
-
-
-
-
+
+
+
+ Whether the file hash is being calculatedSi se está calculando el hash del archivo
-
+ Displayed when the file hash is being calculatedSe muestra cuando se está calculando el hash del archivo
-
+ enter $API_KEYingrese $API_KEY
-
+ File sizeTamaño del archivo
-
+ RAM requiredRAM requerida
-
+ ParametersParámetros
-
+ QuantCuantificación
-
+ TypeTipo
-
+ ERROR: $API_KEY is empty.ERROR: $API_KEY está vacío.
-
+ ERROR: $BASE_URL is empty.ERROR: $BASE_URL está vacío.
-
+ enter $BASE_URLingrese $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.ERROR: $MODEL_NAME está vacío.
-
+ enter $MODEL_NAMEingrese $MODEL_NAME
@@ -356,22 +356,22 @@
ApplicationSettings
-
+ ApplicationAplicación
-
+ Network dialogDiálogo de red
-
+ opt-in to share feedback/conversationsoptar por compartir comentarios/conversaciones
-
+ ERROR: Update system could not find the MaintenanceTool used<br>
to check for updates!<br><br>
Did you install this application using the online installer? If so,<br>
@@ -382,57 +382,57 @@
-
+ Error dialogDiálogo de error
-
+ Application SettingsConfiguración de la aplicación
-
+ GeneralGeneral
-
+ ThemeTema
-
+ The application color scheme.El esquema de colores de la aplicación.
-
+ DarkOscuro
-
+ LightClaro
-
+ LegacyDarkOscuro legado
-
+ Font SizeTamaño de fuente
-
+ The size of text in the application.El tamaño del texto en la aplicación.
-
+ DeviceDispositivo
@@ -441,127 +441,127 @@
El dispositivo de cómputo utilizado para la generación de texto. "Auto" utiliza Vulkan o Metal.
-
+ SmallPequeño
-
+ MediumMediano
-
+ LargeGrande
-
+ Language and LocaleIdioma y configuración regional
-
+ The language and locale you wish to use.El idioma y la configuración regional que deseas usar.
-
+ Default ModelModelo predeterminado
-
+ The preferred model for new chats. Also used as the local server fallback.El modelo preferido para nuevos chats. También se utiliza como respaldo del servidor local.
-
+ Suggestion ModeModo de sugerencia
-
+ Generate suggested follow-up questions at the end of responses.Generar preguntas de seguimiento sugeridas al final de las respuestas.
-
+ When chatting with LocalDocsAl chatear con LocalDocs
-
+ Whenever possibleSiempre que sea posible
-
+ NeverNunca
-
+ Download PathRuta de descarga
-
+ Where to store local models and the LocalDocs database.Dónde almacenar los modelos locales y la base de datos de LocalDocs.
-
+ BrowseExplorar
-
+ Choose where to save model filesElegir dónde guardar los archivos del modelo
-
+ Enable DatalakeHabilitar Datalake
-
+ Send chats and feedback to the GPT4All Open-Source Datalake.Enviar chats y comentarios al Datalake de código abierto de GPT4All.
-
+ AdvancedAvanzado
-
+ CPU ThreadsHilos de CPU
-
+ The number of CPU threads used for inference and embedding.El número de hilos de CPU utilizados para inferencia e incrustación.
-
+ Save Chat ContextGuardar contexto del chat
-
+ Save the chat model's state to disk for faster loading. WARNING: Uses ~2GB per chat.Guardar el estado del modelo de chat en el disco para una carga más rápida. ADVERTENCIA: Usa ~2GB por chat.
-
+ Expose an OpenAI-Compatible server to localhost. WARNING: Results in increased resource usage.Exponer un servidor compatible con OpenAI a localhost. ADVERTENCIA: Resulta en un mayor uso de recursos.
-
+ Enable Local ServerHabilitar servidor local
@@ -572,43 +572,43 @@
en un mayor uso de recursos.
-
+ API Server PortPuerto del servidor API
-
+ The port to use for the local server. Requires restart.El puerto a utilizar para el servidor local. Requiere reinicio.
-
+ Check For UpdatesBuscar actualizaciones
-
+ Manually check for an update to GPT4All.Buscar manualmente una actualización para GPT4All.
-
+ UpdatesActualizaciones
-
+ System LocaleRegional del sistema
-
+ The compute device used for text generation.El dispositivo de cómputo utilizado para la generación de texto.
-
-
+
+ Application defaultPredeterminado de la aplicación
@@ -632,13 +632,13 @@
Chat
-
-
+
+ New ChatNuevo chat
-
+ Server ChatChat del servidor
@@ -646,12 +646,12 @@
ChatAPIWorker
-
+ ERROR: Network error occurred while connecting to the API serverERROR: Ocurrió un error de red al conectar con el servidor API
-
+ ChatAPIWorker::handleFinished got HTTP Error %1 %2ChatAPIWorker::handleFinished obtuvo Error HTTP %1 %2
@@ -659,62 +659,62 @@
ChatDrawer
-
+ DrawerCajón
-
+ Main navigation drawerCajón de navegación principal
-
+ + New Chat+ Nuevo chat
-
+ Create a new chatCrear un nuevo chat
-
+ Select the current chat or edit the chat when in edit modeSeleccionar el chat actual o editar el chat cuando esté en modo de edición
-
+ Edit chat nameEditar nombre del chat
-
+ Save chat nameGuardar nombre del chat
-
+ Delete chatEliminar chat
-
+ Confirm chat deletionConfirmar eliminación del chat
-
+ Cancel chat deletionCancelar eliminación del chat
-
+ List of chatsLista de chats
-
+ List of chats in the drawer dialogLista de chats en el diálogo del cajón
@@ -722,32 +722,32 @@
ChatListModel
-
+ TODAYHOY
-
+ THIS WEEKESTA SEMANA
-
+ THIS MONTHESTE MES
-
+ LAST SIX MONTHSÚLTIMOS SEIS MESES
-
+ THIS YEARESTE AÑO
-
+ LAST YEARAÑO PASADO
@@ -755,148 +755,148 @@
ChatView
-
+ <h3>Warning</h3><p>%1</p><h3>Advertencia</h3><p>%1</p>
-
+ Switch model dialogDiálogo para cambiar de modelo
-
+ Warn the user if they switch models, then context will be erasedAdvertir al usuario si cambia de modelo, entonces se borrará el contexto
-
+ Conversation copied to clipboard.Conversación copiada al portapapeles.
-
+ Code copied to clipboard.Código copiado al portapapeles.
-
+ Chat panelPanel de chat
-
+ Chat panel with optionsPanel de chat con opciones
-
+ Reload the currently loaded modelRecargar el modelo actualmente cargado
-
+ Eject the currently loaded modelExpulsar el modelo actualmente cargado
-
+ No model installed.No hay modelo instalado.
-
+ Model loading error.Error al cargar el modelo.
-
+ Waiting for model...Esperando al modelo...
-
+ Switching context...Cambiando contexto...
-
+ Choose a model...Elige un modelo...
-
+ Not found: %1No encontrado: %1
-
+ The top item is the current modelEl elemento superior es el modelo actual
-
-
+
+ LocalDocsDocumentosLocales
-
+ Add documentsAgregar documentos
-
+ add collections of documents to the chatagregar colecciones de documentos al chat
-
+ Load the default modelCargar el modelo predeterminado
-
+ Loads the default model which can be changed in settingsCarga el modelo predeterminado que se puede cambiar en la configuración
-
+ No Model InstalledNo hay modelo instalado
-
+ <h3>Encountered an error loading model:</h3><br><i>"%1"</i><br><br>Model loading failures can happen for a variety of reasons, but the most common causes include a bad file format, an incomplete or corrupted download, the wrong file type, not enough system RAM or an incompatible model type. Here are some suggestions for resolving the problem:<br><ul><li>Ensure the model file has a compatible format and type<li>Check the model file is complete in the download folder<li>You can find the download folder in the settings dialog<li>If you've sideloaded the model ensure the file is not corrupt by checking md5sum<li>Read more about what models are supported in our <a href="https://docs.gpt4all.io/">documentation</a> for the gui<li>Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help<h3>Se encontró un error al cargar el modelo:</h3><br><i>"%1"</i><br><br>Los fallos en la carga de modelos pueden ocurrir por varias razones, pero las causas más comunes incluyen un formato de archivo incorrecto, una descarga incompleta o corrupta, un tipo de archivo equivocado, RAM del sistema insuficiente o un tipo de modelo incompatible. Aquí hay algunas sugerencias para resolver el problema:<br><ul><li>Asegúrate de que el archivo del modelo tenga un formato y tipo compatibles<li>Verifica que el archivo del modelo esté completo en la carpeta de descargas<li>Puedes encontrar la carpeta de descargas en el diálogo de configuración<li>Si has cargado el modelo manualmente, asegúrate de que el archivo no esté corrupto verificando el md5sum<li>Lee más sobre qué modelos son compatibles en nuestra <a href="https://docs.gpt4all.io/">documentación</a> para la interfaz gráfica<li>Visita nuestro <a href="https://discord.gg/4M2QFmTt2k">canal de discord</a> para obtener ayuda
-
+ Install a ModelInstalar un modelo
-
+ Shows the add model viewMuestra la vista de agregar modelo
-
+ Conversation with the modelConversación con el modelo
-
+ prompt / response pairs from the conversationpares de pregunta / respuesta de la conversación
-
+ GPT4AllGPT4All
-
+ YouTú
@@ -905,129 +905,129 @@
recalculando contexto ...
-
+ response stopped ...respuesta detenida ...
-
+ processing ...procesando ...
-
+ generating response ...generando respuesta ...
-
+ generating questions ...generando preguntas ...
-
-
+
+ CopyCopiar
-
+ Copy MessageCopiar mensaje
-
+ Disable markdownDesactivar markdown
-
+ Enable markdownActivar markdown
-
+ Thumbs upMe gusta
-
+ Gives a thumbs up to the responseDa un me gusta a la respuesta
-
+ Thumbs downNo me gusta
-
+ Opens thumbs down dialogAbre el diálogo de no me gusta
-
+ Suggested follow-upsSeguimientos sugeridos
-
+ Erase and reset chat sessionBorrar y reiniciar sesión de chat
-
+ Copy chat session to clipboardCopiar sesión de chat al portapapeles
-
+ Redo last chat responseRehacer última respuesta del chat
-
+ Stop generatingDetener generación
-
+ Stop the current response generationDetener la generación de la respuesta actual
-
+ Reloads the modelRecarga el modelo
-
-
+
+ Reload · %1Recargar · %1
-
+ Loading · %1Cargando · %1
-
+ Load · %1 (default) →Cargar · %1 (predeterminado) →
-
+ retrieving localdocs: %1 ...recuperando documentos locales: %1 ...
-
+ searching localdocs: %1 ...buscando en documentos locales: %1 ...
-
+ %n Source(s)%n Fuente
@@ -1035,47 +1035,47 @@
-
+ Send a message...Enviar un mensaje...
-
+ Load a model to continue...Carga un modelo para continuar...
-
+ Send messages/prompts to the modelEnviar mensajes/indicaciones al modelo
-
+ CutCortar
-
+ PastePegar
-
+ Select AllSeleccionar todo
-
+ Send messageEnviar mensaje
-
+ Sends the message/prompt contained in textfield to the modelEnvía el mensaje/indicación contenido en el campo de texto al modelo
-
+ GPT4All requires that you install at least one
model to get startedGPT4All requiere que instale al menos un
@@ -1083,7 +1083,7 @@ modelo para comenzar
-
+ restoring from text ...restaurando desde texto ...
@@ -1091,12 +1091,12 @@ modelo para comenzar
CollectionsDrawer
-
+ Warning: searching collections while indexing can return incomplete resultsAdvertencia: buscar en colecciones mientras se indexan puede devolver resultados incompletos
-
+ %n file(s)%n archivo
@@ -1104,7 +1104,7 @@ modelo para comenzar
-
+ %n word(s)%n palabra
@@ -1112,17 +1112,17 @@ modelo para comenzar
-
+ UpdatingActualizando
-
+ + Add Docs+ Agregar documentos
-
+ Select a collection to make it available to the chat model.Seleccione una colección para hacerla disponible al modelo de chat.
@@ -1130,37 +1130,37 @@ modelo para comenzar
Download
-
+ Model "%1" is installed successfully.El modelo "%1" se ha instalado correctamente.
-
+ ERROR: $MODEL_NAME is empty.ERROR: $MODEL_NAME está vacío.
-
+ ERROR: $API_KEY is empty.ERROR: $API_KEY está vacía.
-
+ ERROR: $BASE_URL is invalid.ERROR: $BASE_URL no es válida.
-
+ ERROR: Model "%1 (%2)" is conflict.ERROR: El modelo "%1 (%2)" está en conflicto.
-
+ Model "%1 (%2)" is installed successfully.El modelo "%1 (%2)" se ha instalado correctamente.
-
+ Model "%1" is removed.El modelo "%1" ha sido eliminado.
@@ -1168,92 +1168,92 @@ modelo para comenzar
HomeView
-
+ Welcome to GPT4AllBienvenido a GPT4All
-
+ The privacy-first LLM chat applicationLa aplicación de chat LLM que prioriza la privacidad
-
+ Start chattingComenzar a chatear
-
+ Start ChattingIniciar chat
-
+ Chat with any LLMChatear con cualquier LLM
-
+ LocalDocsDocumentosLocales
-
+ Chat with your local filesChatear con tus archivos locales
-
+ Find ModelsBuscar modelos
-
+ Explore and download modelsExplorar y descargar modelos
-
+ Latest newsÚltimas noticias
-
+ Latest news from GPT4AllÚltimas noticias de GPT4All
-
+ Release NotesNotas de la versión
-
+ DocumentationDocumentación
-
+ DiscordDiscord
-
+ X (Twitter)X (Twitter)
-
+ GithubGithub
-
+ nomic.ainomic.ai
-
+ Subscribe to NewsletterSuscribirse al boletín
@@ -1261,22 +1261,22 @@ modelo para comenzar
LocalDocsSettings
-
+ LocalDocsDocumentosLocales
-
+ LocalDocs SettingsConfiguración de DocumentosLocales
-
+ IndexingIndexación
-
+ Allowed File ExtensionsExtensiones de archivo permitidas
@@ -1287,12 +1287,12 @@ modelo para comenzar
archivos con estas extensiones.
-
+ EmbeddingIncrustación
-
+ Use Nomic Embed APIUsar API de incrustación Nomic
@@ -1303,77 +1303,77 @@ modelo para comenzar
local privado. Requiere reinicio.
-
+ Nomic API KeyClave API de Nomic
-
+ API key to use for Nomic Embed. Get one from the Atlas <a href="https://atlas.nomic.ai/cli-login">API keys page</a>. Requires restart.Clave API para usar con Nomic Embed. Obtén una en la <a href="https://atlas.nomic.ai/cli-login">página de claves API</a> de Atlas. Requiere reinicio.
-
+ Embeddings DeviceDispositivo de incrustaciones
-
+ The compute device used for embeddings. Requires restart.El dispositivo de cómputo utilizado para las incrustaciones. Requiere reinicio.
-
+ Comma-separated list. LocalDocs will only attempt to process files with these extensions.Lista separada por comas. LocalDocs solo intentará procesar archivos con estas extensiones.
-
+ Embed documents using the fast Nomic API instead of a private local model. Requires restart.Incrustar documentos usando la API rápida de Nomic en lugar de un modelo local privado. Requiere reinicio.
-
+ Application defaultPredeterminado de la aplicación
-
+ DisplayVisualización
-
+ Show SourcesMostrar fuentes
-
+ Display the sources used for each response.Mostrar las fuentes utilizadas para cada respuesta.
-
+ AdvancedAvanzado
-
+ Warning: Advanced usage only.Advertencia: Solo para uso avanzado.
-
+ Values too large may cause localdocs failure, extremely slow responses or failure to respond at all. Roughly speaking, the {N chars x N snippets} are added to the model's context window. More info <a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">here</a>.Valores demasiado grandes pueden causar fallos en localdocs, respuestas extremadamente lentas o falta de respuesta. En términos generales, los {N caracteres x N fragmentos} se añaden a la ventana de contexto del modelo. Más información <a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">aquí</a>.
-
+ Number of characters per document snippet. Larger numbers increase likelihood of factual responses, but also result in slower generation.Número de caracteres por fragmento de documento. Números más grandes aumentan la probabilidad de respuestas verídicas, pero también resultan en una generación más lenta.
-
+ Max best N matches of retrieved document snippets to add to the context for prompt. Larger numbers increase likelihood of factual responses, but also result in slower generation.Máximo de N mejores coincidencias de fragmentos de documentos recuperados para añadir al contexto del prompt. Números más grandes aumentan la probabilidad de respuestas verídicas, pero también resultan en una generación más lenta.
@@ -1383,12 +1383,12 @@ modelo para comenzar
Valores demasiado grandes pueden causar fallos en documentos locales, respuestas extremadamente lentas o falta de respuesta. En términos generales, los {N caracteres x N fragmentos} se agregan a la ventana de contexto del modelo. Más información <a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">aquí</a>.
-
+ Document snippet size (characters)Tamaño del fragmento de documento (caracteres)
-
+ Max document snippets per promptMáximo de fragmentos de documento por indicación
@@ -1396,17 +1396,17 @@ modelo para comenzar
LocalDocsView
-
+ LocalDocsDocumentosLocales
-
+ Chat with your local filesChatea con tus archivos locales
-
+ + Add Collection+ Agregar colección
@@ -1415,97 +1415,97 @@ modelo para comenzar
ERROR: La base de datos de DocumentosLocales no es válida.
-
+ No Collections InstalledNo hay colecciones instaladas
-
+ Install a collection of local documents to get started using this featureInstala una colección de documentos locales para comenzar a usar esta función
-
+ + Add Doc Collection+ Agregar colección de documentos
-
+ Shows the add model viewMuestra la vista de agregar modelo
-
+ Indexing progressBarBarra de progreso de indexación
-
+ Shows the progress made in the indexingMuestra el progreso realizado en la indexación
-
+ ERRORERROR
-
+ INDEXINGINDEXANDO
-
+ EMBEDDINGINCRUSTANDO
-
+ REQUIRES UPDATEREQUIERE ACTUALIZACIÓN
-
+ READYLISTO
-
+ INSTALLINGINSTALANDO
-
+ Indexing in progressIndexación en progreso
-
+ Embedding in progressIncrustación en progreso
-
+ This collection requires an update after version changeEsta colección requiere una actualización después del cambio de versión
-
+ Automatically reindexes upon changes to the folderReindexación automática al cambiar la carpeta
-
+ Installation in progressInstalación en progreso
-
+ %%
-
+ %n file(s)%n archivo
@@ -1513,7 +1513,7 @@ modelo para comenzar
-
+ %n word(s)%n palabra
@@ -1521,32 +1521,32 @@ modelo para comenzar
-
+ RemoveEliminar
-
+ RebuildReconstruir
-
+ Reindex this folder from scratch. This is slow and usually not needed.Reindexar esta carpeta desde cero. Esto es lento y generalmente no es necesario.
-
+ UpdateActualizar
-
+ Update the collection to the new version. This is a slow operation.Actualizar la colección a la nueva versión. Esta es una operación lenta.
-
+ <h3>ERROR: The LocalDocs database cannot be accessed or is not valid.</h3><br><i>Note: You will need to restart after trying any of the following suggested fixes.</i><br><ul><li>Make sure that the folder set as <b>Download Path</b> exists on the file system.</li><li>Check ownership as well as read and write permissions of the <b>Download Path</b>.</li><li>If there is a <b>localdocs_v2.db</b> file, check its ownership and read/write permissions, too.</li></ul><br>If the problem persists and there are any 'localdocs_v*.db' files present, as a last resort you can<br>try backing them up and removing them. You will have to recreate your collections, however.<h3>ERROR: No se puede acceder a la base de datos LocalDocs o no es válida.</h3><br><i>Nota: Necesitará reiniciar después de intentar cualquiera de las siguientes soluciones sugeridas.</i><br><ul><li>Asegúrese de que la carpeta establecida como <b>Ruta de Descarga</b> exista en el sistema de archivos.</li><li>Verifique la propiedad y los permisos de lectura y escritura de la <b>Ruta de Descarga</b>.</li><li>Si hay un archivo <b>localdocs_v2.db</b>, verifique también su propiedad y permisos de lectura/escritura.</li></ul><br>Si el problema persiste y hay archivos 'localdocs_v*.db' presentes, como último recurso puede<br>intentar hacer una copia de seguridad y eliminarlos. Sin embargo, tendrá que recrear sus colecciones.
@@ -1554,7 +1554,7 @@ modelo para comenzar
ModelList
-
+ <ul><li>Requires personal OpenAI API key.</li><li>WARNING: Will send your chats to OpenAI!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with OpenAI</li><li>You can apply for an API key <a href="https://platform.openai.com/account/api-keys">here.</a></li><ul><li>Requiere clave API personal de OpenAI.</li><li>ADVERTENCIA: ¡Enviará sus chats a OpenAI!</li><li>Su clave API se almacenará en el disco</li><li>Solo se usará para comunicarse con OpenAI</li><li>Puede solicitar una clave API <a href="https://platform.openai.com/account/api-keys">aquí.</a></li>
@@ -1565,62 +1565,73 @@ modelo para comenzar
OpenAI</strong><br> %1
-
+ <strong>OpenAI's ChatGPT model GPT-3.5 Turbo</strong><br> %1<strong>Modelo ChatGPT GPT-3.5 Turbo de OpenAI</strong><br> %1
-
+ <br><br><i>* Even if you pay OpenAI for ChatGPT-4 this does not guarantee API key access. Contact OpenAI for more info.<br><br><i>* Aunque pagues a OpenAI por ChatGPT-4, esto no garantiza el acceso a la clave API. Contacta a OpenAI para más información.
-
+ <strong>OpenAI's ChatGPT model GPT-4</strong><br> %1 %2<strong>Modelo ChatGPT GPT-4 de OpenAI</strong><br> %1 %2
-
+ <ul><li>Requires personal Mistral API key.</li><li>WARNING: Will send your chats to Mistral!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with Mistral</li><li>You can apply for an API key <a href="https://console.mistral.ai/user/api-keys">here</a>.</li><ul><li>Requiere una clave API personal de Mistral.</li><li>ADVERTENCIA: ¡Enviará tus chats a Mistral!</li><li>Tu clave API se almacenará en el disco</li><li>Solo se usará para comunicarse con Mistral</li><li>Puedes solicitar una clave API <a href="https://console.mistral.ai/user/api-keys">aquí</a>.</li>
-
+ <strong>Mistral Tiny model</strong><br> %1<strong>Modelo Mistral Tiny</strong><br> %1
-
+ <strong>Mistral Small model</strong><br> %1<strong>Modelo Mistral Small</strong><br> %1
-
+ <strong>Mistral Medium model</strong><br> %1<strong>Modelo Mistral Medium</strong><br> %1
-
+ <strong>Created by %1.</strong><br><ul><li>Published on %2.<li>This model has %3 likes.<li>This model has %4 downloads.<li>More info can be found <a href="https://huggingface.co/%5">here.</a></ul><strong>Creado por %1.</strong><br><ul><li>Publicado el %2.<li>Este modelo tiene %3 me gusta.<li>Este modelo tiene %4 descargas.<li>Más información puede encontrarse <a href="https://huggingface.co/%5">aquí.</a></ul>
-
+ %1 (%2)%1 (%2)
-
+
+
+ cannot open "%1": %2
+
+
+
+
+ cannot create "%1": %2
+
+
+
+ <strong>OpenAI-Compatible API Model</strong><br><ul><li>API Key: %1</li><li>Base URL: %2</li><li>Model Name: %3</li></ul><strong>Modelo de API compatible con OpenAI</strong><br><ul><li>Clave API: %1</li><li>URL base: %2</li><li>Nombre del modelo: %3</li></ul>
-
+ <ul><li>Requires personal API key and the API base URL.</li><li>WARNING: Will send your chats to the OpenAI-compatible API Server you specified!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with the OpenAI-compatible API Server</li><ul><li>Requiere una clave API personal y la URL base de la API.</li><li>ADVERTENCIA: ¡Enviará sus chats al servidor de API compatible con OpenAI que especificó!</li><li>Su clave API se almacenará en el disco</li><li>Solo se utilizará para comunicarse con el servidor de API compatible con OpenAI</li>
-
+ <strong>Connect to OpenAI-compatible API server</strong><br> %1<strong>Conectar al servidor de API compatible con OpenAI</strong><br> %1
@@ -1628,87 +1639,87 @@ modelo para comenzar
ModelSettings
-
+ ModelModelo
-
+ Model SettingsConfiguración del modelo
-
+ CloneClonar
-
+ RemoveEliminar
-
+ NameNombre
-
+ Model FileArchivo del modelo
-
+ System PromptIndicación del sistema
-
+ Prefixed at the beginning of every conversation. Must contain the appropriate framing tokens.Prefijado al inicio de cada conversación. Debe contener los tokens de encuadre apropiados.
-
+ Prompt TemplatePlantilla de indicación
-
+ The template that wraps every prompt.La plantilla que envuelve cada indicación.
-
+ Must contain the string "%1" to be replaced with the user's input.Debe contener la cadena "%1" para ser reemplazada con la entrada del usuario.
-
+ Chat Name PromptIndicación para el nombre del chat
-
+ Prompt used to automatically generate chat names.Indicación utilizada para generar automáticamente nombres de chat.
-
+ Suggested FollowUp PromptIndicación de seguimiento sugerida
-
+ Prompt used to generate suggested follow-up questions.Indicación utilizada para generar preguntas de seguimiento sugeridas.
-
+ Context LengthLongitud del contexto
-
+ Number of input and output tokens the model sees.Número de tokens de entrada y salida que el modelo ve.
@@ -1719,128 +1730,128 @@ NOTE: Does not take effect until you reload the model.
NOTA: No tiene efecto hasta que recargues el modelo.
-
+ TemperatureTemperatura
-
+ Randomness of model output. Higher -> more variation.Aleatoriedad de la salida del modelo. Mayor -> más variación.
-
+ Temperature increases the chances of choosing less likely tokens.
NOTE: Higher temperature gives more creative but less predictable outputs.La temperatura aumenta las probabilidades de elegir tokens menos probables.
NOTA: Una temperatura más alta da resultados más creativos pero menos predecibles.
-
+ Top-PTop-P
-
+ Nucleus Sampling factor. Lower -> more predictable.Factor de muestreo de núcleo. Menor -> más predecible.
-
+ Only the most likely tokens up to a total probability of top_p can be chosen.
NOTE: Prevents choosing highly unlikely tokens.Solo se pueden elegir los tokens más probables hasta una probabilidad total de top_p.
NOTA: Evita elegir tokens altamente improbables.
-
+ Min-PMin-P
-
+ Minimum token probability. Higher -> more predictable.Probabilidad mínima del token. Mayor -> más predecible.
-
+ Sets the minimum relative probability for a token to be considered.Establece la probabilidad relativa mínima para que un token sea considerado.
-
+ Top-KTop-K
-
+ Size of selection pool for tokens.Tamaño del grupo de selección para tokens.
-
+ Only the top K most likely tokens will be chosen from.Solo se elegirán los K tokens más probables.
-
+ Max LengthLongitud máxima
-
+ Maximum response length, in tokens.Longitud máxima de respuesta, en tokens.
-
+ Prompt Batch SizeTamaño del lote de indicaciones
-
+ The batch size used for prompt processing.El tamaño del lote utilizado para el procesamiento de indicaciones.
-
+ Amount of prompt tokens to process at once.
NOTE: Higher values can speed up reading prompts but will use more RAM.Cantidad de tokens de prompt a procesar de una vez.
NOTA: Valores más altos pueden acelerar la lectura de prompts, pero usarán más RAM.
-
+ Repeat PenaltyPenalización por repetición
-
+ Repetition penalty factor. Set to 1 to disable.Factor de penalización por repetición. Establecer a 1 para desactivar.
-
+ Repeat Penalty TokensTokens de penalización por repetición
-
+ Number of previous tokens used for penalty.Número de tokens anteriores utilizados para la penalización.
-
+ GPU LayersCapas de GPU
-
+ Number of model layers to load into VRAM.Número de capas del modelo a cargar en la VRAM.
-
+ Maximum combined prompt/response tokens before information is lost.
Using more context than the model was trained on will yield poor results.
NOTE: Does not take effect until you reload the model.
@@ -1849,7 +1860,7 @@ Usar más contexto del que el modelo fue entrenado producirá resultados deficie
NOTA: No surtirá efecto hasta que recargue el modelo.
-
+ How many model layers to load into VRAM. Decrease this if GPT4All runs out of VRAM while loading this model.
Lower values increase CPU load and RAM usage, and make inference slower.
NOTE: Does not take effect until you reload the model.
@@ -1861,217 +1872,217 @@ NOTA: No surte efecto hasta que recargue el modelo.
ModelsView
-
+ No Models InstalledNo hay modelos instalados
-
+ Install a model to get started using GPT4AllInstala un modelo para empezar a usar GPT4All
-
-
+
+ + Add Model+ Agregar modelo
-
+ Shows the add model viewMuestra la vista de agregar modelo
-
+ Installed ModelsModelos instalados
-
+ Locally installed chat modelsModelos de chat instalados localmente
-
+ Model fileArchivo del modelo
-
+ Model file to be downloadedArchivo del modelo a descargar
-
+ DescriptionDescripción
-
+ File descriptionDescripción del archivo
-
+ CancelCancelar
-
+ ResumeReanudar
-
+ Stop/restart/start the downloadDetener/reiniciar/iniciar la descarga
-
+ RemoveEliminar
-
+ Remove model from filesystemEliminar modelo del sistema de archivos
-
-
+
+ InstallInstalar
-
+ Install online modelInstalar modelo en línea
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font><strong><font size="1"><a href="#error">Error</a></strong></font>
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font><strong><font size="2">ADVERTENCIA: No recomendado para su hardware. El modelo requiere más memoria (%1 GB) de la que su sistema tiene disponible (%2).</strong></font>
-
+ %1 GB%1 GB
-
+ ??
-
+ Describes an error that occurred when downloadingDescribe un error que ocurrió durante la descarga
-
+ Error for incompatible hardwareError por hardware incompatible
-
+ Download progressBarBarra de progreso de descarga
-
+ Shows the progress made in the downloadMuestra el progreso realizado en la descarga
-
+ Download speedVelocidad de descarga
-
+ Download speed in bytes/kilobytes/megabytes per secondVelocidad de descarga en bytes/kilobytes/megabytes por segundo
-
+ Calculating...Calculando...
-
-
-
-
+
+
+
+ Whether the file hash is being calculatedSi se está calculando el hash del archivo
-
+ Busy indicatorIndicador de ocupado
-
+ Displayed when the file hash is being calculatedSe muestra cuando se está calculando el hash del archivo
-
+ enter $API_KEYingrese $API_KEY
-
+ File sizeTamaño del archivo
-
+ RAM requiredRAM requerida
-
+ ParametersParámetros
-
+ QuantCuantificación
-
+ TypeTipo
-
+ ERROR: $API_KEY is empty.ERROR: $API_KEY está vacía.
-
+ ERROR: $BASE_URL is empty.ERROR: $BASE_URL está vacía.
-
+ enter $BASE_URLingrese $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.ERROR: $MODEL_NAME está vacío.
-
+ enter $MODEL_NAMEingrese $MODEL_NAME
@@ -2079,12 +2090,12 @@ NOTA: No surte efecto hasta que recargue el modelo.
MyFancyLink
-
+ Fancy linkEnlace elegante
-
+ A stylized linkUn enlace estilizado
@@ -2092,7 +2103,7 @@ NOTA: No surte efecto hasta que recargue el modelo.
MySettingsStack
-
+ Please choose a directoryPor favor, elija un directorio
@@ -2100,12 +2111,12 @@ NOTA: No surte efecto hasta que recargue el modelo.
MySettingsTab
-
+ Restore DefaultsRestaurar valores predeterminados
-
+ Restores settings dialog to a default stateRestaura el diálogo de configuración a su estado predeterminado
@@ -2113,7 +2124,7 @@ NOTA: No surte efecto hasta que recargue el modelo.
NetworkDialog
-
+ Contribute data to the GPT4All Opensource Datalake.Contribuir datos al Datalake de código abierto de GPT4All.
@@ -2131,7 +2142,7 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
NOTA: Al activar esta función, enviará sus datos al Datalake de código abierto de GPT4All. No debe esperar privacidad en el chat cuando esta función esté habilitada. Sin embargo, puede esperar una atribución opcional si lo desea. Sus datos de chat estarán disponibles abiertamente para que cualquiera los descargue y serán utilizados por Nomic AI para mejorar futuros modelos de GPT4All. Nomic AI conservará toda la información de atribución adjunta a sus datos y se le acreditará como contribuyente en cualquier lanzamiento de modelo GPT4All que utilice sus datos.
-
+ By enabling this feature, you will be able to participate in the democratic process of training a large language model by contributing data for future model improvements.
When a GPT4All model responds to you and you have opted-in, your conversation will be sent to the GPT4All Open Source Datalake. Additionally, you can like/dislike its response. If you dislike a response, you can suggest an alternative response. This data will be collected and aggregated in the GPT4All Datalake.
@@ -2144,47 +2155,47 @@ Cuando un modelo GPT4All te responda y hayas aceptado participar, tu conversaci
NOTA: Al activar esta función, estarás enviando tus datos al Datalake de Código Abierto de GPT4All. No debes esperar privacidad en el chat cuando esta función esté habilitada. Sin embargo, puedes esperar una atribución opcional si lo deseas. Tus datos de chat estarán disponibles abiertamente para que cualquiera los descargue y serán utilizados por Nomic AI para mejorar futuros modelos de GPT4All. Nomic AI conservará toda la información de atribución adjunta a tus datos y se te acreditará como contribuyente en cualquier lanzamiento de modelo GPT4All que utilice tus datos.
-
+ Terms for opt-inTérminos para optar por participar
-
+ Describes what will happen when you opt-inDescribe lo que sucederá cuando opte por participar
-
+ Please provide a name for attribution (optional)Por favor, proporcione un nombre para la atribución (opcional)
-
+ Attribution (optional)Atribución (opcional)
-
+ Provide attributionProporcionar atribución
-
+ EnableHabilitar
-
+ Enable opt-inHabilitar participación
-
+ CancelCancelar
-
+ Cancel opt-inCancelar participación
@@ -2192,17 +2203,17 @@ NOTA: Al activar esta función, estarás enviando tus datos al Datalake de Códi
NewVersionDialog
-
+ New version is availableNueva versión disponible
-
+ UpdateActualizar
-
+ Update to new versionActualizar a nueva versión
@@ -2210,17 +2221,17 @@ NOTA: Al activar esta función, estarás enviando tus datos al Datalake de Códi
PopupDialog
-
+ Reveals a shortlived help balloonMuestra un globo de ayuda de corta duración
-
+ Busy indicatorIndicador de ocupado
-
+ Displayed when the popup is showing busySe muestra cuando la ventana emergente está ocupada
@@ -2235,28 +2246,28 @@ NOTA: Al activar esta función, estarás enviando tus datos al Datalake de Códi
SettingsView
-
-
+
+ SettingsConfiguración
-
+ Contains various application settingsContiene varias configuraciones de la aplicación
-
+ ApplicationAplicación
-
+ ModelModelo
-
+ LocalDocsDocumentosLocales
@@ -2264,17 +2275,17 @@ NOTA: Al activar esta función, estarás enviando tus datos al Datalake de Códi
StartupDialog
-
+ Welcome!¡Bienvenido!
-
+ Release notesNotas de la versión
-
+ Release notes for this versionNotas de la versión para esta versión
@@ -2300,76 +2311,76 @@ NOTA: Al activar esta función, estarás enviando tus datos al Datalake de Códi
-
+ Terms for opt-inTérminos para aceptar
-
+ Describes what will happen when you opt-inDescribe lo que sucederá cuando acepte
-
-
+
+ Opt-in for anonymous usage statisticsAceptar estadísticas de uso anónimas
-
-
+
+ YesSí
-
+ Allow opt-in for anonymous usage statisticsPermitir aceptación de estadísticas de uso anónimas
-
-
+
+ NoNo
-
+ Opt-out for anonymous usage statisticsRechazar estadísticas de uso anónimas
-
+ Allow opt-out for anonymous usage statisticsPermitir rechazo de estadísticas de uso anónimas
-
-
+
+ Opt-in for networkAceptar para la red
-
+ Allow opt-in for networkPermitir aceptación para la red
-
+ Allow opt-in anonymous sharing of chats to the GPT4All DatalakePermitir compartir anónimamente los chats con el Datalake de GPT4All
-
+ Opt-out for networkRechazar para la red
-
+ Allow opt-out anonymous sharing of chats to the GPT4All DatalakePermitir rechazar el compartir anónimo de chats con el Datalake de GPT4All
-
+ ### Release notes
%1### Contributors
%2
@@ -2379,7 +2390,7 @@ NOTA: Al activar esta función, estarás enviando tus datos al Datalake de Códi
-
+ ### Opt-ins for anonymous usage analytics and datalake
By enabling these features, you will be able to participate in the democratic process of training a
large language model by contributing data for future model improvements.
@@ -2419,23 +2430,23 @@ lanzamiento de modelo GPT4All que utilice sus datos.
actual. ¿Desea continuar?
-
+ <b>Warning:</b> changing the model will erase the current conversation. Do you wish to continue?<b>Advertencia:</b> cambiar el modelo borrará la conversación actual. ¿Deseas continuar?
-
+ ContinueContinuar
-
+ Continue with model loadingContinuar con la carga del modelo
-
-
+
+ CancelCancelar
@@ -2443,33 +2454,33 @@ lanzamiento de modelo GPT4All que utilice sus datos.
ThumbsDownDialog
-
+ Please edit the text below to provide a better response. (optional)Por favor, edite el texto a continuación para proporcionar una mejor
respuesta. (opcional)
-
+ Please provide a better response...Por favor, proporcione una mejor respuesta...
-
+ SubmitEnviar
-
+ Submits the user's responseEnvía la respuesta del usuario
-
+ CancelCancelar
-
+ Closes the response dialogCierra el diálogo de respuesta
@@ -2477,126 +2488,126 @@ lanzamiento de modelo GPT4All que utilice sus datos.
main
-
+ GPT4All v%1GPT4All v%1
-
+ <h3>Encountered an error starting up:</h3><br><i>"Incompatible hardware detected."</i><br><br>Unfortunately, your CPU does not meet the minimal requirements to run this program. In particular, it does not support AVX intrinsics which this program requires to successfully run a modern large language model. The only solution at this time is to upgrade your hardware to a more modern CPU.<br><br>See here for more information: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a><h3>Se encontró un error al iniciar:</h3><br><i>"Se detectó hardware incompatible."</i><br><br>Desafortunadamente, tu CPU no cumple con los requisitos mínimos para ejecutar este programa. En particular, no soporta instrucciones AVX, las cuales este programa requiere para ejecutar con éxito un modelo de lenguaje grande moderno. La única solución en este momento es actualizar tu hardware a una CPU más moderna.<br><br>Consulta aquí para más información: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a>
-
+ <h3>Encountered an error starting up:</h3><br><i>"Inability to access settings file."</i><br><br>Unfortunately, something is preventing the program from accessing the settings file. This could be caused by incorrect permissions in the local app config directory where the settings file is located. Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help.<h3>Se encontró un error al iniciar:</h3><br><i>"No se puede acceder al archivo de configuración."</i><br><br>Desafortunadamente, algo está impidiendo que el programa acceda al archivo de configuración. Esto podría ser causado por permisos incorrectos en el directorio de configuración local de la aplicación donde se encuentra el archivo de configuración. Visita nuestro <a href="https://discord.gg/4M2QFmTt2k">canal de Discord</a> para obtener ayuda.
-
+ Connection to datalake failed.La conexión al datalake falló.
-
+ Saving chats.Guardando chats.
-
+ Network dialogDiálogo de red
-
+ opt-in to share feedback/conversationsoptar por compartir comentarios/conversaciones
-
+ Home viewVista de inicio
-
+ Home view of applicationVista de inicio de la aplicación
-
+ HomeInicio
-
+ Chat viewVista de chat
-
+ Chat view to interact with modelsVista de chat para interactuar con modelos
-
+ ChatsChats
-
-
+
+ ModelsModelos
-
+ Models view for installed modelsVista de modelos para modelos instalados
-
-
+
+ LocalDocsDocs
Locales
-
+ LocalDocs view to configure and use local docsVista de DocumentosLocales para configurar y usar documentos locales
-
-
+
+ SettingsConfig.
-
+ Settings view for application configurationVista de configuración para la configuración de la aplicación
-
+ The datalake is enabledEl datalake está habilitado
-
+ Using a network modelUsando un modelo de red
-
+ Server mode is enabledEl modo servidor está habilitado
-
+ Installed modelsModelos instalados
-
+ View of installed modelsVista de modelos instalados
diff --git a/gpt4all-chat/translations/gpt4all_it_IT.ts b/gpt4all-chat/translations/gpt4all_it_IT.ts
index 85360148..7a6ec7e1 100644
--- a/gpt4all-chat/translations/gpt4all_it_IT.ts
+++ b/gpt4all-chat/translations/gpt4all_it_IT.ts
@@ -4,62 +4,62 @@
AddCollectionView
-
+ ← Existing Collections← Raccolte esistenti
-
+ Add Document CollectionAggiungi raccolta documenti
-
+ Add a folder containing plain text files, PDFs, or Markdown. Configure additional extensions in Settings.Aggiungi una cartella contenente file di testo semplice, PDF o Markdown. Configura estensioni aggiuntive in Settaggi.
-
+ Please choose a directoryScegli una cartella
-
+ NameNome
-
+ Collection name...Nome della raccolta...
-
+ Name of the collection to add (Required)Nome della raccolta da aggiungere (Obbligatorio)
-
+ FolderCartella
-
+ Folder path...Percorso cartella...
-
+ Folder path to documents (Required)Percorso della cartella dei documenti (richiesto)
-
+ BrowseEsplora
-
+ Create CollectionCrea raccolta
@@ -67,288 +67,288 @@
AddModelView
-
+ ← Existing Models← Modelli esistenti
-
+ Explore ModelsEsplora modelli
-
+ Discover and download models by keyword search...Scopri e scarica i modelli tramite ricerca per parole chiave...
-
+ Text field for discovering and filtering downloadable modelsCampo di testo per scoprire e filtrare i modelli scaricabili
-
+ Initiate model discovery and filteringAvvia rilevamento e filtraggio dei modelli
-
+ Triggers discovery and filtering of modelsAttiva la scoperta e il filtraggio dei modelli
-
+ DefaultPredefinito
-
+ LikesMi piace
-
+ DownloadsScaricamenti
-
+ RecentRecenti
-
+ AscAsc
-
+ DescDisc
-
+ NoneNiente
-
+ Searching · %1Ricerca · %1
-
+ Sort by: %1Ordina per: %1
-
+ Sort dir: %1Direzione ordinamento: %1
-
+ Limit: %1Limite: %1
-
+ Network error: could not retrieve %1Errore di rete: impossibile recuperare %1
-
-
+
+ Busy indicatorIndicatore di occupato
-
+ Displayed when the models request is ongoingVisualizzato quando la richiesta dei modelli è in corso
-
+ Model fileFile del modello
-
+ Model file to be downloadedFile del modello da scaricare
-
+ DescriptionDescrizione
-
+ File descriptionDescrizione del file
-
+ CancelAnnulla
-
+ ResumeRiprendi
-
+ DownloadScarica
-
+ Stop/restart/start the downloadArresta/riavvia/avvia il download
-
+ RemoveRimuovi
-
+ Remove model from filesystemRimuovi il modello dal sistema dei file
-
-
+
+ InstallInstalla
-
+ Install online modelInstalla il modello online
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font><strong><font size="2">AVVERTENZA: non consigliato per il tuo hardware. Il modello richiede più memoria (%1 GB) di quella disponibile nel sistema (%2).</strong></font>
-
+ ERROR: $API_KEY is empty.ERRORE: $API_KEY è vuoto.
-
+ ERROR: $BASE_URL is empty.ERRORE: $BASE_URL non è valido.
-
+ enter $BASE_URLinserisci $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.ERRORE: $MODEL_NAME è vuoto.
-
+ enter $MODEL_NAMEinserisci $MODEL_NAME
-
+ %1 GB
-
-
+
+ ?
-
+ Describes an error that occurred when downloadingDescrive un errore che si è verificato durante lo scaricamento
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font><strong><font size="1"><a href="#error">Errore</a></strong></font>
-
+ Error for incompatible hardwareErrore per hardware incompatibile
-
+ Download progressBarBarra di avanzamento dello scaricamento
-
+ Shows the progress made in the downloadMostra lo stato di avanzamento dello scaricamento
-
+ Download speedVelocità di scaricamento
-
+ Download speed in bytes/kilobytes/megabytes per secondVelocità di scaricamento in byte/kilobyte/megabyte al secondo
-
+ Calculating...Calcolo in corso...
-
-
-
-
+
+
+
+ Whether the file hash is being calculatedSe viene calcolato l'hash del file
-
+ Displayed when the file hash is being calculatedVisualizzato durante il calcolo dell'hash del file
-
+ enter $API_KEYInserire $API_KEY
-
+ File sizeDimensione del file
-
+ RAM requiredRAM richiesta
-
+ ParametersParametri
-
+ QuantQuant
-
+ TypeTipo
@@ -356,22 +356,22 @@
ApplicationSettings
-
+ ApplicationApplicazione
-
+ Network dialogDialogo di rete
-
+ opt-in to share feedback/conversationsaderisci per condividere feedback/conversazioni
-
+ ERROR: Update system could not find the MaintenanceTool used<br>
to check for updates!<br><br>
Did you install this application using the online installer? If so,<br>
@@ -382,87 +382,87 @@
-
+ Error dialogDialogo d'errore
-
+ Application SettingsSettaggi applicazione
-
+ GeneralGenerale
-
+ ThemeTema
-
+ The application color scheme.La combinazione di colori dell'applicazione.
-
+ DarkScuro
-
+ LightChiaro
-
+ LegacyDarkScuro Legacy
-
+ Font SizeDimensioni del Font
-
+ The size of text in the application.La dimensione del testo nell'applicazione.
-
+ SmallPiccolo
-
+ MediumMedio
-
+ LargeGrande
-
+ Language and LocaleLingua e settaggi locali
-
+ The language and locale you wish to use.La lingua e i settaggi locali che vuoi utilizzare.
-
+ System LocaleSettaggi locali del sistema
-
+ DeviceDispositivo
@@ -471,139 +471,139 @@
Il dispositivo di calcolo utilizzato per la generazione del testo. "Auto" utilizza Vulkan o Metal.
-
+ The compute device used for text generation.Il dispositivo di calcolo utilizzato per la generazione del testo.
-
-
+
+ Application defaultApplicazione predefinita
-
+ Default ModelModello predefinito
-
+ The preferred model for new chats. Also used as the local server fallback.Il modello preferito per le nuove chat. Utilizzato anche come ripiego del server locale.
-
+ Suggestion ModeModalità suggerimento
-
+ Generate suggested follow-up questions at the end of responses.Genera le domande di approfondimento suggerite alla fine delle risposte.
-
+ When chatting with LocalDocsQuando chatti con LocalDocs
-
+ Whenever possibleQuando possibile
-
+ NeverMai
-
+ Download PathPercorso di scarico
-
+ Where to store local models and the LocalDocs database.Dove archiviare i modelli locali e il database LocalDocs.
-
+ BrowseEsplora
-
+ Choose where to save model filesScegli dove salvare i file del modello
-
+ Enable DatalakeAbilita Datalake
-
+ Send chats and feedback to the GPT4All Open-Source Datalake.Invia chat e commenti al Datalake Open Source GPT4All.
-
+ AdvancedAvanzate
-
+ CPU ThreadsThread della CPUTread CPU
-
+ The number of CPU threads used for inference and embedding.Il numero di thread della CPU utilizzati per l'inferenza e l'incorporamento.
-
+ Save Chat ContextSalva il contesto della chat
-
+ Save the chat model's state to disk for faster loading. WARNING: Uses ~2GB per chat.Salva lo stato del modello di chat su disco per un caricamento più rapido. ATTENZIONE: utilizza circa 2 GB per chat.
-
+ Enable Local ServerAbilita server locale
-
+ Expose an OpenAI-Compatible server to localhost. WARNING: Results in increased resource usage.Esporre un server compatibile con OpenAI a localhost. ATTENZIONE: comporta un maggiore utilizzo delle risorse.
-
+ API Server PortPorta del server API
-
+ The port to use for the local server. Requires restart.La porta da utilizzare per il server locale. Richiede il riavvio.
-
+ Check For UpdatesControlla gli aggiornamenti
-
+ Manually check for an update to GPT4All.Verifica manualmente l'aggiornamento di GPT4All.
-
+ UpdatesAggiornamenti
@@ -611,13 +611,13 @@
Chat
-
-
+
+ New ChatNuova Chat
-
+ Server ChatChat del server
@@ -625,12 +625,12 @@
ChatAPIWorker
-
+ ERROR: Network error occurred while connecting to the API serverERRORE: si è verificato un errore di rete durante la connessione al server API
-
+ ChatAPIWorker::handleFinished got HTTP Error %1 %2ChatAPIWorker::handleFinished ha ricevuto l'errore HTTP %1 %2
@@ -638,62 +638,62 @@
ChatDrawer
-
+ DrawerCassetto
-
+ Main navigation drawerCassetto di navigazione principale
-
+ + New Chat+ Nuova Chat
-
+ Create a new chatCrea una nuova chat
-
+ Select the current chat or edit the chat when in edit modeSeleziona la chat corrente o modifica la chat in modalità modifica
-
+ Edit chat nameModifica il nome della chat
-
+ Save chat nameSalva il nome della chat
-
+ Delete chatElimina chat
-
+ Confirm chat deletionConferma l'eliminazione della chat
-
+ Cancel chat deletionAnnulla l'eliminazione della chat
-
+ List of chatsElenco delle chat
-
+ List of chats in the drawer dialogElenco delle chat nella finestra di dialogo del cassetto
@@ -701,32 +701,32 @@
ChatListModel
-
+ TODAYOGGI
-
+ THIS WEEKQUESTA SETTIMANA
-
+ THIS MONTHQUESTO MESE
-
+ LAST SIX MONTHSULTIMI SEI MESI
-
+ THIS YEARQUEST'ANNO
-
+ LAST YEARL'ANNO SCORSO
@@ -734,150 +734,150 @@
ChatView
-
+ <h3>Warning</h3><p>%1</p><h3>Avviso</h3><p>%1</p>
-
+ Switch model dialogFinestra di dialogo Cambia modello
-
+ Warn the user if they switch models, then context will be erasedAvvisa l'utente che se cambia modello, il contesto verrà cancellato
-
+ Conversation copied to clipboard.Conversazione copiata negli appunti.
-
+ Code copied to clipboard.Codice copiato negli appunti.
-
+ Chat panelPannello chat
-
+ Chat panel with optionsPannello chat con opzioni
-
+ Reload the currently loaded modelRicarica il modello attualmente caricato
-
+ Eject the currently loaded modelEspelli il modello attualmente caricato
-
+ No model installed.Nessun modello installato.
-
+ Model loading error.Errore di caricamento del modello.
-
+ Waiting for model...In attesa del modello...
-
+ Switching context...Cambio contesto...
-
+ Choose a model...Scegli un modello...
-
+ Not found: %1Non trovato: %1
-
+ The top item is the current modelL'elemento in alto è il modello attuale
-
-
+
+ LocalDocs
-
+ Add documentsAggiungi documenti
-
+ add collections of documents to the chataggiungi raccolte di documenti alla chat
-
+ Load the default modelCarica il modello predefinito
-
+ Loads the default model which can be changed in settingsCarica il modello predefinito che può essere modificato nei settaggi
-
+ No Model InstalledNessun modello installato
-
+ GPT4All requires that you install at least one
model to get startedGPT4All richiede l'installazione di almeno un
modello per iniziare
-
+ Install a ModelInstalla un modello
-
+ Shows the add model viewMostra la vista aggiungi modello
-
+ Conversation with the modelConversazione con il modello
-
+ prompt / response pairs from the conversationcoppie prompt/risposta dalla conversazione
-
+ GPT4All
-
+ YouTu
@@ -886,139 +886,139 @@ modello per iniziare
ricalcolo contesto ...
-
+ response stopped ...risposta interrotta ...
-
+ processing ...elaborazione ...
-
+ generating response ...generazione risposta ...
-
+ generating questions ...generarzione domande ...
-
-
+
+ CopyCopia
-
+ Copy MessageCopia messaggio
-
+ Disable markdownDisabilita Markdown
-
+ Enable markdownAbilita Markdown
-
+ Thumbs upMi piace
-
+ Gives a thumbs up to the responseDà un mi piace alla risposta
-
+ Thumbs downNon mi piace
-
+ Opens thumbs down dialogApre la finestra di dialogo "Non mi piace"
-
+ Suggested follow-upsApprofondimenti suggeriti
-
+ Erase and reset chat sessionCancella e ripristina la sessione di chat
-
+ Copy chat session to clipboardCopia la sessione di chat negli appunti
-
+ Redo last chat responseRiesegui l'ultima risposta della chat
-
+ Stop generatingInterrompi la generazione
-
+ Stop the current response generationArresta la generazione della risposta corrente
-
+ Reloads the modelRicarica il modello
-
+ <h3>Encountered an error loading model:</h3><br><i>"%1"</i><br><br>Model loading failures can happen for a variety of reasons, but the most common causes include a bad file format, an incomplete or corrupted download, the wrong file type, not enough system RAM or an incompatible model type. Here are some suggestions for resolving the problem:<br><ul><li>Ensure the model file has a compatible format and type<li>Check the model file is complete in the download folder<li>You can find the download folder in the settings dialog<li>If you've sideloaded the model ensure the file is not corrupt by checking md5sum<li>Read more about what models are supported in our <a href="https://docs.gpt4all.io/">documentation</a> for the gui<li>Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help<h3>Si è verificato un errore durante il caricamento del modello:</h3><br><i>"%1"</i><br><br>Gli errori di caricamento del modello possono verificarsi per diversi motivi, ma le cause più comuni includono un formato di file non valido, un download incompleto o danneggiato, il tipo di file sbagliato, RAM di sistema insufficiente o un tipo di modello incompatibile. Ecco alcuni suggerimenti per risolvere il problema:<br><ul><li>Assicurati che il file del modello abbia un formato e un tipo compatibili<li>Verifica che il file del modello sia completo nella cartella di download<li>Puoi trovare la cartella di download nella finestra di dialogo dei settaggi<li>Se hai scaricato manualmente il modello, assicurati che il file non sia danneggiato controllando md5sum<li>Leggi ulteriori informazioni su quali modelli sono supportati nella nostra <a href="https://docs.gpt4all.io/ ">documentazione</a> per la GUI<li>Consulta il nostro <a href="https://discord.gg/4M2QFmTt2k">canale Discord</a> per assistenza
-
-
+
+ Reload · %1Ricarica · %1
-
+ Loading · %1Caricamento · %1
-
+ Load · %1 (default) →Carica · %1 (predefinito) →
-
+ restoring from text ...ripristino dal testo ...
-
+ retrieving localdocs: %1 ...recupero documenti locali: %1 ...
-
+ searching localdocs: %1 ...ricerca in documenti locali: %1 ...
-
+ %n Source(s)%n Fonte
@@ -1026,42 +1026,42 @@ modello per iniziare
-
+ Send a message...Manda un messaggio...
-
+ Load a model to continue...Carica un modello per continuare...
-
+ Send messages/prompts to the modelInvia messaggi/prompt al modello
-
+ CutTaglia
-
+ PasteIncolla
-
+ Select AllSeleziona tutto
-
+ Send messageInvia messaggio
-
+ Sends the message/prompt contained in textfield to the modelInvia il messaggio/prompt contenuto nel campo di testo al modello
@@ -1069,12 +1069,12 @@ modello per iniziare
CollectionsDrawer
-
+ Warning: searching collections while indexing can return incomplete resultsAvviso: la ricerca nelle raccolte durante l'indicizzazione può restituire risultati incompleti
-
+ %n file(s)%n file
@@ -1082,7 +1082,7 @@ modello per iniziare
-
+ %n word(s)%n parola
@@ -1090,17 +1090,17 @@ modello per iniziare
-
+ UpdatingIn aggiornamento
-
+ + Add Docs+ Aggiungi documenti
-
+ Select a collection to make it available to the chat model.Seleziona una raccolta per renderla disponibile al modello in chat.
@@ -1108,37 +1108,37 @@ modello per iniziare
Download
-
+ Model "%1" is installed successfully.Il modello "%1" è stato installato correttamente.
-
+ ERROR: $MODEL_NAME is empty.ERRORE: $MODEL_NAME è vuoto.
-
+ ERROR: $API_KEY is empty.ERRORE: $API_KEY è vuoto.
-
+ ERROR: $BASE_URL is invalid.ERRORE: $BASE_URL non è valido.
-
+ ERROR: Model "%1 (%2)" is conflict.ERRORE: il modello "%1 (%2)" è in conflitto.
-
+ Model "%1 (%2)" is installed successfully.Il modello "%1 (%2)" è stato installato correttamente.
-
+ Model "%1" is removed.Il modello "%1" è stato rimosso.
@@ -1146,92 +1146,92 @@ modello per iniziare
HomeView
-
+ Welcome to GPT4AllBenvenuto in GPT4All
-
+ The privacy-first LLM chat applicationL'applicazione di chat LLM che mette al primo posto la privacy
-
+ Start chattingInizia a chattare
-
+ Start ChattingInizia a Chattare
-
+ Chat with any LLMChatta con qualsiasi LLM
-
+ LocalDocs
-
+ Chat with your local filesChatta con i tuoi file locali
-
+ Find ModelsTrova modelli
-
+ Explore and download modelsEsplora e scarica i modelli
-
+ Latest newsUltime notizie
-
+ Latest news from GPT4AllUltime notizie da GPT4All
-
+ Release NotesNote di rilascio
-
+ DocumentationDocumentazione
-
+ Discord
-
+ X (Twitter)
-
+ Github
-
+ nomic.ainomic.ai
-
+ Subscribe to NewsletterIscriviti alla Newsletter
@@ -1239,118 +1239,118 @@ modello per iniziare
LocalDocsSettings
-
+ LocalDocs
-
+ LocalDocs SettingsSettaggi LocalDocs
-
+ IndexingIndicizzazione
-
+ Allowed File ExtensionsEstensioni di file consentite
-
+ Comma-separated list. LocalDocs will only attempt to process files with these extensions.Elenco separato da virgole. LocalDocs tenterà di elaborare solo file con queste estensioni.
-
+ EmbeddingQuesto termine si dovrebbe tradurre come "Incorporamento". This term has been translated in other applications like A1111 and InvokeAI as "Incorporamento"Incorporamento
-
+ Use Nomic Embed APIUtilizza l'API di incorporamento Nomic Embed
-
+ Embed documents using the fast Nomic API instead of a private local model. Requires restart.Incorpora documenti utilizzando la veloce API di Nomic invece di un modello locale privato. Richiede il riavvio.
-
+ Nomic API KeyChiave API di Nomic
-
+ API key to use for Nomic Embed. Get one from the Atlas <a href="https://atlas.nomic.ai/cli-login">API keys page</a>. Requires restart.Chiave API da utilizzare per Nomic Embed. Ottienine una dalla <a href="https://atlas.nomic.ai/cli-login">pagina delle chiavi API</a> di Atlas. Richiede il riavvio.
-
+ Embeddings DeviceDispositivo per incorporamenti
-
+ The compute device used for embeddings. Requires restart.Il dispositivo di calcolo utilizzato per gli incorporamenti. Richiede il riavvio.
-
+ Application defaultApplicazione predefinita
-
+ DisplayMostra
-
+ Show SourcesMostra le fonti
-
+ Display the sources used for each response.Visualizza le fonti utilizzate per ciascuna risposta.
-
+ AdvancedAvanzate
-
+ Warning: Advanced usage only.Avvertenza: solo per uso avanzato.
-
+ Values too large may cause localdocs failure, extremely slow responses or failure to respond at all. Roughly speaking, the {N chars x N snippets} are added to the model's context window. More info <a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">here</a>.Valori troppo grandi possono causare errori di Localdocs, risposte estremamente lente o l'impossibilità di rispondere. In parole povere, {N caratteri x N frammenti} vengono aggiunti alla finestra di contesto del modello. Maggiori informazioni <a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">qui</a>.
-
+ Document snippet size (characters)Dimensioni del frammento di documento (caratteri)
-
+ Number of characters per document snippet. Larger numbers increase likelihood of factual responses, but also result in slower generation.Numero di caratteri per frammento di documento. Numeri più grandi aumentano la probabilità di risposte basate sui fatti, ma comportano anche una generazione più lenta.
-
+ Max document snippets per promptNumero massimo di frammenti di documento per prompt
-
+ Max best N matches of retrieved document snippets to add to the context for prompt. Larger numbers increase likelihood of factual responses, but also result in slower generation.Il numero massimo di frammenti di documento recuperati, che presentano le migliori corrispondenze, da includere nel contesto del prompt. Numeri più alti aumentano la probabilità di ricevere risposte basate sui fatti, ma comportano anche una generazione più lenta.
@@ -1358,17 +1358,17 @@ modello per iniziare
LocalDocsView
-
+ LocalDocs
-
+ Chat with your local filesChatta con i tuoi file locali
-
+ + Add Collection+ Aggiungi raccolta
@@ -1377,102 +1377,102 @@ modello per iniziare
ERRORE: il database di LocalDocs non è valido.
-
+ <h3>ERROR: The LocalDocs database cannot be accessed or is not valid.</h3><br><i>Note: You will need to restart after trying any of the following suggested fixes.</i><br><ul><li>Make sure that the folder set as <b>Download Path</b> exists on the file system.</li><li>Check ownership as well as read and write permissions of the <b>Download Path</b>.</li><li>If there is a <b>localdocs_v2.db</b> file, check its ownership and read/write permissions, too.</li></ul><br>If the problem persists and there are any 'localdocs_v*.db' files present, as a last resort you can<br>try backing them up and removing them. You will have to recreate your collections, however.<h3>ERRORE: Impossibile accedere al database LocalDocs o non è valido.</h3><br><i>Nota: sarà necessario riavviare dopo aver provato una delle seguenti soluzioni suggerite.</i><br><ul><li>Assicurati che la cartella impostata come <b>Percorso di download</b> esista nel file system.</li><li>Controlla la proprietà e i permessi di lettura e scrittura del <b>Percorso di download</b>.</li><li>Se è presente un file <b>localdocs_v2.db</b>, controlla anche la sua proprietà e i permessi di lettura/scrittura.</li></ul><br>Se il problema persiste e sono presenti file 'localdocs_v*.db', come ultima risorsa puoi<br>provare a eseguirne il backup e a rimuoverli. Tuttavia, dovrai ricreare le tue raccolte.
-
+ No Collections InstalledNessuna raccolta installata
-
+ Install a collection of local documents to get started using this featureInstalla una raccolta di documenti locali per iniziare a utilizzare questa funzionalità
-
+ + Add Doc Collection+ Aggiungi raccolta di documenti
-
+ Shows the add model viewMostra la vista aggiungi modello
-
+ Indexing progressBarBarra di avanzamento indicizzazione
-
+ Shows the progress made in the indexingMostra lo stato di avanzamento dell'indicizzazione
-
+ ERRORERRORE
-
+ INDEXINGINDICIZZAZIONE
-
+ EMBEDDINGINCORPORAMENTO
-
+ REQUIRES UPDATERICHIEDE AGGIORNAMENTO
-
+ READYPRONTO
-
+ INSTALLINGINSTALLAZIONE
-
+ Indexing in progressIndicizzazione in corso
-
+ Embedding in progressIncorporamento in corso
-
+ This collection requires an update after version changeQuesta raccolta richiede un aggiornamento dopo il cambio di versione
-
+ Automatically reindexes upon changes to the folderReindicizza automaticamente in caso di modifiche alla cartella
-
+ Installation in progressInstallazione in corso
-
+ %%
-
+ %n file(s)%n file
@@ -1480,7 +1480,7 @@ modello per iniziare
-
+ %n word(s)%n parola
@@ -1488,27 +1488,27 @@ modello per iniziare
-
+ RemoveRimuovi
-
+ RebuildRicostruisci
-
+ Reindex this folder from scratch. This is slow and usually not needed.Reindicizzare questa cartella da zero. Lento e di solito non necessario.
-
+ UpdateAggiorna
-
+ Update the collection to the new version. This is a slow operation.Aggiorna la raccolta alla nuova versione. Questa è un'operazione lenta.
@@ -1516,67 +1516,78 @@ modello per iniziare
ModelList
-
+ <ul><li>Requires personal OpenAI API key.</li><li>WARNING: Will send your chats to OpenAI!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with OpenAI</li><li>You can apply for an API key <a href="https://platform.openai.com/account/api-keys">here.</a></li><ul><li>Richiede una chiave API OpenAI personale.</li><li>ATTENZIONE: invierà le tue chat a OpenAI!</li><li>La tua chiave API verrà archiviata su disco</li><li> Verrà utilizzato solo per comunicare con OpenAI</li><li>Puoi richiedere una chiave API <a href="https://platform.openai.com/account/api-keys">qui.</a> </li>
-
+ <strong>OpenAI's ChatGPT model GPT-3.5 Turbo</strong><br> %1
-
+ <strong>OpenAI's ChatGPT model GPT-4</strong><br> %1 %2
-
+ <strong>Mistral Tiny model</strong><br> %1
-
+ <strong>Mistral Small model</strong><br> %1
-
+ <strong>Mistral Medium model</strong><br> %1
-
+ <br><br><i>* Even if you pay OpenAI for ChatGPT-4 this does not guarantee API key access. Contact OpenAI for more info.<br><br><i>* Anche se paghi OpenAI per ChatGPT-4 questo non garantisce l'accesso alla chiave API. Contatta OpenAI per maggiori informazioni.
-
+
+
+ cannot open "%1": %2
+
+
+
+
+ cannot create "%1": %2
+
+
+
+ %1 (%2)%1 (%2)
-
+ <strong>OpenAI-Compatible API Model</strong><br><ul><li>API Key: %1</li><li>Base URL: %2</li><li>Model Name: %3</li></ul><strong>Modello API compatibile con OpenAI</strong><br><ul><li>Chiave API: %1</li><li>URL di base: %2</li><li>Nome modello: %3</li></ul>
-
+ <ul><li>Requires personal Mistral API key.</li><li>WARNING: Will send your chats to Mistral!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with Mistral</li><li>You can apply for an API key <a href="https://console.mistral.ai/user/api-keys">here</a>.</li><ul><li>Richiede una chiave API Mistral personale.</li><li>ATTENZIONE: invierà le tue chat a Mistral!</li><li>La tua chiave API verrà archiviata su disco</li><li> Verrà utilizzato solo per comunicare con Mistral</li><li>Puoi richiedere una chiave API <a href="https://console.mistral.ai/user/api-keys">qui</a>. </li>
-
+ <ul><li>Requires personal API key and the API base URL.</li><li>WARNING: Will send your chats to the OpenAI-compatible API Server you specified!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with the OpenAI-compatible API Server</li><ul><li>Richiede una chiave API personale e l'URL di base dell'API.</li><li>ATTENZIONE: invierà le tue chat al server API compatibile con OpenAI che hai specificato!</li><li>La tua chiave API verrà archiviata su disco</li><li>Verrà utilizzata solo per comunicare con il server API compatibile con OpenAI</li>
-
+ <strong>Connect to OpenAI-compatible API server</strong><br> %1<strong>Connetti al server API compatibile con OpenAI</strong><br> %1
-
+ <strong>Created by %1.</strong><br><ul><li>Published on %2.<li>This model has %3 likes.<li>This model has %4 downloads.<li>More info can be found <a href="https://huggingface.co/%5">here.</a></ul><strong>Creato da %1.</strong><br><ul><li>Pubblicato il %2.<li>Questo modello ha %3 Mi piace.<li>Questo modello ha %4 download.<li>Altro informazioni possono essere trovate <a href="https://huggingface.co/%5">qui.</a></ul>
@@ -1584,92 +1595,92 @@ modello per iniziare
ModelSettings
-
+ ModelModello
-
+ Model SettingsSettaggi modello
-
+ CloneClona
-
+ RemoveRimuovi
-
+ NameNome
-
+ Model FileFile del modello
-
+ System PromptPrompt di sistema
-
+ Prefixed at the beginning of every conversation. Must contain the appropriate framing tokens.Prefisso all'inizio di ogni conversazione. Deve contenere i token di inquadramento appropriati.
-
+ Prompt TemplateSchema del prompt
-
+ The template that wraps every prompt.Lo schema che incorpora ogni prompt.
-
+ Must contain the string "%1" to be replaced with the user's input.Deve contenere la stringa "%1" da sostituire con l'input dell'utente.
-
+ Chat Name PromptPrompt del nome della chat
-
+ Prompt used to automatically generate chat names.Prompt utilizzato per generare automaticamente nomi di chat.
-
+ Suggested FollowUp PromptPrompt di approfondimento suggerito
-
+ Prompt used to generate suggested follow-up questions.Prompt utilizzato per generare le domande di approfondimento suggerite.
-
+ Context LengthLunghezza del contesto
-
+ Number of input and output tokens the model sees.Numero di token di input e output visualizzati dal modello.
-
+ Maximum combined prompt/response tokens before information is lost.
Using more context than the model was trained on will yield poor results.
NOTE: Does not take effect until you reload the model.
@@ -1678,128 +1689,128 @@ L'utilizzo di un contesto maggiore rispetto a quello su cui è stato addest
NOTA: non ha effetto finché non si ricarica il modello.
-
+ TemperatureTemperatura
-
+ Randomness of model output. Higher -> more variation.Casualità dell'uscita del modello. Più alto -> più variazione.
-
+ Temperature increases the chances of choosing less likely tokens.
NOTE: Higher temperature gives more creative but less predictable outputs.La temperatura aumenta le possibilità di scegliere token meno probabili.
NOTA: una temperatura più elevata offre risultati più creativi ma meno prevedibili.
-
+ Top-P
-
+ Nucleus Sampling factor. Lower -> more predictable.Fattore di campionamento del nucleo. Inferiore -> più prevedibile.
-
+ Only the most likely tokens up to a total probability of top_p can be chosen.
NOTE: Prevents choosing highly unlikely tokens.Solo i token più probabili, fino a un totale di probabilità di top_p, possono essere scelti.
NOTA: impedisce la scelta di token altamente improbabili.
-
+ Min-P
-
+ Minimum token probability. Higher -> more predictable.Probabilità minima del token. Più alto -> più prevedibile.
-
+ Sets the minimum relative probability for a token to be considered.Imposta la probabilità relativa minima affinché un token venga considerato.
-
+ Top-K
-
+ Size of selection pool for tokens.Dimensione del lotto di selezione per i token.
-
+ Only the top K most likely tokens will be chosen from.Saranno scelti solo i primi K token più probabili.
-
+ Max LengthLunghezza massima
-
+ Maximum response length, in tokens.Lunghezza massima della risposta, in token.
-
+ Prompt Batch SizeDimensioni del lotto di prompt
-
+ The batch size used for prompt processing.La dimensione del lotto usata per l'elaborazione dei prompt.
-
+ Amount of prompt tokens to process at once.
NOTE: Higher values can speed up reading prompts but will use more RAM.Numero di token del prompt da elaborare contemporaneamente.
NOTA: valori più alti possono velocizzare la lettura dei prompt ma utilizzeranno più RAM.
-
+ Repeat PenaltyPenalità di ripetizione
-
+ Repetition penalty factor. Set to 1 to disable.Fattore di penalità di ripetizione. Impostare su 1 per disabilitare.
-
+ Repeat Penalty TokensToken di penalità ripetizione
-
+ Number of previous tokens used for penalty.Numero di token precedenti utilizzati per la penalità.
-
+ GPU LayersLivelli GPU
-
+ Number of model layers to load into VRAM.Numero di livelli del modello da caricare nella VRAM.
-
+ How many model layers to load into VRAM. Decrease this if GPT4All runs out of VRAM while loading this model.
Lower values increase CPU load and RAM usage, and make inference slower.
NOTE: Does not take effect until you reload the model.
@@ -1811,217 +1822,217 @@ NOTA: non ha effetto finché non si ricarica il modello.
ModelsView
-
+ No Models InstalledNessun modello installato
-
+ Install a model to get started using GPT4AllInstalla un modello per iniziare a utilizzare GPT4All
-
-
+
+ + Add Model+ Aggiungi Modello
-
+ Shows the add model viewMostra la vista aggiungi modello
-
+ Installed ModelsModelli installati
-
+ Locally installed chat modelsModelli per chat installati localmente
-
+ Model fileFile del modello
-
+ Model file to be downloadedFile del modello da scaricare
-
+ DescriptionDescrizione
-
+ File descriptionDescrizione del file
-
+ CancelAnnulla
-
+ ResumeRiprendi
-
+ Stop/restart/start the downloadArresta/riavvia/avvia il download
-
+ RemoveRimuovi
-
+ Remove model from filesystemRimuovi il modello dal sistema dei file
-
-
+
+ InstallInstalla
-
+ Install online modelInstalla il modello online
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font><strong><font size="1"><a href="#error">Errore</a></strong></font>
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font><strong><font size="2">AVVISO: non consigliato per il tuo hardware. Il modello richiede più memoria (%1 GB) di quella disponibile nel sistema (%2).</strong></font>
-
+ ERROR: $API_KEY is empty.ERRORE: $API_KEY è vuoto.
-
+ ERROR: $BASE_URL is empty.ERRORE: $BASE_URL non è valido.
-
+ enter $BASE_URLinserisci $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.ERRORE: $MODEL_NAME è vuoto.
-
+ enter $MODEL_NAMEinserisci $MODEL_NAME
-
+ %1 GB
-
+ ?
-
+ Describes an error that occurred when downloadingDescrive un errore che si è verificato durante lo scaricamento
-
+ Error for incompatible hardwareErrore per hardware incompatibile
-
+ Download progressBarBarra di avanzamento dello scaricamento
-
+ Shows the progress made in the downloadMostra lo stato di avanzamento dello scaricamento
-
+ Download speedVelocità di scaricamento
-
+ Download speed in bytes/kilobytes/megabytes per secondVelocità di scaricamento in byte/kilobyte/megabyte al secondo
-
+ Calculating...Calcolo in corso...
-
-
-
-
+
+
+
+ Whether the file hash is being calculatedSe viene calcolato l'hash del file
-
+ Busy indicatorIndicatore di occupato
-
+ Displayed when the file hash is being calculatedVisualizzato durante il calcolo dell'hash del file
-
+ enter $API_KEYInserire $API_KEY
-
+ File sizeDimensione del file
-
+ RAM requiredRAM richiesta
-
+ ParametersParametri
-
+ QuantQuant
-
+ TypeTipo
@@ -2029,12 +2040,12 @@ NOTA: non ha effetto finché non si ricarica il modello.
MyFancyLink
-
+ Fancy linkMio link
-
+ A stylized linkUn link d'esempio
@@ -2042,7 +2053,7 @@ NOTA: non ha effetto finché non si ricarica il modello.
MySettingsStack
-
+ Please choose a directoryScegli una cartella
@@ -2050,12 +2061,12 @@ NOTA: non ha effetto finché non si ricarica il modello.
MySettingsTab
-
+ Restore DefaultsRiprista i valori predefiniti
-
+ Restores settings dialog to a default stateRipristina la finestra di dialogo dei settaggi a uno stato predefinito
@@ -2063,12 +2074,12 @@ NOTA: non ha effetto finché non si ricarica il modello.
NetworkDialog
-
+ Contribute data to the GPT4All Opensource Datalake.Contribuisci con i tuoi dati al Datalake Open Source di GPT4All.
-
+ By enabling this feature, you will be able to participate in the democratic process of training a large language model by contributing data for future model improvements.
When a GPT4All model responds to you and you have opted-in, your conversation will be sent to the GPT4All Open Source Datalake. Additionally, you can like/dislike its response. If you dislike a response, you can suggest an alternative response. This data will be collected and aggregated in the GPT4All Datalake.
@@ -2081,47 +2092,47 @@ Quando un modello di GPT4All ti risponde e tu hai aderito, la tua conversazione
NOTA: attivando questa funzione, invierai i tuoi dati al Datalake Open Source di GPT4All. Non dovresti avere aspettative sulla privacy della chat quando questa funzione è abilitata. Dovresti, tuttavia, aspettarti un'attribuzione facoltativa, se lo desideri. I tuoi dati di chat saranno liberamente disponibili per essere scaricati da chiunque e verranno utilizzati da Nomic AI per migliorare i futuri modelli GPT4All. Nomic AI conserverà tutte le informazioni di attribuzione allegate ai tuoi dati e verrai accreditato come collaboratore a qualsiasi versione del modello GPT4All che utilizza i tuoi dati!
-
+ Terms for opt-inTermini per l'adesione
-
+ Describes what will happen when you opt-inDescrive cosa accadrà quando effettuerai l'adesione
-
+ Please provide a name for attribution (optional)Fornisci un nome per l'attribuzione (facoltativo)
-
+ Attribution (optional)Attribuzione (facoltativo)
-
+ Provide attributionFornire attribuzione
-
+ EnableAbilita
-
+ Enable opt-inAbilita l'adesione
-
+ CancelAnnulla
-
+ Cancel opt-inAnnulla l'adesione
@@ -2129,17 +2140,17 @@ NOTA: attivando questa funzione, invierai i tuoi dati al Datalake Open Source di
NewVersionDialog
-
+ New version is availableNuova versione disponibile
-
+ UpdateAggiorna
-
+ Update to new versionAggiorna alla nuova versione
@@ -2147,17 +2158,17 @@ NOTA: attivando questa funzione, invierai i tuoi dati al Datalake Open Source di
PopupDialog
-
+ Reveals a shortlived help balloonRivela un messaggio di aiuto di breve durata
-
+ Busy indicatorIndicatore di occupato
-
+ Displayed when the popup is showing busyVisualizzato quando la finestra a comparsa risulta occupata
@@ -2165,28 +2176,28 @@ NOTA: attivando questa funzione, invierai i tuoi dati al Datalake Open Source di
SettingsView
-
-
+
+ SettingsSettaggi
-
+ Contains various application settingsContiene vari settaggi dell'applicazione
-
+ ApplicationApplicazione
-
+ ModelModello
-
+ LocalDocs
@@ -2194,12 +2205,12 @@ NOTA: attivando questa funzione, invierai i tuoi dati al Datalake Open Source di
StartupDialog
-
+ Welcome!Benvenuto!
-
+ ### Release notes
%1### Contributors
%2
@@ -2208,17 +2219,17 @@ NOTA: attivando questa funzione, invierai i tuoi dati al Datalake Open Source di
%2
-
+ Release notesNote di rilascio
-
+ Release notes for this versionNote di rilascio per questa versione
-
+ ### Opt-ins for anonymous usage analytics and datalake
By enabling these features, you will be able to participate in the democratic process of training a
large language model by contributing data for future model improvements.
@@ -2241,71 +2252,71 @@ Quando un modello di GPT4All ti risponde e tu hai aderito, la tua conversazione
NOTA: attivando questa funzione, invierai i tuoi dati al Datalake Open Source di GPT4All. Non dovresti avere aspettative sulla privacy della chat quando questa funzione è abilitata. Dovresti, tuttavia, aspettarti un'attribuzione facoltativa, se lo desideri, . I tuoi dati di chat saranno liberamente disponibili per essere scaricati da chiunque e verranno utilizzati da Nomic AI per migliorare i futuri modelli GPT4All. Nomic AI conserverà tutte le informazioni di attribuzione allegate ai tuoi dati e verrai accreditato come collaboratore a qualsiasi versione del modello GPT4All che utilizza i tuoi dati!
-
+ Terms for opt-inTermini per l'adesione
-
+ Describes what will happen when you opt-inDescrive cosa accadrà quando effettuerai l'adesione
-
-
+
+ Opt-in for anonymous usage statisticsAttiva le statistiche di utilizzo anonime
-
-
+
+ YesSi
-
+ Allow opt-in for anonymous usage statisticsConsenti l'attivazione di statistiche di utilizzo anonime
-
-
+
+ NoNo
-
+ Opt-out for anonymous usage statisticsDisattiva le statistiche di utilizzo anonime
-
+ Allow opt-out for anonymous usage statisticsConsenti la disattivazione per le statistiche di utilizzo anonime
-
-
+
+ Opt-in for networkAderisci per la rete
-
+ Allow opt-in for networkConsenti l'adesione per la rete
-
+ Allow opt-in anonymous sharing of chats to the GPT4All DatalakeConsenti la condivisione anonima delle chat su GPT4All Datalake
-
+ Opt-out for networkDisattiva per la rete
-
+ Allow opt-out anonymous sharing of chats to the GPT4All DatalakeConsenti la non adesione alla condivisione anonima delle chat nel GPT4All Datalake
@@ -2313,23 +2324,23 @@ NOTA: attivando questa funzione, invierai i tuoi dati al Datalake Open Source di
SwitchModelDialog
-
+ <b>Warning:</b> changing the model will erase the current conversation. Do you wish to continue?<b>Avviso:</b> la modifica del modello cancellerà la conversazione corrente. Vuoi continuare?
-
+ ContinueContinua
-
+ Continue with model loadingContinuare con il caricamento del modello
-
-
+
+ CancelAnnulla
@@ -2337,32 +2348,32 @@ NOTA: attivando questa funzione, invierai i tuoi dati al Datalake Open Source di
ThumbsDownDialog
-
+ Please edit the text below to provide a better response. (optional)Modifica il testo seguente per fornire una risposta migliore. (opzionale)
-
+ Please provide a better response...Si prega di fornire una risposta migliore...
-
+ SubmitInvia
-
+ Submits the user's responseInvia la risposta dell'utente
-
+ CancelAnnulla
-
+ Closes the response dialogChiude la finestra di dialogo della risposta
@@ -2370,125 +2381,125 @@ NOTA: attivando questa funzione, invierai i tuoi dati al Datalake Open Source di
main
-
+ <h3>Encountered an error starting up:</h3><br><i>"Incompatible hardware detected."</i><br><br>Unfortunately, your CPU does not meet the minimal requirements to run this program. In particular, it does not support AVX intrinsics which this program requires to successfully run a modern large language model. The only solution at this time is to upgrade your hardware to a more modern CPU.<br><br>See here for more information: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a><h3>Si è verificato un errore all'avvio:</h3><br><i>"Rilevato hardware incompatibile."</i><br><br>Sfortunatamente, la tua CPU non soddisfa i requisiti minimi per eseguire questo programma. In particolare, non supporta gli elementi intrinseci AVX richiesti da questo programma per eseguire con successo un modello linguistico moderno e di grandi dimensioni. L'unica soluzione in questo momento è aggiornare il tuo hardware con una CPU più moderna.<br><br>Vedi qui per ulteriori informazioni: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https ://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a>
-
+ GPT4All v%1
-
+ <h3>Encountered an error starting up:</h3><br><i>"Inability to access settings file."</i><br><br>Unfortunately, something is preventing the program from accessing the settings file. This could be caused by incorrect permissions in the local app config directory where the settings file is located. Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help.<h3>Si è verificato un errore all'avvio:</h3><br><i>"Impossibile accedere al file dei settaggi."</i><br><br>Sfortunatamente, qualcosa impedisce al programma di accedere al file dei settaggi. Ciò potrebbe essere causato da autorizzazioni errate nella cartella di configurazione locale dell'app in cui si trova il file dei settaggi. Dai un'occhiata al nostro <a href="https://discord.gg/4M2QFmTt2k">canale Discord</a> per ricevere assistenza.
-
+ Connection to datalake failed.La connessione al Datalake non è riuscita.
-
+ Saving chats.Salvataggio delle chat.
-
+ Network dialogDialogo di rete
-
+ opt-in to share feedback/conversationsaderisci per condividere feedback/conversazioni
-
+ Home viewVista iniziale
-
+ Home view of applicationVista iniziale dell'applicazione
-
+ HomeInizia
-
+ Chat viewVista chat
-
+ Chat view to interact with modelsVista chat per interagire con i modelli
-
+ ChatsChat
-
-
+
+ ModelsModelli
-
+ Models view for installed modelsVista modelli per i modelli installati
-
-
+
+ LocalDocs
-
+ LocalDocs view to configure and use local docsVista LocalDocs per configurare e utilizzare i documenti locali
-
-
+
+ SettingsSettaggi
-
+ Settings view for application configurationVista dei settaggi per la configurazione dell'applicazione
-
+ The datalake is enabledIl Datalake è abilitato
-
+ Using a network modelUtilizzando un modello di rete
-
+ Server mode is enabledLa modalità server è abilitata
-
+ Installed modelsModelli installati
-
+ View of installed modelsVista dei modelli installati
diff --git a/gpt4all-chat/translations/gpt4all_pt_BR.ts b/gpt4all-chat/translations/gpt4all_pt_BR.ts
index 38a9177b..76b5b5aa 100644
--- a/gpt4all-chat/translations/gpt4all_pt_BR.ts
+++ b/gpt4all-chat/translations/gpt4all_pt_BR.ts
@@ -4,62 +4,62 @@
AddCollectionView
-
+ ← Existing Collections← Minhas coleções
-
+ Add Document CollectionAdicionar Coleção de Documentos
-
+ Add a folder containing plain text files, PDFs, or Markdown. Configure additional extensions in Settings.Adicione uma pasta contendo arquivos de texto simples, PDFs ou Markdown. Configure extensões adicionais nas Configurações.
-
+ Please choose a directoryEscolha um diretório
-
+ NameNome
-
+ Collection name...Nome da coleção...
-
+ Name of the collection to add (Required)Nome da coleção (obrigatório)
-
+ FolderPasta
-
+ Folder path...Caminho da pasta...
-
+ Folder path to documents (Required)Caminho da pasta com os documentos (obrigatório)
-
+ BrowseProcurar
-
+ Create CollectionCriar Coleção
@@ -67,288 +67,288 @@
AddModelView
-
+ ← Existing Models← Meus Modelos
-
+ Explore ModelsDescobrir Modelos
-
+ Discover and download models by keyword search...Pesquisar modelos...
-
+ Text field for discovering and filtering downloadable modelsCampo de texto para descobrir e filtrar modelos para download
-
+ Initiate model discovery and filteringPesquisar e filtrar modelos
-
+ Triggers discovery and filtering of modelsAciona a descoberta e filtragem de modelos
-
+ DefaultPadrão
-
+ LikesCurtidas
-
+ DownloadsDownloads
-
+ RecentRecentes
-
+ AscAsc
-
+ DescDesc
-
+ NoneNenhum
-
+ Searching · %1Pesquisando · %1
-
+ Sort by: %1Ordenar por: %1
-
+ Sort dir: %1Ordenar diretório: %1
-
+ Limit: %1Limite: %1
-
+ Network error: could not retrieve %1Erro de rede: não foi possível obter %1
-
-
+
+ Busy indicatorIndicador de processamento
-
+ Displayed when the models request is ongoingxibido enquanto os modelos estão sendo carregados
-
+ Model fileArquivo do modelo
-
+ Model file to be downloadedArquivo do modelo a ser baixado
-
+ DescriptionDescrição
-
+ File descriptionDescrição do arquivo
-
+ CancelCancelar
-
+ ResumeRetomar
-
+ DownloadBaixar
-
+ Stop/restart/start the downloadParar/reiniciar/iniciar o download
-
+ RemoveRemover
-
+ Remove model from filesystemRemover modelo do sistema
-
-
+
+ InstallInstalar
-
+ Install online modelInstalar modelo online
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font><strong><font size="2">ATENÇÃO: Este modelo não é recomendado para seu hardware. Ele exige mais memória (%1 GB) do que seu sistema possui (%2).</strong></font>
-
+ ERROR: $API_KEY is empty.ERRO: A $API_KEY está vazia.
-
+ ERROR: $BASE_URL is empty.ERRO: A $BASE_URL está vazia.
-
+ enter $BASE_URLinserir a $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.ERRO: O $MODEL_NAME está vazio.
-
+ enter $MODEL_NAMEinserir o $MODEL_NAME
-
+ %1 GB%1 GB
-
-
+
+ ??
-
+ Describes an error that occurred when downloadingMostra informações sobre o erro no download
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font><strong><font size="1"><a href="#error">Erro</a></strong></font>
-
+ Error for incompatible hardwareAviso: Hardware não compatível
-
+ Download progressBarProgresso do download
-
+ Shows the progress made in the downloadMostra o progresso do download
-
+ Download speedVelocidade de download
-
+ Download speed in bytes/kilobytes/megabytes per secondVelocidade de download em bytes/kilobytes/megabytes por segundo
-
+ Calculating...Calculando...
-
-
-
-
+
+
+
+ Whether the file hash is being calculatedQuando o hash do arquivo está sendo calculado
-
+ Displayed when the file hash is being calculatedExibido durante o cálculo do hash do arquivo
-
+ enter $API_KEYinserir $API_KEY
-
+ File sizeTamanho do arquivo
-
+ RAM requiredRAM necessária
-
+ ParametersParâmetros
-
+ QuantQuant
-
+ TypeTipo
@@ -356,22 +356,22 @@
ApplicationSettings
-
+ ApplicationAplicativo
-
+ Network dialogMensagens de rede
-
+ opt-in to share feedback/conversationsCompartilhar feedback e conversas
-
+ ERROR: Update system could not find the MaintenanceTool used<br>
to check for updates!<br><br>
Did you install this application using the online installer? If so,<br>
@@ -388,225 +388,225 @@
reinstalar o aplicativo.
-
+ Error dialogMensagens de erro
-
+ Application SettingsConfigurações
-
+ GeneralGeral
-
+ ThemeTema
-
+ The application color scheme.Esquema de cores.
-
+ DarkModo Escuro
-
+ LightModo Claro
-
+ LegacyDarkModo escuro (legado)
-
+ Font SizeTamanho da Fonte
-
+ The size of text in the application.Tamanho do texto.
-
+ SmallPequeno
-
+ MediumMédio
-
+ LargeGrande
-
+ Language and LocaleIdioma e Região
-
+ The language and locale you wish to use.Selecione seu idioma e região.
-
+ System LocaleLocal do Sistema
-
+ DeviceProcessador
-
+ The compute device used for text generation.I chose to use "Processador" instead of "Dispositivo" (Device) or "Dispositivo de Computação" (Compute Device) to simplify the terminology and make it more straightforward and understandable. "Dispositivo" can be vague and could refer to various types of hardware, whereas "Processador" clearly and specifically indicates the component responsible for processing tasks. This improves usability by avoiding the ambiguity that might arise from using more generic terms like "Dispositivo."Processador usado para gerar texto.
-
-
+
+ Application defaultAplicativo padrão
-
+ Default ModelModelo Padrão
-
+ The preferred model for new chats. Also used as the local server fallback.Modelo padrão para novos chats e em caso de falha do modelo principal.
-
+ Suggestion ModeModo de sugestões
-
+ Generate suggested follow-up questions at the end of responses.Sugerir perguntas após as respostas.
-
+ When chatting with LocalDocsAo conversar com o LocalDocs
-
+ Whenever possibleSempre que possível
-
+ NeverNunca
-
+ Download PathDiretório de Download
-
+ Where to store local models and the LocalDocs database.Pasta para modelos e banco de dados do LocalDocs.
-
+ BrowseProcurar
-
+ Choose where to save model filesLocal para armazenar os modelos
-
+ Enable DatalakeHabilitar Datalake
-
+ Send chats and feedback to the GPT4All Open-Source Datalake.Contribua para o Datalake de código aberto do GPT4All.
-
+ AdvancedAvançado
-
+ CPU ThreadsThreads de CPU
-
+ The number of CPU threads used for inference and embedding.Quantidade de núcleos (threads) do processador usados para processar e responder às suas perguntas.
-
+ Save Chat ContextI used "Histórico do Chat" (Chat History) instead of "Contexto do Chat" (Chat Context) to clearly convey that it refers to saving past messages, making it more intuitive and avoiding potential confusion with abstract terms.Salvar Histórico do Chat
-
+ Save the chat model's state to disk for faster loading. WARNING: Uses ~2GB per chat.Salvar histórico do chat para carregamento mais rápido. (Usa aprox. 2GB por chat).
-
+ Enable Local ServerAtivar Servidor Local
-
+ Expose an OpenAI-Compatible server to localhost. WARNING: Results in increased resource usage.Ativar servidor local compatível com OpenAI (uso de recursos elevado).
-
+ API Server PortPorta da API
-
+ The port to use for the local server. Requires restart.Porta de acesso ao servidor local. (requer reinicialização).
-
+ Check For UpdatesProcurar por Atualizações
-
+ Manually check for an update to GPT4All.Verifica se há novas atualizações para o GPT4All.
-
+ UpdatesAtualizações
@@ -614,13 +614,13 @@
Chat
-
-
+
+ New ChatNovo Chat
-
+ Server ChatChat com o Servidor
@@ -628,12 +628,12 @@
ChatAPIWorker
-
+ ERROR: Network error occurred while connecting to the API serverERRO: Ocorreu um erro de rede ao conectar-se ao servidor da API
-
+ ChatAPIWorker::handleFinished got HTTP Error %1 %2ChatAPIWorker::handleFinished recebeu erro HTTP %1 %2
@@ -641,62 +641,62 @@
ChatDrawer
-
+ DrawerMenu Lateral
-
+ Main navigation drawerMenu de navegação principal
-
+ + New Chat+ Novo Chat
-
+ Create a new chatCriar um novo chat
-
+ Select the current chat or edit the chat when in edit modeSelecione o chat atual ou edite o chat quando estiver no modo de edição
-
+ Edit chat nameEditar nome do chat
-
+ Save chat nameSalvar nome do chat
-
+ Delete chatExcluir chat
-
+ Confirm chat deletionConfirmar exclusão do chat
-
+ Cancel chat deletionCancelar exclusão do chat
-
+ List of chatsLista de chats
-
+ List of chats in the drawer dialogLista de chats na caixa de diálogo do menu lateral
@@ -704,32 +704,32 @@
ChatListModel
-
+ TODAYHOJE
-
+ THIS WEEKESTA SEMANA
-
+ THIS MONTHESTE MÊS
-
+ LAST SIX MONTHSÚLTIMOS SEIS MESES
-
+ THIS YEARESTE ANO
-
+ LAST YEARANO PASSADO
@@ -737,150 +737,150 @@
ChatView
-
+ <h3>Warning</h3><p>%1</p><h3>Aviso</h3><p>%1</p>
-
+ Switch model dialogMensagem ao troca de modelo
-
+ Warn the user if they switch models, then context will be erasedAo trocar de modelo, o contexto da conversa será apagado
-
+ Conversation copied to clipboard.Conversa copiada.
-
+ Code copied to clipboard.Código copiado.
-
+ Chat panelPainel de chat
-
+ Chat panel with optionsPainel de chat com opções
-
+ Reload the currently loaded modelRecarregar modelo atual
-
+ Eject the currently loaded modelEjetar o modelo carregado atualmente
-
+ No model installed.Nenhum modelo instalado.
-
+ Model loading error.Erro ao carregar o modelo.
-
+ Waiting for model...Aguardando modelo...
-
+ Switching context...Mudando de contexto...
-
+ Choose a model...Escolha um modelo...
-
+ Not found: %1Não encontrado: %1
-
+ The top item is the current modelO modelo atual é exibido no topo
-
-
+
+ LocalDocsLocalDocs
-
+ Add documentsAdicionar documentos
-
+ add collections of documents to the chatAdicionar Coleção de Documentos
-
+ Load the default modelCarregar o modelo padrão
-
+ Loads the default model which can be changed in settingsCarrega o modelo padrão (personalizável nas configurações)
-
+ No Model InstalledNenhum Modelo Instalado
-
+ GPT4All requires that you install at least one
model to get startedO GPT4All precisa de pelo menos um modelo
modelo instalado para funcionar
-
+ Install a ModelInstalar um Modelo
-
+ Shows the add model viewMostra a visualização para adicionar modelo
-
+ Conversation with the modelConversa com o modelo
-
+ prompt / response pairs from the conversationPares de pergunta/resposta da conversa
-
+ GPT4AllGPT4All
-
+ YouVocê
@@ -889,139 +889,139 @@ modelo instalado para funcionar
recalculando contexto...
-
+ response stopped ...resposta interrompida...
-
+ processing ...processando...
-
+ generating response ...gerando resposta...
-
+ generating questions ...gerando perguntas...
-
-
+
+ CopyCopiar
-
+ Copy MessageCopiar Mensagem
-
+ Disable markdownDesativar markdown
-
+ Enable markdownAtivar markdown
-
+ Thumbs upResposta boa
-
+ Gives a thumbs up to the responseCurte a resposta
-
+ Thumbs downResposta ruim
-
+ Opens thumbs down dialogAbrir diálogo de joinha para baixo
-
+ Suggested follow-upsPerguntas relacionadas
-
+ Erase and reset chat sessionApagar e redefinir sessão de chat
-
+ Copy chat session to clipboardCopiar histórico da conversa
-
+ Redo last chat responseRefazer última resposta
-
+ Stop generatingParar de gerar
-
+ Stop the current response generationParar a geração da resposta atual
-
+ Reloads the modelRecarrega modelo
-
+ <h3>Encountered an error loading model:</h3><br><i>"%1"</i><br><br>Model loading failures can happen for a variety of reasons, but the most common causes include a bad file format, an incomplete or corrupted download, the wrong file type, not enough system RAM or an incompatible model type. Here are some suggestions for resolving the problem:<br><ul><li>Ensure the model file has a compatible format and type<li>Check the model file is complete in the download folder<li>You can find the download folder in the settings dialog<li>If you've sideloaded the model ensure the file is not corrupt by checking md5sum<li>Read more about what models are supported in our <a href="https://docs.gpt4all.io/">documentation</a> for the gui<li>Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help<h3>Ocorreu um erro ao carregar o modelo:</h3><br><i>"%1"</i><br><br>Falhas no carregamento do modelo podem acontecer por vários motivos, mas as causas mais comuns incluem um formato de arquivo incorreto, um download incompleto ou corrompido, o tipo de arquivo errado, memória RAM do sistema insuficiente ou um tipo de modelo incompatível. Aqui estão algumas sugestões para resolver o problema:<br><ul><li>Certifique-se de que o arquivo do modelo tenha um formato e tipo compatíveis<li>Verifique se o arquivo do modelo está completo na pasta de download<li>Você pode encontrar a pasta de download na caixa de diálogo de configurações<li>Se você carregou o modelo, certifique-se de que o arquivo não esteja corrompido verificando o md5sum<li>Leia mais sobre quais modelos são suportados em nossa <a href="https://docs.gpt4all.io/">documentação</a> para a interface gráfica<li>Confira nosso <a href="https://discord.gg/4M2QFmTt2k">canal do Discord</a> para obter ajuda
-
-
+
+ Reload · %1Recarregar · %1
-
+ Loading · %1Carregando · %1
-
+ Load · %1 (default) →Carregar · %1 (padrão) →
-
+ restoring from text ...Recuperando do texto...
-
+ retrieving localdocs: %1 ...Recuperando dados em LocalDocs: %1 ...
-
+ searching localdocs: %1 ...Buscando em LocalDocs: %1 ...
-
+ %n Source(s)%n Origem
@@ -1029,42 +1029,42 @@ modelo instalado para funcionar
-
+ Send a message...Enviar uma mensagem...
-
+ Load a model to continue...Carregue um modelo para continuar...
-
+ Send messages/prompts to the modelEnviar mensagens/prompts para o modelo
-
+ CutRecortar
-
+ PasteColar
-
+ Select AllSelecionar tudo
-
+ Send messageEnviar mensagem
-
+ Sends the message/prompt contained in textfield to the modelEnvia a mensagem/prompt contida no campo de texto para o modelo
@@ -1072,12 +1072,12 @@ modelo instalado para funcionar
CollectionsDrawer
-
+ Warning: searching collections while indexing can return incomplete resultsAviso: pesquisar coleções durante a indexação pode retornar resultados incompletos
-
+ %n file(s)%n arquivo(s)
@@ -1085,7 +1085,7 @@ modelo instalado para funcionar
-
+ %n word(s)%n palavra(s)
@@ -1093,17 +1093,17 @@ modelo instalado para funcionar
-
+ UpdatingAtualizando
-
+ + Add Docs+ Adicionar Documentos
-
+ Select a collection to make it available to the chat model.Selecione uma coleção para disponibilizá-la ao modelo de chat.
@@ -1111,37 +1111,37 @@ modelo instalado para funcionar
Download
-
+ Model "%1" is installed successfully.Modelo "%1" instalado com sucesso.
-
+ ERROR: $MODEL_NAME is empty.ERRO: O nome do modelo ($MODEL_NAME) está vazio.
-
+ ERROR: $API_KEY is empty.ERRO: A chave da API ($API_KEY) está vazia.
-
+ ERROR: $BASE_URL is invalid.ERRO: A URL base ($BASE_URL) é inválida.
-
+ ERROR: Model "%1 (%2)" is conflict.ERRO: Conflito com o modelo "%1 (%2)".
-
+ Model "%1 (%2)" is installed successfully.Modelo "%1 (%2)" instalado com sucesso.
-
+ Model "%1" is removed.Modelo "%1" removido.
@@ -1149,92 +1149,92 @@ modelo instalado para funcionar
HomeView
-
+ Welcome to GPT4AllBem-vindo ao GPT4All
-
+ The privacy-first LLM chat applicationO aplicativo de chat LLM que prioriza a privacidade
-
+ Start chattingIniciar chat
-
+ Start ChattingIniciar Chat
-
+ Chat with any LLMConverse com qualquer LLM
-
+ LocalDocsLocalDocs
-
+ Chat with your local filesConverse com seus arquivos locais
-
+ Find ModelsEncontrar Modelos
-
+ Explore and download modelsDescubra e baixe modelos
-
+ Latest newsÚltimas novidades
-
+ Latest news from GPT4AllÚltimas novidades do GPT4All
-
+ Release NotesNotas de versão
-
+ DocumentationDocumentação
-
+ DiscordDiscord
-
+ X (Twitter)X (Twitter)
-
+ GithubGithub
-
+ nomic.ainomic.ai
-
+ Subscribe to NewsletterAssine nossa Newsletter
@@ -1242,118 +1242,118 @@ modelo instalado para funcionar
LocalDocsSettings
-
+ LocalDocsLocalDocs
-
+ LocalDocs SettingsConfigurações do LocalDocs
-
+ IndexingIndexação
-
+ Allowed File ExtensionsExtensões de Arquivo Permitidas
-
+ Comma-separated list. LocalDocs will only attempt to process files with these extensions.Lista separada por vírgulas. O LocalDocs tentará processar apenas arquivos com essas extensões.
-
+ EmbeddingIncorporação
-
+ Use Nomic Embed APIUsar a API Nomic Embed
-
+ Embed documents using the fast Nomic API instead of a private local model. Requires restart.Incorporar documentos usando a API Nomic rápida em vez de um modelo local privado. Requer reinicialização.
-
+ Nomic API KeyChave da API Nomic
-
+ API key to use for Nomic Embed. Get one from the Atlas <a href="https://atlas.nomic.ai/cli-login">API keys page</a>. Requires restart.Chave da API a ser usada para Nomic Embed. Obtenha uma na página de <a href="https://atlas.nomic.ai/cli-login">chaves de API do Atlas</a>. Requer reinicialização.
-
+ Embeddings DeviceProcessamento de Incorporações
-
+ The compute device used for embeddings. Requires restart.Dispositivo usado para processar as incorporações. Requer reinicialização.
-
+ Application defaultAplicativo padrão
-
+ DisplayExibir
-
+ Show SourcesMostrar Fontes
-
+ Display the sources used for each response.Mostra as fontes usadas para cada resposta.
-
+ AdvancedApenas para usuários avançados
-
+ Warning: Advanced usage only.Atenção: Apenas para usuários avançados.
-
+ Values too large may cause localdocs failure, extremely slow responses or failure to respond at all. Roughly speaking, the {N chars x N snippets} are added to the model's context window. More info <a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">here</a>.Valores muito altos podem causar falhas no LocalDocs, respostas extremamente lentas ou até mesmo nenhuma resposta. De forma geral, o valor {Número de Caracteres x Número de Trechos} é adicionado à janela de contexto do modelo. Clique <a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">aqui</a> para mais informações.
-
+ Document snippet size (characters)I translated "snippet" as "trecho" to make the term feel more natural and understandable in Portuguese. "Trecho" effectively conveys the idea of a portion or section of a document, fitting well within the context, whereas a more literal translation might sound less intuitive or awkward for users.Tamanho do trecho de documento (caracteres)
-
+ Number of characters per document snippet. Larger numbers increase likelihood of factual responses, but also result in slower generation.Número de caracteres por trecho de documento. Valores maiores aumentam a chance de respostas factuais, mas também tornam a geração mais lenta.
-
+ Max document snippets per promptMáximo de Trechos de Documento por Prompt
-
+ Max best N matches of retrieved document snippets to add to the context for prompt. Larger numbers increase likelihood of factual responses, but also result in slower generation.Número máximo de trechos de documentos a serem adicionados ao contexto do prompt. Valores maiores aumentam a chance de respostas factuais, mas também tornam a geração mais lenta.
@@ -1361,17 +1361,17 @@ modelo instalado para funcionar
LocalDocsView
-
+ LocalDocsLocalDocs
-
+ Chat with your local filesConverse com seus arquivos locais
-
+ + Add Collection+ Adicionar Coleção
@@ -1380,102 +1380,102 @@ modelo instalado para funcionar
ERRO: O banco de dados do LocalDocs não é válido.
-
+ <h3>ERROR: The LocalDocs database cannot be accessed or is not valid.</h3><br><i>Note: You will need to restart after trying any of the following suggested fixes.</i><br><ul><li>Make sure that the folder set as <b>Download Path</b> exists on the file system.</li><li>Check ownership as well as read and write permissions of the <b>Download Path</b>.</li><li>If there is a <b>localdocs_v2.db</b> file, check its ownership and read/write permissions, too.</li></ul><br>If the problem persists and there are any 'localdocs_v*.db' files present, as a last resort you can<br>try backing them up and removing them. You will have to recreate your collections, however.<h3>ERRO: Não foi possível acessar o banco de dados do LocalDocs ou ele não é válido.</h3><br><i>Observação: Será necessário reiniciar o aplicativo após tentar qualquer uma das seguintes correções sugeridas.</i><br><ul><li>Certifique-se de que a pasta definida como <b>Caminho de Download</b> existe no sistema de arquivos.</li><li>Verifique a propriedade, bem como as permissões de leitura e gravação do <b>Caminho de Download</b>.</li><li>Se houver um arquivo <b>localdocs_v2.db</b>, verifique também sua propriedade e permissões de leitura/gravação.</li></ul><br>Se o problema persistir e houver algum arquivo 'localdocs_v*.db' presente, como último recurso, você pode<br>tentar fazer backup deles e removê-los. No entanto, você terá que recriar suas coleções.
-
+ No Collections InstalledNenhuma Coleção Instalada
-
+ Install a collection of local documents to get started using this featureInstale uma coleção de documentos locais para começar a usar este recurso
-
+ + Add Doc Collection+ Adicionar Coleção de Documentos
-
+ Shows the add model viewMostra a visualização para adicionar modelo
-
+ Indexing progressBarBarra de progresso de indexação
-
+ Shows the progress made in the indexingMostra o progresso da indexação
-
+ ERRORERRO
-
+ INDEXINGINDEXANDO
-
+ EMBEDDINGINCORPORANDO
-
+ REQUIRES UPDATEREQUER ATUALIZAÇÃO
-
+ READYPRONTO
-
+ INSTALLINGINSTALANDO
-
+ Indexing in progressIndexação em andamento
-
+ Embedding in progressIncorporação em andamento
-
+ This collection requires an update after version changeEsta coleção precisa ser atualizada após a mudança de versão
-
+ Automatically reindexes upon changes to the folderReindexa automaticamente após alterações na pasta
-
+ Installation in progressInstalação em andamento
-
+ %%
-
+ %n file(s)%n arquivo(s)
@@ -1483,7 +1483,7 @@ modelo instalado para funcionar
-
+ %n word(s)%n palavra(s)
@@ -1491,27 +1491,27 @@ modelo instalado para funcionar
-
+ RemoveRemover
-
+ RebuildReconstruir
-
+ Reindex this folder from scratch. This is slow and usually not needed.eindexar pasta do zero. Lento e geralmente desnecessário.
-
+ UpdateAtualizar
-
+ Update the collection to the new version. This is a slow operation.Atualizar coleção para nova versão. Pode demorar.
@@ -1519,67 +1519,78 @@ modelo instalado para funcionar
ModelList
-
+ <ul><li>Requires personal OpenAI API key.</li><li>WARNING: Will send your chats to OpenAI!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with OpenAI</li><li>You can apply for an API key <a href="https://platform.openai.com/account/api-keys">here.</a></li><ul><li>É necessária uma chave de API da OpenAI.</li><li>AVISO: Seus chats serão enviados para a OpenAI!</li><li>Sua chave de API será armazenada localmente</li><li>Ela será usada apenas para comunicação com a OpenAI</li><li>Você pode solicitar uma chave de API <a href="https://platform.openai.com/account/api-keys">aqui.</a></li>
-
+ <strong>OpenAI's ChatGPT model GPT-3.5 Turbo</strong><br> %1<strong>Modelo ChatGPT GPT-3.5 Turbo da OpenAI</strong><br> %1
-
+ <strong>OpenAI's ChatGPT model GPT-4</strong><br> %1 %2<strong>Modelo ChatGPT GPT-4 da OpenAI</strong><br> %1 %2
-
+ <strong>Mistral Tiny model</strong><br> %1<strong>Modelo Mistral Tiny</strong><br> %1
-
+ <strong>Mistral Small model</strong><br> %1<strong>Modelo Mistral Small</strong><br> %1
-
+ <strong>Mistral Medium model</strong><br> %1<strong>Modelo Mistral Medium</strong><br> %1
-
+ <br><br><i>* Even if you pay OpenAI for ChatGPT-4 this does not guarantee API key access. Contact OpenAI for more info.<br><br><i>* Mesmo que você pague pelo ChatGPT-4 da OpenAI, isso não garante acesso à chave de API. Contate a OpenAI para mais informações.
-
+
+
+ cannot open "%1": %2
+
+
+
+
+ cannot create "%1": %2
+
+
+
+ %1 (%2)%1 (%2)
-
+ <strong>OpenAI-Compatible API Model</strong><br><ul><li>API Key: %1</li><li>Base URL: %2</li><li>Model Name: %3</li></ul><strong>Modelo de API Compatível com OpenAI</strong><br><ul><li>Chave da API: %1</li><li>URL Base: %2</li><li>Nome do Modelo: %3</li></ul>
-
+ <ul><li>Requires personal Mistral API key.</li><li>WARNING: Will send your chats to Mistral!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with Mistral</li><li>You can apply for an API key <a href="https://console.mistral.ai/user/api-keys">here</a>.</li><ul><li>É necessária uma chave de API da Mistral.</li><li>AVISO: Seus chats serão enviados para a Mistral!</li><li>Sua chave de API será armazenada localmente</li><li>Ela será usada apenas para comunicação com a Mistral</li><li>Você pode solicitar uma chave de API <a href="https://console.mistral.ai/user/api-keys">aqui</a>.</li>
-
+ <ul><li>Requires personal API key and the API base URL.</li><li>WARNING: Will send your chats to the OpenAI-compatible API Server you specified!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with the OpenAI-compatible API Server</li><ul><li>É necessária uma chave de API e a URL da API.</li><li>AVISO: Seus chats serão enviados para o servidor de API compatível com OpenAI que você especificou!</li><li>Sua chave de API será armazenada no disco</li><li>Será usada apenas para comunicação com o servidor de API compatível com OpenAI</li>
-
+ <strong>Connect to OpenAI-compatible API server</strong><br> %1<strong>Conectar a um servidor de API compatível com OpenAI</strong><br> %1
-
+ <strong>Created by %1.</strong><br><ul><li>Published on %2.<li>This model has %3 likes.<li>This model has %4 downloads.<li>More info can be found <a href="https://huggingface.co/%5">here.</a></ul><strong>Criado por %1.</strong><br><ul><li>Publicado em %2.<li>Este modelo tem %3 curtidas.<li>Este modelo tem %4 downloads.<li>Mais informações podem ser encontradas <a href="https://huggingface.co/%5">aqui.</a></ul>
@@ -1587,92 +1598,92 @@ modelo instalado para funcionar
ModelSettings
-
+ ModelModelo
-
+ Model SettingsConfigurações do Modelo
-
+ CloneClonar
-
+ RemoveRemover
-
+ NameNome
-
+ Model FileArquivo do Modelo
-
+ System PromptPrompt do Sistema
-
+ Prefixed at the beginning of every conversation. Must contain the appropriate framing tokens.Prefixado no início de cada conversa. Deve conter os tokens de enquadramento apropriados.
-
+ Prompt TemplateModelo de Prompt
-
+ The template that wraps every prompt.Modelo para cada prompt.
-
+ Must contain the string "%1" to be replaced with the user's input.Deve incluir "%1" para a entrada do usuário.
-
+ Chat Name PromptPrompt para Nome do Chat
-
+ Prompt used to automatically generate chat names.Prompt usado para gerar automaticamente nomes de chats.
-
+ Suggested FollowUp PromptPrompt de Sugestão de Acompanhamento
-
+ Prompt used to generate suggested follow-up questions.Prompt usado para gerar sugestões de perguntas.
-
+ Context LengthTamanho do Contexto
-
+ Number of input and output tokens the model sees.Tamanho da Janela de Contexto.
-
+ Maximum combined prompt/response tokens before information is lost.
Using more context than the model was trained on will yield poor results.
NOTE: Does not take effect until you reload the model.
@@ -1681,128 +1692,128 @@ Usar mais contexto do que o modelo foi treinado pode gerar resultados ruins.
Obs.: Só entrará em vigor após recarregar o modelo.
-
+ TemperatureTemperatura
-
+ Randomness of model output. Higher -> more variation.Aleatoriedade das respostas. Quanto maior, mais variadas.
-
+ Temperature increases the chances of choosing less likely tokens.
NOTE: Higher temperature gives more creative but less predictable outputs.Aumenta a chance de escolher tokens menos prováveis.
Obs.: Uma temperatura mais alta gera resultados mais criativos, mas menos previsíveis.
-
+ Top-PTop-P
-
+ Nucleus Sampling factor. Lower -> more predictable.Amostragem por núcleo. Menor valor, respostas mais previsíveis.
-
+ Only the most likely tokens up to a total probability of top_p can be chosen.
NOTE: Prevents choosing highly unlikely tokens.Apenas tokens com probabilidade total até o valor de top_p serão escolhidos.
Obs.: Evita tokens muito improváveis.
-
+ Min-PMin-P
-
+ Minimum token probability. Higher -> more predictable.Probabilidade mínima do token. Quanto maior -> mais previsível.
-
+ Sets the minimum relative probability for a token to be considered.Define a probabilidade relativa mínima para um token ser considerado.
-
+ Top-KTop-K
-
+ Size of selection pool for tokens.Número de tokens considerados na amostragem.
-
+ Only the top K most likely tokens will be chosen from.Serão escolhidos apenas os K tokens mais prováveis.
-
+ Max LengthComprimento Máximo
-
+ Maximum response length, in tokens.Comprimento máximo da resposta, em tokens.
-
+ Prompt Batch SizeTamanho do Lote de Processamento
-
+ The batch size used for prompt processing.Tokens processados por lote.
-
+ Amount of prompt tokens to process at once.
NOTE: Higher values can speed up reading prompts but will use more RAM.Quantidade de tokens de prompt para processar de uma vez.
OBS.: Valores mais altos podem acelerar a leitura dos prompts, mas usarão mais RAM.
-
+ Repeat PenaltyPenalidade de Repetição
-
+ Repetition penalty factor. Set to 1 to disable.Penalidade de Repetição (1 para desativar).
-
+ Repeat Penalty TokensTokens para penalizar repetição
-
+ Number of previous tokens used for penalty.Número de tokens anteriores usados para penalidade.
-
+ GPU LayersCamadas na GPU
-
+ Number of model layers to load into VRAM.Camadas Carregadas na GPU.
-
+ How many model layers to load into VRAM. Decrease this if GPT4All runs out of VRAM while loading this model.
Lower values increase CPU load and RAM usage, and make inference slower.
NOTE: Does not take effect until you reload the model.
@@ -1814,217 +1825,217 @@ Obs.: Só entrará em vigor após recarregar o modelo.
ModelsView
-
+ No Models InstalledNenhum Modelo Instalado
-
+ Install a model to get started using GPT4AllInstale um modelo para começar a usar o GPT4All
-
-
+
+ + Add Model+ Adicionar Modelo
-
+ Shows the add model viewMostra a visualização para adicionar modelo
-
+ Installed ModelsModelos Instalados
-
+ Locally installed chat modelsModelos de chat instalados localmente
-
+ Model fileArquivo do modelo
-
+ Model file to be downloadedArquivo do modelo a ser baixado
-
+ DescriptionDescrição
-
+ File descriptionDescrição do arquivo
-
+ CancelCancelar
-
+ ResumeRetomar
-
+ Stop/restart/start the downloadParar/reiniciar/iniciar o download
-
+ RemoveRemover
-
+ Remove model from filesystemRemover modelo do sistema de arquivos
-
-
+
+ InstallInstalar
-
+ Install online modelInstalar modelo online
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font><strong><font size="1"><a href="#error">Erro</a></strong></font>
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font><strong><font size="2">AVISO: Não recomendado para seu hardware. O modelo requer mais memória (%1 GB) do que seu sistema tem disponível (%2).</strong></font>
-
+ ERROR: $API_KEY is empty.ERRO: A $API_KEY está vazia.
-
+ ERROR: $BASE_URL is empty.ERRO: A $BASE_URL está vazia.
-
+ enter $BASE_URLinserir a $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.ERRO: O $MODEL_NAME está vazio.
-
+ enter $MODEL_NAMEinserir o $MODEL_NAME
-
+ %1 GB%1 GB
-
+ ??
-
+ Describes an error that occurred when downloadingDescreve um erro que ocorreu durante o download
-
+ Error for incompatible hardwareErro para hardware incompatível
-
+ Download progressBarBarra de progresso do download
-
+ Shows the progress made in the downloadMostra o progresso do download
-
+ Download speedVelocidade de download
-
+ Download speed in bytes/kilobytes/megabytes per secondVelocidade de download em bytes/kilobytes/megabytes por segundo
-
+ Calculating...Calculando...
-
-
-
-
+
+
+
+ Whether the file hash is being calculatedSe o hash do arquivo está sendo calculado
-
+ Busy indicatorIndicador de ocupado
-
+ Displayed when the file hash is being calculatedExibido quando o hash do arquivo está sendo calculado
-
+ enter $API_KEYinserir $API_KEY
-
+ File sizeTamanho do arquivo
-
+ RAM requiredRAM necessária
-
+ ParametersParâmetros
-
+ QuantQuant
-
+ TypeTipo
@@ -2032,12 +2043,12 @@ Obs.: Só entrará em vigor após recarregar o modelo.
MyFancyLink
-
+ Fancy linkLink personalizado
-
+ A stylized linkUm link personalizado
@@ -2045,7 +2056,7 @@ Obs.: Só entrará em vigor após recarregar o modelo.
MySettingsStack
-
+ Please choose a directoryEscolha um diretório
@@ -2053,12 +2064,12 @@ Obs.: Só entrará em vigor após recarregar o modelo.
MySettingsTab
-
+ Restore DefaultsRestaurar Configurações Padrão
-
+ Restores settings dialog to a default stateRestaura as configurações para o estado padrão
@@ -2066,12 +2077,12 @@ Obs.: Só entrará em vigor após recarregar o modelo.
NetworkDialog
-
+ Contribute data to the GPT4All Opensource Datalake.Contribuir com dados para o Datalake de código aberto GPT4All.
-
+ By enabling this feature, you will be able to participate in the democratic process of training a large language model by contributing data for future model improvements.
When a GPT4All model responds to you and you have opted-in, your conversation will be sent to the GPT4All Open Source Datalake. Additionally, you can like/dislike its response. If you dislike a response, you can suggest an alternative response. This data will be collected and aggregated in the GPT4All Datalake.
@@ -2084,47 +2095,47 @@ Quando um modelo GPT4All responder a você e você tiver optado por participar,
OBS.: Ao ativar este recurso, você estará enviando seus dados para o Datalake de Código Aberto do GPT4All. Você não deve ter nenhuma expectativa de privacidade no chat quando este recurso estiver ativado. No entanto, você deve ter a expectativa de uma atribuição opcional, se desejar. Seus dados de chat estarão disponíveis para qualquer pessoa baixar e serão usados pela Nomic AI para melhorar os futuros modelos GPT4All. A Nomic AI manterá todas as informações de atribuição anexadas aos seus dados e você será creditado como colaborador em qualquer versão do modelo GPT4All que utilize seus dados!
-
+ Terms for opt-inTermos de participação
-
+ Describes what will happen when you opt-inDescrição do que acontece ao participar
-
+ Please provide a name for attribution (optional)Forneça um nome para atribuição (opcional)
-
+ Attribution (optional)Atribuição (opcional)
-
+ Provide attributionFornecer atribuição
-
+ EnableHabilitar
-
+ Enable opt-inAtivar participação
-
+ CancelCancelar
-
+ Cancel opt-inCancelar participação
@@ -2132,17 +2143,17 @@ OBS.: Ao ativar este recurso, você estará enviando seus dados para o Datalake
NewVersionDialog
-
+ New version is availableAtualização disponível
-
+ UpdateAtualizar agora
-
+ Update to new versionBaixa e instala a última versão do GPT4All
@@ -2150,18 +2161,18 @@ OBS.: Ao ativar este recurso, você estará enviando seus dados para o Datalake
PopupDialog
-
+ Reveals a shortlived help balloonExibe uma dica rápida
-
+ Busy indicatorThe literal translation of "busy indicator" as "indicador de ocupado" might create ambiguity in Portuguese, as it doesn't clearly convey whether the system is processing something or simply unavailable. "Progresso" (progress) was chosen to more clearly indicate that an activity is in progress and that the user should wait for its completion.Indicador de progresso
-
+ Displayed when the popup is showing busyVisível durante o processamento
@@ -2176,29 +2187,29 @@ OBS.: Ao ativar este recurso, você estará enviando seus dados para o Datalake
SettingsView
-
-
+
+ SettingsI used "Config" instead of "Configurações" to keep the UI concise and visually balanced. "Config" is a widely recognized abbreviation that maintains clarity while saving space, making the interface cleaner and more user-friendly, especially in areas with limited space.Config
-
+ Contains various application settingsAcessar as configurações do aplicativo
-
+ ApplicationAplicativo
-
+ ModelModelo
-
+ LocalDocsLocalDocs
@@ -2206,12 +2217,12 @@ OBS.: Ao ativar este recurso, você estará enviando seus dados para o Datalake
StartupDialog
-
+ Welcome!Bem-vindo(a)!
-
+ ### Release notes
%1### Contributors
%2
@@ -2220,17 +2231,17 @@ OBS.: Ao ativar este recurso, você estará enviando seus dados para o Datalake
%2
-
+ Release notesNotas de lançamento
-
+ Release notes for this versionNotas de lançamento desta versão
-
+ ### Opt-ins for anonymous usage analytics and datalake
By enabling these features, you will be able to participate in the democratic process of training a
large language model by contributing data for future model improvements.
@@ -2261,71 +2272,71 @@ todas as informações de atribuição anexadas aos seus dados e você será cre
versão do modelo GPT4All que utilize seus dados!
-
+ Terms for opt-inTermos de participação
-
+ Describes what will happen when you opt-inDescrição do que acontece ao participar
-
-
+
+ Opt-in for anonymous usage statisticsEnviar estatísticas de uso anônimas
-
-
+
+ YesSim
-
+ Allow opt-in for anonymous usage statisticsPermitir o envio de estatísticas de uso anônimas
-
-
+
+ NoNão
-
+ Opt-out for anonymous usage statisticsRecusar envio de estatísticas de uso anônimas
-
+ Allow opt-out for anonymous usage statisticsPermitir recusar envio de estatísticas de uso anônimas
-
-
+
+ Opt-in for networkAceitar na rede
-
+ Allow opt-in for networkPermitir aceitação na rede
-
+ Allow opt-in anonymous sharing of chats to the GPT4All DatalakePermitir compartilhamento anônimo de chats no Datalake GPT4All
-
+ Opt-out for networkRecusar na rede
-
+ Allow opt-out anonymous sharing of chats to the GPT4All DatalakePermitir recusar compartilhamento anônimo de chats no Datalake GPT4All
@@ -2333,23 +2344,23 @@ versão do modelo GPT4All que utilize seus dados!
SwitchModelDialog
-
+ <b>Warning:</b> changing the model will erase the current conversation. Do you wish to continue?<b>Atenção:</b> Ao trocar o modelo a conversa atual será perdida. Continuar?
-
+ ContinueContinuar
-
+ Continue with model loadingConfirma a troca do modelo
-
-
+
+ CancelCancelar
@@ -2357,32 +2368,32 @@ versão do modelo GPT4All que utilize seus dados!
ThumbsDownDialog
-
+ Please edit the text below to provide a better response. (optional)Editar resposta (opcional)
-
+ Please provide a better response...Digite sua resposta...
-
+ SubmitEnviar
-
+ Submits the user's responseEnviar
-
+ CancelCancelar
-
+ Closes the response dialogFecha a caixa de diálogo de resposta
@@ -2390,125 +2401,125 @@ versão do modelo GPT4All que utilize seus dados!
main
-
+ <h3>Encountered an error starting up:</h3><br><i>"Incompatible hardware detected."</i><br><br>Unfortunately, your CPU does not meet the minimal requirements to run this program. In particular, it does not support AVX intrinsics which this program requires to successfully run a modern large language model. The only solution at this time is to upgrade your hardware to a more modern CPU.<br><br>See here for more information: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a><h3>Ocorreu um erro ao iniciar:</h3><br><i>"Hardware incompatível detectado."</i><br><br>Infelizmente, seu processador não atende aos requisitos mínimos para executar este programa. Especificamente, ele não possui suporte às instruções AVX, que são necessárias para executar modelos de linguagem grandes e modernos. A única solução, no momento, é atualizar seu hardware para um processador mais recente.<br><br>Para mais informações, consulte: <a href="https://pt.wikipedia.org/wiki/Advanced_Vector_Extensions">https://pt.wikipedia.org/wiki/Advanced_Vector_Extensions</a>
-
+ GPT4All v%1GPT4All v%1
-
+ <h3>Encountered an error starting up:</h3><br><i>"Inability to access settings file."</i><br><br>Unfortunately, something is preventing the program from accessing the settings file. This could be caused by incorrect permissions in the local app config directory where the settings file is located. Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help.<h3>Ocorreu um erro ao iniciar:</h3><br><i>"Não foi possível acessar o arquivo de configurações."</i><br><br>Infelizmente, algo está impedindo o programa de acessar o arquivo de configurações. Isso pode acontecer devido a permissões incorretas na pasta de configurações do aplicativo. Para obter ajuda, acesse nosso <a href="https://discord.gg/4M2QFmTt2k">canal no Discord</a>.
-
+ Connection to datalake failed.Falha na conexão com o datalake.
-
+ Saving chats.Salvando chats.
-
+ Network dialogAvisos de rede
-
+ opt-in to share feedback/conversationspermitir compartilhamento de feedback/conversas
-
+ Home viewTela inicial
-
+ Home view of applicationTela inicial do aplicativo
-
+ HomeInício
-
+ Chat viewVisualização do Chat
-
+ Chat view to interact with modelsVisualização do chat para interagir com os modelos
-
+ ChatsChats
-
-
+
+ ModelsModelos
-
+ Models view for installed modelsTela de modelos instalados
-
-
+
+ LocalDocsLocalDocs
-
+ LocalDocs view to configure and use local docsTela de configuração e uso de documentos locais do LocalDocs
-
-
+
+ SettingsConfig
-
+ Settings view for application configurationTela de configurações do aplicativo
-
+ The datalake is enabledO datalake está ativado
-
+ Using a network modelUsando um modelo de rede
-
+ Server mode is enabledModo servidor ativado
-
+ Installed modelsModelos instalados
-
+ View of installed modelsExibe os modelos instalados
diff --git a/gpt4all-chat/translations/gpt4all_ro_RO.ts b/gpt4all-chat/translations/gpt4all_ro_RO.ts
index 703e5b14..a1eefb88 100644
--- a/gpt4all-chat/translations/gpt4all_ro_RO.ts
+++ b/gpt4all-chat/translations/gpt4all_ro_RO.ts
@@ -4,12 +4,12 @@
AddCollectionView
-
+ ← Existing Collections← Colecţiile curente
-
+ Add Document CollectionAdaugă o Colecţie de documente
@@ -19,52 +19,52 @@
Adaugă un folder care conţine fişiere în cu text-simplu, PDF sau Markdown. Extensii suplimentare pot fi specificate în Configurare.
-
+ Add a folder containing plain text files, PDFs, or Markdown. Configure additional extensions in Settings.Adaugă un folder cu fişiere în format text, PDF sau Markdown. Alte extensii pot fi specificate în Configurare.
-
+ Please choose a directorySelectează un director/folder
-
+ NameDenumire
-
+ Collection name...Denumirea Colecţiei...
-
+ Name of the collection to add (Required)Denumirea Colecţiei de adăugat (necesar)
-
+ FolderFolder
-
+ Folder path...Calea spre folder...
-
+ Folder path to documents (Required)Calea spre documente (necesar)
-
+ BrowseCăutare
-
+ Create CollectionCreează Colecţia
@@ -72,174 +72,174 @@
AddModelView
-
+ ← Existing Models← Modelele curente/instalate
-
+ Explore ModelsCaută modele
-
+ Discover and download models by keyword search...Caută şi descarcă modele după un cuvânt-cheie...
-
+ Text field for discovering and filtering downloadable modelsCâmp pentru căutarea şi filtrarea modelelor ce pot fi descărcate
-
+ Initiate model discovery and filteringIniţiază căutarea şi filtrarea modelelor
-
+ Triggers discovery and filtering of modelsActivează căutarea şi filtrarea modelelor
-
+ DefaultImplicit
-
+ LikesLikes
-
+ DownloadsDownload-uri
-
+ RecentRecent/e
-
+ AscAsc. (A->Z)
-
+ DescDesc. (Z->A)
-
+ NoneNiciunul
-
+ Searching · %1Căutare · %1
-
+ Sort by: %1Ordonare după: %1
-
+ Sort dir: %1Sensul ordonării: %1
-
+ Limit: %1Límită: %1
-
+ Network error: could not retrieve %1Eroare de reţea: nu se poate prelua %1
-
-
+
+ Busy indicatorIndicator de activitate
-
+ Displayed when the models request is ongoingAfişat în timpul solicitării modelului
-
+ Model fileFişierul modelului
-
+ Model file to be downloadedFişierul modelului de descărcat
-
+ DescriptionDescriere
-
+ File descriptionDescrierea fişierului
-
+ CancelAnulare
-
+ ResumeContinuare
-
+ DownloadDownload
-
+ Stop/restart/start the downloadOpreşte/Reporneşte/Începe descărcarea
-
+ RemoveŞterge
-
+ Remove model from filesystemŞterge modelul din sistemul de fişiere
-
-
+
+ InstallInstalare
-
+ Install online modelInstalez un model din online
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font><strong><font size="1"><a href="#error">Eroare</a></strong></font>
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font><strong><font size="2">ATENŢIE: Nerecomandat pentru acest hardware. Modelul necesită mai multă memorie (%1 GB) decât are acest sistem (%2).</strong></font>
@@ -250,18 +250,18 @@
<strong><font size="2">ATENţIE: Nerecomandat pentru acest hardware. Modelul necesită mai multă memorie (%1 GB) decât are acest sistem (%2).</strong></font>
-
+ %1 GB%1 GB
-
-
+
+ ??
-
+ Describes an error that occurred when downloadingDescrie eroarea apărută în timpul descărcării
@@ -271,100 +271,100 @@
<strong><font size="1"><a href="#eroare">Eroare</a></strong></font>
-
+ Error for incompatible hardwareEroare: hardware incompatibil
-
+ Download progressBarProgresia descărcării
-
+ Shows the progress made in the downloadAfişează progresia descărcării
-
+ Download speedViteza de download
-
+ Download speed in bytes/kilobytes/megabytes per secondViteza de download în bytes/kilobytes/megabytes pe secundă
-
+ Calculating...Calculare...
-
-
-
-
+
+
+
+ Whether the file hash is being calculatedDacă se calculează hash-ul fişierului
-
+ Displayed when the file hash is being calculatedSe afişează când se calculează hash-ul fişierului
-
+ ERROR: $API_KEY is empty.EROARE: $API_KEY absentă
-
+ enter $API_KEYintrodu cheia $API_KEY
-
+ ERROR: $BASE_URL is empty.EROARE: $BASE_URL absentă
-
+ enter $BASE_URLintrodu $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.EROARE: $MODEL_NAME absent
-
+ enter $MODEL_NAMEintrodu $MODEL_NAME
-
+ File sizeDimensiunea fişierului
-
+ RAM requiredRAM necesară
-
+ ParametersParametri
-
+ QuantQuant(ificare)
-
+ TypeTip
@@ -372,17 +372,17 @@
ApplicationSettings
-
+ ApplicationAplicaţie/Program
-
+ Network dialogReţea
-
+ opt-in to share feedback/conversationsoptional: partajarea (share) de comentarii/conversatii
@@ -398,57 +398,57 @@
EROARE: Sistemul de actualizare nu poate găsi componenta MaintenanceTool<br> necesară căutării de versiuni noi!<br><br> Ai instalat acest program folosind kitul online? Dacă da,<br> atunci MaintenanceTool trebuie să fie un nivel mai sus de folderul<br> unde ai instalat programul.<br><br> Dacă nu poate fi lansată manual, atunci programul trebuie reinstalat.
-
+ Error dialogEroare
-
+ Application SettingsConfigurarea programului
-
+ GeneralGeneral
-
+ ThemeTema pentru interfaţă
-
+ The application color scheme.Schema de culori a programului.
-
+ DarkÎntunecat
-
+ LightLuminos
-
+ LegacyDarkÎntunecat-vechi
-
+ Font SizeDimensiunea textului
-
+ The size of text in the application.Dimensiunea textului în program.
-
+ DeviceDispozitiv/Device
@@ -458,7 +458,7 @@
Dispozitivul de calcul utilizat pentru generarea de text. "Auto" apelează la Vulkan sau la Metal.
-
+ ERROR: Update system could not find the MaintenanceTool used<br>
to check for updates!<br><br>
Did you install this application using the online installer? If so,<br>
@@ -469,138 +469,138 @@
EROARE: Sistemul de Update nu poate găsi componenta MaintenanceTool<br> necesară căutării de versiuni noi!<br><br> Ai instalat acest program folosind kitul online? Dacă da,<br> atunci MaintenanceTool trebuie să fie un nivel mai sus de folderul<br> unde ai instalat programul.<br><br> Dacă nu poate fi lansată manual, atunci programul trebuie reinstalat.
-
+ SmallMic
-
+ MediumMediu
-
+ LargeMare
-
+ Language and LocaleLimbă şi Localizare
-
+ The language and locale you wish to use.Limba şi Localizarea de utilizat.
-
+ System LocaleLocalizare
-
+ The compute device used for text generation.Dispozitivul de calcul utilizat pentru generarea de text.
-
-
+
+ Application defaultImplicit
-
+ Default ModelModelul implicit
-
+ The preferred model for new chats. Also used as the local server fallback.Modelul preferat pentru noile conversaţii. Va fi folosit drept rezervă pentru serverul local.
-
+ Suggestion ModeModul de sugerare
-
+ Generate suggested follow-up questions at the end of responses.Generarea de întrebări pentru continuare, la finalul replicilor.
-
+ When chatting with LocalDocsCând se discută cu LocalDocs
-
+ Whenever possibleOricând e posibil
-
+ NeverNiciodată
-
+ Download PathCalea pentru download
-
+ Where to store local models and the LocalDocs database.Unde să fie plasate modelele şi baza de date LocalDocs.
-
+ BrowseCăutare
-
+ Choose where to save model filesSelectează locul unde vor fi plasate fişierele modelelor
-
+ Enable DatalakeActivează DataLake
-
+ Send chats and feedback to the GPT4All Open-Source Datalake.Trimite conversaţii şi comentarii către componenta Open-source DataLake a GPT4All.
-
+ AdvancedAvansate
-
+ CPU ThreadsThread-uri CPU
-
+ The number of CPU threads used for inference and embedding.Numărul de thread-uri CPU utilizate pentru inferenţă şi embedding.
-
+ Save Chat ContextSalvarea contextului conversaţiei
-
+ Save the chat model's state to disk for faster loading. WARNING: Uses ~2GB per chat.Salvează pe disc starea modelului pentru încărcare mai rapidă. ATENŢIE: Consumă ~2GB/conversaţie.
-
+ Expose an OpenAI-Compatible server to localhost. WARNING: Results in increased resource usage.Activează pe localhost un Server compatibil cu Open-AI. ATENŢIE: Creşte consumul de resurse.
@@ -610,7 +610,7 @@
Salvează pe disc starea modelului pentru încărcare mai rapidă. ATENŢIE: Consumă ~2GB/conversaţie.
-
+ Enable Local ServerActivează Serverul local
@@ -620,27 +620,27 @@
Activează pe localhost un Server compatibil cu Open-AI. ATENŢIE: Creşte consumul de resurse.
-
+ API Server PortPortul Serverului API
-
+ The port to use for the local server. Requires restart.Portul utilizat pentru Serverul local. Necesită repornirea programului.
-
+ Check For UpdatesCaută update-uri
-
+ Manually check for an update to GPT4All.Caută manual update-uri pentru GPT4All.
-
+ UpdatesUpdate-uri/Actualizări
@@ -648,13 +648,13 @@
Chat
-
-
+
+ New ChatConversaţie Nouă
-
+ Server ChatConversaţie cu Serverul
@@ -662,12 +662,12 @@
ChatAPIWorker
-
+ ERROR: Network error occurred while connecting to the API serverEROARE: Eroare de reţea - conectarea la serverul API
-
+ ChatAPIWorker::handleFinished got HTTP Error %1 %2ChatAPIWorker::handleFinished - eroare: HTTP Error %1 %2
@@ -675,62 +675,62 @@
ChatDrawer
-
+ DrawerSertar
-
+ Main navigation drawerSertarul principal de navigare
-
+ + New Chat+ Conversaţie nouă
-
+ Create a new chatCreează o Conversaţie nouă
-
+ Select the current chat or edit the chat when in edit modeSelectează conversaţia curentă sau editeaz-o când eşti în modul editare
-
+ Edit chat nameEditează denumirea conversaţiei
-
+ Save chat nameSalvează denumirea conversaţiei
-
+ Delete chatŞterge conversaţia
-
+ Confirm chat deletionCONFIRMĂ ştergerea conversaţiei
-
+ Cancel chat deletionANULEAZĂ ştergerea conversaţiei
-
+ List of chatsLista conversaţiilor
-
+ List of chats in the drawer dialogLista conversaţiilor în secţiunea-sertar
@@ -738,32 +738,32 @@
ChatListModel
-
+ TODAYASTĂZI
-
+ THIS WEEKSĂPTĂMÂNA ACEASTA
-
+ THIS MONTHLUNA ACEASTA
-
+ LAST SIX MONTHSULTIMELE ŞASE LUNI
-
+ THIS YEARANUL ACESTA
-
+ LAST YEARANUL TRECUT
@@ -771,113 +771,113 @@
ChatView
-
+ <h3>Warning</h3><p>%1</p><h3>Atenţie</h3><p>%1</p>
-
+ Switch model dialogSchimbarea modelului
-
+ Warn the user if they switch models, then context will be erasedAvertizează utilizatorul că la schimbarea modelului va fi şters contextul
-
+ Conversation copied to clipboard.Conversaţia a fost plasată în Clipboard.
-
+ Code copied to clipboard.Codul a fost plasat în Clipboard.
-
+ Chat panelSecţiunea de chat
-
+ Chat panel with optionsSecţiunea de chat cu opţiuni
-
+ Reload the currently loaded modelReîncarcă modelul curent
-
+ Eject the currently loaded modelEjectează modelul curent
-
+ No model installed.Niciun model instalat.
-
+ Model loading error.Eroare la încărcarea modelului.
-
+ Waiting for model...Se aşteaptă modelul...
-
+ Switching context...Se schimbă contextul...
-
+ Choose a model...Selectează un model...
-
+ Not found: %1Absent: %1
-
+ The top item is the current modelPrimul element e modelul curent
-
-
+
+ LocalDocsLocalDocs
-
+ Add documentsAdaug documente
-
+ add collections of documents to the chatadaugă Colecţii de documente la conversaţie
-
+ Load the default modelÎncarcă modelul implicit
-
+ Loads the default model which can be changed in settingsÎncarcă modelul implicit care poate fi stabilit în Configurare
-
+ No Model InstalledNiciun model instalat
@@ -887,7 +887,7 @@
GPT4All necesită cel puţin un model pentru a putea porni
-
+ <h3>Encountered an error loading model:</h3><br><i>"%1"</i><br><br>Model loading failures can happen for a variety of reasons, but the most common causes include a bad file format, an incomplete or corrupted download, the wrong file type, not enough system RAM or an incompatible model type. Here are some suggestions for resolving the problem:<br><ul><li>Ensure the model file has a compatible format and type<li>Check the model file is complete in the download folder<li>You can find the download folder in the settings dialog<li>If you've sideloaded the model ensure the file is not corrupt by checking md5sum<li>Read more about what models are supported in our <a href="https://docs.gpt4all.io/">documentation</a> for the gui<li>Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help<h3>EROARE la încărcarea
modelului:</h3><br><i>"%1"</i><br><br>Astfel
@@ -906,38 +906,38 @@
se oferă ajutor
-
+ GPT4All requires that you install at least one
model to get startedGPT4All necesită cel puţin un model pentru a putea rula
-
+ Install a ModelInstalează un model
-
+ Shows the add model viewAfişează secţiunea de adăugare a unui model
-
+ Conversation with the modelConversaţie cu modelul
-
+ prompt / response pairs from the conversationperechi prompt/replică din conversaţie
-
+ GPT4AllGPT4All
-
+ YouTu
@@ -946,98 +946,98 @@ model to get started
se recalculează contextul...
-
+ response stopped ...replică întreruptă...
-
+ processing ...procesare...
-
+ generating response ...se generează replica...
-
+ generating questions ...se generează întrebări...
-
-
+
+ CopyCopiere
-
+ Copy MessageCopiez mesajul
-
+ Disable markdownDezactivez markdown
-
+ Enable markdownActivez markdown
-
+ Thumbs upBravo
-
+ Gives a thumbs up to the responseDă un Bravo acestei replici
-
+ Thumbs downAiurea
-
+ Opens thumbs down dialogDeschide reacţia Aiurea
-
+ Suggested follow-upsContinuări sugerate
-
+ Erase and reset chat sessionŞterge şi resetează sesiunea de chat
-
+ Copy chat session to clipboardCopiez sesiunea de chat (conversaţia) în Clipboard
-
+ Redo last chat responseReface ultima replică
-
+ Stop generatingOpreşte generarea
-
+ Stop the current response generationOpreşte generarea replicii curente
-
+ Reloads the modelReîncarc modelul
@@ -1072,38 +1072,38 @@ model to get started
se oferă ajutor
-
-
+
+ Reload · %1Reîncărcare · %1
-
+ Loading · %1Încărcare · %1
-
+ Load · %1 (default) →Încarcă · %1 (implicit) →
-
+ restoring from text ...restaurare din text...
-
+ retrieving localdocs: %1 ...se preia din LocalDocs: %1 ...
-
+ searching localdocs: %1 ...se caută în LocalDocs: %1 ...
-
+ %n Source(s)%n Sursa
@@ -1112,42 +1112,42 @@ model to get started
-
+ Send a message...Trimite un mesaj...
-
+ Load a model to continue...Încarcă un model pentru a continua...
-
+ Send messages/prompts to the modelTrimite mesaje/prompt-uri către model
-
+ CutDecupare (Cut)
-
+ PasteAlipire (Paste)
-
+ Select AllSelectez tot
-
+ Send messageTrimit mesajul
-
+ Sends the message/prompt contained in textfield to the modelTrimite modelului mesajul/prompt-ul din câmpul-text
@@ -1155,12 +1155,12 @@ model to get started
CollectionsDrawer
-
+ Warning: searching collections while indexing can return incomplete resultsAtenţie: căutarea în Colecţii în timp ce sunt Indexate poate întoarce rezultate incomplete
-
+ %n file(s)%n fişier
@@ -1169,7 +1169,7 @@ model to get started
-
+ %n word(s)%n cuvânt
@@ -1178,17 +1178,17 @@ model to get started
-
+ UpdatingActualizare
-
+ + Add Docs+ Adaug documente
-
+ Select a collection to make it available to the chat model.Selectează o Colecţie pentru ca modelul să o poată accesa.
@@ -1196,37 +1196,37 @@ model to get started
Download
-
+ Model "%1" is installed successfully.Modelul "%1" - instalat cu succes.
-
+ ERROR: $MODEL_NAME is empty.EROARE: $MODEL_NAME absent.
-
+ ERROR: $API_KEY is empty.EROARE: $API_KEY absentă
-
+ ERROR: $BASE_URL is invalid.EROARE: $API_KEY incorecta
-
+ ERROR: Model "%1 (%2)" is conflict.EROARE: Model "%1 (%2)" conflictual.
-
+ Model "%1 (%2)" is installed successfully.Modelul "%1 (%2)" - instalat cu succes.
-
+ Model "%1" is removed.Modelul "%1" - îndepărtat
@@ -1234,87 +1234,87 @@ model to get started
HomeView
-
+ Welcome to GPT4AllBun venit în GPT4All
-
+ The privacy-first LLM chat applicationProgramul ce prioritizează confidenţialitatea (privacy)
-
+ Start chattingÎncepe o conversaţie
-
+ Start ChattingÎncepe o conversaţie
-
+ Chat with any LLMDialoghează cu orice LLM
-
+ LocalDocsLocalDocs
-
+ Chat with your local filesDialoghează cu fişiere locale
-
+ Find ModelsCaută modele
-
+ Explore and download modelsExplorează şi descarcă modele
-
+ Latest newsUltimele ştiri
-
+ Latest news from GPT4AllUltimele ştiri de la GPT4All
-
+ Release NotesDespre această versiune
-
+ DocumentationDocumentaţie
-
+ DiscordDiscord
-
+ X (Twitter)X (Twitter)
-
+ GithubGitHub
-
+ nomic.ainomic.ai
@@ -1323,7 +1323,7 @@ model to get started
GitHub
-
+ Subscribe to NewsletterAbonare la Newsletter
@@ -1331,22 +1331,22 @@ model to get started
LocalDocsSettings
-
+ LocalDocsLocalDocs
-
+ LocalDocs SettingsConfigurarea LocalDocs
-
+ IndexingIndexare
-
+ Allowed File ExtensionsExtensii compatibile de fişier
@@ -1356,12 +1356,12 @@ model to get started
Extensiile, separate prin virgulă. LocalDocs va încerca procesarea numai a fişierelor cu aceste extensii.
-
+ EmbeddingEmbedding
-
+ Use Nomic Embed APIFolosesc Nomic Embed API
@@ -1371,7 +1371,7 @@ model to get started
Embedding pe documente folosind API de la Nomic în locul unui model local. Necesită repornire.
-
+ Nomic API KeyCheia API Nomic
@@ -1382,72 +1382,72 @@ model to get started
Cheia API de utilizat cu Nomic Embed. Obţine o cheie prin Atlas: <a href="https://atlas.nomic.ai/cli-login">pagina cheilor API</a> Necesită repornire.
-
+ Embeddings DeviceDispozitivul pentru Embeddings
-
+ The compute device used for embeddings. Requires restart.Dispozitivul pentru Embeddings. Necesită repornire.
-
+ Comma-separated list. LocalDocs will only attempt to process files with these extensions.Extensiile, separate prin virgulă. LocalDocs va încerca procesarea numai a fişierelor cu aceste extensii.
-
+ Embed documents using the fast Nomic API instead of a private local model. Requires restart.Embedding pe documente folosind API de la Nomic în locul unui model local. Necesită repornire.
-
+ API key to use for Nomic Embed. Get one from the Atlas <a href="https://atlas.nomic.ai/cli-login">API keys page</a>. Requires restart.Cheia API de utilizat cu Nomic Embed. Obţine o cheie prin Atlas: <a href="https://atlas.nomic.ai/cli-login">pagina cheilor API</a> Necesită repornire.
-
+ Application defaultImplicit
-
+ DisplayVizualizare
-
+ Show SourcesAfişarea Surselor
-
+ Display the sources used for each response.Afişează Sursele utilizate pentru fiecare replică.
-
+ AdvancedAvansate
-
+ Warning: Advanced usage only.Atenţie: Numai pentru utilizare avansată.
-
+ Values too large may cause localdocs failure, extremely slow responses or failure to respond at all. Roughly speaking, the {N chars x N snippets} are added to the model's context window. More info <a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">here</a>.Valori prea mari pot cauza erori cu LocalDocs, replici foarte lente sau chiar absenţa lor. În mare, numărul {N caractere x N citate} este adăugat la Context Window/Size/Length a modelului. Mai multe informaţii: <a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">aici</a>.
-
+ Number of characters per document snippet. Larger numbers increase likelihood of factual responses, but also result in slower generation.Numărul caracterelor din fiecare citat. Numere mari amplifică probabilitatea unor replici corecte, dar de asemenea cauzează generare lentă.
-
+ Max best N matches of retrieved document snippets to add to the context for prompt. Larger numbers increase likelihood of factual responses, but also result in slower generation.Numărul maxim al citatelor ce corespund şi care vor fi adăugate la contextul pentru prompt. Numere mari amplifică probabilitatea unor replici corecte, dar de asemenea cauzează generare lentă.
@@ -1461,7 +1461,7 @@ model to get started
Valori prea mari pot cauza erori cu LocalDocs, replici lente sau absenţa lor completă. în mare, numărul {N caractere x N citate} este adăugat la Context Window/Size/Length a modelului. Mai multe informaţii: <a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">aici</a>.
-
+ Document snippet size (characters)Lungimea (în caractere) a citatelor din documente
@@ -1471,7 +1471,7 @@ model to get started
numărul caracterelor din fiecare citat. Numere mari amplifică probabilitatea unor replici corecte, dar de asemenea pot cauza generare lentă.
-
+ Max document snippets per promptNumărul maxim de citate per prompt
@@ -1485,17 +1485,17 @@ model to get started
LocalDocsView
-
+ LocalDocsLocalDocs
-
+ Chat with your local filesDialoghează cu fişiere locale
-
+ + Add Collection+ Adaugă o Colecţie
@@ -1504,102 +1504,102 @@ model to get started
EROARE: Baza de date LocalDocs nu e validă.
-
+ <h3>ERROR: The LocalDocs database cannot be accessed or is not valid.</h3><br><i>Note: You will need to restart after trying any of the following suggested fixes.</i><br><ul><li>Make sure that the folder set as <b>Download Path</b> exists on the file system.</li><li>Check ownership as well as read and write permissions of the <b>Download Path</b>.</li><li>If there is a <b>localdocs_v2.db</b> file, check its ownership and read/write permissions, too.</li></ul><br>If the problem persists and there are any 'localdocs_v*.db' files present, as a last resort you can<br>try backing them up and removing them. You will have to recreate your collections, however.EROARE: Baza de date LocalDocs nu poate fi accesată sau nu e validă. Programul trebuie repornit după ce se încearcă oricare din următoarele remedii sugerate.</i><br><ul><li>Asigură-te că folderul pentru <b>Download Path</b> există în sistemul de fişiere.</li><li>Verifică permisiunile şi apartenenţa folderului pentru <b>Download Path</b>.</li><li>Dacă există fişierul <b>localdocs_v2.db</b>, verifică-i apartenenţa şi permisiunile citire/scriere (read/write).</li></ul><br>Dacă problema persistă şi există vreun fişier 'localdocs_v*.db', ca ultimă soluţie poţi<br>încerca duplicarea (backup) şi apoi ştergerea lor. Oricum, va trebui să re-creezi Colecţiile.
-
+ No Collections InstalledNu există Colecţii instalate
-
+ Install a collection of local documents to get started using this featureInstalează o Colecţie de documente pentru a putea utiliza funcţionalitatea aceasta
-
+ + Add Doc Collection+ Adaugă o Colecţie de documente
-
+ Shows the add model viewAfişează secţiunea de adăugare a unui model
-
+ Indexing progressBarBara de progresie a Indexării
-
+ Shows the progress made in the indexingAfişează progresia Indexării
-
+ ERROREROARE
-
+ INDEXING...SE INDEXEAZĂ...
-
+ EMBEDDING...EMBEDDINGs...
-
+ REQUIRES UPDATENECESITĂ UPDATE
-
+ READYGATA
-
+ INSTALLING...INSTALARE...
-
+ Indexing in progressSe Indexează...
-
+ Embedding in progress...Se calculează Embeddings...
-
+ This collection requires an update after version changeAceastă Colecţie necesită update după schimbarea versiunii
-
+ Automatically reindexes upon changes to the folderSe reindexează automat după schimbări ale folderului
-
+ Installation in progress...Instalare în curs...
-
+ %%
-
+ %n file(s)%n fişier
@@ -1608,7 +1608,7 @@ model to get started
-
+ %n word(s)%n cuvânt
@@ -1617,27 +1617,27 @@ model to get started
-
+ RemoveŞterg
-
+ RebuildReconstrucţie
-
+ Reindex this folder from scratch. This is slow and usually not needed.Reindexează de la zero acest folder. Procesul e lent şi de obicei inutil.
-
+ UpdateUpdate/Actualizare
-
+ Update the collection to the new version. This is a slow operation.Actualizează Colecţia la noua versiune. Această procedură e lentă.
@@ -1661,67 +1661,78 @@ model to get started
<strong>Modelul ChatGPT GPT-3.5 Turbo al OpenAI</strong><br> %1
-
+
+
+ cannot open "%1": %2
+
+
+
+
+ cannot create "%1": %2
+
+
+
+ %1 (%2)%1 (%2)
-
+ <strong>OpenAI-Compatible API Model</strong><br><ul><li>API Key: %1</li><li>Base URL: %2</li><li>Model Name: %3</li></ul><strong>Model API compatibil cu OpenAI</strong><br><ul><li>Cheia API: %1</li><li>Base URL: %2</li><li>Numele modelului: %3</li></ul>
-
+ <ul><li>Requires personal OpenAI API key.</li><li>WARNING: Will send your chats to OpenAI!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with OpenAI</li><li>You can apply for an API key <a href="https://platform.openai.com/account/api-keys">here.</a></li><ul><li>Necesită o cheie API OpenAI personală. </li><li>ATENŢIE: Conversaţiile tale vor fi trimise la OpenAI!</li><li>Cheia ta API va fi stocată pe disc (local) </li><li>Va fi utilizată numai pentru comunicarea cu OpenAI</li><li>Poţi solicita o cheie API aici: <a href="https://platform.openai.com/account/api-keys">aici.</a></li>
-
+ <strong>OpenAI's ChatGPT model GPT-3.5 Turbo</strong><br> %1<strong>Modelul OpenAI's ChatGPT GPT-3.5 Turbo</strong><br> %1
-
+ <br><br><i>* Even if you pay OpenAI for ChatGPT-4 this does not guarantee API key access. Contact OpenAI for more info.<br><br><i>* Chiar dacă plăteşti la OpenAI pentru ChatGPT-4, aceasta nu garantează accesul la cheia API. Contactează OpenAI pentru mai multe informaţii.
-
+ <strong>OpenAI's ChatGPT model GPT-4</strong><br> %1 %2<strong>Modelul ChatGPT GPT-4 al OpenAI</strong><br> %1 %2
-
+ <ul><li>Requires personal Mistral API key.</li><li>WARNING: Will send your chats to Mistral!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with Mistral</li><li>You can apply for an API key <a href="https://console.mistral.ai/user/api-keys">here</a>.</li><ul><li>Necesită cheia personală Mistral API. </li><li>ATENŢIE: Conversaţiile tale vor fi trimise la Mistral!</li><li>Cheia ta API va fi stocată pe disc (local)</li><li>Va fi utilizată numai pentru comunicarea cu Mistral</li><li>Poţi solicita o cheie API aici: <a href="https://console.mistral.ai/user/api-keys">aici</a>.</li>
-
+ <strong>Mistral Tiny model</strong><br> %1<strong>Modelul Mistral Tiny</strong><br> %1
-
+ <strong>Mistral Small model</strong><br> %1<strong>Modelul Mistral Small</strong><br> %1
-
+ <strong>Mistral Medium model</strong><br> %1<strong>Modelul Mistral Medium</strong><br> %1
-
+ <ul><li>Requires personal API key and the API base URL.</li><li>WARNING: Will send your chats to the OpenAI-compatible API Server you specified!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with the OpenAI-compatible API Server</li><ul><li>Necesită cheia personală API si base-URL a API.</li><li>ATENŢIE: Conversaţiile tale vor fi trimise la serverul API compatibil cu OpenAI specificat!</li><li>Cheia ta API va fi stocată pe disc (local)</li><li>Va fi utilizată numai pentru comunicarea cu serverul API compatibil cu OpenAI</li>
-
+ <strong>Connect to OpenAI-compatible API server</strong><br> %1<strong>Conectare la un server API compatibil cu OpenAI</strong><br> %1
-
+ <strong>Created by %1.</strong><br><ul><li>Published on %2.<li>This model has %3 likes.<li>This model has %4 downloads.<li>More info can be found <a href="https://huggingface.co/%5">here.</a></ul><strong>Creat de către %1.</strong><br><ul><li>Publicat in: %2.<li>Acest model are %3 Likes.<li>Acest model are %4 download-uri.<li>Mai multe informaţii pot fi găsite la: <a href="https://huggingface.co/%5">aici.</a></ul>
@@ -1752,37 +1763,37 @@ model to get started
ModelSettings
-
+ ModelModel
-
+ Model SettingsConfigurez modelul
-
+ CloneClonez
-
+ RemoveŞterg
-
+ NameDenumire
-
+ Model FileFişierul modelului
-
+ System PromptSystem Prompt
@@ -1792,12 +1803,12 @@ model to get started
Plasat la Începutul fiecărei conversaţii. Trebuie să conţină token-uri(le) adecvate de Încadrare.
-
+ Prompt TemplatePrompt Template
-
+ The template that wraps every prompt.Standardul de formulare a fiecărui prompt.
@@ -1807,32 +1818,32 @@ model to get started
Trebuie să conţină textul "%1" care va fi Înlocuit cu ceea ce scrie utilizatorul.
-
+ Chat Name PromptDenumirea conversaţiei
-
+ Prompt used to automatically generate chat names.Standardul de formulare a denumirii conversaţiilor.
-
+ Suggested FollowUp PromptPrompt-ul sugerat pentru a continua
-
+ Prompt used to generate suggested follow-up questions.Prompt-ul folosit pentru generarea întrebărilor de continuare.
-
+ Context LengthLungimea Contextului
-
+ Number of input and output tokens the model sees.Numărul token-urilor de input şi de output văzute de model.
@@ -1843,12 +1854,12 @@ model to get started
Numărul maxim combinat al token-urilor în prompt+replică înainte de a se pierde informaţie. Utilizarea unui context mai mare decât cel cu care a fost instruit modelul va întoarce rezultate mai slabe. NOTĂ: Nu are efect până la reîncărcarea modelului.
-
+ TemperatureTemperatura
-
+ Randomness of model output. Higher -> more variation.Libertatea/Confuzia din replica modelului. Mai mare -> mai multă libertate.
@@ -1858,12 +1869,12 @@ model to get started
Temperatura creşte probabilitatea de alegere a unor token-uri puţin probabile. NOTĂ: O temperatură tot mai înaltă determină replici tot mai creative şi mai puţin predictibile.
-
+ Top-PTop-P
-
+ Nucleus Sampling factor. Lower -> more predictable.Factorul de Nucleus Sampling. Mai mic -> predictibilitate mai mare.
@@ -1873,92 +1884,92 @@ model to get started
Pot fi alese numai cele mai probabile token-uri a căror probabilitate totală este Top-P. NOTĂ: Se evită selectarea token-urilor foarte improbabile.
-
+ Prefixed at the beginning of every conversation. Must contain the appropriate framing tokens.Plasat la începutul fiecărei conversaţii. Trebuie să conţină token-uri(le) adecvate de încadrare.
-
+ Must contain the string "%1" to be replaced with the user's input.Trebuie să conţină textul "%1" care va fi înlocuit cu ceea ce scrie utilizatorul.
-
+ Maximum combined prompt/response tokens before information is lost.
Using more context than the model was trained on will yield poor results.
NOTE: Does not take effect until you reload the model.Numărul maxim combinat al token-urilor în prompt+replică înainte de a se pierde informaţie. Utilizarea unui context mai mare decât cel cu care a fost instruit modelul va întoarce rezultate mai slabe. NOTĂ: Nu are efect până la reîncărcarea modelului.
-
+ Temperature increases the chances of choosing less likely tokens.
NOTE: Higher temperature gives more creative but less predictable outputs.Temperatura creşte probabilitatea de alegere a unor token-uri puţin probabile. NOTĂ: O temperatură tot mai înaltă determină replici tot mai creative şi mai puţin predictibile.
-
+ Only the most likely tokens up to a total probability of top_p can be chosen.
NOTE: Prevents choosing highly unlikely tokens.Pot fi alese numai cele mai probabile token-uri a căror probabilitate totală este Top-P. NOTĂ: Se evită selectarea token-urilor foarte improbabile.
-
+ Min-PMin-P
-
+ Minimum token probability. Higher -> more predictable.Probabilitatea mínimă a unui token. Mai mare -> mai predictibil.
-
+ Sets the minimum relative probability for a token to be considered.Stabileşte probabilitatea minimă relativă a unui token de luat în considerare.
-
+ Top-KTop-K
-
+ Size of selection pool for tokens.Dimensiunea setului de token-uri.
-
+ Only the top K most likely tokens will be chosen from.Se va alege numai din cele mai probabile K token-uri.
-
+ Max LengthLungimea maximă
-
+ Maximum response length, in tokens.Lungimea maximă - în token-uri - a replicii.
-
+ Prompt Batch SizePrompt Batch Size
-
+ The batch size used for prompt processing.Dimensiunea setului de token-uri citite simultan din prompt.
-
+ Amount of prompt tokens to process at once.
NOTE: Higher values can speed up reading prompts but will use more RAM.Numărul token-urilor procesate simultan. NOTĂ: Valori tot mai mari pot accelera citirea prompt-urilor, dar şi utiliza mai multă RAM.
-
+ How many model layers to load into VRAM. Decrease this if GPT4All runs out of VRAM while loading this model.
Lower values increase CPU load and RAM usage, and make inference slower.
NOTE: Does not take effect until you reload the model.
@@ -1970,32 +1981,32 @@ NOTE: Does not take effect until you reload the model.
numărul token-urilor procesate simultan. NOTĂ: Valori tot mai mari pot accelera citirea prompt-urilor, dar şi utiliza mai multă RAM.
-
+ Repeat PenaltyPenalizarea pentru repetare
-
+ Repetition penalty factor. Set to 1 to disable.Factorul de penalizare a repetării ce se dezactivează cu valoarea 1.
-
+ Repeat Penalty TokensToken-uri pentru penalizare a repetării
-
+ Number of previous tokens used for penalty.Numărul token-urilor anterioare considerate pentru penalizare.
-
+ GPU LayersLayere în GPU
-
+ Number of model layers to load into VRAM.Numărul layerelor modelului ce vor fi Încărcate în VRAM.
@@ -2010,89 +2021,89 @@ NOTE: Does not take effect until you reload the model.
ModelsView
-
+ No Models InstalledNu există modele instalate
-
+ Install a model to get started using GPT4AllInstalează un model pentru a începe să foloseşti GPT4All
-
-
+
+ + Add Model+ Adaugă un model
-
+ Shows the add model viewAfişează secţiunea de adăugare a unui model
-
+ Installed ModelsModele instalate
-
+ Locally installed chat modelsModele conversaţionale instalate local
-
+ Model fileFişierul modelului
-
+ Model file to be downloadedFişierul modelului ce va fi descărcat
-
+ DescriptionDescriere
-
+ File descriptionDescrierea fişierului
-
+ CancelAnulare
-
+ ResumeContinuare
-
+ Stop/restart/start the downloadOprirea/Repornirea/Iniţierea descărcării
-
+ RemoveŞterg
-
+ Remove model from filesystemŞterg modelul din sistemul de fişiere
-
-
+
+ InstallInstalează
-
+ Install online modelInstalez un model din online
@@ -2108,130 +2119,130 @@ NOTE: Does not take effect until you reload the model.
<strong><font size="2">ATENţIE: Nerecomandat pentru acest hardware. Modelul necesită mai multă memorie (%1 GB) decât are sistemul tău (%2).</strong></font>
-
+ %1 GB%1 GB
-
+ ??
-
+ Describes an error that occurred when downloadingDescrie o eroare apărută la download
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font><strong><font size="1"><a href="#eroare">Error</a></strong></font>
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font><strong><font size="2">ATENŢIE: Nerecomandat pentru acest hardware. Modelul necesită mai multă memorie (%1 GB) decât are sistemul tău(%2).</strong></font>
-
+ Error for incompatible hardwareEroare - hardware incompatibil
-
+ Download progressBarBara de progresie a descărcării
-
+ Shows the progress made in the downloadAfişează progresia descărcării
-
+ Download speedViteza de download
-
+ Download speed in bytes/kilobytes/megabytes per secondViteza de download în bytes/kilobytes/megabytes pe secundă
-
+ Calculating......Se calculează...
-
-
-
-
+
+
+
+ Whether the file hash is being calculatedDacă se va calcula hash-ul fişierului
-
+ Busy indicatorIndicator de activitate
-
+ Displayed when the file hash is being calculatedAfişat când se calculează hash-ul unui fişier
-
+ ERROR: $API_KEY is empty.EROARE: $API_KEY absentă.
-
+ enter $API_KEYintrodu cheia $API_KEY
-
+ ERROR: $BASE_URL is empty.EROARE: $BASE_URL absentă.
-
+ enter $BASE_URLintrodu $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.EROARE: $MODEL_NAME absent.
-
+ enter $MODEL_NAMEintrodu $MODEL_NAME
-
+ File sizeDimensiunea fişierului
-
+ RAM requiredRAM necesară
-
+ ParametersParametri
-
+ QuantQuant(ificare)
-
+ TypeTip
@@ -2239,12 +2250,12 @@ NOTE: Does not take effect until you reload the model.
MyFancyLink
-
+ Fancy linkLink haios
-
+ A stylized linkUn link cu stil
@@ -2252,7 +2263,7 @@ NOTE: Does not take effect until you reload the model.
MySettingsStack
-
+ Please choose a directorySelectează un director (folder)
@@ -2260,12 +2271,12 @@ NOTE: Does not take effect until you reload the model.
MySettingsTab
-
+ Restore DefaultsRestaurez valorile implicite
-
+ Restores settings dialog to a default stateRestaurez secţiunea de configurare la starea sa implicită
@@ -2273,7 +2284,7 @@ NOTE: Does not take effect until you reload the model.
NetworkDialog
-
+ Contribute data to the GPT4All Opensource Datalake.Contribuie cu date/informaţii la componenta Open-source DataLake a GPT4All.
@@ -2311,7 +2322,7 @@ NOTE: Does not take effect until you reload the model.
participant contribuitor la orice lansare a unui model GPT4All care foloseşte datele tale!
-
+ By enabling this feature, you will be able to participate in the democratic process of training a large language model by contributing data for future model improvements.
When a GPT4All model responds to you and you have opted-in, your conversation will be sent to the GPT4All Open Source Datalake. Additionally, you can like/dislike its response. If you dislike a response, you can suggest an alternative response. This data will be collected and aggregated in the GPT4All Datalake.
@@ -2334,47 +2345,47 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
participant contribuitor la orice lansare a unui model GPT4All care foloseşte datele tale!
-
+ Terms for opt-inTermenii pentru participare
-
+ Describes what will happen when you opt-inDescrie ce se întâmplă când participi
-
+ Please provide a name for attribution (optional)Specifică o denumire pentru această apreciere (opţional)
-
+ Attribution (optional)Apreciere (opţional)
-
+ Provide attributionApreciază
-
+ EnableActivează
-
+ Enable opt-inActivează participarea
-
+ CancelAnulare
-
+ Cancel opt-inAnulează participarea
@@ -2382,17 +2393,17 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
NewVersionDialog
-
+ New version is availableO nouă versiune disponibilă!
-
+ UpdateUpdate/Actualizare
-
+ Update to new versionActualizează la noua versiune
@@ -2400,17 +2411,17 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
PopupDialog
-
+ Reveals a shortlived help balloonAfişează un mesaj scurt de asistenţă
-
+ Busy indicatorIndicator de activitate
-
+ Displayed when the popup is showing busySe afişează când procedura este în desfăşurare
@@ -2418,28 +2429,28 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
SettingsView
-
-
+
+ SettingsConfigurare
-
+ Contains various application settingsConţine setări ale programului
-
+ ApplicationProgram
-
+ ModelModel
-
+ LocalDocsLocalDocs
@@ -2447,7 +2458,7 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
StartupDialog
-
+ Welcome!Bun venit!
@@ -2460,12 +2471,12 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
%2
-
+ Release notesDespre versiune
-
+ Release notes for this versionDespre această versiune
@@ -2517,7 +2528,7 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
care foloseşte datele tale!
-
+ ### Release notes
%1### Contributors
%2
@@ -2526,7 +2537,7 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
%2
-
+ ### Opt-ins for anonymous usage analytics and datalake
By enabling these features, you will be able to participate in the democratic process of training a
large language model by contributing data for future model improvements.
@@ -2562,71 +2573,71 @@ participant contribuitor la orice lansare a unui model GPT4All
care foloseşte datele tale!
-
+ Terms for opt-inTermenii pentru participare
-
+ Describes what will happen when you opt-inDescrie ce se întâmplă când participi
-
-
+
+ Opt-in for anonymous usage statisticsAcceptă colectarea de statistici despre utilizare -anonimă-
-
-
+
+ YesDa
-
+ Allow opt-in for anonymous usage statisticsAcceptă participarea la colectarea de statistici despre utilizare -anonimă-
-
-
+
+ NoNu
-
+ Opt-out for anonymous usage statisticsAnulează participarea la colectarea de statistici despre utilizare -anonimă-
-
+ Allow opt-out for anonymous usage statisticsPermite anularea participării la colectarea de statistici despre utilizare -anonimă-
-
-
+
+ Opt-in for networkAcceptă pentru reţea
-
+ Allow opt-in for networkPermite participarea pentru reţea
-
+ Allow opt-in anonymous sharing of chats to the GPT4All DatalakePermite participarea la partajarea (share) -anonimă- a conversaţiilor către DataLake a GPT4All
-
+ Opt-out for networkRefuz participarea, pentru reţea
-
+ Allow opt-out anonymous sharing of chats to the GPT4All DatalakePermite anularea participării la partajarea -anonimă- a conversaţiilor către DataLake a GPT4All
@@ -2639,23 +2650,23 @@ care foloseşte datele tale!
<b>Atenţie:</b> schimbarea modelului va şterge conversaţia curentă. Confirmi aceasta?
-
+ <b>Warning:</b> changing the model will erase the current conversation. Do you wish to continue?<b>Atenţie:</b> schimbarea modelului va şterge conversaţia curentă. Confirmi aceasta?
-
+ ContinueContinuă
-
+ Continue with model loadingContinuă încărcarea modelului
-
-
+
+ CancelAnulare
@@ -2663,32 +2674,32 @@ care foloseşte datele tale!
ThumbsDownDialog
-
+ Please edit the text below to provide a better response. (optional)Te rog, editează textul de mai jos pentru a oferi o replică mai bună (opţional).
-
+ Please provide a better response...Te rog, oferă o replică mai bună...
-
+ SubmitTrimite
-
+ Submits the user's responseTrimite răspunsul dat de utilizator
-
+ CancelAnulare
-
+ Closes the response dialogÎnchide afişarea răspunsului
@@ -2709,7 +2720,7 @@ care foloseşte datele tale!
<h3>A apărut o eroare la iniţializare:; </h3><br><i>"Hardware incompatibil. "</i><br><br>Din păcate, procesorul (CPU) nu întruneşte condiţiile minime pentru a rula acest program. În particular, nu suportă instrucţiunile AVX pe care programul le necesită pentru a integra un model conversaţional modern. În acest moment, unica soluţie este să îţi aduci la zi sistemul hardware cu un CPU mai recent.<br><br>Aici sunt mai multe informaţii: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a>
-
+ GPT4All v%1GPT4All v%1
@@ -2728,120 +2739,120 @@ care foloseşte datele tale!
<h3>A apărut o eroare la iniţializare:; </h3><br><i>"Hardware incompatibil. "</i><br><br>Din păcate, procesorul (CPU) nu întruneşte condiţiile minime pentru a rula acest program. În particular, nu suportă instrucţiunile AVX pe care programul le necesită pentru a integra un model conversaţional modern. În acest moment, unica soluţie este să îţi aduci la zi sistemul hardware cu un CPU mai recent.<br><br>Aici sunt mai multe informaţii: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a>
-
+ <h3>Encountered an error starting up:</h3><br><i>"Incompatible hardware detected."</i><br><br>Unfortunately, your CPU does not meet the minimal requirements to run this program. In particular, it does not support AVX intrinsics which this program requires to successfully run a modern large language model. The only solution at this time is to upgrade your hardware to a more modern CPU.<br><br>See here for more information: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a><h3>A apărut o eroare la iniţializare:; </h3><br><i>"Hardware incompatibil. "</i><br><br>Din păcate, procesorul (CPU) nu întruneşte condiţiile minime pentru a rula acest program. În particular, nu suportă instrucţiunile AVX pe care programul le necesită pentru a integra un model conversaţional modern. În acest moment, unica soluţie este să îţi aduci la zi sistemul hardware cu un CPU mai recent.<br><br>Aici sunt mai multe informaţii: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a>
-
+ <h3>Encountered an error starting up:</h3><br><i>"Inability to access settings file."</i><br><br>Unfortunately, something is preventing the program from accessing the settings file. This could be caused by incorrect permissions in the local app config directory where the settings file is located. Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help.<h3>A apărut o eroare la iniţializare:; </h3><br><i>"Nu poate fi accesat fişierul de configurare a programului."</i><br><br>Din păcate, ceva împiedică programul în a accesa acel fişier. Cauza poate fi un set de permisiuni incorecte pe directorul/folderul local de configurare unde se află acel fişier. Poţi parcurge canalul nostru <a href="https://discord.gg/4M2QFmTt2k">Discord</a> unde vei putea primi asistenţă.
-
+ Connection to datalake failed.Conectarea la DataLake a eşuat.
-
+ Saving chats.Se salvează conversaţiile.
-
+ Network dialogDialogul despre reţea
-
+ opt-in to share feedback/conversationsacceptă partajarea (share) de comentarii/conversaţii
-
+ Home viewSecţiunea de Început
-
+ Home view of applicationSecţiunea de Început a programului
-
+ HomePrima<br>pagină
-
+ Chat viewSecţiunea conversaţiilor
-
+ Chat view to interact with modelsSecţiunea de chat pentru interacţiune cu modele
-
+ ChatsConversaţii
-
-
+
+ ModelsModele
-
+ Models view for installed modelsSecţiunea modelelor instalate
-
-
+
+ LocalDocsLocalDocs
-
+ LocalDocs view to configure and use local docsSecţiunea LocalDocs de configurare şi folosire a Documentelor Locale
-
-
+
+ SettingsConfigurare
-
+ Settings view for application configurationSecţiunea de configurare a programului
-
+ The datalake is enabledDataLake: ACTIV
-
+ Using a network modelSe foloseşte un model pe reţea
-
+ Server mode is enabledModul Server: ACTIV
-
+ Installed modelsModele instalate
-
+ View of installed modelsSecţiunea modelelor instalate
diff --git a/gpt4all-chat/translations/gpt4all_zh_CN.ts b/gpt4all-chat/translations/gpt4all_zh_CN.ts
index 4bd6c395..5ee7e9b4 100644
--- a/gpt4all-chat/translations/gpt4all_zh_CN.ts
+++ b/gpt4all-chat/translations/gpt4all_zh_CN.ts
@@ -4,62 +4,62 @@
AddCollectionView
-
+ ← Existing Collections← 存在集合
-
+ Add Document Collection添加文档集合
-
+ Add a folder containing plain text files, PDFs, or Markdown. Configure additional extensions in Settings.添加一个包含纯文本文件、PDF或Markdown的文件夹。在“设置”中配置其他扩展。
-
+ Please choose a directory请选择一个目录
-
+ Name名称
-
+ Collection name...集合名称
-
+ Name of the collection to add (Required)集合名称 (必须)
-
+ Folder目录
-
+ Folder path...目录地址
-
+ Folder path to documents (Required)文档的目录地址(必须)
-
+ Browse查看
-
+ Create Collection创建集合
@@ -67,22 +67,22 @@
AddModelView
-
+ ← Existing Models← 存在的模型
-
+ Explore Models发现模型
-
+ Discover and download models by keyword search...通过关键词查找并下载模型 ...
-
+ Text field for discovering and filtering downloadable models用于发现和筛选可下载模型的文本字段
@@ -91,32 +91,32 @@
搜索中
-
+ Initiate model discovery and filtering启动模型发现和过滤
-
+ Triggers discovery and filtering of models触发模型的发现和筛选
-
+ Default默认
-
+ Likes喜欢
-
+ Downloads下载
-
+ Recent近期
@@ -125,12 +125,12 @@
排序:
-
+ Asc升序
-
+ Desc倒序
@@ -139,7 +139,7 @@
排序目录:
-
+ None无
@@ -152,145 +152,145 @@
网络问题:无法访问 http://gpt4all.io/models/models3.json
-
+ Searching · %1搜索中 · %1
-
+ Sort by: %1排序: %1
-
+ Sort dir: %1排序目录: %1
-
+ Limit: %1数量: %1
-
+ Network error: could not retrieve %1网络错误:无法检索 %1
-
-
+
+ Busy indicator繁忙程度
-
+ Displayed when the models request is ongoing在模型请求进行中时显示
-
+ Model file模型文件
-
+ Model file to be downloaded待下载模型
-
+ Description描述
-
+ File description文件描述
-
+ Cancel取消
-
+ Resume继续
-
+ Download下载
-
+ Stop/restart/start the download停止/重启/开始下载
-
+ Remove删除
-
+ Remove model from filesystem从系统中删除模型
-
-
+
+ Install安装
-
+ Install online model安装在线模型
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font><strong><font size="1"><a href="#error">错误</a></strong></font>
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font><strong><font size="2">警告: 你的设备硬件不推荐 ,模型需要的内存 (%1 GB)比你的系统还要多 (%2).</strong></font>
-
+ ERROR: $API_KEY is empty.错误:$API_KEY 为空
-
+ ERROR: $BASE_URL is empty.错误:$BASE_URL 为空
-
+ enter $BASE_URL输入 $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.错误:$MODEL_NAME为空
-
+ enter $MODEL_NAME输入:$MODEL_NAME
-
+ %1 GB%1 GB
-
-
+
+ ??
@@ -299,7 +299,7 @@
<a href="#error">错误</a>
-
+ Describes an error that occurred when downloading描述下载过程中发生的错误
@@ -316,60 +316,60 @@
你的系统需要 (
-
+ Error for incompatible hardware硬件不兼容的错误
-
+ Download progressBar下载进度
-
+ Shows the progress made in the download显示下载进度
-
+ Download speed下载速度
-
+ Download speed in bytes/kilobytes/megabytes per second下载速度 b/kb/mb /s
-
+ Calculating...计算中
-
-
-
-
+
+
+
+ Whether the file hash is being calculated是否正在计算文件哈希
-
+ Displayed when the file hash is being calculated在计算文件哈希时显示
-
+ enter $API_KEY输入$API_KEY
-
+ File size文件大小
-
+ RAM requiredRAM 需要
@@ -378,17 +378,17 @@
GB
-
+ Parameters参数
-
+ Quant量化
-
+ Type类型
@@ -396,22 +396,22 @@
ApplicationSettings
-
+ Application应用
-
+ Network dialog网络对话
-
+ opt-in to share feedback/conversations选择加入以共享反馈/对话
-
+ ERROR: Update system could not find the MaintenanceTool used<br>
to check for updates!<br><br>
Did you install this application using the online installer? If so,<br>
@@ -425,87 +425,87 @@
如果无法手动启动它,那么恐怕您需要重新安装。
-
+ Error dialog错误对话
-
+ Application Settings应用设置
-
+ General通用设置
-
+ Theme主题
-
+ The application color scheme.应用的主题颜色
-
+ Dark深色
-
+ Light亮色
-
+ LegacyDarkLegacyDark
-
+ Font Size字体大小
-
+ The size of text in the application.应用中的文本大小。
-
+ Small小
-
+ Medium中
-
+ Large大
-
+ Language and Locale语言和本地化
-
+ The language and locale you wish to use.你想使用的语言
-
+ System Locale系统语言
-
+ Device设备
@@ -514,138 +514,138 @@
用于文本生成的计算设备. "自动" 使用 Vulkan or Metal.
-
+ The compute device used for text generation.设备用于文本生成
-
-
+
+ Application default程序默认
-
+ Default Model默认模型
-
+ The preferred model for new chats. Also used as the local server fallback.新聊天的首选模式。也用作本地服务器回退。
-
+ Suggestion Mode建议模式
-
+ Generate suggested follow-up questions at the end of responses.在答复结束时生成建议的后续问题。
-
+ When chatting with LocalDocs本地文档检索
-
+ Whenever possible只要有可能
-
+ Never从不
-
+ Download Path下载目录
-
+ Where to store local models and the LocalDocs database.本地模型和本地文档数据库存储目录
-
+ Browse查看
-
+ Choose where to save model files模型下载目录
-
+ Enable Datalake开启数据湖
-
+ Send chats and feedback to the GPT4All Open-Source Datalake.发送对话和反馈给GPT4All 的开源数据湖。
-
+ Advanced高级
-
+ CPU ThreadsCPU线程
-
+ The number of CPU threads used for inference and embedding.用于推理和嵌入的CPU线程数
-
+ Save Chat Context保存对话上下文
-
+ Save the chat model's state to disk for faster loading. WARNING: Uses ~2GB per chat.保存模型's 状态以提供更快加载速度. 警告: 需用 ~2GB 每个对话.
-
+ Enable Local Server开启本地服务
-
+ Expose an OpenAI-Compatible server to localhost. WARNING: Results in increased resource usage.将OpenAI兼容服务器暴露给本地主机。警告:导致资源使用量增加。
-
+ API Server PortAPI 服务端口
-
+ The port to use for the local server. Requires restart.使用本地服务的端口,需要重启
-
+ Check For Updates检查更新
-
+ Manually check for an update to GPT4All.手动检查更新
-
+ Updates更新
@@ -653,13 +653,13 @@
Chat
-
-
+
+ New Chat新对话
-
+ Server Chat服务器对话
@@ -675,12 +675,12 @@
ChatAPIWorker
-
+ ERROR: Network error occurred while connecting to the API server错误:连接到 API 服务器时发生网络错误
-
+ ChatAPIWorker::handleFinished got HTTP Error %1 %2ChatAPIWorker::handleFinished 收到 HTTP 错误 %1 %2
@@ -688,62 +688,62 @@
ChatDrawer
-
+ Drawer抽屉
-
+ Main navigation drawer导航
-
+ + New Chat+ 新对话
-
+ Create a new chat新对话
-
+ Select the current chat or edit the chat when in edit mode选择当前的聊天或在编辑模式下编辑聊天
-
+ Edit chat name修改对话名称
-
+ Save chat name保存对话名称
-
+ Delete chat删除对话
-
+ Confirm chat deletion确认删除对话
-
+ Cancel chat deletion取消删除对话
-
+ List of chats对话列表
-
+ List of chats in the drawer dialog对话框中的聊天列表
@@ -751,32 +751,32 @@
ChatListModel
-
+ TODAY今天
-
+ THIS WEEK本周
-
+ THIS MONTH本月
-
+ LAST SIX MONTHS半年内
-
+ THIS YEAR今年内
-
+ LAST YEAR去年
@@ -792,27 +792,27 @@
lt;br><br>模型加载失败可能由多种原因造成,但最常见的原因包括文件格式错误、下载不完整或损坏、文件类型错误、系统 RAM 不足或模型类型不兼容。以下是解决该问题的一些建议:<br><ul><li>确保模型文件具有兼容的格式和类型<li>检查下载文件夹中的模型文件是否完整<li>您可以在设置对话框中找到下载文件夹<li>如果您已侧载模型,请通过检查 md5sum 确保文件未损坏<li>在我们的<a href="https://docs.gpt4all.io/">文档</a>中了解有关支持哪些模型的更多信息对于 gui<li>查看我们的<a href="https://discord.gg/4M2QFmTt2k">discord 频道</a> 获取帮助
-
+ <h3>Warning</h3><p>%1</p><h3>警告</h3><p>%1</p>
-
+ Switch model dialog切换模型对话
-
+ Warn the user if they switch models, then context will be erased如果用户切换模型,则警告用户,然后上下文将被删除
-
+ Conversation copied to clipboard.复制对话到剪切板
-
+ Code copied to clipboard.复制代码到剪切板
@@ -821,52 +821,52 @@
响应:
-
+ Chat panel对话面板
-
+ Chat panel with options对话面板选项
-
+ Reload the currently loaded model重载当前模型
-
+ Eject the currently loaded model弹出当前加载的模型
-
+ No model installed.没有安装模型
-
+ Model loading error.模型加载错误
-
+ Waiting for model...稍等片刻
-
+ Switching context...切换上下文
-
+ Choose a model...选择模型
-
+ Not found: %1没找到: %1
@@ -879,23 +879,23 @@
载入中·
-
+ The top item is the current model当前模型的最佳选项
-
-
+
+ LocalDocs本地文档
-
+ Add documents添加文档
-
+ add collections of documents to the chat将文档集合添加到聊天中
@@ -908,53 +908,53 @@
(默认) →
-
+ Load the default model载入默认模型
-
+ Loads the default model which can be changed in settings加载默认模型,可以在设置中更改
-
+ No Model Installed没有下载模型
-
+ GPT4All requires that you install at least one
model to get startedGPT4All要求您至少安装一个模型才能开始
-
+ Install a Model下载模型
-
+ Shows the add model view查看添加的模型
-
+ Conversation with the model使用此模型对话
-
+ prompt / response pairs from the conversation对话中的提示/响应对
-
+ GPT4AllGPT4All
-
+ You您
@@ -971,7 +971,7 @@ model to get started
重新生成上下文...
-
+ response stopped ...响应停止...
@@ -984,176 +984,176 @@ model to get started
检索本地文档:
-
+ processing ...处理中
-
+ generating response ...响应中...
-
+ generating questions ...生成响应
-
-
+
+ Copy复制
-
+ Copy Message复制内容
-
+ Disable markdown不允许markdown
-
+ Enable markdown允许markdown
-
+ Thumbs up点赞
-
+ Gives a thumbs up to the response点赞响应
-
+ Thumbs down点踩
-
+ Opens thumbs down dialog打开点踩对话框
-
+ Suggested follow-ups建议的后续行动
-
+ Erase and reset chat session擦除并重置聊天会话
-
+ Copy chat session to clipboard复制对话到剪切板
-
+ Redo last chat response重新生成上个响应
-
+ Stop generating停止生成
-
+ Stop the current response generation停止当前响应
-
+ Reloads the model重载模型
-
+ <h3>Encountered an error loading model:</h3><br><i>"%1"</i><br><br>Model loading failures can happen for a variety of reasons, but the most common causes include a bad file format, an incomplete or corrupted download, the wrong file type, not enough system RAM or an incompatible model type. Here are some suggestions for resolving the problem:<br><ul><li>Ensure the model file has a compatible format and type<li>Check the model file is complete in the download folder<li>You can find the download folder in the settings dialog<li>If you've sideloaded the model ensure the file is not corrupt by checking md5sum<li>Read more about what models are supported in our <a href="https://docs.gpt4all.io/">documentation</a> for the gui<li>Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help<h3>加载模型时遇到错误:</h3><br><i><%1></i><br><br>模型加载失败可能由多种原因引起,但最常见的原因包括文件格式错误、下载不完整或损坏、文件类型错误、系统 RAM 不足或模型类型不兼容。以下是一些解决问题的建议:<br><ul><li>确保模型文件具有兼容的格式和类型<li>检查下载文件夹中的模型文件是否完整<li>您可以在设置对话框中找到下载文件夹<li>如果您已侧载模型,请通过检查 md5sum 确保文件未损坏<li>在我们的 <a href="https://docs.gpt4all.io/">文档</a> 中了解有关 gui 支持哪些模型的更多信息<li>查看我们的 <a href="https://discord.gg/4M2QFmTt2k">discord 频道</a> 以获取帮助
-
-
+
+ Reload · %1重载 · %1
-
+ Loading · %1载入中 · %1
-
+ Load · %1 (default) →载入 · %1 (默认) →
-
+ restoring from text ...从文本恢复中
-
+ retrieving localdocs: %1 ...检索本地文档: %1 ...
-
+ searching localdocs: %1 ...搜索本地文档: %1 ...
-
+ %n Source(s)%n 资源
-
+ Send a message...发送消息...
-
+ Load a model to continue...选择模型并继续
-
+ Send messages/prompts to the model发送消息/提示词给模型
-
+ Cut剪切
-
+ Paste粘贴
-
+ Select All全选
-
+ Send message发送消息
-
+ Sends the message/prompt contained in textfield to the model将文本框中包含的消息/提示发送给模型
@@ -1161,36 +1161,36 @@ model to get started
CollectionsDrawer
-
+ Warning: searching collections while indexing can return incomplete results提示: 索引时搜索集合可能会返回不完整的结果
-
+ %n file(s)
-
+ %n word(s)
-
+ Updating更新中
-
+ + Add Docs+ 添加文档
-
+ Select a collection to make it available to the chat model.选择一个集合,使其可用于聊天模型。
@@ -1198,37 +1198,37 @@ model to get started
Download
-
+ Model "%1" is installed successfully.模型 "%1" 安装成功
-
+ ERROR: $MODEL_NAME is empty.错误:$MODEL_NAME 为空
-
+ ERROR: $API_KEY is empty.错误:$API_KEY为空
-
+ ERROR: $BASE_URL is invalid.错误:$BASE_URL 非法
-
+ ERROR: Model "%1 (%2)" is conflict.错误: 模型 "%1 (%2)" 有冲突.
-
+ Model "%1 (%2)" is installed successfully.模型 "%1 (%2)" 安装成功.
-
+ Model "%1" is removed.模型 "%1" 已删除.
@@ -1236,92 +1236,92 @@ model to get started
HomeView
-
+ Welcome to GPT4All欢迎
-
+ The privacy-first LLM chat application隐私至上的大模型咨询应用程序
-
+ Start chatting开始聊天
-
+ Start Chatting开始聊天
-
+ Chat with any LLM大预言模型聊天
-
+ LocalDocs本地文档
-
+ Chat with your local files本地文件聊天
-
+ Find Models查找模型
-
+ Explore and download models发现并下载模型
-
+ Latest news新闻
-
+ Latest news from GPT4AllGPT4All新闻
-
+ Release Notes发布日志
-
+ Documentation文档
-
+ DiscordDiscord
-
+ X (Twitter)X (Twitter)
-
+ GithubGithub
-
+ nomic.ainomic.ai
-
+ Subscribe to Newsletter订阅信息
@@ -1329,117 +1329,117 @@ model to get started
LocalDocsSettings
-
+ LocalDocs本地文档
-
+ LocalDocs Settings本地文档设置
-
+ Indexing索引中
-
+ Allowed File Extensions添加文档扩展名
-
+ Comma-separated list. LocalDocs will only attempt to process files with these extensions.逗号分隔的列表。LocalDocs 只会尝试处理具有这些扩展名的文件
-
+ EmbeddingEmbedding
-
+ Use Nomic Embed API使用 Nomic 内部 API
-
+ Embed documents using the fast Nomic API instead of a private local model. Requires restart.使用快速的 Nomic API 嵌入文档,而不是使用私有本地模型
-
+ Nomic API KeyNomic API Key
-
+ API key to use for Nomic Embed. Get one from the Atlas <a href="https://atlas.nomic.ai/cli-login">API keys page</a>. Requires restart.Nomic Embed 使用的 API 密钥。请访问官网获取,需要重启。
-
+ Embeddings DeviceEmbeddings 设备
-
+ The compute device used for embeddings. Requires restart.技术设备用于embeddings. 需要重启.
-
+ Application default程序默认
-
+ Display显示
-
+ Show Sources查看源码
-
+ Display the sources used for each response.显示每个响应所使用的源。
-
+ Advanced高级
-
+ Warning: Advanced usage only.提示: 仅限高级使用。
-
+ Values too large may cause localdocs failure, extremely slow responses or failure to respond at all. Roughly speaking, the {N chars x N snippets} are added to the model's context window. More info <a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">here</a>.值过大可能会导致 localdocs 失败、响应速度极慢或根本无法响应。粗略地说,{N 个字符 x N 个片段} 被添加到模型的上下文窗口中。更多信息请见<a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">此处</a>。
-
+ Document snippet size (characters)文档粘贴大小 (字符)
-
+ Number of characters per document snippet. Larger numbers increase likelihood of factual responses, but also result in slower generation.每个文档片段的字符数。较大的数值增加了事实性响应的可能性,但也会导致生成速度变慢。
-
+ Max document snippets per prompt每个提示的最大文档片段数
-
+ Max best N matches of retrieved document snippets to add to the context for prompt. Larger numbers increase likelihood of factual responses, but also result in slower generation.检索到的文档片段最多添加到提示上下文中的前 N 个最佳匹配项。较大的数值增加了事实性响应的可能性,但也会导致生成速度变慢。
@@ -1447,17 +1447,17 @@ model to get started
LocalDocsView
-
+ LocalDocs本地文档
-
+ Chat with your local files和本地文件对话
-
+ + Add Collection+ 添加集合
@@ -1472,136 +1472,136 @@ model to get started
<h3>错误:无法访问 LocalDocs 数据库或该数据库无效。</h3><br><i>注意:尝试以下任何建议的修复方法后,您将需要重新启动。</i><br><ul><li>确保设置为<b>下载路径</b>的文件夹存在于文件系统中。</li><li>检查<b>下载路径</b>的所有权以及读写权限。</li><li>如果有<b>localdocs_v2.db</b>文件,请检查其所有权和读/写权限。</li></ul><br>如果问题仍然存在,并且存在任何“localdocs_v*.db”文件,作为最后的手段,您可以<br>尝试备份并删除它们。但是,您必须重新创建您的收藏。
-
+ <h3>ERROR: The LocalDocs database cannot be accessed or is not valid.</h3><br><i>Note: You will need to restart after trying any of the following suggested fixes.</i><br><ul><li>Make sure that the folder set as <b>Download Path</b> exists on the file system.</li><li>Check ownership as well as read and write permissions of the <b>Download Path</b>.</li><li>If there is a <b>localdocs_v2.db</b> file, check its ownership and read/write permissions, too.</li></ul><br>If the problem persists and there are any 'localdocs_v*.db' files present, as a last resort you can<br>try backing them up and removing them. You will have to recreate your collections, however.<h3>错误:无法访问 LocalDocs 数据库或该数据库无效。</h3><br><i>注意:尝试以下任何建议的修复方法后,您将需要重新启动。</i><br><ul><li>确保设置为<b>下载路径</b>的文件夹存在于文件系统中。</li><li>检查<b>下载路径</b>的所有权以及读写权限。</li><li>如果有<b>localdocs_v2.db</b>文件,请检查其所有权和读/写权限。</li></ul><br>如果问题仍然存在,并且存在任何“localdocs_v*.db”文件,作为最后的手段,您可以<br>尝试备份并删除它们。但是,您必须重新创建您的收藏。
-
+ No Collections Installed没有集合
-
+ Install a collection of local documents to get started using this feature安装一组本地文档以开始使用此功能
-
+ + Add Doc Collection+ 添加文档集合
-
+ Shows the add model view查看添加的模型
-
+ Indexing progressBar索引进度
-
+ Shows the progress made in the indexing显示索引进度
-
+ ERROR错误
-
+ INDEXING索引
-
+ EMBEDDINGEMBEDDING
-
+ REQUIRES UPDATE需更新
-
+ READY准备
-
+ INSTALLING安装中
-
+ Indexing in progress构建索引中
-
+ Embedding in progressEmbedding进度
-
+ This collection requires an update after version change此集合需要在版本更改后进行更新
-
+ Automatically reindexes upon changes to the folder在文件夹变动时自动重新索引
-
+ Installation in progress安装进度
-
+ %%
-
+ %n file(s)%n 文件
-
+ %n word(s)%n 词
-
+ Remove删除
-
+ Rebuild重新构建
-
+ Reindex this folder from scratch. This is slow and usually not needed.从头开始重新索引此文件夹。这个过程较慢,通常情况下不需要。
-
+ Update更新
-
+ Update the collection to the new version. This is a slow operation.将集合更新为新版本。这是一个缓慢的操作。
@@ -1609,52 +1609,63 @@ model to get started
ModelList
-
+
+
+ cannot open "%1": %2
+
+
+
+
+ cannot create "%1": %2
+
+
+
+ %1 (%2)%1 (%2)
-
+ <strong>OpenAI-Compatible API Model</strong><br><ul><li>API Key: %1</li><li>Base URL: %2</li><li>Model Name: %3</li></ul><strong>与 OpenAI 兼容的 API 模型</strong><br><ul><li>API 密钥:%1</li><li>基本 URL:%2</li><li>模型名称:%3</li></ul>
-
+ <ul><li>Requires personal OpenAI API key.</li><li>WARNING: Will send your chats to OpenAI!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with OpenAI</li><li>You can apply for an API key <a href="https://platform.openai.com/account/api-keys">here.</a></li><ul><li>需要个人 OpenAI API 密钥。</li><li>警告:将把您的聊天内容发送给 OpenAI!</li><li>您的 API 密钥将存储在磁盘上</li><li>仅用于与 OpenAI 通信</li><li>您可以在此处<a href="https://platform.openai.com/account/api-keys">申请 API 密钥。</a></li>
-
+ <strong>OpenAI's ChatGPT model GPT-3.5 Turbo</strong><br> %1<strong>OpenAI's ChatGPT model GPT-3.5 Turbo</strong><br> %1
-
+ <strong>OpenAI's ChatGPT model GPT-4</strong><br> %1 %2<strong>OpenAI's ChatGPT model GPT-4</strong><br> %1 %2
-
+ <strong>Mistral Tiny model</strong><br> %1<strong>Mistral Tiny model</strong><br> %1
-
+ <strong>Mistral Small model</strong><br> %1<strong>Mistral Small model</strong><br> %1
-
+ <strong>Mistral Medium model</strong><br> %1<strong>Mistral Medium model</strong><br> %1
-
+ <ul><li>Requires personal API key and the API base URL.</li><li>WARNING: Will send your chats to the OpenAI-compatible API Server you specified!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with the OpenAI-compatible API Server</li><ul><li>需要个人 API 密钥和 API 基本 URL。</li><li>警告:将把您的聊天内容发送到您指定的与 OpenAI 兼容的 API 服务器!</li><li>您的 API 密钥将存储在磁盘上</li><li>仅用于与与 OpenAI 兼容的 API 服务器通信</li>
-
+ <strong>Connect to OpenAI-compatible API server</strong><br> %1<strong>连接到与 OpenAI 兼容的 API 服务器</strong><br> %1
@@ -1663,7 +1674,7 @@ model to get started
<strong>OpenAI's ChatGPT model GPT-3.5 Turbo</strong><br>
-
+ <br><br><i>* Even if you pay OpenAI for ChatGPT-4 this does not guarantee API key access. Contact OpenAI for more info.<br><br><i>* 即使您为ChatGPT-4向OpenAI付款,这也不能保证API密钥访问。联系OpenAI获取更多信息。
@@ -1672,7 +1683,7 @@ model to get started
<strong>OpenAI's ChatGPT model GPT-4</strong><br>
-
+ <ul><li>Requires personal Mistral API key.</li><li>WARNING: Will send your chats to Mistral!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with Mistral</li><li>You can apply for an API key <a href="https://console.mistral.ai/user/api-keys">here</a>.</li><ul><li>Requires personal Mistral API key.</li><li>WARNING: Will send your chats to Mistral!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with Mistral</li><li>You can apply for an API key <a href="https://console.mistral.ai/user/api-keys">here</a>.</li>
@@ -1689,7 +1700,7 @@ model to get started
<strong>Mistral Medium model</strong><br>
-
+ <strong>Created by %1.</strong><br><ul><li>Published on %2.<li>This model has %3 likes.<li>This model has %4 downloads.<li>More info can be found <a href="https://huggingface.co/%5">here.</a></ul><strong>Created by %1.</strong><br><ul><li>Published on %2.<li>This model has %3 likes.<li>This model has %4 downloads.<li>More info can be found <a href="https://huggingface.co/%5">here.</a></ul>
@@ -1697,57 +1708,57 @@ model to get started
ModelSettings
-
+ Model模型
-
+ Model Settings模型设置
-
+ Clone克隆
-
+ Remove删除
-
+ Name名称
-
+ Model File模型文件
-
+ System Prompt系统提示词
-
+ Prefixed at the beginning of every conversation. Must contain the appropriate framing tokens.每次对话开始时的前缀
-
+ Prompt Template提示词模版
-
+ The template that wraps every prompt.包装每个提示的模板
-
+ Must contain the string "%1" to be replaced with the user's input.必须包含字符串 "%1" 替换为用户的's 输入.
@@ -1757,37 +1768,37 @@ optional image
添加可选图片
-
+ Chat Name Prompt聊天名称提示
-
+ Prompt used to automatically generate chat names.用于自动生成聊天名称的提示。
-
+ Suggested FollowUp Prompt建议的后续提示
-
+ Prompt used to generate suggested follow-up questions.用于生成建议的后续问题的提示。
-
+ Context Length上下文长度
-
+ Number of input and output tokens the model sees.模型看到的输入和输出令牌的数量。
-
+ Maximum combined prompt/response tokens before information is lost.
Using more context than the model was trained on will yield poor results.
NOTE: Does not take effect until you reload the model.
@@ -1796,128 +1807,128 @@ NOTE: Does not take effect until you reload the model.
注意:在重新加载模型之前不会生效。
-
+ Temperature温度
-
+ Randomness of model output. Higher -> more variation.模型输出的随机性。更高->更多的变化。
-
+ Temperature increases the chances of choosing less likely tokens.
NOTE: Higher temperature gives more creative but less predictable outputs.温度增加了选择不太可能的token的机会。
注:温度越高,输出越有创意,但预测性越低。
-
+ Top-PTop-P
-
+ Nucleus Sampling factor. Lower -> more predictable.核子取样系数。较低->更具可预测性。
-
+ Only the most likely tokens up to a total probability of top_p can be chosen.
NOTE: Prevents choosing highly unlikely tokens.只能选择总概率高达top_p的最有可能的令牌。
注意:防止选择极不可能的token。
-
+ Min-PMin-P
-
+ Minimum token probability. Higher -> more predictable.最小令牌概率。更高 -> 更可预测。
-
+ Sets the minimum relative probability for a token to be considered.设置被考虑的标记的最小相对概率。
-
+ Top-KTop-K
-
+ Size of selection pool for tokens.令牌选择池的大小。
-
+ Only the top K most likely tokens will be chosen from.仅从最可能的前 K 个标记中选择
-
+ Max Length最大长度
-
+ Maximum response length, in tokens.最大响应长度(以令牌为单位)
-
+ Prompt Batch Size提示词大小
-
+ The batch size used for prompt processing.用于快速处理的批量大小。
-
+ Amount of prompt tokens to process at once.
NOTE: Higher values can speed up reading prompts but will use more RAM.一次要处理的提示令牌数量。
注意:较高的值可以加快读取提示,但会使用更多的RAM。
-
+ Repeat Penalty重复惩罚
-
+ Repetition penalty factor. Set to 1 to disable.重复处罚系数。设置为1可禁用。
-
+ Repeat Penalty Tokens重复惩罚数
-
+ Number of previous tokens used for penalty.用于惩罚的先前令牌数量。
-
+ GPU LayersGPU 层
-
+ Number of model layers to load into VRAM.要加载到VRAM中的模型层数。
-
+ How many model layers to load into VRAM. Decrease this if GPT4All runs out of VRAM while loading this model.
Lower values increase CPU load and RAM usage, and make inference slower.
NOTE: Does not take effect until you reload the model.
@@ -1929,134 +1940,134 @@ NOTE: Does not take effect until you reload the model.
ModelsView
-
+ No Models Installed无模型
-
+ Install a model to get started using GPT4All安装模型并开始使用
-
-
+
+ + Add Model+ 添加模型
-
+ Shows the add model view查看增加到模型
-
+ Installed Models已安装的模型
-
+ Locally installed chat models本地安装的聊天
-
+ Model file模型文件
-
+ Model file to be downloaded待下载的模型
-
+ Description描述
-
+ File description文件描述
-
+ Cancel取消
-
+ Resume继续
-
+ Stop/restart/start the download停止/重启/开始下载
-
+ Remove删除
-
+ Remove model from filesystem从系统中删除模型
-
-
+
+ Install按照
-
+ Install online model安装在线模型
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font><strong><font size="1"><a href="#error">Error</a></strong></font>
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font><strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font>
-
+ ERROR: $API_KEY is empty.错误:$API_KEY 为空
-
+ ERROR: $BASE_URL is empty.错误:$BASE_URL 为空
-
+ enter $BASE_URL输入 $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.错误:$MODEL_NAME为空
-
+ enter $MODEL_NAME输入:$MODEL_NAME
-
+ %1 GB%1 GB
-
+ ??
@@ -2065,7 +2076,7 @@ NOTE: Does not take effect until you reload the model.
<a href="#错误">错误</a>
-
+ Describes an error that occurred when downloading描述下载时发生的错误
@@ -2082,65 +2093,65 @@ NOTE: Does not take effect until you reload the model.
GB) 你的系统需要 (
-
+ Error for incompatible hardware硬件不兼容的错误
-
+ Download progressBar下载进度
-
+ Shows the progress made in the download显示下载进度
-
+ Download speed下载速度
-
+ Download speed in bytes/kilobytes/megabytes per second下载速度 b/kb/mb /s
-
+ Calculating...计算中...
-
-
-
-
+
+
+
+ Whether the file hash is being calculated是否正在计算文件哈希
-
+ Busy indicator繁忙程度
-
+ Displayed when the file hash is being calculated在计算文件哈希时显示
-
+ enter $API_KEY输入 $API_KEY
-
+ File size文件大小
-
+ RAM required需要 RAM
@@ -2149,17 +2160,17 @@ NOTE: Does not take effect until you reload the model.
GB
-
+ Parameters参数
-
+ Quant量化
-
+ Type类型
@@ -2167,12 +2178,12 @@ NOTE: Does not take effect until you reload the model.
MyFancyLink
-
+ Fancy link精选链接
-
+ A stylized link样式化链接
@@ -2180,7 +2191,7 @@ NOTE: Does not take effect until you reload the model.
MySettingsStack
-
+ Please choose a directory请选择目录
@@ -2188,12 +2199,12 @@ NOTE: Does not take effect until you reload the model.
MySettingsTab
-
+ Restore Defaults恢复初始化
-
+ Restores settings dialog to a default state将设置对话框恢复为默认状态
@@ -2201,12 +2212,12 @@ NOTE: Does not take effect until you reload the model.
NetworkDialog
-
+ Contribute data to the GPT4All Opensource Datalake.向GPT4All开源数据湖贡献数据
-
+ By enabling this feature, you will be able to participate in the democratic process of training a large language model by contributing data for future model improvements.
When a GPT4All model responds to you and you have opted-in, your conversation will be sent to the GPT4All Open Source Datalake. Additionally, you can like/dislike its response. If you dislike a response, you can suggest an alternative response. This data will be collected and aggregated in the GPT4All Datalake.
@@ -2219,47 +2230,47 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
注意:通过启用此功能,您将把数据发送到 GPT4All 开源数据湖。启用此功能后,您不应该期望聊天隐私。但是,如果您愿意,您应该期望可选的归因。您的聊天数据将公开供任何人下载,并将被 Nomic AI 用于改进未来的 GPT4All 模型。Nomic AI 将保留与您的数据相关的所有归因信息,并且您将被视为使用您的数据的任何 GPT4All 模型发布的贡献者!
-
+ Terms for opt-in选择加入的条款
-
+ Describes what will happen when you opt-in描述选择加入时会发生的情况
-
+ Please provide a name for attribution (optional)填写名称属性 (可选)
-
+ Attribution (optional)属性 (可选)
-
+ Provide attribution提供属性
-
+ Enable启用
-
+ Enable opt-in启用选择加入
-
+ Cancel取消
-
+ Cancel opt-in取消加入
@@ -2267,17 +2278,17 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
NewVersionDialog
-
+ New version is available新版本可选
-
+ Update更新
-
+ Update to new version更新到新版本
@@ -2285,17 +2296,17 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
PopupDialog
-
+ Reveals a shortlived help balloon显示一个短暂的帮助气球
-
+ Busy indicator繁忙程度
-
+ Displayed when the popup is showing busy在弹出窗口显示忙碌时显示
@@ -2310,28 +2321,28 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
SettingsView
-
-
+
+ Settings设置
-
+ Contains various application settings包含各种应用程序设置
-
+ Application应用
-
+ Model模型
-
+ LocalDocs本地文档
@@ -2339,7 +2350,7 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
StartupDialog
-
+ Welcome!欢迎!
@@ -2354,7 +2365,7 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
### 贡献者
-
+ ### Release notes
%1### Contributors
%2
@@ -2363,17 +2374,17 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
%2
-
+ Release notes发布日志
-
+ Release notes for this version本版本发布日志
-
+ ### Opt-ins for anonymous usage analytics and datalake
By enabling these features, you will be able to participate in the democratic process of training a
large language model by contributing data for future model improvements.
@@ -2398,71 +2409,71 @@ model release that uses your data!
模型发布的贡献者!
-
+ Terms for opt-in选择加入选项
-
+ Describes what will happen when you opt-in描述选择加入时会发生的情况
-
-
+
+ Opt-in for anonymous usage statistics允许选择加入匿名使用统计数据
-
-
+
+ Yes是
-
+ Allow opt-in for anonymous usage statistics允许选择加入匿名使用统计数据
-
-
+
+ No否
-
+ Opt-out for anonymous usage statistics退出匿名使用统计数据
-
+ Allow opt-out for anonymous usage statistics允许选择退出匿名使用统计数据
-
-
+
+ Opt-in for network加入网络
-
+ Allow opt-in for network允许选择加入网络
-
+ Allow opt-in anonymous sharing of chats to the GPT4All Datalake允许选择加入匿名共享聊天至 GPT4All 数据湖
-
+ Opt-out for network取消网络
-
+ Allow opt-out anonymous sharing of chats to the GPT4All Datalake允许选择退出将聊天匿名共享至 GPT4All 数据湖
@@ -2470,23 +2481,23 @@ model release that uses your data!
SwitchModelDialog
-
+ <b>Warning:</b> changing the model will erase the current conversation. Do you wish to continue?<b>警告:</b> 更改模型将删除当前对话。您想继续吗?
-
+ Continue继续
-
+ Continue with model loading模型载入时继续
-
-
+
+ Cancel取消
@@ -2494,32 +2505,32 @@ model release that uses your data!
ThumbsDownDialog
-
+ Please edit the text below to provide a better response. (optional)请编辑下方文本以提供更好的回复。(可选)
-
+ Please provide a better response...提供更好回答...
-
+ Submit提交
-
+ Submits the user's response提交用户响应
-
+ Cancel取消
-
+ Closes the response dialog关闭的对话
@@ -2583,125 +2594,125 @@ model release that uses your data!
检查链接 <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> 寻求.
-
+ GPT4All v%1GPT4All v%1
-
+ <h3>Encountered an error starting up:</h3><br><i>"Incompatible hardware detected."</i><br><br>Unfortunately, your CPU does not meet the minimal requirements to run this program. In particular, it does not support AVX intrinsics which this program requires to successfully run a modern large language model. The only solution at this time is to upgrade your hardware to a more modern CPU.<br><br>See here for more information: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a><h3>启动时遇到错误:</h3><br><i>“检测到不兼容的硬件。”</i><br><br>很遗憾,您的 CPU 不满足运行此程序的最低要求。特别是,它不支持此程序成功运行现代大型语言模型所需的 AVX 内在函数。目前唯一的解决方案是将您的硬件升级到更现代的 CPU。<br><br>有关更多信息,请参阅此处:<a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions>>https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a>
-
+ <h3>Encountered an error starting up:</h3><br><i>"Inability to access settings file."</i><br><br>Unfortunately, something is preventing the program from accessing the settings file. This could be caused by incorrect permissions in the local app config directory where the settings file is located. Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help.<h3>启动时遇到错误:</h3><br><i>“无法访问设置文件。”</i><br><br>不幸的是,某些东西阻止程序访问设置文件。这可能是由于设置文件所在的本地应用程序配置目录中的权限不正确造成的。请查看我们的<a href="https://discord.gg/4M2QFmTt2k">discord 频道</a> 以获取帮助。
-
+ Connection to datalake failed.链接数据湖失败
-
+ Saving chats.保存对话
-
+ Network dialog网络对话
-
+ opt-in to share feedback/conversations选择加入以共享反馈/对话
-
+ Home view主页
-
+ Home view of application主页
-
+ Home主页
-
+ Chat view对话视图
-
+ Chat view to interact with models聊天视图可与模型互动
-
+ Chats对话
-
-
+
+ Models模型
-
+ Models view for installed models已安装模型的页面
-
-
+
+ LocalDocs本地文档
-
+ LocalDocs view to configure and use local docsLocalDocs视图可配置和使用本地文档
-
-
+
+ Settings设置
-
+ Settings view for application configuration设置页面
-
+ The datalake is enabled数据湖已开启
-
+ Using a network model使用联网模型
-
+ Server mode is enabled服务器模式已开
-
+ Installed models安装模型
-
+ View of installed models查看已安装模型
diff --git a/gpt4all-chat/translations/gpt4all_zh_TW.ts b/gpt4all-chat/translations/gpt4all_zh_TW.ts
index f0fd630e..e6473e0e 100644
--- a/gpt4all-chat/translations/gpt4all_zh_TW.ts
+++ b/gpt4all-chat/translations/gpt4all_zh_TW.ts
@@ -4,62 +4,62 @@
AddCollectionView
-
+ ← Existing Collections← 現有收藏
-
+ Add Document Collection新增收藏文件
-
+ Add a folder containing plain text files, PDFs, or Markdown. Configure additional extensions in Settings.新增一個含有純文字檔案、PDF 與 Markdown 文件的資料夾。可在設定上增加文件副檔名。
-
+ Please choose a directory請選擇一個資料夾
-
+ Name名稱
-
+ Collection name...收藏名稱......
-
+ Name of the collection to add (Required)新增的收藏名稱(必填)
-
+ Folder資料夾
-
+ Folder path...資料夾路徑......
-
+ Folder path to documents (Required)文件所屬的資料夾路徑(必填)
-
+ Browse瀏覽
-
+ Create Collection建立收藏
@@ -67,289 +67,289 @@
AddModelView
-
+ ← Existing Models← 現有模型
-
+ Explore Models探索模型
-
+ Discover and download models by keyword search...透過關鍵字搜尋探索並下載模型......
-
+ Text field for discovering and filtering downloadable models用於探索與過濾可下載模型的文字字段
-
+ Searching · %1搜尋 · %1
-
+ Initiate model discovery and filtering探索與過濾模型
-
+ Triggers discovery and filtering of models觸發探索與過濾模型
-
+ Default預設
-
+ Likes讚
-
+ Downloads下載次數
-
+ Recent最新
-
+ Sort by: %1排序依據:%1
-
+ Asc升序
-
+ Desc降序
-
+ Sort dir: %1排序順序:%1
-
+ None無
-
+ Limit: %1上限:%1
-
+ Network error: could not retrieve %1網路錯誤:無法取得 %1
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font><strong><font size="1"><a href="#error">錯誤</a></strong></font>
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font><strong><font size="2">警告:不推薦在您的硬體上運作。模型需要比較多的記憶體(%1 GB),但您的系統記憶體空間不足(%2)。</strong></font>
-
+ %1 GB%1 GB
-
-
+
+ ??
-
-
+
+ Busy indicator參考自 https://terms.naer.edu.tw忙線指示器
-
+ Displayed when the models request is ongoing當模型請求正在進行時顯示
-
+ Model file模型檔案
-
+ Model file to be downloaded即將下載的模型檔案
-
+ Description描述
-
+ File description檔案描述
-
+ Cancel取消
-
+ Resume恢復
-
+ Download下載
-
+ Stop/restart/start the download停止/重啟/開始下載
-
+ Remove移除
-
+ Remove model from filesystem從檔案系統移除模型
-
-
+
+ Install安裝
-
+ Install online model安裝線上模型
-
+ Describes an error that occurred when downloading解釋下載時發生的錯誤
-
+ Error for incompatible hardware錯誤,不相容的硬體
-
+ Download progressBar下載進度條
-
+ Shows the progress made in the download顯示下載進度
-
+ Download speed下載速度
-
+ Download speed in bytes/kilobytes/megabytes per second下載速度每秒 bytes/kilobytes/megabytes
-
+ Calculating...計算中......
-
-
-
-
+
+
+
+ Whether the file hash is being calculated是否正在計算檔案雜湊
-
+ Displayed when the file hash is being calculated計算檔案雜湊值時顯示
-
+ ERROR: $API_KEY is empty.錯誤:$API_KEY 未填寫。
-
+ enter $API_KEY請輸入 $API_KEY
-
+ ERROR: $BASE_URL is empty.錯誤:$BASE_URL 未填寫。
-
+ enter $BASE_URL請輸入 $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.錯誤:$MODEL_NAME 未填寫。
-
+ enter $MODEL_NAME請輸入 $MODEL_NAME
-
+ File size檔案大小
-
+ RAM required所需的記憶體
-
+ Parameters參數
-
+ Quant量化
-
+ Type類型
@@ -357,22 +357,22 @@
ApplicationSettings
-
+ Application應用程式
-
+ Network dialog資料湖泊計畫對話視窗
-
+ opt-in to share feedback/conversations分享回饋/對話計畫
-
+ ERROR: Update system could not find the MaintenanceTool used<br>
to check for updates!<br><br>
Did you install this application using the online installer? If so,<br>
@@ -387,223 +387,223 @@
如果您無法順利啟動,您可能得重新安裝本應用程式。
-
+ Error dialog錯誤對話視窗
-
+ Application Settings應用程式設定
-
+ General一般
-
+ Theme主題
-
+ The application color scheme.應用程式的配色方案。
-
+ Dark暗色
-
+ Light亮色
-
+ LegacyDark傳統暗色
-
+ Font Size字體大小
-
+ The size of text in the application.應用程式中的字體大小。
-
+ Small小
-
+ Medium中
-
+ Large大
-
+ Language and Locale語言與區域設定
-
+ The language and locale you wish to use.您希望使用的語言與區域設定。
-
+ System Locale系統語系
-
+ Device裝置
-
+ Default Model預設模型
-
+ The preferred model for new chats. Also used as the local server fallback.用於新交談的預設模型。也用於作為本機伺服器後援使用。
-
+ Suggestion Mode建議模式
-
+ When chatting with LocalDocs當使用「我的文件」交談時
-
+ Whenever possible視情況允許
-
+ Never永不
-
+ Generate suggested follow-up questions at the end of responses.在回覆末尾生成後續建議的問題。
-
+ The compute device used for text generation.用於生成文字的計算裝置。
-
-
+
+ Application default應用程式預設值
-
+ Download Path下載路徑
-
+ Where to store local models and the LocalDocs database.儲存本機模型與「我的文件」資料庫的位置。
-
+ Browse瀏覽
-
+ Choose where to save model files選擇儲存模型檔案的位置
-
+ Enable Datalake啟用資料湖泊
-
+ Send chats and feedback to the GPT4All Open-Source Datalake.將交談與回饋傳送到 GPT4All 開放原始碼資料湖泊。
-
+ Advanced進階
-
+ CPU Threads中央處理器(CPU)線程
-
+ The number of CPU threads used for inference and embedding.用於推理與嵌入的中央處理器線程數。
-
+ Save Chat Context儲存交談語境
-
+ Save the chat model's state to disk for faster loading. WARNING: Uses ~2GB per chat.將交談模型的狀態儲存到磁碟以加快載入速度。警告:每次交談使用約 2GB。
-
+ Enable Local Server啟用本機伺服器
-
+ Expose an OpenAI-Compatible server to localhost. WARNING: Results in increased resource usage.將 OpenAI 相容伺服器公開給本機。警告:導致資源使用增加。
-
+ API Server PortAPI 伺服器埠口
-
+ The port to use for the local server. Requires restart.用於本機伺服器的埠口。需要重新啟動。
-
+ Check For Updates檢查更新
-
+ Manually check for an update to GPT4All.手動檢查 GPT4All 的更新。
-
+ Updates更新
@@ -611,13 +611,13 @@
Chat
-
-
+
+ New Chat新的交談
-
+ Server Chat伺服器交談
@@ -625,12 +625,12 @@
ChatAPIWorker
-
+ ERROR: Network error occurred while connecting to the API server錯誤:網路錯誤,無法連線到目標 API 伺服器
-
+ ChatAPIWorker::handleFinished got HTTP Error %1 %2ChatAPIWorker::handleFinished 遇到一個 HTTP 錯誤 %1 %2
@@ -638,62 +638,62 @@
ChatDrawer
-
+ Drawer側邊欄
-
+ Main navigation drawer主要導航側邊欄
-
+ + New Chat+ 新的交談
-
+ Create a new chat建立新的交談
-
+ Select the current chat or edit the chat when in edit mode選擇目前交談或在編輯模式下編輯交談
-
+ Edit chat name修改對話名稱
-
+ Save chat name儲存對話名稱
-
+ Delete chat刪除對話
-
+ Confirm chat deletion確定刪除對話
-
+ Cancel chat deletion取消刪除對話
-
+ List of chats交談列表
-
+ List of chats in the drawer dialog側邊欄對話視窗的交談列表
@@ -701,32 +701,32 @@
ChatListModel
-
+ TODAY今天
-
+ THIS WEEK這星期
-
+ THIS MONTH這個月
-
+ LAST SIX MONTHS前六個月
-
+ THIS YEAR今年
-
+ LAST YEAR去年
@@ -734,329 +734,329 @@
ChatView
-
+ <h3>Warning</h3><p>%1</p><h3>警告</h3><p>%1</p>
-
+ Switch model dialog切換模型對話視窗
-
+ Warn the user if they switch models, then context will be erased警告使用者如果切換模型,則語境將被刪除
-
+ Conversation copied to clipboard.對話已複製到剪貼簿。
-
+ Code copied to clipboard.程式碼已複製到剪貼簿。
-
+ Chat panel交談面板
-
+ Chat panel with options具有選項的交談面板
-
+ Reload the currently loaded model重新載入目前已載入的模型
-
+ Eject the currently loaded model彈出目前載入的模型
-
+ No model installed.沒有已安裝的模型。
-
+ Model loading error.模型載入時發生錯誤。
-
+ Waiting for model...等待模型中......
-
+ Switching context...切換語境中......
-
+ Choose a model...選擇一個模型......
-
+ Not found: %1不存在:%1
-
-
+
+ Reload · %1重新載入 · %1
-
+ Loading · %1載入中 · %1
-
+ Load · %1 (default) →載入 · %1 (預設) →
-
+ The top item is the current model最上面的那項是目前使用的模型
-
-
+
+ LocalDocs我的文件
-
+ Add documents新增文件
-
+ add collections of documents to the chat將文件集合新增至交談中
-
+ Load the default model載入預設模型
-
+ Loads the default model which can be changed in settings預設模型可於設定中變更
-
+ No Model Installed沒有已安裝的模型
-
+ GPT4All requires that you install at least one
model to get startedGPT4All 要求您至少安裝一個
模型開始
-
+ Install a Model安裝一個模型
-
+ Shows the add model view顯示新增模型視圖
-
+ Conversation with the model與模型對話
-
+ prompt / response pairs from the conversation對話中的提示詞 / 回覆組合
-
+ GPT4AllGPT4All
-
+ You您
-
+ response stopped ...回覆停止......
-
+ retrieving localdocs: %1 ...檢索本機文件中:%1 ......
-
+ searching localdocs: %1 ...搜尋本機文件中:%1 ......
-
+ processing ...處理中......
-
+ generating response ...生成回覆......
-
+ generating questions ...生成問題......
-
-
+
+ Copy複製
-
+ Copy Message複製訊息
-
+ Disable markdown停用 Markdown
-
+ Enable markdown啟用 Markdown
-
+ Thumbs up讚
-
+ Gives a thumbs up to the response對這則回覆比讚
-
+ Thumbs down倒讚
-
+ Opens thumbs down dialog開啟倒讚對話視窗
-
+ Suggested follow-ups後續建議
-
+ Erase and reset chat session刪除並重置交談會話
-
+ Copy chat session to clipboard複製交談會議到剪貼簿
-
+ Redo last chat response復原上一個交談回覆
-
+ Stop generating停止生成
-
+ Stop the current response generation停止當前回覆生成
-
+ Reloads the model重新載入模型
-
+ <h3>Encountered an error loading model:</h3><br><i>"%1"</i><br><br>Model loading failures can happen for a variety of reasons, but the most common causes include a bad file format, an incomplete or corrupted download, the wrong file type, not enough system RAM or an incompatible model type. Here are some suggestions for resolving the problem:<br><ul><li>Ensure the model file has a compatible format and type<li>Check the model file is complete in the download folder<li>You can find the download folder in the settings dialog<li>If you've sideloaded the model ensure the file is not corrupt by checking md5sum<li>Read more about what models are supported in our <a href="https://docs.gpt4all.io/">documentation</a> for the gui<li>Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help<h3>載入模型時發生錯誤:</h3><br><i>"%1"</i><br><br>導致模型載入失敗的原因可能有很多種,但絕大多數的原因是檔案格式損毀、下載的檔案不完整、檔案類型錯誤、系統RAM空間不足或不相容的模型類型。這裡有些建議可供疑難排解:<br><ul><li>確保使用的模型是相容的格式與類型<li>檢查位於下載資料夾的檔案是否完整<li>您可以從設定中找到您所設定的「下載資料夾路徑」<li>如果您有側載模型,請利用 md5sum 等工具確保您的檔案是完整的<li>想了解更多關於我們所支援的模型資訊,煩請詳閱<a href="https://docs.gpt4all.io/">本文件</a>。<li>歡迎洽詢我們的 <a href="https://discord.gg/4M2QFmTt2k">Discord 伺服器</a> 以尋求幫助
-
+ restoring from text ...從文字中恢復......
-
+ %n Source(s)%n 來源
-
+ Send a message...傳送一則訊息......
-
+ Load a model to continue...載入模型以繼續......
-
+ Send messages/prompts to the model向模型傳送訊息/提示詞
-
+ Cut剪下
-
+ Paste貼上
-
+ Select All全選
-
+ Send message傳送訊息
-
+ Sends the message/prompt contained in textfield to the model將文字欄位中包含的訊息/提示詞傳送到模型
@@ -1064,36 +1064,36 @@ model to get started
CollectionsDrawer
-
+ Warning: searching collections while indexing can return incomplete results警告:在索引時搜尋收藏可能會傳回不完整的結果
-
+ %n file(s)%n 個檔案
-
+ %n word(s)%n 個字
-
+ Updating更新中
-
+ + Add Docs+ 新增文件
-
+ Select a collection to make it available to the chat model.選擇一個收藏以使其可供交談模型使用。
@@ -1101,37 +1101,37 @@ model to get started
Download
-
+ Model "%1" is installed successfully.模型「%1」已安裝成功。
-
+ ERROR: $MODEL_NAME is empty.錯誤:$MODEL_NAME 未填寫。
-
+ ERROR: $API_KEY is empty.錯誤:$API_KEY 未填寫。
-
+ ERROR: $BASE_URL is invalid.錯誤:$BASE_URL 無效。
-
+ ERROR: Model "%1 (%2)" is conflict.錯誤:模型「%1 (%2)」發生衝突。
-
+ Model "%1 (%2)" is installed successfully.模型「%1(%2)」已安裝成功。
-
+ Model "%1" is removed.模型「%1」已移除。
@@ -1139,92 +1139,92 @@ model to get started
HomeView
-
+ Welcome to GPT4All歡迎使用 GPT4All
-
+ The privacy-first LLM chat application隱私第一的大型語言模型交談應用程式
-
+ Start chatting開始交談
-
+ Start Chatting開始交談
-
+ Chat with any LLM與任何大型語言模型交談
-
+ LocalDocs我的文件
-
+ Chat with your local files使用「我的文件」來交談
-
+ Find Models搜尋模型
-
+ Explore and download models瀏覽與下載模型
-
+ Latest news最新消息
-
+ Latest news from GPT4All從 GPT4All 來的最新消息
-
+ Release Notes版本資訊
-
+ Documentation文件
-
+ DiscordDiscord
-
+ X (Twitter)X (Twitter)
-
+ GithubGithub
-
+ nomic.ainomic.ai
-
+ Subscribe to Newsletter訂閱電子報
@@ -1232,117 +1232,117 @@ model to get started
LocalDocsSettings
-
+ LocalDocs我的文件
-
+ LocalDocs Settings我的文件設定
-
+ Indexing索引中
-
+ Allowed File Extensions允許的副檔名
-
+ Comma-separated list. LocalDocs will only attempt to process files with these extensions.以逗號分隔的列表。「我的文件」將僅嘗試處理具有這些副檔名的檔案。
-
+ Embedding嵌入
-
+ Use Nomic Embed API使用 Nomic 嵌入 API
-
+ Embed documents using the fast Nomic API instead of a private local model. Requires restart.使用快速的 Nomic API 而不是本機私有模型嵌入文件。需要重新啟動。
-
+ Nomic API KeyNomic API 金鑰
-
+ API key to use for Nomic Embed. Get one from the Atlas <a href="https://atlas.nomic.ai/cli-login">API keys page</a>. Requires restart.用於 Nomic Embed 的 API 金鑰。從 Atlas <a href="https://atlas.nomic.ai/cli-login">API 金鑰頁面</a>取得一個。需要重新啟動。
-
+ Embeddings Device嵌入裝置
-
+ The compute device used for embeddings. Requires restart.用於嵌入的計算裝置。需要重新啟動。
-
+ Application default應用程式預設值
-
+ Display顯示
-
+ Show Sources查看來源
-
+ Display the sources used for each response.顯示每則回覆所使用的來源。
-
+ Advanced進階
-
+ Warning: Advanced usage only.警告:僅限進階使用。
-
+ Values too large may cause localdocs failure, extremely slow responses or failure to respond at all. Roughly speaking, the {N chars x N snippets} are added to the model's context window. More info <a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">here</a>.設定太大的數值可能會導致「我的文件」處理失敗、反應速度極慢或根本無法回覆。簡單地說,這會將 {N 個字元 x N 個片段} 被添加到模型的語境視窗中。更多資訊<a href="https://docs.gpt4all.io/gpt4all_desktop/localdocs.html">此處</a>。
-
+ Document snippet size (characters)文件片段大小(字元)
-
+ Number of characters per document snippet. Larger numbers increase likelihood of factual responses, but also result in slower generation.每個文件片段的字元數。較大的數字會增加實際反應的可能性,但也會導致生成速度變慢。
-
+ Max document snippets per prompt每個提示詞的最大文件片段
-
+ Max best N matches of retrieved document snippets to add to the context for prompt. Larger numbers increase likelihood of factual responses, but also result in slower generation.新增至提示詞語境中的檢索到的文件片段的最大 N 個符合的項目。較大的數字會增加實際反應的可能性,但也會導致生成速度變慢。
@@ -1350,151 +1350,151 @@ model to get started
LocalDocsView
-
+ LocalDocs我的文件
-
+ Chat with your local files使用「我的文件」來交談
-
+ + Add Collection+ 新增收藏
-
+ <h3>ERROR: The LocalDocs database cannot be accessed or is not valid.</h3><br><i>Note: You will need to restart after trying any of the following suggested fixes.</i><br><ul><li>Make sure that the folder set as <b>Download Path</b> exists on the file system.</li><li>Check ownership as well as read and write permissions of the <b>Download Path</b>.</li><li>If there is a <b>localdocs_v2.db</b> file, check its ownership and read/write permissions, too.</li></ul><br>If the problem persists and there are any 'localdocs_v*.db' files present, as a last resort you can<br>try backing them up and removing them. You will have to recreate your collections, however.<h3>錯誤:「我的文件」資料庫已無法存取或已損壞。</h3><br><i>提醒:執行完以下任何疑難排解的動作後,請務必重新啟動應用程式。</i><br><ul><li>請確保<b>「下載路徑」</b>所指向的資料夾確實存在於檔案系統當中。</li><li>檢查 <b>「下載路徑」</b>所指向的資料夾,確保其「擁有者」為您本身,以及確保您對該資料夾擁有讀寫權限。</li><li>如果該資料夾內存在一份名為 <b>localdocs_v2.db</b> 的檔案,請同時確保您對其擁有讀寫權限。</li></ul><br>如果問題依舊存在,且該資料夾內存在與「localdocs_v*.db」名稱相關的檔案,請嘗試備份並移除它們。<br>雖然這樣一來,您恐怕得著手重建您的收藏,但這將或許能夠解決這份錯誤。
-
+ No Collections Installed沒有已安裝的收藏
-
+ Install a collection of local documents to get started using this feature安裝本機文件收藏以開始使用此功能
-
+ + Add Doc Collection+ 新增文件收藏
-
+ Shows the add model view查看新增的模型視圖
-
+ Indexing progressBar索引進度條
-
+ Shows the progress made in the indexing顯示索引進度
-
+ ERROR錯誤
-
+ INDEXING索引中
-
+ EMBEDDING嵌入中
-
+ REQUIRES UPDATE必須更新
-
+ READY已就緒
-
+ INSTALLING安裝中
-
+ Indexing in progress正在索引
-
+ Embedding in progress正在嵌入
-
+ This collection requires an update after version change該收藏需要在版本變更後更新
-
+ Automatically reindexes upon changes to the folder若資料夾有變動,會自動重新索引
-
+ Installation in progress正在安裝中
-
+ %%
-
+ %n file(s)%n 個檔案
-
+ %n word(s)%n 個字
-
+ Remove移除
-
+ Rebuild重建
-
+ Reindex this folder from scratch. This is slow and usually not needed.重新索引該資料夾。這將會耗費許多時間並且通常不太需要這樣做。
-
+ Update更新
-
+ Update the collection to the new version. This is a slow operation.更新收藏。這將會耗費許多時間。
@@ -1502,67 +1502,78 @@ model to get started
ModelList
-
+
+
+ cannot open "%1": %2
+
+
+
+
+ cannot create "%1": %2
+
+
+
+ %1 (%2)%1(%2)
-
+ <strong>OpenAI-Compatible API Model</strong><br><ul><li>API Key: %1</li><li>Base URL: %2</li><li>Model Name: %3</li></ul><strong>OpenAI API 相容模型</strong><br><ul><li>API 金鑰:%1</li><li>基底 URL:%2</li><li>模型名稱:%3</li></ul>
-
+ <ul><li>Requires personal OpenAI API key.</li><li>WARNING: Will send your chats to OpenAI!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with OpenAI</li><li>You can apply for an API key <a href="https://platform.openai.com/account/api-keys">here.</a></li><ul><li>需要個人的 OpenAI API 金鑰。</li><li>警告:這將會傳送您的交談紀錄到 OpenAI</li><li>您的 API 金鑰將被儲存在硬碟上</li><li>它只被用於與 OpenAI 進行通訊</li><li>您可以在<a href="https://platform.openai.com/account/api-keys">此處</a>申請一個 API 金鑰。</li>
-
+ <strong>OpenAI's ChatGPT model GPT-3.5 Turbo</strong><br> %1<strong>OpenAI 的 ChatGPT 模型 GPT-3.5 Turbo</strong><br> %1
-
+ <br><br><i>* Even if you pay OpenAI for ChatGPT-4 this does not guarantee API key access. Contact OpenAI for more info.<br><br><i>* 即使您已向 OpenAI 付費購買了 ChatGPT 的 GPT-4 模型使用權,但這也不能保證您能擁有 API 金鑰的使用權限。請聯繫 OpenAI 以查閱更多資訊。
-
+ <strong>OpenAI's ChatGPT model GPT-4</strong><br> %1 %2<strong>OpenAI 的 ChatGPT 模型 GPT-4</strong><br> %1 %2
-
+ <ul><li>Requires personal Mistral API key.</li><li>WARNING: Will send your chats to Mistral!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with Mistral</li><li>You can apply for an API key <a href="https://console.mistral.ai/user/api-keys">here</a>.</li><ul><li>需要個人的 Mistral API 金鑰。</li><li>警告:這將會傳送您的交談紀錄到 Mistral!</li><li>您的 API 金鑰將被儲存在硬碟上</li><li>它只被用於與 Mistral 進行通訊</li><li>您可以在<a href="https://console.mistral.ai/user/api-keys">此處</a>申請一個 API 金鑰。</li>
-
+ <strong>Mistral Tiny model</strong><br> %1<strong>Mistral 迷你模型</strong><br> %1
-
+ <strong>Mistral Small model</strong><br> %1<strong>Mistral 小型模型</strong><br> %1
-
+ <strong>Mistral Medium model</strong><br> %1<strong>Mistral 中型模型</strong><br> %1
-
+ <ul><li>Requires personal API key and the API base URL.</li><li>WARNING: Will send your chats to the OpenAI-compatible API Server you specified!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with the OpenAI-compatible API Server</li><ul><li>需要個人的 API 金鑰和 API 的基底 URL(Base URL)。</li><li>警告:這將會傳送您的交談紀錄到您所指定的 OpenAI API 相容伺服器</li><li>您的 API 金鑰將被儲存在硬碟上</li><li>它只被用於與其 OpenAI API 相容伺服器進行通訊</li>
-
+ <strong>Connect to OpenAI-compatible API server</strong><br> %1<strong>連線到 OpenAI API 相容伺服器</strong><br> %1
-
+ <strong>Created by %1.</strong><br><ul><li>Published on %2.<li>This model has %3 likes.<li>This model has %4 downloads.<li>More info can be found <a href="https://huggingface.co/%5">here.</a></ul><strong>模型作者:%1</strong><br><ul><li>發佈日期:%2<li>累積讚數:%3 個讚<li>下載次數:%4 次<li>更多資訊請查閱<a href="https://huggingface.co/%5">此處</a>。</ul>
@@ -1570,92 +1581,92 @@ model to get started
ModelSettings
-
+ Model模型
-
+ Model Settings模型設定
-
+ Clone複製
-
+ Remove移除
-
+ Name名稱
-
+ Model File模型檔案
-
+ System Prompt系統提示詞
-
+ Prefixed at the beginning of every conversation. Must contain the appropriate framing tokens.在每個對話的開頭加上前綴。必須包含適當的構建符元(framing tokens)。
-
+ Prompt Template提示詞模板
-
+ The template that wraps every prompt.包裝每個提示詞的模板。
-
+ Must contain the string "%1" to be replaced with the user's input.必須包含要替換為使用者輸入的字串「%1」。
-
+ Chat Name Prompt交談名稱提示詞
-
+ Prompt used to automatically generate chat names.用於自動生成交談名稱的提示詞。
-
+ Suggested FollowUp Prompt後續建議提示詞
-
+ Prompt used to generate suggested follow-up questions.用於生成後續建議問題的提示詞。
-
+ Context Length語境長度(Context Length)
-
+ Number of input and output tokens the model sees.模型看見的輸入與輸出的符元數量。
-
+ Maximum combined prompt/response tokens before information is lost.
Using more context than the model was trained on will yield poor results.
NOTE: Does not take effect until you reload the model.
@@ -1664,128 +1675,128 @@ NOTE: Does not take effect until you reload the model.
注意:重新載入模型後才會生效。
-
+ Temperature語境溫度(Temperature)
-
+ Randomness of model output. Higher -> more variation.模型輸出的隨機性。更高 -> 更多變化。
-
+ Temperature increases the chances of choosing less likely tokens.
NOTE: Higher temperature gives more creative but less predictable outputs.語境溫度會提高選擇不容易出現的符元機率。
注意:較高的語境溫度會生成更多創意,但輸出的可預測性會相對較差。
-
+ Top-P核心採樣(Top-P)
-
+ Nucleus Sampling factor. Lower -> more predictable.核心採樣因子。更低 -> 更可預測。
-
+ Only the most likely tokens up to a total probability of top_p can be chosen.
NOTE: Prevents choosing highly unlikely tokens.只選擇總機率約為核心採樣,最有可能性的符元。
注意:用於避免選擇不容易出現的符元。
-
+ Min-P最小符元機率(Min-P)
-
+ Minimum token probability. Higher -> more predictable.最小符元機率。更高 -> 更可預測。
-
+ Sets the minimum relative probability for a token to be considered.設定要考慮的符元的最小相對機率。
-
+ Top-K高頻率採樣機率(Top-K)
-
+ Size of selection pool for tokens.符元選擇池的大小。
-
+ Only the top K most likely tokens will be chosen from.只選擇前 K 個最有可能性的符元。
-
+ Max Length最大長度(Max Length)
-
+ Maximum response length, in tokens.最大響應長度(以符元為單位)。
-
+ Prompt Batch Size提示詞批次大小(Prompt Batch Size)
-
+ The batch size used for prompt processing.用於即時處理的批量大小。
-
+ Amount of prompt tokens to process at once.
NOTE: Higher values can speed up reading prompts but will use more RAM.一次處理的提示詞符元數量。
注意:較高的值可以加快讀取提示詞的速度,但會使用比較多的記憶體。
-
+ Repeat Penalty重複處罰(Repeat Penalty)
-
+ Repetition penalty factor. Set to 1 to disable.重複懲罰因子。設定為 1 以停用。
-
+ Repeat Penalty Tokens重複懲罰符元(Repeat Penalty Tokens)
-
+ Number of previous tokens used for penalty.之前用於懲罰的符元數量。
-
+ GPU Layers圖形處理器負載層(GPU Layers)
-
+ Number of model layers to load into VRAM.要載入到顯示記憶體中的模型層數。
-
+ How many model layers to load into VRAM. Decrease this if GPT4All runs out of VRAM while loading this model.
Lower values increase CPU load and RAM usage, and make inference slower.
NOTE: Does not take effect until you reload the model.
@@ -1797,218 +1808,218 @@ NOTE: Does not take effect until you reload the model.
ModelsView
-
+ No Models Installed沒有已安裝的模型
-
+ Install a model to get started using GPT4All安裝模型以開始使用 GPT4All
-
-
+
+ + Add Model+ 新增模型
-
+ Shows the add model view顯示新增模型視圖
-
+ Installed Models已安裝的模型
-
+ Locally installed chat models本機已安裝的交談模型
-
+ Model file模型檔案
-
+ Model file to be downloaded即將下載的模型檔案
-
+ Description描述
-
+ File description檔案描述
-
+ Cancel取消
-
+ Resume恢復
-
+ Stop/restart/start the download停止/重啟/開始下載
-
+ Remove移除
-
+ Remove model from filesystem從檔案系統移除模型
-
-
+
+ Install安裝
-
+ Install online model安裝線上模型
-
+ <strong><font size="1"><a href="#error">Error</a></strong></font><strong><font size="1"><a href="#error">錯誤</a></strong></font>
-
+ <strong><font size="2">WARNING: Not recommended for your hardware. Model requires more memory (%1 GB) than your system has available (%2).</strong></font><strong><font size="2">警告:不推薦在您的硬體上運作。模型需要比較多的記憶體(%1 GB),但您的系統記憶體空間不足(%2)。</strong></font>
-
+ %1 GB%1 GB
-
+ ??
-
+ Describes an error that occurred when downloading解釋下載時發生的錯誤
-
+ Error for incompatible hardware錯誤,不相容的硬體
-
+ Download progressBar下載進度條
-
+ Shows the progress made in the download顯示下載進度
-
+ Download speed下載速度
-
+ Download speed in bytes/kilobytes/megabytes per second下載速度每秒 bytes/kilobytes/megabytes
-
+ Calculating...計算中......
-
-
-
-
+
+
+
+ Whether the file hash is being calculated是否正在計算檔案雜湊
-
+ Busy indicator參考自 https://terms.naer.edu.tw忙線指示器
-
+ Displayed when the file hash is being calculated計算檔案雜湊值時顯示
-
+ ERROR: $API_KEY is empty.錯誤:$API_KEY 未填寫。
-
+ enter $API_KEY請輸入 $API_KEY
-
+ ERROR: $BASE_URL is empty.錯誤:$BASE_URL 未填寫。
-
+ enter $BASE_URL請輸入 $BASE_URL
-
+ ERROR: $MODEL_NAME is empty.錯誤:$MODEL_NAME 未填寫。
-
+ enter $MODEL_NAME請輸入 $MODEL_NAME
-
+ File size檔案大小
-
+ RAM required所需的記憶體
-
+ Parameters參數
-
+ Quant量化
-
+ Type類型
@@ -2016,12 +2027,12 @@ NOTE: Does not take effect until you reload the model.
MyFancyLink
-
+ Fancy link精緻網址
-
+ A stylized link個性化網址
@@ -2029,7 +2040,7 @@ NOTE: Does not take effect until you reload the model.
MySettingsStack
-
+ Please choose a directory請選擇一個資料夾
@@ -2037,12 +2048,12 @@ NOTE: Does not take effect until you reload the model.
MySettingsTab
-
+ Restore Defaults恢復預設值
-
+ Restores settings dialog to a default state恢復設定對話視窗到預設狀態
@@ -2050,12 +2061,12 @@ NOTE: Does not take effect until you reload the model.
NetworkDialog
-
+ Contribute data to the GPT4All Opensource Datalake.貢獻資料到 GPT4All 的開放原始碼資料湖泊。
-
+ By enabling this feature, you will be able to participate in the democratic process of training a large language model by contributing data for future model improvements.
When a GPT4All model responds to you and you have opted-in, your conversation will be sent to the GPT4All Open Source Datalake. Additionally, you can like/dislike its response. If you dislike a response, you can suggest an alternative response. This data will be collected and aggregated in the GPT4All Datalake.
@@ -2073,47 +2084,47 @@ NOTE: By turning on this feature, you will be sending your data to the GPT4All O
Nomic AI 將保留附加在您的資料上的所有署名訊息,並且您將被認可為任何使用您的資料的 GPT4All 模型版本的貢獻者!
-
+ Terms for opt-in計畫規範
-
+ Describes what will happen when you opt-in解釋當您加入計畫後,會發生什麼事情
-
+ Please provide a name for attribution (optional)請提供署名(非必填)
-
+ Attribution (optional)署名(非必填)
-
+ Provide attribution提供署名
-
+ Enable啟用
-
+ Enable opt-in加入計畫
-
+ Cancel取消
-
+ Cancel opt-in拒絕計畫
@@ -2121,17 +2132,17 @@ Nomic AI 將保留附加在您的資料上的所有署名訊息,並且您將
NewVersionDialog
-
+ New version is available發現新版本
-
+ Update更新
-
+ Update to new version更新版本
@@ -2139,18 +2150,18 @@ Nomic AI 將保留附加在您的資料上的所有署名訊息,並且您將
PopupDialog
-
+ Reveals a shortlived help balloon呼叫提示小幫手
-
+ Busy indicator參考自 https://terms.naer.edu.tw忙線指示器
-
+ Displayed when the popup is showing busy當彈出視窗忙碌時顯示
@@ -2158,28 +2169,28 @@ Nomic AI 將保留附加在您的資料上的所有署名訊息,並且您將
SettingsView
-
-
+
+ Settings設定
-
+ Contains various application settings內含多種應用程式設定
-
+ Application應用程式
-
+ Model模型
-
+ LocalDocs我的文件
@@ -2187,22 +2198,22 @@ Nomic AI 將保留附加在您的資料上的所有署名訊息,並且您將
StartupDialog
-
+ Welcome!歡迎使用!
-
+ Release notes版本資訊
-
+ Release notes for this version這個版本的版本資訊
-
+ ### Opt-ins for anonymous usage analytics and datalake
By enabling these features, you will be able to participate in the democratic process of training a
large language model by contributing data for future model improvements.
@@ -2230,35 +2241,35 @@ model release that uses your data!
Nomic AI 將保留附加在您的資料上的所有署名訊息,並且您將被認可為任何使用您的資料的 GPT4All 模型版本的貢獻者!
-
+ Terms for opt-in計畫規範
-
+ Describes what will happen when you opt-in解釋當您加入計畫後,會發生什麼事情
-
-
+
+ Yes是
-
-
+
+ No否
-
-
+
+ Opt-in for anonymous usage statistics匿名使用統計計畫
-
+ ### Release notes
%1### Contributors
%2
@@ -2267,43 +2278,43 @@ Nomic AI 將保留附加在您的資料上的所有署名訊息,並且您將
%2
-
+ Allow opt-in for anonymous usage statistics加入匿名使用統計計畫
-
+ Opt-out for anonymous usage statistics退出匿名使用統計計畫
-
+ Allow opt-out for anonymous usage statistics終止並退出匿名使用統計計畫
-
-
+
+ Opt-in for network資料湖泊計畫
-
+ Allow opt-in for network加入資料湖泊計畫
-
+ Opt-out for network退出資料湖泊計畫
-
+ Allow opt-in anonymous sharing of chats to the GPT4All Datalake開始將交談內容匿名分享到 GPT4All 資料湖泊
-
+ Allow opt-out anonymous sharing of chats to the GPT4All Datalake終止將交談內容匿名分享到 GPT4All 資料湖泊
@@ -2311,23 +2322,23 @@ Nomic AI 將保留附加在您的資料上的所有署名訊息,並且您將
SwitchModelDialog
-
+ <b>Warning:</b> changing the model will erase the current conversation. Do you wish to continue?<b>警告:</b> 變更模型將會清除目前對話內容。您真的想要繼續嗎?
-
+ Continue繼續
-
+ Continue with model loading繼續載入模型
-
-
+
+ Cancel取消
@@ -2335,32 +2346,32 @@ Nomic AI 將保留附加在您的資料上的所有署名訊息,並且您將
ThumbsDownDialog
-
+ Please edit the text below to provide a better response. (optional)請編輯以下文字,以提供更好的回覆。(非必填)
-
+ Please provide a better response...請提供一則更好的回覆......
-
+ Submit送出
-
+ Submits the user's response送出使用者的回覆
-
+ Cancel取消
-
+ Closes the response dialog關閉回覆對話視窗
@@ -2368,125 +2379,125 @@ Nomic AI 將保留附加在您的資料上的所有署名訊息,並且您將
main
-
+ GPT4All v%1GPT4All v%1
-
+ <h3>Encountered an error starting up:</h3><br><i>"Incompatible hardware detected."</i><br><br>Unfortunately, your CPU does not meet the minimal requirements to run this program. In particular, it does not support AVX intrinsics which this program requires to successfully run a modern large language model. The only solution at this time is to upgrade your hardware to a more modern CPU.<br><br>See here for more information: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a><h3>啟動時發生錯誤:</h3><br><i>「偵測到不相容的硬體。」</i><br><br>糟糕!您的中央處理器不符合運行所需的最低需求。尤其,它不支援本程式運行現代大型語言模型所需的 AVX 指令集。目前唯一的解決方案,只有更新您的中央處理器及其相關硬體裝置。<br><br>更多資訊請查閱:<a href="https://zh.wikipedia.org/wiki/AVX指令集">AVX 指令集 - 維基百科</a>
-
+ <h3>Encountered an error starting up:</h3><br><i>"Inability to access settings file."</i><br><br>Unfortunately, something is preventing the program from accessing the settings file. This could be caused by incorrect permissions in the local app config directory where the settings file is located. Check out our <a href="https://discord.gg/4M2QFmTt2k">discord channel</a> for help.<h3>啟動時發生錯誤:</h3><br><i>「無法存取設定檔。」</i><br><br>糟糕!有些東西正在阻止程式存取設定檔。這極為可能是由於設定檔所在的本機應用程式設定資料夾中的權限設定不正確所造成的。煩請洽詢我們的 <a href="https://discord.gg/4M2QFmTt2k">Discord 伺服器</a> 以尋求協助。
-
+ Connection to datalake failed.連線資料湖泊失敗。
-
+ Saving chats.儲存交談。
-
+ Network dialog資料湖泊計畫對話視窗
-
+ opt-in to share feedback/conversations分享回饋/對話計畫
-
+ Home view首頁視圖
-
+ Home view of application應用程式首頁視圖
-
+ Home首頁
-
+ Chat view查看交談
-
+ Chat view to interact with models模型互動交談視圖
-
+ Chats交談
-
-
+
+ Models模型
-
+ Models view for installed models已安裝模型的模型視圖
-
-
+
+ LocalDocs我的文件
-
+ LocalDocs view to configure and use local docs用於設定與使用我的文件的「我的文件」視圖
-
-
+
+ Settings設定
-
+ Settings view for application configuration應用程式設定視圖
-
+ The datalake is enabled資料湖泊已啟用
-
+ Using a network model使用一個網路模型
-
+ Server mode is enabled伺服器模式已啟用
-
+ Installed models已安裝的模型
-
+ View of installed models已安裝的模型視圖