mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2025-09-23 04:21:45 +00:00
Establish basic compiler warnings, and fix a few style issues (#3039)
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
This commit is contained in:
@@ -33,7 +33,7 @@ class Chat : public QObject
|
||||
Q_PROPERTY(ResponseState responseState READ responseState NOTIFY responseStateChanged)
|
||||
Q_PROPERTY(QList<QString> collectionList READ collectionList NOTIFY collectionListChanged)
|
||||
Q_PROPERTY(QString modelLoadingError READ modelLoadingError NOTIFY modelLoadingErrorChanged)
|
||||
Q_PROPERTY(QString tokenSpeed READ tokenSpeed NOTIFY tokenSpeedChanged);
|
||||
Q_PROPERTY(QString tokenSpeed READ tokenSpeed NOTIFY tokenSpeedChanged)
|
||||
Q_PROPERTY(QString deviceBackend READ deviceBackend NOTIFY loadedModelInfoChanged)
|
||||
Q_PROPERTY(QString device READ device NOTIFY loadedModelInfoChanged)
|
||||
Q_PROPERTY(QString fallbackReason READ fallbackReason NOTIFY loadedModelInfoChanged)
|
||||
|
@@ -585,7 +585,7 @@ bool ChatLLM::loadNewModel(const ModelInfo &modelInfo, QVariantMap &modelLoadPro
|
||||
|
||||
modelLoadProps.insert("$duration", modelLoadTimer.elapsed() / 1000.);
|
||||
return true;
|
||||
};
|
||||
}
|
||||
|
||||
bool ChatLLM::isModelLoaded() const
|
||||
{
|
||||
|
@@ -65,8 +65,8 @@ struct ChatItem
|
||||
Q_PROPERTY(bool thumbsDownState MEMBER thumbsDownState)
|
||||
Q_PROPERTY(QList<ResultInfo> sources MEMBER sources)
|
||||
Q_PROPERTY(QList<ResultInfo> consolidatedSources MEMBER consolidatedSources)
|
||||
Q_PROPERTY(QList<PromptAttachment> promptAttachments MEMBER promptAttachments);
|
||||
Q_PROPERTY(QString promptPlusAttachments READ promptPlusAttachments);
|
||||
Q_PROPERTY(QList<PromptAttachment> promptAttachments MEMBER promptAttachments)
|
||||
Q_PROPERTY(QString promptPlusAttachments READ promptPlusAttachments)
|
||||
|
||||
public:
|
||||
QString promptPlusAttachments() const
|
||||
|
@@ -296,10 +296,12 @@ static bool selectAllUncompletedChunks(QSqlQuery &q, QHash<IncompleteChunk, QStr
|
||||
while (q.next()) {
|
||||
QString collection = q.value(0).toString();
|
||||
IncompleteChunk ic {
|
||||
/*embedding_model =*/ q.value(1).toString(),
|
||||
/*chunk_id =*/ q.value(2).toInt(),
|
||||
/*folder_id =*/ q.value(3).toInt(),
|
||||
/*text =*/ q.value(4).toString(),
|
||||
/*EmbeddingKey*/ {
|
||||
.embedding_model = q.value(1).toString(),
|
||||
.chunk_id = q.value(2).toInt(),
|
||||
},
|
||||
/*folder_id =*/ q.value(3).toInt(),
|
||||
/*text =*/ q.value(4).toString(),
|
||||
};
|
||||
chunks[ic] << collection;
|
||||
}
|
||||
@@ -1659,7 +1661,7 @@ void Database::scanQueue()
|
||||
if (info.isPdf()) {
|
||||
QPdfDocument doc;
|
||||
if (doc.load(document_path) != QPdfDocument::Error::None) {
|
||||
qWarning() << "ERROR: Could not load pdf" << document_id << document_path;;
|
||||
qWarning() << "ERROR: Could not load pdf" << document_id << document_path;
|
||||
return updateFolderToIndex(folder_id, countForFolder);
|
||||
}
|
||||
title = doc.metaData(QPdfDocument::MetaDataField::Title).toString();
|
||||
|
@@ -176,7 +176,6 @@ private:
|
||||
QString m_author;
|
||||
QString m_subject;
|
||||
QString m_keywords;
|
||||
bool m_atStart;
|
||||
|
||||
// working state
|
||||
QString m_chunk; // has a trailing space for convenience
|
||||
|
@@ -504,7 +504,7 @@ ModelList::ModelList()
|
||||
connect(MySettings::globalInstance(), &MySettings::contextLengthChanged, this, &ModelList::updateDataForSettings);
|
||||
connect(MySettings::globalInstance(), &MySettings::gpuLayersChanged, this, &ModelList::updateDataForSettings);
|
||||
connect(MySettings::globalInstance(), &MySettings::repeatPenaltyChanged, this, &ModelList::updateDataForSettings);
|
||||
connect(MySettings::globalInstance(), &MySettings::repeatPenaltyTokensChanged, this, &ModelList::updateDataForSettings);;
|
||||
connect(MySettings::globalInstance(), &MySettings::repeatPenaltyTokensChanged, this, &ModelList::updateDataForSettings);
|
||||
connect(MySettings::globalInstance(), &MySettings::promptTemplateChanged, this, &ModelList::updateDataForSettings);
|
||||
connect(MySettings::globalInstance(), &MySettings::systemPromptChanged, this, &ModelList::updateDataForSettings);
|
||||
connect(&m_networkManager, &QNetworkAccessManager::sslErrors, this, &ModelList::handleSslErrors);
|
||||
@@ -520,12 +520,12 @@ QString ModelList::compatibleModelNameHash(QUrl baseUrl, QString modelName) {
|
||||
QCryptographicHash sha256(QCryptographicHash::Sha256);
|
||||
sha256.addData((baseUrl.toString() + "_" + modelName).toUtf8());
|
||||
return sha256.result().toHex();
|
||||
};
|
||||
}
|
||||
|
||||
QString ModelList::compatibleModelFilename(QUrl baseUrl, QString modelName) {
|
||||
QString hash(compatibleModelNameHash(baseUrl, modelName));
|
||||
return QString(u"gpt4all-%1-capi.rmodel"_s).arg(hash);
|
||||
};
|
||||
}
|
||||
|
||||
bool ModelList::eventFilter(QObject *obj, QEvent *ev)
|
||||
{
|
||||
@@ -2103,7 +2103,7 @@ void ModelList::parseDiscoveryJsonFile(const QByteArray &jsonData)
|
||||
emit discoverProgressChanged();
|
||||
if (!m_discoverNumberOfResults) {
|
||||
m_discoverInProgress = false;
|
||||
emit discoverInProgressChanged();;
|
||||
emit discoverInProgressChanged();
|
||||
}
|
||||
}
|
||||
|
||||
@@ -2180,7 +2180,7 @@ void ModelList::handleDiscoveryItemFinished()
|
||||
|
||||
if (discoverProgress() >= 1.0) {
|
||||
m_discoverInProgress = false;
|
||||
emit discoverInProgressChanged();;
|
||||
emit discoverInProgressChanged();
|
||||
}
|
||||
|
||||
reply->deleteLater();
|
||||
|
@@ -186,7 +186,7 @@ void MySettings::restoreModelDefaults(const ModelInfo &info)
|
||||
setModelTemperature(info, info.m_temperature);
|
||||
setModelTopP(info, info.m_topP);
|
||||
setModelMinP(info, info.m_minP);
|
||||
setModelTopK(info, info.m_topK);;
|
||||
setModelTopK(info, info.m_topK);
|
||||
setModelMaxLength(info, info.m_maxLength);
|
||||
setModelPromptBatchSize(info, info.m_promptBatchSize);
|
||||
setModelContextLength(info, info.m_contextLength);
|
||||
|
Reference in New Issue
Block a user