Fix regenerate button to be deterministic and bump the llama version to latest we have for gguf.

This commit is contained in:
Adam Treat
2023-10-03 12:42:31 -04:00
parent 40c78d2f78
commit 12f943e966
4 changed files with 5 additions and 15 deletions

View File

@@ -371,7 +371,7 @@ void ChatLLM::regenerateResponse()
else
m_ctx.n_past -= m_promptResponseTokens;
m_ctx.n_past = std::max(0, m_ctx.n_past);
m_ctx.tokens.erase(m_ctx.tokens.end() -= m_promptResponseTokens, m_ctx.tokens.end());
m_ctx.tokens.erase(m_ctx.tokens.end() - m_promptResponseTokens, m_ctx.tokens.end());
m_promptResponseTokens = 0;
m_promptTokens = 0;
m_response = std::string();