From 3f8257c5630d39574a07aaaba7539f9cbddda486 Mon Sep 17 00:00:00 2001 From: Jared Van Bortel Date: Fri, 12 Apr 2024 17:25:15 -0400 Subject: [PATCH] llamamodel: fix semantic typo in nomic client dynamic mode (#2216) Signed-off-by: Jared Van Bortel --- gpt4all-backend/llamamodel.cpp | 2 +- gpt4all-bindings/python/setup.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/gpt4all-backend/llamamodel.cpp b/gpt4all-backend/llamamodel.cpp index 795f1e7e..2e2d3792 100644 --- a/gpt4all-backend/llamamodel.cpp +++ b/gpt4all-backend/llamamodel.cpp @@ -302,8 +302,8 @@ bool LLamaModel::loadModel(const std::string &modelPath, int n_ctx, int ngl) if (llama_verbose()) { std::cerr << "llama.cpp: using Metal" << std::endl; - d_ptr->backend_name = "metal"; } + d_ptr->backend_name = "metal"; // always fully offload on Metal // TODO(cebtenzzre): use this parameter to allow using more than 53% of system RAM to load a model diff --git a/gpt4all-bindings/python/setup.py b/gpt4all-bindings/python/setup.py index 4d4f6f8c..86e5e6f3 100644 --- a/gpt4all-bindings/python/setup.py +++ b/gpt4all-bindings/python/setup.py @@ -68,7 +68,7 @@ def get_long_description(): setup( name=package_name, - version="2.5.0", + version="2.5.1", description="Python bindings for GPT4All", long_description=get_long_description(), long_description_content_type="text/markdown",