fix non-AVX CPU detection (#2141)

* chat: fix non-AVX CPU detection on Windows
* bindings: throw exception instead of logging to console

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
This commit is contained in:
Jared Van Bortel 2024-03-19 10:56:14 -04:00 committed by GitHub
parent 6c2542e540
commit 699410014a
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
6 changed files with 79 additions and 84 deletions

View File

@ -19,33 +19,27 @@
std::string s_implementations_search_path = "."; std::string s_implementations_search_path = ".";
static bool has_at_least_minimal_hardware() { #if !(defined(__x86_64__) || defined(_M_X64))
#if defined(__x86_64__) || defined(_M_X64) // irrelevant on non-x86_64
#ifndef _MSC_VER #define cpu_supports_avx() -1
return __builtin_cpu_supports("avx"); #define cpu_supports_avx2() -1
#else #elif defined(_MSC_VER)
int cpuInfo[4]; // MSVC
__cpuid(cpuInfo, 1); static int get_cpu_info(int func_id, int reg_id) {
return cpuInfo[2] & (1 << 28); int info[4];
#endif __cpuid(info, func_id);
#else return info[reg_id];
return true; // Don't know how to handle non-x86_64 }
#endif
}
static bool requires_avxonly() { // AVX via EAX=1: Processor Info and Feature Bits, bit 28 of ECX
#if defined(__x86_64__) || defined(_M_X64) #define cpu_supports_avx() (get_cpu_info(1, 2) & (1 << 28))
#ifndef _MSC_VER // AVX2 via EAX=7, ECX=0: Extended Features, bit 5 of EBX
return !__builtin_cpu_supports("avx2"); #define cpu_supports_avx2() (get_cpu_info(7, 1) & (1 << 5))
#else
int cpuInfo[4];
__cpuidex(cpuInfo, 7, 0);
return !(cpuInfo[1] & (1 << 5));
#endif
#else #else
return false; // Don't know how to handle non-x86_64 // gcc/clang
#define cpu_supports_avx() __builtin_cpu_supports("avx")
#define cpu_supports_avx2() __builtin_cpu_supports("avx2")
#endif #endif
}
LLModel::Implementation::Implementation(Dlhandle &&dlhandle_) LLModel::Implementation::Implementation(Dlhandle &&dlhandle_)
: m_dlhandle(new Dlhandle(std::move(dlhandle_))) { : m_dlhandle(new Dlhandle(std::move(dlhandle_))) {
@ -71,21 +65,25 @@ LLModel::Implementation::Implementation(Implementation &&o)
} }
LLModel::Implementation::~Implementation() { LLModel::Implementation::~Implementation() {
if (m_dlhandle) delete m_dlhandle; delete m_dlhandle;
} }
bool LLModel::Implementation::isImplementation(const Dlhandle &dl) { static bool isImplementation(const Dlhandle &dl) {
return dl.get<bool(uint32_t)>("is_g4a_backend_model_implementation"); return dl.get<bool(uint32_t)>("is_g4a_backend_model_implementation");
} }
const std::vector<LLModel::Implementation> &LLModel::Implementation::implementationList() { const std::vector<LLModel::Implementation> &LLModel::Implementation::implementationList() {
if (cpu_supports_avx() == 0) {
throw std::runtime_error("CPU does not support AVX");
}
// NOTE: allocated on heap so we leak intentionally on exit so we have a chance to clean up the // NOTE: allocated on heap so we leak intentionally on exit so we have a chance to clean up the
// individual models without the cleanup of the static list interfering // individual models without the cleanup of the static list interfering
static auto* libs = new std::vector<Implementation>([] () { static auto* libs = new std::vector<Implementation>([] () {
std::vector<Implementation> fres; std::vector<Implementation> fres;
std::string impl_name_re = "(bert|gptj|llamamodel-mainline)"; std::string impl_name_re = "(gptj|llamamodel-mainline)";
if (requires_avxonly()) { if (cpu_supports_avx2() == 0) {
impl_name_re += "-avxonly"; impl_name_re += "-avxonly";
} else { } else {
impl_name_re += "-(default|metal)"; impl_name_re += "-(default|metal)";
@ -107,9 +105,8 @@ const std::vector<LLModel::Implementation> &LLModel::Implementation::implementat
// Add to list if model implementation // Add to list if model implementation
try { try {
Dlhandle dl(p.string()); Dlhandle dl(p.string());
if (!Implementation::isImplementation(dl)) { if (!isImplementation(dl))
continue; continue;
}
fres.emplace_back(Implementation(std::move(dl))); fres.emplace_back(Implementation(std::move(dl)));
} catch (...) {} } catch (...) {}
} }
@ -134,18 +131,13 @@ const LLModel::Implementation* LLModel::Implementation::implementation(const cha
return &i; return &i;
} }
if (!buildVariantMatched) { if (!buildVariantMatched)
std::cerr << "LLModel ERROR: Could not find any implementations for build variant: " << buildVariant << "\n"; throw std::runtime_error("Could not find any implementations for build variant: " + buildVariant);
}
return nullptr; return nullptr; // unsupported model format
} }
LLModel *LLModel::Implementation::construct(const std::string &modelPath, std::string buildVariant, int n_ctx) { LLModel *LLModel::Implementation::construct(const std::string &modelPath, std::string buildVariant, int n_ctx) {
if (!has_at_least_minimal_hardware()) {
std::cerr << "LLModel ERROR: CPU does not support AVX\n";
return nullptr;
}
// Get correct implementation // Get correct implementation
const Implementation* impl = nullptr; const Implementation* impl = nullptr;
@ -178,7 +170,7 @@ LLModel *LLModel::Implementation::construct(const std::string &modelPath, std::s
if (!impl) { if (!impl) {
//TODO: Auto-detect CUDA/OpenCL //TODO: Auto-detect CUDA/OpenCL
if (buildVariant == "auto") { if (buildVariant == "auto") {
if (requires_avxonly()) { if (cpu_supports_avx2() == 0) {
buildVariant = "avxonly"; buildVariant = "avxonly";
} else { } else {
buildVariant = "default"; buildVariant = "default";
@ -196,15 +188,24 @@ LLModel *LLModel::Implementation::construct(const std::string &modelPath, std::s
LLModel *LLModel::Implementation::constructDefaultLlama() { LLModel *LLModel::Implementation::constructDefaultLlama() {
static std::unique_ptr<LLModel> llama([]() -> LLModel * { static std::unique_ptr<LLModel> llama([]() -> LLModel * {
const std::vector<LLModel::Implementation> *impls;
try {
impls = &implementationList();
} catch (const std::runtime_error &e) {
std::cerr << __func__ << ": implementationList failed: " << e.what() << "\n";
return nullptr;
}
const LLModel::Implementation *impl = nullptr; const LLModel::Implementation *impl = nullptr;
for (const auto &i : implementationList()) { for (const auto &i: *impls) {
if (i.m_buildVariant == "metal" || i.m_modelType != "LLaMA") continue; if (i.m_buildVariant == "metal" || i.m_modelType != "LLaMA") continue;
impl = &i; impl = &i;
} }
if (!impl) { if (!impl) {
std::cerr << "LLModel ERROR: Could not find CPU LLaMA implementation\n"; std::cerr << __func__ << ": could not find llama.cpp implementation\n";
return nullptr; return nullptr;
} }
auto fres = impl->m_construct(); auto fres = impl->m_construct();
fres->m_implementation = impl; fres->m_implementation = impl;
return fres; return fres;
@ -240,3 +241,7 @@ void LLModel::Implementation::setImplementationsSearchPath(const std::string& pa
const std::string& LLModel::Implementation::implementationsSearchPath() { const std::string& LLModel::Implementation::implementationsSearchPath() {
return s_implementations_search_path; return s_implementations_search_path;
} }
bool LLModel::Implementation::hasSupportedCPU() {
return cpu_supports_avx() != 0;
}

View File

@ -30,7 +30,6 @@ public:
class Implementation { class Implementation {
public: public:
Implementation(Dlhandle &&);
Implementation(const Implementation &) = delete; Implementation(const Implementation &) = delete;
Implementation(Implementation &&); Implementation(Implementation &&);
~Implementation(); ~Implementation();
@ -38,9 +37,6 @@ public:
std::string_view modelType() const { return m_modelType; } std::string_view modelType() const { return m_modelType; }
std::string_view buildVariant() const { return m_buildVariant; } std::string_view buildVariant() const { return m_buildVariant; }
static bool isImplementation(const Dlhandle &dl);
static const std::vector<Implementation> &implementationList();
static const Implementation *implementation(const char *fname, const std::string &buildVariant);
static LLModel *construct(const std::string &modelPath, std::string buildVariant = "auto", int n_ctx = 2048); static LLModel *construct(const std::string &modelPath, std::string buildVariant = "auto", int n_ctx = 2048);
static std::vector<GPUDevice> availableGPUDevices(); static std::vector<GPUDevice> availableGPUDevices();
static int32_t maxContextLength(const std::string &modelPath); static int32_t maxContextLength(const std::string &modelPath);
@ -48,8 +44,13 @@ public:
static bool isEmbeddingModel(const std::string &modelPath); static bool isEmbeddingModel(const std::string &modelPath);
static void setImplementationsSearchPath(const std::string &path); static void setImplementationsSearchPath(const std::string &path);
static const std::string &implementationsSearchPath(); static const std::string &implementationsSearchPath();
static bool hasSupportedCPU();
private: private:
Implementation(Dlhandle &&);
static const std::vector<Implementation> &implementationList();
static const Implementation *implementation(const char *fname, const std::string &buildVariant);
static LLModel *constructDefaultLlama(); static LLModel *constructDefaultLlama();
bool (*m_magicMatch)(const char *fname); bool (*m_magicMatch)(const char *fname);

View File

@ -13,8 +13,6 @@ struct LLModelWrapper {
~LLModelWrapper() { delete llModel; } ~LLModelWrapper() { delete llModel; }
}; };
thread_local static std::string last_error_message;
llmodel_model llmodel_model_create(const char *model_path) { llmodel_model llmodel_model_create(const char *model_path) {
const char *error; const char *error;
auto fres = llmodel_model_create2(model_path, "auto", &error); auto fres = llmodel_model_create2(model_path, "auto", &error);
@ -24,24 +22,30 @@ llmodel_model llmodel_model_create(const char *model_path) {
return fres; return fres;
} }
static void llmodel_set_error(const char **errptr, const char *message) {
thread_local static std::string last_error_message;
if (errptr) {
last_error_message = message;
*errptr = last_error_message.c_str();
}
}
llmodel_model llmodel_model_create2(const char *model_path, const char *build_variant, const char **error) { llmodel_model llmodel_model_create2(const char *model_path, const char *build_variant, const char **error) {
auto wrapper = new LLModelWrapper; LLModel *llModel;
try { try {
wrapper->llModel = LLModel::Implementation::construct(model_path, build_variant); llModel = LLModel::Implementation::construct(model_path, build_variant);
if (!wrapper->llModel) {
last_error_message = "Model format not supported (no matching implementation found)";
}
} catch (const std::exception& e) { } catch (const std::exception& e) {
last_error_message = e.what(); llmodel_set_error(error, e.what());
return nullptr;
} }
if (!wrapper->llModel) { if (!llModel) {
delete std::exchange(wrapper, nullptr); llmodel_set_error(error, "Model format not supported (no matching implementation found)");
if (error) { return nullptr;
*error = last_error_message.c_str();
}
} }
auto wrapper = new LLModelWrapper;
wrapper->llModel = llModel;
return wrapper; return wrapper;
} }
@ -159,8 +163,7 @@ float *llmodel_embed(
auto *wrapper = static_cast<LLModelWrapper *>(model); auto *wrapper = static_cast<LLModelWrapper *>(model);
if (!texts || !*texts) { if (!texts || !*texts) {
if (error) llmodel_set_error(error, "'texts' is NULL or empty");
*error = strdup("'texts' is NULL or empty");
return nullptr; return nullptr;
} }
@ -183,8 +186,7 @@ float *llmodel_embed(
embedding = new float[embd_size]; embedding = new float[embd_size];
wrapper->llModel->embed(textsVec, embedding, prefixStr, dimensionality, do_mean, atlas); wrapper->llModel->embed(textsVec, embedding, prefixStr, dimensionality, do_mean, atlas);
} catch (std::exception const &e) { } catch (std::exception const &e) {
if (error) llmodel_set_error(error, e.what());
*error = strdup(e.what());
return nullptr; return nullptr;
} }

View File

@ -185,7 +185,7 @@ class LLModel:
model = llmodel.llmodel_model_create2(self.model_path, b"auto", ctypes.byref(err)) model = llmodel.llmodel_model_create2(self.model_path, b"auto", ctypes.byref(err))
if model is None: if model is None:
s = err.value s = err.value
raise ValueError(f"Unable to instantiate model: {'null' if s is None else s.decode()}") raise RuntimeError(f"Unable to instantiate model: {'null' if s is None else s.decode()}")
self.model = model self.model = model
def __del__(self): def __del__(self):

View File

@ -1,4 +1,5 @@
#include "llm.h" #include "llm.h"
#include "../gpt4all-backend/llmodel.h"
#include "../gpt4all-backend/sysinfo.h" #include "../gpt4all-backend/sysinfo.h"
#include <QCoreApplication> #include <QCoreApplication>
@ -25,22 +26,8 @@ LLM *LLM::globalInstance()
LLM::LLM() LLM::LLM()
: QObject{nullptr} : QObject{nullptr}
, m_compatHardware(true) , m_compatHardware(LLModel::Implementation::hasSupportedCPU())
{ {
#if defined(__x86_64__)
#ifndef _MSC_VER
const bool minimal(__builtin_cpu_supports("avx"));
#else
int cpuInfo[4];
__cpuid(cpuInfo, 1);
const bool minimal(cpuInfo[2] & (1 << 28));
#endif
#else
const bool minimal = true; // Don't know how to handle non-x86_64
#endif
m_compatHardware = minimal;
QNetworkInformation::loadDefaultBackend(); QNetworkInformation::loadDefaultBackend();
auto * netinfo = QNetworkInformation::instance(); auto * netinfo = QNetworkInformation::instance();
if (netinfo) { if (netinfo) {

View File

@ -228,11 +228,11 @@ int ModelInfo::maxContextLength() const
if (!installed || isOnline) return -1; if (!installed || isOnline) return -1;
if (m_maxContextLength != -1) return m_maxContextLength; if (m_maxContextLength != -1) return m_maxContextLength;
auto path = (dirpath + filename()).toStdString(); auto path = (dirpath + filename()).toStdString();
int layers = LLModel::Implementation::maxContextLength(path); int n_ctx = LLModel::Implementation::maxContextLength(path);
if (layers < 0) { if (n_ctx < 0) {
layers = 4096; // fallback value n_ctx = 4096; // fallback value
} }
m_maxContextLength = layers; m_maxContextLength = n_ctx;
return m_maxContextLength; return m_maxContextLength;
} }