mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2025-09-07 03:20:26 +00:00
Add q8_0 kernels to kompute shaders and bump to latest llama/gguf.
This commit is contained in:
Submodule gpt4all-backend/llama.cpp-mainline updated: 7f8f62880c...2b7b3f37f9
@@ -239,6 +239,7 @@ if (LLAMA_KOMPUTE)
|
||||
kompute/op_rmsnorm.comp
|
||||
kompute/op_diagmask.comp
|
||||
kompute/op_mul_mat_f16.comp
|
||||
kompute/op_mul_mat_q8_0.comp
|
||||
kompute/op_mul_mat_q4_0.comp
|
||||
kompute/op_mul_mat_q4_1.comp
|
||||
kompute/op_mul_mat_q6_k.comp
|
||||
@@ -268,6 +269,7 @@ if (LLAMA_KOMPUTE)
|
||||
shaderop_rmsnorm.h
|
||||
shaderop_diagmask.h
|
||||
shaderop_mul_mat_f16.h
|
||||
shaderop_mul_mat_q8_0.h
|
||||
shaderop_mul_mat_q4_0.h
|
||||
shaderop_mul_mat_q4_1.h
|
||||
shaderop_mul_mat_q6_k.h
|
||||
|
Reference in New Issue
Block a user