diff --git a/gpt4all-backend/CMakeLists.txt b/gpt4all-backend/CMakeLists.txt index c2190674..878385d6 100644 --- a/gpt4all-backend/CMakeLists.txt +++ b/gpt4all-backend/CMakeLists.txt @@ -65,7 +65,7 @@ foreach(BUILD_VARIANT IN LISTS BUILD_VARIANTS) message(STATUS "Configuring model implementation target ${TARGET_NAME}") # Link to ggml/llama target_link_libraries(${TARGET_NAME} - PUBLIC ${BASE_LIB}-${BUILD_VARIANT}) + PRIVATE ${BASE_LIB}-${BUILD_VARIANT}) # Let it know about its build variant target_compile_definitions(${TARGET_NAME} PRIVATE GGML_BUILD_VARIANT="${BUILD_VARIANT}") diff --git a/gpt4all-backend/llama.cpp.cmake b/gpt4all-backend/llama.cpp.cmake index db85d4c6..c7e59eb3 100644 --- a/gpt4all-backend/llama.cpp.cmake +++ b/gpt4all-backend/llama.cpp.cmake @@ -338,7 +338,7 @@ function(include_ggml DIRECTORY SUFFIX WITH_LLAMA) set(LLAMA_UTIL_SOURCE_FILE llama_util.h) endif() - add_library(llama${SUFFIX} + add_library(llama${SUFFIX} STATIC ${DIRECTORY}/llama.cpp ${DIRECTORY}/llama.h ${DIRECTORY}/${LLAMA_UTIL_SOURCE_FILE})