File tree Expand file tree Collapse file tree 3 files changed +774
-2
lines changed Expand file tree Collapse file tree 3 files changed +774
-2
lines changed Original file line number Diff line number Diff line change @@ -10,6 +10,7 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
10
10
## [ 0.3.9]
11
11
12
12
- feat: Update llama.cpp to ggerganov/llama.cpp@8733e0cf6eefc7c7752297cc22d0836706f4222c
13
+ - Make building mtmd optional by setting ` CMAKE_ARGS="-DMTMD_BUILD=OFF" ` and using ` MTMD_CPP_LIB ` to specify alternative path to shared library
13
14
14
15
## [ 0.3.8]
15
16
Original file line number Diff line number Diff line change @@ -3,7 +3,7 @@ cmake_minimum_required(VERSION 3.21)
3
3
project (llama_cpp )
4
4
5
5
option (LLAMA_BUILD "Build llama.cpp shared library and install alongside python package" ON )
6
- option (LLAVA_BUILD "Build llava shared library and install alongside python package" ON )
6
+ option (MTMD_BUILD "Build mtmd shared library and install alongside python package" ON )
7
7
8
8
function (llama_cpp_python_install_target target )
9
9
if (NOT TARGET ${target} )
@@ -62,6 +62,9 @@ if (LLAMA_BUILD)
62
62
# Enable building of the common library
63
63
set (LLAMA_BUILD_COMMON ON CACHE BOOL "Build llama.cpp common library" FORCE )
64
64
65
+ # Disable building curl support
66
+ set (LLAMA_CURL OFF CACHE BOOL "llama.cpp: enable curl" FORCE )
67
+
65
68
# Architecture detection and settings for Apple platforms
66
69
if (APPLE )
67
70
# Get the target architecture
@@ -133,7 +136,7 @@ if (LLAMA_BUILD)
133
136
)
134
137
endif ()
135
138
136
- if (LLAVA_BUILD )
139
+ if (MTMD_BUILD )
137
140
if (LLAMA_CUBLAS OR LLAMA_CUDA )
138
141
add_compile_definitions (GGML_USE_CUBLAS )
139
142
add_compile_definitions (GGML_USE_CUDA )
You can’t perform that action at this time.
0 commit comments