Skip to content
This repository was archived by the owner on Jul 4, 2025. It is now read-only.

Commit d358274

Browse files
authored
Merge pull request #101 from janhq/41-feat-batch-inference-for-nitro
Merge batching inference to nitro + fix cuda on windows issues (shouldn't use shared libs)
2 parents d175680 + a463cdb commit d358274

File tree

8 files changed

+1602
-873
lines changed

8 files changed

+1602
-873
lines changed

.github/workflows/build.yml

Lines changed: 1 addition & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -292,8 +292,6 @@ jobs:
292292
run: |
293293
robocopy build_deps\_install\bin .\build\Release zlib.dll
294294
robocopy build\bin\Release .\build\Release llama.dll
295-
robocopy ext_libs .\build\Release libcrypto-3-x64.dll
296-
robocopy ext_libs .\build\Release libssl-3-x64.dll
297295
7z a nitro.zip .\build\Release\*
298296
cd .\build\Release
299297
..\..\.github\scripts\e2e-test-windows.bat .\nitro.exe ${{ env.MODEL_URL }}
@@ -343,7 +341,7 @@ jobs:
343341
cmake --build ./build_deps/nitro_deps --config Release
344342
mkdir -p build
345343
cd build
346-
cmake .. -DLLAMA_NATIVE=OFF -DLLAMA_BUILD_SERVER=ON -DLLAMA_CUBLAS=ON -DBUILD_SHARED_LIBS=ON
344+
cmake .. -DLLAMA_NATIVE=OFF -DLLAMA_BUILD_SERVER=ON -DLLAMA_CUBLAS=ON
347345
cmake --build . --config Release -j 4
348346
349347
- name: Pack artifacts
@@ -354,8 +352,6 @@ jobs:
354352
echo %PATH%
355353
robocopy build_deps\_install\bin .\build\Release zlib.dll
356354
robocopy build\bin\Release .\build\Release llama.dll
357-
robocopy ext_libs .\build\Release libcrypto-3-x64.dll
358-
robocopy ext_libs .\build\Release libssl-3-x64.dll
359355
7z a nitro.zip .\build\Release\*
360356
cd .\build\Release
361357
..\..\.github\scripts\e2e-test-windows.bat .\nitro.exe ${{ env.MODEL_URL }}

CMakeLists.txt

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,6 +16,9 @@ else()
1616
set(CMAKE_CXX_STANDARD 14)
1717
endif()
1818

19+
# llama cpp server need llava example to work, this is for llama cpp server
20+
set(LLAMA_BUILD_EXAMPLES ON)
21+
1922
set(CMAKE_CXX_STANDARD_REQUIRED ON)
2023
set(CMAKE_CXX_EXTENSIONS OFF)
2124
set(OPENSSL_USE_STATIC_LIBS TRUE)
@@ -48,7 +51,7 @@ add_executable(${PROJECT_NAME} main.cc)
4851
#
4952
# and comment out the following lines
5053
find_package(Drogon CONFIG REQUIRED)
51-
target_link_libraries(${PROJECT_NAME} PRIVATE Drogon::Drogon common llama
54+
target_link_libraries(${PROJECT_NAME} PRIVATE Drogon::Drogon common llama clip
5255
${CMAKE_THREAD_LIBS_INIT})
5356

5457
# ##############################################################################

0 commit comments

Comments
 (0)