Skip to content

[WIP] feat: build llama cpp externally #5790

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Draft
wants to merge 6 commits into
base: master
Choose a base branch
from

Conversation

mudler
Copy link
Owner

@mudler mudler commented Jul 4, 2025

Description

This PR fixes #

Notes for Reviewers

Signed commits

  • Yes, I signed my commits.

Copy link

netlify bot commented Jul 4, 2025

Deploy Preview for localai ready!

Name Link
🔨 Latest commit 6543c3f
🔍 Latest deploy log https://app.netlify.com/projects/localai/deploys/686eaf9ae0b0cc00085d189c
😎 Deploy Preview https://deploy-preview-5790--localai.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify project configuration.

@mudler mudler force-pushed the feat/build-llama-cpp-externally branch 3 times, most recently from 4980a37 to 608264c Compare July 7, 2025 09:38
@mudler mudler mentioned this pull request Jul 8, 2025
1 task
@mudler mudler force-pushed the feat/build-llama-cpp-externally branch 2 times, most recently from d1569f2 to f3b1c38 Compare July 8, 2025 17:22
@richiejp
Copy link
Collaborator

richiejp commented Jul 8, 2025

So a completely separate Dockerfile and Makefile? This will be a major improvement!

@mudler
Copy link
Owner Author

mudler commented Jul 9, 2025

So a completely separate Dockerfile and Makefile? This will be a major improvement!

yup! my plan is to isolate everything, one backend at a time. Currently the llama.cpp one is the most heavy, having also lots of specific code in the golang part - ideally I want to get rid of all of the specific llama.cpp code and the binary bundling bits out of the main code.

This is how I'm testing things now with #5816 in:

docker build --build-arg BACKEND=llama-cpp -t llama-cpp-backend -f backend/Dockerfile.llama-cpp .
docker save llama-cpp-backend -o llama-backend.tar
local-ai backends install "ocifile://$PWD/llama-backend.tar"

mudler added 4 commits July 9, 2025 18:44
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
@mudler mudler force-pushed the feat/build-llama-cpp-externally branch from f3b1c38 to 5885711 Compare July 9, 2025 16:45
mudler added 2 commits July 9, 2025 19:39
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants