-
Notifications
You must be signed in to change notification settings - Fork 2.9k
add support for Ollama num_gpu #4353
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
✅ Deploy Preview for continuedev ready!
To edit notification comments on pull requests, go to your Netlify site configuration. |
I think this is failing due to an unrelated timeout? |
This is a welcome addition. It will allow the use of larger models or deeper context without forcing those changes on all of the models. |
Hi @Fmstrat, yesterday we shared some updates with our contributors about how we're aiming to improve the contribution process. Part of this included the addition of a Contributor License Agreement (CLA) to protect both contributors and the project. We're reaching out to ask that previous contributors sign it. Could you please take a moment to sign, or if you have any questions send me a message? (either here or nate@continue.dev would work) To do so, you just need to post a comment below with the following text:
❤️ Thank you for the work you've done on Continue, and let me know if you have any suggestions on how we can make the project even better! |
I have read the CLA Document and I hereby sign the CLA 1 out of 2 committers have signed the CLA. |
Description
Added support for Ollama
num_gpu
. This allows for forcing models to use the GPU in memory limited situations. Modeled after existing Ollama specific options such askeepAlive
.Checklist
Testing instructions
Set your config up with
numGpu
:Watch is pass through to Ollama via the Ollama logs (
--n-gpu-layers 1000
):