-
Notifications
You must be signed in to change notification settings - Fork 10.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
docker: add support for CUDA in docker #1461
Conversation
I should add a note, I haven't been able to get the:
...part of the examples in I just wanted to keep it in line with the other Docker examples. |
f4ac04e
to
fdad997
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I was thinking something like this would be nice to have but ultimately prioritized fixing the current issues like memory management first. Thanks.
aeb43a1
to
9d121b7
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm not particularly knowledgeable about docker but to me it looks like everything is in order now.
Fixed a trailing white space error from the |
Could I get another run of CI on this? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Small changes to make it slightly less confusing, otherwise LGTM, it works.
This looks really useful for integration/deployment of llama.cpp into docker hosted services on cloud! Is there anything (apart from the vast amount of other activity ;-) ) holding up this being merged now? Um, on which note, I see that total open pull-requests, not issues(!), has risen from, iirc, 53 to 65 since I last checked in.. which is kinda great but also kinda scary because of implication for level of stress on supervision bandwidth.. |
The upstream |
I didn't merge this PR because I wanted someone else to check it as well; as I said, I'm not very knowledgeable about Docker. |
0484ffb
to
8c6c334
Compare
I have rebased on the latest changes in |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Will wait for an extra approve before merging as I'm unfamiliar with Docker
Hi folks, just a very gentle nudge again on this front - last time I checked here the total open pull-requests on this repo were about 65, a month later they are over 80.. as before I'm worried that more project approval/supervision bandwidth needs to be allocated to clear backlog a bit so older 'finished' PRs like this don't age stuck in limbo, suffer new integration issues and need revisiting.. It would be a shame if good work gets buried/lost/stuck with PRs not folded into main (and if sprawling PR numbers mean stale PRs are not being explicitly closed or duplicates being coalesced). The docker support stuff ought to be relatively independent of other changes - maybe worth a final CI check and folding in this one now - any problems for nvidia-docker use can be sorted via bug-reports by people trying it (who currently don't have the easy option). Best, M. |
Assuming one has the nvidia-container-toolkit installed on Linux, or is using a GPU enabled cloud,
cuBLAS
should be accessible inside the container.I'm not very familiar with Github Actions, nor the available execution environments available on GitHub. I wouldn't suggest pre-building these and putting them in the registry, unless there's a CI path for them.
I'm not too sure what the right path for putting these into a registry is. But I did want to contribute so people could try it locally!