-
Notifications
You must be signed in to change notification settings - Fork 95
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: update torch and modify cuda graphs #220
Conversation
@@ -1,5 +1,5 @@ | |||
triton==2.0.0.dev20221202 | |||
torch==1.14.0.dev20221029+cu117 | |||
torch== 2.0.0.dev20221214+cu117 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
note to remember: version in docker needs to be changed
simple test pass
|
all tests pass
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
needs docker change
@@ -22,15 +22,11 @@ | |||
from kernl.optimizer.dynamo_backend import dynamo_backend_ofi | |||
|
|||
|
|||
# single shared pool by default | |||
_pool: (int, int) = torch.cuda.graph_pool_handle() |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
should we open an issue to add pool back ?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
it will be covered here:
- Reduce memory overhead of CUDA graphs #205
- no copy of recycled K/V cache in CUDA graphs #206
(my guess is that both will be managed at the same time)
@gaetansnl docker updated |
fix #197