-
Notifications
You must be signed in to change notification settings - Fork 27.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support compilation via Torchdynamo, AOT Autograd, NVFuser #17308
Changes from 1 commit
2396218
0b4c279
52503f2
298b3ba
6756fa5
28f80ec
bf41704
7925fa3
ce47619
9daaf93
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
- Loading branch information
There are no files selected for viewing
Original file line number | Diff line number | Diff line change | ||||
---|---|---|---|---|---|---|
|
@@ -450,7 +450,7 @@ class TrainingArguments: | |||||
full_determinism (`bool`, *optional*, defaults to `False`) | ||||||
If `True`, [`enable_full_determinism`] is called instead of [`set_seed`] to ensure reproducible results in | ||||||
distributed training | ||||||
use_torchdynamo ('bool`, `str`, defaults to `False`): | ||||||
torchdynamo (`str`, *optional*): | ||||||
If `True`, TorchDynamo is called with AOT Autograd and nvfuser compiler to compile the appropriate portions | ||||||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. this doesn't match the actual usage. Definitely not |
||||||
of the model. | ||||||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more.
Suggested change
|
||||||
""" | ||||||
|
@@ -884,15 +884,16 @@ class TrainingArguments: | |||||
) | ||||||
}, | ||||||
) | ||||||
use_torchdynamo: bool = field( | ||||||
default=False, | ||||||
torchdynamo: Optional[str] = field( | ||||||
default=None, | ||||||
metadata={ | ||||||
"help": ( | ||||||
"Whether or not to use TorchDynamo. TorchDynamo is a Python level JIT compilers designed to make" | ||||||
"Whether or not to use TorchDynamo. TorchDynamo is a Python level JIT compiler designed to make" | ||||||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. same as the comment above - not whether or not, but how - via |
||||||
" unmodified PyTorch programs faster. TorchDynamo dynamically modifies the Python bytecode right" | ||||||
" before its executed. It rewrites Python bytecode in order to extract sequences of PyTorch operations" | ||||||
" and lift them up into Fx fraph. We can then pass these Fx graphs to other backend compilers. Here" | ||||||
" we use AOT Autograd and nvfuser compiler." | ||||||
" and lift them up into Fx graph. We can then pass these Fx graphs to other backend compilers. There" | ||||||
" are two options - eager and nvfuser. Eager defaults to pytorch eager and is useful for debugging." | ||||||
" nvfuser path uses AOT Autograd and nvfuser compiler to optimize the models." | ||||||
), | ||||||
}, | ||||||
) | ||||||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Not sure that
else
is needed since we set it to this by default.