-
Notifications
You must be signed in to change notification settings - Fork 350
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add Clipping schedulers #556
Conversation
This pull request was exported from Phabricator. Differential Revision: D42644261 |
Summary: Pull Request resolved: pytorch#556 This diff introduces gradient clipping schedulers that can be used to vary gradient clipping throughout training. Addresses pytorch#375 in OSS. Differential Revision: D42644261 fbshipit-source-id: 7480e2f81432bd4a05d58af420ee4a783bdd63f1
15d5d51
to
6358c20
Compare
This pull request was exported from Phabricator. Differential Revision: D42644261 |
Summary: Pull Request resolved: pytorch#556 This diff introduces gradient clipping schedulers that can be used to vary gradient clipping throughout training. Addresses pytorch#375 in OSS. Differential Revision: D42644261 fbshipit-source-id: 0af54b977d41b164531c18d4804a2506a359ce28
6358c20
to
ec71f8b
Compare
This pull request was exported from Phabricator. Differential Revision: D42644261 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for taking this one on. The changes look good to me, but be sure to fix the lint before merging. :)
) | ||
@settings(deadline=None) | ||
def test_checkpoints(self, noise_scheduler: Optional[Type[StepNoise]]): | ||
def test_checkpoints( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could you break this function into multiple smaller functions to make flake8 happy?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yeah, looks like that function was right at the limit for complexity and I crossed it :D Will refactor
Summary: Pull Request resolved: pytorch#556 This diff introduces gradient clipping schedulers that can be used to vary gradient clipping throughout training. Addresses pytorch#375 in OSS. Reviewed By: karthikprasad Differential Revision: D42644261 fbshipit-source-id: 91bedb4c3dd68f336917d16cec42f939ace02406
ec71f8b
to
28cfe8c
Compare
This pull request was exported from Phabricator. Differential Revision: D42644261 |
Summary: Pull Request resolved: pytorch#556 This diff introduces gradient clipping schedulers that can be used to vary gradient clipping throughout training. Addresses pytorch#375 in OSS. Reviewed By: karthikprasad Differential Revision: D42644261 fbshipit-source-id: 94f5fb756dce8ec576cce2ff003c2054eb926e27
This pull request was exported from Phabricator. Differential Revision: D42644261 |
28cfe8c
to
3412d26
Compare
Summary: Pull Request resolved: pytorch#556 This diff introduces gradient clipping schedulers that can be used to vary gradient clipping throughout training. Addresses pytorch#375 in OSS. Reviewed By: karthikprasad Differential Revision: D42644261 fbshipit-source-id: a571569785697b369c4c6f9709d35a5d3ff7e73c
This pull request was exported from Phabricator. Differential Revision: D42644261 |
3412d26
to
5d3ed10
Compare
Summary: Pull Request resolved: pytorch#556 This diff introduces gradient clipping schedulers that can be used to vary gradient clipping throughout training. Addresses pytorch#375 in OSS. Reviewed By: karthikprasad Differential Revision: D42644261 fbshipit-source-id: 57c87ba5f0b012359761fa015d046edc5c13da88
This pull request was exported from Phabricator. Differential Revision: D42644261 |
5d3ed10
to
8c8e540
Compare
This pull request has been merged in d888fd0. |
Summary:
This diff introduces gradient clipping schedulers that can be used to vary gradient clipping throughout training.
Addresses #375 in OSS.
Differential Revision: D42644261