-
-
Notifications
You must be signed in to change notification settings - Fork 608
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
dropout function always active #1263
Conversation
I don't think we should do it this way, dropout is usually omitted at test time, this way it will make it difficult to turn it off when most cases all for it to be off |
this is what the |
In that case better to just remove the starting |
to avoid silent change of behavior, I introduced the mandatory keyword argument |
NEWS.md
Outdated
@@ -17,6 +17,7 @@ | |||
* Testing suite improvements now test for gradients of all layers along with GPU support. | |||
* Functors have now moved to [Functors.jl](https://github.com/FluxML/Flux.jl/pull/1174) to allow for their use outside of Flux. | |||
* Added [helper functions](https://github.com/FluxML/Flux.jl/pull/873) `Flux.convfilter` and `Flux.depthwiseconvfilter` to construct weight arrays for convolutions outside of layer constructors so as to not have to depend on the default layers for custom implementations. | |||
* `dropout` function is now [always active](https://github.com/FluxML/Flux.jl/pull/1263) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is out of date with the new keyword?
also maybe mention the Dropout
layer here ?
@DhairyaLGandhi merge? |
NEWS.md
Outdated
@@ -17,6 +17,8 @@ | |||
* Testing suite improvements now test for gradients of all layers along with GPU support. | |||
* Functors have now moved to [Functors.jl](https://github.com/FluxML/Flux.jl/pull/1174) to allow for their use outside of Flux. | |||
* Added [helper functions](https://github.com/FluxML/Flux.jl/pull/873) `Flux.convfilter` and `Flux.depthwiseconvfilter` to construct weight arrays for convolutions outside of layer constructors so as to not have to depend on the default layers for custom implementations. | |||
* `dropout` function now has a mandatory [active](https://github.com/FluxML/Flux.jl/pull/1263) | |||
keyword argument. The `Dropout` struct is the recommended choice for common usage. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Maybe say the behaviour of the layer has not changed.
@DhairyaLGandhi done |
bors r+ In general, we should try to remove the dependency on the |
Build succeeded: |
Fix #1084 according to the suggestions of @AzamatB and @jondeuce
PR Checklist
@MikeInnes
or@dhairyagandhi96
(for API changes).