Skip to content
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.

update docs for lrn #5913

Merged
merged 3 commits into from
Apr 27, 2017
Merged

update docs for lrn #5913

merged 3 commits into from
Apr 27, 2017

Conversation

nswamy
Copy link
Member

@nswamy nswamy commented Apr 20, 2017

Copy link
Contributor

@zackchase zackchase left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks @nswamy for the contributions. Please see the few minor comments and let me know if they are sensible or if you disagree.

@@ -31,11 +31,11 @@ struct LRNParam : public dmlc::Parameter<LRNParam> {
uint32_t nsize;
DMLC_DECLARE_PARAMETER(LRNParam) {
DMLC_DECLARE_FIELD(alpha).set_default(1e-4f)
.describe("value of the alpha variance scaling parameter in the normalization formula");
.describe("Alpha variance scaling parameter in the normalization formula.");
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Here "alpha is used as an adjective. I think it might be more intuitive to write:

"The variance scaling parameter alpha from the normalization formula."

Where is the normalization formula? How does the user find what alpha refers to? Should we have some guidance for where this reference "normalization formula" points to? What are your thoughts??

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hey any luck resolving this one? Thanks!

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LRN expression is in the description of the function

DMLC_DECLARE_FIELD(beta).set_default(0.75f)
.describe("value of the beta power parameter in the normalization formula");
.describe("Beta power parameter in the normalization formula.");
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Same issue - let's make it consistent with however we do alpha.

DMLC_DECLARE_FIELD(knorm).set_default(2.0f)
.describe("value of the k parameter in normalization formula");
.describe("k parameter in normalization formula.");
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Same here.

.add_arguments(LRNParam::__FIELDS__())
.describe("Apply convolution to input then add a bias.");
.describe(R"code(Applies Local Response Normalization to the input.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Let's not have this in capital letters. Just

"Applies local response normalization to the input."

The main reasons to use capital letters would be if we are introducing a proper noun or an acronym.

.describe("Apply convolution to input then add a bias.");
.describe(R"code(Applies Local Response Normalization to the input.

Normalization helps in generalization and used to prevent neurons from
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Maybe be more descriptive here. Not all neurons can saturate. Also there is a grammatical bug on "used to" (suggests in the past, not utility).

Perhaps: "Normalization helps to prevent squashing neurons (like tanh) from saturating. It can also help for transfer learning, as the inputs for each task will exhibit the same dynamic range."

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LRN is applied on a ReLU neuron. Are ReLU neurons also considered squashing neurons ?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

what I was aiming for was to write a sentence about Normalization and then describe LRN. I will remove it if this leads to confusion.
You are right that ReLU neurons don't saturate and might turn into on negative inputs.

@zackchase
Copy link
Contributor

zackchase commented Apr 20, 2017 via email

@madjam madjam added the Doc label Apr 20, 2017
Naveen Swamy and others added 3 commits April 27, 2017 13:47
Copy link
Member Author

@nswamy nswamy left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks @zackchase, I have made changes per your suggestions.

@@ -31,11 +31,11 @@ struct LRNParam : public dmlc::Parameter<LRNParam> {
uint32_t nsize;
DMLC_DECLARE_PARAMETER(LRNParam) {
DMLC_DECLARE_FIELD(alpha).set_default(1e-4f)
.describe("value of the alpha variance scaling parameter in the normalization formula");
.describe("Alpha variance scaling parameter in the normalization formula.");
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LRN expression is in the description of the function

@piiswrong piiswrong merged commit b5903c3 into apache:master Apr 27, 2017
@nswamy nswamy deleted the lrn branch April 27, 2017 22:11
Guneet-Dhillon pushed a commit to Guneet-Dhillon/mxnet that referenced this pull request Sep 13, 2017
* update doc for lrn

* [Documentation] Ndarray instancenorm doc modified (apache#6008)

* InstanceNorm doc modified

* minor change

* xx
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants