Skip to content

Conversation

@tippisum
Copy link
Contributor

@tippisum tippisum commented Jul 1, 2018

It implements per-channel alpha_dropout. It also creates corresponding function classes and unifies the process of dropout and alpha_dropout.

soumith
soumith previously requested changes Jul 1, 2018
Copy link
Contributor

@soumith soumith left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

please add tests in test_nn.py for the 2d and 3d variants.

@ssnl
Copy link
Collaborator

ssnl commented Jul 1, 2018

I would much rather having this in ATen, but unfortunately JIT blocks that currently (#9069).

@soumith Can we drop the *d thing and just call it FeatureAlphaDropout? I was deprecating Dropout*d with FeatureDropout in #9008 , but that PR's not happening soon due to #9069 .

@apaszke
Copy link
Contributor

apaszke commented Jul 1, 2018

+1 for not going with Nd names

@soumith
Copy link
Contributor

soumith commented Jul 2, 2018

agree, let's not introduce *Nd names, so this PR only has FeatureAlphaDropout, the rest should be removed.

@tippisum
Copy link
Contributor Author

tippisum commented Jul 2, 2018

Names are now corrected (existing droput*d names are untouched). Tests also seem OK.

Copy link
Contributor

@facebook-github-bot facebook-github-bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@ezyang has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator.

@ezyang ezyang dismissed soumith’s stale review July 14, 2018 19:01

names corrected

b = ctx.noise
else:
ctx.noise.bernoulli_(1 - ctx.p)
a = ((3.09091329228798008 * ctx.p + 1) * (1 - ctx.p)) ** (-0.5)

This comment was marked as off-topic.

@ezyang
Copy link
Contributor

ezyang commented Jul 14, 2018

math checks out

Copy link
Contributor

@facebook-github-bot facebook-github-bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@ezyang has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator.

jramseyer pushed a commit to jramseyer/pytorch that referenced this pull request Jul 30, 2018
Summary:
It implements per-channel alpha_dropout. It also creates corresponding function classes and unifies the process of dropout and alpha_dropout.
Pull Request resolved: pytorch#9073

Differential Revision: D8727008

Pulled By: ezyang

fbshipit-source-id: 9d509f9c5db4e98f7b698cdfc4443505a4d2b331
goodlux pushed a commit to goodlux/pytorch that referenced this pull request Aug 15, 2018
Summary:
It implements per-channel alpha_dropout. It also creates corresponding function classes and unifies the process of dropout and alpha_dropout.
Pull Request resolved: pytorch#9073

Differential Revision: D8727008

Pulled By: ezyang

fbshipit-source-id: 9d509f9c5db4e98f7b698cdfc4443505a4d2b331
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants