Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feature request: Weight normalization #942

Open
ararslan opened this issue Nov 22, 2019 · 1 comment · May be fixed by #2053
Open

Feature request: Weight normalization #942

ararslan opened this issue Nov 22, 2019 · 1 comment · May be fixed by #2053

Comments

@ararslan
Copy link
Contributor

ararslan commented Nov 22, 2019

Weight normalization is a reparameterization of the weights in terms of two quantities which are optimized instead of the weights directly. The authors provide a reference implementation in Python using Theano, and the functionality is available in PyTorch as a hook.

I've toyed just a tiny bit with implementing this, but I think my knowledge of Flux's internals is insufficient. I was looking at implementing it as a layer that wraps another layer, for which the params are the g and v parameterization of the weights of the wrapped layer. Perhaps there's a better way.

@CarloLucibello
Copy link
Member

another attempt here #1005

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants