Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Effect of bias in linear layers #16

Closed
ptrcarta opened this issue Dec 14, 2020 · 1 comment
Closed

Effect of bias in linear layers #16

ptrcarta opened this issue Dec 14, 2020 · 1 comment

Comments

@ptrcarta
Copy link

I've been experimenting with SELUs, and found they provide an improvement in terms of computation time during training with respect to batchnorm, thank you for your work.

I just have a question regarding the effect of bias in linear layers. As I understand it, every neuron should have mean zero in order to stay in the self regularizing zone, but bias precisely shifts that mean. In my experiments however I didn't see much of an effect either removing or adding biases. I see that in the tutorial notebook bias is used, and I wonder wether you've considered the issue.

@gklambauer
Copy link
Member

Dear ptrcarta, thanks, good point! We have experimented a lot with SNNs with and without bias units. In wide networks they hardly play a role. My hypothesis is that it is due to the following: a) SELUs counter the bias shift well and keep activations close to zero mean which is good for learning and b) in wide layers, any unit can learn to take the role of a bias unit. However, at the output layer, bias units can help especially if you have unbalanced data. Hope this helps!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants