Skip to content

myoons/test-activation-functions

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Activation Functions Test

Test 6 activation functions for training a single node. (1 weight, 1 bias)

  • Input : 1.
  • Initial weight : -1.
  • Initial bias : -1.
  • Target : 1.

I have done this test because I was curious that since GeLU and swish function has negative derivative (in some part), training could be messed. So I manipulated initial output (before activation function) to have negative derivatives and could see that it isn't trained. Actually, it is trained to opposite side. Below are images of Sigmoid, ReLU, GeLU. You can find more cases in images folder.

Sigmoid

sigmoid sigmoid_derivative sigmoid_result


ReLU (Dying)

relu relu_derivative relu_result

We can see Dying ReLU.


GeLU

gelu gelu_derivative gelu_result

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages