Skip to content

togheppi/compare_initializers

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 

Repository files navigation

compare_initializers

Performance comparison of various weight and bias initializers using MNIST dataset

Network architecture

  • hidden layers: Two fully-connected layer (256 nodes, respectively) with ReLU
  • output layer: Fully-connected layer (10 nodes = # of class for MNIST)
  • Batch normalization is used for hidden layers

Used initializers

  • Weight initializer: Normal, Truncate normal, Xavier, He initializer
  • Bias initializer: Constant (zero), Normal initializer

Results

Training accuracies of classifying MNIST data are compared.

References

  1. https://github.com/hwalsuklee/tensorflow-mnist-MLP-batch_normalization-weight_initializers

About

Comparison of various weight and bias initializers

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages