Skip to content

Commit

Permalink
Changed bias initialization to avoid NaN errors during training
Browse files Browse the repository at this point in the history
  • Loading branch information
srigas committed May 19, 2023
1 parent 9135372 commit d00c408
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions modules.py
Expand Up @@ -57,7 +57,7 @@ def __init__(self, n_features, window_size, dropout, alpha, embed_dim=None, use_
nn.init.xavier_uniform_(self.a.data, gain=1.414)

if self.use_bias:
self.bias = nn.Parameter(torch.empty(n_features, n_features))
self.bias = nn.Parameter(torch.zeros(n_features, n_features))

self.leakyrelu = nn.LeakyReLU(alpha)
self.sigmoid = nn.Sigmoid()
Expand Down Expand Up @@ -158,7 +158,7 @@ def __init__(self, n_features, window_size, dropout, alpha, embed_dim=None, use_
nn.init.xavier_uniform_(self.a.data, gain=1.414)

if self.use_bias:
self.bias = nn.Parameter(torch.empty(window_size, window_size))
self.bias = nn.Parameter(torch.zeros(window_size, window_size))

self.leakyrelu = nn.LeakyReLU(alpha)
self.sigmoid = nn.Sigmoid()
Expand Down

0 comments on commit d00c408

Please sign in to comment.