Skip to content

Commit

Permalink
.
Browse files Browse the repository at this point in the history
  • Loading branch information
srush committed Mar 11, 2018
1 parent bad80e8 commit a38f3ea
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 3 deletions.
4 changes: 2 additions & 2 deletions onmt/modules/Embeddings.py
Expand Up @@ -2,7 +2,7 @@
import torch.nn as nn
from torch.autograd import Variable

from onmt.modules import BottleLinear, Elementwise
from onmt.modules import Elementwise
from onmt.Utils import aeq


Expand Down Expand Up @@ -138,7 +138,7 @@ def __init__(self, word_vec_size,
if feat_merge == 'mlp':
in_dim = sum(emb_dims)
out_dim = word_vec_size
mlp = nn.Sequential(BottleLinear(in_dim, out_dim), nn.ReLU())
mlp = nn.Sequential(nn.Linear(in_dim, out_dim), nn.ReLU())
self.make_embedding.add_module('mlp', mlp)

if position_encoding:
Expand Down
1 change: 0 additions & 1 deletion onmt/modules/MultiHeadedAttn.py
Expand Up @@ -4,7 +4,6 @@
from torch.autograd import Variable

from onmt.Utils import aeq
from onmt.modules.UtilClass import BottleLinear, BottleSoftmax


class MultiHeadedAttention(nn.Module):
Expand Down

0 comments on commit a38f3ea

Please sign in to comment.