Skip to content
This repository has been archived by the owner on Jan 15, 2024. It is now read-only.

Commit

Permalink
Fix parameter share regex (#1267)
Browse files Browse the repository at this point in the history
  • Loading branch information
leezu committed Jul 17, 2020
1 parent 70a1887 commit 9d83fe6
Show file tree
Hide file tree
Showing 2 changed files with 6 additions and 6 deletions.
6 changes: 3 additions & 3 deletions src/gluonnlp/models/transformer_xl.py
Original file line number Diff line number Diff line change
Expand Up @@ -277,11 +277,11 @@ def __init__(self, cfg=None):
weight_initializer=weight_initializer,
bias_initializer=bias_initializer)
if cfg.MODEL.tie_weights and cfg.MODEL.tie_projs:
crit_params = self.word_emb.collect_params('(.*_embed|.*_inter_proj)')
crit_params = self.word_emb.collect_params('(embed|inter_proj)')
elif cfg.MODEL.tie_weights and not cfg.MODEL.tie_projs:
crit_params = self.word_emb.collect_params('.*_embed')
crit_params = self.word_emb.collect_params('embed')
elif not cfg.MODEL.tie_weights and cfg.MODEL.tie_projs:
crit_params = self.word_emb.collect_params('.*_inter_proj')
crit_params = self.word_emb.collect_params('inter_proj')
else:
crit_params = None
self.crit = ProjectedAdaptiveLogSoftmaxWithLoss(
Expand Down
6 changes: 3 additions & 3 deletions tests/test_layers.py
Original file line number Diff line number Diff line change
Expand Up @@ -143,21 +143,21 @@ def test_projected_adaptive_softmax(vocab_size, cutoffs, embed_size, in_units, d
embed_size=embed_size,
in_units=in_units,
div_val=div_val)
layer_with_shared_proj.share_parameters(embed_layer.collect_params('.*_inter_proj'))
layer_with_shared_proj.share_parameters(embed_layer.collect_params('inter_proj'))
layer_with_shared_embed = \
ProjectedAdaptiveLogSoftmaxWithLoss(vocab_size=vocab_size,
cutoffs=cutoffs,
embed_size=embed_size,
in_units=in_units,
div_val=div_val)
layer_with_shared_embed.share_parameters(embed_layer.collect_params('.*_embed'))
layer_with_shared_embed.share_parameters(embed_layer.collect_params('embed'))
layer_with_shared_proj_embed = \
ProjectedAdaptiveLogSoftmaxWithLoss(vocab_size=vocab_size,
cutoffs=cutoffs,
embed_size=embed_size,
in_units=in_units,
div_val=div_val)
layer_with_shared_proj_embed.share_parameters(embed_layer.collect_params('(.*_embed|.*_inter_proj)'))
layer_with_shared_proj_embed.share_parameters(embed_layer.collect_params('(embed|inter_proj)'))
embed_layer.initialize()
embed_layer.hybridize()
layer_with_shared_proj.initialize()
Expand Down

0 comments on commit 9d83fe6

Please sign in to comment.