Skip to content
This repository has been archived by the owner on Dec 16, 2022. It is now read-only.

Add scalar_mix_parameters to ElmoTokenEmbedder #1955

Merged

Conversation

nelson-liu
Copy link
Contributor

No description provided.

Copy link
Contributor

@matt-gardner matt-gardner left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM.

scalar_mix_parameters : ``List[int]``, optional, (default=None)
If not ``None``, use these scalar mix parameters to weight the representations
produced by different layers. These mixing weights are not updated during
training.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is there an option to just initialize to these values, but still train them? Just something to think about.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

yeah, you're right that this is currently missing. I didn't think anyone would use it / it would add extra complexity, so i just went with this for now.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yeah, definitely don't add it until there's a justification for it; I was just wondering about your experiments.

@nelson-liu nelson-liu merged commit 0264002 into allenai:master Oct 24, 2018
@nelson-liu nelson-liu deleted the fixed_scalar_mix_elmo_token_embedder branch October 24, 2018 20:23
nelson-liu added a commit that referenced this pull request Oct 24, 2018
Sorry, continuation of #1955 . Was able to run a model with this setting, so pretty sure I got it right this time around...
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants