Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some bugs about model's architecture #3761

Closed
trestad opened this issue Aug 4, 2021 · 1 comment
Closed

Some bugs about model's architecture #3761

trestad opened this issue Aug 4, 2021 · 1 comment

Comments

@trestad
Copy link

trestad commented Aug 4, 2021

🐛 Bug

To Reproduce

Steps to reproduce the behavior (always include the command you ran):

I find that when I use cmd "git clone https://github.com/pytorch/fairseq" to install fairseq, there are some bugs. For example, in
“transformer_iwslt_de_en”, the parameter "--encoder/decoder-ffn-embed-dim" should be 1024, but even though I have used "-arch transformer_iwslt_de_en" in the train command, I still got a model with architecture of "base_architecture" where "--encoder-ffn-embed-dim" is 2048.
When I turn to the stable release, this bug never shows.
I hope you can fix this.

  1. See error

image

image

@trestad trestad changed the title Transformer Some bugs about model's architecture Aug 4, 2021
@freewym
Copy link
Contributor

freewym commented Aug 9, 2021

I have the same problem with the master branch: command line arguments for the Transformer's architecture seem not to override the default values. I think this bug was introduced in the commit 129d859

facebook-github-bot pushed a commit that referenced this issue Sep 20, 2021
#3773)

Summary:
…verride the defaults

# Before submitting

- [x] Was this discussed/approved via a Github issue? (no need for typos, doc improvements)
- [x] Did you read the [contributor guideline](https://github.com/pytorch/fairseq/blob/master/CONTRIBUTING.md)?
- [ ] Did you make sure to update the docs?
- [ ] Did you write any new necessary tests?

## What does this PR do?
Fixes #3761.

## PR review
Anyone in the community is free to review the PR once the tests have passed.
If we didn't discuss your PR in Github issues there's a high chance it will not be merged.

## Did you have fun?
Make sure you had fun coding �

Pull Request resolved: #3773

Reviewed By: yuntang

Differential Revision: D30310383

Pulled By: kahne

fbshipit-source-id: cbfcbc032dbf53490a25ffdebe57f65c42d52e71
sorenmulli pushed a commit to sorenmulli/fairseq that referenced this issue Oct 4, 2021
facebookresearch#3773)

Summary:
…verride the defaults

# Before submitting

- [x] Was this discussed/approved via a Github issue? (no need for typos, doc improvements)
- [x] Did you read the [contributor guideline](https://github.com/pytorch/fairseq/blob/master/CONTRIBUTING.md)?
- [ ] Did you make sure to update the docs?
- [ ] Did you write any new necessary tests?

## What does this PR do?
Fixes facebookresearch#3761.

## PR review
Anyone in the community is free to review the PR once the tests have passed.
If we didn't discuss your PR in Github issues there's a high chance it will not be merged.

## Did you have fun?
Make sure you had fun coding �

Pull Request resolved: facebookresearch#3773

Reviewed By: yuntang

Differential Revision: D30310383

Pulled By: kahne

fbshipit-source-id: cbfcbc032dbf53490a25ffdebe57f65c42d52e71
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants