Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Ideal max_length for macaw-11b #15

Open
sanxchep opened this issue May 25, 2022 · 2 comments
Open

Ideal max_length for macaw-11b #15

sanxchep opened this issue May 25, 2022 · 2 comments

Comments

@sanxchep
Copy link

Hi,
I just wanted to know what would be the ideal tokenizer model_max_length/max_length during inference of the model.
Does max_length affect generation quality of questions? If yes, then can you briefly explain me why.

Thanksss

P.S - I've been using 2048 as my max length.

@yeswanthkuruba
Copy link

Ideal max length doesn't depends in the parameter size of the model. it depends on the trained data. As long as your computation support you can increase the max length but, the better output would be generated based on the trained data max length.

So I prefer to use the trained data max length as inference max length.

@sanxchep
Copy link
Author

Ideal max length doesn't depends in the parameter size of the model. it depends on the trained data. As long as your computation support you can increase the max length but, the better output would be generated based on the trained data max length.

So I prefer to use the trained data max length as inference max length.

Okay so then what was the max_length used during training?
Are any training scripts provided?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants