-
Notifications
You must be signed in to change notification settings - Fork 712
Fix eval #5955
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix eval #5955
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/5955
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 4158d4c with merge base 36a5bc6 ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
|
This pull request was exported from Phabricator. Differential Revision: D62198560 |
Summary: This PR fixes a bunch of issues in the eval pipeline: - Use the right token for `eot_token_id` - Do not add `bos` and `eos` during `tok_encode` based on this [discussion](https://fburl.com/code/uifmt746). - Update `executorch/examples/models/llama2/tokenizer/tiktoken.py` to be synced with llama 3.1's official [code](https://github.com/meta-llama/llama-models/blob/main/models/llama3/api/tokenizer.py). Majorly updated set of special tokens. - Update `--limit`'s default value to none based on `lm_eval`'s [doc](https://github.com/EleutherAI/lm-evaluation-harness/blob/main/docs/interface.md). It should only be used during test. - Update `--dtype-override`'s default value to none such that we don't override mode's dtype by default. For the context, we observed a gap between ExecuTorch's eval result and SpinQuant's eval result: 19 vs 14. After these fixes, ExecuTorch's eval result is closer to SpinQuant's eval. Differential Revision: D62198560
f105c39 to
bf05031
Compare
|
This pull request was exported from Phabricator. Differential Revision: D62198560 |
bf05031 to
004045b
Compare
Summary: This PR fixes a bunch of issues in the eval pipeline: - Use the right token for `eot_token_id` - Do not add `bos` and `eos` during `tok_encode` based on this [discussion](https://fburl.com/code/uifmt746). - Update `executorch/examples/models/llama2/tokenizer/tiktoken.py` to be synced with llama 3.1's official [code](https://github.com/meta-llama/llama-models/blob/main/models/llama3/api/tokenizer.py). Majorly updated set of special tokens. - Update `--limit`'s default value to none based on `lm_eval`'s [doc](https://github.com/EleutherAI/lm-evaluation-harness/blob/main/docs/interface.md). It should only be used during test. - Update `--dtype-override`'s default value to none such that we don't override mode's dtype by default. For the context, we observed a gap between ExecuTorch's eval result and SpinQuant's eval result: 19 vs 14. After these fixes, ExecuTorch's eval result is closer to SpinQuant's eval. Differential Revision: D62198560
|
This pull request was exported from Phabricator. Differential Revision: D62198560 |
Summary: This PR fixes a bunch of issues in the eval pipeline: - Use the right token for `eot_token_id` - Do not add `bos` and `eos` during `tok_encode` based on this [discussion](https://fburl.com/code/uifmt746). - Update `executorch/examples/models/llama2/tokenizer/tiktoken.py` to be synced with llama 3.1's official [code](https://github.com/meta-llama/llama-models/blob/main/models/llama3/api/tokenizer.py). Majorly updated set of special tokens. Differential Revision: D62198560
004045b to
1ecad67
Compare
|
This pull request was exported from Phabricator. Differential Revision: D62198560 |
Summary: This PR fixes a bunch of issues in the eval pipeline: - Use the right token for `eot_token_id` - Do not add `bos` and `eos` during `tok_encode` based on this [discussion](https://fburl.com/code/uifmt746). - Update `executorch/examples/models/llama2/tokenizer/tiktoken.py` to be synced with llama 3.1's official [code](https://github.com/meta-llama/llama-models/blob/main/models/llama3/api/tokenizer.py). Majorly updated set of special tokens. Reviewed By: mergennachin Differential Revision: D62198560
1ecad67 to
4158d4c
Compare
|
This pull request was exported from Phabricator. Differential Revision: D62198560 |
|
This pull request has been merged in 2027a14. |
Summary:
This PR fixes a bunch of issues in the eval pipeline:
eot_token_idbosandeosduringtok_encodebased on this discussion.executorch/examples/models/llama2/tokenizer/tiktoken.pyto be synced with llama 3.1's official code. Majorly updated set of special tokens.Differential Revision: D62198560