-
Notifications
You must be signed in to change notification settings - Fork 175
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reproduce perplexity #49
Comments
Hi!
Hope this helps. If you have any additional questions, please feel free to ask. |
Very clear, thx so much. I will try to reproduce it. |
@deciding The current Llama-2-7b checkpoint with wikitext2 ppl=5.91 was obtained as follows. Quantization with blockwise finetuning yields 6.22 ppl. Compared to the version in the
The final model was obtained via end-to-end finetuning (script
|
@Godofnothing really appreciate the tuning details! Besides, may I know the number of a100 GPU hours required for this finetune script? |
@deciding I do not remember exact numbers, I think the first part took 1 day on 2 A 100 and the second one 6 hours on single A100 |
@Godofnothing Cool. Thx a lot for the information 👍 |
In the readme the ppl is
Llama-2-7b | 1x16 | 5.92 | 2.4
In the paper it is:
Llama-2-7b AQLM 2.29 6.29 8.11
When I run locally using the same command as in the readme
it gives me
Llama-2-7b AQLM 2.29 6.45 8.39
Can I know why there is such a mismatch? Thanks for any clarifications.
The text was updated successfully, but these errors were encountered: