-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Create h2oGPT 40B based on tiiuae/falcon-40b #216
Comments
|
|
Improved LoRA coverage:
|
8x A100 80GB tiiuae/falcon-40b + oasst1_h2ogpt_graded INSTRUCT TUNING (4-bit)Note: failed with OOM for
|
8x A100 80GB tiiuae/falcon-40b + h2ogpt-oig-oasst1-instruct-cleaned-v3 INSTRUCT TUNING (4-bit)
|
16-bit/80GB across 2x48GB+1x24GB cards |
8xA100 Eval suitehttps://github.com/EleutherAI/lm-evaluation-harness
|
8xA100 ShareGPT Eval 40B
|
1x A6000 Ada ShareGPT Eval 40B 4bit
|
Attempt to improve h2oGPT 40B slightly, based on findings from h2ogpt-gm modelsChanges:
|
Eval Suitesame as #216 (comment)
|
https://huggingface.co/tiiuae/falcon-40b Apache 2.0 model (can't use the -instruct one, since trained on Alpaca)
The text was updated successfully, but these errors were encountered: