-
Notifications
You must be signed in to change notification settings - Fork 3.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
SFT of LLama 2 #3582
Comments
Absolutely. We are already evaluating the models. We will start own fine-tuning runs soon (in a couple of hours). |
That's good to know.
Le mer. 19 juil. 2023 à 00:49, Andreas Köpf ***@***.***> a
écrit :
… Absolutely. We are already evaluating the models. We will start own
fine-tuning runs soon (in a couple of hours).
—
Reply to this email directly, view it on GitHub
<#3582 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ACVZJTUM2EH3LVCVXGH7TO3XQ4HH5ANCNFSM6AAAAAA2O4RWHI>
.
You are receiving this because you are subscribed to this thread.Message
ID: ***@***.***>
|
This is really great. Please keep us posted if possible, on the training and evaluation progress. |
Is there any place to track the training ? |
I'm assuming the training progress will show up here as soon as the process is started, but I might be wrong. |
We publish/copy successful runs afterwards to the public-sft wandb project. |
So we just have to wait for the release :). Do you think it might happen next week? |
well, 13b model is here https://huggingface.co/OpenAssistant/llama2-13b-orca-8k-3319 and it is with 8k context length 🤯 |
Here is the first oasst dataset SFT model for Llama2 70B. As usual details are on the HF page and future models will also be on HF, so I will close this issue now https://huggingface.co/OpenAssistant/llama2-70b-oasst-sft-v10 |
Hello,
Just today, Meta open sourced Llama 2 models. Wondering if OA team is considering these.
The text was updated successfully, but these errors were encountered: