Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

SFT of LLama 2 #3582

Closed
gsaivinay opened this issue Jul 18, 2023 · 9 comments
Closed

SFT of LLama 2 #3582

gsaivinay opened this issue Jul 18, 2023 · 9 comments

Comments

@gsaivinay
Copy link
Contributor

Hello,

Just today, Meta open sourced Llama 2 models. Wondering if OA team is considering these.

@andreaskoepf
Copy link
Collaborator

Absolutely. We are already evaluating the models. We will start own fine-tuning runs soon (in a couple of hours).

@Billyroot
Copy link

Billyroot commented Jul 18, 2023 via email

@gsaivinay
Copy link
Contributor Author

We will start own fine-tuning runs soon (in a couple of hours).

This is really great. Please keep us posted if possible, on the training and evaluation progress.

@gsaivinay gsaivinay changed the title LLama 2 is released SFT of LLama 2 Jul 19, 2023
@flozi00
Copy link

flozi00 commented Jul 19, 2023

Is there any place to track the training ?
I running some experiments using lora, no resized embeddings for OA special tokens, sequence length of 4096 tokens
The loss of 13b model is stable and lower than 2048 tokens falcon 7b

@gsaivinay
Copy link
Contributor Author

Is there any place to track the training ?

I'm assuming the training progress will show up here as soon as the process is started, but I might be wrong.

@andreaskoepf
Copy link
Collaborator

I'm assuming the training progress will show up here as soon as the process is started, but I might be wrong.

We publish/copy successful runs afterwards to the public-sft wandb project.

@Billyroot
Copy link

So we just have to wait for the release :). Do you think it might happen next week?

@gsaivinay
Copy link
Contributor Author

well, 13b model is here https://huggingface.co/OpenAssistant/llama2-13b-orca-8k-3319 and it is with 8k context length 🤯

@olliestanley
Copy link
Collaborator

Here is the first oasst dataset SFT model for Llama2 70B. As usual details are on the HF page and future models will also be on HF, so I will close this issue now

https://huggingface.co/OpenAssistant/llama2-70b-oasst-sft-v10

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants