Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training data #2

Open
NtaylorOX opened this issue Oct 31, 2023 · 6 comments
Open

Training data #2

NtaylorOX opened this issue Oct 31, 2023 · 6 comments

Comments

@NtaylorOX
Copy link

Great work and repo.

Whilst I'm aware the actual training likely follows general LLM training scripts/flow. It would be nice to see the training scripts. Is there any plan to upload?

@SteveKGYang
Copy link
Owner

Thank you very much for your interest. We mostly modified the training architecture of FastChat for current released parts (mostly SFT) of MentaLLaMA, so I'll point you to their repo for now. But we are working towards further enhancing MentaLLaMA with other techniques such as RLHF, and we will release these codes. Stay tuned!

@NtaylorOX
Copy link
Author

Thanks for the reply. Very helpful, and looking forward to what is to come.

@NtaylorOX
Copy link
Author

Actually, one quick question. To perform the SFT for MentalLLaMA, with the instruction training data for say, the DR task. Do you treat this as a standard auto-regressive objective and combine the "query" and the "gpt-3.5-turbo" response? Just hoping to play around with training some smaller models/architectures myself to have a play

@SteveKGYang
Copy link
Owner

Yes. This is the standard instruction tuning paradigm. I suggest you base on foundation models with SFT/RLHF (e.g. LLaMA2-chat, Vicuna) as they will facilitate your training process, especially with small training datasets.

@NtaylorOX
Copy link
Author

Thought so, was just double checking. Thanks for prompt reply! I'll keep you posted if I develop anything that could be brought into this repo.

@SteveKGYang
Copy link
Owner

Thanks! Any contributions will be appreciated!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants