Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Distill Mistral 7B? #3

Open
ojus1 opened this issue Dec 30, 2023 · 1 comment
Open

Distill Mistral 7B? #3

ojus1 opened this issue Dec 30, 2023 · 1 comment
Labels
enhancement New feature or request wontfix This will not be worked on

Comments

@ojus1
Copy link

ojus1 commented Dec 30, 2023

Mistral-7b is a much better model (and perhaps a teacher) than Llama-2-7b. Would you kindly release checkpoints for a distilled mistral? Would greatly appreciate it!

@GeneZC
Copy link
Owner

GeneZC commented Dec 31, 2023

Thanks for your interests, we will consider using mistral-7b as an alternative teacher.

However, we are concerned that mistral-7b would make no difference from llama-2-7b since we cannot tell which pretraining data has been used by mistral-7b. And the data used for distillation would largely impact the results.

@GeneZC GeneZC added the enhancement New feature or request label Dec 31, 2023
@GeneZC GeneZC added the wontfix This will not be worked on label Apr 15, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request wontfix This will not be worked on
Projects
None yet
Development

No branches or pull requests

2 participants