-
Notifications
You must be signed in to change notification settings - Fork 9.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Confused about the "hf" meaning. #612
Comments
hugging-face-format |
Isn't it human feedback? The 70b-hf is the chat tuned one with RLHF, the 70b base, isn't RLHF'ed and not trained for conversational data. |
That would make sense to me if there wasn't already a Llama-2-chat and a Llama-2-chat-hf |
AFAIK, the chat model is trained for conversations, but they went further and refined the model to be more helpful using RLHF. |
Gotcha. That makes a bit more sense. Thanks. |
Seems like this issue has been answered, thanks. Closing |
so the highest rated answer here says "hugging-face-format" but another answer says the HF is cause its RLHF (Reinforced Learned Human Feedback). Which is it? |
I can't say for sure but, I can tell that ie... it looks like |
The model card on HF is clear on this: meta-llama/Llama-2-7b: meta-llama/Llama-2-7b-hf: |
Right answer, i guess you are the first one who notices the difference in model cards. |
So, could any one tell me the "hf" mean in Llama-2-70b-hf? What's the difference between Llama-2-70b-hf and Llama-2-70b in hugging face?
"hf" means fp16? or hugging-face-format?
The text was updated successfully, but these errors were encountered: