Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Chatllama] it seems you don't support flan_t5_xl to generate rewards training data #241

Open
lonelydancer opened this issue Mar 10, 2023 · 3 comments

Comments

@lonelydancer
Copy link

you claim"If you prefer avoiding external paid APIs, we suggest using HuggingFace’s models (e.g. flan_t5_xl) as described in more detail in the Supported models section."
howerver,
python artifacts/generate_rewards.py ./artifacts/datasets/reward_training_data.json --model flan_t5_xl

self.llm = LLMChain(llm=openai_llm, prompt=prompt_template)

@PierpaoloSorbellini
Copy link
Collaborator

PierpaoloSorbellini commented Mar 10, 2023

Hi @lonelydancer thank you for the feedback.
Yep you are right, we actually added the support to T5 only in extend_rlhf_dataset.py . The error in the readme is due to the fact that we wanted to update the documentation and code as soon as possible to fix as many issues as possible in the shortest amount of time. Unfortunately, we didn't make to add all the things that we want to, as a result that sentence wasn't properly removed from the readme.md. Thank you for spotting the mistake. We will adjust this inconsistency as soon as possible and possibly add the feature.

@avacaondata
Copy link

I've faced the same issue

@PierpaoloSorbellini PierpaoloSorbellini changed the title it seems you don't support flan_t5_xl to generate rewards training data [Chatllama] it seems you don't support flan_t5_xl to generate rewards training data Mar 14, 2023
@pengwei-iie
Copy link

I've faced the same issue, have you fixed it?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants