Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About training code and scripts #5

Open
wytbwytb opened this issue Jan 17, 2024 · 6 comments
Open

About training code and scripts #5

wytbwytb opened this issue Jan 17, 2024 · 6 comments

Comments

@wytbwytb
Copy link

It's a nice work. When will you upload the training code and scripts ?

@SteveKGYang
Copy link
Owner

We mostly modified the scripts of FastChat (https://github.com/lm-sys/FastChat) for the fine-tuning process. You can look into that.

@NirmalManoj
Copy link

Hi @SteveKGYang, great work! Can you please release the code used for training bart-large and T5?

@NirmalManoj
Copy link

NirmalManoj commented Feb 6, 2024

@SteveKGYang I want to fine-tune with bart-base, but with the same code, processing, etc. that your team used.

@biirving
Copy link

Why not just release it bro

@Zuhashaik
Copy link

Which fine-tuning method you guys used for this ?

  • Unfreezing only last few layers
  • LoRA
  • Attached an extra decoder layer at last ?

@SteveKGYang
Copy link
Owner

@Zuhashaik We used full fine-tuning, which means all parameters are tuned.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants