Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add LoRA to BLOOM #3

Merged
merged 5 commits into from
Jun 29, 2022
Merged

Add LoRA to BLOOM #3

merged 5 commits into from
Jun 29, 2022

Conversation

haileyschoelkopf
Copy link
Collaborator

This PR adds LoRA to adapter-transformers' BLOOM implementation.

I have verified that with no LoRA adapter added to the model (just using pfeiffer+inv config for adapters), the performance is exactly the same (with a seed set) as before I added this change.

so this can be safely merged for us to try out.

TODO:

  • test madx_run_clm.py using a LoRA adapter (+ determine appropriate config values.)

@haileyschoelkopf haileyschoelkopf changed the base branch from master to bloom June 28, 2022 14:51
@yongzx
Copy link
Owner

yongzx commented Jun 29, 2022

@haileyschoelkopf can you separate prefix tuning from this PR (by creating another PR)?

I have tested LoRA and the implementation looks correct! Once I have obtained the training curves from madx_run_clm.py (currently running), I would just merge this PR for LoRA.

@haileyschoelkopf
Copy link
Collaborator Author

I've just done this! Will open a separate PR for prefix tuning

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants