Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add support for GPT-neo and GPT-J #2

Closed
chijames opened this issue Feb 20, 2022 · 1 comment
Closed

Add support for GPT-neo and GPT-J #2

chijames opened this issue Feb 20, 2022 · 1 comment
Labels
enhancement New feature or request

Comments

@chijames
Copy link

Hi,

Thanks for the great project! Do you plan to support GPT-neo and GPT-J down the road?

@LiuYuemei111 LiuYuemei111 added the enhancement New feature or request label Feb 21, 2022
@ShengdingHu
Copy link
Collaborator

Thanks for your comment, the project is basically compatible with any PyTorch-based models.
To apply OpenDelta to a new model that we haven't tested, please follow the instructions in the docs, basically,

  1. visualize and understand the parameter/modular structure of the pre-trained model using Visualization
  2. specify the modified modules according to your knowledge about the delta methods, e.g., if applying Lora, any linear can be specified, if applying adapters, any submodules that have hidden state as the output can be specified.
  3. freeze the irrelevant parameters.

However, as a kind reminder, prefix-tuning is not supported due to its compicated nature to process the attention mask.

If you still have any other questions, feel free to let us know.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants