Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[REQUEST] Ability to switch out weights when using deepspeed inference #2619

Open
jihan-yin opened this issue Dec 16, 2022 · 0 comments
Open
Labels
enhancement New feature or request

Comments

@jihan-yin
Copy link

Is your feature request related to a problem? Please describe.
If I have multiple fine-tuned versions of the same base model where only a small number of weights are different, I'd like to be able to change out the weights quickly without relaunching deepspeed. Not sure if this is already possible, so marking this issue as a feature request.

Describe the solution you'd like
Have a straightforward and fast way to reload weights to specific modules within a pytorch model, that has already been prepared by deepspeed for multi-gpu single-node inference, across multiple processes.

Describe alternatives you've considered
If not using deepspeed for inference, it is quite simple to reload weights for a small portion of the model.

Additional context
I'm wondering if this is possible for models for which deepspeed inference does not support custom kernels (T5), and models that deepspeed inference does support custom kernels for (GPT)

@jihan-yin jihan-yin added the enhancement New feature or request label Dec 16, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant