-
Notifications
You must be signed in to change notification settings - Fork 2.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to train Stable Video Diffusion model? #239
Comments
sorry,ask me |
|
find a non official one #267 |
Our team has released the SVD training script: |
What are the GPU requirements for fine tuning @hadipash ? |
Currently a lot (64 GB), but we're working on reducing VRAM usage. |
@hadipash thats not bad really. You tested it with sharding across multiple GPU's (eg multiple GPU's with <64gb of VRAM)? |
Currently, a sequence of 4 frames can be trained on a single 64GB GPU. We are working on 1) optimizing memory usage on a single device and 2) implementing distributed training to allow for longer sequences (e.g., 30+ frames). |
@hadipash hello,I tried using LoRA to fine-tune the U-Net with SVD, and even with a batch size of 1, memory overflow occurs on the A100 GPU when the dataset consists of 25-frame videos. Does this mean that model parallel training must be employed, distributing the model parameters across multiple GPUs? |
@DataAIPlayer Not sure about LoRA, haven't integrated it yet. However, for vanilla training - yes, need distributed training as a single 64GB GPU can only fit 4 frames. |
No description provided.
The text was updated successfully, but these errors were encountered: