-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Sampling with multiple GPUs? #29
Comments
We don't currently support it. One way to do so would be to keep half the layers of your model on one GPU, half on another, and spin up the process with two visible GPU's. For eg like here: https://pytorch.org/tutorials/intermediate/model_parallel_tutorial.html You might have to insert .to("cuda:0/1") calls at the correct places for it to work |
@tmlbl Did you ever figure out a workable solution? |
@prafullasd - we need this to have a chance of running 2 cards - to get 48gb of VRAM. related - #142 Could you please reopen this ticket? as an investigation. can use this shell script to find-forks.sh chmod +x find-forks.sh
/find-forks.sh -do there is actual an official branch / this should be mentioned on readme. so did you guys get anywhere with this? I don't have multiple gpus to support @heewooj - do you know if multiple gpus can access more VRAM? if I buy 2 x 3090 nvidia cards / 24gb - will it give 48gb and allow fine tune of 5 billion parameters? |
I am running on a machine at home with 2x8GB GPU and the
5b_lyrics
model runs out of GPU memory, but it appears to only be using device 0. Is there a way to distribute the sampling across 2 physical GPUs?The text was updated successfully, but these errors were encountered: