Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training on multiple gpu #92

Open
iedmrc opened this issue Jul 29, 2019 · 4 comments
Open

Training on multiple gpu #92

iedmrc opened this issue Jul 29, 2019 · 4 comments

Comments

@iedmrc
Copy link

iedmrc commented Jul 29, 2019

Hi,
What would be the best way of training gpt-2-simple on multiple GPUs? Is there any built-in way to do multi-gpu training? If not, would you consider adding this feature as an enhancement? At least, splitting batches on to different GPUs would be nice especially in the '345M' case which needs a lot of memory and compute power.

Thanks.

@saippuakauppias
Copy link

@huntrontrakkr , Is your fork still in development? Or how to run it so that it uses all the GPU?

@huntrontrakkr
Copy link
Contributor

@saippuakauppias Sorry for the delay - been gone for a bit. Fork works fine for my uses, may need refinement otherwise. There is a new setting called "Multi_GPU" that one can turn on to activate it - will go through and index all the gpus then split up the blocks appropriately. The CLIs still need to be extended to allow for the option, but otherwise it works great.

@huntrontrakkr
Copy link
Contributor

huntrontrakkr commented Oct 4, 2019

Just pushed up, should be in a good place now.

I may put in a pull request.

@saippuakauppias
Copy link

I think @minimaxir will be happy to get PR!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants