Skip to content
This repository has been archived by the owner on Oct 5, 2023. It is now read-only.

Running on Multiple GPU? #66

Open
Joematpal opened this issue Dec 9, 2019 · 5 comments
Open

Running on Multiple GPU? #66

Joematpal opened this issue Dec 9, 2019 · 5 comments
Labels
question Further information is requested

Comments

@Joematpal
Copy link

I have two gtx 1070s, non-sli. Has anyone experienced running the aidungeon on two gpus? What were some of the issues or challenges? The program suggests a beefy amount of gpu ram. Would a configuration with two gpus work well?

@ben-bay
Copy link
Contributor

ben-bay commented Dec 9, 2019

That's a good point. Definitely something we should support.

@ben-bay ben-bay added the question Further information is requested label Dec 9, 2019
@Joematpal
Copy link
Author

Should as in currently do support? Or as in we should get to supporting it?

@WAUthethird
Copy link

WAUthethird commented Dec 10, 2019

Should as in currently do support? Or as in we should get to supporting it?

I would assume he means that it’ll be supported later. There is no multi-GPU support at the moment.

@Akababa
Copy link
Contributor

Akababa commented Dec 10, 2019

I don't know if GPT-2 can be split across GPUs. Probably the best bet is to get quantization working or train a smaller model that uses less VRAM.

@bryanlyon
Copy link

Multi GPU cannot split a model like that without a LOT of work, instead multiGPU would be used to handle multiple inputs at once (not something that this project would benefit from).

Essentially using model-parallelism is much more difficult to implement than data-parallelism and neither would work terribly well with GPT-2 without some significant changes.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

5 participants