You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.
What I mean is I have 2 GPUs (12GB each).A single Caffe model (Resnet152 with some changes and additions) exceeds 12GB memory and does not fit in a single GPU for training. How can I solve that problem with 2 GPUs?
Can I split this huge model to both the GPUs and make sure they communicate gradients with each other?
If so what all should I change in the solver/train files in MXNET?
The text was updated successfully, but these errors were encountered:
Does MXNET support model parallelism?
What I mean is I have 2 GPUs (12GB each).A single Caffe model (Resnet152 with some changes and additions) exceeds 12GB memory and does not fit in a single GPU for training. How can I solve that problem with 2 GPUs?
Can I split this huge model to both the GPUs and make sure they communicate gradients with each other?
If so what all should I change in the solver/train files in MXNET?
The text was updated successfully, but these errors were encountered: