Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to improve the use of cpu #29

Closed
shushan2017 opened this issue Apr 21, 2018 · 1 comment
Closed

How to improve the use of cpu #29

shushan2017 opened this issue Apr 21, 2018 · 1 comment

Comments

@shushan2017
Copy link

I run the dnc example and found that in 32-core machines, the CPU usage is not high. Is there any way to make the CPU work at full capacity and improve efficiency? Thank you

@dm-jrae
Copy link
Contributor

dm-jrae commented May 14, 2018

I would recommend data-parallelism versus model parallelism here, see https://www.tensorflow.org/deploy/distributed#replicated_training. The general idea is that you want to split your input training batch across multiple groups of CPUs and have several instantiations of the DNC train on these partitions, instead of hoping the model can parallelize very well over 32 machines for a single minibatch.

@dm-jrae dm-jrae closed this as completed May 14, 2018
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants