Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

why slow when using iter_size. #3808

Closed
cheer37 opened this issue Mar 12, 2016 · 1 comment
Closed

why slow when using iter_size. #3808

cheer37 opened this issue Mar 12, 2016 · 1 comment

Comments

@cheer37
Copy link

cheer37 commented Mar 12, 2016

I used to use small batch size due to small sized gpu memory.
with help of iter_size, i could increase the number of images to calculate the gradient.
But its slow than original one without iter_size.
What makes this?
Thanks.

@seanbell
Copy link

I assume you mean that with the same number of images being computed, it's slower? A larger iter_size (and correspondingly smaller batchsize) has more overhead in synchronizing CUDA threads and launching CUDA kernels. Rather than batching all the images together, instead there are several passes through the data.

Also, from https://github.com/BVLC/caffe/blob/master/CONTRIBUTING.md:

Please do not post usage, installation, or modeling questions, or other requests for help to Issues.
Use the caffe-users list instead. This helps developers maintain a clear, uncluttered, and efficient view of the state of Caffe.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants