Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Using 128 threads by default for cuda kernels #599

Merged
merged 1 commit into from
Mar 22, 2023
Merged

Conversation

coreylowman
Copy link
Owner

Resolves #526

This shaves off a bit of time during forward & backward pass by allowing cuda to split operations across multiple groups.

Shaves off a millisecond from conv2d bench, and when benchmarking with https://github.com/coreylowman/image-classification on batch size 64, the forward time goes from approximately 36ms -> 28ms, and backward from 60ms -> 51ms

@coreylowman coreylowman merged commit d0bdc75 into main Mar 22, 2023
@coreylowman coreylowman deleted the 526-launch-cfg branch March 22, 2023 14:48
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Figure out a better number of threads to launch kernels with
1 participant