This repository has been archived by the owner on Jan 3, 2023. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 222
Select ResNet-50 model for 2017-Q3 #13
Projects
Comments
Summary
Comparision of two implementationsCandidate 1: Google's benchmark implementation (link)This implementation is based on (the first paper of ResNet), also called ResNetV1. It includes ResNet 18, 34, 50, 101, 152.
Candidate 2: Google's model implementation (link)This implementation is based on (the second paper of ResNet), also called ResNetV2. It includes ResNet 32, 110, 164, 1001.
Running and getting the XLA Ops
TODO
|
While we're here, I also did some benchmarks on the Candidate 1 above.
CPU
CPU MKL
GPU
2xGPU
XLA CPU
XLA GPU
XLA 2xGPU
The CPU results are very poor, this is quite strange. With Candidate 2, when running with Non-XLA CPU I see all CPU threads used to full, but somehow with Candidate 1, we only see 30% usage of CPU threads. Maybe some settings problem. |
Model chosen. |
silee2
pushed a commit
that referenced
this issue
Mar 14, 2018
mbrookhart
pushed a commit
that referenced
this issue
Sep 24, 2018
* add scalar ops, repeat * add a TODO
This was referenced May 11, 2019
Sign up for free
to subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Choose a specific ResNet-50 model implementation for our Q3 work.
This choice is subject to change in the future, for example based on future discussions with the Benchmark team. But it needs to be a reasonable starting point for our development work.
The tangible work-product of this Issue should be a file (perhaps some scripts, perhaps a README, or both) giving the details of this choice.
The text was updated successfully, but these errors were encountered: