-
Notifications
You must be signed in to change notification settings - Fork 2.3k
Closed
Labels
triagedIssue has been triaged by maintainersIssue has been triaged by maintainers
Description
Description
Hi. I upgraded TensorRT from 8.2.1 to 8.6.1 recently. The inference time for the same onnx model takes 2x time more than before. After overriding IProfiler class to print the time consumption for each layer, it shows that ConvTranspose + Add in 8.6.1 takes much more time than which in 8.2.1.
8.6.1:
Reformatting CopyNode for Input Tensor 0 to ConvTranspose_102 + Add_103 0.006144ms
ConvTranspose_102 + Add_103 2.00294ms
8.2.1
ConvTranspose_102 0.072704ms
Add_103 0.0256ms
Is that a bug? How can I solve this problem?
Thanks for any help.
Environment
TensorRT Version: 8.6.1
NVIDIA GPU: NVIDIA GeForce RTX 3060 Laptop GPU
NVIDIA Driver Version: 528.79
CUDA Version: 11.8
CUDNN Version: 8.9.0
Operating System: Windows 11 home 22h2
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
triagedIssue has been triaged by maintainersIssue has been triaged by maintainers