Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Port Tensor-RT and model compatibility tests to FX #1634

Closed
Christina-Young-NVIDIA opened this issue Jan 31, 2023 · 2 comments
Closed

Port Tensor-RT and model compatibility tests to FX #1634

Christina-Young-NVIDIA opened this issue Jan 31, 2023 · 2 comments
Assignees

Comments

@Christina-Young-NVIDIA
Copy link
Collaborator

No description provided.

@Christina-Young-NVIDIA
Copy link
Collaborator Author

To find any FX2TRT / dynamo gaps.

@gs-olive
Copy link
Collaborator

Currently, the TorchScript path routinely tests model compatibility for ResNet18, MobileNetV2, EfficientNetB0, ResNet18 with Half precision, and HuggingFace BERT base uncased. I have added a PR which tests the FX path (via specifying ir="fx" directly in torch_tensorrt.compile(...)) for all of these models except BERT, which currently fails to compile with this method and also does not compile using torch_tensorrt.fx.compile(...) with either is_aten=True or is_aten=False, via the current main branch (deda87b). The model also fails to compile with torch._dynamo.optimize(...). An issue will be raised regarding compilation of HuggingFace BERT in FX. This (current) investigative issue will be closed and the PR will remain as a draft until BERT model compilation in FX is supported.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants