diff --git a/py/torch_tensorrt/dynamo/utils.py b/py/torch_tensorrt/dynamo/utils.py index a47d21a3d7..0a29c2f743 100644 --- a/py/torch_tensorrt/dynamo/utils.py +++ b/py/torch_tensorrt/dynamo/utils.py @@ -859,18 +859,6 @@ def get_output_dtypes(output: Any, truncate_double: bool = False) -> List[dtype] return output_dtypes -def is_tegra_platform() -> bool: - if torch.cuda.get_device_capability() in [(8, 7), (7, 2)]: - return True - return False - - -def is_thor() -> bool: - if torch.cuda.get_device_capability() in [(11, 0)]: - return True - return False - - def get_cpu_memory_usage() -> Any: return psutil.Process().memory_info().rss / 1024 / 1024 diff --git a/setup.py b/setup.py index d151e0df0c..348ec6f989 100644 --- a/setup.py +++ b/setup.py @@ -729,6 +729,7 @@ def run(self): "packaging>=23", "typing-extensions>=4.7.0", "dllist", + "psutil", # dummy package as a WAR for the tensorrt dependency on nvidia-cuda-runtime-cu13 "nvidia-cuda-runtime-cu13==0.0.0a0", ]