Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

No me permite instalar TENSORrt #3946

Open
MAFLIXD opened this issue Jun 15, 2024 · 2 comments
Open

No me permite instalar TENSORrt #3946

MAFLIXD opened this issue Jun 15, 2024 · 2 comments

Comments

@MAFLIXD
Copy link

MAFLIXD commented Jun 15, 2024

Description

Environment

TensorRT Version:

NVIDIA GPU:

NVIDIA Driver Version:

CUDA Version:

CUDNN Version:

Operating System:

Python Version (if applicable):

Tensorflow Version (if applicable):

PyTorch Version (if applicable):

Baremetal or Container (if so, version):

Relevant Files

Model link:

Steps To Reproduce

Commands or scripts:

Have you tried the latest release?:

Can this model run on other frameworks? For example run ONNX model with ONNXRuntime (polygraphy run <model.onnx> --onnxrt):

@Essa79
Copy link

Essa79 commented Jun 15, 2024

I'd be glad to assist you with installing TensorRT, even without the specific error message #3946. Here's a comprehensive guide that incorporates insights from previous responses and addresses potential issues:

Prerequisites:

NVIDIA GPU: TensorRT requires an NVIDIA GPU with compute capability 3.0 or higher. You can check your GPU model using nvidia-smi in the terminal (Linux/macOS) or NVIDIA Control Panel (Windows).
CUDA Toolkit and NVIDIA Driver: Install the latest CUDA Toolkit and NVIDIA Driver version that's compatible with your GPU from the NVIDIA website (https://developer.nvidia.com/cuda-toolkit).
Installation Steps:

Choose Installation Method:

Package Manager (Recommended): For Linux distributions like Ubuntu, Debian, or CentOS, you can often use the package manager:
Bash
sudo apt-get install libnvinfer-dev libnvinfer-plugins-dev # For Ubuntu/Debian
sudo yum install cuda-toolkit nvidia-tensorrt # For CentOS/RHEL
From Source (Advanced): If packages aren't available or you need a specific version, download the source code from the NVIDIA NGC catalog (https://developer.nvidia.com/tensorrt) and follow the build instructions.
Verify Installation:

Open a Python interpreter and run import tensorrt. If there's no error, TensorRT is installed correctly.
Troubleshooting Common Issues:

Incorrect CUDA or Driver Version: Ensure they're compatible with TensorRT. Refer to the TensorRT documentation for supported versions.
Missing Dependencies: Install required dependencies like python3-dev or build-essential (Linux) using the package manager.
Permissions: Make sure you have the necessary permissions to install packages. Use sudo if required.
Cache Issues: Sometimes deleting package cache files can help:
Bash
sudo apt-get clean # Ubuntu/Debian
sudo yum clean all # CentOS/RHEL
Use code with caution.
content_copy
Additional Tips:

If you encounter specific errors, search online for solutions related to that error code.
Check TensorRT documentation and forums for known issues and troubleshooting steps (https://docs.nvidia.com/deeplearning/tensorrt/).
If you're still facing issues after trying these steps:

Provide More Details: Share more information about your environment (OS, GPU model, CUDA version, etc.) and the exact error message you're encountering. This will help in providing more tailored assistance.
Consider Using Docker: Docker can simplify installation and environment management by creating a container with pre-configured dependencies.

@lix19937
Copy link

trt install doc see https://docs.nvidia.com/deeplearning/tensorrt/install-guide/index.html#installing
You can choose between the following installation options when installing TensorRT; Debian or RPM packages, a Python wheel file, a tar file, or a zip file.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants