trtexec can't compile ONNX model with !n->candidateRequirements.empty() failed. No supported formats for Unsqueeze
#3688
Labels
triaged
Issue has been triaged by maintainers
Description
As reported in huggingface/optimum#1735, a valid ONNX model fails with the latest TRT release:
I tried both with int32 & int64 input dtype and it does not seem to matter.
Environment
TensorRT Version:
nvcr.io/nvidia/tensorrt:24.01-py3
NVIDIA GPU: A100-80GB
NVIDIA Driver Version:
CUDA_DRIVER_VERSION=545.23.08
CUDA Version:
CUDA_VERSION=12.3.2.001
CUDNN Version:
CUDNN_VERSION=8.9.7.29+cuda12.2
Relevant Files
It is 125 MB larger than 25 MB, so uploading here: https://huggingface.co/fxmarty/tiny-gemma-onnx-quantized-trt
Please use
git clone https://huggingface.co/fxmarty/tiny-gemma-onnx-quantized-trt
Steps To Reproduce
Download the above model and run:
Commands or scripts:
trtexec --onnx=model_quantized.onnx --saveEngine=model.plan --minShapes=input_ids:1x400,attention_mask:1x400,position_ids:1x400 --optShapes=input_ids:16x400,attention_mask:16x400,position_ids:16x400 --maxShapes=input_ids:32x400,attention_mask:32x400,position_ids:32x400 --verbose --int8
Can this model run on other frameworks? For example run ONNX model with ONNXRuntime (
polygraphy run <model.onnx> --onnxrt
): Yes it works.The text was updated successfully, but these errors were encountered: