Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Error int8 support for jetson tx2 #26

Open
engineer1109 opened this issue Nov 4, 2018 · 8 comments
Open

Error int8 support for jetson tx2 #26

engineer1109 opened this issue Nov 4, 2018 · 8 comments

Comments

@engineer1109
Copy link

Error: CUDNN_STATUS_ARCH_MISMATCH - This GPU doesn't support DP4A (INT8 weights and input)

cudnnstat = 6

So jetson tx2 not support for quantized?

@AlexeyAB
Copy link
Owner

AlexeyAB commented Nov 4, 2018

It seems yes, jetson tx2 doesn't support INT8-quantization (DP4A).
This is strange, because jetson tx2 is Pascal architecture and compute capability (CC) = 6.2 that is higher than 6.0: https://en.wikipedia.org/wiki/CUDA#GPUs_supported

What CUDA, cuDNN do you use?
Can you show output of command nvidia-smi?

@engineer1109
Copy link
Author

nvidia-smi don't support jetson tx2?
cuda 9.0 cudnn 7.1.5
Is there any way to make jetson tx2 support for fp16?

@AlexeyAB
Copy link
Owner

AlexeyAB commented Nov 4, 2018

nvidia-smi don't support jetson tx2?

Any desktop GPU supports nvidia-smi.

Is there any way to make jetson tx2 support for fp16?

jetson tx2 supports fp16, but it doesn't have Tensor Cores, so fp16 will not be faster than fp32 on jetson tx2.

@engineer1109
Copy link
Author

jetson tx2 really not support nvidia-smi
it is not a dependent gpu.
Its gpu memory shares with cpu memory.

@AlexeyAB
Copy link
Owner

AlexeyAB commented Nov 4, 2018

oh yeah nvidia-smi doesn't work on tegra (jetson tx2)

so I think it doesn't support DP4A (INT8).

You can only try to use XNOR (1-bit) quantization by training these models:

@lynnw123
Copy link

lynnw123 commented Nov 9, 2018

I met the same issue, it can compile successfully with GPU, but when do the INT8 inference, it complains about the mismatch. I also use cuda 9.0 and cudnn 7.1.5 on TX2. In TX2, I use "sudo ~/tegrastats" to monitor the GPU usage since nvidia-smi is not working.
error

@AlexeyAB
Copy link
Owner

AlexeyAB commented Nov 9, 2018

@Yinling-123 TX2 doesn't support INT8 optimizations.

@aniketvartak
Copy link

oh yeah nvidia-smi doesn't work on tegra (jetson tx2)

so I think it doesn't support DP4A (INT8).

You can only try to use XNOR (1-bit) quantization by training these models:

@AlexeyAB will you share models trained with XNOR quantization with us?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants