-
Notifications
You must be signed in to change notification settings - Fork 84
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to use cudnn backend API to do int8x32 convolution calculation on Ampere? #8
Comments
Thank you raising this issue. A sample with int8x32 has been created for you to look at.
Note: There are numerical issues with engine_id=0 and int8x32 vectorCount and has been added to errata. |
Thanks for reply. For project application, I need test the conv_op+scale_op+bias_op+activation_op case(same as using cudnnFusedOpsExecute), in which According to the cudnn8 developer guide manual, it supports Convolution_Pointwise flexibly when the compute capability is above 7.5. |
Hi Zhao, Apologies for the delayed response. We found we do not support the above data type combination because of an internal bug. We have a fix for this and will be part of our future 8.3.0 release and will supported through the run time fusion. Thanks |
Hi Zhao, We have fixed this issue in cudnn v8.3 and have released a sample ConvScaleBiasAct_int8 sample for the same. Let us know if it addresses your use case. Thanks |
Hi @ZhaoJob hope the responses above answer your questions! I'm closing the issue for now. If you have additional questions, please feel free to open a new issue! Thanks |
Can give samples about int8x32 convolution calculation using cudnn backend API?
1、How to create xTensor, wTensor and so on?
2、How to create conv_op node?
3、Other creation.
The text was updated successfully, but these errors were encountered: