[QNN] Add hardswish int8 impl using table lookup#11700
[QNN] Add hardswish int8 impl using table lookup#11700AndrewZhaoLuo merged 13 commits intoapache:mainfrom
Conversation
|
I'll take a look today or tomorrow |
By now 2 gpu ci tasks and Hexagon tasks failed. Could you please help to figure out the error? Thanks in advance. |
AndrewZhaoLuo
left a comment
There was a problem hiding this comment.
LGTM, the hexagon was probably just flaky so you can rerun CI with an empty commit:
git commit -m 'jostle ci' --allow-empty and git push
As for the other errors, see the comment
Thanks for your kind help. I have fix the typo and recommit with It seems ci doesn't skip tasks for Hexagon? @AndrewZhaoLuo |
|
@zhaoyang-star I checked our CI on TVM main and I don't see this error. Can you reproduce the error locally using ci_hexagon docker image. It might be related to your PR |
|
|
Have fixed error and triggle ci again. But the status of ci is pending for a few hours. |
|
Nah, it's just flaky -- I think there was some planned work with the CI machines so you got hit with the outage. Just jostle one more time and let's get this merged! Edit: #11914 <-- due to this. |
|
Also apparently you can do this: |
|
@tvm-bot rerun |
|
One more question: |
|
Hmm, do you have a profiler report? I am curious since I would expect runtimes to be better vs dq - fp32 - q. Do you have a repo to reproduce? |
Based on
About 3% speedup by using LUT. I also tried yolov5 with hswish model, which is about 9% speedup by LUT. |
|
Hmm yeah this makes sense, i would expect LUT to be slower than ReLU as it requires more memory access. I suspect perhaps the activation functions just don't take much time? Really ReLU is close to the fastest you can go probably. You can maybe see the upper value for speedup by removing all activations. Still technically a little bit of improvement! |
* v1 * [QNN] Add hardswish int8 impl using table lookup * format * format * fix * fix utest * fix ci error * jostle ci * triggle ci * remote nn * jostle ci * fix
Maybe there is something wrong when I created the resnet50 with hswish.
|
* v1 * [QNN] Add hardswish int8 impl using table lookup * format * format * fix * fix utest * fix ci error * jostle ci * triggle ci * remote nn * jostle ci * fix
* v1 * [QNN] Add hardswish int8 impl using table lookup * format * format * fix * fix utest * fix ci error * jostle ci * triggle ci * remote nn * jostle ci * fix
Registered the hardswish unary elementwise op
@AndrewZhaoLuo @mbrookhart