Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

resenet50 ONNX int8 model inference run in ONNXRUNTIME vs IREE-COMPILED CPU run differ #430

Open
kumardeepakamd opened this issue Feb 12, 2024 · 0 comments
Assignees

Comments

@kumardeepakamd
Copy link

There is 11% match only between ONNXRUNTIME run and IREE compiled inference for VAI quantized int8 model of resenet50.

To reproduce the issue:

git clone https://github.com/nod-ai/SHARK-TestSuite.git
cd e2eshark
(You may need to read the https://github.com/nod-ai/SHARK-TestSuite/blob/main/e2eshark/README.md and make sure you have the right pip install etc.)

Replace the path names for hfhome, -c and -i to point to your Hugging Face Home, Torch MLIR build and IREE builds respectively and run as:

python ./run.py --hfhome /proj/gdba/kumar/HF_HOME -c ../../torch-mlir/build -i ../../mainireee/iree-build --torchtolinalg --tests onnx/models/resnet50_vaiq_int8

cd test-run/onnx/models/resnet50_vaiq_int8 and examine the logs, you can 'cat commands.log' and rerun any step you like. You will see that failedinference.log shows you difference in results

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants