We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We鈥檒l occasionally send you account related emails.
Already on GitHub? Sign in to your account
mobilenet_v3_large PTQ/QAT performance regression
SW info
Repro:
git clone -b chuanqiw/inductor_quant https://github.com/pytorch/benchmark.git cd benchmark pip install --no-deps -r requirements.txt pip install --no-cache Jinja2==3.1.2 markupsafe==2.0.1 beartype==0.15.0 && pip install mpmath==1.3.0 python install.py --continue_on_fail export LD_PRELOAD=${CONDA_PREFIX:-"$(dirname $(which conda))/../"}/lib/libiomp5.so:${CONDA_PREFIX:-"$(dirname $(which conda))/../"}/lib/libjemalloc.so export MALLOC_CONF="oversize_threshold:1,background_thread:true,metadata_thp:auto,dirty_decay_ms:-1,muzzy_decay_ms:-1" #QAT TORCHINDUCTOR_FREEZING=1 python run_benchmark.py cpu -m mobilenet_v3_large --torchdynamo inductor --quantize --is_qat --launcher --launcher-args="--throughput-mode" -b 128 --metrics throughputs mv .userbenchmark/cpu qat cat qat/metric* # to see the results #PTQ TORCHINDUCTOR_FREEZING=1 python run_benchmark.py cpu -m mobilenet_v3_large --torchdynamo inductor --quantize --launcher --launcher-args="--throughput-mode" -b 128 --metrics throughputs mv .userbenchmark/cpu ptq cat ptq/metric* # to see the results
Suspected guilty commit: e592a60 torchbench-mobilenet_v3_large-inference-qat-performance-drop_guilty_commit.log cc @ezyang @msaroufim @bdhirsh @anijain2305 @chauhang @WeizhuoZhang-intel @chuanqi129
The text was updated successfully, but these errors were encountered:
4575d3b
Xia-Weiwen
Successfully merging a pull request may close this issue.
馃悰 Describe the bug
mobilenet_v3_large PTQ/QAT performance regression
SW info
Repro:
Suspected guilty commit: e592a60
torchbench-mobilenet_v3_large-inference-qat-performance-drop_guilty_commit.log
cc @ezyang @msaroufim @bdhirsh @anijain2305 @chauhang @WeizhuoZhang-intel @chuanqi129
The text was updated successfully, but these errors were encountered: