[Inductor][Quant] Change the QConv output scale name #124246
Dr.CI classification results
{"FAILED":[],"FLAKY":[{"workflowId":9011432287,"workflowUniqueId":37481293,"id":24759895061,"runnerName":"i-037cd09029b16abd9","authorEmail":"leslie.fang@intel.com","name":"inductor / cuda12.1-py3.10-gcc9-sm86 / test (dynamic_inductor_timm, 2, 2, linux.g5.4xlarge.nvidia.gpu)","jobName":"cuda12.1-py3.10-gcc9-sm86 / test (dynamic_inductor_timm, 2, 2, linux.g5.4xlarge.nvidia.gpu)","conclusion":"failure","completed_at":"2024-05-09T05:07:04Z","html_url":"https://github.com/pytorch/pytorch/actions/runs/9011432287/job/24759895061","head_branch":"ciflow/inductor/124246","pr_number":124246,"head_sha":"b291167c577dce7258d76b0893d822e5c73ce935","head_sha_timestamp":"2024-05-08T19:25:08-07:00","failure_captures":["sebotnet33ts_256","FAIL: accuracy=fail_accuracy, expected=pass"],"failure_lines":["sebotnet33ts_256 FAIL: accuracy=fail_accuracy, expected=pass"],"failure_context":["+ python benchmarks/dynamo/check_accuracy.py --actual /var/lib/jenkins/workspace/test/test-reports/training_timm_models.csv --expected benchmarks/dynamo/ci_expected_accuracy/dynamic_inductor_timm_training.csv","+ python benchmarks/dynamo/timm_models.py --ci --accuracy --timing --explain --inductor --dynamic-shapes --dynamic-batch-only --device cuda --training --amp --total-partitions 2 --partition-id 1 --output /var/lib/jenkins/workspace/test/test-reports/training_timm_models.csv","+ [[ dynamic_inductor_timm == aot_inductor ]]","+ [[ dynamic_inductor_timm == perf ]]","+ [[ dynamic_inductor_timm == perf_compare ]]","+ partition_flags=(--total-partitions "$NUM_TEST_SHARDS" --partition-id "$shard_id")","+ [[ -n 1 ]]","+ [[ -n 2 ]]","+ local partition_flags","+ partition_flags=()","+ shift","+ local shard_id=1"],"time":"2024-05-09T05:07:08.772534Z"}],"BROKEN_TRUNK":[{"workflowId":9011429989,"workflowUniqueId":16521569,"id":24759592320,"runnerName":"i-00630b7c7529856cd","authorEmail":"leslie.fang@intel.com","name":"pull / linux-focal-cuda11.8-py3.10-gcc9 / test (distributed, 3, 3, linux.8xlarge.nvidia.gpu)","jobName":"linux-focal-cuda11.8-py3.10-gcc9 / test (distributed, 3, 3, linux.8xlarge.nvidia.gpu)","conclusion":"failure","completed_at":"2024-05-09T04:51:38Z","html_url":"https://github.com/pytorch/pytorch/actions/runs/9011429989/job/24759592320","head_branch":"gh/leslie-fang-intel/93/head","pr_number":124246,"head_sha":"b291167c577dce7258d76b0893d822e5c73ce935","head_sha_timestamp":"2024-05-08T19:25:08-07:00","failure_captures":["!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! KeyboardInterrupt !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!"],"failure_lines":["!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! KeyboardInterrupt !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!"],"failure_context":["+ python test/run_test.py --distributed-tests --shard 3 3 --verbose","+ echo 'Testing distributed python tests'","+ test_distributed","+ [[ distributed == distributed ]]","+ [[ linux-focal-cuda11.8-py3.10-gcc9 == libtorch ]]","+ [[ distributed == \j\i\t\_\l\e\g\a\c\y ]]","+ [[ distributed == executorch ]]","+ [[ distributed == xla ]]","+ [[ distributed == backward ]]","+ [[ linux-focal-cuda11.8-py3.10-gcc9 == aarch64 ]]","+ python -c 'import torch; print(torch.config.parallel_info())'","+ cd test"],"time":"2024-05-09T04:51:43.091416Z"}],"UNSTABLE":[]}