diff --git a/etc/ci/performance/runner.py b/etc/ci/performance/runner.py index 6f396d80dc4a..9a09399f09d9 100644 --- a/etc/ci/performance/runner.py +++ b/etc/ci/performance/runner.py @@ -9,6 +9,7 @@ import itertools import json import os +import platform import subprocess from datetime import datetime from functools import partial @@ -17,6 +18,8 @@ DATE = datetime.now().strftime("%Y%m%d") +MACHINE = platform.machine() +SYSTEM = platform.system() def load_manifest(filename): @@ -144,6 +147,9 @@ def valid_timing(timing, url=None): # able to identify failed tests (successful tests have time >=0). def create_placeholder(testcase): return { + "system": SYSTEM, + "machine": MACHINE, + "date": DATE, "testcase": testcase, "title": "", "navigationStart": 0, @@ -173,6 +179,8 @@ def create_placeholder(testcase): # rather than the url. def set_testcase(timing, testcase=None): timing['testcase'] = testcase + timing['system'] = SYSTEM + timing['machine'] = MACHINE timing['date'] = DATE return timing @@ -245,6 +253,8 @@ def save_result_json(results, filename, manifest, expected_runs, base): def save_result_csv(results, filename, manifest, expected_runs, base): fieldnames = [ + 'system', + 'machine', 'date', 'testcase', 'title', diff --git a/etc/ci/performance/test_all.sh b/etc/ci/performance/test_all.sh index b5d3e1e82c9c..44493f522205 100755 --- a/etc/ci/performance/test_all.sh +++ b/etc/ci/performance/test_all.sh @@ -48,7 +48,7 @@ python3 -m http.server > /dev/null 2>&1 & # MANIFEST="page_load_test/tp5n/20160509.manifest" MANIFEST="page_load_test/test.manifest" # A manifest that excludes # timeout test cases -PERF_FILE="output/perf-$(date +%s).csv" +PERF_FILE="output/perf-$(uname -s)-$(uname -m)-$(date +%s).csv" echo "Running tests" python3 runner.py ${engine} --runs 4 --timeout "${timeout}" --base "${base}" \