-
Notifications
You must be signed in to change notification settings - Fork 243
Description
The benchmark results feel more and more like a mess due to a number of issues in the way the benchmark is organized and how the results are shown.
-
As the title says,
Cold Runresults make no sense with in-memory and remote storage, so ideally they should not be shown on the generated html page. -
The benchmark rules mention the following:
run.sh: a loop for running the queries; every query is run three times; if it's a database with local on-disk storage, the first query should be run after dropping the page cache;
This rule is not applied to some of the contestants that run on the local disk like Arc which produces wrong results in Combined and Cold Run categories. For instance, Arc uses DuckDB as its query engine: it outperforms DuckDB+parquet in Cold Run by 3.08x, but it's slower than DuckDB in Hot Run by 2.7x - that's quite puzzling. More than that, many if the Arc's Cold Run query execution times are worse than the corresponding Hot Run times.
- The benchmark rules should specify the disk/volume type and its settings. The
machinefield used to havegp2 500GBin the text - now it doesn't, so it looks like contestants can use any volume type, e.g. io2, making the Cold Run results questionable.