As a Concourse dev, it would be nice to know the impact that a particular release has on the overall performance of Concourse. @jama22 pointed out that since a lot of the runtime backlog revolves around stability, this would help us understand whether the changes made in a particular release improve stability or have unforeseen consequences that impact resiliency or efficiency.
Here are some of the benchmarking metrics that we would like to measure against some predefined workload:
System load average
Min, max and average number of containers
Min, max and average number of volumes
Standard deviation of the above values among all workers (esp. containers and volumes)
HTTP response duration
Number of DB connections
Time taken to schedule a build
This issue is to discuss and track work related to building out a benchmarking tool or process.
The text was updated successfully, but these errors were encountered:
I'd also add Brendan Gregg's book, Systems Performance: Enterprise and the Cloud to a reading list. It's very good at breaking down the way different components and systems can manifest performance problems.