Skip to content

Latest commit

 

History

History
79 lines (54 loc) · 4.97 KB

release-v1-perf-scale-benchmarks.md

File metadata and controls

79 lines (54 loc) · 4.97 KB

Benchmarks

This document shares some of the performance benchmarks observed as part of the v1 release. It will talk about what this means for an end-user's perf and scale story.

Background: How to interpret the numbers?

KubeVirt is an extension for Kubernetes that includes a collection of custom resource definitions (CRDs) served by kubevirt-apiserver. These CRDs are managed by controllers. Due to the distributed nature of the system, understanding performance and scalability data becomes challenging without taking specific assumptions into account. This section aims to provide clarity on those assumptions.

  1. The data presented in the document is collected by periodic-kubevirt-e2e-k8s-1.25-sig-performance.
  2. The test suite includes three tests:
    1. It creates 100 minimal VMIs, with a small pause of 100 ms between creation of 2 VMIs. The definition of minimal VMIs can be found here. This is represented in the graphs as for VMI, for example vmiCreationToRunningSecondsP50 for VMI
    2. It creates 100 minimal VMs, with a small pause of 100 ms between creation of 2 VMIs. The definition of minimal VMs created can be found here. This is represented in the graphs as for VM, for example vmiCreationToRunningSecondsP50 for VM
    3. It creates VMs with instancetype and preference, the definition can be found here. The benchmarks for this will be added in future releases.
  3. The test waits for the VMIs to go into running state and collects a bunch of metrics
  4. The collected metrics are categorized into two buckets, performance and scale
    1. Performance Metrics: This tells users how KubeVirt stack is performing. Examples include vmiCreationToRunningSecondsP50 and vmiCreationToRunningSecondsP95. This helps users understand how KubeVirt performance evolved over the releases; depending on the user deployment, the numbers will vary, because a real production workload could use other KubeVirt extension points like the device plugins, custom scheduler, different version of kubelet etc. These numbers are just a guidance for how the KubeVirt codebase is performing with minimal VMIs, provided all other variables(hardware, kubernetes version, cluster-size etc) remain the same.
    2. Scalability metrics: This helps users understand the KubeVirt scaling behaviors. Examples include, PATCH-pods-count for VMI, PATCH-virtualmachineinstances-count for VMI and UPDATE-virtualmachineinstances-count for VMI. These metrics are measured on the client side to understand the load generated to apiserver by the KubeVirt stack. This will help users and developers understand the cost of new features going into KubeVirt. It will also make end-users aware about the most expensive calls coming from KubeVirt in their deployment and potentially act on it.
  5. The performance job is run 3 times a day and metrics are collected.
  6. The blue dots on the graphs are individual measurements, and orange line is weekly average

Performance benchmarks for V1 release

vmiCreationToRunningSecondsP50

vmiCreationToRunningSecondsP50 for VMI

vmiCreationToRunningSecondsP50 for VM

vmiCreationToRunningSecondsP95

vmiCreationToRunningSecondsP95 for VMI

vmiCreationToRunningSecondsP95 for VM

Scalability benchmarks for V1 release

PATCH-pods-count

PATCH-pods-count for VMI

PATCH-pods-count for VM

UPDATE-vmis-count

UPDATE-vmis-count for VMI

UPDATE-vmis-count for VM

PATCH-vmis-count

PATCH-vmis-count for VMI

PATCH-vmis-count for VM

PATCH-nodes-count

PATCH-nodes-count for VMI

PATCH-nodes-count for VM