Forked from Predictive Horizontal Pod Autoscaler, this is a Custom Pod Autoscaler; aiming to have identical functionality to the Horizontal Pod Autoscaler, however with added predictive elements using statistical models and ability to handle bursty-load.
This uses the Horizontal Pod Autoscaler Custom Pod Autoscaler extensively to provide most functionality for the Horizontal Pod Autoscaler parts.
This autoscaler lets you choose models and fine tune them in order to predict how many replicas a resource should have, preempting events such as regular, repeated high load.
However, statistical models are not perfect and can be easily overfit. To adapt to real world scenarios with bursty-load, ability to override replica count recommendation from predictive model is added on top of the original Predictive Horizontal Pod Autoscaler.
- Functionally identical to Horizontal Pod Autoscaler for calculating replica counts without prediction.
- Choice of statistical models to apply over Horizontal Pod Autoscaler replica counting logic.
- Holt-Winters Smoothing
- Linear Regression
- Allows customisation of Kubernetes autoscaling options without master node access. Can therefore work on managed
solutions such as EKS or GCP.
- CPU Initialization Period.
- Downscale Stabilization.
- Sync Period.
- Initial Readiness Delay.
- Runs in Kubernetes as a standard Pod.
- In work and experimental: Provide burst-awareness based on configured Service Level Objective
This project works by calculating the number of replicas a resource should have, then storing these values and using statistical models against them to produce predictions for the future. These predictions are compared and can be used instead of the raw replica count calculated by the Horizontal Pod Autoscaler logic. The evaluator will then query Prometheus based on a defined SLO (Service Level Objective) and if the SLO is not met, the evaluator will increase the replica count over the amount predicted by the model.
See the wiki for more information, such as guides and references.
See the examples/
directory for working code samples.
Developing this project requires these dependencies:
Burst-awareness functionality requires the following dependencies:
- A service mesh, currently supports Istio >=
1.11.0
- Prometheus >=
2.17.0
Any Python dependencies must be installed by running:
pip install -r requirements-dev.txt
To view docs locally you need some Python dependencies, run:
pip install -r docs/requirements.txt
It is recommended to test locally using a local Kubernetes managment system, such as k3d (allows running a small Kubernetes cluster locally using Docker).
Once you have a cluster available, you should install the Custom Pod Autoscaler Operator (CPAO) onto the cluster to let you install the PHPA.
With the CPAO installed you can install your development builds of the PHPA onto the cluster by building the image
locally, and then pushing the image to the K8s cluster's registry (to do that with k3d you can use the
k3d image import
command).
Finally you can deploy a PHPA example (see the examples/
directory for choices) to test your changes.
Note that the examples generally use
ImagePullPolicy: Always
, you may need to change this toImagePullPolicy: IfNotPresent
to use your local build.
go mod vendor
- generates a vendor folder.make
- builds the Predictive HPA binary.make docker
- builds the Predictive HPA image.make lint
- lints the code.make beautify
- beautifies the code, must be run to pass the CI.make unittest
- runs the unit tests.make doc
- hosts the documentation locally, at127.0.0.1:8000
.make view_coverage
- opens up any generated coverage reports in the browser.