This project is supported by:
This is the Custom Pod Autoscaler (CPA) code and base images.
A Custom Pod Autoscaler is a Kubernetes autoscaler that is customised and user created. Custom Pod Autoscalers are designed to be similar to the Kubernetes Horizontal Pod Autoscaler. The Custom Pod Autoscaler framework allows easier and faster development of Kubernetes autoscalers. A Custom Pod Autoscaler can be created by using this project, extending the Docker base images provided and inserting your own logic; see the examples for more information.
- Supports any language, environment and framework; the only requirement is it must be startable by a shell command or HTTP request.
- Supports all configuration options of the Horizontal Pod Autoscaler (downscale stabilisation, sync period etc.)
- Allows fast and easy prototyping and development.
- Abstracts away all complicated Kubernetes API interactions.
- Exposes a HTTP REST API for integration with wider systems/manual intervention.
- Can write autoscalers with limited Kubernetes API or lifecycle knowledge.
- Configuration at build time or deploy time.
- Allows scaling to and from zero.
- Can be configured without master node access, can be configured on managed providers such as EKS or GKE.
- Supports Kubernetes metrics that the Horizontal Pod Autoscaler uses, can be configured using a similar syntax and used in custom scaling decisions.
A Custom Pod Autoscaler has a base program (defined in this repository) that handles interacting with user logic, for example by using shell commands and piping data into them. When developing a Custom Pod Autoscaler you define logic for two stages:
- Metric gathering - collecting or generating metrics; can be calling metrics APIs, running calculations locally, making HTTP requests.
- Evaluating metrics - taking these gathered metrics and using them to decide how many replicas a resource should have.
These two pieces of logic are all the custom logic required to build a Custom Pod Autoscaler, the base program will handle all Kubernetes API interactions for scaling/retrieving resources.
Check out this getting started guide for a quick start for developers.
See the wiki for more information, such as guides and references.
Developing this project requires these dependencies:
To view the docs, you need Python 3 installed:
- Python ==
3.8.5
To view docs locally you need some Python dependencies, run:
pip install -r docs/requirements.txt
It is recommended to test locally using a local Kubernetes managment system, such as k3d (allows running a small Kubernetes cluster locally using Docker).
Once you have a cluster available, you should install the Custom Pod Autoscaler Operator (CPAO) onto the cluster to let you install Custom Pod Autoscalers.
With the CPAO installed you can install your development builds of the CPA onto the cluster by building the image locally, and then build CPAs using the new development image.
Finally you can build a CPA example (see the example/
directory for choices), and then
push the image to the K8s cluster's registry (to do that with k3d you can use the k3d image import
command). Once
the autoscaler's image is available in the registry it can be deployed using kubectl
.
make
- builds the CPA binary.make docker
- builds the CPA base images.make test
- runs the unit tests.make lint
- lints the code.make beautify
- beautifies the code, must be run to pass the CI.make view_coverage
- opens up any generated coverage reports in the browser.make doc
- hosts the documentation locally, at127.0.0.1:8000
.