As a result:
kubectl topcan collect Prometheus metrics via the adapter;
- autoscaling/v2 can be used for scaling applications (HPA);
- Prometheus data can be obtained using the Kubernetes API and utilized in other modules (Vertical Pod Autoscaler, etc.).
The following parameters serve as a basis for scaling:
- CPU (of the Pod),
- memory (of the Pod),
- rps (of the Ingress) — over 1,5,15 minutes (
- CPU (of the Pod) — over 1,5,15 minutes (
cpu_Nm) — average CPU utilization over N minutes,
- memory (of the Pod) — over 1,5,15 minutes (
memory_Nm) — average Memory utilization over N minutes,
- any Prometheus metrics and any queries based on them.
How does it work?
This module registers
k8s-prometheus-adapter as an external API service that extends the capabilities of the Kubernetes API. When some Kubernetes component (VPA, HPA) needs information about the resources consumed, it requests the Kubernetes API, which, in turn, proxies that request to the adapter. The adapter figures out (using its configuration file) how to calculate the metric and sends a request to Prometheus.