Understanding Azure Kubernetes Services Autoscaler
SantoshKumar V (He/Him)
#Opentowork |Actively looking for new Roles | Multi-Cloud (Azure, AWS, GCP) | DevOps | SRE | DataOps | AI | Cloud Native | Security| Ex- Samsung, Morgan Stanley, Goldman Sachs
What is the Azure Kubernetes Services Autoscaler?
Azure Kubernetes Service Autoscaler (AKSA) is a feature of Azure Kubernetes Service (AKS) that automatically scales, or changes, the number of nodes in a Kubernetes cluster based on the resource usage of the pods in the cluster. It monitors, or watches, the resource usage of the pods and adjusts the number of nodes in the cluster to ensure that the pods have enough resources to run. This helps to ensure that your team does not have to manually intervene when there are changes in workloads.
How does it work?
The AKSA automatically changes the number of nodes in an AKS cluster based on how many resources the pods are using. It uses a Kubernetes controller to watch for changes in resource usage and then adjusts the number of nodes up or down as needed. The AKSA uses an algorithm to decide when to scale the number of nodes. The algorithm takes into account the current resource usage, as well as the desired resource usage that is specified in each pod’s settings. If the pods are using more resources than desired, then additional nodes will be added to provide the extra resources that are needed.
When the resource usage of the pods falls below the desired resource usage, the AKSA will remove nodes from the cluster to reduce the number of resources being used.
The AKSA can also be configured to use custom metrics, this way the system can be scaled based on other criteria than just CPU and Memory usage.
It is important to note that AKSA is not a replacement for monitoring and alerting, it should be used as an additional tool to help ensure that the cluster has the resources it needs to run the pods.
When should you use the AKSA?
When the workload running on an AKS cluster is anticipated to fluctuate dynamically and manual scaling would be challenging or impracticable, the AKSA should be employed. For example, you might use the AKSA in the following situations:
In these circumstances, the AKSA can assist in making sure that the cluster has the resources required to run the pods, even as the demand varies. This could increase accessibility and
It’s also important to note that AKSA can work together with other Azure services to provide a more holistic solution, for example, AKSA can scale the AKS cluster based on the custom metric that comes from Azure Monitor, or it can scale the underlying VMs based on Azure Monitor metric.
领英推荐
Limitation of AKSA
There are a few restrictions with the Azure Kubernetes Service Autoscaler (AKSA) to be aware of.
In conclusion, AKSA can be an excellent tool for ensuring that the AKS cluster has the resources required to run the pods, but it’s important to be aware of its limitations and to use it in conjunction with other monitoring and scaling tools as necessary.
How to get started with AKSA?
There are two ways you can configure AKSA:
2. Or you can update the Autoscaler cluster profile by CLI
and the same way we can use --enable-cluster-autoscaler,?--min-count, and?--max-count?parameters. to enable it again.
Conclusion
Azure Kubernetes Service Autoscaler (AKSA) is a powerful tool that can help ensure that your AKS cluster has the resources it needs to run your pods. By automatically scaling the number of nodes in the cluster based on CPU and memory usage, the AKSA can help prevent resource contention and ensure that your pods have the resources they need to run. However, it’s important to be aware of the limitations of the AKSA, such as its delay in scaling based on pod resource usage and its inability to scale the number of pods. To get started with the AKSA, you’ll need to create an AKS cluster, enable the AKSA, set the minimum and the maximum number of nodes, and configure the AKSA to use custom metrics. It is important to monitor the AKSA and the AKS cluster to test the AKSA with realistic loads before deploying it to production.
DevOps | Cloud | Infra | Agile | CI/CD | Fintech | AI
6 个月according to this documentation: https://learn.microsoft.com/en-us/azure/aks/cluster-autoscaler?tabs=azure-cli new node creation is triggered if there are pods pending creation, yet the existing nodes have run out of resources to create the pods How do you leverage the aks autoscaler to scale nodes based on pod resource usage?