

In this approach, you combine multiple workloads in a single cluster. Adding nodes is not instantaneous and could take minutes before those nodes become available for scheduling.Īn alternative is Pod Priority and Preemption.


This approach often works, but costs more as you would have to pay for the resources that are idle most of the time. One obvious solution to this problem is to over-provision your cluster resources to have some amount of slack resources available for scale-up situations. When the application is critical for your product, you want to make sure that these new instances are scheduled even when your cluster is under resource pressure. When a workload is scaled up, more instances of the application get created. It scales your workloads based on their resource usage. Kubernetes is well-known for running scalable workloads.
