Summary
This was a chapter with tons of information. Our goal was to show you how to scale deployments with Kubernetes. We did this by showing you how to create multiple instances of your application.
We started the chapter by looking at how to define the use of a load balancer and leverage the deployment scale feature in Kubernetes to achieve scalability. With this type of scalability, we also achieve failover by using a load balancer and multiple instances of the software for stateless applications. We also looked into using the HPA to automatically scale our deployment based on load.
After that, we also looked into how we can scale the cluster itself. First, we manually scaled our cluster, and afterward we used a cluster autoscaler to scale our cluster based on application demand.
We finished the chapter by looking into different ways to upgrade a deployed application. First, we explored manually updating YAML files. Then, we delved into two additional kubectl
commands...