Scaling Explained Through Kubernetes HPA, VPA, KEDA & Cluster Autoscaler

preview_player
Показать описание
Discover the secrets behind effectively scaling your applications and servers with this dive into Kubernetes scaling mechanisms. This video demystifies the concepts of Horizontal Pod Autoscaler (HPA), Vertical Pod Autoscaler (VPA), Kubernetes Event-driven Autoscaling (KEDA), and Cluster Autoscaler. Learn when and how to use each tool, understand their differences, and catch a glimpse of real-world scenarios that showcase their capabilities.

#KubernetesScaling #PodAutoscaling #KubernetesTutorial #ClusterAutoscaler

▬▬▬▬▬▬ 🔗 Additional Info 🔗 ▬▬▬▬▬▬

▬▬▬▬▬▬ 💰 Sponsorships 💰 ▬▬▬▬▬▬

▬▬▬▬▬▬ 👋 Contact me 👋 ▬▬▬▬▬▬

▬▬▬▬▬▬ 🚀 Other Channels 🚀 ▬▬▬▬▬▬

▬▬▬▬▬▬ ⏱ Timecodes ⏱ ▬▬▬▬▬▬
00:00 Scaling Introduction
01:04 Port (Sponsor)
02:08 Scaling Introduction (cont.)
05:38 Vertical Scaling Applications with VerticalPodAutoscaler
13:58 Horizontal Scaling Applications with HorizontalPodAutoscaler
16:04 Horizontal Scaling Applications with KEDA
18:38 Vertical Scaling Nodes
19:33 Horizontal Scaling Nodes Cluster Autoscaler
24:25 What to Use and When to Use It
Рекомендации по теме
Комментарии
Автор

I would like to see a video on dealing with SIGTERM signals. Also, I would be interested in learning how to implement services within a distributed system that can be started in any order. Fantastic video and please do keep them coming.

conradtaylor
Автор

I would really love to hear about graceful shutdowns, because we did that after implementing autoscaling, which was a pain we had in the last year. Pretty please from your fan.

TheVasanthbuddy
Автор

I am experimenting with using a VPA "recreate" for development environments in an GKE Autopilot cluster while setting it to "off" for production, while the HPA is set to 1 replica for development. The theory is that a under utilized development deployments should be placed onto smaller nodes by GKE Autopilot as the pod are re-created by the VPA. Time will tell if it works as planned 😅 I am still developing my process for setting resource requests/limits and scaling.

I agree with some of the other comments, I'd like to hear your thoughts on graceful shutdowns for pods. I am often concerned that my auto scaler will interrupt some work 😅

mzimmerman
Автор

I use the Datadog operator that allows me to scale based on any metrics that I send to Datadog. Anyway, nice video 🙂

civilapalyan
Автор

Hi Viktor thanks for another great video 🎉 I think that autoscaling is one of the most underestimated tasks in the journey towards kubernetes adoption. I've seen many companies struggling with pods restarting continuously, working for a while and then stopping and all sorts of problems which makes them wondering why they adopted kubernetes in the first place! Instead understanding these concepts is fundamental thanks for clarifying them!
I'd also like to see a video in graceful shutdown thank you

IvanRizzante
Автор

I would like to see a video on SIGTERM signal and handling gracefully shut down

hassaanjavaid
Автор

Your book is one of the best out there. Thanks for sharing knowledge

obasoroolakunle
Автор

The most annoying thing about HPA is that it doesn't scale well(pun obviously intended) for low amounts of resources. Lets say you have a web app that uses 400mb idling and does great up to 600mb. When it reaches this point i know site is actually in real use and i want to start some new pods that _start_ at 4gig

DennisHaney
Автор

Is there a very simple way to scale a homelab/non-production environment deploys to 0 when there is no traffic? Yet I found ctrox/zeropod.

eastming
Автор

It would be nice if there were a distributed possibly global and sharable operating system to abstract, reuse and resource the various hardware systems, architectures and providers. And some economic model to make it more affordable for everyone. I will try to learn more; try to do that or find something that does already. I want to create a system that carefully applies my focus; then start creating on that.

hewhointheearthlydomainsee
Автор

Can any of the autoscalers work by committing to my Git repo instead of mutating my Kubernetes resources? I'd like to use a declarative GitOps approach even for scaling.

RobertWinter