Kubernetes pod autoscaling for beginners

preview_player
Показать описание

In this episode, were taking a look at how to scale pods on Kubernetes based on CPU or Memory usage. This feature in Kubernetes is called the Horizontal Pod autoscaler.
Before scaling its important to understand your resource usage for the service you wish to scale.
We take a look at resource requests and limits and how they play a key role in autoscaling.

Checkout the source code below 👇🏽 and follow along 🤓

Also if you want to support the channel further, become a member 😎

Checkout "That DevOps Community" too

Source Code 🧐
--------------------------------------------------------------

If you are new to Kubernetes, check out my getting started playlist on Kubernetes below :)

Kubernetes Guide for Beginners:
---------------------------------------------------

Kubernetes Monitoring Guide:
-----------------------------------------------

Kubernetes Secret Management Guide:
--------------------------------------------------------------

Like and Subscribe for more :)

Follow me on socials!

Music:

Рекомендации по теме
Комментарии
Автор

In this episode we learn how to scale pods with the horizontal pod autoscaler.

MarcelDempers
Автор

the best video I ever watched on the internet explaining HPA

ibrahemazad
Автор

I read many articles on many sites and watch many videos to understand pod autoscaler, but all this time, I just needed to watch this video. Thank you.

tiagomedeiros
Автор

Great content as usual, and the production quality is constantly getting better too! Awesome

yovangrbovich
Автор

Nice discover I like the way you explaining dude thanks for effort.I subscribe and will let other people know you

emergirie
Автор

Such a well-done video! Can't believe you haven't gone huge yet. I don't usually comment on YouTube but I felt compelled this time. Looking forward to going through more of your library of content as I get more into Kubernetes and DevOps in general.

happy
Автор

Clearly explained and really useful for beginners, excellent work! May you kindly reply my small question: how can we estimate the resources request and limit for some specific pods?

dangvu
Автор

Not having to provision infrastructure is awesome, thank you for the great video.

DevsLikeUs
Автор

Good lecture. Good presentation. Interesting fast and to the point. Good job man!!! Keep it coming and thanks. Deserved my SUB definitely. :)

nikoladacic
Автор

Thank you so much for making this concept easy to understand. Actually, I was also struggling setting the values of cpu requests and limits in the deployment, because in my Kubernetes even when the replicas increase, it starts running all pods with same load and didn't distribute evenly among the pods to make it come down and I have faced bad behaviour of scaling in my cluster. I have no clue what is happening

prabhatnagpal
Автор

Again, great content delivered in an easy way and also essy to reproduce. Thanks!

torbendury
Автор

Such great work deserves like and comment))

inf
Автор

Absolutely useful video, you saved my job 🤣 thanks a ton mate!

ankitguhe
Автор

Amazing, thank you very much, loved the edition and the concise way o explaining

elmeroranchero
Автор

Thank you very much! Please make a video on kubernetes e2e testing.

Gandolfof
Автор

You're awesome! kudos to your efforts

AmjadW.
Автор

Absolutely killer video my man, much appreciated. Noob question: does the metrics server require a separate node for a production deployment? Or does it just run in the same k8s service process, the way a plugin would? It would be useful to have a better idea of how this maps to actual cloud infra in terms of VMs/nodes, etc.

martinzen
Автор

How do you select a good minimum pod count for the hpa? I see this constant oscillation of it scaling up and down. Should i set my minimum above my normal load?

janco
Автор

Congrats on the excellent and well-explained video. However as your example at 7:39, the only resource scaled is CPU, not MEMORY (after scaling up to 4 replicas the memory of each pod remain unchanged). I wonder is this something obvious? And if so how can we actually scale base on memory consumed?

vuhaiang
Автор

Thanks Marcel. This all is load based - is there a way where I can define it time based e.g. if there is a heavylifting job runs on my cluster between 2-4 AM and I can not afford to miss it?

sachin-sachdeva