KL Divergence - CLEARLY EXPLAINED!

preview_player
Показать описание
This tutorial explains what KL Divergence is as well as show its derivation using log-likelihood ratio.

Friend link to "What is Entropy?" article:

#kldivergence
#variationalinference
#bayesianstatistics
Рекомендации по теме
Комментарии
Автор

These videos should be highly recommended by YouTube algorithm

orjihvy
Автор

Fantastic tutorial! What I find great is that you anticipate the questions arising in the student's mind and then address them with very satisfying explanations!

Vikram-wxhg
Автор

Your explanations and visualizations are very good! Also you teaching style has the perfect tempo. Thank you very much for this great explanation

hannes
Автор

The best explanation I've heard about KL-Divergence. Keep up the great work.

kaiponel
Автор

I cannot express the gratitude I have for your explanation. What a beautiful soul you are .wow

paedrufernando
Автор

It was fantastic. The most informative video of KL divergence

homakashefiamiri
Автор

Amazing value, Kapil. I like several of the things you do when you teach: refreshing necessary concepts (expectation), the precision of your language and notation, equivalent expressions, and so on. The pace is also great. Thank you very much!

aclapes
Автор

These videos are pure gold. Thank you so much. You can explain incredible well.

TheProblembaer
Автор

was searching for some tutorials on approximate inference and the pre-reqs for it and stumbled upon this, literally my mind got blown with the way you explained the concepts here.

shantanudixit
Автор

this is the most simple and clear explanation of KL divergence, thank you

matiascaceres
Автор

He is good! Very good to say, for example, we want the average difference, but when talking about rv we talk about expected value ... . And many other very careful explanations.

pauledam
Автор

Thank you for the video. I am preparing a paper for my math stats class and after many videos yours gave me the best total explanation with terminology I am familiar with so far.

puck
Автор

Hi, I had one doubt, in 5:18, why do we multiply p_theta(xi) to log(p_theta(xi)/q_phi(xi)) and not multiply q_phi(xi) to log(p_theta(xi)/q_phi(xi))?
In 7:30 you show the variation with q_phi(xi). It seems like the probability distribution that is multiplied with the function of the random variable log(P1(x)/P2(x)) is the probability distribution that appears in the numerator. Is there a reason to not multiple the probability distribution in the denominator?

nikhilsrajan
Автор

i loved this small session on KL Divergence. Thank you sir for this beautiful lecture.

mdekramnazar
Автор

Thank you very much! Your explanations are really clear and neat. Thanks to your video, now I understand KL divergence much much better than I did before.

bluepeace
Автор

This is amaaaazing! What a nicely paced and deep explanation!

sergiobromberg
Автор

You have a talent for teaching. Good explanation.

gauravkumarshah
Автор

Great content! Definitely need more views. Please keep uploading videos.

raghav
Автор

I fell in love with the explanation. Thanks a lot Kapil.

spandanbasu
Автор

Thank youu for great explanation. @9:27 I can't get how come reverse KL divergence has mode seeking behaviour and forward has mean seeking. I understood that P(x) is multimodal gaussian distribution, but what is Q(X) as we needed both distribution for finding K-L divergence.

TJ-zssv