Closed Form of the Covariance Matrix : Data Science Basics

preview_player
Показать описание
Can we find a closed form of the covariance matrix?

---

Like, Subscribe, and Hit that Bell to get all the latest videos from ritvikmath ~

---

Check out my Medium:
Рекомендации по теме
Комментарии
Автор

I have studied Linear Algebra. Went though Gilbert Strang's book(s). This is the first time someone addressed intuition and related the matrices to real world meaning. Congrats, now I am glad to know what Linear algebra means to me and what it should mean to everyone else. You are😇 a rare and gifted lecturer with deep insight.

rhke
Автор

The lack of views on your videos surprises me. Ritvik, thank you for doing this.
You take away the anxiety of starting fresh.

rohitbastian
Автор

Awesome video, but I think what makes you great is the simple examples you share in your videos. I was able to follow everything, but it would be even better if you showed us how to take the apple/banana example and create the closed form of the covariance matrix here. Just my personal feedback. Other than that, it was wonderful!

uafiewn
Автор

Really loving your videos! Impressive how you boil it down to the most essential stuff and giving context too

paddygthatsme
Автор

Hi Ritvik, I love your videos, but this one took me a while to understand. I actually had to figure out for myself how this formula makes sense. My suggestion: Start with the better known matrix formula for Covariance X'X and from there derive the closed form. Define X_k as [X_ik; X_jk] at the start of the derivation.

gioebinger
Автор

Covariance matrix has the shape of (features, features). I think we should put the (Xi - mu).T before (Xi-mu). That means the matrix (features, samples) @ matrix (samples, features) = matrix (features, features)
And also, we need to divide by (samples - 1) instead of (samples) to avoid underestimating variance

rcketRacoon
Автор

Hi Ritvik, could you post a link to a more formal or detailed version of the proof of the closed form of covariance matrix formula?
Theres several components that are unclear to me:
1. The x_bar_i on the LHS is the i-th component of average across vectors whereas the x_bar_i on the right is the average within the i-th vector. These are two different quantities.
2. When you want the i-th element on the LHS at 6:20, you subscript to get your "i-th element" using the column, instead of the row. I'd imagine that if youre trying to get the i-th element of the k-th column vector that it should be X_{i, k} instead of X_{k, i}

archidar
Автор

really amazing what you do thank you for the help

vishwajithsubhash
Автор

this is very confusing because first it mentioned x_i is a dx1 matrix but later in the calculation we have x_ki, x_kj which is not clear what i and j is. i, j are living ing the R^d which are the components of kth observation, and k is in R^N. So the notation should include i, j to avoid confusion like this. the product on the second line was for S_ij and it was erased later. (i was very confused and it took me 30 mins to figure out why i was confused)

googlesong
Автор

Really nice, I like your new video style!

fyaa
Автор

Why do we not use 1/(N - 1) instead of 1/N to account for sample bias?

erlint
Автор

solve examples for us to understand better

infinityturkson
Автор

Hello sir, request if you can help me with pattern recognition and machine learning basics .

eshanverma
Автор

Love those videos!
Definitely my favorite source of intuitive explanations for Data Science and Statistics (together with Josh from StatQuest)

However, I am slightly confused why we divide by N here. Aren't we actually computing the sample covariance here? That would mean that we have to divide by N-1, right?
As far as I can tell, the wiki page for sample covariance uses almost the same notation, while dividing by N - 1.

sejmou