Deep Learning Lecture 5: Regularization, model complexity and data complexity (part 2)

preview_player
Показать описание
Course taught in 2015 at the University of Oxford by Nando de Freitas with great help from Brendan Shillingford.
Рекомендации по теме
Комментарии
Автор

Was reading about how Coursera and other portals carry out significant data analysis on how to make lecture videos more engaging - having the right background colors, etc. etc.
They should just look at Professor Nando de Freitas. With his outstanding lecture series, he shows how its done and in the end what really matters - his teaching style/method, and passion for the subject, and striking a really good balance between mathematical concepts and programming. As he is teaching in this lecture - he neither "underfits" math for a machine learning course nor does he "overfit"...
"Simple" slides, extremely knowledgable teacher, a very humble teaching style, accessible, explaining history behind certain concepts, mentioning advanced concepts he will not cover but giving enough pointers to go chase them down on our own!...

Some of the best lecture series like - EE263 by Stephen Boyd,  Statistics 110 by Joe Blitzstein and MIT 18.06 by Prof. Gilbert Strang all share similar characteristics.

sbansal
Автор

Hello Professor I have followed all your courses CPSC 340, CPSC 540 and now this. I will say that you have helped me a lot in strengthening my understanding of Machine Learning and I really enjoy watching your lectures.

gauravpathak
Автор

hello

at 35:29 the matrix xTrain has been accessed by the notation xTrain[{{i}}] . any reason for the extra { { } } ? I did run the code and the notation given in lecture slides does work, but whats the reason?

roar
Автор

Hm, I actually chose detla = 1, because both sides were most equal. I expected them to be roughly equal if the errors were only coming from the actual data noise and not from systematic model errors..

GuillermoValleCosmos
Автор

Shouldn't the MSE on the training set always decreases when delta decreases (i.e. when the constraint on theta loosens and we get closer to the MLE)?

LionelJourdain
Автор

at 43:15 doesn't this depend on the train/test split? If the split is 90/10 that would change your decision from a 50/50 split no?

mrrodbruno
Автор

Definitely an excellent series so far! First lecturer I've had to recommend reading the wiki articles and not be so dismissive of easy information. These lectures are definitely easy to follow as they appeal to all backgrounds in Mathematics, Statistics, Computer Science. By the way, does anyone know if it is possible to run torch on OSX 10.10? So far Yosemite has done me no favours.

harryritchie
Автор

In the plot of MSE to the size of training set, how did we pick the baseline of the error to be 4 (the black line)? Is that picked arbitrarily?

Raj-zpiw
Автор

i have the matlab code look in matlab site file exchange "deep learning lecture"

michaelscheinfeild
welcome to shbcf.ru