Neural networks [6.7] : Autoencoder - contractive autoencoder

preview_player
Показать описание

Рекомендации по теме
Комментарии
Автор

I just want to say you, 'Thank u; in person.An awesome representation!!

NikhilYadala
Автор

Hi Hugo! Great video..I really appreciate that you have taken some time out of your busy schedule for educating people like us :-)

Автор

Thanks for this video, it's very clear with you representation!

junjieli
Автор

@4:30 I don't understand why you say that the regularizer throws away all information in x_k. It throws away only the *noise* in x_k and it does that by flattening out h near x_k, AFAICT.
To clarify, your explanation is 90% aligned with my intuition, but there's that "throws away all information" part which I don't get.
On second thought, I think I get what you mean. You mean that if the regularizer is not contrasted by the reconstruction term then it will throw away all information, right?
But that doesn't need to be the case as it's enough that the jacobian is 0 at the samples (edit: but we're using continuous functions...). But maybe SGD is good at finding degenerate/uninteresting solutions, which is why we needed the regularizer in the first place (i.e. to avoid the "copying").

kiuhnmmnhuik
Автор

Hİ Hugo, Are the contractive autoencoders related with contraction mappings? Since the regularization term could be thought as a squared Lİpschitz constant of the h which may give some degree of control on the learned manifold? If not, why? Thanks!

emirceyani
Автор

Thank you so much for this video. Love the visualisations!!

laimeilin
Автор

Thank you Mr. Larochelle for this really helpful presentation.

behradtajalli
Автор

I have a question -- Can auto encoders be used for building good word embeddings (like word2vec, GloVe etc.)? I have a TREC style text corpus. My target is to find contextually similar words
using cosine similarity between their embedding vectors. For example, I have a word "emergency" and I want to find words like "crisis", disaster" etc. which are supposed to have high contextual similarity
(hence high cosine similarity) with "emergency". Thanks in advance!

Автор

Great explanation thank you! Which book do you recommend (apart from Deep Learning Goodfellow book) to deep dive into this concepts?

jamespaz
Автор

thank you, this explains the intuition behinds the CAE quite well..I finally get it!

planktonscollective
Автор

So for those inputs (x_ks) which dont change much you do not want significant weights. But that partial derivate of h(x_t) w.r.t x_k is just weight isnt it? And if so this sounds a lot like regular L2 norm for the purposes of regularization. Am I missing something?

kimishpatel
Автор

Hi Hugo, what happens when you combine contractive AE and DAE? So you perturb your input, compute the loss with respect to the untouched input, but have the jacobian in the loss function as well?

Herrgolani
Автор

But isn't it violating the purpose of dimensionality reduction?

TechFinTv
Автор

Great video ! Intuition explained to the point. Thanks a lot

MohitTare
join shbcf.ru