NLP Demystified 12: Capturing Word Meaning with Embeddings

preview_player
Показать описание

We'll learn a method to vectorize words such that words with similar meanings have closer vectors (aka "embeddings"). This was a breakthrough in NLP and boosted performance on a variety of NLP problems while addressing the shortcomings of previous approaches. We'll look at how to create these word embeddings and how to use them in our models.

Timestamps
00:00:00 Word Vectors
00:00:37 One-Hot Encoding and its shortcomings
00:02:07 What embeddings are and why they're useful
00:05:12 Similar words share similar contexts
00:06:15 Word2Vec, a way to automatically create word embeddings
00:08:08 Skip-Gram With Negative Sampling (SGNS)
00:17:11 Three ways to use word vectors in models
00:18:48 DEMO: Training and using word vectors
00:41:29 The weaknesses of static word embeddings

This video is part of Natural Language Processing Demystified --a free, accessible course on NLP.

Рекомендации по теме
Комментарии
Автор

CORRECTION: at 1:33, that should be "20, 000-element vector" .

Timestamps
00:00:00 Word Vectors
00:00:37 One-Hot Encoding and its shortcomings
00:02:07 What embeddings are and why they're useful
00:05:12 Similar words share similar contexts
00:06:15 Word2Vec, a way to automatically create word embeddings
00:08:08 Skip-Gram With Negative Sampling (SGNS)
00:17:11 Three ways to use word vectors in models
00:18:48 DEMO: Training and using word vectors
00:41:29 The weaknesses of static word embeddings

futuremojo
Автор

So I attend a really, REALLY prestigious university in the US and I took a course on Neural Networks this last term--this video series has higher lecture quality than that. You are very good at teaching these concepts

moistnar
Автор

I have watched 5 videos on this subject in the last 2 days, and browsed dozens. This one is OUTSTANDING!!! By far the best i have seen. Wow!
I will do the whole NLP course. Very grateful for Huge effort it took

michaelm
Автор

In SGNS, when you are talking about matrices of context and target embeddings (10000 * 300), what do these matrices have/contain before the training has started (collection of one hot encodings or arbitrary numbers)? At 17:00, I also did not understand how only taking the target word embeddings would be sufficient to capture similarity between words.

lochanaemandi
Автор

Are you planning to do courses on other machine learning topics, such as computer vision?

dmytrokulaiev
Автор

I saw openai also provide embedding tool. It seems that this make easier than the old library such as NLTK, spacy, making them outdated? It make these concepts as a black box for us. We do not need to know in detail if only to use it.

caiyu