How did the Attention Mechanism start an AI frenzy? | LM3

preview_player
Показать описание
The attention mechanism is well known for its use in Transformers. But where does it come from? It's origins lie in fixing a strange problems of RNNs.

The source code for the animations can be found here:

These animation in this video was made using 3blue1brown's library, manim:

Chapters
0:00 Introduction
0:22 Machine Translation
2:01 Attention Mechanism
8:04 Outro

Music (In Order):
Helynt - Route 10
Helynt - Bo-Omb Battlefield
Helynt - Underwater
Helynt - Twinleaf Town

Follow me!
Рекомендации по теме
Комментарии
Автор

With that, these are the three videos I had planned out. Do check out the previous ones if you missed them!

What kind of videos would you guys like to see next?

vcubingx
Автор

This is one of the best explanations of attention I have seen so far. Understanding the bottleneck motivation really makes this clear right around 3:15.

scottmcevoy
Автор

you’re doing god’s work brother, thank you for the series

blackveganarchist
Автор

I really like how easy you make it to understand the why of things. I think you've accomplished your goal of making it seem like I could come up with this!

Please cover multi headed self attention next! :)

I am worried that this simple approach skips important pieces of the puzzle though. Transformers do have a lot of moving parts it seems. But it seems like you're only getting started!

antoineberkani
Автор

What a great video mfv I paid attention the whole time

calix-tang
Автор

Thank you for your work! Your videos were very helpful for understanding the evolution of transformers 👍

FlyingHenroxx
Автор

Best explanation when i have found so far, thank you

shukurullomeliboyev
Автор

Thank you, Vivek. Absolutely love your content. Please also keep adding Math content, though. Maybe create a playlist about different functions, limits etc? Whatever suits you.

kevindave
Автор

Nice, time to boost this video in the algorithm by typing out a comment

TheRoganExperienceJoe
Автор

Great material and presentation, thanks a lot for your work! I'd like to see some deep dive into how embeddings work, as we can get embeddings from decoder-only models like GPTs, Llamas, etc. and they use some form of embeddings for their internal representations, right? But there are also encoder-only models like BERT and others (OpenAIs text-embedding models) which are actually used instead. What is their difference and why does one work better than the other? Is it just because of computer differences or are there some inherent differences?

Fussfackel
Автор

What about Q, K a d V matrixes meaning?

maurogdilalla
Автор

If you hate others, your really just hating yourself, because we are all one with god source

aidanthompson
Автор

Weird 3b1b has the same series going on now.

OBGynKenobi