How positional encoding in transformers works?

preview_player
Показать описание

Рекомендации по теме
Комментарии
Автор

In my opinion, best explanation so far of positional encoding! Super clear and concise! Thank you very much sir!

emitate
Автор

The best explanation of transformer positional encoding on the internet. Awesome video. Thanks!

cybermanaudiobooks
Автор

Great explanation. Short enough. Detailed enough. Enough talking. Enough showing. Loved the examples.

atabhatti
Автор

I like very concise graphical explanation with the similarity to binary coding and basic linear algebra!

Wesleykoonps
Автор

Fantastic. This was amazing! Best explanation.

JesseNerio
Автор

I couldn't find anywhere why creators of transformer decied to encode the positions in this way and last minute of your video was what I was looking for. Thanks for good explanation

marcinstrzesak
Автор

I'm eternally grateful for this concise explanation, other sources made the positional encoding concept sound so counter-intuitive to grasp

mohammedelsiddig
Автор

Just when I was about to pull the last hair on top of my head, I came across this video. Beautifully Explained. Thank You !

ea_
Автор

1. Why positional encoding is added to the word embedding? Will it changes the semantic value?

2. Why positional encoding use random number produce by sin and cosine... I think it must be simple if we add the one dimension to word embedding storing the position as integer.

Why use such a hard, random, and unpredictable algorithm to encode positions!

temanangka
Автор

How can adding positional encoding to word embedding doesnt change the word semantic meaning?

Example:
Word embedding of "Cat" is [1, 2, 3],
Word embedding of Money is [2, 3, 4].

If the positional encoding is [2, 1, 0] for word "Cat",
positional encoding for word "Money" is [1, 0, -1]

then the positional encoded of both word is [3, 3, 3]

How can "Cat" equal to "Money"?

temanangka