filmov
tv
Lecture 13: Introduction to the Attention Mechanism in Large Language Models (LLMs)
![preview_player](https://i.ytimg.com/vi/XN7sevVxyUM/maxresdefault.jpg)
Показать описание
In this lecture, we learn about the attention mechanism
In particular, we look at 5 aspects:
(1) Why we care about “attention”
(2) RNNs and their limitations
(3) The working of the attention mechanism
(4) History of RNNs, LSTMs, Bahdanau Attention and Transformers
(5) Self attention
0:00 Why we care about “attention”
6:19 4 types of attention mechanism
10:21 Problems with modeling long sequences
16:12 How RNNs work
23:35 RNN Limitations
27:12 Bahdanau Attention Mechanism
42:03 History of RNNs, LSTMs, Attention and Transformers
44:08 Self attention
48:00 Lecture recap
=================================================
=================================================
Vizuara philosophy:
As we learn AI/ML/DL the material, we will share thoughts on what is actually useful in industry and what has become irrelevant. We will also share a lot of information on which subject contains open areas of research. Interested students can also start their research journey there.
Students who are confused or stuck in their ML journey, maybe courses and offline videos are not inspiring enough. What might inspire you is if you see someone else learning and implementing machine learning from scratch.
No cost. No hidden charges. Pure old school teaching and learning.
=================================================
🌟 Meet Our Team: 🌟
🎓 Dr. Raj Dandekar (MIT PhD, IIT Madras department topper)
🎓 Dr. Rajat Dandekar (Purdue PhD, IIT Madras department gold medalist)
🎓 Dr. Sreedath Panat (MIT PhD, IIT Madras department gold medalist)
🎓 Sahil Pocker (Machine Learning Engineer at Vizuara)
🎓 Abhijeet Singh (Software Developer at Vizuara, GSOC 24, SOB 23)
🎓 Sourav Jana (Software Developer at Vizuara)
In particular, we look at 5 aspects:
(1) Why we care about “attention”
(2) RNNs and their limitations
(3) The working of the attention mechanism
(4) History of RNNs, LSTMs, Bahdanau Attention and Transformers
(5) Self attention
0:00 Why we care about “attention”
6:19 4 types of attention mechanism
10:21 Problems with modeling long sequences
16:12 How RNNs work
23:35 RNN Limitations
27:12 Bahdanau Attention Mechanism
42:03 History of RNNs, LSTMs, Attention and Transformers
44:08 Self attention
48:00 Lecture recap
=================================================
=================================================
Vizuara philosophy:
As we learn AI/ML/DL the material, we will share thoughts on what is actually useful in industry and what has become irrelevant. We will also share a lot of information on which subject contains open areas of research. Interested students can also start their research journey there.
Students who are confused or stuck in their ML journey, maybe courses and offline videos are not inspiring enough. What might inspire you is if you see someone else learning and implementing machine learning from scratch.
No cost. No hidden charges. Pure old school teaching and learning.
=================================================
🌟 Meet Our Team: 🌟
🎓 Dr. Raj Dandekar (MIT PhD, IIT Madras department topper)
🎓 Dr. Rajat Dandekar (Purdue PhD, IIT Madras department gold medalist)
🎓 Dr. Sreedath Panat (MIT PhD, IIT Madras department gold medalist)
🎓 Sahil Pocker (Machine Learning Engineer at Vizuara)
🎓 Abhijeet Singh (Software Developer at Vizuara, GSOC 24, SOB 23)
🎓 Sourav Jana (Software Developer at Vizuara)
Комментарии