filmov
tv
Lecture 16: Causal Self Attention Mechanism | Coded from scratch in Python
Показать описание
In this lecture, we learn and code the casual attention mechanism from scratch. We learn about masking, dropout and key concepts involved in causal attention.
This is a very dense lecture consisting of detailed whiteboard notes, mathematics intuition and hands on Python coding.
0:00 Self attention recap
9:30 What is causal attention?
14:25 Coding the casual attention mask in Python
24:45 Data leakage
26:39 Negative infinity masking and softmax
32:33 Dropout in causal attention
36:13 Coding causal attention dropout in Python
40:45 Coding the Causal Attention Class in Python
51:27 register_buffer in PyTorch
53:53 Next steps
PyTorch Upper and Lower Triangular Matrix implementation:
PyTorch Masked_fill implementation:
PyTorch Dropout implementation:
PyTorch register_buffer implementation:
=================================================
=================================================
Vizuara philosophy:
As we learn AI/ML/DL the material, we will share thoughts on what is actually useful in industry and what has become irrelevant. We will also share a lot of information on which subject contains open areas of research. Interested students can also start their research journey there.
Students who are confused or stuck in their ML journey, maybe courses and offline videos are not inspiring enough. What might inspire you is if you see someone else learning and implementing machine learning from scratch.
No cost. No hidden charges. Pure old school teaching and learning.
=================================================
🌟 Meet Our Team: 🌟
🎓 Dr. Raj Dandekar (MIT PhD, IIT Madras department topper)
🎓 Dr. Rajat Dandekar (Purdue PhD, IIT Madras department gold medalist)
🎓 Dr. Sreedath Panat (MIT PhD, IIT Madras department gold medalist)
🎓 Sahil Pocker (Machine Learning Engineer at Vizuara)
🎓 Abhijeet Singh (Software Developer at Vizuara, GSOC 24, SOB 23)
🎓 Sourav Jana (Software Developer at Vizuara)
This is a very dense lecture consisting of detailed whiteboard notes, mathematics intuition and hands on Python coding.
0:00 Self attention recap
9:30 What is causal attention?
14:25 Coding the casual attention mask in Python
24:45 Data leakage
26:39 Negative infinity masking and softmax
32:33 Dropout in causal attention
36:13 Coding causal attention dropout in Python
40:45 Coding the Causal Attention Class in Python
51:27 register_buffer in PyTorch
53:53 Next steps
PyTorch Upper and Lower Triangular Matrix implementation:
PyTorch Masked_fill implementation:
PyTorch Dropout implementation:
PyTorch register_buffer implementation:
=================================================
=================================================
Vizuara philosophy:
As we learn AI/ML/DL the material, we will share thoughts on what is actually useful in industry and what has become irrelevant. We will also share a lot of information on which subject contains open areas of research. Interested students can also start their research journey there.
Students who are confused or stuck in their ML journey, maybe courses and offline videos are not inspiring enough. What might inspire you is if you see someone else learning and implementing machine learning from scratch.
No cost. No hidden charges. Pure old school teaching and learning.
=================================================
🌟 Meet Our Team: 🌟
🎓 Dr. Raj Dandekar (MIT PhD, IIT Madras department topper)
🎓 Dr. Rajat Dandekar (Purdue PhD, IIT Madras department gold medalist)
🎓 Dr. Sreedath Panat (MIT PhD, IIT Madras department gold medalist)
🎓 Sahil Pocker (Machine Learning Engineer at Vizuara)
🎓 Abhijeet Singh (Software Developer at Vizuara, GSOC 24, SOB 23)
🎓 Sourav Jana (Software Developer at Vizuara)
Комментарии