filmov
tv
From Attention to Generative Language Models - One line of code at a time!
![preview_player](https://i.ytimg.com/vi/s3OUzmUDdg8/maxresdefault.jpg)
Показать описание
This in-depth tutorial is about writing Causal Generative Language models from scratch in Pytorch. I explain all the concepts you need to understand what goes under the hood in Transformers and Attention Models. I also tried to add some visualizations to help explain each concept, and walk you through every line of code to explain how all the math works. I aimed for the right balance of complexity of material and simplicity of explanation to make the best learning experience I can give. Follow-up topics such as Multi-Headed Attention, Multi-Query Attention, Grouped Query Attention will be covered in a future episode.
Join the channel on Patreon to receive updates about the channel, and get access to bonus content used in all my videos. Here is the link:
All the codes, slides, and notebooks used in this video have all been shared in the Patreon. Go have fun! :)
Videos you might like:
#pytorch #deeplearning #machinelearning
0:00 - Intro
1:00 - Semantic Similarity
3:00 - Matrix Multiplication
7:00 - Attention Scores
13:20 - Contextual Embeddings
21:18 - Attention with Pytorch
28:43 - Self Attention
30:37 - Causal Masked Attention
33:37 - Transformer Decoder Blocks
38:40 - Next Word Prediction
Join the channel on Patreon to receive updates about the channel, and get access to bonus content used in all my videos. Here is the link:
All the codes, slides, and notebooks used in this video have all been shared in the Patreon. Go have fun! :)
Videos you might like:
#pytorch #deeplearning #machinelearning
0:00 - Intro
1:00 - Semantic Similarity
3:00 - Matrix Multiplication
7:00 - Attention Scores
13:20 - Contextual Embeddings
21:18 - Attention with Pytorch
28:43 - Self Attention
30:37 - Causal Masked Attention
33:37 - Transformer Decoder Blocks
38:40 - Next Word Prediction
Комментарии