filmov
tv
Lecture 12.1 Self-attention
![preview_player](https://i.ytimg.com/vi/KmAISyVvE1Y/maxresdefault.jpg)
Показать описание
ERRATA:
- In slide 23, the indices are incorrect. The index of the key and value should match (j) and theindex of the query should be different (i).
- In slide 25, the diagram illustrating how multi-head self-attention is computed is a slight departure from how it's usually done (the implementation in the subsequent slide is correct, but these are not quite functionally equivalent). See the slides PDF below for an updates diagram.
In this video, we discuss the self-attention mechanism. A very simple and powerful sequence-to-sequence layer that is at the heart of transformer architectures.
Lecturer: Peter Bloem
- In slide 23, the indices are incorrect. The index of the key and value should match (j) and theindex of the query should be different (i).
- In slide 25, the diagram illustrating how multi-head self-attention is computed is a slight departure from how it's usually done (the implementation in the subsequent slide is correct, but these are not quite functionally equivalent). See the slides PDF below for an updates diagram.
In this video, we discuss the self-attention mechanism. A very simple and powerful sequence-to-sequence layer that is at the heart of transformer architectures.
Lecturer: Peter Bloem
Lecture 12.1 Self-attention
Self-attention in deep learning (transformers) - Part 1
Self Attention in Transformer Neural Networks (with Code!)
[ML 2021 (English version)] Lecture 10: Self-attention (1/2)
CS 182: Lecture 12: Part 1: Transformers
Lecture 12: Flash Attention
Rasa Algorithm Whiteboard - Transformers & Attention 1: Self Attention
Attention in transformers, visually explained | Chapter 6, Deep Learning
Multi Head Attention Part 2: Entire mathematics explained
EfficientML.ai Lecture 12 - Transformer and LLM (Part I) (MIT 6.5940, Fall 2023)
EE599 Project 12: Transformer and Self-Attention mechanism
Transformers - Part 1 - Self-attention: an introduction
[ML 2021 (English version)] Lecture 11: Self-attention (2/2)
[ML 2021 (English version)] Lecture 12: Transformer (1/2)
5 tasks transformers can solve?
Self-Attention Equations - Math + Illustrations
BE544 Lecture 12 - Vision Transformers (ViTs) Explained
#12 QANet: Combining Local Convolution With Global Self-Attention for Reading Comprehension
Coding the self attention mechanism with key, query and value matrices
Attention is all you need. A Transformer Tutorial. 1: Self-Attention
Ali Ghodsi, Deep Learning, Attention mechanism, self-attention, S2S, Fall 2023, Lecture 9
Why is Self Attention called 'Self'? | Self Attention Vs Luong Attention in Depth Lecture ...
DeepMind x UCL | Deep Learning Lectures | 8/12 | Attention and Memory in Deep Learning
Self Attention and Transformers (ASU Intro AI Fall 2023 Final Lecture)
Комментарии