filmov
tv
Stanford CS25: V1 I Self Attention and Non-parametric transformers (NPTs)
Показать описание
Aidan is a PhD student at Oxford supervised by Yarin Gal, and one of the cofounders of Cohere. He is fascinated by building massive neural networks and getting them into the hands of more engineers and researchers.
Jannik is a PhD student at the University of Oxford. He is supervised by Yarin Gal and Tom Rainforth. His research interests include Bayesian Deep Learning, Active Learning, and, apparently, building non-parametric models with Transformers.
Neil is a Masters by Research student at the University of Oxford, supervised by Yarin Gal. He is interested in models with relational inductive biases such as Transformers and graph neural networks, as well as Bayesian deep learning.
#NPT
Stanford CS25: V1 I Self Attention and Non-parametric transformers (NPTs)
Stanford CS25: V1 I Transformers United: DL Models that have revolutionized NLP, CV, RL
Stanford CS25: V1 I DeepMind's Perceiver and Perceiver IO: new data family architecture
Stanford CS25: V1 I Mixture of Experts (MoE) paradigm and the Switch Transformer
Stanford CS25: V1 I Transformers in Vision: Tackling problems in Computer Vision
Stanford CS25: V1 I Transformer Circuits, Induction Heads, In-Context Learning
Stanford CS25: V1 I Decision Transformer: Reinforcement Learning via Sequence Modeling
clear voice CS25 Transformers United 2023 Introduction to Transformers w Andrej Karpathy
Stanford CS25: V2 I Neuroscience-Inspired Artificial Intelligence
Cross Attention vs Self Attention
Transformers - Part 1 - Self-attention: an introduction
5 Free LLM Courses RoadMap in 2023
Better Attention is All You Need
Stanford Seminar - Universal Intelligent Systems by 2030 - Carl Hewitt and John Perry
Transformers for Structural Extraction
Mechanistic Interpretability - Stella Biderman | Stanford MLSys #70
What the hell is going on inside neural networks? | Chris Olah
Learning to Throw with a Handful of Samples using Decision Transformers
MedAI Session 26: Towards Generalist Imaging Using Multimodal Self-supervised Learning | Mars Huang
Drew Jaegle | Perceivers: Towards General-Purpose Neural Network Architectures
25. Transformers
Improving Transfer and Robustness of Supervised Contrastive Learning - Dan Fu | Stanford MLSys #62
Transformers: limitations and Vision Transformers [part 3]
Meta-Reinforcement Learning for Causal Discovery - Pitch
Комментарии