filmov
tv
Jay Alammar on LLMs, RAG, and AI Engineering
Показать описание
Jay Alammar, renowned AI educator and researcher at Cohere, discusses the latest developments in large language models (LLMs) and their applications in industry. Jay shares his expertise on retrieval augmented generation (RAG), semantic search, and the future of AI architectures.
MLST is sponsored by Brave:
Jay Alamaar:
Buy Jay's new book here!
Hands-On Large Language Models: Language Understanding and Generation
TOC:
00:00:00 Introduction to Jay Alammar and AI Education
00:01:47 Cohere's Approach to RAG and AI Re-ranking
00:07:15 Implementing AI in Enterprise: Challenges and Solutions
00:09:26 Jay's Role at Cohere and the Importance of Learning in Public
00:15:16 The Evolution of AI in Industry: From Deep Learning to LLMs
00:26:12 Expert Advice for Newcomers in Machine Learning
00:32:39 The Power of Semantic Search and Embeddings in AI Systems
00:37:59 Jay Alammar's Journey as an AI Educator and Visualizer
00:43:36 Visual Learning in AI: Making Complex Concepts Accessible
00:47:38 Strategies for Keeping Up with Rapid AI Advancements
00:49:12 The Future of Transformer Models and AI Architectures
00:51:40 Evolution of the Transformer: From 2017 to Present
00:54:19 Preview of Jay's Upcoming Book on Large Language Models
Disclaimer: This is the fourth video from our Cohere partnership. We were not told what to say in the interview, and didn't edit anything out from the interview. Note also that this combines several previously unpublished interviews from Jay into one, the earlier one at Tim's house was shot in Aug 2023, and the more recent one in Toronto in May 2024.
Refs:
The Illustrated Transformer
Attention Is All You Need
The Unreasonable Effectiveness of Recurrent Neural Networks
Neural Networks in 11 Lines of Code
Understanding LSTM Networks (Chris Olah's blog post)
Luis Serrano's YouTube Channel
Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks
GPT (Generative Pre-trained Transformer) models
BERT (Bidirectional Encoder Representations from Transformers)
RoPE (Rotary Positional Encoding)
Grouped Query Attention
RLHF (Reinforcement Learning from Human Feedback)
DPO (Direct Preference Optimization)
MLST is sponsored by Brave:
Jay Alamaar:
Buy Jay's new book here!
Hands-On Large Language Models: Language Understanding and Generation
TOC:
00:00:00 Introduction to Jay Alammar and AI Education
00:01:47 Cohere's Approach to RAG and AI Re-ranking
00:07:15 Implementing AI in Enterprise: Challenges and Solutions
00:09:26 Jay's Role at Cohere and the Importance of Learning in Public
00:15:16 The Evolution of AI in Industry: From Deep Learning to LLMs
00:26:12 Expert Advice for Newcomers in Machine Learning
00:32:39 The Power of Semantic Search and Embeddings in AI Systems
00:37:59 Jay Alammar's Journey as an AI Educator and Visualizer
00:43:36 Visual Learning in AI: Making Complex Concepts Accessible
00:47:38 Strategies for Keeping Up with Rapid AI Advancements
00:49:12 The Future of Transformer Models and AI Architectures
00:51:40 Evolution of the Transformer: From 2017 to Present
00:54:19 Preview of Jay's Upcoming Book on Large Language Models
Disclaimer: This is the fourth video from our Cohere partnership. We were not told what to say in the interview, and didn't edit anything out from the interview. Note also that this combines several previously unpublished interviews from Jay into one, the earlier one at Tim's house was shot in Aug 2023, and the more recent one in Toronto in May 2024.
Refs:
The Illustrated Transformer
Attention Is All You Need
The Unreasonable Effectiveness of Recurrent Neural Networks
Neural Networks in 11 Lines of Code
Understanding LSTM Networks (Chris Olah's blog post)
Luis Serrano's YouTube Channel
Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks
GPT (Generative Pre-trained Transformer) models
BERT (Bidirectional Encoder Representations from Transformers)
RoPE (Rotary Positional Encoding)
Grouped Query Attention
RLHF (Reinforcement Learning from Human Feedback)
DPO (Direct Preference Optimization)
Комментарии