Jay Alammar on LLMs, RAG, and AI Engineering

preview_player
Показать описание
Jay Alammar, renowned AI educator and researcher at Cohere, discusses the latest developments in large language models (LLMs) and their applications in industry. Jay shares his expertise on retrieval augmented generation (RAG), semantic search, and the future of AI architectures.

MLST is sponsored by Brave:

Jay Alamaar:

Buy Jay's new book here!
Hands-On Large Language Models: Language Understanding and Generation

TOC:
00:00:00 Introduction to Jay Alammar and AI Education
00:01:47 Cohere's Approach to RAG and AI Re-ranking
00:07:15 Implementing AI in Enterprise: Challenges and Solutions
00:09:26 Jay's Role at Cohere and the Importance of Learning in Public
00:15:16 The Evolution of AI in Industry: From Deep Learning to LLMs
00:26:12 Expert Advice for Newcomers in Machine Learning
00:32:39 The Power of Semantic Search and Embeddings in AI Systems
00:37:59 Jay Alammar's Journey as an AI Educator and Visualizer
00:43:36 Visual Learning in AI: Making Complex Concepts Accessible
00:47:38 Strategies for Keeping Up with Rapid AI Advancements
00:49:12 The Future of Transformer Models and AI Architectures
00:51:40 Evolution of the Transformer: From 2017 to Present
00:54:19 Preview of Jay's Upcoming Book on Large Language Models

Disclaimer: This is the fourth video from our Cohere partnership. We were not told what to say in the interview, and didn't edit anything out from the interview. Note also that this combines several previously unpublished interviews from Jay into one, the earlier one at Tim's house was shot in Aug 2023, and the more recent one in Toronto in May 2024.

Refs:
The Illustrated Transformer

Attention Is All You Need

The Unreasonable Effectiveness of Recurrent Neural Networks

Neural Networks in 11 Lines of Code

Understanding LSTM Networks (Chris Olah's blog post)

Luis Serrano's YouTube Channel

Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks

GPT (Generative Pre-trained Transformer) models

BERT (Bidirectional Encoder Representations from Transformers)

RoPE (Rotary Positional Encoding)

Grouped Query Attention

RLHF (Reinforcement Learning from Human Feedback)

DPO (Direct Preference Optimization)
Рекомендации по теме
Комментарии
Автор

Great to see a summary like this of how a real practitioner in the field is solving real problems with real companies. We do the same in my role and my teams. Not as sexy as the latest AGI hype or LLM research, but where the bulk of the “journeyman” GenAI work is for the next several years at least! Thanks for sharing Jay’s insights with us.

snarkyboojum
Автор

Roughly around 46:00 regarding layered levels of complexity in a paper. I can't emphasize enough how many times I come back to specific papers and reread them, then walk away knowing more than the first time I read them because they sent me down a path that caused me to learn more, then 3 months later, I'm coming back and finding new gems, 3 months after that, doing the same thing. Every AI researcher should be taking notes.

manslaughterinc.
Автор

aistructuralreview AI fixes this. s, RAG, and AI Engineering.

DeniseMilaTeresa
Автор

Love this guy. Great discussion as usual, unlucky it was so "short"

Darkon
Автор

Jay is a wonderful knowledgable person. Makes things simple . Thanks you.

nasrawi
Автор

Most amazing, Tim, how you seem to be right on top of what I am currently investigating 😄. I just started reading Jay's book on O'Reilly (pre-release) and you come out with this terrific interview, so I can get a more nuanced view of the author. Excited to read his latest, having gotten my earlier understanding of how transformers work from his 2017 paper. Thanks once again, Tim.

toadlguy
Автор

Top quality content as always. My favorite AI channel.

CodexPermutatio
Автор

Excellent video, touches on many different topics, has great practical advice, pointers to education material, etc… Jay is talented and doing this purely out of passion for technology and innovation.

muhannadobeidat
Автор

Can you please focus again on more complex topics that are like u previously? I still appreciate ur content but somehow it gets quite repetitive and deviates from the technical side to more and more soft topics

paulk
Автор

Thank you so much for all these knowledge so well explained 🎉❤

saulyarhi
Автор

Thank you, Tim. This video brought me some illusion regarding the future uses of LLMs; sometimes is easy to think they're just stupid and useless

arowindahouse
Автор

I usually like these conversations because they allow you to meet the person instead of their sales mask. But in this one he never really took his mask off at all. I just listened to an hour of just marketing and I am honestly pretty disappointed.

lexer_
Автор

So basic is the material that while I don't think anything in here is wrong I didn't learn anything either

tacticalgaryvrgamer
Автор

Sobering content. AGI's nowhere to be seen, Cohere building a path for soft landing from the heights of the hypeness, real business cases still missing

LuigiSimoncini