Code an LLM Tokenizer from Scratch in Python

preview_player
Показать описание
In this lecture, we will build a simple tokenizer from scratch in Python.

0:00 Lecture goals
2:48 2 steps of tokenisation
7:22 Importing the dataset
12:43 Tokenising the text
24:18 Converting tokens into token IDs
31:33 Simple Tokenizer class in Python
43:33 Special Context Tokens
58:33 Additional context tokens
1:02:37 Lecture recap

=================================================

=================================================
Vizuara philosophy:

As we learn AI/ML/DL the material, we will share thoughts on what is actually useful in industry and what has become irrelevant. We will also share a lot of information on which subject contains open areas of research. Interested students can also start their research journey there.

Students who are confused or stuck in their ML journey, maybe courses and offline videos are not inspiring enough. What might inspire you is if you see someone else learning and implementing machine learning from scratch.

No cost. No hidden charges. Pure old school teaching and learning.

=================================================

🌟 Meet Our Team: 🌟

🎓 Dr. Raj Dandekar (MIT PhD, IIT Madras department topper)

🎓 Dr. Rajat Dandekar (Purdue PhD, IIT Madras department gold medalist)

🎓 Dr. Sreedath Panat (MIT PhD, IIT Madras department gold medalist)

🎓 Sahil Pocker (Machine Learning Engineer at Vizuara)

🎓 Abhijeet Singh (Software Developer at Vizuara, GSOC 24, SOB 23)

🎓 Sourav Jana (Software Developer at Vizuara)
Рекомендации по теме
Комментарии
Автор

This series of lectures is highly underrated. It looks under the hood in an easy to understand manner unlike so many other courses out there.

helrod
Автор

This is Lecture 7, just not labeled as such. Thank you for your time putting this together!!!

helrod
Автор

Excited!! Raj please also share white board notes

nitesh
Автор

Thank you for creating this fantastic playlist!!

aseemasthana
Автор

One interesting fact: "The quick brown fox jumps over the lazy dog" contains all 26 letters of English Alphabet

nsipubc
Автор

how are token IDs assigned? is it strictly their alphabetical order? or is there a standard as part of the Python libraries? Thx.

helrod
Автор

Thank You, Sir For This Amazing Lecture :), NGL This One Is Quite Heavy 😅

Omunamantech
Автор

Is Tokenizer Specific to Each model, i.e. does every AI Model has its own Specific Tokenizer or any Tokenizer can work with any AI Model

tripchowdhry