Fine-Tuning T5 on Question Answer dataset using Huggingface Transformer & PyTorch

preview_player
Показать описание
📝 In this video, we explore Text-to-Text Transfer Transformers, or T5 for short. T5 takes NLP tasks and converts them into a text-to-text format, making it incredibly versatile. Whether it's text classification, language translation, or text summarization, T5 can handle it all without modifying its architecture.

⚠️Note: Due to the model's size, we recommend using Google Colab or a GPU with 10GB+ VRAM.

Here's what we cover in the video:

* Importing libraries for machine learning, creating a dataset pipeline, and using transformers for a pre-trained model and tokenizer.
* Loading the CoQA dataset into memory and preparing the data for training.
* Setting up the T5 model, optimizer, and defining parameters.
* Creating a data preprocessing pipeline and splitting the dataset for training and validation.
*Training the model, monitoring training and validation loss, and saving the model and tokenizer.
* Evaluating the model using the BLEU score, a metric for generated sentences compared to reference sentences.
* Demonstrating the model's performance on a sample from the dataset.

We're excited to continue exploring transformer models, fine-tuning, creating new models, and embarking on creative projects in this series. Drop your ideas for upcoming videos in the comments, and stay tuned for more exciting content.

Facing issue? Discuss with me on social media

Thank you ❤️ for watching and joining us on this Transformers journey!

Credits:

Chapters:
0:00 Overview
2:04 Pre-requisites
2:12 Dataset
2:21 Dataset Overview
2:45 Coding
8:14 Wrapping up
Рекомендации по теме
Комментарии
Автор

Thanks this video is so helpful. Just having a question, what part are we fine-tuning? are we fine-tuning all layers?

yvonnewu
Автор

Do i have to provide context (input) for asking at the inference stage?

superfreiheit
Автор

Thanks for the video. Can we download the fine tuned tokenizer and model from google colab for later use. If yes how?

Ananya-jg
Автор

Great video, i'm kind of new to pytorch. I've already trained and saved the model, how do i load it again for inferencing?

tranhuy
Автор

Can i fine tune the model to generate sentence from a set of keywords

arf_atelier
Автор

Hey its a very nice video,

can you guide if i want to create a project of Natural Language to SQL query conversion? Using hugging face and NLTK how can i build this, I am not getting how can I make this..

sauravfarkade
Автор

Can you make a video on document question answering?

dhruvtiwari
Автор

Hi there,

Thank you! I need your help with translating from Urdu to English using T5 small. Could you please guide me? I'm willing to tip $50 for your assistance. I'm quite new to this. I also have a dataset ready.

Thanks!

MuhammadZubair-cucx
Автор

can we fine tune this model to generate interview questions from the job description (as context) or is their any other model that can do such thing?

umaisimran
Автор

with another dataset do like this or i need to cusstome the dataset

haulu