Fine-tuning Llama 2 on Your Own Dataset | Train an LLM for Your Use Case with QLoRA on a Single GPU

preview_player
Показать описание

Learn how to fine-tune the Llama 2 7B base model on a custom dataset (using a single T4 GPU). We'll use the QLoRa technique to train an LLM for text summarization of conversations between support agents and customers over Twitter.

Join this channel to get access to the perks and support my work:

00:00 - When to Fine-tune an LLM?
00:30 - Fine-tune vs Retrieval Augmented Generation (Custom Knowledge Base)
03:38 - Text Summarization (our example)
04:47 - Dataset Selection
05:36 - Choose a Model (Llama 2)
06:22 - Google Colab Setup
07:26 - Process data
10:08 - Load Llama 2 Model & Tokenizer
11:18 - Training
14:49 - Compare Base Model with Fine-tuned Model
18:08 - Conclusion

#llama2 #llm #promptengineering #chatgpt #chatbot #langchain #gpt4 #summarization
Рекомендации по теме
Комментарии
Автор

This is great. A version for question answering would be helpful too.

christopherbader
Автор

Good stuff coming, thank you in advance ❤

stawils
Автор

Can you provide the Google Collab notebook?

vivekjyotibhowmik
Автор

Excellent video! What changes in the input we need to make to use 8 bit quantization instead of 4 bit. Thanks.

krishchatterjee
Автор

Do you have or plan to make a tutorial for something like bellow?
Tutorial for the plane text fine-tuning and then tuning that model to make it an instruct tuned one?

AbdulBasit-fftq
Автор

Thank you for this! Is finetuning a good approach for a private/proprietary documentation Q&A?

GregMatoga
Автор

Fantastic video! It will be nice to see a full tutorial on how to do it with pdf locally...

fabsync
Автор

Any idea how can we deploy llama-2 on huggingface api? just like the falcon one, has some issue with the handler.

DawnWillTurn
Автор

Incredible video!! Thank you very much, I have a question: isn't it mandatory to put characters like EOS at the end of the summary? for the LLM to finish the instruction?

williamgomezsantana
Автор

will you be able to add a tutorial for llama2-chat model

jensonjoy
Автор

Great!! Do some videos regarding RLHF.

experiment
Автор

Thanks for sharing, really helpful. Waiting for my Llama model access to follow it step by step. Can I use any other model in place of this?

techtraversal
Автор

Great video!

Is there anyway to build my instruction dataset for instruct fine-tuning from classical text books?

ikurious
Автор

can you train the model on german data?

sasukeuchiha-ckhy
Автор

I still don't get it i have my data locally, how should start finetuning it please tell

tarunku
Автор

Do you have an idea how GPT4 is so good with its responses from its base model when I upload documents to it?
Could it be the parameter. size only or do you think other technologies are what determine the quality difference?

shopbc
Автор

can i download the finetuned model after finetuning?
is it in format .bin or .safetensor or else?
cuz im current trying to do finetuning on textgen, but having troubles. with dataset (format) i guess.

GooBello-grls
Автор

Hi there, I am just reading through the repo and Im pretty sure this is the answer...i just wanted to make sure...
The actual input to the model is only from the [text] field, is that correct? As the [text] field contains the prompt, the conversation and the summary...

williamfussell
Автор

Hola, For me the validation log show No log with mistral instruct model. Please help anyone.

xyreqhd
Автор

I need help please. I just want to be pointed in the right direction since I'm new to this and since I couldn't really find any proper guide to summarize the steps for what I want to accomplish.

I want to integrate a LLama 2 70B chatbot into my website. I have no idea where to start. I looked into setting up the environment on one of my cloud servers(Has to be private). Now I'm looking into training/fine-tuneing the chat model using our data from our DBs(It's not clear for me here but I assume it involves two steps, first I have to have the data in a CSV format since it's easier for me, second I will need to format it in Alpaca or Openassistant formats). After that, the result should be a deployment-ready model ?

Just bullet points I'd highly appreciate that.

vitocorleon