Fine-Tune Language Models with LoRA! OobaBooga Walkthrough and Explanation.

preview_player
Показать описание
In this video, we dive into the world of LoRA (Low-Rank Approximation) to fine-tune large language models. We'll explore how LoRA works, its significance in reducing memory usage, and how to implement it using oobabooga's text generation web UI. Whether you're a beginner or a pro, this step-by-step tutorial will help you harness the power of LoRA to improve your language model's performance. Don't miss out on our explanation of the underlying linear algebra concepts, as well as a detailed breakdown of the hyperparameters involved in LoRA training. Join us in our quest for efficient language model fine-tuning!

#LoRA #LanguageModel #FineTuning #NLP #AI #machinelearning

0:00 Intro
0:30 What are LoRA's
4:48 How to use LoRA's in OobaBooga

Рекомендации по теме
Комментарии
Автор

That is exactly what I was waiting for!! Thank you so much for the insight!

swannschilling
Автор

Just keep these coming. Tons of helpful information.

kaymcneely
Автор

Thank you for this. I am sick of seeing click bait or low quality content when Im trying to self educate about llms and get some good knowledge .

MichaelSalaA
Автор

Didn't know I wanted Kurt Cobain to teach me ML but I'm here for it!

victarion
Автор

Hey thanks for sharing your time and knowledge with us!

nartrab
Автор

I am very hopeful for local OpenSource AI this year in particular. Having LoRA's fine tuned to certain characters would be incredible rather than just using the old prompt method.

infini_ryu
Автор

This was AWSOME!!!! Thank you, Thank you, Thank you, this was a total mystery to me you managed to clear it up completely. I truly thank you for this video, have a blessed year!

timothymaggenti
Автор

I'd love to see a video on building these data sets. I am currently working on an AI safety and morality project, and this is super helpful.

manslaughterinc.
Автор

Nice. Will check it out over the weekend.

ArjunKrishnaUserProfile
Автор

Thanks for also explaining math "basics" in a nutshell. Unfortunately I've forgotten about everything in the regards what I learned some 20 years ago when I studied because I never had to use it. Who would have thought that one day I actually coule make use of it? But it's as one of my teachers once said: Used it or lose it.

testales
Автор

Really helpful and informative. Thank you! Could you talk about using these for query, key, and value matrices in attention? Also, explanations of where in particular it can be used in models that aren’t transformers would be interesting 😊

amelieschreiber
Автор

I think this video was very informative! It just boils down to math which is kind of cool

blondiedawn
Автор

I love your content, very tangible, but still ambitious with theory, thanks a lot!

cz
Автор

Great video! Do expand on the parameters section. We are all out of ram.

wpyx
Автор

Could you make a video going through an example from start to finish? For example, expanding on a base model by creating a Lora with specific knowledge that it can only know with the Lora trained (e.g. the contents of a certain book that you feed it) and have it answer questions about the knowledge that you added. A lot of people are searching for this information, especially now that models have come on the market that can be used for commercial use (e.g. MPT-7B). Easy 10.000+ subscribers if you were to make such a video where you for example take MPT-7B or other commercially usable model and show how to make the data set and instructions and feed it, I don't know, scripts on every Simpsons episode and then ask the model to make a list of all the characters featured in episode "Marge Vs The Monorail".

netwar
Автор

Just found your channel. Excellent Content - Another sub for you sir!

andre-le-bone-aparte
Автор

Can you give examples of training data and results?
that would help a lot

cryptoape
Автор

Great work, you are doing a great job at helping the community with your videos. I had one question or confusion. At the start in the architecture diagram, the talk was about fine tuning using Lora on historical medical appeals however during the hands on part the example was to generate question and answer on some other book dataset. I wasn’t able to relate to it.

karrtikiyer
Автор

This was great. I would love to see you do this in python code at a simple level to set these parameters instead of the GUI

logan
Автор

Just subscribed. Would love a video on training and fine-tuning to clone characters or yourself.

trackerprince