What is LoRA? Low-Rank Adaptation for finetuning LLMs EXPLAINED

preview_player
Показать описание
How does LoRA work? Low-Rank Adaptation for Parameter-Efficient LLM Finetuning explained. Works for any other neural network as well, not just for LLMs.

Thanks to our Patrons who support us in Tier 2, 3, 4: 🙏
Dres. Trost GbR, Siltax, Vignesh Valliappan, Mutual Information, Kshitij

Outline:
00:00 LoRA explained
00:59 Why finetuning LLMs is costly
01:44 How LoRA works
03:45 Low-rank adaptation
06:14 LoRA vs other approaches

▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀
🔥 Optionally, pay us a coffee to help with our Coffee Bean production! ☕
▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀

🔗 Links:

#AICoffeeBreak #MsCoffeeBean #MachineLearning #AI #research​
Music 🎵 : Meadows - Ramzoid

Video editing: Nils Trost
Рекомендации по теме
Комментарии
Автор

Insightful : Especially the comparison from LORA to prefix tuning and adapters at the end!

MikeTon
Автор

Perfect. This exactly what I wanted to know. "Bite-sized" is right!

rockapedra
Автор

I’ve been using LoRAs for a while now but didn’t have a great understanding of how they work. Thank you for the explainer!

wholenutsanddonuts
Автор

By far the clearest explanation on youtube

michelcusteau
Автор

Very clear and straightforward. The explanation of matrix rank was especially helpful. Thank you for the video.

SoulessGinge
Автор

Thanks Letitia. Your explanation was very clear and helpful to understand the paper.

minkijung
Автор

So simple explanation, thank you soo much!!

keshavsingh
Автор

Thank you much for this video. I started reading the paper, was very terrified by it, then I thought I should watch some YouTube video, watch one video, was asleep half-way through the video. Woke up again and stumbled across your video, your coffee woke me up and now I got the LoRA. Thanks for your efforts.

thecodest
Автор

Thanks again for amazing video. I would also request a detailed video on Flash Attention. Thanks

karndeepsingh
Автор

Very comprehensive explanation! Thank you

Lanc
Автор

LoRA is awesome! It also helps with overfitting in protein language models as well. Cool video!

amelieschreiber
Автор

Thank you for clearing my concepts regarding LoRA

m.rr.c.
Автор

Firstly thanks for the amazing video. Can you also make a video about QLoRA.

igrpfkc
Автор

If we knew what abstractions were handled layer by layer, we could make sure that the individual layers were trained to completely learn those abstractions. Let's hope Max Tegmark's work on introspection get us there.

dr.mikeybee
Автор

Thanks for the simple and educating video!
If I'm not mistaken, prefix tuning is pretty much the same as embedding vectors in diffusion models! How cool is that? 😀

alirezafarzaneh
Автор

Fantastic video as always. QLora is even better if you are GPU poor like me.

bdennyw
Автор

You are just amazing >>> so beautiful so elegant just wow😇😇

deepak_kori
Автор

Concept of rank of a matrix, tauught in such an effective way

soulfuljourney
Автор

how is Lora fine-tuning track changes from creating two decomposition matrix? How the ΔW is determined?

ArunkumarMTamil
Автор

Great explanation, thanks for the video!
I have a lingering question about LoRA: Is it necessary to approximate the low-rank matrices of the difference weights (the Delta W in the video). Or can we reduce the size of the original weight matrices? If I understood the video correctly, at the end of LoRA training, I have the full parameters of the roginal model + the difference weights (in reduced size). My question is why can't I learn low rank matrices for the original weights as well?

yacinegaci