QLoRA: Efficient Finetuning of Quantized LLMs Explained

preview_player
Показать описание
Рекомендации по теме
Комментарии
Автор

This is the best explanation of QLoRA on YouTube.

wryltxw
Автор

Thanks for this enlightening video. I guess, the first c-value that you mentioned should be one magnitude smaller (0.062 instead of 0.62).

Aruuuq
Автор

is there intuition behind which modules to LoRA adapt?
idk, i saw all, just qv, ... and if i just apply to qv, it means im diminishing the % drop of training weights when compared to full lora, as only Lora adapted layers get quantized..correct?

iFastee
Автор

Amazing explanation. Is that ok if I show your video in my podcast and show the link?

kiwillm