Understanding: AI Model Quantization, GGML vs GPTQ!

preview_player
Показать описание
Learning Resources:

❤️ If you want to support the channel ❤️
Support here:
Рекомендации по теме
Комментарии
Автор

This series(back to basics) needs a boost. Love the way you explained all fundamentals. Keep them coming

explorer
Автор

This is such a good video but the number of likes/views don't reflect it. Thank you so much.

rupjitchakraborty
Автор

Always my go to channel to understand concepts clearly. Can't thankyou enough brother. 🙌

vivekraj
Автор

Well done mate. Thank you for your thorough and clear explanation.

vamp
Автор

I was exactly wondering how quantization works this morning. Thank you, such a good video 🎉

megamehdi
Автор

Great and informative video dude! Well done, and I always appreciate your content!

MaJetiGizzle
Автор

Great explanation about the differences about GGPQ and GGML, thanks once again!

echofloripa
Автор

Excellent and most accurate explanation. Thank you!

tarun
Автор

Wow, amazing video, everything was well explained and detailed, thanks!

luisxd
Автор

Wonderful explanation! Keep up the great content!

mokanin
Автор

Good work, great explanation. Thanks!

fredrik-ekelund
Автор

I knew the number of bits had to do with the accuracy and how powerful the hardware was required to run a LLM but beyond that i had no idea what it meant. Your explanation was super clear, so thanks.

MonkeySimius
Автор

Thanks mate for the great explanation!

Semion.
Автор

Thanks this is a nice video. Can GPTQ models run on apple metal framework? Also, I have seen some GGML models use CPU and GPU together. How is this different from the other approach?

harry
Автор

Never knew there was a different but now I know, thank you!

Nerdimo
Автор

Can you explain how to figure out what settings to use to run models in textui. such as transformer, qlora, i usually just end up trying every combination untill it works or i give up. And usualy no instructions on huggingface repo.

cdb
Автор

A GPTQ quantized model inherits from the nn.Module class in pytorch? How can I integrate a GPTQ model with my pytorch code?

aurkom
Автор

have you planed to do some more videos on gptq and ggml, where finetuning the quatized model or converting fp16 models to quantized model

im-notai
Автор

How can we utilize both GPU and CPU for training a model. Like somehow break the model and store half of it in CPU RAM and the other half in GPU RAM

kalilinux
Автор

Great explanation! I needed this...lol

sytekdd