What is Post Training Quantization - GGUF, AWQ, GPTQ - LLM Concepts ( EP - 4 ) #ai #llm #genai #ml

preview_player
Показать описание


All thanks to -

@MaartenGrootendorst

For his amazing work - In this video, I have used the colab file built by him to explain this concept. Maarten is quite talented, do check out his work!

🎌🎌 Join this channel to get access to perks:

Рекомендации по теме
Комментарии
Автор

well expalined.. Lot many concepts got cleared

sqlsql
Автор

Hey akhil, could you make this course in such a way that after doing this one can at least apply at your company for an internship.


If not then atleast make a roadmap mentioning all the keywords one can search and learn from the internet. Since i am a full stack I don't have much idea of the ai landscape.

mitejmadan
Автор

Hi Akhi, Hoping that you can help, I have an Alienware m18 R2 with an Intel i9-14900HX, NVIDIA RTX 4090 (24GB), 64GB RAM, and 8TB storage. For extra information I don't plan to use this for high intensive tasks like model training or any other such high intensive computing tasks, i only mainly will be using it for analysing my business documents and also writing 20 minute elaborate stories based on a five step story structure. I wanted to use the 70B model to generate the best possible results for these smaller less intensive type tasks. Based on my system specs which which optimisation method would you recommend. GPT Q, GGUF, or AWQ ad would you have any additional advice on the best way to optimise based on my use case requirements?

theuniversityofthemind
Автор

very fast explanation. pls be slow from next time. hard to follow.

rahuldebdas