How to Run Llama 3 Locally? 🦙

preview_player
Показать описание
Llama 3 8B is better than Llama 2 70B, and that is crazy!
Here's how to run Llama 3 model (4-bit quantized) on Google Colab - Free tier.
It requires around 6 GBs of GPU memory, which could be easily handled by Colab’s T4 GPU.

#llama3 #metaai #ai
Рекомендации по теме
Комментарии
Автор

Locally literally means opposite to colab

tuna
Автор

This is excellent. Thanks very much. Saved me loads of time :)

clamr
Автор

Gradio is giving me an error:


Error: No API found

Has this code rotted already?

izzlenizzle
Автор

Which is better Gemini 1.5 pro llama3 and does llama 3 have any special feature or is it same as Gemini and chat gpt??

tech_informer