Run Your Own LLM Locally: LLaMa, Mistral & More

preview_player
Показать описание
Today I show you how you can run your own LLM instance locally at home with a tool called Ollama.

◾◾◾◾◾◾◾◾◾◾◾◾◾◾◾◾◾
📚 Programming Books & Merch 📚

💼 Services 💼

🌐 Social Media & Contact 🌐
Рекомендации по теме
Комментарии
Автор

I'd really like to see a video how you can train those models on your own documents.

matthias_schumm
Автор

I love how I used Lm studio 1 day before you posted this video 😂

Deltaforce_
Автор

Hi, thanks for the video, it is really great but since you do tutorial for python as well and I’m really interested about it, can you please provide us with a tutorial on how to run the model in python? It would be pretty useful for making own applications or voice interacting.

devox
Автор

would be nice if you could give a demo on how to run custom/fine-tuned open source models on cloud services (or even non fine-tuned models)

mmzzzmeemee
Автор

Awsome! How could i run this in VSC as to extract the response generated?

diogomarques
Автор

Do you have any videos about post-processing LLM outputs and achieving the best accuracy (e.g. we have math tasks with an exact and only one answer)?

mak_kry
Автор

Looks cool. Now all I need is a mega-spec machine :-(

daveys
Автор

I've just switched to nvidia and the model runs so much faster. I definitely can't recommend using this if you don't have a nvidia gpu.

seasong
Автор

Thanks for video, can you please also let us know, how to uninstall any module, it is eating up huge space,

malikrajat
Автор

one month later and there is still no support for windows and also no support for AMD gpu

seasong
Автор

The Flood gates just got breached. 😂 kidding, but this is awesome
i wish you mentioned ollama-webui as well. it might be more helpful for people with home lab/lab setup

HomelessDeamon
Автор

Thanks for the video! Even on my 16 Gb ram it runs but answers slowly but it s ok anyway

ekszmui
Автор

You can run most of the models without a GPU; however, it isn't as fast.

maka
Автор

is there a way to hire a VPS with things like T4 GPU by the query so we can use the high-end hardware on a query-based pay-as-you-go plan?

oliverli
Автор

Do I need to have GPU to run these models?

erodotosdemetriou
Автор

Are you using any gpu? If so what is the configuration?

mdbayazid
Автор

I am a java developer . I domt have idea about these AI stuff. How to train these systems or models when deployed locally . With documents??. From whom.and where to get . Whats the aurhenticity of those docs 🤔🤔

subramanianchenniappan
Автор

Got the Windows version, how to know that it is the 70B version of LLama2?

luminaire
Автор

If you run in Windows via WSL, can you use GPU? Since virtualization doesn't use native devices

dominiks
Автор

How to interact with these models programatically? Like in a Java server application???

WikiPeoples