How to Run Any LLM using Cloud GPUs and Ollama with Runpod.io

preview_player
Показать описание

Don't forget to sign up for the newsletter below to give updates in AI, what I'm working on and struggles I've dealt with (which you may have too!):

=========================================================
=========================================================

📖 Chapters:
00:00 Intro
01:11 How to setup
03:26 Install Ollama
05:40 Run Example
06:26 Outro

💬 If you have any issues, let me know in the comments and I will help you out!
Рекомендации по теме
Комментарии
Автор

Can you do a new guide for text gen ui as well please? The bloke doesnt work anymore.

Larimuss
Автор

My question is why would u use runpod and still pay their rate when you can just throw a llama 405b model or whatever model in a aws server and deploy it yourself and only being charged for hosting that aws server which would be probably cheaper and probably is what runpod is doing anyways.

ZodakZach
Автор

Since you can run a python file there in runpod, I’m assuming you can also serve a gradio ui from there? Kinda like in your YouTube service video. I really appreciate all of your hard work on your channel. One of my favorite ag centric channels.

BradDStephensAIFPV
Автор

How using Runpod serverless and pods differ in this use case, considering eg. costs? How can we minimize our costs eg. with stopping running the pod after usage?

attilavass
Автор

Hi what is the difference between this method and using vllm I saw in runpod data centric video which way is better

jarad
Автор

Is it possible to host the server here, or is the run pod just used for fine tuning and training models

lololoololdudusoejdhdjswkk
Автор

Is it possible to use a model on the server and parse it to the local Ollama to use it in any software locally?

MichaelTrader