How to Self-Host an LLM | Fly GPUs + Ollama

preview_player
Показать описание


Video written and produced by Annie Sexton.
Рекомендации по теме
Комментарии
Автор

These videos are really cool. I'm not a beginner, far from it, but it is soooo nice to get this information in such a distilled manner, and from a person that clearly knows what they are talking about. So natural!

MrManafon
Автор

I see alot of explainer videos and yours are the best!, just grate content delivery and tone, prefection at all!

יהוידעשטינמץ
Автор

I don't understand, how is this self hosting, isn't this cloud hosting?

hassenalhandy
Автор

Can we use any llama based model ? In the destination xan we use the llm we have downloaded ? Imean the custom llm based on llama ?

sidiocity
Автор

Can you provide wireguard instructions you mentioned? Btw perfect tutorial :)

miro
Автор

That looks like a nice way to run an LLM for my personal use, but I’d like to also try out one of the bigger LLM models.
Is that doable at all?
Or will I need to stick to models that fit within the 40gb GPU memory of the a100 for instance?

thedavymac
Автор

Hey, I tried setting this up but I have this error:

2024-08-24T00:27:36.386 runner[***] ord [info] Machine started in 3.517s
2024-08-24T00:27:37.133 app[***] ord [info] INFO Main child exited normally with code: 1
2024-08-24T00:27:37.152 app[***] ord [info] INFO Starting clean up.
2024-08-24T00:27:37.266 app[***] ord [info] INFO Umounting /dev/vdc from /root/.ollama
2024-08-24T00:27:37.268 app[***] ord [info] WARN could not unmount /rootfs: EINVAL: Invalid argument
2024-08-24T00:27:37.269 app[***] ord [info] [ 3.718685] reboot: Power down

any ideas on what would cause this?>

TheloniousBird