NVIDIA 5000 Series for AI: Is it worth it vs 4000 or 3000? Entire Lineup Analysis!

preview_player
Показать описание
Is the 5090 the BEST GPU EVER for local AI self-hosting? What are some of the marketing benchmarks used by NVIDIA and how do they relate to running something like Ollama and llama3.3 or Qwen LLMs? The announcement created a lot of FOMO so here is some analysis comparing the entire 5000 series GPUs against the entire 4000 and 3000 GPU series, focusing on their impact on local Ai performance with actionable insights. Learn which GPU best suits your needs for inferencing, training, or generating AI content! Every GPU is covered 5090, 5080, 5070s, 4090, 4080, 4070, 4060, 3090, 3080, 3070, 3060 from my anticipated price impact on used GPUs also.

GPU Price Watch:
3090
4090

Be sure to 👍✅Subscribe✅👍 for more content like this!

Please share this video to help spread the word and drop a comment below with your thoughts or questions. Thanks for watching!

Digital Spaceport Website

*****
As an Amazon Associate I earn from qualifying purchases.

When you click on links to various merchants on this site and make a purchase, this can result in this site earning a commission. Affiliate programs and affiliations include, but are not limited to, the eBay Partner Network.
*****
Рекомендации по теме
Комментарии
Автор

the DIGITS supercomputer is what I will get later this year, everything else is too expensive 128GB of VRAM for $3k is a no brainer

TheDjnatronic
Автор

I'd stand in line with you in Houston. I think the cap on the price will be the system builders though. I didn't worry about buying a 4090 when the price spiked, I just bought a $3000 PC with a 4090. I did this throughout the GPU shortage era. Because Nvidia will feed MSI, Asus, Alienware, and even iBuyPower; your best bet will be to buy a PC with a 5090 instead of feeding the scalpers if you can't get one near MSRP.

ruzaroos
Автор

Been waiting with excitement for your thoughts on the value of the new cards for AI!

Thank You! For taking the time to/ energy to post em!


😬

sondrax
Автор

I just wanted to say Kudos for the effort and enthusiasm in your videos. You have been tremendously helpful in fighuring out the pros and cons of current hardware. I am sure this channel will be huge in the future.

fawal.
Автор

I hope we will get a 24GB 5080 Ti later this year!

CsendesMark
Автор

I am just feeling more and more that I made the right choice buying (at MSRP) and then keeping my 4090. I think it will be my six year card. It's so good for everything.

IndieAuthorX
Автор

Agreed that the 5090 will be Unobtainium. I picked up four 3090's pretty cheap 650~ ish but had to repaste and replace a few fans on them. Totally worth it. I'm running TabbyAPI so I can do speculative decoding and it cranks out around 15 t/s using llama 3.3 70b 5.0bpw with a 1b 3.1 model in front I get around 30 t/s for smaller conversational prompts. I haven't done any fine tuning yet and haven't run long conversations chains but very happy with the quality of the responses and the performance so far. I'm still using open webui. I'd love to get your opinion on Lebrechat. It has been kind of a PITA to get up and running and I don't plan on using any outside services so I don't know how much value it offers over open webui.

WesBrownStories
Автор

The 50 series looks interesting for AI on paper, but I don't think they're even worth looking at for the average person until supply chain catches up to eventually saturate the market (probably right before the 60 series drops). Personally, I'm more interested in investing in a pair of RTX A8000s. Should still be very performant (especially compared to my pair of Tesla P40s currently in use), and 96GB of VRAM to boot instead of "just" 72 with a pair of 5090s. Good luck securing your slice of unobtanium, I'll be curious to see actual performance metrics comparing it to the 3090 and 4090.

BrokenGlytch
Автор

A 5090 @ $2000 is crazy and when the gamer gets it home and wonders why his PC sudleny turns off or blue screens he's going to realize his 800W PS wasn't enough...

JoeVSvolcano
Автор

26:15 you can actually replace the RAM chips on a 2080ti and double the RAM. I don't know if it's cost-effective, but it's definitely a thing.

michaelgleason
Автор

I want a 5060Ti Super with 24Gb of VRAM. Between chip claimshelling and 3Gb GDDR7 memory modules, it can be done! Could help with prices on the used 3090 and 4090 cards too. Lots of us just need the memory on a modern RTX card. Half the speed is fine if it means a third to a quarter of the price.

tungstentaco
Автор

I love how partners could have added extra ram on 5080 but the big brother Nvidia would sue them for going against their specs.

Elyygee
Автор

I want a rackmount Project DIGITS Pro with 512G.

drdiesel
Автор

The nvlink is so good on the the 3090s for training/fine tuning! It's 300GB/s which is more than double PCIe x16 5.0 (128 GB/s). The hidden benefit is you can use fewer PCI lanes which opens motherboard/cpu options quite a bit. Shame they are getting rid of that in 50 series, I didn't realize that from watching the CES coverage...

differentmoves
Автор

15:50 Do people just have no memory? They said a $500 3070 would be like a 2080ti, and it was. I don't know why people are so wuick to dimiss this. Im very intrigued by the 5070ti for selfhosting AI.

michaelgleason
Автор

Super new to wanting to play around with local AI. Can you use a 3090 and 3090ti together?

HoldingHeavy
Автор

I hope they drill some ventilation holes in that Digits box.

aflury
Автор

I wonder whats going to be the difference in AI Tops - Nvidia showing 2x-3x increase over the 40 series - how will that translate into Ollama workflow?

ActveSpn
Автор

Do you guys just buy expensive GPUs like this for testing or do you actually make money with LLMs ( by building SaaS or something like that) ?

MihaiMangu
Автор

Is the $200 ring still worth building, or should I wait for 5060/70?

Meltyhead