Building a GPU cluster for AI

preview_player
Показать описание

Learn, from start to finish, how to build a GPU cluster for deep learning. We'll cover the entire process, including cluster level design, rack level design, node level design, CPU and GPU selection, power distribution, storage, and networking.

This talk is based on the Lambda Echelon GPU Cluster whitepaper. The whitepaper can be found above.

Slides for the talk can be found here:

Errata:
- Slide 46 contains an erroneous diagram with a connection from the storage server to the compute fabric network, the storage server does not connect ot the compute fabric network. The correct diagram is available in the whitepaper.
Рекомендации по теме
Комментарии
Автор

Thanks. I’m planning on building a “massive” 2 GPU system for home use.

peterxyz
Автор

Extraordinary presentation. Covered all the important topics in depth and with real teaching talent. Many thanks!!

randahan
Автор

One of the best presentations on GPU cluster design, even at 3 years old. Great teaching skills!

fundyourhustle
Автор

Ground level details with all the critical aspects covered nice for GPU Cluster to the last cable length calculation.

onlooker
Автор

Its nice to see a holistic explanation of designing / building / installing a complex multi-rack system...As someone that has spent years working on both sides of the "analog/digital divide" (physical data center world / digital world's various segments), the un-sexy physical aspects of available rack space / power / cooling / floor loading / network uplink bandwidth are often overlooked (often assumed)...A semi arrives with a pallet: "Hey Carl, you can have this online in a couple days, right?"

carlschumacher
Автор

Thank you for highlighting an underrated topic/options that company should re-consider within their compute infrastructure.

yassinebouchoucha
Автор

Best of the best presentation on server clusters. Author presented deep understanding of server clusters so that he explains things in an easy way. thank you!!!

lovanda
Автор

Thank you. You got me started years ago with your lambda stack -- the only way I could get TensorFlow installed on Linux.

dr.mikeybee
Автор

Lots and lots of A100 GPUs. Every single one of them is a monster, almost 2x faster memory than the next best GPU. An entire room full of A100 racks... holy cow.

ProjectPhysX
Автор

What an amazing presentation - one of the better videos I have watched. Great breadth and depth.

AjaySimha-sy
Автор

Most professional and holistic explanation I heard about this topic.
Thank you so much!!

randahan
Автор

Highly appreciated...Youtube should have a separate category called Founder's video.

cyberspider
Автор

Very expert suggestions for hpc and compute sizing.

NSPK-
Автор

Hey Stephen, this is highly informative. I work on this clustering. Now am able to connect the dots and get the bigger picture.
where can i read about the relationship between numa topology and GPU peering capability.

HarishN.J
Автор

I want to build a multi dual epyc 7742 based system for goofing around learning this stuff.

loadmastergod
Автор

My machine learning team consists of me baby

HankGallows
Автор

Really good analysis and presentation!

ilyboc
Автор

Still most relevant today, 2 years later. Thanks.

julianfiacconi
Автор

I have three computers, and a nas, and a external hub. I think that I don’t need a another server because of the NAS. As far as my architecture goes, is there anything else that you can advise?

glennisholcomb
Автор

Tell me how difficult it is so i can buy your solution kind of talk

rosenangelow