filmov
tv
1 Million Tiny Experts in an AI? Fine-Grained MoE Explained
Показать описание
Mixture of Experts explained, well, re-explained. We are in the Fine-Grain era of Mixture of Experts and it's about to get even more interesting as we further scale it up.
This video was sponsored by Brilliant
Check out my newsletter:
Special thanks to LDJ for helping me with this video
Mixtral 8x7B Paper
Sparse MoE (2017)
Adaptive Mixtures of Local Experts (1991)
Gshard
Branch-Train Mix
DeepSeek-MoE
MoWE (from the meme at 7:51)
Mixture of A Million Experts
This video is supported by the kind Patrons & YouTube Members:
🙏Andrew Lescelius, alex j, Chris LeDoux, Alex Maurice, Miguilim, Deagan, FiFaŁ, Robert Zawiasa, Daddy Wen, Tony Jimenez, Panther Modern, Jake Disco, Demilson Quintao, Shuhong Chen, Hongbo Men, happi nyuu nyaa, Carol Lo, Mose Sakashita, Miguel, Bandera, Gennaro Schiano, gunwoo, Ravid Freedman, Mert Seftali, Mrityunjay, Richárd Nagyfi, Timo Steiner, Henrik G Sundt, projectAnthony, Brigham Hall, Kyle Hudson, Kalila, Jef Come, Jvari Williams, Tien Tien, BIll Mangrum, owned, Janne Kytölä, SO, Richárd Nagyfi, Hector, Drexon, Claxvii 177th, Inferencer, Michael Brenner, Akkusativ, Oleg Wock, FantomBloth
[Music] massobeats - daydream
[Video Editor] @Askejm
Комментарии