2-D Parallelism using DistributedTensor and PyTorch DistributedTensor

preview_player
Показать описание
PyTorch 2.0 Q&A:
🗓️ March 1
⏰ 11am PT

Introduction to 2-D Parallelism (FSDP + Tensor Parallel) to train large scale ViT models and Introduction to PyTorch DistributedTensor, a fundamental tensor level primitives that expresses tensor distribution & computation across devices/hosts.

Join Wanchao Liang & Junjie Wang
Host DA: Shashank Prasanna

Рекомендации по теме
Комментарии
Автор

could you just edit out the troubleshooting the audio part out of the video?

foxdog
Автор

hi does DTensor also work with databricks + pyspark running a cluster with only CPUs?

Gerald-izmv
join shbcf.ru