Добавить
Уведомления

PYTORCH DISTRIBUTED | YANLI ZHAO

The PyTorch Distributed Stack is a set of PyTorch features that facilitate training models in distributed systems. PyTorch has introduced multiple new features in the distributed package to help support larger scale and higher efficiency for both data and model parallel. In this talk, Yanli Zhao (Software Engineer, Meta AI) shares tips on how to reduce memory footprint, fit larger models and achieve significant speedup with distributed systems with features like Zero Redundancy Optimizer, DistributedDataParallel, FullyShardedDataParallel, CUDA RDMA and ShardedTensor.

Иконка канала Python обучение
11 подписчиков
12+
16 просмотров
2 года назад
12+
16 просмотров
2 года назад

The PyTorch Distributed Stack is a set of PyTorch features that facilitate training models in distributed systems. PyTorch has introduced multiple new features in the distributed package to help support larger scale and higher efficiency for both data and model parallel. In this talk, Yanli Zhao (Software Engineer, Meta AI) shares tips on how to reduce memory footprint, fit larger models and achieve significant speedup with distributed systems with features like Zero Redundancy Optimizer, DistributedDataParallel, FullyShardedDataParallel, CUDA RDMA and ShardedTensor.

, чтобы оставлять комментарии