Skip to content

Latest commit

 

History

History
18 lines (14 loc) · 643 Bytes

distributed-training.md

File metadata and controls

18 lines (14 loc) · 643 Bytes
title order snippet summary-home featured-home
Distributed Training
2
```python import torch.distributed as dist from torch.nn.parallel import DistributedDataParallel dist.init_process_group(backend='gloo') model = DistributedDataParallel(model) ```
Scalable distributed training and performance optimization in research and production is enabled by the torch.distributed backend.
true

Optimize performance in both research and production by taking advantage of native support for asynchronous execution of collective operations and peer-to-peer communication that is accessible from Python and C++.