Skip to content

Commit 3dca121

Browse files
jvmncschsasank
authored andcommitted
pointer to NCCL in docs + add barrier + typo fixes (pytorch#248)
1 parent 6bc51b9 commit 3dca121

File tree

1 file changed

+5
-4
lines changed

1 file changed

+5
-4
lines changed

intermediate_source/dist_tuto.rst

+5-4
Original file line numberDiff line numberDiff line change
@@ -234,6 +234,7 @@ of 6 collectives currently implemented in PyTorch.
234234
from all processes in ``dst``.
235235
- ``dist.all_gather(tensor_list, tensor, group)``: Copies ``tensor``
236236
from all processes to ``tensor_list``, on all processes.
237+
- ``dist.barrier(group)``: block all processes in `group` until each one has entered this function.
237238

238239
Distributed Training
239240
--------------------
@@ -421,9 +422,9 @@ slightly different signature than the ones in PyTorch. It takes a
421422
``recv`` tensor and will store the sum of all ``send`` tensors in it. As
422423
an exercise left to the reader, there is still one difference between
423424
our version and the one in DeepSpeech: their implementation divide the
424-
gradient tensor into *chunks*, so as to optimially utilize the
425+
gradient tensor into *chunks*, so as to optimally utilize the
425426
communication bandwidth. (Hint:
426-
`toch.chunk <http://pytorch.org/docs/master/torch.html#torch.chunk>`__)
427+
`torch.chunk <http://pytorch.org/docs/master/torch.html#torch.chunk>`__)
427428

428429
Advanced Topics
429430
---------------
@@ -446,7 +447,7 @@ there are currently three backends implemented in PyTorch: TCP, MPI, and
446447
Gloo. They each have different specifications and tradeoffs, depending
447448
on the desired use-case. A comparative table of supported functions can
448449
be found
449-
`here <http://pytorch.org/docs/master/distributed.html#module-torch.distributed>`__.
450+
`here <http://pytorch.org/docs/master/distributed.html#module-torch.distributed>`__. Note that a fourth backend, NCCL, has been added since the creation of this tutorial. See `this section <https://pytorch.org/docs/master/distributed.html#multi-gpu-collective-functions>`__ of the ``torch.distributed`` docs for more information about its use and value.
450451

451452
**TCP Backend**
452453

@@ -508,7 +509,7 @@ and we'll have to recompile it by hand. Fortunately, this process is
508509
fairly simple given that upon compilation, PyTorch will look *by itself*
509510
for an available MPI implementation. The following steps install the MPI
510511
backend, by installing PyTorch `from
511-
sources <https://github.com/pytorch/pytorch#from-source>`__.
512+
source <https://github.com/pytorch/pytorch#from-source>`__.
512513

513514
1. Create and activate your Anaconda environment, install all the
514515
pre-requisites following `the

0 commit comments

Comments
 (0)