Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Transformer building blocks tutorial #3075

Draft
wants to merge 6 commits into
base: main
Choose a base branch
from
Draft
Show file tree
Hide file tree
Changes from 4 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions .jenkins/validate_tutorials_built.py
Original file line number Diff line number Diff line change
Expand Up @@ -28,6 +28,7 @@
"intermediate_source/mnist_train_nas", # used by ax_multiobjective_nas_tutorial.py
"intermediate_source/fx_conv_bn_fuser",
"intermediate_source/_torch_export_nightly_tutorial", # does not work on release
"intermediate_source/transformer_building_blocks.py", # does not work on release
svekars marked this conversation as resolved.
Show resolved Hide resolved
"advanced_source/super_resolution_with_onnxruntime",
"advanced_source/usb_semisup_learn", # fails with CUDA OOM error, should try on a different worker
"prototype_source/fx_graph_mode_ptq_dynamic",
Expand Down
10 changes: 9 additions & 1 deletion index.rst
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ Welcome to PyTorch Tutorials
* `torch.export AOTInductor Tutorial for Python runtime (Beta) <https://pytorch.org/tutorials/recipes/torch_export_aoti_python.html>`__
* `A guide on good usage of non_blocking and pin_memory() in PyTorch <https://pytorch.org/tutorials/intermediate/pinmem_nonblock.html>`__
* `Introduction to Distributed Pipeline Parallelism <https://pytorch.org/tutorials/intermediate/pipelining_tutorial.html>`__
* `Introduction to Libuv TCPStore Backend <https://pytorch.org/tutorials/intermediate/TCPStore_libuv_backend.html>`__
* `Introduction to Libuv TCPStore Backend <https://pytorch.org/tutorials/intermediate/TCPStore_libuv_backend.html>`__
* `Asynchronous Saving with Distributed Checkpoint (DCP) <https://pytorch.org/tutorials/recipes/distributed_async_checkpoint_recipe.html>`__
* `Python Custom Operators <https://pytorch.org/tutorials/advanced/python_custom_ops.html>`__
* Updated `Getting Started with DeviceMesh <https://pytorch.org/tutorials/recipes/distributed_device_mesh.html>`__
Expand Down Expand Up @@ -686,6 +686,14 @@ Welcome to PyTorch Tutorials
:link: beginner/knowledge_distillation_tutorial.html
:tags: Model-Optimization,Image/Video


.. customcarditem::
:header: [Title TBD] Unbundling nn.Transformer modules for gains and profits
:card_description: This tutorial goes over recommended best practices for implementing Transformers.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
:card_description: This tutorial goes over recommended best practices for implementing Transformers.
:card_description: This tutorial goes over recommended best practices for implementing Transformers with native PyTorch.

:image: _static/img/thumbnails/cropped/pytorch-logo.png
:link: intermediate/transformer_building_blocks.html
:tags: Transformer

.. Parallel-and-Distributed-Training


Expand Down
Loading
Loading