Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Minor path release 2.5.1 [wip] #20589

Open
wants to merge 13 commits into
base: release/stable
Choose a base branch
from
Open
Next Next commit
chore: remove redundant words in comment (#20510)
Signed-off-by: withbest <seekseat@outlook.com.>
(cherry picked from commit afe5708)
  • Loading branch information
withbest authored and Borda committed Feb 14, 2025
commit 87fd8cb1302b34dde1ef2f5161a560155ced69ff
4 changes: 2 additions & 2 deletions docs/source-pytorch/tuning/profiler_intermediate.rst
Original file line number Diff line number Diff line change
Expand Up @@ -55,7 +55,7 @@ The profiler will generate an output like this:
Self CPU time total: 1.681ms

.. note::
When using the PyTorch Profiler, wall clock time will not not be representative of the true wall clock time.
When using the PyTorch Profiler, wall clock time will not be representative of the true wall clock time.
This is due to forcing profiled operations to be measured synchronously, when many CUDA ops happen asynchronously.
It is recommended to use this Profiler to find bottlenecks/breakdowns, however for end to end wall clock time use
the ``SimpleProfiler``.
Expand Down Expand Up @@ -142,7 +142,7 @@ This profiler will record ``training_step``, ``validation_step``, ``test_step``,
The output above shows the profiling for the action ``training_step``.

.. note::
When using the PyTorch Profiler, wall clock time will not not be representative of the true wall clock time.
When using the PyTorch Profiler, wall clock time will not be representative of the true wall clock time.
This is due to forcing profiled operations to be measured synchronously, when many CUDA ops happen asynchronously.
It is recommended to use this Profiler to find bottlenecks/breakdowns, however for end to end wall clock time use
the ``SimpleProfiler``.
Expand Down
2 changes: 1 addition & 1 deletion src/lightning/fabric/strategies/deepspeed.py
Original file line number Diff line number Diff line change
Expand Up @@ -144,7 +144,7 @@ def __init__(
nvme_path: Filesystem path for NVMe device for optimizer/parameter state offloading.

optimizer_buffer_count: Number of buffers in buffer pool for optimizer state offloading
when ``offload_optimizer_device`` is set to to ``nvme``.
when ``offload_optimizer_device`` is set to ``nvme``.
This should be at least the number of states maintained per parameter by the optimizer.
For example, Adam optimizer has 4 states (parameter, gradient, momentum, and variance).

Expand Down
2 changes: 1 addition & 1 deletion src/lightning/pytorch/core/module.py
Original file line number Diff line number Diff line change
Expand Up @@ -979,7 +979,7 @@ def configure_optimizers(self) -> OptimizerLRScheduler:
# `scheduler.step()`. 1 corresponds to updating the learning
# rate after every epoch/step.
"frequency": 1,
# Metric to to monitor for schedulers like `ReduceLROnPlateau`
# Metric to monitor for schedulers like `ReduceLROnPlateau`
"monitor": "val_loss",
# If set to `True`, will enforce that the value specified 'monitor'
# is available when the scheduler is updated, thus stopping
Expand Down
2 changes: 1 addition & 1 deletion src/lightning/pytorch/strategies/deepspeed.py
Original file line number Diff line number Diff line change
Expand Up @@ -166,7 +166,7 @@ def __init__(
nvme_path: Filesystem path for NVMe device for optimizer/parameter state offloading.

optimizer_buffer_count: Number of buffers in buffer pool for optimizer state offloading
when ``offload_optimizer_device`` is set to to ``nvme``.
when ``offload_optimizer_device`` is set to ``nvme``.
This should be at least the number of states maintained per parameter by the optimizer.
For example, Adam optimizer has 4 states (parameter, gradient, momentum, and variance).

Expand Down