You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: CHANGELOG.md
+34-1Lines changed: 34 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -9,7 +9,10 @@ The format is based on [Keep a Changelog](http://keepachangelog.com/en/1.0.0/).
9
9
10
10
### Added
11
11
12
-
- Add new `DETAIL` log level to provide useful logs for improving monitoring and debugging of batch jobs
12
+
- Enable gradient accumulation using Horovod's `backward_passes_per_step` ([#11911](https://github.com/PyTorchLightning/pytorch-lightning/pull/11911))
13
+
14
+
15
+
- Add new `DETAIL` log level to provide useful logs for improving monitoring and debugging of batch jobs ([#11008](https://github.com/PyTorchLightning/pytorch-lightning/pull/11008))
13
16
14
17
15
18
- Added a flag `SLURMEnvironment(auto_requeue=True|False)` to control whether Lightning handles the requeuing ([#10601](https://github.com/PyTorchLightning/pytorch-lightning/pull/10601))
@@ -33,6 +36,9 @@ The format is based on [Keep a Changelog](http://keepachangelog.com/en/1.0.0/).
33
36
- Added a function to validate if fault tolerant training is supported. ([#10465](https://github.com/PyTorchLightning/pytorch-lightning/pull/10465))
34
37
35
38
39
+
- Added a private callback to manage the creation and deletion of fault-tolerance checkpoints ([#11862](https://github.com/PyTorchLightning/pytorch-lightning/pull/11862))
40
+
41
+
36
42
- Show a better error message when a custom `DataLoader` implementation is not well implemented and we need to reconstruct it ([#10719](https://github.com/PyTorchLightning/pytorch-lightning/pull/10719))
37
43
38
44
@@ -66,6 +72,8 @@ The format is based on [Keep a Changelog](http://keepachangelog.com/en/1.0.0/).
66
72
- Added a `LOGGER_REGISTRY` instance to register custom loggers to the `LightningCLI` ([#11533](https://github.com/PyTorchLightning/pytorch-lightning/pull/11533))
67
73
68
74
75
+
- Added info message when the `Trainer` arguments `limit_*_batches`, `overfit_batches`, or `val_check_interval` are set to `1` or `1.0` ([#11950](https://github.com/PyTorchLightning/pytorch-lightning/pull/11950))
76
+
69
77
- Added a `PrecisionPlugin.teardown` method ([#10990](https://github.com/PyTorchLightning/pytorch-lightning/pull/10990))
70
78
71
79
@@ -117,9 +125,13 @@ The format is based on [Keep a Changelog](http://keepachangelog.com/en/1.0.0/).
117
125
- Added `Accelerator.is_available` to check device availability ([#11797](https://github.com/PyTorchLightning/pytorch-lightning/pull/11797))
118
126
119
127
128
+
- Enabled static type-checking on the signature of `Trainer` ([#11888](https://github.com/PyTorchLightning/pytorch-lightning/pull/11888))
129
+
130
+
120
131
- Added utility functions for moving optimizers to devices ([#11758](https://github.com/PyTorchLightning/pytorch-lightning/pull/11758))
121
132
122
133
134
+
123
135
### Changed
124
136
125
137
- Implemented a new native and rich format in `_print_results` method of the `EvaluationLoop` ([#11332](https://github.com/PyTorchLightning/pytorch-lightning/pull/11332))
@@ -296,6 +308,9 @@ The format is based on [Keep a Changelog](http://keepachangelog.com/en/1.0.0/).
296
308
297
309
- Changed default logger name to `lightning_logs` for consistency ([#11762](https://github.com/PyTorchLightning/pytorch-lightning/pull/11762))
- Deprecated `training_type_plugin` property in favor of `strategy` in `Trainer` and updated the references ([#11141](https://github.com/PyTorchLightning/pytorch-lightning/pull/11141))
@@ -400,6 +415,12 @@ The format is based on [Keep a Changelog](http://keepachangelog.com/en/1.0.0/).
400
415
- Deprecated `pytorch_lightning.utilities.warnings.LightningDeprecationWarning` in favor of `pytorch_lightning.utilities.rank_zero.LightningDeprecationWarning`
401
416
402
417
418
+
- Deprecated `agg_key_funcs` and `agg_default_func` parameters from `LightningLoggerBase` ([#11871](https://github.com/PyTorchLightning/pytorch-lightning/pull/11871))
- Deprecated `LightningLoggerBase.agg_and_log_metrics` in favor of `LightningLoggerBase.log_metrics` ([#11832](https://github.com/PyTorchLightning/pytorch-lightning/pull/11832))
404
425
405
426
@@ -553,6 +574,12 @@ The format is based on [Keep a Changelog](http://keepachangelog.com/en/1.0.0/).
553
574
- Removed `log_text` and `log_image` from the `LightningLoggerBase` API ([#11857](https://github.com/PyTorchLightning/pytorch-lightning/pull/11857))
554
575
555
576
577
+
- Removed calls to `profile("model_forward")` in favor of profiling `training_step` ([#12032](https://github.com/PyTorchLightning/pytorch-lightning/pull/12032))
578
+
579
+
580
+
- Removed `get_mp_spawn_kwargs` from `DDPSpawnStrategy` and `TPUSpawnStrategy` in favor of configuration in the `_SpawnLauncher` ([#11966](https://github.com/PyTorchLightning/pytorch-lightning/pull/11966))
581
+
582
+
556
583
### Fixed
557
584
558
585
- Fixed an issue where `HorovodStrategy.teardown()` did not complete gracefully if an exception was thrown during callback setup [#11752](https://github.com/PyTorchLightning/pytorch-lightning/pull/11752)
@@ -605,6 +632,9 @@ The format is based on [Keep a Changelog](http://keepachangelog.com/en/1.0.0/).
605
632
- Configure native Deepspeed schedulers with interval='step' ([#11788](https://github.com/PyTorchLightning/pytorch-lightning/pull/11788))
606
633
607
634
635
+
- Update `RichProgressBarTheme` styles after detecting light theme on colab ([#10993](https://github.com/PyTorchLightning/pytorch-lightning/pull/10993))
636
+
637
+
608
638
## [1.5.10] - 2022-02-08
609
639
610
640
### Fixed
@@ -641,6 +671,9 @@ The format is based on [Keep a Changelog](http://keepachangelog.com/en/1.0.0/).
641
671
- Disabled sampler replacement when using `IterableDataset` ([#11507](https://github.com/PyTorchLightning/pytorch-lightning/pull/11507))
642
672
643
673
674
+
- Disable loading dataloades if corresponding `limit_batches=0` ([#11576](https://github.com/PyTorchLightning/pytorch-lightning/pull/11576))
:class:`~pytorch_lightning.profiler.xla.XLAProfiler` will help you debug and optimize training
215
+
workload performance for your models using Cloud TPU performance tools.
216
+
217
+
.. code-block:: python
218
+
219
+
# by passing the `XLAProfiler` alias
220
+
trainer = Trainer(..., profiler="xla")
221
+
222
+
# or by passing an instance
223
+
from pytorch_lightning.profiler import XLAProfiler
224
+
225
+
profiler = XLAProfiler(port=9001)
226
+
trainer = Trainer(..., profiler=profiler)
227
+
228
+
229
+
Manual Capture via TensorBoard
230
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
231
+
232
+
The following instructions are for capturing traces from a running program:
233
+
234
+
0. This `guide <https://cloud.google.com/tpu/docs/pytorch-xla-performance-profiling-tpu-vm#tpu-vm>`_ will
235
+
help you with the Cloud TPU setup with the required installations.
236
+
237
+
1. Start a `TensorBoard <https://www.tensorflow.org/tensorboard>`_ server. You could view the TensorBoard output at ``http://localhost:9001`` on your local machine, and then open the
238
+
``PROFILE`` plugin from the top right dropdown or open ``http://localhost:9001/#profile``
239
+
240
+
.. code-block:: bash
241
+
242
+
tensorboard --logdir ./tensorboard --port 9001
243
+
244
+
2. Once the code you'd like to profile is running, click on the ``CAPTURE PROFILE`` button. Enter
245
+
``localhost:9001`` (default port for XLA Profiler) as the Profile Service URL. Then, enter
246
+
the number of milliseconds for the profiling duration, and click ``CAPTURE``
247
+
248
+
3. Make sure the code is running while you are trying to capture the traces. Also, it would lead to better
249
+
performance insights if the profiling duration is longer than the step time.
250
+
251
+
4. Once the capture is finished, the page will refresh and you can browse through the insights using the
0 commit comments