diff --git a/pytorch_lightning/callbacks/model_checkpoint.py b/pytorch_lightning/callbacks/model_checkpoint.py index 5d1f6cedb52ff..cb0f36db44493 100644 --- a/pytorch_lightning/callbacks/model_checkpoint.py +++ b/pytorch_lightning/callbacks/model_checkpoint.py @@ -659,7 +659,7 @@ def _save_top_k_checkpoint(self, trainer, monitor_candidates: Dict[str, Any]): if self.check_monitor_top_k(trainer, current): self._update_best_and_save(current, epoch, step, trainer, monitor_candidates) elif self.verbose: - rank_zero_info(f"Epoch {epoch:d}, step {step:d}: {self.monitor} was not in top {self.save_top_k}") + rank_zero_info(f"Epoch {epoch:d}, global step {step:d}: {self.monitor} was not in top {self.save_top_k}") def _save_none_monitor_checkpoint(self, trainer, monitor_candidates: Dict[str, Any]): if self.monitor is not None or self.save_top_k == 0: