diff --git a/pytorch_lightning/callbacks/__init__.py b/pytorch_lightning/callbacks/__init__.py index a9dc6ef7d99fe..24193fe3debf3 100644 --- a/pytorch_lightning/callbacks/__init__.py +++ b/pytorch_lightning/callbacks/__init__.py @@ -1,9 +1,7 @@ from pytorch_lightning.callbacks.base import Callback from pytorch_lightning.callbacks.early_stopping import EarlyStopping from pytorch_lightning.callbacks.gpu_stats_monitor import GPUStatsMonitor -from pytorch_lightning.callbacks.gpu_usage_logger import GpuUsageLogger from pytorch_lightning.callbacks.gradient_accumulation_scheduler import GradientAccumulationScheduler -from pytorch_lightning.callbacks.lr_logger import LearningRateLogger from pytorch_lightning.callbacks.lr_monitor import LearningRateMonitor from pytorch_lightning.callbacks.model_checkpoint import ModelCheckpoint from pytorch_lightning.callbacks.progress import ProgressBar, ProgressBarBase @@ -13,9 +11,7 @@ 'Callback', 'EarlyStopping', 'GPUStatsMonitor', - 'GpuUsageLogger', 'GradientAccumulationScheduler', - 'LearningRateLogger', 'LearningRateMonitor', 'ModelCheckpoint', 'ProgressBar', diff --git a/pytorch_lightning/callbacks/early_stopping.py b/pytorch_lightning/callbacks/early_stopping.py index 1f66caf0e3c91..9566542f2ece4 100644 --- a/pytorch_lightning/callbacks/early_stopping.py +++ b/pytorch_lightning/callbacks/early_stopping.py @@ -203,10 +203,3 @@ def _run_early_stopping_check(self, trainer, pl_module): # stop every ddp process if any world process decides to stop should_stop = trainer.accelerator_backend.early_stopping_should_stop(pl_module) trainer.should_stop = should_stop - - def on_train_end(self, trainer, pl_module): - if self.stopped_epoch > 0 and self.verbose > 0: - # todo: remove this old warning - rank_zero_warn('Displayed epoch numbers by `EarlyStopping` start from "1" until v0.6.x,' - ' but will start from "0" in v0.8.0.', DeprecationWarning) - log.info(f'Epoch {self.stopped_epoch + 1:05d}: early stopping triggered.') diff --git a/pytorch_lightning/callbacks/gpu_usage_logger.py b/pytorch_lightning/callbacks/gpu_usage_logger.py deleted file mode 100644 index 3a0793887f7ab..0000000000000 --- a/pytorch_lightning/callbacks/gpu_usage_logger.py +++ /dev/null @@ -1,9 +0,0 @@ -from pytorch_lightning.callbacks.gpu_stats_monitor import GPUStatsMonitor -from pytorch_lightning.utilities import rank_zero_warn - - -class GpuUsageLogger(GPUStatsMonitor): - def __init__(self, *args, **kwargs): - rank_zero_warn("`GpuUsageLogger is now `GPUStatsMonitor`" - " and it will be removed in v0.11.0", DeprecationWarning) - super().__init__(*args, **kwargs) diff --git a/pytorch_lightning/callbacks/lr_logger.py b/pytorch_lightning/callbacks/lr_logger.py deleted file mode 100644 index 76ade47087743..0000000000000 --- a/pytorch_lightning/callbacks/lr_logger.py +++ /dev/null @@ -1,9 +0,0 @@ -from pytorch_lightning.callbacks.lr_monitor import LearningRateMonitor -from pytorch_lightning.utilities import rank_zero_warn - - -class LearningRateLogger(LearningRateMonitor): - def __init__(self, *args, **kwargs): - rank_zero_warn("`LearningRateLogger` is now `LearningRateMonitor`" - " and this will be removed in v0.11.0", DeprecationWarning) - super().__init__(*args, **kwargs) diff --git a/tests/test_deprecated.py b/tests/test_deprecated.py index 00f142e79e71e..730fa3ea0ba74 100644 --- a/tests/test_deprecated.py +++ b/tests/test_deprecated.py @@ -1,12 +1,8 @@ """Test deprecated functionality which will be removed in vX.Y.Z""" -import random import sys -import pytest import torch -from pytorch_lightning import Trainer -from pytorch_lightning.callbacks import GpuUsageLogger, LearningRateLogger from tests.base import EvalModelTemplate