From 15d67887515a81de39e224b81fce28ac03cef8c8 Mon Sep 17 00:00:00 2001 From: Justus Schock <12886177+justusschock@users.noreply.github.com> Date: Wed, 17 Feb 2021 20:14:59 +0100 Subject: [PATCH] Fix Wrapping optimizers upon assignment (#6006) * Update properties.py * pep8 --- pytorch_lightning/trainer/properties.py | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/pytorch_lightning/trainer/properties.py b/pytorch_lightning/trainer/properties.py index feb0b3f4d8fb3..282b4539df0be 100644 --- a/pytorch_lightning/trainer/properties.py +++ b/pytorch_lightning/trainer/properties.py @@ -373,6 +373,11 @@ def optimizers(self) -> Optional[List[Optimizer]]: @optimizers.setter def optimizers(self, new_optims: Optional[List[Optimizer]]) -> None: + # Necessary to rewrap optimizers to lightning + # They will be re-created when accessing + # the `lightning_optimizers` trainer property + self._lightning_optimizers = None + self.accelerator.optimizers = new_optims @property