Skip to content

Commit

Permalink
ref: finish decoupling apex, LM and backward
Browse files Browse the repository at this point in the history
  • Loading branch information
williamFalcon committed Oct 10, 2020
1 parent 7adbe90 commit 2f78993
Show file tree
Hide file tree
Showing 2 changed files with 8 additions and 4 deletions.
4 changes: 3 additions & 1 deletion pytorch_lightning/accelerators/accelerator_connector.py
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,9 @@ def on_trainer_init(
self.accelerator = accelerator
if isinstance(accelerator, Accelerator):
self.accelerator.trainer = self
distributed_backend = self.accelerator.nickname
distributed_backend = self.accelerator.nickname
else:
distributed_backend = accelerator

self.trainer.deterministic = deterministic

Expand Down
8 changes: 5 additions & 3 deletions pytorch_lightning/utilities/argparse_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -105,9 +105,11 @@ def get_init_arguments_and_types(cls) -> List[Tuple[str, Tuple, Any]]:
>>> args = get_init_arguments_and_types(Trainer)
>>> import pprint
>>> pprint.pprint(sorted(args)) # doctest: +ELLIPSIS +NORMALIZE_WHITESPACE
[('accumulate_grad_batches',
(<class 'int'>, typing.Dict[int, int], typing.List[list]),
1),
[('accelerator',
(<class 'str'>,
<class 'pytorch_lightning.accelerators.base_accelerator.Accelerator'>,
<class 'NoneType'>),
None),
...
('callbacks',
(typing.List[pytorch_lightning.callbacks.base.Callback],
Expand Down

0 comments on commit 2f78993

Please sign in to comment.