Skip to content
This repository has been archived by the owner on Sep 28, 2022. It is now read-only.

Update local pytorch-lightning master #1

Merged
merged 2,310 commits into from
Mar 30, 2021
This pull request is big! We’re only showing the most recent 250 commits.

Commits on Feb 17, 2021

  1. Configuration menu
    Copy the full SHA
    ad36c7b View commit details
    Browse the repository at this point in the history
  2. Prevent flickering progress bar (#6009)

    * add padding
    
    * fix
    
    * fix
    
    * Update pytorch_lightning/callbacks/progress.py
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    
    * updated based on suggestion
    
    * changelog
    
    * add test
    
    * fix pep8
    
    * resolve test
    
    * fix code format
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Co-authored-by: tchaton <thomas@grid.ai>
    3 people committed Feb 17, 2021
    Configuration menu
    Copy the full SHA
    68fd308 View commit details
    Browse the repository at this point in the history
  3. Fix Wrapping optimizers upon assignment (#6006)

    * Update properties.py
    
    * pep8
    justusschock committed Feb 17, 2021
    Configuration menu
    Copy the full SHA
    15d6788 View commit details
    Browse the repository at this point in the history
  4. [Bugfix] Apply untoggle_optimizer when result is None (#5983)

    * update changelog
    
    * apply untoggle_optimizer when result is None
    
    * update tests
    
    * still return loss sometimes
    
    * Update CHANGELOG.md
    
    Co-authored-by: deng-cy <dcy1996@gmail.com>
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    3 people committed Feb 17, 2021
    Configuration menu
    Copy the full SHA
    a121fd3 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    6a409c7 View commit details
    Browse the repository at this point in the history
  6. DeepSpeed Integration (#5954)

    * Add initial deepspeed changes
    
    * Address code review
    
    * Move static method outside of function
    
    * Fixes
    
    * Add missing annotation
    
    * Remove seed setting
    
    * Doc changes
    
    * Doc changes, add address reviews
    
    * Fix docs
    
    * Try fixing issue by moving to torch adam
    
    * Clean up check
    
    * Changes, better APIs!
    
    * Add wrapper, swap to git install revision
    
    * Add special test
    
    * Add warning
    
    * Address review
    
    * Add better disclaimer
    
    * Turn off ZeRO for testing due to compilation
    
    * Add description on modifying parameters via the plugin
    
    * Doc strings clear
    
    * Small doc fixes
    
    * Fix hash, reduce test
    
    * Added CI change
    
    * Move to azure pipeline
    
    * Fix test name
    
    * Add missing flag
    
    * Remove sudo...
    
    * Try conda instead
    
    * Swap to conda base
    
    * Try suggested install
    
    * Apply suggestions from code review
    
    * Apply suggestions from code review
    
    * Revert "Apply suggestions from code review"
    
    This reverts commit 41cca05
    
    * Revert "Apply suggestions from code review"
    
    This reverts commit e06ec29
    
    * Remove setter
    
    * Address most review
    
    * Move out function, remove DeepSpeed from requirements
    
    * Install deepspeed/mpi4py within container
    
    * Use special tests, move to master commit for deepspeed
    
    * Export path
    
    * Force compile to happen first
    
    * Remove!
    
    * Debugging ninja
    
    * Fix error in optimizer step logic
    
    * Attempt to fix symbolic link
    
    * Reverse to aid debugging
    
    * Export path again
    
    * Clean up mess
    
    * var
    
    * Revert "var"
    
    This reverts commit 3450eac
    
    * Address review, add todo
    
    * Add note about unsupported functionality
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: tchaton <thomas@grid.ai>
    Co-authored-by: Jirka Borovec <jirka.borovec@seznam.cz>
    4 people committed Feb 17, 2021
    Configuration menu
    Copy the full SHA
    7189d67 View commit details
    Browse the repository at this point in the history
  7. Trainer only references accelerator (#6039)

    * Trainer only references accelerator where it can
    
    * Move teardown to the trainer, as it is reponsible for the accelerator
    SeanNaren committed Feb 17, 2021
    Configuration menu
    Copy the full SHA
    b7c2e0a View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    8d7ac8f View commit details
    Browse the repository at this point in the history
  9. [feat] Add Trainer(stochastic_weight_avg=True/False) (#6038)

    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: Kaushik B <45285388+kaushikb11@users.noreply.github.com>
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    4 people committed Feb 17, 2021
    Configuration menu
    Copy the full SHA
    c9622ba View commit details
    Browse the repository at this point in the history
  10. [CI] Move DeepSpeed into CUDA image, remove DeepSpeed install from az…

    …ure (#6043)
    
    * Move to CUDA image
    
    * Remove deepspeed install as deepspeed now in the cuda image
    
    * Remove path setting, as ninja should be in the container now
    SeanNaren committed Feb 17, 2021
    Configuration menu
    Copy the full SHA
    8440595 View commit details
    Browse the repository at this point in the history
  11. drop deprecated result object 1/n (#5005)

    * ro1
    
    * ro2
    Borda committed Feb 17, 2021
    Configuration menu
    Copy the full SHA
    bac617f View commit details
    Browse the repository at this point in the history

Commits on Feb 18, 2021

  1. Add option for weight tying on TPU's (#5441)

    * added on_post_move_to_device
    
    * added tests
    
    * docs and refactors
    
    * Update tests/backends/test_tpu_backend.py
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    
    * Update docs/source/tpu.rst
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    
    * Update docs/source/tpu.rst
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    
    * Update pytorch_lightning/core/decorators.py
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    
    * Update pytorch_lightning/core/decorators.py
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    
    * Update docs/source/tpu.rst
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    * Update pytorch_lightning/core/decorators.py
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    * Update pytorch_lightning/core/decorators.py
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    * Update pytorch_lightning/core/decorators.py
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    * Update pytorch_lightning/core/decorators.py
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    * Update pytorch_lightning/core/hooks.py
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    * moved weight sharing module back to test
    
    updated tpu available
    
    * add count to warning
    
    * fix doctest
    
    * import trainer in doctest
    
    * import trainer in doctest
    
    * do not test code as no TPU device
    
    * param count to layer count
    
    * formatting
    
    * update docs
    
    * update import
    
    * update
    
    * resolve tests
    
    * remove legacy accelerator
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    Co-authored-by: tchaton <thomas@grid.ai>
    Co-authored-by: Your Name <you@example.com>
    5 people committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    d2cd7cb View commit details
    Browse the repository at this point in the history
  2. Delete tests.helpers.TrialMNISTDataModule (#5999)

    * Remove TrialMNISTDataModule
    
    * Allow using TrialMNIST in the MNISTDataModule
    
    * Update tests/helpers/datasets.py
    carmocca committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    bfcfac4 View commit details
    Browse the repository at this point in the history
  3. Fix: Allow hashing of metrics with lists in their state (#5939)

    * Fix: Allow hashing of metrics with lists in their state
    
    * Add test case and modify semantics of Metric __hash__ in order to be compatible with structural equality checks
    
    * Fix pep8 style issue
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: mergify[bot] <37929162+mergify[bot]@users.noreply.github.com>
    3 people committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    77f6aa4 View commit details
    Browse the repository at this point in the history
  4. et al. (#6050)

    * et al.
    
    * Apply suggestions from code review
    
    * Apply suggestions from code review
    
    Co-authored-by: chaton <thomas@grid.ai>
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: chaton <thomas@grid.ai>
    3 people committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    6de8dca View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    38ad9e0 View commit details
    Browse the repository at this point in the history
  6. fix/test quant (#6040)

    * fix/test quant
    
    * ...
    
    * ---
    Borda committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    049006a View commit details
    Browse the repository at this point in the history
  7. Add descriptions to accelerator broadcast function/clean up all_gather (

    #6044)
    
    * Add descriptions to accelerator broadcast function/clean up all_gather
    
    * Remove todo
    SeanNaren committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    b019c25 View commit details
    Browse the repository at this point in the history
  8. Add before_batch_transfer and after_batch_transfer hooks (#3671)

    * add hooks
    
    * comment
    
    * docs
    
    * add tests
    
    * make it private
    
    * fix tests
    
    * docs
    
    * chlog
    
    * testcode
    
    * codefactor
    
    * fix doctest
    
    * fix doctest
    
    * suggestions
    
    * is always overriden
    
    * pep and BoringModel
    
    * BoringModel
    
    * docs
    
    * docs
    
    * docs
    
    * fix
    
    * rebase
    
    * rebase
    
    * suggestions
    
    * docs
    
    * suggestions
    
    * try fix docs
    
    * docs
    
    * update name
    
    * yapf
    
    * docs
    
    * rebase
    
    * yapf
    rohitgr7 committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    bcc0004 View commit details
    Browse the repository at this point in the history
  9. Make parallel devices optional across all plugins (#6051)

    * Make parallel devices optional across all plugins so that they can be instantiated
    
    * Add any to types to capture vars passed in
    SeanNaren committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    ffdcb62 View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    115e58a View commit details
    Browse the repository at this point in the history
  11. Fix docs typo (#6055)

    Put .test() in  code blocks
    ieshreya committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    f48a933 View commit details
    Browse the repository at this point in the history
  12. Docs for Pruning, Quantization, and SWA (#6041)

    Co-authored-by: chaton <thomas@grid.ai>
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Co-authored-by: Sean Naren <sean.narenthiran@gmail.com>
    5 people committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    3449e2d View commit details
    Browse the repository at this point in the history
  13. Replace .get_model() with explicit .lightning_module (#6035)

    * rename get_model -> lightning_module
    
    * update references to get_model
    
    * pep8
    
    * add proper deprecation
    
    * remove outdated _get_reference_model
    
    * fix cyclic import
    awaelchli committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    02ac4b0 View commit details
    Browse the repository at this point in the history
  14. rename accelerator_backend -> accelerator (#6034)

    * rename accelerator backend
    
    * rename new additions from master
    
    * add proper deprecation
    
    * pep8
    
    * warning match
    
    * add missing warning type
    awaelchli committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    6cc1a06 View commit details
    Browse the repository at this point in the history
  15. fix flake8 for new plugins (#5951)

    * flake8
    
    * fix cyclic import
    
    * isort
    awaelchli committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    fc9bb53 View commit details
    Browse the repository at this point in the history
  16. fix docs links (#6057)

    Borda committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    d3a31bc View commit details
    Browse the repository at this point in the history
  17. Add warnings to on_before/after_batch_transfer hooks (#6059)

    * Add warnings to hooks
    
    * Add default idx to prevent signature change in the future
    
    * Nothing to see here
    
    * Add default val to transfer_batch_to_device hook
    
    * Apply suggestions from code review
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    
    * Revert "Add default val to transfer_batch_to_device hook"
    
    This reverts commit 5c6a68f
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    SeanNaren and Borda committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    2cf39dc View commit details
    Browse the repository at this point in the history
  18. v1.2.0rc2 (#6063)

    * v1.2.0rc2
    
    * chlogs
    
    * chlogs
    
    * format
    
    * Apply suggestions from code review
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    Borda and rohitgr7 committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    c46c23a View commit details
    Browse the repository at this point in the history
  19. Update auto-opt docs (#6037)

    * fix docs
    
    * update on comments
    
    * Apply suggestions from code review
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    
    * Apply suggestions from code review
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    
    * Apply suggestions from code review
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    
    * rm comment
    
    * Update docs/source/common/lightning_module.rst
    
    Co-authored-by: chaton <thomas@grid.ai>
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Co-authored-by: chaton <thomas@grid.ai>
    5 people committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    b0074a4 View commit details
    Browse the repository at this point in the history
  20. Raise AttributeError in lightning_getattr and lightning_setattr when …

    …attribute not found (#6024)
    
    * Empty commit
    
    * Raise AttributeError instead of ValueError
    
    * Make functions private
    
    * Update tests
    
    * Add match string
    
    * Apply suggestions from code review
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * lightning to Lightning
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    akihironitta and awaelchli committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    8f82823 View commit details
    Browse the repository at this point in the history
  21. default sched (#6062)

    rohitgr7 committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    5d6a091 View commit details
    Browse the repository at this point in the history
  22. v1.2.0 (#6065)

    * v1.2.0
    
    * docs
    Borda committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    4574023 View commit details
    Browse the repository at this point in the history
  23. add Azure tags trigger (#6066)

    * add Azure tags trigger
    
    * fix
    
    * mnodes
    Borda committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    e12c8a7 View commit details
    Browse the repository at this point in the history
  24. pypi azure badges - tags (#6068)

    * pypi azure badges - tags
    
    * pep8
    
    * id
    Borda committed Feb 18, 2021
    Configuration menu
    Copy the full SHA
    3645eb1 View commit details
    Browse the repository at this point in the history

Commits on Feb 19, 2021

  1. continue towards 1.3 (#6069)

    Borda committed Feb 19, 2021
    Configuration menu
    Copy the full SHA
    0b27147 View commit details
    Browse the repository at this point in the history
  2. Fix amp autocast (#6080)

    * precision fixes
    
    * add amp test model
    
    * fix test
    
    * revert
    
    * move assert to training step
    
    * fix test
    
    * fix test
    
    * remove unrelated changes
    
    * add changelog
    
    * remove unused import
    awaelchli committed Feb 19, 2021
    Configuration menu
    Copy the full SHA
    4b7c0fa View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    f2660ac View commit details
    Browse the repository at this point in the history

Commits on Feb 20, 2021

  1. consistent behavior for reduce method across all Plugins (#6011)

    * reduction docs
    
    * docs for abstract base method
    
    * make mean the default
    
    * add preliminary chlog
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    awaelchli and Borda committed Feb 20, 2021
    Configuration menu
    Copy the full SHA
    3bdc067 View commit details
    Browse the repository at this point in the history
  2. [Hot Fix] Give priority to plugins to set distributed mode, and then …

    …accelerator (#6089)
    
    * Give priority to plugins to set distributed mode, and then accelerator
    
    * Add CHANGELOG.md
    
    * Update CHANGELOG.md
    
    * Remove very scary line
    
    * Ensure we set cluster environment after slurm configured if necessary
    
    * Simplify the fix with a reset
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    SeanNaren and carmocca committed Feb 20, 2021
    Configuration menu
    Copy the full SHA
    97a81c3 View commit details
    Browse the repository at this point in the history

Commits on Feb 21, 2021

  1. Enable ZeRO tests for CI, fix to/half function calls (#6070)

    * Enable ZeRO optimization, and make sure that the lightning module hook is called when we move to half precision
    
    * Added test, update to function
    SeanNaren committed Feb 21, 2021
    Configuration menu
    Copy the full SHA
    3b0e4e0 View commit details
    Browse the repository at this point in the history
  2. Expose DeepSpeed FP16 parameters due to loss instability (#6115)

    * Expose deepspeed config parameters to init function due to instability in parameters
    
    * See if tests can run on normal CI, without special tests
    
    * Add changelog
    
    * Update pytorch_lightning/plugins/training_type/deepspeed.py
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    SeanNaren and carmocca committed Feb 21, 2021
    Configuration menu
    Copy the full SHA
    432e563 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    97b4b3e View commit details
    Browse the repository at this point in the history

Commits on Feb 22, 2021

  1. fix amp/apex misconfiguration error for cpu (#6107)

    * fix weird test
    
    * fix apex plugin test
    
    * fix raise
    
    * cpu test
    
    * fix type
    
    * add changelog
    awaelchli committed Feb 22, 2021
    Configuration menu
    Copy the full SHA
    ae6ce17 View commit details
    Browse the repository at this point in the history
  2. Update Contributing Guide (#6118)

    * Update Contributing Guide
    
    * update docs
    kaushikb11 committed Feb 22, 2021
    Configuration menu
    Copy the full SHA
    9b99328 View commit details
    Browse the repository at this point in the history
  3. Minor fixes/improvements in Metric docs (#6114)

    * Fix wrong render
    
    * Improve classification metrics docs
    
    * Improve other domain metrics docs
    
    * Change the structure level in the docs
    akihironitta committed Feb 22, 2021
    Configuration menu
    Copy the full SHA
    1d28d11 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    57215b7 View commit details
    Browse the repository at this point in the history
  5. Feature/5275 clean progress bar print (#5470)

    * Trainer.test should return only test metrics (#5214)
    
    * resolve bug
    
    * merge tests
    
    * Fix metric state reset (#5273)
    
    * Fix metric state reset
    
    * Fix test
    
    * Improve formatting
    
    Co-authored-by: Ananya Harsh Jha <ananya@pytorchlightning.ai>
    
    * print() method added to ProgressBar
    
    * printing alongside progress bar added to LightningModule.print()
    
    * LightningModule.print() method documentation updated
    
    * ProgressBarBase.print() stub added
    
    * stub
    
    * add progress bar tests
    
    * fix isort
    
    * Progress Callback fixes
    
    * test_metric.py duplicate DummyList removed
    
    * PEP and isort fixes
    
    * CHANGELOG updated
    
    * test_progress_bar_print win linesep fix
    
    * test_progress_bar.py remove whitespaces
    
    * Update CHANGELOG.md
    
    Co-authored-by: chaton <thomas@grid.ai>
    Co-authored-by: Tadej Svetina <tadej.svetina@gmail.com>
    Co-authored-by: Ananya Harsh Jha <ananya@pytorchlightning.ai>
    Co-authored-by: Alexander Snorkin <Alexander.Snorkin@acronis.com>
    Co-authored-by: rohitgr7 <rohitgr1998@gmail.com>
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    8 people committed Feb 22, 2021
    Configuration menu
    Copy the full SHA
    423ecf9 View commit details
    Browse the repository at this point in the history
  6. mini refactor for _running_stage access (#5724)

    * running stage
    
    * circular import
    
    * running stage cleanup
    
    * fix unused import
    
    * fix running stage access
    
    * add return type
    
    * Revert "add return type"
    
    This reverts commit 65b0fe2.
    
    * try fix typing
    awaelchli committed Feb 22, 2021
    Configuration menu
    Copy the full SHA
    0456b45 View commit details
    Browse the repository at this point in the history
  7. Add specifics around DeepSpeed docs (#6142)

    * Be more specific with DeepSpeed compatibility
    
    * Better wording
    SeanNaren committed Feb 22, 2021
    Configuration menu
    Copy the full SHA
    863a70c View commit details
    Browse the repository at this point in the history

Commits on Feb 23, 2021

  1. Ensure accelerator is valid if running interactively (#5970)

    Co-authored-by: chaton <thomas@grid.ai>
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    Co-authored-by: Carlos Mocholi <carlossmocholi@gmail.com>
    4 people committed Feb 23, 2021
    Configuration menu
    Copy the full SHA
    ebabe56 View commit details
    Browse the repository at this point in the history
  2. fixing miss-leading tested acc values (#5876)

    * fixing tested values
    
    * .
    
    * tests
    
    * yapf
    
    * softmax
    
    * hvd
    
    * rename
    
    * lr
    
    * duplicate
    
    * drop
    
    * classif
    
    * rm EvalModel
    
    * Revert "rm EvalModel"
    
    This reverts commit 6c3fb39.
    
    * update tests
    
    * fix
    
    * azure
    
    * azure
    
    * self
    
    * cpu
    
    * Apply suggestions from code review
    
    Co-authored-by: rohitgr7 <rohitgr1998@gmail.com>
    Borda and rohitgr7 committed Feb 23, 2021
    Configuration menu
    Copy the full SHA
    1c851b8 View commit details
    Browse the repository at this point in the history
  3. Update CHANGELOG (#6156)

    carmocca committed Feb 23, 2021
    Configuration menu
    Copy the full SHA
    45158aa View commit details
    Browse the repository at this point in the history

Commits on Feb 24, 2021

  1. prune deprecated profiler as bool (#6164)

    * prune profiler
    
    * chlog
    Borda committed Feb 24, 2021
    Configuration menu
    Copy the full SHA
    09baf29 View commit details
    Browse the repository at this point in the history
  2. prune deprecated Trainer arg enable_pl_optimizer (#6163)

    * prune enable_pl_optimizer
    
    * prune automatic_optimization
    Borda committed Feb 24, 2021
    Configuration menu
    Copy the full SHA
    1d9c553 View commit details
    Browse the repository at this point in the history
  3. Prune deprecated metrics for 1.3 (#6161)

    * prune deprecated metrics for 1.3
    
    * isort / yapf
    Borda committed Feb 24, 2021
    Configuration menu
    Copy the full SHA
    a731269 View commit details
    Browse the repository at this point in the history
  4. [Bugfix] Fixed epoch level schedulers not being called when val_check…

    …_interval < 1.0 (#6075)
    
    * fix bug
    
    * fix tests
    
    * changelog
    
    * fix pep8
    
    * fix tests
    
    * fix and add some tests
    
    * add test for rlop
    
    * chlog
    
    * Update CHANGELOG.md
    
    Co-authored-by: rohitgr7 <rohitgr1998@gmail.com>
    SkafteNicki and rohitgr7 committed Feb 24, 2021
    Configuration menu
    Copy the full SHA
    1b498d1 View commit details
    Browse the repository at this point in the history
  5. Prune deprecated checkpoint arguments (#6162)

    * prune prefix
    
    * prune mode=auto
    
    * chlog
    Borda committed Feb 24, 2021
    Configuration menu
    Copy the full SHA
    46617d9 View commit details
    Browse the repository at this point in the history
  6. Prune deprecated EarlyStopping(mode='auto') (#6167)

    Co-authored-by: Roger Shieh <sh.rog@protonmail.ch>
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    3 people committed Feb 24, 2021
    Configuration menu
    Copy the full SHA
    8b47527 View commit details
    Browse the repository at this point in the history
  7. Fix typo (#6178)

    akihironitta committed Feb 24, 2021
    Configuration menu
    Copy the full SHA
    5cf892b View commit details
    Browse the repository at this point in the history
  8. Update issue template to use discussions for questions (#6155)

    * add issue config
    
    * remove question template
    
    * update URL
    
    * Update README.md
    
    * Update README.md
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    * Update .github/ISSUE_TEMPLATE/config.yml
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    edenlightning and rohitgr7 committed Feb 24, 2021
    Configuration menu
    Copy the full SHA
    c33fd52 View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    c7130b7 View commit details
    Browse the repository at this point in the history
  10. Update gpu warning (#6181)

    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: Kaushik Bokka <kaushikbokka@gmail.com>
    3 people committed Feb 24, 2021
    Configuration menu
    Copy the full SHA
    b0d1996 View commit details
    Browse the repository at this point in the history

Commits on Feb 25, 2021

  1. Configuration menu
    Copy the full SHA
    3ed8ef8 View commit details
    Browse the repository at this point in the history
  2. Fix for multiple callbacks (#6197)

    * Fix for multiple callbacks
    
    * Add CHANGELOG.md
    
    * Remove old params
    
    * Skip tests on windows using ddp
    
    * Change name of the variable to not clash with should stop, which is separate
    
    * Apply suggestions from code review
    
    * Fix params
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    SeanNaren and Borda committed Feb 25, 2021
    Configuration menu
    Copy the full SHA
    dd2f5a0 View commit details
    Browse the repository at this point in the history
  3. Add checkpoint parameter to on_save_checkpoint (#6072)

    Co-authored-by: Kaushik B <45285388+kaushikb11@users.noreply.github.com>
    carmocca and kaushikb11 committed Feb 25, 2021
    Configuration menu
    Copy the full SHA
    3df02b8 View commit details
    Browse the repository at this point in the history
  4. Document exceptions in loggers (#6171)

    * Document exceptions in loggers
    
    * minor formatting
    
    * docstring changed in comet.py
    
    * Apply suggestions from code review
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    AlKun25 and rohitgr7 committed Feb 25, 2021
    Configuration menu
    Copy the full SHA
    4d96f19 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    ddf55a2 View commit details
    Browse the repository at this point in the history

Commits on Feb 26, 2021

  1. Configuration menu
    Copy the full SHA
    e7298b5 View commit details
    Browse the repository at this point in the history
  2. Add mypy typing to precision plugins. (#6149)

    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: Jirka Borovec <jirka.borovec@seznam.cz>
    Co-authored-by: Akihiro Nitta <nitta@akihironitta.com>
    4 people committed Feb 26, 2021
    Configuration menu
    Copy the full SHA
    0647340 View commit details
    Browse the repository at this point in the history
  3. apply_func.py: from torchtext.legacy.data import Batch (#6211)

    * Update apply_func.py
    
    The name Batch is no longer located under torchtext.data
    --Error message--
    File "/home/daniel/py38/lib/python3.8/site-packages/pytorch_lightning/utilities/apply_func.py", line 25, in <module>                                                      
        from torchtext.data import Batch                                                  
    ImportError: cannot import name 'Batch' from 'torchtext.data' (/home/daniel/py38/lib/p
    ython3.8/site-packages/torchtext/data/__init__.py)
    You can fix this by changing line line 28 to:
        from torchtext.legacy.data import Batch
    
    * Update apply_func.py
    
    * Update apply_func.py
    
    * Update apply_func.py
    
    * Update apply_func.py
    
    * Update apply_func.py
    dbonner committed Feb 26, 2021
    Configuration menu
    Copy the full SHA
    ee5032a View commit details
    Browse the repository at this point in the history

Commits on Feb 27, 2021

  1. fix(wandb): prevent WandbLogger from dropping values (#5931)

    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Co-authored-by: chaton <thomas@grid.ai>
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    4 people committed Feb 27, 2021
    Configuration menu
    Copy the full SHA
    40d5a9d View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    111d9c7 View commit details
    Browse the repository at this point in the history

Commits on Feb 28, 2021

  1. document exceptions for metrics/regression (#6202)

    Co-authored-by: Akihiro Nitta <nitta@akihironitta.com>
    Co-authored-by: Prajakta Phadke <pphadke@iu.edu>
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    4 people committed Feb 28, 2021
    Configuration menu
    Copy the full SHA
    15c477e View commit details
    Browse the repository at this point in the history

Commits on Mar 1, 2021

  1. simplify skip-if tests >> 0/n (#5920)

    * skipif + yapf + isort
    
    * tests
    
    * docs
    
    * pp
    Borda committed Mar 1, 2021
    Configuration menu
    Copy the full SHA
    58a6d59 View commit details
    Browse the repository at this point in the history
  2. update (#6237)

    awaelchli committed Mar 1, 2021
    Configuration menu
    Copy the full SHA
    ce05687 View commit details
    Browse the repository at this point in the history
  3. Document Exceptions in profilers (#6229)

    * docstring changes in profilers
    
    * minor changes in profilers.py
    AlKun25 committed Mar 1, 2021
    Configuration menu
    Copy the full SHA
    8aba885 View commit details
    Browse the repository at this point in the history
  4. Call optimizer.zero_grad() before backward inside closure in AutoOpt (

    #6147)
    
    Co-authored-by: Carlos Mocholi <carlossmocholi@gmail.com>
    akihironitta and carmocca committed Mar 1, 2021
    Configuration menu
    Copy the full SHA
    925f082 View commit details
    Browse the repository at this point in the history
  5. Fix for incorrect usage of detach(), cpu(), to() (#6216)

    * Fix for incorrect detach/cpu calls (#6214)
    
    * Fix incorrect use of detach(), to(), and cpu(), #6214
    
    * Fix incorrect use of detach() and cpu(), #6214
    
    * update pr
    
    * add typing
    
    * chlog
    
    * more...
    
    * revert on module
    
    * update on comments
    
    * revert changes on model
    
    Co-authored-by: tchaton <thomas@grid.ai>
    Co-authored-by: Jirka Borovec <jirka.borovec@seznam.cz>
    3 people committed Mar 1, 2021
    Configuration menu
    Copy the full SHA
    651c25f View commit details
    Browse the repository at this point in the history
  6. add skipif warpper (#6258)

    Borda committed Mar 1, 2021
    Configuration menu
    Copy the full SHA
    352e8f0 View commit details
    Browse the repository at this point in the history
  7. cleaning SWA (#6259)

    * rename
    
    * if
    
    * test
    
    * chlog
    Borda committed Mar 1, 2021
    Configuration menu
    Copy the full SHA
    ed67490 View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    412a7d8 View commit details
    Browse the repository at this point in the history
  9. switch agents pool (#6270)

    Borda committed Mar 1, 2021
    Configuration menu
    Copy the full SHA
    6788dba View commit details
    Browse the repository at this point in the history

Commits on Mar 2, 2021

  1. docstring changes in tuner (#6264)

    * docstring changes in tuner
    
    * added full stop
    AlKun25 committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    3371d32 View commit details
    Browse the repository at this point in the history
  2. Disable CPU Offload as default for DeepSpeed (#6262)

    * Change default for CPU offload to false for best throughput/memory efficiency
    
    * Add changelog
    
    * default
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    SeanNaren and Borda committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    efda48f View commit details
    Browse the repository at this point in the history
  3. split profilers (#6261)

    Borda committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    dc8647e View commit details
    Browse the repository at this point in the history
  4. Refactor: skipif for multi - gpus 1/n (#6266)

    * ngpus
    
    * gpu
    
    * isort
    
    * pt
    
    * flake8
    Borda committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    eb81500 View commit details
    Browse the repository at this point in the history
  5. Improved EarlyStopping.patience documentation (#6278)

    * Improved early stopping documentation
    
    * Changed to 120 column format
    
    * doc
    
    * doc
    
    * doc
    
    Co-authored-by: Jirka Borovec <jirka.borovec@seznam.cz>
    turian and Borda committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    22985d2 View commit details
    Browse the repository at this point in the history
  6. Refactor: skipif for Windows 2/n (#6268)

    * win
    
    * isort
    
    * flake8
    Borda committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    0f9134e View commit details
    Browse the repository at this point in the history
  7. fix duplicate console logging bug v2 (#6275)

    Co-authored-by: chaton <thomas@grid.ai>
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    3 people committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    bc577ca View commit details
    Browse the repository at this point in the history
  8. Refactor: skipif for AMPs 3/n (#6293)

    * args
    
    * native
    
    * apex
    
    * isort
    Borda committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    b46d221 View commit details
    Browse the repository at this point in the history
  9. [fix] Ensure we check deepspeed/sharded in multinode DDP (#6297)

    * Ensure we check deepspeed/sharded in multinode
    
    * Add CHANGELOG.md
    
    * Add CHANGELOG.md
    
    * Drop mock, use actual multi-gpu node
    SeanNaren committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    8001987 View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    38274b9 View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    24c3a3f View commit details
    Browse the repository at this point in the history
  12. try to fix imports for parsing (#6256)

    * try to fix imports
    
    * legacy 1.2.1
    Borda committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    7e8f4b9 View commit details
    Browse the repository at this point in the history
  13. Refactor: Runif for TPU and Horovod 5/n (#6301)

    * TPU
    
    * horovod
    
    * extra
    
    * fix
    
    * Apply suggestions from code review
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    
    * doc
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    Borda and SkafteNicki committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    ac58378 View commit details
    Browse the repository at this point in the history
  14. Refactor: runif for spec 6/6 (#6307)

    * special
    
    * rpc
    Borda committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    d1a0315 View commit details
    Browse the repository at this point in the history
  15. Add fairscale & deepspeed to skipif 4/n (#6281)

    * add fairscale & windows to skipif
    
    * add deepspeed to runif
    
    * fairscale
    
    * deepspeed
    
    * flake8
    
    Co-authored-by: Jirka Borovec <jirka.borovec@seznam.cz>
    kaushikb11 and Borda committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    4157b35 View commit details
    Browse the repository at this point in the history
  16. [bugfix] TPU test hangs to barrier on 1 process (#6272)

    * update
    
    * resolve flake8
    
    * update
    
    * update
    
    * update changelog
    
    * update
    
    * resolve flake8
    
    Co-authored-by: Your Name <you@example.com>
    tchaton and Your Name committed Mar 2, 2021
    Configuration menu
    Copy the full SHA
    1aac481 View commit details
    Browse the repository at this point in the history

Commits on Mar 3, 2021

  1. Configuration menu
    Copy the full SHA
    bf6ba83 View commit details
    Browse the repository at this point in the history
  2. Simplify test for AMP plugins (#6311)

    * AMP
    
    * fuse
    
    * yapf
    Borda committed Mar 3, 2021
    Configuration menu
    Copy the full SHA
    dcec4ef View commit details
    Browse the repository at this point in the history
  3. Fix ModelPruning(make_pruning_permanent=True) buffers getting removed…

    … when saved during training (#6073)
    
    Co-authored-by: chaton <thomas@grid.ai>
    carmocca and tchaton committed Mar 3, 2021
    Configuration menu
    Copy the full SHA
    4a8422c View commit details
    Browse the repository at this point in the history
  4. [bugfix] TPU + all_gather + SingleTPU shouldn't call xm.all_gather (#…

    …6296)
    
    * resolve an issue with TPU
    
    * update
    
    * add changelog
    tchaton committed Mar 3, 2021
    Configuration menu
    Copy the full SHA
    484dce1 View commit details
    Browse the repository at this point in the history

Commits on Mar 4, 2021

  1. drop unused variable in API (#6308)

    * drop unused pl model in ckpt
    
    * irelevant
    
    * on_evaluation_batch_start
    
    * evaluation_epoch_end
    
    * attach_datamodule
    Borda committed Mar 4, 2021
    Configuration menu
    Copy the full SHA
    6166f46 View commit details
    Browse the repository at this point in the history
  2. hotfix for PT1.6 and torchtext (#6323)

    * ci: azure reinstall torchtext
    
    * move
    
    * todos
    
    * 0.6.0
    
    * skip examples
    
    * formatter
    
    * skip
    
    * todo
    
    * Apply suggestions from code review
    Borda committed Mar 4, 2021
    Configuration menu
    Copy the full SHA
    e038e74 View commit details
    Browse the repository at this point in the history
  3. [fix] Use training type plugin hook when saving (FSDP 1/n) (#6321)

    * Rely on training type plugin when saving
    
    * Add better typing to training type plugin
    SeanNaren committed Mar 4, 2021
    Configuration menu
    Copy the full SHA
    d01e8fd View commit details
    Browse the repository at this point in the history
  4. leaving lezwon (#6347)

    lezwon committed Mar 4, 2021
    Configuration menu
    Copy the full SHA
    577323c View commit details
    Browse the repository at this point in the history
  5. Add tests/utilities/test_parsing.py (#4460)

    * Create branch tests/4400_parsing
    
    * Rename test file for parsing.py
    
    * Fix lightning_hasattr
    
    * Fix lightning_hasattr
    
    * Fix lightning_setattr
    
    * Add empty lines and remove rubbish spaces
    
    * Raise AttributeError not ValueError
    
    * Use getattr in hasattr
    
    * Remove rubbish spaces
    
    * Fix getattr
    
    * Fix by flake8
    
    * Add tests for str_to_bool_or_str
    
    * Fix by flake8
    
    * Add tests for str_to_bool
    
    * Add tests for is_picklable
    
    * Add tests for clean_namespace
    
    * Fix typo
    
    * Fix lightning_getattr
    
    * Add tests for AttributeDict
    
    * Add tests for flatten_dict
    
    * Fix by flake8
    
    * Apply suggestions from code review
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    
    * Apply isort
    
    * Revert "Apply suggestions from code review"
    
    * Define unpicklable_function outside
    
    * Add comment to test_clean_namespace
    
    * Add tests for parse_class_init_keys
    
    * Add tests for get_init_args and collect_init_args
    
    * Share objects across the tests
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: Ethan Harris <ewah1g13@soton.ac.uk>
    3 people committed Mar 4, 2021
    Configuration menu
    Copy the full SHA
    48a10f1 View commit details
    Browse the repository at this point in the history
  6. Add ignore param to save_hyperparameters (#6056)

    * add ignore param to save_hyperparameters
    
    * add docstring for ignore
    
    * add type for frame object
    
    * Update pytorch_lightning/core/lightning.py
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    
    * Update pytorch_lightning/core/lightning.py
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    
    * fix whitespace
    
    * Update pytorch_lightning/core/lightning.py
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    
    * Parametrize tests
    
    * Update pytorch_lightning/core/lightning.py
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    * Update pytorch_lightning/core/lightning.py
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    * seq
    
    * fix docs
    
    * Update lightning.py
    
    * Update lightning.py
    
    * fix docs errors
    
    * add example keyword
    
    * update docstring
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    Co-authored-by: Carlos Mocholi <carlossmocholi@gmail.com>
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    4 people committed Mar 4, 2021
    Configuration menu
    Copy the full SHA
    59acf57 View commit details
    Browse the repository at this point in the history
  7. Fix when _stable_1d_sort to work when n >= N (#6177)

    * Fix when _stable_1d_sort to work when n >= N
    
    * Apply suggestions
    
    Co-authored-by: Carlos Mocholi <carlossmocholi@gmail.com>
    frankier and carmocca committed Mar 4, 2021
    Configuration menu
    Copy the full SHA
    5d7388d View commit details
    Browse the repository at this point in the history
  8. Update docs on arg train_dataloader in fit (#6076)

    * add to docs
    
    * update docs
    
    * Apply suggestions from code review
    
    * Update pytorch_lightning/core/hooks.py
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    * nested loaders
    
    * Apply suggestions from code review
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * shorten text length
    
    * Update pytorch_lightning/core/hooks.py
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    4 people committed Mar 4, 2021
    Configuration menu
    Copy the full SHA
    4f90455 View commit details
    Browse the repository at this point in the history
  9. missing tests default_root_dir=tmpdir (#6314)

    * default_root_dir=tmpdir
    
    * miss
    Borda committed Mar 4, 2021
    Configuration menu
    Copy the full SHA
    b9cf122 View commit details
    Browse the repository at this point in the history
  10. Document exception for metrics/classification (#6190)

    * document exception for metrics/classification
    
    * minor formatting fixes
    
    * fix trailing whitespaces
    
    * document exception for metrics
    
    * Apply suggestions from code review
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    
    * Apply suggestions from code review
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    
    * Apply suggestions from code review
    
    Co-authored-by: Akihiro Nitta <nitta@akihironitta.com>
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: Akihiro Nitta <nitta@akihironitta.com>
    5 people committed Mar 4, 2021
    Configuration menu
    Copy the full SHA
    8e3524d View commit details
    Browse the repository at this point in the history
  11. [Fix] Call clip gradients if clip val greater than 0 (#6330)

    * Call clip gradients if clip val greater than 0
    
    * format
    
    * Format
    
    * Move to top of file
    SeanNaren committed Mar 4, 2021
    Configuration menu
    Copy the full SHA
    39231ae View commit details
    Browse the repository at this point in the history
  12. Configuration menu
    Copy the full SHA
    7acbd65 View commit details
    Browse the repository at this point in the history
  13. docstring changes in accelerators (#6327)

    * docstring changes in accelerators
    
    * docstrings moved
    
    * whitespaces removed
    
    * PEP8 correction[1]
    AlKun25 committed Mar 4, 2021
    Configuration menu
    Copy the full SHA
    49c579f View commit details
    Browse the repository at this point in the history
  14. [bugfix] Perform reduction for dict in training_step and DP (#6324)

    * fix
    
    * update
    
    * update
    
    * add changelog
    
    * Update CHANGELOG.md
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    
    * Update tests/accelerators/test_dp.py
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    
    * update changelog
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    tchaton and carmocca committed Mar 4, 2021
    Configuration menu
    Copy the full SHA
    248a8e8 View commit details
    Browse the repository at this point in the history

Commits on Mar 5, 2021

  1. introduce default cluster environment for lightning-specific ddp (#5915)

    * handle distributed_sampler_kwargs
    
    * move emptying cache to accelertor
    
    * fix a few tests
    
    * restoring the result from subprocess
    
    * fix queue.get() order for results
    
    * add missing "block_backward_sync" context manager
    
    * add missing "block_backward_sync" context manager
    
    * fix sync_batchnorm
    
    * fix supported gpu-ids for tuple
    
    * fix clip gradients and inf recursion
    
    * accelerator selection: added cluster_environment plugin
    
    * fix torchelastic test
    
    * fix reduce early stopping decision for DDP
    
    * fix tests: callbacks, conversion to lightning optimizer
    
    * fix lightning optimizer does not pickle
    
    * fix setting benchmark and deterministic option
    
    * fix slurm amp test
    
    * fix prepare_data test and determine node_rank
    
    * fix retrieving last path when testing
    
    * remove obsolete plugin argument
    
    * fix test: test_trainer_config
    
    * fix torchscript tests
    
    * fix trainer.model access
    
    * move properties
    
    * fix test_transfer_batch_hook
    
    * fix auto_select_gpus
    
    * fix omegaconf test
    
    * fix test that needs to simulate slurm ddp
    
    * add horovod plugin
    
    * fix test with named arguments
    
    * clean up whitespace
    
    * fix datamodules test
    
    * remove old accelerators
    
    * fix naming
    
    * move old plugins
    
    * move to plugins
    
    * create precision subpackage
    
    * create training_type subpackage
    
    * fix all new import errors
    
    * fix wrong arguments order passed to test
    
    * fix LR finder
    
    * Added sharded training type and amp plugin
    
    * Move clip grad to precision plugin
    
    * Added sharded spawn, select accelerators based on distributed_backend + enable custom fp16 plugin automatically
    
    * Fix import issue, attempting to fix tests
    
    * Fix initial test
    
    * Reflect hook logic from master, should wrap model after move to device
    
    * Optional state consolidation, since master has optimizers not wrapped
    
    * change attribute for instance test
    
    * reset optimizers
    
    optimizers are not used in main process, so state would be wrong.
    
    * legacy
    
    * imports in accel
    
    * legacy2
    
    * trainer imports
    
    * fix import errors after rebase
    
    * move hook to new setup location
    
    * provide unwrapping logic
    
    * fix trainer callback system
    
    * added ddp2 implementation
    
    * fix imports .legacy
    
    * move plugins
    
    * restore legacy
    
    * drop test.py from root
    
    * add tpu accelerator and plugins
    
    * fixes
    
    * fix lightning optimizer merge
    
    * reset bugreportmodel
    
    * unwrapping
    
    * step routing forward
    
    * model access
    
    * unwrap
    
    * opt
    
    * integrate distrib_type
    
    * sync changes
    
    * sync
    
    * fixes
    
    * add forgotten generators
    
    * add missing logic
    
    * update
    
    * import
    
    * missed imports
    
    * import fixes
    
    * isort
    
    * mv f
    
    * changelog
    
    * format
    
    * move helper to parallel plugin
    
    * d
    
    * add world size
    
    * clean up
    
    * duplicate
    
    * activate ddp_sharded and tpu
    
    * set nvidia flags
    
    * remove unused colab var
    
    * use_tpu <-> on_tpu attrs
    
    * make some ddp_cpu and clusterplugin tests pass
    
    * Ref/accelerator connector (#5742)
    
    * final cleanup
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * connector cleanup
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * trainer cleanup
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * accelerator cleanup + missing logic in accelerator connector
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * add missing changes to callbacks
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * reflect accelerator changes to lightning module
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * clean cluster envs
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * cleanup plugins
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * add broadcasting
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * yapf
    
    * remove plugin connector
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * plugins
    
    * manual optimization
    
    * update optimizer routing
    
    * add rank to torchelastic
    
    * fix memory mixed precision
    
    * setstate on trainer for pickling in ddp spawn
    
    * add predict method
    
    * add back commented accelerator code
    
    * adapt test for sync_batch_norm to new plugin
    
    * fix deprecated tests
    
    * fix ddp cpu choice when no num_processes are given
    
    * yapf format
    
    * skip a memory test that cannot pass anymore
    
    * fix pickle error in spawn plugin
    
    * x
    
    * avoid
    
    * x
    
    * fix cyclic import in docs build
    
    * add support for sharded
    
    * update typing
    
    * add sharded and sharded_spawn to distributed types
    
    * make unwrap model default
    
    * refactor LightningShardedDataParallel similar to LightningDistributedDataParallel
    
    * update sharded spawn to reflect changes
    
    * update sharded to reflect changes
    
    * Merge 1.1.5 changes
    
    * fix merge
    
    * fix merge
    
    * yapf isort
    
    * fix merge
    
    * yapf isort
    
    * fix indentation in test
    
    * copy over reinit scheduler implementation from dev1.2
    
    * fix apex tracking calls with dev_debugger
    
    * reduce diff to dev1.2, clean up
    
    * fix trainer config test  when gpus>0 and num_processes >0 and ddp_cpu
    
    * sort plugin tests legacy/new
    
    * fix error handling for amp on cpu
    
    * fix merge
    
    
    fix merge
    
    
    fix merge
    
    * [Feat] Resolve manual_backward (#5837)
    
    * resolve manual_backward
    
    * resolve flake8
    
    * update
    
    * resolve for ddp_spawn
    
    * resolve flake8
    
    * resolve flake8
    
    * resolve flake8
    
    Co-authored-by: Ubuntu <ubuntu@ip-172-31-88-60.ec2.internal>
    
    * fix tests/accelerator tests on cpu
    
    * [BugFix] Resolve manual optimization (#5852)
    
    * resolve manual_optimization
    
    * update
    
    * update
    
    Co-authored-by: Ubuntu <ubuntu@ip-172-31-88-60.ec2.internal>
    
    * Remove copy trainer parameters to happen earlier within the loop and add safe guard to get ref model (#5856)
    
    * resovle a bug
    
    * Accelerator refactor sharded rpc (#5854)
    
    * rpc branch
    
    * merge
    
    * update handling of rpc
    
    * make devices etc. Optional in RPC
    
    * set devices etc. later if necessary
    
    * remove devices from sequential
    
    * make devices optional in rpc
    
    * fix import
    
    * uncomment everything
    
    * fix cluster selection
    
    Co-authored-by: Ubuntu <ubuntu@ip-172-31-88-60.ec2.internal>
    
    * resolve bug
    
    * fix assert in rpc test
    
    * resolve a test
    
    * fix docs compilation
    
    * accelerator refactor - fix for sharded parity test (#5866)
    
    * fix memory issue with ddp_spawn
    
    * x
    
    
    x
    
    
    x
    
    
    x
    
    
    x
    
    
    x
    
    
    x
    
    
    x
    
    
    x
    
    * x
    
    * Remove DDP2 as this does not apply
    
    * Add missing pre optimizer hook to ensure lambda closure is called
    
    * fix apex docstring
    
    * [accelerator][BugFix] Resolve some test for 1 gpu (#5863)
    
    * update
    
    * revert init
    
    * resolve a bug
    
    * update
    
    * resolve flake8
    
    * update
    
    * update
    
    * update
    
    * revert init
    
    * resolve a bug
    
    * update
    
    * resolve flake8
    
    * update
    
    * update
    
    * update
    
    * update
    
    * update
    
    * revert init
    
    * resolve a bug
    
    * update
    
    * resolve flake8
    
    * update
    
    * update
    
    * update
    
    * revert init
    
    * update
    
    * resolve flake8
    
    * update
    
    * update
    
    * update
    
    * update
    
    * update
    
    * all_gather
    
    * update
    
    * make plugins work, add misconfig for RPC
    
    * update
    
    * update
    
    * remove breaking test
    
    * resolve some tests
    
    * resolve flake8
    
    * revert to ddp_spawn
    
    Co-authored-by: root <root@ip-172-31-88-60.ec2.internal>
    Co-authored-by: Ubuntu <ubuntu@ip-172-31-88-60.ec2.internal>
    Co-authored-by: Justus Schock <justus.schock@rwth-aachen.de>
    
    * yapf isort
    
    * resolve flake8
    
    * fix apex doctests
    
    * fix apex doctests 2
    
    * resolve docs
    
    * update drone
    
    * clean env
    
    * update
    
    * update
    
    * update
    
    * update
    
    * merge
    
    * Fix RPC related tests, clean out old API, update for new accelerator API [skip ci] (#5881)
    
    * Fix RPC related tests, clean out old API, update for new accelerator API
    
    * Move tests out of legacy folder, update paths and names
    
    * Update test_remove_1-4.py
    
    * Expose properties for tpu cores/gpus/num_gpus
    
    * Add root GPU property
    
    * Move properties to properties.py
    
    * move tests that were previously in drone
    
    * Fix root GPU property (#5908)
    
    * Move root GPU to property, remove horovod set as this is handled in horovod plugin, ensure we mock correctly to set GPU accelerator
    
    * Add missing tests back
    
    * fix best model path transfer when no checkpoint callback available
    
    * Fix setup hook order [wip] (#5858)
    
    * Call trainer setup hook before accelerator setup
    
    * Add test case
    
    * add new test
    
    * typo
    
    * fix callback order in test
    
    Co-authored-by: tchaton <thomas@grid.ai>
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * rename ddp sequential -> rpc sequential for special test
    
    * revert
    
    * fix stupid merge problem
    
    * abstract the cluster plugins
    
    * default plugin
    
    * integrate default environment
    
    * fix property
    
    * adapt tests
    
    * adjust test
    
    * fix world size access
    
    * base cluster env
    
    * revert rebase errors
    
    * revert rebase errors
    
    * missing import
    
    * revert unrelated change
    
    * remove unused cluster local rank
    
    * remove unrelated changes
    
    * fix unrelated changes
    
    * fix pep8
    
    * remove unused var
    
    * reset permissions
    
    * ypaf
    
    * test default environment
    
    * test torchelastic environment
    
    * world  size as int
    
    * tests for slurm environment
    
    * changelog
    
    * test comments
    
    * remove unintended change
    
    * keep master port fixed after it is generated
    
    * test random master port
    
    * yapf
    
    * add missing default environment
    
    * move helper function
    
    * rename default environment
    
    * rename
    
    * rename
    
    * yapf
    
    * Update pytorch_lightning/plugins/environments/lightning_environment.py
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    
    * Update CHANGELOG.md
    
    Co-authored-by: Justus Schock <12886177+justusschock@users.noreply.github.com>
    
    * spawn -> create
    
    Co-authored-by: justusschock <justus.schock@posteo.de>
    Co-authored-by: SeanNaren <sean@grid.ai>
    Co-authored-by: Justus Schock <12886177+justusschock@users.noreply.github.com>
    Co-authored-by: Jirka Borovec <jirka.borovec@seznam.cz>
    Co-authored-by: Justus Schock <justus.schock@rwth-aachen.de>
    Co-authored-by: chaton <thomas@grid.ai>
    Co-authored-by: Ubuntu <ubuntu@ip-172-31-88-60.ec2.internal>
    Co-authored-by: Sean Naren <sean.narenthiran@gmail.com>
    Co-authored-by: root <root@ip-172-31-88-60.ec2.internal>
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    11 people committed Mar 5, 2021
    Configuration menu
    Copy the full SHA
    ec8d46e View commit details
    Browse the repository at this point in the history
  2. [bugfix] Resolve memory leak for evaluation (#6326)

    * resolve bug
    
    * resolve flake8
    
    * revert name
    tchaton committed Mar 5, 2021
    Configuration menu
    Copy the full SHA
    46540ee View commit details
    Browse the repository at this point in the history
  3. Update changelog for v1.2.2 (#6325)

    * update changelog for v1.2.2
    
    * ckpr 1.2.2
    
    Co-authored-by: Jirka Borovec <jirka.borovec@seznam.cz>
    kaushikb11 and Borda committed Mar 5, 2021
    Configuration menu
    Copy the full SHA
    b6aa350 View commit details
    Browse the repository at this point in the history
  4. CI: fix examples - patch download MNIST (#6357)

    * patch download
    
    * CI
    
    * isort
    
    * extra
    Borda committed Mar 5, 2021
    Configuration menu
    Copy the full SHA
    e848542 View commit details
    Browse the repository at this point in the history
  5. [bug] Fix Pytorch profiler with emit_nvtx (#6260)

    * resolve bug
    
    * update changelog
    
    * Update tests/trainer/test_trainer.py
    
    * Update pytorch_lightning/profiler/profilers.py
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    
    * resolve comments
    
    * resolve flake8
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    3 people committed Mar 5, 2021
    Configuration menu
    Copy the full SHA
    2ec67a4 View commit details
    Browse the repository at this point in the history
  6. fix importing torchtext batch (#6365)

    * copy torchtext batch
    
    * update
    
    * rev
    
    * rev
    Borda committed Mar 5, 2021
    Configuration menu
    Copy the full SHA
    2a3ab67 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    4f391bc View commit details
    Browse the repository at this point in the history

Commits on Mar 6, 2021

  1. Refactor RunningStage usage in advance of implementing Trainer.valida…

    …te() (#4945)
    
    * Update code
    
    Co-authored-by: EliaCereda
    
    * More property updates
    
    * Move properties. Introduce trainer._fitting
    
    * Use trainer.fitting
    
    * Fix reset dataloaders
    
    * Unused code
    
    * RunningStage.SANITY_CHECKING
    
    * Use setters
    
    * Fix bugs
    
    * Fix bugs
    
    * TrainerState.{FITTING,VALIDATING,TESTING,PREDICTING,TUNING}
    
    * Fix bugs
    
    * Fix bugs
    
    * Fix tests
    
    * Update CHANGELOG. Add deprecation warning. Fix tests
    
    * Unused imports
    
    * Optional trainer
    
    * More deprecation. More refactoring
    
    * Correct version
    
    * Use properties
    
    * Address comments
    
    * flake8
    
    * Missed renamings
    
    * Typo
    
    * is -> ==
    
    It is recommended to use  for Enums since they are singletons, however, since the LightningEnum subclasses str, it's not a good idea in case a user sets the state/stage with a str
    
    * Also for tests
    
    * Typo
    
    * Address @tchaton's comments
    
    * PEP8
    
    * Correct property
    
    * Update CHANGELOG
    
    * Apply suggestions from code review
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * Update pytorch_lightning/trainer/trainer.py
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * Remove called sanity check
    
    Co-authored-by: Carlos Mocholi <carlossmocholi@gmail.com>
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    3 people committed Mar 6, 2021
    Configuration menu
    Copy the full SHA
    d0596fa View commit details
    Browse the repository at this point in the history
  2. require: adjust versions (#6363)

    * adjust versions
    
    * release
    
    * manifest
    
    * pep8
    
    * CI
    
    * fix
    
    * build
    Borda committed Mar 6, 2021
    Configuration menu
    Copy the full SHA
    85c8074 View commit details
    Browse the repository at this point in the history
  3. Use f-"""-string in a Trainer comment (#6377)

    * Use f-"""-string
    
    * Add r
    
    * Use Trainer.
    
    * r -> noqa: W605
    carmocca committed Mar 6, 2021
    Configuration menu
    Copy the full SHA
    217470b View commit details
    Browse the repository at this point in the history
  4. Remove no return warning from val/test step (#6139)

    * remove warning
    
    * auto_opt
    
    * chlog
    
    * auto_opt
    
    * no_warning_call
    
    * rm old code
    
    * add warning for predict
    
    * Apply suggestions from code review
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    rohitgr7 and awaelchli committed Mar 6, 2021
    Configuration menu
    Copy the full SHA
    facfda8 View commit details
    Browse the repository at this point in the history
  5. Fix manual optimization in pl_example (#6373)

    * Fix automatic_optimization
    
    * Fix automatic_optimization
    
    * Uncomment fairscale
    akihironitta committed Mar 6, 2021
    Configuration menu
    Copy the full SHA
    34b733b View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    966184a View commit details
    Browse the repository at this point in the history

Commits on Mar 7, 2021

  1. Remove optimizer_idx arg in manual optimization (#6093)

    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Co-authored-by: chaton <thomas@grid.ai>
    3 people committed Mar 7, 2021
    Configuration menu
    Copy the full SHA
    38a5fe7 View commit details
    Browse the repository at this point in the history
  2. [doc] Improve Multiple Val/Test Dataloaders with simultaneous batches…

    … option (#6320)
    
    * improve doc to describe how to combine batches of multiple test and val dataloaders simultaneously
    
    * fix typo
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * use paramref
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    mees and awaelchli committed Mar 7, 2021
    Configuration menu
    Copy the full SHA
    2708c39 View commit details
    Browse the repository at this point in the history
  3. [doc] Fix closure in manual optimization (#6374)

    * Fix manual optimization docs
    
    * Fix typo. Thanks @import-antigravity
    akihironitta committed Mar 7, 2021
    Configuration menu
    Copy the full SHA
    c7f30a2 View commit details
    Browse the repository at this point in the history
  4. Fix ModelCheckpoint(monitor=None, save_last=True) not saving checkpoi…

    …nts (#6136)
    
    Co-authored-by: ananthsub <ananth.subramaniam@gmail.com>
    carmocca and ananthsub committed Mar 7, 2021
    Configuration menu
    Copy the full SHA
    826375e View commit details
    Browse the repository at this point in the history

Commits on Mar 8, 2021

  1. Update TBLogger docs (#6315)

    * Update tensorboard.py
    
    * Update logging.rst
    
    * pep8
    
    * Update logging.rst
    
    * Update logging.rst
    
    * Apply suggestions from code review
    
    * add code sample
    
    * Update logging.rst
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    s-rog and Borda committed Mar 8, 2021
    Configuration menu
    Copy the full SHA
    ff16104 View commit details
    Browse the repository at this point in the history
  2. Fix trainer not resetting lightning_optimizers (#6372)

    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    awaelchli and carmocca committed Mar 8, 2021
    Configuration menu
    Copy the full SHA
    718074b View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    0ec7a23 View commit details
    Browse the repository at this point in the history
  4. Fix AttributeError: 'NoneType' object has no attribute 'finalize' on …

    …TPU (#6221)
    
    * Fix bug
    
    Fix AttributeError: 'NoneType' object has no attribute 'finalize'
    
    * Update CHANGELOG.md
    
    * deleted a period
    
    * Update CHANGELOG.md
    
    Co-authored-by: Akihiro Nitta <nitta@akihironitta.com>
    
    * Update CHANGELOG.md
    
    * Update pytorch_lightning/plugins/training_type/tpu_spawn.py
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    Co-authored-by: Akihiro Nitta <nitta@akihironitta.com>
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    4 people committed Mar 8, 2021
    Configuration menu
    Copy the full SHA
    a6c98c4 View commit details
    Browse the repository at this point in the history
  5. Run CI (#6402)

    carmocca committed Mar 8, 2021
    Configuration menu
    Copy the full SHA
    8dabc30 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    efd272a View commit details
    Browse the repository at this point in the history
  7. fix dp reduction test (#6404)

    * fix
    
    * update
    
    * fix
    
    * move the class outside
    awaelchli committed Mar 8, 2021
    Configuration menu
    Copy the full SHA
    e1f5eac View commit details
    Browse the repository at this point in the history
  8. Add check for verbose attribute of ModelCheckpoint (#6419)

    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    ashleve and awaelchli committed Mar 8, 2021
    Configuration menu
    Copy the full SHA
    9eded7f View commit details
    Browse the repository at this point in the history

Commits on Mar 9, 2021

  1. fixed bug where tuner would not tune lr if also tuning batch_size (#4688

    )
    
    * fixed bug where tuner would not tune lr if also tuning batch_size
    
    * added a '+1' to computing the smoothed loss. This maintains the behavior for the smoothed loss as before the bug fix
    
    * pep8 fix
    
    * add changelog
    
    Co-authored-by: chaton <thomas@grid.ai>
    Co-authored-by: Carlos Mocholi <carlossmocholi@gmail.com>
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    4 people committed Mar 9, 2021
    Configuration menu
    Copy the full SHA
    523c59b View commit details
    Browse the repository at this point in the history
  2. update (#6403)

    awaelchli committed Mar 9, 2021
    Configuration menu
    Copy the full SHA
    75c6486 View commit details
    Browse the repository at this point in the history
  3. fix logger creating directory structure too early in DDP (#6380)

    * fix
    
    * add simple test
    
    * fix imports
    
    * add changelog
    
    * tighter test with on_fit_start hook closer to the dispatch call
    
    * move class inside test f unction
    
    * add a comment
    awaelchli committed Mar 9, 2021
    Configuration menu
    Copy the full SHA
    fc6d402 View commit details
    Browse the repository at this point in the history
  4. Typing for tests 1/n (#6313)

    * typing
    
    * yapf
    
    * typing
    Borda committed Mar 9, 2021
    Configuration menu
    Copy the full SHA
    55dd3a4 View commit details
    Browse the repository at this point in the history
  5. [changelog] Update Changelog on release v1.2.3 (#6444)

    * update changelog
    
    * legacy 1.2.3
    
    Co-authored-by: Jirka Borovec <jirka.borovec@seznam.cz>
    tchaton and Borda committed Mar 9, 2021
    Configuration menu
    Copy the full SHA
    30d649b View commit details
    Browse the repository at this point in the history
  6. Improve DummyLogger (#6398)

    * fix dummy logger
    
    * docs
    
    * update docs
    
    * add changelog
    
    * add none return annotation
    
    * return empty string for name, version
    awaelchli committed Mar 9, 2021
    Configuration menu
    Copy the full SHA
    615b2f7 View commit details
    Browse the repository at this point in the history

Commits on Mar 10, 2021

  1. Raise an exception if check_val_every_n_epoch is not an integer (#6411)

    * raise an exception if check_val_every_n_epoch is not an integer
    
    * remove unused object
    
    * add type hints
    
    * add return type
    
    * update exception message
    
    * update exception message
    kaushikb11 committed Mar 10, 2021
    Configuration menu
    Copy the full SHA
    74d79e7 View commit details
    Browse the repository at this point in the history
  2. Set find unused parameters to True by default to fix breaking compati…

    …bility (#6438)
    
    * Set find unused parameters to True by default to fix breaking models, add suggestion to re-enable
    
    * Add changelog
    SeanNaren committed Mar 10, 2021
    Configuration menu
    Copy the full SHA
    c81b2a8 View commit details
    Browse the repository at this point in the history
  3. [bug] All_gather support tensor on cpu (#6416)

    * add test
    
    * update changelog
    
    * update
    
    * rename function
    tchaton committed Mar 10, 2021
    Configuration menu
    Copy the full SHA
    7d4e74c View commit details
    Browse the repository at this point in the history
  4. [Fix] Ensure we set the default device before initializing deepspeed (#…

    …6460)
    
    * Ensure we set the default device before initializing deepspeed
    
    * Add CHANGELOG.md
    
    * Update pytorch_lightning/plugins/training_type/deepspeed.py
    
    Co-authored-by: Kaushik B <45285388+kaushikb11@users.noreply.github.com>
    
    Co-authored-by: Kaushik B <45285388+kaushikb11@users.noreply.github.com>
    SeanNaren and kaushikb11 committed Mar 10, 2021
    Configuration menu
    Copy the full SHA
    1c013b4 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    d1db604 View commit details
    Browse the repository at this point in the history

Commits on Mar 11, 2021

  1. Add Trainer.validate(…) method to run one validation epoch (#4948)

    Co-authored-by: Carlos Mocholi <carlossmocholi@gmail.com>
    Co-authored-by: chaton <thomas@grid.ai>
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    4 people committed Mar 11, 2021
    Configuration menu
    Copy the full SHA
    f4cc745 View commit details
    Browse the repository at this point in the history
  2. Allow user to disable the automatic formatting of checkpoint file nam…

    …es. (#6277)
    
    * cleaning SWA (#6259)
    
    * rename
    
    * if
    
    * test
    
    * chlog
    
    * Remove opt from manual_backward in docs (#6267)
    
    * switch agents pool (#6270)
    
    * Allow user to disable the automatic formatting of checkpoint file names.
    
    * Added changelog entry.
    
    * Made flake8 happy.
    
    * Applied review suggestion: quotes for special characters in docstring
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    
    * Fixed example in docstring.
    
    * Fixed syntax error in docstring.
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: Akihiro Nitta <nitta@akihironitta.com>
    Co-authored-by: thomas chaton <thomas@grid.ai>
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    5 people committed Mar 11, 2021
    Configuration menu
    Copy the full SHA
    2ecda5d View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    079fe9b View commit details
    Browse the repository at this point in the history
  4. cover subproc coverage (#6477)

    Borda committed Mar 11, 2021
    Configuration menu
    Copy the full SHA
    afe0ede View commit details
    Browse the repository at this point in the history
  5. argparse: Add use_argument_group=True (#6088)

    * argparse: Add inplace option
    
    Replicate in GAN model
    
    * datamodule: Deduplicate logic w/ argparser utilities
    
    * Update pl_examples/domain_templates/generative_adversarial_net.py
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    
    * Apply suggestions from code review
    
    Co-authored-by: Akihiro Nitta <nitta@akihironitta.com>
    
    * Keep docstrings
    
    * Correct name
    
    * Whitespace
    
    * Consistency
    
    * fix weird type stuff
    
    * try alt - use_argument_group
    
    * fix syntax + lint
    
    * fix ci errs
    
    * fix ci
    
    * change examples... still failing w/ "unrecognized arguments: --batch_size"
    
    * address review
    
    * mnist_datamodule: add some docstrings
    
    * argparse: check cls or cls.__init__ for param
    
    didn't capture issue, but meh
    
    * fix lint
    
    * fix no-doc edge case
    
    * address review
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: Akihiro Nitta <nitta@akihironitta.com>
    Co-authored-by: Carlos Mocholi <carlossmocholi@gmail.com>
    4 people committed Mar 11, 2021
    Configuration menu
    Copy the full SHA
    e886d55 View commit details
    Browse the repository at this point in the history
  6. Disable batch transfer in DP mode (#6098)

    * add exceptions and test
    
    * hook
    
    * fix
    
    * clean up
    
    * clean up
    
    * regex
    
    * regex
    
    * docs
    
    * rev
    
    * comment and docs
    
    * chlog
    
    * Apply suggestions from code review
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    
    * Apply suggestions from code review
    
    Co-authored-by: chaton <thomas@grid.ai>
    
    * Monkey-patch device count
    
    * docs
    
    * pep
    
    * api_change
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Co-authored-by: chaton <thomas@grid.ai>
    3 people committed Mar 11, 2021
    Configuration menu
    Copy the full SHA
    c53edce View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    62d4304 View commit details
    Browse the repository at this point in the history
  8. [feat] Support iteration-based checkpointing in model checkpoint call…

    …back (#6146)
    
    * Update model_checkpoint.py
    
    * add tests
    
    * Update model_checkpoint.py
    
    * Update test_model_checkpoint.py
    
    * fix tests
    
    * every_n_batches
    
    * Update test_model_checkpoint.py
    
    * defaults
    
    * rm tests
    
    * Update model_checkpoint.py
    
    * Update test_model_checkpoint.py
    
    * Prune deprecated metrics for 1.3 (#6161)
    
    * prune deprecated metrics for 1.3
    
    * isort / yapf
    
    * Update model_checkpoint.py
    
    * add tests
    
    * defaults
    
    * Update CHANGELOG.md
    
    * pre-commit
    
    * Update model_checkpoint.py
    
    * update defaults
    
    * Update test_remove_1-5.py
    
    * Update model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * fix tests
    
    * Update test_model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * Update test_model_checkpoint.py
    
    * ckpt-callback
    
    * Update test_model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * validation-end
    
    * Update model_checkpoint.py
    
    * Update test_model_checkpoint.py
    
    * Update test_model_checkpoint.py
    
    * Update test_model_checkpoint.py
    
    * Update test_model_checkpoint.py
    
    * clarify-names
    
    - Make names explicit as to which hooks they apply to
    - Use step instead of batch for consistency with global step
    
    * Update model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * Update model_checkpoint.py
    
    * mutual-exclusive
    
    Make every_n_train_steps and every_n_val_epochs mutually exclusive
    
    * fix-default-0
    
    * Update CHANGELOG.md
    
    * formatting
    
    * make-private
    
    make attributes private to the class
    
    * rebase
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    ananthsub and Borda committed Mar 11, 2021
    Configuration menu
    Copy the full SHA
    cea170e View commit details
    Browse the repository at this point in the history

Commits on Mar 12, 2021

  1. update xla version (#6464)

    awaelchli committed Mar 12, 2021
    Configuration menu
    Copy the full SHA
    6596447 View commit details
    Browse the repository at this point in the history
  2. Remove unused mixin attributes (#6487)

    * Remove unused mixing attributes
    
    * Missing import
    carmocca committed Mar 12, 2021
    Configuration menu
    Copy the full SHA
    518c7e4 View commit details
    Browse the repository at this point in the history
  3. [doc] Update the order of zero_grad and backward (#6478)

    * Fix zero_grad in docs
    
    * Fix zero_grad in docs
    akihironitta committed Mar 12, 2021
    Configuration menu
    Copy the full SHA
    680e83a View commit details
    Browse the repository at this point in the history

Commits on Mar 14, 2021

  1. Configuration menu
    Copy the full SHA
    b2bcad1 View commit details
    Browse the repository at this point in the history
  2. Update docs for limit_predict_batches (#6507)

    * add docs and minor updates
    
    * docs
    
    * fraction
    rohitgr7 committed Mar 14, 2021
    Configuration menu
    Copy the full SHA
    dcd9dd8 View commit details
    Browse the repository at this point in the history
  3. [bug] Update broadcast + reduce decision ModelCheckpoint] (#6410)

    * resolve bug
    
    * update
    
    * update changelog
    
    * update PR
    
    * Update pytorch_lightning/trainer/connectors/logger_connector/epoch_result_store.py
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    
    * add todo
    
    * resolve issues
    
    * resolve flake8
    
    * update
    
    * add coverage for reduce
    
    * wip
    
    * restore back to brodbact
    
    * remove test.py
    
    * resolve flake8
    
    * update
    
    * check world size
    
    * resolve test
    
    * update
    
    * use pytorch version when defined
    
    * update on comments
    
    * update on comments
    
    * flake8
    
    * resolve bugs
    
    * Update CHANGELOG.md
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    
    * update
    
    * update
    
    * update
    
    * update
    
    * remove test
    
    * update
    
    * resolve flake8
    
    * update
    
    * update
    
    * update
    
    * proxy
    
    * update
    
    * update
    
    * resolve typo
    
    * prune
    
    * update parallel
    
    * update
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    tchaton and carmocca committed Mar 14, 2021
    Configuration menu
    Copy the full SHA
    0544efd View commit details
    Browse the repository at this point in the history

Commits on Mar 15, 2021

  1. Configuration menu
    Copy the full SHA
    02fa32b View commit details
    Browse the repository at this point in the history
  2. CI: resume testing with py3.8 (#6516)

    * testing on python 3.8
    
    * req
    Borda committed Mar 15, 2021
    Configuration menu
    Copy the full SHA
    156847b View commit details
    Browse the repository at this point in the history
  3. document exceptions for metrics/functional (#6273)

    * document exceptions for metrics/functional
    
    * Apply suggestions from code review
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    * Apply suggestions from code review
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    Co-authored-by: Akihiro Nitta <nitta@akihironitta.com>
    4 people committed Mar 15, 2021
    Configuration menu
    Copy the full SHA
    06756a8 View commit details
    Browse the repository at this point in the history
  4. Mean Average Precision metric for Information Retrieval (1/5) (#5032)

    * init information retrieval metrics
    
    * changed retrieval metrics names, expanded arguments and fixed typo
    
    * added 'Retrieval' prefix to metrics and fixed conflict with already-present 'average_precision' file
    
    * improved code formatting
    
    * pep8 code compatibility
    
    * features/implemented new Mean Average Precision metrics for Information Retrieval + doc
    
    * fixed pep8 compatibility
    
    * removed threshold parameter and fixed typo on types in RetrievalMAP and improved doc
    
    * improved doc, put first class-specific args in RetrievalMetric and transformed RetrievalMetric in abstract class
    
    * implemented tests for functional and class metric. fixed typo when input tensors are empty or when all targets are False
    
    * fixed typos in doc and changed torch.true_divide to torch.div
    
    * fixed typos pep8 compatibility
    
    * fixed types in long division in ir_average_precision and example in mean_average_precision
    
    * RetrievalMetric states are not lists and _metric method accepts predictions and targets for easier extension
    
    * updated CHANGELOG file
    
    * added '# noqa: F401' flag to not used imports
    
    * added double space before '# noqa: F401' flag
    
    * Update CHANGELOG.md
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    
    * change get_mini_groups in get_group_indexes
    
    * added checks on target inputs
    
    * minor refactoring for code cleanness
    
    * split tests over exception raising in separate function && refactored test code into multiple functions
    
    * fixed pep8 compatibility
    
    * implemented suggestions of @SkafteNicki
    
    * fixed imports for isort and added types annontations to functions in test_map.py
    
    * isort on test_map and fixed typing
    
    * isort on retrieval and on __init__.py and utils.py in metrics package
    
    * fixed typo in pytorch_lightning/metrics/__init__.py regarding code style
    
    * fixed yapf compatibility
    
    * fixed yapf compatibility
    
    * fixed typo in doc
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    Co-authored-by: mergify[bot] <37929162+mergify[bot]@users.noreply.github.com>
    4 people committed Mar 15, 2021
    Configuration menu
    Copy the full SHA
    5d73fbb View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    eb3ff41 View commit details
    Browse the repository at this point in the history
  6. deprecate metrics pkg (#6505)

    * deprecate metrics
    
    * examples
    
    * req
    
    * docs
    
    * Apply suggestions from code review
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    
    * pep8
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    3 people committed Mar 15, 2021
    Configuration menu
    Copy the full SHA
    b341b53 View commit details
    Browse the repository at this point in the history
  7. [test] lr_find with bs_scale (#6422)

    * init test: test_lr_find_with_bs_scale
    
    * Update test_lr_finder.py
    
    * remove gpu req
    
    * try boring model
    
    * custom boring model
    
    * pep8
    
    * fix typo
    
    * Update test_lr_finder.py
    
    * typo
    
    * typo
    s-rog committed Mar 15, 2021
    Configuration menu
    Copy the full SHA
    c48fc6a View commit details
    Browse the repository at this point in the history
  8. Update DeepSpeed docs (#6528)

    * Clean up docs and add some explicitness around stages
    
    * Apply suggestions from code review
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    SeanNaren and rohitgr7 committed Mar 15, 2021
    Configuration menu
    Copy the full SHA
    383565d View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    ea36ee3 View commit details
    Browse the repository at this point in the history
  10. Update hook lifecycle (#6538)

    * Update hook lifecycle
    
    * Update docs/source/common/lightning_module.rst
    carmocca committed Mar 15, 2021
    Configuration menu
    Copy the full SHA
    9c59733 View commit details
    Browse the repository at this point in the history
  11. Prune metrics base classes 2/n (#6530)

    * base class
    
    * extensions
    
    * chlog
    
    * _stable_1d_sort
    
    * _check_same_shape
    
    * _input_format_classification_one_hot
    
    * utils
    
    * to_onehot
    
    * select_topk
    
    * to_categorical
    
    * get_num_classes
    
    * reduce
    
    * class_reduce
    
    * tests
    Borda committed Mar 15, 2021
    Configuration menu
    Copy the full SHA
    6453091 View commit details
    Browse the repository at this point in the history
  12. Custom Plugin is_distributed (#6537)

    * return from plugin
    
    * dont return for tpu
    amogkam committed Mar 15, 2021
    Configuration menu
    Copy the full SHA
    6a14146 View commit details
    Browse the repository at this point in the history

Commits on Mar 16, 2021

  1. refactor reading env defaults (#6510)

    * change tests
    
    * fix
    
    * test
    
    * _defaults_from_env_vars
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Borda and carmocca committed Mar 16, 2021
    Configuration menu
    Copy the full SHA
    0f07eaf View commit details
    Browse the repository at this point in the history
  2. Prune metric: helpers and inputs 3/n (#6547)

    * _basic_input_validation
    
    * _check_shape_and_type_consistency
    
    * _check_num_classes_binary
    
    * _check_num_classes_mc
    
    * _check_num_classes_ml
    
    * _check_top_k
    
    * _check_classification_inputs
    
    * _input_format_classification
    
    * _reduce_stat_scores
    
    * DataType
    
    * rest
    
    * flake8
    
    * chlog
    Borda committed Mar 16, 2021
    Configuration menu
    Copy the full SHA
    a312219 View commit details
    Browse the repository at this point in the history
  3. prune warning & deprecation wrapper (#6540)

    * docs
    
    * wrapper
    
    * test
    
    * count
    
    * flake8
    Borda committed Mar 16, 2021
    Configuration menu
    Copy the full SHA
    555a6fe View commit details
    Browse the repository at this point in the history
  4. Add outputs param for on_val/test_epoch_end hooks (#6120)

    * add outputs param for on_val/test_epoch_end hooks
    
    * update changelog
    
    * fix warning message
    
    * add custom call hook
    
    * cache logged metrics
    
    * add args to docstrings
    
    * use warning cache
    
    * add utility method for param in sig check
    
    * Update CHANGELOG.md
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    
    * update docstring
    
    * add test for eval epoch end hook
    
    * add types and replace model ref
    
    * add deprecation test
    
    * fix test fx name
    
    * add model hooks warning
    
    * add old signature model to tests
    
    * add clear warning cache
    
    * sopport args param
    
    * update tests
    
    * add tests for model hooks
    
    * code suggestions
    
    * add signature utils
    
    * fix pep8 issues
    
    * fix pep8 issues
    
    * fix outputs issue
    
    * fix tests
    
    * code fixes
    
    * fix validate test
    
    * test
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    kaushikb11 and Borda committed Mar 16, 2021
    Configuration menu
    Copy the full SHA
    b190403 View commit details
    Browse the repository at this point in the history
  5. [doc] Add Zero Grad set_to_none=True trick (#6548)

    * add trick to doc
    
    * update
    
    * update path
    
    * Update docs/source/benchmarking/performance.rst
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    tchaton and rohitgr7 committed Mar 16, 2021
    Configuration menu
    Copy the full SHA
    00cd918 View commit details
    Browse the repository at this point in the history

Commits on Mar 17, 2021

  1. fix deprecation wrapper & tests (#6553)

    * fix deprecation wrapper & tests
    
    * flake8
    Borda committed Mar 17, 2021
    Configuration menu
    Copy the full SHA
    297e438 View commit details
    Browse the repository at this point in the history
  2. prune metric: accuracy 4/n (#6515)

    * prune accuracy
    
    * chlog
    
    * flake8
    
    * Apply suggestions from code review
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    
    * wrap
    
    * test
    
    * test
    
    * fix
    
    Co-authored-by: Nicki Skafte <skaftenicki@gmail.com>
    Borda and SkafteNicki committed Mar 17, 2021
    Configuration menu
    Copy the full SHA
    2f6ce1a View commit details
    Browse the repository at this point in the history

Commits on Mar 18, 2021

  1. Prune metrics: AUC & AUROC (#6572)

    * class: AUC AUROC
    
    * func: auc auroc
    
    * format
    
    * tests
    Borda committed Mar 18, 2021
    Configuration menu
    Copy the full SHA
    9e35f97 View commit details
    Browse the repository at this point in the history
  2. [doc] Update Dict Train Loader doc. (#6579)

    * update doc
    
    * update example
    tchaton committed Mar 18, 2021
    Configuration menu
    Copy the full SHA
    8853a36 View commit details
    Browse the repository at this point in the history
  3. Prune metrics: precision & recall 6/n (#6573)

    * avg precision
    
    * precision
    * recall
    
    * curve
    
    * tests
    
    * chlog
    
    * isort
    
    * fix
    Borda committed Mar 18, 2021
    Configuration menu
    Copy the full SHA
    38a2119 View commit details
    Browse the repository at this point in the history
  4. Update Changelog for v1.2.4 (#6581)

    * Update changelog for v1.2.4
    
    * lagacy v1.2.4
    
    * prune duplicates from changelog
    
    Co-authored-by: Jirka Borovec <jirka.borovec@seznam.cz>
    kaushikb11 and Borda committed Mar 18, 2021
    Configuration menu
    Copy the full SHA
    b606171 View commit details
    Browse the repository at this point in the history
  5. [Fix] Move init dist connection into the setup function (#6506)

    * Move connection setup into the setup function. Call setup hook after we set up the accelerator
    
    * Added CHANGELOG.md
    
    * fix setup order in callback test
    
    * fix input arguments in test
    
    * Mock distributed function, remove protection to turn into training type hook
    
    * Remove import
    
    * Add missing mock, ensure custom plugin does not create children process
    
    * Skip test on windows
    
    * Update deepspeed to init connection in setup
    
    * Do not initialize distributed module
    
    * Move DeepSpeed tests to special tests since dist communication is being set up
    
    * Special the test to see if this fixes CI
    
    * Delete accelerator connector test to see if its causing build to fail
    
    * Delete deepspeed test
    
    * Revert "Delete accelerator connector test to see if its causing build to fail"
    
    This reverts commit edde60b
    
    * Revert "Delete deepspeed test"
    
    This reverts commit 9d317429
    
    * Reverse hook
    
    * Reverse setup hooks to debug again
    
    * Add todo so i know where i left off
    
    * For single device move in pre_dispatch after setup function
    
    * Add additional model to device hook if any additional parameters have been set
    
    * See if we can enable deepspeed tests
    
    * Revert "See if we can enable deepspeed tests"
    
    This reverts commit b5450de
    
    * See if this hook approach works
    
    * Introduce new granular hooks
    
    * Remove import, fix tpu spawn by moving the function to setup
    
    * Added missing special test
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    SeanNaren and awaelchli committed Mar 18, 2021
    Configuration menu
    Copy the full SHA
    4e9b453 View commit details
    Browse the repository at this point in the history

Commits on Mar 19, 2021

  1. Configuration menu
    Copy the full SHA
    983a888 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    87c03b1 View commit details
    Browse the repository at this point in the history
  3. NGC container PoC (#6187)

    * add NVIDIA flows
    
    * push
    
    * pull
    
    * ...
    
    * extras
    
    * ci prune
    
    * fix
    
    * tag
    
    * .
    
    * list
    Borda committed Mar 19, 2021
    Configuration menu
    Copy the full SHA
    5780796 View commit details
    Browse the repository at this point in the history
  4. Automatically set sync_batchnorm for training_type_plugin (#6536)

    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Co-authored-by: Roger Shieh <sh.rog@protonmail.ch>
    Co-authored-by: Kaushik Bokka <kaushikbokka@gmail.com>
    4 people committed Mar 19, 2021
    Configuration menu
    Copy the full SHA
    3b72bcc View commit details
    Browse the repository at this point in the history
  5. Prune metrics: other classification 7/n (#6584)

    * confusion_matrix
    
    * iou
    
    * f_beta
    
    * hamming_distance
    
    * stat_scores
    
    * tests
    
    * flake8
    
    * chlog
    Borda committed Mar 19, 2021
    Configuration menu
    Copy the full SHA
    3a56a60 View commit details
    Browse the repository at this point in the history

Commits on Mar 20, 2021

  1. fixing examples (#6600)

    * try Azure
    
    * -e
    
    * path
    Borda committed Mar 20, 2021
    Configuration menu
    Copy the full SHA
    cb59039 View commit details
    Browse the repository at this point in the history
  2. Add AMP for validation, prediction and testing (#6565)

    * Add Tests for val and test-steps
    
    * Add native AMP
    
    * pep8 tests
    
    * pep8 plugin
    
    * changelog
    justusschock committed Mar 20, 2021
    Configuration menu
    Copy the full SHA
    634d831 View commit details
    Browse the repository at this point in the history

Commits on Mar 21, 2021

  1. Add trainer.predict config validation (#6543)

    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    kaushikb11 and carmocca committed Mar 21, 2021
    Configuration menu
    Copy the full SHA
    37f22c9 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    42a7b70 View commit details
    Browse the repository at this point in the history
  3. Move profiler tests (#6619)

    carmocca committed Mar 21, 2021
    Configuration menu
    Copy the full SHA
    51c9260 View commit details
    Browse the repository at this point in the history

Commits on Mar 22, 2021

  1. Configuration menu
    Copy the full SHA
    870247f View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    853523e View commit details
    Browse the repository at this point in the history
  3. Allow training type plugin to delay optimizer creation (FSDP 2/n) (#6331

    )
    
    * Allow training_type_plugin to delay optimizer configure
    
    * Add missing references to trainer, add a CPU accelerator based test
    SeanNaren committed Mar 22, 2021
    Configuration menu
    Copy the full SHA
    58c9fa7 View commit details
    Browse the repository at this point in the history
  4. Add teardown method to BaseProfiler. (#6370)

    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Co-authored-by: ananthsub <ananth.subramaniam@gmail.com>
    3 people committed Mar 22, 2021
    Configuration menu
    Copy the full SHA
    e2e1de0 View commit details
    Browse the repository at this point in the history
  5. refactoring setup (#6590)

    * refactoring setup
    
    * .
    
    * docs
    
    * flake8
    Borda committed Mar 22, 2021
    Configuration menu
    Copy the full SHA
    1fae10a View commit details
    Browse the repository at this point in the history
  6. hotfix: mock examples (#6632)

    * mock examples
    
    * drop from GA
    Borda committed Mar 22, 2021
    Configuration menu
    Copy the full SHA
    e62c7c7 View commit details
    Browse the repository at this point in the history
  7. [refactor] Add setup to profilers + _run_stage_setup to trainer 2/5 (#…

    …6633)
    
    * add setup
    
    * update
    
    * updates on comment
    
    * Minor changes
    
    * Extra import
    
    * Docs
    
    Co-authored-by: Carlos Mocholi <carlossmocholi@gmail.com>
    tchaton and carmocca committed Mar 22, 2021
    Configuration menu
    Copy the full SHA
    2064ece View commit details
    Browse the repository at this point in the history

Commits on Mar 23, 2021

  1. fix comparing versions (#6434)

    * fix comparing versions
    
    * chlog
    
    * .
    
    * ...
    
    * datasets
    Borda committed Mar 23, 2021
    Configuration menu
    Copy the full SHA
    8cd75a4 View commit details
    Browse the repository at this point in the history
  2. Prune metrics: regression 8/n (#6636)

    * explained_variance
    
    * tests
    
    * mean_absolute_error
    
    * mean_squared_error
    
    * mean_relative_error
    
    * mean_squared_log_error
    
    * chlog
    Borda committed Mar 23, 2021
    Configuration menu
    Copy the full SHA
    efce2b7 View commit details
    Browse the repository at this point in the history
  3. Prune metyrics: regression 9/n (#6637)

    * psnr
    
    * r2score
    
    * ssim
    
    * chlog
    Borda committed Mar 23, 2021
    Configuration menu
    Copy the full SHA
    f93414d View commit details
    Browse the repository at this point in the history
  4. Refactor base profilers 3/5 (#6621)

    Co-authored-by: tchaton <thomas@grid.ai>
    carmocca and tchaton committed Mar 23, 2021
    Configuration menu
    Copy the full SHA
    36d180e View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    a74909a View commit details
    Browse the repository at this point in the history
  6. Flash predict step (#6577)

    * add predict_step
    
    * Update predict_loop.py
    
    * Update trainer.py
    
    * Update trainer.py
    
    * resolve bugs
    
    * update
    
    * update
    
    * update
    
    * resolve bug
    
    * resolve some failing tests
    
    * udpate tests
    
    * update
    
    * resolve tests
    
    * add a test
    
    * remove typo
    
    * add a test for attachement
    
    * update
    
    * changed to on_train_dataloader
    
    * remove __flash_special_attr__
    
    * resolve tests
    
    * update
    
    * update
    
    * update
    
    * update on comments
    
    * Update pytorch_lightning/trainer/data_loading.py
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    
    Co-authored-by: Justus Schock <12886177+justusschock@users.noreply.github.com>
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    3 people committed Mar 23, 2021
    Configuration menu
    Copy the full SHA
    0995d30 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    3cf0c31 View commit details
    Browse the repository at this point in the history
  8. Refactor PyTorch profiler 4/5 (#6349)

    Co-authored-by: thomas chaton <thomas@grid.ai>
    carmocca and tchaton committed Mar 23, 2021
    Configuration menu
    Copy the full SHA
    51b10f7 View commit details
    Browse the repository at this point in the history
  9. Add PyTorch 1.8 Profiler 5/5 (#6618)

    * Refactor profilers
    
    * Update PassThrough
    
    * WIP - This is broken and will change
    
    * Update pytorch_lightning/profiler/pytorch.py
    
    Co-authored-by: thomas chaton <thomas@grid.ai>
    
    * resolve tests
    
    * resolve tests
    
    * find output
    
    * try something
    
    * update
    
    * add support for test and predict
    
    * update
    
    * update
    
    * use getattr
    
    * test
    
    * test
    
    * update
    
    * tests
    
    * update
    
    * update
    
    * update
    
    * update
    
    * update
    
    * remove file
    
    * update
    
    * update
    
    * update
    
    * update
    
    * update
    
    * test
    
    * update#
    
    * update
    
    * update tests
    
    * update
    
    * add suport for 1.8
    
    * rename records
    
    * add support for 1.8
    
    * update
    
    * resolve flake8
    
    * resolve test
    
    * Refactor basic profilers
    
    * Fixes
    
    * Unused import
    
    * Introduce setup
    
    * Profile on all ranks. Print to stdout on 0
    
    * Introduce dirpath + filename
    
    * CHANGELOG
    
    * Add tests. Address comments
    
    * add `on_run_stage_setup`
    
    * add on_run_stage_setup function
    
    * update
    
    * add test for RegisterRecordFunction
    
    * update lightnng flow direction
    
    * move variable to private
    
    * remove trace
    
    * Undo code that should be in 3/4
    
    * Multi-stage multi-rank
    
    * 2/5 changes
    
    * Pass stage in __del__
    
    * Remove TODOs
    
    * Describe on_evaluation_end. Add tests
    
    * Typo
    
    * Address comments
    
    * deepcopy tests
    
    * Advanced teardown
    
    * Fix teardown test
    
    * Fix tests
    
    * Minor change
    
    * Update CHANGELOG.md
    
    * Fix test
    
    * Quick fixes
    
    * Fix 6522
    
    * resolve ddp tests
    
    * resolve tests
    
    * resolve some tests
    
    * update tests
    
    * resolve tests
    
    * update
    
    * resolve tests
    
    * resolve some tests
    
    * Missed fixes from 3/5
    
    * Fixes
    
    * resolve some tests
    
    * resolve test for 1.7.1
    
    * Broken refactor
    
    * Missed stage
    
    * Minor changes
    
    * resolve tests
    
    * Update CHANGELOG
    
    * resolve bug
    
    * remove print
    
    * Typo
    
    * Cleanup
    
    * resolve ddp test
    
    * remove barrier
    
    * update profiler
    
    * update
    
    * Smaller model
    
    * update
    
    * resolve tests
    
    * update
    
    * Minor changes. CHANGELOG
    
    * Minimize diff
    
    * update to 1.8.1
    
    * RunIf. Extra code. Check segfault
    
    * resolve tests
    
    * Typo. Bad merge
    
    * Fixing a bad merge
    
    * replace for kineto
    
    * Update pytorch_lightning/profiler/pytorch.py
    
    Co-authored-by: ananthsub <ananth.subramaniam@gmail.com>
    
    * Update pytorch_lightning/profiler/pytorch.py
    
    Co-authored-by: ananthsub <ananth.subramaniam@gmail.com>
    
    * Minor changes
    
    * Bad merge
    
    * Use lists for flexibility
    
    * Use sets
    
    * predict_step
    
    * Ananth's suggestion
    
    * update
    
    * Docs
    
    * Update pl_examples/basic_examples/profiler_example.py
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    
    * update example
    
    * update example
    
    Co-authored-by: Carlos Mocholi <carlossmocholi@gmail.com>
    Co-authored-by: ananthsub <ananth.subramaniam@gmail.com>
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    4 people committed Mar 23, 2021
    Configuration menu
    Copy the full SHA
    fd5cb7f View commit details
    Browse the repository at this point in the history
  10. update coverage config (#6524)

    * update coverage config
    
    * parallel
    
    * parallel
    
    * Apply suggestions from code review
    
    * Apply suggestions from code review
    
    * paralel
    
    * paralel
    
    * paralel
    
    * combine
    
    * combine
    
    * .
    
    * ..
    
    * ..
    
    * ..
    
    * rev
    
    * cb
    
    * cb
    
    * drop
    
    * drop
    
    * .
    
    * ..
    
    * ...
    
    * ...
    
    * ...
    
    * .
    Borda committed Mar 23, 2021
    Configuration menu
    Copy the full SHA
    64d0fa4 View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    741c452 View commit details
    Browse the repository at this point in the history
  12. Configuration menu
    Copy the full SHA
    b1e3dcc View commit details
    Browse the repository at this point in the history

Commits on Mar 24, 2021

  1. Prune metrics: others 11/DoNe (#6659)

    * classif
    
    * grad_img
    
    * nlp
    
    * ssl
    
    * format
    Borda committed Mar 24, 2021
    Configuration menu
    Copy the full SHA
    70beddf View commit details
    Browse the repository at this point in the history
  2. fix: update example autoencoder.py to reflect args (#6638)

    * fix: update example autoencoder.py to reflect args
    
    * Update pl_examples/basic_examples/autoencoder.py
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    bmahlbrand and carmocca committed Mar 24, 2021
    Configuration menu
    Copy the full SHA
    cbca6cd View commit details
    Browse the repository at this point in the history
  3. Docs/robots (#6658)

    Borda committed Mar 24, 2021
    Configuration menu
    Copy the full SHA
    5733889 View commit details
    Browse the repository at this point in the history
  4. Feature/double precision (#6595)

    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    Co-authored-by: Rohit Gupta <rohitgr1998@gmail.com>
    Co-authored-by: Justus Schock <12886177+justusschock@users.noreply.github.com>
    4 people committed Mar 24, 2021
    Configuration menu
    Copy the full SHA
    d02fe34 View commit details
    Browse the repository at this point in the history
  5. Follow E231 [flake8] (#6110)

    * Remove E231 from ignore list
    
    * Follow E231
    
    * Update pytorch_lightning/trainer/data_loading.py
    akihironitta committed Mar 24, 2021
    Configuration menu
    Copy the full SHA
    ac60536 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    ab4c838 View commit details
    Browse the repository at this point in the history
  7. add copyr (#6661)

    Borda committed Mar 24, 2021
    Configuration menu
    Copy the full SHA
    d471fa3 View commit details
    Browse the repository at this point in the history
  8. MetricsHolder clean-up + typing (#6645)

    * Metrics holder cleanup and better error message
    
    * Update pytorch_lightning/trainer/connectors/logger_connector/logger_connector.py
    
    * _VALUE -> _METRIC_TYPE
    carmocca committed Mar 24, 2021
    Configuration menu
    Copy the full SHA
    2dd6f9e View commit details
    Browse the repository at this point in the history

Commits on Mar 25, 2021

  1. Configuration menu
    Copy the full SHA
    b8ef52b View commit details
    Browse the repository at this point in the history
  2. Fix checkpoint callback & Trainer.test(_) issue for TPUs (#6654)

    * Fix checkpoint callback issue for TPUs
    
    * update changelog
    
    * add barrier
    
    * apply code suggestions
    
    * update trainer test
    
    * remove spaces
    
    * fix tpu tests
    
    * Apply suggestions from code review
    
    * add comment
    
    Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
    kaushikb11 and Borda committed Mar 25, 2021
    Configuration menu
    Copy the full SHA
    2cbdc01 View commit details
    Browse the repository at this point in the history
  3. Update CODEOWNERS (#6220)

    Borda committed Mar 25, 2021
    Configuration menu
    Copy the full SHA
    92a1671 View commit details
    Browse the repository at this point in the history
  4. Support teardown hook on DataModule (#4673)

    Co-authored-by: Carlos Mocholi <carlossmocholi@gmail.com>
    Co-authored-by: chaton <thomas@grid.ai>
    3 people committed Mar 25, 2021
    Configuration menu
    Copy the full SHA
    40976e4 View commit details
    Browse the repository at this point in the history
  5. Add on_epoch_start to run at the beginning of every loop irrespective…

    … of train/val/test (#6498)
    
    * update docs
    
    * add hook and update docs
    
    * update tests
    
    * chlog
    
    * Update CHANGELOG.md
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    
    * chlog
    
    Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
    rohitgr7 and awaelchli committed Mar 25, 2021
    Configuration menu
    Copy the full SHA
    9be092d View commit details
    Browse the repository at this point in the history
  6. Simplify deprecations (#6620)

    * use external deprecate
    
    * simplify
    
    * simplify
    
    * simplify
    
    * flake8
    
    * .
    
    * others
    
    * .
    Borda committed Mar 25, 2021
    Configuration menu
    Copy the full SHA
    217c12a View commit details
    Browse the repository at this point in the history
  7. Resolve schedule step bug for PyTorch Profiler (#6674)

    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    tchaton and carmocca committed Mar 25, 2021
    Configuration menu
    Copy the full SHA
    0ea8f39 View commit details
    Browse the repository at this point in the history
  8. Add artifcact_location arg to MLFlow logger (#6677)

    * Add artifcact_location arg to MLFlow logger
    
    * Add CHANGELOG URL
    
    * Update test
    ethanwharris committed Mar 25, 2021
    Configuration menu
    Copy the full SHA
    6b990f3 View commit details
    Browse the repository at this point in the history

Commits on Mar 26, 2021

  1. Configuration menu
    Copy the full SHA
    bc61361 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    b730a5a View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    21fc5eb View commit details
    Browse the repository at this point in the history
  4. [warning] Add warning when values are not being reduced (#6417)

    * add warning non reduced
    
    * add test
    
    * update test
    
    * update changelog
    
    * Update pytorch_lightning/trainer/connectors/logger_connector/epoch_result_store.py
    
    Co-authored-by: Justus Schock <12886177+justusschock@users.noreply.github.com>
    
    * update
    
    Co-authored-by: Kaushik B <45285388+kaushikb11@users.noreply.github.com>
    Co-authored-by: Justus Schock <12886177+justusschock@users.noreply.github.com>
    3 people committed Mar 26, 2021
    Configuration menu
    Copy the full SHA
    0e45220 View commit details
    Browse the repository at this point in the history

Commits on Mar 28, 2021

  1. Configuration menu
    Copy the full SHA
    f0c5479 View commit details
    Browse the repository at this point in the history

Commits on Mar 29, 2021

  1. remake nvidia docker (#6686)

    * use latest
    
    * remake
    
    * examples
    Borda committed Mar 29, 2021
    Configuration menu
    Copy the full SHA
    dcf6e4e View commit details
    Browse the repository at this point in the history
  2. More explicit exception message when testing with fast_dev_run=True (#…

    …6667)
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    ashleve and carmocca committed Mar 29, 2021
    Configuration menu
    Copy the full SHA
    cca0eca View commit details
    Browse the repository at this point in the history
  3. support python 3.9 (#4944)

    * support python 3.9
    
    * update CI
    
    * onnxruntime
    
    * .
    
    * .
    
    * onnxruntime
    
    * t 55
    
    * t 75
    
    * add script
    
    * use
    
    * onnx
    
    * onnx
    
    * onnx
    
    * whl
    
    * np
    
    * find
    
    * 21
    
    * Apply suggestions from code review
    
    * Apply suggestions from code review
    
    * onnx
    
    * CI
    
    * req
    
    * ~ dockers
    
    * min
    
    * .
    
    * drop horovod
    
    * drop horovod
    
    * drop horovod
    
    * fix
    
    * fix
    
    * .
    Borda committed Mar 29, 2021
    Configuration menu
    Copy the full SHA
    5b5a5cc View commit details
    Browse the repository at this point in the history
  4. [TPU] update is_tpu_exists utils internal logic to rely on xmp.spawn (#…

    …6719)
    
    * update_logic
    
    * update
    
    * Update tests/utilities/test_xla_device_utils.py
    
    * Update pytorch_lightning/utilities/xla_device.py
    
    Co-authored-by: Kaushik B <45285388+kaushikb11@users.noreply.github.com>
    
    * Update pytorch_lightning/utilities/xla_device.py
    
    Co-authored-by: Kaushik B <45285388+kaushikb11@users.noreply.github.com>
    
    * update test
    
    * Update tests/utilities/test_xla_device_utils.py
    
    * update
    
    * Apply fix
    
    * Docstring
    
    * flake8
    
    * update
    
    Co-authored-by: Your Name <you@example.com>
    Co-authored-by: Kaushik B <45285388+kaushikb11@users.noreply.github.com>
    Co-authored-by: Carlos Mocholi <carlossmocholi@gmail.com>
    4 people committed Mar 29, 2021
    Configuration menu
    Copy the full SHA
    3a4c424 View commit details
    Browse the repository at this point in the history
  5. [refactor] Move save_function to accelerator 1/n [DeepSpeed] (#6689)

    * move save_checkpoint responsability to accelerator
    
    * update
    tchaton committed Mar 29, 2021
    Configuration menu
    Copy the full SHA
    646cf2f View commit details
    Browse the repository at this point in the history
  6. [Model Parallel] Add configure sharded model hook (#6679)

    * Add base hook for model parallel
    
    * fix callback signature
    
    * Simplify hook
    
    * Add hook logic
    
    * add tests
    
    * add property setter
    
    * add logic for being called once
    
    * Update changelog
    
    * Fix
    
    * fix return type
    
    * fix lambda callback test
    
    * Fix tests
    
    * Apply code suggestions
    
    * add logic for setup_optimizers_predispatch
    
    * add common dummy model
    
    * Swap call order
    
    * Remove test that isn't needed anymore
    
    * Update tests
    
    * Add a bit more doc
    
    * Few code review fixes
    
    * Update pytorch_lightning/accelerators/accelerator.py
    
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    
    * Change hook name
    
    * Fix test
    
    * Test setup hook, refactor names
    
    * Swap call order of callbacks and model initialization
    
    * Change name of context manager
    
    Co-authored-by: SeanNaren <sean@grid.ai>
    Co-authored-by: Sean Naren <sean.narenthiran@gmail.com>
    Co-authored-by: Carlos Mocholí <carlossmocholi@gmail.com>
    4 people committed Mar 29, 2021
    Configuration menu
    Copy the full SHA
    f79a13e View commit details
    Browse the repository at this point in the history
  7. update readme by v1.2.x (#6728)

    Borda committed Mar 29, 2021
    Configuration menu
    Copy the full SHA
    3c86193 View commit details
    Browse the repository at this point in the history

Commits on Mar 30, 2021

  1. Configuration menu
    Copy the full SHA
    9044470 View commit details
    Browse the repository at this point in the history
  2. update chlog v1.2.5 (#6742)

    * update chlog v1.2.5
    
    * legacy
    Borda committed Mar 30, 2021
    Configuration menu
    Copy the full SHA
    583fcf2 View commit details
    Browse the repository at this point in the history