Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bump pytorch-lightning from 1.6.5 to 1.7.0 in /requirements #697

Merged
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 0 additions & 2 deletions .github/dependabot.yml
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,5 @@ updates:
# Allow up to 2 open pull requests at a time
open-pull-requests-limit: 2
ignore:
# torch, tensorboard require protobuf < 4
- dependency-name: "protobuf"
# segmentation-models-pytorch requires older timm, can't update
- dependency-name: "timm"
1 change: 1 addition & 0 deletions docs/conf.py
Original file line number Diff line number Diff line change
Expand Up @@ -60,6 +60,7 @@
("py:class", ".."),
# TODO: can't figure out why this isn't found
("py:class", "LightningDataModule"),
("py:class", "pytorch_lightning.core.module.LightningModule"),
# Undocumented class
("py:class", "torchvision.models.resnet.ResNet"),
("py:class", "segmentation_models_pytorch.base.model.SegmentationModel"),
Expand Down
4 changes: 2 additions & 2 deletions experiments/test_chesapeakecvpr_models.py
Original file line number Diff line number Diff line change
Expand Up @@ -89,8 +89,8 @@ def main(args: argparse.Namespace) -> None:
trainer = pl.Trainer(
gpus=[args.device] if torch.cuda.is_available() else None,
logger=False,
progress_bar_refresh_rate=0,
checkpoint_callback=False,
Comment on lines -92 to -93
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Both of these parameters have been deprecated since v1.5 and removed in v1.7.

enable_progress_bar=False,
enable_checkpointing=False,
)

for experiment_dir in os.listdir(args.input_dir):
Expand Down
2 changes: 2 additions & 0 deletions pyproject.toml
Original file line number Diff line number Diff line change
Expand Up @@ -71,6 +71,8 @@ filterwarnings = [
# https://github.com/PyTorchLightning/pytorch-lightning/issues/13256
# https://github.com/PyTorchLightning/pytorch-lightning/pull/13261
"ignore:torch.distributed._sharded_tensor will be deprecated:DeprecationWarning:torch.distributed._sharded_tensor",
# https://github.com/Lightning-AI/lightning/issues/13989
"ignore:SelectableGroups dict interface is deprecated. Use select.:DeprecationWarning:pytorch_lightning.trainer.connectors.callback_connector",
# https://github.com/rasterio/rasterio/issues/1742
# https://github.com/rasterio/rasterio/pull/1753
"ignore:Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated:DeprecationWarning:rasterio.crs",
Expand Down
1 change: 0 additions & 1 deletion requirements/required.old
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,6 @@ numpy==1.21.6;python_version=='3.7'
omegaconf==2.2.2
packaging==21.3
pillow==9.2.0
protobuf==3.20.1
pyproj==3.3.1;python_version>='3.8'
pyproj==3.2.0;python_version=='3.7'
pytorch-lightning==1.6.4
Expand Down
3 changes: 1 addition & 2 deletions requirements/required.txt
Original file line number Diff line number Diff line change
Expand Up @@ -10,9 +10,8 @@ numpy==1.23.1;python_version>='3.8'
omegaconf==2.2.2
packaging==21.3
pillow==9.2.0
protobuf==3.20.1
pyproj==3.3.1;python_version>='3.8'
pytorch-lightning==1.6.5
pytorch-lightning==1.7.0
rasterio==1.3.0;python_version>='3.8'
rtree==1.0.0
scikit-learn==1.1.1;python_version>='3.8'
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/bigearthnet.py
Original file line number Diff line number Diff line change
Expand Up @@ -90,7 +90,7 @@ def __init__(
batch_size: The batch size to use in all created DataLoaders
num_workers: The number of workers to use in all created DataLoaders
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.bands = bands
self.num_classes = num_classes
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/chesapeake.py
Original file line number Diff line number Diff line change
Expand Up @@ -64,7 +64,7 @@ def __init__(
Raises:
ValueError: if ``use_prior_labels`` is used with ``class_set==7``
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
for state in train_splits + val_splits + test_splits:
assert state in ChesapeakeCVPR.splits
assert class_set in [5, 7]
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/cowc.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ def __init__(
batch_size: The batch size to use in all created DataLoaders
num_workers: The number of workers to use in all created DataLoaders
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.seed = seed
self.batch_size = batch_size
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/cyclone.py
Original file line number Diff line number Diff line change
Expand Up @@ -44,7 +44,7 @@ def __init__(
api_key: The RadiantEarth MLHub API key to use if the dataset needs to be
downloaded
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.seed = seed
self.batch_size = batch_size
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/deepglobelandcover.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ def __init__(
num_workers: The number of workers to use in all created DataLoaders
val_split_pct: What percentage of the dataset to use as a validation set
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.batch_size = batch_size
self.num_workers = num_workers
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/etci2021.py
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ def __init__(
batch_size: The batch size to use in all created DataLoaders
num_workers: The number of workers to use in all created DataLoaders
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.seed = seed
self.batch_size = batch_size
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/eurosat.py
Original file line number Diff line number Diff line change
Expand Up @@ -68,7 +68,7 @@ def __init__(
batch_size: The batch size to use in all created DataLoaders
num_workers: The number of workers to use in all created DataLoaders
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.batch_size = batch_size
self.num_workers = num_workers
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/fair1m.py
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ def __init__(
val_split_pct: What percentage of the dataset to use as a validation set
test_split_pct: What percentage of the dataset to use as a test set
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.batch_size = batch_size
self.num_workers = num_workers
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/inria.py
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@ def __init__(
num_patches_per_tile: Number of random patches per sample
predict_on: Directory/Dataset of images to run inference on
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.batch_size = batch_size
self.num_workers = num_workers
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/landcoverai.py
Original file line number Diff line number Diff line change
Expand Up @@ -33,7 +33,7 @@ def __init__(
batch_size: The batch size to use in all created DataLoaders
num_workers: The number of workers to use in all created DataLoaders
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.batch_size = batch_size
self.num_workers = num_workers
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/loveda.py
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ def __init__(
batch_size: The batch size to use in all created DataLoaders
num_workers: The number of workers to use in all created DataLoaders
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.scene = scene
self.batch_size = batch_size
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/naip.py
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,7 @@ def __init__(
num_workers: The number of workers to use in all created DataLoaders
patch_size: size of patches to sample
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.naip_root_dir = naip_root_dir
self.chesapeake_root_dir = chesapeake_root_dir
self.batch_size = batch_size
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/nasa_marine_debris.py
Original file line number Diff line number Diff line change
Expand Up @@ -57,7 +57,7 @@ def __init__(
val_split_pct: What percentage of the dataset to use as a validation set
test_split_pct: What percentage of the dataset to use as a test set
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.batch_size = batch_size
self.num_workers = num_workers
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/oscd.py
Original file line number Diff line number Diff line change
Expand Up @@ -87,7 +87,7 @@ def __init__(
num_patches_per_tile: number of random patches per sample
pad_size: size to pad images to during val/test steps
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.bands = bands
self.train_batch_size = train_batch_size
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/potsdam.py
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ def __init__(
num_workers: The number of workers to use in all created DataLoaders
val_split_pct: What percentage of the dataset to use as a validation set
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.batch_size = batch_size
self.num_workers = num_workers
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/resisc45.py
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ def __init__(
batch_size: The batch size to use in all created DataLoaders
num_workers: The number of workers to use in all created DataLoaders
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.batch_size = batch_size
self.num_workers = num_workers
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/sen12ms.py
Original file line number Diff line number Diff line change
Expand Up @@ -72,7 +72,7 @@ def __init__(
batch_size: The batch size to use in all created DataLoaders
num_workers: The number of workers to use in all created DataLoaders
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
assert band_set in SEN12MS.BAND_SETS.keys()

self.root_dir = root_dir
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/so2sat.py
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,7 @@ def __init__(
unsupervised_mode: Makes the train dataloader return imagery from the train,
val, and test sets
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.batch_size = batch_size
self.num_workers = num_workers
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/ucmerced.py
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ def __init__(
batch_size: The batch size to use in all created DataLoaders
num_workers: The number of workers to use in all created DataLoaders
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.batch_size = batch_size
self.num_workers = num_workers
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/vaihingen.py
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ def __init__(
num_workers: The number of workers to use in all created DataLoaders
val_split_pct: What percentage of the dataset to use as a validation set
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.batch_size = batch_size
self.num_workers = num_workers
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/datamodules/xview.py
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ def __init__(
num_workers: The number of workers to use in all created DataLoaders
val_split_pct: What percentage of the dataset to use as a validation set
"""
super().__init__() # type: ignore[no-untyped-call]
super().__init__()
self.root_dir = root_dir
self.batch_size = batch_size
self.num_workers = num_workers
Expand Down
4 changes: 2 additions & 2 deletions torchgeo/trainers/byol.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,14 +6,14 @@
import random
from typing import Any, Callable, Dict, Optional, Tuple, cast

import pytorch_lightning as pl
import torch
import torch.nn.functional as F
import torchvision
from kornia import augmentation as K
from kornia import filters
from kornia.geometry import transform as KorniaTransform
from packaging.version import parse
from pytorch_lightning.core.lightning import LightningModule
from torch import Tensor, optim
from torch.autograd import Variable
from torch.nn.modules import BatchNorm1d, Conv2d, Linear, Module, ReLU, Sequential
Expand Down Expand Up @@ -304,7 +304,7 @@ def update_target(self) -> None:
pt.data = self.beta * pt.data + (1 - self.beta) * p.data


class BYOLTask(LightningModule):
class BYOLTask(pl.LightningModule):
"""Class for pre-training any PyTorch model using BYOL."""

def config_task(self) -> None:
Expand Down
4 changes: 2 additions & 2 deletions torchgeo/trainers/classification.py
Original file line number Diff line number Diff line change
Expand Up @@ -189,7 +189,7 @@ def validation_step(self, *args: Any, **kwargs: Any) -> None:

if batch_idx < 10:
try:
datamodule = self.trainer.datamodule # type: ignore[union-attr]
datamodule = self.trainer.datamodule # type: ignore[attr-defined]
batch["prediction"] = y_hat_hard
for key in ["image", "label", "prediction"]:
batch[key] = batch[key].cpu()
Expand Down Expand Up @@ -358,7 +358,7 @@ def validation_step(self, *args: Any, **kwargs: Any) -> None:

if batch_idx < 10:
try:
datamodule = self.trainer.datamodule # type: ignore[union-attr]
datamodule = self.trainer.datamodule # type: ignore[attr-defined]
batch["prediction"] = y_hat_hard
for key in ["image", "label", "prediction"]:
batch[key] = batch[key].cpu()
Expand Down
2 changes: 1 addition & 1 deletion torchgeo/trainers/regression.py
Original file line number Diff line number Diff line change
Expand Up @@ -127,7 +127,7 @@ def validation_step(self, *args: Any, **kwargs: Any) -> None:

if batch_idx < 10:
try:
datamodule = self.trainer.datamodule # type: ignore[union-attr]
datamodule = self.trainer.datamodule # type: ignore[attr-defined]
batch["prediction"] = y_hat
for key in ["image", "label", "prediction"]:
batch[key] = batch[key].cpu()
Expand Down
6 changes: 3 additions & 3 deletions torchgeo/trainers/segmentation.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,10 +6,10 @@
import warnings
from typing import Any, Dict, cast

import pytorch_lightning as pl
import segmentation_models_pytorch as smp
import torch
import torch.nn as nn
from pytorch_lightning.core.lightning import LightningModule
from torch import Tensor
from torch.optim.lr_scheduler import ReduceLROnPlateau
from torch.utils.data import DataLoader
Expand All @@ -23,7 +23,7 @@
DataLoader.__module__ = "torch.utils.data"


class SemanticSegmentationTask(LightningModule):
class SemanticSegmentationTask(pl.LightningModule):
"""LightningModule for semantic segmentation of images."""

def config_task(self) -> None:
Expand Down Expand Up @@ -184,7 +184,7 @@ def validation_step(self, *args: Any, **kwargs: Any) -> None:

if batch_idx < 10:
try:
datamodule = self.trainer.datamodule # type: ignore[union-attr]
datamodule = self.trainer.datamodule # type: ignore[attr-defined]
batch["prediction"] = y_hat_hard
for key in ["image", "mask", "prediction"]:
batch[key] = batch[key].cpu()
Expand Down