-
Notifications
You must be signed in to change notification settings - Fork 3.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comet logger cannot be pickled after creating an experiment #1682
Comments
@ceyzaguirre4 pls ^^ |
I don't know if it can help or if it is the right place, but a similar error occurswhen running in ddp mode with the WandB logger. WandB uses a lambda function at some point. Does the logger have to pickled ? Couldn't it log only on rank 0 at epoch_end ?
also related: |
I had the same error as @jeremyjordan
I could make the problem go away by directly overwriting the members of
|
Same issue as @F-Barto using a wandb logger across 2 nodes with |
same issue when using wandb logger with ddp |
same here.. @joseluisvaz your workaround doesn't solve the callback issue.. when I try to add a callback like this it is simply being ignored :/ but adding it the Trainer init call normally works.. so I'm pretty sure the error is thrown by the logger (I'm using TB) not the callbacks. |
Same issue, using wandb logger with 8 gpus in an AWS p2.8xlarge machine |
With CometLogger, I get this error only when the experiment name is declared. If it is not declared, I get no issue. |
I still have this error with 1.5.10 on macOS
|
I still see this bug as well with WandB logger. |
Currently having this issue with wandbLogger. |
Having same issue with wandb |
@ebalogun01 Were you able to solve this issue? I'm also seeing the same issue with WandbLogger |
What version are you using? |
@Borda I'm using Lightning 2.1.0.post0 version. Another detail I'd like to add is that, I find that |
🐛 Bug
The Comet logger cannot be pickled after an experiment (at least an OfflineExperiment) has been created.
To Reproduce
Steps to reproduce the behavior:
initialize the logger object (works fine)
initialize a Trainer object with the logger (works fine)
access the
experiment
attribute which creates the OfflineExperiment object (fails)Expected behavior
We should be able to pickle loggers for distributed training.
Environment
- GPU:
- available: False
- version: None
- numpy: 1.18.1
- pyTorch_debug: False
- pyTorch_version: 1.4.0
- pytorch-lightning: 0.7.5
- tensorboard: 2.1.0
- tqdm: 4.42.0
- OS: Darwin
- architecture:
- 64bit
-
- processor: i386
- python: 3.7.6
- version: Darwin Kernel Version 19.3.0: Thu Jan 9 20:58:23 PST 2020; root:xnu-6153.81.5~1/RELEASE_X86_64
The text was updated successfully, but these errors were encountered: