-
Notifications
You must be signed in to change notification settings - Fork 3.6k
Closed
Labels
3rd partyRelated to a 3rd-partyRelated to a 3rd-partybugSomething isn't workingSomething isn't workinghelp wantedOpen to be worked onOpen to be worked onwaiting on authorWaiting on user action, correction, or updateWaiting on user action, correction, or updatewon't fixThis will not be worked onThis will not be worked on
Description
🐛 Bug
WandbLogger log_model argument does not work as expected. The model checkpoint is always being saved to wandb logs (project_name/run/checkpoints/checkpoint_name.ckpt). The model is not uploaded to wandb, only saved on the drive. It should not be saved on disk nor uploaded.
It is especially important during sweeping (hundreds of runs exhaust available space quickly). I don't use ModelCheckpoint here so the weights should not be saved at all.
To Reproduce
wandb_logger = WandbLogger(project='private_example', config={
**model_args.dict, **data_args.dict, **training_args.dict
}, log_model=False)
trainer = pl.Trainer(...,
logger=[wandb_logger])
Expected behavior
Checkpoints are saved if log_model True and not saved otherwise.
Environment
- PyTorch Version (e.g., 1.0): 1.7.1
- OS (e.g., Linux): Debian
- How you installed PyTorch (
conda
,pip
, source): conda - Python version: 3.6.9
- CUDA/cuDNN version: cuda 11
- wandb version: 0.10.15 (same problem on 0.10.17)
- lightning version: 1.1.4 (same problem on 1.1.8)
Metadata
Metadata
Assignees
Labels
3rd partyRelated to a 3rd-partyRelated to a 3rd-partybugSomething isn't workingSomething isn't workinghelp wantedOpen to be worked onOpen to be worked onwaiting on authorWaiting on user action, correction, or updateWaiting on user action, correction, or updatewon't fixThis will not be worked onThis will not be worked on