From 8fe60902bc46c2088f832ffe0ee2b4381e50b390 Mon Sep 17 00:00:00 2001 From: Ir1d Date: Tue, 3 Dec 2019 22:35:48 +0800 Subject: [PATCH 1/5] use print for INFO and lower levels summarize() --- pytorch_lightning/core/lightning.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/pytorch_lightning/core/lightning.py b/pytorch_lightning/core/lightning.py index 95a98e51e9233..54c77ae73481c 100644 --- a/pytorch_lightning/core/lightning.py +++ b/pytorch_lightning/core/lightning.py @@ -1022,7 +1022,8 @@ def load_from_checkpoint(cls, checkpoint_path): def summarize(self, mode): model_summary = ModelSummary(self, mode=mode) - logging.info(model_summary) + if logging.getLogger().level <= 20: + print(model_summary) def freeze(self): """Freeze all params for inference From 2828be032068d9b210d438e41018010a2749bfe2 Mon Sep 17 00:00:00 2001 From: Ir1d Date: Tue, 3 Dec 2019 22:39:13 +0800 Subject: [PATCH 2/5] use logging.INFO instead of magic number --- pytorch_lightning/core/lightning.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/pytorch_lightning/core/lightning.py b/pytorch_lightning/core/lightning.py index 54c77ae73481c..2b2ec241bde86 100644 --- a/pytorch_lightning/core/lightning.py +++ b/pytorch_lightning/core/lightning.py @@ -1022,7 +1022,7 @@ def load_from_checkpoint(cls, checkpoint_path): def summarize(self, mode): model_summary = ModelSummary(self, mode=mode) - if logging.getLogger().level <= 20: + if logging.getLogger().level <= logging.INFO: print(model_summary) def freeze(self): From 93d8dbfc44c85c8269c6358825436cd2aff542d9 Mon Sep 17 00:00:00 2001 From: Ir1d Date: Tue, 3 Dec 2019 23:28:45 +0800 Subject: [PATCH 3/5] bring logging.info back for other cases --- pytorch_lightning/core/lightning.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/pytorch_lightning/core/lightning.py b/pytorch_lightning/core/lightning.py index 2b2ec241bde86..f7ba613d8ae82 100644 --- a/pytorch_lightning/core/lightning.py +++ b/pytorch_lightning/core/lightning.py @@ -1022,8 +1022,8 @@ def load_from_checkpoint(cls, checkpoint_path): def summarize(self, mode): model_summary = ModelSummary(self, mode=mode) - if logging.getLogger().level <= logging.INFO: - print(model_summary) + summary_logger = print if logging.getLogger().level <= logging.INFO else logging.info + summary_logger(model_summary) def freeze(self): """Freeze all params for inference From 9347e2e593f19194c0fcda3edf70566e9d74eb6e Mon Sep 17 00:00:00 2001 From: Ir1d Date: Wed, 4 Dec 2019 15:21:47 +0800 Subject: [PATCH 4/5] move logging config to __init__.py --- pytorch_lightning/__init__.py | 2 ++ pytorch_lightning/core/lightning.py | 6 ++++-- pytorch_lightning/trainer/trainer.py | 3 --- 3 files changed, 6 insertions(+), 5 deletions(-) diff --git a/pytorch_lightning/__init__.py b/pytorch_lightning/__init__.py index 2897d19bd0cb5..6252e34326492 100644 --- a/pytorch_lightning/__init__.py +++ b/pytorch_lightning/__init__.py @@ -28,9 +28,11 @@ from .trainer.trainer import Trainer from .core.lightning import LightningModule from .core.decorators import data_loader + import logging __all__ = [ 'Trainer', 'LightningModule', 'data_loader', ] + logging.basicConfig(level=logging.INFO) diff --git a/pytorch_lightning/core/lightning.py b/pytorch_lightning/core/lightning.py index f7ba613d8ae82..0461b20cee545 100644 --- a/pytorch_lightning/core/lightning.py +++ b/pytorch_lightning/core/lightning.py @@ -1022,8 +1022,10 @@ def load_from_checkpoint(cls, checkpoint_path): def summarize(self, mode): model_summary = ModelSummary(self, mode=mode) - summary_logger = print if logging.getLogger().level <= logging.INFO else logging.info - summary_logger(model_summary) + if logging.getLogger().level <= logging.INFO: + print(model_summary) + else: + logging.info('\n' + model_summary.__str__()) def freeze(self): """Freeze all params for inference diff --git a/pytorch_lightning/trainer/trainer.py b/pytorch_lightning/trainer/trainer.py index eb00cda4690c1..6c844f53b68a7 100644 --- a/pytorch_lightning/trainer/trainer.py +++ b/pytorch_lightning/trainer/trainer.py @@ -240,9 +240,6 @@ def __init__(self, self.amp_level = amp_level self.init_amp(use_amp) - # set logging options - logging.basicConfig(level=logging.INFO) - @property def slurm_job_id(self): try: From 9c08d771b19ff9f730126e920caecbebd45c863b Mon Sep 17 00:00:00 2001 From: Ir1d Date: Wed, 4 Dec 2019 15:22:12 +0800 Subject: [PATCH 5/5] prepend the model summary with a newline --- pytorch_lightning/core/lightning.py | 5 +---- 1 file changed, 1 insertion(+), 4 deletions(-) diff --git a/pytorch_lightning/core/lightning.py b/pytorch_lightning/core/lightning.py index 0461b20cee545..97fb772e4890d 100644 --- a/pytorch_lightning/core/lightning.py +++ b/pytorch_lightning/core/lightning.py @@ -1022,10 +1022,7 @@ def load_from_checkpoint(cls, checkpoint_path): def summarize(self, mode): model_summary = ModelSummary(self, mode=mode) - if logging.getLogger().level <= logging.INFO: - print(model_summary) - else: - logging.info('\n' + model_summary.__str__()) + logging.info('\n' + model_summary.__str__()) def freeze(self): """Freeze all params for inference