Skip to content

Commit e7e9f1f

Browse files
committed
change default logger to dedicated one
1 parent 5e013f6 commit e7e9f1f

14 files changed

+19
-15
lines changed

pytorch_lightning/callbacks/early_stopping.py

+1-2
Original file line numberDiff line numberDiff line change
@@ -5,13 +5,12 @@
55
66
"""
77

8-
import logging as log
98
import warnings
109

1110
import numpy as np
1211

1312
from .base import Callback
14-
13+
from ..core import root_logger
1514

1615
class EarlyStopping(Callback):
1716
r"""

pytorch_lightning/callbacks/model_checkpoint.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -7,13 +7,13 @@
77

88
import os
99
import shutil
10-
import logging as log
1110
import warnings
1211
import re
1312

1413
import numpy as np
1514

1615
from pytorch_lightning.callbacks.base import Callback
16+
from pytorch_lightning.core import root_logger as log
1717

1818

1919
class ModelCheckpoint(Callback):

pytorch_lightning/core/__init__.py

+3
Original file line numberDiff line numberDiff line change
@@ -312,6 +312,9 @@ def test_dataloader(self):
312312
313313
"""
314314

315+
import logging
316+
root_logger = logging.getLogger("lightning")
317+
315318
from .decorators import data_loader
316319
from .lightning import LightningModule
317320

pytorch_lightning/core/lightning.py

+1-2
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,5 @@
11
import collections
22
import inspect
3-
import logging as log
43
import os
54
import warnings
65
from abc import ABC, abstractmethod
@@ -11,6 +10,7 @@
1110
import torch.distributed as dist
1211
from torch.optim import Adam
1312

13+
from pytorch_lightning.core import root_logger as log
1414
from pytorch_lightning.core.grads import GradInformation
1515
from pytorch_lightning.core.hooks import ModelHooks
1616
from pytorch_lightning.core.saving import ModelIO, load_hparams_from_tags_csv
@@ -25,7 +25,6 @@
2525
except ImportError:
2626
XLA_AVAILABLE = False
2727

28-
2928
class LightningModule(ABC, GradInformation, ModelIO, ModelHooks):
3029

3130
def __init__(self, *args, **kwargs):

pytorch_lightning/core/memory.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -3,14 +3,14 @@
33
'''
44

55
import gc
6-
import logging as log
76
import os
87
import subprocess
98
from subprocess import PIPE
109

1110
import numpy as np
1211
import torch
1312

13+
from ..core import root_logger as log
1414

1515
class ModelSummary(object):
1616

pytorch_lightning/core/saving.py

+3-2
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,9 @@
1-
import os
21
import csv
3-
import logging as log
2+
import os
43
from argparse import Namespace
54

5+
from ..core import root_logger as log
6+
67

78
class ModelIO(object):
89

pytorch_lightning/profiler/profiler.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66
import pstats
77
import io
88
from abc import ABC, abstractmethod
9-
import logging as log
9+
from ..core import root_logger as log
1010

1111

1212
class BaseProfiler(ABC):

pytorch_lightning/trainer/auto_mix_precision.py

+2-1
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,14 @@
11

22
from abc import ABC
33

4+
from ..core import root_logger as log
5+
46
try:
57
from apex import amp
68

79
APEX_AVAILABLE = True
810
except ImportError:
911
APEX_AVAILABLE = False
10-
import logging as log
1112

1213

1314
class TrainerAMPMixin(ABC):

pytorch_lightning/trainer/distrib_data_parallel.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -113,14 +113,14 @@ def train_fx(trial_hparams, cluster_manager, _):
113113
114114
"""
115115

116-
import logging as log
117116
import os
118117
import re
119118
import warnings
120119
from abc import ABC, abstractmethod
121120
from typing import Union
122121

123122
import torch
123+
from pytorch_lightning.core import root_logger as log
124124
from pytorch_lightning.loggers import LightningLoggerBase
125125

126126
from pytorch_lightning.utilities.debugging import MisconfigurationException

pytorch_lightning/trainer/distrib_parts.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -335,12 +335,12 @@
335335
"""
336336

337337
from abc import ABC, abstractmethod
338-
import logging as log
339338
import os
340339
import signal
341340

342341
import torch
343342

343+
from pytorch_lightning.core import root_logger as log
344344
from pytorch_lightning.overrides.data_parallel import (
345345
LightningDistributedDataParallel,
346346
LightningDataParallel,

pytorch_lightning/trainer/trainer.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,6 @@
11
import os
22
import sys
33
import warnings
4-
import logging as log
54
from typing import Union, Optional, List, Dict, Tuple, Iterable
65
from argparse import ArgumentParser
76

@@ -13,6 +12,7 @@
1312
from tqdm.auto import tqdm
1413
from torch.optim.optimizer import Optimizer
1514

15+
from pytorch_lightning.core import root_logger as log
1616
from pytorch_lightning.callbacks import ModelCheckpoint, EarlyStopping
1717
from pytorch_lightning.loggers import LightningLoggerBase
1818
from pytorch_lightning.profiler.profiler import BaseProfiler

pytorch_lightning/trainer/training_io.py

+1
Original file line numberDiff line numberDiff line change
@@ -102,6 +102,7 @@
102102
import torch
103103
import torch.distributed as dist
104104

105+
from pytorch_lightning.core import root_logger as log
105106
from pytorch_lightning.core.lightning import LightningModule
106107
from pytorch_lightning.loggers import LightningLoggerBase
107108
from pytorch_lightning.overrides.data_parallel import (

pytorch_lightning/trainer/training_loop.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -125,13 +125,13 @@ def training_step(self, batch, batch_idx):
125125

126126
import copy
127127
import warnings
128-
import logging as log
129128
from abc import ABC, abstractmethod
130129
from typing import Union, List
131130

132131
import numpy as np
133132
from torch.utils.data import DataLoader
134133

134+
from pytorch_lightning.core import root_logger as log
135135
from pytorch_lightning.core.lightning import LightningModule
136136
from pytorch_lightning.loggers import LightningLoggerBase
137137
from pytorch_lightning.utilities.debugging import MisconfigurationException

pytorch_lightning/trainer/training_tricks.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,9 @@
1-
import logging as log
21
from abc import ABC, abstractmethod
32

43
import torch
54
import math
65

6+
from pytorch_lightning.core import root_logger as log
77
from pytorch_lightning.callbacks import GradientAccumulationScheduler
88

99
EPSILON = 1e-6

0 commit comments

Comments
 (0)