Skip to content

Commit 3d18099

Browse files
removed decorators (#1079)
1 parent 2bc01a0 commit 3d18099

File tree

5 files changed

+0
-12
lines changed

5 files changed

+0
-12
lines changed

docs/source/tpu.rst

-1
Original file line numberDiff line numberDiff line change
@@ -101,7 +101,6 @@ train_dataloader (and val, train) code as follows.
101101
102102
import torch_xla.core.xla_model as xm
103103
104-
@pl.data_loader
105104
def train_dataloader(self):
106105
dataset = MNIST(
107106
os.getcwd(),

pl_examples/domain_templates/gan.py

-1
Original file line numberDiff line numberDiff line change
@@ -167,7 +167,6 @@ def configure_optimizers(self):
167167
opt_d = torch.optim.Adam(self.discriminator.parameters(), lr=lr, betas=(b1, b2))
168168
return [opt_g, opt_d], []
169169

170-
@data_loader
171170
def train_dataloader(self):
172171
transform = transforms.Compose([transforms.ToTensor(),
173172
transforms.Normalize([0.5], [0.5])])

pl_examples/full_examples/imagenet/imagenet_example.py

-3
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,6 @@
2020

2121
import pytorch_lightning as pl
2222
from pytorch_lightning.core import LightningModule
23-
from pytorch_lightning.core import data_loader
2423

2524
# pull out resnet names from torchvision models
2625
MODEL_NAMES = sorted(
@@ -132,7 +131,6 @@ def configure_optimizers(self):
132131
scheduler = lr_scheduler.ExponentialLR(optimizer, gamma=0.1)
133132
return [optimizer], [scheduler]
134133

135-
@data_loader
136134
def train_dataloader(self):
137135
normalize = transforms.Normalize(
138136
mean=[0.485, 0.456, 0.406],
@@ -163,7 +161,6 @@ def train_dataloader(self):
163161
)
164162
return train_loader
165163

166-
@data_loader
167164
def val_dataloader(self):
168165
normalize = transforms.Normalize(
169166
mean=[0.485, 0.456, 0.406],

pytorch_lightning/core/lightning.py

-4
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,6 @@
1111
import torch.distributed as dist
1212
from torch.optim import Adam
1313

14-
from pytorch_lightning.core.decorators import data_loader
1514
from pytorch_lightning.core.grads import GradInformation
1615
from pytorch_lightning.core.hooks import ModelHooks
1716
from pytorch_lightning.core.saving import ModelIO, load_hparams_from_tags_csv
@@ -1139,7 +1138,6 @@ def train_dataloader(self):
11391138
"""
11401139
return None
11411140

1142-
@data_loader
11431141
def tng_dataloader(self): # todo: remove in v1.0.0
11441142
"""Implement a PyTorch DataLoader.
11451143
@@ -1239,7 +1237,6 @@ def val_dataloader(self):
12391237
12401238
.. code-block:: python
12411239
1242-
@pl.data_loader
12431240
def val_dataloader(self):
12441241
transform = transforms.Compose([transforms.ToTensor(),
12451242
transforms.Normalize((0.5,), (1.0,))])
@@ -1254,7 +1251,6 @@ def val_dataloader(self):
12541251
return loader
12551252
12561253
# can also return multiple dataloaders
1257-
@pl.data_loader
12581254
def val_dataloader(self):
12591255
return [loader_a, loader_b, ..., loader_n]
12601256

tests/models/debug.py

-3
Original file line numberDiff line numberDiff line change
@@ -41,14 +41,11 @@ def validation_epoch_end(self, outputs):
4141
def configure_optimizers(self):
4242
return [torch.optim.Adam(self.parameters(), lr=0.02)]
4343

44-
@pl.data_loader
4544
def train_dataloader(self):
4645
return DataLoader(MNIST('path/to/save', train=True), batch_size=32)
4746

48-
@pl.data_loader
4947
def val_dataloader(self):
5048
return DataLoader(MNIST('path/to/save', train=False), batch_size=32)
5149

52-
@pl.data_loader
5350
def test_dataloader(self):
5451
return DataLoader(MNIST('path/to/save', train=False), batch_size=32)

0 commit comments

Comments
 (0)