You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* removes need to unsqueeze from dp
* removes need to unsqueeze from dp
* fixed examples
* added auto unsqueeze
* added auto unsqueeze
* added auto unsqueeze
* added auto unsqueeze
* Update pytorch_lightning/overrides/data_parallel.py
Co-Authored-By: Adrian Wälchli <[email protected]>
* fixed dp parse
* fixed dp parse
Co-authored-by: Adrian Wälchli <[email protected]>
Copy file name to clipboardexpand all lines: CHANGELOG.md
+1
Original file line number
Diff line number
Diff line change
@@ -25,6 +25,7 @@ The format is based on [Keep a Changelog](http://keepachangelog.com/en/1.0.0/).
25
25
26
26
### Changed
27
27
28
+
- On DP and DDP2 unsqueeze is automated now ([#1319](https://github.com/PyTorchLightning/pytorch-lightning/pull/1319))
28
29
- Does not interfere with a default sampler ([#1318](https://github.com/PyTorchLightning/pytorch-lightning/pull/1318))
29
30
- Enhanced load_from_checkpoint to also forward params to the model ([#1307](https://github.com/PyTorchLightning/pytorch-lightning/pull/1307))
30
31
- Made `evalaute` method private >> `Trainer._evaluate(...)`. ([#1260](https://github.com/PyTorchLightning/pytorch-lightning/pull/1260))
0 commit comments