Skip to content

Commit 9b742d2

Browse files
committed
training batch clean up
1 parent 3676b6c commit 9b742d2

File tree

1 file changed

+6
-0
lines changed

1 file changed

+6
-0
lines changed

pytorch_lightning/trainer/distrib_data_parallel.py

+6
Original file line numberDiff line numberDiff line change
@@ -427,7 +427,13 @@ def ddp_train(self, process_idx, model, is_master=False, proc_offset=0):
427427
# try to init for 20 times at max in case ports are taken
428428
# where to store ip_table
429429
model.trainer = self
430+
print('-'*100)
431+
print('starting ddp')
432+
print('-'*100)
430433
model.init_ddp_connection(self.proc_rank, self.world_size, self.is_slurm_managing_tasks)
434+
print('-'*100)
435+
print('ddp started')
436+
print('-'*100)
431437

432438
# CHOOSE OPTIMIZER
433439
# allow for lr schedulers as well

0 commit comments

Comments
 (0)