Skip to content

The changed things -- DataParallel removed from train.py script #310

Answered by rwightman
jinseok-karl asked this question in Q&A
Discussion options

You must be logged in to vote

@jinseok-karl yes, I removed support for DataParallel in the train script. It wasn't worth maintaining as it conflicts with a number of the useful other training options and seems to be a lower priority for PyTorch team these days. It is slower than DDP and all around not so useful. DDP is really easy to use via the shell script here for multi-gpu single machine training. It's still used for validation because it's hard to get 100% correct multi-gpu validation for ALL samples in a validation set without using it (or writing some extra fiddly code), the DDP default data setup involves padding the last few samples.

See:
https://pytorch.org/docs/stable/generated/torch.nn.DataParallel.html
ht…

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@jinseok-karl
Comment options

Answer selected by jinseok-karl
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants