PyTorch

For multiple GPUs, use distributed data parallel

Related libraries

Tutorials

Tips