Training on multiple GPUs and multi-node training with PyTorch DistributedDataParallel
In this video we’ll cover how multi-GPU and multi-node training works in general.
We’ll also show how to do this using PyTorch DistributedDataParallel and how PyTorch Lightning automates this for you.
Follow along with this notebook: https://bit.ly/33YqQ3O
To learn more about Lightning, please visit the official website: https://pytorchlightning.ai
👉 Read our docs: https://bit.ly/3515yBU
👉 Github: https://github.com/PyTorchLightning/p…
👉 Join our slack: https://bit.ly/3j52N7y
👉 Twitter: https://twitter.com/PyTorchLightnin
👉 SUBSCRIBE!