![Multi-GPU distributed deep learning training at scale with Ubuntu18 DLAMI, EFA on P3dn instances, and Amazon FSx for Lustre | AWS Machine Learning Blog Multi-GPU distributed deep learning training at scale with Ubuntu18 DLAMI, EFA on P3dn instances, and Amazon FSx for Lustre | AWS Machine Learning Blog](https://d2908q01vomqb2.cloudfront.net/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59/2020/05/08/ease-of-running-bert-1.png)
Multi-GPU distributed deep learning training at scale with Ubuntu18 DLAMI, EFA on P3dn instances, and Amazon FSx for Lustre | AWS Machine Learning Blog
![Multi-GPU training. Example using two GPUs, but scalable to all GPUs... | Download Scientific Diagram Multi-GPU training. Example using two GPUs, but scalable to all GPUs... | Download Scientific Diagram](https://www.researchgate.net/profile/Andres-Milioto/publication/323410760/figure/fig1/AS:598487393636352@1519701922416/Multi-GPU-training-Example-using-two-GPUs-but-scalable-to-all-GPUs-available-in.png)
Multi-GPU training. Example using two GPUs, but scalable to all GPUs... | Download Scientific Diagram
![DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research](https://www.microsoft.com/en-us/research/uploads/prod/2021/05/1400x788_deepspeed_no_logo_still-1-scaled.jpg)