Scaling Deep Learning Training: Fast Inter-GPU Communication with NCCL

, Distinguished Engineer, NVIDIA
高度评价
Learn why fast inter-GPU communication is critical to accelerate deep learning training, and how to make sure your system has the right level of performance for your model. Discover NCCL, the inter-GPU communication library used by all deep learning frameworks for inter-GPU communication, and how it combines NVLink with high-speed networks like Infiniband to accelerate communication by an order of magnitude, allowing training to be run on hundreds, or even thousands, of GPUs. See how new technologies in Hopper GPUs and ConnectX-7 allow for NCCL performance to reach new highs on the latest generation of DGX and HGX systems. Finally, get updates on the latest improvements in NCCL, and what should come in the near future.
活动: GTC Digital Spring
日期: March 2023
话题: Deep Learning - Training
行业: HPC / Supercomputing
级别: 中级技术
语言: 英语
话题: Deep Learning
所在地: