Colossal-AI: Scaling AI Models in the Big Model Era

, AI System Engineer, HPC-AI Technology, Inc.

Deep learning (DL) has shown impressive capabilities across various fields. However, the growing size of DL models is outpacing hardware capacity, resulting in a demand for large models to be trained and inferenced more efficiently and easily. We'll introduce a user-friendly DL system, Colossal-AI, that enables users to maximize the efficiency of AI training and inference with drastically reduced costs. It integrates advanced techniques such as efficient multidimensional parallelism, heterogeneous memory management, adaptive task scheduling, and more. You'll walk away with a better understanding of parallelism and memory optimization techniques behind large model training and inference, learn practical applications of DL system, including natural language processing, computer vision, bioinformatics, and more, and be able to contribute to the large AI model era of the future. While it's not required, basic understanding of DL and distributed systems will help you better learn the design of Colossal-AI. 

 

Prerequisite(s)

  • While it is not required, basic understanding of deep learning and distributed systems will help you better learn the design of Colossal-AI. 

 

Please disregard any reference to "Event Code" for access to training materials. "Event Codes" are only valid during the original live session. Explore more training options offered by the NVIDIA Deep Learning Institute (DLI). Choose from an extensive catalog of self-paced, online courses or instructor-led virtual workshops to help you develop key skills in AI, HPC, graphics & simulation, and more.

活动: GTC Digital Spring
日期: March 2023
话题: Deep Learning - Training
行业: HPC / Supercomputing
级别: 中级技术
语言: 英语
话题: Deep Learning
所在地: