Video Player is loading.
Current Time 0:00
Duration 0:00
Loaded: 0%
Stream Type LIVE
Remaining Time 0:00
 
1x
    • Chapters
    • descriptions off, selected
    • subtitles off, selected
      • Quality

      Training Optimization for LLM with NVIDIA NeMo and AWS

      , Senior Applied Scientist, Amazon Store Foundational AI
      , Senior Manager, Amazon Store Foundational AI
      Training a large language model at scale while ensuring efficiency and reliability poses numerous challenges. During this presentation, we'll share our experience training LLMs at Amazon Search, utilizing NVIDIA's Nemo Framework in collaboration with AWS. We'll discuss the process of selecting the appropriate training framework, establishing the training infrastructure by harnessing the power of Nemo and AWS , and implementing zero-touch training through automated job monitoring and recovery mechanisms. Additionally, we'll share practical insights into fine-tuning hyperparameters and selecting model architectures to optimize training efficiency. Finally, we'll examine potential paths to further streamline the training process of Large Language Models.
      活动: GTC 24
      日期: March 2024
      级别: 高级技术
      行业: 所有行业
      NVIDIA 技术: NeMo,Triton
      话题: Training AI Models
      语言: 英语
      所在地: