Video Player is loading.
Current Time 
Duration 
Loaded: 0%
Stream Type LIVE
Remaining Time 
 
1x
    • Chapters
    • descriptions off, selected
    • captions and subtitles off, selected

      Large Language Model Fine-Tuning using Parameter Efficient Fine-Tuning (PEFT)

      , Senior Product Manager, AI, Domino Data Lab
      , Director of Solution Architecture, Tech Alliances, Domino Data Lab
      We all recognize the immense business opportunity from generative AI and large language models (LLMs) — particularly those trained or developed on proprietary company data. However, developing them is resource-intensive, time-consuming, and requires deep technical expertise. The NVIDIA NeMo Framework accelerates LLM training by up to 30% (models ranging from 22 billion to 1 trillion parameters) — a quick, efficient, containerized framework for model training, evaluation, and inference. In this session, we'll use NVIDIA NeMo Megatron — a powerful transformer developed by the Applied Deep Learning Research team at NVIDIA — to fine-tune using parameter efficient fine tuning (PEFT) on Domino’s Enterprise AI Platform. We’ll walk through the end-to-end model life cycle, starting with NVIDIA’s NeMo Toolkit in Domino’s AI Project Hub, customizing a data science environment for fine-tuning in Domino, then use NeMo Megatron to encode text prompts and generate task-specific virtual tokens.
      活动: GTC 24
      日期: March 2024
      级别: 高级技术
      行业: 所有行业
      NVIDIA 技术: DGX,EGX,NeMo
      话题: MLOps
      语言: 英语
      所在地: