Video Player is loading.
Current Time 0:00
Duration 0:00
Loaded: 0%
Stream Type LIVE
Remaining Time 0:00
 
1x
    • Chapters
    • descriptions off, selected
    • subtitles off, selected
      • Quality

      Customizing Foundation Large Language Models in Diverse Languages With NVIDIA NeMo

      , Senior Solutions Architect - GenAI&Inference, NVIDIA
      , Solutions Architect, NVIDIA
      , Sr. Deep Learning Data Scientist, NVIDIA
      , Sr. Deep Learning Data Scientist, NVIDIA
      , Solution Architect, NVIDIA
      We'll focus on customizing foundation large language models (LLMs) for languages other than English. We'll go through techniques like prompt-engineering, prompt-tuning, parameter-efficient fine-tuning, and supervised instruction fine-tuning (SFT), enabling LLMs to adapt to diverse use cases. We'll showcase some of these techniques using NVIDIA NeMo Framework for both NVIDIA Foundation Models and other community models, such as Llama-2. Finally, we'll demonstrate how to efficiently deploy the customized models using NVIDIA TensorRT-LLM and NVIDIA Triton Inference Server.
      活动: GTC 24
      日期: March 2024
      行业: 所有行业
      级别: 中级技术
      NVIDIA 技术: NeMo,TensorRT,Triton
      话题: Text Generation
      语言: 英语
      所在地: