Video Player is loading.
Current Time 0:00
Duration 0:00
Loaded: 0%
Stream Type LIVE
Remaining Time 0:00
 
1x
    • Chapters
    • descriptions off, selected
    • subtitles off, selected
      • Quality

      Optimizing & Deploying PyTorch Models for High-Performance Inference

      , Product Manager for Deep Learning Inference - TensorRT, NVIDIA
      , Staff Software Engineer, Meta
      , Software Engineer, Meta
      Learn about optimizing & deploying dynamic PyTorch models in Python for production. We’ll cover the new `torch.package` and `torch::deploy` interfaces as well as tools for extracting performance out of models like compression toolkits, torch.fx, and more. Then we’ll give the latest updates on Torch-TensorRT for maximizing performance on GPUs, including a technical deep dive into how torch.fx is being applied to go directly from a PyTorch model to TensorRT, entirely in Python. Participants will understand the software stack for PyTorch provides uncompromised flexibility, usability, and performance today on NVIDIA GPUs, and future plans for delivering.
      活动: GTC Digital Spring
      日期: March 2022
      行业: 所有行业
      级别: 初级技术
      话题: Deep Learning - Inference
      语言: 英语
      所在地: