Video Player is loading.
Current Time 0:00
Duration 0:00
Loaded: 0%
Stream Type LIVE
Remaining Time 0:00
 
1x
    • Chapters
    • descriptions off, selected
    • subtitles off, selected
      • Quality

      Distributed Agentic Multi-Modal LLM Deployments

      , Product CTO, Dell Technologies
      , Global Head of Edge Computing Sales, NVIDIA
      Future enterprise architectures for edge AI inference will be multi-modal. With core foundation model (FM)/LLM functions centralized (attention, MLP-FFNN, layer normalization, residual connection) and some functions distributed to edge (retrieval-augmented generation, input processing-tokenization/vector embedding/position encoding [maybe], guardrails, caching, inference optimization-sampling/beam search). In addition, in multi-modal architectures, convolutional neural networks (CNNs) will be integrated with the FM/LLM.


      活动: GTC 25
      日期: March 2025
      行业: 所有行业
      话题: Edge Computing - Edge Management & Orchestration
      NVIDIA 技术: NVIDIA NIM,NVIDIA AI Enterprise
      级别: 技术 - 高级
      语言: 英语
      所在地: