Video Player is loading.
Current Time 0:00
Duration 0:00
Loaded: 0%
Stream Type LIVE
Remaining Time 0:00
 
1x
    • Chapters
    • descriptions off, selected
    • subtitles off, selected
      • Quality

      Turbocharge Cloud-Native AI Workloads With DPU-Accelerated Service Proxy for Kubernetes

      , VP and GM, Service Providers, F5
      , Sr. Director, AI Networking and Security Solutions, Ecosystem and Marketing, NVIDIA
      , Senior Network Architect and Senior Director, SoftBank Corp.
      AI clouds deploy AI applications at a massive scale, typically as containerized workloads accelerated for cloud-native environments. To meet the needs of LLMs, RAGs, and other AI applications, AI cloud infrastructure relies on high-bandwidth, low-latency interconnections between CPU, GPU, memory, and storage.

      AI cloud architecture is built using Kubernetes clusters. Kubernetes networking lets containerized applications communicate within the GPU cluster directly. Deployment of AI applications at scale needs high performance, efficient load balancing, application security, auto scaling, and multi-tenancy. An efficient proxy infrastructure service provides ingress/egress access to cluster resources without burdening the CPU and GPU.

      Learn how to meet AI cloud demands with a cloud-native, hardware-accelerated proxy. Offloading and isolating application delivery services to a DPU dramatically reduces CPU overhead and boosts infrastructure efficiency while minimizing attack surface.
      活动: GTC 25
      日期: March 2025
      行业: 所有行业
      NVIDIA 技术: BlueField DPU
      话题: Networking & Communications - AI Networking
      级别: Technical – Intermediate
      语言: 英语
      所在地: