As LLMs grow, efficient inference requires multi-node execution—introducing challenges in orchestration, scheduling, and low-latency GPU-to-GPU data transfers. Hardware like the GB200 NVL72 delivers massive scale-up compute, but truly scalable inference also depends on advanced software. Explore how open-source frameworks like NVIDIA Dynamo, combined with Azure’s AKS managed Kubernetes service, unlock new levels of performance and cost-efficiency.
talk-data.com
A
Speaker
Anish Maddipoti
1
talks
Anish Maddipoti is a product manager at NVIDIA. He currently works on building AI/ML frameworks, such as NVIDIA Dynamo and NVIDIA Grove. Previously, he was a founding team member of Brev.dev (acquired by NVIDIA) and co-founder of Agora Labs. He studied in the Plan II program at the University of Texas at Austin.
Bio from: Microsoft Ignite 2025
Filtering by:
Microsoft Ignite 2025
×
Filter by Event / Source
Talks & appearances
Showing 1 of 1 activities