As LLMs grow, efficient inference requires multi-node execution—introducing challenges in orchestration, scheduling, and low-latency GPU-to-GPU data transfers. Hardware like the GB200 NVL72 delivers massive scale-up compute, but truly scalable inference also depends on advanced software. Explore how open-source frameworks like NVIDIA Dynamo, combined with Azure’s AKS managed Kubernetes service, unlock new levels of performance and cost-efficiency.
talk-data.com
Topic
DynamoDB
database
nosql
aws
cloud
2
tagged
Activity Trend
12
peak/qtr
2020-Q1
2026-Q1
Top Events
AWS re:Invent 2024
21
O'Reilly Data Engineering Books
6
Databricks DATA + AI Summit 2023
3
O'Reilly Data Visualization Books
2
Microsoft Ignite 2025
2
Data Engineering Podcast
2
Grafana & Friends Berlin + AWS UG Berlin Collaborative Session - October
1
Data Unchained
1
dbt Coalesce 2022
1
O'Reilly Data Science Books
1
DataTopics: All Things Data, AI & Tech
1
Data + AI Summit 2025
1
Filtering by:
Microsoft Ignite 2025
×
Discover how Dynamo AI, a fast-growing AI security startup in financial services and government, uses Azure AI to confidently deploy generative AI at scale. Attendees will learn how Dynamo’s auditable guardrails, hallucination checks, and observability keep applications reliable and compliant, while gaining insights on building and deploying responsible AI solutions that balance innovation, trust, and speed.