Struggling to monitor the performance and health of your large language model (LLM) deployments on Google Kubernetes Engine (GKE)? This session unveils how the Google Cloud Observability suite provides a comprehensive solution for monitoring leading AI model servers like Ray, NVIDIA Triton, vLLM, TGI, and others. Learn how our one-click setup automatically configures dashboards, alerts, and critical metrics – including GPU and TPU utilization, latency, throughput, and error analysis – to enable faster troubleshooting and optimized performance. Discover how to gain complete visibility into your LLM infrastructure.
talk-data.com
J
Speaker
James Maffey
2
talks
Product Manager
Google Cloud
Filter by Event / Source
Talks & appearances
2 activities · Newest first
OpenTelemetry is causing a groundswell of change in the observability space. Learn about OpenTelemetry as the new telemetry standard, how we leverage it on Google Cloud and beyond, and how to get started ingesting telemetry data from anywhere.
Click the blue “Learn more” button above to tap into special offers designed to help you implement what you are learning at Google Cloud Next 25.