Scale your AI training and achieve peak performance with AI Hypercomputer. Gain actionable insights into optimizing your AI workloads for maximum goodput. Learn how to leverage our robust infrastructure for diverse models, including dense, Mixture of Experts, and diffusion. Discover how to customize your workflows with custom kernels and developer tools, facilitating seamless interactive development. You'll learn firsthand how Pathways, developed by Google Deepmind, enables large scale training resiliency, flexibility to express architecture.
talk-data.com
Speaker
Kirat Pandya
4
talks
Frequent Collaborators
Filter by Event / Source
Talks & appearances
4 activities · Newest first
Learn how to run high-throughput and low-latency inference on Google Cloud to maximize price-performance on TPUs and GPUs, leveraging JetStream and vLLM.
Deploying AI models at scale demands high-performance inference capabilities. Google Cloud offers a range of cloud tensor processing units (TPUs) and NVIDIA-powered graphics processing unit (GPU) VMs. This session will guide you through the key considerations for choosing TPUs and GPUs for your inference needs. Explore the strengths of each accelerator for various workloads like large language models and generative AI models. Discover how to deploy and optimize your inference pipeline on Google Cloud using TPUs or GPUs. Understand the cost implications and explore cost-optimization strategies.
Click the blue “Learn more” button above to tap into special offers designed to help you implement what you are learning at Google Cloud Next 25.
Learn how to run high-throughput and low-latency inference on Google Cloud to maximize price-performance on TPUs and GPUs, leveraging JetStream and vLLM.