talk-data.com talk-data.com

Google Cloud Next session 2025-04-11 at 17:45

Architecting for energy-efficient AI Inference

Topics

Description

The growing demand for AI inference raises crucial questions about sustainability. For developers, that means looking for energy efficiency everywhere. This session will explore deployment models (PaaS vs IaaS), capabilities you can optimize under the hood (architectural best practices and measurement), and data center wide optimizations Google makes and manages (liquid cooling, custom hardware) to be considered in the context of an AI inference application.