Facing challenges with the cost and performance of your AI inference workloads? This talk presents TPUs and Google Kubernetes Engine (GKE) as a solution for achieving both high throughput and low latency while optimizing costs with open source models and libraries. Learn how to leverage TPUs to scale massive inference workloads efficiently.
talk-data.com
J
Speaker
Jon Li
1
talks
Software Engineer
Google Cloud
Filter by Event / Source
Talks & appearances
1 activities · Newest first