Hands-On Lab (Part 2): Containerize and deploy your agent to Cloud Run with GPU acceleration, focusing on production configurations and live testing.
talk-data.com
Topic
gpu
3
tagged
Activity Trend
2
peak/qtr
2020-Q1
2026-Q1
Hands-On Lab (Part 1): Secure and deploy a Model Context Protocol (MCP) server and start integrating the Agent Development Kit (ADK) to empower an intelligent agent.
Running models locally on the CPU and possibly a GPU means we can experiment with the latest quantised models on real client data without anything leaving the machine. We can explore text question answering, image analysis and calling these tools via a Python API for rapid PoC experimentation. This quickly exposes the ways that LLMs go weird and maybe that helps us avoid some of the examples of early LLM deployments making embarrassing mistakes!