Step through training and deploying a custom reasoning model with Azure ML and Microsoft Foundry. See the magic happen in real time and learn how to bring it to your own projects with Foundry.
Speaker
3
talks
Wei Wei is currently a Managing Director and the Global Lead of AI Platform and Research at Accenture. He leads a global team of in research, engineering, and delivery to build and scale the AI refinery platform, the industry’s first delivery-centric platform for Agentic AI development. .
Bio from: Microsoft Ignite 2025
Talks & appearances
3 activities · Newest first
Step through training and deploying a custom reasoning model with Azure ML and Microsoft Foundry. See the magic happen in real time and learn how to bring it to your own projects with Foundry.
Unleash the full potential of large language models (LLMs) on your edge devices, even when there’s spotty internet. This session explores a hybrid approach that combines the power of cloud-based LLMs with the efficiency of on-device models. Learn how to intelligently route queries, enabling laptops and mobile phones to perform complex tasks while maintaining snappy performance. View demos of efficient task routing that optimizes for quality and cost to ensure your apps run smoothly, even during network disruptions.
Learn how to fast train your LLMs by maximizing utilization of TPUs. In this talk, we are going to start with training a GPT2 model with JAX from scratch on Colab and Kaggle TPUs for free. To scale it up, we will transition to the latest TPUs and the cutting edge MaxText library to train the state-of-the-art LLMs.