Deploying ML models doesn’t have to mean spinning up servers and writing backend code. This talk shows how to run machine learning inference directly in the browser—using ONNX and WebAssembly—to go from prototype to interactive demo in minutes, not weeks.