talk-data.com talk-data.com

Event

Data + AI Summit 2025

2025-06-09 – 2025-06-13 Databricks Summit Visit website ↗

Activities tracked

715

Sessions & talks

Showing 476–500 of 715 · Newest first

Search within this event →

Genie for Engineering: Optimizing HVAC Design and Operational Insights With Data and AI

2025-06-10
talk

In this session, we will explore how Genie, an AI-driven platform transformed HVAC operational insights by leveraging Databricks offerings like Apache Spark, Delta Lake and the Databricks Data Intelligence Platform.Key contributions: Real-time data processing: Lakeflow Declarative Pipelines and Apache Spark™ for efficient data ingestion and real-time analysis. Workflow orchestration: Databricks Data Intelligence Platform to orchestrate complex workflows and integrate various data sources and analytical tools. Field Data Integration: Incorporating real-time field data into design and algorithm development, enabling engineers to make informed adjustments and optimize performance. By analyzing real-time data from HVAC installations, Genie identified discrepancies between design specs and field performance, allowing engineers to optimize algorithms, reduce inefficiencies and improve customer satisfaction. Discover how Genie revolutionized HVAC management and apply to your projects.

Geo-Powering Insights: The Art of Spatial Data Integration and Visualization

Geo-Powering Insights: The Art of Spatial Data Integration and Visualization

2025-06-10 Watch
talk
Mathieu Pelletier (Databricks)

In this presentation, we will explore how to leverage Databricks' SQL engine to efficiently ingest and transform geospatial data. We'll demonstrate the seamless process of connecting to external systems such as ArcGIS to retrieve datasets, showcasing the platform's versatility in handling diverse data sources. We'll then delve into the power of Databricks Apps, illustrating how you can create custom geospatial dashboards using various frameworks like Streamlit and Flask, or any framework of your choice. This flexibility allows you to tailor your visualizations to your specific needs and preferences. Furthermore, we'll highlight the Databricks Lakehouse's integration capabilities with popular dashboarding tools such as Tableau and Power BI. This integration enables you to combine the robust data processing power of Databricks with the advanced visualization features of these specialized tools.

High-Throughput ML: Mastering Efficient Model Serving at Enterprise Scale

High-Throughput ML: Mastering Efficient Model Serving at Enterprise Scale

2025-06-10 Watch
talk
Mingyang Ge (Databricks) , Yucheng Qian (Databricks)

Ever wondered how industry leaders handle thousands of ML predictions per second? This session reveals the architecture behind high-performance model serving systems on Databricks. We'll explore how to build inference pipelines that efficiently scale to handle massive request volumes while maintaining low latency. You'll learn how to leverage Feature Store for consistent, low-latency feature lookups and implement auto-scaling strategies that optimize both performance and cost. Key takeaways: Determining optimal compute capacity using the QPS × model execution time formula Configuring Feature Store for high-throughput, low-latency feature retrieval Managing cold starts and scaling strategies for latency-sensitive applications Implementing monitoring systems that provide visibility into inference performance Whether you're serving recommender systems or real-time fraud detection models, you'll gain practical strategies for building enterprise-grade ML serving systems.

How HP Is Optimizing the 3D Printing Supply Chain Using Delta Sharing

How HP Is Optimizing the 3D Printing Supply Chain Using Delta Sharing

2025-06-10 Watch
talk

HP’s 3D Print division empowers manufacturers with telemetry data to optimize operations and streamline maintenance. Using Delta Sharing, Unity Catalog and AI/BI dashboards, HP provides a secure, scalable solution for data sharing and analytics. Delta Sharing D2O enables seamless data access, even for customers not on Databricks. Apigee masks private URLs, and Unity Catalog enhances security by managing data assets. Predictive maintenance with Mosaic AI boosts uptime by identifying issues early and alerting support teams. Custom dashboards and sample code let customers run analytics using any supported client, while Apigee simplifies access by abstracting complexity. Insights from A/BI dashboards help HP refines data strategy, aligning solutions with customer needs despite the complexity of diverse technologies, fragmented systems and customer-specific requirements. This fosters trust, drives innovation,and strengthens HP as a trusted partner for scalable, secure data solutions.

IQVIA's Analytics for Patient Support Services: Transforming Scalability, Performance and Governance

IQVIA's Analytics for Patient Support Services: Transforming Scalability, Performance and Governance

2025-06-10 Watch
talk
Dmytro Kobryn (Customertimes) , Sudha Ragothaman (IQVIA)

This presentation will explore the transformation of IQVIA's decade-old patient support platform through the implementation of Databricks Data Intelligence Platform. Facing scalability challenges, performance bottlenecks and rising costs, the existing platform required significant redesign to handle growing data volumes and complex analytics. Key issues included static metrics limiting workflow optimization, fragmented data governance and heightened compliance and security demands. By partnering with Customertimes (a Databricks Partner) and adopting Databricks' centralized, scalable analytics solution with enhanced self-service capabilities, IQVIA achieved improved query performance, cost efficiency and robust governance, ensuring operational effectiveness and regulatory compliance in an increasingly complex environment.

Managing Data and AI Security Risks With DASF 2.0 — and a Customer Story

Managing Data and AI Security Risks With DASF 2.0 — and a Customer Story

2025-06-10 Watch
talk
Arun Pamulapati (Databricks) , Joseph Raetano (US AI)

The Databricks Security team led a broad working group that significantly evolved the Databricks AI Security Framework (DASF) to its 2.0 version since its first release by closely collaborating with the top cyber security researchers at industry organizations such as OWASP, Gartner, NIST, HITRUST, FAIR Institute and several Fortune 100 companies to address the evolving risks and associated controls of AI systems in enterprises. Join us to to learn how The CLEVER GenAI pipeline, an AI-driven innovation in healthcare, processes over 1.5 million clinical notes daily to classify social determinants impacting veteran care while adhering to robust security measures like NIST 800-53 controls and by leveraging Databricks AI Security Framework. We will discuss robust AI security guidelines to help data and AI teams understand how to deploy their AI applications securely. This session will give a security framework for security teams, AI practitioners, data engineers and governance teams.

Moody's AI Screening Agent: Automating Compliance Decisions

Moody's AI Screening Agent: Automating Compliance Decisions

2025-06-10 Watch
talk
Nishant Gurunath (Moody's)

The AI Screening Agent automates Level 1 (L1) screening process, essential for Know Your Customer (KYC) and compliance due diligence during customer onboarding. This system aims to minimize false positives, significantly reducing human review time and costs. Beyond typical Retrieval-Augmented Generation (RAG) applications like summarization and chat-with-your-data (CWYD), the AI Screening Agent employs a ReAct architecture with intelligent tools, enabling it to perform complex compliance decision-making with human-like accuracy and greater consistency. In this talk, I will explore the screening agent architecture, demonstrating its ability to meet evolving client policies. I will discuss evaluation and configuration management using MLflow LLM-as-judge and Unity Catalog, and discuss challenges, such as, data fidelity and customization. This session underscores the transformative potential of AI agents in compliance workflows, emphasizing their adaptability, accuracy, and consistency.

Real-Time Market Insights — Powering Optiver’s Live Trading Dashboard with Databricks Apps and Dash

Real-Time Market Insights — Powering Optiver’s Live Trading Dashboard with Databricks Apps and Dash

2025-06-10 Watch
talk
Huy Nguyen (Optiver)

In the fast-paced world of trading, real-time insights are critical for making informed decisions. This presentation explores how Optiver, a leading high-frequency trading firm, harnesses Databricks apps to power its live trading dashboards. The technology enables traders to analyze market data, detect patterns and respond instantly. In this talk, we will showcase how our system leverages Databricks’ scalable infrastructures such as Structured Streaming to efficiently handle vast streams of financial data while ensuring low-latency performance. In addition, we will show how the integration of Databricks apps with Dash has empowered traders to rapidly develop and deploy custom dashboards, minimizing dependency on developers. Attendees will gain insights into our architecture, data processing techniques and lessons learned in integrating Databricks apps with Dash in order to drive rapid, data-driven trading decisions.

ServiceNow ‘Walks the Talk’ With Databricks: Revolutionizing Go-To-Market With AI

ServiceNow ‘Walks the Talk’ With Databricks: Revolutionizing Go-To-Market With AI

2025-06-10 Watch
talk
Mili Merchant (ServiceNow) , Amulya Gupta (ServiceNow)

At ServiceNow, we’re not just talking about AI innovation — we’re delivering it. By harnessing the power of Databricks, we’re reimagining Go-To-Market (GTM) strategies, seamlessly integrating AI at every stage of the deal journey — from identifying high-value leads to generating hyper-personalized outreach and pitch materials. In this session, learn how we’ve slashed data processing times by over 90%, reducing workflows from an entire day to just 30 minutes with Databricks. This unprecedented speed enables us to deploy AI-driven GTM initiatives faster, empowering our sellers with real-time insights that accelerate deal velocity and drive business growth. As Agentic AI becomes a game-changer in enterprise GTM, ServiceNow and Databricks are leading the charge — paving the way for a smarter, more efficient future in AI-powered sales.

Sponsored by: Accenture & Avanade | Enterprise Scaling and Value of Generative AI and Agentic AI

Sponsored by: Accenture & Avanade | Enterprise Scaling and Value of Generative AI and Agentic AI

2025-06-10 Watch
talk
Venkatesh Rao (Accenture (HQ))

In this talk, we will explore the transformative potential of Generative AI and Agentic AI in driving enterprise-scale innovation and delivering substantial business value. As organizations increasingly recognize the power of AI to move beyond automation towards true augmentation and intelligent decision-making, understanding the nuances of scaling these advanced AI paradigms becomes critical. We will delve into practical strategies for deploying, managing, and optimizing Agentic AI frameworks showcasing how autonomous, goal-directed AI systems can unlock new efficiencies, enhance customer experiences, and foster continuous innovation. Through real-world case studies and actionable insights, attendees will gain a comprehensive understanding of the key considerations to architect, implement, and measure the ROI of large-scale Generative and Agentic AI initiatives, positioning their enterprises for sustained growth and competitive advantage in the AI-first era.

Sponsored by: Deloitte | Advancing AI in Cybersecurity with Databricks & Deloitte: Data Management & Analytics

Sponsored by: Deloitte | Advancing AI in Cybersecurity with Databricks & Deloitte: Data Management & Analytics

2025-06-10 Watch
talk
Kieran Norton (Deloitte Consulting (HQ)) , Chris Knackstedt (Deloitte & Touche LLP)

Deloitte is observing a growing trend among cybersecurity organizations to develop big data management and analytics solutions beyond traditional Security Information and Event Management (SIEM) systems. Leveraging Databricks to extend these SIEM capabilities, Deloitte can help clients lower the cost of cyber data management while enabling scalable, cloud-native architectures. Deloitte helps clients design and implement cybersecurity data meshes, using Databricks as a foundational data lake platform to unify and govern security data at scale. Additionally, Deloitte extends clients’ cybersecurity capabilities by integrating advanced AI and machine learning solutions on Databricks, driving more proactive and automated cybersecurity solutions. Attendees will gain insight into how Deloitte is utilizing Databricks to manage enterprise cyber risks and deliver performant and innovative analytics and AI insights that traditional security tools and data platforms aren’t able to deliver.

Sponsored by: Prophecy | Reinventing Data Prep in the Age of AI: Build an Agent-Driven Pipeline in 7 Minutes

Sponsored by: Prophecy | Reinventing Data Prep in the Age of AI: Build an Agent-Driven Pipeline in 7 Minutes

2025-06-10 Watch
talk
Maciej Szpakowski (Prophecy) , Raj Bains (Prophecy)

Still coding data transformations by hand? Struggling with rigid, proprietary data prep tools? AI agents are flipping the script, reshaping data teams and delivering production-ready data preparation. Join this session to see how analysts, data scientists, and data engineers can build powerful, production-ready data pipelines simply by describing their intent in natural language. All in under 7 minutes. No complex UI or coding is required. Select datasets, join tables, apply filters, perform calculations - all just by chatting - and watch the pipeline materialize in real time, ready for deployment with documentation, testing, lineage, and versioning. Ready to leave slow, traditional data prep behind and be part of the next wave of innovation? You won’t want to miss this session.

SQL-First ETL: Building Easy, Efficient Data Pipelines With Lakeflow Declarative Pipelines

SQL-First ETL: Building Easy, Efficient Data Pipelines With Lakeflow Declarative Pipelines

2025-06-10 Watch
talk
Paul Lappas (Databricks) , Ritwik Yadav (Databricks) , Meixian Li (Databricks)

This session explores how SQL-based ETL can accelerate development, simplify maintenance and make data transformation more accessible to both engineers and analysts. We'll walk through how Databricks Lakeflow Declarative Pipelines and Databricks SQL warehouse support building production-grade pipelines using familiar SQL constructs.Topics include: Using streaming tables for real-time ingestion and processing Leveraging materialized views to deliver fast, pre-computed datasets Integrating with tools like dbt to manage batch and streaming workflows at scale By the end of the session, you’ll understand how SQL-first approaches can streamline ETL development and support both operational and analytical use cases.

Unifying Data Delivery: Using Databricks as Your Enterprise Serving Layer

Unifying Data Delivery: Using Databricks as Your Enterprise Serving Layer

2025-06-10 Watch
talk
Ivan Spiriev (The World Bank) , Ivan Donev (The World Bank)

This session will take you on our journey of integrating Databricks as the core serving layer in a large enterprise, demonstrating how you can build a unified data platform that meets diverse business needs. We will walk through the steps for constructing a central serving layer by leveraging Databricks’ SQL Warehouse to efficiently deliver data to analytics tools and downstream applications. To tackle low latency requirements, we’ll show you how to incorporate an interim scalable relational database layer that delivers sub-second performance for hot data scenarios. Additionally, we’ll explore how Delta Sharing enables secure and cost-effective data distribution beyond your organization, eliminating silos and unnecessary duplication for a truly end-to-end centralized solution. This session is perfect for data architects, engineers and decision-makers looking to unlock the full potential of Databricks as a centralized serving hub.

Unifying Human-Curated Data Ingestion and Real-Time Updates with Databricks Lakeflow Declarative Pipelines, Protobuf and BSR

2025-06-10
talk
Dwight Whitlock (Clinician Nexus)

Red Stapler is a streaming-native system on Databricks that merges file-based ingestion and real-time user edits into one Lakeflow Declarative Pipelines for near real-time feedback. Protobuf definitions, managed in the Buf Schema Registry (BSR), govern schema and data-quality rules, ensuring backward compatibility. All records — valid or not — are stored in an SCD Type 2 table, capturing every version for full history and immediate quarantine views of invalid data. This unified approach boosts data governance, simplifies auditing and streamlines error fixes.Running on Lakeflow Declarative Pipelines Serverless and the Kafka-compatible Bufstream keeps costs low by scaling down to zero when idle. Red Stapler’s configuration-driven Protobuf logic adapts easily to evolving survey definitions without risking production. The result is consistent validation, quick updates and a complete audit trail — all critical for trustworthy, flexible data pipelines.

Unity Catalog Upgrades Made Easy. Step-by-Step Guide for Databricks Labs UCX

Unity Catalog Upgrades Made Easy. Step-by-Step Guide for Databricks Labs UCX

2025-06-10 Watch
talk
Vuong ‎ (Databricks) , Liran Bareket (Databricks)

The Databricks labs project UCX aims to optimize the Unity Catalog (UC) upgrade process, ensuring a seamless transition for businesses. This session will delve into various aspects of the UCX project including the installation and configuration of UCX, the use of the UCX Assessment Dashboard to reduce upgrade risks and prepare effectively for a UC upgrade, and the automation of key components such as group, table and code migration. Attendees will gain comprehensive insights into leveraging UCX and Lakehouse Federation for a streamlined and efficient upgrade process. This session is aimed at customers new to UCX as well as veterans.

Unlocking Data Intelligence: A Beginner’s Guide to Unity Catalog

Unlocking Data Intelligence: A Beginner’s Guide to Unity Catalog

2025-06-10 Watch
talk
Chris Grabiel (Databricks) , Sachin Thakur (Databricks)

Getting started with data and AI governance in the modern data stack? Unity Catalog is your gateway to secure, discoverable and well-governed data and AI assets. In this session, we’ll break down what Unity Catalog is, why it matters and how it simplifies access control, lineage, discovery, auditing, business semantics and secure, open collaboration — all from a single place. We’ll explore how it enables open interoperability across formats, tools and platforms, helping you avoid lock-in and build on open standards. Most importantly, you’ll learn how Unity Catalog lays the foundation for data intelligence — by unifying governance across data and AI, enabling AI tuned to your business. It helps build a deep understanding of your data and delivers contextual, domain-specific insights that boost productivity for both technical and business users across any workload.

Using Catalogs for a Well-Governed and Efficient Data Ecosystem

Using Catalogs for a Well-Governed and Efficient Data Ecosystem

2025-06-10 Watch
talk
Kajal Woods (Capital One Financial) , jim Lebonitte (Capital One)

The ability to enforce data management controls at scale and reduce the effort required to manage data pipelines is critical to operating efficiently. Capital One has scaled its data management capabilities and invested in platforms to help address this need. In the past couple of years, the role of “the catalog” in a data platform architecture has transitioned from just providing SQL to providing a full suite of capabilities that can help solve this problem at scale. This talk will give insight into how Capital One is thinking about leveraging Databricks Unity Catalog to help tackle these challenges.

Break the Ice: Your Guide to the AccuWeather Data Suite in Databricks

Break the Ice: Your Guide to the AccuWeather Data Suite in Databricks

2025-06-10 Watch
lightning_talk
Crystal Camron (AccuWeather)

AccuWeather harnesses cutting-edge technology, industry-leading weather data, and expert insights to empower businesses and individuals worldwide. In this session, we will explore how AccuWeather’s comprehensive datasets—ranging from historical and current conditions to forecasts and climate normals—can drive real-world impact across diverse industries. By showcasing scenario-based examples, we’ll demonstrate how AccuWeather’s hourly and daily weather data can address the unique needs of your organization, whether for operational planning, risk management, or strategic decision-making. This session is ideal for both newcomers to AccuWeather’s offerings and experienced users seeking to unlock the full potential of our weather data to optimize performance, improve efficiency, and boost overall success.

Dealing With Sensitive Data on Databricks at Natura

Dealing With Sensitive Data on Databricks at Natura

2025-06-10 Watch
lightning_talk
Daniel Shimura (Natura)

Ensuring the protection of sensitive data within a Databricks environment requires robust mechanisms to prevent unauthorized access, even by high-privileged roles such as Databricks Administrators: Account Console Admins, Workspace Admins, and Unity Catalog Admins. To address this, a comprehensive data governance and access control strategy can be implemented, leveraging encryption, secret scope, column mask, fine-grained access on tables and auditing capabilities.

Delta-rs Turning Five: Growing Pains and Life Lessons

Delta-rs Turning Five: Growing Pains and Life Lessons

2025-06-10 Watch
lightning_talk
Robert Pack (Databricks)

Five years ago, the delta-rs project embarked on a journey to bring Delta Lake's robust capabilities to the Rust & Python ecosystem. In this talk, we'll delve into the triumphs, tribulations and lessons learned along the way. We'll explore how delta-rs has matured alongside the thriving Rust data ecosystem, adapting to its evolving landscape and overcoming the challenges of maintaining a complex data project. Join us as we share insights into the project's evolution, the symbiotic relationship between delta-rs and the Rust community, and the current hurdles and future directions that lie ahead. Audio for this session is delivered in the conference mobile app, you must bring your own headphones to listen.

Gaining Insight From Image Data in Databricks Using Multi-Modal Foundation Model API

Gaining Insight From Image Data in Databricks Using Multi-Modal Foundation Model API

2025-06-10 Watch
lightning_talk
Ankit Mathur (Databricks)

Unlock the hidden potential in your image data without specialized computer vision expertise! This session explores how to leverage Databricks' multi-modal Foundation Model APIs to analyze, classify and extract insights from visual content. Learn how Databricks provides a unified API to understand images using powerful foundation models within your data workflows. Key takeaways: Implementing efficient workflows for image data processing within your Databricks lakehouse Understanding multi-modal foundation models for image understanding Integrating image analysis with other data types for business insights Using OpenAI-compatible APIs to query multi-modal models Building end-to-end pipelines from image ingestion to model deployment Whether analyzing product images, processing visual documents or building content moderation systems, you'll discover how to extract valuable insights from your image data within the Databricks ecosystem.

Improving User Experience and Efficiency Using DBSQL

Improving User Experience and Efficiency Using DBSQL

2025-06-10 Watch
lightning_talk
Renato Suarez (PicPay) , Gustavo Tadao Okida (PicPay)

To scale Databricks SQL to 2,000 users efficiently and cost-effectively, we adopted serverless, ensuring dynamic scalability and resource optimization. During peak times, resources scale up automatically; during low demand, they scale down, preventing waste. Additionally, we implemented a strong content governance model. We created continuous monitoring to assess query and dashboard performance, notifying users about adjustments and ensuring only relevant content remains active. If a query exceeds time or impact limits, access is reviewed and, if necessary, deactivated. This approach brought greater efficiency, cost reduction and an improved user experience, keeping the platform well-organized and high-performing.

Powering Personalization at Scale with Data: How T-Mobile and Deep Sync Help Brands Connect with Consumers

Powering Personalization at Scale with Data: How T-Mobile and Deep Sync Help Brands Connect with Consumers

2025-06-10 Watch
lightning_talk
Jeff Frantz (T-Mobile) , Pieter De Temmerman (Deep Sync)

Discover how T-Mobile and Deep Sync are redefining personalized marketing through the power of Databricks. Deep Sync, a leader in deterministic identity solutions, has brought its identity spine to Databricks Lakehouse, which covers over 97% of U.S. households with the most current and accurate attribute data available. T-Mobile is bringing to market for the first time a new data services business that introduces privacy-compliant, consent-based consumer data. Together, T-Mobile and Deep Sync are transforming how brands engage with consumers—enabling bespoke, hyper-personalized workflows, identity-driven insights, and closed-loop measurement through Databricks’ Multi-Party Cleanrooms. Join this session to learn how data and identity are converging to solve today’s modern marketing challenges so consumers can rediscover what it feels like to be seen, not targeted

Revolutionizing Counterparty Credit Risk (SACCR) – How Morgan Stanley Scaled With Databricks

Revolutionizing Counterparty Credit Risk (SACCR) – How Morgan Stanley Scaled With Databricks

2025-06-10 Watch
lightning_talk
Naeem Rehman (Databricks) , Alistair MacDonald (Morgan Stanley)

Learn how Morgan Stanley scaled one of their most significant regulatory calculators (SACCR) by leveraging Databricks for horizontal and vertical scaling. Discover how we harnessed Databricks to improve performance, improve calculation accuracy, regulatory compliance and more.