talk-data.com talk-data.com

Topic

Databricks

big_data analytics spark

1286

tagged

Activity Trend

515 peak/qtr
2020-Q1 2026-Q1

Activities

1286 activities · Newest first

Accelerating Growth in Capital Markets: Data-Driven Strategies for Success

Growth in capital markets thrives on innovation, agility and real-time insights. This session highlights how leading firms use Databricks’ Data Intelligence Platform to uncover opportunities, optimize trading strategies and deliver personalized client experiences. Learn how advanced analytics and AI help organizations expand their reach, improve decision-making and unlock new revenue streams. Industry leaders share how unified data platforms break down silos, deepen insights and drive success in a fast-changing market. Key takeaways: Predictive analytics and machine learning strategies for growth Real-world examples of optimized trading and enhanced client engagement Tools to innovate while ensuring operational efficiency Discover how data intelligence empowers capital markets firms to thrive in today’s competitive landscape!

AI-Powered Data Discovery and Curation With Unity Catalog

This session is repeated. In today’s data landscape, the challenge isn’t just storing or processing data — it’s enabling every user, from data stewards to analysts, to find and trust the right data, fast. This session explores how Databricks is reimagining data discovery with the new Discover Page Experience — an intuitive, curated interface showcasing key data and workspace assets. We’ll dive into AI-assisted governance and AI-powered discovery features like AI-generated metadata, AI-assisted lineage and natural language data exploration in Unity Catalog. Plus, see how new certifications and deprecations bring clarity to complex data environments. Whether you’re a data steward highlighting trusted assets or an analyst navigating data without deep schema knowledge, this session will show how Databricks is making data discovery seamless for everyone.

Cutting Costs, Not Performance: Optimizing Databricks at Scale

As Databricks transforms data processing, analytics and machine learning, managing platform costs has become crucial for organizations aiming to maximize value while staying within budget. While Databricks offers unmatched scalability and performance, inefficient usage can lead to unexpected cost overruns. This presentation will explore common challenges organizations face in controlling Databricks costs and provide actionable best practices for optimizing resource allocation, preventing over-provisioning and eliminating underutilization. Drawing from NTT DATA’s experience, I'll share how we reduced Databricks costs by up to 50% through strategies like choosing the right compute resource, leveraging manage tables and using Unity Catalog features, such as system tables, to monitor consumption. Join this session to gain practical insights and tools that will empower your team to optimize Databricks without overspending.

Databricks AI Factory Transforming Seven West Media

The implementation of the Databricks AI Factory enabled Seven West Media to transform its business by accelerating the launch of AI-driven use cases, fostering innovation and reducing time to market. By leveraging a unified data and AI platform, the company achieved better ROI through optimized workflows, improved operational efficiency and scalable machine learning models. The AI Factory empowered data teams to experiment faster, unlocking deeper audience insights that enhanced engagement and content personalization. This transformation positioned Seven West Media as a leader in AI-driven media, driving measurable business impact and future-proofing its data strategy.

Databricks Lakeflow: the Foundation of Data + AI Innovation for Your Industry

Every analytics, BI and AI project relies on high-quality data. This is why data engineering, the practice of building reliable data pipelines that ingest and transform data, is consequential to the success of these projects. In this session, we'll show how you can use Lakeflow to accelerate innovation in multiple parts of the organization. We'll review real-world examples of Databricks customers using Lakeflow in different industries such as automotive, healthcare and retail. We'll touch on how the foundational data engineering capabilities Lakeflow provides help power initiatives that improve customer experiences, make real-time decisions and drive business results.

From Days to Minutes - AI Transforms Audit at KPMG

Imagine performing complex regulatory checks in minutes instead of days. We made this a reality using GenAI on the Databricks Data Intelligence Platform. Join us for a deep dive into our journey from POC to a production-ready AI audit tool. Discover how we automated thousands of legal requirement checks in annual reports with remarkable speed and accuracy. Learn our blueprint for: High-Performance AI: Building a scalable, >90% accurate AI system with an optimized RAG pipeline that auditors praise. Robust Productionization: Achieving secure, governed deployment using Unity Catalog, MLflow, LLM-based evaluation, and MLOps best practices. This session provides actionable insights for deploying impactful, compliant GenAI in the enterprise.

This session is repeated. This introductory workshop caters to data engineers seeking hands-on experience and data architects looking to deepen their knowledge. The workshop is structured to provide a solid understanding of the following data engineering and streaming concepts: Introduction to Lakeflow and the Data Intelligence Platform Getting started with Lakeflow Declarative Pipelines for declarative data pipelines in SQL using Streaming Tables and Materialized Views Mastering Databricks Workflows with advanced control flow and triggers Understanding serverless compute Data governance and lineage with Unity Catalog Generative AI for Data Engineers: Genie and Databricks Assistant We believe you can only become an expert if you work on real problems and gain hands-on experience. Therefore, we will equip you with your own lab environment in this workshop and guide you through practical exercises like using GitHub, ingesting data from various sources, creating batch and streaming data pipelines, and more.

Want to learn how to build your own custom data intelligence applications directly in Databricks? In this workshop, we’ll guide you through a hands-on tutorial for building a Streamlit web app that leverages many of the key products at Databricks as building blocks. You’ll integrate a live DB SQL warehouse, use Genie to ask questions in natural language, and embed AI/BI dashboards for interactive visualizations. In addition, we’ll discuss key concepts and best practices for building production-ready apps, including logging and observability, scalability, different authorization models, and deployment. By the end, you'll have a working AI app—and the skills to build more.

Health Data, Delivered: How Lakeflow Declarative Pipelines Powers the HealthVerity Marketplace

Building scalable, reliable ETL pipelines is a challenge for organizations managing large, diverse data sources. Theseus, our custom ETL framework, streamlines data ingestion and transformation by fully leveraging Databricks-native capabilities, including Lakeflow Declarative Pipelines, auto loader and event-driven orchestration. By decoupling supplier logic and implementing structured bronze, silver, and gold layers, Theseus ensures high-performance, fault-tolerant data processing with minimal operational overhead. The result? Faster time-to-value, simplified governance and improved data quality — all within a declarative framework that reduces engineering effort. In this session, we’ll explore how Theseus automates complex data workflows, optimizes cost efficiency and enhances scalability, showcasing how Databricks-native tools drive real business outcomes.

How We Transformed Two Businesses With Databricks as the Cornerstone

In this talk, we will discuss the lessons learned and future vision of transforming two business units to a modern financial data platform at Nasdaq. We'll highlight the transition from disjointed systems to a unified platform using Databricks. Our target audience includes financial engineers, data architects and technical leaders. The agenda covers challenges of legacy systems, reasons for choosing Databricks and key architectural decisions.

Igniting Innovation at Gilead: Convergence of Cloud, Data, AI and Agents

The convergence of cloud, data and AI is revolutionizing the pharmaceutical industry, creating a powerful ecosystem that drives innovation at scale across the entire value chain. At Gilead, teams harness these technologies on a unified cloud, data, & AI platform, accelerating business processes in pre-clinical and clinical stage, enabling smarter manufacturing and commercial processes, and deliver AI initiatives by reusing data products. Gilead will discuss how they have leveraged AWS, Databricks, and Data Mesh to manage vast amounts of heterogeneous data. Also, showcase use cases of traditional AI/ML, and Generative AI, and a Marketplace approach to drive adoption of AI Agents, demonstrating how this cloud-based, AI-powered platform is transforming the entire value chain. Gilead will also discuss how they are exploring the future of pharmaceutical innovation through Agentic AI, where the synergy of cloud, data and AI is unlocking new possibilities for a healthier world. In the second part, Muddu Sudhakar, Founder and Investor, will discuss how organizations can build and buy solutions for AI, Agents with Data Platforms. AWS and Databricks provide industry-leading platforms to build Agentic AI solutions. We will also cover Agentic AI Platform, Agent orchestration, Agent Interoperability, Agent Guardrails and Agentic workflows. This discussion also covers challenges in deploying and managing Agentic AI platforms. Enterprises need impactful AI initiatives & Agents to realize the promise and vision of AI and drive significant ROI.

Introduction to Unity Catalog Metrics: Define Your Business Metrics Once, Trust Everywhere

Today’s organizations need faster, more reliable insights — but metric sprawl and inconsistent KPIs make that difficult. In this session, you’ll learn how Unity Catalog Metrics helps unify business semantics across your organization. Define your KPIs once, apply enterprise-grade governance with fine-grained access controls, auditing and lineage, and use them across any Databricks tool — from AI/BI Dashboards and Genie to notebooks and Lakeflow. You’ll learn how to eliminate metric chaos by centrally defining and governing metrics with Unity Catalog. You’ll walk away with strategies to boost trust through built-in governance and empower every team — regardless of technical skill — to work from the same certified metrics.

Intro to the Mosaic AI Platform: Building Data Intelligence Into Your AI Solutions

Take a front-row seat for a comprehensive, high-level introduction to Mosaic AI through the lens of Data Intelligence. In this session, we’ll spotlight the Databricks Platform’s newest features and announcements, showcase how Mosaic AI transforms raw enterprise data into actionable insights and share real-world examples of success. Whether you’re beginning your AI journey or scaling your existing efforts, this talk will provide you with the foundational knowledge and inspiration to fully leverage Mosaic AI for Data Intelligence and next-generation GenAI solutions.

IQVIA’s Serverless Journey: Enabling Data and AI in a Regulated World

Your data and AI use-cases are multiplying. At the same time, there is increased focus and scrutiny to meet sophisticated security and regulatory requirements. IQVIA utilizes serverless use-cases across data engineering, data analytics, and ML and AI, to empower their customers to make informed decisions, support their R&D processes and improve patient outcomes. By leveraging native controls on the platform, serverless enables them to streamline their use cases while maintaining a strong security posture, top performance and optimized costs. This session will go over IQVIA’s journey to serverless, how they met their security and regulatory requirements, and the latest and upcoming enhancements to the Databricks Platform.

As cheat developers evolve, so must detection techniques. This session will explore our methodologies, challenges and future directions, demonstrating how machine learning is transforming anti-cheat strategies and preserving competitive integrity in online gaming and how Databricks is enabling us to do so. As online gaming grows, maintaining fair play is an ongoing challenge. Call of Duty, a highly competitive first-person action game, faces aimbot usage—cheats that enable near-perfect accuracy, undermining fair play. Additionally, traditional detection methods are increasingly becoming less effective against advanced cheats that mimic human behavior. Machine learning presents a scalable and adaptive solution to this. We developed a data pipeline that collects features such as angle velocity, acceleration, etc. to train a deep neural network and deployed it. We are processing 30 million rows of data per hour for this detection on Databricks Platform.

MLflow 3.0: AI and MLOps on Databricks

Ready to streamline your ML lifecycle? Join us to explore MLflow 3.0 on Databricks, where we'll show you how to manage everything from experimentation to production with less effort and better results. See how this powerful platform provides comprehensive tracking, evaluation, and deployment capabilities for traditional ML models and cutting-edge generative AI applications. Key takeaways: Track experiments automatically to compare model performance Monitor models throughout their lifecycle across environments Manage deployments with robust versioning and governance Implement proven MLOps workflows across development stages Build and deploy generative AI applications at scale Whether you're an MLOps novice or veteran, you'll walk away with practical techniques to accelerate your ML development and deployment.

Raising the Stakes: Enhancing Player Experience using ML/AI

At Second Dinner, delivering fast, personalized gameplay experiences is key to player engagement. In this session, Justin Wu shares how the team implemented real-time feature serving using Databricks to power responsive, data-driven game mechanics at scale. He’ll dive into the architecture, technical decisions, and trade-offs behind their solution—highlighting how they balance performance, scalability, and cost. Whether you're building live features or rethinking your game data stack, this session offers practical insights to accelerate your journey.

Sponsored by: Immuta | Protecting People Data: How Shell Empowers HR to Drive a Brighter Future

HR departments increasingly rely on data to improve workforce planning and experiences. However, managing and getting value from this data can be challenging, especially given the complex technology landscape and the need to ensure data security and compliance. Shell has placed a high priority on safeguarding its people data while empowering its HR department with the tools and access they need to make informed decisions. This session will explore the transformation of Shell's Central Data Platform, starting with their HR use case. You’ll hear about:- The role of automation and data governance, quality, and literacy in Shell’s strategy.- Why they chose Databricks and Immuta for enhanced policy-based access control.- The future for Shell and their vision for a data marketplace to truly embrace a culture of global data sharing.The result? A robust, scalable HR Data Platform that is securely driving a brighter future for Shell and its employees.

Tech Industry Session: Building Collaborative Ecosystems With Openness and Portability

Join us to discover how leading tech companies accelerate growth using open ecosystems and built-on solutions to foster collaboration, accelerate innovation and create scalable data products. This session will explore how organizations use Databricks to securely share data, integrate with partners and enable teams to build impactful applications powered by AI and analytics. Topics include: Using Delta Sharing for secure, real-time data collaboration across teams and partners Embedding analytics and creating marketplaces to extend product capabilities Building with open standards and governance frameworks to ensure compliance without sacrificing agility Hear real-world examples of how open ecosystems empower organizations to widen the aperture on collaboration, driving better business outcomes. Walk away with insights into how open data sharing and built-on solutions can help your teams innovate faster at scale.

Telecom Innovation Exchange: Demos and Dialogues

Join us for an interactive breakout session designed to explore scalable, real-world solutions powered by Partners with Databricks. In this high-energy session, you'll hear from three of our leading partners — Accenture, Capgemini and Wipro — as they each deliver rapid-fire, 5-minute demos of their most impactful, production-grade solutions built for the telecom industry. From network intelligence to customer experience to AI-driven automation, these solutions are already driving tangible outcomes at scale. After the demos, you’ll have the unique opportunity to engage directly with each partner in a “speed dating” style format. Dive deep into the solutions, ask your questions and explore how these approaches can be tailored to your organization’s needs. Whether you're solving for churn, fraud, network ops or enterprise AI use cases, this session is your chance to connect, collaborate and walk away with practical ideas you can take back to your teams.