talk-data.com talk-data.com

Filter by Source

Select conferences and events

People (475 results)

See all 475 →

Companies (1 result)

Databricks 34 speakers
AI/ML Product Specialist Developer Advocate Director, Manager Product Management
Showing 10 results

Activities & events

Title & Speakers Event

Want to grow your organization with Geospatial insights through Databricks?

Learn the fundamentals to successfully work with Geospatial data in Databricks - in our one-day training!

Besides the basics, we will cover topics such as when to migrate to Databricks, and the interoperability between FME, ArcGIS, and Databricks. You will gain hands-on experience with location allocation, sky view factor, and so much more.

More information can be found here: https://revodata.nl/databricks-geospatial/

Please note this is a paid event, tickets and registration are required. As this is a hands-on training with interactive labs, we have space for max 15 participants.

Geospatial data in Databricks

Most organizations capture huge volumes of spatial data, including addresses, coordinates, routes, and catchments, but struggle to operationalize it at scale. Traditional GIS (Geographic Information Systems) tools are powerful but isolated; unlocking value requires integrating spatial analytics directly within your data platform. In this session, we’ll cover: - Geospatial fundamentals on Databricks: understanding geometry vs. geography, coordinate systems, and H3 grids. - Scaling challenges: Combining spatial and business data, processing millions of coordinates efficiently, and maintaining real-time freshness. - Databricks capabilities: How Spatial SQL, Lakeflow, and Unity Catalog enable native spatial processing, federated access, and governed sharing across teams. - Applied use cases: From network optimisation to asset tracking and location-based insights across industries. We'll finish with a live demo, see how raw coordinates become actionable intelligence within the Lakehouse. Why Attend: - Learn how to bring geospatial analytics natively into Databricks. - Discover best practices for scaling spatial workloads efficiently. - Understand how Unity Catalog underpins governance and reusability. - See real-world examples and a live demo in action. Join us to learn how Databricks unifies spatial and analytical workloads, delivering governed, high-performance geospatial insight at enterprise scale. This session will be delivered by Unifeye's CDO and Databricks Champion Bianca Stratulat, and Senior Data Engineers, Jordan Begg and Hasnat Abdul

Geospatial Data on Databricks
Chris Crawford – Sr. Solutions Archtect @ Databricks , Hobson Bryan – Associate Director of Technology @ Global Water Security Center

The Global Water Security Center translates environmental science into actionable insights for the U.S. Department of Defense. Prior to incorporating Databricks, responding to these requests required querying approximately five hundred thousand raster files representing over five hundred billion points. By leveraging lakehouse architecture, Databricks Auto Loader, Spark Streaming, Databricks Spatial SQL, H3 geospatial indexing and Databricks Liquid Clustering, we were able to drastically reduce our “time to analysis” from multiple business days to a matter of seconds. Now, our data scientists execute queries on pre-computed tables in Databricks, resulting in a “time to analysis” that is 99% faster, giving our teams more time for deeper analysis of the data. Additionally, we’ve incorporated Databricks Workflows, Databricks Asset Bundles, Git and Git Actions to support CI/CD across workspaces. We completed this work in close partnership with Databricks.

CI/CD Data Lakehouse Databricks Git Cyber Security Spark SQL Data Streaming
Data + AI Summit 2025

🎙️ Speaker: William Dean\, Corey Abshire\, Thomas Wiecki \| ⏰ Time: 5 PM UTC / 9 AM PT / 12 PM ET / 6 PM Berlin

In this event, we will discuss how customers can use Databricks to develop and productionize MMM models for their companies. By combining Databricks capabilities in consolidating, organizing and securing data pipelines and manage ML models and pipelines with PyMC-Marketing’s easy to use modelling capabilities, companies can bring develop sophisticated MMM models to help understand, optimize and forecast their marketing budgets.

📜 Outline of Talk / Agenda:

  • 5 min: Intro to PyMC Labs and speakers
  • 45 min: Presentation, panel discussion
  • 10 min: Q&A

💼 About the speaker:

  1. Corey Abshire, Senior AI Specialist Solutions Architect, DataBricks Corey Abshire is a Senior Specialist Solutions Architect focused on GenAI and ML in Communications, Media and Entertainment. Prior to Databricks, Corey was a Principal Machine Learning Engineer at Cummins Inc., working on critical data science and machine learning initiatives for quality, engineering, and manufacturing. He holds an M.S. in data science from Indiana University in Bloomington, IN.

🔗 Connect with Corey: 👉 Linkedin: https://www.linkedin.com/in/coreyabshire/

  1. Will Dean, Principal Data Scientist, PyMC Labs Will Dean is a Statistician and Data Scientist with experience in geospatial and user analytics. He is passionate about Bayesian methods and using data visualization to tell a story. He is interested in software design and how it can make data problems easier and more enjoyable to solve.

🔗 Connect with Will Dean: 👉 Linkedin: https://www.linkedin.com/in/williambdean/ 👉 Github: https://github.com/wd60622/

💼 About the Host:

  1. Thomas Wiecki (Founder of PyMC Labs) Dr. Thomas Wiecki is an author of PyMC, the leading platform for statistical data science. To help businesses solve some of their trickiest data science problems, he assembled a world-class team of Bayesian modelers and founded PyMC Labs -- the Bayesian consultancy. He did his PhD at Brown University studying cognitive neuroscience. 🔗 Connect with Thomas: 👉 Linkedin: https://www.linkedin.com/in/twiecki/ 👉 Website: https://www.pymc-labs.com/, https://twiecki.io/ 👉 GitHub: https://github.com/twiecki 👉 Twitter: https://twitter.com/twiecki

📖 Code of Conduct: Please note that participants are expected to abide by PyMC's Code of Conduct.

🔗 Connecting with PyMC Labs: 🌐 Website: https://www.pymc-labs.com/ 👥 LinkedIn: https://www.linkedin.com/company/pymc-labs/ 🐦 Twitter: https://twitter.com/pymc_labs 🎥 YouTube: https://www.youtube.com/c/PyMCLabs 🤝 Meetup: https://www.meetup.com/pymc-labs-online-meetup/

[Online] MMM with PyMC Marketing and Databricks

Welcome to the London Databricks Meetup May 2024 Gen AI Edition! Join us for an exciting evening of Gen AI insights, expert sessions, and networking.

Event Schedule: 18:00 - 18:15: Arrival & Networking Mingle with fellow data enthusiasts and professionals. 18:15 - 18:30: Opening Remarks & Introductions Get to know your fellow attendees and the organizers. 18:30 - 19:00: How we built DBRX - an open high-quality LLM -- Jonathan Frankle -- Chief Scientist\, Neural Networks @ Databricks and Co-Founder of MosaicML

We are excited to introduce DBRX, an open, leading, general-purpose LLM created by Databricks. It provides the open community and enterprises building their own LLMs with capabilities previously limited to closed model APIs. According to our measurements, it surpasses GPT-3.5 and is competitive with Gemini 1.0 Pro. It is an especially capable code model, surpassing specialized models like CodeLLaMA-70B on programming, in addition to its strength as a general-purpose LLM.

19:00 - 19:30: Streamlining Large Language Model (LLM) Operations with MLflow: A Comprehensive Overview -- Sepideh Ebrahimi -- Senior Specialist Solution Architect @ Databricks 19:30 - 20:00: Geospatial Data within Databricks -- Luke Menzies -- Principal AI Consultant @ Advancing Analytics 20:00 onwards: Pizza, Drinks & Networking

Enjoy some delicious pizza and beverages while networking with peers.

Date and Location: Date: 2nd May Venue: IDEAL London Join us for a fantastic evening of learning and networking at the London Databricks meetup!

London Databricks Meetup May 2024

Geospatial data analysis is critical to understanding the impact of agricultural operations on environmental sustainability with respect to water quality, soil health, greenhouse gasses, and more. Outside of a few specialized software products, however, support for spatial data types is often limited or missing from analytics and visualization platforms. In this session, we show how Truterra is using Databricks, Apache Sedona, and R to analyze spatial data at scale. Additionally, learn how Truterra uses spatial insights to educate and promote practices that optimize profitability, sustainability, and stewardship outcomes at the farm.

In this session, you will see how Databricks and Apache Sedona are used to process large spatial datasets including field, watershed, and hydrologic boundaries. You will see dynamic widgets, SQL and R used in tandem to generate map visuals, display them, and enable download all from a Databricks notebook.

Talk by: Nara Khou and Cort Lunke

Connect with us: Website: https://databricks.com Twitter: https://twitter.com/databricks LinkedIn: https://www.linkedin.com/company/databricks Instagram: https://www.instagram.com/databricksinc Facebook: https://www.facebook.com/databricksinc

Analytics Databricks SQL
Aayush Patel – Data & Platform Engineer @ SkyWatch

SkyWatch is on a mission to democratize earth observation data and make it simple for anyone to use.

In this session, you will learn about how SkyWatch aggregates demand signals for the EO market and turns them into monetizable recommendations for satellite operators. Skywatch’s Data & Platform Engineer, Aayush will share how the team built a serverless architecture that synthesizes customer requests for satellite images and identifies geographic locations with high demand, helping satellite operators maximize revenue and satisfying a broad range of EO data hungry consumers.

This session will cover:

  • Challenges with Fulfillment in Earth Observation ecosystem
  • Processing large scale GeoSpatial Data with Databricks
  • Databricks in-built H3 functions
  • Delta Lake to efficiently store data leveraging optimization techniques like Z-Ordering
  • Data LakeHouse Architecture with Serverless SQL Endpoints and AWS Step Functions
  • Building Tasking Recommendations for Satellite Operators

Talk by: Aayush Patel

Connect with us: Website: https://databricks.com Twitter: https://twitter.com/databricks LinkedIn: https://www.linkedin.com/company/databricks Instagram: https://www.instagram.com/databricksinc Facebook: https://www.facebook.com/databricksinc

AWS Data Lakehouse Databricks Delta SQL

In this session we’ll present Mosaic, a new Databricks Labs project with a geospatial flavour.

Mosaic provides users of Spark and Databricks with a unified framework for distributing geospatial analytics. Users can choose to employ existing Java-based tools such as JTS or Esri's Geometry API for Java and Mosaic will handle the task of parallelizing these tools' operations: e.g. efficiently reading and writing geospatial data and performing spatial functions on geometries. Mosaic helps users scale these operations by providing spatial indexing capabilities (using, for example, Uber's H3 library) and advanced techniques for optimising common point-in-polygon and polygon-polygon intersection operations.

The development of Mosaic builds upon techniques developed with Ordnance Survey (the central hub for geospatial data across UK Government) and described in this blog post: https://databricks.com/blog/2021/10/11/efficient-point-in-polygon-joins-via-pyspark-and-bng-geospatial-indexing.html

Connect with us: Website: https://databricks.com Facebook: https://www.facebook.com/databricksinc Twitter: https://twitter.com/databricks LinkedIn: https://www.linkedin.com/company/data... Instagram: https://www.instagram.com/databricksinc/

Analytics API Databricks HTML Java PySpark Spark
Yanqing Zeng – Lead Data Scientist @ JLL , Luis Sanz – CEO @ CARTO

Luis Sanz, CEO of CARTO and Yanqing Zeng, Lead Data Scientist at JLL, take us through how cloud native geospatial analytics can be unlocked on the Databricks Lakehouse platform with CARTO. Yanqing will showcase her work on large scale spatial analytics projects to address some of the most critical analysis use cases in Real Estate. Taking a geospatial perspective, Yanqing will share practical examples of how large-scale spatial data and analytics can be used for property portfolio mapping, AI-driven risk assessment, real estate valuation and more.

Connect with us: Website: https://databricks.com Facebook: https://www.facebook.com/databricksinc Twitter: https://twitter.com/databricks LinkedIn: https://www.linkedin.com/company/data... Instagram: https://www.instagram.com/databricksinc/

AI/ML Analytics Cloud Computing Data Lakehouse Databricks

In this talk, we'll review the major milestones that have defined Spatial SQL as the powerful tool for geospatial analytics that it is today.

From the early foundations of the JTS Topology Suite and GEOS and its application on the PostGIS extension for PostgreSQL, to the latest implementation in Spark SQL using libraries such as the CARTO Analytics Toolbox for Databricks, Spatial SQL has been a key component of many geospatial analytics products and solutions, leveraging the computing power of different databases with SQL as lingua franca, allowing easy adoption by data scientists, analysts and engineers.

The latest innovation in this area is the CARTO Spatial Extension for Databricks, which makes the most of the near-unlimited scalability provided by Spark and the cutting-edge geospatial capabilities that CARTO offers.

Connect with us: Website: https://databricks.com Facebook: https://www.facebook.com/databricksinc Twitter: https://twitter.com/databricks LinkedIn: https://www.linkedin.com/company/data... Instagram: https://www.instagram.com/databricksinc/

Analytics Databricks postgresql Spark SQL
Showing 10 results