talk-data.com talk-data.com

Filter by Source

Select conferences and events

People (174 results)

See all 174 →
Showing 7 results

Activities & events

Title & Speakers Event

We are pleased to invite you to our upcoming meetup co-organized with Databricks on January 21 at the Databricks Amsterdam Office. Get ready for a deep dive into the cutting edge of data intelligence and innovation! First, join Principal Engineer Martin Grund as he reveals how to supercharge your development workflows - making it dramatically easier to build and run large-scale data applications from anywhere. Then, discover how ABN AMRO’s Christian te Riet (Senior Data Scientist) and Ivan Liono (Machine Learning Engineer) orchestrate high-impact, data-driven marketing campaigns using databricks and MLflow. From streamlined developer and application experiences to scaling personalized campaigns, these talks offer practical takeaways for engineers, data scientists, developers and data & AI strategists and practitioners who want to make a real-world impact - fast! We look forward to seeing you then 👋.

SCHEDULE

17:30 - 18:15: Walk-in with drinks and bites (🍕 /🍺) 18:15 - 19:00: 10x Easier Development for Spark and Python using Databricks Connect! by Martin Grund 19:00 - 19:15: Break 19:15 - 20:00: Orchestrating and Scaling Marketing Campaigns using Databricks & MLflow by Christian te Riet & Ivan Liono 20:00 - 20:30: Networking and drinks

TALKS

[Talk 1]: 10x Easier Development for Spark and Python using Databricks Connect! by Martin Grund Databricks Connect allows customers to connect to Databricks natively using Python, Scala, or R from any IDEs and data applications leveraging the power of serverless Spark. Using the serverless Spark platform has never been easier. All you need to connect to Spark is a Databricks workspace and user credential, and within seconds, you can leverage the full elasticity of the serverless Spark infrastructure. In this talk, Martin Grund, Principal Engineer at Databricks will walk you through live examples of developing and running applications using DB Connect with Databricks serverless Spark directly from different applications like Ploty, RStudio, integration testing, or even any other hosted notebook offering. Using Databricks Spark instantaneously from anywhere will dramatically improve your developer and application experience.

[Talk 2]: Orchestrating and Scaling Marketing Campaigns using Databricks & MLflow by Christian te Riet and Ivan Liono

In this talk, Christian te Riet and Ivan Liono will explore how ABN AMRO leverages Databricks and MLflow to effectively scale data-driven marketing campaigns. By centralizing and automating the creation of machine learning models for marketing campaigns, they aim to deliver more personalized experiences, boost conversion rates, and reduce operational costs. Their system generates uplift, propensity-to-buy, and churn models for various products (or product groups), simplifying model management and compliance.

Christian and Ivan will share insights from both the data science and engineering perspectives, discussing their technical implementation, design decisions, and the successes and challenges they’ve encountered. This session will inspire attendees and invite feedback, offering practical takeaways for anyone looking to implement similar solutions.

NOTE: Since we can only host a limited amount of attendees for this meetup event, if you cannot join the event, update your status to "not going" so you can give your spot to the folks on the waitlist. Thanks for your help!

DIRECTIONS

Address: Databricks Amsterdam Office Barbara Strozzilaan 350 2nd floor, Eurocenter Amsterdam 1083 HN Amsterdam

By Bike If you’re already in Amsterdam or nearby, biking is a convenient option. Use Google Maps or a similar app for a bike-friendly route. Most paths will guide you to the business district near Amsterdam RAI Station. Once you arrive at Barbara Strozzilaan, look for the Eurocenter Amsterdam building at number 350. The Databricks office is on the 2nd floor.

By Train & Metro Take a train or metro to Amsterdam RAI Station (many intercity and local trains stop here). Exit the station and follow signs toward Barbara Strozzilaan - it’s roughly a 5–10 minute walk. Databricks is located at Barbara Strozzilaan 350, in the Eurocenter Amsterdam building, on the 2nd floor.

By Tram or Bus Tram: Lines such as 4 and 12 typically serve the Amsterdam RAI area (check current schedules). Bus: Multiple bus routes also stop near Amsterdam RAI Station (e.g., 62, 65, 125, depending on direction).

Take the tram or bus to Amsterdam RAI Station. Exit the station and walk to Barbara Strozzilaan, following the same route as above. Look for Eurocenter Amsterdam at number 350, then go to the 2nd floor for Databricks.

By Car If you’re driving, take the A10 ring road around Amsterdam and use the exit for Amsterdam RAI (often exit 9 or 10, depending on your direction). Follow signs toward Barbara Strozzilaan.

Parking: There are paid parking facilities near the RAI and within the Zuidas area. You may also find parking options in the Eurocenter garage.

Driving value forward: rapid development to real-world impact
Neil Metzler – Career Coach, Senior Account Executive , Beyza Ari – Programm-Manager HR EMEA / AWS @ AWS , Martin Grund – Engineering Lead / Site Lead @ Databricks , Kendy Rannenberg – Manager Business Dev. CE @ Databricks
Empowerment of women in tech 2024-06-20 · 19:00

Karlsruher Data Community Treffen Dezember 2023

Im Dezember steht das nächste Treffen in Karlsruhe an! Wir planen mit insgesamt 6 Veranstaltungen, die geplanten Termine und Themen könnt ihr hier nachlesen: https://www.sqlpass.de/groups/rg-karlsruhe/

Auch dieses Mal haben wir wieder 2 spannende Vorträge: Martin Guth stellt Tools fürs Performance Tooling im SQL Server vor und Stefan Kirner Apache Airflow in der Data Factory. Dazwischen gibt es wie immer die Gelegenheit sich mit anderen Data Monstern bei Pizza & Bier auszutauschen. Wir treffen uns wie gehabt im Karlsruher Büro des Raumsponsors scieneers.

Session 1: Profi-Werkzeuge für Performancetuning: Monitoring, Baselines & Workloadtests

Lernt in dieser Session wie ihr durch Profi-Werkzeuge effizienter beim Performance Tuning werden könnt. Was geschieht gerade? - Einsatz von Monitoring zum Aufdecken von Engpässen Was hat sich verändert? - Nutzung von Baselines zur Kenntnis eurer Workloads und dem Wissen wann sich proaktives Tuning empfiehlt Was kann beim Change schiefgehen? - Nutzung von Workloadtests für entspannte Releases Grundkenntnisse über Performancetuning von SQL Server Abfragen werden vorausgesetzt. In der Session geht es jedoch explizit nicht um Beispiele für das Tuning einzelner Abfragen, sondern vielmehr um Techniken relevante Abfragen aufzudecken und das Performancetuning durch umfangreiche Tests abzusichern.

Speaker: Martin Guth (40) ist seit mehr als 14 Jahren ein leidenschaftlicher BI-Entwickler. Seit 8 Jahren fasziniert ihn auch die Datenbankadministration. Er arbeitet bei der 3C Deutschland GmbH (Experian) in Heilbronn. Highlights seiner Arbeit als BI-Entwickler sind der Aufbau eines Data Warehouse von Grund auf und die Auslieferung von analytischen Produkten. Als Datenbankadministrator ist ihm das Performance-Tuning besonders wichtig. Martin teilt seine Erfahrungen (absichtlich unregelmäßig) in seinem Blog unter martinguth.de.

Pause mit Pizza & Getränken

An dem Abend ist wie gewohnt auch für das leibliche Wohl gesorgt, nerd typisch gibt es vom Raumsponsor Pizza and Bier (& andere kalte Getränke).

Session 2: Apache Airflow in der Azure Data Factory

Seit vielen Jahren gerade bei Data Monstern mit Python Kenntnissen beliebt ist nun auch Apache Airflow teil der Azure Data Factory Famlie, und zwar in einer managed Variante! Wir schauen uns an wie die DAGs in der Data Factory genutzt werden können, wie es um die Integration und Skalierung in der ADF bestellt ist. Natürlich ziehen wir auch den Vergleich zur Orchestrierung mit klassischen Pipelines.

Speaker: Stefan Kirner ist Director Business Intelligence bei der scieneers GmbH und beschäftigt sich seit dem SQL Server 2000 mit der Microsoft Data Platform. Stefan ist zertifiziert zum Azure Data Engineer & Power BI Associate and Azure Solution Architect Expert. Seit 2006 leitet er die SQL PASS Regionalgruppe Karlsruhe und engagiert sich im Beirat des Vereins. Ihr erreicht ihn unter [email protected] oder auf Twitter unter @KirnerKa

Organisatorisches Bitte meldet euch vorab über Meetup an, da die Teilnehmerzahl begrenzt ist.

Wir starten um 18 Uhr mit dem Event, die Vorträge dauern beide ungefähr eine Stunde +/-.

Wir freuen uns auf einen interessanten Abend und natürlich darauf, viele von euch wiederzusehen!

Stefan Kirner / Michael Riedmüller

Profi-Werkzeuge für SQL Performancetuning // Airflow in der Azure Data Factory
Martin Grund – Engineering Lead / Site Lead @ Databricks , Stefania Leone – Director, Product Management @ Databricks

Over the past decade, developers, researchers, and the community at large have successfully built tens of thousands of data applications using Apache Spark™. Since then, use cases and requirements of data applications have evolved. Today, every application, from web services that run in application servers, interactive environments such as notebooks and IDEs, to phones and edge devices such as smart home devices, want to leverage the power of data. However, Spark's driver architecture is monolithic, running client applications on top of a scheduler, optimizer and analyzer. This architecture makes it hard to address these new requirements as there is no built-in capability to remotely connect to a Spark cluster from languages other than SQL.

Spark Connect introduces a decoupled client-server architecture for Apache Spark that allows remote connectivity to Spark clusters using the DataFrame API and unresolved logical plans as the protocol. The separation between client and server allows Spark and its open ecosystem to be leveraged from everywhere. It can be embedded in modern data applications, in IDEs, notebooks and programming languages. This session highlights how simple it is to connect to Spark using Spark Connect from any data applications or IDEs. We will do a deep dive into the architecture of Spark Connect and provide an outlook on how the community can participate in the extension of Spark Connect for new programming languages and frameworks bringing the power of Spark everywhere.

Talk by: Martin Grund and Stefania Leone

Connect with us: Website: https://databricks.com Twitter: https://twitter.com/databricks LinkedIn: https://www.linkedin.com/company/databricks Instagram: https://www.instagram.com/databricksinc Facebook: https://www.facebook.com/databricksinc

API Databricks Spark SQL
Lakehouse / Spark AMA 2023-07-26 · 21:04
Hyukjin Kwon – Staff Software Engineer @ Databricks , Martin Grund – Engineering Lead / Site Lead @ Databricks , Wenchen Fan – Senior Staff Software Engineer @ Databricks

Have some great questions about Apache Spark™ and Lakehouses?  Well, come by and ask the experts your questions!

Talk by: Martin Grund, Hyukjin Kwon, and Wenchen Fan

Connect with us: Website: https://databricks.com Twitter: https://twitter.com/databricks LinkedIn: https://www.linkedin.com/company/databricks Instagram: https://www.instagram.com/databricksinc Facebook: https://www.facebook.com/databricksinc

Data Lakehouse Databricks Spark
Martin Grund – Engineering Lead / Site Lead @ Databricks , Stefania Leone – Director, Product Management @ Databricks

Spark developers want to develop and debug their code using their tools of choice and development best practices while ensuring high-production fidelity on the target remote cluster. However, Spark's driver architecture is monolithic, with no built-in capability to directly connect to a remote Spark cluster from languages other than SQL. This makes it hard to enable such interactive developer experiences from a user’s local IDE of choice. Spark Connect’s decoupled client-server architecture introduces remote connectivity to Spark clusters and with that, enables interactive development experience - Spark and its open ecosystem can be leveraged from everywhere.

In this session, we show how we leverage Spark Connect to build a completely redesigned version of Databricks Connect, a first-class IDE-based developer experience that offers interactive debugging from any IDE. We show how developers can easily ensure consistency between their local and remote environments. We walk the audience through real-live examples of how to locally debug code running on Databrick. We also show how Databricks Connect integrates into the Databricks Visual Studio Code extension for an even better developer experience.

Talk by: Martin Grund and Stefania Leone

Connect with us: Website: https://databricks.com Twitter: https://twitter.com/databricks LinkedIn: https://www.linkedin.com/company/databricks Instagram: https://www.instagram.com/databricksinc Facebook: https://www.facebook.com/databricksinc

Databricks Spark SQL
Showing 7 results