talk-data.com talk-data.com

Filter by Source

Select conferences and events

People (774 results)

See all 774 →

Activities & events

Title & Speakers Event

Real-Time Intelligence session: A hands-on look at end-to-end data round-trip from creation to insights. The presenter demonstrates building a weather-prediction solution using Real-Time Intelligence within Microsoft Fabric. Topics include collecting weather data on a Raspberry Pi, sending it to Fabric via Azure IoT Hub, and connecting to the data in Microsoft Fabric to transform, predict, and alert in real time about whether to go outside today.

raspberry pi azure iot hub Fabric
February Monthly CASSUG Meeting
Meetup @ Axxes 2026-01-28 · 16:30

2026 is here and we continue with hosting the best AWS meetups in The Netherlands, to kick off this year we are hosting a meetup together with Axxes. We have amazing talks lined up for this event, so make sure to register yourself soon.

Information

  • Paid parking is available (street or parking garage)
  • Dinner, drinks and borrel are included

Agenda 17:30 - Food 🍴 18:30 - Rob Kenis - No more long lived credentials 19:00 - Amer Grgic - Kiro, Agentic AI development environment from prototype to production 19:00 - 🚰 Break 19:15 - Yannick van Rooyen & Joeri Malmberg - AWS Platform Engineering at Europe's Largest Tendering Platform 19:45 - Drinks 🍻 & Networking

No more long lived credentials When connecting AWS and other services, we still see the issue of using long lived credentials for authentication. In this talk, we will solve the issue using AWS IAM OIDC providers and IAM outbound identity federation.

Kiro, Agentic AI development environment from prototype to production In this talk, we'll discuss Kiro - an agentic AI development environment that seamlessly takes your projects from prototype to production. It's designed to streamline the AI development lifecycle, making the transition from experimental code to production-ready systems smoother than ever. Kiro isn't just another development tool - it's your AI project's companion from concept to deployment.

AWS Platform Engineering at Europe's Largest Tendering Platform Building an AWS platform for Europe's largest tendering system means supporting a constantly growing number of workloads, teams, and architectural styles—from legacy solutions inherited through acquisitions to modern, event-driven and serverless systems. It often feels like fixing the plane while flying it. In this talk, we'll share how we built and evolved our AWS platform to reduce complexity for developers, provide a safe foundation for change, and enable all kinds of workloads to land and scale on the same platform—while keeping delivery fast.

Meetup @ Axxes
New Year, New GolangNYC 2026-01-15 · 23:30

The holidays are over, the champagne's gone flat, and resolutions have been set. It's time for the first GolangNYC meetup of the year. Join us as we kick of 2026 the right way and `defer fallingBackToBadHabits()` by learning from some great talks, and enjoying some food, drinks, and great company at the Copia Automation offices.

Where: Copia Automation 43 W 24th St 6th Floor New York, NY 10010

When: January 15th, 2026 6:30pm-9:00pm

Agenda 6:30pm - 7:00pm: Doors Open, Networking 7:00pm - 7:10pm: Opening Remarks 7:10pm - 7:30pm: Pong in Go with Nick Golebiewski 7:30pm - 8:10pm: Git as an S3 Replacement? with Tom Elliot 8:10pm - 8:30pm: Open Lighting Round Talks (Sign up in Person) or Networking 8:30pm - 9:00pm: Networking

New Year, New GolangNYC

Welcome to day two of the Best of NeurIPS series, your virtual pass to some of the groundbreaking research, insights, and innovations that defined this year’s conference. Live streaming from the authors to you.

Time and Location

Jan 15, 2026 9:00-11:00 AM Pacific Online. Register for the Zoom!

Why Diffusion Models Don't Memorize: The Role of Implicit Dynamical Regularization in Training

Diffusion models have achieved impressive results across many generative tasks, yet the mechanisms that prevent memorization and enable generalization remain unclear. In this talk, I will focus on how training dynamics shape the transition from generalization to memorization. Our experiments and theory reveal two key timescales: an early time when high-quality generation emerges and a later one when memorization begins. Notably, the memorization timescale grows linearly with the size of the training set, while the generalization timescale stays constant, creating an increasingly wide window where models generalize well. These results highlight an implicit dynamical regularization that helps diffusion models avoid memorization even in highly overparameterized regimes.

About the Author

Raphaël Urfin is a PhD student at École Normale Supérieure – PSL in Paris, supervised by Giulio Biroli (ENS) and Marc Mézard (Bocconi University). His work focuses on applying ideas and tools of statistical physics to better understand diffusion models and their generalization properties.

Open-Insect: Benchmarking Open-Set Recognition of Novel Species in Biodiversity Monitoring

Global biodiversity is declining at an unprecedented rate, yet little information is known about most species and how their populations are changing. Indeed, some 90% Earth’s species are estimated to be completely unknown. Machine learning has recently emerged as a promising tool to facilitate long-term, large-scale biodiversity monitoring, including algorithms for fine-grained classification of species from images. However, such algorithms typically are not designed to detect examples from categories unseen during training – the problem of open-set recognition (OSR) – limiting their applicability for highly diverse, poorly studied taxa such as insects. To address this gap, we introduce Open-Insect, a large-scale, fine-grained dataset to evaluate unknown species detection across different geographic regions with varying difficulty. We benchmark 38 OSR algorithms across three categories: post-hoc, training-time regularization, and training with auxiliary data, finding that simple post-hoc approaches remain a strong baseline. We also demonstrate how to leverage auxiliary data to improve species discovery in regions with limited data. Our results provide timely insights to guide the development of computer vision methods for biodiversity monitoring and species discovery.

About the Speaker

Yuyan Chen is a PhD student in Computer Science at McGill University and Mila - Quebec AI Institute, supervised by Prof. David Rolnick. My research focuses on machine learning for biodiversity monitoring.

GuideFlow3D: Optimization-Guided Rectified Flow For Appearance Transfer

Transferring appearance to 3D assets using different representations of the appearance object - such as images or text - has garnered interest due to its wide range of applications in industries like gaming, augmented reality, and digital content creation. However, state-of-the-art methods still fail when the geometry between the input and appearance objects is significantly different. A straightforward approach is to directly apply a 3D generative model, but we show that this ultimately fails to produce appealing results.

Instead, we propose a principled approach inspired by universal guidance. Given a pretrained rectified flow model conditioned on image or text, our training-free method interacts with the sampling process by periodically adding guidance. This guidance can be modeled as a differentiable loss function, and we experiment with two different types of guidance including part-aware losses for appearance and self-similarity. Our experiments show that our approach successfully transfers texture and geometric details to the input 3D asset, outperforming baselines both qualitatively and quantitatively.

We also show that traditional metrics are not suitable for evaluating the task due to their inability of focusing on local details and comparing dissimilar inputs, in absence of ground truth data. We thus evaluate appearance transfer quality with a GPT-based system objectively ranking outputs, ensuring robust and human-like assessment, as further confirmed by our user study. Beyond showcased scenarios, our method is general and could be extended to different types of diffusion models and guidance functions.

About the Speaker

Sayan Deb Sarkar is a 2nd-year PhD student at Stanford University in the Gradient Spaces Group, advised by Prof. Iro Armeni, part of the Stanford Vision Lab (SVL). His research interests are on multimodal 3D scene understanding and interactive editing. Past summer, he interned with the Microsoft Spatial AI Lab, hosted by Prof. Marc Pollefeys, working on efficient video understanding in spatial context. Before starting PhD, he was a CS master student at ETH Zürich, in the Computer Vision and Geometry Group (CVG), working on aligning real-world 3D environments from multi-modal data. In the past, he has been a Research Intern at Qualcomm XR labs, Computer Vision Engineer at Mercedes Benz R & D and Research Engineer at ICG, TU Graz. Website: https://sayands.github.io/

HouseLayout3D: A Benchmark and Baseline Method for 3D Layout Estimation in the Wild​ ​

Current 3D layout estimation models are primarily trained on synthetic datasets containing simple single room or single floor environments. As a consequence, they cannot natively handle large multi floor buildings and require scenes to be split into individual floors before processing, which removes global spatial context that is essential for reasoning about structures such as staircases that connect multiple levels. In this work, we introduce HouseLayout3D, a real world benchmark designed to support progress toward full building scale layout estimation, including multiple floors and architecturally intricate spaces. We also present MultiFloor3D, a simple training free baseline that leverages recent scene understanding methods and already outperforms existing 3D layout estimation models on both our benchmark and prior datasets, highlighting the need for further research in this direction.

About the Speaker

Valentin Bieri is a Machine Learning Engineer and Researcher specializing in the intersection of 3D Computer Vision and Natural Language Processing. Building on his applied research in SLAM and Vision-Language Models at ETH Zurich, he now develops AI agents for manufacturing at EthonAI.

Jan 15 - Best of NeurIPS (Day 2)

Welcome to day two of the Best of NeurIPS series, your virtual pass to some of the groundbreaking research, insights, and innovations that defined this year’s conference. Live streaming from the authors to you.

Time and Location

Jan 15, 2026 9:00-11:00 AM Pacific Online. Register for the Zoom!

Why Diffusion Models Don't Memorize: The Role of Implicit Dynamical Regularization in Training

Diffusion models have achieved impressive results across many generative tasks, yet the mechanisms that prevent memorization and enable generalization remain unclear. In this talk, I will focus on how training dynamics shape the transition from generalization to memorization. Our experiments and theory reveal two key timescales: an early time when high-quality generation emerges and a later one when memorization begins. Notably, the memorization timescale grows linearly with the size of the training set, while the generalization timescale stays constant, creating an increasingly wide window where models generalize well. These results highlight an implicit dynamical regularization that helps diffusion models avoid memorization even in highly overparameterized regimes.

About the Author

Raphaël Urfin is a PhD student at École Normale Supérieure – PSL in Paris, supervised by Giulio Biroli (ENS) and Marc Mézard (Bocconi University). His work focuses on applying ideas and tools of statistical physics to better understand diffusion models and their generalization properties.

Open-Insect: Benchmarking Open-Set Recognition of Novel Species in Biodiversity Monitoring

Global biodiversity is declining at an unprecedented rate, yet little information is known about most species and how their populations are changing. Indeed, some 90% Earth’s species are estimated to be completely unknown. Machine learning has recently emerged as a promising tool to facilitate long-term, large-scale biodiversity monitoring, including algorithms for fine-grained classification of species from images. However, such algorithms typically are not designed to detect examples from categories unseen during training – the problem of open-set recognition (OSR) – limiting their applicability for highly diverse, poorly studied taxa such as insects. To address this gap, we introduce Open-Insect, a large-scale, fine-grained dataset to evaluate unknown species detection across different geographic regions with varying difficulty. We benchmark 38 OSR algorithms across three categories: post-hoc, training-time regularization, and training with auxiliary data, finding that simple post-hoc approaches remain a strong baseline. We also demonstrate how to leverage auxiliary data to improve species discovery in regions with limited data. Our results provide timely insights to guide the development of computer vision methods for biodiversity monitoring and species discovery.

About the Speaker

Yuyan Chen is a PhD student in Computer Science at McGill University and Mila - Quebec AI Institute, supervised by Prof. David Rolnick. My research focuses on machine learning for biodiversity monitoring.

GuideFlow3D: Optimization-Guided Rectified Flow For Appearance Transfer

Transferring appearance to 3D assets using different representations of the appearance object - such as images or text - has garnered interest due to its wide range of applications in industries like gaming, augmented reality, and digital content creation. However, state-of-the-art methods still fail when the geometry between the input and appearance objects is significantly different. A straightforward approach is to directly apply a 3D generative model, but we show that this ultimately fails to produce appealing results.

Instead, we propose a principled approach inspired by universal guidance. Given a pretrained rectified flow model conditioned on image or text, our training-free method interacts with the sampling process by periodically adding guidance. This guidance can be modeled as a differentiable loss function, and we experiment with two different types of guidance including part-aware losses for appearance and self-similarity. Our experiments show that our approach successfully transfers texture and geometric details to the input 3D asset, outperforming baselines both qualitatively and quantitatively.

We also show that traditional metrics are not suitable for evaluating the task due to their inability of focusing on local details and comparing dissimilar inputs, in absence of ground truth data. We thus evaluate appearance transfer quality with a GPT-based system objectively ranking outputs, ensuring robust and human-like assessment, as further confirmed by our user study. Beyond showcased scenarios, our method is general and could be extended to different types of diffusion models and guidance functions.

About the Speaker

Sayan Deb Sarkar is a 2nd-year PhD student at Stanford University in the Gradient Spaces Group, advised by Prof. Iro Armeni, part of the Stanford Vision Lab (SVL). His research interests are on multimodal 3D scene understanding and interactive editing. Past summer, he interned with the Microsoft Spatial AI Lab, hosted by Prof. Marc Pollefeys, working on efficient video understanding in spatial context. Before starting PhD, he was a CS master student at ETH Zürich, in the Computer Vision and Geometry Group (CVG), working on aligning real-world 3D environments from multi-modal data. In the past, he has been a Research Intern at Qualcomm XR labs, Computer Vision Engineer at Mercedes Benz R & D and Research Engineer at ICG, TU Graz. Website: https://sayands.github.io/

HouseLayout3D: A Benchmark and Baseline Method for 3D Layout Estimation in the Wild​ ​

Current 3D layout estimation models are primarily trained on synthetic datasets containing simple single room or single floor environments. As a consequence, they cannot natively handle large multi floor buildings and require scenes to be split into individual floors before processing, which removes global spatial context that is essential for reasoning about structures such as staircases that connect multiple levels. In this work, we introduce HouseLayout3D, a real world benchmark designed to support progress toward full building scale layout estimation, including multiple floors and architecturally intricate spaces. We also present MultiFloor3D, a simple training free baseline that leverages recent scene understanding methods and already outperforms existing 3D layout estimation models on both our benchmark and prior datasets, highlighting the need for further research in this direction.

About the Speaker

Valentin Bieri is a Machine Learning Engineer and Researcher specializing in the intersection of 3D Computer Vision and Natural Language Processing. Building on his applied research in SLAM and Vision-Language Models at ETH Zurich, he now develops AI agents for manufacturing at EthonAI.

Jan 15 - Best of NeurIPS (Day 2)

Welcome to day two of the Best of NeurIPS series, your virtual pass to some of the groundbreaking research, insights, and innovations that defined this year’s conference. Live streaming from the authors to you.

Time and Location

Jan 15, 2026 9:00-11:00 AM Pacific Online. Register for the Zoom!

Why Diffusion Models Don't Memorize: The Role of Implicit Dynamical Regularization in Training

Diffusion models have achieved impressive results across many generative tasks, yet the mechanisms that prevent memorization and enable generalization remain unclear. In this talk, I will focus on how training dynamics shape the transition from generalization to memorization. Our experiments and theory reveal two key timescales: an early time when high-quality generation emerges and a later one when memorization begins. Notably, the memorization timescale grows linearly with the size of the training set, while the generalization timescale stays constant, creating an increasingly wide window where models generalize well. These results highlight an implicit dynamical regularization that helps diffusion models avoid memorization even in highly overparameterized regimes.

About the Author

Raphaël Urfin is a PhD student at École Normale Supérieure – PSL in Paris, supervised by Giulio Biroli (ENS) and Marc Mézard (Bocconi University). His work focuses on applying ideas and tools of statistical physics to better understand diffusion models and their generalization properties.

Open-Insect: Benchmarking Open-Set Recognition of Novel Species in Biodiversity Monitoring

Global biodiversity is declining at an unprecedented rate, yet little information is known about most species and how their populations are changing. Indeed, some 90% Earth’s species are estimated to be completely unknown. Machine learning has recently emerged as a promising tool to facilitate long-term, large-scale biodiversity monitoring, including algorithms for fine-grained classification of species from images. However, such algorithms typically are not designed to detect examples from categories unseen during training – the problem of open-set recognition (OSR) – limiting their applicability for highly diverse, poorly studied taxa such as insects. To address this gap, we introduce Open-Insect, a large-scale, fine-grained dataset to evaluate unknown species detection across different geographic regions with varying difficulty. We benchmark 38 OSR algorithms across three categories: post-hoc, training-time regularization, and training with auxiliary data, finding that simple post-hoc approaches remain a strong baseline. We also demonstrate how to leverage auxiliary data to improve species discovery in regions with limited data. Our results provide timely insights to guide the development of computer vision methods for biodiversity monitoring and species discovery.

About the Speaker

Yuyan Chen is a PhD student in Computer Science at McGill University and Mila - Quebec AI Institute, supervised by Prof. David Rolnick. My research focuses on machine learning for biodiversity monitoring.

GuideFlow3D: Optimization-Guided Rectified Flow For Appearance Transfer

Transferring appearance to 3D assets using different representations of the appearance object - such as images or text - has garnered interest due to its wide range of applications in industries like gaming, augmented reality, and digital content creation. However, state-of-the-art methods still fail when the geometry between the input and appearance objects is significantly different. A straightforward approach is to directly apply a 3D generative model, but we show that this ultimately fails to produce appealing results.

Instead, we propose a principled approach inspired by universal guidance. Given a pretrained rectified flow model conditioned on image or text, our training-free method interacts with the sampling process by periodically adding guidance. This guidance can be modeled as a differentiable loss function, and we experiment with two different types of guidance including part-aware losses for appearance and self-similarity. Our experiments show that our approach successfully transfers texture and geometric details to the input 3D asset, outperforming baselines both qualitatively and quantitatively.

We also show that traditional metrics are not suitable for evaluating the task due to their inability of focusing on local details and comparing dissimilar inputs, in absence of ground truth data. We thus evaluate appearance transfer quality with a GPT-based system objectively ranking outputs, ensuring robust and human-like assessment, as further confirmed by our user study. Beyond showcased scenarios, our method is general and could be extended to different types of diffusion models and guidance functions.

About the Speaker

Sayan Deb Sarkar is a 2nd-year PhD student at Stanford University in the Gradient Spaces Group, advised by Prof. Iro Armeni, part of the Stanford Vision Lab (SVL). His research interests are on multimodal 3D scene understanding and interactive editing. Past summer, he interned with the Microsoft Spatial AI Lab, hosted by Prof. Marc Pollefeys, working on efficient video understanding in spatial context. Before starting PhD, he was a CS master student at ETH Zürich, in the Computer Vision and Geometry Group (CVG), working on aligning real-world 3D environments from multi-modal data. In the past, he has been a Research Intern at Qualcomm XR labs, Computer Vision Engineer at Mercedes Benz R & D and Research Engineer at ICG, TU Graz. Website: https://sayands.github.io/

HouseLayout3D: A Benchmark and Baseline Method for 3D Layout Estimation in the Wild​ ​

Current 3D layout estimation models are primarily trained on synthetic datasets containing simple single room or single floor environments. As a consequence, they cannot natively handle large multi floor buildings and require scenes to be split into individual floors before processing, which removes global spatial context that is essential for reasoning about structures such as staircases that connect multiple levels. In this work, we introduce HouseLayout3D, a real world benchmark designed to support progress toward full building scale layout estimation, including multiple floors and architecturally intricate spaces. We also present MultiFloor3D, a simple training free baseline that leverages recent scene understanding methods and already outperforms existing 3D layout estimation models on both our benchmark and prior datasets, highlighting the need for further research in this direction.

About the Speaker

Valentin Bieri is a Machine Learning Engineer and Researcher specializing in the intersection of 3D Computer Vision and Natural Language Processing. Building on his applied research in SLAM and Vision-Language Models at ETH Zurich, he now develops AI agents for manufacturing at EthonAI.

Jan 15 - Best of NeurIPS (Day 2)

Welcome to day two of the Best of NeurIPS series, your virtual pass to some of the groundbreaking research, insights, and innovations that defined this year’s conference. Live streaming from the authors to you.

Time and Location

Jan 15, 2026 9:00-11:00 AM Pacific Online. Register for the Zoom!

Why Diffusion Models Don't Memorize: The Role of Implicit Dynamical Regularization in Training

Diffusion models have achieved impressive results across many generative tasks, yet the mechanisms that prevent memorization and enable generalization remain unclear. In this talk, I will focus on how training dynamics shape the transition from generalization to memorization. Our experiments and theory reveal two key timescales: an early time when high-quality generation emerges and a later one when memorization begins. Notably, the memorization timescale grows linearly with the size of the training set, while the generalization timescale stays constant, creating an increasingly wide window where models generalize well. These results highlight an implicit dynamical regularization that helps diffusion models avoid memorization even in highly overparameterized regimes.

About the Author

Raphaël Urfin is a PhD student at École Normale Supérieure – PSL in Paris, supervised by Giulio Biroli (ENS) and Marc Mézard (Bocconi University). His work focuses on applying ideas and tools of statistical physics to better understand diffusion models and their generalization properties.

Open-Insect: Benchmarking Open-Set Recognition of Novel Species in Biodiversity Monitoring

Global biodiversity is declining at an unprecedented rate, yet little information is known about most species and how their populations are changing. Indeed, some 90% Earth’s species are estimated to be completely unknown. Machine learning has recently emerged as a promising tool to facilitate long-term, large-scale biodiversity monitoring, including algorithms for fine-grained classification of species from images. However, such algorithms typically are not designed to detect examples from categories unseen during training – the problem of open-set recognition (OSR) – limiting their applicability for highly diverse, poorly studied taxa such as insects. To address this gap, we introduce Open-Insect, a large-scale, fine-grained dataset to evaluate unknown species detection across different geographic regions with varying difficulty. We benchmark 38 OSR algorithms across three categories: post-hoc, training-time regularization, and training with auxiliary data, finding that simple post-hoc approaches remain a strong baseline. We also demonstrate how to leverage auxiliary data to improve species discovery in regions with limited data. Our results provide timely insights to guide the development of computer vision methods for biodiversity monitoring and species discovery.

About the Speaker

Yuyan Chen is a PhD student in Computer Science at McGill University and Mila - Quebec AI Institute, supervised by Prof. David Rolnick. My research focuses on machine learning for biodiversity monitoring.

GuideFlow3D: Optimization-Guided Rectified Flow For Appearance Transfer

Transferring appearance to 3D assets using different representations of the appearance object - such as images or text - has garnered interest due to its wide range of applications in industries like gaming, augmented reality, and digital content creation. However, state-of-the-art methods still fail when the geometry between the input and appearance objects is significantly different. A straightforward approach is to directly apply a 3D generative model, but we show that this ultimately fails to produce appealing results.

Instead, we propose a principled approach inspired by universal guidance. Given a pretrained rectified flow model conditioned on image or text, our training-free method interacts with the sampling process by periodically adding guidance. This guidance can be modeled as a differentiable loss function, and we experiment with two different types of guidance including part-aware losses for appearance and self-similarity. Our experiments show that our approach successfully transfers texture and geometric details to the input 3D asset, outperforming baselines both qualitatively and quantitatively.

We also show that traditional metrics are not suitable for evaluating the task due to their inability of focusing on local details and comparing dissimilar inputs, in absence of ground truth data. We thus evaluate appearance transfer quality with a GPT-based system objectively ranking outputs, ensuring robust and human-like assessment, as further confirmed by our user study. Beyond showcased scenarios, our method is general and could be extended to different types of diffusion models and guidance functions.

About the Speaker

Sayan Deb Sarkar is a 2nd-year PhD student at Stanford University in the Gradient Spaces Group, advised by Prof. Iro Armeni, part of the Stanford Vision Lab (SVL). His research interests are on multimodal 3D scene understanding and interactive editing. Past summer, he interned with the Microsoft Spatial AI Lab, hosted by Prof. Marc Pollefeys, working on efficient video understanding in spatial context. Before starting PhD, he was a CS master student at ETH Zürich, in the Computer Vision and Geometry Group (CVG), working on aligning real-world 3D environments from multi-modal data. In the past, he has been a Research Intern at Qualcomm XR labs, Computer Vision Engineer at Mercedes Benz R & D and Research Engineer at ICG, TU Graz. Website: https://sayands.github.io/

HouseLayout3D: A Benchmark and Baseline Method for 3D Layout Estimation in the Wild​ ​

Current 3D layout estimation models are primarily trained on synthetic datasets containing simple single room or single floor environments. As a consequence, they cannot natively handle large multi floor buildings and require scenes to be split into individual floors before processing, which removes global spatial context that is essential for reasoning about structures such as staircases that connect multiple levels. In this work, we introduce HouseLayout3D, a real world benchmark designed to support progress toward full building scale layout estimation, including multiple floors and architecturally intricate spaces. We also present MultiFloor3D, a simple training free baseline that leverages recent scene understanding methods and already outperforms existing 3D layout estimation models on both our benchmark and prior datasets, highlighting the need for further research in this direction.

About the Speaker

Valentin Bieri is a Machine Learning Engineer and Researcher specializing in the intersection of 3D Computer Vision and Natural Language Processing. Building on his applied research in SLAM and Vision-Language Models at ETH Zurich, he now develops AI agents for manufacturing at EthonAI.

Jan 15 - Best of NeurIPS (Day 2)

Welcome to day two of the Best of NeurIPS series, your virtual pass to some of the groundbreaking research, insights, and innovations that defined this year’s conference. Live streaming from the authors to you.

Time and Location

Jan 15, 2026 9:00-11:00 AM Pacific Online. Register for the Zoom!

Why Diffusion Models Don't Memorize: The Role of Implicit Dynamical Regularization in Training

Diffusion models have achieved impressive results across many generative tasks, yet the mechanisms that prevent memorization and enable generalization remain unclear. In this talk, I will focus on how training dynamics shape the transition from generalization to memorization. Our experiments and theory reveal two key timescales: an early time when high-quality generation emerges and a later one when memorization begins. Notably, the memorization timescale grows linearly with the size of the training set, while the generalization timescale stays constant, creating an increasingly wide window where models generalize well. These results highlight an implicit dynamical regularization that helps diffusion models avoid memorization even in highly overparameterized regimes.

About the Author

Raphaël Urfin is a PhD student at École Normale Supérieure – PSL in Paris, supervised by Giulio Biroli (ENS) and Marc Mézard (Bocconi University). His work focuses on applying ideas and tools of statistical physics to better understand diffusion models and their generalization properties.

Open-Insect: Benchmarking Open-Set Recognition of Novel Species in Biodiversity Monitoring

Global biodiversity is declining at an unprecedented rate, yet little information is known about most species and how their populations are changing. Indeed, some 90% Earth’s species are estimated to be completely unknown. Machine learning has recently emerged as a promising tool to facilitate long-term, large-scale biodiversity monitoring, including algorithms for fine-grained classification of species from images. However, such algorithms typically are not designed to detect examples from categories unseen during training – the problem of open-set recognition (OSR) – limiting their applicability for highly diverse, poorly studied taxa such as insects. To address this gap, we introduce Open-Insect, a large-scale, fine-grained dataset to evaluate unknown species detection across different geographic regions with varying difficulty. We benchmark 38 OSR algorithms across three categories: post-hoc, training-time regularization, and training with auxiliary data, finding that simple post-hoc approaches remain a strong baseline. We also demonstrate how to leverage auxiliary data to improve species discovery in regions with limited data. Our results provide timely insights to guide the development of computer vision methods for biodiversity monitoring and species discovery.

About the Speaker

Yuyan Chen is a PhD student in Computer Science at McGill University and Mila - Quebec AI Institute, supervised by Prof. David Rolnick. My research focuses on machine learning for biodiversity monitoring.

GuideFlow3D: Optimization-Guided Rectified Flow For Appearance Transfer

Transferring appearance to 3D assets using different representations of the appearance object - such as images or text - has garnered interest due to its wide range of applications in industries like gaming, augmented reality, and digital content creation. However, state-of-the-art methods still fail when the geometry between the input and appearance objects is significantly different. A straightforward approach is to directly apply a 3D generative model, but we show that this ultimately fails to produce appealing results.

Instead, we propose a principled approach inspired by universal guidance. Given a pretrained rectified flow model conditioned on image or text, our training-free method interacts with the sampling process by periodically adding guidance. This guidance can be modeled as a differentiable loss function, and we experiment with two different types of guidance including part-aware losses for appearance and self-similarity. Our experiments show that our approach successfully transfers texture and geometric details to the input 3D asset, outperforming baselines both qualitatively and quantitatively.

We also show that traditional metrics are not suitable for evaluating the task due to their inability of focusing on local details and comparing dissimilar inputs, in absence of ground truth data. We thus evaluate appearance transfer quality with a GPT-based system objectively ranking outputs, ensuring robust and human-like assessment, as further confirmed by our user study. Beyond showcased scenarios, our method is general and could be extended to different types of diffusion models and guidance functions.

About the Speaker

Sayan Deb Sarkar is a 2nd-year PhD student at Stanford University in the Gradient Spaces Group, advised by Prof. Iro Armeni, part of the Stanford Vision Lab (SVL). His research interests are on multimodal 3D scene understanding and interactive editing. Past summer, he interned with the Microsoft Spatial AI Lab, hosted by Prof. Marc Pollefeys, working on efficient video understanding in spatial context. Before starting PhD, he was a CS master student at ETH Zürich, in the Computer Vision and Geometry Group (CVG), working on aligning real-world 3D environments from multi-modal data. In the past, he has been a Research Intern at Qualcomm XR labs, Computer Vision Engineer at Mercedes Benz R & D and Research Engineer at ICG, TU Graz. Website: https://sayands.github.io/

HouseLayout3D: A Benchmark and Baseline Method for 3D Layout Estimation in the Wild​ ​

Current 3D layout estimation models are primarily trained on synthetic datasets containing simple single room or single floor environments. As a consequence, they cannot natively handle large multi floor buildings and require scenes to be split into individual floors before processing, which removes global spatial context that is essential for reasoning about structures such as staircases that connect multiple levels. In this work, we introduce HouseLayout3D, a real world benchmark designed to support progress toward full building scale layout estimation, including multiple floors and architecturally intricate spaces. We also present MultiFloor3D, a simple training free baseline that leverages recent scene understanding methods and already outperforms existing 3D layout estimation models on both our benchmark and prior datasets, highlighting the need for further research in this direction.

About the Speaker

Valentin Bieri is a Machine Learning Engineer and Researcher specializing in the intersection of 3D Computer Vision and Natural Language Processing. Building on his applied research in SLAM and Vision-Language Models at ETH Zurich, he now develops AI agents for manufacturing at EthonAI.

Jan 15 - Best of NeurIPS (Day 2)

Welcome to day two of the Best of NeurIPS series, your virtual pass to some of the groundbreaking research, insights, and innovations that defined this year’s conference. Live streaming from the authors to you.

Time and Location

Jan 15, 2026 9:00-11:00 AM Pacific Online. Register for the Zoom!

Why Diffusion Models Don't Memorize: The Role of Implicit Dynamical Regularization in Training

Diffusion models have achieved impressive results across many generative tasks, yet the mechanisms that prevent memorization and enable generalization remain unclear. In this talk, I will focus on how training dynamics shape the transition from generalization to memorization. Our experiments and theory reveal two key timescales: an early time when high-quality generation emerges and a later one when memorization begins. Notably, the memorization timescale grows linearly with the size of the training set, while the generalization timescale stays constant, creating an increasingly wide window where models generalize well. These results highlight an implicit dynamical regularization that helps diffusion models avoid memorization even in highly overparameterized regimes.

About the Author

Raphaël Urfin is a PhD student at École Normale Supérieure – PSL in Paris, supervised by Giulio Biroli (ENS) and Marc Mézard (Bocconi University). His work focuses on applying ideas and tools of statistical physics to better understand diffusion models and their generalization properties.

Open-Insect: Benchmarking Open-Set Recognition of Novel Species in Biodiversity Monitoring

Global biodiversity is declining at an unprecedented rate, yet little information is known about most species and how their populations are changing. Indeed, some 90% Earth’s species are estimated to be completely unknown. Machine learning has recently emerged as a promising tool to facilitate long-term, large-scale biodiversity monitoring, including algorithms for fine-grained classification of species from images. However, such algorithms typically are not designed to detect examples from categories unseen during training – the problem of open-set recognition (OSR) – limiting their applicability for highly diverse, poorly studied taxa such as insects. To address this gap, we introduce Open-Insect, a large-scale, fine-grained dataset to evaluate unknown species detection across different geographic regions with varying difficulty. We benchmark 38 OSR algorithms across three categories: post-hoc, training-time regularization, and training with auxiliary data, finding that simple post-hoc approaches remain a strong baseline. We also demonstrate how to leverage auxiliary data to improve species discovery in regions with limited data. Our results provide timely insights to guide the development of computer vision methods for biodiversity monitoring and species discovery.

About the Speaker

Yuyan Chen is a PhD student in Computer Science at McGill University and Mila - Quebec AI Institute, supervised by Prof. David Rolnick. My research focuses on machine learning for biodiversity monitoring.

GuideFlow3D: Optimization-Guided Rectified Flow For Appearance Transfer

Transferring appearance to 3D assets using different representations of the appearance object - such as images or text - has garnered interest due to its wide range of applications in industries like gaming, augmented reality, and digital content creation. However, state-of-the-art methods still fail when the geometry between the input and appearance objects is significantly different. A straightforward approach is to directly apply a 3D generative model, but we show that this ultimately fails to produce appealing results.

Instead, we propose a principled approach inspired by universal guidance. Given a pretrained rectified flow model conditioned on image or text, our training-free method interacts with the sampling process by periodically adding guidance. This guidance can be modeled as a differentiable loss function, and we experiment with two different types of guidance including part-aware losses for appearance and self-similarity. Our experiments show that our approach successfully transfers texture and geometric details to the input 3D asset, outperforming baselines both qualitatively and quantitatively.

We also show that traditional metrics are not suitable for evaluating the task due to their inability of focusing on local details and comparing dissimilar inputs, in absence of ground truth data. We thus evaluate appearance transfer quality with a GPT-based system objectively ranking outputs, ensuring robust and human-like assessment, as further confirmed by our user study. Beyond showcased scenarios, our method is general and could be extended to different types of diffusion models and guidance functions.

About the Speaker

Sayan Deb Sarkar is a 2nd-year PhD student at Stanford University in the Gradient Spaces Group, advised by Prof. Iro Armeni, part of the Stanford Vision Lab (SVL). His research interests are on multimodal 3D scene understanding and interactive editing. Past summer, he interned with the Microsoft Spatial AI Lab, hosted by Prof. Marc Pollefeys, working on efficient video understanding in spatial context. Before starting PhD, he was a CS master student at ETH Zürich, in the Computer Vision and Geometry Group (CVG), working on aligning real-world 3D environments from multi-modal data. In the past, he has been a Research Intern at Qualcomm XR labs, Computer Vision Engineer at Mercedes Benz R & D and Research Engineer at ICG, TU Graz. Website: https://sayands.github.io/

HouseLayout3D: A Benchmark and Baseline Method for 3D Layout Estimation in the Wild​ ​

Current 3D layout estimation models are primarily trained on synthetic datasets containing simple single room or single floor environments. As a consequence, they cannot natively handle large multi floor buildings and require scenes to be split into individual floors before processing, which removes global spatial context that is essential for reasoning about structures such as staircases that connect multiple levels. In this work, we introduce HouseLayout3D, a real world benchmark designed to support progress toward full building scale layout estimation, including multiple floors and architecturally intricate spaces. We also present MultiFloor3D, a simple training free baseline that leverages recent scene understanding methods and already outperforms existing 3D layout estimation models on both our benchmark and prior datasets, highlighting the need for further research in this direction.

About the Speaker

Valentin Bieri is a Machine Learning Engineer and Researcher specializing in the intersection of 3D Computer Vision and Natural Language Processing. Building on his applied research in SLAM and Vision-Language Models at ETH Zurich, he now develops AI agents for manufacturing at EthonAI.

Jan 15 - Best of NeurIPS (Day 2)

Welcome to day two of the Best of NeurIPS series, your virtual pass to some of the groundbreaking research, insights, and innovations that defined this year’s conference. Live streaming from the authors to you.

Time and Location

Jan 15, 2026 9:00-11:00 AM Pacific Online. Register for the Zoom!

Why Diffusion Models Don't Memorize: The Role of Implicit Dynamical Regularization in Training

Diffusion models have achieved impressive results across many generative tasks, yet the mechanisms that prevent memorization and enable generalization remain unclear. In this talk, I will focus on how training dynamics shape the transition from generalization to memorization. Our experiments and theory reveal two key timescales: an early time when high-quality generation emerges and a later one when memorization begins. Notably, the memorization timescale grows linearly with the size of the training set, while the generalization timescale stays constant, creating an increasingly wide window where models generalize well. These results highlight an implicit dynamical regularization that helps diffusion models avoid memorization even in highly overparameterized regimes.

About the Author

Raphaël Urfin is a PhD student at École Normale Supérieure – PSL in Paris, supervised by Giulio Biroli (ENS) and Marc Mézard (Bocconi University). His work focuses on applying ideas and tools of statistical physics to better understand diffusion models and their generalization properties.

Open-Insect: Benchmarking Open-Set Recognition of Novel Species in Biodiversity Monitoring

Global biodiversity is declining at an unprecedented rate, yet little information is known about most species and how their populations are changing. Indeed, some 90% Earth’s species are estimated to be completely unknown. Machine learning has recently emerged as a promising tool to facilitate long-term, large-scale biodiversity monitoring, including algorithms for fine-grained classification of species from images. However, such algorithms typically are not designed to detect examples from categories unseen during training – the problem of open-set recognition (OSR) – limiting their applicability for highly diverse, poorly studied taxa such as insects. To address this gap, we introduce Open-Insect, a large-scale, fine-grained dataset to evaluate unknown species detection across different geographic regions with varying difficulty. We benchmark 38 OSR algorithms across three categories: post-hoc, training-time regularization, and training with auxiliary data, finding that simple post-hoc approaches remain a strong baseline. We also demonstrate how to leverage auxiliary data to improve species discovery in regions with limited data. Our results provide timely insights to guide the development of computer vision methods for biodiversity monitoring and species discovery.

About the Speaker

Yuyan Chen is a PhD student in Computer Science at McGill University and Mila - Quebec AI Institute, supervised by Prof. David Rolnick. My research focuses on machine learning for biodiversity monitoring.

GuideFlow3D: Optimization-Guided Rectified Flow For Appearance Transfer

Transferring appearance to 3D assets using different representations of the appearance object - such as images or text - has garnered interest due to its wide range of applications in industries like gaming, augmented reality, and digital content creation. However, state-of-the-art methods still fail when the geometry between the input and appearance objects is significantly different. A straightforward approach is to directly apply a 3D generative model, but we show that this ultimately fails to produce appealing results.

Instead, we propose a principled approach inspired by universal guidance. Given a pretrained rectified flow model conditioned on image or text, our training-free method interacts with the sampling process by periodically adding guidance. This guidance can be modeled as a differentiable loss function, and we experiment with two different types of guidance including part-aware losses for appearance and self-similarity. Our experiments show that our approach successfully transfers texture and geometric details to the input 3D asset, outperforming baselines both qualitatively and quantitatively.

We also show that traditional metrics are not suitable for evaluating the task due to their inability of focusing on local details and comparing dissimilar inputs, in absence of ground truth data. We thus evaluate appearance transfer quality with a GPT-based system objectively ranking outputs, ensuring robust and human-like assessment, as further confirmed by our user study. Beyond showcased scenarios, our method is general and could be extended to different types of diffusion models and guidance functions.

About the Speaker

Sayan Deb Sarkar is a 2nd-year PhD student at Stanford University in the Gradient Spaces Group, advised by Prof. Iro Armeni, part of the Stanford Vision Lab (SVL). His research interests are on multimodal 3D scene understanding and interactive editing. Past summer, he interned with the Microsoft Spatial AI Lab, hosted by Prof. Marc Pollefeys, working on efficient video understanding in spatial context. Before starting PhD, he was a CS master student at ETH Zürich, in the Computer Vision and Geometry Group (CVG), working on aligning real-world 3D environments from multi-modal data. In the past, he has been a Research Intern at Qualcomm XR labs, Computer Vision Engineer at Mercedes Benz R & D and Research Engineer at ICG, TU Graz. Website: https://sayands.github.io/

HouseLayout3D: A Benchmark and Baseline Method for 3D Layout Estimation in the Wild​ ​

Current 3D layout estimation models are primarily trained on synthetic datasets containing simple single room or single floor environments. As a consequence, they cannot natively handle large multi floor buildings and require scenes to be split into individual floors before processing, which removes global spatial context that is essential for reasoning about structures such as staircases that connect multiple levels. In this work, we introduce HouseLayout3D, a real world benchmark designed to support progress toward full building scale layout estimation, including multiple floors and architecturally intricate spaces. We also present MultiFloor3D, a simple training free baseline that leverages recent scene understanding methods and already outperforms existing 3D layout estimation models on both our benchmark and prior datasets, highlighting the need for further research in this direction.

About the Speaker

Valentin Bieri is a Machine Learning Engineer and Researcher specializing in the intersection of 3D Computer Vision and Natural Language Processing. Building on his applied research in SLAM and Vision-Language Models at ETH Zurich, he now develops AI agents for manufacturing at EthonAI.

Jan 15 - Best of NeurIPS (Day 2)

Welcome to day two of the Best of NeurIPS series, your virtual pass to some of the groundbreaking research, insights, and innovations that defined this year’s conference. Live streaming from the authors to you.

Time and Location

Jan 15, 2026 9:00-11:00 AM Pacific Online. Register for the Zoom!

Why Diffusion Models Don't Memorize: The Role of Implicit Dynamical Regularization in Training

Diffusion models have achieved impressive results across many generative tasks, yet the mechanisms that prevent memorization and enable generalization remain unclear. In this talk, I will focus on how training dynamics shape the transition from generalization to memorization. Our experiments and theory reveal two key timescales: an early time when high-quality generation emerges and a later one when memorization begins. Notably, the memorization timescale grows linearly with the size of the training set, while the generalization timescale stays constant, creating an increasingly wide window where models generalize well. These results highlight an implicit dynamical regularization that helps diffusion models avoid memorization even in highly overparameterized regimes.

About the Author

Raphaël Urfin is a PhD student at École Normale Supérieure – PSL in Paris, supervised by Giulio Biroli (ENS) and Marc Mézard (Bocconi University). His work focuses on applying ideas and tools of statistical physics to better understand diffusion models and their generalization properties.

Open-Insect: Benchmarking Open-Set Recognition of Novel Species in Biodiversity Monitoring

Global biodiversity is declining at an unprecedented rate, yet little information is known about most species and how their populations are changing. Indeed, some 90% Earth’s species are estimated to be completely unknown. Machine learning has recently emerged as a promising tool to facilitate long-term, large-scale biodiversity monitoring, including algorithms for fine-grained classification of species from images. However, such algorithms typically are not designed to detect examples from categories unseen during training – the problem of open-set recognition (OSR) – limiting their applicability for highly diverse, poorly studied taxa such as insects. To address this gap, we introduce Open-Insect, a large-scale, fine-grained dataset to evaluate unknown species detection across different geographic regions with varying difficulty. We benchmark 38 OSR algorithms across three categories: post-hoc, training-time regularization, and training with auxiliary data, finding that simple post-hoc approaches remain a strong baseline. We also demonstrate how to leverage auxiliary data to improve species discovery in regions with limited data. Our results provide timely insights to guide the development of computer vision methods for biodiversity monitoring and species discovery.

About the Speaker

Yuyan Chen is a PhD student in Computer Science at McGill University and Mila - Quebec AI Institute, supervised by Prof. David Rolnick. My research focuses on machine learning for biodiversity monitoring.

GuideFlow3D: Optimization-Guided Rectified Flow For Appearance Transfer

Transferring appearance to 3D assets using different representations of the appearance object - such as images or text - has garnered interest due to its wide range of applications in industries like gaming, augmented reality, and digital content creation. However, state-of-the-art methods still fail when the geometry between the input and appearance objects is significantly different. A straightforward approach is to directly apply a 3D generative model, but we show that this ultimately fails to produce appealing results.

Instead, we propose a principled approach inspired by universal guidance. Given a pretrained rectified flow model conditioned on image or text, our training-free method interacts with the sampling process by periodically adding guidance. This guidance can be modeled as a differentiable loss function, and we experiment with two different types of guidance including part-aware losses for appearance and self-similarity. Our experiments show that our approach successfully transfers texture and geometric details to the input 3D asset, outperforming baselines both qualitatively and quantitatively.

We also show that traditional metrics are not suitable for evaluating the task due to their inability of focusing on local details and comparing dissimilar inputs, in absence of ground truth data. We thus evaluate appearance transfer quality with a GPT-based system objectively ranking outputs, ensuring robust and human-like assessment, as further confirmed by our user study. Beyond showcased scenarios, our method is general and could be extended to different types of diffusion models and guidance functions.

About the Speaker

Sayan Deb Sarkar is a 2nd-year PhD student at Stanford University in the Gradient Spaces Group, advised by Prof. Iro Armeni, part of the Stanford Vision Lab (SVL). His research interests are on multimodal 3D scene understanding and interactive editing. Past summer, he interned with the Microsoft Spatial AI Lab, hosted by Prof. Marc Pollefeys, working on efficient video understanding in spatial context. Before starting PhD, he was a CS master student at ETH Zürich, in the Computer Vision and Geometry Group (CVG), working on aligning real-world 3D environments from multi-modal data. In the past, he has been a Research Intern at Qualcomm XR labs, Computer Vision Engineer at Mercedes Benz R & D and Research Engineer at ICG, TU Graz. Website: https://sayands.github.io/

HouseLayout3D: A Benchmark and Baseline Method for 3D Layout Estimation in the Wild​ ​

Current 3D layout estimation models are primarily trained on synthetic datasets containing simple single room or single floor environments. As a consequence, they cannot natively handle large multi floor buildings and require scenes to be split into individual floors before processing, which removes global spatial context that is essential for reasoning about structures such as staircases that connect multiple levels. In this work, we introduce HouseLayout3D, a real world benchmark designed to support progress toward full building scale layout estimation, including multiple floors and architecturally intricate spaces. We also present MultiFloor3D, a simple training free baseline that leverages recent scene understanding methods and already outperforms existing 3D layout estimation models on both our benchmark and prior datasets, highlighting the need for further research in this direction.

About the Speaker

Valentin Bieri is a Machine Learning Engineer and Researcher specializing in the intersection of 3D Computer Vision and Natural Language Processing. Building on his applied research in SLAM and Vision-Language Models at ETH Zurich, he now develops AI agents for manufacturing at EthonAI.

Jan 15 - Best of NeurIPS (Day 2)

📢 Rencontre Hybride Date : 10 décembre 2025 Lieu : Bureaux de Larochelle Groupe Conseil (plan disponible dans la section photo de l’événement) Format : Hybride (présentiel + virtuel) ✅ 35 places disponibles en personneParticipation virtuelle illimitée Larochelle Groupe Conseil nous accueille dans ses bureaux et offrira des bouchées pour les participants sur place. Inscription obligatoire pour réserver votre place (présentiel ou virtuel).

Plan de la rencontre

🕔 17 h 00 – 17 h 20 \| Accueil des participants Arrivée, enregistrement et réseautage informel. 🕔 17 h 20 – 17 h 25 \| Mot de bienvenue Par le comité d’organisation. 🕔 17 h 25 – 17 h 35 \| Présentation de Larochelle Mot d’un représentant de l’entreprise hôte. 🕔 17 h 35 – 18 h 10 \| Conférence #1 (voir la description ci-dessous) 🕔 18 h 10 – 18 h 45 \| Conférence #2 (voir la description ci-dessous) 🕔 18 h 45 – 20 h 00 \| Réseautage et bouchées Moment convivial pour échanger et créer des liens. 🍴 Bouchées offertes gracieuseté de Larochelle.

CONFÉRENCE #1 Conférencier : Nick Desbarats

Pourquoi les gens n'utilisent pas mon tableau de bord? Vous avez mis sur pied un tout nouveau tableau de bord pour aider votre organisation à prendre de meilleures décisions qui sont supportées par des données. Vous avez bien cerné leurs besoins, vous vous êtes inspiré des exemples des experts, puis vous avez lancé votre création… et la réaction a été plutôt silencieuse.

Tout le monde s’est connecté dès le lancement, mais quelques semaines ou quelques mois plus tard, presque plus personne ne l’utilisait. Quand vous demandez aux utilisateurs pourquoi, vous obtenez des réponses floues comme « Je manque de temps » (ce qui veut souvent dire « Ça ne m’apporte pas assez de valeur ») ou « Il manque des choses » (même si vous avez livré exactement ce qu’ils avaient demandé).

Que se passe-t-il au juste? Quelles sont les vraies raisons pour lesquelles les gens n’utilisent pas votre tableau de bord? Découvrez-les dans cette présentation de Nick Desbarats, expert en conception de tableaux de bord et auteur à succès.

CONFÉRENCE #2: Conférencier : Joel Gagnon

Repenser la gestion axée sur les données : comment Fabric et l’IA impacteront votre organisation De Microsoft Excel à Copilot, cette session explore l’évolution de la prise de décision guidée par les données. Vous y verrez comment passent les données de l’ingestion à la visualisation, en incluant toutes les étapes essentielles : préparation, transformation et modélisation. L’objectif : passer de rapports manuels à des décisions automatisées – voire prescriptives. Découvrez comment Power BI, Fabric et Copilot permettent de transformer des décisions de gestion récurrentes en processus reproductibles et data-driven, et comment préparer efficacement votre organisation à cette transition.

Créer de la valeur avec vos données : adoption, IA et automatisation
Nick Schrock – guest , Joe Reis – founder @ Ternary Data

Data engineering is undergoing a fundamental shift. In this episode, I sit down with Nick Schrock, founder and CTO of Dagster, to discuss why he went from being an "AI moderate" to believing 90% of code will be written by AI. Being hands on also led to a massive pivot in Dagster’s roadmap and a new focus on managing and engineering context. We dive deep into why simply feeding data to LLMs isn't enough. Nick explains why real-time context tools (like MCPs) can become "token hogs" that lack precision and why the future belongs to "context pipelines": offline, batch-computed context that is governed, versioned, and treated like code. We also explore Compass, Dagster’s new collaborative agent that lives in Slack, bridging the gap between business stakeholders and data teams. If you’re wondering how your role as a data engineer will evolve in an agentic world, this conversation maps out the territory Dagster: dagster.io Nick Schrock on X: @schrockn

AI/ML Dagster Data Engineering LLM
The Joe Reis Show
PyData x Pipple Meetup 2025-11-18 · 16:30

Please register here: https://pipple.nl/en/pydata-x-pipple-meetup/

The power of data and AI in practice

AI goes beyond experiments and isolated analyses. During this meetup, speakers from Pipple and PyData Eindhoven will share concrete cases and technical learnings from practice, from data model to deployment. Discover how organisations are making the transition from smart analyses to solutions that are actually used: reliable, scalable and with practical impact.

Meet the community, share your knowledge

Together with PyData Eindhoven, we are bringing data scientists, engineers and AI professionals together for an evening full of content and inspiration. Expect three technical talks, practical examples and valuable conversations with colleagues who work daily on the theme of from data to production.

| 17:30-18:00 | Walk-in & welcome | | ----------- | ----------------- | | 18:00-18:15 | Intro PyData Eindhoven & Pipple | | 18:15-18:45 | From code to production: how to keep pipelines running smoothly at scale

In this talk, Yannick and Joshua will share how they designed and evolved the CI/CD pipelines that power a global retailer’s data platform across 30+ countries. From the first setup to today’s architecture, they’ll dive into what worked, what didn’t, and why. Expect real-world lessons around dbt, Terraform, and security – from failed builds to the automation wins that finally made life easier.

This session offers practical insights for anyone working with data:from engineers maintaining pipelines to analysts and ML developers relying on them. | | 18:45-19:00 | Break & pizza 🍕 | | 19:00-19:30 | Learnings from integrating AI within CyberBench: the good, the bad and the ugly

In this talk, Lucas shares his journey of building CyberBench: a framework that automatically benchmarks Large Language Models (LLMs) for direct security risks such as prompt injections and data leaks.

After a brief introduction to the cybersecurity risks of GenAI systems and the motivation behind CyberBench, Lucas dives into the technical lessons learned from integrating LLM-based components into a software system. Expect practical insights into building reliable and efficient LLM pipelines and lessons on what worked (and what didn’t).

In short: learn to maximize the good, minimize the bad, and make the ugly a little prettier when building LLM systems in production. | | 19:30-20:00 | Speaker 3 – to be announced! |

PyData x Pipple Meetup

For two years, virtual threads had full support in the Java platform. In this presentation, we will have a second look at virtual thread performance, memory usage, platform support, and more advanced features, in particular, structured concurrency and scoped values. You will learn about effective patterns for structured concurrency, how to monitor virtual threads in practice, how to detect overuse of thread locals, and how to use scoped values as a more performant replacement. This topic should be of interest to all developers who want to reduce the cost of deploying business applications that \"mostly block\" (waiting for results from databases or other services), without the pain of implementing business logic in a reactive style.

Java virtual threads structured concurrency
Alexander Matveev – Technology Management Lead, Investment Bank Technologies Domain @ Deutsche Bank

In this talk, we’ll demonstrate how we manage application properties and component topology as plain Java code. We’ll show how this approach brings all the benefits of a type-safe language to configuration — including compilation checks, full IDE support, and even unit tests. We’ll also explain how it enables seamless management of a complex system topology, allowing changes to be distributed effortlessly while ensuring its consistency.

Java configuration type-safety
Global Day of Code Retreat 2025-11-07 · 09:00

⚠️ Aktueller Hinweis ⚠️

Das Event findet in kleinerem Rahmen als Coding Dojo von 10:00 bis 12:30 Uhr statt. Im Anschluss können wir gerne gemeinsam Mittag essen (Selbstverpflegung). Wir freuen uns auf euch!

⚓ Ahoy, ihr Landratten! Macht euch bereit für einen Coderetreat der besonderen Art – am Global Day of Coderetreat!

🌴 Arrr! Monkey Island wartet! Auf der berüchtigten Insel werden wir in Teams Rätsel lösen, Schätze entdecken und gegen furchterregende Affen, Piraten und tückische Anforderungen kämpfen. Ihr schärft eure Fähigkeiten in Test-Driven Development (TDD), Agile Development, Code-Design und Kommunikation. Wie in der Realität werden die Product Owner die Anforderungen genau kennen – und es kann jeder mitmachen, egal ob Entwickler, Tester, Scrum Master oder Product Owner. Vorkenntnisse? Braucht ihr nicht! Anmerkung: wenn ihr letztes Jahr dabei wart, kennt ihr das Thema schon. Die einzelnen Sessions werden aber etwas anders sein. Also kommt gerne trotzdem dazu!

Was ist ein Coderetreat? Ein Coderetreat ist ein ganztägiger, praxisorientierter Coding-Workshop. Ihr werdet die Möglichkeit haben, neue Wege im Testen, Design und Software-Architektur auszuprobieren und zu vertiefen – und das alles, während ihr euch durch die Rätsel von Monkey Island kämpft. Mehr Infos: coderetreat.org.

💻 Was solltet ihr mitbringen? Bringt euren Laptop mit einer funktionsfähigen Entwicklungsumgebung mit. Ihr solltet in der Lage sein, einfache Tests auszuführen, wie z.B. ein roter Test für `assertFalse(true)` und ein grüner Test für `assertTrue(true)`.

Initiale Setups für die unterschiedlichsten Sprachen findet ihr hier: https://github.com/rradczewski/kata-bootstraps

🧭 Anreise und Ablauf

  • Ort: codecentric AG, Köpenicker Str. 31, 10179 Berlin
  • Agenda:
  • 09:45 – Türen öffnen
  • 10:00 – Beginn der Sessions (Piratenkämpfe!)
  • 12:30 – Offizielles Ende
  • danach optional gemeinsames Mittagessen

🚩 Sprachen und Programmierung Unser Code Retreat wird auf Deutsch moderiert, aber auch Englisch-Sprechende sind herzlich willkommen! Wir unterstützen euch gerne, egal in welcher Sprache ihr euch wohler fühlt. Und was Programmiersprachen angeht: Wählt, was euch am meisten Spaß macht, solange ihr Tests darin schreiben könnt. Für die meisten Sprachen findet ihr einen Partner.

🥰 Seid gut zueinander! Wir werden den ganzen Tag über in Pairs programmieren. Seid freundlich und respektvoll. Braucht ihr eine Pause, dann lasst es uns wissen. Lasst uns gemeinsam einen spaßigen und lehrreichen Tag erleben!

Global Day of Code Retreat
Michaela Truchanová (Miši) – Analytics Engineer @ Slido

Using email personalisation tools is not enough. True impact comes when your efforts are powered by data that is always consistent and unified across systems. In this session, Miši will show us how Slido leverages dbt models to drive engaging client communication while ensuring that data remains accurate and harmonised throughout all systems.

dbt
Lukáš Kozelnický – Senior Data Engineer @ Muziker

Lukáš Kozelnický will walk us through the production-grade dbt implementation that powers Muziker, Europe's leading musical instrument e-commerce platform operating across 30+ countries. He will explore the data warehouse that: Handles complex multi-currency operations; Integrates 50+ data sources; Features sophisticated AI-powered inventory optimisation. Let’s compare Muziker’s setup 4 years ago with their data platform today to see what has changed in the analytics world, how the new approach empowers the business, and how the data culture at the company has changed to support this shift.

dbt ClickHouse

Venue: Carnival House, 100 Harbour Parade, Southampton, SO15 1ST 📢 Want to speak 📢: submit your talk proposal

Main Talks 1️⃣ Unlocking the Black Box: Demystifying ML Models with Shapley Values - Philip Le Model explainability is key to help us build trust and enable decision-making with ML/AI models. We will dig deeper into the theoretical background of the Shapley value to help us address the complexity and bias challenges of black box models. There will be several examples to showcase the practical side of how we could use these techniques in practice.

2️⃣ Predicting Extreme Weather Events: Augmenting AI Models to Improve Reliability - Austen Wallis Machine learning thrives on data, and the common wisdom is simple: the more, the better. A widely used approach to expand training datasets is data augmentation, which can enhance model robustness. Yet, augmentation is not without risk—it can also degrade performance by introducing noise. More critically, there are cases where the required training data simply does not exist, especially when models are applied to previously unseen regimes.

Nowhere is this more relevant than in the context of improving weather model predictions for extreme events, such as hurricanes. Climate change is driving our atmosphere into uncharted territory, producing extremes absent from recorded history. How, then, can we trust model predictions under such conditions? One promising avenue is through the generation of synthetic data. Hence, in this talk, we will explore how weather data can be simulated, the ML model architectures we use that are designed to forecast extreme events, and the growing competition to build the most powerful foundational weather model.

⚡Lightning Talks ⚡ 1️⃣ Exploration of Bayesian Template Selection for Tracking Arbitrary Objects - Tim Trew 2️⃣ From Issue to PR with Claude Code - Nick Thorne

Please note:

  1. 🚨🚨🚨A valid photo ID is required by building security. You MUST use your initial/first name and surname on your meetup profile, otherwise, you will NOT make it on the guest list! 🚨🚨🚨
  2. This event follows the NumFOCUS Code of Conduct, please familiarise yourself with it before the event.

If your RSVP status says "You're going" you will be able to get in. No further confirmation required. You will NOT need to show your RSVP confirmation when signing in. If you can no longer make it, please unRSVP as soon as you know so we can assign your place to someone on the waiting list.

*** Code of Conduct: This event follows the NumFOCUS Code of Conduct, please familiarise yourself with it before the event. Please get in touch with the organisers with any questions or concerns regarding the Code of Conduct. *** There will be pizza & drinks, generously provided by our host, Carnival UK. ***

Logistics Doors open at 6.30 pm, talks start at 7 pm. For those who wish to continue networking and chatting we will move to a nearby pub/bar for drinks from 9 pm.

Please unRSVP in good time if you realise you can't make it. We're limited by building security on the number of attendees, so please free up your place for your fellow community members!

Follow @pydatasoton (https://twitter.com/pydatasoton) for updates and early announcements. We are also on Instagram/Threads as @pydatasoton, and find us on LinkedIn.

PyData Southampton - 19th Meetup